++ echo 'Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/logs/demand-backup-sharded.log' Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/logs/demand-backup-sharded.log ++ '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/cloud-secret.yml ']' ++ SKIP_BACKUPS_TO_AWS_GCP_AZURE= ++ oc get projects ++ kubectl get nodes ++ grep '^minikube' +++ kubectl version -o json +++ jq -r .serverVersion.gitVersion +++ grep '\-eks\-' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ '[' ']' ++ EKS=0 +++ kubectl version -o json +++ jq -r .serverVersion.gitVersion +++ grep gke WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ '[' v1.26.15-gke.1469001 ']' ++ GKE=1 +++ kubectl version -o json +++ jq -r '.serverVersion.major + "." + .serverVersion.minor' +++ /usr/bin/sed -r 's/[^0-9.]+//g' WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 ++ KUBE_VERSION=1.26 + set_debug + [[ 1 == 1 ]] + set -o xtrace + create_infra demand-backup-sharded-7957 + local ns=demand-backup-sharded-7957 + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.OeAQxtz5Sw ++ mktemp + local LAST_ERR=/tmp/tmp.IqEHwgYoDE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OeAQxtz5Sw customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.IqEHwgYoDE + rm /tmp/tmp.OeAQxtz5Sw /tmp/tmp.IqEHwgYoDE + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/crd.yaml ++ grep -v '\-\-\-' + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n demand-backup-sharded-30529 backup-aws-s3 --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodbbackup.psmdb.percona.com/backup-aws-s3 patched + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n demand-backup-sharded-30529 backup-azure-blob --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodbbackup.psmdb.percona.com/backup-azure-blob patched + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n demand-backup-sharded-30529 backup-minio --type=merge -p '{"metadata":{"finalizers":[]}}' perconaservermongodbbackup.psmdb.percona.com/backup-minio patched + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.yhAXgc27s0 ++ mktemp + local LAST_ERR=/tmp/tmp.9xQGDI3cMX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.yhAXgc27s0 customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com condition met + cat /tmp/tmp.9xQGDI3cMX + rm /tmp/tmp.yhAXgc27s0 /tmp/tmp.9xQGDI3cMX + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.AWfyKZYpaR ++ mktemp + local LAST_ERR=/tmp/tmp.6FbJAqPro0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.AWfyKZYpaR + cat /tmp/tmp.6FbJAqPro0 + rm /tmp/tmp.AWfyKZYpaR /tmp/tmp.6FbJAqPro0 + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.3c7CcjHjgs ++ mktemp + local LAST_ERR=/tmp/tmp.jPSWB7iHH8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.3c7CcjHjgs + cat /tmp/tmp.jPSWB7iHH8 + rm /tmp/tmp.3c7CcjHjgs /tmp/tmp.jPSWB7iHH8 + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.JWTn3fSTuq ++ mktemp + local LAST_ERR=/tmp/tmp.Is2kyp905U + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.JWTn3fSTuq clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.Is2kyp905U + rm /tmp/tmp.JWTn3fSTuq /tmp/tmp.Is2kyp905U + return 0 + check_crd_for_deletion PR-1573-f54074df + local git_tag=PR-1573-f54074df ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-1573-f54074df/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/bin/sed ':a;N;$!ba;s/\n/ /g' ++ /usr/bin/sed s/---//g + for crd_name in '$(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '\''.metadata.name'\'' | $sed '\''s/---//g'\'' | $sed '\'':a;N;$!ba;s/\n/ /g'\'')' ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cf2clcb0m4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.iSAZlxqaTI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.cf2clcb0m4 ++ cat /tmp/tmp.iSAZlxqaTI Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.cf2clcb0m4 ++ cat /tmp/tmp.iSAZlxqaTI Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.cf2clcb0m4 ++ cat /tmp/tmp.iSAZlxqaTI Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.cf2clcb0m4 ++ cat /tmp/tmp.iSAZlxqaTI Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.cf2clcb0m4 /tmp/tmp.iSAZlxqaTI ++ return 1 + [[ '' == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get clusterrolebinding ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|gke-mcs|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace + kubectl_bin get ns ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + xargs kubectl delete ns + kubectl_bin delete namespace psmdb-operator --ignore-not-found ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.OrBwbe4l0Y + local LAST_OUT=/tmp/tmp.6TCofR1OLH ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.lfjBwK5uoQ + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.zrZ4FWE3tE + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + for i in '$(seq 0 2)' + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OrBwbe4l0Y + cat /tmp/tmp.lfjBwK5uoQ + rm /tmp/tmp.OrBwbe4l0Y /tmp/tmp.lfjBwK5uoQ + return 0 namespace "demand-backup-sharded-30529" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6TCofR1OLH namespace "psmdb-operator" deleted + cat /tmp/tmp.zrZ4FWE3tE + rm /tmp/tmp.6TCofR1OLH /tmp/tmp.zrZ4FWE3tE + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.6wQKaIv39C ++ mktemp + local LAST_ERR=/tmp/tmp.zwUeSB3a9C + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6wQKaIv39C + cat /tmp/tmp.zwUeSB3a9C + rm /tmp/tmp.6wQKaIv39C /tmp/tmp.zwUeSB3a9C + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.QFmgjZ4OtW ++ mktemp + local LAST_ERR=/tmp/tmp.blY4T3N5Nn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.QFmgjZ4OtW namespace/psmdb-operator created + cat /tmp/tmp.blY4T3N5Nn + rm /tmp/tmp.QFmgjZ4OtW /tmp/tmp.blY4T3N5Nn + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.baz6kOTQSL +++ mktemp ++ local LAST_ERR=/tmp/tmp.J4w8FA8kTI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.baz6kOTQSL ++ cat /tmp/tmp.J4w8FA8kTI ++ rm /tmp/tmp.baz6kOTQSL /tmp/tmp.J4w8FA8kTI ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1573-f54074df-4-cluster1 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.RxGqHjQbtj ++ mktemp + local LAST_ERR=/tmp/tmp.08zdhRUfUv + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1573-f54074df-4-cluster1 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.RxGqHjQbtj Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1573-f54074df-4-cluster1" modified. + cat /tmp/tmp.08zdhRUfUv + rm /tmp/tmp.RxGqHjQbtj /tmp/tmp.08zdhRUfUv + return 0 + deploy_operator + desc 'start PSMDB operator' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.4WxeizXzjm ++ mktemp + local LAST_ERR=/tmp/tmp.Q820LNJsAA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4WxeizXzjm customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.Q820LNJsAA + rm /tmp/tmp.4WxeizXzjm /tmp/tmp.Q820LNJsAA + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.kBpoYGYHmJ ++ mktemp + local LAST_ERR=/tmp/tmp.Z6Ht9D9uh3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.kBpoYGYHmJ clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.Z6Ht9D9uh3 + rm /tmp/tmp.kBpoYGYHmJ /tmp/tmp.Z6Ht9D9uh3 + return 0 + yq eval ' (.spec.template.spec.containers[].image = "perconalab/percona-server-mongodb-operator:PR-1573-f54074df") | ((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") | ((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/cw-operator.yaml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.oIW13ffMnr ++ mktemp + local LAST_ERR=/tmp/tmp.m2SJjJ2yL1 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.oIW13ffMnr deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.m2SJjJ2yL1 + rm /tmp/tmp.oIW13ffMnr /tmp/tmp.m2SJjJ2yL1 + return 0 + sleep 2 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.nMTGrcPeOl +++ mktemp ++ local LAST_ERR=/tmp/tmp.EP8cRnuQ0B ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nMTGrcPeOl ++ cat /tmp/tmp.EP8cRnuQ0B ++ rm /tmp/tmp.nMTGrcPeOl /tmp/tmp.EP8cRnuQ0B ++ return 0 + wait_pod percona-server-mongodb-operator-7bb67465c5-95xbd + local pod=percona-server-mongodb-operator-7bb67465c5-95xbd + set +o xtrace waiting for pod/percona-server-mongodb-operator-7bb67465c5-95xbd to be ready.OK + create_namespace demand-backup-sharded-7957 + local namespace=demand-backup-sharded-7957 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|gke-mcs|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces demand-backup-sharded-7957' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces demand-backup-sharded-7957 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace demand-backup-sharded-7957 --ignore-not-found + xargs kubectl delete ns + awk '{print$1}' + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.XH2hygNhh6 ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.xpcHlTz4Qf + local LAST_ERR=/tmp/tmp.0DIyGsETNn + local exit_status=0 + local timeout=4 ++ mktemp ++ seq 0 2 + local LAST_ERR=/tmp/tmp.0Sk4Hf8M4T + local exit_status=0 + local timeout=4 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-sharded-7957 --ignore-not-found ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.xpcHlTz4Qf + cat /tmp/tmp.0Sk4Hf8M4T + rm /tmp/tmp.xpcHlTz4Qf /tmp/tmp.0Sk4Hf8M4T + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.XH2hygNhh6 + cat /tmp/tmp.0DIyGsETNn + rm /tmp/tmp.XH2hygNhh6 /tmp/tmp.0DIyGsETNn + return 0 + kubectl_bin wait --for=delete namespace demand-backup-sharded-7957 ++ mktemp + local LAST_OUT=/tmp/tmp.FNKO2DeRsL ++ mktemp + local LAST_ERR=/tmp/tmp.nmQIay6xAc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete namespace demand-backup-sharded-7957 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FNKO2DeRsL + cat /tmp/tmp.nmQIay6xAc + rm /tmp/tmp.FNKO2DeRsL /tmp/tmp.nmQIay6xAc + return 0 + desc 'create namespace demand-backup-sharded-7957' + set +o xtrace ----------------------------------------------------------------------------------- create namespace demand-backup-sharded-7957 ----------------------------------------------------------------------------------- + kubectl_bin create namespace demand-backup-sharded-7957 ++ mktemp + local LAST_OUT=/tmp/tmp.SWmYhWcR5a ++ mktemp + local LAST_ERR=/tmp/tmp.ida2pNyGRZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace demand-backup-sharded-7957 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.SWmYhWcR5a namespace/demand-backup-sharded-7957 created + cat /tmp/tmp.ida2pNyGRZ + rm /tmp/tmp.SWmYhWcR5a /tmp/tmp.ida2pNyGRZ + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.0afh92ykbd +++ mktemp ++ local LAST_ERR=/tmp/tmp.IGXY5EGqso ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.0afh92ykbd ++ cat /tmp/tmp.IGXY5EGqso ++ rm /tmp/tmp.0afh92ykbd /tmp/tmp.IGXY5EGqso ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1573-f54074df-4-cluster1 --namespace=demand-backup-sharded-7957 ++ mktemp + local LAST_OUT=/tmp/tmp.MwIgvSZCN5 ++ mktemp + local LAST_ERR=/tmp/tmp.oNQG9xvTpb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1573-f54074df-4-cluster1 --namespace=demand-backup-sharded-7957 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MwIgvSZCN5 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1573-f54074df-4-cluster1" modified. + cat /tmp/tmp.oNQG9xvTpb + rm /tmp/tmp.MwIgvSZCN5 /tmp/tmp.oNQG9xvTpb + return 0 + deploy_minio + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + helm repo remove minio "minio" has been removed from your repositories + helm repo add minio https://charts.min.io/ "minio" has been added to your repositories + retry 10 60 helm install minio-service --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.0.14 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio NAME: minio-service LAST DEPLOYED: Sun Jul 7 17:30:37 2024 NAMESPACE: demand-backup-sharded-7957 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.demand-backup-sharded-7957.svc.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace demand-backup-sharded-7957 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace demand-backup-sharded-7957 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace demand-backup-sharded-7957 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace demand-backup-sharded-7957 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GPvyXdj5Tn +++ mktemp ++ local LAST_ERR=/tmp/tmp.m4kILzt1jI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GPvyXdj5Tn ++ cat /tmp/tmp.m4kILzt1jI ++ rm /tmp/tmp.GPvyXdj5Tn /tmp/tmp.m4kILzt1jI ++ return 0 + MINIO_POD=minio-service-57dd49b-9br22 + wait_pod minio-service-57dd49b-9br22 + local pod=minio-service-57dd49b-9br22 + set +o xtrace waiting for pod/minio-service-57dd49b-9br22 to be ready.OK + '[' -n psmdb-operator ']' + kubectl_bin create svc -n psmdb-operator externalname minio-service --external-name=minio-service.demand-backup-sharded-7957.svc.cluster.local --tcp=9000 ++ mktemp + local LAST_OUT=/tmp/tmp.Ws9kRZ5SFh ++ mktemp + local LAST_ERR=/tmp/tmp.l5KQgEaP4w + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create svc -n psmdb-operator externalname minio-service --external-name=minio-service.demand-backup-sharded-7957.svc.cluster.local --tcp=9000 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Ws9kRZ5SFh service/minio-service created + cat /tmp/tmp.l5KQgEaP4w + rm /tmp/tmp.Ws9kRZ5SFh /tmp/tmp.l5KQgEaP4w + return 0 + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 \ /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' ++ mktemp + local LAST_OUT=/tmp/tmp.2ufzu6wEXo ++ mktemp + local LAST_ERR=/tmp/tmp.3OQypKClpK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 \ /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2ufzu6wEXo make_bucket: operator-testing pod "aws-cli" deleted + cat /tmp/tmp.3OQypKClpK + rm /tmp/tmp.2ufzu6wEXo /tmp/tmp.3OQypKClpK + return 0 + desc 'create first PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- + cluster=some-name + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.tfZ3IAZo3G ++ mktemp + local LAST_ERR=/tmp/tmp.a4jcEpbLg6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.tfZ3IAZo3G secret/some-users created deployment.apps/psmdb-client created + cat /tmp/tmp.a4jcEpbLg6 + rm /tmp/tmp.tfZ3IAZo3G /tmp/tmp.a4jcEpbLg6 + return 0 + apply_s3_storage_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.OFUaZH9N7v ++ mktemp + local LAST_ERR=/tmp/tmp.GaTxa3R8Fb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OFUaZH9N7v secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.GaTxa3R8Fb + rm /tmp/tmp.OFUaZH9N7v /tmp/tmp.GaTxa3R8Fb + return 0 + version_gt 1.19 ++ bc -l ++ echo '1.26 >= 1.19' + '[' 1 -eq 1 ']' + return 0 + '[' 0 -ne 1 ']' + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/container-rc.yaml + /usr/bin/sed s/docker/runc/g + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.SEl10rAUEz ++ mktemp + local LAST_ERR=/tmp/tmp.rezBvihsUe + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.SEl10rAUEz runtimeclass.node.k8s.io/container-rc unchanged + cat /tmp/tmp.rezBvihsUe + rm /tmp/tmp.SEl10rAUEz /tmp/tmp.rezBvihsUe + return 0 + apply_cluster /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/some-name-rs0.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/some-name-rs0.yml + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/some-name-rs0.yml + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab/percona-server-mongodb-operator:PR-1573-f54074df"' + yq eval '(.spec | select(has("backup"))).backup.image = "perconalab/percona-server-mongodb-operator:main-backup"' ++ mktemp + local LAST_OUT=/tmp/tmp.uopz6LB4rO ++ mktemp + local LAST_ERR=/tmp/tmp.hdqEH87dsi + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + yq eval '(.spec | select(.image == null)).image = "perconalab/percona-server-mongodb-operator:main-mongod7.0"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "perconalab/pmm-client:dev-latest"' + yq eval '.spec.upgradeOptions.apply="Never"' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.uopz6LB4rO perconaservermongodb.psmdb.percona.com/some-name created + cat /tmp/tmp.hdqEH87dsi + rm /tmp/tmp.uopz6LB4rO /tmp/tmp.hdqEH87dsi + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- + wait_for_running some-name-rs0 3 + local name=some-name-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=some-name ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + [[ 0 -eq 2 ]] + wait_pod some-name-rs0-0 + local pod=some-name-rs0-0 + set +o xtrace waiting for pod/some-name-rs0-0 to be ready......OK + for i in '$(seq 0 $last_pod)' + [[ 1 -eq 2 ]] + wait_pod some-name-rs0-1 + local pod=some-name-rs0-1 + set +o xtrace waiting for pod/some-name-rs0-1 to be ready......OK + for i in '$(seq 0 $last_pod)' + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XqWvPLpeVj +++ mktemp ++ local LAST_ERR=/tmp/tmp.7gDSKTxW9N ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XqWvPLpeVj ++ cat /tmp/tmp.7gDSKTxW9N ++ rm /tmp/tmp.XqWvPLpeVj /tmp/tmp.7gDSKTxW9N ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-rs0-2 + local pod=some-name-rs0-2 + set +o xtrace waiting for pod/some-name-rs0-2 to be ready....OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bp08jl0Z40 +++ mktemp ++ local LAST_ERR=/tmp/tmp.JyC7t5vrvw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].non_voting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bp08jl0Z40 ++ cat /tmp/tmp.JyC7t5vrvw ++ rm /tmp/tmp.bp08jl0Z40 /tmp/tmp.JyC7t5vrvw ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness........................................... + wait_for_running some-name-cfg 3 false + local name=some-name-cfg + let last_pod=2 + local check_cluster_readyness=false + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=cfg + local cluster_name=some-name ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + [[ 0 -eq 2 ]] + wait_pod some-name-cfg-0 + local pod=some-name-cfg-0 + set +o xtrace waiting for pod/some-name-cfg-0 to be ready.OK + for i in '$(seq 0 $last_pod)' + [[ 1 -eq 2 ]] + wait_pod some-name-cfg-1 + local pod=some-name-cfg-1 + set +o xtrace waiting for pod/some-name-cfg-1 to be ready.OK + for i in '$(seq 0 $last_pod)' + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Yh09UUlrRK +++ mktemp ++ local LAST_ERR=/tmp/tmp.99BQnH3BNn ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Yh09UUlrRK ++ cat /tmp/tmp.99BQnH3BNn ++ rm /tmp/tmp.Yh09UUlrRK /tmp/tmp.99BQnH3BNn ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-cfg-2 + local pod=some-name-cfg-2 + set +o xtrace waiting for pod/some-name-cfg-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].non_voting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sWsGIBwTi7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.imvvT4BSD9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="cfg")].non_voting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sWsGIBwTi7 ++ cat /tmp/tmp.imvvT4BSD9 ++ rm /tmp/tmp.sWsGIBwTi7 /tmp/tmp.imvvT4BSD9 ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ false == \t\r\u\e ]] + wait_for_running some-name-mongos 3 + local name=some-name-mongos + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=mongos + local cluster_name=some-name ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + [[ 0 -eq 2 ]] + wait_pod some-name-mongos-0 + local pod=some-name-mongos-0 + set +o xtrace waiting for pod/some-name-mongos-0 to be ready.OK + for i in '$(seq 0 $last_pod)' + [[ 1 -eq 2 ]] + wait_pod some-name-mongos-1 + local pod=some-name-mongos-1 + set +o xtrace waiting for pod/some-name-mongos-1 to be ready.OK + for i in '$(seq 0 $last_pod)' + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tkTMepFzDB +++ mktemp ++ local LAST_ERR=/tmp/tmp.S3Itc2M8bc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.tkTMepFzDB ++ cat /tmp/tmp.S3Itc2M8bc ++ rm /tmp/tmp.tkTMepFzDB /tmp/tmp.S3Itc2M8bc ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-mongos-2 + local pod=some-name-mongos-2 + set +o xtrace waiting for pod/some-name-mongos-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].non_voting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CQ1JSzl2JQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.y75OOBbOc9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].non_voting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CQ1JSzl2JQ ++ cat /tmp/tmp.y75OOBbOc9 ++ rm /tmp/tmp.CQ1JSzl2JQ /tmp/tmp.y75OOBbOc9 ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness + sleep 10 + desc 'check if service and statefulset created with expected config' + set +o xtrace ----------------------------------------------------------------------------------- check if service and statefulset created with expected config ----------------------------------------------------------------------------------- + compare_kubectl statefulset/some-name-rs0 + local resource=statefulset/some-name-rs0 + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs0.yml + local new_result=/tmp/tmp.bLUg6fPDea/statefulset_some-name-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs0-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-rs0 + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("demand-backup-sharded-7957", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.GDF97R9bi9 ++ mktemp + local LAST_ERR=/tmp/tmp.d3MMZ6kCwP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.GDF97R9bi9 + cat /tmp/tmp.d3MMZ6kCwP + rm /tmp/tmp.GDF97R9bi9 /tmp/tmp.d3MMZ6kCwP + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs0.yml + version_gt 1.22 ++ bc -l ++ echo '1.26 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs0.yml == */cronjob* ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs0.yml /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs0.yml + compare_kubectl statefulset/some-name-rs1 + local resource=statefulset/some-name-rs1 + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs1.yml + local new_result=/tmp/tmp.bLUg6fPDea/statefulset_some-name-rs1.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs1-oc.yml ']' + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("demand-backup-sharded-7957", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - + kubectl_bin get -o yaml statefulset/some-name-rs1 ++ mktemp + local LAST_OUT=/tmp/tmp.ROm09ryxvN ++ mktemp + local LAST_ERR=/tmp/tmp.ewLk5qZuFk + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-rs1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ROm09ryxvN + cat /tmp/tmp.ewLk5qZuFk + rm /tmp/tmp.ROm09ryxvN /tmp/tmp.ewLk5qZuFk + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs1.yml + version_gt 1.22 ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs1.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs1.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs1.yml == */cronjob* ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs1.yml /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs1.yml + compare_kubectl statefulset/some-name-rs2 + local resource=statefulset/some-name-rs2 + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs2.yml + local new_result=/tmp/tmp.bLUg6fPDea/statefulset_some-name-rs2.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs2-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-rs2 ++ mktemp + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("demand-backup-sharded-7957", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - + local LAST_OUT=/tmp/tmp.ImLcSXosn7 ++ mktemp + local LAST_ERR=/tmp/tmp.d2Fz9ppQSD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-rs2 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ImLcSXosn7 + cat /tmp/tmp.d2Fz9ppQSD + rm /tmp/tmp.ImLcSXosn7 /tmp/tmp.d2Fz9ppQSD + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs2.yml + version_gt 1.22 ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs2.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs2.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs2.yml == */cronjob* ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-rs2.yml /tmp/tmp.bLUg6fPDea/statefulset_some-name-rs2.yml + compare_kubectl statefulset/some-name-cfg + local resource=statefulset/some-name-cfg + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-cfg.yml + local new_result=/tmp/tmp.bLUg6fPDea/statefulset_some-name-cfg.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-cfg-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-cfg ++ mktemp + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("demand-backup-sharded-7957", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - + local LAST_OUT=/tmp/tmp.n3ATe9snMo ++ mktemp + local LAST_ERR=/tmp/tmp.40U0n3x8nS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-cfg + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.n3ATe9snMo + cat /tmp/tmp.40U0n3x8nS + rm /tmp/tmp.n3ATe9snMo /tmp/tmp.40U0n3x8nS + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-cfg.yml + version_gt 1.22 ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-cfg.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-cfg.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-cfg.yml == */cronjob* ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-cfg.yml /tmp/tmp.bLUg6fPDea/statefulset_some-name-cfg.yml + compare_kubectl statefulset/some-name-mongos '' + local resource=statefulset/some-name-mongos + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-mongos.yml + local new_result=/tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-mongos-oc.yml ']' + kubectl_bin get -o yaml statefulset/some-name-mongos + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("demand-backup-sharded-7957", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.HNjxs4jMHO ++ mktemp + local LAST_ERR=/tmp/tmp.UgKudjgKuj + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-mongos + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.HNjxs4jMHO + cat /tmp/tmp.UgKudjgKuj + rm /tmp/tmp.HNjxs4jMHO /tmp/tmp.UgKudjgKuj + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + version_gt 1.22 ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-mongos.yml == */cronjob* ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-mongos.yml /tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/mongos-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.OoH3WUXeB3 ++ mktemp + local LAST_ERR=/tmp/tmp.4FiFB4HBwD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/mongos-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OoH3WUXeB3 secret/some-name-mongos created + cat /tmp/tmp.4FiFB4HBwD + rm /tmp/tmp.OoH3WUXeB3 /tmp/tmp.4FiFB4HBwD + return 0 + sleep 10 + wait_for_running some-name-mongos 3 + local name=some-name-mongos + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=mongos + local cluster_name=some-name ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + [[ 0 -eq 2 ]] + wait_pod some-name-mongos-0 + local pod=some-name-mongos-0 + set +o xtrace waiting for pod/some-name-mongos-0 to be ready.OK + for i in '$(seq 0 $last_pod)' + [[ 1 -eq 2 ]] + wait_pod some-name-mongos-1 + local pod=some-name-mongos-1 + set +o xtrace waiting for pod/some-name-mongos-1 to be ready.OK + for i in '$(seq 0 $last_pod)' + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zcNOjd0EBr +++ mktemp ++ local LAST_ERR=/tmp/tmp.Oi4HqMFw0t ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zcNOjd0EBr ++ cat /tmp/tmp.Oi4HqMFw0t ++ rm /tmp/tmp.zcNOjd0EBr /tmp/tmp.Oi4HqMFw0t ++ return 0 + [[ '' == \t\r\u\e ]] + wait_pod some-name-mongos-2 + local pod=some-name-mongos-2 + set +o xtrace waiting for pod/some-name-mongos-2 to be ready.OK ++ kubectl_bin get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].non_voting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yrRT83itMC +++ mktemp ++ local LAST_ERR=/tmp/tmp.6ODJv42QRh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.spec.replsets[?(@.name=="mongos")].non_voting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.yrRT83itMC ++ cat /tmp/tmp.6ODJv42QRh ++ rm /tmp/tmp.yrRT83itMC /tmp/tmp.6ODJv42QRh ++ return 0 + [[ '' == \t\r\u\e ]] + sleep 10 + [[ true == \t\r\u\e ]] + set +x Waiting for cluster readyness........................................................................................... + compare_kubectl statefulset/some-name-mongos -secret + local resource=statefulset/some-name-mongos + local postfix=-secret + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-mongos-secret.yml + local new_result=/tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-mongos-secret-oc.yml ']' + yq eval ' del(.metadata.ownerReferences[].apiVersion) | del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) | del(.metadata.selfLink) | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.status) | (.. | select(tag == "!!str")) |= sub("demand-backup-sharded-7957", "NAME_SPACE") | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "extensions/v1beta1")) = "apps/v1" | (.. | select(. == "batch/v1beta1")) = "batch/v1" ' - + kubectl_bin get -o yaml statefulset/some-name-mongos ++ mktemp + local LAST_OUT=/tmp/tmp.XHymfuMnLb ++ mktemp + local LAST_ERR=/tmp/tmp.u9tnYWjSTx + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-mongos + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.XHymfuMnLb + cat /tmp/tmp.u9tnYWjSTx + rm /tmp/tmp.XHymfuMnLb /tmp/tmp.u9tnYWjSTx + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + version_gt 1.22 ++ bc -l ++ echo '1.26 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-mongos-secret.yml == */cronjob* ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/statefulset_some-name-mongos-secret.yml /tmp/tmp.bLUg6fPDea/statefulset_some-name-mongos.yml + desc 'write data, read from all' + set +o xtrace ----------------------------------------------------------------------------------- write data, read from all ----------------------------------------------------------------------------------- + run_mongos 'db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"},{db:"myApp1",role:"readWrite"},{db:"myApp2",role:"readWrite"}]})' userAdmin:userAdmin123456@some-name-mongos.demand-backup-sharded-7957 + local 'command=db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"},{db:"myApp1",role:"readWrite"},{db:"myApp2",role:"readWrite"}]})' + local uri=userAdmin:userAdmin123456@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DGG6qtAXVw +++ mktemp ++ local LAST_ERR=/tmp/tmp.WhubpYYyi9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DGG6qtAXVw ++ cat /tmp/tmp.WhubpYYyi9 ++ rm /tmp/tmp.DGG6qtAXVw /tmp/tmp.WhubpYYyi9 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"},{db:"myApp1",role:"readWrite"},{db:"myApp2",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.OS7pU7SciU ++ mktemp + local LAST_ERR=/tmp/tmp.EVE5PHEEHG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"},{db:"myApp1",role:"readWrite"},{db:"myApp2",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OS7pU7SciU Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("266d41ac-2cde-42b2-a660-f496d9ba326a") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" }, { "db" : "myApp1", "role" : "readWrite" }, { "db" : "myApp2", "role" : "readWrite" } ] } bye + cat /tmp/tmp.EVE5PHEEHG + rm /tmp/tmp.OS7pU7SciU /tmp/tmp.EVE5PHEEHG + return 0 + run_mongos 'sh.enableSharding("myApp","rs0")' clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957 + local 'command=sh.enableSharding("myApp","rs0")' + local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dup7YEFCML +++ mktemp ++ local LAST_ERR=/tmp/tmp.6Zs9bUeyAk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dup7YEFCML ++ cat /tmp/tmp.6Zs9bUeyAk ++ rm /tmp/tmp.dup7YEFCML /tmp/tmp.6Zs9bUeyAk ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''sh.enableSharding("myApp","rs0")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.MMGUeZ3YxR ++ mktemp + local LAST_ERR=/tmp/tmp.syYOwwshtH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''sh.enableSharding("myApp","rs0")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MMGUeZ3YxR Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("e27fa3ae-e435-42cb-8b76-abbf5db3b03a") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match { "ok" : 1, "$clusterTime" : { "clusterTime" : Timestamp(1720373946, 14), "signature" : { "hash" : BinData(0,"hzqVmq21MW0dUuDBWvbw2HqGcCs="), "keyId" : NumberLong("7388948138448388119") } }, "operationTime" : Timestamp(1720373946, 8) } bye + cat /tmp/tmp.syYOwwshtH + rm /tmp/tmp.MMGUeZ3YxR /tmp/tmp.syYOwwshtH + return 0 + run_mongos 'sh.enableSharding("myApp1","rs1")' clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957 + local 'command=sh.enableSharding("myApp1","rs1")' + local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WjYSU8hFAt +++ mktemp ++ local LAST_ERR=/tmp/tmp.PU3aydY1pF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.WjYSU8hFAt ++ cat /tmp/tmp.PU3aydY1pF ++ rm /tmp/tmp.WjYSU8hFAt /tmp/tmp.PU3aydY1pF ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''sh.enableSharding("myApp1","rs1")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.7P99IXzrcr ++ mktemp + local LAST_ERR=/tmp/tmp.2hm03shOi0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''sh.enableSharding("myApp1","rs1")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7P99IXzrcr Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("d14a9319-5c86-427e-9222-745ce5f51f2e") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match { "ok" : 1, "$clusterTime" : { "clusterTime" : Timestamp(1720373949, 9), "signature" : { "hash" : BinData(0,"NU9IQYY45Apm56+AHF+CkXNrgAM="), "keyId" : NumberLong("7388948138448388119") } }, "operationTime" : Timestamp(1720373949, 4) } bye + cat /tmp/tmp.2hm03shOi0 + rm /tmp/tmp.7P99IXzrcr /tmp/tmp.2hm03shOi0 + return 0 + run_mongos 'sh.enableSharding("myApp2","rs2")' clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957 + local 'command=sh.enableSharding("myApp2","rs2")' + local uri=clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GpJPYtSqA8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.eA1QSURs2D ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.GpJPYtSqA8 ++ cat /tmp/tmp.eA1QSURs2D ++ rm /tmp/tmp.GpJPYtSqA8 /tmp/tmp.eA1QSURs2D ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''sh.enableSharding("myApp2","rs2")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.zUv4CyMuMn ++ mktemp + local LAST_ERR=/tmp/tmp.k0vYJmHCcX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''sh.enableSharding("myApp2","rs2")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.zUv4CyMuMn Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("1a5b9554-c37d-4ca8-aa54-e2cce3f93144") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match { "ok" : 1, "$clusterTime" : { "clusterTime" : Timestamp(1720373952, 9), "signature" : { "hash" : BinData(0,"DnglnFHEoEEHt+X2mwvXl3PB+JY="), "keyId" : NumberLong("7388948138448388119") } }, "operationTime" : Timestamp(1720373952, 4) } bye + cat /tmp/tmp.k0vYJmHCcX + rm /tmp/tmp.zUv4CyMuMn /tmp/tmp.k0vYJmHCcX + return 0 + insert_data_mongos 100500 myApp + local data=100500 + local db_name=myApp + local flags= + run_mongos 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.veAl63sdz1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.O2LBKz1owM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.veAl63sdz1 ++ cat /tmp/tmp.O2LBKz1owM ++ rm /tmp/tmp.veAl63sdz1 /tmp/tmp.O2LBKz1owM ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.tQMAR1vgqM ++ mktemp + local LAST_ERR=/tmp/tmp.lFvgnybWEQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.tQMAR1vgqM Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("c53245d5-5e75-4b05-956c-eb5f64274d72") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.lFvgnybWEQ + rm /tmp/tmp.tQMAR1vgqM /tmp/tmp.lFvgnybWEQ + return 0 + insert_data_mongos 100500 myApp1 + local data=100500 + local db_name=myApp1 + local flags= + run_mongos 'use myApp1\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp1\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o9N7lgSmeh +++ mktemp ++ local LAST_ERR=/tmp/tmp.EEY1fn5gmh ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.o9N7lgSmeh ++ cat /tmp/tmp.EEY1fn5gmh ++ rm /tmp/tmp.o9N7lgSmeh /tmp/tmp.EEY1fn5gmh ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.T3sJTFL6kY ++ mktemp + local LAST_ERR=/tmp/tmp.PSmh0OmmgY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.T3sJTFL6kY Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("56298e10-e538-4cb0-9d12-deed03efcfed") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp1 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.PSmh0OmmgY + rm /tmp/tmp.T3sJTFL6kY /tmp/tmp.PSmh0OmmgY + return 0 + insert_data_mongos 100500 myApp2 + local data=100500 + local db_name=myApp2 + local flags= + run_mongos 'use myApp2\n db.test.insert({ x: 100500 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp2\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rzn7PUbD9P +++ mktemp ++ local LAST_ERR=/tmp/tmp.J7miJc92st ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rzn7PUbD9P ++ cat /tmp/tmp.J7miJc92st ++ rm /tmp/tmp.rzn7PUbD9P /tmp/tmp.J7miJc92st ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.wGJfuzBLoQ ++ mktemp + local LAST_ERR=/tmp/tmp.iJZLl3RQoA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wGJfuzBLoQ Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("01c5f29c-ed8e-40a5-b8d4-c5ef0247e50d") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp2 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.iJZLl3RQoA + rm /tmp/tmp.wGJfuzBLoQ /tmp/tmp.iJZLl3RQoA + return 0 + minikube_sleep + sleep_time=10 + [[ '' == 1 ]] + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix= + local suffix= + local database=myApp + local collection=test + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb '' + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hCidXUFFf1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cKai1tGtxs ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hCidXUFFf1 ++ cat /tmp/tmp.cKai1tGtxs ++ rm /tmp/tmp.hCidXUFFf1 /tmp/tmp.cKai1tGtxs ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.fyauwBXTl6 ++ mktemp + local LAST_ERR=/tmp/tmp.nhS86Nd5tH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fyauwBXTl6 + cat /tmp/tmp.nhS86Nd5tH + rm /tmp/tmp.fyauwBXTl6 /tmp/tmp.nhS86Nd5tH + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find.json /tmp/tmp.bLUg6fPDea/find + wait_backup_agent some-name-rs0-0 + local agent_pod=some-name-rs0-0 + set +o xtrace some-name-rs0-0 + wait_backup_agent some-name-rs0-1 + local agent_pod=some-name-rs0-1 + set +o xtrace some-name-rs0-1 + wait_backup_agent some-name-rs0-2 + local agent_pod=some-name-rs0-2 + set +o xtrace some-name-rs0-2 + wait_backup_agent some-name-rs1-0 + local agent_pod=some-name-rs1-0 + set +o xtrace some-name-rs1-0 + wait_backup_agent some-name-rs1-1 + local agent_pod=some-name-rs1-1 + set +o xtrace some-name-rs1-1 + wait_backup_agent some-name-rs1-2 + local agent_pod=some-name-rs1-2 + set +o xtrace some-name-rs1-2 + wait_backup_agent some-name-rs2-0 + local agent_pod=some-name-rs2-0 + set +o xtrace some-name-rs2-0 + wait_backup_agent some-name-rs2-1 + local agent_pod=some-name-rs2-1 + set +o xtrace some-name-rs2-1 + wait_backup_agent some-name-rs2-2 + local agent_pod=some-name-rs2-2 + set +o xtrace some-name-rs2-2 + backup_name_aws=backup-aws-s3 + backup_name_minio=backup-minio + backup_name_gcp=backup-gcp-cs + backup_name_azure=backup-azure-blob + desc 'run backups' + set +o xtrace ----------------------------------------------------------------------------------- run backups ----------------------------------------------------------------------------------- + run_backup minio + local storage=minio + local backup_name=backup-minio + desc 'run backup backup-minio' + set +o xtrace ----------------------------------------------------------------------------------- run backup backup-minio ----------------------------------------------------------------------------------- + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.21e4qrrnBF ++ mktemp + local LAST_ERR=/tmp/tmp.kMSi39HV5Q + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + yq eval '.metadata.name = "backup-minio" | .spec.storageName = "minio"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/backup-minio.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.21e4qrrnBF perconaservermongodbbackup.psmdb.percona.com/backup-minio created + cat /tmp/tmp.kMSi39HV5Q + rm /tmp/tmp.21e4qrrnBF /tmp/tmp.kMSi39HV5Q + return 0 + '[' -z '' ']' + run_backup aws-s3 + local storage=aws-s3 + local backup_name=backup-aws-s3 + desc 'run backup backup-aws-s3' + set +o xtrace ----------------------------------------------------------------------------------- run backup backup-aws-s3 ----------------------------------------------------------------------------------- + yq eval '.metadata.name = "backup-aws-s3" | .spec.storageName = "aws-s3"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/backup-aws-s3.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.wiWq5KQpKq ++ mktemp + local LAST_ERR=/tmp/tmp.KwTsk0XKC9 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wiWq5KQpKq perconaservermongodbbackup.psmdb.percona.com/backup-aws-s3 created + cat /tmp/tmp.KwTsk0XKC9 + rm /tmp/tmp.wiWq5KQpKq /tmp/tmp.KwTsk0XKC9 + return 0 + run_backup gcp-cs + local storage=gcp-cs + local backup_name=backup-gcp-cs + desc 'run backup backup-gcp-cs' + set +o xtrace ----------------------------------------------------------------------------------- run backup backup-gcp-cs ----------------------------------------------------------------------------------- + yq eval '.metadata.name = "backup-gcp-cs" | .spec.storageName = "gcp-cs"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/backup-gcp-cs.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.gnkju2SVlc ++ mktemp + local LAST_ERR=/tmp/tmp.72HRurIiVB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.gnkju2SVlc perconaservermongodbbackup.psmdb.percona.com/backup-gcp-cs created + cat /tmp/tmp.72HRurIiVB + rm /tmp/tmp.gnkju2SVlc /tmp/tmp.72HRurIiVB + return 0 + run_backup azure-blob + local storage=azure-blob + local backup_name=backup-azure-blob + desc 'run backup backup-azure-blob' + set +o xtrace ----------------------------------------------------------------------------------- run backup backup-azure-blob ----------------------------------------------------------------------------------- + yq eval '.metadata.name = "backup-azure-blob" | .spec.storageName = "azure-blob"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/backup-azure-blob.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.CrQcC9ZMs7 ++ mktemp + local LAST_ERR=/tmp/tmp.fnWGIKTh6H + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CrQcC9ZMs7 perconaservermongodbbackup.psmdb.percona.com/backup-azure-blob created + cat /tmp/tmp.fnWGIKTh6H + rm /tmp/tmp.CrQcC9ZMs7 /tmp/tmp.fnWGIKTh6H + return 0 + wait_backup backup-aws-s3 + local backup_name=backup-aws-s3 + set +o xtrace backup-aws-s3..................................... + wait_backup backup-gcp-cs + local backup_name=backup-gcp-cs + set +o xtrace backup-gcp-cs................ + wait_backup backup-azure-blob + local backup_name=backup-azure-blob + set +o xtrace backup-azure-blob....................................... + wait_backup backup-minio + local backup_name=backup-minio + set +o xtrace backup-minio. + sleep 5 + '[' -z '' ']' + desc 'check backup and restore -- aws-s3' + set +o xtrace ----------------------------------------------------------------------------------- check backup and restore -- aws-s3 ----------------------------------------------------------------------------------- ++ get_backup_dest backup-aws-s3 ++ local backup_name=backup-aws-s3 ++ kubectl_bin get psmdb-backup backup-aws-s3 -o 'jsonpath={.status.destination}' ++ sed -e 's/.json$//' +++ mktemp ++ sed 's|s3://||' ++ sed 's|azure://||' ++ local LAST_OUT=/tmp/tmp.8xVxAzO21x +++ mktemp ++ local LAST_ERR=/tmp/tmp.cmxklX7tFx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb-backup backup-aws-s3 -o 'jsonpath={.status.destination}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8xVxAzO21x ++ cat /tmp/tmp.cmxklX7tFx ++ rm /tmp/tmp.8xVxAzO21x /tmp/tmp.cmxklX7tFx ++ return 0 + backup_dest_aws=operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:40:35Z + curl -s https://s3.amazonaws.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:40:35Z/rs0/myApp.test.gz + gunzip + gunzip + curl -s https://s3.amazonaws.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:40:35Z/rs1/myApp1.test.gz + curl -s https://s3.amazonaws.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:40:35Z/rs2/myApp2.test.gz + gunzip + insert_data_mongos 100501 myApp + local data=100501 + local db_name=myApp + local flags= + run_mongos 'use myApp\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sSyNSB3IU5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.3RrrfVuqQE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.sSyNSB3IU5 ++ cat /tmp/tmp.3RrrfVuqQE ++ rm /tmp/tmp.sSyNSB3IU5 /tmp/tmp.3RrrfVuqQE ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.P4AqtdB8Fu ++ mktemp + local LAST_ERR=/tmp/tmp.y6Xoin06po + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.P4AqtdB8Fu Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("916049ae-15ab-4448-af20-460b9daacf66") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.y6Xoin06po + rm /tmp/tmp.P4AqtdB8Fu /tmp/tmp.y6Xoin06po + return 0 + insert_data_mongos 100501 myApp1 + local data=100501 + local db_name=myApp1 + local flags= + run_mongos 'use myApp1\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp1\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MWuWr0WIp2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.N3VtHEgZPX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MWuWr0WIp2 ++ cat /tmp/tmp.N3VtHEgZPX ++ rm /tmp/tmp.MWuWr0WIp2 /tmp/tmp.N3VtHEgZPX ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.BmvusMBpiO ++ mktemp + local LAST_ERR=/tmp/tmp.AcehfCgnKs + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.BmvusMBpiO Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("931a5f2d-8442-4843-b12a-7ae3dba8ad60") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp1 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.AcehfCgnKs + rm /tmp/tmp.BmvusMBpiO /tmp/tmp.AcehfCgnKs + return 0 + insert_data_mongos 100501 myApp2 + local data=100501 + local db_name=myApp2 + local flags= + run_mongos 'use myApp2\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp2\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.w3YBbSoeZE +++ mktemp ++ local LAST_ERR=/tmp/tmp.6KnYifOOGc ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.w3YBbSoeZE ++ cat /tmp/tmp.6KnYifOOGc ++ rm /tmp/tmp.w3YBbSoeZE /tmp/tmp.6KnYifOOGc ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.y4zRthnIPN ++ mktemp + local LAST_ERR=/tmp/tmp.xuGiNWDIaA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.y4zRthnIPN Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("8583a6ef-3821-4cbe-8661-c0c3b7e1bcfd") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp2 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.xuGiNWDIaA + rm /tmp/tmp.y4zRthnIPN /tmp/tmp.xuGiNWDIaA + return 0 + check_data -2nd + local postfix=-2nd ++ seq 0 2 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd + local suffix=.svc.cluster.local + local database=myApp + local collection=test + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp\n db.test.find()' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5awsCb5wHk +++ mktemp ++ local LAST_ERR=/tmp/tmp.jQ0khHCJkk ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5awsCb5wHk ++ cat /tmp/tmp.jQ0khHCJkk ++ rm /tmp/tmp.5awsCb5wHk /tmp/tmp.jQ0khHCJkk ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.vcz0NIG8lw ++ mktemp + local LAST_ERR=/tmp/tmp.nR9wVXRsI8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.vcz0NIG8lw + cat /tmp/tmp.nR9wVXRsI8 + rm /tmp/tmp.vcz0NIG8lw /tmp/tmp.nR9wVXRsI8 + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd.json /tmp/tmp.bLUg6fPDea/find-2nd + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd1 .svc.cluster.local myApp1 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd1 + local suffix=.svc.cluster.local + local database=myApp1 + local collection=test + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + run_mongos 'use myApp1\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp1\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.y1KSOyQtTC +++ mktemp ++ local LAST_ERR=/tmp/tmp.TvVULEvl2s ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.y1KSOyQtTC ++ cat /tmp/tmp.TvVULEvl2s ++ rm /tmp/tmp.y1KSOyQtTC /tmp/tmp.TvVULEvl2s ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.Q0ob2wgYhQ ++ mktemp + local LAST_ERR=/tmp/tmp.HZjE1kVnfC + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Q0ob2wgYhQ + cat /tmp/tmp.HZjE1kVnfC + rm /tmp/tmp.Q0ob2wgYhQ /tmp/tmp.HZjE1kVnfC + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd1.json /tmp/tmp.bLUg6fPDea/find-2nd1 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd2 .svc.cluster.local myApp2 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd2 + local suffix=.svc.cluster.local + local database=myApp2 + local collection=test + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + run_mongos 'use myApp2\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp2\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ local LAST_OUT=/tmp/tmp.ijUJJihEMS +++ mktemp ++ local LAST_ERR=/tmp/tmp.trOtXii07K ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ijUJJihEMS ++ cat /tmp/tmp.trOtXii07K ++ rm /tmp/tmp.ijUJJihEMS /tmp/tmp.trOtXii07K ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.GSQRBOJ4x4 ++ mktemp + local LAST_ERR=/tmp/tmp.xpdlozxfNt + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.GSQRBOJ4x4 + cat /tmp/tmp.xpdlozxfNt + rm /tmp/tmp.GSQRBOJ4x4 /tmp/tmp.xpdlozxfNt + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd2.json /tmp/tmp.bLUg6fPDea/find-2nd2 + run_restore backup-aws-s3 + local backup_name=backup-aws-s3 + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/restore.yml + /usr/bin/sed -e 's/name:/name: restore-backup-aws-s3/' + /usr/bin/sed -e 's/backupName:/backupName: backup-aws-s3/' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.6MtI5aQyFJ ++ mktemp + local LAST_ERR=/tmp/tmp.U1DrNSX6Az + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6MtI5aQyFJ perconaservermongodbrestore.psmdb.percona.com/restore-backup-aws-s3 created + cat /tmp/tmp.U1DrNSX6Az + rm /tmp/tmp.6MtI5aQyFJ /tmp/tmp.U1DrNSX6Az + return 0 + wait_restore backup-aws-s3 some-name + local backup_name=backup-aws-s3 + local cluster_name=some-name + local target_state=ready + local wait_cluster_consistency=1 + local wait_time=780 + set +o xtrace waiting psmdb-restore/backup-aws-s3 to reach ready state.............. + '[' 1 -eq 1 ']' + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Wl0VYWVCCq +++ mktemp ++ local LAST_ERR=/tmp/tmp.Vn0KGp4gUm ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Wl0VYWVCCq ++ cat /tmp/tmp.Vn0KGp4gUm ++ rm /tmp/tmp.Wl0VYWVCCq /tmp/tmp.Vn0KGp4gUm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y7yBwKFf1n +++ mktemp ++ local LAST_ERR=/tmp/tmp.ryw05dHCkM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Y7yBwKFf1n ++ cat /tmp/tmp.ryw05dHCkM ++ rm /tmp/tmp.Y7yBwKFf1n /tmp/tmp.ryw05dHCkM ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MpAQJ2aBnH +++ mktemp ++ local LAST_ERR=/tmp/tmp.akg0xdpdxL ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MpAQJ2aBnH ++ cat /tmp/tmp.akg0xdpdxL ++ rm /tmp/tmp.MpAQJ2aBnH /tmp/tmp.akg0xdpdxL ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2aSxxyJQrb +++ mktemp ++ local LAST_ERR=/tmp/tmp.7NUHOKGJdX ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2aSxxyJQrb ++ cat /tmp/tmp.7NUHOKGJdX ++ rm /tmp/tmp.2aSxxyJQrb /tmp/tmp.7NUHOKGJdX ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dvUkdyTMaI +++ mktemp ++ local LAST_ERR=/tmp/tmp.2X963CgCo3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dvUkdyTMaI ++ cat /tmp/tmp.2X963CgCo3 ++ rm /tmp/tmp.dvUkdyTMaI /tmp/tmp.2X963CgCo3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EHzjIun1ua +++ mktemp ++ local LAST_ERR=/tmp/tmp.D6fxsUPvna ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.EHzjIun1ua ++ cat /tmp/tmp.D6fxsUPvna ++ rm /tmp/tmp.EHzjIun1ua /tmp/tmp.D6fxsUPvna ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CTmNBKAZOB +++ mktemp ++ local LAST_ERR=/tmp/tmp.GYPJaKS5Ob ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.CTmNBKAZOB ++ cat /tmp/tmp.GYPJaKS5Ob ++ rm /tmp/tmp.CTmNBKAZOB /tmp/tmp.GYPJaKS5Ob ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 7 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rn4YdpZLMf +++ mktemp ++ local LAST_ERR=/tmp/tmp.K1wIlq4cm1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.rn4YdpZLMf ++ cat /tmp/tmp.K1wIlq4cm1 ++ rm /tmp/tmp.rn4YdpZLMf /tmp/tmp.K1wIlq4cm1 ++ return 0 + [[ ready == \r\e\a\d\y ]] + check_data + local postfix= ++ seq 0 2 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix= + local suffix=.svc.cluster.local + local database=myApp + local collection=test + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' +++ mktemp + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ local LAST_OUT=/tmp/tmp.04QuqULapC +++ mktemp ++ local LAST_ERR=/tmp/tmp.GGBE8k3wB4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.04QuqULapC ++ cat /tmp/tmp.GGBE8k3wB4 ++ rm /tmp/tmp.04QuqULapC /tmp/tmp.GGBE8k3wB4 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.2p1Xij9PUq ++ mktemp + local LAST_ERR=/tmp/tmp.clroDPuWCP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2p1Xij9PUq + cat /tmp/tmp.clroDPuWCP + rm /tmp/tmp.2p1Xij9PUq /tmp/tmp.clroDPuWCP + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find.json /tmp/tmp.bLUg6fPDea/find + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 1 .svc.cluster.local myApp1 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=1 + local suffix=.svc.cluster.local + local database=myApp1 + local collection=test + run_mongos 'use myApp1\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp1\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ local LAST_OUT=/tmp/tmp.UzQlzdu8sZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.OQNDVuWaxb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.UzQlzdu8sZ ++ cat /tmp/tmp.OQNDVuWaxb ++ rm /tmp/tmp.UzQlzdu8sZ /tmp/tmp.OQNDVuWaxb ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.Eo6P05Jl4M ++ mktemp + local LAST_ERR=/tmp/tmp.n0laBbzD9c + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Eo6P05Jl4M + cat /tmp/tmp.n0laBbzD9c + rm /tmp/tmp.Eo6P05Jl4M /tmp/tmp.n0laBbzD9c + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find1.json /tmp/tmp.bLUg6fPDea/find1 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 2 .svc.cluster.local myApp2 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=2 + local suffix=.svc.cluster.local + local database=myApp2 + local collection=test + run_mongos 'use myApp2\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp2\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ih9aSQEm6G +++ mktemp ++ local LAST_ERR=/tmp/tmp.VhWcq6vFCA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ih9aSQEm6G ++ cat /tmp/tmp.VhWcq6vFCA ++ rm /tmp/tmp.Ih9aSQEm6G /tmp/tmp.VhWcq6vFCA ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.EI1jhBOK4c ++ mktemp + local LAST_ERR=/tmp/tmp.iNiGofiNBE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.EI1jhBOK4c + cat /tmp/tmp.iNiGofiNBE + rm /tmp/tmp.EI1jhBOK4c /tmp/tmp.iNiGofiNBE + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find2.json /tmp/tmp.bLUg6fPDea/find2 + desc 'check backup and restore -- gcp-cs' + set +o xtrace ----------------------------------------------------------------------------------- check backup and restore -- gcp-cs ----------------------------------------------------------------------------------- ++ get_backup_dest backup-gcp-cs ++ local backup_name=backup-gcp-cs ++ kubectl_bin get psmdb-backup backup-gcp-cs -o 'jsonpath={.status.destination}' ++ sed -e 's/.json$//' ++ sed 's|azure://||' ++ sed 's|s3://||' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XvQ3lJvjuD +++ mktemp ++ local LAST_ERR=/tmp/tmp.wzrxjK6fXW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb-backup backup-gcp-cs -o 'jsonpath={.status.destination}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XvQ3lJvjuD ++ cat /tmp/tmp.wzrxjK6fXW ++ rm /tmp/tmp.XvQ3lJvjuD /tmp/tmp.wzrxjK6fXW ++ return 0 + backup_dest_gcp=operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:17Z + gunzip + curl -s https://storage.googleapis.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:17Z/rs0/myApp.test.gz + curl -s https://storage.googleapis.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:17Z/rs1/myApp1.test.gz + gunzip + gunzip + curl -s https://storage.googleapis.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:17Z/rs2/myApp2.test.gz + insert_data_mongos 100501 myApp + local data=100501 + local db_name=myApp + local flags= + run_mongos 'use myApp\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oT6qvzPSDL +++ mktemp ++ local LAST_ERR=/tmp/tmp.U5VYzLXPMV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.oT6qvzPSDL ++ cat /tmp/tmp.U5VYzLXPMV ++ rm /tmp/tmp.oT6qvzPSDL /tmp/tmp.U5VYzLXPMV ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.teyd7Bu748 ++ mktemp + local LAST_ERR=/tmp/tmp.hj9Y0UpcMR + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.teyd7Bu748 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("9a63ca7f-3ad8-4b18-a1c9-61fbd4ef0b67") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.hj9Y0UpcMR + rm /tmp/tmp.teyd7Bu748 /tmp/tmp.hj9Y0UpcMR + return 0 + insert_data_mongos 100501 myApp1 + local data=100501 + local db_name=myApp1 + local flags= + run_mongos 'use myApp1\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp1\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3TntQWxruN +++ mktemp ++ local LAST_ERR=/tmp/tmp.X94hwYSneS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3TntQWxruN ++ cat /tmp/tmp.X94hwYSneS ++ rm /tmp/tmp.3TntQWxruN /tmp/tmp.X94hwYSneS ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.o1roboCItm ++ mktemp + local LAST_ERR=/tmp/tmp.1otCT2Pv1H + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.o1roboCItm Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("22c08a34-35b2-449f-a103-47e0311adc6b") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp1 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.1otCT2Pv1H + rm /tmp/tmp.o1roboCItm /tmp/tmp.1otCT2Pv1H + return 0 + insert_data_mongos 100501 myApp2 + local data=100501 + local db_name=myApp2 + local flags= + run_mongos 'use myApp2\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp2\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iWHF4zaotu +++ mktemp ++ local LAST_ERR=/tmp/tmp.7y5fYxHMr3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iWHF4zaotu ++ cat /tmp/tmp.7y5fYxHMr3 ++ rm /tmp/tmp.iWHF4zaotu /tmp/tmp.7y5fYxHMr3 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.N5XJ0KJaB1 ++ mktemp + local LAST_ERR=/tmp/tmp.xLgZOj635U + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.N5XJ0KJaB1 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("67c6dc53-a362-4217-a7c7-1fc4972e5b54") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp2 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.xLgZOj635U + rm /tmp/tmp.N5XJ0KJaB1 /tmp/tmp.xLgZOj635U + return 0 + check_data -2nd + local postfix=-2nd ++ seq 0 2 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd + local suffix=.svc.cluster.local + local database=myApp + local collection=test + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9xoyXYdRlE +++ mktemp ++ local LAST_ERR=/tmp/tmp.GspRKOayuW ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9xoyXYdRlE ++ cat /tmp/tmp.GspRKOayuW ++ rm /tmp/tmp.9xoyXYdRlE /tmp/tmp.GspRKOayuW ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.PT2aZZyeBt ++ mktemp + local LAST_ERR=/tmp/tmp.cbTXprwu1E + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PT2aZZyeBt + cat /tmp/tmp.cbTXprwu1E + rm /tmp/tmp.PT2aZZyeBt /tmp/tmp.cbTXprwu1E + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd.json /tmp/tmp.bLUg6fPDea/find-2nd + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd1 .svc.cluster.local myApp1 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd1 + local suffix=.svc.cluster.local + local database=myApp1 + local collection=test + run_mongos 'use myApp1\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp1\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8IznfG12j9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.eW4WMeEVRU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8IznfG12j9 ++ cat /tmp/tmp.eW4WMeEVRU ++ rm /tmp/tmp.8IznfG12j9 /tmp/tmp.eW4WMeEVRU ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.z2kvN1wtOW ++ mktemp + local LAST_ERR=/tmp/tmp.UE9EaYqrjT + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.z2kvN1wtOW + cat /tmp/tmp.UE9EaYqrjT + rm /tmp/tmp.z2kvN1wtOW /tmp/tmp.UE9EaYqrjT + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd1.json /tmp/tmp.bLUg6fPDea/find-2nd1 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd2 .svc.cluster.local myApp2 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd2 + local suffix=.svc.cluster.local + local database=myApp2 + local collection=test + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + run_mongos 'use myApp2\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + local 'command=use myApp2\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QbSKe16l2L +++ mktemp ++ local LAST_ERR=/tmp/tmp.J3cxaquirO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.QbSKe16l2L ++ cat /tmp/tmp.J3cxaquirO ++ rm /tmp/tmp.QbSKe16l2L /tmp/tmp.J3cxaquirO ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.c7bSUBDwCS ++ mktemp + local LAST_ERR=/tmp/tmp.PzyWlJ5VLT + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.c7bSUBDwCS + cat /tmp/tmp.PzyWlJ5VLT + rm /tmp/tmp.c7bSUBDwCS /tmp/tmp.PzyWlJ5VLT + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd2.json /tmp/tmp.bLUg6fPDea/find-2nd2 + run_restore backup-gcp-cs + local backup_name=backup-gcp-cs + /usr/bin/sed -e 's/name:/name: restore-backup-gcp-cs/' + /usr/bin/sed -e 's/backupName:/backupName: backup-gcp-cs/' + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/restore.yml ++ mktemp + local LAST_OUT=/tmp/tmp.SLqhDhhfVm ++ mktemp + local LAST_ERR=/tmp/tmp.7224h287xx + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.SLqhDhhfVm perconaservermongodbrestore.psmdb.percona.com/restore-backup-gcp-cs created + cat /tmp/tmp.7224h287xx + rm /tmp/tmp.SLqhDhhfVm /tmp/tmp.7224h287xx + return 0 + wait_restore backup-gcp-cs some-name + local backup_name=backup-gcp-cs + local cluster_name=some-name + local target_state=ready + local wait_cluster_consistency=1 + local wait_time=780 + set +o xtrace waiting psmdb-restore/backup-gcp-cs to reach ready state........... + '[' 1 -eq 1 ']' + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5LWj6ZzFv8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9RC6zYEVA5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5LWj6ZzFv8 ++ cat /tmp/tmp.9RC6zYEVA5 ++ rm /tmp/tmp.5LWj6ZzFv8 /tmp/tmp.9RC6zYEVA5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5Pl7xiVTWB +++ mktemp ++ local LAST_ERR=/tmp/tmp.ddxLtqpJ8e ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5Pl7xiVTWB ++ cat /tmp/tmp.ddxLtqpJ8e ++ rm /tmp/tmp.5Pl7xiVTWB /tmp/tmp.ddxLtqpJ8e ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NyIxHJ1hQO +++ mktemp ++ local LAST_ERR=/tmp/tmp.brJBDeDIF0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.NyIxHJ1hQO ++ cat /tmp/tmp.brJBDeDIF0 ++ rm /tmp/tmp.NyIxHJ1hQO /tmp/tmp.brJBDeDIF0 ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SNsncSLucI +++ mktemp ++ local LAST_ERR=/tmp/tmp.XDgd0DbRuJ ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SNsncSLucI ++ cat /tmp/tmp.XDgd0DbRuJ ++ rm /tmp/tmp.SNsncSLucI /tmp/tmp.XDgd0DbRuJ ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hD0ia1PAMH +++ mktemp ++ local LAST_ERR=/tmp/tmp.sF5famdM8Z ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hD0ia1PAMH ++ cat /tmp/tmp.sF5famdM8Z ++ rm /tmp/tmp.hD0ia1PAMH /tmp/tmp.sF5famdM8Z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RtyfVeDRBR +++ mktemp ++ local LAST_ERR=/tmp/tmp.zSbaX7aLxD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RtyfVeDRBR ++ cat /tmp/tmp.zSbaX7aLxD ++ rm /tmp/tmp.RtyfVeDRBR /tmp/tmp.zSbaX7aLxD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.26FS4Aadec +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZRU5WGetVu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.26FS4Aadec ++ cat /tmp/tmp.ZRU5WGetVu ++ rm /tmp/tmp.26FS4Aadec /tmp/tmp.ZRU5WGetVu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 7 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bGJ9QUZ2M8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.KWsQgioqPP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bGJ9QUZ2M8 ++ cat /tmp/tmp.KWsQgioqPP ++ rm /tmp/tmp.bGJ9QUZ2M8 /tmp/tmp.KWsQgioqPP ++ return 0 + [[ ready == \r\e\a\d\y ]] + check_data + local postfix= ++ seq 0 2 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix= + local suffix=.svc.cluster.local + local database=myApp + local collection=test + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bGTcZTqC7D +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZbSKAjS773 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bGTcZTqC7D ++ cat /tmp/tmp.ZbSKAjS773 ++ rm /tmp/tmp.bGTcZTqC7D /tmp/tmp.ZbSKAjS773 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.sSsIGfhD7h ++ mktemp + local LAST_ERR=/tmp/tmp.TmyszwrOw3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sSsIGfhD7h + cat /tmp/tmp.TmyszwrOw3 + rm /tmp/tmp.sSsIGfhD7h /tmp/tmp.TmyszwrOw3 + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find.json /tmp/tmp.bLUg6fPDea/find + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 1 .svc.cluster.local myApp1 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=1 + local suffix=.svc.cluster.local + local database=myApp1 + local collection=test + run_mongos 'use myApp1\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp1\n db.test.find()' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uDn8i3vrMl +++ mktemp ++ local LAST_ERR=/tmp/tmp.QsvrToDZWO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.uDn8i3vrMl ++ cat /tmp/tmp.QsvrToDZWO ++ rm /tmp/tmp.uDn8i3vrMl /tmp/tmp.QsvrToDZWO ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.5QzwGE8Akv ++ mktemp + local LAST_ERR=/tmp/tmp.1HjtuQ0cg7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5QzwGE8Akv + cat /tmp/tmp.1HjtuQ0cg7 + rm /tmp/tmp.5QzwGE8Akv /tmp/tmp.1HjtuQ0cg7 + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find1.json /tmp/tmp.bLUg6fPDea/find1 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 2 .svc.cluster.local myApp2 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=2 + local suffix=.svc.cluster.local + local database=myApp2 + local collection=test + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + run_mongos 'use myApp2\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + local 'command=use myApp2\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5ardgFo3hX +++ mktemp ++ local LAST_ERR=/tmp/tmp.og9q9IR3s3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5ardgFo3hX ++ cat /tmp/tmp.og9q9IR3s3 ++ rm /tmp/tmp.5ardgFo3hX /tmp/tmp.og9q9IR3s3 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.MmyztzCpCA ++ mktemp + local LAST_ERR=/tmp/tmp.uRcd1zdOc0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MmyztzCpCA + cat /tmp/tmp.uRcd1zdOc0 + rm /tmp/tmp.MmyztzCpCA /tmp/tmp.uRcd1zdOc0 + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find2.json /tmp/tmp.bLUg6fPDea/find2 + desc 'check backup and restore -- azure-blob' + set +o xtrace ----------------------------------------------------------------------------------- check backup and restore -- azure-blob ----------------------------------------------------------------------------------- ++ get_backup_dest backup-azure-blob ++ local backup_name=backup-azure-blob ++ kubectl_bin get psmdb-backup backup-azure-blob -o 'jsonpath={.status.destination}' ++ sed -e 's/.json$//' ++ sed 's|s3://||' +++ mktemp ++ sed 's|azure://||' ++ local LAST_OUT=/tmp/tmp.DiTsA7kZgZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.eRDtQQYthM ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb-backup backup-azure-blob -o 'jsonpath={.status.destination}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DiTsA7kZgZ ++ cat /tmp/tmp.eRDtQQYthM ++ rm /tmp/tmp.DiTsA7kZgZ /tmp/tmp.eRDtQQYthM ++ return 0 + backup_dest_azure=operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:49Z + curl -s https://engk8soperators.blob.core.windows.net/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:49Z/rs0/myApp.test.gz + gunzip + gunzip + curl -s https://engk8soperators.blob.core.windows.net/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:49Z/rs1/myApp1.test.gz + gunzip + curl -s https://engk8soperators.blob.core.windows.net/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:49Z/rs2/myApp2.test.gz + insert_data_mongos 100501 myApp + local data=100501 + local db_name=myApp + local flags= + run_mongos 'use myApp\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y7L4lXzbNn +++ mktemp ++ local LAST_ERR=/tmp/tmp.FW8eTLp5gG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Y7L4lXzbNn ++ cat /tmp/tmp.FW8eTLp5gG ++ rm /tmp/tmp.Y7L4lXzbNn /tmp/tmp.FW8eTLp5gG ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.FuBGf8W5PL ++ mktemp + local LAST_ERR=/tmp/tmp.osZRT6NVXh + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FuBGf8W5PL Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("60ab3300-2a6f-4310-adbf-5d190a454fdf") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.osZRT6NVXh + rm /tmp/tmp.FuBGf8W5PL /tmp/tmp.osZRT6NVXh + return 0 + insert_data_mongos 100501 myApp1 + local data=100501 + local db_name=myApp1 + local flags= + run_mongos 'use myApp1\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp1\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nWRwPebPbZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.9pycKEWXN2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nWRwPebPbZ ++ cat /tmp/tmp.9pycKEWXN2 ++ rm /tmp/tmp.nWRwPebPbZ /tmp/tmp.9pycKEWXN2 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.H4i29U7RAA ++ mktemp + local LAST_ERR=/tmp/tmp.H4Tgnq8CUL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.H4i29U7RAA Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("68927ca2-8037-415f-8e95-b21c39840032") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp1 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.H4Tgnq8CUL + rm /tmp/tmp.H4i29U7RAA /tmp/tmp.H4Tgnq8CUL + return 0 + insert_data_mongos 100501 myApp2 + local data=100501 + local db_name=myApp2 + local flags= + run_mongos 'use myApp2\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp2\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gsSZO9QL3F +++ mktemp ++ local LAST_ERR=/tmp/tmp.hWtTvV8Ik1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gsSZO9QL3F ++ cat /tmp/tmp.hWtTvV8Ik1 ++ rm /tmp/tmp.gsSZO9QL3F /tmp/tmp.hWtTvV8Ik1 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.AcBOZ4TRpR ++ mktemp + local LAST_ERR=/tmp/tmp.PCW48e5YfM + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.AcBOZ4TRpR Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("6a5ce04a-88ee-4154-80cb-9ebe273d116c") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp2 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.PCW48e5YfM + rm /tmp/tmp.AcBOZ4TRpR /tmp/tmp.PCW48e5YfM + return 0 + check_data -2nd + local postfix=-2nd ++ seq 0 2 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd + local suffix=.svc.cluster.local + local database=myApp + local collection=test + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp\n db.test.find()' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.X0noZL5PlW +++ mktemp ++ local LAST_ERR=/tmp/tmp.bRPVAFGSLI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.X0noZL5PlW ++ cat /tmp/tmp.bRPVAFGSLI ++ rm /tmp/tmp.X0noZL5PlW /tmp/tmp.bRPVAFGSLI ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.4c6afwlY0H ++ mktemp + local LAST_ERR=/tmp/tmp.kToJrUhElN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4c6afwlY0H + cat /tmp/tmp.kToJrUhElN + rm /tmp/tmp.4c6afwlY0H /tmp/tmp.kToJrUhElN + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd.json /tmp/tmp.bLUg6fPDea/find-2nd + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd1 .svc.cluster.local myApp1 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd1 + local suffix=.svc.cluster.local + local database=myApp1 + local collection=test + run_mongos 'use myApp1\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp1\n db.test.find()' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.33n7r0KIBs +++ mktemp ++ local LAST_ERR=/tmp/tmp.VFsujCAXwP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.33n7r0KIBs ++ cat /tmp/tmp.VFsujCAXwP ++ rm /tmp/tmp.33n7r0KIBs /tmp/tmp.VFsujCAXwP ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.361ReHriU5 ++ mktemp + local LAST_ERR=/tmp/tmp.2qM29XG6pL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.361ReHriU5 + cat /tmp/tmp.2qM29XG6pL + rm /tmp/tmp.361ReHriU5 /tmp/tmp.2qM29XG6pL + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd1.json /tmp/tmp.bLUg6fPDea/find-2nd1 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd2 .svc.cluster.local myApp2 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd2 + local suffix=.svc.cluster.local + local database=myApp2 + local collection=test + run_mongos 'use myApp2\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp2\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t6oroVSDm4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.QOnNhuZji1 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.t6oroVSDm4 ++ cat /tmp/tmp.QOnNhuZji1 ++ rm /tmp/tmp.t6oroVSDm4 /tmp/tmp.QOnNhuZji1 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.me0JrtsYY2 ++ mktemp + local LAST_ERR=/tmp/tmp.YNxUaXFWdP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.me0JrtsYY2 + cat /tmp/tmp.YNxUaXFWdP + rm /tmp/tmp.me0JrtsYY2 /tmp/tmp.YNxUaXFWdP + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd2.json /tmp/tmp.bLUg6fPDea/find-2nd2 + run_restore backup-azure-blob + local backup_name=backup-azure-blob + /usr/bin/sed -e 's/name:/name: restore-backup-azure-blob/' + /usr/bin/sed -e 's/backupName:/backupName: backup-azure-blob/' + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/restore.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.8F09wZ3LYO ++ mktemp + local LAST_ERR=/tmp/tmp.hVNVoUO25U + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.8F09wZ3LYO perconaservermongodbrestore.psmdb.percona.com/restore-backup-azure-blob created + cat /tmp/tmp.hVNVoUO25U + rm /tmp/tmp.8F09wZ3LYO /tmp/tmp.hVNVoUO25U + return 0 + wait_restore backup-azure-blob some-name + local backup_name=backup-azure-blob + local cluster_name=some-name + local target_state=ready + local wait_cluster_consistency=1 + local wait_time=780 + set +o xtrace waiting psmdb-restore/backup-azure-blob to reach ready state............ + '[' 1 -eq 1 ']' + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8ypWskzAH8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.iyuU3lZzkq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8ypWskzAH8 ++ cat /tmp/tmp.iyuU3lZzkq ++ rm /tmp/tmp.8ypWskzAH8 /tmp/tmp.iyuU3lZzkq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YD36ttPfuA +++ mktemp ++ local LAST_ERR=/tmp/tmp.IU6m2bzj2M ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.YD36ttPfuA ++ cat /tmp/tmp.IU6m2bzj2M ++ rm /tmp/tmp.YD36ttPfuA /tmp/tmp.IU6m2bzj2M ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nQKhjeeVs8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FyRCVHMAB6 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.nQKhjeeVs8 ++ cat /tmp/tmp.FyRCVHMAB6 ++ rm /tmp/tmp.nQKhjeeVs8 /tmp/tmp.FyRCVHMAB6 ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4LNScDUzjJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.dC9KNjG9KI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4LNScDUzjJ ++ cat /tmp/tmp.dC9KNjG9KI ++ rm /tmp/tmp.4LNScDUzjJ /tmp/tmp.dC9KNjG9KI ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cSlZllz4DZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.sXSVMFYl9m ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.cSlZllz4DZ ++ cat /tmp/tmp.sXSVMFYl9m ++ rm /tmp/tmp.cSlZllz4DZ /tmp/tmp.sXSVMFYl9m ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jvIi7rQr8c +++ mktemp ++ local LAST_ERR=/tmp/tmp.8igXtQNZKA ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jvIi7rQr8c ++ cat /tmp/tmp.8igXtQNZKA ++ rm /tmp/tmp.jvIi7rQr8c /tmp/tmp.8igXtQNZKA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xadImgQKbe +++ mktemp ++ local LAST_ERR=/tmp/tmp.DjJ7ahDRXb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.xadImgQKbe ++ cat /tmp/tmp.DjJ7ahDRXb ++ rm /tmp/tmp.xadImgQKbe /tmp/tmp.DjJ7ahDRXb ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 7 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HR8O5YNBP2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.KlE3MBkIM2 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.HR8O5YNBP2 ++ cat /tmp/tmp.KlE3MBkIM2 ++ rm /tmp/tmp.HR8O5YNBP2 /tmp/tmp.KlE3MBkIM2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 8 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MDKWQojCJ5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.3V3BCYi2TK ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MDKWQojCJ5 ++ cat /tmp/tmp.3V3BCYi2TK ++ rm /tmp/tmp.MDKWQojCJ5 /tmp/tmp.3V3BCYi2TK ++ return 0 + [[ ready == \r\e\a\d\y ]] + check_data + local postfix= ++ seq 0 2 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix= + local suffix=.svc.cluster.local + local database=myApp + local collection=test + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LwpTxYoFbs +++ mktemp ++ local LAST_ERR=/tmp/tmp.MDncWTLcvF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.LwpTxYoFbs ++ cat /tmp/tmp.MDncWTLcvF ++ rm /tmp/tmp.LwpTxYoFbs /tmp/tmp.MDncWTLcvF ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.FonsPRkxRv ++ mktemp + local LAST_ERR=/tmp/tmp.i8CJQjw5VX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FonsPRkxRv + cat /tmp/tmp.i8CJQjw5VX + rm /tmp/tmp.FonsPRkxRv /tmp/tmp.i8CJQjw5VX + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find.json /tmp/tmp.bLUg6fPDea/find + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 1 .svc.cluster.local myApp1 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=1 + local suffix=.svc.cluster.local + local database=myApp1 + local collection=test + run_mongos 'use myApp1\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp1\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZafKPMFNxd +++ mktemp ++ local LAST_ERR=/tmp/tmp.Wvj164cAK8 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZafKPMFNxd ++ cat /tmp/tmp.Wvj164cAK8 ++ rm /tmp/tmp.ZafKPMFNxd /tmp/tmp.Wvj164cAK8 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.4iHOM1kD46 ++ mktemp + local LAST_ERR=/tmp/tmp.zzd1uMEUGD + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4iHOM1kD46 + cat /tmp/tmp.zzd1uMEUGD + rm /tmp/tmp.4iHOM1kD46 /tmp/tmp.zzd1uMEUGD + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find1.json /tmp/tmp.bLUg6fPDea/find1 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 2 .svc.cluster.local myApp2 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=2 + local suffix=.svc.cluster.local + local database=myApp2 + local collection=test + run_mongos 'use myApp2\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + local 'command=use myApp2\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fEAXDOAcTg +++ mktemp ++ local LAST_ERR=/tmp/tmp.yWID1b1gYV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.fEAXDOAcTg ++ cat /tmp/tmp.yWID1b1gYV ++ rm /tmp/tmp.fEAXDOAcTg /tmp/tmp.yWID1b1gYV ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.Zt6CoOmxCC ++ mktemp + local LAST_ERR=/tmp/tmp.6iq6O10XF7 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Zt6CoOmxCC + cat /tmp/tmp.6iq6O10XF7 + rm /tmp/tmp.Zt6CoOmxCC /tmp/tmp.6iq6O10XF7 + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find2.json /tmp/tmp.bLUg6fPDea/find2 + desc 'check backup and restore -- minio' + set +o xtrace ----------------------------------------------------------------------------------- check backup and restore -- minio ----------------------------------------------------------------------------------- ++ get_backup_dest backup-minio ++ local backup_name=backup-minio ++ kubectl_bin get psmdb-backup backup-minio -o 'jsonpath={.status.destination}' ++ sed -e 's/.json$//' ++ sed 's|s3://||' ++ sed 's|azure://||' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gyI4e3ep1f +++ mktemp ++ local LAST_ERR=/tmp/tmp.RQ2aXwOrAB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb-backup backup-minio -o 'jsonpath={.status.destination}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gyI4e3ep1f ++ cat /tmp/tmp.RQ2aXwOrAB ++ rm /tmp/tmp.gyI4e3ep1f /tmp/tmp.RQ2aXwOrAB ++ return 0 + backup_dest_minio=operator-testing/2024-07-07T17:39:57Z + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 ls s3://operator-testing/2024-07-07T17:39:57Z/rs0/ + grep myApp.test.gz ++ mktemp + local LAST_OUT=/tmp/tmp.YtU9N26UlD ++ mktemp + local LAST_ERR=/tmp/tmp.OHDUJFeE0z + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 ls s3://operator-testing/2024-07-07T17:39:57Z/rs0/ + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.YtU9N26UlD + cat /tmp/tmp.OHDUJFeE0z If you don't see a command prompt, try pressing enter. warning: couldn't attach to pod/aws-cli, falling back to streaming logs: unable to upgrade connection: container aws-cli not found in pod aws-cli_demand-backup-sharded-7957 + rm /tmp/tmp.YtU9N26UlD /tmp/tmp.OHDUJFeE0z + return 0 2024-07-07 17:40:01 55 myApp.test.gz + insert_data_mongos 100501 myApp + local data=100501 + local db_name=myApp + local flags= + run_mongos 'use myApp\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zYm5Jj7C8N +++ mktemp ++ local LAST_ERR=/tmp/tmp.rxFkRADk36 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.zYm5Jj7C8N ++ cat /tmp/tmp.rxFkRADk36 ++ rm /tmp/tmp.zYm5Jj7C8N /tmp/tmp.rxFkRADk36 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.5hGvMl99Rp ++ mktemp + local LAST_ERR=/tmp/tmp.MWRwIwnN7l + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5hGvMl99Rp Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("952bee4a-63dd-4380-8967-c49c50795912") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.MWRwIwnN7l + rm /tmp/tmp.5hGvMl99Rp /tmp/tmp.MWRwIwnN7l + return 0 + insert_data_mongos 100501 myApp1 + local data=100501 + local db_name=myApp1 + local flags= + run_mongos 'use myApp1\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp1\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9Gn5lmZMAk +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ywg5Uki3ZC ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.9Gn5lmZMAk ++ cat /tmp/tmp.Ywg5Uki3ZC ++ rm /tmp/tmp.9Gn5lmZMAk /tmp/tmp.Ywg5Uki3ZC ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.2iiq0qe1N3 ++ mktemp + local LAST_ERR=/tmp/tmp.F5mEXUziEU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2iiq0qe1N3 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("b3e1675e-61b5-45ce-b91c-024ac57adf43") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp1 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.F5mEXUziEU + rm /tmp/tmp.2iiq0qe1N3 /tmp/tmp.F5mEXUziEU + return 0 + insert_data_mongos 100501 myApp2 + local data=100501 + local db_name=myApp2 + local flags= + run_mongos 'use myApp2\n db.test.insert({ x: 100501 })' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' '' '' + local 'command=use myApp2\n db.test.insert({ x: 100501 })' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eYHRxhivNU +++ mktemp ++ local LAST_ERR=/tmp/tmp.hlLrws5CK5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.eYHRxhivNU ++ cat /tmp/tmp.hlLrws5CK5 ++ rm /tmp/tmp.eYHRxhivNU /tmp/tmp.hlLrws5CK5 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.0rolq24eEw ++ mktemp + local LAST_ERR=/tmp/tmp.jAU5Y8yPet + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.insert({ x: 100501 })\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0rolq24eEw Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.demand-backup-sharded-7957.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("13f16f85-4448-4559-9f5f-8e95fc8d92b2") } Percona Server for MongoDB server version: v7.0.11-6 WARNING: shell and server versions do not match switched to db myApp2 WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.jAU5Y8yPet + rm /tmp/tmp.0rolq24eEw /tmp/tmp.jAU5Y8yPet + return 0 + check_data -2nd + local postfix=-2nd ++ seq 0 2 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd + local suffix=.svc.cluster.local + local database=myApp + local collection=test + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.secYYdtAxg +++ mktemp ++ local LAST_ERR=/tmp/tmp.rxeyKX7se0 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.secYYdtAxg ++ cat /tmp/tmp.rxeyKX7se0 ++ rm /tmp/tmp.secYYdtAxg /tmp/tmp.rxeyKX7se0 ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.1DJu79bIbR ++ mktemp + local LAST_ERR=/tmp/tmp.abM4YrLzlB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1DJu79bIbR + cat /tmp/tmp.abM4YrLzlB + rm /tmp/tmp.1DJu79bIbR /tmp/tmp.abM4YrLzlB + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd.json /tmp/tmp.bLUg6fPDea/find-2nd + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd1 .svc.cluster.local myApp1 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd1 + local suffix=.svc.cluster.local + local database=myApp1 + local collection=test + run_mongos 'use myApp1\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp1\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3ieoZpalxo +++ mktemp ++ local LAST_ERR=/tmp/tmp.Si7JgpfYEO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.3ieoZpalxo ++ cat /tmp/tmp.Si7JgpfYEO ++ rm /tmp/tmp.3ieoZpalxo /tmp/tmp.Si7JgpfYEO ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.ypy1fQYm8Z ++ mktemp + local LAST_ERR=/tmp/tmp.DI3lYnxuLh + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ypy1fQYm8Z + cat /tmp/tmp.DI3lYnxuLh + rm /tmp/tmp.ypy1fQYm8Z /tmp/tmp.DI3lYnxuLh + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd1.json /tmp/tmp.bLUg6fPDea/find-2nd1 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 -2nd2 .svc.cluster.local myApp2 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=-2nd2 + local suffix=.svc.cluster.local + local database=myApp2 + local collection=test + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + run_mongos 'use myApp2\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + local 'command=use myApp2\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l5UWWflH9P +++ mktemp ++ local LAST_ERR=/tmp/tmp.HH2PpZICzP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.l5UWWflH9P ++ cat /tmp/tmp.HH2PpZICzP ++ rm /tmp/tmp.l5UWWflH9P /tmp/tmp.HH2PpZICzP ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.20vjGt2TFN ++ mktemp + local LAST_ERR=/tmp/tmp.HKnxLuylqJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.20vjGt2TFN + cat /tmp/tmp.HKnxLuylqJ + rm /tmp/tmp.20vjGt2TFN /tmp/tmp.HKnxLuylqJ + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find-2nd2.json /tmp/tmp.bLUg6fPDea/find-2nd2 + run_restore backup-minio + local backup_name=backup-minio + /usr/bin/sed -e 's/name:/name: restore-backup-minio/' + /usr/bin/sed -e 's/backupName:/backupName: backup-minio/' + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/conf/restore.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.jhU48HtL84 ++ mktemp + local LAST_ERR=/tmp/tmp.h5TTAvU2GR + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jhU48HtL84 perconaservermongodbrestore.psmdb.percona.com/restore-backup-minio created + cat /tmp/tmp.h5TTAvU2GR + rm /tmp/tmp.jhU48HtL84 /tmp/tmp.h5TTAvU2GR + return 0 + wait_restore backup-minio some-name + local backup_name=backup-minio + local cluster_name=some-name + local target_state=ready + local wait_cluster_consistency=1 + local wait_time=780 + set +o xtrace waiting psmdb-restore/backup-minio to reach ready state........... + '[' 1 -eq 1 ']' + wait_cluster_consistency some-name + local cluster_name=some-name + local wait_time=32 + retry=0 + sleep 7 + echo -n 'waiting for cluster readyness' waiting for cluster readyness++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q94k4fwXVL +++ mktemp ++ local LAST_ERR=/tmp/tmp.0B8qZzD6h9 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.q94k4fwXVL ++ cat /tmp/tmp.0B8qZzD6h9 ++ rm /tmp/tmp.q94k4fwXVL /tmp/tmp.0B8qZzD6h9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 1 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2HNvi8Bhjy +++ mktemp ++ local LAST_ERR=/tmp/tmp.XFQW6wI75N ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2HNvi8Bhjy ++ cat /tmp/tmp.XFQW6wI75N ++ rm /tmp/tmp.2HNvi8Bhjy /tmp/tmp.XFQW6wI75N ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 2 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2n2WzbtOX5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.oFCu5qGrxy ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.2n2WzbtOX5 ++ cat /tmp/tmp.oFCu5qGrxy ++ rm /tmp/tmp.2n2WzbtOX5 /tmp/tmp.oFCu5qGrxy ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 3 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ngv3r8Ly4Z +++ mktemp ++ local LAST_ERR=/tmp/tmp.B8hMymupC4 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ngv3r8Ly4Z ++ cat /tmp/tmp.B8hMymupC4 ++ rm /tmp/tmp.ngv3r8Ly4Z /tmp/tmp.B8hMymupC4 ++ return 0 + [[ error == \r\e\a\d\y ]] + let retry+=1 + '[' 4 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.T5Sn3Kky61 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bSDE1qqv5P ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.T5Sn3Kky61 ++ cat /tmp/tmp.bSDE1qqv5P ++ rm /tmp/tmp.T5Sn3Kky61 /tmp/tmp.bSDE1qqv5P ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 5 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jsCTLBHu63 +++ mktemp ++ local LAST_ERR=/tmp/tmp.8A1KgaQks3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.jsCTLBHu63 ++ cat /tmp/tmp.8A1KgaQks3 ++ rm /tmp/tmp.jsCTLBHu63 /tmp/tmp.8A1KgaQks3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 6 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pxsaE2AHUo +++ mktemp ++ local LAST_ERR=/tmp/tmp.9rA8nSp4My ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.pxsaE2AHUo ++ cat /tmp/tmp.9rA8nSp4My ++ rm /tmp/tmp.pxsaE2AHUo /tmp/tmp.9rA8nSp4My ++ return 0 + [[ initializing == \r\e\a\d\y ]] + let retry+=1 + '[' 7 -ge 32 ']' + echo -n . .+ sleep 10 ++ kubectl_bin get psmdb some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8wVPCArHFb +++ mktemp ++ local LAST_ERR=/tmp/tmp.gAfenil7es ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get psmdb some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.8wVPCArHFb ++ cat /tmp/tmp.gAfenil7es ++ rm /tmp/tmp.8wVPCArHFb /tmp/tmp.gAfenil7es ++ return 0 + [[ ready == \r\e\a\d\y ]] + check_data + local postfix= ++ seq 0 2 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 '' .svc.cluster.local myApp test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix= + local suffix=.svc.cluster.local + local database=myApp + local collection=test + run_mongos 'use myApp\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp\n db.test.find()' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KCgWtU9tGQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.9CI2aBMyXb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KCgWtU9tGQ ++ cat /tmp/tmp.9CI2aBMyXb ++ rm /tmp/tmp.KCgWtU9tGQ /tmp/tmp.9CI2aBMyXb ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.DDPnzbztFQ ++ mktemp + local LAST_ERR=/tmp/tmp.hKmXZWgxVK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DDPnzbztFQ + cat /tmp/tmp.hKmXZWgxVK + rm /tmp/tmp.DDPnzbztFQ /tmp/tmp.hKmXZWgxVK + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find.json /tmp/tmp.bLUg6fPDea/find + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 1 .svc.cluster.local myApp1 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=1 + local suffix=.svc.cluster.local + local database=myApp1 + local collection=test + run_mongos 'use myApp1\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp1\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ybHzo9ACJf +++ mktemp ++ local LAST_ERR=/tmp/tmp.lAslFezods ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ybHzo9ACJf ++ cat /tmp/tmp.lAslFezods ++ rm /tmp/tmp.ybHzo9ACJf /tmp/tmp.lAslFezods ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.pPe7rvyHEn ++ mktemp + local LAST_ERR=/tmp/tmp.kwsJZHkx1G + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp1\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.pPe7rvyHEn + cat /tmp/tmp.kwsJZHkx1G + rm /tmp/tmp.pPe7rvyHEn /tmp/tmp.kwsJZHkx1G + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find1.json /tmp/tmp.bLUg6fPDea/find1 + for i in '$(seq 0 2)' + compare_mongos_cmd find myApp:myPass@some-name-mongos.demand-backup-sharded-7957 2 .svc.cluster.local myApp2 test + local command=find + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local postfix=2 + local suffix=.svc.cluster.local + local database=myApp2 + local collection=test + run_mongos 'use myApp2\n db.test.find()' myApp:myPass@some-name-mongos.demand-backup-sharded-7957 mongodb .svc.cluster.local + local 'command=use myApp2\n db.test.find()' + local uri=myApp:myPass@some-name-mongos.demand-backup-sharded-7957 + local driver=mongodb + local suffix=.svc.cluster.local + /usr/bin/sed -re 's/ObjectId\("[0-9a-f]+"\)//; s/-[0-9]+.svc/-xxx.svc/' + egrep -v 'I NETWORK|W NETWORK|Error saving history file|Percona Server for MongoDB|connecting to:|Unable to reach primary for set|Implicit session:|versions do not match|Error saving history file:' ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.izqdpevg0G +++ mktemp ++ local LAST_ERR=/tmp/tmp.Sftbqo3UIF ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.izqdpevg0G ++ cat /tmp/tmp.Sftbqo3UIF ++ rm /tmp/tmp.izqdpevg0G /tmp/tmp.Sftbqo3UIF ++ return 0 + local client_container=psmdb-client-7469665986-fbs6j + local mongo_flag= + kubectl_bin exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' ++ mktemp + local LAST_OUT=/tmp/tmp.wXkFkfBNkH ++ mktemp + local LAST_ERR=/tmp/tmp.rbzqJ0J3zH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec psmdb-client-7469665986-fbs6j -- bash -c 'printf '\''use myApp2\n db.test.find()\n'\'' | mongo mongodb://myApp:myPass@some-name-mongos.demand-backup-sharded-7957.svc.cluster.local/admin ' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wXkFkfBNkH + cat /tmp/tmp.rbzqJ0J3zH + rm /tmp/tmp.wXkFkfBNkH /tmp/tmp.rbzqJ0J3zH + return 0 + diff /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/demand-backup-sharded/compare/find2.json /tmp/tmp.bLUg6fPDea/find2 + desc 'delete backup and check if it is removed from bucket -- minio' + set +o xtrace ----------------------------------------------------------------------------------- delete backup and check if it is removed from bucket -- minio ----------------------------------------------------------------------------------- + kubectl_bin delete psmdb-backup --all ++ mktemp + local LAST_OUT=/tmp/tmp.CsewnuAfzV ++ mktemp + local LAST_ERR=/tmp/tmp.1shLwpti9z + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete psmdb-backup --all + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.CsewnuAfzV perconaservermongodbbackup.psmdb.percona.com "backup-aws-s3" deleted perconaservermongodbbackup.psmdb.percona.com "backup-azure-blob" deleted perconaservermongodbbackup.psmdb.percona.com "backup-gcp-cs" deleted perconaservermongodbbackup.psmdb.percona.com "backup-minio" deleted + cat /tmp/tmp.1shLwpti9z + rm /tmp/tmp.CsewnuAfzV /tmp/tmp.1shLwpti9z + return 0 ++ kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 ls s3://operator-testing/ ++ grep -c operator-testing/2024-07-07T17:39:57Z_ ++ cat +++ mktemp ++ local LAST_OUT=/tmp/tmp.mraTzWkG0m +++ mktemp ++ local LAST_ERR=/tmp/tmp.opJapPu3u7 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- /usr/bin/env AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 ls s3://operator-testing/ ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mraTzWkG0m ++ cat /tmp/tmp.opJapPu3u7 If you don't see a command prompt, try pressing enter. warning: couldn't attach to pod/aws-cli, falling back to streaming logs: unable to upgrade connection: container aws-cli not found in pod aws-cli_demand-backup-sharded-7957 ++ rm /tmp/tmp.mraTzWkG0m /tmp/tmp.opJapPu3u7 ++ return 0 + backup_exists=0 + [[ 0 -eq 1 ]] + '[' -z '' ']' + check_backup_deletion https://s3.amazonaws.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:40:35Z aws-s3 + path=https://s3.amazonaws.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:40:35Z + storage_name=aws-s3 + retry=0 ++ curl -sw '%{http_code}' -o /dev/null https://s3.amazonaws.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:40:35Z + [[ 403 -eq 403 ]] + check_backup_deletion https://storage.googleapis.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:17Z gcp-cs + path=https://storage.googleapis.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:17Z + storage_name=gcp-cs + retry=0 ++ curl -sw '%{http_code}' -o /dev/null https://storage.googleapis.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:17Z + [[ 404 -eq 403 ]] ++ curl -sw '%{http_code}' -o /dev/null https://storage.googleapis.com/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:17Z + [[ 404 -eq 404 ]] + check_backup_deletion https://engk8soperators.blob.core.windows.net/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:49Z azure-blob + path=https://engk8soperators.blob.core.windows.net/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:49Z + storage_name=azure-blob + retry=0 ++ curl -sw '%{http_code}' -o /dev/null https://engk8soperators.blob.core.windows.net/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:49Z + [[ 404 -eq 403 ]] ++ curl -sw '%{http_code}' -o /dev/null https://engk8soperators.blob.core.windows.net/operator-testing/psmdb-demand-backup-sharded/2024-07-07T17:41:49Z + [[ 404 -eq 404 ]] + desc 'check for passwords leak' + set +o xtrace ----------------------------------------------------------------------------------- check for passwords leak ----------------------------------------------------------------------------------- + check_passwords_leak + local secrets + local passwords + local pods ++ kubectl_bin get secrets -o json +++ mktemp ++ local LAST_OUT=/tmp/tmp.MBG6Lox0Xd +++ mktemp ++ local LAST_ERR=/tmp/tmp.razELQ6fGV ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets -o json ++ jq -r '.items[].data | to_entries | .[] | select(.key | (contains("_PASSWORD"))) | .value' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.MBG6Lox0Xd ++ cat /tmp/tmp.razELQ6fGV ++ rm /tmp/tmp.MBG6Lox0Xd /tmp/tmp.razELQ6fGV ++ return 0 + secrets='YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2' + echo secrets=YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 secrets=YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 ++ for i in '$secrets' ++ base64 -d ++ echo ++ for i in '$secrets' ++ base64 -d ++ echo ++ for i in '$secrets' ++ base64 -d ++ echo ++ for i in '$secrets' ++ base64 -d ++ echo ++ for i in '$secrets' ++ base64 -d ++ echo ++ for i in '$secrets' ++ base64 -d ++ echo ++ for i in '$secrets' ++ base64 -d ++ echo ++ for i in '$secrets' ++ base64 -d ++ echo ++ for i in '$secrets' ++ base64 -d ++ echo ++ for i in '$secrets' ++ base64 -d ++ echo + passwords='backup123456 clusterAdmin123456 clusterMonitor123456 databaseAdmin123456 userAdmin123456 backup123456 clusterAdmin123456 clusterMonitor123456 databaseAdmin123456 userAdmin123456 YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2' + echo passwords=backup123456 clusterAdmin123456 clusterMonitor123456 databaseAdmin123456 userAdmin123456 backup123456 clusterAdmin123456 clusterMonitor123456 databaseAdmin123456 userAdmin123456 YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 passwords=backup123456 clusterAdmin123456 clusterMonitor123456 databaseAdmin123456 userAdmin123456 backup123456 clusterAdmin123456 clusterMonitor123456 databaseAdmin123456 userAdmin123456 YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 YmFja3VwMTIzNDU2 Y2x1c3RlckFkbWluMTIzNDU2 Y2x1c3Rlck1vbml0b3IxMjM0NTY= ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== dXNlckFkbWluMTIzNDU2 ++ kubectl_bin get pods -o name ++ awk -F / '{print $2}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TwxRJdcuEc +++ mktemp ++ local LAST_ERR=/tmp/tmp.m40RMqoxr3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TwxRJdcuEc ++ cat /tmp/tmp.m40RMqoxr3 ++ rm /tmp/tmp.TwxRJdcuEc /tmp/tmp.m40RMqoxr3 ++ return 0 + pods='minio-service-57dd49b-9br22 psmdb-client-7469665986-fbs6j some-name-cfg-0 some-name-cfg-1 some-name-cfg-2 some-name-mongos-0 some-name-mongos-1 some-name-mongos-2 some-name-rs0-0 some-name-rs0-1 some-name-rs0-2 some-name-rs1-0 some-name-rs1-1 some-name-rs1-2 some-name-rs2-0 some-name-rs2-1 some-name-rs2-2' + echo pods=minio-service-57dd49b-9br22 psmdb-client-7469665986-fbs6j some-name-cfg-0 some-name-cfg-1 some-name-cfg-2 some-name-mongos-0 some-name-mongos-1 some-name-mongos-2 some-name-rs0-0 some-name-rs0-1 some-name-rs0-2 some-name-rs1-0 some-name-rs1-1 some-name-rs1-2 some-name-rs2-0 some-name-rs2-1 some-name-rs2-2 pods=minio-service-57dd49b-9br22 psmdb-client-7469665986-fbs6j some-name-cfg-0 some-name-cfg-1 some-name-cfg-2 some-name-mongos-0 some-name-mongos-1 some-name-mongos-2 some-name-rs0-0 some-name-rs0-1 some-name-rs0-2 some-name-rs1-0 some-name-rs1-1 some-name-rs1-2 some-name-rs2-0 some-name-rs2-1 some-name-rs2-2 + collect_logs demand-backup-sharded-7957 + local containers + local count + NS=demand-backup-sharded-7957 + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod minio-service-57dd49b-9br22 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o6jO70ybFB +++ mktemp ++ local LAST_ERR=/tmp/tmp.zSgHwYpPve ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod minio-service-57dd49b-9br22 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.o6jO70ybFB ++ cat /tmp/tmp.zSgHwYpPve ++ rm /tmp/tmp.o6jO70ybFB /tmp/tmp.zSgHwYpPve ++ return 0 + containers=minio + for c in '$containers' + [[ minio =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs minio-service-57dd49b-9br22 -c minio ++ mktemp + local LAST_OUT=/tmp/tmp.dWG5gbtpbG ++ mktemp + local LAST_ERR=/tmp/tmp.ZAhlkxyryb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs minio-service-57dd49b-9br22 -c minio + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.dWG5gbtpbG + cat /tmp/tmp.ZAhlkxyryb + rm /tmp/tmp.dWG5gbtpbG /tmp/tmp.ZAhlkxyryb + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-minio-service-57dd49b-9br22-minio.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod psmdb-client-7469665986-fbs6j -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iRrR21ZkMR +++ mktemp ++ local LAST_ERR=/tmp/tmp.joqNXnFp8n ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod psmdb-client-7469665986-fbs6j -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.iRrR21ZkMR ++ cat /tmp/tmp.joqNXnFp8n ++ rm /tmp/tmp.iRrR21ZkMR /tmp/tmp.joqNXnFp8n ++ return 0 + containers=psmdb-client + for c in '$containers' + [[ psmdb-client =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs psmdb-client-7469665986-fbs6j -c psmdb-client ++ mktemp + local LAST_OUT=/tmp/tmp.vEMmusUtus ++ mktemp + local LAST_ERR=/tmp/tmp.EGzELh4qAZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs psmdb-client-7469665986-fbs6j -c psmdb-client + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.vEMmusUtus + cat /tmp/tmp.EGzELh4qAZ + rm /tmp/tmp.vEMmusUtus /tmp/tmp.EGzELh4qAZ + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-psmdb-client-7469665986-fbs6j-psmdb-client.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-cfg-0 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gQcVHM0nVj +++ mktemp ++ local LAST_ERR=/tmp/tmp.mZNsZeI9RD ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-cfg-0 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gQcVHM0nVj ++ cat /tmp/tmp.mZNsZeI9RD ++ rm /tmp/tmp.gQcVHM0nVj /tmp/tmp.mZNsZeI9RD ++ return 0 + containers='mongod cfg-sidecar-1 backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-cfg-0 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.zUop8gSa25 ++ mktemp + local LAST_ERR=/tmp/tmp.jZ5Yo03n1b + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-cfg-0 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.zUop8gSa25 + cat /tmp/tmp.jZ5Yo03n1b + rm /tmp/tmp.zUop8gSa25 /tmp/tmp.jZ5Yo03n1b + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ cfg-sidecar-1 =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-cfg-0 -c cfg-sidecar-1 ++ mktemp + local LAST_OUT=/tmp/tmp.XPyPHS4RQU ++ mktemp + local LAST_ERR=/tmp/tmp.tL5VlatxLc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-cfg-0 -c cfg-sidecar-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.XPyPHS4RQU + cat /tmp/tmp.tL5VlatxLc + rm /tmp/tmp.XPyPHS4RQU /tmp/tmp.tL5VlatxLc + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-cfg-0 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.iCGJKyCy9F ++ mktemp + local LAST_ERR=/tmp/tmp.TkDTbpLUzQ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-cfg-0 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.iCGJKyCy9F + cat /tmp/tmp.TkDTbpLUzQ + rm /tmp/tmp.iCGJKyCy9F /tmp/tmp.TkDTbpLUzQ + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-cfg-1 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qAuEdU3z3i +++ mktemp ++ local LAST_ERR=/tmp/tmp.sVDr6enIGj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-cfg-1 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.qAuEdU3z3i ++ cat /tmp/tmp.sVDr6enIGj ++ rm /tmp/tmp.qAuEdU3z3i /tmp/tmp.sVDr6enIGj ++ return 0 + containers='mongod cfg-sidecar-1 backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-cfg-1 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.6kzhZxlp8p ++ mktemp + local LAST_ERR=/tmp/tmp.UfRoFufNVA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-cfg-1 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6kzhZxlp8p + cat /tmp/tmp.UfRoFufNVA + rm /tmp/tmp.6kzhZxlp8p /tmp/tmp.UfRoFufNVA + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ cfg-sidecar-1 =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-cfg-1 -c cfg-sidecar-1 ++ mktemp + local LAST_OUT=/tmp/tmp.WtDYF2NEzQ ++ mktemp + local LAST_ERR=/tmp/tmp.B8o0OvQCk6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-cfg-1 -c cfg-sidecar-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.WtDYF2NEzQ + cat /tmp/tmp.B8o0OvQCk6 + rm /tmp/tmp.WtDYF2NEzQ /tmp/tmp.B8o0OvQCk6 + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-cfg-1 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.R9mRCfVG9T ++ mktemp + local LAST_ERR=/tmp/tmp.caAkqfeWqj + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-cfg-1 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.R9mRCfVG9T + cat /tmp/tmp.caAkqfeWqj + rm /tmp/tmp.R9mRCfVG9T /tmp/tmp.caAkqfeWqj + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-cfg-2 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DyRBjhnPKx +++ mktemp ++ local LAST_ERR=/tmp/tmp.pxybh6bKJI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-cfg-2 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DyRBjhnPKx ++ cat /tmp/tmp.pxybh6bKJI ++ rm /tmp/tmp.DyRBjhnPKx /tmp/tmp.pxybh6bKJI ++ return 0 + containers='mongod cfg-sidecar-1 backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-cfg-2 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.rTiPMCOSYh ++ mktemp + local LAST_ERR=/tmp/tmp.QqQ0FdYwEb + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-cfg-2 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rTiPMCOSYh + cat /tmp/tmp.QqQ0FdYwEb + rm /tmp/tmp.rTiPMCOSYh /tmp/tmp.QqQ0FdYwEb + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ cfg-sidecar-1 =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-cfg-2 -c cfg-sidecar-1 ++ mktemp + local LAST_OUT=/tmp/tmp.4zrLP6DzlK ++ mktemp + local LAST_ERR=/tmp/tmp.vChVrxIvjU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-cfg-2 -c cfg-sidecar-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.4zrLP6DzlK + cat /tmp/tmp.vChVrxIvjU + rm /tmp/tmp.4zrLP6DzlK /tmp/tmp.vChVrxIvjU + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-cfg-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-cfg-2 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.abolW3XWuU ++ mktemp + local LAST_ERR=/tmp/tmp.C6KrWY9QDo + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-cfg-2 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.abolW3XWuU + cat /tmp/tmp.C6KrWY9QDo + rm /tmp/tmp.abolW3XWuU /tmp/tmp.C6KrWY9QDo + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-cfg-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-mongos-0 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BuXRzeHj19 +++ mktemp ++ local LAST_ERR=/tmp/tmp.snGIb6T0eT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-mongos-0 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.BuXRzeHj19 ++ cat /tmp/tmp.snGIb6T0eT ++ rm /tmp/tmp.BuXRzeHj19 /tmp/tmp.snGIb6T0eT ++ return 0 + containers='mongos mongos-sidecar-1' + for c in '$containers' + [[ mongos =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-mongos-0 -c mongos ++ mktemp + local LAST_OUT=/tmp/tmp.a2kiDB2s4l ++ mktemp + local LAST_ERR=/tmp/tmp.jV1GzrifI8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-mongos-0 -c mongos + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.a2kiDB2s4l + cat /tmp/tmp.jV1GzrifI8 + rm /tmp/tmp.a2kiDB2s4l /tmp/tmp.jV1GzrifI8 + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ mongos-sidecar-1 =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-mongos-0 -c mongos-sidecar-1 ++ mktemp + local LAST_OUT=/tmp/tmp.Y0qR9GQ4ji ++ mktemp + local LAST_ERR=/tmp/tmp.XzbUVOQhNf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-mongos-0 -c mongos-sidecar-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Y0qR9GQ4ji + cat /tmp/tmp.XzbUVOQhNf + rm /tmp/tmp.Y0qR9GQ4ji /tmp/tmp.XzbUVOQhNf + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-0-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-mongos-1 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lxq6o8slDE +++ mktemp ++ local LAST_ERR=/tmp/tmp.5YrWCXMSLI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-mongos-1 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.lxq6o8slDE ++ cat /tmp/tmp.5YrWCXMSLI ++ rm /tmp/tmp.lxq6o8slDE /tmp/tmp.5YrWCXMSLI ++ return 0 + containers='mongos mongos-sidecar-1' + for c in '$containers' + [[ mongos =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-mongos-1 -c mongos ++ mktemp + local LAST_OUT=/tmp/tmp.0cE4g7sBS0 ++ mktemp + local LAST_ERR=/tmp/tmp.03UBV1VcfS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-mongos-1 -c mongos + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.0cE4g7sBS0 + cat /tmp/tmp.03UBV1VcfS + rm /tmp/tmp.0cE4g7sBS0 /tmp/tmp.03UBV1VcfS + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ mongos-sidecar-1 =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-mongos-1 -c mongos-sidecar-1 ++ mktemp + local LAST_OUT=/tmp/tmp.1oPw0KmhY6 ++ mktemp + local LAST_ERR=/tmp/tmp.fDEyoYA12m + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-mongos-1 -c mongos-sidecar-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1oPw0KmhY6 + cat /tmp/tmp.fDEyoYA12m + rm /tmp/tmp.1oPw0KmhY6 /tmp/tmp.fDEyoYA12m + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-1-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-mongos-2 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4cQM5J6gNW +++ mktemp ++ local LAST_ERR=/tmp/tmp.MV6wjqCES3 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-mongos-2 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.4cQM5J6gNW ++ cat /tmp/tmp.MV6wjqCES3 ++ rm /tmp/tmp.4cQM5J6gNW /tmp/tmp.MV6wjqCES3 ++ return 0 + containers='mongos mongos-sidecar-1' + for c in '$containers' + [[ mongos =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-mongos-2 -c mongos ++ mktemp + local LAST_OUT=/tmp/tmp.H0P2lvjUEt ++ mktemp + local LAST_ERR=/tmp/tmp.hTs6LCumsI + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-mongos-2 -c mongos + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.H0P2lvjUEt + cat /tmp/tmp.hTs6LCumsI + rm /tmp/tmp.H0P2lvjUEt /tmp/tmp.hTs6LCumsI + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ mongos-sidecar-1 =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-mongos-2 -c mongos-sidecar-1 ++ mktemp + local LAST_OUT=/tmp/tmp.HJUNFLvLo7 ++ mktemp + local LAST_ERR=/tmp/tmp.0t8V1mWYcw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-mongos-2 -c mongos-sidecar-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.HJUNFLvLo7 + cat /tmp/tmp.0t8V1mWYcw + rm /tmp/tmp.HJUNFLvLo7 /tmp/tmp.0t8V1mWYcw + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-mongos-2-mongos-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-rs0-0 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZsJBsBRfbl +++ mktemp ++ local LAST_ERR=/tmp/tmp.FgVAx29wOE ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-rs0-0 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZsJBsBRfbl ++ cat /tmp/tmp.FgVAx29wOE ++ rm /tmp/tmp.ZsJBsBRfbl /tmp/tmp.FgVAx29wOE ++ return 0 + containers='mongod backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs0-0 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.C8LVAgszfH ++ mktemp + local LAST_ERR=/tmp/tmp.YTHVLUh864 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs0-0 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.C8LVAgszfH + cat /tmp/tmp.YTHVLUh864 + rm /tmp/tmp.C8LVAgszfH /tmp/tmp.YTHVLUh864 + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs0-0 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.NwLvGqFf1E ++ mktemp + local LAST_ERR=/tmp/tmp.KonexeEBSf + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs0-0 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.NwLvGqFf1E + cat /tmp/tmp.KonexeEBSf + rm /tmp/tmp.NwLvGqFf1E /tmp/tmp.KonexeEBSf + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-rs0-1 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VRBTBQMPwq +++ mktemp ++ local LAST_ERR=/tmp/tmp.MMjvLsRXgP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-rs0-1 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.VRBTBQMPwq ++ cat /tmp/tmp.MMjvLsRXgP ++ rm /tmp/tmp.VRBTBQMPwq /tmp/tmp.MMjvLsRXgP ++ return 0 + containers='mongod backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs0-1 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.7JaFKu3YiF ++ mktemp + local LAST_ERR=/tmp/tmp.lWift3gKB3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs0-1 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7JaFKu3YiF + cat /tmp/tmp.lWift3gKB3 + rm /tmp/tmp.7JaFKu3YiF /tmp/tmp.lWift3gKB3 + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs0-1 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.AWYZsn8xkY ++ mktemp + local LAST_ERR=/tmp/tmp.NB59ckUI0H + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs0-1 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.AWYZsn8xkY + cat /tmp/tmp.NB59ckUI0H + rm /tmp/tmp.AWYZsn8xkY /tmp/tmp.NB59ckUI0H + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-rs0-2 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ScDdEOV8q1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.gwNepbmQSN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-rs0-2 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ScDdEOV8q1 ++ cat /tmp/tmp.gwNepbmQSN ++ rm /tmp/tmp.ScDdEOV8q1 /tmp/tmp.gwNepbmQSN ++ return 0 + containers='mongod backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs0-2 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.oIGZeyYG2V ++ mktemp + local LAST_ERR=/tmp/tmp.LfRpsFUXNL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs0-2 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.oIGZeyYG2V + cat /tmp/tmp.LfRpsFUXNL + rm /tmp/tmp.oIGZeyYG2V /tmp/tmp.LfRpsFUXNL + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs0-2 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.rsbFY6Q5sG ++ mktemp + local LAST_ERR=/tmp/tmp.oyWyQhSkMZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs0-2 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.rsbFY6Q5sG + cat /tmp/tmp.oyWyQhSkMZ + rm /tmp/tmp.rsbFY6Q5sG /tmp/tmp.oyWyQhSkMZ + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs0-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-rs1-0 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5dcESWp371 +++ mktemp ++ local LAST_ERR=/tmp/tmp.jttvjZn0Ym ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-rs1-0 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5dcESWp371 ++ cat /tmp/tmp.jttvjZn0Ym ++ rm /tmp/tmp.5dcESWp371 /tmp/tmp.jttvjZn0Ym ++ return 0 + containers='mongod backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs1-0 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.TOfi3byOSV ++ mktemp + local LAST_ERR=/tmp/tmp.AvHmmpFgOL + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs1-0 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.TOfi3byOSV + cat /tmp/tmp.AvHmmpFgOL + rm /tmp/tmp.TOfi3byOSV /tmp/tmp.AvHmmpFgOL + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs1-0 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.60yeOMbSGj ++ mktemp + local LAST_ERR=/tmp/tmp.EqYEtx2fNs + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs1-0 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.60yeOMbSGj + cat /tmp/tmp.EqYEtx2fNs + rm /tmp/tmp.60yeOMbSGj /tmp/tmp.EqYEtx2fNs + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-rs1-1 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mmD4IYMwsb +++ mktemp ++ local LAST_ERR=/tmp/tmp.kvIzcefb0u ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-rs1-1 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.mmD4IYMwsb ++ cat /tmp/tmp.kvIzcefb0u ++ rm /tmp/tmp.mmD4IYMwsb /tmp/tmp.kvIzcefb0u ++ return 0 + containers='mongod backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs1-1 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.JPSMPhJ0By ++ mktemp + local LAST_ERR=/tmp/tmp.4OEMa9qZXt + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs1-1 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.JPSMPhJ0By + cat /tmp/tmp.4OEMa9qZXt + rm /tmp/tmp.JPSMPhJ0By /tmp/tmp.4OEMa9qZXt + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs1-1 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.DFANOsn4Mr ++ mktemp + local LAST_ERR=/tmp/tmp.2LB2VfLVpW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs1-1 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DFANOsn4Mr + cat /tmp/tmp.2LB2VfLVpW + rm /tmp/tmp.DFANOsn4Mr /tmp/tmp.2LB2VfLVpW + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-rs1-2 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5u20O6kuIC +++ mktemp ++ local LAST_ERR=/tmp/tmp.24oN9sNlys ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-rs1-2 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5u20O6kuIC ++ cat /tmp/tmp.24oN9sNlys ++ rm /tmp/tmp.5u20O6kuIC /tmp/tmp.24oN9sNlys ++ return 0 + containers='mongod backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs1-2 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.S82WqNzHkI ++ mktemp + local LAST_ERR=/tmp/tmp.t9CjoONENy + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs1-2 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.S82WqNzHkI + cat /tmp/tmp.t9CjoONENy + rm /tmp/tmp.S82WqNzHkI /tmp/tmp.t9CjoONENy + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs1-2 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.94zxzWmHPK ++ mktemp + local LAST_ERR=/tmp/tmp.jXPtwGtPHS + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs1-2 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.94zxzWmHPK + cat /tmp/tmp.jXPtwGtPHS + rm /tmp/tmp.94zxzWmHPK /tmp/tmp.jXPtwGtPHS + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs1-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-rs2-0 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KJO95GQHyb +++ mktemp ++ local LAST_ERR=/tmp/tmp.0ZrI4oPrws ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-rs2-0 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.KJO95GQHyb ++ cat /tmp/tmp.0ZrI4oPrws ++ rm /tmp/tmp.KJO95GQHyb /tmp/tmp.0ZrI4oPrws ++ return 0 + containers='mongod rs-sidecar-1 backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs2-0 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.sBUOlyoRKT ++ mktemp + local LAST_ERR=/tmp/tmp.of2eqFeBCw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs2-0 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sBUOlyoRKT + cat /tmp/tmp.of2eqFeBCw + rm /tmp/tmp.sBUOlyoRKT /tmp/tmp.of2eqFeBCw + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ rs-sidecar-1 =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs2-0 -c rs-sidecar-1 ++ mktemp + local LAST_OUT=/tmp/tmp.Mnnp7HbfPh ++ mktemp + local LAST_ERR=/tmp/tmp.xVqBBJ12LG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs2-0 -c rs-sidecar-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Mnnp7HbfPh + cat /tmp/tmp.xVqBBJ12LG + rm /tmp/tmp.Mnnp7HbfPh /tmp/tmp.xVqBBJ12LG + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs2-0 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.y9eIO1nqcE ++ mktemp + local LAST_ERR=/tmp/tmp.ccHLFZNrvu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs2-0 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.y9eIO1nqcE + cat /tmp/tmp.ccHLFZNrvu + rm /tmp/tmp.y9eIO1nqcE /tmp/tmp.ccHLFZNrvu + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-0-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-rs2-1 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.j4e0TbjX15 +++ mktemp ++ local LAST_ERR=/tmp/tmp.FrhwROBgIO ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-rs2-1 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.j4e0TbjX15 ++ cat /tmp/tmp.FrhwROBgIO ++ rm /tmp/tmp.j4e0TbjX15 /tmp/tmp.FrhwROBgIO ++ return 0 + containers='mongod rs-sidecar-1 backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs2-1 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.TpXbPoOQz6 ++ mktemp + local LAST_ERR=/tmp/tmp.eoSAiUUl1C + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs2-1 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.TpXbPoOQz6 + cat /tmp/tmp.eoSAiUUl1C + rm /tmp/tmp.TpXbPoOQz6 /tmp/tmp.eoSAiUUl1C + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ rs-sidecar-1 =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs2-1 -c rs-sidecar-1 ++ mktemp + local LAST_OUT=/tmp/tmp.xg5bcqUqp5 ++ mktemp + local LAST_ERR=/tmp/tmp.89c1zSdfm6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs2-1 -c rs-sidecar-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.xg5bcqUqp5 + cat /tmp/tmp.89c1zSdfm6 + rm /tmp/tmp.xg5bcqUqp5 /tmp/tmp.89c1zSdfm6 + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs2-1 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.w7CBKasheu ++ mktemp + local LAST_ERR=/tmp/tmp.R4T8uWaptX + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs2-1 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.w7CBKasheu + cat /tmp/tmp.R4T8uWaptX + rm /tmp/tmp.w7CBKasheu /tmp/tmp.R4T8uWaptX + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-1-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + for p in '$pods' ++ kubectl_bin -n demand-backup-sharded-7957 get pod some-name-rs2-2 -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S66wqvu16L +++ mktemp ++ local LAST_ERR=/tmp/tmp.ieROs1qFWz ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n demand-backup-sharded-7957 get pod some-name-rs2-2 -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.S66wqvu16L ++ cat /tmp/tmp.ieROs1qFWz ++ rm /tmp/tmp.S66wqvu16L /tmp/tmp.ieROs1qFWz ++ return 0 + containers='mongod rs-sidecar-1 backup-agent' + for c in '$containers' + [[ mongod =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs2-2 -c mongod ++ mktemp + local LAST_OUT=/tmp/tmp.QFZzvW3LFI ++ mktemp + local LAST_ERR=/tmp/tmp.nFPUGKfb1I + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs2-2 -c mongod + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.QFZzvW3LFI + cat /tmp/tmp.nFPUGKfb1I + rm /tmp/tmp.QFZzvW3LFI /tmp/tmp.nFPUGKfb1I + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-mongod.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ rs-sidecar-1 =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs2-2 -c rs-sidecar-1 ++ mktemp + local LAST_OUT=/tmp/tmp.T6HRRjyFIA ++ mktemp + local LAST_ERR=/tmp/tmp.6S2ybpAUJ4 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs2-2 -c rs-sidecar-1 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.T6HRRjyFIA + cat /tmp/tmp.6S2ybpAUJ4 + rm /tmp/tmp.T6HRRjyFIA /tmp/tmp.6S2ybpAUJ4 + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-rs-sidecar-1.txt ++ : + count=0 + [[ 0 != 0 ]] + for c in '$containers' + [[ backup-agent =~ pmm ]] + kubectl_bin -n demand-backup-sharded-7957 logs some-name-rs2-2 -c backup-agent ++ mktemp + local LAST_OUT=/tmp/tmp.t57oU2nLq9 ++ mktemp + local LAST_ERR=/tmp/tmp.1W2QSCgmnP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n demand-backup-sharded-7957 logs some-name-rs2-2 -c backup-agent + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.t57oU2nLq9 + cat /tmp/tmp.1W2QSCgmnP + rm /tmp/tmp.t57oU2nLq9 /tmp/tmp.1W2QSCgmnP + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-some-name-rs2-2-backup-agent.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + '[' -n psmdb-operator ']' ++ kubectl_bin -n psmdb-operator get pods -o name ++ awk -F / '{print $2}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q8m9C21G1e +++ mktemp ++ local LAST_ERR=/tmp/tmp.iyBrfEKFQo ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n psmdb-operator get pods -o name ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.q8m9C21G1e ++ cat /tmp/tmp.iyBrfEKFQo ++ rm /tmp/tmp.q8m9C21G1e /tmp/tmp.iyBrfEKFQo ++ return 0 + pods=percona-server-mongodb-operator-7bb67465c5-95xbd + collect_logs psmdb-operator + local containers + local count + NS=psmdb-operator + for p in '$pods' ++ kubectl_bin -n psmdb-operator get pod percona-server-mongodb-operator-7bb67465c5-95xbd -o 'jsonpath={.spec.containers[*].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.O4Yo4ukzCm +++ mktemp ++ local LAST_ERR=/tmp/tmp.e1Na6XVKtI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl -n psmdb-operator get pod percona-server-mongodb-operator-7bb67465c5-95xbd -o 'jsonpath={.spec.containers[*].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.O4Yo4ukzCm ++ cat /tmp/tmp.e1Na6XVKtI ++ rm /tmp/tmp.O4Yo4ukzCm /tmp/tmp.e1Na6XVKtI ++ return 0 + containers=percona-server-mongodb-operator + for c in '$containers' + [[ percona-server-mongodb-operator =~ pmm ]] + kubectl_bin -n psmdb-operator logs percona-server-mongodb-operator-7bb67465c5-95xbd -c percona-server-mongodb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.t0TLpwwTlY ++ mktemp + local LAST_ERR=/tmp/tmp.XWZeP0erAe + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl -n psmdb-operator logs percona-server-mongodb-operator-7bb67465c5-95xbd -c percona-server-mongodb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.t0TLpwwTlY + cat /tmp/tmp.XWZeP0erAe + rm /tmp/tmp.t0TLpwwTlY /tmp/tmp.XWZeP0erAe + return 0 + echo logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt logs saved in: /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- backup123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- clusterMonitor123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- databaseAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- userAdmin123456 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- YmFja3VwMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3RlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- Y2x1c3Rlck1vbml0b3IxMjM0NTY= /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- ZGF0YWJhc2VBZG1pbjEyMzQ1Ng== /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + for pass in '$passwords' ++ grep -c --fixed-strings -- dXNlckFkbWluMTIzNDU2 /tmp/tmp.bLUg6fPDea/logs_output-percona-server-mongodb-operator-7bb67465c5-95xbd-percona-server-mongodb-operator.txt ++ : + count=0 + [[ 0 != 0 ]] + echo + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/container-rc.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.k9h5IrOve5 ++ mktemp + local LAST_ERR=/tmp/tmp.2l94zBGCDc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/e2e-tests/conf/container-rc.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.k9h5IrOve5 runtimeclass.node.k8s.io "container-rc" deleted + cat /tmp/tmp.2l94zBGCDc + rm /tmp/tmp.k9h5IrOve5 /tmp/tmp.2l94zBGCDc + return 0 + destroy demand-backup-sharded-7957 + local namespace=demand-backup-sharded-7957 + local ignore_logs=true + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.jrGV7rqkm1 ++ mktemp + local LAST_ERR=/tmp/tmp.ZC7poAt8Uy + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jrGV7rqkm1 customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.ZC7poAt8Uy + rm /tmp/tmp.jrGV7rqkm1 /tmp/tmp.ZC7poAt8Uy + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/crd.yaml ++ grep -v '\-\-\-' + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.jCVQfius2c ++ mktemp + local LAST_ERR=/tmp/tmp.uCY9HtZM01 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.jCVQfius2c + cat /tmp/tmp.uCY9HtZM01 + rm /tmp/tmp.jCVQfius2c /tmp/tmp.uCY9HtZM01 + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + grep -v NAMESPACE + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.cUYGTwEKod ++ mktemp + local LAST_ERR=/tmp/tmp.wi8mlKWPYt + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.cUYGTwEKod + cat /tmp/tmp.wi8mlKWPYt + rm /tmp/tmp.cUYGTwEKod /tmp/tmp.wi8mlKWPYt + return 0 + for crd_name in '$(yq eval '\''.metadata.name'\'' "${src_dir}/deploy/crd.yaml" | grep -v '\''\-\-\-'\'')' + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.3ykYVmEqzh ++ mktemp + local LAST_ERR=/tmp/tmp.71OW8BAnyn + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.3ykYVmEqzh + cat /tmp/tmp.71OW8BAnyn + rm /tmp/tmp.3ykYVmEqzh /tmp/tmp.71OW8BAnyn + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.71OujLhzdq ++ mktemp + local LAST_ERR=/tmp/tmp.OmhnHqooqN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1573/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.71OujLhzdq clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.OmhnHqooqN + rm /tmp/tmp.71OujLhzdq /tmp/tmp.OmhnHqooqN + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.X34ULBU1lj ++ mktemp + local LAST_ERR=/tmp/tmp.C0S15QN93H + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.X34ULBU1lj + cat /tmp/tmp.C0S15QN93H Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.X34ULBU1lj + cat /tmp/tmp.C0S15QN93H Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in '$(seq 0 2)' + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.X34ULBU1lj + cat /tmp/tmp.C0S15QN93H Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.X34ULBU1lj + cat /tmp/tmp.C0S15QN93H Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": namespaces "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.14.5/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.X34ULBU1lj /tmp/tmp.C0S15QN93H + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + rm -rf /tmp/tmp.bLUg6fPDea + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.9d5ZOm6Vz6 + kubectl_bin delete --grace-period=0 --force=true namespace demand-backup-sharded-7957 ++ mktemp + local LAST_ERR=/tmp/tmp.Wsl1RJLxMe + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.gKxbSP9oUX ++ mktemp + local LAST_ERR=/tmp/tmp.eWdQTDj0o8 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace demand-backup-sharded-7957