Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/logs/monitoring-pmm3.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 + create_infra monitoring-pmm3-31037 + local ns=monitoring-pmm3-31037 + [[ 1 == 1 ]] + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.skjNPXHOKr ++ mktemp + local LAST_ERR=/tmp/tmp.wA2yY1cqIK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.skjNPXHOKr customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.wA2yY1cqIK + rm /tmp/tmp.skjNPXHOKr /tmp/tmp.wA2yY1cqIK + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.ck0zN0rk8G ++ mktemp + local LAST_ERR=/tmp/tmp.fyjdOUZkd3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ck0zN0rk8G + cat /tmp/tmp.fyjdOUZkd3 + rm /tmp/tmp.ck0zN0rk8G /tmp/tmp.fyjdOUZkd3 + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.Odt2sLMHRW ++ mktemp + local LAST_ERR=/tmp/tmp.SpX9Hc5SEU + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Odt2sLMHRW + cat /tmp/tmp.SpX9Hc5SEU + rm /tmp/tmp.Odt2sLMHRW /tmp/tmp.SpX9Hc5SEU + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.HIbDLHfXnL ++ mktemp + local LAST_ERR=/tmp/tmp.i4bvo7Uliv + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.HIbDLHfXnL + cat /tmp/tmp.i4bvo7Uliv + rm /tmp/tmp.HIbDLHfXnL /tmp/tmp.i4bvo7Uliv + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.sWoRQUykQz ++ mktemp + local LAST_ERR=/tmp/tmp.o2syvOpZ6S + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.sWoRQUykQz clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.o2syvOpZ6S + rm /tmp/tmp.sWoRQUykQz /tmp/tmp.o2syvOpZ6S + return 0 + check_crd_for_deletion PR-2256-98b7d85f + local git_tag=PR-2256-98b7d85f ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/PR-2256-98b7d85f/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/sbin/sed s/---//g ++ /usr/sbin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in $(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '.metadata.name' | $sed 's/---//g' | $sed ':a;N;$!ba;s/\n/ /g') ++ kubectl_bin get crd/null -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5c5beDeAfM +++ mktemp ++ local LAST_ERR=/tmp/tmp.mVcKcV6iTH ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.5c5beDeAfM ++ cat /tmp/tmp.mVcKcV6iTH Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.5c5beDeAfM ++ cat /tmp/tmp.mVcKcV6iTH Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get crd/null -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.5c5beDeAfM ++ cat /tmp/tmp.mVcKcV6iTH Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ sleep 8 ++ cat /tmp/tmp.5c5beDeAfM ++ cat /tmp/tmp.mVcKcV6iTH Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ++ rm /tmp/tmp.5c5beDeAfM /tmp/tmp.mVcKcV6iTH ++ return 1 + [[ '' == Terminating ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get clusterrole ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + awk '{print$1}' ++ mktemp + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + local LAST_OUT=/tmp/tmp.ur95FL9FPY ++ mktemp + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found + xargs kubectl delete ns + local LAST_ERR=/tmp/tmp.nlFIywnsJc + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ mktemp + for i in $(seq 0 2) + set +e + kubectl get ns + local LAST_OUT=/tmp/tmp.v5zasYgp9z ++ mktemp + local LAST_ERR=/tmp/tmp.6mN6Oswm4N + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ur95FL9FPY + cat /tmp/tmp.nlFIywnsJc + rm /tmp/tmp.ur95FL9FPY /tmp/tmp.nlFIywnsJc + return 0 namespace "cert-manager" deleted namespace "monitoring-pmm3-9753" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.v5zasYgp9z namespace "psmdb-operator" deleted + cat /tmp/tmp.6mN6Oswm4N + rm /tmp/tmp.v5zasYgp9z /tmp/tmp.6mN6Oswm4N + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.7i4JjpKdiz ++ mktemp + local LAST_ERR=/tmp/tmp.Hq9eRjJAuI + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.7i4JjpKdiz + cat /tmp/tmp.Hq9eRjJAuI + rm /tmp/tmp.7i4JjpKdiz /tmp/tmp.Hq9eRjJAuI + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.etiMwlq43d ++ mktemp + local LAST_ERR=/tmp/tmp.eqlstlLUxP + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.etiMwlq43d namespace/psmdb-operator created + cat /tmp/tmp.eqlstlLUxP + rm /tmp/tmp.etiMwlq43d /tmp/tmp.eqlstlLUxP + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.6Yk6VrrX82 +++ mktemp ++ local LAST_ERR=/tmp/tmp.J3paO0Pnm5 ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.6Yk6VrrX82 ++ cat /tmp/tmp.J3paO0Pnm5 ++ rm /tmp/tmp.6Yk6VrrX82 /tmp/tmp.J3paO0Pnm5 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2256-98b7d85f-7-cluster7 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.DQaFGQnmMa ++ mktemp + local LAST_ERR=/tmp/tmp.YR7nbTNndM + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2256-98b7d85f-7-cluster7 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DQaFGQnmMa Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2256-98b7d85f-7-cluster7" modified. + cat /tmp/tmp.YR7nbTNndM + rm /tmp/tmp.DQaFGQnmMa /tmp/tmp.YR7nbTNndM + return 0 + deploy_operator + desc 'start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2256-98b7d85f' + set +o xtrace ----------------------------------------------------------------------------------- start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2256-98b7d85f ----------------------------------------------------------------------------------- + local cr_file + '[' -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/conf/crd.yaml ']' + cr_file=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/crd.yaml + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.xYoIKdaBFP ++ mktemp + local LAST_ERR=/tmp/tmp.KlRONVGclY + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.xYoIKdaBFP customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.KlRONVGclY + rm /tmp/tmp.xYoIKdaBFP /tmp/tmp.KlRONVGclY + return 0 + '[' -n psmdb-operator ']' + apply_rbac cw-rbac + local operator_namespace=psmdb-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: psmdb-operator^' + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.3KDWOJiUPV ++ mktemp + local LAST_ERR=/tmp/tmp.ck7Muc0Xx2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.3KDWOJiUPV clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created + cat /tmp/tmp.ck7Muc0Xx2 + rm /tmp/tmp.3KDWOJiUPV /tmp/tmp.ck7Muc0Xx2 + return 0 + yq eval $'\n\t\t\t(.spec.template.spec.containers[].image = "docker.io/perconalab/percona-server-mongodb-operator:PR-2256-98b7d85f") |\n\t\t\t((.. | select(.[] == "DISABLE_TELEMETRY")) |= .value="true") |\n\t\t\t((.. | select(.[] == "LOG_LEVEL")) |= .value="DEBUG")' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/cw-operator.yaml + kubectl_bin apply -n psmdb-operator -f - ++ mktemp + local LAST_OUT=/tmp/tmp.L5RPlnbb1W ++ mktemp + local LAST_ERR=/tmp/tmp.aLLJbpdTPE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -n psmdb-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.L5RPlnbb1W deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.aLLJbpdTPE + rm /tmp/tmp.L5RPlnbb1W /tmp/tmp.aLLJbpdTPE + return 0 + sleep 20 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.euEfEQZhaB +++ mktemp ++ local LAST_ERR=/tmp/tmp.SQ3cBI5NYP ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.euEfEQZhaB ++ cat /tmp/tmp.SQ3cBI5NYP ++ rm /tmp/tmp.euEfEQZhaB /tmp/tmp.SQ3cBI5NYP ++ return 0 + wait_operator_pod percona-server-mongodb-operator-7d9c7747d9-vcwsm + local pod=percona-server-mongodb-operator-7d9c7747d9-vcwsm + set +o xtrace waiting for pod/percona-server-mongodb-operator-7d9c7747d9-vcwsm to be ready.OK + echo 'Print operator info from log' Print operator info from log ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator + grep 'Manager starting up' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hZcxsgPwaM +++ mktemp ++ local LAST_ERR=/tmp/tmp.oBUNYwITdu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.hZcxsgPwaM ++ cat /tmp/tmp.oBUNYwITdu ++ rm /tmp/tmp.hZcxsgPwaM /tmp/tmp.oBUNYwITdu ++ return 0 + kubectl_bin logs -n psmdb-operator percona-server-mongodb-operator-7d9c7747d9-vcwsm ++ mktemp + local LAST_OUT=/tmp/tmp.elyUVW6gEn ++ mktemp + local LAST_ERR=/tmp/tmp.l0a6ZiEKwN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl logs -n psmdb-operator percona-server-mongodb-operator-7d9c7747d9-vcwsm + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.elyUVW6gEn + cat /tmp/tmp.l0a6ZiEKwN + rm /tmp/tmp.elyUVW6gEn /tmp/tmp.l0a6ZiEKwN + return 0 2026-04-14T11:10:07.409Z INFO setup Manager starting up {"gitCommit": "98b7d85f6fc9316ce5207960e26f4bbf2b33bfd3", "gitBranch": "PR-2256-98b7d85f", "buildTime": "", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} + create_namespace monitoring-pmm3-31037 + local namespace=monitoring-pmm3-31037 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' ++ mktemp + xargs kubectl delete ns + '[' -n '' ']' + desc 'cleaned up old namespaces monitoring-pmm3-31037' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces monitoring-pmm3-31037 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace monitoring-pmm3-31037 --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.krDWj1m77u ++ mktemp + local LAST_OUT=/tmp/tmp.2ToOrjS8Y3 ++ mktemp + local LAST_ERR=/tmp/tmp.gpTYtQQeJd + local exit_status=0 + local timeout=4 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.NVtqZOHjjW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get ns + for i in $(seq 0 2) + set +e + kubectl delete namespace monitoring-pmm3-31037 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.krDWj1m77u + cat /tmp/tmp.gpTYtQQeJd + rm /tmp/tmp.krDWj1m77u /tmp/tmp.gpTYtQQeJd + return 0 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2ToOrjS8Y3 + cat /tmp/tmp.NVtqZOHjjW + rm /tmp/tmp.2ToOrjS8Y3 /tmp/tmp.NVtqZOHjjW + return 0 + kubectl_bin wait --for=delete namespace monitoring-pmm3-31037 ++ mktemp + local LAST_OUT=/tmp/tmp.6ErjR6bZCN ++ mktemp + local LAST_ERR=/tmp/tmp.lYww3yFhOo + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete namespace monitoring-pmm3-31037 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.6ErjR6bZCN + cat /tmp/tmp.lYww3yFhOo + rm /tmp/tmp.6ErjR6bZCN /tmp/tmp.lYww3yFhOo + return 0 + desc 'create namespace monitoring-pmm3-31037' + set +o xtrace ----------------------------------------------------------------------------------- create namespace monitoring-pmm3-31037 ----------------------------------------------------------------------------------- + kubectl_bin create namespace monitoring-pmm3-31037 ++ mktemp + local LAST_OUT=/tmp/tmp.uc6j8QacIG ++ mktemp + local LAST_ERR=/tmp/tmp.UX9yWupJTN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace monitoring-pmm3-31037 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.uc6j8QacIG namespace/monitoring-pmm3-31037 created + cat /tmp/tmp.UX9yWupJTN + rm /tmp/tmp.uc6j8QacIG /tmp/tmp.UX9yWupJTN + return 0 + set_kube_ctx monitoring-pmm3-31037 + local namespace=monitoring-pmm3-31037 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.bI8lhCWYWW +++ mktemp ++ local LAST_ERR=/tmp/tmp.K5ZHbfw63t ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.bI8lhCWYWW ++ cat /tmp/tmp.K5ZHbfw63t ++ rm /tmp/tmp.bI8lhCWYWW /tmp/tmp.K5ZHbfw63t ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2256-98b7d85f-7-cluster7 --namespace=monitoring-pmm3-31037 ++ mktemp + local LAST_OUT=/tmp/tmp.fponOBWTn4 ++ mktemp + local LAST_ERR=/tmp/tmp.vfe26MqhOi + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-2256-98b7d85f-7-cluster7 --namespace=monitoring-pmm3-31037 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fponOBWTn4 Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2256-98b7d85f-7-cluster7" modified. + cat /tmp/tmp.vfe26MqhOi + rm /tmp/tmp.fponOBWTn4 /tmp/tmp.vfe26MqhOi + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.EdM3aj8B4D ++ mktemp + local LAST_ERR=/tmp/tmp.XTZmaOsLsA + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.EdM3aj8B4D namespace/cert-manager created + cat /tmp/tmp.XTZmaOsLsA + rm /tmp/tmp.EdM3aj8B4D /tmp/tmp.XTZmaOsLsA + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.ow9M4T95QD ++ mktemp + local LAST_ERR=/tmp/tmp.8nHkXhd137 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ow9M4T95QD namespace/cert-manager labeled + cat /tmp/tmp.8nHkXhd137 + rm /tmp/tmp.ow9M4T95QD /tmp/tmp.8nHkXhd137 + return 0 + kubectl_bin apply -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.3tTt3GqU5V ++ mktemp + local LAST_ERR=/tmp/tmp.2zY6SM4arH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.3tTt3GqU5V namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.2zY6SM4arH Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.3tTt3GqU5V /tmp/tmp.2zY6SM4arH + return 0 + kubectl_bin -n cert-manager wait pod -l app.kubernetes.io/instance=cert-manager --for=condition=ready ++ mktemp + local LAST_OUT=/tmp/tmp.YyDR6nXSGA ++ mktemp + local LAST_ERR=/tmp/tmp.nBlq5Isc7i + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl -n cert-manager wait pod -l app.kubernetes.io/instance=cert-manager --for=condition=ready + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.YyDR6nXSGA pod/cert-manager-559d798845-rsxxq condition met pod/cert-manager-cainjector-64958d9c7c-gz8c8 condition met pod/cert-manager-webhook-7fb6f99b56-gnt2w condition met + cat /tmp/tmp.nBlq5Isc7i + rm /tmp/tmp.YyDR6nXSGA /tmp/tmp.nBlq5Isc7i + return 0 + sleep 120 + desc 'install PMM Server' + set +o xtrace ----------------------------------------------------------------------------------- install PMM Server ----------------------------------------------------------------------------------- + deploy_pmm3_server + helm uninstall monitoring Error: uninstall: Release not loaded: monitoring: release: not found + : + helm repo remove percona "percona" has been removed from your repositories + kubectl delete clusterrole monitoring --ignore-not-found + kubectl delete clusterrolebinding monitoring --ignore-not-found + helm repo add percona https://percona.github.io/percona-helm-charts/ "percona" has been added to your repositories + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "percona" chart repository ...Successfully got an update from the "stable" chart repository Update Complete. ⎈Happy Helming!⎈ + [[ -n '' ]] + retry 10 60 helm install monitoring percona/pmm --set fullnameOverride=monitoring-server --set image.tag=3-dev-latest --set image.repository=docker.io/perconalab/pmm-server --set service.type=LoadBalancer --force + local max=10 + local delay=60 + shift 2 + local n=1 + helm install monitoring percona/pmm --set fullnameOverride=monitoring-server --set image.tag=3-dev-latest --set image.repository=docker.io/perconalab/pmm-server --set service.type=LoadBalancer --force NAME: monitoring LAST DEPLOYED: Tue Apr 14 11:13:15 2026 NAMESPACE: monitoring-pmm3-31037 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: Percona Monitoring and Management (PMM) An open source database monitoring, observability and management tool Check more info here: https://docs.percona.com/percona-monitoring-and-management/index.html Get the application URL: NOTE: It may take a few minutes for the LoadBalancer IP to be available. You can watch the status of by running 'kubectl get --namespace monitoring-pmm3-31037 svc -w monitoring-service' export SERVICE_IP=$(kubectl get svc --namespace monitoring-pmm3-31037 monitoring-service -o jsonpath="{.status.loadBalancer.ingress[0].ip}") echo https://$SERVICE_IP: Get password for the "admin" user: export ADMIN_PASS=$(kubectl get secret pmm-secret --namespace monitoring-pmm3-31037 -o jsonpath='{.data.PMM_ADMIN_PASSWORD}' | base64 --decode) echo $ADMIN_PASS + sleep 20 + kubectl_bin exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' ++ mktemp + local LAST_OUT=/tmp/tmp.4wMc1dcxoY ++ mktemp + local LAST_ERR=/tmp/tmp.zJuLyk0rgG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.4wMc1dcxoY + cat /tmp/tmp.zJuLyk0rgG Error from server (BadRequest): pod monitoring-server-0 does not have a host assigned + sleep 0 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.4wMc1dcxoY + cat /tmp/tmp.zJuLyk0rgG Error from server (BadRequest): pod monitoring-server-0 does not have a host assigned + sleep 4 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.4wMc1dcxoY + cat /tmp/tmp.zJuLyk0rgG Error from server (BadRequest): pod monitoring-server-0 does not have a host assigned + sleep 8 + cat /tmp/tmp.4wMc1dcxoY + cat /tmp/tmp.zJuLyk0rgG Error from server (BadRequest): pod monitoring-server-0 does not have a host assigned + rm /tmp/tmp.4wMc1dcxoY /tmp/tmp.zJuLyk0rgG + return 1 + echo 'Retry 0' Retry 0 + sleep 5 + let retry+=1 + '[' 1 -ge 20 ']' + kubectl_bin exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' ++ mktemp + local LAST_OUT=/tmp/tmp.HZdNBiNKqF ++ mktemp + local LAST_ERR=/tmp/tmp.Wi5KNlAZKx + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.HZdNBiNKqF + cat /tmp/tmp.Wi5KNlAZKx error: Internal error occurred: error sending request: Post "https://10.214.0.42:10250/exec/monitoring-pmm3-31037/monitoring-server-0/pmm?command=bash&command=-c&command=ls+-l+%2Fproc%2F%2A%2Fexe+2%3E%2Fdev%2Fnull%7C+grep+postgres+%3E%2Fdev%2Fnull&error=1&output=1": No agent available + sleep 0 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.HZdNBiNKqF + cat /tmp/tmp.Wi5KNlAZKx error: Internal error occurred: error sending request: Post "https://10.214.0.42:10250/exec/monitoring-pmm3-31037/monitoring-server-0/pmm?command=bash&command=-c&command=ls+-l+%2Fproc%2F%2A%2Fexe+2%3E%2Fdev%2Fnull%7C+grep+postgres+%3E%2Fdev%2Fnull&error=1&output=1": No agent available + sleep 4 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.HZdNBiNKqF + cat /tmp/tmp.Wi5KNlAZKx error: Internal error occurred: error sending request: Post "https://10.214.0.42:10250/exec/monitoring-pmm3-31037/monitoring-server-0/pmm?command=bash&command=-c&command=ls+-l+%2Fproc%2F%2A%2Fexe+2%3E%2Fdev%2Fnull%7C+grep+postgres+%3E%2Fdev%2Fnull&error=1&output=1": No agent available + sleep 8 + cat /tmp/tmp.HZdNBiNKqF + cat /tmp/tmp.Wi5KNlAZKx error: Internal error occurred: error sending request: Post "https://10.214.0.42:10250/exec/monitoring-pmm3-31037/monitoring-server-0/pmm?command=bash&command=-c&command=ls+-l+%2Fproc%2F%2A%2Fexe+2%3E%2Fdev%2Fnull%7C+grep+postgres+%3E%2Fdev%2Fnull&error=1&output=1": No agent available + rm /tmp/tmp.HZdNBiNKqF /tmp/tmp.Wi5KNlAZKx + return 1 + echo 'Retry 1' Retry 1 + sleep 5 + let retry+=1 + '[' 2 -ge 20 ']' + kubectl_bin exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' ++ mktemp + local LAST_OUT=/tmp/tmp.8V6SLxHS8l ++ mktemp + local LAST_ERR=/tmp/tmp.Ewrt5VlwHz + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.8V6SLxHS8l + cat /tmp/tmp.Ewrt5VlwHz error: Internal error occurred: unable to upgrade connection: container not found ("pmm") + sleep 0 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.8V6SLxHS8l + cat /tmp/tmp.Ewrt5VlwHz error: Internal error occurred: unable to upgrade connection: container not found ("pmm") + sleep 4 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.8V6SLxHS8l + cat /tmp/tmp.Ewrt5VlwHz error: Internal error occurred: unable to upgrade connection: container not found ("pmm") + sleep 8 + cat /tmp/tmp.8V6SLxHS8l + cat /tmp/tmp.Ewrt5VlwHz error: Internal error occurred: unable to upgrade connection: container not found ("pmm") + rm /tmp/tmp.8V6SLxHS8l /tmp/tmp.Ewrt5VlwHz + return 1 + echo 'Retry 2' Retry 2 + sleep 5 + let retry+=1 + '[' 3 -ge 20 ']' + kubectl_bin exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' ++ mktemp + local LAST_OUT=/tmp/tmp.1hWcZa99R4 ++ mktemp + local LAST_ERR=/tmp/tmp.MBeab5amuV + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.1hWcZa99R4 + cat /tmp/tmp.MBeab5amuV command terminated with exit code 1 + sleep 0 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.1hWcZa99R4 + cat /tmp/tmp.MBeab5amuV command terminated with exit code 1 + sleep 4 + for i in $(seq 0 2) + set +e + kubectl exec monitoring-server-0 -- bash -c 'ls -l /proc/*/exe 2>/dev/null| grep postgres >/dev/null' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.1hWcZa99R4 + cat /tmp/tmp.MBeab5amuV + rm /tmp/tmp.1hWcZa99R4 /tmp/tmp.MBeab5amuV + return 0 + cluster=monitoring-pmm3 + desc 'create secrets and start client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and start client ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.RMHisHwVCD ++ mktemp + local LAST_ERR=/tmp/tmp.O9QlTWkjY6 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/conf/secrets.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.RMHisHwVCD secret/some-users created secret/some-users unchanged + cat /tmp/tmp.O9QlTWkjY6 + rm /tmp/tmp.RMHisHwVCD /tmp/tmp.O9QlTWkjY6 + return 0 + yq '.spec.template.spec.volumes[0].secret.secretName="monitoring-pmm3-ssl"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/conf/client_with_tls.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.9M8IR35KKY ++ mktemp + local LAST_ERR=/tmp/tmp.9tdfpsx7EE + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.9M8IR35KKY deployment.apps/psmdb-client created + cat /tmp/tmp.9tdfpsx7EE + rm /tmp/tmp.9M8IR35KKY /tmp/tmp.9tdfpsx7EE + return 0 + sleep 90 + desc 'create first PSMDB cluster monitoring-pmm3' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster monitoring-pmm3 ----------------------------------------------------------------------------------- + custom_cluster_name=super-custom + yq eval '(.spec | select(.image == null)).image = "docker.io/perconalab/percona-server-mongodb-operator:main-mongod8.0"' /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/conf/monitoring-pmm3-rs0.yml + yq eval '(.spec | select(has("pmm"))).pmm.image = "docker.io/perconalab/pmm-client:3-dev-latest"' - + yq eval '(.spec | select(has("pmm"))).pmm.customClusterName = "super-custom"' - + yq eval '(.spec | select(has("initImage"))).initImage = "docker.io/perconalab/percona-server-mongodb-operator:PR-2256-98b7d85f"' - + yq eval '(.spec | select(has("backup"))).backup.image = "docker.io/perconalab/percona-server-mongodb-operator:main-backup"' - + yq eval '.spec.upgradeOptions.apply = "Never"' - + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.TiIgnmmHP3 ++ mktemp + local LAST_ERR=/tmp/tmp.CXaU9zaUTZ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.TiIgnmmHP3 perconaservermongodb.psmdb.percona.com/monitoring-pmm3 created + cat /tmp/tmp.CXaU9zaUTZ + rm /tmp/tmp.TiIgnmmHP3 /tmp/tmp.CXaU9zaUTZ + return 0 + wait_for_running monitoring-pmm3-rs0 3 + local name=monitoring-pmm3-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=monitoring-pmm3 ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod monitoring-pmm3-rs0-0 + local pod=monitoring-pmm3-rs0-0 + set +o xtrace waiting for pod/monitoring-pmm3-rs0-0 to be ready...................OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod monitoring-pmm3-rs0-1 + local pod=monitoring-pmm3-rs0-1 + set +o xtrace waiting for pod/monitoring-pmm3-rs0-1 to be ready.............OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ojQaBJfIhZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.X5fbaa40Gq ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ojQaBJfIhZ ++ cat /tmp/tmp.X5fbaa40Gq ++ rm /tmp/tmp.ojQaBJfIhZ /tmp/tmp.X5fbaa40Gq ++ return 0 + [[ '' == true ]] + wait_pod monitoring-pmm3-rs0-2 + local pod=monitoring-pmm3-rs0-2 + set +o xtrace waiting for pod/monitoring-pmm3-rs0-2 to be ready..................OK ++ kubectl_bin get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Dd8xLDA8uJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.DrcfGX2KlI ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Dd8xLDA8uJ ++ cat /tmp/tmp.DrcfGX2KlI ++ rm /tmp/tmp.Dd8xLDA8uJ /tmp/tmp.DrcfGX2KlI ++ return 0 + [[ '' == true ]] ++ kubectl_bin get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gGkfkQmeow +++ mktemp ++ local LAST_ERR=/tmp/tmp.eJB76GILqT ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gGkfkQmeow ++ cat /tmp/tmp.eJB76GILqT ++ rm /tmp/tmp.gGkfkQmeow /tmp/tmp.eJB76GILqT ++ return 0 + [[ '' == true ]] + sleep 10 + [[ true == true ]] + set +x Waiting for cluster readyness...................... + desc 'check if pmm-client container is not enabled' + set +o xtrace ----------------------------------------------------------------------------------- check if pmm-client container is not enabled ----------------------------------------------------------------------------------- + compare_kubectl statefulset/monitoring-pmm3-rs0 -no-pmm + local resource=statefulset/monitoring-pmm3-rs0 + local postfix=-no-pmm + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-rs0-no-pmm.yml + local new_result=/tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-rs0-no-pmm-oc.yml ']' + kubectl_bin get -o yaml statefulset/monitoring-pmm3-rs0 ++ mktemp + yq eval $'\n\t\t\tdel(.metadata.ownerReferences[].apiVersion) |\n\t\t\tdel(.metadata.managedFields) |\n\t\t\tdel(.. | select(has("creationTimestamp")).creationTimestamp) |\n\t\t\tdel(.. | select(has("namespace")).namespace) |\n\t\t\tdel(.. | select(has("uid")).uid) |\n\t\t\tdel(.metadata.resourceVersion) |\n\t\t\tdel(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) |\n\t\t\tdel(.metadata.selfLink) |\n\t\t\tdel(.metadata.annotations."cloud.google.com/neg") |\n\t\t\tdel(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") |\n\t\t\tdel(.. | select(has("image")).image) |\n\t\t\tdel(.. | select(has("clusterIP")).clusterIP) |\n\t\t\tdel(.. | select(has("clusterIPs")).clusterIPs) |\n\t\t\tdel(.. | select(has("dataSource")).dataSource) |\n\t\t\tdel(.. | select(has("procMount")).procMount) |\n\t\t\tdel(.. | select(has("storageClassName")).storageClassName) |\n\t\t\tdel(.. | select(has("finalizers")).finalizers) |\n\t\t\tdel(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") |\n\t\t\tdel(.. | select(has("volumeName")).volumeName) |\n\t\t\tdel(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.spec.volumeMode) |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") |\n\t\t\tdel(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") |\n\t\t\tdel(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") |\n\t\t\tdel(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) |\n\t\t\tdel(.. | select(has("healthCheckNodePort")).healthCheckNodePort) |\n\t\t\tdel(.. | select(has("nodePort")).nodePort) |\n\t\t\tdel(.status) |\n\t\t\t(.. | select(tag == "!!str")) |= sub("monitoring-pmm3-31037", "NAME_SPACE") |\n\t\t\tdel(.spec.volumeClaimTemplates[].apiVersion) |\n\t\t\tdel(.spec.volumeClaimTemplates[].kind) |\n\t\t\tdel(.spec.ipFamilies) |\n\t\t\tdel(.spec.ipFamilyPolicy) |\n\t\t\t(.. | select(. == "extensions/v1beta1")) = "apps/v1" |\n\t\t\t(.. | select(. == "batch/v1beta1")) = "batch/v1" ' - + local LAST_OUT=/tmp/tmp.aaqLS7ZJSt ++ mktemp + local LAST_ERR=/tmp/tmp.ZZyJwZesFr + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml statefulset/monitoring-pmm3-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.aaqLS7ZJSt + cat /tmp/tmp.ZZyJwZesFr + rm /tmp/tmp.aaqLS7ZJSt /tmp/tmp.ZZyJwZesFr + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-rs0-no-pmm.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-rs0-no-pmm.yml /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + log 'compare_kubectl: statefulset/monitoring-pmm3-rs0 OK' + set +o xtrace [2026-04-14T11:19:13+0000] compare_kubectl: statefulset/monitoring-pmm3-rs0 OK + sleep 10 + custom_port=27019 + run_mongos 'db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' userAdmin:userAdmin123456@monitoring-pmm3-mongos.monitoring-pmm3-31037 '' '' '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' 27019 + local 'command=db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})' + local uri=userAdmin:userAdmin123456@monitoring-pmm3-mongos.monitoring-pmm3-31037 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27019 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27019 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ePAmENk3yO +++ mktemp ++ local LAST_ERR=/tmp/tmp.FWjH81DiKj ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ePAmENk3yO ++ cat /tmp/tmp.FWjH81DiKj ++ rm /tmp/tmp.ePAmENk3yO /tmp/tmp.FWjH81DiKj ++ return 0 + local client_container=psmdb-client-7c8f77fb65-rw769 + kubectl_bin exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ mktemp + local LAST_OUT=/tmp/tmp.mYXJfBxXXe ++ mktemp + local LAST_ERR=/tmp/tmp.2y8cyMRJrC + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''db.createUser({user:"myApp",pwd:"myPass",roles:[{db:"myApp",role:"readWrite"}]})\n'\'' | mongo mongodb://userAdmin:userAdmin123456@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mYXJfBxXXe Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-04-14T11:19:25.457Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("f3a1754c-d4b3-4ab5-afb5-471279f7a4c4") } Percona Server for MongoDB server version: v8.0.20-8 WARNING: shell and server versions do not match Successfully added user: { "user" : "myApp", "roles" : [ { "db" : "myApp", "role" : "readWrite" } ] } bye + cat /tmp/tmp.2y8cyMRJrC + rm /tmp/tmp.mYXJfBxXXe /tmp/tmp.2y8cyMRJrC + return 0 + run_mongos 'sh.enableSharding("myApp")' clusterAdmin:clusterAdmin123456@monitoring-pmm3-mongos.monitoring-pmm3-31037 '' '' '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' 27019 + local 'command=sh.enableSharding("myApp")' + local uri=clusterAdmin:clusterAdmin123456@monitoring-pmm3-mongos.monitoring-pmm3-31037 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27019 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27019 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.R3j5h0A5pH +++ mktemp ++ local LAST_ERR=/tmp/tmp.WVd6tu7iJR ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.R3j5h0A5pH ++ cat /tmp/tmp.WVd6tu7iJR ++ rm /tmp/tmp.R3j5h0A5pH /tmp/tmp.WVd6tu7iJR ++ return 0 + local client_container=psmdb-client-7c8f77fb65-rw769 + kubectl_bin exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''sh.enableSharding("myApp")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ mktemp + local LAST_OUT=/tmp/tmp.xrxGcoPoIO ++ mktemp + local LAST_ERR=/tmp/tmp.86hqnTWCNj + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''sh.enableSharding("myApp")\n'\'' | mongo mongodb://clusterAdmin:clusterAdmin123456@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.xrxGcoPoIO Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-04-14T11:19:27.554Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("48aae856-d674-427a-8802-996b2d69a236") } Percona Server for MongoDB server version: v8.0.20-8 WARNING: shell and server versions do not match { "ok" : 1, "$clusterTime" : { "clusterTime" : Timestamp(1776165567, 8), "signature" : { "hash" : BinData(0,"v5srl6PBVPJP5meBjH1gXzhdTwo="), "keyId" : NumberLong("7628572618819371032") } }, "operationTime" : Timestamp(1776165567, 5) } bye + cat /tmp/tmp.86hqnTWCNj + rm /tmp/tmp.xrxGcoPoIO /tmp/tmp.86hqnTWCNj + return 0 + insert_data_mongos 100500 myApp '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' 27019 + local data=100500 + local db_name=myApp + local 'flags=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27019 + run_mongos 'use myApp\n db.test.insert({ x: 100500 })' myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037 '' '' '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' 27019 + local 'command=use myApp\n db.test.insert({ x: 100500 })' + local uri=myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27019 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27019 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DGgi97vb2b +++ mktemp ++ local LAST_ERR=/tmp/tmp.tOdUsJqg6D ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.DGgi97vb2b ++ cat /tmp/tmp.tOdUsJqg6D ++ rm /tmp/tmp.DGgi97vb2b /tmp/tmp.tOdUsJqg6D ++ return 0 + local client_container=psmdb-client-7c8f77fb65-rw769 + kubectl_bin exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ mktemp + local LAST_OUT=/tmp/tmp.JnsEVjn9kB ++ mktemp + local LAST_ERR=/tmp/tmp.TexMsNwuKw + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100500 })\n'\'' | mongo mongodb://myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.JnsEVjn9kB Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-04-14T11:19:29.685Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("83fcaf9d-ae15-427d-930f-07f148b72204") } Percona Server for MongoDB server version: v8.0.20-8 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.TexMsNwuKw + rm /tmp/tmp.JnsEVjn9kB /tmp/tmp.TexMsNwuKw + return 0 + insert_data_mongos 100600 myApp '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' 27019 + local data=100600 + local db_name=myApp + local 'flags=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27019 + run_mongos 'use myApp\n db.test.insert({ x: 100600 })' myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037 '' '' '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' 27019 + local 'command=use myApp\n db.test.insert({ x: 100600 })' + local uri=myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27019 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27019 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gIBa60YDK7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.OA8JPjD1RG ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.gIBa60YDK7 ++ cat /tmp/tmp.OA8JPjD1RG ++ rm /tmp/tmp.gIBa60YDK7 /tmp/tmp.OA8JPjD1RG ++ return 0 + local client_container=psmdb-client-7c8f77fb65-rw769 + kubectl_bin exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100600 })\n'\'' | mongo mongodb://myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ mktemp + local LAST_OUT=/tmp/tmp.fZlaeiRWA0 ++ mktemp + local LAST_ERR=/tmp/tmp.jhWjf7yzke + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100600 })\n'\'' | mongo mongodb://myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.fZlaeiRWA0 Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-04-14T11:19:32.303Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("c9e1ac2f-e935-4c86-9969-9a4a810a1b5f") } Percona Server for MongoDB server version: v8.0.20-8 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.jhWjf7yzke + rm /tmp/tmp.fZlaeiRWA0 /tmp/tmp.jhWjf7yzke + return 0 + insert_data_mongos 100700 myApp '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' 27019 + local data=100700 + local db_name=myApp + local 'flags=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27019 + run_mongos 'use myApp\n db.test.insert({ x: 100700 })' myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037 '' '' '--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' 27019 + local 'command=use myApp\n db.test.insert({ x: 100700 })' + local uri=myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037 + local driver=mongodb + local suffix=.svc.cluster.local + local 'mongo_flag=--tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + local port=27019 + local mongo_bin=mongo ++ echo .svc.cluster.local ++ awk -F: '{print $2}' + suffix_port= + [[ -z '' ]] + suffix=.svc.cluster.local:27019 ++ kubectl_bin get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dTyu0Jla7L +++ mktemp ++ local LAST_ERR=/tmp/tmp.wovKlcdXbb ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get pods --selector=name=psmdb-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.dTyu0Jla7L ++ cat /tmp/tmp.wovKlcdXbb ++ rm /tmp/tmp.dTyu0Jla7L /tmp/tmp.wovKlcdXbb ++ return 0 + local client_container=psmdb-client-7c8f77fb65-rw769 + kubectl_bin exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100700 })\n'\'' | mongo mongodb://myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' ++ mktemp + local LAST_OUT=/tmp/tmp.vvrzWEyftk ++ mktemp + local LAST_ERR=/tmp/tmp.CbrzbsL6bu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl exec psmdb-client-7c8f77fb65-rw769 -- bash -c 'printf '\''use myApp\n db.test.insert({ x: 100700 })\n'\'' | mongo mongodb://myApp:myPass@monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin --tlsCertificateKeyFile /tmp/tls.pem --tlsCAFile /etc/mongodb-ssl/ca.crt --tls' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.vvrzWEyftk Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://monitoring-pmm3-mongos.monitoring-pmm3-31037.svc.cluster.local:27019/admin?compressors=disabled&gssapiServiceName=mongodb {"t":{"$date":"2026-04-14T11:19:35.027Z"},"s":"I", "c":"NETWORK", "id":5490002, "ctx":"thread1","msg":"Started a new thread for the timer service"} Implicit session: session { "id" : UUID("8985dbf8-5131-4d41-9158-08da07cbdd1a") } Percona Server for MongoDB server version: v8.0.20-8 WARNING: shell and server versions do not match switched to db myApp WriteResult({ "nInserted" : 1 }) bye + cat /tmp/tmp.CbrzbsL6bu + rm /tmp/tmp.vvrzWEyftk /tmp/tmp.CbrzbsL6bu + return 0 + desc 'add PMM3 token to secret' + set +o xtrace ----------------------------------------------------------------------------------- add PMM3 token to secret ----------------------------------------------------------------------------------- ++ get_pmm_server_token operator ++ local key_name=operator ++ [[ -z operator ]] ++ local ADMIN_PASSWORD +++ base64 --decode +++ kubectl get secret pmm-secret -o 'jsonpath={.data.PMM_ADMIN_PASSWORD}' ++ ADMIN_PASSWORD=KyEo.de6w,DsTnRv ++ [[ -z KyEo.de6w,DsTnRv ]] ++ local create_response create_status_code create_json_response ++++ get_service_endpoint monitoring-service ++++ local service=monitoring-service +++++ kubectl_bin get service/monitoring-service -o json +++++ jq '.status.loadBalancer.ingress[].hostname' +++++ sed -e 's/^"//; s/"$//;' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.1kCI9azz1p ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.7zDIuEs7dB +++++ local exit_status=0 +++++ local timeout=4 ++++++ seq 0 2 +++++ for i in $(seq 0 2) +++++ set +e +++++ kubectl get service/monitoring-service -o json +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 -a -n 1 ']' +++++ break +++++ cat /tmp/tmp.1kCI9azz1p +++++ cat /tmp/tmp.7zDIuEs7dB +++++ rm /tmp/tmp.1kCI9azz1p /tmp/tmp.7zDIuEs7dB +++++ return 0 ++++ local hostname=null ++++ '[' -n null -a null '!=' null ']' +++++ kubectl_bin get service/monitoring-service -o json +++++ jq '.status.loadBalancer.ingress[].ip' +++++ sed -e 's/^"//; s/"$//;' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.JssUEusc2m ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.5IPXG7tqDS +++++ local exit_status=0 +++++ local timeout=4 ++++++ seq 0 2 +++++ for i in $(seq 0 2) +++++ set +e +++++ kubectl get service/monitoring-service -o json +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 -a -n 1 ']' +++++ break +++++ cat /tmp/tmp.JssUEusc2m +++++ cat /tmp/tmp.5IPXG7tqDS +++++ rm /tmp/tmp.JssUEusc2m /tmp/tmp.5IPXG7tqDS +++++ return 0 ++++ local ip=34.133.92.251 ++++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++++ echo 34.133.92.251 ++++ return +++ curl --insecure -s -X POST -H 'Content-Type: application/json' -H 'Accept: application/json' -d '{"name":"operator", "role":"Admin", "isDisabled":false}' --user admin:KyEo.de6w,DsTnRv https://34.133.92.251/graph/api/serviceaccounts -w '\n%{http_code}' ++ create_response=$'{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}\n201' +++ echo $'{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}\n201' +++ tail -n1 ++ create_status_code=201 +++ echo $'{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}\n201' +++ sed '$ d' ++ create_json_response='{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}' ++ [[ 201 -ne 201 ]] ++ local service_account_id +++ echo '{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}' +++ jq -r .id ++ service_account_id=2 ++ [[ -z 2 ]] ++ [[ 2 == null ]] ++ local token_response token_status_code token_json_response ++++ get_service_endpoint monitoring-service ++++ local service=monitoring-service +++++ kubectl_bin get service/monitoring-service -o json +++++ jq '.status.loadBalancer.ingress[].hostname' +++++ sed -e 's/^"//; s/"$//;' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.ymTadnIH9F ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.HemhSv8m5Z +++++ local exit_status=0 +++++ local timeout=4 ++++++ seq 0 2 +++++ for i in $(seq 0 2) +++++ set +e +++++ kubectl get service/monitoring-service -o json +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 -a -n 1 ']' +++++ break +++++ cat /tmp/tmp.ymTadnIH9F +++++ cat /tmp/tmp.HemhSv8m5Z +++++ rm /tmp/tmp.ymTadnIH9F /tmp/tmp.HemhSv8m5Z +++++ return 0 ++++ local hostname=null ++++ '[' -n null -a null '!=' null ']' +++++ kubectl_bin get service/monitoring-service -o json +++++ jq '.status.loadBalancer.ingress[].ip' +++++ sed -e 's/^"//; s/"$//;' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.tCktf6RLTN ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.E8u3LyuHKk +++++ local exit_status=0 +++++ local timeout=4 ++++++ seq 0 2 +++++ for i in $(seq 0 2) +++++ set +e +++++ kubectl get service/monitoring-service -o json +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 -a -n 1 ']' +++++ break +++++ cat /tmp/tmp.tCktf6RLTN +++++ cat /tmp/tmp.E8u3LyuHKk +++++ rm /tmp/tmp.tCktf6RLTN /tmp/tmp.E8u3LyuHKk +++++ return 0 ++++ local ip=34.133.92.251 ++++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++++ echo 34.133.92.251 ++++ return +++ curl --insecure -s -X POST -H 'Content-Type: application/json' -d '{"name":"operator"}' --user admin:KyEo.de6w,DsTnRv https://34.133.92.251/graph/api/serviceaccounts/2/tokens -w '\n%{http_code}' ++ token_response=$'{"id":1,"name":"operator","key":"glsa_Rz1HV5FArehXRjYNxkFJK3k71Wywd0kp_8acfcd7c"}\n200' +++ echo $'{"id":1,"name":"operator","key":"glsa_Rz1HV5FArehXRjYNxkFJK3k71Wywd0kp_8acfcd7c"}\n200' +++ tail -n1 ++ token_status_code=200 +++ echo $'{"id":1,"name":"operator","key":"glsa_Rz1HV5FArehXRjYNxkFJK3k71Wywd0kp_8acfcd7c"}\n200' +++ sed '$ d' ++ token_json_response='{"id":1,"name":"operator","key":"glsa_Rz1HV5FArehXRjYNxkFJK3k71Wywd0kp_8acfcd7c"}' ++ [[ 200 -ne 200 ]] ++ echo '{"id":1,"name":"operator","key":"glsa_Rz1HV5FArehXRjYNxkFJK3k71Wywd0kp_8acfcd7c"}' ++ jq -r .key + TOKEN=glsa_Rz1HV5FArehXRjYNxkFJK3k71Wywd0kp_8acfcd7c + kubectl_bin patch secret some-users --type merge --patch '{"stringData": {"PMM_SERVER_TOKEN": "glsa_Rz1HV5FArehXRjYNxkFJK3k71Wywd0kp_8acfcd7c"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.FsVaDK2JX5 ++ mktemp + local LAST_ERR=/tmp/tmp.IuN346Ahpu + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch secret some-users --type merge --patch '{"stringData": {"PMM_SERVER_TOKEN": "glsa_Rz1HV5FArehXRjYNxkFJK3k71Wywd0kp_8acfcd7c"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.FsVaDK2JX5 secret/some-users patched + cat /tmp/tmp.IuN346Ahpu + rm /tmp/tmp.FsVaDK2JX5 /tmp/tmp.IuN346Ahpu + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- + wait_for_running monitoring-pmm3-rs0 3 + local name=monitoring-pmm3-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=monitoring-pmm3 ++ seq 0 2 + for i in $(seq 0 $last_pod) + [[ 0 -eq 2 ]] + wait_pod monitoring-pmm3-rs0-0 + local pod=monitoring-pmm3-rs0-0 + set +o xtrace waiting for pod/monitoring-pmm3-rs0-0 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 1 -eq 2 ]] + wait_pod monitoring-pmm3-rs0-1 + local pod=monitoring-pmm3-rs0-1 + set +o xtrace waiting for pod/monitoring-pmm3-rs0-1 to be ready.OK + for i in $(seq 0 $last_pod) + [[ 2 -eq 2 ]] ++ kubectl_bin get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RLRDTRvOk9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.2AXtjpyipU ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].arbiter.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.RLRDTRvOk9 ++ cat /tmp/tmp.2AXtjpyipU ++ rm /tmp/tmp.RLRDTRvOk9 /tmp/tmp.2AXtjpyipU ++ return 0 + [[ '' == true ]] + wait_pod monitoring-pmm3-rs0-2 + local pod=monitoring-pmm3-rs0-2 + set +o xtrace waiting for pod/monitoring-pmm3-rs0-2 to be ready.OK ++ kubectl_bin get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZokC0VmOPf +++ mktemp ++ local LAST_ERR=/tmp/tmp.OaIUT53gtp ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].nonvoting.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.ZokC0VmOPf ++ cat /tmp/tmp.OaIUT53gtp ++ rm /tmp/tmp.ZokC0VmOPf /tmp/tmp.OaIUT53gtp ++ return 0 + [[ '' == true ]] ++ kubectl_bin get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XdntVfvZjl +++ mktemp ++ local LAST_ERR=/tmp/tmp.VVDteTRjJS ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb monitoring-pmm3 -o 'jsonpath={.spec.replsets[?(@.name=="rs0")].hidden.enabled}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.XdntVfvZjl ++ cat /tmp/tmp.VVDteTRjJS ++ rm /tmp/tmp.XdntVfvZjl /tmp/tmp.VVDteTRjJS ++ return 0 + [[ '' == true ]] + sleep 10 + [[ true == true ]] + set +x Waiting for cluster readyness........................................................................................................................................................................................................................................................................................................................ + sleep 90 + desc 'check if pmm-client container enabled' + set +o xtrace ----------------------------------------------------------------------------------- check if pmm-client container enabled ----------------------------------------------------------------------------------- + compare_kubectl statefulset/monitoring-pmm3-rs0 + local resource=statefulset/monitoring-pmm3-rs0 + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-rs0.yml + local new_result=/tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-rs0-oc.yml ']' + kubectl_bin get -o yaml statefulset/monitoring-pmm3-rs0 + yq eval $'\n\t\t\tdel(.metadata.ownerReferences[].apiVersion) |\n\t\t\tdel(.metadata.managedFields) |\n\t\t\tdel(.. | select(has("creationTimestamp")).creationTimestamp) |\n\t\t\tdel(.. | select(has("namespace")).namespace) |\n\t\t\tdel(.. | select(has("uid")).uid) |\n\t\t\tdel(.metadata.resourceVersion) |\n\t\t\tdel(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) |\n\t\t\tdel(.metadata.selfLink) |\n\t\t\tdel(.metadata.annotations."cloud.google.com/neg") |\n\t\t\tdel(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") |\n\t\t\tdel(.. | select(has("image")).image) |\n\t\t\tdel(.. | select(has("clusterIP")).clusterIP) |\n\t\t\tdel(.. | select(has("clusterIPs")).clusterIPs) |\n\t\t\tdel(.. | select(has("dataSource")).dataSource) |\n\t\t\tdel(.. | select(has("procMount")).procMount) |\n\t\t\tdel(.. | select(has("storageClassName")).storageClassName) |\n\t\t\tdel(.. | select(has("finalizers")).finalizers) |\n\t\t\tdel(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") |\n\t\t\tdel(.. | select(has("volumeName")).volumeName) |\n\t\t\tdel(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.spec.volumeMode) |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") |\n\t\t\tdel(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") |\n\t\t\tdel(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") |\n\t\t\tdel(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) |\n\t\t\tdel(.. | select(has("healthCheckNodePort")).healthCheckNodePort) |\n\t\t\tdel(.. | select(has("nodePort")).nodePort) |\n\t\t\tdel(.status) |\n\t\t\t(.. | select(tag == "!!str")) |= sub("monitoring-pmm3-31037", "NAME_SPACE") |\n\t\t\tdel(.spec.volumeClaimTemplates[].apiVersion) |\n\t\t\tdel(.spec.volumeClaimTemplates[].kind) |\n\t\t\tdel(.spec.ipFamilies) |\n\t\t\tdel(.spec.ipFamilyPolicy) |\n\t\t\t(.. | select(. == "extensions/v1beta1")) = "apps/v1" |\n\t\t\t(.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.cCkYmpedMR ++ mktemp + local LAST_ERR=/tmp/tmp.hW2qVYY7Gl + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml statefulset/monitoring-pmm3-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.cCkYmpedMR + cat /tmp/tmp.hW2qVYY7Gl + rm /tmp/tmp.cCkYmpedMR /tmp/tmp.hW2qVYY7Gl + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-rs0.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-rs0.yml /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-rs0.yml + log 'compare_kubectl: statefulset/monitoring-pmm3-rs0 OK' + set +o xtrace [2026-04-14T11:31:56+0000] compare_kubectl: statefulset/monitoring-pmm3-rs0 OK + compare_kubectl service/monitoring-pmm3-rs0 + local resource=service/monitoring-pmm3-rs0 + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/service_monitoring-pmm3-rs0.yml + local new_result=/tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-rs0.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/service_monitoring-pmm3-rs0-oc.yml ']' + kubectl_bin get -o yaml service/monitoring-pmm3-rs0 + yq eval $'\n\t\t\tdel(.metadata.ownerReferences[].apiVersion) |\n\t\t\tdel(.metadata.managedFields) |\n\t\t\tdel(.. | select(has("creationTimestamp")).creationTimestamp) |\n\t\t\tdel(.. | select(has("namespace")).namespace) |\n\t\t\tdel(.. | select(has("uid")).uid) |\n\t\t\tdel(.metadata.resourceVersion) |\n\t\t\tdel(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) |\n\t\t\tdel(.metadata.selfLink) |\n\t\t\tdel(.metadata.annotations."cloud.google.com/neg") |\n\t\t\tdel(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") |\n\t\t\tdel(.. | select(has("image")).image) |\n\t\t\tdel(.. | select(has("clusterIP")).clusterIP) |\n\t\t\tdel(.. | select(has("clusterIPs")).clusterIPs) |\n\t\t\tdel(.. | select(has("dataSource")).dataSource) |\n\t\t\tdel(.. | select(has("procMount")).procMount) |\n\t\t\tdel(.. | select(has("storageClassName")).storageClassName) |\n\t\t\tdel(.. | select(has("finalizers")).finalizers) |\n\t\t\tdel(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") |\n\t\t\tdel(.. | select(has("volumeName")).volumeName) |\n\t\t\tdel(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.spec.volumeMode) |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") |\n\t\t\tdel(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") |\n\t\t\tdel(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") |\n\t\t\tdel(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) |\n\t\t\tdel(.. | select(has("healthCheckNodePort")).healthCheckNodePort) |\n\t\t\tdel(.. | select(has("nodePort")).nodePort) |\n\t\t\tdel(.status) |\n\t\t\t(.. | select(tag == "!!str")) |= sub("monitoring-pmm3-31037", "NAME_SPACE") |\n\t\t\tdel(.spec.volumeClaimTemplates[].apiVersion) |\n\t\t\tdel(.spec.volumeClaimTemplates[].kind) |\n\t\t\tdel(.spec.ipFamilies) |\n\t\t\tdel(.spec.ipFamilyPolicy) |\n\t\t\t(.. | select(. == "extensions/v1beta1")) = "apps/v1" |\n\t\t\t(.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.X0PFqImk1w ++ mktemp + local LAST_ERR=/tmp/tmp.mtqDznIsoc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml service/monitoring-pmm3-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.X0PFqImk1w + cat /tmp/tmp.mtqDznIsoc + rm /tmp/tmp.X0PFqImk1w /tmp/tmp.mtqDznIsoc + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-rs0.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-rs0.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-rs0.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/service_monitoring-pmm3-rs0.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/service_monitoring-pmm3-rs0.yml /tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-rs0.yml + log 'compare_kubectl: service/monitoring-pmm3-rs0 OK' + set +o xtrace [2026-04-14T11:31:57+0000] compare_kubectl: service/monitoring-pmm3-rs0 OK + compare_kubectl service/monitoring-pmm3-mongos + local resource=service/monitoring-pmm3-mongos + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/service_monitoring-pmm3-mongos.yml + local new_result=/tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-mongos.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/service_monitoring-pmm3-mongos-oc.yml ']' + kubectl_bin get -o yaml service/monitoring-pmm3-mongos + yq eval $'\n\t\t\tdel(.metadata.ownerReferences[].apiVersion) |\n\t\t\tdel(.metadata.managedFields) |\n\t\t\tdel(.. | select(has("creationTimestamp")).creationTimestamp) |\n\t\t\tdel(.. | select(has("namespace")).namespace) |\n\t\t\tdel(.. | select(has("uid")).uid) |\n\t\t\tdel(.metadata.resourceVersion) |\n\t\t\tdel(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) |\n\t\t\tdel(.metadata.selfLink) |\n\t\t\tdel(.metadata.annotations."cloud.google.com/neg") |\n\t\t\tdel(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") |\n\t\t\tdel(.. | select(has("image")).image) |\n\t\t\tdel(.. | select(has("clusterIP")).clusterIP) |\n\t\t\tdel(.. | select(has("clusterIPs")).clusterIPs) |\n\t\t\tdel(.. | select(has("dataSource")).dataSource) |\n\t\t\tdel(.. | select(has("procMount")).procMount) |\n\t\t\tdel(.. | select(has("storageClassName")).storageClassName) |\n\t\t\tdel(.. | select(has("finalizers")).finalizers) |\n\t\t\tdel(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") |\n\t\t\tdel(.. | select(has("volumeName")).volumeName) |\n\t\t\tdel(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.spec.volumeMode) |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") |\n\t\t\tdel(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") |\n\t\t\tdel(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") |\n\t\t\tdel(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) |\n\t\t\tdel(.. | select(has("healthCheckNodePort")).healthCheckNodePort) |\n\t\t\tdel(.. | select(has("nodePort")).nodePort) |\n\t\t\tdel(.status) |\n\t\t\t(.. | select(tag == "!!str")) |= sub("monitoring-pmm3-31037", "NAME_SPACE") |\n\t\t\tdel(.spec.volumeClaimTemplates[].apiVersion) |\n\t\t\tdel(.spec.volumeClaimTemplates[].kind) |\n\t\t\tdel(.spec.ipFamilies) |\n\t\t\tdel(.spec.ipFamilyPolicy) |\n\t\t\t(.. | select(. == "extensions/v1beta1")) = "apps/v1" |\n\t\t\t(.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.lKUdH90RNI ++ mktemp + local LAST_ERR=/tmp/tmp.rFsJWQpmKp + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml service/monitoring-pmm3-mongos + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lKUdH90RNI + cat /tmp/tmp.rFsJWQpmKp + rm /tmp/tmp.lKUdH90RNI /tmp/tmp.rFsJWQpmKp + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-mongos.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-mongos.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-mongos.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/service_monitoring-pmm3-mongos.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/service_monitoring-pmm3-mongos.yml /tmp/tmp.KtaPi3a7ND/service_monitoring-pmm3-mongos.yml + log 'compare_kubectl: service/monitoring-pmm3-mongos OK' + set +o xtrace [2026-04-14T11:31:57+0000] compare_kubectl: service/monitoring-pmm3-mongos OK + compare_kubectl statefulset/monitoring-pmm3-cfg + local resource=statefulset/monitoring-pmm3-cfg + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-cfg.yml + local new_result=/tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-cfg.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-cfg-oc.yml ']' + kubectl_bin get -o yaml statefulset/monitoring-pmm3-cfg + yq eval $'\n\t\t\tdel(.metadata.ownerReferences[].apiVersion) |\n\t\t\tdel(.metadata.managedFields) |\n\t\t\tdel(.. | select(has("creationTimestamp")).creationTimestamp) |\n\t\t\tdel(.. | select(has("namespace")).namespace) |\n\t\t\tdel(.. | select(has("uid")).uid) |\n\t\t\tdel(.metadata.resourceVersion) |\n\t\t\tdel(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) |\n\t\t\tdel(.metadata.selfLink) |\n\t\t\tdel(.metadata.annotations."cloud.google.com/neg") |\n\t\t\tdel(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") |\n\t\t\tdel(.. | select(has("image")).image) |\n\t\t\tdel(.. | select(has("clusterIP")).clusterIP) |\n\t\t\tdel(.. | select(has("clusterIPs")).clusterIPs) |\n\t\t\tdel(.. | select(has("dataSource")).dataSource) |\n\t\t\tdel(.. | select(has("procMount")).procMount) |\n\t\t\tdel(.. | select(has("storageClassName")).storageClassName) |\n\t\t\tdel(.. | select(has("finalizers")).finalizers) |\n\t\t\tdel(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") |\n\t\t\tdel(.. | select(has("volumeName")).volumeName) |\n\t\t\tdel(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.spec.volumeMode) |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") |\n\t\t\tdel(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") |\n\t\t\tdel(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") |\n\t\t\tdel(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) |\n\t\t\tdel(.. | select(has("healthCheckNodePort")).healthCheckNodePort) |\n\t\t\tdel(.. | select(has("nodePort")).nodePort) |\n\t\t\tdel(.status) |\n\t\t\t(.. | select(tag == "!!str")) |= sub("monitoring-pmm3-31037", "NAME_SPACE") |\n\t\t\tdel(.spec.volumeClaimTemplates[].apiVersion) |\n\t\t\tdel(.spec.volumeClaimTemplates[].kind) |\n\t\t\tdel(.spec.ipFamilies) |\n\t\t\tdel(.spec.ipFamilyPolicy) |\n\t\t\t(.. | select(. == "extensions/v1beta1")) = "apps/v1" |\n\t\t\t(.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.3MGQm9xeoA ++ mktemp + local LAST_ERR=/tmp/tmp.jWxH2sAHfF + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml statefulset/monitoring-pmm3-cfg + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.3MGQm9xeoA + cat /tmp/tmp.jWxH2sAHfF + rm /tmp/tmp.3MGQm9xeoA /tmp/tmp.jWxH2sAHfF + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-cfg.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-cfg.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-cfg.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-cfg.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-cfg.yml /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-cfg.yml + log 'compare_kubectl: statefulset/monitoring-pmm3-cfg OK' + set +o xtrace [2026-04-14T11:31:58+0000] compare_kubectl: statefulset/monitoring-pmm3-cfg OK + compare_kubectl statefulset/monitoring-pmm3-mongos + local resource=statefulset/monitoring-pmm3-mongos + local postfix= + local skip_generation_check= + local expected_result=/mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-mongos.yml + local new_result=/tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-mongos.yml + '[' -n '' -a -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-mongos-oc.yml ']' + kubectl_bin get -o yaml statefulset/monitoring-pmm3-mongos + yq eval $'\n\t\t\tdel(.metadata.ownerReferences[].apiVersion) |\n\t\t\tdel(.metadata.managedFields) |\n\t\t\tdel(.. | select(has("creationTimestamp")).creationTimestamp) |\n\t\t\tdel(.. | select(has("namespace")).namespace) |\n\t\t\tdel(.. | select(has("uid")).uid) |\n\t\t\tdel(.metadata.resourceVersion) |\n\t\t\tdel(.spec.template.spec.containers[].env[] | select(.name == "NAMESPACE")) |\n\t\t\tdel(.metadata.selfLink) |\n\t\t\tdel(.metadata.annotations."cloud.google.com/neg") |\n\t\t\tdel(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") |\n\t\t\tdel(.. | select(has("image")).image) |\n\t\t\tdel(.. | select(has("clusterIP")).clusterIP) |\n\t\t\tdel(.. | select(has("clusterIPs")).clusterIPs) |\n\t\t\tdel(.. | select(has("dataSource")).dataSource) |\n\t\t\tdel(.. | select(has("procMount")).procMount) |\n\t\t\tdel(.. | select(has("storageClassName")).storageClassName) |\n\t\t\tdel(.. | select(has("finalizers")).finalizers) |\n\t\t\tdel(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") |\n\t\t\tdel(.. | select(has("volumeName")).volumeName) |\n\t\t\tdel(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") |\n\t\t\tdel(.spec.volumeMode) |\n\t\t\tdel(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") |\n\t\t\tdel(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") |\n\t\t\tdel(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") |\n\t\t\tdel(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") |\n\t\t\tdel(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) |\n\t\t\tdel(.. | select(has("healthCheckNodePort")).healthCheckNodePort) |\n\t\t\tdel(.. | select(has("nodePort")).nodePort) |\n\t\t\tdel(.status) |\n\t\t\t(.. | select(tag == "!!str")) |= sub("monitoring-pmm3-31037", "NAME_SPACE") |\n\t\t\tdel(.spec.volumeClaimTemplates[].apiVersion) |\n\t\t\tdel(.spec.volumeClaimTemplates[].kind) |\n\t\t\tdel(.spec.ipFamilies) |\n\t\t\tdel(.spec.ipFamilyPolicy) |\n\t\t\t(.. | select(. == "extensions/v1beta1")) = "apps/v1" |\n\t\t\t(.. | select(. == "batch/v1beta1")) = "batch/v1" ' - ++ mktemp + local LAST_OUT=/tmp/tmp.Nvywu5NuOg ++ mktemp + local LAST_ERR=/tmp/tmp.egWATs0N5H + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get -o yaml statefulset/monitoring-pmm3-mongos + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Nvywu5NuOg + cat /tmp/tmp.egWATs0N5H + rm /tmp/tmp.Nvywu5NuOg /tmp/tmp.egWATs0N5H + return 0 + yq -i eval 'del(.spec.persistentVolumeClaimRetentionPolicy)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-mongos.yml + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + yq -i eval 'del(.spec.internalTrafficPolicy)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-mongos.yml + yq -i eval 'del(.spec.allocateLoadBalancerNodePorts)' /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-mongos.yml + [[ /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-mongos.yml == */cronjob* ]] + '[' -n '' ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/e2e-tests/monitoring-pmm3/compare/statefulset_monitoring-pmm3-mongos.yml /tmp/tmp.KtaPi3a7ND/statefulset_monitoring-pmm3-mongos.yml + log 'compare_kubectl: statefulset/monitoring-pmm3-mongos OK' + set +o xtrace [2026-04-14T11:31:59+0000] compare_kubectl: statefulset/monitoring-pmm3-mongos OK + desc 'create new PMM token and add it to the secret' + set +o xtrace ----------------------------------------------------------------------------------- create new PMM token and add it to the secret ----------------------------------------------------------------------------------- ++ get_pmm_server_token operator_new ++ local key_name=operator_new ++ [[ -z operator_new ]] ++ local ADMIN_PASSWORD +++ kubectl get secret pmm-secret -o 'jsonpath={.data.PMM_ADMIN_PASSWORD}' +++ base64 --decode ++ ADMIN_PASSWORD=KyEo.de6w,DsTnRv ++ [[ -z KyEo.de6w,DsTnRv ]] ++ local create_response create_status_code create_json_response ++++ get_service_endpoint monitoring-service ++++ local service=monitoring-service +++++ kubectl_bin get service/monitoring-service -o json +++++ jq '.status.loadBalancer.ingress[].hostname' +++++ sed -e 's/^"//; s/"$//;' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.P2hPWVajcq ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.4lDQDpneFu +++++ local exit_status=0 +++++ local timeout=4 ++++++ seq 0 2 +++++ for i in $(seq 0 2) +++++ set +e +++++ kubectl get service/monitoring-service -o json +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 -a -n 1 ']' +++++ break +++++ cat /tmp/tmp.P2hPWVajcq +++++ cat /tmp/tmp.4lDQDpneFu +++++ rm /tmp/tmp.P2hPWVajcq /tmp/tmp.4lDQDpneFu +++++ return 0 ++++ local hostname=null ++++ '[' -n null -a null '!=' null ']' +++++ kubectl_bin get service/monitoring-service -o json ++++++ mktemp +++++ jq '.status.loadBalancer.ingress[].ip' +++++ sed -e 's/^"//; s/"$//;' +++++ local LAST_OUT=/tmp/tmp.207XCiXQhB ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.woy5YibpaV +++++ local exit_status=0 +++++ local timeout=4 ++++++ seq 0 2 +++++ for i in $(seq 0 2) +++++ set +e +++++ kubectl get service/monitoring-service -o json +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 -a -n 1 ']' +++++ break +++++ cat /tmp/tmp.207XCiXQhB +++++ cat /tmp/tmp.woy5YibpaV +++++ rm /tmp/tmp.207XCiXQhB /tmp/tmp.woy5YibpaV +++++ return 0 ++++ local ip=34.133.92.251 ++++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++++ echo 34.133.92.251 ++++ return +++ curl --insecure -s -X POST -H 'Content-Type: application/json' -H 'Accept: application/json' -d '{"name":"operator_new", "role":"Admin", "isDisabled":false}' --user admin:KyEo.de6w,DsTnRv https://34.133.92.251/graph/api/serviceaccounts -w '\n%{http_code}' ++ create_response=$'{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}\n201' +++ echo $'{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}\n201' +++ tail -n1 ++ create_status_code=201 +++ echo $'{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}\n201' +++ sed '$ d' ++ create_json_response='{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}' ++ [[ 201 -ne 201 ]] ++ local service_account_id +++ echo '{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":0,"avatarUrl":""}' +++ jq -r .id ++ service_account_id=3 ++ [[ -z 3 ]] ++ [[ 3 == null ]] ++ local token_response token_status_code token_json_response ++++ get_service_endpoint monitoring-service ++++ local service=monitoring-service +++++ kubectl_bin get service/monitoring-service -o json +++++ sed -e 's/^"//; s/"$//;' +++++ jq '.status.loadBalancer.ingress[].hostname' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.TGA9S15543 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.huLfCg2CJw +++++ local exit_status=0 +++++ local timeout=4 ++++++ seq 0 2 +++++ for i in $(seq 0 2) +++++ set +e +++++ kubectl get service/monitoring-service -o json +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 -a -n 1 ']' +++++ break +++++ cat /tmp/tmp.TGA9S15543 +++++ cat /tmp/tmp.huLfCg2CJw +++++ rm /tmp/tmp.TGA9S15543 /tmp/tmp.huLfCg2CJw +++++ return 0 ++++ local hostname=null ++++ '[' -n null -a null '!=' null ']' +++++ kubectl_bin get service/monitoring-service -o json +++++ jq '.status.loadBalancer.ingress[].ip' +++++ sed -e 's/^"//; s/"$//;' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.L8yaXrlzIX ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.OdpzNvEsSp +++++ local exit_status=0 +++++ local timeout=4 ++++++ seq 0 2 +++++ for i in $(seq 0 2) +++++ set +e +++++ kubectl get service/monitoring-service -o json +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 -a -n 1 ']' +++++ break +++++ cat /tmp/tmp.L8yaXrlzIX +++++ cat /tmp/tmp.OdpzNvEsSp +++++ rm /tmp/tmp.L8yaXrlzIX /tmp/tmp.OdpzNvEsSp +++++ return 0 ++++ local ip=34.133.92.251 ++++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++++ echo 34.133.92.251 ++++ return +++ curl --insecure -s -X POST -H 'Content-Type: application/json' -d '{"name":"operator_new"}' --user admin:KyEo.de6w,DsTnRv https://34.133.92.251/graph/api/serviceaccounts/3/tokens -w '\n%{http_code}' ++ token_response=$'{"id":2,"name":"operator_new","key":"glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c"}\n200' +++ echo $'{"id":2,"name":"operator_new","key":"glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c"}\n200' +++ tail -n1 ++ token_status_code=200 +++ echo $'{"id":2,"name":"operator_new","key":"glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c"}\n200' +++ sed '$ d' ++ token_json_response='{"id":2,"name":"operator_new","key":"glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c"}' ++ [[ 200 -ne 200 ]] ++ echo '{"id":2,"name":"operator_new","key":"glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c"}' ++ jq -r .key + NEW_TOKEN=glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + kubectl_bin patch secret some-users --type merge --patch '{"stringData": {"PMM_SERVER_TOKEN": "glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c"}}' ++ mktemp + local LAST_OUT=/tmp/tmp.OGxDMVq8wP ++ mktemp + local LAST_ERR=/tmp/tmp.ePgGldMYOB + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch secret some-users --type merge --patch '{"stringData": {"PMM_SERVER_TOKEN": "glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c"}}' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.OGxDMVq8wP secret/some-users patched + cat /tmp/tmp.ePgGldMYOB + rm /tmp/tmp.OGxDMVq8wP /tmp/tmp.ePgGldMYOB + return 0 + desc 'delete old PMM token' + set +o xtrace ----------------------------------------------------------------------------------- delete old PMM token ----------------------------------------------------------------------------------- + delete_pmm_server_token operator + local key_name=operator + [[ -z operator ]] + local ADMIN_PASSWORD ++ kubectl get secret pmm-secret -o 'jsonpath={.data.PMM_ADMIN_PASSWORD}' ++ base64 --decode + ADMIN_PASSWORD=KyEo.de6w,DsTnRv + [[ -z KyEo.de6w,DsTnRv ]] + local user_credentials=admin:KyEo.de6w,DsTnRv + local service_accounts_response service_accounts_status +++ get_service_endpoint monitoring-service +++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o json ++++ jq '.status.loadBalancer.ingress[].hostname' ++++ sed -e 's/^"//; s/"$//;' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.Qmg1rbPWYA +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.GL5CFhVAtJ ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o json ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.Qmg1rbPWYA ++++ cat /tmp/tmp.GL5CFhVAtJ ++++ rm /tmp/tmp.Qmg1rbPWYA /tmp/tmp.GL5CFhVAtJ ++++ return 0 +++ local hostname=null +++ '[' -n null -a null '!=' null ']' ++++ kubectl_bin get service/monitoring-service -o json ++++ jq '.status.loadBalancer.ingress[].ip' ++++ sed -e 's/^"//; s/"$//;' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.wFxRZuDQ97 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.fW4qItmG6Z ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o json ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.wFxRZuDQ97 ++++ cat /tmp/tmp.fW4qItmG6Z ++++ rm /tmp/tmp.wFxRZuDQ97 /tmp/tmp.fW4qItmG6Z ++++ return 0 +++ local ip=34.133.92.251 +++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' +++ echo 34.133.92.251 +++ return ++ curl --insecure -s -X GET --user admin:KyEo.de6w,DsTnRv https://34.133.92.251/graph/api/serviceaccounts/search -w '\n%{http_code}' + service_accounts_response=$'{"totalCount":2,"serviceAccounts":[{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/f741e7abdd8def4ed30a37f602429df3"},{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/6d1b2db67a8f3e06db4ae076406fece5"}],"page":1,"perPage":1000}\n200' ++ echo $'{"totalCount":2,"serviceAccounts":[{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/f741e7abdd8def4ed30a37f602429df3"},{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/6d1b2db67a8f3e06db4ae076406fece5"}],"page":1,"perPage":1000}\n200' ++ tail -n1 + service_accounts_status=200 ++ echo $'{"totalCount":2,"serviceAccounts":[{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/f741e7abdd8def4ed30a37f602429df3"},{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/6d1b2db67a8f3e06db4ae076406fece5"}],"page":1,"perPage":1000}\n200' ++ sed '$ d' + service_accounts_json='{"totalCount":2,"serviceAccounts":[{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/f741e7abdd8def4ed30a37f602429df3"},{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/6d1b2db67a8f3e06db4ae076406fece5"}],"page":1,"perPage":1000}' + [[ 200 -ne 200 ]] + local service_account_id ++ echo '{"totalCount":2,"serviceAccounts":[{"id":2,"uid":"ffj261311dr7kc","name":"operator","login":"sa-1-operator","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/f741e7abdd8def4ed30a37f602429df3"},{"id":3,"uid":"cfj274x9cn37kb","name":"operator_new","login":"sa-1-operator_new","orgId":1,"isDisabled":false,"role":"Admin","tokens":1,"avatarUrl":"/graph/avatar/6d1b2db67a8f3e06db4ae076406fece5"}],"page":1,"perPage":1000}' ++ jq -r '.serviceAccounts[] | select(.name == "operator").id' + service_account_id=2 + [[ -z 2 ]] + [[ 2 == null ]] + local tokens_response tokens_status tokens_json +++ get_service_endpoint monitoring-service +++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o json ++++ jq '.status.loadBalancer.ingress[].hostname' ++++ sed -e 's/^"//; s/"$//;' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.4cxGutHdbu +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.qeV41ujm0X ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o json ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.4cxGutHdbu ++++ cat /tmp/tmp.qeV41ujm0X ++++ rm /tmp/tmp.4cxGutHdbu /tmp/tmp.qeV41ujm0X ++++ return 0 +++ local hostname=null +++ '[' -n null -a null '!=' null ']' ++++ kubectl_bin get service/monitoring-service -o json ++++ jq '.status.loadBalancer.ingress[].ip' ++++ sed -e 's/^"//; s/"$//;' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.v7p6JFGYiP +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.eijZD0Y3n2 ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o json ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.v7p6JFGYiP ++++ cat /tmp/tmp.eijZD0Y3n2 ++++ rm /tmp/tmp.v7p6JFGYiP /tmp/tmp.eijZD0Y3n2 ++++ return 0 +++ local ip=34.133.92.251 +++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' +++ echo 34.133.92.251 +++ return ++ curl --insecure -s -X GET --user admin:KyEo.de6w,DsTnRv https://34.133.92.251/graph/api/serviceaccounts/2/tokens -w '\n%{http_code}' + tokens_response=$'[{"id":1,"name":"operator","created":"2026-04-14T11:19:40Z","lastUsedAt":"2026-04-14T11:27:57Z","expiration":null,"secondsUntilExpiration":0,"hasExpired":false,"isRevoked":false}]\n200' ++ echo $'[{"id":1,"name":"operator","created":"2026-04-14T11:19:40Z","lastUsedAt":"2026-04-14T11:27:57Z","expiration":null,"secondsUntilExpiration":0,"hasExpired":false,"isRevoked":false}]\n200' ++ tail -n1 + tokens_status=200 ++ echo $'[{"id":1,"name":"operator","created":"2026-04-14T11:19:40Z","lastUsedAt":"2026-04-14T11:27:57Z","expiration":null,"secondsUntilExpiration":0,"hasExpired":false,"isRevoked":false}]\n200' ++ sed '$ d' + tokens_json='[{"id":1,"name":"operator","created":"2026-04-14T11:19:40Z","lastUsedAt":"2026-04-14T11:27:57Z","expiration":null,"secondsUntilExpiration":0,"hasExpired":false,"isRevoked":false}]' + [[ 200 -ne 200 ]] + local token_id ++ echo '[{"id":1,"name":"operator","created":"2026-04-14T11:19:40Z","lastUsedAt":"2026-04-14T11:27:57Z","expiration":null,"secondsUntilExpiration":0,"hasExpired":false,"isRevoked":false}]' ++ jq -r '.[] | select(.name == "operator").id' + token_id=1 + [[ -z 1 ]] + [[ 1 == null ]] + local delete_response delete_status +++ get_service_endpoint monitoring-service +++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o json ++++ jq '.status.loadBalancer.ingress[].hostname' ++++ sed -e 's/^"//; s/"$//;' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.hLWzJTz0A0 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.Y2QVF4kneW ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o json ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.hLWzJTz0A0 ++++ cat /tmp/tmp.Y2QVF4kneW ++++ rm /tmp/tmp.hLWzJTz0A0 /tmp/tmp.Y2QVF4kneW ++++ return 0 +++ local hostname=null +++ '[' -n null -a null '!=' null ']' ++++ jq '.status.loadBalancer.ingress[].ip' ++++ kubectl_bin get service/monitoring-service -o json ++++ sed -e 's/^"//; s/"$//;' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.1pMHVF3AfP +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.lIKc0OxEAe ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o json ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.1pMHVF3AfP ++++ cat /tmp/tmp.lIKc0OxEAe ++++ rm /tmp/tmp.1pMHVF3AfP /tmp/tmp.lIKc0OxEAe ++++ return 0 +++ local ip=34.133.92.251 +++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' +++ echo 34.133.92.251 +++ return ++ curl --insecure -s -X DELETE --user admin:KyEo.de6w,DsTnRv https://34.133.92.251/graph/api/serviceaccounts/2/tokens/1 -w '\n%{http_code}' + delete_response=$'{"message":"Service account token deleted"}\n200' ++ echo $'{"message":"Service account token deleted"}\n200' ++ tail -n1 + delete_status=200 + [[ 200 -ne 200 ]] + desc 'check for authentication errors in PMM client logs' + set +o xtrace ----------------------------------------------------------------------------------- check for authentication errors in PMM client logs ----------------------------------------------------------------------------------- ++ kubectl get pods --selector=app.kubernetes.io/replset=rs0 -o 'jsonpath={.items[*].metadata.name}' + pods='monitoring-pmm3-rs0-0 monitoring-pmm3-rs0-1 monitoring-pmm3-rs0-2' + for pod in $pods + for i in {1..3} + kubectl logs monitoring-pmm3-rs0-0 pmm-client + grep -q 'Invalid username or password' + sleep 2 + for i in {1..3} + kubectl logs monitoring-pmm3-rs0-0 pmm-client + grep -q 'Invalid username or password' + sleep 2 + for i in {1..3} + kubectl logs monitoring-pmm3-rs0-0 pmm-client + grep -q 'Invalid username or password' + sleep 2 + for pod in $pods + for i in {1..3} + kubectl logs monitoring-pmm3-rs0-1 pmm-client + grep -q 'Invalid username or password' + sleep 2 + for i in {1..3} + kubectl logs monitoring-pmm3-rs0-1 pmm-client + grep -q 'Invalid username or password' + sleep 2 + for i in {1..3} + kubectl logs monitoring-pmm3-rs0-1 pmm-client + grep -q 'Invalid username or password' + sleep 2 + for pod in $pods + for i in {1..3} + kubectl logs monitoring-pmm3-rs0-2 pmm-client + grep -q 'Invalid username or password' + sleep 2 + for i in {1..3} + kubectl logs monitoring-pmm3-rs0-2 pmm-client + grep -q 'Invalid username or password' + sleep 2 + for i in {1..3} + kubectl logs monitoring-pmm3-rs0-2 pmm-client + grep -q 'Invalid username or password' + sleep 2 + desc 'check mongod metrics' + set +o xtrace ----------------------------------------------------------------------------------- check mongod metrics ----------------------------------------------------------------------------------- + get_metric_values node_boot_time_seconds monitoring-pmm3-31037-monitoring-pmm3-rs0-1 glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + local metric=node_boot_time_seconds + local instance=monitoring-pmm3-31037-monitoring-pmm3-rs0-1 + local token=glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ++ /usr/sbin/date -u +%s -d '-1 minute' + local start=1776166305 ++ /usr/sbin/date -u +%s + local end=1776166365 ++ get_service_endpoint monitoring-service ++ local service=monitoring-service +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].hostname' +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.tUsqHu9Zng ++++ mktemp +++ local LAST_ERR=/tmp/tmp.r23b7qurzw +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.tUsqHu9Zng +++ cat /tmp/tmp.r23b7qurzw +++ rm /tmp/tmp.tUsqHu9Zng /tmp/tmp.r23b7qurzw +++ return 0 ++ local hostname=null ++ '[' -n null -a null '!=' null ']' +++ jq '.status.loadBalancer.ingress[].ip' +++ kubectl_bin get service/monitoring-service -o json +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.uJWK08BcGS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.PI6q0smYE5 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.uJWK08BcGS +++ cat /tmp/tmp.PI6q0smYE5 +++ rm /tmp/tmp.uJWK08BcGS /tmp/tmp.PI6q0smYE5 +++ return 0 ++ local ip=34.133.92.251 ++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++ echo 34.133.92.251 ++ return + local endpoint=34.133.92.251 + '[' -z node_boot_time_seconds ']' + '[' -z glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ']' + local wait_count=30 + local retry=0 ++ curl -s -k -H 'Authorization: Bearer glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c' 'https://34.133.92.251/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28node_boot_time_seconds%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-rs0-1%22%7d%20or%20node_boot_time_seconds%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-rs0-1%22%7D%29&start=1776166305&end=1776166365&step=60' ++ jq '.data.result[0].values[][1]' ++ grep '^"[0-9]' + [[ -n "1776165681" "1776165681" ]] + get_metric_values mongodb_connections monitoring-pmm3-31037-monitoring-pmm3-rs0-1 glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + local metric=mongodb_connections + local instance=monitoring-pmm3-31037-monitoring-pmm3-rs0-1 + local token=glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ++ /usr/sbin/date -u +%s -d '-1 minute' + local start=1776166307 ++ /usr/sbin/date -u +%s + local end=1776166367 ++ get_service_endpoint monitoring-service ++ local service=monitoring-service +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].hostname' +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Hu8a9UFqrI ++++ mktemp +++ local LAST_ERR=/tmp/tmp.a83jiQMsyM +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.Hu8a9UFqrI +++ cat /tmp/tmp.a83jiQMsyM +++ rm /tmp/tmp.Hu8a9UFqrI /tmp/tmp.a83jiQMsyM +++ return 0 ++ local hostname=null ++ '[' -n null -a null '!=' null ']' +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].ip' ++++ mktemp +++ sed -e 's/^"//; s/"$//;' +++ local LAST_OUT=/tmp/tmp.QHK9qmwySu ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5gCA4SPNkE +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.QHK9qmwySu +++ cat /tmp/tmp.5gCA4SPNkE +++ rm /tmp/tmp.QHK9qmwySu /tmp/tmp.5gCA4SPNkE +++ return 0 ++ local ip=34.133.92.251 ++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++ echo 34.133.92.251 ++ return + local endpoint=34.133.92.251 + '[' -z mongodb_connections ']' + '[' -z glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ']' + local wait_count=30 + local retry=0 ++ curl -s -k -H 'Authorization: Bearer glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c' 'https://34.133.92.251/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28mongodb_connections%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-rs0-1%22%7d%20or%20mongodb_connections%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-rs0-1%22%7D%29&start=1776166307&end=1776166367&step=60' ++ jq '.data.result[0].values[][1]' ++ grep '^"[0-9]' + [[ -n "0" "0" ]] + desc 'check mongo config metrics' + set +o xtrace ----------------------------------------------------------------------------------- check mongo config metrics ----------------------------------------------------------------------------------- + get_metric_values node_boot_time_seconds monitoring-pmm3-31037-monitoring-pmm3-cfg-1 glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + local metric=node_boot_time_seconds + local instance=monitoring-pmm3-31037-monitoring-pmm3-cfg-1 + local token=glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ++ /usr/sbin/date -u +%s -d '-1 minute' + local start=1776166309 ++ /usr/sbin/date -u +%s + local end=1776166369 ++ get_service_endpoint monitoring-service ++ local service=monitoring-service +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].hostname' +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.3feT8waFo5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.kEkyiUXPul +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.3feT8waFo5 +++ cat /tmp/tmp.kEkyiUXPul +++ rm /tmp/tmp.3feT8waFo5 /tmp/tmp.kEkyiUXPul +++ return 0 ++ local hostname=null ++ '[' -n null -a null '!=' null ']' +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].ip' ++++ mktemp +++ sed -e 's/^"//; s/"$//;' +++ local LAST_OUT=/tmp/tmp.FA3Le9LH6r ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Y9b7RP0zuR +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.FA3Le9LH6r +++ cat /tmp/tmp.Y9b7RP0zuR +++ rm /tmp/tmp.FA3Le9LH6r /tmp/tmp.Y9b7RP0zuR +++ return 0 ++ local ip=34.133.92.251 ++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++ echo 34.133.92.251 ++ return + local endpoint=34.133.92.251 + '[' -z node_boot_time_seconds ']' + '[' -z glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ']' + local wait_count=30 + local retry=0 ++ curl -s -k -H 'Authorization: Bearer glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c' 'https://34.133.92.251/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28node_boot_time_seconds%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-cfg-1%22%7d%20or%20node_boot_time_seconds%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-cfg-1%22%7D%29&start=1776166309&end=1776166369&step=60' ++ jq '.data.result[0].values[][1]' ++ grep '^"[0-9]' + [[ -n "1776165244" "1776165244" ]] + get_metric_values mongodb_connections monitoring-pmm3-31037-monitoring-pmm3-cfg-1 glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + local metric=mongodb_connections + local instance=monitoring-pmm3-31037-monitoring-pmm3-cfg-1 + local token=glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ++ /usr/sbin/date -u +%s -d '-1 minute' + local start=1776166311 ++ /usr/sbin/date -u +%s + local end=1776166371 ++ get_service_endpoint monitoring-service ++ local service=monitoring-service +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].hostname' ++++ mktemp +++ sed -e 's/^"//; s/"$//;' +++ local LAST_OUT=/tmp/tmp.7LhAhzfBRC ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ckdJjIksms +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.7LhAhzfBRC +++ cat /tmp/tmp.ckdJjIksms +++ rm /tmp/tmp.7LhAhzfBRC /tmp/tmp.ckdJjIksms +++ return 0 ++ local hostname=null ++ '[' -n null -a null '!=' null ']' +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].ip' +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.nGN4G1rPzZ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.0fTKRvlPy6 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.nGN4G1rPzZ +++ cat /tmp/tmp.0fTKRvlPy6 +++ rm /tmp/tmp.nGN4G1rPzZ /tmp/tmp.0fTKRvlPy6 +++ return 0 ++ local ip=34.133.92.251 ++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++ echo 34.133.92.251 ++ return + local endpoint=34.133.92.251 + '[' -z mongodb_connections ']' + '[' -z glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ']' + local wait_count=30 + local retry=0 ++ curl -s -k -H 'Authorization: Bearer glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c' 'https://34.133.92.251/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28mongodb_connections%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-cfg-1%22%7d%20or%20mongodb_connections%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-cfg-1%22%7D%29&start=1776166311&end=1776166371&step=60' ++ jq '.data.result[0].values[][1]' ++ grep '^"[0-9]' + [[ -n "0" "0" ]] + desc 'check mongos metrics' + set +o xtrace ----------------------------------------------------------------------------------- check mongos metrics ----------------------------------------------------------------------------------- ++ kubectl get pod -l app.kubernetes.io/component=mongos -o 'jsonpath={.items[0].metadata.name}' + MONGOS_POD_NAME=monitoring-pmm3-mongos-0 + get_metric_values node_boot_time_seconds monitoring-pmm3-31037-monitoring-pmm3-mongos-0 glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + local metric=node_boot_time_seconds + local instance=monitoring-pmm3-31037-monitoring-pmm3-mongos-0 + local token=glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ++ /usr/sbin/date -u +%s -d '-1 minute' + local start=1776166315 ++ /usr/sbin/date -u +%s + local end=1776166375 ++ get_service_endpoint monitoring-service ++ local service=monitoring-service +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].hostname' ++++ mktemp +++ sed -e 's/^"//; s/"$//;' +++ local LAST_OUT=/tmp/tmp.xPk70wGY6A ++++ mktemp +++ local LAST_ERR=/tmp/tmp.nLgjIBqYtN +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.xPk70wGY6A +++ cat /tmp/tmp.nLgjIBqYtN +++ rm /tmp/tmp.xPk70wGY6A /tmp/tmp.nLgjIBqYtN +++ return 0 ++ local hostname=null ++ '[' -n null -a null '!=' null ']' +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].ip' +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.xD8TdXlWPq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.7Y1uqp84vP +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.xD8TdXlWPq +++ cat /tmp/tmp.7Y1uqp84vP +++ rm /tmp/tmp.xD8TdXlWPq /tmp/tmp.7Y1uqp84vP +++ return 0 ++ local ip=34.133.92.251 ++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++ echo 34.133.92.251 ++ return + local endpoint=34.133.92.251 + '[' -z node_boot_time_seconds ']' + '[' -z glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c ']' + local wait_count=30 + local retry=0 ++ curl -s -k -H 'Authorization: Bearer glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c' 'https://34.133.92.251/graph/api/datasources/proxy/1/api/v1/query_range?query=min%28node_boot_time_seconds%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-mongos-0%22%7d%20or%20node_boot_time_seconds%7Bnode_name%3D%7E%22monitoring-pmm3-31037-monitoring-pmm3-mongos-0%22%7D%29&start=1776166315&end=1776166375&step=60' ++ jq '.data.result[0].values[][1]' ++ grep '^"[0-9]' + [[ -n "1776166150" "1776166150" ]] + sleep 90 + desc 'check QAN data' + set +o xtrace ----------------------------------------------------------------------------------- check QAN data ----------------------------------------------------------------------------------- + get_qan_values mongodb dev-mongod glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + local service_type=mongodb + local environment=dev-mongod + local token=glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + local start + local end + local endpoint ++ /usr/sbin/date -u -d '-12 hour' +%Y-%m-%dT%H:%M:%S%:z + start=2026-04-13T23:34:28+00:00 ++ /usr/sbin/date -u +%Y-%m-%dT%H:%M:%S%:z + end=2026-04-14T11:34:28+00:00 ++ get_service_endpoint monitoring-service ++ local service=monitoring-service +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].hostname' +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.S8G9wEYSz0 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.eELziLuuo9 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.S8G9wEYSz0 +++ cat /tmp/tmp.eELziLuuo9 +++ rm /tmp/tmp.S8G9wEYSz0 /tmp/tmp.eELziLuuo9 +++ return 0 ++ local hostname=null ++ '[' -n null -a null '!=' null ']' +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].ip' +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rimuFDAZpq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.BaYs9y3xHg +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.rimuFDAZpq +++ cat /tmp/tmp.BaYs9y3xHg +++ rm /tmp/tmp.rimuFDAZpq /tmp/tmp.BaYs9y3xHg +++ return 0 ++ local ip=34.133.92.251 ++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++ echo 34.133.92.251 ++ return + endpoint=34.133.92.251 + cat + local response + local wait_count=30 + local retry=0 ++ curl -s -k -H 'Authorization: Bearer glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c' -XPOST -d @payload.json https://34.133.92.251/v1/qan/metrics:getReport ++ jq '.rows[].fingerprint' + [[ -n "TOTAL" "db.runCommand({\"$configTime\":\"?\",\"$topologyTime\":\"?\",\"batchSize\":\"?\",\"collection\":\"?\",\"getMore\":\"?\",\"maxTimeMS\":\"?\",\"term\":\"?\"})" "db.oplog.rs.find({}).sort({\"$natural\":1}).limit(?)" "db.runCommand({\"$audit\":{\"$impersonatedRoles\":[{\"db\":\"?\",\"role\":\"?\"},{\"db\":\"?\",\"role\":\"?\"},{\"db\":\"?\",\"role\":\"?\"},{\"db\":\"?\",\"role\":\"?\"}],\"$impersonatedUser\":{\"db\":\"?\",\"user\":\"?\"}},\"$client\":{\"application\":{\"name\":\"?\"},\"driver\":{\"name\":\"?\",\"version\":\"?\"},\"mongos\":{\"client\":\"?\",\"host\":\"?\",\"version\":\"?\"},\"os\":{\"architecture\":\"?\",\"type\":\"?\"},\"platform\":\"?\"},\"$configTime\":\"?\",\"$topologyTime\":\"?\",\"dbStats\":\"?\",\"mayBypassWriteBlocking\":\"?\",\"readConcern\":{\"level\":\"?\",\"provenance\":\"?\"},\"scale\":\"?\"})" "db.oplog.rs.find({}).sort({\"$natural\":-1}).limit(?)" "db.system.sessions.update({\"_id\":{\"id\":\"?\",\"uid\":\"?\"}}, [{\"$set\":{\"lastUse\":\"?\"}}], {\"upsert\":true})" "db.system.sessions.deleteOne({\"_id\":{\"id\":\"?\",\"uid\":\"?\"}})" "db.runCommand({\"$audit\":{\"$impersonatedRoles\":[{\"db\":\"?\",\"role\":\"?\"},{\"db\":\"?\",\"role\":\"?\"},{\"db\":\"?\",\"role\":\"?\"},{\"db\":\"?\",\"role\":\"?\"}],\"$impersonatedUser\":{\"db\":\"?\",\"user\":\"?\"}},\"$client\":{\"application\":{\"name\":\"?\"},\"driver\":{\"name\":\"?\",\"version\":\"?\"},\"mongos\":{\"client\":\"?\",\"host\":\"?\",\"version\":\"?\"},\"os\":{\"architecture\":\"?\",\"type\":\"?\"},\"platform\":\"?\"},\"$configTime\":\"?\",\"$topologyTime\":\"?\",\"dbStats\":\"?\",\"mayBypassWriteBlocking\":\"?\",\"readConcern\":{\"level\":\"?\",\"provenance\":\"?\"},\"scale\":\"?\"})" "db.system.version.find({\"_id\":\"?\"}).limit(?)" "db.system.sessions.update({\"_id\":{\"id\":\"?\",\"uid\":\"?\"}}, [{\"$set\":{\"lastUse\":\"?\"}},{\"$set\":{\"user\":{\"name\":\"?\"}}}], {\"upsert\":true})" ]] + rm -f payload.json + get_qan_values mongodb dev-mongos glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + local service_type=mongodb + local environment=dev-mongos + local token=glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + local start + local end + local endpoint ++ /usr/sbin/date -u -d '-12 hour' +%Y-%m-%dT%H:%M:%S%:z + start=2026-04-13T23:34:31+00:00 ++ /usr/sbin/date -u +%Y-%m-%dT%H:%M:%S%:z + end=2026-04-14T11:34:31+00:00 ++ get_service_endpoint monitoring-service ++ local service=monitoring-service +++ kubectl_bin get service/monitoring-service -o json +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ jq '.status.loadBalancer.ingress[].hostname' +++ local LAST_OUT=/tmp/tmp.MMWpHyydRB ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bEmbbV2zAf +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.MMWpHyydRB +++ cat /tmp/tmp.bEmbbV2zAf +++ rm /tmp/tmp.MMWpHyydRB /tmp/tmp.bEmbbV2zAf +++ return 0 ++ local hostname=null ++ '[' -n null -a null '!=' null ']' +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].ip' +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OwU5BtaQaE ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jw42Mgw6tf +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.OwU5BtaQaE +++ cat /tmp/tmp.jw42Mgw6tf +++ rm /tmp/tmp.OwU5BtaQaE /tmp/tmp.jw42Mgw6tf +++ return 0 ++ local ip=34.133.92.251 ++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++ echo 34.133.92.251 ++ return + endpoint=34.133.92.251 + cat + local response + local wait_count=30 + local retry=0 ++ curl -s -k -H 'Authorization: Bearer glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c' -XPOST -d @payload.json https://34.133.92.251/v1/qan/metrics:getReport ++ jq '.rows[].fingerprint' + [[ -n null ]] + rm -f payload.json + desc 'verify that the custom cluster name is configured' + set +o xtrace ----------------------------------------------------------------------------------- verify that the custom cluster name is configured ----------------------------------------------------------------------------------- + verify_custom_cluster_name super-custom glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c monitoring-pmm3-31037-monitoring-pmm3-mongos-0 monitoring-pmm3-31037-monitoring-pmm3-cfg-0 monitoring-pmm3-31037-monitoring-pmm3-rs0-0 + local expected_cluster=super-custom + local token=glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c + shift 2 + service_names=('monitoring-pmm3-31037-monitoring-pmm3-mongos-0' 'monitoring-pmm3-31037-monitoring-pmm3-cfg-0' 'monitoring-pmm3-31037-monitoring-pmm3-rs0-0') + local service_names + local endpoint ++ get_service_endpoint monitoring-service ++ local service=monitoring-service +++ kubectl_bin get service/monitoring-service -o json ++++ mktemp +++ jq '.status.loadBalancer.ingress[].hostname' +++ sed -e 's/^"//; s/"$//;' +++ local LAST_OUT=/tmp/tmp.Pl0uuBztac ++++ mktemp +++ local LAST_ERR=/tmp/tmp.8HLCvUA7i5 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.Pl0uuBztac +++ cat /tmp/tmp.8HLCvUA7i5 +++ rm /tmp/tmp.Pl0uuBztac /tmp/tmp.8HLCvUA7i5 +++ return 0 ++ local hostname=null ++ '[' -n null -a null '!=' null ']' +++ kubectl_bin get service/monitoring-service -o json +++ jq '.status.loadBalancer.ingress[].ip' +++ sed -e 's/^"//; s/"$//;' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.wLHJKO9FHR ++++ mktemp +++ local LAST_ERR=/tmp/tmp.poEPW4XdxF +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get service/monitoring-service -o json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.wLHJKO9FHR +++ cat /tmp/tmp.poEPW4XdxF +++ rm /tmp/tmp.wLHJKO9FHR /tmp/tmp.poEPW4XdxF +++ return 0 ++ local ip=34.133.92.251 ++ '[' -n 34.133.92.251 -a 34.133.92.251 '!=' null ']' ++ echo 34.133.92.251 ++ return + endpoint=34.133.92.251 + local response ++ curl -s -k -H 'Authorization: Bearer glsa_UBj9vIpeX0X1MrwWBzWwitAj2qtgu8OI_174d578c' 'https://34.133.92.251/v1/inventory/services?service_type=SERVICE_TYPE_MONGODB_SERVICE' + response=$'{\n "mysql": [],\n "mongodb": [\n {\n "service_id": "0b632a0d-0120-43c3-a1a2-78ef2cc0b909",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-0",\n "node_id": "9bd0200e-5913-40eb-b73d-d7c2dabf1c9b",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "195d0871-681a-4ecc-b523-926f9083b1a2",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-2",\n "node_id": "c4b64abb-c01b-4f21-8a28-1ad1423b1945",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "27df9434-425e-4676-a6a9-4b986991ac24",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-1",\n "node_id": "9b02add2-b915-4fcd-9ad8-9b08f8f8d701",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "2a3e5541-9f3d-417e-88bc-27bf19a71693",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-2",\n "node_id": "ced1678f-8786-43fe-bc96-cdbe81b1330b",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "51bac060-84d6-4d50-9a5b-150dd0d1e8c0",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-0",\n "node_id": "a3c1225e-8867-45ec-ab09-a8c24fee8f52",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "aabb3a50-7bf3-423d-ba58-8428081cc65b",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-1",\n "node_id": "323e006f-a2dd-4255-bb74-58c2618e1313",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "ca42a63d-f8d8-4b0f-b858-c5ef06cc5569",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-2",\n "node_id": "e21d39cb-99de-40bb-8f0a-2b342df379df",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "d1b1fe5f-ee05-4b6f-82fd-d0bdfcb0adeb",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-0",\n "node_id": "942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "d984828e-9147-4047-9a7f-0d9f0ec608ab",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-1",\n "node_id": "2f443660-febf-4fe1-9d79-919e88cfe5d6",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n }\n ],\n "postgresql": [],\n "proxysql": [],\n "haproxy": [],\n "external": [],\n "valkey": []\n}' + local verified=0 + for service_name in "${service_names[@]}" + local actual_cluster ++ jq -r --arg name monitoring-pmm3-31037-monitoring-pmm3-mongos-0 $'\n\t\t\t.mongodb[] | select(.service_name == $name) | .cluster\n\t\t' ++ echo $'{\n "mysql": [],\n "mongodb": [\n {\n "service_id": "0b632a0d-0120-43c3-a1a2-78ef2cc0b909",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-0",\n "node_id": "9bd0200e-5913-40eb-b73d-d7c2dabf1c9b",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "195d0871-681a-4ecc-b523-926f9083b1a2",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-2",\n "node_id": "c4b64abb-c01b-4f21-8a28-1ad1423b1945",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "27df9434-425e-4676-a6a9-4b986991ac24",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-1",\n "node_id": "9b02add2-b915-4fcd-9ad8-9b08f8f8d701",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "2a3e5541-9f3d-417e-88bc-27bf19a71693",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-2",\n "node_id": "ced1678f-8786-43fe-bc96-cdbe81b1330b",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "51bac060-84d6-4d50-9a5b-150dd0d1e8c0",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-0",\n "node_id": "a3c1225e-8867-45ec-ab09-a8c24fee8f52",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "aabb3a50-7bf3-423d-ba58-8428081cc65b",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-1",\n "node_id": "323e006f-a2dd-4255-bb74-58c2618e1313",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "ca42a63d-f8d8-4b0f-b858-c5ef06cc5569",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-2",\n "node_id": "e21d39cb-99de-40bb-8f0a-2b342df379df",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "d1b1fe5f-ee05-4b6f-82fd-d0bdfcb0adeb",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-0",\n "node_id": "942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "d984828e-9147-4047-9a7f-0d9f0ec608ab",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-1",\n "node_id": "2f443660-febf-4fe1-9d79-919e88cfe5d6",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n }\n ],\n "postgresql": [],\n "proxysql": [],\n "haproxy": [],\n "external": [],\n "valkey": []\n}' + actual_cluster=super-custom + [[ -z super-custom ]] + [[ super-custom == null ]] + [[ super-custom != super\-custom ]] + for service_name in "${service_names[@]}" + local actual_cluster ++ jq -r --arg name monitoring-pmm3-31037-monitoring-pmm3-cfg-0 $'\n\t\t\t.mongodb[] | select(.service_name == $name) | .cluster\n\t\t' ++ echo $'{\n "mysql": [],\n "mongodb": [\n {\n "service_id": "0b632a0d-0120-43c3-a1a2-78ef2cc0b909",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-0",\n "node_id": "9bd0200e-5913-40eb-b73d-d7c2dabf1c9b",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "195d0871-681a-4ecc-b523-926f9083b1a2",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-2",\n "node_id": "c4b64abb-c01b-4f21-8a28-1ad1423b1945",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "27df9434-425e-4676-a6a9-4b986991ac24",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-1",\n "node_id": "9b02add2-b915-4fcd-9ad8-9b08f8f8d701",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "2a3e5541-9f3d-417e-88bc-27bf19a71693",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-2",\n "node_id": "ced1678f-8786-43fe-bc96-cdbe81b1330b",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "51bac060-84d6-4d50-9a5b-150dd0d1e8c0",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-0",\n "node_id": "a3c1225e-8867-45ec-ab09-a8c24fee8f52",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "aabb3a50-7bf3-423d-ba58-8428081cc65b",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-1",\n "node_id": "323e006f-a2dd-4255-bb74-58c2618e1313",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "ca42a63d-f8d8-4b0f-b858-c5ef06cc5569",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-2",\n "node_id": "e21d39cb-99de-40bb-8f0a-2b342df379df",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "d1b1fe5f-ee05-4b6f-82fd-d0bdfcb0adeb",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-0",\n "node_id": "942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "d984828e-9147-4047-9a7f-0d9f0ec608ab",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-1",\n "node_id": "2f443660-febf-4fe1-9d79-919e88cfe5d6",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n }\n ],\n "postgresql": [],\n "proxysql": [],\n "haproxy": [],\n "external": [],\n "valkey": []\n}' + actual_cluster=super-custom + [[ -z super-custom ]] + [[ super-custom == null ]] + [[ super-custom != super\-custom ]] + for service_name in "${service_names[@]}" + local actual_cluster ++ echo $'{\n "mysql": [],\n "mongodb": [\n {\n "service_id": "0b632a0d-0120-43c3-a1a2-78ef2cc0b909",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-0",\n "node_id": "9bd0200e-5913-40eb-b73d-d7c2dabf1c9b",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "195d0871-681a-4ecc-b523-926f9083b1a2",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-2",\n "node_id": "c4b64abb-c01b-4f21-8a28-1ad1423b1945",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "27df9434-425e-4676-a6a9-4b986991ac24",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-1",\n "node_id": "9b02add2-b915-4fcd-9ad8-9b08f8f8d701",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "2a3e5541-9f3d-417e-88bc-27bf19a71693",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-2",\n "node_id": "ced1678f-8786-43fe-bc96-cdbe81b1330b",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "51bac060-84d6-4d50-9a5b-150dd0d1e8c0",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-0",\n "node_id": "a3c1225e-8867-45ec-ab09-a8c24fee8f52",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "aabb3a50-7bf3-423d-ba58-8428081cc65b",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-rs0-1",\n "node_id": "323e006f-a2dd-4255-bb74-58c2618e1313",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "ca42a63d-f8d8-4b0f-b858-c5ef06cc5569",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-2",\n "node_id": "e21d39cb-99de-40bb-8f0a-2b342df379df",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "d1b1fe5f-ee05-4b6f-82fd-d0bdfcb0adeb",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-mongos-0",\n "node_id": "942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongos",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n },\n {\n "service_id": "d984828e-9147-4047-9a7f-0d9f0ec608ab",\n "service_name": "monitoring-pmm3-31037-monitoring-pmm3-cfg-1",\n "node_id": "2f443660-febf-4fe1-9d79-919e88cfe5d6",\n "address": "localhost",\n "port": 27019,\n "socket": "",\n "environment": "dev-mongod",\n "cluster": "super-custom",\n "replication_set": "",\n "custom_labels": {},\n "version": ""\n }\n ],\n "postgresql": [],\n "proxysql": [],\n "haproxy": [],\n "external": [],\n "valkey": []\n}' ++ jq -r --arg name monitoring-pmm3-31037-monitoring-pmm3-rs0-0 $'\n\t\t\t.mongodb[] | select(.service_name == $name) | .cluster\n\t\t' + actual_cluster=super-custom + [[ -z super-custom ]] + [[ super-custom == null ]] + [[ super-custom != super\-custom ]] + return 0 + nodeList=($(get_node_id_from_pmm)) ++ get_node_id_from_pmm ++ nodeList=() ++ local -a nodeList +++ kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns=NAME:.metadata.name ++++ mktemp +++ local LAST_OUT=/tmp/tmp.10bAwlHe4T ++++ mktemp +++ local LAST_ERR=/tmp/tmp.7aDuSWAn9c +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns=NAME:.metadata.name +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.10bAwlHe4T +++ cat /tmp/tmp.7aDuSWAn9c +++ rm /tmp/tmp.10bAwlHe4T /tmp/tmp.7aDuSWAn9c +++ return 0 ++ for instance in $(kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns='NAME:.metadata.name') ++ nodeList+=($(kubectl_bin exec -n "$namespace" $instance -c pmm-client -- pmm-admin status --json | jq -r '.pmm_agent_status.node_id')) +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-pmm3-cfg-0 -c pmm-client -- pmm-admin status --json +++ jq -r .pmm_agent_status.node_id ++++ mktemp +++ local LAST_OUT=/tmp/tmp.PqFpg7lEaX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.GQFwHK79oh +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-pmm3-cfg-0 -c pmm-client -- pmm-admin status --json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.PqFpg7lEaX +++ cat /tmp/tmp.GQFwHK79oh +++ rm /tmp/tmp.PqFpg7lEaX /tmp/tmp.GQFwHK79oh +++ return 0 ++ for instance in $(kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns='NAME:.metadata.name') ++ nodeList+=($(kubectl_bin exec -n "$namespace" $instance -c pmm-client -- pmm-admin status --json | jq -r '.pmm_agent_status.node_id')) +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-pmm3-cfg-1 -c pmm-client -- pmm-admin status --json +++ jq -r .pmm_agent_status.node_id ++++ mktemp +++ local LAST_OUT=/tmp/tmp.11eyoXyc7A ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5bkN6cpwA6 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-pmm3-cfg-1 -c pmm-client -- pmm-admin status --json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.11eyoXyc7A +++ cat /tmp/tmp.5bkN6cpwA6 +++ rm /tmp/tmp.11eyoXyc7A /tmp/tmp.5bkN6cpwA6 +++ return 0 ++ for instance in $(kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns='NAME:.metadata.name') ++ nodeList+=($(kubectl_bin exec -n "$namespace" $instance -c pmm-client -- pmm-admin status --json | jq -r '.pmm_agent_status.node_id')) +++ jq -r .pmm_agent_status.node_id +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-pmm3-cfg-2 -c pmm-client -- pmm-admin status --json ++++ mktemp +++ local LAST_OUT=/tmp/tmp.6dogfikG5o ++++ mktemp +++ local LAST_ERR=/tmp/tmp.oWnhY4omPH +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-pmm3-cfg-2 -c pmm-client -- pmm-admin status --json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.6dogfikG5o +++ cat /tmp/tmp.oWnhY4omPH +++ rm /tmp/tmp.6dogfikG5o /tmp/tmp.oWnhY4omPH +++ return 0 ++ for instance in $(kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns='NAME:.metadata.name') ++ nodeList+=($(kubectl_bin exec -n "$namespace" $instance -c pmm-client -- pmm-admin status --json | jq -r '.pmm_agent_status.node_id')) +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-pmm3-mongos-0 -c pmm-client -- pmm-admin status --json +++ jq -r .pmm_agent_status.node_id ++++ mktemp +++ local LAST_OUT=/tmp/tmp.68E19lR1zZ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.osYV3z8Gg2 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-pmm3-mongos-0 -c pmm-client -- pmm-admin status --json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.68E19lR1zZ +++ cat /tmp/tmp.osYV3z8Gg2 +++ rm /tmp/tmp.68E19lR1zZ /tmp/tmp.osYV3z8Gg2 +++ return 0 ++ for instance in $(kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns='NAME:.metadata.name') ++ nodeList+=($(kubectl_bin exec -n "$namespace" $instance -c pmm-client -- pmm-admin status --json | jq -r '.pmm_agent_status.node_id')) +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-pmm3-mongos-1 -c pmm-client -- pmm-admin status --json +++ jq -r .pmm_agent_status.node_id ++++ mktemp +++ local LAST_OUT=/tmp/tmp.XIYtPWbh8E ++++ mktemp +++ local LAST_ERR=/tmp/tmp.IWrRjgMPuf +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-pmm3-mongos-1 -c pmm-client -- pmm-admin status --json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.XIYtPWbh8E +++ cat /tmp/tmp.IWrRjgMPuf +++ rm /tmp/tmp.XIYtPWbh8E /tmp/tmp.IWrRjgMPuf +++ return 0 ++ for instance in $(kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns='NAME:.metadata.name') ++ nodeList+=($(kubectl_bin exec -n "$namespace" $instance -c pmm-client -- pmm-admin status --json | jq -r '.pmm_agent_status.node_id')) +++ jq -r .pmm_agent_status.node_id +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-pmm3-mongos-2 -c pmm-client -- pmm-admin status --json ++++ mktemp +++ local LAST_OUT=/tmp/tmp.ff4sFRKrBO ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jlyegYCUF1 +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-pmm3-mongos-2 -c pmm-client -- pmm-admin status --json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.ff4sFRKrBO +++ cat /tmp/tmp.jlyegYCUF1 +++ rm /tmp/tmp.ff4sFRKrBO /tmp/tmp.jlyegYCUF1 +++ return 0 ++ for instance in $(kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns='NAME:.metadata.name') ++ nodeList+=($(kubectl_bin exec -n "$namespace" $instance -c pmm-client -- pmm-admin status --json | jq -r '.pmm_agent_status.node_id')) +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-pmm3-rs0-0 -c pmm-client -- pmm-admin status --json +++ jq -r .pmm_agent_status.node_id ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OaJU6b2YrT ++++ mktemp +++ local LAST_ERR=/tmp/tmp.8zK7ySpqXc +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-pmm3-rs0-0 -c pmm-client -- pmm-admin status --json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.OaJU6b2YrT +++ cat /tmp/tmp.8zK7ySpqXc +++ rm /tmp/tmp.OaJU6b2YrT /tmp/tmp.8zK7ySpqXc +++ return 0 ++ for instance in $(kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns='NAME:.metadata.name') ++ nodeList+=($(kubectl_bin exec -n "$namespace" $instance -c pmm-client -- pmm-admin status --json | jq -r '.pmm_agent_status.node_id')) +++ jq -r .pmm_agent_status.node_id +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-pmm3-rs0-1 -c pmm-client -- pmm-admin status --json ++++ mktemp +++ local LAST_OUT=/tmp/tmp.RU0sXvhskq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.r4np6magMg +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-pmm3-rs0-1 -c pmm-client -- pmm-admin status --json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.RU0sXvhskq +++ cat /tmp/tmp.r4np6magMg +++ rm /tmp/tmp.RU0sXvhskq /tmp/tmp.r4np6magMg +++ return 0 ++ for instance in $(kubectl_bin get pods --no-headers -l app.kubernetes.io/name=percona-server-mongodb --output=custom-columns='NAME:.metadata.name') ++ nodeList+=($(kubectl_bin exec -n "$namespace" $instance -c pmm-client -- pmm-admin status --json | jq -r '.pmm_agent_status.node_id')) +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-pmm3-rs0-2 -c pmm-client -- pmm-admin status --json +++ jq -r .pmm_agent_status.node_id ++++ mktemp +++ local LAST_OUT=/tmp/tmp.dbExKOCuJg ++++ mktemp +++ local LAST_ERR=/tmp/tmp.w7CJ5oIk0M +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-pmm3-rs0-2 -c pmm-client -- pmm-admin status --json +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.dbExKOCuJg +++ cat /tmp/tmp.w7CJ5oIk0M +++ rm /tmp/tmp.dbExKOCuJg /tmp/tmp.w7CJ5oIk0M +++ return 0 ++ echo 9bd0200e-5913-40eb-b73d-d7c2dabf1c9b 2f443660-febf-4fe1-9d79-919e88cfe5d6 c4b64abb-c01b-4f21-8a28-1ad1423b1945 942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e 9b02add2-b915-4fcd-9ad8-9b08f8f8d701 e21d39cb-99de-40bb-8f0a-2b342df379df a3c1225e-8867-45ec-ab09-a8c24fee8f52 323e006f-a2dd-4255-bb74-58c2618e1313 ced1678f-8786-43fe-bc96-cdbe81b1330b + nodeList_from_pmm=($(does_node_id_exists "${nodeList[@]}")) ++ does_node_id_exists 9bd0200e-5913-40eb-b73d-d7c2dabf1c9b 2f443660-febf-4fe1-9d79-919e88cfe5d6 c4b64abb-c01b-4f21-8a28-1ad1423b1945 942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e 9b02add2-b915-4fcd-9ad8-9b08f8f8d701 e21d39cb-99de-40bb-8f0a-2b342df379df a3c1225e-8867-45ec-ab09-a8c24fee8f52 323e006f-a2dd-4255-bb74-58c2618e1313 ced1678f-8786-43fe-bc96-cdbe81b1330b ++ nodeList=('9bd0200e-5913-40eb-b73d-d7c2dabf1c9b' '2f443660-febf-4fe1-9d79-919e88cfe5d6' 'c4b64abb-c01b-4f21-8a28-1ad1423b1945' '942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e' '9b02add2-b915-4fcd-9ad8-9b08f8f8d701' 'e21d39cb-99de-40bb-8f0a-2b342df379df' 'a3c1225e-8867-45ec-ab09-a8c24fee8f52' '323e006f-a2dd-4255-bb74-58c2618e1313' 'ced1678f-8786-43fe-bc96-cdbe81b1330b') ++ local -a nodeList ++ nodeList_from_pmm=() ++ local -a nodeList_from_pmm ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep 9bd0200e-5913-40eb-b73d-d7c2dabf1c9b +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.neSZVcDNdE +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.V9FqbtZijJ ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.neSZVcDNdE ++++ cat /tmp/tmp.V9FqbtZijJ ++++ rm /tmp/tmp.neSZVcDNdE /tmp/tmp.V9FqbtZijJ ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.YMg73kagoV +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.VBtUA4T2xX ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.YMg73kagoV ++++ cat /tmp/tmp.VBtUA4T2xX ++++ rm /tmp/tmp.YMg73kagoV /tmp/tmp.VBtUA4T2xX ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.WcPvc3WVnX ++++ mktemp +++ local LAST_ERR=/tmp/tmp.fgrxycgbeD +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.WcPvc3WVnX +++ cat /tmp/tmp.fgrxycgbeD command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.WcPvc3WVnX +++ cat /tmp/tmp.fgrxycgbeD command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.WcPvc3WVnX +++ cat /tmp/tmp.fgrxycgbeD command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.WcPvc3WVnX +++ cat /tmp/tmp.fgrxycgbeD command terminated with exit code 1 +++ rm /tmp/tmp.WcPvc3WVnX /tmp/tmp.fgrxycgbeD +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep 2f443660-febf-4fe1-9d79-919e88cfe5d6 +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.QEniQhivBc +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.9nrzrcE7XE ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.QEniQhivBc ++++ cat /tmp/tmp.9nrzrcE7XE ++++ rm /tmp/tmp.QEniQhivBc /tmp/tmp.9nrzrcE7XE ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.ZzTZCkkImG +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.0o4vU2Npax ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.ZzTZCkkImG ++++ cat /tmp/tmp.0o4vU2Npax ++++ rm /tmp/tmp.ZzTZCkkImG /tmp/tmp.0o4vU2Npax ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.C7OYu7vPx6 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.KlDHkUB5SJ +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.C7OYu7vPx6 +++ cat /tmp/tmp.KlDHkUB5SJ command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.C7OYu7vPx6 +++ cat /tmp/tmp.KlDHkUB5SJ command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.C7OYu7vPx6 +++ cat /tmp/tmp.KlDHkUB5SJ command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.C7OYu7vPx6 +++ cat /tmp/tmp.KlDHkUB5SJ command terminated with exit code 1 +++ rm /tmp/tmp.C7OYu7vPx6 /tmp/tmp.KlDHkUB5SJ +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) ++++ get_pmm_service_ip monitoring-service +++ grep c4b64abb-c01b-4f21-8a28-1ad1423b1945 ++++ local service=monitoring-service +++ awk '{print $4}' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.uC6QOafcRu +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.SyVx8BZR1s ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.uC6QOafcRu ++++ cat /tmp/tmp.SyVx8BZR1s ++++ rm /tmp/tmp.uC6QOafcRu /tmp/tmp.SyVx8BZR1s ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.GCAUplrQh6 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.2izhbEQ6Rt ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.GCAUplrQh6 ++++ cat /tmp/tmp.2izhbEQ6Rt ++++ rm /tmp/tmp.GCAUplrQh6 /tmp/tmp.2izhbEQ6Rt ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.NKm83Gg6Rk ++++ mktemp +++ local LAST_ERR=/tmp/tmp.p4YHnTh7dg +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.NKm83Gg6Rk +++ cat /tmp/tmp.p4YHnTh7dg command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.NKm83Gg6Rk +++ cat /tmp/tmp.p4YHnTh7dg command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.NKm83Gg6Rk +++ cat /tmp/tmp.p4YHnTh7dg command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.NKm83Gg6Rk +++ cat /tmp/tmp.p4YHnTh7dg command terminated with exit code 1 +++ rm /tmp/tmp.NKm83Gg6Rk /tmp/tmp.p4YHnTh7dg +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep 942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service +++ awk '{print $4}' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.4GlwIXkAuZ +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.A1mCXbpT8U ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.4GlwIXkAuZ ++++ cat /tmp/tmp.A1mCXbpT8U ++++ rm /tmp/tmp.4GlwIXkAuZ /tmp/tmp.A1mCXbpT8U ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.2YyyfeKmAi +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.5Etdq3OKkx ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.2YyyfeKmAi ++++ cat /tmp/tmp.5Etdq3OKkx ++++ rm /tmp/tmp.2YyyfeKmAi /tmp/tmp.5Etdq3OKkx ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.Uo3JQT3wls ++++ mktemp +++ local LAST_ERR=/tmp/tmp.kSQj7YgJWI +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.Uo3JQT3wls +++ cat /tmp/tmp.kSQj7YgJWI command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.Uo3JQT3wls +++ cat /tmp/tmp.kSQj7YgJWI command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.Uo3JQT3wls +++ cat /tmp/tmp.kSQj7YgJWI command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.Uo3JQT3wls +++ cat /tmp/tmp.kSQj7YgJWI command terminated with exit code 1 +++ rm /tmp/tmp.Uo3JQT3wls /tmp/tmp.kSQj7YgJWI +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep 9b02add2-b915-4fcd-9ad8-9b08f8f8d701 +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.T4fmD2di3r +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.evPGygKdsh ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.T4fmD2di3r ++++ cat /tmp/tmp.evPGygKdsh ++++ rm /tmp/tmp.T4fmD2di3r /tmp/tmp.evPGygKdsh ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.D4GjxE9xmP +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.2Q2O1XkwJP ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.D4GjxE9xmP ++++ cat /tmp/tmp.2Q2O1XkwJP ++++ rm /tmp/tmp.D4GjxE9xmP /tmp/tmp.2Q2O1XkwJP ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.smFgElQJEv ++++ mktemp +++ local LAST_ERR=/tmp/tmp.wqoTdXmGep +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.smFgElQJEv +++ cat /tmp/tmp.wqoTdXmGep command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.smFgElQJEv +++ cat /tmp/tmp.wqoTdXmGep command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.smFgElQJEv +++ cat /tmp/tmp.wqoTdXmGep command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.smFgElQJEv +++ cat /tmp/tmp.wqoTdXmGep command terminated with exit code 1 +++ rm /tmp/tmp.smFgElQJEv /tmp/tmp.wqoTdXmGep +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep e21d39cb-99de-40bb-8f0a-2b342df379df +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.r2aLJuQXcB +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.9C4eajhQYj ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.r2aLJuQXcB ++++ cat /tmp/tmp.9C4eajhQYj ++++ rm /tmp/tmp.r2aLJuQXcB /tmp/tmp.9C4eajhQYj ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.URIs7vdCUk +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.8dplGYKDXq ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.URIs7vdCUk ++++ cat /tmp/tmp.8dplGYKDXq ++++ rm /tmp/tmp.URIs7vdCUk /tmp/tmp.8dplGYKDXq ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.TWpvFmmOaI ++++ mktemp +++ local LAST_ERR=/tmp/tmp.VEHcanBkdV +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.TWpvFmmOaI +++ cat /tmp/tmp.VEHcanBkdV command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.TWpvFmmOaI +++ cat /tmp/tmp.VEHcanBkdV command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.TWpvFmmOaI +++ cat /tmp/tmp.VEHcanBkdV command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.TWpvFmmOaI +++ cat /tmp/tmp.VEHcanBkdV command terminated with exit code 1 +++ rm /tmp/tmp.TWpvFmmOaI /tmp/tmp.VEHcanBkdV +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep a3c1225e-8867-45ec-ab09-a8c24fee8f52 ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service +++ awk '{print $4}' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.OF54D4jkqQ +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.jel5ttzYGA ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.OF54D4jkqQ ++++ cat /tmp/tmp.jel5ttzYGA ++++ rm /tmp/tmp.OF54D4jkqQ /tmp/tmp.jel5ttzYGA ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.gvpP4GNo9P +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.YKbcIFriYo ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.gvpP4GNo9P ++++ cat /tmp/tmp.YKbcIFriYo ++++ rm /tmp/tmp.gvpP4GNo9P /tmp/tmp.YKbcIFriYo ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.z4bK72k33Z ++++ mktemp +++ local LAST_ERR=/tmp/tmp.LAvAlmcCxZ +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.z4bK72k33Z +++ cat /tmp/tmp.LAvAlmcCxZ command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.z4bK72k33Z +++ cat /tmp/tmp.LAvAlmcCxZ command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.z4bK72k33Z +++ cat /tmp/tmp.LAvAlmcCxZ command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.z4bK72k33Z +++ cat /tmp/tmp.LAvAlmcCxZ command terminated with exit code 1 +++ rm /tmp/tmp.z4bK72k33Z /tmp/tmp.LAvAlmcCxZ +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ awk '{print $4}' +++ grep 323e006f-a2dd-4255-bb74-58c2618e1313 ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.souD6girM7 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.JvviMGnooF ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.souD6girM7 ++++ cat /tmp/tmp.JvviMGnooF ++++ rm /tmp/tmp.souD6girM7 /tmp/tmp.JvviMGnooF ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.Cj4BHr9ABJ +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.HTVFaqcIaJ ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.Cj4BHr9ABJ ++++ cat /tmp/tmp.HTVFaqcIaJ ++++ rm /tmp/tmp.Cj4BHr9ABJ /tmp/tmp.HTVFaqcIaJ ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.mW5ptVARbK ++++ mktemp +++ local LAST_ERR=/tmp/tmp.jnpBv8hVam +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.mW5ptVARbK +++ cat /tmp/tmp.jnpBv8hVam command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.mW5ptVARbK +++ cat /tmp/tmp.jnpBv8hVam command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.mW5ptVARbK +++ cat /tmp/tmp.jnpBv8hVam command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.mW5ptVARbK +++ cat /tmp/tmp.jnpBv8hVam command terminated with exit code 1 +++ rm /tmp/tmp.mW5ptVARbK /tmp/tmp.jnpBv8hVam +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep ced1678f-8786-43fe-bc96-cdbe81b1330b +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.kxz4S6hAP4 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.Ys93SOYYyg ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.kxz4S6hAP4 ++++ cat /tmp/tmp.Ys93SOYYyg ++++ rm /tmp/tmp.kxz4S6hAP4 /tmp/tmp.Ys93SOYYyg ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.MokarJ8mRG +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.ve4kSceGkc ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.MokarJ8mRG ++++ cat /tmp/tmp.ve4kSceGkc ++++ rm /tmp/tmp.MokarJ8mRG /tmp/tmp.ve4kSceGkc ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.nHlbKIZxaq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.81raHXfn3w +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.nHlbKIZxaq +++ cat /tmp/tmp.81raHXfn3w command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.nHlbKIZxaq +++ cat /tmp/tmp.81raHXfn3w command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.nHlbKIZxaq +++ cat /tmp/tmp.81raHXfn3w command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.nHlbKIZxaq +++ cat /tmp/tmp.81raHXfn3w command terminated with exit code 1 +++ rm /tmp/tmp.nHlbKIZxaq /tmp/tmp.81raHXfn3w +++ return 1 ++ echo + kubectl_bin patch psmdb monitoring-pmm3 --type json '-p=[{"op":"add","path":"/spec/pause","value":true}]' ++ mktemp + local LAST_OUT=/tmp/tmp.GGIVSWg7em ++ mktemp + local LAST_ERR=/tmp/tmp.mnW5uhgWSN + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl patch psmdb monitoring-pmm3 --type json '-p=[{"op":"add","path":"/spec/pause","value":true}]' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.GGIVSWg7em perconaservermongodb.psmdb.percona.com/monitoring-pmm3 patched + cat /tmp/tmp.mnW5uhgWSN + rm /tmp/tmp.GGIVSWg7em /tmp/tmp.mnW5uhgWSN + return 0 + wait_for_delete pod/monitoring-pmm3-mongos-0 + local res=pod/monitoring-pmm3-mongos-0 + local wait_time=60 + set +o xtrace waiting for pod/monitoring-pmm3-mongos-0 to be deleted........................Error from server (NotFound): pods "monitoring-pmm3-mongos-0" not found Error from server (NotFound): pods "monitoring-pmm3-mongos-0" not found Error from server (NotFound): pods "monitoring-pmm3-mongos-0" not found Error from server (NotFound): pods "monitoring-pmm3-mongos-0" not found + wait_for_delete pod/monitoring-pmm3-rs0-0 + local res=pod/monitoring-pmm3-rs0-0 + local wait_time=60 + set +o xtrace waiting for pod/monitoring-pmm3-rs0-0 to be deleted.......Error from server (NotFound): pods "monitoring-pmm3-rs0-0" not found Error from server (NotFound): pods "monitoring-pmm3-rs0-0" not found Error from server (NotFound): pods "monitoring-pmm3-rs0-0" not found Error from server (NotFound): pods "monitoring-pmm3-rs0-0" not found + wait_for_delete pod/monitoring-pmm3-cfg-0 + local res=pod/monitoring-pmm3-cfg-0 + local wait_time=60 + set +o xtrace waiting for pod/monitoring-pmm3-cfg-0 to be deleted..........Error from server (NotFound): pods "monitoring-pmm3-cfg-0" not found Error from server (NotFound): pods "monitoring-pmm3-cfg-0" not found Error from server (NotFound): pods "monitoring-pmm3-cfg-0" not found Error from server (NotFound): pods "monitoring-pmm3-cfg-0" not found + desc 'check if services are not deleted' + set +o xtrace ----------------------------------------------------------------------------------- check if services are not deleted ----------------------------------------------------------------------------------- + kubectl_bin get svc monitoring-pmm3-rs0 ++ mktemp + local LAST_OUT=/tmp/tmp.5FqYVCzaV5 ++ mktemp + local LAST_ERR=/tmp/tmp.WqFDG1RAHl + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get svc monitoring-pmm3-rs0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5FqYVCzaV5 NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE monitoring-pmm3-rs0 ClusterIP None 27019/TCP 23m + cat /tmp/tmp.WqFDG1RAHl + rm /tmp/tmp.5FqYVCzaV5 /tmp/tmp.WqFDG1RAHl + return 0 + kubectl_bin get svc monitoring-pmm3-cfg ++ mktemp + local LAST_OUT=/tmp/tmp.PzpGjaAk0c ++ mktemp + local LAST_ERR=/tmp/tmp.IRGhOsXTWG + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get svc monitoring-pmm3-cfg + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.PzpGjaAk0c NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE monitoring-pmm3-cfg ClusterIP None 27019/TCP 23m + cat /tmp/tmp.IRGhOsXTWG + rm /tmp/tmp.PzpGjaAk0c /tmp/tmp.IRGhOsXTWG + return 0 + kubectl_bin get svc monitoring-pmm3-mongos ++ mktemp + local LAST_OUT=/tmp/tmp.MUm9bsbn57 ++ mktemp + local LAST_ERR=/tmp/tmp.wdcJrhmXvH + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl get svc monitoring-pmm3-mongos + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.MUm9bsbn57 NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE monitoring-pmm3-mongos ClusterIP 34.118.226.87 27019/TCP 23m + cat /tmp/tmp.wdcJrhmXvH + rm /tmp/tmp.MUm9bsbn57 /tmp/tmp.wdcJrhmXvH + return 0 + does_node_id_exists_in_pmm=($(does_node_id_exists "${nodeList[@]}")) ++ does_node_id_exists 9bd0200e-5913-40eb-b73d-d7c2dabf1c9b 2f443660-febf-4fe1-9d79-919e88cfe5d6 c4b64abb-c01b-4f21-8a28-1ad1423b1945 942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e 9b02add2-b915-4fcd-9ad8-9b08f8f8d701 e21d39cb-99de-40bb-8f0a-2b342df379df a3c1225e-8867-45ec-ab09-a8c24fee8f52 323e006f-a2dd-4255-bb74-58c2618e1313 ced1678f-8786-43fe-bc96-cdbe81b1330b ++ nodeList=('9bd0200e-5913-40eb-b73d-d7c2dabf1c9b' '2f443660-febf-4fe1-9d79-919e88cfe5d6' 'c4b64abb-c01b-4f21-8a28-1ad1423b1945' '942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e' '9b02add2-b915-4fcd-9ad8-9b08f8f8d701' 'e21d39cb-99de-40bb-8f0a-2b342df379df' 'a3c1225e-8867-45ec-ab09-a8c24fee8f52' '323e006f-a2dd-4255-bb74-58c2618e1313' 'ced1678f-8786-43fe-bc96-cdbe81b1330b') ++ local -a nodeList ++ nodeList_from_pmm=() ++ local -a nodeList_from_pmm ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep 9bd0200e-5913-40eb-b73d-d7c2dabf1c9b ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service +++ awk '{print $4}' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.L6DQvhN1kR +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.ZoXa0MAD6W ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.L6DQvhN1kR ++++ cat /tmp/tmp.ZoXa0MAD6W ++++ rm /tmp/tmp.L6DQvhN1kR /tmp/tmp.ZoXa0MAD6W ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.Pw4cwD1e5q +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.UT9kx2WMfb ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.Pw4cwD1e5q ++++ cat /tmp/tmp.UT9kx2WMfb ++++ rm /tmp/tmp.Pw4cwD1e5q /tmp/tmp.UT9kx2WMfb ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.SUhA3RFoyL ++++ mktemp +++ local LAST_ERR=/tmp/tmp.ihYEorTWUm +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.SUhA3RFoyL +++ cat /tmp/tmp.ihYEorTWUm command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.SUhA3RFoyL +++ cat /tmp/tmp.ihYEorTWUm command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.SUhA3RFoyL +++ cat /tmp/tmp.ihYEorTWUm command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.SUhA3RFoyL +++ cat /tmp/tmp.ihYEorTWUm command terminated with exit code 1 +++ rm /tmp/tmp.SUhA3RFoyL /tmp/tmp.ihYEorTWUm +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep 2f443660-febf-4fe1-9d79-919e88cfe5d6 ++++ get_pmm_service_ip monitoring-service +++ awk '{print $4}' ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.u3utcsTP22 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.y04L7ozloY ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.u3utcsTP22 ++++ cat /tmp/tmp.y04L7ozloY ++++ rm /tmp/tmp.u3utcsTP22 /tmp/tmp.y04L7ozloY ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.hLnMX3HROK +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.dNyDCr8fG5 ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.hLnMX3HROK ++++ cat /tmp/tmp.dNyDCr8fG5 ++++ rm /tmp/tmp.hLnMX3HROK /tmp/tmp.dNyDCr8fG5 ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.xsh0iKpeei ++++ mktemp +++ local LAST_ERR=/tmp/tmp.h0bAIKxvXK +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.xsh0iKpeei +++ cat /tmp/tmp.h0bAIKxvXK command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.xsh0iKpeei +++ cat /tmp/tmp.h0bAIKxvXK command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.xsh0iKpeei +++ cat /tmp/tmp.h0bAIKxvXK command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.xsh0iKpeei +++ cat /tmp/tmp.h0bAIKxvXK command terminated with exit code 1 +++ rm /tmp/tmp.xsh0iKpeei /tmp/tmp.h0bAIKxvXK +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep c4b64abb-c01b-4f21-8a28-1ad1423b1945 +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.vgkeTcn830 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.LnCbhn4dR3 ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.vgkeTcn830 ++++ cat /tmp/tmp.LnCbhn4dR3 ++++ rm /tmp/tmp.vgkeTcn830 /tmp/tmp.LnCbhn4dR3 ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.L2wVfQexSW +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.HLpwuP1mzn ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.L2wVfQexSW ++++ cat /tmp/tmp.HLpwuP1mzn ++++ rm /tmp/tmp.L2wVfQexSW /tmp/tmp.HLpwuP1mzn ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.fzGofzxlp9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.T1ArnfK09C +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.fzGofzxlp9 +++ cat /tmp/tmp.T1ArnfK09C command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.fzGofzxlp9 +++ cat /tmp/tmp.T1ArnfK09C command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.fzGofzxlp9 +++ cat /tmp/tmp.T1ArnfK09C command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.fzGofzxlp9 +++ cat /tmp/tmp.T1ArnfK09C command terminated with exit code 1 +++ rm /tmp/tmp.fzGofzxlp9 /tmp/tmp.T1ArnfK09C +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep 942bbb2a-dbd1-404c-a7b4-1c0a53c64d3e +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.mW4za7i5FJ +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.ZpVYJgyNRH ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.mW4za7i5FJ ++++ cat /tmp/tmp.ZpVYJgyNRH ++++ rm /tmp/tmp.mW4za7i5FJ /tmp/tmp.ZpVYJgyNRH ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.c6FSBGK2nL +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.lGBGFDJ6AG ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.c6FSBGK2nL ++++ cat /tmp/tmp.lGBGFDJ6AG ++++ rm /tmp/tmp.c6FSBGK2nL /tmp/tmp.lGBGFDJ6AG ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.t5qMBdpkBL ++++ mktemp +++ local LAST_ERR=/tmp/tmp.3LFH8j6YjF +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.t5qMBdpkBL +++ cat /tmp/tmp.3LFH8j6YjF command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.t5qMBdpkBL +++ cat /tmp/tmp.3LFH8j6YjF command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.t5qMBdpkBL +++ cat /tmp/tmp.3LFH8j6YjF command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.t5qMBdpkBL +++ cat /tmp/tmp.3LFH8j6YjF command terminated with exit code 1 +++ rm /tmp/tmp.t5qMBdpkBL /tmp/tmp.3LFH8j6YjF +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep 9b02add2-b915-4fcd-9ad8-9b08f8f8d701 +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.djGEWwwZqV +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.alQa3YwLVK ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.djGEWwwZqV ++++ cat /tmp/tmp.alQa3YwLVK ++++ rm /tmp/tmp.djGEWwwZqV /tmp/tmp.alQa3YwLVK ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.x2w3jfzhM9 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.KKlwcqUxgG ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.x2w3jfzhM9 ++++ cat /tmp/tmp.KKlwcqUxgG ++++ rm /tmp/tmp.x2w3jfzhM9 /tmp/tmp.KKlwcqUxgG ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OI6UaNi9FS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1rBMDHximn +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.OI6UaNi9FS +++ cat /tmp/tmp.1rBMDHximn command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.OI6UaNi9FS +++ cat /tmp/tmp.1rBMDHximn command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.OI6UaNi9FS +++ cat /tmp/tmp.1rBMDHximn command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.OI6UaNi9FS +++ cat /tmp/tmp.1rBMDHximn command terminated with exit code 1 +++ rm /tmp/tmp.OI6UaNi9FS /tmp/tmp.1rBMDHximn +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep e21d39cb-99de-40bb-8f0a-2b342df379df +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.y7947rH3AP +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.Ny1GAudBN5 ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.y7947rH3AP ++++ cat /tmp/tmp.Ny1GAudBN5 ++++ rm /tmp/tmp.y7947rH3AP /tmp/tmp.Ny1GAudBN5 ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.ZC56UPithT +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.Tbb3oRqbVI ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.ZC56UPithT ++++ cat /tmp/tmp.Tbb3oRqbVI ++++ rm /tmp/tmp.ZC56UPithT /tmp/tmp.Tbb3oRqbVI ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.w9k9J4HsUJ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.j08zG1cXqQ +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.w9k9J4HsUJ +++ cat /tmp/tmp.j08zG1cXqQ command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.w9k9J4HsUJ +++ cat /tmp/tmp.j08zG1cXqQ command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.w9k9J4HsUJ +++ cat /tmp/tmp.j08zG1cXqQ command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.w9k9J4HsUJ +++ cat /tmp/tmp.j08zG1cXqQ command terminated with exit code 1 +++ rm /tmp/tmp.w9k9J4HsUJ /tmp/tmp.j08zG1cXqQ +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep a3c1225e-8867-45ec-ab09-a8c24fee8f52 +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.2c0086vEYk +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.YGGjnLQI16 ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.2c0086vEYk ++++ cat /tmp/tmp.YGGjnLQI16 ++++ rm /tmp/tmp.2c0086vEYk /tmp/tmp.YGGjnLQI16 ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.5TQ3deOhr3 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.etNAMNBwad ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.5TQ3deOhr3 ++++ cat /tmp/tmp.etNAMNBwad ++++ rm /tmp/tmp.5TQ3deOhr3 /tmp/tmp.etNAMNBwad ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.oy5QDpOncx ++++ mktemp +++ local LAST_ERR=/tmp/tmp.i6evt7MBed +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.oy5QDpOncx +++ cat /tmp/tmp.i6evt7MBed command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.oy5QDpOncx +++ cat /tmp/tmp.i6evt7MBed command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.oy5QDpOncx +++ cat /tmp/tmp.i6evt7MBed command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.oy5QDpOncx +++ cat /tmp/tmp.i6evt7MBed command terminated with exit code 1 +++ rm /tmp/tmp.oy5QDpOncx /tmp/tmp.i6evt7MBed +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep 323e006f-a2dd-4255-bb74-58c2618e1313 +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.tLmpUOiVUL +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.Ox2TegL8c7 ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.tLmpUOiVUL ++++ cat /tmp/tmp.Ox2TegL8c7 ++++ rm /tmp/tmp.tLmpUOiVUL /tmp/tmp.Ox2TegL8c7 ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.3JfZtVEoLU +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.gcaRNRvNlA ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.3JfZtVEoLU ++++ cat /tmp/tmp.gcaRNRvNlA ++++ rm /tmp/tmp.3JfZtVEoLU /tmp/tmp.gcaRNRvNlA ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.JlsOacqoaJ ++++ mktemp +++ local LAST_ERR=/tmp/tmp.SqtluAIThU +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.JlsOacqoaJ +++ cat /tmp/tmp.SqtluAIThU command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.JlsOacqoaJ +++ cat /tmp/tmp.SqtluAIThU command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.JlsOacqoaJ +++ cat /tmp/tmp.SqtluAIThU command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.JlsOacqoaJ +++ cat /tmp/tmp.SqtluAIThU command terminated with exit code 1 +++ rm /tmp/tmp.JlsOacqoaJ /tmp/tmp.SqtluAIThU +++ return 1 ++ for node_id in "${nodeList[@]}" ++ nodeList_from_pmm+=($(kubectl_bin exec -n "${namespace}" monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@$(get_pmm_service_ip monitoring-service)/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE | grep $node_id | awk '{print $4}')) +++ grep ced1678f-8786-43fe-bc96-cdbe81b1330b +++ awk '{print $4}' ++++ get_pmm_service_ip monitoring-service ++++ local service=monitoring-service ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.spec.type}' ++++ grep -q NotFound ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[]}' ++++ grep -E -q 'hostname|ip' ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.uRlYtD7siF +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.oxUQUzkZbL ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].ip}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.uRlYtD7siF ++++ cat /tmp/tmp.oxUQUzkZbL ++++ rm /tmp/tmp.uRlYtD7siF /tmp/tmp.oxUQUzkZbL ++++ return 0 ++++ kubectl_bin get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.OKKRd0G8mo +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.J0x5AZ0i2Z ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in $(seq 0 2) ++++ set +e ++++ kubectl get service/monitoring-service -o 'jsonpath={.status.loadBalancer.ingress[].hostname}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.OKKRd0G8mo ++++ cat /tmp/tmp.J0x5AZ0i2Z ++++ rm /tmp/tmp.OKKRd0G8mo /tmp/tmp.J0x5AZ0i2Z ++++ return 0 +++ kubectl_bin exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE ++++ mktemp +++ local LAST_OUT=/tmp/tmp.k7LrjJuKdV ++++ mktemp +++ local LAST_ERR=/tmp/tmp.3ntwPuCMIQ +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.k7LrjJuKdV +++ cat /tmp/tmp.3ntwPuCMIQ command terminated with exit code 1 +++ sleep 0 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.k7LrjJuKdV +++ cat /tmp/tmp.3ntwPuCMIQ command terminated with exit code 1 +++ sleep 4 +++ for i in $(seq 0 2) +++ set +e +++ kubectl exec -n monitoring-pmm3-31037 monitoring-server-0 -- pmm-admin --server-url=https://admin:admin@34.133.92.251/ --server-insecure-tls inventory list nodes --node-type=CONTAINER_NODE +++ exit_status=1 +++ set -e +++ '[' 1 '!=' 0 -a -n 1 ']' +++ cat /tmp/tmp.k7LrjJuKdV +++ cat /tmp/tmp.3ntwPuCMIQ command terminated with exit code 1 +++ sleep 8 +++ cat /tmp/tmp.k7LrjJuKdV +++ cat /tmp/tmp.3ntwPuCMIQ command terminated with exit code 1 +++ rm /tmp/tmp.k7LrjJuKdV /tmp/tmp.3ntwPuCMIQ +++ return 1 ++ echo + [[ -n '' ]] ++ kubectl_bin logs monitoring-pmm3-rs0-0 pmm-client ++ grep -c 'cannot auto discover databases and collections' +++ mktemp ++ local LAST_OUT=/tmp/tmp.00Ct19ilL2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bPnZqdq7zB ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl logs monitoring-pmm3-rs0-0 pmm-client ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.00Ct19ilL2 ++ cat /tmp/tmp.bPnZqdq7zB error: error from server (NotFound): pods "monitoring-pmm3-rs0-0" not found in namespace "monitoring-pmm3-31037" ++ sleep 0 ++ for i in $(seq 0 2) ++ set +e ++ kubectl logs monitoring-pmm3-rs0-0 pmm-client ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.00Ct19ilL2 ++ cat /tmp/tmp.bPnZqdq7zB error: error from server (NotFound): pods "monitoring-pmm3-rs0-0" not found in namespace "monitoring-pmm3-31037" ++ sleep 4 ++ for i in $(seq 0 2) ++ set +e ++ kubectl logs monitoring-pmm3-rs0-0 pmm-client ++ exit_status=1 ++ set -e ++ '[' 1 '!=' 0 -a -n 1 ']' ++ cat /tmp/tmp.00Ct19ilL2 ++ cat /tmp/tmp.bPnZqdq7zB error: error from server (NotFound): pods "monitoring-pmm3-rs0-0" not found in namespace "monitoring-pmm3-31037" ++ sleep 8 ++ cat /tmp/tmp.00Ct19ilL2 ++ cat /tmp/tmp.bPnZqdq7zB error: error from server (NotFound): pods "monitoring-pmm3-rs0-0" not found in namespace "monitoring-pmm3-31037" ++ rm /tmp/tmp.00Ct19ilL2 /tmp/tmp.bPnZqdq7zB ++ return 1 + [[ 0 != 0 ]] + helm uninstall monitoring release "monitoring" uninstalled + destroy monitoring-pmm3-31037 + local namespace=monitoring-pmm3-31037 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false ']' + delete_backups + desc 'Delete psmdb-backup' + set +o xtrace ----------------------------------------------------------------------------------- Delete psmdb-backup ----------------------------------------------------------------------------------- ++ kubectl_bin get psmdb-backup --no-headers ++ wc -l +++ mktemp ++ local LAST_OUT=/tmp/tmp.elPSFRG2Sx +++ mktemp ++ local LAST_ERR=/tmp/tmp.KDaMZzANUu ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in $(seq 0 2) ++ set +e ++ kubectl get psmdb-backup --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.elPSFRG2Sx ++ cat /tmp/tmp.KDaMZzANUu No resources found in monitoring-pmm3-31037 namespace. ++ rm /tmp/tmp.elPSFRG2Sx /tmp/tmp.KDaMZzANUu ++ return 0 + '[' 0 '!=' 0 ']' + delete_crd + desc 'get and delete old CRDs and RBAC' + set +o xtrace ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/crd.yaml --ignore-not-found --wait=false ++ mktemp + local LAST_OUT=/tmp/tmp.5ASjUsTnMW ++ mktemp + local LAST_ERR=/tmp/tmp.YZg9NhjNu0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/crd.yaml --ignore-not-found --wait=false + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5ASjUsTnMW customresourcedefinition.apiextensions.k8s.io "perconaservermongodbbackups.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbrestores.psmdb.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaservermongodbs.psmdb.percona.com" deleted + cat /tmp/tmp.YZg9NhjNu0 + rm /tmp/tmp.5ASjUsTnMW /tmp/tmp.YZg9NhjNu0 + return 0 ++ yq eval .metadata.name /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/crd.yaml ++ grep -v '\-\-\-' grep: warning: stray \ before - grep: warning: stray \ before - + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbbackups.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbbackups.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" + : + kubectl_bin wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.S2Dxl2F8OH ++ mktemp + local LAST_ERR=/tmp/tmp.xgeyyWrkhW + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbbackups.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.S2Dxl2F8OH + cat /tmp/tmp.xgeyyWrkhW + rm /tmp/tmp.S2Dxl2F8OH /tmp/tmp.xgeyyWrkhW + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbrestores.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbrestores.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" + : + kubectl_bin wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.Y3mPP6drlK ++ mktemp + local LAST_ERR=/tmp/tmp.qroabi3ZJT + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbrestores.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.Y3mPP6drlK + cat /tmp/tmp.qroabi3ZJT + rm /tmp/tmp.Y3mPP6drlK /tmp/tmp.qroabi3ZJT + return 0 + for crd_name in $(yq eval '.metadata.name' "${src_dir}/deploy/crd.yaml" | grep -v '\-\-\-') + kubectl get perconaservermongodbs.psmdb.percona.com --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch perconaservermongodbs.psmdb.percona.com -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" + : + kubectl_bin wait --for=delete crd perconaservermongodbs.psmdb.percona.com ++ mktemp + local LAST_OUT=/tmp/tmp.DKIZmAycdf ++ mktemp + local LAST_ERR=/tmp/tmp.imidUr7QnV + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl wait --for=delete crd perconaservermongodbs.psmdb.percona.com + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.DKIZmAycdf + cat /tmp/tmp.imidUr7QnV + rm /tmp/tmp.DKIZmAycdf /tmp/tmp.imidUr7QnV + return 0 + local rbac_yaml=rbac.yaml + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac.yaml + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/cw-rbac.yaml --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.ZyWuVCVl1i ++ mktemp + local LAST_ERR=/tmp/tmp.GY2kQSXcDi + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2256/deploy/cw-rbac.yaml --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ZyWuVCVl1i clusterrole.rbac.authorization.k8s.io "percona-server-mongodb-operator" deleted clusterrolebinding.rbac.authorization.k8s.io "service-account-percona-server-mongodb-operator" deleted + cat /tmp/tmp.GY2kQSXcDi + rm /tmp/tmp.ZyWuVCVl1i /tmp/tmp.GY2kQSXcDi + return 0 + destroy_cert_manager + kubectl_bin delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.FUFs8izLYJ ++ mktemp + local LAST_ERR=/tmp/tmp.BqkdAtGpHi + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.FUFs8izLYJ namespace "cert-manager" deleted customresourcedefinition.apiextensions.k8s.io "challenges.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "orders.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificaterequests.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificates.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "clusterissuers.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "issuers.cert-manager.io" deleted serviceaccount "cert-manager-cainjector" deleted from cert-manager namespace serviceaccount "cert-manager" deleted from cert-manager namespace serviceaccount "cert-manager-webhook" deleted from cert-manager namespace clusterrole.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-cluster-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-edit" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted role.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace role.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace role.rbac.authorization.k8s.io "cert-manager-tokenrequest" deleted from cert-manager namespace role.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" deleted from cert-manager namespace rolebinding.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager-tokenrequest" deleted from cert-manager namespace rolebinding.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" deleted from cert-manager namespace mutatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted validatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted + cat /tmp/tmp.BqkdAtGpHi Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found + sleep 0 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.FUFs8izLYJ namespace "cert-manager" deleted + cat /tmp/tmp.BqkdAtGpHi Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 4 + for i in $(seq 0 2) + set +e + kubectl delete -f https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml + exit_status=1 + set -e + '[' 1 '!=' 0 -a -n 1 ']' + cat /tmp/tmp.FUFs8izLYJ namespace "cert-manager" deleted + cat /tmp/tmp.BqkdAtGpHi Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + sleep 8 + cat /tmp/tmp.FUFs8izLYJ namespace "cert-manager" deleted + cat /tmp/tmp.BqkdAtGpHi Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "challenges.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "orders.acme.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificaterequests.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "certificates.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "clusterissuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": customresourcedefinitions.apiextensions.k8s.io "issuers.cert-manager.io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": serviceaccounts "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-cluster-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-view" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-edit" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterroles.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-issuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificates" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-orders" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-challenges" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": clusterrolebindings.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": roles.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager:leaderelection" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-tokenrequest" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": rolebindings.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": services "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-cainjector" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": deployments.apps "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": mutatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found Error from server (NotFound): error when deleting "https://github.com/cert-manager/cert-manager/releases/download/v1.19.3/cert-manager.yaml": validatingwebhookconfigurations.admissionregistration.k8s.io "cert-manager-webhook" not found + rm /tmp/tmp.FUFs8izLYJ /tmp/tmp.BqkdAtGpHi + return 1 + true + '[' -n '' ']' + '[' -n psmdb-operator ']' + rm -rf /tmp/tmp.KtaPi3a7ND + kubectl_bin delete --grace-period=0 --force=true namespace psmdb-operator + kubectl_bin delete --grace-period=0 --force=true namespace monitoring-pmm3-31037 ++ mktemp ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.9BvrIdmFpx + local LAST_OUT=/tmp/tmp.aSqpaS5vmh ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.EOrHub68nl + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace monitoring-pmm3-31037 + local LAST_ERR=/tmp/tmp.l31GCyoifc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in $(seq 0 2) + set +e + kubectl delete --grace-period=0 --force=true namespace psmdb-operator