Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1961/e2e-tests/logs/upgrade.log WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 + cluster=upgrade + CLUSTER_SIZE=3 + TARGET_OPERATOR_VER=1.21.0 + TARGET_IMAGE=perconalab/percona-server-mongodb-operator:PR-1961-970eeaa4 + TARGET_IMAGE_MONGOD=perconalab/percona-server-mongodb-operator:main-mongod7.0 + TARGET_IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest + TARGET_IMAGE_BACKUP=perconalab/percona-server-mongodb-operator:main-backup ++ get_mongod_ver_from_image perconalab/percona-server-mongodb-operator:main-mongod7.0 ++ local image=perconalab/percona-server-mongodb-operator:main-mongod7.0 +++ run_simple_cli_inside_image perconalab/percona-server-mongodb-operator:main-mongod7.0 'mongod --version' +++ local image=perconalab/percona-server-mongodb-operator:main-mongod7.0 +++ local 'cli=mongod --version' +++ local pod_name=23490 +++ /usr/bin/sed -r 's/^.*db version v(([0-9]+\.){2}[0-9]+-[0-9]+).*$/\1/g' +++ kubectl_bin -n default run 23490 --image=perconalab/percona-server-mongodb-operator:main-mongod7.0 --restart=Never --command -- sleep infinity ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DfZOkkuO2V ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zUC4ElmLpY +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl -n default run 23490 --image=perconalab/percona-server-mongodb-operator:main-mongod7.0 --restart=Never --command -- sleep infinity +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.DfZOkkuO2V +++ cat /tmp/tmp.zUC4ElmLpY +++ rm /tmp/tmp.DfZOkkuO2V /tmp/tmp.zUC4ElmLpY +++ return 0 +++ kubectl_bin -n default wait --for=condition=Ready pod/23490 ++++ mktemp +++ local LAST_OUT=/tmp/tmp.SAgb071j2l ++++ mktemp +++ local LAST_ERR=/tmp/tmp.trEE6NCPwa +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl -n default wait --for=condition=Ready pod/23490 +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.SAgb071j2l +++ cat /tmp/tmp.trEE6NCPwa +++ rm /tmp/tmp.SAgb071j2l /tmp/tmp.trEE6NCPwa +++ return 0 ++++ kubectl_bin -n default exec 23490 -- bash -c 'mongod --version 2>&1' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.prdnB0iGwh +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.HNZA1ILbAj ++++ local exit_status=0 ++++ local timeout=4 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl -n default exec 23490 -- bash -c 'mongod --version 2>&1' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 -a -n 1 ']' ++++ break ++++ cat /tmp/tmp.prdnB0iGwh ++++ cat /tmp/tmp.HNZA1ILbAj ++++ rm /tmp/tmp.prdnB0iGwh /tmp/tmp.HNZA1ILbAj ++++ return 0 +++ local 'output=db version v7.0.22-12 Build Info: { "version": "7.0.22-12", "gitVersion": "dba13d69ddf3a3457b6e155d33709fd8bbf41a96", "openSSLVersion": "OpenSSL 3.2.2 4 Jun 2024", "modules": [], "proFeatures": [], "allocator": "tcmalloc", "environment": { "distarch": "x86_64", "target_arch": "x86_64" } }' +++ kubectl_bin -n default delete pod/23490 --grace-period=0 --force ++++ mktemp +++ local LAST_OUT=/tmp/tmp.CukMZb4OKM ++++ mktemp +++ local LAST_ERR=/tmp/tmp.YhHaefyQha +++ local exit_status=0 +++ local timeout=4 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl -n default delete pod/23490 --grace-period=0 --force +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 -a -n 1 ']' +++ break +++ cat /tmp/tmp.CukMZb4OKM +++ cat /tmp/tmp.YhHaefyQha Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. +++ rm /tmp/tmp.CukMZb4OKM /tmp/tmp.YhHaefyQha +++ return 0 +++ echo db version v7.0.22-12 Build Info: '{' '"version":' '"7.0.22-12",' '"gitVersion":' '"dba13d69ddf3a3457b6e155d33709fd8bbf41a96",' '"openSSLVersion":' '"OpenSSL' 3.2.2 4 Jun '2024",' '"modules":' '[],' '"proFeatures":' '[],' '"allocator":' '"tcmalloc",' '"environment":' '{' '"distarch":' '"x86_64",' '"target_arch":' '"x86_64"' '}' '}' ++ version_info=7.0.22-12 ++ [[ ! 7.0.22-12 =~ ^([0-9]+\.){2}[0-9]+-[0-9]+$ ]] ++ echo 7.0.22-12 + FULL_VER=7.0.22-12 + MONGO_VER=7.0 ++ curl -s https://check.percona.com/versions/v1/psmdb-operator ++ jq -r '.versions[].operator' ++ sort -V ++ tail -n1 + INIT_OPERATOR_VER=1.20.1 + [[ 1.20.1 == \1\.\2\1\.\0 ]] + GIT_TAG=v1.20.1 + case $(curl -s -o /dev/null -w "%{http_code}" 'https://check.percona.com/versions/v1/psmdb-operator/'${INIT_OPERATOR_VER}'/latest?databaseVersion='${MONGO_VER}'') in ++ curl -s -o /dev/null -w '%{http_code}' 'https://check.percona.com/versions/v1/psmdb-operator/1.20.1/latest?databaseVersion=7.0' ++ curl -s 'https://check.percona.com/versions/v1/psmdb-operator/1.20.1/latest?databaseVersion=7.0' + INIT_OPERATOR_IMAGES='{"versions":[{"product":"psmdb-operator", "operator":"1.20.1", "matrix":{"mongod":{"7.0.18-11":{"imagePath":"percona/percona-server-mongodb:7.0.18-11", "imageHash":"0115a72f5e60d86cb4f4b7eae32118c0910e8c96831e013de12798a1771c4c91", "imageHashArm64":"86c17067f3e233f522612389ed2500231cbb22ce93524c476b9aa8d464d06f0b", "status":"recommended", "critical":false}}, "pxc":{}, "pmm":{"2.44.1":{"imagePath":"percona/pmm-client:2.44.1", "imageHash":"8b2eaddffd626f02a2d5318ffebc0c277fe8457da6083b8cfcada9b6e6168616", "imageHashArm64":"337fecd4afdb3f6daf2caa2b341b9fe41d0418a0e4ec76980c7f29be9d08b5ea", "status":"available", "critical":false}}, "proxysql":{}, "haproxy":{}, "backup":{"2.9.1":{"imagePath":"percona/percona-backup-mongodb:2.9.1", "imageHash":"976bfbaa548eb70dd90bf0bd2dcfe40b2994d749ef644af3a0590f4856e4d7e2", "imageHashArm64":"ebc6e5c5aa3ed97991d3fd90e9201597b485ddc0eae8d7ee4311ecb785c03bf0", "status":"recommended", "critical":false}}, "operator":{"1.20.1":{"imagePath":"percona/percona-server-mongodb-operator:1.20.1", "imageHash":"b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46", "imageHashArm64":"5a66e497dd1650e5a1123659292fe4c615e0ab5ce7e5d8437bf2101f91b625e1", "status":"recommended", "critical":false}}, "logCollector":{}, "postgresql":{}, "pgbackrest":{}, "pgbackrestRepo":{}, "pgbadger":{}, "pgbouncer":{}, "pxcOperator":{}, "psmdbOperator":{}, "pgOperatorApiserver":{}, "pgOperatorEvent":{}, "pgOperatorRmdata":{}, "pgOperatorScheduler":{}, "pgOperator":{}, "pgOperatorDeployer":{}, "psOperator":{}, "mysql":{}, "router":{}, "orchestrator":{}, "toolkit":{}, "postgis":{}}}]}' + OPERATOR_NAME=percona-server-mongodb-operator ++ echo '{"versions":[{"product":"psmdb-operator", "operator":"1.20.1", "matrix":{"mongod":{"7.0.18-11":{"imagePath":"percona/percona-server-mongodb:7.0.18-11", "imageHash":"0115a72f5e60d86cb4f4b7eae32118c0910e8c96831e013de12798a1771c4c91", "imageHashArm64":"86c17067f3e233f522612389ed2500231cbb22ce93524c476b9aa8d464d06f0b", "status":"recommended", "critical":false}}, "pxc":{}, "pmm":{"2.44.1":{"imagePath":"percona/pmm-client:2.44.1", "imageHash":"8b2eaddffd626f02a2d5318ffebc0c277fe8457da6083b8cfcada9b6e6168616", "imageHashArm64":"337fecd4afdb3f6daf2caa2b341b9fe41d0418a0e4ec76980c7f29be9d08b5ea", "status":"available", "critical":false}}, "proxysql":{}, "haproxy":{}, "backup":{"2.9.1":{"imagePath":"percona/percona-backup-mongodb:2.9.1", "imageHash":"976bfbaa548eb70dd90bf0bd2dcfe40b2994d749ef644af3a0590f4856e4d7e2", "imageHashArm64":"ebc6e5c5aa3ed97991d3fd90e9201597b485ddc0eae8d7ee4311ecb785c03bf0", "status":"recommended", "critical":false}}, "operator":{"1.20.1":{"imagePath":"percona/percona-server-mongodb-operator:1.20.1", "imageHash":"b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46", "imageHashArm64":"5a66e497dd1650e5a1123659292fe4c615e0ab5ce7e5d8437bf2101f91b625e1", "status":"recommended", "critical":false}}, "logCollector":{}, "postgresql":{}, "pgbackrest":{}, "pgbackrestRepo":{}, "pgbadger":{}, "pgbouncer":{}, "pxcOperator":{}, "psmdbOperator":{}, "pgOperatorApiserver":{}, "pgOperatorEvent":{}, "pgOperatorRmdata":{}, "pgOperatorScheduler":{}, "pgOperator":{}, "pgOperatorDeployer":{}, "psOperator":{}, "mysql":{}, "router":{}, "orchestrator":{}, "toolkit":{}, "postgis":{}}}]}' ++ jq -r '.versions[].matrix.operator[].imagePath' + IMAGE=percona/percona-server-mongodb-operator:1.20.1 ++ echo perconalab/percona-server-mongodb-operator:PR-1961-970eeaa4 ++ cut -d/ -f1 + [[ perconalab == \p\e\r\c\o\n\a\l\a\b ]] + IMAGE='perconalab\/percona-server-mongodb-operator:1.20.1' ++ echo '{"versions":[{"product":"psmdb-operator", "operator":"1.20.1", "matrix":{"mongod":{"7.0.18-11":{"imagePath":"percona/percona-server-mongodb:7.0.18-11", "imageHash":"0115a72f5e60d86cb4f4b7eae32118c0910e8c96831e013de12798a1771c4c91", "imageHashArm64":"86c17067f3e233f522612389ed2500231cbb22ce93524c476b9aa8d464d06f0b", "status":"recommended", "critical":false}}, "pxc":{}, "pmm":{"2.44.1":{"imagePath":"percona/pmm-client:2.44.1", "imageHash":"8b2eaddffd626f02a2d5318ffebc0c277fe8457da6083b8cfcada9b6e6168616", "imageHashArm64":"337fecd4afdb3f6daf2caa2b341b9fe41d0418a0e4ec76980c7f29be9d08b5ea", "status":"available", "critical":false}}, "proxysql":{}, "haproxy":{}, "backup":{"2.9.1":{"imagePath":"percona/percona-backup-mongodb:2.9.1", "imageHash":"976bfbaa548eb70dd90bf0bd2dcfe40b2994d749ef644af3a0590f4856e4d7e2", "imageHashArm64":"ebc6e5c5aa3ed97991d3fd90e9201597b485ddc0eae8d7ee4311ecb785c03bf0", "status":"recommended", "critical":false}}, "operator":{"1.20.1":{"imagePath":"percona/percona-server-mongodb-operator:1.20.1", "imageHash":"b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46", "imageHashArm64":"5a66e497dd1650e5a1123659292fe4c615e0ab5ce7e5d8437bf2101f91b625e1", "status":"recommended", "critical":false}}, "logCollector":{}, "postgresql":{}, "pgbackrest":{}, "pgbackrestRepo":{}, "pgbadger":{}, "pgbouncer":{}, "pxcOperator":{}, "psmdbOperator":{}, "pgOperatorApiserver":{}, "pgOperatorEvent":{}, "pgOperatorRmdata":{}, "pgOperatorScheduler":{}, "pgOperator":{}, "pgOperatorDeployer":{}, "psOperator":{}, "mysql":{}, "router":{}, "orchestrator":{}, "toolkit":{}, "postgis":{}}}]}' ++ jq -r '.versions[].matrix.mongod[].imagePath' + IMAGE_MONGOD=percona/percona-server-mongodb:7.0.18-11 ++ echo '{"versions":[{"product":"psmdb-operator", "operator":"1.20.1", "matrix":{"mongod":{"7.0.18-11":{"imagePath":"percona/percona-server-mongodb:7.0.18-11", "imageHash":"0115a72f5e60d86cb4f4b7eae32118c0910e8c96831e013de12798a1771c4c91", "imageHashArm64":"86c17067f3e233f522612389ed2500231cbb22ce93524c476b9aa8d464d06f0b", "status":"recommended", "critical":false}}, "pxc":{}, "pmm":{"2.44.1":{"imagePath":"percona/pmm-client:2.44.1", "imageHash":"8b2eaddffd626f02a2d5318ffebc0c277fe8457da6083b8cfcada9b6e6168616", "imageHashArm64":"337fecd4afdb3f6daf2caa2b341b9fe41d0418a0e4ec76980c7f29be9d08b5ea", "status":"available", "critical":false}}, "proxysql":{}, "haproxy":{}, "backup":{"2.9.1":{"imagePath":"percona/percona-backup-mongodb:2.9.1", "imageHash":"976bfbaa548eb70dd90bf0bd2dcfe40b2994d749ef644af3a0590f4856e4d7e2", "imageHashArm64":"ebc6e5c5aa3ed97991d3fd90e9201597b485ddc0eae8d7ee4311ecb785c03bf0", "status":"recommended", "critical":false}}, "operator":{"1.20.1":{"imagePath":"percona/percona-server-mongodb-operator:1.20.1", "imageHash":"b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46", "imageHashArm64":"5a66e497dd1650e5a1123659292fe4c615e0ab5ce7e5d8437bf2101f91b625e1", "status":"recommended", "critical":false}}, "logCollector":{}, "postgresql":{}, "pgbackrest":{}, "pgbackrestRepo":{}, "pgbadger":{}, "pgbouncer":{}, "pxcOperator":{}, "psmdbOperator":{}, "pgOperatorApiserver":{}, "pgOperatorEvent":{}, "pgOperatorRmdata":{}, "pgOperatorScheduler":{}, "pgOperator":{}, "pgOperatorDeployer":{}, "psOperator":{}, "mysql":{}, "router":{}, "orchestrator":{}, "toolkit":{}, "postgis":{}}}]}' ++ jq -r '.versions[].matrix.pmm[].imagePath' + IMAGE_PMM_CLIENT=percona/pmm-client:2.44.1 ++ echo '{"versions":[{"product":"psmdb-operator", "operator":"1.20.1", "matrix":{"mongod":{"7.0.18-11":{"imagePath":"percona/percona-server-mongodb:7.0.18-11", "imageHash":"0115a72f5e60d86cb4f4b7eae32118c0910e8c96831e013de12798a1771c4c91", "imageHashArm64":"86c17067f3e233f522612389ed2500231cbb22ce93524c476b9aa8d464d06f0b", "status":"recommended", "critical":false}}, "pxc":{}, "pmm":{"2.44.1":{"imagePath":"percona/pmm-client:2.44.1", "imageHash":"8b2eaddffd626f02a2d5318ffebc0c277fe8457da6083b8cfcada9b6e6168616", "imageHashArm64":"337fecd4afdb3f6daf2caa2b341b9fe41d0418a0e4ec76980c7f29be9d08b5ea", "status":"available", "critical":false}}, "proxysql":{}, "haproxy":{}, "backup":{"2.9.1":{"imagePath":"percona/percona-backup-mongodb:2.9.1", "imageHash":"976bfbaa548eb70dd90bf0bd2dcfe40b2994d749ef644af3a0590f4856e4d7e2", "imageHashArm64":"ebc6e5c5aa3ed97991d3fd90e9201597b485ddc0eae8d7ee4311ecb785c03bf0", "status":"recommended", "critical":false}}, "operator":{"1.20.1":{"imagePath":"percona/percona-server-mongodb-operator:1.20.1", "imageHash":"b559cdd349916d806f6b13b4ac43fdbae982298fad2088b649631a356020ee46", "imageHashArm64":"5a66e497dd1650e5a1123659292fe4c615e0ab5ce7e5d8437bf2101f91b625e1", "status":"recommended", "critical":false}}, "logCollector":{}, "postgresql":{}, "pgbackrest":{}, "pgbackrestRepo":{}, "pgbadger":{}, "pgbouncer":{}, "pxcOperator":{}, "psmdbOperator":{}, "pgOperatorApiserver":{}, "pgOperatorEvent":{}, "pgOperatorRmdata":{}, "pgOperatorScheduler":{}, "pgOperator":{}, "pgOperatorDeployer":{}, "psOperator":{}, "mysql":{}, "router":{}, "orchestrator":{}, "toolkit":{}, "postgis":{}}}]}' ++ jq -r '.versions[].matrix.backup[].imagePath' + IMAGE_BACKUP=percona/percona-backup-mongodb:2.9.1 + [[ 1.21.0 == \1\.\2\0\.\1 ]] + main + rbac=rbac + '[' -n psmdb-operator ']' + rbac=cw-rbac + create_infra_gh upgrade-27940 v1.20.1 + local ns=upgrade-27940 + local git_tag=v1.20.1 + check_crd_for_deletion v1.20.1 + local git_tag=v1.20.1 ++ curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/v1.20.1/deploy/crd.yaml ++ yq eval .metadata.name ++ /usr/bin/sed s/---//g ++ /usr/bin/sed ':a;N;$!ba;s/\n/ /g' + for crd_name in '$(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '\''.metadata.name'\'' | $sed '\''s/---//g'\'' | $sed '\'':a;N;$!ba;s/\n/ /g'\'')' ++ kubectl_bin get crd/perconaservermongodbbackups.psmdb.percona.com -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Hh4WCATF02 +++ mktemp ++ local LAST_ERR=/tmp/tmp.1jTFfZUctN ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/perconaservermongodbbackups.psmdb.percona.com -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Hh4WCATF02 ++ cat /tmp/tmp.1jTFfZUctN ++ rm /tmp/tmp.Hh4WCATF02 /tmp/tmp.1jTFfZUctN ++ return 0 + [[ Established == \T\e\r\m\i\n\a\t\i\n\g ]] + for crd_name in '$(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '\''.metadata.name'\'' | $sed '\''s/---//g'\'' | $sed '\'':a;N;$!ba;s/\n/ /g'\'')' ++ kubectl_bin get crd/perconaservermongodbrestores.psmdb.percona.com -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5Qt7wFYj0w +++ mktemp ++ local LAST_ERR=/tmp/tmp.3QIhkBImiw ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/perconaservermongodbrestores.psmdb.percona.com -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.5Qt7wFYj0w ++ cat /tmp/tmp.3QIhkBImiw ++ rm /tmp/tmp.5Qt7wFYj0w /tmp/tmp.3QIhkBImiw ++ return 0 + [[ Established == \T\e\r\m\i\n\a\t\i\n\g ]] + for crd_name in '$(curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/${git_tag}/deploy/crd.yaml | yq eval '\''.metadata.name'\'' | $sed '\''s/---//g'\'' | $sed '\'':a;N;$!ba;s/\n/ /g'\'')' ++ kubectl_bin get crd/perconaservermongodbs.psmdb.percona.com -o 'jsonpath={.status.conditions[-1].type}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kAGFSkzYeL +++ mktemp ++ local LAST_ERR=/tmp/tmp.XK63Jif83G ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/perconaservermongodbs.psmdb.percona.com -o 'jsonpath={.status.conditions[-1].type}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.kAGFSkzYeL ++ cat /tmp/tmp.XK63Jif83G ++ rm /tmp/tmp.kAGFSkzYeL /tmp/tmp.XK63Jif83G ++ return 0 + [[ Established == \T\e\r\m\i\n\a\t\i\n\g ]] + '[' -n psmdb-operator ']' + create_namespace psmdb-operator + local namespace=psmdb-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces psmdb-operator' + set +o xtrace + xargs kubectl delete ns ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace psmdb-operator --ignore-not-found ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.lqf2anOIf7 + local LAST_OUT=/tmp/tmp.5dN2T9yFq8 ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.5YU4XQncpG + local exit_status=0 + local timeout=4 + local LAST_ERR=/tmp/tmp.AoHJfczZQl + local exit_status=0 + local timeout=4 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + for i in '$(seq 0 2)' + set +e + kubectl delete namespace psmdb-operator --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.lqf2anOIf7 + cat /tmp/tmp.5YU4XQncpG + rm /tmp/tmp.lqf2anOIf7 /tmp/tmp.5YU4XQncpG + return 0 namespace "upgrade-21880" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5dN2T9yFq8 namespace "psmdb-operator" deleted + cat /tmp/tmp.AoHJfczZQl + rm /tmp/tmp.5dN2T9yFq8 /tmp/tmp.AoHJfczZQl + return 0 + kubectl_bin wait --for=delete namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.eJqVUnCAEc ++ mktemp + local LAST_ERR=/tmp/tmp.MD9XH01YIx + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.eJqVUnCAEc + cat /tmp/tmp.MD9XH01YIx + rm /tmp/tmp.eJqVUnCAEc /tmp/tmp.MD9XH01YIx + return 0 + desc 'create namespace psmdb-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.aGgUsRCp5x ++ mktemp + local LAST_ERR=/tmp/tmp.W08xyahoXK + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.aGgUsRCp5x namespace/psmdb-operator created + cat /tmp/tmp.W08xyahoXK + rm /tmp/tmp.aGgUsRCp5x /tmp/tmp.W08xyahoXK + return 0 + set_kube_ctx psmdb-operator + local namespace=psmdb-operator ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.g68v7AaJG4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.vaDEpUFcce ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.g68v7AaJG4 ++ cat /tmp/tmp.vaDEpUFcce ++ rm /tmp/tmp.g68v7AaJG4 /tmp/tmp.vaDEpUFcce ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1961-970eeaa4-5-cluster2 --namespace=psmdb-operator ++ mktemp + local LAST_OUT=/tmp/tmp.tYZvdoeaCZ ++ mktemp + local LAST_ERR=/tmp/tmp.eOGGpaL0zO + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1961-970eeaa4-5-cluster2 --namespace=psmdb-operator + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.tYZvdoeaCZ Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1961-970eeaa4-5-cluster2" modified. + cat /tmp/tmp.eOGGpaL0zO + rm /tmp/tmp.tYZvdoeaCZ /tmp/tmp.eOGGpaL0zO + return 0 + deploy_operator_gh v1.20.1 + local git_tag=v1.20.1 + desc 'start operator' + set +o xtrace ----------------------------------------------------------------------------------- start operator ----------------------------------------------------------------------------------- + kubectl_bin apply -f https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/v1.20.1/deploy/crd.yaml --server-side ++ mktemp + local LAST_OUT=/tmp/tmp.5wNe4uIx6A ++ mktemp + local LAST_ERR=/tmp/tmp.adMLvDq6sh + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/v1.20.1/deploy/crd.yaml --server-side + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.5wNe4uIx6A customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied + cat /tmp/tmp.adMLvDq6sh + rm /tmp/tmp.5wNe4uIx6A /tmp/tmp.adMLvDq6sh + return 0 + local rbac_yaml=rbac + local operator_yaml=operator + '[' -n psmdb-operator ']' + rbac_yaml=cw-rbac + operator_yaml=cw-operator + kubectl_bin apply -f https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/v1.20.1/deploy/cw-rbac.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.YNF9asbQl4 ++ mktemp + local LAST_ERR=/tmp/tmp.xkm1IMgyo3 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/v1.20.1/deploy/cw-rbac.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.YNF9asbQl4 clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator unchanged serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator unchanged + cat /tmp/tmp.xkm1IMgyo3 + rm /tmp/tmp.YNF9asbQl4 /tmp/tmp.xkm1IMgyo3 + return 0 + curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/v1.20.1/deploy/cw-operator.yaml + /usr/bin/sed -i -e 's^image: .*^image: perconalab\/percona-server-mongodb-operator:1.20.1^' /tmp/tmp.hPMIWUh8lp/cw-operator_v1.20.1.yaml + kubectl_bin apply -f /tmp/tmp.hPMIWUh8lp/cw-operator_v1.20.1.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.wFEUN9WpZ5 ++ mktemp + local LAST_ERR=/tmp/tmp.tilbIUJndc + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /tmp/tmp.hPMIWUh8lp/cw-operator_v1.20.1.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.wFEUN9WpZ5 deployment.apps/percona-server-mongodb-operator created + cat /tmp/tmp.tilbIUJndc + rm /tmp/tmp.wFEUN9WpZ5 /tmp/tmp.tilbIUJndc + return 0 + sleep 2 ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.TaUztBHknE +++ mktemp ++ local LAST_ERR=/tmp/tmp.Yq7epJ9kSi ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.TaUztBHknE ++ cat /tmp/tmp.Yq7epJ9kSi ++ rm /tmp/tmp.TaUztBHknE /tmp/tmp.Yq7epJ9kSi ++ return 0 + wait_pod percona-server-mongodb-operator-78d4cb8b56-k2m98 + local pod=percona-server-mongodb-operator-78d4cb8b56-k2m98 + set +o xtrace waiting for pod/percona-server-mongodb-operator-78d4cb8b56-k2m98 to be ready.OK + echo 'Print operator info from log' Print operator info from log + grep 'Manager starting up' ++ get_operator_pod ++ kubectl_bin get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.SmpY0KtIIs +++ mktemp ++ local LAST_ERR=/tmp/tmp.6DiIt0t5OY ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=percona-server-mongodb-operator -o 'jsonpath={.items[].metadata.name}' -n psmdb-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.SmpY0KtIIs ++ cat /tmp/tmp.6DiIt0t5OY ++ rm /tmp/tmp.SmpY0KtIIs /tmp/tmp.6DiIt0t5OY ++ return 0 + kubectl_bin logs percona-server-mongodb-operator-78d4cb8b56-k2m98 ++ mktemp + local LAST_OUT=/tmp/tmp.2Nv1tVqOTI ++ mktemp + local LAST_ERR=/tmp/tmp.xHVJMSdlOJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs percona-server-mongodb-operator-78d4cb8b56-k2m98 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.2Nv1tVqOTI + cat /tmp/tmp.xHVJMSdlOJ + rm /tmp/tmp.2Nv1tVqOTI /tmp/tmp.xHVJMSdlOJ + return 0 2025-08-06T22:15:22.588Z INFO setup Manager starting up {"gitCommit": "30d9ec941baf57619c8973249a3c5d3fd5cc08f4", "gitBranch": "release-1-20-1", "buildTime": "", "goVersion": "go1.24.3", "os": "linux", "arch": "amd64"} + create_namespace upgrade-27940 + local namespace=upgrade-27940 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces upgrade-27940' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- ++ mktemp cleaned up old namespaces upgrade-27940 ----------------------------------------------------------------------------------- + egrep -v '^kube-|^default|Terminating|psmdb-operator|openshift|^gke-|^gmp-|^NAME' + kubectl_bin delete namespace upgrade-27940 --ignore-not-found ++ mktemp + local LAST_OUT=/tmp/tmp.WxxTSEpLlY ++ mktemp + local LAST_OUT=/tmp/tmp.zSqMK0vWOG ++ mktemp + local LAST_ERR=/tmp/tmp.ieReNSmtam + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + local LAST_ERR=/tmp/tmp.u2pmX4QOHy + kubectl get ns + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-27940 --ignore-not-found + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.WxxTSEpLlY + cat /tmp/tmp.ieReNSmtam + rm /tmp/tmp.WxxTSEpLlY /tmp/tmp.ieReNSmtam + return 0 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.zSqMK0vWOG + cat /tmp/tmp.u2pmX4QOHy + rm /tmp/tmp.zSqMK0vWOG /tmp/tmp.u2pmX4QOHy + return 0 + kubectl_bin wait --for=delete namespace upgrade-27940 ++ mktemp + local LAST_OUT=/tmp/tmp.gy75gJakQG ++ mktemp + local LAST_ERR=/tmp/tmp.qQR3TH3wqJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=delete namespace upgrade-27940 error: resource(s) were provided, but no name was specified + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.gy75gJakQG + cat /tmp/tmp.qQR3TH3wqJ + rm /tmp/tmp.gy75gJakQG /tmp/tmp.qQR3TH3wqJ + return 0 + desc 'create namespace upgrade-27940' + set +o xtrace ----------------------------------------------------------------------------------- create namespace upgrade-27940 ----------------------------------------------------------------------------------- + kubectl_bin create namespace upgrade-27940 ++ mktemp + local LAST_OUT=/tmp/tmp.GD8MR63Amc ++ mktemp + local LAST_ERR=/tmp/tmp.ReqFHDiEvJ + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace upgrade-27940 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.GD8MR63Amc namespace/upgrade-27940 created + cat /tmp/tmp.ReqFHDiEvJ + rm /tmp/tmp.GD8MR63Amc /tmp/tmp.ReqFHDiEvJ + return 0 + set_kube_ctx upgrade-27940 + local namespace=upgrade-27940 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.17bXLT55a4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.UQCbE5SAwf ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.17bXLT55a4 ++ cat /tmp/tmp.UQCbE5SAwf ++ rm /tmp/tmp.17bXLT55a4 /tmp/tmp.UQCbE5SAwf ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1961-970eeaa4-5-cluster2 --namespace=upgrade-27940 ++ mktemp + local LAST_OUT=/tmp/tmp.3pVZX4qpHM ++ mktemp + local LAST_ERR=/tmp/tmp.s04FWvUzyj + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-psmdb-1961-970eeaa4-5-cluster2 --namespace=upgrade-27940 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.3pVZX4qpHM Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1961-970eeaa4-5-cluster2" modified. + cat /tmp/tmp.s04FWvUzyj + rm /tmp/tmp.3pVZX4qpHM /tmp/tmp.s04FWvUzyj + return 0 + apply_s3_storage_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1961/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1961/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.AxicXRzmmH ++ mktemp + local LAST_ERR=/tmp/tmp.80PkOOZLza + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1961/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1961/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.AxicXRzmmH secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.80PkOOZLza + rm /tmp/tmp.AxicXRzmmH /tmp/tmp.80PkOOZLza + return 0 + deploy_minio + desc 'install Minio' + set +o xtrace ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- + helm uninstall minio-service Error: uninstall: Release not loaded: minio-service: release: not found + : + helm repo remove minio "minio" has been removed from your repositories + helm repo add minio https://charts.min.io/ "minio" has been added to your repositories + retry 10 60 helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio + local max=10 + local delay=60 + shift 2 + local n=1 + helm install minio-service --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio NAME: minio-service LAST DEPLOYED: Wed Aug 6 22:15:40 2025 NAMESPACE: upgrade-27940 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.upgrade-27940.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace upgrade-27940 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace upgrade-27940 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace upgrade-27940 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace upgrade-27940 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local ++ kubectl_bin get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ai1ZJZR12h +++ mktemp ++ local LAST_ERR=/tmp/tmp.V1HMgeMxKx ++ local exit_status=0 ++ local timeout=4 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 -a -n 1 ']' ++ break ++ cat /tmp/tmp.Ai1ZJZR12h ++ cat /tmp/tmp.V1HMgeMxKx ++ rm /tmp/tmp.Ai1ZJZR12h /tmp/tmp.V1HMgeMxKx ++ return 0 + MINIO_POD=minio-service-86dfccd949-rhlpx + wait_pod minio-service-86dfccd949-rhlpx + local pod=minio-service-86dfccd949-rhlpx + set +o xtrace waiting for pod/minio-service-86dfccd949-rhlpx to be ready.OK + '[' -n psmdb-operator ']' + kubectl_bin create svc -n psmdb-operator externalname minio-service --external-name=minio-service.upgrade-27940.svc.cluster.local --tcp=9000 ++ mktemp + local LAST_OUT=/tmp/tmp.ysIf0fOTA8 ++ mktemp + local LAST_ERR=/tmp/tmp.CO5syJcBG2 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create svc -n psmdb-operator externalname minio-service --external-name=minio-service.upgrade-27940.svc.cluster.local --tcp=9000 + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.ysIf0fOTA8 service/minio-service created + cat /tmp/tmp.CO5syJcBG2 + rm /tmp/tmp.ysIf0fOTA8 /tmp/tmp.CO5syJcBG2 + return 0 + create_minio_bucket operator-testing + local bucket=operator-testing + kubectl_bin run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' ++ mktemp + local LAST_OUT=/tmp/tmp.YAX9J0WaeE ++ mktemp + local LAST_ERR=/tmp/tmp.ZNqIVP02b0 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID=some-access-key AWS_SECRET_ACCESS_KEY=some-secret-key AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.YAX9J0WaeE make_bucket: operator-testing pod "aws-cli" deleted + cat /tmp/tmp.ZNqIVP02b0 If you don't see a command prompt, try pressing enter. warning: couldn't attach to pod/aws-cli, falling back to streaming logs: Internal error occurred: unable to upgrade connection: container aws-cli not found in pod aws-cli_upgrade-27940 + rm /tmp/tmp.YAX9J0WaeE /tmp/tmp.ZNqIVP02b0 + return 0 + desc 'create secrets and start client' + set +o xtrace ----------------------------------------------------------------------------------- create secrets and start client ----------------------------------------------------------------------------------- + curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/v1.20.1/deploy/secrets.yaml + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1961/e2e-tests/conf/client.yml -f /tmp/tmp.hPMIWUh8lp/secrets.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.mkiUNluzaX ++ mktemp + local LAST_ERR=/tmp/tmp.3yldebC3b1 + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1961/e2e-tests/conf/client.yml -f /tmp/tmp.hPMIWUh8lp/secrets.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.mkiUNluzaX deployment.apps/psmdb-client created secret/my-cluster-name-secrets created + cat /tmp/tmp.3yldebC3b1 + rm /tmp/tmp.mkiUNluzaX /tmp/tmp.3yldebC3b1 + return 0 + local cr_yaml=/tmp/tmp.hPMIWUh8lp/cr_v1.20.1.yaml + prepare_cr_yaml /tmp/tmp.hPMIWUh8lp/cr_v1.20.1.yaml + local cr_yaml=/tmp/tmp.hPMIWUh8lp/cr_v1.20.1.yaml + curl -s https://raw.githubusercontent.com/percona/percona-server-mongodb-operator/v1.20.1/deploy/cr.yaml + yq eval ' .metadata.name = "upgrade" | .spec.upgradeOptions.apply = "disabled" | .spec.replsets[].size = 3 | .spec.replsets[].arbiter.enabled = false | .spec.backup.enabled = true | .spec.backup.pitr.enabled = false | .spec.backup.storages.minio.type = "s3" | .spec.backup.storages.minio.s3.credentialsSecret = "minio-secret" | .spec.backup.storages.minio.s3.region = "us-east-1" | .spec.backup.storages.minio.s3.bucket = "operator-testing" | .spec.backup.storages.minio.s3.endpointUrl = "http://minio-service:9000/" | .spec.sharding.enabled = false | .spec.image="" | .spec.image tag="!!null" | .spec.backup.image = "-backup" | .spec.pmm.image = "-pmm"' + desc 'create first PSMDB cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- + apply_cluster /tmp/tmp.hPMIWUh8lp/cr_v1.20.1.yaml + '[' -z '' ']' + cat_config /tmp/tmp.hPMIWUh8lp/cr_v1.20.1.yaml + kubectl_bin apply -f - + cat /tmp/tmp.hPMIWUh8lp/cr_v1.20.1.yaml + yq eval '(.spec | select(.image == null)).image = "percona/percona-server-mongodb:7.0.18-11"' + yq eval '(.spec | select(has("initImage"))).initImage = "perconalab\/percona-server-mongodb-operator:1.20.1"' + yq eval '(.spec | select(has("pmm"))).pmm.image = "percona/pmm-client:2.44.1"' + yq eval '(.spec | select(has("backup"))).backup.image = "percona/percona-backup-mongodb:2.9.1"' ++ mktemp + yq eval '.spec.upgradeOptions.apply="Never"' + local LAST_OUT=/tmp/tmp.NiJNGCn0tM ++ mktemp + local LAST_ERR=/tmp/tmp.fn1UBZCADm + local exit_status=0 + local timeout=4 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 -a -n 1 ']' + break + cat /tmp/tmp.NiJNGCn0tM perconaservermongodb.psmdb.percona.com/upgrade created + cat /tmp/tmp.fn1UBZCADm + rm /tmp/tmp.NiJNGCn0tM /tmp/tmp.fn1UBZCADm + return 0 + desc 'check if Pod is started' + set +o xtrace ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- + wait_for_running upgrade-rs0 3 + local name=upgrade-rs0 + let last_pod=2 + local check_cluster_readyness=true + set_debug + [[ 1 == 1 ]] + set -o xtrace + local rs_name=rs0 + local cluster_name=upgrade ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + [[ 0 -eq 2 ]] + wait_pod upgrade-rs0-0 + local pod=upgrade-rs0-0 + set +o xtrace waiting for pod/upgrade-rs0-0 to be ready..........OK + for i in '$(seq 0 $last_pod)' + [[ 1 -eq 2 ]] + wait_pod upgrade-rs0-1 + local pod=upgrade-rs0-1 + set +o xtrace waiting for pod/upgrade-rs0-1 to be ready