Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/logs/upgrade-proxysql-8-0.log Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.32) exceeds the supported minor version skew of +/-1 + CLUSTER=upgrade-proxysql + CLUSTER_SIZE=3 + TARGET_OPERATOR_VER=1.20.0 + TARGET_IMAGE=perconalab/percona-xtradb-cluster-operator:PR-2407-e9e16845 + TARGET_IMAGE_PXC=perconalab/percona-xtradb-cluster-operator:main-pxc8.0 + TARGET_IMAGE_PMM_CLIENT=perconalab/pmm-client:dev-latest + TARGET_IMAGE_PROXY=perconalab/percona-xtradb-cluster-operator:main-proxysql + TARGET_IMAGE_HAPROXY=perconalab/percona-xtradb-cluster-operator:main-haproxy + TARGET_IMAGE_BACKUP=perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup + TARGET_IMAGE_LOGCOLLECTOR=perconalab/fluentbit:main-logcollector + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 == *\p\e\r\c\o\n\a\-\x\t\r\a\d\b\-\c\l\u\s\t\e\r\-\o\p\e\r\a\t\o\r* ]] ++ /usr/bin/sed -r 's/.*([0-9].[0-9])$/\1/' ++ echo -n perconalab/percona-xtradb-cluster-operator:main-pxc8.0 + PXC_VER=8.0 ++ curl -s https://check.percona.com/versions/v1/pxc-operator ++ sort -V ++ tail -n1 ++ jq -r '.versions[].operator' + INIT_OPERATOR_VER=1.19.1 + [[ 1.19.1 == \1\.\2\0\.\0 ]] + GIT_TAG=v1.19.1 ++ curl -s 'https://check.percona.com/versions/v1/pxc-operator/1.19.1/latest?databaseVersion=8.0' + INIT_OPERATOR_IMAGES='{"versions":[{"product":"pxc-operator","operator":"1.19.1","matrix":{"mongod":{},"pxc":{"8.0.44-35.1":{"imagePath":"percona/percona-xtradb-cluster:8.0.44-35.1","imageHash":"34d7e23a42b9d009039a1a51f3068623320a002eac3498b0030cc3c6be0e36ac","imageHashArm64":"4e1bb902cc8fb973134f828d05c785dd4fd32a94e30da773aadc8aab1f7e358a","status":"recommended","critical":false}},"pmm":{"2.44.1-1":{"imagePath":"percona/pmm-client:2.44.1-1","imageHash":"52a8fb5e8f912eef1ff8a117ea323c401e278908ce29928dafc23fac1db4f1e3","imageHashArm64":"390bfd12f981e8b3890550c4927a3ece071377065e001894458047602c744e3b","status":"recommended","critical":false},"3.5.0":{"imagePath":"percona/pmm-client:3.5.0","imageHash":"352aee74f25b3c1c4cd9dff1f378a0c3940b315e551d170c09953bf168531e4a","imageHashArm64":"cbbb074d51d90a5f2d6f1d98a05024f6de2ffdcb5acab632324cea4349a820bd","status":"recommended","critical":false}},"proxysql":{"2.7.3-1.2":{"imagePath":"percona/proxysql2:2.7.3-1.2","imageHash":"719d0ab363c65c7f75431bbed7ec0d9f2af7e691765c489da954813c552359a2","imageHashArm64":"4c4d094652c9f2eb097be5d92dcc05da61c9e8699ac7321def959d5a205a89f7","status":"recommended","critical":false}},"haproxy":{"2.8.17":{"imagePath":"percona/haproxy:2.8.17","imageHash":"ef8486b39a1e8dca97b5cdf1135e6282be1757ad188517b889d12c5a3470eeda","imageHashArm64":"bbc5b3b66ac985d1a4500195539e7dff5196245a5a842a6858ea0848ec089967","status":"recommended","critical":false}},"backup":{"8.0.35-34.1":{"imagePath":"percona/percona-xtrabackup:8.0.35-34.1","imageHash":"967bafa0823c90aa8fa9c25a9012be36b0deef64e255294a09148d77ce6aea68","imageHashArm64":"83f814dca9ed398b585938baa86508bda796ba301e34c948a5106095d27bf86e","status":"recommended","critical":false}},"operator":{"1.19.1":{"imagePath":"percona/percona-xtradb-cluster-operator:1.19.1","imageHash":"937a583f0031f45175aee2e5c7a380bbaa9a35046f0e7cae8cd900c44732c4f8","imageHashArm64":"d1939659a5d331db9fae970dee42fd25f083d1955330e9043fe4cb862ff19326","status":"recommended","critical":false}},"logCollector":{"4.0.1-1":{"imagePath":"percona/fluentbit:4.0.1-1","imageHash":"65bdf7d38cbceed6b6aa6412aea3fb4a196000ac6c66185f114a0a62c4a442ad","imageHashArm64":"dabda77b298b67d30d7f53b5cdb7215ad19dabb22b9543e3fd8aedb74ab24733","status":"recommended","critical":false}},"postgresql":{},"pgbackrest":{},"pgbackrestRepo":{},"pgbadger":{},"pgbouncer":{},"pxcOperator":{},"psmdbOperator":{},"pgOperatorApiserver":{},"pgOperatorEvent":{},"pgOperatorRmdata":{},"pgOperatorScheduler":{},"pgOperator":{},"pgOperatorDeployer":{},"psOperator":{},"mysql":{},"router":{},"orchestrator":{},"toolkit":{},"postgis":{},"binlogServer":{}}}]}' + OPERATOR_NAME=percona-xtradb-cluster-operator ++ jq -r '.versions[].matrix.operator[].imagePath' ++ echo '{"versions":[{"product":"pxc-operator","operator":"1.19.1","matrix":{"mongod":{},"pxc":{"8.0.44-35.1":{"imagePath":"percona/percona-xtradb-cluster:8.0.44-35.1","imageHash":"34d7e23a42b9d009039a1a51f3068623320a002eac3498b0030cc3c6be0e36ac","imageHashArm64":"4e1bb902cc8fb973134f828d05c785dd4fd32a94e30da773aadc8aab1f7e358a","status":"recommended","critical":false}},"pmm":{"2.44.1-1":{"imagePath":"percona/pmm-client:2.44.1-1","imageHash":"52a8fb5e8f912eef1ff8a117ea323c401e278908ce29928dafc23fac1db4f1e3","imageHashArm64":"390bfd12f981e8b3890550c4927a3ece071377065e001894458047602c744e3b","status":"recommended","critical":false},"3.5.0":{"imagePath":"percona/pmm-client:3.5.0","imageHash":"352aee74f25b3c1c4cd9dff1f378a0c3940b315e551d170c09953bf168531e4a","imageHashArm64":"cbbb074d51d90a5f2d6f1d98a05024f6de2ffdcb5acab632324cea4349a820bd","status":"recommended","critical":false}},"proxysql":{"2.7.3-1.2":{"imagePath":"percona/proxysql2:2.7.3-1.2","imageHash":"719d0ab363c65c7f75431bbed7ec0d9f2af7e691765c489da954813c552359a2","imageHashArm64":"4c4d094652c9f2eb097be5d92dcc05da61c9e8699ac7321def959d5a205a89f7","status":"recommended","critical":false}},"haproxy":{"2.8.17":{"imagePath":"percona/haproxy:2.8.17","imageHash":"ef8486b39a1e8dca97b5cdf1135e6282be1757ad188517b889d12c5a3470eeda","imageHashArm64":"bbc5b3b66ac985d1a4500195539e7dff5196245a5a842a6858ea0848ec089967","status":"recommended","critical":false}},"backup":{"8.0.35-34.1":{"imagePath":"percona/percona-xtrabackup:8.0.35-34.1","imageHash":"967bafa0823c90aa8fa9c25a9012be36b0deef64e255294a09148d77ce6aea68","imageHashArm64":"83f814dca9ed398b585938baa86508bda796ba301e34c948a5106095d27bf86e","status":"recommended","critical":false}},"operator":{"1.19.1":{"imagePath":"percona/percona-xtradb-cluster-operator:1.19.1","imageHash":"937a583f0031f45175aee2e5c7a380bbaa9a35046f0e7cae8cd900c44732c4f8","imageHashArm64":"d1939659a5d331db9fae970dee42fd25f083d1955330e9043fe4cb862ff19326","status":"recommended","critical":false}},"logCollector":{"4.0.1-1":{"imagePath":"percona/fluentbit:4.0.1-1","imageHash":"65bdf7d38cbceed6b6aa6412aea3fb4a196000ac6c66185f114a0a62c4a442ad","imageHashArm64":"dabda77b298b67d30d7f53b5cdb7215ad19dabb22b9543e3fd8aedb74ab24733","status":"recommended","critical":false}},"postgresql":{},"pgbackrest":{},"pgbackrestRepo":{},"pgbadger":{},"pgbouncer":{},"pxcOperator":{},"psmdbOperator":{},"pgOperatorApiserver":{},"pgOperatorEvent":{},"pgOperatorRmdata":{},"pgOperatorScheduler":{},"pgOperator":{},"pgOperatorDeployer":{},"psOperator":{},"mysql":{},"router":{},"orchestrator":{},"toolkit":{},"postgis":{},"binlogServer":{}}}]}' + IMAGE=percona/percona-xtradb-cluster-operator:1.19.1 ++ echo perconalab/percona-xtradb-cluster-operator:PR-2407-e9e16845 ++ cut -d/ -f1 + [[ perconalab == \p\e\r\c\o\n\a\l\a\b ]] + IMAGE=perconalab/percona-xtradb-cluster-operator:1.19.1 ++ jq -r '.versions[].matrix.pxc[].imagePath' ++ echo '{"versions":[{"product":"pxc-operator","operator":"1.19.1","matrix":{"mongod":{},"pxc":{"8.0.44-35.1":{"imagePath":"percona/percona-xtradb-cluster:8.0.44-35.1","imageHash":"34d7e23a42b9d009039a1a51f3068623320a002eac3498b0030cc3c6be0e36ac","imageHashArm64":"4e1bb902cc8fb973134f828d05c785dd4fd32a94e30da773aadc8aab1f7e358a","status":"recommended","critical":false}},"pmm":{"2.44.1-1":{"imagePath":"percona/pmm-client:2.44.1-1","imageHash":"52a8fb5e8f912eef1ff8a117ea323c401e278908ce29928dafc23fac1db4f1e3","imageHashArm64":"390bfd12f981e8b3890550c4927a3ece071377065e001894458047602c744e3b","status":"recommended","critical":false},"3.5.0":{"imagePath":"percona/pmm-client:3.5.0","imageHash":"352aee74f25b3c1c4cd9dff1f378a0c3940b315e551d170c09953bf168531e4a","imageHashArm64":"cbbb074d51d90a5f2d6f1d98a05024f6de2ffdcb5acab632324cea4349a820bd","status":"recommended","critical":false}},"proxysql":{"2.7.3-1.2":{"imagePath":"percona/proxysql2:2.7.3-1.2","imageHash":"719d0ab363c65c7f75431bbed7ec0d9f2af7e691765c489da954813c552359a2","imageHashArm64":"4c4d094652c9f2eb097be5d92dcc05da61c9e8699ac7321def959d5a205a89f7","status":"recommended","critical":false}},"haproxy":{"2.8.17":{"imagePath":"percona/haproxy:2.8.17","imageHash":"ef8486b39a1e8dca97b5cdf1135e6282be1757ad188517b889d12c5a3470eeda","imageHashArm64":"bbc5b3b66ac985d1a4500195539e7dff5196245a5a842a6858ea0848ec089967","status":"recommended","critical":false}},"backup":{"8.0.35-34.1":{"imagePath":"percona/percona-xtrabackup:8.0.35-34.1","imageHash":"967bafa0823c90aa8fa9c25a9012be36b0deef64e255294a09148d77ce6aea68","imageHashArm64":"83f814dca9ed398b585938baa86508bda796ba301e34c948a5106095d27bf86e","status":"recommended","critical":false}},"operator":{"1.19.1":{"imagePath":"percona/percona-xtradb-cluster-operator:1.19.1","imageHash":"937a583f0031f45175aee2e5c7a380bbaa9a35046f0e7cae8cd900c44732c4f8","imageHashArm64":"d1939659a5d331db9fae970dee42fd25f083d1955330e9043fe4cb862ff19326","status":"recommended","critical":false}},"logCollector":{"4.0.1-1":{"imagePath":"percona/fluentbit:4.0.1-1","imageHash":"65bdf7d38cbceed6b6aa6412aea3fb4a196000ac6c66185f114a0a62c4a442ad","imageHashArm64":"dabda77b298b67d30d7f53b5cdb7215ad19dabb22b9543e3fd8aedb74ab24733","status":"recommended","critical":false}},"postgresql":{},"pgbackrest":{},"pgbackrestRepo":{},"pgbadger":{},"pgbouncer":{},"pxcOperator":{},"psmdbOperator":{},"pgOperatorApiserver":{},"pgOperatorEvent":{},"pgOperatorRmdata":{},"pgOperatorScheduler":{},"pgOperator":{},"pgOperatorDeployer":{},"psOperator":{},"mysql":{},"router":{},"orchestrator":{},"toolkit":{},"postgis":{},"binlogServer":{}}}]}' + IMAGE_PXC=percona/percona-xtradb-cluster:8.0.44-35.1 ++ jq -r '.versions[].matrix.pmm[].imagePath' ++ echo '{"versions":[{"product":"pxc-operator","operator":"1.19.1","matrix":{"mongod":{},"pxc":{"8.0.44-35.1":{"imagePath":"percona/percona-xtradb-cluster:8.0.44-35.1","imageHash":"34d7e23a42b9d009039a1a51f3068623320a002eac3498b0030cc3c6be0e36ac","imageHashArm64":"4e1bb902cc8fb973134f828d05c785dd4fd32a94e30da773aadc8aab1f7e358a","status":"recommended","critical":false}},"pmm":{"2.44.1-1":{"imagePath":"percona/pmm-client:2.44.1-1","imageHash":"52a8fb5e8f912eef1ff8a117ea323c401e278908ce29928dafc23fac1db4f1e3","imageHashArm64":"390bfd12f981e8b3890550c4927a3ece071377065e001894458047602c744e3b","status":"recommended","critical":false},"3.5.0":{"imagePath":"percona/pmm-client:3.5.0","imageHash":"352aee74f25b3c1c4cd9dff1f378a0c3940b315e551d170c09953bf168531e4a","imageHashArm64":"cbbb074d51d90a5f2d6f1d98a05024f6de2ffdcb5acab632324cea4349a820bd","status":"recommended","critical":false}},"proxysql":{"2.7.3-1.2":{"imagePath":"percona/proxysql2:2.7.3-1.2","imageHash":"719d0ab363c65c7f75431bbed7ec0d9f2af7e691765c489da954813c552359a2","imageHashArm64":"4c4d094652c9f2eb097be5d92dcc05da61c9e8699ac7321def959d5a205a89f7","status":"recommended","critical":false}},"haproxy":{"2.8.17":{"imagePath":"percona/haproxy:2.8.17","imageHash":"ef8486b39a1e8dca97b5cdf1135e6282be1757ad188517b889d12c5a3470eeda","imageHashArm64":"bbc5b3b66ac985d1a4500195539e7dff5196245a5a842a6858ea0848ec089967","status":"recommended","critical":false}},"backup":{"8.0.35-34.1":{"imagePath":"percona/percona-xtrabackup:8.0.35-34.1","imageHash":"967bafa0823c90aa8fa9c25a9012be36b0deef64e255294a09148d77ce6aea68","imageHashArm64":"83f814dca9ed398b585938baa86508bda796ba301e34c948a5106095d27bf86e","status":"recommended","critical":false}},"operator":{"1.19.1":{"imagePath":"percona/percona-xtradb-cluster-operator:1.19.1","imageHash":"937a583f0031f45175aee2e5c7a380bbaa9a35046f0e7cae8cd900c44732c4f8","imageHashArm64":"d1939659a5d331db9fae970dee42fd25f083d1955330e9043fe4cb862ff19326","status":"recommended","critical":false}},"logCollector":{"4.0.1-1":{"imagePath":"percona/fluentbit:4.0.1-1","imageHash":"65bdf7d38cbceed6b6aa6412aea3fb4a196000ac6c66185f114a0a62c4a442ad","imageHashArm64":"dabda77b298b67d30d7f53b5cdb7215ad19dabb22b9543e3fd8aedb74ab24733","status":"recommended","critical":false}},"postgresql":{},"pgbackrest":{},"pgbackrestRepo":{},"pgbadger":{},"pgbouncer":{},"pxcOperator":{},"psmdbOperator":{},"pgOperatorApiserver":{},"pgOperatorEvent":{},"pgOperatorRmdata":{},"pgOperatorScheduler":{},"pgOperator":{},"pgOperatorDeployer":{},"psOperator":{},"mysql":{},"router":{},"orchestrator":{},"toolkit":{},"postgis":{},"binlogServer":{}}}]}' ++ tail -n1 + IMAGE_PMM_CLIENT=percona/pmm-client:3.5.0 ++ jq -r '.versions[].matrix.proxysql[].imagePath' ++ echo '{"versions":[{"product":"pxc-operator","operator":"1.19.1","matrix":{"mongod":{},"pxc":{"8.0.44-35.1":{"imagePath":"percona/percona-xtradb-cluster:8.0.44-35.1","imageHash":"34d7e23a42b9d009039a1a51f3068623320a002eac3498b0030cc3c6be0e36ac","imageHashArm64":"4e1bb902cc8fb973134f828d05c785dd4fd32a94e30da773aadc8aab1f7e358a","status":"recommended","critical":false}},"pmm":{"2.44.1-1":{"imagePath":"percona/pmm-client:2.44.1-1","imageHash":"52a8fb5e8f912eef1ff8a117ea323c401e278908ce29928dafc23fac1db4f1e3","imageHashArm64":"390bfd12f981e8b3890550c4927a3ece071377065e001894458047602c744e3b","status":"recommended","critical":false},"3.5.0":{"imagePath":"percona/pmm-client:3.5.0","imageHash":"352aee74f25b3c1c4cd9dff1f378a0c3940b315e551d170c09953bf168531e4a","imageHashArm64":"cbbb074d51d90a5f2d6f1d98a05024f6de2ffdcb5acab632324cea4349a820bd","status":"recommended","critical":false}},"proxysql":{"2.7.3-1.2":{"imagePath":"percona/proxysql2:2.7.3-1.2","imageHash":"719d0ab363c65c7f75431bbed7ec0d9f2af7e691765c489da954813c552359a2","imageHashArm64":"4c4d094652c9f2eb097be5d92dcc05da61c9e8699ac7321def959d5a205a89f7","status":"recommended","critical":false}},"haproxy":{"2.8.17":{"imagePath":"percona/haproxy:2.8.17","imageHash":"ef8486b39a1e8dca97b5cdf1135e6282be1757ad188517b889d12c5a3470eeda","imageHashArm64":"bbc5b3b66ac985d1a4500195539e7dff5196245a5a842a6858ea0848ec089967","status":"recommended","critical":false}},"backup":{"8.0.35-34.1":{"imagePath":"percona/percona-xtrabackup:8.0.35-34.1","imageHash":"967bafa0823c90aa8fa9c25a9012be36b0deef64e255294a09148d77ce6aea68","imageHashArm64":"83f814dca9ed398b585938baa86508bda796ba301e34c948a5106095d27bf86e","status":"recommended","critical":false}},"operator":{"1.19.1":{"imagePath":"percona/percona-xtradb-cluster-operator:1.19.1","imageHash":"937a583f0031f45175aee2e5c7a380bbaa9a35046f0e7cae8cd900c44732c4f8","imageHashArm64":"d1939659a5d331db9fae970dee42fd25f083d1955330e9043fe4cb862ff19326","status":"recommended","critical":false}},"logCollector":{"4.0.1-1":{"imagePath":"percona/fluentbit:4.0.1-1","imageHash":"65bdf7d38cbceed6b6aa6412aea3fb4a196000ac6c66185f114a0a62c4a442ad","imageHashArm64":"dabda77b298b67d30d7f53b5cdb7215ad19dabb22b9543e3fd8aedb74ab24733","status":"recommended","critical":false}},"postgresql":{},"pgbackrest":{},"pgbackrestRepo":{},"pgbadger":{},"pgbouncer":{},"pxcOperator":{},"psmdbOperator":{},"pgOperatorApiserver":{},"pgOperatorEvent":{},"pgOperatorRmdata":{},"pgOperatorScheduler":{},"pgOperator":{},"pgOperatorDeployer":{},"psOperator":{},"mysql":{},"router":{},"orchestrator":{},"toolkit":{},"postgis":{},"binlogServer":{}}}]}' + IMAGE_PROXY=percona/proxysql2:2.7.3-1.2 ++ jq -r '.versions[].matrix.haproxy[].imagePath' ++ echo '{"versions":[{"product":"pxc-operator","operator":"1.19.1","matrix":{"mongod":{},"pxc":{"8.0.44-35.1":{"imagePath":"percona/percona-xtradb-cluster:8.0.44-35.1","imageHash":"34d7e23a42b9d009039a1a51f3068623320a002eac3498b0030cc3c6be0e36ac","imageHashArm64":"4e1bb902cc8fb973134f828d05c785dd4fd32a94e30da773aadc8aab1f7e358a","status":"recommended","critical":false}},"pmm":{"2.44.1-1":{"imagePath":"percona/pmm-client:2.44.1-1","imageHash":"52a8fb5e8f912eef1ff8a117ea323c401e278908ce29928dafc23fac1db4f1e3","imageHashArm64":"390bfd12f981e8b3890550c4927a3ece071377065e001894458047602c744e3b","status":"recommended","critical":false},"3.5.0":{"imagePath":"percona/pmm-client:3.5.0","imageHash":"352aee74f25b3c1c4cd9dff1f378a0c3940b315e551d170c09953bf168531e4a","imageHashArm64":"cbbb074d51d90a5f2d6f1d98a05024f6de2ffdcb5acab632324cea4349a820bd","status":"recommended","critical":false}},"proxysql":{"2.7.3-1.2":{"imagePath":"percona/proxysql2:2.7.3-1.2","imageHash":"719d0ab363c65c7f75431bbed7ec0d9f2af7e691765c489da954813c552359a2","imageHashArm64":"4c4d094652c9f2eb097be5d92dcc05da61c9e8699ac7321def959d5a205a89f7","status":"recommended","critical":false}},"haproxy":{"2.8.17":{"imagePath":"percona/haproxy:2.8.17","imageHash":"ef8486b39a1e8dca97b5cdf1135e6282be1757ad188517b889d12c5a3470eeda","imageHashArm64":"bbc5b3b66ac985d1a4500195539e7dff5196245a5a842a6858ea0848ec089967","status":"recommended","critical":false}},"backup":{"8.0.35-34.1":{"imagePath":"percona/percona-xtrabackup:8.0.35-34.1","imageHash":"967bafa0823c90aa8fa9c25a9012be36b0deef64e255294a09148d77ce6aea68","imageHashArm64":"83f814dca9ed398b585938baa86508bda796ba301e34c948a5106095d27bf86e","status":"recommended","critical":false}},"operator":{"1.19.1":{"imagePath":"percona/percona-xtradb-cluster-operator:1.19.1","imageHash":"937a583f0031f45175aee2e5c7a380bbaa9a35046f0e7cae8cd900c44732c4f8","imageHashArm64":"d1939659a5d331db9fae970dee42fd25f083d1955330e9043fe4cb862ff19326","status":"recommended","critical":false}},"logCollector":{"4.0.1-1":{"imagePath":"percona/fluentbit:4.0.1-1","imageHash":"65bdf7d38cbceed6b6aa6412aea3fb4a196000ac6c66185f114a0a62c4a442ad","imageHashArm64":"dabda77b298b67d30d7f53b5cdb7215ad19dabb22b9543e3fd8aedb74ab24733","status":"recommended","critical":false}},"postgresql":{},"pgbackrest":{},"pgbackrestRepo":{},"pgbadger":{},"pgbouncer":{},"pxcOperator":{},"psmdbOperator":{},"pgOperatorApiserver":{},"pgOperatorEvent":{},"pgOperatorRmdata":{},"pgOperatorScheduler":{},"pgOperator":{},"pgOperatorDeployer":{},"psOperator":{},"mysql":{},"router":{},"orchestrator":{},"toolkit":{},"postgis":{},"binlogServer":{}}}]}' + IMAGE_HAPROXY=percona/haproxy:2.8.17 ++ echo '{"versions":[{"product":"pxc-operator","operator":"1.19.1","matrix":{"mongod":{},"pxc":{"8.0.44-35.1":{"imagePath":"percona/percona-xtradb-cluster:8.0.44-35.1","imageHash":"34d7e23a42b9d009039a1a51f3068623320a002eac3498b0030cc3c6be0e36ac","imageHashArm64":"4e1bb902cc8fb973134f828d05c785dd4fd32a94e30da773aadc8aab1f7e358a","status":"recommended","critical":false}},"pmm":{"2.44.1-1":{"imagePath":"percona/pmm-client:2.44.1-1","imageHash":"52a8fb5e8f912eef1ff8a117ea323c401e278908ce29928dafc23fac1db4f1e3","imageHashArm64":"390bfd12f981e8b3890550c4927a3ece071377065e001894458047602c744e3b","status":"recommended","critical":false},"3.5.0":{"imagePath":"percona/pmm-client:3.5.0","imageHash":"352aee74f25b3c1c4cd9dff1f378a0c3940b315e551d170c09953bf168531e4a","imageHashArm64":"cbbb074d51d90a5f2d6f1d98a05024f6de2ffdcb5acab632324cea4349a820bd","status":"recommended","critical":false}},"proxysql":{"2.7.3-1.2":{"imagePath":"percona/proxysql2:2.7.3-1.2","imageHash":"719d0ab363c65c7f75431bbed7ec0d9f2af7e691765c489da954813c552359a2","imageHashArm64":"4c4d094652c9f2eb097be5d92dcc05da61c9e8699ac7321def959d5a205a89f7","status":"recommended","critical":false}},"haproxy":{"2.8.17":{"imagePath":"percona/haproxy:2.8.17","imageHash":"ef8486b39a1e8dca97b5cdf1135e6282be1757ad188517b889d12c5a3470eeda","imageHashArm64":"bbc5b3b66ac985d1a4500195539e7dff5196245a5a842a6858ea0848ec089967","status":"recommended","critical":false}},"backup":{"8.0.35-34.1":{"imagePath":"percona/percona-xtrabackup:8.0.35-34.1","imageHash":"967bafa0823c90aa8fa9c25a9012be36b0deef64e255294a09148d77ce6aea68","imageHashArm64":"83f814dca9ed398b585938baa86508bda796ba301e34c948a5106095d27bf86e","status":"recommended","critical":false}},"operator":{"1.19.1":{"imagePath":"percona/percona-xtradb-cluster-operator:1.19.1","imageHash":"937a583f0031f45175aee2e5c7a380bbaa9a35046f0e7cae8cd900c44732c4f8","imageHashArm64":"d1939659a5d331db9fae970dee42fd25f083d1955330e9043fe4cb862ff19326","status":"recommended","critical":false}},"logCollector":{"4.0.1-1":{"imagePath":"percona/fluentbit:4.0.1-1","imageHash":"65bdf7d38cbceed6b6aa6412aea3fb4a196000ac6c66185f114a0a62c4a442ad","imageHashArm64":"dabda77b298b67d30d7f53b5cdb7215ad19dabb22b9543e3fd8aedb74ab24733","status":"recommended","critical":false}},"postgresql":{},"pgbackrest":{},"pgbackrestRepo":{},"pgbadger":{},"pgbouncer":{},"pxcOperator":{},"psmdbOperator":{},"pgOperatorApiserver":{},"pgOperatorEvent":{},"pgOperatorRmdata":{},"pgOperatorScheduler":{},"pgOperator":{},"pgOperatorDeployer":{},"psOperator":{},"mysql":{},"router":{},"orchestrator":{},"toolkit":{},"postgis":{},"binlogServer":{}}}]}' ++ jq -r '.versions[].matrix.backup[].imagePath' + IMAGE_BACKUP=percona/percona-xtrabackup:8.0.35-34.1 ++ jq -r '.versions[].matrix.logCollector[].imagePath' ++ echo '{"versions":[{"product":"pxc-operator","operator":"1.19.1","matrix":{"mongod":{},"pxc":{"8.0.44-35.1":{"imagePath":"percona/percona-xtradb-cluster:8.0.44-35.1","imageHash":"34d7e23a42b9d009039a1a51f3068623320a002eac3498b0030cc3c6be0e36ac","imageHashArm64":"4e1bb902cc8fb973134f828d05c785dd4fd32a94e30da773aadc8aab1f7e358a","status":"recommended","critical":false}},"pmm":{"2.44.1-1":{"imagePath":"percona/pmm-client:2.44.1-1","imageHash":"52a8fb5e8f912eef1ff8a117ea323c401e278908ce29928dafc23fac1db4f1e3","imageHashArm64":"390bfd12f981e8b3890550c4927a3ece071377065e001894458047602c744e3b","status":"recommended","critical":false},"3.5.0":{"imagePath":"percona/pmm-client:3.5.0","imageHash":"352aee74f25b3c1c4cd9dff1f378a0c3940b315e551d170c09953bf168531e4a","imageHashArm64":"cbbb074d51d90a5f2d6f1d98a05024f6de2ffdcb5acab632324cea4349a820bd","status":"recommended","critical":false}},"proxysql":{"2.7.3-1.2":{"imagePath":"percona/proxysql2:2.7.3-1.2","imageHash":"719d0ab363c65c7f75431bbed7ec0d9f2af7e691765c489da954813c552359a2","imageHashArm64":"4c4d094652c9f2eb097be5d92dcc05da61c9e8699ac7321def959d5a205a89f7","status":"recommended","critical":false}},"haproxy":{"2.8.17":{"imagePath":"percona/haproxy:2.8.17","imageHash":"ef8486b39a1e8dca97b5cdf1135e6282be1757ad188517b889d12c5a3470eeda","imageHashArm64":"bbc5b3b66ac985d1a4500195539e7dff5196245a5a842a6858ea0848ec089967","status":"recommended","critical":false}},"backup":{"8.0.35-34.1":{"imagePath":"percona/percona-xtrabackup:8.0.35-34.1","imageHash":"967bafa0823c90aa8fa9c25a9012be36b0deef64e255294a09148d77ce6aea68","imageHashArm64":"83f814dca9ed398b585938baa86508bda796ba301e34c948a5106095d27bf86e","status":"recommended","critical":false}},"operator":{"1.19.1":{"imagePath":"percona/percona-xtradb-cluster-operator:1.19.1","imageHash":"937a583f0031f45175aee2e5c7a380bbaa9a35046f0e7cae8cd900c44732c4f8","imageHashArm64":"d1939659a5d331db9fae970dee42fd25f083d1955330e9043fe4cb862ff19326","status":"recommended","critical":false}},"logCollector":{"4.0.1-1":{"imagePath":"percona/fluentbit:4.0.1-1","imageHash":"65bdf7d38cbceed6b6aa6412aea3fb4a196000ac6c66185f114a0a62c4a442ad","imageHashArm64":"dabda77b298b67d30d7f53b5cdb7215ad19dabb22b9543e3fd8aedb74ab24733","status":"recommended","critical":false}},"postgresql":{},"pgbackrest":{},"pgbackrestRepo":{},"pgbadger":{},"pgbouncer":{},"pxcOperator":{},"psmdbOperator":{},"pgOperatorApiserver":{},"pgOperatorEvent":{},"pgOperatorRmdata":{},"pgOperatorScheduler":{},"pgOperator":{},"pgOperatorDeployer":{},"psOperator":{},"mysql":{},"router":{},"orchestrator":{},"toolkit":{},"postgis":{},"binlogServer":{}}}]}' + IMAGE_LOGCOLLECTOR=percona/fluentbit:4.0.1-1 + [[ -n '' ]] + [[ 1.20.0 == \1\.\1\9\.\1 ]] + main + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.sKHf7uqm2I ++ mktemp + local LAST_ERR=/tmp/tmp.Rz8RK2NUgm + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sKHf7uqm2I namespace/cert-manager created + cat /tmp/tmp.Rz8RK2NUgm + rm /tmp/tmp.sKHf7uqm2I /tmp/tmp.Rz8RK2NUgm + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.kWd15vWA9y ++ mktemp + local LAST_ERR=/tmp/tmp.ix0s2KQJy8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.kWd15vWA9y namespace/cert-manager labeled + cat /tmp/tmp.ix0s2KQJy8 + rm /tmp/tmp.kWd15vWA9y /tmp/tmp.ix0s2KQJy8 + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.ZHr5TYd53K ++ mktemp + local LAST_ERR=/tmp/tmp.Xj2OuGwdM4 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZHr5TYd53K namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.Xj2OuGwdM4 Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.ZHr5TYd53K /tmp/tmp.Xj2OuGwdM4 + return 0 + '[' '' == 4.10 ']' + sleep 70 + create_infra_gh upgrade-proxysql-17292 v1.19.1 + local ns=upgrade-proxysql-17292 + local git_tag=v1.19.1 + '[' -n pxc-operator ']' + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ grep chaos-mesh ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl api-resources ++ awk '{print $1}' ++ awk '{print $1}' ++ grep chaos-mesh.org ++ kubectl get crd + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrolebinding ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + kubectl_bin get ns + xargs kubectl delete ns + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp + local LAST_OUT=/tmp/tmp.jnh47hu9z5 + awk '{print$1}' ++ mktemp + local LAST_OUT=/tmp/tmp.RMCT77m9xf ++ mktemp + local LAST_ERR=/tmp/tmp.9izAQmGgkd + local exit_status=0 ++ mktemp + local LAST_ERR=/tmp/tmp.FrrUOVdCry + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.RMCT77m9xf + cat /tmp/tmp.FrrUOVdCry + rm /tmp/tmp.RMCT77m9xf /tmp/tmp.FrrUOVdCry + return 0 namespace "cert-manager" deleted namespace "upgrade-proxysql-15631" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jnh47hu9z5 namespace "pxc-operator" deleted + cat /tmp/tmp.9izAQmGgkd + rm /tmp/tmp.jnh47hu9z5 /tmp/tmp.9izAQmGgkd + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.sAJZELuxPt ++ mktemp + local LAST_ERR=/tmp/tmp.o6t4HX2Dwv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sAJZELuxPt namespace/pxc-operator created + cat /tmp/tmp.o6t4HX2Dwv + rm /tmp/tmp.sAJZELuxPt /tmp/tmp.o6t4HX2Dwv + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.PHZzZ9yJzA +++ mktemp ++ local LAST_ERR=/tmp/tmp.U9Xzqj1363 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PHZzZ9yJzA ++ cat /tmp/tmp.U9Xzqj1363 ++ rm /tmp/tmp.PHZzZ9yJzA /tmp/tmp.U9Xzqj1363 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.g3PVrW451y ++ mktemp + local LAST_ERR=/tmp/tmp.OOJVgQNhx6 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.g3PVrW451y Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9" modified. + cat /tmp/tmp.OOJVgQNhx6 + rm /tmp/tmp.g3PVrW451y /tmp/tmp.OOJVgQNhx6 + return 0 + deploy_operator_gh v1.19.1 + local git_tag=v1.19.1 + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- ++ kubectl_bin get crds -o 'jsonpath={.items[?(@.metadata.name == "perconaxtradbclusters.pxc.percona.com")].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.i7GAH0xYln +++ mktemp ++ local LAST_ERR=/tmp/tmp.L79BqktBrT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crds -o 'jsonpath={.items[?(@.metadata.name == "perconaxtradbclusters.pxc.percona.com")].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.i7GAH0xYln ++ cat /tmp/tmp.L79BqktBrT ++ rm /tmp/tmp.i7GAH0xYln /tmp/tmp.L79BqktBrT ++ return 0 + [[ -n perconaxtradbclusters.pxc.percona.com ]] ++ kubectl_bin get crd/perconaxtradbclusters.pxc.percona.com -o 'jsonpath={.spec.versions[?(@.name == "v1-19-1")].name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DWjpirvDM3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.9kuSRGBDPN ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get crd/perconaxtradbclusters.pxc.percona.com -o 'jsonpath={.spec.versions[?(@.name == "v1-19-1")].name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DWjpirvDM3 ++ cat /tmp/tmp.9kuSRGBDPN ++ rm /tmp/tmp.DWjpirvDM3 /tmp/tmp.9kuSRGBDPN ++ return 0 + [[ -n '' ]] + kubectl_bin apply --server-side --force-conflicts -f https://raw.githubusercontent.com/percona/percona-xtradb-cluster-operator/v1.19.1/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.uzzF6snygl ++ mktemp + local LAST_ERR=/tmp/tmp.lKw8uMtGDV + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f https://raw.githubusercontent.com/percona/percona-xtradb-cluster-operator/v1.19.1/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uzzF6snygl + cat /tmp/tmp.lKw8uMtGDV + rm /tmp/tmp.uzzF6snygl /tmp/tmp.lKw8uMtGDV + return 0 + local rbac_yaml=rbac + local operator_yaml=operator.yaml + '[' -n pxc-operator ']' + rbac_yaml=cw-rbac + operator_yaml=cw-operator.yaml + apply_rbac_gh cw-rbac v1.19.1 + local operator_namespace=pxc-operator + local rbac=cw-rbac + local git_tag=v1.19.1 + curl -s https://raw.githubusercontent.com/percona/percona-xtradb-cluster-operator/v1.19.1/deploy/cw-rbac.yaml + /usr/bin/sed -i -e 's^namespace: .*^namespace: pxc-operator^' /tmp/tmp.TZccA2lPpt/rbac_v1.19.1.yaml + kubectl_bin apply -f /tmp/tmp.TZccA2lPpt/rbac_v1.19.1.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.suDhb6Sq5J ++ mktemp + local LAST_ERR=/tmp/tmp.sDXs5FqjUo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /tmp/tmp.TZccA2lPpt/rbac_v1.19.1.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.suDhb6Sq5J clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.sDXs5FqjUo + rm /tmp/tmp.suDhb6Sq5J /tmp/tmp.sDXs5FqjUo + return 0 + curl -s https://raw.githubusercontent.com/percona/percona-xtradb-cluster-operator/v1.19.1/deploy/cw-operator.yaml + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:1.19.1^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "DEBUG"' + kubectl_bin apply -n pxc-operator -f - + cat /tmp/tmp.TZccA2lPpt/cw-operator.yaml_v1.19.1.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.KlNlZZDdPH ++ mktemp + local LAST_ERR=/tmp/tmp.F03d8JLdVM + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -n pxc-operator -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KlNlZZDdPH deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.F03d8JLdVM + rm /tmp/tmp.KlNlZZDdPH /tmp/tmp.F03d8JLdVM + return 0 + sleep 10 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ head -1 +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZMOwdRIfFI +++ mktemp ++ local LAST_ERR=/tmp/tmp.VRoh9TNgDH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZMOwdRIfFI ++ cat /tmp/tmp.VRoh9TNgDH ++ rm /tmp/tmp.ZMOwdRIfFI /tmp/tmp.VRoh9TNgDH ++ return 0 + wait_pod percona-xtradb-cluster-operator-5f848dfbb7-trcrx + local pod=percona-xtradb-cluster-operator-5f848dfbb7-trcrx + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo percona-xtradb-cluster-operator-5f848dfbb7-trcrx ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-5f848dfbb7-trcrx condition met waiting for pod/percona-xtradb-cluster-operator-5f848dfbb7-trcrx to become Ready.Ok + create_namespace upgrade-proxysql-17292 + local namespace=upgrade-proxysql-17292 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ sed s/NAMESPACE// ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get MutatingWebhookConfiguration + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + awk '{print$1}' + '[' -n '' ']' + xargs kubectl delete ns + desc 'cleaned up old namespaces upgrade-proxysql-17292' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces upgrade-proxysql-17292 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace upgrade-proxysql-17292 ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.SsZGmJUuzk + local LAST_OUT=/tmp/tmp.FjNhTDRnTf ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.owRngDSYMr + local exit_status=0 + local LAST_ERR=/tmp/tmp.YNgXfAprpo + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-proxysql-17292 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-proxysql-17292 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FjNhTDRnTf + cat /tmp/tmp.YNgXfAprpo + rm /tmp/tmp.FjNhTDRnTf /tmp/tmp.YNgXfAprpo + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-proxysql-17292 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.SsZGmJUuzk + cat /tmp/tmp.owRngDSYMr Error from server (NotFound): namespaces "upgrade-proxysql-17292" not found + rm /tmp/tmp.SsZGmJUuzk /tmp/tmp.owRngDSYMr + return 1 + : + wait_for_delete namespace/upgrade-proxysql-17292 + local res=namespace/upgrade-proxysql-17292 + echo -n 'waiting for namespace/upgrade-proxysql-17292 to be deleted' waiting for namespace/upgrade-proxysql-17292 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "upgrade-proxysql-17292" not found + desc 'create namespace upgrade-proxysql-17292' + set +o xtrace ----------------------------------------------------------------------------------- create namespace upgrade-proxysql-17292 ----------------------------------------------------------------------------------- + kubectl_bin create namespace upgrade-proxysql-17292 ++ mktemp + local LAST_OUT=/tmp/tmp.EdVUqH2hUE ++ mktemp + local LAST_ERR=/tmp/tmp.PKqsVm2ktc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace upgrade-proxysql-17292 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EdVUqH2hUE namespace/upgrade-proxysql-17292 created + cat /tmp/tmp.PKqsVm2ktc + rm /tmp/tmp.EdVUqH2hUE /tmp/tmp.PKqsVm2ktc + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.lckpUlS1vq +++ mktemp ++ local LAST_ERR=/tmp/tmp.WzIcc7n9Lq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lckpUlS1vq ++ cat /tmp/tmp.WzIcc7n9Lq ++ rm /tmp/tmp.lckpUlS1vq /tmp/tmp.WzIcc7n9Lq ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9 --namespace=upgrade-proxysql-17292 ++ mktemp + local LAST_OUT=/tmp/tmp.NmWeCIif0a ++ mktemp + local LAST_ERR=/tmp/tmp.FGKD91EQPX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9 --namespace=upgrade-proxysql-17292 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NmWeCIif0a Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9" modified. + cat /tmp/tmp.FGKD91EQPX + rm /tmp/tmp.NmWeCIif0a /tmp/tmp.FGKD91EQPX + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.NQugr4Yvjj ++ mktemp + local LAST_ERR=/tmp/tmp.GoktxzbH74 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NQugr4Yvjj secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.GoktxzbH74 + rm /tmp/tmp.NQugr4Yvjj /tmp/tmp.GoktxzbH74 + return 0 + local proxy=proxysql + local cr_yaml=/tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml + prepare_cr_yaml /tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml proxysql upgrade-proxysql 3 v1.19.1 + local cr_yaml=/tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml + local proxy=proxysql + local cluster=upgrade-proxysql + local cluster_size=3 + local git_tag=v1.19.1 + curl -s https://raw.githubusercontent.com/percona/percona-xtradb-cluster-operator/v1.19.1/deploy/cr.yaml + yq eval ' .metadata.name = "upgrade-proxysql" | .spec.secretsName = "my-cluster-secrets" | .spec.vaultSecretName = "some-name-vault" | .spec.sslSecretName = "some-name-ssl" | .spec.sslInternalSecretName = "some-name-ssl-internal" | .spec.upgradeOptions.apply = "disabled" | .spec.pxc.size = 3 | .spec.proxysql.size = 3 | .spec.haproxy.size = 3 | .spec.pxc.image = "-pxc" | .spec.proxysql.image = "-proxysql" | .spec.haproxy.image = "-haproxy" | .spec.backup.image = "-backup" | .spec.backup.storages.minio.s3.credentialsSecret = "minio-secret" | .spec.backup.storages.minio.s3.region = "us-east-1" | .spec.backup.storages.minio.s3.bucket = "operator-testing" | .spec.backup.storages.minio.s3.endpointUrl = "http://minio-service.#namespace:9000/" | .spec.backup.storages.minio.type = "s3" | .spec.pmm.image = "-pmm" ' - + [[ proxysql == \h\a\p\r\o\x\y ]] + yq -i eval ' .spec.haproxy.enabled = false | .spec.proxysql.enabled = true ' /tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml + [[ -n '' ]] + spinup_pxc upgrade-proxysql /tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml 3 30 /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/secrets_without_tls.yml + local cluster=upgrade-proxysql + local config=/tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml + local size=3 + local sleep=30 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/secrets_without_tls.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/secrets_without_tls.yml ++ mktemp + local LAST_OUT=/tmp/tmp.ULRedpqvGF ++ mktemp + local LAST_ERR=/tmp/tmp.e5S7SHj3hG + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/secrets_without_tls.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ULRedpqvGF secret/my-cluster-secrets created + cat /tmp/tmp.e5S7SHj3hG + rm /tmp/tmp.ULRedpqvGF /tmp/tmp.e5S7SHj3hG + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.fhKoDm7GkE + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/client.yml + local pvc_name= + /usr/bin/sed -e 's#image:.*-backup$#image: percona/percona-xtrabackup:8.0.35-34.1#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ mktemp + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:1.19.1#' + local LAST_ERR=/tmp/tmp.vDVOHdgAlB + local exit_status=0 + /usr/bin/sed -e 's#image:.*-proxysql$#image: percona/proxysql2:2.7.3-1.2#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: percona/haproxy:2.8.17#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: percona/fluentbit:4.0.1-1#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.upgrade-proxysql-17292~ + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: percona/percona-xtradb-cluster:8.0.44-35.1#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: percona/percona-xtradb-cluster:8.0.44-35.1#' ++ seq 0 2 + /usr/bin/sed -e 's#image:.*-pmm$#image: percona/pmm-client:3.5.0#' + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.fhKoDm7GkE deployment.apps/pxc-client created + cat /tmp/tmp.vDVOHdgAlB + rm /tmp/tmp.fhKoDm7GkE /tmp/tmp.vDVOHdgAlB + return 0 + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + apply_config /tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml + local config_file=/tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml + local pvc_name= + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml '' + local input_file=/tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml + local pvc_name= + cat /tmp/tmp.TZccA2lPpt/cr_1.19.1_proxysql.yaml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: percona/percona-xtradb-cluster:8.0.44-35.1#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:1.19.1#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: percona/haproxy:2.8.17#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: percona/proxysql2:2.7.3-1.2#' + /usr/bin/sed -e 's#image:.*-backup$#image: percona/percona-xtrabackup:8.0.35-34.1#' + /usr/bin/sed -e 's#image:.*-pmm$#image: percona/pmm-client:3.5.0#' ++ mktemp + local LAST_OUT=/tmp/tmp.dGjizB2GyH ++ mktemp + /usr/bin/sed -e 's#image:.*-logcollector$#image: percona/fluentbit:4.0.1-1#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.upgrade-proxysql-17292~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + local LAST_ERR=/tmp/tmp.GMUUcvKVvl + local exit_status=0 + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: percona/percona-xtradb-cluster:8.0.44-35.1#' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dGjizB2GyH perconaxtradbcluster.pxc.percona.com/upgrade-proxysql created + cat /tmp/tmp.GMUUcvKVvl + rm /tmp/tmp.dGjizB2GyH /tmp/tmp.GMUUcvKVvl + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy upgrade-proxysql ++ local target_cluster=upgrade-proxysql +++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.O0DD0oxtSq ++++ mktemp +++ local LAST_ERR=/tmp/tmp.41nhG3MZiP +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.O0DD0oxtSq +++ cat /tmp/tmp.41nhG3MZiP +++ rm /tmp/tmp.O0DD0oxtSq /tmp/tmp.41nhG3MZiP +++ return 0 ++ [[ false == \t\r\u\e ]] +++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.4pd3TPz1Ds ++++ mktemp +++ local LAST_ERR=/tmp/tmp.1VOVCthg3B +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.4pd3TPz1Ds +++ cat /tmp/tmp.1VOVCthg3B +++ rm /tmp/tmp.4pd3TPz1Ds /tmp/tmp.1VOVCthg3B +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo upgrade-proxysql-proxysql ++ return + local proxy=upgrade-proxysql-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n upgrade-proxysql-17292 ++ mktemp + local LAST_OUT=/tmp/tmp.5aUu4imazq ++ mktemp + local LAST_ERR=/tmp/tmp.NLjtVgL63L + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n upgrade-proxysql-17292 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n upgrade-proxysql-17292 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n upgrade-proxysql-17292 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.5aUu4imazq + cat /tmp/tmp.NLjtVgL63L error: no matching resources found + rm /tmp/tmp.5aUu4imazq /tmp/tmp.NLjtVgL63L + return 1 + true + wait_for_running upgrade-proxysql-proxysql 1 + local name=upgrade-proxysql-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-proxysql-0 480 + local pod=upgrade-proxysql-proxysql-0 + local max_retry=480 + local ns= ++ echo upgrade-proxysql-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/upgrade-proxysql-proxysql-0 condition met waiting for pod/upgrade-proxysql-proxysql-0 to become Ready.Ok + wait_for_running upgrade-proxysql-pxc 3 + local name=upgrade-proxysql-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-pxc-0 480 + local pod=upgrade-proxysql-pxc-0 + local max_retry=480 + local ns= ++ echo upgrade-proxysql-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/upgrade-proxysql-pxc-0 condition met waiting for pod/upgrade-proxysql-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-pxc-1 480 + local pod=upgrade-proxysql-pxc-1 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo upgrade-proxysql-pxc-1 ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/upgrade-proxysql-pxc-1 condition met waiting for pod/upgrade-proxysql-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-pxc-2 480 + local pod=upgrade-proxysql-pxc-2 + local max_retry=480 + local ns= ++ echo upgrade-proxysql-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/upgrade-proxysql-pxc-2 condition met waiting for pod/upgrade-proxysql-pxc-2 to become Ready.Ok + sleep 30 ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.h2pQRpfEET +++ mktemp ++ local LAST_ERR=/tmp/tmp.P0RGkhl1hn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.h2pQRpfEET ++ cat /tmp/tmp.P0RGkhl1hn ++ rm /tmp/tmp.h2pQRpfEET /tmp/tmp.P0RGkhl1hn ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h upgrade-proxysql-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h upgrade-proxysql-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zRMe1HbBOG +++ mktemp ++ local LAST_ERR=/tmp/tmp.kafOLocv1Z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zRMe1HbBOG ++ cat /tmp/tmp.kafOLocv1Z ++ rm /tmp/tmp.zRMe1HbBOG /tmp/tmp.kafOLocv1Z ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-6d7bb97fcf-4hj7m + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h upgrade-proxysql-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h upgrade-proxysql-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yF49oBNl52 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DHp7s68ReL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yF49oBNl52 ++ cat /tmp/tmp.DHp7s68ReL ++ rm /tmp/tmp.yF49oBNl52 /tmp/tmp.DHp7s68ReL ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ echo pxc-client-6d7bb97fcf-4hj7m ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1-80.sql ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V8gEvUO7gH +++ mktemp ++ local LAST_ERR=/tmp/tmp.iWZs64TCyb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.V8gEvUO7gH ++ cat /tmp/tmp.iWZs64TCyb ++ rm /tmp/tmp.V8gEvUO7gH /tmp/tmp.iWZs64TCyb ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ echo pxc-client-6d7bb97fcf-4hj7m ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.TZccA2lPpt/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.TZccA2lPpt/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql /tmp/tmp.TZccA2lPpt/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1-80.sql ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ngD3l34VWB +++ mktemp ++ local LAST_ERR=/tmp/tmp.bsXg86tj6c ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ngD3l34VWB ++ cat /tmp/tmp.bsXg86tj6c ++ rm /tmp/tmp.ngD3l34VWB /tmp/tmp.bsXg86tj6c ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-6d7bb97fcf-4hj7m ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.TZccA2lPpt/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.TZccA2lPpt/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql /tmp/tmp.TZccA2lPpt/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1-80.sql ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DhldVbUBAQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.3y9WD0G5ft ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DhldVbUBAQ ++ cat /tmp/tmp.3y9WD0G5ft ++ rm /tmp/tmp.DhldVbUBAQ /tmp/tmp.3y9WD0G5ft ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ echo pxc-client-6d7bb97fcf-4hj7m ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.TZccA2lPpt/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.TZccA2lPpt/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql /tmp/tmp.TZccA2lPpt/select-1.sql + is_keyring_plugin_in_use upgrade-proxysql + local cluster=upgrade-proxysql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + kubectl exec upgrade-proxysql-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + grep -E -o 'early-plugin-load=keyring_\w+.so' + return 1 + compare_generation 1 proxysql upgrade-proxysql + local generation=1 + local proxy=proxysql + local cluster=upgrade-proxysql + local current_generation + [[ proxysql == \h\a\p\r\o\x\y ]] + containers=(pxc proxysql) + for container in '"${containers[@]}"' + check_generation 1 pxc upgrade-proxysql + local generation=1 + local container=pxc + local cluster=upgrade-proxysql + local current_generation ++ kubectl_bin get statefulset upgrade-proxysql-pxc -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.00RogK1XwY +++ mktemp ++ local LAST_ERR=/tmp/tmp.KgU15ufLV1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset upgrade-proxysql-pxc -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.00RogK1XwY ++ cat /tmp/tmp.KgU15ufLV1 ++ rm /tmp/tmp.00RogK1XwY /tmp/tmp.KgU15ufLV1 ++ return 0 + current_generation=1 + [[ 1 != \1 ]] + for container in '"${containers[@]}"' + check_generation 1 proxysql upgrade-proxysql + local generation=1 + local container=proxysql + local cluster=upgrade-proxysql + local current_generation ++ kubectl_bin get statefulset upgrade-proxysql-proxysql -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sZpIYNapht +++ mktemp ++ local LAST_ERR=/tmp/tmp.r7iaJhymQe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset upgrade-proxysql-proxysql -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sZpIYNapht ++ cat /tmp/tmp.r7iaJhymQe ++ rm /tmp/tmp.sZpIYNapht /tmp/tmp.r7iaJhymQe ++ return 0 + current_generation=1 + [[ 1 != \1 ]] + desc 'upgrade operator' + set +o xtrace ----------------------------------------------------------------------------------- upgrade operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.5W9WzCq6K2 ++ mktemp + local LAST_ERR=/tmp/tmp.ubKqJaAhDR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.5W9WzCq6K2 customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.ubKqJaAhDR + rm /tmp/tmp.5W9WzCq6K2 /tmp/tmp.ubKqJaAhDR + return 0 + [[ -n pxc-operator ]] + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/deploy/cw-rbac.yaml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.LmvbeNasSt ++ mktemp + local LAST_ERR=/tmp/tmp.0UCJgTM4gp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + sed -e 's^namespace: .*^namespace: pxc-operator^' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LmvbeNasSt clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.0UCJgTM4gp + rm /tmp/tmp.LmvbeNasSt /tmp/tmp.0UCJgTM4gp + return 0 + kubectl_bin patch deployment percona-xtradb-cluster-operator '-p{"spec":{"template":{"spec":{"containers":[{"name":"percona-xtradb-cluster-operator","image":"perconalab/percona-xtradb-cluster-operator:PR-2407-e9e16845"}]}}}}' -n pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.TPfWWbhygI ++ mktemp + local LAST_ERR=/tmp/tmp.bqD2h0akwp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch deployment percona-xtradb-cluster-operator '-p{"spec":{"template":{"spec":{"containers":[{"name":"percona-xtradb-cluster-operator","image":"perconalab/percona-xtradb-cluster-operator:PR-2407-e9e16845"}]}}}}' -n pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.TPfWWbhygI deployment.apps/percona-xtradb-cluster-operator patched + cat /tmp/tmp.bqD2h0akwp + rm /tmp/tmp.TPfWWbhygI /tmp/tmp.bqD2h0akwp + return 0 + kubectl_bin rollout status deployment/percona-xtradb-cluster-operator -n pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.r93kstEBQA ++ mktemp + local LAST_ERR=/tmp/tmp.D2SyihLnRf + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl rollout status deployment/percona-xtradb-cluster-operator -n pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.r93kstEBQA Waiting for deployment "percona-xtradb-cluster-operator" rollout to finish: 0 of 1 updated replicas are available... deployment "percona-xtradb-cluster-operator" successfully rolled out + cat /tmp/tmp.D2SyihLnRf + rm /tmp/tmp.r93kstEBQA /tmp/tmp.D2SyihLnRf + return 0 + sleep 10 + desc 'wait for operator upgrade' + set +o xtrace ----------------------------------------------------------------------------------- wait for operator upgrade ----------------------------------------------------------------------------------- + local i=0 + local max=60 ++ grep -vc NAME ++ awk '{print $1}' ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'custom-columns=NAME:.metadata.name,IMAGE:.spec.containers[0].image' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.jxKUlBywyg +++ mktemp ++ local LAST_ERR=/tmp/tmp.bHNmv9wcz5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'custom-columns=NAME:.metadata.name,IMAGE:.spec.containers[0].image' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jxKUlBywyg ++ cat /tmp/tmp.bHNmv9wcz5 ++ rm /tmp/tmp.jxKUlBywyg /tmp/tmp.bHNmv9wcz5 ++ return 0 + [[ 1 -eq 1 ]] + '[' -n pxc-operator ']' ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Szrbuuhw3u +++ mktemp ++ local LAST_ERR=/tmp/tmp.0ewkbCrznG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Szrbuuhw3u ++ cat /tmp/tmp.0ewkbCrznG ++ rm /tmp/tmp.Szrbuuhw3u /tmp/tmp.0ewkbCrznG ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.KSbN66mOis ++ mktemp + local LAST_ERR=/tmp/tmp.6nUrpPLYKo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KSbN66mOis Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9" modified. + cat /tmp/tmp.6nUrpPLYKo + rm /tmp/tmp.KSbN66mOis /tmp/tmp.6nUrpPLYKo + return 0 ++ grep perconalab/percona-xtradb-cluster-operator:PR-2407-e9e16845 ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'custom-columns=NAME:.metadata.name,IMAGE:.spec.containers[0].image' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HADAWJEFGS +++ mktemp ++ local LAST_ERR=/tmp/tmp.wbnUG7B54D ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'custom-columns=NAME:.metadata.name,IMAGE:.spec.containers[0].image' ++ awk '{print $1}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HADAWJEFGS ++ cat /tmp/tmp.wbnUG7B54D ++ rm /tmp/tmp.HADAWJEFGS /tmp/tmp.wbnUG7B54D ++ return 0 + wait_pod percona-xtradb-cluster-operator-585c67699f-7mwv4 + local pod=percona-xtradb-cluster-operator-585c67699f-7mwv4 + local max_retry=480 + local ns= ++ echo percona-xtradb-cluster-operator-585c67699f-7mwv4 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-585c67699f-7mwv4 condition met waiting for pod/percona-xtradb-cluster-operator-585c67699f-7mwv4 to become Ready.Ok ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.TVJhBAdvMu +++ mktemp ++ local LAST_ERR=/tmp/tmp.WqTX3PpjLW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TVJhBAdvMu ++ cat /tmp/tmp.WqTX3PpjLW ++ rm /tmp/tmp.TVJhBAdvMu /tmp/tmp.WqTX3PpjLW ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9 --namespace=upgrade-proxysql-17292 ++ mktemp + local LAST_OUT=/tmp/tmp.V61vPykg37 ++ mktemp + local LAST_ERR=/tmp/tmp.lzLQxoZQez + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9 --namespace=upgrade-proxysql-17292 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.V61vPykg37 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2407-e9e16845-10-cluster9" modified. + cat /tmp/tmp.lzLQxoZQez + rm /tmp/tmp.V61vPykg37 /tmp/tmp.lzLQxoZQez + return 0 + desc 'check images and generation after operator upgrade' + set +o xtrace ----------------------------------------------------------------------------------- check images and generation after operator upgrade ----------------------------------------------------------------------------------- + check_pxc_liveness upgrade-proxysql 3 + local cluster=upgrade-proxysql + local cluster_size=3 + wait_cluster_consistency upgrade-proxysql 3 + local cluster_name=upgrade-proxysql + local cluster_size=3 + local proxy_size= + '[' -z '' ']' ++ get_proxy_size upgrade-proxysql ++ local cluster=upgrade-proxysql +++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.cD9adtcS51 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.rciefZ6tCD +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.cD9adtcS51 +++ cat /tmp/tmp.rciefZ6tCD +++ rm /tmp/tmp.cD9adtcS51 /tmp/tmp.rciefZ6tCD +++ return 0 ++ [[ false == \t\r\u\e ]] +++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qyHI9yesVh ++++ mktemp +++ local LAST_ERR=/tmp/tmp.qBIVRthcvH +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.qyHI9yesVh +++ cat /tmp/tmp.qBIVRthcvH +++ rm /tmp/tmp.qyHI9yesVh /tmp/tmp.qBIVRthcvH +++ return 0 ++ [[ true == \t\r\u\e ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.size}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LqRHrIUql2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.WfPwAGlOMV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.size}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LqRHrIUql2 ++ cat /tmp/tmp.WfPwAGlOMV ++ rm /tmp/tmp.LqRHrIUql2 /tmp/tmp.WfPwAGlOMV ++ return 0 ++ return + proxy_size=3 + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/upgrade-proxysql to be ready' waiting for pxc/upgrade-proxysql to be ready++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ES0XPguwcb +++ mktemp ++ local LAST_ERR=/tmp/tmp.oiyK05RKhO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ES0XPguwcb ++ cat /tmp/tmp.oiyK05RKhO ++ rm /tmp/tmp.ES0XPguwcb /tmp/tmp.oiyK05RKhO ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.s9lM9L9xC3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.mFOgEEfJGz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.s9lM9L9xC3 ++ cat /tmp/tmp.mFOgEEfJGz ++ rm /tmp/tmp.s9lM9L9xC3 /tmp/tmp.mFOgEEfJGz ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine upgrade-proxysql +++ local cluster_name=upgrade-proxysql ++++ get_proxy upgrade-proxysql ++++ local target_cluster=upgrade-proxysql +++++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.51OrqBYSj4 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.8LulPGxcme +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.51OrqBYSj4 +++++ cat /tmp/tmp.8LulPGxcme +++++ rm /tmp/tmp.51OrqBYSj4 /tmp/tmp.8LulPGxcme +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.ThyewE8n0C ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.PYiyeJ3BWe +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.ThyewE8n0C +++++ cat /tmp/tmp.PYiyeJ3BWe +++++ rm /tmp/tmp.ThyewE8n0C /tmp/tmp.PYiyeJ3BWe +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo upgrade-proxysql-proxysql ++++ return +++ local cluster_proxy=upgrade-proxysql-proxysql +++ echo proxysql ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t9zy2fOk5H +++ mktemp ++ local LAST_ERR=/tmp/tmp.wK0ca7BxOI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.t9zy2fOk5H ++ cat /tmp/tmp.wK0ca7BxOI ++ rm /tmp/tmp.t9zy2fOk5H /tmp/tmp.wK0ca7BxOI ++ return 0 + [[ 3 == \3 ]] + echo + wait_for_running upgrade-proxysql-pxc 3 + local name=upgrade-proxysql-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-pxc-0 480 + local pod=upgrade-proxysql-pxc-0 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo upgrade-proxysql-pxc-0 + local container=pxc + set +o xtrace pod/upgrade-proxysql-pxc-0 condition met waiting for pod/upgrade-proxysql-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-pxc-1 480 + local pod=upgrade-proxysql-pxc-1 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo upgrade-proxysql-pxc-1 + local container=pxc + set +o xtrace pod/upgrade-proxysql-pxc-1 condition met waiting for pod/upgrade-proxysql-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-pxc-2 480 + local pod=upgrade-proxysql-pxc-2 + local max_retry=480 + local ns= ++ echo upgrade-proxysql-pxc-2 ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/upgrade-proxysql-pxc-2 condition met waiting for pod/upgrade-proxysql-pxc-2 to become Ready.Ok ++ seq 0 2 + for i in '$(seq 0 $((cluster_size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1-80.sql ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rjUwDrWec2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZYoMuyTjb9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rjUwDrWec2 ++ cat /tmp/tmp.ZYoMuyTjb9 ++ rm /tmp/tmp.rjUwDrWec2 /tmp/tmp.ZYoMuyTjb9 ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-6d7bb97fcf-4hj7m + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.TZccA2lPpt/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.TZccA2lPpt/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql /tmp/tmp.TZccA2lPpt/select-1.sql + for i in '$(seq 0 $((cluster_size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1-80.sql ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BNbfnxpOoX +++ mktemp ++ local LAST_ERR=/tmp/tmp.3iBeFABxO5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BNbfnxpOoX ++ cat /tmp/tmp.3iBeFABxO5 ++ rm /tmp/tmp.BNbfnxpOoX /tmp/tmp.3iBeFABxO5 ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-6d7bb97fcf-4hj7m + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.TZccA2lPpt/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.TZccA2lPpt/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql /tmp/tmp.TZccA2lPpt/select-1.sql + for i in '$(seq 0 $((cluster_size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1-80.sql ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P1eWMS4Lh6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.c85hjK57AZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.P1eWMS4Lh6 ++ cat /tmp/tmp.c85hjK57AZ ++ rm /tmp/tmp.P1eWMS4Lh6 /tmp/tmp.c85hjK57AZ ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-6d7bb97fcf-4hj7m ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.TZccA2lPpt/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.TZccA2lPpt/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql /tmp/tmp.TZccA2lPpt/select-1.sql ++ kubectl_bin get pod -n pxc-operator --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[*].spec.containers[?(@.name == "percona-xtradb-cluster-operator")].image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P8keFhpY3N +++ mktemp ++ local LAST_ERR=/tmp/tmp.qWTuEgCTqx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pod -n pxc-operator --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[*].spec.containers[?(@.name == "percona-xtradb-cluster-operator")].image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.P8keFhpY3N ++ cat /tmp/tmp.qWTuEgCTqx ++ rm /tmp/tmp.P8keFhpY3N /tmp/tmp.qWTuEgCTqx ++ return 0 + [[ perconalab/percona-xtradb-cluster-operator:PR-2407-e9e16845 == perconalab/percona-xtradb-cluster-operator:PR-2407-e9e16845 ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.VlStmXDyIc +++ mktemp ++ local LAST_ERR=/tmp/tmp.gmux6ZqGxi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.VlStmXDyIc ++ cat /tmp/tmp.gmux6ZqGxi ++ rm /tmp/tmp.VlStmXDyIc /tmp/tmp.gmux6ZqGxi ++ return 0 + [[ percona/proxysql2:2.7.3-1.2 == percona/proxysql2:2.7.3-1.2 ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r51HBsyJpy +++ mktemp ++ local LAST_ERR=/tmp/tmp.0HeBYofiZM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.r51HBsyJpy ++ cat /tmp/tmp.0HeBYofiZM ++ rm /tmp/tmp.r51HBsyJpy /tmp/tmp.0HeBYofiZM ++ return 0 + [[ percona/haproxy:2.8.17 == percona/haproxy:2.8.17 ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.backup.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.AwrnmVgqRO +++ mktemp ++ local LAST_ERR=/tmp/tmp.5OzKZxVYP7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.backup.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.AwrnmVgqRO ++ cat /tmp/tmp.5OzKZxVYP7 ++ rm /tmp/tmp.AwrnmVgqRO /tmp/tmp.5OzKZxVYP7 ++ return 0 + [[ percona/percona-xtrabackup:8.0.35-34.1 == percona/percona-xtrabackup:8.0.35-34.1 ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.pmm.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nSetsSwmDk +++ mktemp ++ local LAST_ERR=/tmp/tmp.vuTvsi1Lz3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.pmm.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nSetsSwmDk ++ cat /tmp/tmp.vuTvsi1Lz3 ++ rm /tmp/tmp.nSetsSwmDk /tmp/tmp.vuTvsi1Lz3 ++ return 0 + [[ percona/pmm-client:3.5.0 == percona/pmm-client:3.5.0 ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.logcollector.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fmD6fhB1ll +++ mktemp ++ local LAST_ERR=/tmp/tmp.agxMehtR5X ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.logcollector.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fmD6fhB1ll ++ cat /tmp/tmp.agxMehtR5X ++ rm /tmp/tmp.fmD6fhB1ll /tmp/tmp.agxMehtR5X ++ return 0 + [[ percona/fluentbit:4.0.1-1 == percona/fluentbit:4.0.1-1 ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.pxc.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.r9sAn4c3k2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.310NNRbQHK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.pxc.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.r9sAn4c3k2 ++ cat /tmp/tmp.310NNRbQHK ++ rm /tmp/tmp.r9sAn4c3k2 /tmp/tmp.310NNRbQHK ++ return 0 + [[ percona/percona-xtradb-cluster:8.0.44-35.1 == percona/percona-xtradb-cluster:8.0.44-35.1 ]] + : Operator image has been updated correctly + compare_generation 1 proxysql upgrade-proxysql + local generation=1 + local proxy=proxysql + local cluster=upgrade-proxysql + local current_generation + [[ proxysql == \h\a\p\r\o\x\y ]] + containers=(pxc proxysql) + for container in '"${containers[@]}"' + check_generation 1 pxc upgrade-proxysql + local generation=1 + local container=pxc + local cluster=upgrade-proxysql + local current_generation ++ kubectl_bin get statefulset upgrade-proxysql-pxc -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.C2nUxyqqth +++ mktemp ++ local LAST_ERR=/tmp/tmp.5vtnkTdFxH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset upgrade-proxysql-pxc -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.C2nUxyqqth ++ cat /tmp/tmp.5vtnkTdFxH ++ rm /tmp/tmp.C2nUxyqqth /tmp/tmp.5vtnkTdFxH ++ return 0 + current_generation=1 + [[ 1 != \1 ]] + for container in '"${containers[@]}"' + check_generation 1 proxysql upgrade-proxysql + local generation=1 + local container=proxysql + local cluster=upgrade-proxysql + local current_generation ++ kubectl_bin get statefulset upgrade-proxysql-proxysql -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4UOKCClEZY +++ mktemp ++ local LAST_ERR=/tmp/tmp.7AcQVv835P ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset upgrade-proxysql-proxysql -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4UOKCClEZY ++ cat /tmp/tmp.7AcQVv835P ++ rm /tmp/tmp.4UOKCClEZY /tmp/tmp.7AcQVv835P ++ return 0 + current_generation=1 + [[ 1 != \1 ]] + desc 'patch pxc images and upgrade' + set +o xtrace ----------------------------------------------------------------------------------- patch pxc images and upgrade ----------------------------------------------------------------------------------- + kubectl_bin patch pxc upgrade-proxysql --type=merge --patch '{ "spec": { "crVersion": "1.20.0", "pxc": { "image": "perconalab/percona-xtradb-cluster-operator:main-pxc8.0" }, "pmm": { "image": "perconalab/pmm-client:dev-latest" }, "haproxy": { "image": "perconalab/percona-xtradb-cluster-operator:main-haproxy" }, "proxysql": { "image": "perconalab/percona-xtradb-cluster-operator:main-proxysql" }, "logcollector": { "image": "perconalab/fluentbit:main-logcollector" }, "backup": { "image": "perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup" } }}' ++ mktemp + local LAST_OUT=/tmp/tmp.BO62b6jVOx ++ mktemp + local LAST_ERR=/tmp/tmp.yVfiFv8ejG + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc upgrade-proxysql --type=merge --patch '{ "spec": { "crVersion": "1.20.0", "pxc": { "image": "perconalab/percona-xtradb-cluster-operator:main-pxc8.0" }, "pmm": { "image": "perconalab/pmm-client:dev-latest" }, "haproxy": { "image": "perconalab/percona-xtradb-cluster-operator:main-haproxy" }, "proxysql": { "image": "perconalab/percona-xtradb-cluster-operator:main-proxysql" }, "logcollector": { "image": "perconalab/fluentbit:main-logcollector" }, "backup": { "image": "perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup" } }}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.BO62b6jVOx perconaxtradbcluster.pxc.percona.com/upgrade-proxysql patched + cat /tmp/tmp.yVfiFv8ejG + rm /tmp/tmp.BO62b6jVOx /tmp/tmp.yVfiFv8ejG + return 0 + sleep 10 + desc 'check images and generation after full upgrade' + set +o xtrace ----------------------------------------------------------------------------------- check images and generation after full upgrade ----------------------------------------------------------------------------------- + check_pxc_liveness upgrade-proxysql 3 + local cluster=upgrade-proxysql + local cluster_size=3 + wait_cluster_consistency upgrade-proxysql 3 + local cluster_name=upgrade-proxysql + local cluster_size=3 + local proxy_size= + '[' -z '' ']' ++ get_proxy_size upgrade-proxysql ++ local cluster=upgrade-proxysql +++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.DztI9O1Gy5 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.zT25UQ5Glv +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.DztI9O1Gy5 +++ cat /tmp/tmp.zT25UQ5Glv +++ rm /tmp/tmp.DztI9O1Gy5 /tmp/tmp.zT25UQ5Glv +++ return 0 ++ [[ false == \t\r\u\e ]] +++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.KOJj1bZZwS ++++ mktemp +++ local LAST_ERR=/tmp/tmp.maZzz6Phs4 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.KOJj1bZZwS +++ cat /tmp/tmp.maZzz6Phs4 +++ rm /tmp/tmp.KOJj1bZZwS /tmp/tmp.maZzz6Phs4 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.size}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZzPhf6w90t +++ mktemp ++ local LAST_ERR=/tmp/tmp.ze4Na01Lmt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.size}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZzPhf6w90t ++ cat /tmp/tmp.ze4Na01Lmt ++ rm /tmp/tmp.ZzPhf6w90t /tmp/tmp.ze4Na01Lmt ++ return 0 ++ return + proxy_size=3 + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/upgrade-proxysql to be ready' waiting for pxc/upgrade-proxysql to be ready++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JZ7yXUay0b +++ mktemp ++ local LAST_ERR=/tmp/tmp.vIixZhZK7Z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JZ7yXUay0b ++ cat /tmp/tmp.vIixZhZK7Z ++ rm /tmp/tmp.JZ7yXUay0b /tmp/tmp.vIixZhZK7Z ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DC6KxfVr4P +++ mktemp ++ local LAST_ERR=/tmp/tmp.O4BRksME9G ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DC6KxfVr4P ++ cat /tmp/tmp.O4BRksME9G ++ rm /tmp/tmp.DC6KxfVr4P /tmp/tmp.O4BRksME9G ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XddQMi5I3w +++ mktemp ++ local LAST_ERR=/tmp/tmp.s1KXNvZaJg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XddQMi5I3w ++ cat /tmp/tmp.s1KXNvZaJg ++ rm /tmp/tmp.XddQMi5I3w /tmp/tmp.s1KXNvZaJg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NtRJx48zSF +++ mktemp ++ local LAST_ERR=/tmp/tmp.A8drIPzd6c ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NtRJx48zSF ++ cat /tmp/tmp.A8drIPzd6c ++ rm /tmp/tmp.NtRJx48zSF /tmp/tmp.A8drIPzd6c ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pRZO2WeRho +++ mktemp ++ local LAST_ERR=/tmp/tmp.csY5Wa3QAb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pRZO2WeRho ++ cat /tmp/tmp.csY5Wa3QAb ++ rm /tmp/tmp.pRZO2WeRho /tmp/tmp.csY5Wa3QAb ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kZJx57W9yp +++ mktemp ++ local LAST_ERR=/tmp/tmp.S5qpXF0K8M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kZJx57W9yp ++ cat /tmp/tmp.S5qpXF0K8M ++ rm /tmp/tmp.kZJx57W9yp /tmp/tmp.S5qpXF0K8M ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lGSxU7KLqo +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ftjp0fDd63 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lGSxU7KLqo ++ cat /tmp/tmp.Ftjp0fDd63 ++ rm /tmp/tmp.lGSxU7KLqo /tmp/tmp.Ftjp0fDd63 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4V2UvlDoSu +++ mktemp ++ local LAST_ERR=/tmp/tmp.JiZTVOIYDi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4V2UvlDoSu ++ cat /tmp/tmp.JiZTVOIYDi ++ rm /tmp/tmp.4V2UvlDoSu /tmp/tmp.JiZTVOIYDi ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oiX25dYAW7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.PmPM4bNKZ9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oiX25dYAW7 ++ cat /tmp/tmp.PmPM4bNKZ9 ++ rm /tmp/tmp.oiX25dYAW7 /tmp/tmp.PmPM4bNKZ9 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Yz13H23RFc +++ mktemp ++ local LAST_ERR=/tmp/tmp.wSTnvSHHNt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Yz13H23RFc ++ cat /tmp/tmp.wSTnvSHHNt ++ rm /tmp/tmp.Yz13H23RFc /tmp/tmp.wSTnvSHHNt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ZI7pwHucRV +++ mktemp ++ local LAST_ERR=/tmp/tmp.uhJeBCHPXt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ZI7pwHucRV ++ cat /tmp/tmp.uhJeBCHPXt ++ rm /tmp/tmp.ZI7pwHucRV /tmp/tmp.uhJeBCHPXt ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8o3yyJUvbB +++ mktemp ++ local LAST_ERR=/tmp/tmp.VuEehL5pX0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8o3yyJUvbB ++ cat /tmp/tmp.VuEehL5pX0 ++ rm /tmp/tmp.8o3yyJUvbB /tmp/tmp.VuEehL5pX0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JcEMDUIDmO +++ mktemp ++ local LAST_ERR=/tmp/tmp.4Mab7ks4jc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JcEMDUIDmO ++ cat /tmp/tmp.4Mab7ks4jc ++ rm /tmp/tmp.JcEMDUIDmO /tmp/tmp.4Mab7ks4jc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UJ9p8kYfGn +++ mktemp ++ local LAST_ERR=/tmp/tmp.HJpF6vMAXl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UJ9p8kYfGn ++ cat /tmp/tmp.HJpF6vMAXl ++ rm /tmp/tmp.UJ9p8kYfGn /tmp/tmp.HJpF6vMAXl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RcxAx9LvgN +++ mktemp ++ local LAST_ERR=/tmp/tmp.mlvnioNybP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RcxAx9LvgN ++ cat /tmp/tmp.mlvnioNybP ++ rm /tmp/tmp.RcxAx9LvgN /tmp/tmp.mlvnioNybP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NFtbhcSykW +++ mktemp ++ local LAST_ERR=/tmp/tmp.mfw0L61fBL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NFtbhcSykW ++ cat /tmp/tmp.mfw0L61fBL ++ rm /tmp/tmp.NFtbhcSykW /tmp/tmp.mfw0L61fBL ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2375Mh0Ms3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.RfYy8RO39y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2375Mh0Ms3 ++ cat /tmp/tmp.RfYy8RO39y ++ rm /tmp/tmp.2375Mh0Ms3 /tmp/tmp.RfYy8RO39y ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8TXpJawjhI +++ mktemp ++ local LAST_ERR=/tmp/tmp.y8Wy1TDWzu ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8TXpJawjhI ++ cat /tmp/tmp.y8Wy1TDWzu ++ rm /tmp/tmp.8TXpJawjhI /tmp/tmp.y8Wy1TDWzu ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XwQuGCTOyr +++ mktemp ++ local LAST_ERR=/tmp/tmp.OMSUrNHfxh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XwQuGCTOyr ++ cat /tmp/tmp.OMSUrNHfxh ++ rm /tmp/tmp.XwQuGCTOyr /tmp/tmp.OMSUrNHfxh ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V2F2GYqzvA +++ mktemp ++ local LAST_ERR=/tmp/tmp.h3JrpY1h2w ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.V2F2GYqzvA ++ cat /tmp/tmp.h3JrpY1h2w ++ rm /tmp/tmp.V2F2GYqzvA /tmp/tmp.h3JrpY1h2w ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.HzMQs9lDSL +++ mktemp ++ local LAST_ERR=/tmp/tmp.2oagjgfXt0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.HzMQs9lDSL ++ cat /tmp/tmp.2oagjgfXt0 ++ rm /tmp/tmp.HzMQs9lDSL /tmp/tmp.2oagjgfXt0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Pd6dm87a1O +++ mktemp ++ local LAST_ERR=/tmp/tmp.i3eTSgT02X ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Pd6dm87a1O ++ cat /tmp/tmp.i3eTSgT02X ++ rm /tmp/tmp.Pd6dm87a1O /tmp/tmp.i3eTSgT02X ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pSGIJlak3i +++ mktemp ++ local LAST_ERR=/tmp/tmp.Vg8vfz7UKq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pSGIJlak3i ++ cat /tmp/tmp.Vg8vfz7UKq ++ rm /tmp/tmp.pSGIJlak3i /tmp/tmp.Vg8vfz7UKq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bkW5I24ehT +++ mktemp ++ local LAST_ERR=/tmp/tmp.W1Y4gp0Q4b ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bkW5I24ehT ++ cat /tmp/tmp.W1Y4gp0Q4b ++ rm /tmp/tmp.bkW5I24ehT /tmp/tmp.W1Y4gp0Q4b ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lA1wrrFl5z +++ mktemp ++ local LAST_ERR=/tmp/tmp.FUjVIogXfe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lA1wrrFl5z ++ cat /tmp/tmp.FUjVIogXfe ++ rm /tmp/tmp.lA1wrrFl5z /tmp/tmp.FUjVIogXfe ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.YbTbbmjOXs +++ mktemp ++ local LAST_ERR=/tmp/tmp.FQZ2dg27xl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YbTbbmjOXs ++ cat /tmp/tmp.FQZ2dg27xl ++ rm /tmp/tmp.YbTbbmjOXs /tmp/tmp.FQZ2dg27xl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yS7bqetqau +++ mktemp ++ local LAST_ERR=/tmp/tmp.tK2G7GjynM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yS7bqetqau ++ cat /tmp/tmp.tK2G7GjynM ++ rm /tmp/tmp.yS7bqetqau /tmp/tmp.tK2G7GjynM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 26 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.K1Ny4O6bQe +++ mktemp ++ local LAST_ERR=/tmp/tmp.FySYm2WIlG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.K1Ny4O6bQe ++ cat /tmp/tmp.FySYm2WIlG ++ rm /tmp/tmp.K1Ny4O6bQe /tmp/tmp.FySYm2WIlG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 27 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9O2WGCeoYw +++ mktemp ++ local LAST_ERR=/tmp/tmp.85BiWqTjzJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9O2WGCeoYw ++ cat /tmp/tmp.85BiWqTjzJ ++ rm /tmp/tmp.9O2WGCeoYw /tmp/tmp.85BiWqTjzJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 28 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6uaYQWDgkN +++ mktemp ++ local LAST_ERR=/tmp/tmp.tLDLcyrDwF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6uaYQWDgkN ++ cat /tmp/tmp.tLDLcyrDwF ++ rm /tmp/tmp.6uaYQWDgkN /tmp/tmp.tLDLcyrDwF ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 29 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2NGV9iXUoZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.3ozwv0lUCZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2NGV9iXUoZ ++ cat /tmp/tmp.3ozwv0lUCZ ++ rm /tmp/tmp.2NGV9iXUoZ /tmp/tmp.3ozwv0lUCZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 30 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.djt6M2UdYO +++ mktemp ++ local LAST_ERR=/tmp/tmp.mYdLAJTo2G ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.djt6M2UdYO ++ cat /tmp/tmp.mYdLAJTo2G ++ rm /tmp/tmp.djt6M2UdYO /tmp/tmp.mYdLAJTo2G ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 31 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B8gqrYRxUg +++ mktemp ++ local LAST_ERR=/tmp/tmp.qZQS9nv49a ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.B8gqrYRxUg ++ cat /tmp/tmp.qZQS9nv49a ++ rm /tmp/tmp.B8gqrYRxUg /tmp/tmp.qZQS9nv49a ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 32 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kj1zgih1cG +++ mktemp ++ local LAST_ERR=/tmp/tmp.1nkBwLfVQx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kj1zgih1cG ++ cat /tmp/tmp.1nkBwLfVQx ++ rm /tmp/tmp.kj1zgih1cG /tmp/tmp.1nkBwLfVQx ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 33 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FHaRklHl6n +++ mktemp ++ local LAST_ERR=/tmp/tmp.29Twr7MSQE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FHaRklHl6n ++ cat /tmp/tmp.29Twr7MSQE ++ rm /tmp/tmp.FHaRklHl6n /tmp/tmp.29Twr7MSQE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 34 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bW8EHOQacY +++ mktemp ++ local LAST_ERR=/tmp/tmp.oJuNRoP9Gw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bW8EHOQacY ++ cat /tmp/tmp.oJuNRoP9Gw ++ rm /tmp/tmp.bW8EHOQacY /tmp/tmp.oJuNRoP9Gw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 35 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.FwHaib5VLR +++ mktemp ++ local LAST_ERR=/tmp/tmp.s6wRGmWcZR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FwHaib5VLR ++ cat /tmp/tmp.s6wRGmWcZR ++ rm /tmp/tmp.FwHaib5VLR /tmp/tmp.s6wRGmWcZR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 36 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.swdkuHZ65N +++ mktemp ++ local LAST_ERR=/tmp/tmp.p7NSQk3jqY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.swdkuHZ65N ++ cat /tmp/tmp.p7NSQk3jqY ++ rm /tmp/tmp.swdkuHZ65N /tmp/tmp.p7NSQk3jqY ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.H8hojD8Txz +++ mktemp ++ local LAST_ERR=/tmp/tmp.tUQPrtjBwb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.H8hojD8Txz ++ cat /tmp/tmp.tUQPrtjBwb ++ rm /tmp/tmp.H8hojD8Txz /tmp/tmp.tUQPrtjBwb ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine upgrade-proxysql +++ local cluster_name=upgrade-proxysql ++++ get_proxy upgrade-proxysql ++++ local target_cluster=upgrade-proxysql +++++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.cLmMjstbfw ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.F4EU21AoEM +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.cLmMjstbfw +++++ cat /tmp/tmp.F4EU21AoEM +++++ rm /tmp/tmp.cLmMjstbfw /tmp/tmp.F4EU21AoEM +++++ return 0 ++++ [[ false == \t\r\u\e ]] +++++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.hdenR8KrDq ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.BdamnU6isx +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.hdenR8KrDq +++++ cat /tmp/tmp.BdamnU6isx +++++ rm /tmp/tmp.hdenR8KrDq /tmp/tmp.BdamnU6isx +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo upgrade-proxysql-proxysql ++++ return +++ local cluster_proxy=upgrade-proxysql-proxysql +++ echo proxysql ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.F0Bv54ZVlQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.Hfjcu8DhTM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.F0Bv54ZVlQ ++ cat /tmp/tmp.Hfjcu8DhTM ++ rm /tmp/tmp.F0Bv54ZVlQ /tmp/tmp.Hfjcu8DhTM ++ return 0 + [[ 3 == \3 ]] + echo + wait_for_running upgrade-proxysql-pxc 3 + local name=upgrade-proxysql-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-pxc-0 480 + local pod=upgrade-proxysql-pxc-0 + local max_retry=480 + local ns= ++ echo upgrade-proxysql-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/upgrade-proxysql-pxc-0 condition met waiting for pod/upgrade-proxysql-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-pxc-1 480 + local pod=upgrade-proxysql-pxc-1 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo upgrade-proxysql-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/upgrade-proxysql-pxc-1 condition met waiting for pod/upgrade-proxysql-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod upgrade-proxysql-pxc-2 480 + local pod=upgrade-proxysql-pxc-2 + local max_retry=480 + local ns= ++ echo upgrade-proxysql-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/upgrade-proxysql-pxc-2 condition met waiting for pod/upgrade-proxysql-pxc-2 to become Ready.Ok ++ seq 0 2 + for i in '$(seq 0 $((cluster_size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1-80.sql ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-0.upgrade-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cGlMYcBzTR +++ mktemp ++ local LAST_ERR=/tmp/tmp.bstbbUOzpn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cGlMYcBzTR ++ cat /tmp/tmp.bstbbUOzpn ++ rm /tmp/tmp.cGlMYcBzTR /tmp/tmp.bstbbUOzpn ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ echo pxc-client-6d7bb97fcf-4hj7m ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.TZccA2lPpt/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.TZccA2lPpt/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql /tmp/tmp.TZccA2lPpt/select-1.sql + for i in '$(seq 0 $((cluster_size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1-80.sql ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-1.upgrade-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.20lyUahnai +++ mktemp ++ local LAST_ERR=/tmp/tmp.FRBkW3Ghox ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.20lyUahnai ++ cat /tmp/tmp.FRBkW3Ghox ++ rm /tmp/tmp.20lyUahnai /tmp/tmp.FRBkW3Ghox ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ echo pxc-client-6d7bb97fcf-4hj7m ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.TZccA2lPpt/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.TZccA2lPpt/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql /tmp/tmp.TZccA2lPpt/select-1.sql + for i in '$(seq 0 $((cluster_size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1-80.sql ]] + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h upgrade-proxysql-pxc-2.upgrade-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hMXSHiLUuH +++ mktemp ++ local LAST_ERR=/tmp/tmp.51UGmlvG8a ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hMXSHiLUuH ++ cat /tmp/tmp.51UGmlvG8a ++ rm /tmp/tmp.hMXSHiLUuH /tmp/tmp.51UGmlvG8a ++ return 0 + client_pod=pxc-client-6d7bb97fcf-4hj7m + wait_pod pxc-client-6d7bb97fcf-4hj7m + local pod=pxc-client-6d7bb97fcf-4hj7m + local max_retry=480 + local ns= ++ echo pxc-client-6d7bb97fcf-4hj7m ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-6d7bb97fcf-4hj7m condition met waiting for pod/pxc-client-6d7bb97fcf-4hj7m to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.TZccA2lPpt/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.TZccA2lPpt/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/select-1.sql /tmp/tmp.TZccA2lPpt/select-1.sql ++ kubectl_bin get pod -n pxc-operator --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[*].spec.containers[?(@.name == "percona-xtradb-cluster-operator")].image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NJ4jcFFM87 +++ mktemp ++ local LAST_ERR=/tmp/tmp.gF48T72Q61 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pod -n pxc-operator --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[*].spec.containers[?(@.name == "percona-xtradb-cluster-operator")].image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NJ4jcFFM87 ++ cat /tmp/tmp.gF48T72Q61 ++ rm /tmp/tmp.NJ4jcFFM87 /tmp/tmp.gF48T72Q61 ++ return 0 + [[ perconalab/percona-xtradb-cluster-operator:PR-2407-e9e16845 == perconalab/percona-xtradb-cluster-operator:PR-2407-e9e16845 ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QHc3K3pt6g +++ mktemp ++ local LAST_ERR=/tmp/tmp.YYTfVsWkDh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.proxysql.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QHc3K3pt6g ++ cat /tmp/tmp.YYTfVsWkDh ++ rm /tmp/tmp.QHc3K3pt6g /tmp/tmp.YYTfVsWkDh ++ return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-proxysql == perconalab/percona-xtradb-cluster-operator:main-proxysql ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Apf7PQjqQ5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.KkumuKqnK3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.haproxy.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Apf7PQjqQ5 ++ cat /tmp/tmp.KkumuKqnK3 ++ rm /tmp/tmp.Apf7PQjqQ5 /tmp/tmp.KkumuKqnK3 ++ return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-haproxy == perconalab/percona-xtradb-cluster-operator:main-haproxy ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.backup.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ARwkP15nno +++ mktemp ++ local LAST_ERR=/tmp/tmp.s4IJxrwVrh ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.backup.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ARwkP15nno ++ cat /tmp/tmp.s4IJxrwVrh ++ rm /tmp/tmp.ARwkP15nno /tmp/tmp.s4IJxrwVrh ++ return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup == perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.pmm.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sVOUHvvo72 +++ mktemp ++ local LAST_ERR=/tmp/tmp.gjjr61Ao7Y ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.pmm.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sVOUHvvo72 ++ cat /tmp/tmp.gjjr61Ao7Y ++ rm /tmp/tmp.sVOUHvvo72 /tmp/tmp.gjjr61Ao7Y ++ return 0 + [[ perconalab/pmm-client:dev-latest == perconalab/pmm-client:dev-latest ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.logcollector.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hbaG9nTjP2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.22ge5r7H7I ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.logcollector.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hbaG9nTjP2 ++ cat /tmp/tmp.22ge5r7H7I ++ rm /tmp/tmp.hbaG9nTjP2 /tmp/tmp.22ge5r7H7I ++ return 0 + [[ perconalab/fluentbit:main-logcollector == perconalab/fluentbit:main-logcollector ]] ++ kubectl_bin get pxc upgrade-proxysql -o 'jsonpath={.spec.pxc.image}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7nH7RCpnvd +++ mktemp ++ local LAST_ERR=/tmp/tmp.cKag1sT1Zp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc upgrade-proxysql -o 'jsonpath={.spec.pxc.image}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7nH7RCpnvd ++ cat /tmp/tmp.cKag1sT1Zp ++ rm /tmp/tmp.7nH7RCpnvd /tmp/tmp.cKag1sT1Zp ++ return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 == perconalab/percona-xtradb-cluster-operator:main-pxc8.0 ]] + : Cluster images have been updated correctly + compare_generation 2 proxysql upgrade-proxysql + local generation=2 + local proxy=proxysql + local cluster=upgrade-proxysql + local current_generation + [[ proxysql == \h\a\p\r\o\x\y ]] + containers=(pxc proxysql) + for container in '"${containers[@]}"' + check_generation 2 pxc upgrade-proxysql + local generation=2 + local container=pxc + local cluster=upgrade-proxysql + local current_generation ++ kubectl_bin get statefulset upgrade-proxysql-pxc -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lcaISnFP3t +++ mktemp ++ local LAST_ERR=/tmp/tmp.w3XEZfN4HT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset upgrade-proxysql-pxc -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lcaISnFP3t ++ cat /tmp/tmp.w3XEZfN4HT ++ rm /tmp/tmp.lcaISnFP3t /tmp/tmp.w3XEZfN4HT ++ return 0 + current_generation=2 + [[ 2 != \2 ]] + for container in '"${containers[@]}"' + check_generation 2 proxysql upgrade-proxysql + local generation=2 + local container=proxysql + local cluster=upgrade-proxysql + local current_generation ++ kubectl_bin get statefulset upgrade-proxysql-proxysql -o 'jsonpath={.metadata.generation}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.n1SL2pRftw +++ mktemp ++ local LAST_ERR=/tmp/tmp.AmtFpRJpsE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get statefulset upgrade-proxysql-proxysql -o 'jsonpath={.metadata.generation}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.n1SL2pRftw ++ cat /tmp/tmp.AmtFpRJpsE ++ rm /tmp/tmp.n1SL2pRftw /tmp/tmp.AmtFpRJpsE ++ return 0 + current_generation=2 + [[ 2 != \2 ]] + compare_kubectl statefulset/upgrade-proxysql-pxc + local resource=statefulset/upgrade-proxysql-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc.yml + local new_result=/tmp/tmp.TZccA2lPpt/statefulset_upgrade-proxysql-pxc.yml + desc 'compare statefulset/upgrade-proxysql-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/upgrade-proxysql-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-eks.yml ']' + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-80.yml ']' + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.32 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.32 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.32 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-k127.yml ']' + version_gt 1.24 ++ echo '1.32 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-k124.yml ']' + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-k122.yml ']' + version_gt 1.21 ++ echo '1.32 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-oc.yml ']' + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc-aks.yml ']' + kubectl_bin get -o yaml statefulset/upgrade-proxysql-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-proxysql-17292", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.v2XLs0E02w ++ mktemp + local LAST_ERR=/tmp/tmp.oL1JVGnFf0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/upgrade-proxysql-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.v2XLs0E02w + cat /tmp/tmp.oL1JVGnFf0 + rm /tmp/tmp.v2XLs0E02w /tmp/tmp.oL1JVGnFf0 + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-pxc.yml /tmp/tmp.TZccA2lPpt/statefulset_upgrade-proxysql-pxc.yml + log 'compare_kubectl: statefulset/upgrade-proxysql-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-30T16:38:07+0000]' compare_kubectl: statefulset/upgrade-proxysql-pxc OK [2026-04-30T16:38:07+0000] compare_kubectl: statefulset/upgrade-proxysql-pxc OK + compare_kubectl statefulset/upgrade-proxysql-proxysql + local resource=statefulset/upgrade-proxysql-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql.yml + local new_result=/tmp/tmp.TZccA2lPpt/statefulset_upgrade-proxysql-proxysql.yml + desc 'compare statefulset/upgrade-proxysql-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/upgrade-proxysql-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-eks.yml ']' + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-80.yml ']' + [[ percona/percona-xtradb-cluster:8.0.44-35.1 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.32 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-k129.yml ']' + version_gt 1.27 ++ echo '1.32 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-k127.yml ']' + version_gt 1.24 ++ echo '1.32 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-k124.yml ']' + version_gt 1.22 ++ echo '1.32 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-k122.yml ']' + version_gt 1.21 ++ echo '1.32 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-oc.yml ']' + version_gt 1.29 ++ echo '1.32 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql-aks.yml ']' + kubectl_bin get -o yaml statefulset/upgrade-proxysql-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-proxysql-17292", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.4ytgODGGQI ++ mktemp + local LAST_ERR=/tmp/tmp.sr64uUS3s0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/upgrade-proxysql-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4ytgODGGQI + cat /tmp/tmp.sr64uUS3s0 + rm /tmp/tmp.4ytgODGGQI /tmp/tmp.sr64uUS3s0 + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2407/e2e-tests/upgrade-proxysql/compare/statefulset_upgrade-proxysql-proxysql.yml /tmp/tmp.TZccA2lPpt/statefulset_upgrade-proxysql-proxysql.yml + log 'compare_kubectl: statefulset/upgrade-proxysql-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-04-30T16:38:08+0000]' compare_kubectl: statefulset/upgrade-proxysql-proxysql OK [2026-04-30T16:38:08+0000] compare_kubectl: statefulset/upgrade-proxysql-proxysql OK + desc cleanup + set +o xtrace ----------------------------------------------------------------------------------- cleanup ----------------------------------------------------------------------------------- + destroy upgrade-proxysql-17292 + local namespace=upgrade-proxysql-17292 + local ignore_logs=true + [[ 0 == 1 ]] + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' + grep -v 'get backup status: Job.batch' + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + grep -v level=info ++ get_operator_pod + tee /tmp/tmp.TZccA2lPpt/operator.log ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator + sort -u + grep -v 'the object has been modified' ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.XgVDegIyC6 +++ mktemp ++ head -1 ++ local LAST_ERR=/tmp/tmp.L3yohN17LR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XgVDegIyC6 ++ cat /tmp/tmp.L3yohN17LR ++ rm /tmp/tmp.XgVDegIyC6 /tmp/tmp.L3yohN17LR ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-585c67699f-7mwv4 ++ mktemp + local LAST_OUT=/tmp/tmp.4IACe4F28b ++ mktemp + local LAST_ERR=/tmp/tmp.8Ofc2iaMKQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-585c67699f-7mwv4 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4IACe4F28b + cat /tmp/tmp.8Ofc2iaMKQ + rm /tmp/tmp.4IACe4F28b /tmp/tmp.8Ofc2iaMKQ + return 0 2026-04-30T16:31:09.729Z INFO setup Feature gates {"PXCO_FEATURE_GATES": "", "enabled": ""} 2026-04-30T16:31:09.729Z INFO setup Manager starting up {"gitCommit": "e9e168459e04a086e6783ce380d39942aa5d957c", "gitBranch": "PR-2407-e9e16845", "buildTime": "2026-04-30T12:56:58Z", "goVersion": "go1.25.9", "os": "linux", "arch": "amd64"} 2026-04-30T16:31:09.729Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.32.13-gke.1362000"} 2026-04-30T16:31:09.732Z INFO setup Registering Components. 2026-04-30T16:31:10.513Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2026-04-30T16:31:10.513Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2026-04-30T16:31:10.513Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2026-04-30T16:31:10.513Z INFO controller-runtime.metrics Starting metrics server 2026-04-30T16:31:10.513Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2026-04-30T16:31:10.513Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2026-04-30T16:31:10.513Z INFO controller-runtime.webhook Starting webhook server 2026-04-30T16:31:10.513Z INFO setup Starting the Cmd. 2026-04-30T16:31:10.513Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2026-04-30T16:31:10.713Z INFO Attempting to acquire leader lease... {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-30T16:31:27.453Z DEBUG events percona-xtradb-cluster-operator-585c67699f-7mwv4_c4460ed2-5ad6-4f3a-b12f-0e7038a03a1b became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"40da14ec-e088-4c30-af7f-4317cad47339","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1777566687447103009"}, "reason": "LeaderElection"} 2026-04-30T16:31:27.453Z INFO Starting EventSource {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2026-04-30T16:31:27.453Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.PerconaXtraDBCluster"} 2026-04-30T16:31:27.453Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.Secret"} 2026-04-30T16:31:27.453Z INFO Starting EventSource {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2026-04-30T16:31:27.453Z INFO Successfully acquired lease {"lock": "pxc-operator/08db1feb.percona.com"} 2026-04-30T16:31:27.553Z INFO Starting Controller {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore"} 2026-04-30T16:31:27.553Z INFO Starting workers {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "worker count": 1} 2026-04-30T16:31:27.554Z INFO Starting Controller {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup"} 2026-04-30T16:31:27.554Z INFO Starting workers {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "worker count": 1} 2026-04-30T16:31:27.653Z INFO Starting Controller {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster"} 2026-04-30T16:31:27.653Z INFO Starting workers {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "worker count": 1} 2026-04-30T16:31:29.173Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "fc4c5fd4-4417-488f-8463-00f72cec2c1d", "object": "upgrade-proxysql-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}", "hashChanged": true, "metaChanged": true} 2026-04-30T16:31:29.282Z INFO Creating or updating backup job {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "fc4c5fd4-4417-488f-8463-00f72cec2c1d", "name": "3519a-daily-backup", "schedule": "0 0 * * *"} 2026-04-30T16:31:29.740Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "fc4c5fd4-4417-488f-8463-00f72cec2c1d"} 2026-04-30T16:31:36.360Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "106bbee7-2849-4e2c-b37a-1dfcbb5bbe0e"} 2026-04-30T16:31:42.643Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "010bc991-2573-4390-bcd9-12f8faaa935d"} 2026-04-30T16:31:49.344Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "fd6ed06a-1fb9-4d55-b884-a08208192296"} 2026-04-30T16:31:55.929Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "ee180246-d418-4651-9c17-59dd421942de"} 2026-04-30T16:32:01.921Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "c7f1c6dc-9a9b-4eb9-a95d-d82907b74578"} 2026-04-30T16:32:08.098Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "f9725825-922a-4d08-af93-5b65a9ba9e78"} 2026-04-30T16:32:14.401Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "cb130891-8cdb-4165-ac84-f8f7197b89d2"} 2026-04-30T16:32:20.738Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "dcbe0cce-27c1-4df5-a84d-3b57580b05f0"} 2026-04-30T16:32:27.640Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "9e58eb1b-5f68-4b87-a2bc-0bff4a6dee89"} 2026-04-30T16:32:33.696Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "238f2a33-d6d3-495d-95dc-b1d9d1223e11"} 2026-04-30T16:32:40.017Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "25a4f0e6-1c5e-4c6c-85c8-6166e541e0fc"} 2026-04-30T16:32:46.428Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "80a87b47-010f-4503-b2f2-2ff92a098126"} 2026-04-30T16:32:52.640Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "19d20c9e-453e-44ef-964b-90d4f49c49e6"} 2026-04-30T16:32:54.811Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "object": "upgrade-proxysql-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-30T16:32:55.430Z DEBUG Updating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "object": "upgrade-proxysql-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}", "hashChanged": true, "metaChanged": true} 2026-04-30T16:32:56.778Z INFO statefulSet was changed, run smart update {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d"} 2026-04-30T16:32:56.784Z INFO apply changes to secondary pod {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2"} 2026-04-30T16:32:56.784Z INFO primary pod {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0.upgrade-proxysql-pxc.upgrade-proxysql-17292.svc.cluster.local"} 2026-04-30T16:32:57.644Z INFO Pod is not updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "currentRevision": "upgrade-proxysql-pxc-5b86c6c44d", "targetRevision": "upgrade-proxysql-pxc-85f9ff46d5"} 2026-04-30T16:32:58.111Z ERROR sync users {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "error": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / ", "errorVerbose": "exec syncusers: failed to execute command in pod: unable to upgrade connection: container not found (\"proxysql\") / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:993\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:837\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2026-04-30T16:33:07.774Z INFO Pod is not updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "currentRevision": "upgrade-proxysql-pxc-5b86c6c44d", "targetRevision": "upgrade-proxysql-pxc-85f9ff46d5"} 2026-04-30T16:33:17.737Z INFO Pod is not running {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "phase": "Pending"} 2026-04-30T16:33:17.737Z INFO pod is waiting {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "reason": "PodInitializing"} 2026-04-30T16:33:57.770Z INFO Pod is updated, running and ready {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2"} 2026-04-30T16:33:57.786Z INFO waiting for pod to be present in hostgroup {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "host": "upgrade-proxysql-pxc-2.upgrade-proxysql-pxc.upgrade-proxysql-17292", "hostgroup": 10} 2026-04-30T16:34:07.788Z INFO waiting for pod to be present in hostgroup {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "host": "upgrade-proxysql-pxc-2.upgrade-proxysql-pxc.upgrade-proxysql-17292", "hostgroup": 10} 2026-04-30T16:34:17.788Z INFO waiting for pod to be present in hostgroup {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "host": "upgrade-proxysql-pxc-2.upgrade-proxysql-pxc.upgrade-proxysql-17292", "hostgroup": 10} 2026-04-30T16:34:27.788Z INFO waiting for pod to be present in hostgroup {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "host": "upgrade-proxysql-pxc-2.upgrade-proxysql-pxc.upgrade-proxysql-17292", "hostgroup": 10} 2026-04-30T16:34:37.788Z INFO waiting for pod to be present in hostgroup {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "host": "upgrade-proxysql-pxc-2.upgrade-proxysql-pxc.upgrade-proxysql-17292", "hostgroup": 10} 2026-04-30T16:34:47.788Z INFO waiting for pod to be present in hostgroup {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "host": "upgrade-proxysql-pxc-2.upgrade-proxysql-pxc.upgrade-proxysql-17292", "hostgroup": 10} 2026-04-30T16:34:57.788Z INFO pod present in hostgroup {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2", "host": "upgrade-proxysql-pxc-2.upgrade-proxysql-pxc.upgrade-proxysql-17292", "hostgroup": 10} 2026-04-30T16:34:57.792Z INFO apply changes to secondary pod {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1"} 2026-04-30T16:34:57.792Z INFO pod is online {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-2"} 2026-04-30T16:34:58.026Z INFO Pod is not updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1", "currentRevision": "upgrade-proxysql-pxc-5b86c6c44d", "targetRevision": "upgrade-proxysql-pxc-85f9ff46d5"} 2026-04-30T16:35:08.151Z INFO Pod is not updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1", "currentRevision": "upgrade-proxysql-pxc-5b86c6c44d", "targetRevision": "upgrade-proxysql-pxc-85f9ff46d5"} 2026-04-30T16:35:18.122Z INFO Pod is not running {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1", "phase": "Pending"} 2026-04-30T16:35:18.122Z INFO pod is waiting {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1", "reason": "PodInitializing"} 2026-04-30T16:35:28.117Z INFO Pod is not running {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1", "phase": "Pending"} 2026-04-30T16:35:28.117Z INFO pod is waiting {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1", "reason": "PodInitializing"} 2026-04-30T16:36:08.153Z INFO Pod is updated, running and ready {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1"} 2026-04-30T16:36:08.170Z INFO pod present in hostgroup {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1", "host": "upgrade-proxysql-pxc-1.upgrade-proxysql-pxc.upgrade-proxysql-17292", "hostgroup": 10} 2026-04-30T16:36:08.172Z INFO apply changes to primary pod {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0"} 2026-04-30T16:36:08.172Z INFO pod is online {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-1"} 2026-04-30T16:36:08.365Z INFO Pod is not updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0", "currentRevision": "upgrade-proxysql-pxc-5b86c6c44d", "targetRevision": "upgrade-proxysql-pxc-85f9ff46d5"} 2026-04-30T16:36:18.499Z INFO Pod is not updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0", "currentRevision": "upgrade-proxysql-pxc-5b86c6c44d", "targetRevision": "upgrade-proxysql-pxc-85f9ff46d5"} 2026-04-30T16:36:28.453Z INFO Pod is not running {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0", "phase": "Pending"} 2026-04-30T16:36:28.453Z INFO pod is waiting {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0", "reason": "PodInitializing"} 2026-04-30T16:36:38.452Z INFO Pod is not running {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0", "phase": "Pending"} 2026-04-30T16:36:38.452Z INFO pod is waiting {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0", "reason": "PodInitializing"} 2026-04-30T16:37:18.498Z INFO Pod is updated, running and ready {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0"} 2026-04-30T16:37:18.514Z INFO pod present in hostgroup {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0", "host": "upgrade-proxysql-pxc-0.upgrade-proxysql-pxc.upgrade-proxysql-17292", "hostgroup": 11} 2026-04-30T16:37:18.517Z INFO pod is online {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "pod": "upgrade-proxysql-pxc-0"} 2026-04-30T16:37:18.517Z INFO smart update finished {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d"} 2026-04-30T16:37:18.591Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e87698e3-db43-4d56-83c9-b2967d0d912d", "new version": "8.0.43-34.1"} 2026-04-30T16:37:21.639Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "f8df3442-1937-4191-9c91-22e7ed554a78"} 2026-04-30T16:37:27.765Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "4ff95e34-6d0e-4c5a-b310-f1044b3df5f0"} 2026-04-30T16:37:33.958Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "d7d4e587-f7c7-4cb8-88b2-24d3a22b1144"} 2026-04-30T16:37:40.289Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "d9ecd4d6-b99a-4d42-b878-e59575f35652"} 2026-04-30T16:37:46.621Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "e0c2bfa6-3f55-459f-8011-2a3f364c12ee"} 2026-04-30T16:37:52.923Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "83e1d992-df52-413a-a536-4e3f56dcd0f9"} 2026-04-30T16:37:59.247Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "0e3708a0-96b9-40bc-b767-ce60f7c6aa4a"} 2026-04-30T16:38:05.446Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"upgrade-proxysql","namespace":"upgrade-proxysql-17292"}, "namespace": "upgrade-proxysql-17292", "name": "upgrade-proxysql", "reconcileID": "83a0fb47-4c46-41d5-824a-3b5ceeda2009"} github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:839 + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n upgrade-proxysql-17292 upgrade-proxysql --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/upgrade-proxysql patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.GBQ708jK9c ++ mktemp + local LAST_ERR=/tmp/tmp.cw9S1ILVt0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.GBQ708jK9c perconaxtradbcluster.pxc.percona.com "upgrade-proxysql" deleted from upgrade-proxysql-17292 namespace + cat /tmp/tmp.cw9S1ILVt0 + rm /tmp/tmp.GBQ708jK9c /tmp/tmp.cw9S1ILVt0 + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.dWfc9emgD8 ++ mktemp + local LAST_ERR=/tmp/tmp.x6nFujeNWl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dWfc9emgD8 No resources found + cat /tmp/tmp.x6nFujeNWl + rm /tmp/tmp.dWfc9emgD8 /tmp/tmp.x6nFujeNWl + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.y2XKwkxBRJ ++ mktemp + local LAST_ERR=/tmp/tmp.suvjRizzYG + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.y2XKwkxBRJ No resources found + cat /tmp/tmp.suvjRizzYG + rm /tmp/tmp.y2XKwkxBRJ /tmp/tmp.suvjRizzYG + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.sivyZOTvlZ ++ mktemp + local LAST_ERR=/tmp/tmp.IcykxVVu82 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sivyZOTvlZ validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.IcykxVVu82 + rm /tmp/tmp.sivyZOTvlZ /tmp/tmp.IcykxVVu82 + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + rm -rf /tmp/tmp.TZccA2lPpt + kubectl_bin delete --grace-period=0 --force=true namespace upgrade-proxysql-17292 + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.832OJhaIZW ++ mktemp + local LAST_ERR=/tmp/tmp.Q1ZANN3SHp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace upgrade-proxysql-17292 ++ mktemp + local LAST_OUT=/tmp/tmp.XrLx8sXciV ++ tr '\n' ' ' ++ kubectl get crd ++ grep perconaxtradbcluster ++ mktemp ++ awk '{print $1}' + local LAST_ERR=/tmp/tmp.wOXbuZRXOY + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator + kubectl delete crd perconaxtradbclusterbackups.pxc.percona.com perconaxtradbclusterrestores.pxc.percona.com perconaxtradbclusters.pxc.percona.com customresourcedefinition.apiextensions.k8s.io "perconaxtradbclusterbackups.pxc.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaxtradbclusterrestores.pxc.percona.com" deleted customresourcedefinition.apiextensions.k8s.io "perconaxtradbclusters.pxc.percona.com" deleted + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------