=== RUN kuttl harness.go:459: starting setup harness.go:254: running tests using configured kubeconfig. harness.go:277: Successful connection to cluster at: https://136.113.206.41 harness.go:362: running tests harness.go:74: going to run test suite with timeout of 180 seconds for each step harness.go:374: testsuite: e2e-tests/tests has 46 tests === RUN kuttl/harness === RUN kuttl/harness/one-pod === PAUSE kuttl/harness/one-pod === CONT kuttl/harness/one-pod logger.go:42: 01:30:09 | one-pod | Creating namespace: kuttl-test-hardy-elf logger.go:42: 01:30:09 | one-pod/0-deploy-operator | starting test step 0-deploy-operator logger.go:42: 01:30:09 | one-pod/0-deploy-operator | running command: [sh -c set -o errexit set -o xtrace source ../../functions init_temp_dir # do this only in the first TestStep apply_s3_storage_secrets deploy_operator deploy_tls_cluster_secrets deploy_client deploy_minio] logger.go:42: 01:30:09 | one-pod/0-deploy-operator | + source ../../functions logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ realpath ../../.. logger.go:42: 01:30:09 | one-pod/0-deploy-operator | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | ++++ pwd logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/tests/one-pod logger.go:42: 01:30:09 | one-pod/0-deploy-operator | ++ test_name=one-pod logger.go:42: 01:30:09 | one-pod/0-deploy-operator | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/vars.sh logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 01:30:09 | one-pod/0-deploy-operator | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export GIT_BRANCH=PR-1149 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ GIT_BRANCH=PR-1149 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export VERSION=PR-1149-9302e4b5 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ VERSION=PR-1149-9302e4b5 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-1149-9302e4b5 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-1149-9302e4b5 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ [[ -z 8.4 ]] logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export MYSQL_VERSION=8.4 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ MYSQL_VERSION=8.4 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql8.4 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql8.4 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup8.4 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup8.4 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router8.4 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router8.4 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export PMM_SERVER_VERSION=1.4.3 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ PMM_SERVER_VERSION=1.4.3 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export CERT_MANAGER_VER=1.18.2 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ CERT_MANAGER_VER=1.18.2 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export MINIO_VER=5.4.0 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ MINIO_VER=5.4.0 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export CHAOS_MESH_VER=2.7.2 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ CHAOS_MESH_VER=2.7.2 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ export VAULT_VER=0.16.1 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ VAULT_VER=0.16.1 logger.go:42: 01:30:09 | one-pod/0-deploy-operator | ++++ which gdate logger.go:42: 01:30:09 | one-pod/0-deploy-operator | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 01:30:09 | one-pod/0-deploy-operator | ++++ which date logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ date=/usr/sbin/date logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ oc get projects logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ : logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ kubectl get nodes logger.go:42: 01:30:09 | one-pod/0-deploy-operator | +++ grep '^minikube' logger.go:42: 01:30:10 | one-pod/0-deploy-operator | +++ which gsed logger.go:42: 01:30:10 | one-pod/0-deploy-operator | which: no gsed in (/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 01:30:10 | one-pod/0-deploy-operator | +++ which sed logger.go:42: 01:30:10 | one-pod/0-deploy-operator | ++ sed=/usr/sbin/sed logger.go:42: 01:30:10 | one-pod/0-deploy-operator | ++ oc get projects logger.go:42: 01:30:10 | one-pod/0-deploy-operator | +++ kubectl version -o json logger.go:42: 01:30:10 | one-pod/0-deploy-operator | +++ jq -r .serverVersion.gitVersion logger.go:42: 01:30:10 | one-pod/0-deploy-operator | +++ grep '\-eks\-' logger.go:42: 01:30:10 | one-pod/0-deploy-operator | grep: warning: stray \ before - logger.go:42: 01:30:10 | one-pod/0-deploy-operator | Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 logger.go:42: 01:30:10 | one-pod/0-deploy-operator | ++ '[' ']' logger.go:42: 01:30:10 | one-pod/0-deploy-operator | ++ EKS=0 logger.go:42: 01:30:10 | one-pod/0-deploy-operator | + init_temp_dir logger.go:42: 01:30:10 | one-pod/0-deploy-operator | + rm -rf /tmp/kuttl/ps/one-pod logger.go:42: 01:30:10 | one-pod/0-deploy-operator | + mkdir -p /tmp/kuttl/ps/one-pod logger.go:42: 01:30:10 | one-pod/0-deploy-operator | + apply_s3_storage_secrets logger.go:42: 01:30:10 | one-pod/0-deploy-operator | + apply_minio_secret logger.go:42: 01:30:10 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-hardy-elf apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf/minio-secret.yml logger.go:42: 01:30:11 | one-pod/0-deploy-operator | secret/minio-secret created logger.go:42: 01:30:11 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-hardy-elf apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf/cloud-secret.yml logger.go:42: 01:30:12 | one-pod/0-deploy-operator | secret/aws-s3-secret created logger.go:42: 01:30:13 | one-pod/0-deploy-operator | secret/gcp-cs-secret created logger.go:42: 01:30:13 | one-pod/0-deploy-operator | secret/azure-secret created logger.go:42: 01:30:13 | one-pod/0-deploy-operator | + deploy_operator logger.go:42: 01:30:13 | one-pod/0-deploy-operator | + destroy_operator logger.go:42: 01:30:13 | one-pod/0-deploy-operator | + kubectl -n ps-operator delete deployment percona-server-mysql-operator --force --grace-period=0 logger.go:42: 01:30:13 | one-pod/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 01:30:14 | one-pod/0-deploy-operator | Error from server (NotFound): deployments.apps "percona-server-mysql-operator" not found logger.go:42: 01:30:14 | one-pod/0-deploy-operator | + true logger.go:42: 01:30:14 | one-pod/0-deploy-operator | + [[ -n ps-operator ]] logger.go:42: 01:30:14 | one-pod/0-deploy-operator | + kubectl delete namespace ps-operator --force --grace-period=0 logger.go:42: 01:30:14 | one-pod/0-deploy-operator | Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. logger.go:42: 01:30:14 | one-pod/0-deploy-operator | Error from server (NotFound): namespaces "ps-operator" not found logger.go:42: 01:30:14 | one-pod/0-deploy-operator | + true logger.go:42: 01:30:14 | one-pod/0-deploy-operator | + [[ -n ps-operator ]] logger.go:42: 01:30:14 | one-pod/0-deploy-operator | + create_namespace ps-operator logger.go:42: 01:30:14 | one-pod/0-deploy-operator | + local namespace=ps-operator logger.go:42: 01:30:14 | one-pod/0-deploy-operator | + [[ -n '' ]] logger.go:42: 01:30:14 | one-pod/0-deploy-operator | + kubectl delete namespace ps-operator --ignore-not-found logger.go:42: 01:30:15 | one-pod/0-deploy-operator | + kubectl wait --for=delete namespace ps-operator logger.go:42: 01:30:15 | one-pod/0-deploy-operator | + kubectl create namespace ps-operator logger.go:42: 01:30:16 | one-pod/0-deploy-operator | namespace/ps-operator created logger.go:42: 01:30:16 | one-pod/0-deploy-operator | + apply_crd logger.go:42: 01:30:16 | one-pod/0-deploy-operator | + kubectl -n ps-operator apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy/crd.yaml logger.go:42: 01:30:17 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlbackups.ps.percona.com serverside-applied logger.go:42: 01:30:18 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqlrestores.ps.percona.com serverside-applied logger.go:42: 01:30:19 | one-pod/0-deploy-operator | customresourcedefinition.apiextensions.k8s.io/perconaservermysqls.ps.percona.com serverside-applied logger.go:42: 01:30:19 | one-pod/0-deploy-operator | + apply_rbac logger.go:42: 01:30:19 | one-pod/0-deploy-operator | + local rbac_file logger.go:42: 01:30:19 | one-pod/0-deploy-operator | + '[' -n ps-operator ']' logger.go:42: 01:30:19 | one-pod/0-deploy-operator | + rbac_file=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy/cw-rbac.yaml logger.go:42: 01:30:19 | one-pod/0-deploy-operator | + kubectl -n ps-operator apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy/cw-rbac.yaml logger.go:42: 01:30:20 | one-pod/0-deploy-operator | serviceaccount/percona-server-mysql-operator created logger.go:42: 01:30:21 | one-pod/0-deploy-operator | role.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 01:30:21 | one-pod/0-deploy-operator | clusterrole.rbac.authorization.k8s.io/percona-server-mysql-operator unchanged logger.go:42: 01:30:21 | one-pod/0-deploy-operator | rolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator-leaderelection created logger.go:42: 01:30:21 | one-pod/0-deploy-operator | clusterrolebinding.rbac.authorization.k8s.io/percona-server-mysql-operator unchanged logger.go:42: 01:30:21 | one-pod/0-deploy-operator | + local operator_file logger.go:42: 01:30:21 | one-pod/0-deploy-operator | + '[' -n ps-operator ']' logger.go:42: 01:30:21 | one-pod/0-deploy-operator | + operator_file=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy/cw-operator.yaml logger.go:42: 01:30:21 | one-pod/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="DISABLE_TELEMETRY").value) = "true"' logger.go:42: 01:30:21 | one-pod/0-deploy-operator | + yq eval '(select(documentIndex==1).spec.template.spec.containers[] | select(.name=="manager").env[] | select(.name=="LOG_LEVEL").value) = "VERBOSE"' logger.go:42: 01:30:21 | one-pod/0-deploy-operator | ++ printf 'select(documentIndex==1).spec.template.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:PR-1149-9302e4b5 logger.go:42: 01:30:21 | one-pod/0-deploy-operator | + kubectl -n ps-operator apply -f - logger.go:42: 01:30:21 | one-pod/0-deploy-operator | + yq eval 'select(documentIndex==1).spec.template.spec.containers[0].image="perconalab/percona-server-mysql-operator:PR-1149-9302e4b5"' /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy/cw-operator.yaml logger.go:42: 01:30:23 | one-pod/0-deploy-operator | configmap/percona-server-mysql-operator-config created logger.go:42: 01:30:23 | one-pod/0-deploy-operator | deployment.apps/percona-server-mysql-operator created logger.go:42: 01:30:23 | one-pod/0-deploy-operator | + deploy_tls_cluster_secrets logger.go:42: 01:30:23 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-hardy-elf apply -f /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf/ssl-secret.yaml logger.go:42: 01:30:24 | one-pod/0-deploy-operator | secret/test-ssl created logger.go:42: 01:30:24 | one-pod/0-deploy-operator | + deploy_client logger.go:42: 01:30:24 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-hardy-elf apply -f - logger.go:42: 01:30:24 | one-pod/0-deploy-operator | ++ printf '.spec.containers[0].image="%s"' perconalab/percona-server-mysql-operator:main-psmysql8.4 logger.go:42: 01:30:24 | one-pod/0-deploy-operator | + yq eval '.spec.containers[0].image="perconalab/percona-server-mysql-operator:main-psmysql8.4"' /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf/client.yaml logger.go:42: 01:30:26 | one-pod/0-deploy-operator | pod/mysql-client created logger.go:42: 01:30:26 | one-pod/0-deploy-operator | + deploy_minio logger.go:42: 01:30:26 | one-pod/0-deploy-operator | + local storage=2G logger.go:42: 01:30:26 | one-pod/0-deploy-operator | + local access_key logger.go:42: 01:30:26 | one-pod/0-deploy-operator | + local secret_key logger.go:42: 01:30:26 | one-pod/0-deploy-operator | ++ base64 -d logger.go:42: 01:30:26 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-hardy-elf get secret minio-secret -o 'jsonpath={.data.AWS_ACCESS_KEY_ID}' logger.go:42: 01:30:26 | one-pod/0-deploy-operator | + access_key=some-access-key logger.go:42: 01:30:26 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-hardy-elf get secret minio-secret -o 'jsonpath={.data.AWS_SECRET_ACCESS_KEY}' logger.go:42: 01:30:26 | one-pod/0-deploy-operator | ++ base64 -d logger.go:42: 01:30:27 | one-pod/0-deploy-operator | + secret_key=some-secret-key logger.go:42: 01:30:27 | one-pod/0-deploy-operator | + helm uninstall -n kuttl-test-hardy-elf minio-service logger.go:42: 01:30:27 | one-pod/0-deploy-operator | Error: uninstall: Release not loaded: minio-service: release: not found logger.go:42: 01:30:27 | one-pod/0-deploy-operator | + : logger.go:42: 01:30:27 | one-pod/0-deploy-operator | + helm repo remove minio logger.go:42: 01:30:28 | one-pod/0-deploy-operator | "minio" has been removed from your repositories logger.go:42: 01:30:28 | one-pod/0-deploy-operator | + helm repo add minio https://charts.min.io/ logger.go:42: 01:30:28 | one-pod/0-deploy-operator | "minio" has been added to your repositories logger.go:42: 01:30:28 | one-pod/0-deploy-operator | +++ printf %q some-access-key logger.go:42: 01:30:28 | one-pod/0-deploy-operator | ++ printf %q some-access-key logger.go:42: 01:30:28 | one-pod/0-deploy-operator | +++ printf %q some-secret-key logger.go:42: 01:30:28 | one-pod/0-deploy-operator | ++ printf %q some-secret-key logger.go:42: 01:30:28 | one-pod/0-deploy-operator | + retry 10 60 helm install minio-service -n kuttl-test-hardy-elf --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio logger.go:42: 01:30:28 | one-pod/0-deploy-operator | + local max=10 logger.go:42: 01:30:28 | one-pod/0-deploy-operator | + local delay=60 logger.go:42: 01:30:28 | one-pod/0-deploy-operator | + shift 2 logger.go:42: 01:30:28 | one-pod/0-deploy-operator | + local n=1 logger.go:42: 01:30:28 | one-pod/0-deploy-operator | + helm install minio-service -n kuttl-test-hardy-elf --version 5.4.0 --set replicas=1 --set mode=standalone --set resources.requests.memory=256Mi --set rootUser=rootuser --set rootPassword=rootpass123 --set 'users[0].accessKey=some-access-key' --set 'users[0].secretKey=some-secret-key' --set 'users[0].policy=consoleAdmin' --set service.type=ClusterIP --set configPathmc=/tmp/.minio/ --set persistence.size=2G --set securityContext.enabled=false minio/minio logger.go:42: 01:30:58 | one-pod/0-deploy-operator | NAME: minio-service logger.go:42: 01:30:58 | one-pod/0-deploy-operator | LAST DEPLOYED: Sat Nov 1 01:30:29 2025 logger.go:42: 01:30:58 | one-pod/0-deploy-operator | NAMESPACE: kuttl-test-hardy-elf logger.go:42: 01:30:58 | one-pod/0-deploy-operator | STATUS: deployed logger.go:42: 01:30:58 | one-pod/0-deploy-operator | REVISION: 1 logger.go:42: 01:30:58 | one-pod/0-deploy-operator | TEST SUITE: None logger.go:42: 01:30:58 | one-pod/0-deploy-operator | NOTES: logger.go:42: 01:30:58 | one-pod/0-deploy-operator | MinIO can be accessed via port 9000 on the following DNS name from within your cluster: logger.go:42: 01:30:58 | one-pod/0-deploy-operator | minio-service.kuttl-test-hardy-elf.cluster.local logger.go:42: 01:30:58 | one-pod/0-deploy-operator | logger.go:42: 01:30:58 | one-pod/0-deploy-operator | To access MinIO from localhost, run the below commands: logger.go:42: 01:30:58 | one-pod/0-deploy-operator | logger.go:42: 01:30:58 | one-pod/0-deploy-operator | 1. export POD_NAME=$(kubectl get pods --namespace kuttl-test-hardy-elf -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") logger.go:42: 01:30:58 | one-pod/0-deploy-operator | logger.go:42: 01:30:58 | one-pod/0-deploy-operator | 2. kubectl port-forward $POD_NAME 9000 --namespace kuttl-test-hardy-elf logger.go:42: 01:30:58 | one-pod/0-deploy-operator | logger.go:42: 01:30:58 | one-pod/0-deploy-operator | Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ logger.go:42: 01:30:58 | one-pod/0-deploy-operator | logger.go:42: 01:30:58 | one-pod/0-deploy-operator | You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: logger.go:42: 01:30:58 | one-pod/0-deploy-operator | logger.go:42: 01:30:58 | one-pod/0-deploy-operator | 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart logger.go:42: 01:30:58 | one-pod/0-deploy-operator | logger.go:42: 01:30:58 | one-pod/0-deploy-operator | 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace kuttl-test-hardy-elf minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace kuttl-test-hardy-elf minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 logger.go:42: 01:30:58 | one-pod/0-deploy-operator | logger.go:42: 01:30:58 | one-pod/0-deploy-operator | 3. mc ls minio-service-local logger.go:42: 01:30:58 | one-pod/0-deploy-operator | ++ kubectl -n kuttl-test-hardy-elf get pods --selector=release=minio-service -o 'jsonpath={.items[].metadata.name}' logger.go:42: 01:30:59 | one-pod/0-deploy-operator | + MINIO_POD=minio-service-d9589b474-m48lj logger.go:42: 01:30:59 | one-pod/0-deploy-operator | + wait_pod minio-service-d9589b474-m48lj logger.go:42: 01:30:59 | one-pod/0-deploy-operator | + local pod=minio-service-d9589b474-m48lj logger.go:42: 01:30:59 | one-pod/0-deploy-operator | + local ns=kuttl-test-hardy-elf logger.go:42: 01:30:59 | one-pod/0-deploy-operator | + set +o xtrace logger.go:42: 01:30:59 | one-pod/0-deploy-operator | minio-service-d9589b474-m48ljtrue logger.go:42: 01:30:59 | one-pod/0-deploy-operator | + kubectl -n kuttl-test-hardy-elf run -i --rm aws-cli --image=perconalab/awscli --restart=Never -- bash -c 'AWS_ACCESS_KEY_ID='\''some-access-key'\'' AWS_SECRET_ACCESS_KEY='\''some-secret-key'\'' AWS_DEFAULT_REGION=us-east-1 /usr/bin/aws --endpoint-url http://minio-service:9000 s3 mb s3://operator-testing' logger.go:42: 01:31:03 | one-pod/0-deploy-operator | make_bucket: operator-testing logger.go:42: 01:31:03 | one-pod/0-deploy-operator | pod "aws-cli" deleted from kuttl-test-hardy-elf namespace logger.go:42: 01:31:04 | one-pod/0-deploy-operator | running command: [sh -c kubectl assert exist-enhanced deployment percona-server-mysql-operator -n ${OPERATOR_NS:-$NAMESPACE} --field-selector status.readyReplicas=1] logger.go:42: 01:31:04 | one-pod/0-deploy-operator | ASSERT deployment percona-server-mysql-operator matching field criteria 'status.readyReplicas=1' should exist. logger.go:42: 01:31:04 | one-pod/0-deploy-operator | INFO Found 1 resource(s). logger.go:42: 01:31:04 | one-pod/0-deploy-operator | NAME NAMESPACE COL0 logger.go:42: 01:31:04 | one-pod/0-deploy-operator | percona-server-mysql-operator ps-operator 1 logger.go:42: 01:31:04 | one-pod/0-deploy-operator | ASSERT PASS logger.go:42: 01:31:04 | one-pod/0-deploy-operator | test step completed 0-deploy-operator logger.go:42: 01:31:04 | one-pod/1-create-cluster | starting test step 1-create-cluster logger.go:42: 01:31:04 | one-pod/1-create-cluster | running command: [sh -c set -o errexit set -o xtrace source ../../functions get_cr \ | yq eval '.spec.mysql.clusterType="async"' - \ | yq eval ".spec.unsafeFlags.mysqlSize=true" - \ | yq eval ".spec.unsafeFlags.proxySize=true" - \ | yq eval ".spec.unsafeFlags.orchestratorSize=true" - \ | yq eval ".spec.mysql.size=1" - \ | yq eval ".spec.proxy.haproxy.enabled=true" - \ | yq eval ".spec.proxy.haproxy.size=1" - \ | yq eval ".spec.orchestrator.enabled=true" - \ | yq eval ".spec.orchestrator.size=1" - \ | yq eval '.spec.backup.storages.minio.type="s3"' - \ | yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - \ | yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - \ | yq eval ".spec.backup.storages.minio.s3.endpointUrl=\"http://minio-service.${NAMESPACE}:9000\"" - \ | yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - \ | kubectl -n "${NAMESPACE}" apply -f -] logger.go:42: 01:31:04 | one-pod/1-create-cluster | + source ../../functions logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ realpath ../../.. logger.go:42: 01:31:04 | one-pod/1-create-cluster | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149 logger.go:42: 01:31:04 | one-pod/1-create-cluster | ++++ pwd logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/tests/one-pod logger.go:42: 01:31:04 | one-pod/1-create-cluster | ++ test_name=one-pod logger.go:42: 01:31:04 | one-pod/1-create-cluster | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/vars.sh logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 01:31:04 | one-pod/1-create-cluster | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export GIT_BRANCH=PR-1149 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ GIT_BRANCH=PR-1149 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export VERSION=PR-1149-9302e4b5 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ VERSION=PR-1149-9302e4b5 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-1149-9302e4b5 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-1149-9302e4b5 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ [[ -z 8.4 ]] logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export MYSQL_VERSION=8.4 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ MYSQL_VERSION=8.4 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql8.4 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql8.4 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup8.4 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup8.4 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router8.4 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router8.4 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export PMM_SERVER_VERSION=1.4.3 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ PMM_SERVER_VERSION=1.4.3 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export CERT_MANAGER_VER=1.18.2 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ CERT_MANAGER_VER=1.18.2 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export MINIO_VER=5.4.0 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ MINIO_VER=5.4.0 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export CHAOS_MESH_VER=2.7.2 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ CHAOS_MESH_VER=2.7.2 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ export VAULT_VER=0.16.1 logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ VAULT_VER=0.16.1 logger.go:42: 01:31:04 | one-pod/1-create-cluster | ++++ which gdate logger.go:42: 01:31:04 | one-pod/1-create-cluster | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 01:31:04 | one-pod/1-create-cluster | ++++ which date logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ date=/usr/sbin/date logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ oc get projects logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ : logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ kubectl get nodes logger.go:42: 01:31:04 | one-pod/1-create-cluster | +++ grep '^minikube' logger.go:42: 01:31:05 | one-pod/1-create-cluster | +++ which gsed logger.go:42: 01:31:05 | one-pod/1-create-cluster | which: no gsed in (/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 01:31:05 | one-pod/1-create-cluster | +++ which sed logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ sed=/usr/sbin/sed logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ oc get projects logger.go:42: 01:31:05 | one-pod/1-create-cluster | +++ kubectl version -o json logger.go:42: 01:31:05 | one-pod/1-create-cluster | +++ jq -r .serverVersion.gitVersion logger.go:42: 01:31:05 | one-pod/1-create-cluster | +++ grep '\-eks\-' logger.go:42: 01:31:05 | one-pod/1-create-cluster | grep: warning: stray \ before - logger.go:42: 01:31:05 | one-pod/1-create-cluster | Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ '[' ']' logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ EKS=0 logger.go:42: 01:31:05 | one-pod/1-create-cluster | + get_cr logger.go:42: 01:31:05 | one-pod/1-create-cluster | + local name_suffix= logger.go:42: 01:31:05 | one-pod/1-create-cluster | + local image_mysql=perconalab/percona-server-mysql-operator:main-psmysql8.4 logger.go:42: 01:31:05 | one-pod/1-create-cluster | + local image_backup=perconalab/percona-server-mysql-operator:main-backup8.4 logger.go:42: 01:31:05 | one-pod/1-create-cluster | + local image_orchestrator=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + local image_router=perconalab/percona-server-mysql-operator:main-router8.4 logger.go:42: 01:31:05 | one-pod/1-create-cluster | + local image_toolkit=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 01:31:05 | one-pod/1-create-cluster | + local image_haproxy=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 01:31:05 | one-pod/1-create-cluster | + local image_pmm_client=perconalab/pmm-client:3-dev-latest logger.go:42: 01:31:05 | one-pod/1-create-cluster | + local cr_file=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy/cr.yaml logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.unsafeFlags.mysqlSize=true - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.unsafeFlags.proxySize=true - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.unsafeFlags.orchestratorSize=true - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.mysql.size=1 - logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ printf '.metadata.name="%s"' one-pod logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.metadata.name="one-pod"' /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy/cr.yaml logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.upgradeOptions.apply="disabled"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval 'del(.spec.secretsName)' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.mysql.clusterType="async"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.mysql.gracePeriod=30 - logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ printf '.spec.initContainer.image="%s"' perconalab/percona-server-mysql-operator:PR-1149-9302e4b5 logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ printf '.spec.backup.image="%s"' perconalab/percona-server-mysql-operator:main-backup8.4 logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ printf '.spec.mysql.image="%s"' perconalab/percona-server-mysql-operator:main-psmysql8.4 logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.orchestrator.size=1 - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.proxy.haproxy.size=1 - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.backup.image="perconalab/percona-server-mysql-operator:main-backup8.4"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.sslSecretName="test-ssl"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.endpointUrl="http://minio-service.kuttl-test-hardy-elf:9000"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.proxy.haproxy.enabled=true - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.bucket="operator-testing"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval .spec.orchestrator.enabled=true - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.credentialsSecret="minio-secret"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + kubectl -n kuttl-test-hardy-elf apply -f - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.s3.region="us-east-1"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.backup.storages.minio.type="s3"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.initContainer.image="perconalab/percona-server-mysql-operator:PR-1149-9302e4b5"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.mysql.image="perconalab/percona-server-mysql-operator:main-psmysql8.4"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ printf '.spec.proxy.router.image="%s"' perconalab/percona-server-mysql-operator:main-router8.4 logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ printf '.spec.orchestrator.image="%s"' perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ printf '.spec.toolkit.image="%s"' perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ printf '.spec.proxy.haproxy.image="%s"' perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.proxy.router.image="perconalab/percona-server-mysql-operator:main-router8.4"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.orchestrator.image="perconalab/percona-server-mysql-operator:main-orchestrator"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.proxy.haproxy.image="perconalab/percona-server-mysql-operator:main-haproxy"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.toolkit.image="perconalab/percona-server-mysql-operator:main-toolkit"' - logger.go:42: 01:31:05 | one-pod/1-create-cluster | + '[' -n '' ']' logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval - logger.go:42: 01:31:05 | one-pod/1-create-cluster | ++ printf '.spec.pmm.image="%s"' perconalab/pmm-client:3-dev-latest logger.go:42: 01:31:05 | one-pod/1-create-cluster | + yq eval '.spec.pmm.image="perconalab/pmm-client:3-dev-latest"' - logger.go:42: 01:31:07 | one-pod/1-create-cluster | perconaservermysql.ps.percona.com/one-pod created logger.go:42: 01:32:28 | one-pod/1-create-cluster | test step completed 1-create-cluster logger.go:42: 01:32:28 | one-pod/2-write-data | starting test step 2-write-data logger.go:42: 01:32:28 | one-pod/2-write-data | running command: [sh -c set -o errexit set -o xtrace source ../../functions sleep 5 run_mysql \ "CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)" \ "-h $(get_haproxy_svc $(get_cluster_name))" run_mysql \ "INSERT myDB.myTable (id) VALUES (100500)" \ "-h $(get_haproxy_svc $(get_cluster_name))"] logger.go:42: 01:32:28 | one-pod/2-write-data | + source ../../functions logger.go:42: 01:32:28 | one-pod/2-write-data | +++ realpath ../../.. logger.go:42: 01:32:28 | one-pod/2-write-data | ++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149 logger.go:42: 01:32:28 | one-pod/2-write-data | ++++ pwd logger.go:42: 01:32:28 | one-pod/2-write-data | +++ basename /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/tests/one-pod logger.go:42: 01:32:28 | one-pod/2-write-data | ++ test_name=one-pod logger.go:42: 01:32:28 | one-pod/2-write-data | ++ source /mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/vars.sh logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ ROOT_REPO=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy logger.go:42: 01:32:28 | one-pod/2-write-data | +++ DEPLOY_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/deploy logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests logger.go:42: 01:32:28 | one-pod/2-write-data | +++ TESTS_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf logger.go:42: 01:32:28 | one-pod/2-write-data | +++ TESTS_CONFIG_DIR=/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/e2e-tests/conf logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 01:32:28 | one-pod/2-write-data | +++ TEMP_DIR=/tmp/kuttl/ps/one-pod logger.go:42: 01:32:28 | one-pod/2-write-data | ++++ git rev-parse --abbrev-ref HEAD logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export GIT_BRANCH=PR-1149 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ GIT_BRANCH=PR-1149 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export VERSION=PR-1149-9302e4b5 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ VERSION=PR-1149-9302e4b5 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export IMAGE=perconalab/percona-server-mysql-operator:PR-1149-9302e4b5 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ IMAGE=perconalab/percona-server-mysql-operator:PR-1149-9302e4b5 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ [[ -z 8.4 ]] logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export MYSQL_VERSION=8.4 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ MYSQL_VERSION=8.4 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql8.4 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ IMAGE_MYSQL=perconalab/percona-server-mysql-operator:main-psmysql8.4 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup8.4 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ IMAGE_BACKUP=perconalab/percona-server-mysql-operator:main-backup8.4 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 01:32:28 | one-pod/2-write-data | +++ IMAGE_ORCHESTRATOR=perconalab/percona-server-mysql-operator:main-orchestrator logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router8.4 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ IMAGE_ROUTER=perconalab/percona-server-mysql-operator:main-router8.4 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 01:32:28 | one-pod/2-write-data | +++ IMAGE_TOOLKIT=perconalab/percona-server-mysql-operator:main-toolkit logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 01:32:28 | one-pod/2-write-data | +++ IMAGE_HAPROXY=perconalab/percona-server-mysql-operator:main-haproxy logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export PMM_SERVER_VERSION=1.4.3 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ PMM_SERVER_VERSION=1.4.3 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 01:32:28 | one-pod/2-write-data | +++ IMAGE_PMM_CLIENT=perconalab/pmm-client:3-dev-latest logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 01:32:28 | one-pod/2-write-data | +++ IMAGE_PMM_SERVER=perconalab/pmm-server:3-dev-latest logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export CERT_MANAGER_VER=1.18.2 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ CERT_MANAGER_VER=1.18.2 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export MINIO_VER=5.4.0 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ MINIO_VER=5.4.0 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export CHAOS_MESH_VER=2.7.2 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ CHAOS_MESH_VER=2.7.2 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ export VAULT_VER=0.16.1 logger.go:42: 01:32:28 | one-pod/2-write-data | +++ VAULT_VER=0.16.1 logger.go:42: 01:32:28 | one-pod/2-write-data | ++++ which gdate logger.go:42: 01:32:28 | one-pod/2-write-data | which: no gdate in (/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 01:32:28 | one-pod/2-write-data | ++++ which date logger.go:42: 01:32:28 | one-pod/2-write-data | +++ date=/usr/sbin/date logger.go:42: 01:32:28 | one-pod/2-write-data | +++ oc get projects logger.go:42: 01:32:28 | one-pod/2-write-data | +++ : logger.go:42: 01:32:28 | one-pod/2-write-data | +++ kubectl get nodes logger.go:42: 01:32:28 | one-pod/2-write-data | +++ grep '^minikube' logger.go:42: 01:32:28 | one-pod/2-write-data | +++ which gsed logger.go:42: 01:32:28 | one-pod/2-write-data | which: no gsed in (/mnt/jenkins/workspace/cloud-ps-operator_PR-1149/bin/:/root/.krew/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin) logger.go:42: 01:32:28 | one-pod/2-write-data | +++ which sed logger.go:42: 01:32:28 | one-pod/2-write-data | ++ sed=/usr/sbin/sed logger.go:42: 01:32:28 | one-pod/2-write-data | ++ oc get projects logger.go:42: 01:32:28 | one-pod/2-write-data | +++ kubectl version -o json logger.go:42: 01:32:28 | one-pod/2-write-data | +++ jq -r .serverVersion.gitVersion logger.go:42: 01:32:28 | one-pod/2-write-data | +++ grep '\-eks\-' logger.go:42: 01:32:28 | one-pod/2-write-data | grep: warning: stray \ before - logger.go:42: 01:32:29 | one-pod/2-write-data | Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 logger.go:42: 01:32:29 | one-pod/2-write-data | ++ '[' ']' logger.go:42: 01:32:29 | one-pod/2-write-data | ++ EKS=0 logger.go:42: 01:32:29 | one-pod/2-write-data | + sleep 5 logger.go:42: 01:32:34 | one-pod/2-write-data | +++ get_cluster_name logger.go:42: 01:32:34 | one-pod/2-write-data | +++ kubectl -n kuttl-test-hardy-elf get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 01:32:34 | one-pod/2-write-data | ++ get_haproxy_svc one-pod logger.go:42: 01:32:34 | one-pod/2-write-data | ++ local cluster=one-pod logger.go:42: 01:32:34 | one-pod/2-write-data | ++ echo one-pod-haproxy logger.go:42: 01:32:34 | one-pod/2-write-data | + run_mysql 'CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' '-h one-pod-haproxy' logger.go:42: 01:32:34 | one-pod/2-write-data | + local 'command=CREATE DATABASE IF NOT EXISTS myDB; CREATE TABLE IF NOT EXISTS myDB.myTable (id int PRIMARY KEY)' logger.go:42: 01:32:34 | one-pod/2-write-data | + local 'host=-h one-pod-haproxy' logger.go:42: 01:32:34 | one-pod/2-write-data | ++ get_user_pass root logger.go:42: 01:32:34 | one-pod/2-write-data | ++ local user=root logger.go:42: 01:32:34 | one-pod/2-write-data | +++ get_cluster_name logger.go:42: 01:32:34 | one-pod/2-write-data | +++ kubectl -n kuttl-test-hardy-elf get ps -o 'jsonpath={.items[0].metadata.name}' logger.go:42: 01:32:35 | one-pod/2-write-data | ++ local secret=one-pod-secrets logger.go:42: 01:32:35 | one-pod/2-write-data | ++ kubectl -n kuttl-test-hardy-elf get secret one-pod-secrets -o 'jsonpath={.data.root}' logger.go:42: 01:32:35 | one-pod/2-write-data | ++ base64 --decode logger.go:42: 01:32:35 | one-pod/2-write-data | + local 'user=-uroot -p'\''*bVf Initializing ps-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:09 +0000 UTC Normal Pod one-pod-orc-0 Binding Scheduled Successfully assigned kuttl-test-hardy-elf/one-pod-orc-0 to gke-jen-ps-1149-9302e4b5-default-pool-f65a9c18-6n78 default-scheduler logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:09 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orchestrator-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:09 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orchestrator-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" in 130ms (130ms including waiting). Image size: 110257640 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:09 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orchestrator-init} Created Created container: orchestrator-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:09 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orchestrator-init} Started Started container orchestrator-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:11 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:11 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 98ms (98ms including waiting). Image size: 72389345 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:11 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Created Created container: orchestrator kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:11 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Started Started container orchestrator kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:11 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:12 +0000 UTC Normal PersistentVolumeClaim datadir-one-pod-mysql-0 ProvisioningSucceeded Successfully provisioned volume pvc-6d9d4358-17b9-42b5-b26a-37354c72263e pd.csi.storage.gke.io_gke-bf638b65679845438b1b-94af-ce18-vm_29f72643-e2c7-4880-918c-d70f956a05f8 logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:12 +0000 UTC Normal Pod one-pod-mysql-0 Binding Scheduled Successfully assigned kuttl-test-hardy-elf/one-pod-mysql-0 to gke-jen-ps-1149-9302e4b5-default-pool-f65a9c18-94pb default-scheduler logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:12 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 98ms (98ms including waiting). Image size: 72389345 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:12 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Created Created container: mysql-monit kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:12 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:19 +0000 UTC Normal Pod one-pod-mysql-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-6d9d4358-17b9-42b5-b26a-37354c72263e" attachdetach-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:21 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:21 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" in 147ms (147ms including waiting). Image size: 110257640 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:21 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Created Created container: mysql-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:21 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql8.4" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql8.4" in 91ms (91ms including waiting). Image size: 433128099 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Created Created container: mysql kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup8.4" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup8.4" in 125ms (125ms including waiting). Image size: 543257981 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Created Created container: xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 92ms (92ms including waiting). Image size: 133854467 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Created Created container: pt-heartbeat kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:31:23 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:01 +0000 UTC Normal Pod one-pod-haproxy-0 Binding Scheduled Successfully assigned kuttl-test-hardy-elf/one-pod-haproxy-0 to gke-jen-ps-1149-9302e4b5-default-pool-f65a9c18-6n78 default-scheduler logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:01 +0000 UTC Normal PodDisruptionBudget.policy one-pod-haproxy NoPods No matching pods found controllermanager logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:01 +0000 UTC Normal StatefulSet.apps one-pod-haproxy SuccessfulCreate create Pod one-pod-haproxy-0 in StatefulSet one-pod-haproxy successful statefulset-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:02 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:02 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" in 120ms (120ms including waiting). Image size: 110257640 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:02 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Created Created container: haproxy-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:02 +0000 UTC Normal Pod one-pod-haproxy-0.spec.initContainers{haproxy-init} Started Started container haproxy-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:04 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:04 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 85ms (85ms including waiting). Image size: 105313584 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:04 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Created Created container: haproxy kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:04 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{haproxy} Started Started container haproxy kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:05 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-haproxy" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:05 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-haproxy" in 79ms (79ms including waiting). Image size: 105313584 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:05 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Created Created container: mysql-monit kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:05 +0000 UTC Normal Pod one-pod-haproxy-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:26 +0000 UTC Warning PerconaServerMySQL.ps.percona.com one-pod ClusterStateChanged Initializing -> Ready ps-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:41 +0000 UTC Normal Job.batch xb-one-pod-minio-minio SuccessfulCreate Created pod: xb-one-pod-minio-minio-jnctt job-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-jnctt Binding Scheduled Successfully assigned kuttl-test-hardy-elf/xb-one-pod-minio-minio-jnctt to gke-jen-ps-1149-9302e4b5-default-pool-f65a9c18-3rdg default-scheduler logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-jnctt.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-jnctt.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" in 95ms (95ms including waiting). Image size: 110257640 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-jnctt.spec.initContainers{xtrabackup-init} Created Created container: xtrabackup-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:42 +0000 UTC Normal Pod xb-one-pod-minio-minio-jnctt.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-jnctt.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup8.4" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-jnctt.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup8.4" in 96ms (96ms including waiting). Image size: 543257981 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-jnctt.spec.containers{xtrabackup} Created Created container: xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:44 +0000 UTC Normal Pod xb-one-pod-minio-minio-jnctt.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:32:51 +0000 UTC Normal Job.batch xb-one-pod-minio-minio Completed Job completed job-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:03 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:03 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:03 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:03 +0000 UTC Normal StatefulSet.apps one-pod-mysql SuccessfulDelete delete Pod one-pod-mysql-0 in StatefulSet one-pod-mysql successful statefulset-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:03 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Killing Stopping container orchestrator kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:03 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:03 +0000 UTC Normal StatefulSet.apps one-pod-orc SuccessfulDelete delete Pod one-pod-orc-0 in StatefulSet one-pod-orc successful statefulset-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:03 +0000 UTC Warning PerconaServerMySQL.ps.percona.com one-pod ClusterStateChanged Ready -> Stopping ps-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:03 +0000 UTC Warning PerconaServerMySQL.ps.percona.com one-pod ClusterStateChanged Stopping -> Paused ps-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:06 +0000 UTC Warning Pod one-pod-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2025/11/01 01:33:06 MySQL state is not ready... kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:11 +0000 UTC Warning Pod one-pod-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:16 +0000 UTC Warning Pod one-pod-haproxy-0.spec.containers{haproxy} Unhealthy Readiness probe errored: command timed out: "/opt/percona/haproxy_readiness_check.sh" timed out after 1s kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:28 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm Binding Scheduled Successfully assigned kuttl-test-hardy-elf/xb-restore-one-pod-restore-minio-wcsvm to gke-jen-ps-1149-9302e4b5-default-pool-f65a9c18-94pb default-scheduler logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:28 +0000 UTC Normal Job.batch xb-restore-one-pod-restore-minio SuccessfulCreate Created pod: xb-restore-one-pod-restore-minio-wcsvm job-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:34 +0000 UTC Warning Pod one-pod-haproxy-0.spec.containers{haproxy} Unhealthy Readiness probe failed: ERROR 2013 (HY000): Lost connection to MySQL server at 'reading initial communication packet', system error: 2 kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:35 +0000 UTC Warning Pod one-pod-haproxy-0.spec.containers{haproxy} Unhealthy Liveness probe failed: ERROR 2013 (HY000): Lost connection to MySQL server at 'reading initial communication packet', system error: 2 kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:47 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-6d9d4358-17b9-42b5-b26a-37354c72263e" attachdetach-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:48 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm.spec.initContainers{xtrabackup-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:48 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm.spec.initContainers{xtrabackup-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" in 114ms (114ms including waiting). Image size: 110257640 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:48 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm.spec.initContainers{xtrabackup-init} Created Created container: xtrabackup-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:48 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm.spec.initContainers{xtrabackup-init} Started Started container xtrabackup-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:50 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup8.4" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:50 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup8.4" in 113ms (113ms including waiting). Image size: 543257981 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:50 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm.spec.containers{xtrabackup} Created Created container: xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:33:50 +0000 UTC Normal Pod xb-restore-one-pod-restore-minio-wcsvm.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:00 +0000 UTC Normal Job.batch xb-restore-one-pod-restore-minio Completed Job completed job-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:01 +0000 UTC Normal Pod one-pod-mysql-0 Binding Scheduled Successfully assigned kuttl-test-hardy-elf/one-pod-mysql-0 to gke-jen-ps-1149-9302e4b5-default-pool-f65a9c18-94pb default-scheduler logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:01 +0000 UTC Warning PerconaServerMySQL.ps.percona.com one-pod ClusterStateChanged Paused -> Initializing ps-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:02 +0000 UTC Normal Pod one-pod-orc-0 Binding Scheduled Successfully assigned kuttl-test-hardy-elf/one-pod-orc-0 to gke-jen-ps-1149-9302e4b5-default-pool-f65a9c18-3rdg default-scheduler logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:03 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orchestrator-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:03 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orchestrator-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" in 155ms (155ms including waiting). Image size: 110257640 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:03 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orchestrator-init} Created Created container: orchestrator-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:03 +0000 UTC Normal Pod one-pod-orc-0.spec.initContainers{orchestrator-init} Started Started container orchestrator-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:04 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:04 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 102ms (102ms including waiting). Image size: 72389345 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:04 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Created Created container: orchestrator kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:04 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Started Started container orchestrator kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:04 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-orchestrator" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:04 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-orchestrator" in 85ms (85ms including waiting). Image size: 72389345 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:04 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Created Created container: mysql-monit kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:05 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Started Started container mysql-monit kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:07 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulling Pulling image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:07 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:PR-1149-9302e4b5" in 121ms (121ms including waiting). Image size: 110257640 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:07 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Created Created container: mysql-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:07 +0000 UTC Normal Pod one-pod-mysql-0.spec.initContainers{mysql-init} Started Started container mysql-init kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:09 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-psmysql8.4" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:09 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-psmysql8.4" in 100ms (100ms including waiting). Image size: 433128099 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:09 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Created Created container: mysql kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:09 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Started Started container mysql kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:09 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-backup8.4" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:09 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-backup8.4" in 88ms (88ms including waiting). Image size: 543257981 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:09 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Created Created container: xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:10 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Started Started container xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:10 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulling Pulling image "perconalab/percona-server-mysql-operator:main-toolkit" kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:10 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Pulled Successfully pulled image "perconalab/percona-server-mysql-operator:main-toolkit" in 125ms (125ms including waiting). Image size: 133854467 bytes. kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:10 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Created Created container: pt-heartbeat kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:10 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Started Started container pt-heartbeat kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:34:46 +0000 UTC Warning PerconaServerMySQL.ps.percona.com one-pod AsyncReplicationNotReady one-pod-mysql-0: [replication_lag] ps-controller logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:35:02 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{mysql} Killing Stopping container mysql kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:35:02 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{pt-heartbeat} Killing Stopping container pt-heartbeat kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:35:02 +0000 UTC Normal Pod one-pod-mysql-0.spec.containers{xtrabackup} Killing Stopping container xtrabackup kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:35:02 +0000 UTC Warning Pod one-pod-mysql-0.spec.containers{mysql} Unhealthy Readiness probe failed: 2025/11/01 01:35:02 MySQL state is not ready... kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:35:02 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{orchestrator} Killing Stopping container orchestrator kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:35:02 +0000 UTC Normal Pod one-pod-orc-0.spec.containers{mysql-monit} Killing Stopping container mysql-monit kubelet logger.go:42: 01:35:14 | one-pod | 2025-11-01 01:35:07 +0000 UTC Warning Pod one-pod-mysql-0.spec.containers{mysql} Unhealthy Readiness probe errored: rpc error: code = Unknown desc = failed to exec in container: container is in CONTAINER_EXITED state kubelet logger.go:42: 01:35:14 | one-pod | Deleting namespace: kuttl-test-hardy-elf === NAME kuttl harness.go:403: run tests finished harness.go:510: cleaning up harness.go:567: removing temp folder: "" --- PASS: kuttl (327.65s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/one-pod (326.90s) PASS