Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-2271/e2e-tests/logs/balancer.log Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.35) and server (1.32) exceeds the supported minor version skew of +/-1 ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- grep: warning: stray \ before - grep: warning: stray \ before - error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- namespace/psmdb-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2271-c31e8c20-4-cluster2" modified. ----------------------------------------------------------------------------------- start PSMDB operator: docker.io/perconalab/percona-server-mongodb-operator:PR-2271-c31e8c20 ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created deployment.apps/percona-server-mongodb-operator created waiting for pod/percona-server-mongodb-operator-59f5f7f8b4-fz7nz to be ready.OK Print operator info from log 2026-03-26T08:06:38.537Z INFO setup Manager starting up {"gitCommit": "c31e8c20f2a920146c4a2410dc38a83148f57a2b", "gitBranch": "PR-2271-c31e8c20", "buildTime": "", "goVersion": "go1.25.8", "os": "linux", "arch": "amd64"} ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces balancer-9639 ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- create namespace balancer-9639 ----------------------------------------------------------------------------------- namespace/balancer-9639 created Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-2271-c31e8c20-4-cluster2" modified. ----------------------------------------------------------------------------------- install MinIO: minio-service ----------------------------------------------------------------------------------- "minio" has been added to your repositories NAME: minio-service LAST DEPLOYED: Thu Mar 26 08:07:11 2026 NAMESPACE: balancer-9639 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.balancer-9639.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace balancer-9639 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace balancer-9639 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace balancer-9639 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace balancer-9639 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local waiting for pod/minio-service-6d5f646cdc-thnwx to be ready.OK service/minio-service created make_bucket: operator-testing pod "aws-cli" deleted from balancer-9639 namespace All commands and output from this session will be recorded in container logs, including credentials and sensitive information passed through the command prompt. If you don't see a command prompt, try pressing enter. warning: couldn't attach to pod/aws-cli, falling back to streaming logs: Internal error occurred: unable to upgrade connection: container aws-cli not found in pod aws-cli_balancer-9639 ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created secret/gcp-cs-sa-key-secret created [2026-03-26T08:07:56+0000] create PSMDB cluster secret/some-users created deployment.apps/psmdb-client created perconaservermongodb.psmdb.percona.com/some-name created [2026-03-26T08:08:01+0000] check if cfg pods started waiting for pod/some-name-cfg-0 to be ready...............OK waiting for pod/some-name-cfg-1 to be ready...........OK waiting for pod/some-name-cfg-2 to be ready...................OK [2026-03-26T08:10:13+0000] check if all shards started waiting for pod/some-name-rs0-0 to be ready.OK waiting for pod/some-name-rs0-1 to be ready.OK waiting for pod/some-name-rs0-2 to be ready.OK waiting for pod/some-name-rs1-0 to be ready.OK waiting for pod/some-name-rs1-1 to be ready.OK waiting for pod/some-name-rs1-2 to be ready.OK waiting for pod/some-name-rs2-0 to be ready.OK waiting for pod/some-name-rs2-1 to be ready.OK waiting for pod/some-name-rs2-2 to be ready.OK [2026-03-26T08:11:03+0000] check if mongos pods started waiting for pod/some-name-mongos-0 to be ready.OK waiting for pod/some-name-mongos-1 to be ready.OK waiting for pod/some-name-mongos-2 to be ready.OK Waiting for cluster readyness [2026-03-26T08:11:20+0000] sleeping for 10 seconds... checking if balancer status is true...OK [2026-03-26T08:11:34+0000] create user [direct: mongos] admin> { ok: 1, '$clusterTime': { clusterTime: Timestamp({ t: 1774512697, i: 3 }), signature: { hash: Binary.createFromBase64('CXMeSDz/UEYD8+9CQAsqrvAcn6s=', 0), keyId: Long('7621473471670779928') } }, operationTime: Timestamp({ t: 1774512697, i: 3 }) } [direct: mongos] admin> [2026-03-26T08:11:39+0000] enable sharding [direct: mongos] admin> { ok: 1, '$clusterTime': { clusterTime: Timestamp({ t: 1774512703, i: 9 }), signature: { hash: Binary.createFromBase64('FLwxfzyyEId7lfq67TbZ2fiGqXY=', 0), keyId: Long('7621473471670779928') } }, operationTime: Timestamp({ t: 1774512703, i: 6 }) } [direct: mongos] admin> [2026-03-26T08:11:45+0000] shard collection [direct: mongos] admin> { collectionsharded: 'app.city', ok: 1, '$clusterTime': { clusterTime: Timestamp({ t: 1774512709, i: 38 }), signature: { hash: Binary.createFromBase64('N0amJiHXXUHSdZbv0qLhv3GnpuU=', 0), keyId: Long('7621473471670779928') } }, operationTime: Timestamp({ t: 1774512709, i: 37 }) } [direct: mongos] admin> [2026-03-26T08:11:49+0000] write data (this can take some time, be patient) [2026-03-26T08:13:56+0000] running backup: backup-minio-0 [2026-03-26T08:13:56+0000] running backup backup-minio-0 perconaservermongodbbackup.psmdb.percona.com/backup-minio-0 created waiting for backup-minio-0 to reach requested state..OK [2026-03-26T08:14:00+0000] checking if balancer is disabled [2026-03-26T08:14:00+0000] sleeping for 0 seconds... checking if balancer status is false...OK waiting for backup-minio-0 to reach ready state....................OK [2026-03-26T08:14:43+0000] checking if balancer is true after backup [2026-03-26T08:14:43+0000] sleeping for 10 seconds... checking if balancer status is true...OK [2026-03-26T08:14:56+0000] running restore: restore-backup-minio-0 [2026-03-26T08:14:56+0000] running restore restore-backup-minio-0 perconaservermongodbrestore.psmdb.percona.com/restore-backup-minio-0 created Waiting for the psmdb-restore/restore-backup-minio-0 object to be created.OK Waiting psmdb-restore/restore-backup-minio-0 to reach state "requested" .OK after 0 minutes [2026-03-26T08:15:20+0000] checking if balancer is disabled [2026-03-26T08:15:20+0000] sleeping for 0 seconds... checking if balancer status is false...OK Waiting for the psmdb-restore/restore-backup-minio-0 object to be created.OK Waiting psmdb-restore/restore-backup-minio-0 to reach state "ready" ..OK after 1 minutes waiting for cluster readyness................OK [2026-03-26T08:20:15+0000] checking if balancer is true after restore [2026-03-26T08:20:15+0000] sleeping for 10 seconds... checking if balancer status is true...OK [2026-03-26T08:20:29+0000] disabling balancer perconaservermongodb.psmdb.percona.com/some-name patched [2026-03-26T08:20:30+0000] sleeping for 10 seconds... checking if balancer status is false...OK [2026-03-26T08:20:44+0000] running backup: backup-minio-1 [2026-03-26T08:20:44+0000] running backup backup-minio-1 perconaservermongodbbackup.psmdb.percona.com/backup-minio-1 created waiting for backup-minio-1 to reach requested state..OK [2026-03-26T08:20:48+0000] checking if balancer is disabled [2026-03-26T08:20:48+0000] sleeping for 0 seconds... checking if balancer status is false...OK waiting for backup-minio-1 to reach ready state...............OK [2026-03-26T08:21:21+0000] checking if balancer is false after backup [2026-03-26T08:21:21+0000] sleeping for 10 seconds... error: Internal error occurred: unable to upgrade connection: container not found ("psmdb-client") error: Internal error occurred: unable to upgrade connection: container not found ("psmdb-client") error: Internal error occurred: unable to upgrade connection: container not found ("psmdb-client") error: Internal error occurred: unable to upgrade connection: container not found ("psmdb-client")