Log: /mnt/jenkins/workspace/cloud-psmdb-operator_PR-1914/e2e-tests/logs/custom-users-roles-sharded.log WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.30) exceeds the supported minor version skew of +/-1 ----------------------------------------------------------------------------------- get and delete old CRDs and RBAC ----------------------------------------------------------------------------------- error: the server doesn't have a resource type "perconaservermongodbbackups" + kubectl patch perconaservermongodbbackups.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbbackups" error: the server doesn't have a resource type "perconaservermongodbrestores" + kubectl patch perconaservermongodbrestores.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbrestores" error: the server doesn't have a resource type "perconaservermongodbs" + kubectl patch perconaservermongodbs.psmdb.percona.com -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "perconaservermongodbs" Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "null" not found ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces psmdb-operator ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create namespace psmdb-operator ----------------------------------------------------------------------------------- namespace "gke-managed-cim" deleted namespace "gke-managed-system" deleted namespace "gmp-public" deleted namespace "gmp-system" deleted namespace/psmdb-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1914-868a9e61-1-cluster7" modified. ----------------------------------------------------------------------------------- start PSMDB operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaservermongodbbackups.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbrestores.psmdb.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaservermongodbs.psmdb.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-server-mongodb-operator created serviceaccount/percona-server-mongodb-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-server-mongodb-operator created deployment.apps/percona-server-mongodb-operator created waiting for pod/percona-server-mongodb-operator-76fddbf487-vmllj to be ready.OK ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces custom-users-roles-sharded-27551 ----------------------------------------------------------------------------------- namespace "gke-managed-cim" deleted ----------------------------------------------------------------------------------- create namespace custom-users-roles-sharded-27551 ----------------------------------------------------------------------------------- namespace "gke-managed-system" deleted namespace "gmp-public" deleted namespace "gmp-system" deleted namespace/custom-users-roles-sharded-27551 created Context "gke_cloud-dev-112233_us-central1-a_jen-psmdb-1914-868a9e61-1-cluster7" modified. ----------------------------------------------------------------------------------- create secrets and start client ----------------------------------------------------------------------------------- deployment.apps/psmdb-client created secret/some-users created secret/user-one created secret/user-two created ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created runtimeclass.node.k8s.io/container-rc created ----------------------------------------------------------------------------------- create first PSMDB cluster ----------------------------------------------------------------------------------- perconaservermongodb.psmdb.percona.com/some-name created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- waiting for pod/some-name-rs0-0 to be ready...................OK waiting for pod/some-name-rs0-1 to be ready.............OK waiting for pod/some-name-rs0-2 to be ready................OK Waiting for cluster readyness........... waiting for pod/some-name-cfg-0 to be ready.OK waiting for pod/some-name-cfg-1 to be ready.OK waiting for pod/some-name-cfg-2 to be ready.OK waiting for pod/some-name-mongos-0 to be ready.OK waiting for pod/some-name-mongos-1 to be ready.OK waiting for pod/some-name-mongos-2 to be ready.OK Waiting for cluster readyness waiting for cluster readyness ----------------------------------------------------------------------------------- check if service and statefulset created with expected config ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check user created on cluster creation ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- ping return ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- ping return ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- delete initial user from CR and create a new one ----------------------------------------------------------------------------------- perconaservermongodb.psmdb.percona.com/some-name patched waiting for pod/some-name-rs0-0 to be ready.OK waiting for pod/some-name-rs0-1 to be ready.OK waiting for pod/some-name-rs0-2 to be ready.OK Waiting for cluster readyness ----------------------------------------------------------------------------------- ping return ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- ping return ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check password change ----------------------------------------------------------------------------------- secret/user-two patched ----------------------------------------------------------------------------------- ping return ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check user roles update from CR ----------------------------------------------------------------------------------- perconaservermongodb.psmdb.percona.com/some-name patched waiting for pod/some-name-rs0-0 to be ready.OK waiting for pod/some-name-rs0-1 to be ready.OK waiting for pod/some-name-rs0-2 to be ready.OK Waiting for cluster readyness ----------------------------------------------------------------------------------- check user roles update from DB ----------------------------------------------------------------------------------- Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.custom-users-roles-sharded-27551.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("c35d7a44-f1ae-4715-a1db-edd8b77ee3fb") } Percona Server for MongoDB server version: v8.0.8-3 WARNING: shell and server versions do not match switched to db admin bye ----------------------------------------------------------------------------------- check user recreated after deleted from DB ----------------------------------------------------------------------------------- Percona Server for MongoDB shell version v4.4.29-28 connecting to: mongodb://some-name-mongos.custom-users-roles-sharded-27551.svc.cluster.local:27017/admin?compressors=disabled&gssapiServiceName=mongodb Implicit session: session { "id" : UUID("d52de1f7-6bd2-4665-aa96-7c701205de3d") } Percona Server for MongoDB server version: v8.0.8-3 WARNING: shell and server versions do not match switched to db admin true bye ----------------------------------------------------------------------------------- check new user created after updated user name via CR ----------------------------------------------------------------------------------- perconaservermongodb.psmdb.percona.com/some-name patched waiting for pod/some-name-rs0-0 to be ready.OK waiting for pod/some-name-rs0-1 to be ready.OK waiting for pod/some-name-rs0-2 to be ready.OK Waiting for cluster readyness ----------------------------------------------------------------------------------- ping return ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- ping return ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check new user created after updated user db via CR ----------------------------------------------------------------------------------- perconaservermongodb.psmdb.percona.com/some-name patched waiting for pod/some-name-rs0-0 to be ready.OK waiting for pod/some-name-rs0-1 to be ready.OK waiting for pod/some-name-rs0-2 to be ready.OK Waiting for cluster readyness command terminated with exit code 1 command terminated with exit code 1 error: Internal error occurred: error sending request: Post "https://10.214.0.85:10250/exec/custom-users-roles-sharded-27551/psmdb-client-66f577db5f-69bcr/psmdb-client?command=bash&command=-c&command=printf+%27use+admin%5Cn+%28function%28%29+%7B%0A%09var+user+%3D+db.getUser%28%22user-three%22%29%3B%0A%09var+roles+%3D+user.roles%3B%0A%09roles.sort%28%28a%2C+b%29+%3D%3E+%7B%0A%09%09if+%28a.role+%3C+b.role%29+return+-1%3B%0A%09%09if+%28a.role+%3E+b.role%29+return+1%3B%0A%09%09return+0%3B%0A%09%7D%29%3B%0A%09user.roles+%3D+roles%3B%0A%09printjson%28user%29%3B%0A%09%7D%29%28%29%3B%5Cn%27+%7C+mongo+mongodb%3A%2F%2FuserAdmin%3AuserAdmin123456%40some-name-mongos.custom-users-roles-sharded-27551.svc.cluster.local%3A27017%2Fadmin+&error=1&output=1": No agent available error: Internal error occurred: error sending request: Post "https://10.214.0.85:10250/exec/custom-users-roles-sharded-27551/psmdb-client-66f577db5f-69bcr/psmdb-client?command=bash&command=-c&command=printf+%27use+admin%5Cn+%28function%28%29+%7B%0A%09var+user+%3D+db.getUser%28%22user-three%22%29%3B%0A%09var+roles+%3D+user.roles%3B%0A%09roles.sort%28%28a%2C+b%29+%3D%3E+%7B%0A%09%09if+%28a.role+%3C+b.role%29+return+-1%3B%0A%09%09if+%28a.role+%3E+b.role%29+return+1%3B%0A%09%09return+0%3B%0A%09%7D%29%3B%0A%09user.roles+%3D+roles%3B%0A%09printjson%28user%29%3B%0A%09%7D%29%28%29%3B%5Cn%27+%7C+mongo+mongodb%3A%2F%2FuserAdmin%3AuserAdmin123456%40some-name-mongos.custom-users-roles-sharded-27551.svc.cluster.local%3A27017%2Fadmin+&error=1&output=1": No agent available 1,17c1,10 < switched to db admin < { < "_id" : "admin.user-three", < "user" : "user-three", < "db" : "admin", < "roles" : [ < { < "role" : "clusterAdmin", < "db" : "admin" < } < ], < "mechanisms" : [ < "SCRAM-SHA-1", < "SCRAM-SHA-256" < ] < } < bye --- > Error: Mongos is in quiesce mode and will shut down : > connect@src/mongo/shell/mongo.js:374:17 > @(connect):2:6 > exception: connect failed > exiting with code 1 > Error: Mongos is in quiesce mode and will shut down : > connect@src/mongo/shell/mongo.js:374:17 > @(connect):2:6 > exception: connect failed > exiting with code 1