Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2036/e2e-tests/logs/demand-backup-parallel-8-0.log WARNING: version difference between client (1.33) and server (1.31) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.31) exceeds the supported minor version skew of +/-1 error: the server doesn't have a resource type "pxc" + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "pxc" error: the server doesn't have a resource type "pxc" error: the server doesn't have a resource type "pxc-backup" error: the server doesn't have a resource type "pxc-restore" error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "pxc-operator" not found waiting for namespace/pxc-operator to be deletedError from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2036-c42c1c6c-4-cluster9" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator created serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator created deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-779b89dbf-9vc9j condition met pod/percona-xtradb-cluster-operator-779b89dbf-9vc9j condition met waiting for pod/percona-xtradb-cluster-operator-779b89dbf-9vc9j to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces demand-backup-parallel-23125 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "demand-backup-parallel-23125" not found waiting for namespace/demand-backup-parallel-23125 to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "demand-backup-parallel-23125" not found ----------------------------------------------------------------------------------- create namespace demand-backup-parallel-23125 ----------------------------------------------------------------------------------- namespace/demand-backup-parallel-23125 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2036-c42c1c6c-4-cluster9" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created "hashicorp" already exists with the same configuration, skipping "minio" already exists with the same configuration, skipping Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "percona" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- Error: uninstall: Release not loaded: minio-service: release: not found NAME: minio-service LAST DEPLOYED: Tue Jul 29 19:31:08 2025 NAMESPACE: demand-backup-parallel-23125 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.demand-backup-parallel-23125.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace demand-backup-parallel-23125 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace demand-backup-parallel-23125 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace demand-backup-parallel-23125 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace demand-backup-parallel-23125 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local pod/minio-service-55fcc5d75f-8xdx7 condition met waiting for pod/minio-service-55fcc5d75f-8xdx7 to become Ready.Ok pod "aws-cli" deleted If you don't see a command prompt, try pressing enter. [2025-07-29T19:32:21+0000] creating PXC client deployment.apps/pxc-client created [2025-07-29T19:32:22+0000] creating cluster secrets secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created [2025-07-29T19:32:25+0000] create PXC cluster: demand-backup-parallel perconaxtradbcluster.pxc.percona.com/demand-backup-parallel created ----------------------------------------------------------------------------------- creating backups ----------------------------------------------------------------------------------- perconaxtradbclusterbackup.pxc.percona.com/backup1 created perconaxtradbclusterbackup.pxc.percona.com/backup2 created perconaxtradbclusterbackup.pxc.percona.com/backup3 created perconaxtradbclusterbackup.pxc.percona.com/backup4 created ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/demand-backup-parallel to be ready................................................... [2025-07-29T19:38:35+0000] Backup lock holder: backup1 waiting for pxc-backup/backup1 to reach Succeeded state...........Succeeded [2025-07-29T19:39:12+0000] Backup lock holder: backup2 waiting for pxc-backup/backup2 to reach Succeeded state............Succeeded [2025-07-29T19:39:41+0000] Backup lock holder: backup3 waiting for pxc-backup/backup3 to reach Succeeded state.............Succeeded [2025-07-29T19:40:15+0000] Backup lock holder: backup4 waiting for pxc-backup/backup4 to reach Succeeded state.........Succeeded waiting for pxc-backup/backup1 to reach Succeeded stateSucceeded waiting for pxc-backup/backup2 to reach Succeeded stateSucceeded waiting for pxc-backup/backup3 to reach Succeeded stateSucceeded waiting for pxc-backup/backup4 to reach Succeeded stateSucceeded [2025-07-29T19:40:48+0000] test passed ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n demand-backup-parallel-23125 demand-backup-parallel --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/demand-backup-parallel patched perconaxtradbcluster.pxc.percona.com "demand-backup-parallel" deleted perconaxtradbclusterbackup.pxc.percona.com "backup1" deleted perconaxtradbclusterbackup.pxc.percona.com "backup2" deleted perconaxtradbclusterbackup.pxc.percona.com "backup3" deleted perconaxtradbclusterbackup.pxc.percona.com "backup4" deleted No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted