Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2036/e2e-tests/logs/demand-backup-flow-control-8-0.log WARNING: version difference between client (1.33) and server (1.31) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.33) and server (1.31) exceeds the supported minor version skew of +/-1 error: the server doesn't have a resource type "pxc" + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: the server doesn't have a resource type "pxc" error: the server doesn't have a resource type "pxc" error: the server doesn't have a resource type "pxc-backup" error: the server doesn't have a resource type "pxc-restore" error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "pxc-operator" not found waiting for namespace/pxc-operator to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2036-c42c1c6c-4-cluster8" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator created serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator created deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-779b89dbf-5bh6j condition met pod/percona-xtradb-cluster-operator-779b89dbf-5bh6j condition met waiting for pod/percona-xtradb-cluster-operator-779b89dbf-5bh6j to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces demand-backup-flow-control-29424 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "demand-backup-flow-control-29424" not found waiting for namespace/demand-backup-flow-control-29424 to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "demand-backup-flow-control-29424" not found ----------------------------------------------------------------------------------- create namespace demand-backup-flow-control-29424 ----------------------------------------------------------------------------------- namespace/demand-backup-flow-control-29424 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2036-c42c1c6c-4-cluster8" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created "hashicorp" already exists with the same configuration, skipping "minio" already exists with the same configuration, skipping Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- Error: uninstall: Release not loaded: minio-service: release: not found NAME: minio-service LAST DEPLOYED: Tue Jul 29 19:30:30 2025 NAMESPACE: demand-backup-flow-control-29424 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.demand-backup-flow-control-29424.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace demand-backup-flow-control-29424 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace demand-backup-flow-control-29424 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace demand-backup-flow-control-29424 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace demand-backup-flow-control-29424 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local pod/minio-service-55fcc5d75f-wpprx condition met waiting for pod/minio-service-55fcc5d75f-wpprx to become Ready.Ok pod "aws-cli" deleted If you don't see a command prompt, try pressing enter. [2025-07-29T19:31:41+0000] creating PXC client deployment.apps/pxc-client created [2025-07-29T19:31:43+0000] creating cluster secrets secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created [2025-07-29T19:31:46+0000] create PXC cluster: some-name perconaxtradbcluster.pxc.percona.com/some-name created ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/some-name to be ready......................................................... ----------------------------------------------------------------------------------- CASE 1: startingDeadlineSeconds ----------------------------------------------------------------------------------- [2025-07-29T19:38:08+0000] setting startingDeadlineSeconds to 20 perconaxtradbcluster.pxc.percona.com/some-name patched [2025-07-29T19:38:10+0000] create dummy lock to block backup lease.coordination.k8s.io/pxc-some-name-backup-lock created [2025-07-29T19:38:13+0000] creating pxc-backup/backup1 perconaxtradbclusterbackup.pxc.percona.com/backup1 created waiting for pxc-backup/backup1 to reach Failed state...........Failed [2025-07-29T19:38:37+0000] operator should fail backup starting deadline seconds exceeded [2025-07-29T19:38:39+0000] operator successfully failed the backup job lease.coordination.k8s.io "pxc-some-name-backup-lock" deleted ----------------------------------------------------------------------------------- CASE 1: startingDeadlineSeconds PASSED ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- CASE 2: suspend and resume ----------------------------------------------------------------------------------- [2025-07-29T19:38:39+0000] setting suspendedDeadlineSeconds to 300 perconaxtradbcluster.pxc.percona.com/some-name patched [2025-07-29T19:38:41+0000] creating pxc-backup/backup2 perconaxtradbclusterbackup.pxc.percona.com/backup2 created waiting for pxc-backup/backup2 to reach Running stateRunning [2025-07-29T19:38:47+0000] deleting pod/some-name-pxc-2 pod "some-name-pxc-2" force deleted Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. waiting for pxc-backup/backup2 to reach Suspended state.Suspended ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/some-name to be ready........ waiting for pxc-backup/backup2 to reach Succeeded state.............Succeeded ----------------------------------------------------------------------------------- CASE 2: suspend and resume PASSED ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- CASE 3: suspendedDeadlineSeconds ----------------------------------------------------------------------------------- [2025-07-29T19:40:33+0000] setting suspendedDeadlineSeconds to 10 perconaxtradbcluster.pxc.percona.com/some-name patched [2025-07-29T19:40:35+0000] creating pxc-backup/backup3 perconaxtradbclusterbackup.pxc.percona.com/backup3 created waiting for pxc-backup/backup3 to reach Running stateRunning [2025-07-29T19:40:41+0000] deleting pod/some-name-pxc-2 pod "some-name-pxc-2" force deleted Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. waiting for pxc-backup/backup3 to reach Suspended state....Suspended ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/some-name to be ready...... waiting for pxc-backup/backup3 to reach Failed stateFailed [2025-07-29T19:41:57+0000] operator should fail backup suspended deadline seconds exceeded ----------------------------------------------------------------------------------- CASE 3: suspendedDeadlineSeconds PASSED ----------------------------------------------------------------------------------- [2025-07-29T19:42:00+0000] test passed ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n demand-backup-flow-control-29424 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched perconaxtradbcluster.pxc.percona.com "some-name" deleted perconaxtradbclusterbackup.pxc.percona.com "backup1" deleted perconaxtradbclusterbackup.pxc.percona.com "backup2" deleted perconaxtradbclusterbackup.pxc.percona.com "backup3" deleted No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted