Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2473/e2e-tests/logs/init-deploy-5-7.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "pxc-operator" not found waiting for namespace/pxc-operator to be deletedError from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster3" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-55d95dc9d8-d5qqv condition met E0516 23:39:02.681257 17712 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-55d95dc9d8-d5qqv&resourceVersion=1778974742280577000&timeoutSeconds=582&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/percona-xtradb-cluster-operator-55d95dc9d8-d5qqv condition met E0516 23:39:07.550135 18392 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-55d95dc9d8-d5qqv&resourceVersion=1778974745257208000&timeoutSeconds=326&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-55d95dc9d8-d5qqv to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces init-deploy-9911 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "init-deploy-9911" not found waiting for namespace/init-deploy-9911 to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "init-deploy-9911" not found ----------------------------------------------------------------------------------- create namespace init-deploy-9911 ----------------------------------------------------------------------------------- namespace/init-deploy-9911 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2473-6d392bea-4-cluster3" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- namespace/cert-manager created namespace/cert-manager labeled namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io created serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view created clusterrole.rbac.authorization.k8s.io/cert-manager-view created clusterrole.rbac.authorization.k8s.io/cert-manager-edit created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created role.rbac.authorization.k8s.io/cert-manager:leaderelection created role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. ----------------------------------------------------------------------------------- apply my-env-var-secrets to add PEER_LIST_SRV_PROTOCOL ----------------------------------------------------------------------------------- secret/my-env-var-secrets created ----------------------------------------------------------------------------------- update operator config to set MAX_CONCURRENT_RECONCILES ----------------------------------------------------------------------------------- deployment.apps/percona-xtradb-cluster-operator configured ----------------------------------------------------------------------------------- wait for operator deployment rollout to complete ----------------------------------------------------------------------------------- Waiting for deployment "percona-xtradb-cluster-operator" rollout to finish: 0 of 1 updated replicas are available... deployment "percona-xtradb-cluster-operator" successfully rolled out ----------------------------------------------------------------------------------- wait for old pods to terminate and only new pod to be available ----------------------------------------------------------------------------------- deployment.apps/percona-xtradb-cluster-operator condition met E0516 23:41:15.338595 1527 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/apis/apps/v1/namespaces/pxc-operator/deployments?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator&resourceVersion=1778974873673839020&timeoutSeconds=550&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" deployment.apps/percona-xtradb-cluster-operator condition met E0516 23:41:16.161197 1739 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/apis/apps/v1/namespaces/pxc-operator/deployments?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator&resourceVersion=1778974873673839020&timeoutSeconds=306&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" ----------------------------------------------------------------------------------- verify MAX_CONCURRENT_RECONCILES is applied correctly ----------------------------------------------------------------------------------- Using operator pod: percona-xtradb-cluster-operator-84cc8dc696-bp5kq verifying resource: PerconaXtraDBCluster No logs found for PerconaXtraDBCluster yet (attempt 1/30) verifying resource: PerconaXtraDBCluster No logs found for PerconaXtraDBCluster yet (attempt 2/30) verifying resource: PerconaXtraDBCluster No logs found for PerconaXtraDBCluster yet (attempt 3/30) verifying resource: PerconaXtraDBCluster No logs found for PerconaXtraDBCluster yet (attempt 4/30) verifying resource: PerconaXtraDBCluster No logs found for PerconaXtraDBCluster yet (attempt 5/30) verifying resource: PerconaXtraDBCluster PerconaXtraDBCluster controller has correct worker count: 5 verifying resource: PerconaXtraDBClusterBackup PerconaXtraDBClusterBackup controller has correct worker count: 5 verifying resource: PerconaXtraDBClusterRestore PerconaXtraDBClusterRestore controller has correct worker count: 5 All worker count checks passed ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/some-name created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- Error from server (NotFound): pods "some-name-proxysql-0" not found waiting for pod/some-name-proxysql-0 to become Ready..........Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:46:18.930348 9547 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975176942481000&timeoutSeconds=501&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:46:27.943407 10832 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975186421191000&timeoutSeconds=593&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:47:06.048869 15133 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975224389310000&timeoutSeconds=513&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:47:13.920387 16234 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975232499842000&timeoutSeconds=593&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:47:22.573117 17390 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975240575795000&timeoutSeconds=596&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check if cert-manager issued ssl certificates ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check if service and statefulset created with expected config ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc- ----------------------------------------------------------------------------------- [2026-05-16T23:47:33+0000] compare_kubectl: statefulset/some-name-pxc OK ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql- ----------------------------------------------------------------------------------- [2026-05-16T23:47:34+0000] compare_kubectl: statefulset/some-name-proxysql OK ----------------------------------------------------------------------------------- compare service/some-name-pxc- ----------------------------------------------------------------------------------- [2026-05-16T23:47:36+0000] compare_kubectl: service/some-name-pxc OK ----------------------------------------------------------------------------------- compare service/some-name-proxysql- ----------------------------------------------------------------------------------- [2026-05-16T23:47:37+0000] compare_kubectl: service/some-name-proxysql OK ----------------------------------------------------------------------------------- compare service/some-name-proxysql-unready- ----------------------------------------------------------------------------------- [2026-05-16T23:47:38+0000] compare_kubectl: service/some-name-proxysql-unready OK ----------------------------------------------------------------------------------- check PodDisruptionBudget ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare pdb/some-name-pxc- ----------------------------------------------------------------------------------- [2026-05-16T23:47:40+0000] compare_kubectl: pdb/some-name-pxc OK ----------------------------------------------------------------------------------- compare pdb/some-name-proxysql- ----------------------------------------------------------------------------------- [2026-05-16T23:47:42+0000] compare_kubectl: pdb/some-name-proxysql OK ----------------------------------------------------------------------------------- check if MySQL users created ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:47:45.591175 20441 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975263757362000&timeoutSeconds=302&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:47:56.166583 22036 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975274189226000&timeoutSeconds=421&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:48:09.105709 23951 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975286607527000&timeoutSeconds=474&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check that pmm server user don't have access ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:48:21.303777 25833 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975299579914000&timeoutSeconds=599&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check if ProxySQL users created ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:48:31.387269 27321 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975309462342000&timeoutSeconds=427&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:48:41.417458 28999 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975320438278000&timeoutSeconds=439&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check that pmm server user don't have access ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:48:51.546743 30308 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975329605214000&timeoutSeconds=531&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:49:01.901230 31905 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975340111803000&timeoutSeconds=346&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- write data directly, read from all ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:49:11.527919 819 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975349646266000&timeoutSeconds=390&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:49:20.891771 2424 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975360055110000&timeoutSeconds=456&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:49:31.475028 3818 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975369467049000&timeoutSeconds=316&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:49:44.194181 5688 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975381531430000&timeoutSeconds=308&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- kill Primary Pod, check reelection, check data ----------------------------------------------------------------------------------- pod "some-name-pxc-2" force deleted from init-deploy-9911 namespace Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:50:13.606031 10104 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975412544095003&timeoutSeconds=304&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/some-name-pxc-0 condition met E0516 23:50:23.491929 11456 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dsome-name-pxc-0&resourceVersion=1778975418480111000&timeoutSeconds=417&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/some-name-pxc-0 to become Ready.Ok pod/some-name-pxc-1 condition met E0516 23:50:31.714612 12600 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dsome-name-pxc-1&resourceVersion=1778975428480327000&timeoutSeconds=387&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/some-name-pxc-1 to become Ready.Ok pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:50:56.923962 16235 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975454371811000&timeoutSeconds=552&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:51:09.259129 18212 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975468237010000&timeoutSeconds=310&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:51:25.257978 20253 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975482867617000&timeoutSeconds=310&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:51:40.349735 22459 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975498512883000&timeoutSeconds=549&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check if possible to create no-proxysql cluster ----------------------------------------------------------------------------------- perconaxtradbcluster.pxc.percona.com/no-proxysql created ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- Error from server (NotFound): pods "no-proxysql-pxc-0" not found waiting for pod/no-proxysql-pxc-0 to become Ready......................Ok pod/no-proxysql-pxc-1 condition met waiting for pod/no-proxysql-pxc-1 to become Ready.Ok pod/no-proxysql-pxc-2 condition met waiting for pod/no-proxysql-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- check if service and statefulset created ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare service/no-proxysql-pxc- ----------------------------------------------------------------------------------- [2026-05-16T23:55:07+0000] compare_kubectl: service/no-proxysql-pxc OK ----------------------------------------------------------------------------------- compare service/no-proxysql-proxysql- ----------------------------------------------------------------------------------- Error from server (NotFound): services "no-proxysql-proxysql" not found [2026-05-16T23:55:10+0000] compare_kubectl: service/no-proxysql-proxysql OK ----------------------------------------------------------------------------------- compare configmap/no-proxysql-pxc- ----------------------------------------------------------------------------------- [2026-05-16T23:55:12+0000] compare_kubectl: configmap/no-proxysql-pxc OK ----------------------------------------------------------------------------------- check if storageClass changed ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/no-proxysql-pxc- ----------------------------------------------------------------------------------- [2026-05-16T23:55:14+0000] compare_kubectl: statefulset/no-proxysql-pxc OK ----------------------------------------------------------------------------------- compare statefulset/no-proxysql-proxysql- ----------------------------------------------------------------------------------- Error from server (NotFound): statefulsets.apps "no-proxysql-proxysql" not found [2026-05-16T23:55:17+0000] compare_kubectl: statefulset/no-proxysql-proxysql OK ----------------------------------------------------------------------------------- compare issuer/no-proxysql-pxc-issuer- ----------------------------------------------------------------------------------- [2026-05-16T23:55:19+0000] compare_kubectl: issuer/no-proxysql-pxc-issuer OK ----------------------------------------------------------------------------------- compare certificate/no-proxysql-ssl- ----------------------------------------------------------------------------------- [2026-05-16T23:55:21+0000] compare_kubectl: certificate/no-proxysql-ssl OK ----------------------------------------------------------------------------------- compare certificate/no-proxysql-ssl-internal- ----------------------------------------------------------------------------------- [2026-05-16T23:55:23+0000] compare_kubectl: certificate/no-proxysql-ssl-internal OK ----------------------------------------------------------------------------------- write data, read from all ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:55:26.216907 22391 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975725251519000&timeoutSeconds=486&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:55:38.981540 24049 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975737028567000&timeoutSeconds=345&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:56:01.357704 27241 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975759770964000&timeoutSeconds=316&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:56:10.630173 28543 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975768422555000&timeoutSeconds=387&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:56:18.060492 29822 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975777112288000&timeoutSeconds=421&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-g4689 condition met E0516 23:56:27.762519 31075 reflector.go:227] "Failed to watch" err="Get \"https://34.56.49.150/api/v1/namespaces/init-deploy-9911/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-g4689&resourceVersion=1778975785467669000&timeoutSeconds=471&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-g4689 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok perconaxtradbcluster.pxc.percona.com "no-proxysql" deleted from init-deploy-9911 namespace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n init-deploy-9911 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched perconaxtradbcluster.pxc.percona.com "some-name" deleted from init-deploy-9911 namespace No resources found No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted namespace "cert-manager" deleted customresourcedefinition.apiextensions.k8s.io "challenges.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "orders.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificaterequests.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificates.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "clusterissuers.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "issuers.cert-manager.io" deleted serviceaccount "cert-manager-cainjector" deleted from cert-manager namespace serviceaccount "cert-manager" deleted from cert-manager namespace serviceaccount "cert-manager-webhook" deleted from cert-manager namespace clusterrole.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-cluster-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-edit" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted role.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace role.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace role.rbac.authorization.k8s.io "cert-manager-tokenrequest" deleted from cert-manager namespace role.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" deleted from cert-manager namespace rolebinding.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager-tokenrequest" deleted from cert-manager namespace rolebinding.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" deleted from cert-manager namespace service "cert-manager-cainjector" deleted from cert-manager namespace service "cert-manager" deleted from cert-manager namespace service "cert-manager-webhook" deleted from cert-manager namespace deployment.apps "cert-manager-cainjector" deleted from cert-manager namespace deployment.apps "cert-manager" deleted from cert-manager namespace deployment.apps "cert-manager-webhook" deleted from cert-manager namespace mutatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted validatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------