Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/logs/init-deploy-5-7.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "pxc-operator" not found waiting for namespace/pxc-operator to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster8" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-9d9fbdb5-mgqgm condition met E0516 19:21:30.651445 4308 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-9d9fbdb5-mgqgm&resourceVersion=1778959290290223000&timeoutSeconds=348&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/percona-xtradb-cluster-operator-9d9fbdb5-mgqgm condition met E0516 19:21:34.555781 4879 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-9d9fbdb5-mgqgm&resourceVersion=1778959292944948000&timeoutSeconds=427&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-9d9fbdb5-mgqgm to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces init-deploy-6792 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "init-deploy-6792" not found waiting for namespace/init-deploy-6792 to be deletederror: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "init-deploy-6792" not found ----------------------------------------------------------------------------------- create namespace init-deploy-6792 ----------------------------------------------------------------------------------- namespace/init-deploy-6792 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster8" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- namespace/cert-manager created namespace/cert-manager labeled namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io created serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view created clusterrole.rbac.authorization.k8s.io/cert-manager-view created clusterrole.rbac.authorization.k8s.io/cert-manager-edit created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created role.rbac.authorization.k8s.io/cert-manager:leaderelection created role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. ----------------------------------------------------------------------------------- apply my-env-var-secrets to add PEER_LIST_SRV_PROTOCOL ----------------------------------------------------------------------------------- secret/my-env-var-secrets created ----------------------------------------------------------------------------------- update operator config to set MAX_CONCURRENT_RECONCILES ----------------------------------------------------------------------------------- deployment.apps/percona-xtradb-cluster-operator configured ----------------------------------------------------------------------------------- wait for operator deployment rollout to complete ----------------------------------------------------------------------------------- Waiting for deployment "percona-xtradb-cluster-operator" rollout to finish: 0 of 1 updated replicas are available... deployment "percona-xtradb-cluster-operator" successfully rolled out ----------------------------------------------------------------------------------- wait for old pods to terminate and only new pod to be available ----------------------------------------------------------------------------------- deployment.apps/percona-xtradb-cluster-operator condition met E0516 19:23:59.991102 24018 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/apis/apps/v1/namespaces/pxc-operator/deployments?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator&resourceVersion=1778959437222303020&timeoutSeconds=583&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" deployment.apps/percona-xtradb-cluster-operator condition met E0516 19:24:01.516385 24300 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/apis/apps/v1/namespaces/pxc-operator/deployments?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator&resourceVersion=1778959437222303020&timeoutSeconds=344&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" ----------------------------------------------------------------------------------- verify MAX_CONCURRENT_RECONCILES is applied correctly ----------------------------------------------------------------------------------- Using operator pod: percona-xtradb-cluster-operator-558f9c85b5-q9vf7 verifying resource: PerconaXtraDBCluster No logs found for PerconaXtraDBCluster yet (attempt 1/30) verifying resource: PerconaXtraDBCluster No logs found for PerconaXtraDBCluster yet (attempt 2/30) verifying resource: PerconaXtraDBCluster No logs found for PerconaXtraDBCluster yet (attempt 3/30) verifying resource: PerconaXtraDBCluster No logs found for PerconaXtraDBCluster yet (attempt 4/30) verifying resource: PerconaXtraDBCluster PerconaXtraDBCluster controller has correct worker count: 5 verifying resource: PerconaXtraDBClusterBackup PerconaXtraDBClusterBackup controller has correct worker count: 5 verifying resource: PerconaXtraDBClusterRestore PerconaXtraDBClusterRestore controller has correct worker count: 5 All worker count checks passed ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/some-name created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:28:50.994927 25468 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959729576924000&timeoutSeconds=301&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:28:58.088795 26153 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959736403057000&timeoutSeconds=471&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:29:36.371945 29741 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959775562148000&timeoutSeconds=364&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:29:45.256778 30809 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959783712127000&timeoutSeconds=442&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:29:52.458265 31849 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959791672765000&timeoutSeconds=538&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check if cert-manager issued ssl certificates ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- check if service and statefulset created with expected config ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc- ----------------------------------------------------------------------------------- [2026-05-16T19:30:04+0000] compare_kubectl: statefulset/some-name-pxc OK ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql- ----------------------------------------------------------------------------------- [2026-05-16T19:30:05+0000] compare_kubectl: statefulset/some-name-proxysql OK ----------------------------------------------------------------------------------- compare service/some-name-pxc- ----------------------------------------------------------------------------------- [2026-05-16T19:30:06+0000] compare_kubectl: service/some-name-pxc OK ----------------------------------------------------------------------------------- compare service/some-name-proxysql- ----------------------------------------------------------------------------------- [2026-05-16T19:30:08+0000] compare_kubectl: service/some-name-proxysql OK ----------------------------------------------------------------------------------- compare service/some-name-proxysql-unready- ----------------------------------------------------------------------------------- [2026-05-16T19:30:09+0000] compare_kubectl: service/some-name-proxysql-unready OK ----------------------------------------------------------------------------------- check PodDisruptionBudget ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare pdb/some-name-pxc- ----------------------------------------------------------------------------------- [2026-05-16T19:30:09+0000] compare_kubectl: pdb/some-name-pxc OK ----------------------------------------------------------------------------------- compare pdb/some-name-proxysql- ----------------------------------------------------------------------------------- [2026-05-16T19:30:11+0000] compare_kubectl: pdb/some-name-proxysql OK ----------------------------------------------------------------------------------- check if MySQL users created ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:30:13.329982 2304 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959812541759000&timeoutSeconds=568&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:30:20.081413 3096 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959819318269000&timeoutSeconds=528&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:30:25.733088 3774 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959824932226000&timeoutSeconds=302&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check that pmm server user don't have access ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:30:36.803857 5448 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959835776407000&timeoutSeconds=491&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check if ProxySQL users created ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:30:51.655777 7278 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959849163823000&timeoutSeconds=377&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:31:02.911936 9001 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959860802794000&timeoutSeconds=415&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check that pmm server user don't have access ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:31:14.513147 10642 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959872234998000&timeoutSeconds=378&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:31:24.605947 12257 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959883632692000&timeoutSeconds=470&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- write data directly, read from all ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:31:32.352335 13264 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959891519675000&timeoutSeconds=349&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:31:40.149591 14201 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959898568140000&timeoutSeconds=367&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:31:47.959570 15436 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959906833688000&timeoutSeconds=411&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:31:59.341788 16823 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959917364133000&timeoutSeconds=368&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- kill Primary Pod, check reelection, check data ----------------------------------------------------------------------------------- pod "some-name-pxc-2" force deleted from init-deploy-6792 namespace Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:32:23.765683 20002 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959943024540000&timeoutSeconds=459&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/some-name-pxc-0 condition met E0516 19:32:29.197360 20517 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dsome-name-pxc-0&resourceVersion=1778959948515007007&timeoutSeconds=392&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/some-name-pxc-0 to become Ready.Ok pod/some-name-pxc-1 condition met E0516 19:32:34.626772 21139 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dsome-name-pxc-1&resourceVersion=1778959953520010000&timeoutSeconds=556&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/some-name-pxc-1 to become Ready.Ok pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:33:09.301867 25223 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959986666405000&timeoutSeconds=501&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:33:20.007111 26890 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778959999218670000&timeoutSeconds=492&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:33:28.378282 27891 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778960006856630000&timeoutSeconds=396&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:33:36.718944 28979 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778960014960797000&timeoutSeconds=575&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check if possible to create no-proxysql cluster ----------------------------------------------------------------------------------- perconaxtradbcluster.pxc.percona.com/no-proxysql created ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- Error from server (NotFound): pods "no-proxysql-pxc-0" not found waiting for pod/no-proxysql-pxc-0 to become Ready..............................Ok pod/no-proxysql-pxc-1 condition met waiting for pod/no-proxysql-pxc-1 to become Ready.Ok pod/no-proxysql-pxc-2 condition met waiting for pod/no-proxysql-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- check if service and statefulset created ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare service/no-proxysql-pxc- ----------------------------------------------------------------------------------- [2026-05-16T19:36:51+0000] compare_kubectl: service/no-proxysql-pxc OK ----------------------------------------------------------------------------------- compare service/no-proxysql-proxysql- ----------------------------------------------------------------------------------- Error from server (NotFound): services "no-proxysql-proxysql" not found [2026-05-16T19:36:52+0000] compare_kubectl: service/no-proxysql-proxysql OK ----------------------------------------------------------------------------------- compare configmap/no-proxysql-pxc- ----------------------------------------------------------------------------------- [2026-05-16T19:36:53+0000] compare_kubectl: configmap/no-proxysql-pxc OK ----------------------------------------------------------------------------------- check if storageClass changed ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/no-proxysql-pxc- ----------------------------------------------------------------------------------- [2026-05-16T19:36:54+0000] compare_kubectl: statefulset/no-proxysql-pxc OK ----------------------------------------------------------------------------------- compare statefulset/no-proxysql-proxysql- ----------------------------------------------------------------------------------- Error from server (NotFound): statefulsets.apps "no-proxysql-proxysql" not found [2026-05-16T19:36:55+0000] compare_kubectl: statefulset/no-proxysql-proxysql OK ----------------------------------------------------------------------------------- compare issuer/no-proxysql-pxc-issuer- ----------------------------------------------------------------------------------- [2026-05-16T19:36:56+0000] compare_kubectl: issuer/no-proxysql-pxc-issuer OK ----------------------------------------------------------------------------------- compare certificate/no-proxysql-ssl- ----------------------------------------------------------------------------------- [2026-05-16T19:36:57+0000] compare_kubectl: certificate/no-proxysql-ssl OK ----------------------------------------------------------------------------------- compare certificate/no-proxysql-ssl-internal- ----------------------------------------------------------------------------------- [2026-05-16T19:36:58+0000] compare_kubectl: certificate/no-proxysql-ssl-internal OK ----------------------------------------------------------------------------------- write data, read from all ----------------------------------------------------------------------------------- pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:37:00.445412 25549 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778960219627173000&timeoutSeconds=561&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:37:09.910287 26665 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778960228038136000&timeoutSeconds=420&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:37:27.188032 29262 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778960246323809000&timeoutSeconds=359&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:37:35.080039 30386 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778960254071845000&timeoutSeconds=370&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:37:45.903603 31751 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778960263933279000&timeoutSeconds=478&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-97cb9c68b-dxk78 condition met E0516 19:37:54.415366 474 reflector.go:227] "Failed to watch" err="Get \"https://34.123.251.145/api/v1/namespaces/init-deploy-6792/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-97cb9c68b-dxk78&resourceVersion=1778960272875406000&timeoutSeconds=386&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-97cb9c68b-dxk78 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok perconaxtradbcluster.pxc.percona.com "no-proxysql" deleted from init-deploy-6792 namespace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n init-deploy-6792 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched perconaxtradbcluster.pxc.percona.com "some-name" deleted from init-deploy-6792 namespace No resources found No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted namespace "cert-manager" deleted customresourcedefinition.apiextensions.k8s.io "challenges.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "orders.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificaterequests.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificates.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "clusterissuers.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "issuers.cert-manager.io" deleted serviceaccount "cert-manager-cainjector" deleted from cert-manager namespace serviceaccount "cert-manager" deleted from cert-manager namespace serviceaccount "cert-manager-webhook" deleted from cert-manager namespace clusterrole.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-cluster-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-edit" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted role.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace role.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace role.rbac.authorization.k8s.io "cert-manager-tokenrequest" deleted from cert-manager namespace role.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" deleted from cert-manager namespace rolebinding.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager-tokenrequest" deleted from cert-manager namespace rolebinding.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" deleted from cert-manager namespace service "cert-manager-cainjector" deleted from cert-manager namespace service "cert-manager" deleted from cert-manager namespace service "cert-manager-webhook" deleted from cert-manager namespace deployment.apps "cert-manager-cainjector" deleted from cert-manager namespace deployment.apps "cert-manager" deleted from cert-manager namespace deployment.apps "cert-manager-webhook" deleted from cert-manager namespace mutatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted validatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------