Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1774/e2e-tests/logs/smart-update2-8-0.log WARNING: version difference between client (1.30) and server (1.27) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.30) and server (1.27) exceeds the supported minor version skew of +/-1 E0802 20:55:54.816410 7956 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:55.131956 7956 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:55.242677 7956 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:55.370973 7956 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:55.489641 7956 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: the server doesn't have a resource type "pxc" + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' E0802 20:55:56.709066 8095 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:57.032598 8095 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:57.139296 8095 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:57.246498 8095 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:57.585510 8095 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:57.778902 8095 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:57.887376 8095 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: the server doesn't have a resource type "pxc" E0802 20:55:59.356081 8253 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:59.665869 8253 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:59.772211 8253 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:55:59.878084 8253 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:00.202647 8253 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:00.410757 8253 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:00.520071 8253 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: the server doesn't have a resource type "pxc" E0802 20:56:01.684630 8501 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:01.995240 8501 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:02.101159 8501 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:02.207621 8501 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:02.534601 8501 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:02.743288 8501 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:02.851222 8501 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: the server doesn't have a resource type "pxc-backup" E0802 20:56:04.028067 8769 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:04.345204 8769 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:04.453026 8769 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:04.561944 8769 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:04.888703 8769 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:05.124471 8769 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:05.239478 8769 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: the server doesn't have a resource type "pxc-restore" E0802 20:56:07.806085 9292 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:08.127377 9292 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:08.247312 9292 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:08.356662 9292 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:09.756716 9528 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:10.066893 9528 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:10.173141 9528 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:10.278999 9528 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: resource(s) were provided, but no name was specified E0802 20:56:11.502458 9706 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:11.821649 9706 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:11.928706 9706 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:12.036280 9706 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:13.510294 9880 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:13.637879 9880 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:13.770726 9880 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:13.909097 9880 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: resource(s) were provided, but no name was specified E0802 20:56:15.056880 10105 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:15.365923 10105 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:15.470862 10105 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:15.575735 10105 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:17.090296 10312 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:17.403899 10312 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:17.509880 10312 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:17.615560 10312 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: resource(s) were provided, but no name was specified E0802 20:56:18.868359 10498 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:19.187383 10498 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: unable to retrieve the complete list of server APIs: metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:20.488364 10706 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:20.639578 10706 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:20.748237 10706 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:20.855896 10706 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:22.196092 10893 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:22.416463 10893 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:22.524608 10893 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:22.632008 10893 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: resource(s) were provided, but no name was specified E0802 20:56:24.317083 11124 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:24.631544 11124 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:24.737317 11124 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:24.843600 11124 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:26.306167 11417 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:26.525022 11417 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:26.631071 11417 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:26.737099 11417 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: resource(s) were provided, but no name was specified E0802 20:56:28.401353 11672 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:28.541297 11672 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:28.649790 11672 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:28.761851 11672 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:30.627996 11873 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:30.946854 11873 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:31.052644 11873 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:31.164534 11873 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- E0802 20:56:32.662443 12187 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:32.874229 12187 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:32.980407 12187 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:33.086777 12187 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:32.693450 12194 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:33.007921 12194 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:33.114394 12194 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:33.222126 12194 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request Error from server (NotFound): namespaces "pxc-operator" not found namespace/pxc-operator - E0802 20:56:34.717637 12419 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:35.031289 12419 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:35.141654 12419 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request Error from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- E0802 20:56:35.248597 12419 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request namespace "gmp-public" deleted namespace "gmp-system" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1774-70b9684b-2-cluster4" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied E0802 20:56:38.466136 12823 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:38.586927 12823 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator created serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator created E0802 20:56:41.908481 13227 memcache.go:287] couldn't get resource list for pxc.percona.com/v1-10-0: the server could not find the requested resource E0802 20:56:42.021440 13227 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:42.022773 13227 memcache.go:287] couldn't get resource list for pxc.percona.com/v1-11-0: the server could not find the requested resource E0802 20:56:42.153552 13227 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created E0802 20:56:47.626806 13773 memcache.go:287] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request E0802 20:56:47.762418 13773 memcache.go:121] couldn't get resource list for metrics.k8s.io/v1beta1: the server is currently unable to handle the request pod/percona-xtradb-cluster-operator-56bc5d9fb9-pqc2q condition met pod/percona-xtradb-cluster-operator-56bc5d9fb9-pqc2q condition met percona-xtradb-cluster-operator-56bc5d9fb9-pqc2q.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces smart-update2-16051 ----------------------------------------------------------------------------------- Error from server (NotFound): namespaces "smart-update2-16051" not found namespace/smart-update2-16051 - namespace "gmp-public" deleted Error from server (NotFound): namespaces "smart-update2-16051" not found ----------------------------------------------------------------------------------- create namespace smart-update2-16051 ----------------------------------------------------------------------------------- namespace "gmp-system" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted namespace/smart-update2-16051 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1774-70b9684b-2-cluster4" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created ----------------------------------------------------------------------------------- install version service ----------------------------------------------------------------------------------- configmap/versions created deployment.apps/version-service created service/version-service created ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- namespace/cert-manager created namespace/cert-manager labeled namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io created serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view created clusterrole.rbac.authorization.k8s.io/cert-manager-view created clusterrole.rbac.authorization.k8s.io/cert-manager-edit created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created role.rbac.authorization.k8s.io/cert-manager:leaderelection created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Unable to use a TTY - input is not a terminal or the right kind of file customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com patched deployment.apps/percona-xtradb-cluster-operator env updated ----------------------------------------------------------------------------------- Starting telemetry testing ----------------------------------------------------------------------------------- deployment.apps/version-service-cr created service/version-service-cr created pod "version-service-65d97b645d-d2q2g" deleted ----------------------------------------------------------------------------------- Enable telemetry on operator level ----------------------------------------------------------------------------------- deployment.apps/percona-xtradb-cluster-operator configured pod/percona-xtradb-cluster-operator-7d8dd467f5-l7spg condition met percona-xtradb-cluster-operator-7d8dd467f5-l7spg.Ok ----------------------------------------------------------------------------------- create PXC minimal cluster ----------------------------------------------------------------------------------- deployment.apps/pxc-client created secret/minimal-cluster created secret/some-name-ssl created secret/some-name-ssl-internal created perconaxtradbcluster.pxc.percona.com/minimal-cluster created ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- Error from server (NotFound): pods "minimal-cluster-pxc-0" not found minimal-cluster-pxc-0...........................................................Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/minimal-cluster-haproxy-0 condition met minimal-cluster-haproxy-0Defaulted container "haproxy" out of: haproxy, pxc-monit, pxc-init (init), haproxy-init (init) .Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-894587b8-8857q condition met pxc-client-894587b8-8857q.Ok [ERROR] mysql: option '-P' requires an argument command terminated with exit code 5 pod/pxc-client-894587b8-8857q condition met pxc-client-894587b8-8857q.Ok [ERROR] mysql: option '-P' requires an argument command terminated with exit code 5 ----------------------------------------------------------------------------------- telemetry was disabled in CR but in operator not ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- operator fallback VS should have telemetry ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- CR VS should not have telemetry ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- telemetry was disabled in operator but not in CR ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- telemetry was disabled in CR as well as in operator ----------------------------------------------------------------------------------- perconaxtradbcluster.pxc.percona.com/minimal-cluster patched pod "percona-xtradb-cluster-operator-7d8dd467f5-l7spg" deleted perconaxtradbcluster.pxc.percona.com "minimal-cluster" deleted deployment.apps "pxc-client" deleted ----------------------------------------------------------------------------------- Disabling telemetry on the operator level ----------------------------------------------------------------------------------- pod "version-service-cr-6f7b987d8b-wmrvc" deleted pod "version-service-65d97b645d-j8wvv" deleted deployment.apps/percona-xtradb-cluster-operator configured pod/percona-xtradb-cluster-operator-6dd8f5d74-v8b65 condition met percona-xtradb-cluster-operator-6dd8f5d74-v8b65.Ok ----------------------------------------------------------------------------------- create PXC minimal cluster ----------------------------------------------------------------------------------- deployment.apps/pxc-client created secret/minimal-cluster unchanged secret/some-name-ssl unchanged secret/some-name-ssl-internal unchanged perconaxtradbcluster.pxc.percona.com/minimal-cluster created ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/minimal-cluster-pxc-0 condition met minimal-cluster-pxc-0.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/minimal-cluster-haproxy-0 condition met minimal-cluster-haproxy-0Defaulted container "haproxy" out of: haproxy, pxc-monit, pxc-init (init), haproxy-init (init) .Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-894587b8-qj6j2 condition met pxc-client-894587b8-qj6j2.Ok [ERROR] mysql: option '-P' requires an argument command terminated with exit code 5 pod/pxc-client-894587b8-qj6j2 condition met pxc-client-894587b8-qj6j2.Ok [ERROR] mysql: option '-P' requires an argument command terminated with exit code 5 ----------------------------------------------------------------------------------- telemetry was disabled in CR but in operator not ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- telemetry was disabled in operator but not in CR ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cr VS should have telemetry ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- operator VS should not have telemetry ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- telemetry was disabled in CR as well as in operator ----------------------------------------------------------------------------------- perconaxtradbcluster.pxc.percona.com/minimal-cluster patched pod "percona-xtradb-cluster-operator-6dd8f5d74-v8b65" deleted perconaxtradbcluster.pxc.percona.com "minimal-cluster" deleted deployment.apps "pxc-client" deleted pod "version-service-cr-6f7b987d8b-8ttml" deleted pod "version-service-65d97b645d-9tzrl" deleted ----------------------------------------------------------------------------------- create PXC minimal cluster ----------------------------------------------------------------------------------- deployment.apps/pxc-client created secret/minimal-cluster unchanged secret/some-name-ssl unchanged secret/some-name-ssl-internal unchanged perconaxtradbcluster.pxc.percona.com/minimal-cluster created ----------------------------------------------------------------------------------- check if Pod is started ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/minimal-cluster-pxc-0 condition met minimal-cluster-pxc-0.Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/minimal-cluster-haproxy-0 condition met minimal-cluster-haproxy-0Defaulted container "haproxy" out of: haproxy, pxc-monit, pxc-init (init), haproxy-init (init) .Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-894587b8-lt6cp condition met pxc-client-894587b8-lt6cp.Ok [ERROR] mysql: option '-P' requires an argument command terminated with exit code 5 pod/pxc-client-894587b8-lt6cp condition met pxc-client-894587b8-lt6cp.Ok [ERROR] mysql: option '-P' requires an argument command terminated with exit code 5 ----------------------------------------------------------------------------------- telemetry was disabled in CR but in operator not ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- telemetry was disabled in operator but not in CR ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- telemetry was disabled in CR as well as in operator ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- CR VS should not have telemetry ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- operator VS should not have telemetry ----------------------------------------------------------------------------------- perconaxtradbcluster.pxc.percona.com/minimal-cluster patched pod "percona-xtradb-cluster-operator-6dd8f5d74-855hg" deleted perconaxtradbcluster.pxc.percona.com "minimal-cluster" deleted deployment.apps "pxc-client" deleted deployment.apps "version-service-cr" deleted ----------------------------------------------------------------------------------- Telemetry testing finished ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- PXC cluster with version service offline ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created secret/some-name-ssl unchanged secret/some-name-ssl-internal unchanged deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/smart-update created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- Error from server (NotFound): pods "smart-update-haproxy-0" not found smart-update-haproxy-0.......................................Defaulted container "haproxy" out of: haproxy, pxc-monit, pxc-init (init), haproxy-init (init) .Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/smart-update-pxc-0 condition met smart-update-pxc-0.Ok pod/smart-update-pxc-1 condition met smart-update-pxc-1.Ok pod/smart-update-pxc-2 condition met smart-update-pxc-2.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok Unable to use a TTY - input is not a terminal or the right kind of file ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- PXC cluster update with recommended image by version service ----------------------------------------------------------------------------------- pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok perconaxtradbcluster.pxc.percona.com/smart-update patched Waiting for the last pod to update+ wait_cluster_consistency smart-update 3 2 + local cluster_name=smart-update + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok perconaxtradbcluster.pxc.percona.com "smart-update" deleted persistentvolumeclaim "datadir-smart-update-pxc-0" deleted persistentvolumeclaim "datadir-smart-update-pxc-1" deleted persistentvolumeclaim "datadir-smart-update-pxc-2" deleted ----------------------------------------------------------------------------------- PXC cluster update with the latest image by version service ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets unchanged secret/some-name-ssl unchanged secret/some-name-ssl-internal unchanged deployment.apps/pxc-client unchanged perconaxtradbcluster.pxc.percona.com/smart-update created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/smart-update-haproxy-0 condition met smart-update-haproxy-0Defaulted container "haproxy" out of: haproxy, pxc-monit, pxc-init (init), haproxy-init (init) .Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/smart-update-pxc-0 condition met smart-update-pxc-0.Ok pod/smart-update-pxc-1 condition met smart-update-pxc-1.Ok pod/smart-update-pxc-2 condition met smart-update-pxc-2.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok Unable to use a TTY - input is not a terminal or the right kind of file pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok perconaxtradbcluster.pxc.percona.com/smart-update patched Waiting for the last pod to update+ wait_cluster_consistency smart-update 3 2 + local cluster_name=smart-update + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok perconaxtradbcluster.pxc.percona.com "smart-update" deleted persistentvolumeclaim "datadir-smart-update-pxc-0" deleted persistentvolumeclaim "datadir-smart-update-pxc-1" deleted persistentvolumeclaim "datadir-smart-update-pxc-2" deleted ----------------------------------------------------------------------------------- PXC cluster update with explicitly specified image inside version service ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets unchanged secret/some-name-ssl unchanged secret/some-name-ssl-internal unchanged deployment.apps/pxc-client unchanged perconaxtradbcluster.pxc.percona.com/smart-update created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/smart-update-haproxy-0 condition met smart-update-haproxy-0Defaulted container "haproxy" out of: haproxy, pxc-monit, pxc-init (init), haproxy-init (init) .Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/smart-update-pxc-0 condition met smart-update-pxc-0.Ok pod/smart-update-pxc-1 condition met smart-update-pxc-1.Ok pod/smart-update-pxc-2 condition met smart-update-pxc-2.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok Unable to use a TTY - input is not a terminal or the right kind of file Unable to use a TTY - input is not a terminal or the right kind of file pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok perconaxtradbcluster.pxc.percona.com/smart-update patched Waiting for the last pod to update+ wait_cluster_consistency smart-update 3 2 + local cluster_name=smart-update + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok pod/pxc-client-65c795cbdf-wm8hr condition met pxc-client-65c795cbdf-wm8hr.Ok perconaxtradbcluster.pxc.percona.com "smart-update" deleted persistentvolumeclaim "datadir-smart-update-pxc-0" deleted persistentvolumeclaim "datadir-smart-update-pxc-1" deleted persistentvolumeclaim "datadir-smart-update-pxc-2" deleted ----------------------------------------------------------------------------------- cleanup ----------------------------------------------------------------------------------- deployment.apps "version-service" deleted service "version-service" deleted ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted namespace "cert-manager" deleted customresourcedefinition.apiextensions.k8s.io "certificaterequests.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificates.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "challenges.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "clusterissuers.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "issuers.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "orders.acme.cert-manager.io" deleted serviceaccount "cert-manager-cainjector" deleted serviceaccount "cert-manager" deleted serviceaccount "cert-manager-webhook" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-cluster-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-edit" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted role.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted role.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted role.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" deleted rolebinding.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted rolebinding.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted mutatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted validatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------