Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/logs/security-context-8-0.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name was specified No resources found No resources found No resources found error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified namespace "pxc-operator" deleted waiting for namespace/pxc-operator to be deletedError from server (NotFound): namespaces "pxc-operator" not found ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- namespace/pxc-operator created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster4" modified. ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created pod/percona-xtradb-cluster-operator-9d9fbdb5-v9bpq condition met E0516 20:50:47.982601 20839 reflector.go:227] "Failed to watch" err="Get \"https://34.132.40.237/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-9d9fbdb5-v9bpq&resourceVersion=1778964647622923000&timeoutSeconds=464&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" pod/percona-xtradb-cluster-operator-9d9fbdb5-v9bpq condition met E0516 20:50:52.415803 21542 reflector.go:227] "Failed to watch" err="Get \"https://34.132.40.237/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-9d9fbdb5-v9bpq&resourceVersion=1778964650167391000&timeoutSeconds=530&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-9d9fbdb5-v9bpq to become Ready.Ok error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified error: resource(s) were provided, but no name was specified ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- cleaned up old namespaces security-context-26504 ----------------------------------------------------------------------------------- error: resource(s) were provided, but no name was specified Error from server (NotFound): namespaces "security-context-26504" not found waiting for namespace/security-context-26504 to be deletedError from server (NotFound): namespaces "security-context-26504" not found ----------------------------------------------------------------------------------- create namespace security-context-26504 ----------------------------------------------------------------------------------- namespace/security-context-26504 created Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster4" modified. ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- namespace/cert-manager created namespace/cert-manager labeled namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io created serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view created clusterrole.rbac.authorization.k8s.io/cert-manager-view created clusterrole.rbac.authorization.k8s.io/cert-manager-edit created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created role.rbac.authorization.k8s.io/cert-manager:leaderelection created role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. serviceaccount/percona-xtradb-cluster-operator-workload created ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- secret/my-cluster-secrets created deployment.apps/pxc-client created perconaxtradbcluster.pxc.percona.com/sec-context created ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- error: no matching resources found ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- Error from server (NotFound): pods "sec-context-proxysql-0" not found waiting for pod/sec-context-proxysql-0 to become Ready..........Ok ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- pod/sec-context-pxc-0 condition met waiting for pod/sec-context-pxc-0 to become Ready.Ok pod/sec-context-pxc-1 condition met waiting for pod/sec-context-pxc-1 to become Ready.Ok pod/sec-context-pxc-2 condition met waiting for pod/sec-context-pxc-2 to become Ready.Ok ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- pod/pxc-client-67fc4995bb-prrjj condition met E0516 20:57:02.670131 4508 reflector.go:227] "Failed to watch" err="Get \"https://34.132.40.237/api/v1/namespaces/security-context-26504/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-prrjj&resourceVersion=1778965020623828000&timeoutSeconds=411&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-prrjj to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-prrjj condition met E0516 20:57:12.773063 6051 reflector.go:227] "Failed to watch" err="Get \"https://34.132.40.237/api/v1/namespaces/security-context-26504/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-prrjj&resourceVersion=1778965031096578000&timeoutSeconds=347&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-prrjj to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-prrjj condition met E0516 20:57:51.434357 10758 reflector.go:227] "Failed to watch" err="Get \"https://34.132.40.237/api/v1/namespaces/security-context-26504/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-prrjj&resourceVersion=1778965069497547000&timeoutSeconds=323&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-prrjj to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-prrjj condition met E0516 20:58:01.940706 12217 reflector.go:227] "Failed to watch" err="Get \"https://34.132.40.237/api/v1/namespaces/security-context-26504/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-prrjj&resourceVersion=1778965079588202000&timeoutSeconds=539&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-prrjj to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok pod/pxc-client-67fc4995bb-prrjj condition met E0516 20:58:11.969815 13818 reflector.go:227] "Failed to watch" err="Get \"https://34.132.40.237/api/v1/namespaces/security-context-26504/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-prrjj&resourceVersion=1778965091025981000&timeoutSeconds=460&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-prrjj to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok ----------------------------------------------------------------------------------- check if service and statefulset created with expected config ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/sec-context-pxc- ----------------------------------------------------------------------------------- [2026-05-16T20:58:22+0000] compare_kubectl: statefulset/sec-context-pxc OK ----------------------------------------------------------------------------------- compare statefulset/sec-context-proxysql- ----------------------------------------------------------------------------------- [2026-05-16T20:58:23+0000] compare_kubectl: statefulset/sec-context-proxysql OK ----------------------------------------------------------------------------------- change security context in PXC cluster ----------------------------------------------------------------------------------- perconaxtradbcluster.pxc.percona.com/sec-context configured ----------------------------------------------------------------------------------- check if service and statefulset changed to expected config ----------------------------------------------------------------------------------- ----------------------------------------------------------------------------------- compare statefulset/sec-context-pxc--changes ----------------------------------------------------------------------------------- [2026-05-16T20:59:00+0000] compare_kubectl: statefulset/sec-context-pxc OK ----------------------------------------------------------------------------------- compare statefulset/sec-context-proxysql--changes ----------------------------------------------------------------------------------- [2026-05-16T20:59:01+0000] compare_kubectl: statefulset/sec-context-proxysql OK ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/sec-context to be ready..................... ----------------------------------------------------------------------------------- run pvc backup ----------------------------------------------------------------------------------- perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-pvc created waiting for pxc-backup/on-demand-backup-pvc to reach Succeeded state.......................Succeeded /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/security-context/compare/ ----------------------------------------------------------------------------------- compare job.batch/xb-on-demand-backup-pvc- ----------------------------------------------------------------------------------- [2026-05-16T21:02:18+0000] compare_kubectl: job.batch/xb-on-demand-backup-pvc OK Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster4" modified. ----------------------------------------------------------------------------------- run pvc restore ----------------------------------------------------------------------------------- perconaxtradbclusterrestore.pxc.percona.com/restore-pvc created Error from server (NotFound): pods "restore-src-restore-pvc-sec-context" not found waiting for pod/restore-src-restore-pvc-sec-context to become Ready...........................Defaulted container "ncat" out of: ncat, backup-init (init) .Ok apiVersion: v1 kind: Pod metadata: annotations: openshift.io/scc: privileged creationTimestamp: "2026-05-16T21:03:17Z" generation: 1 labels: app.kubernetes.io/instance: sec-context app.kubernetes.io/managed-by: percona-xtradb-cluster-operator app.kubernetes.io/name: percona-xtradb-cluster app.kubernetes.io/part-of: percona-xtradb-cluster percona.com/restore-svc-name: restore-src-restore-pvc-sec-context name: restore-src-restore-pvc-sec-context namespace: security-context-26504 ownerReferences: - apiVersion: pxc.percona.com/v1 blockOwnerDeletion: true controller: true kind: PerconaXtraDBClusterRestore name: restore-pvc uid: 737bca48-5f4a-4b56-8347-3d62740b003c resourceVersion: "1778965411123439004" uid: c650e299-cc5c-401a-a6c5-fae5bd68c46b spec: containers: - command: - /opt/percona/backup/recovery-pvc-donor.sh image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup imagePullPolicy: Always name: ncat resources: {} securityContext: privileged: true terminationMessagePath: /dev/termination-log terminationMessagePolicy: File volumeMounts: - mountPath: /backup name: backup - mountPath: /etc/mysql/ssl name: ssl - mountPath: /etc/mysql/ssl-internal name: ssl-internal - mountPath: /etc/mysql/vault-keyring-secret name: vault-keyring-secret - mountPath: /opt/percona name: bin - mountPath: /var/run/secrets/kubernetes.io/serviceaccount name: kube-api-access-btnbm readOnly: true dnsPolicy: ClusterFirst enableServiceLinks: true initContainers: - command: - /backup-init-entrypoint.sh image: perconalab/percona-xtradb-cluster-operator:PR-2467-3dc7f023 imagePullPolicy: Always name: backup-init resources: limits: cpu: 50m memory: 50M requests: cpu: 50m memory: 50M securityContext: privileged: true terminationMessagePath: /dev/termination-log terminationMessagePolicy: File volumeMounts: - mountPath: /opt/percona name: bin - mountPath: /var/run/secrets/kubernetes.io/serviceaccount name: kube-api-access-btnbm readOnly: true nodeName: gke-jen-pxc-2467-3dc7f02-default-pool-00808929-vs65 preemptionPolicy: PreemptLowerPriority priority: 0 restartPolicy: Always schedulerName: default-scheduler securityContext: fsGroup: 1001 supplementalGroups: - 1001 - 1002 - 1003 serviceAccount: percona-xtradb-cluster-operator-workload serviceAccountName: percona-xtradb-cluster-operator-workload terminationGracePeriodSeconds: 30 tolerations: - effect: NoExecute key: node.kubernetes.io/not-ready operator: Exists tolerationSeconds: 300 - effect: NoExecute key: node.kubernetes.io/unreachable operator: Exists tolerationSeconds: 300 volumes: - name: backup persistentVolumeClaim: claimName: xb-on-demand-backup-pvc-20260516210135-204c2363 - name: ssl-internal secret: defaultMode: 420 optional: true secretName: some-name-ssl-internal - name: ssl secret: defaultMode: 420 optional: false secretName: some-name-ssl - name: vault-keyring-secret secret: defaultMode: 420 optional: true secretName: sec-context-vault - emptyDir: {} name: bin - name: kube-api-access-btnbm projected: defaultMode: 420 sources: - serviceAccountToken: expirationSeconds: 3607 path: token - configMap: items: - key: ca.crt path: ca.crt name: kube-root-ca.crt - downwardAPI: items: - fieldRef: apiVersion: v1 fieldPath: metadata.namespace path: namespace status: conditions: - lastProbeTime: null lastTransitionTime: "2026-05-16T21:03:29Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-05-16T21:03:30Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-05-16T21:03:31Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-05-16T21:03:31Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-05-16T21:03:17Z" status: "True" type: PodScheduled containerStatuses: - containerID: containerd://28f9df53110c6c9a521f2d1bf56203a162f70d1ebfbd0baec3b657f24b321580 image: docker.io/perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup imageID: docker.io/perconalab/percona-xtradb-cluster-operator@sha256:63fb184385f000c63fd08e4185b973aa0b7ceb2b4a1b48e4ffb2aa9fd49df232 lastState: {} name: ncat ready: true resources: {} restartCount: 0 started: true state: running: startedAt: "2026-05-16T21:03:30Z" user: linux: gid: 1001 supplementalGroups: - 1001 - 1002 - 1003 uid: 1001 volumeMounts: - mountPath: /backup name: backup - mountPath: /etc/mysql/ssl name: ssl - mountPath: /etc/mysql/ssl-internal name: ssl-internal - mountPath: /etc/mysql/vault-keyring-secret name: vault-keyring-secret - mountPath: /opt/percona name: bin - mountPath: /var/run/secrets/kubernetes.io/serviceaccount name: kube-api-access-btnbm readOnly: true recursiveReadOnly: Disabled hostIP: 10.211.0.6 hostIPs: - ip: 10.211.0.6 initContainerStatuses: - allocatedResources: cpu: 50m memory: 50M containerID: containerd://40bb7972e211ac4fd787d26f4f898b8536028474b5be560e02b724815001ffd5 image: docker.io/perconalab/percona-xtradb-cluster-operator:PR-2467-3dc7f023 imageID: docker.io/perconalab/percona-xtradb-cluster-operator@sha256:97503a5e0ad1e33c3f4870c7268ab66737e0fc2679627b9ca80b864aad1feb6b lastState: {} name: backup-init ready: true resources: limits: cpu: 50m memory: 50M requests: cpu: 50m memory: 50M restartCount: 0 started: false state: terminated: containerID: containerd://40bb7972e211ac4fd787d26f4f898b8536028474b5be560e02b724815001ffd5 exitCode: 0 finishedAt: "2026-05-16T21:03:29Z" reason: Completed startedAt: "2026-05-16T21:03:28Z" user: linux: gid: 2 supplementalGroups: - 2 - 1001 - 1002 - 1003 uid: 2 volumeMounts: - mountPath: /opt/percona name: bin - mountPath: /var/run/secrets/kubernetes.io/serviceaccount name: kube-api-access-btnbm readOnly: true recursiveReadOnly: Disabled phase: Running podIP: 10.183.121.47 podIPs: - ip: 10.183.121.47 qosClass: Burstable startTime: "2026-05-16T21:03:17Z" ----------------------------------------------------------------------------------- compare pod/restore-src-restore-pvc-sec-context- ----------------------------------------------------------------------------------- [2026-05-16T21:03:39+0000] compare_kubectl: pod/restore-src-restore-pvc-sec-context OK waiting for pxc-restore/restore-pvc to reach Succeeded state 2026-05-16T21:03:40 pxc-restore/restore-pvc state: Restoring 2026-05-16T21:03:43 pxc-restore/restore-pvc state: Restoring 2026-05-16T21:03:46 pxc-restore/restore-pvc state: Restoring 2026-05-16T21:03:49 pxc-restore/restore-pvc state: Restoring 2026-05-16T21:03:51 pxc-restore/restore-pvc state: Restoring 2026-05-16T21:03:54 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:03:57 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:00 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:03 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:05 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:07 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:10 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:12 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:15 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:17 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:19 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:21 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:23 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:25 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:27 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:29 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:32 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:34 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:36 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:38 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:41 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:43 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:45 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:47 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:49 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:51 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:53 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:55 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:57 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:04:59 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:05:01 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:05:04 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:05:06 pxc-restore/restore-pvc state: Preparing Cluster 2026-05-16T21:05:08 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:10 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:12 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:14 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:16 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:19 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:21 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:24 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:27 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:30 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:33 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:38 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:41 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:44 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:49 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:53 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:05:56 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:00 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:03 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:07 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:10 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:13 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:16 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:19 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:21 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:23 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:25 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:27 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:30 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:32 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:35 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:37 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:39 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:41 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:42 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:44 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:47 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:49 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:51 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:53 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:55 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:57 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:06:59 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:01 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:03 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:06 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:08 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:11 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:13 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:16 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:20 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:22 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:25 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:28 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:30 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:32 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:35 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:37 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:39 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:42 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:44 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:47 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:49 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:52 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:54 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:07:57 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:00 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:02 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:04 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:07 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:09 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:12 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:14 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:16 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:18 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:21 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:24 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:26 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:29 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:32 pxc-restore/restore-pvc state: Starting Cluster 2026-05-16T21:08:35 pxc-restore/restore-pvc state: Succeeded ----------------------------------------------------------------------------------- compare job.batch/restore-job-restore-pvc-sec-context- ----------------------------------------------------------------------------------- [2026-05-16T21:08:37+0000] compare_kubectl: job.batch/restore-job-restore-pvc-sec-context OK ----------------------------------------------------------------------------------- run s3 backup ----------------------------------------------------------------------------------- secret/minio-secret unchanged "hashicorp" already exists with the same configuration, skipping "minio" already exists with the same configuration, skipping Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "percona" chart repository ...Successfully got an update from the "hashicorp" chart repository Update Complete. ⎈Happy Helming!⎈ ----------------------------------------------------------------------------------- install Minio ----------------------------------------------------------------------------------- Error: uninstall: Release not loaded: minio-service: release: not found NAME: minio-service LAST DEPLOYED: Sat May 16 21:08:44 2026 NAMESPACE: security-context-26504 STATUS: deployed REVISION: 1 TEST SUITE: None NOTES: MinIO can be accessed via port 9000 on the following DNS name from within your cluster: minio-service.security-context-26504.cluster.local To access MinIO from localhost, run the below commands: 1. export POD_NAME=$(kubectl get pods --namespace security-context-26504 -l "release=minio-service" -o jsonpath="{.items[0].metadata.name}") 2. kubectl port-forward $POD_NAME 9000 --namespace security-context-26504 Read more about port forwarding here: http://kubernetes.io/docs/user-guide/kubectl/kubectl_port-forward/ You can now access MinIO server on http://localhost:9000. Follow the below steps to connect to MinIO server with mc client: 1. Download the MinIO mc client - https://min.io/docs/minio/linux/reference/minio-mc.html#quickstart 2. export MC_HOST_minio-service-local=http://$(kubectl get secret --namespace security-context-26504 minio-service -o jsonpath="{.data.rootUser}" | base64 --decode):$(kubectl get secret --namespace security-context-26504 minio-service -o jsonpath="{.data.rootPassword}" | base64 --decode)@localhost:9000 3. mc ls minio-service-local pod/minio-service-5fd5489bdc-vdsf4 condition met E0516 21:09:45.060190 2085 reflector.go:227] "Failed to watch" err="Get \"https://34.132.40.237/api/v1/namespaces/security-context-26504/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dminio-service-5fd5489bdc-vdsf4&resourceVersion=1778965783176826000&timeoutSeconds=358&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/minio-service-5fd5489bdc-vdsf4 to become Ready.Ok make_bucket: operator-testing pod "aws-cli" deleted from security-context-26504 namespace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- waiting for pxc/sec-context to be ready perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-s3 created waiting for pxc-backup/on-demand-backup-s3 to reach Succeeded state....................Succeeded ----------------------------------------------------------------------------------- compare job.batch/xb-on-demand-backup-s3- ----------------------------------------------------------------------------------- [2026-05-16T21:10:49+0000] compare_kubectl: job.batch/xb-on-demand-backup-s3 OK ----------------------------------------------------------------------------------- run s3 restore ----------------------------------------------------------------------------------- perconaxtradbclusterrestore.pxc.percona.com/restore-s3 created waiting for pxc-restore/restore-s3 to reach Succeeded state 2026-05-16T21:10:54 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:10:57 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:01 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:04 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:08 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:11 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:14 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:17 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:20 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:23 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:26 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:29 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:31 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:34 pxc-restore/restore-s3 state: Stopping Cluster 2026-05-16T21:11:36 pxc-restore/restore-s3 state: Restoring 2026-05-16T21:11:39 pxc-restore/restore-s3 state: Restoring 2026-05-16T21:11:42 pxc-restore/restore-s3 state: Restoring 2026-05-16T21:11:44 pxc-restore/restore-s3 state: Restoring 2026-05-16T21:11:46 pxc-restore/restore-s3 state: Restoring 2026-05-16T21:11:49 pxc-restore/restore-s3 state: Restoring 2026-05-16T21:11:52 pxc-restore/restore-s3 state: Restoring 2026-05-16T21:11:54 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:11:57 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:11:59 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:01 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:04 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:06 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:09 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:11 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:13 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:16 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:19 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:22 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:25 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:27 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:30 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:32 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:35 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:38 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:41 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:44 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:47 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:51 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:54 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:56 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:12:59 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:13:01 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:13:04 pxc-restore/restore-s3 state: Preparing Cluster 2026-05-16T21:13:07 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:11 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:15 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:18 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:21 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:24 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:27 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:30 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:32 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:35 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:38 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:41 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:43 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:46 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:49 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:51 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:53 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:55 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:13:58 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:00 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:02 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:04 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:07 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:09 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:12 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:14 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:17 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:19 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:22 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:25 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:27 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:31 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:34 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:37 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:40 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:43 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:47 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:50 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:54 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:14:56 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:00 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:03 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:06 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:10 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:14 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:17 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:20 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:23 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:26 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:29 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:31 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:35 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:38 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:40 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:43 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:45 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:48 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:50 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:53 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:56 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:15:59 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:03 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:06 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:09 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:12 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:15 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:18 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:22 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:26 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:29 pxc-restore/restore-s3 state: Starting Cluster 2026-05-16T21:16:31 pxc-restore/restore-s3 state: Succeeded ----------------------------------------------------------------------------------- compare job.batch/restore-job-restore-s3-sec-context- ----------------------------------------------------------------------------------- [2026-05-16T21:16:34+0000] compare_kubectl: job.batch/restore-job-restore-s3-sec-context OK ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + kubectl patch pxc -n security-context-26504 sec-context --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/sec-context patched perconaxtradbcluster.pxc.percona.com "sec-context" deleted from security-context-26504 namespace perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-pvc" deleted from security-context-26504 namespace perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-s3" deleted from security-context-26504 namespace perconaxtradbclusterrestore.pxc.percona.com "restore-pvc" deleted from security-context-26504 namespace perconaxtradbclusterrestore.pxc.percona.com "restore-s3" deleted from security-context-26504 namespace validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted namespace "cert-manager" deleted customresourcedefinition.apiextensions.k8s.io "challenges.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "orders.acme.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificaterequests.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "certificates.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "clusterissuers.cert-manager.io" deleted customresourcedefinition.apiextensions.k8s.io "issuers.cert-manager.io" deleted serviceaccount "cert-manager-cainjector" deleted from cert-manager namespace serviceaccount "cert-manager" deleted from cert-manager namespace serviceaccount "cert-manager-webhook" deleted from cert-manager namespace clusterrole.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-cluster-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-view" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-edit" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrole.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-cainjector" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-issuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-clusterissuers" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificates" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-orders" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-challenges" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-ingress-shim" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-approve:cert-manager-io" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-controller-certificatesigningrequests" deleted clusterrolebinding.rbac.authorization.k8s.io "cert-manager-webhook:subjectaccessreviews" deleted role.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace role.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace role.rbac.authorization.k8s.io "cert-manager-tokenrequest" deleted from cert-manager namespace role.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" deleted from cert-manager namespace rolebinding.rbac.authorization.k8s.io "cert-manager-cainjector:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager:leaderelection" deleted from kube-system namespace rolebinding.rbac.authorization.k8s.io "cert-manager-tokenrequest" deleted from cert-manager namespace rolebinding.rbac.authorization.k8s.io "cert-manager-webhook:dynamic-serving" deleted from cert-manager namespace service "cert-manager-cainjector" deleted from cert-manager namespace service "cert-manager" deleted from cert-manager namespace service "cert-manager-webhook" deleted from cert-manager namespace deployment.apps "cert-manager-cainjector" deleted from cert-manager namespace deployment.apps "cert-manager" deleted from cert-manager namespace deployment.apps "cert-manager-webhook" deleted from cert-manager namespace mutatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted validatingwebhookconfiguration.admissionregistration.k8s.io "cert-manager-webhook" deleted ----------------------------------------------------------------------------------- test passed -----------------------------------------------------------------------------------