++ echo 'Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/logs/tls-issue-cert-manager-ref.log' Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/logs/tls-issue-cert-manager-ref.log ++ '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/cloud-secret.yml ']' ++ SKIP_BACKUPS_TO_AWS_GCP= ++ oc get projects error: No Auth Provider found for name "gcp" +++ kubectl version -o json +++ grep '\-eks\-' +++ jq -r .serverVersion.gitVersion ++ '[' ']' ++ EKS=0 +++ kubectl version -o json +++ jq -r '.serverVersion.major + "." + .serverVersion.minor' +++ /usr/bin/sed -r 's/[^0-9.]+//g' ++ KUBE_VERSION=1.20 +++ helm version -c +++ /usr/bin/sed -re 's/.*SemVer:"([^"]+)".*/\1/; s/.*\bVersion:"([^"]+)".*/\1/' ++ HELM_VERSION=v3.8.1 ++ '[' v3 == v2 ']' + main + create_infra tls-issue-cert-manager-ref-21277 + local ns=tls-issue-cert-manager-ref-21277 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE No resources found + kubectl patch pxc -n sh --type=merge -p '{"metadata":{"finalizers":[]}}' error: resource(s) were provided, but no name, label selector, or --all flag specified + : + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.3hWNByyCgu ++ mktemp + local LAST_ERR=/tmp/tmp.V7V2MOxWps + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.3hWNByyCgu No resources found + cat /tmp/tmp.V7V2MOxWps + rm /tmp/tmp.3hWNByyCgu /tmp/tmp.V7V2MOxWps + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.ptVRJ4xQcQ ++ mktemp + local LAST_ERR=/tmp/tmp.JLsftyEjf1 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.ptVRJ4xQcQ No resources found + cat /tmp/tmp.JLsftyEjf1 + rm /tmp/tmp.ptVRJ4xQcQ /tmp/tmp.JLsftyEjf1 + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.7JOAOHg4xI ++ mktemp + local LAST_ERR=/tmp/tmp.fbJ5JsTPPQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.7JOAOHg4xI No resources found + cat /tmp/tmp.fbJ5JsTPPQ + rm /tmp/tmp.7JOAOHg4xI /tmp/tmp.fbJ5JsTPPQ + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^NAME' + '[' '!' -z '' ']' + kubectl_bin delete namespace pxc-operator + kubectl_bin get ns ++ mktemp + awk '{print$1}' + xargs kubectl delete ns + local LAST_OUT=/tmp/tmp.dNSFmFF6Gw ++ mktemp + local LAST_OUT=/tmp/tmp.eXYP47QooP ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.DRBXNLbl3g + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.zFgs3sxM23 + local exit_status=0 + for i in '$(seq 0 2)' + kubectl delete namespace pxc-operator ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl get ns + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.eXYP47QooP + cat /tmp/tmp.zFgs3sxM23 + rm /tmp/tmp.eXYP47QooP /tmp/tmp.zFgs3sxM23 + return 0 error: resource(s) were provided, but no name, label selector, or --all flag specified + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.dNSFmFF6Gw namespace "pxc-operator" deleted + cat /tmp/tmp.DRBXNLbl3g + rm /tmp/tmp.dNSFmFF6Gw /tmp/tmp.DRBXNLbl3g + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + set +o xtrace namespace/pxc-operator - Error from server (NotFound): namespaces "pxc-operator" not found + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.yVWMWXOZzH ++ mktemp + local LAST_ERR=/tmp/tmp.2jpxlHisi9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl create namespace pxc-operator + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.yVWMWXOZzH namespace/pxc-operator created + cat /tmp/tmp.2jpxlHisi9 + rm /tmp/tmp.yVWMWXOZzH /tmp/tmp.2jpxlHisi9 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.gUrQd2PFj2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.pEiLNaZUuD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.gUrQd2PFj2 ++ cat /tmp/tmp.pEiLNaZUuD ++ rm /tmp/tmp.gUrQd2PFj2 /tmp/tmp.pEiLNaZUuD ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.nKlnHqyBFo ++ mktemp + local LAST_ERR=/tmp/tmp.4gxIjfa3pd + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic --namespace=pxc-operator + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.nKlnHqyBFo Context "gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic" modified. + cat /tmp/tmp.4gxIjfa3pd + rm /tmp/tmp.nKlnHqyBFo /tmp/tmp.4gxIjfa3pd + return 0 + deploy_operator + desc 'start operator' + set +o xtrace ----------------------------------------------------------------------------------- start operator ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.PG99GbscsA ++ mktemp + local LAST_ERR=/tmp/tmp.SRK8i46ShD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/crd.yaml + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.PG99GbscsA customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com unchanged customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com unchanged customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com unchanged customresourcedefinition.apiextensions.k8s.io/perconaxtradbbackups.pxc.percona.com configured + cat /tmp/tmp.SRK8i46ShD + rm /tmp/tmp.PG99GbscsA /tmp/tmp.SRK8i46ShD + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/cw-rbac.yaml + kubectl_bin apply -f - + sed -e 's^namespace: .*^namespace: pxc-operator^' ++ mktemp + local LAST_OUT=/tmp/tmp.kdfmUrUcmx ++ mktemp + local LAST_ERR=/tmp/tmp.tta1ZFVrsl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.kdfmUrUcmx clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.tta1ZFVrsl + rm /tmp/tmp.kdfmUrUcmx /tmp/tmp.tta1ZFVrsl + return 0 + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-1125-706f792a^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/deploy/cw-operator.yaml + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.xK1klw3Evz ++ mktemp + local LAST_ERR=/tmp/tmp.p9FJilWVfN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.xK1klw3Evz deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.p9FJilWVfN + rm /tmp/tmp.xK1klw3Evz /tmp/tmp.p9FJilWVfN + return 0 + sleep 10 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.0bkuASEcVN +++ mktemp ++ local LAST_ERR=/tmp/tmp.dFUkheeSzR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.0bkuASEcVN ++ cat /tmp/tmp.dFUkheeSzR ++ rm /tmp/tmp.0bkuASEcVN /tmp/tmp.dFUkheeSzR ++ return 0 + wait_pod percona-xtradb-cluster-operator-5699d7755d-bc2ck 480 pxc-operator + local pod=percona-xtradb-cluster-operator-5699d7755d-bc2ck + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-5699d7755d-bc2ck ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace percona-xtradb-cluster-operator-5699d7755d-bc2ck.Ok + sleep 3 + create_namespace tls-issue-cert-manager-ref-21277 + local namespace=tls-issue-cert-manager-ref-21277 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^NAME' + '[' '!' -z '' ']' + kubectl_bin delete namespace tls-issue-cert-manager-ref-21277 + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.zlaVtkLmvK + awk '{print$1}' + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.eSExwpzFXp ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.VcB2rjf6KN + local exit_status=0 + local LAST_ERR=/tmp/tmp.UHwW9aBM7Q + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl get ns + for i in '$(seq 0 2)' + kubectl delete namespace tls-issue-cert-manager-ref-21277 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + kubectl delete namespace tls-issue-cert-manager-ref-21277 + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.zlaVtkLmvK + cat /tmp/tmp.VcB2rjf6KN + rm /tmp/tmp.zlaVtkLmvK /tmp/tmp.VcB2rjf6KN + return 0 error: resource(s) were provided, but no name, label selector, or --all flag specified + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + for i in '$(seq 0 2)' + kubectl delete namespace tls-issue-cert-manager-ref-21277 + exit_status=1 + [[ 1 != 0 ]] + sleep 0 + cat /tmp/tmp.eSExwpzFXp + cat /tmp/tmp.UHwW9aBM7Q Error from server (NotFound): namespaces "tls-issue-cert-manager-ref-21277" not found + rm /tmp/tmp.eSExwpzFXp /tmp/tmp.UHwW9aBM7Q + return 1 + : + wait_for_delete namespace/tls-issue-cert-manager-ref-21277 + local res=namespace/tls-issue-cert-manager-ref-21277 + set +o xtrace namespace/tls-issue-cert-manager-ref-21277 - Error from server (NotFound): namespaces "tls-issue-cert-manager-ref-21277" not found + kubectl_bin create namespace tls-issue-cert-manager-ref-21277 ++ mktemp + local LAST_OUT=/tmp/tmp.y3vm001Qqk ++ mktemp + local LAST_ERR=/tmp/tmp.xNe4WnAn78 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl create namespace tls-issue-cert-manager-ref-21277 + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.y3vm001Qqk namespace/tls-issue-cert-manager-ref-21277 created + cat /tmp/tmp.xNe4WnAn78 + rm /tmp/tmp.y3vm001Qqk /tmp/tmp.xNe4WnAn78 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.aZE80yo7FI +++ mktemp ++ local LAST_ERR=/tmp/tmp.EN1NQ5TcBE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl config current-context ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.aZE80yo7FI ++ cat /tmp/tmp.EN1NQ5TcBE ++ rm /tmp/tmp.aZE80yo7FI /tmp/tmp.EN1NQ5TcBE ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic --namespace=tls-issue-cert-manager-ref-21277 ++ mktemp + local LAST_OUT=/tmp/tmp.b9zvGUIZ7s ++ mktemp + local LAST_ERR=/tmp/tmp.K3uaBRUxTq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic --namespace=tls-issue-cert-manager-ref-21277 + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.b9zvGUIZ7s Context "gke_cloud-dev-112233_us-central1-a_jenkins-pxc-706f792a-basic" modified. + cat /tmp/tmp.K3uaBRUxTq + rm /tmp/tmp.b9zvGUIZ7s /tmp/tmp.K3uaBRUxTq + return 0 + apply_secrets + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.I2ETa1j5PG ++ mktemp + local LAST_ERR=/tmp/tmp.bucJKuypqL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/cloud-secret.yml + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.I2ETa1j5PG secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.bucJKuypqL + rm /tmp/tmp.I2ETa1j5PG /tmp/tmp.bucJKuypqL + return 0 + cluster=some-name-tls-issueref + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + deploy_cert_manager + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.3TVkyQBIHP ++ mktemp + local LAST_ERR=/tmp/tmp.SUxM55XgCR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl create namespace cert-manager + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.3TVkyQBIHP namespace/cert-manager created + cat /tmp/tmp.SUxM55XgCR + rm /tmp/tmp.3TVkyQBIHP /tmp/tmp.SUxM55XgCR + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.XI1wX3aiVV ++ mktemp + local LAST_ERR=/tmp/tmp.pjEdz761jI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.XI1wX3aiVV namespace/cert-manager labeled + cat /tmp/tmp.pjEdz761jI + rm /tmp/tmp.XI1wX3aiVV /tmp/tmp.pjEdz761jI + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.5.4/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.nP8J9F6LRi ++ mktemp + local LAST_ERR=/tmp/tmp.W9z9MLcvXn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.5.4/cert-manager.yaml --validate=false + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.nP8J9F6LRi customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io created customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io created namespace/cert-manager configured serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrole.rbac.authorization.k8s.io/cert-manager-view created clusterrole.rbac.authorization.k8s.io/cert-manager-edit created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests created clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews created role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created role.rbac.authorization.k8s.io/cert-manager:leaderelection created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook created + cat /tmp/tmp.W9z9MLcvXn Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.nP8J9F6LRi /tmp/tmp.W9z9MLcvXn + return 0 + sleep 60 + desc 'create issuer' + set +o xtrace ----------------------------------------------------------------------------------- create issuer ----------------------------------------------------------------------------------- + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/issuer.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/issuer.yml + kubectl_bin apply -f - ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/issuer.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1-11-0#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_OUT=/tmp/tmp.oBgihYWuuG + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' ++ mktemp + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.tls-issue-cert-manager-ref-21277~ + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#initImage:.*-init$#initImage: perconalab/percona-xtradb-cluster-operator:PR-1125-706f792a#' + local LAST_ERR=/tmp/tmp.q7AoEmjzVP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.oBgihYWuuG clusterissuer.cert-manager.io/special-selfsigned-issuer created + cat /tmp/tmp.q7AoEmjzVP + rm /tmp/tmp.oBgihYWuuG /tmp/tmp.q7AoEmjzVP + return 0 + sleep 10 + desc 'create pxc cluster' + set +o xtrace ----------------------------------------------------------------------------------- create pxc cluster ----------------------------------------------------------------------------------- + spinup_pxc some-name-tls-issueref /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/some-name-tls-issueref.yml 3 10 /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/secrets_without_tls.yml /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/client.yml + local cluster=some-name-tls-issueref + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/some-name-tls-issueref.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/secrets_without_tls.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/client.yml + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/secrets_without_tls.yml ++ mktemp + local LAST_OUT=/tmp/tmp.pFPfRd42Rd ++ mktemp + local LAST_ERR=/tmp/tmp.hazDKflQOg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/conf/secrets_without_tls.yml + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.pFPfRd42Rd secret/my-cluster-secrets created + cat /tmp/tmp.hazDKflQOg + rm /tmp/tmp.pFPfRd42Rd /tmp/tmp.hazDKflQOg + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/client.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/client.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1-11-0#' ++ mktemp + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#initImage:.*-init$#initImage: perconalab/percona-xtradb-cluster-operator:PR-1125-706f792a#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + local LAST_OUT=/tmp/tmp.XtkIQQzVH0 + /usr/bin/sed -e s~minio-service.#namespace~minio-service.tls-issue-cert-manager-ref-21277~ + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + local LAST_ERR=/tmp/tmp.edkESpcdV0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.XtkIQQzVH0 deployment.apps/pxc-client created + cat /tmp/tmp.edkESpcdV0 + rm /tmp/tmp.XtkIQQzVH0 /tmp/tmp.edkESpcdV0 + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/some-name-tls-issueref.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/some-name-tls-issueref.yml ++ mktemp + local LAST_OUT=/tmp/tmp.x5N4wwuASf + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1-11-0#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.tls-issue-cert-manager-ref-21277~ + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#initImage:.*-init$#initImage: perconalab/percona-xtradb-cluster-operator:PR-1125-706f792a#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + local LAST_ERR=/tmp/tmp.TcNTNhh4vv + local exit_status=0 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/conf/some-name-tls-issueref.yml ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl apply -f - + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.x5N4wwuASf perconaxtradbcluster.pxc.percona.com/some-name-tls-issueref created + cat /tmp/tmp.TcNTNhh4vv + rm /tmp/tmp.x5N4wwuASf /tmp/tmp.TcNTNhh4vv + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name-tls-issueref ++ local target_cluster=some-name-tls-issueref +++ kubectl_bin get pxc some-name-tls-issueref -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.34rBDGnmga ++++ mktemp +++ local LAST_ERR=/tmp/tmp.lqO8ZrLTf1 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ kubectl get pxc some-name-tls-issueref -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ cat /tmp/tmp.34rBDGnmga +++ cat /tmp/tmp.lqO8ZrLTf1 +++ rm /tmp/tmp.34rBDGnmga /tmp/tmp.lqO8ZrLTf1 +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name-tls-issueref -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.rAiBgGHOvG ++++ mktemp +++ local LAST_ERR=/tmp/tmp.U7N7YoRkbz +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ kubectl get pxc some-name-tls-issueref -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ [[ 0 != 0 ]] +++ break +++ cat /tmp/tmp.rAiBgGHOvG +++ cat /tmp/tmp.U7N7YoRkbz +++ rm /tmp/tmp.rAiBgGHOvG /tmp/tmp.U7N7YoRkbz +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-tls-issueref-proxysql ++ return + local proxy=some-name-tls-issueref-proxysql + wait_for_running some-name-tls-issueref-proxysql 1 + local name=some-name-tls-issueref-proxysql + let last_pod=0 + : + local max_retry=480 ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-tls-issueref-proxysql-0 480 + local pod=some-name-tls-issueref-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-tls-issueref-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace some-name-tls-issueref-proxysql-0..........Ok + wait_for_running some-name-tls-issueref-pxc 3 + local name=some-name-tls-issueref-pxc + let last_pod=2 + local max_retry=480 ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-tls-issueref-pxc-0 480 + local pod=some-name-tls-issueref-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-tls-issueref-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace some-name-tls-issueref-pxc-0...............................Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-tls-issueref-pxc-1 480 + local pod=some-name-tls-issueref-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-tls-issueref-pxc-1 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace some-name-tls-issueref-pxc-1..................................Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-tls-issueref-pxc-2 480 + local pod=some-name-tls-issueref-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-tls-issueref-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace some-name-tls-issueref-pxc-2.........................................Ok + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-tls-issueref-proxysql -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-tls-issueref-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Nmj7Bpju4f +++ mktemp ++ local LAST_ERR=/tmp/tmp.2hXkXk9Wg1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.Nmj7Bpju4f ++ cat /tmp/tmp.2hXkXk9Wg1 ++ rm /tmp/tmp.Nmj7Bpju4f /tmp/tmp.2hXkXk9Wg1 ++ return 0 + client_pod=pxc-client-5df7b6757f-wwkdx + wait_pod pxc-client-5df7b6757f-wwkdx + local pod=pxc-client-5df7b6757f-wwkdx + local max_retry=480 + local ns= ++ echo pxc-client-5df7b6757f-wwkdx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5df7b6757f-wwkdx.Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-tls-issueref-proxysql -uroot -proot_password' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-tls-issueref-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PUkpjNa2Gd +++ mktemp ++ local LAST_ERR=/tmp/tmp.UBsuUg8Svp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.PUkpjNa2Gd ++ cat /tmp/tmp.UBsuUg8Svp ++ rm /tmp/tmp.PUkpjNa2Gd /tmp/tmp.UBsuUg8Svp ++ return 0 + client_pod=pxc-client-5df7b6757f-wwkdx + wait_pod pxc-client-5df7b6757f-wwkdx + local pod=pxc-client-5df7b6757f-wwkdx + local max_retry=480 + local ns= ++ echo pxc-client-5df7b6757f-wwkdx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5df7b6757f-wwkdx.Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-tls-issueref-pxc-0.some-name-tls-issueref-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-tls-issueref-pxc-0.some-name-tls-issueref-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-tls-issueref-pxc-0.some-name-tls-issueref-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-tls-issueref-pxc-0.some-name-tls-issueref-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EstyI7Zez3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.cYCrj5gjV1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.EstyI7Zez3 ++ cat /tmp/tmp.cYCrj5gjV1 ++ rm /tmp/tmp.EstyI7Zez3 /tmp/tmp.cYCrj5gjV1 ++ return 0 + client_pod=pxc-client-5df7b6757f-wwkdx + wait_pod pxc-client-5df7b6757f-wwkdx + local pod=pxc-client-5df7b6757f-wwkdx + local max_retry=480 + local ns= ++ echo pxc-client-5df7b6757f-wwkdx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5df7b6757f-wwkdx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Dipb9FFYMI/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/select-1.sql /tmp/tmp.Dipb9FFYMI/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-tls-issueref-pxc-1.some-name-tls-issueref-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-tls-issueref-pxc-1.some-name-tls-issueref-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-tls-issueref-pxc-1.some-name-tls-issueref-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-tls-issueref-pxc-1.some-name-tls-issueref-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ipPgJpXd2N +++ mktemp ++ local LAST_ERR=/tmp/tmp.aoohFnroz0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.ipPgJpXd2N ++ cat /tmp/tmp.aoohFnroz0 ++ rm /tmp/tmp.ipPgJpXd2N /tmp/tmp.aoohFnroz0 ++ return 0 + client_pod=pxc-client-5df7b6757f-wwkdx + wait_pod pxc-client-5df7b6757f-wwkdx + local pod=pxc-client-5df7b6757f-wwkdx + local max_retry=480 + local ns= ++ echo pxc-client-5df7b6757f-wwkdx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5df7b6757f-wwkdx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Dipb9FFYMI/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/select-1.sql /tmp/tmp.Dipb9FFYMI/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-tls-issueref-pxc-2.some-name-tls-issueref-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-tls-issueref-pxc-2.some-name-tls-issueref-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-tls-issueref-pxc-2.some-name-tls-issueref-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-tls-issueref-pxc-2.some-name-tls-issueref-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3SUZw9b1Ym +++ mktemp ++ local LAST_ERR=/tmp/tmp.JXid1EvQ2m ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.3SUZw9b1Ym ++ cat /tmp/tmp.JXid1EvQ2m ++ rm /tmp/tmp.3SUZw9b1Ym /tmp/tmp.JXid1EvQ2m ++ return 0 + client_pod=pxc-client-5df7b6757f-wwkdx + wait_pod pxc-client-5df7b6757f-wwkdx + local pod=pxc-client-5df7b6757f-wwkdx + local max_retry=480 + local ns= ++ echo pxc-client-5df7b6757f-wwkdx ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pxc-client-5df7b6757f-wwkdx.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.Dipb9FFYMI/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/select-1.sql /tmp/tmp.Dipb9FFYMI/select-1.sql ++ is_keyring_plugin_in_use some-name-tls-issueref ++ local cluster=some-name-tls-issueref ++ kubectl_bin exec -it some-name-tls-issueref-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.21MOUiTKKz +++ mktemp ++ local LAST_ERR=/tmp/tmp.MjYLGAxfFP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl exec -it some-name-tls-issueref-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.21MOUiTKKz ++ cat /tmp/tmp.MjYLGAxfFP Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.21MOUiTKKz /tmp/tmp.MjYLGAxfFP ++ return 0 + '[' '' ']' + desc 'check if certificates issued with certmanager' + set +o xtrace ----------------------------------------------------------------------------------- check if certificates issued with certmanager ----------------------------------------------------------------------------------- + tlsSecretsShouldExist some-name-tls-issueref-ssl + local secretName=some-name-tls-issueref-ssl + checkTLSSecret some-name-tls-issueref-ssl ca.crt + local secretName=some-name-tls-issueref-ssl + local dataKey=ca.crt ++ kubectl_bin get secrets/some-name-tls-issueref-ssl -o json ++ jq '.data["ca.crt"]' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mk0CuFwrJy +++ mktemp ++ local LAST_ERR=/tmp/tmp.KvLXIClbQC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get secrets/some-name-tls-issueref-ssl -o json ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.mk0CuFwrJy ++ cat /tmp/tmp.KvLXIClbQC ++ rm /tmp/tmp.mk0CuFwrJy /tmp/tmp.KvLXIClbQC ++ return 0 + local 'secretData="LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURsVENDQW4yZ0F3SUJBZ0lSQUp0UEV2M3ljUWFkRkUrM1VIeXFmV0V3RFFZSktvWklodmNOQVFFTEJRQXcKS2pFb01DWUdBMVVFQXhNZmMyOXRaUzF1WVcxbExYUnNjeTFwYzNOMVpYSmxaaTF3Y205NGVYTnhiREFlRncweQpNakEwTVRJeE16UXdOVEJhRncweU1qQTNNVEV4TXpRd05UQmFNQ294S0RBbUJnTlZCQU1USDNOdmJXVXRibUZ0ClpTMTBiSE10YVhOemRXVnlaV1l0Y0hKdmVIbHpjV3d3Z2dFaU1BMEdDU3FHU0liM0RRRUJBUVVBQTRJQkR3QXcKZ2dFS0FvSUJBUURLbHhXYjVyS0ZyKytkWUJCcFl5VzZJOG8zbml2QXpIZVVCeDhuengxT1RPOFNNcEVvUTNqMApxdDVsVVdRaEhQaTcxc3hXRi9JcWZaZmZmN1BsTHd4emdkT0dlUTBQWk1BRVQzUmpIcUx1OW5DMmlBcXJ1NlZPCmFhaGh2NisyWkJ3MjY4ZWRVazhHZGtRWGgzUW9NNXFMZ3NiT2NMbHBnbUFGeXByYU82bHhaOEIzR1lzQzI5YzYKZFBIOUJGL2svajRvWE5wT3VFOGZueXhyVW8zY0xqTE0rT0IwaEx3NHFpUmFzZXRLL3F3bkRoNlVibTFEMDA1NApCQytUTzZYVGhHait4MVhYVWl2eTlpLzNWaFUvWEZ1RHhPR2h1YjgyWG1WZ0QvSitVSElZRkphZnB2a3NaRWU3CjBJVytRRGNZeDA0aHdQVDQrOHh1cEhYRE1vZHcvb1IzQWdNQkFBR2pnYlV3Z2JJd0RnWURWUjBQQVFIL0JBUUQKQWdLa01BOEdBMVVkRXdFQi93UUZNQU1CQWY4d0hRWURWUjBPQkJZRUZINUN2V3RxTlBxc0FObXhyTTRRRjFzdwplczhJTUhBR0ExVWRFUVJwTUdlQ0duTnZiV1V0Ym1GdFpTMTBiSE10YVhOemRXVnlaV1l0Y0hoamdod3FMbk52CmJXVXRibUZ0WlMxMGJITXRhWE56ZFdWeVpXWXRjSGhqZ2lFcUxuTnZiV1V0Ym1GdFpTMTBiSE10YVhOemRXVnkKWldZdGNISnZlSGx6Y1d5Q0NIUmxjM1F1WTI5dE1BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQTdCUjE5M2V1VwpPQ1NUTzBRU3dJREEzdGltdEc3THBvUUJXSk5jYXl2RnEySkZGcGV0MnNpS21Qc1N3K3dOMWxvUUFHMTNpTzhjCkVwYkdEN2MvVklENnRMemhhODBuRkJ4SGsyZ3EvaVBlVDZiQU5EZWE4Y2pZU005TVRwVnhSYitIS2o2SmxsSnQKdFQzTmtSUzhiZEZBMTRjUlhNaUJRRHRiTGRONjh3VkEwYXhwMFNjV25aMWpMUjNKdEo2V2tOOXovcjhuOTNQdQpIT0gvTmpXRzVLOFlhZmFRVXhJUnNIcHRFWEppWVYvSHZjeVlBS1Y1S1djNGFvWnpxUWVjMkp5VGZ0MzdVanVwClBLeXhQZHQ0NG9YUDlwWWRoWlpqL0l3YjRHTTlJTXJWWExHTmxqNnNRcWJoNy8rcnMxUVpQd2xLanZ3V2xSL3EKdzFZZGhOb3FFZFA3Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K"' + '[' -z '"LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURsVENDQW4yZ0F3SUJBZ0lSQUp0UEV2M3ljUWFkRkUrM1VIeXFmV0V3RFFZSktvWklodmNOQVFFTEJRQXcKS2pFb01DWUdBMVVFQXhNZmMyOXRaUzF1WVcxbExYUnNjeTFwYzNOMVpYSmxaaTF3Y205NGVYTnhiREFlRncweQpNakEwTVRJeE16UXdOVEJhRncweU1qQTNNVEV4TXpRd05UQmFNQ294S0RBbUJnTlZCQU1USDNOdmJXVXRibUZ0ClpTMTBiSE10YVhOemRXVnlaV1l0Y0hKdmVIbHpjV3d3Z2dFaU1BMEdDU3FHU0liM0RRRUJBUVVBQTRJQkR3QXcKZ2dFS0FvSUJBUURLbHhXYjVyS0ZyKytkWUJCcFl5VzZJOG8zbml2QXpIZVVCeDhuengxT1RPOFNNcEVvUTNqMApxdDVsVVdRaEhQaTcxc3hXRi9JcWZaZmZmN1BsTHd4emdkT0dlUTBQWk1BRVQzUmpIcUx1OW5DMmlBcXJ1NlZPCmFhaGh2NisyWkJ3MjY4ZWRVazhHZGtRWGgzUW9NNXFMZ3NiT2NMbHBnbUFGeXByYU82bHhaOEIzR1lzQzI5YzYKZFBIOUJGL2svajRvWE5wT3VFOGZueXhyVW8zY0xqTE0rT0IwaEx3NHFpUmFzZXRLL3F3bkRoNlVibTFEMDA1NApCQytUTzZYVGhHait4MVhYVWl2eTlpLzNWaFUvWEZ1RHhPR2h1YjgyWG1WZ0QvSitVSElZRkphZnB2a3NaRWU3CjBJVytRRGNZeDA0aHdQVDQrOHh1cEhYRE1vZHcvb1IzQWdNQkFBR2pnYlV3Z2JJd0RnWURWUjBQQVFIL0JBUUQKQWdLa01BOEdBMVVkRXdFQi93UUZNQU1CQWY4d0hRWURWUjBPQkJZRUZINUN2V3RxTlBxc0FObXhyTTRRRjFzdwplczhJTUhBR0ExVWRFUVJwTUdlQ0duTnZiV1V0Ym1GdFpTMTBiSE10YVhOemRXVnlaV1l0Y0hoamdod3FMbk52CmJXVXRibUZ0WlMxMGJITXRhWE56ZFdWeVpXWXRjSGhqZ2lFcUxuTnZiV1V0Ym1GdFpTMTBiSE10YVhOemRXVnkKWldZdGNISnZlSGx6Y1d5Q0NIUmxjM1F1WTI5dE1BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQTdCUjE5M2V1VwpPQ1NUTzBRU3dJREEzdGltdEc3THBvUUJXSk5jYXl2RnEySkZGcGV0MnNpS21Qc1N3K3dOMWxvUUFHMTNpTzhjCkVwYkdEN2MvVklENnRMemhhODBuRkJ4SGsyZ3EvaVBlVDZiQU5EZWE4Y2pZU005TVRwVnhSYitIS2o2SmxsSnQKdFQzTmtSUzhiZEZBMTRjUlhNaUJRRHRiTGRONjh3VkEwYXhwMFNjV25aMWpMUjNKdEo2V2tOOXovcjhuOTNQdQpIT0gvTmpXRzVLOFlhZmFRVXhJUnNIcHRFWEppWVYvSHZjeVlBS1Y1S1djNGFvWnpxUWVjMkp5VGZ0MzdVanVwClBLeXhQZHQ0NG9YUDlwWWRoWlpqL0l3YjRHTTlJTXJWWExHTmxqNnNRcWJoNy8rcnMxUVpQd2xLanZ3V2xSL3EKdzFZZGhOb3FFZFA3Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K"' ']' + checkTLSSecret some-name-tls-issueref-ssl tls.crt + local secretName=some-name-tls-issueref-ssl + local dataKey=tls.crt ++ kubectl_bin get secrets/some-name-tls-issueref-ssl -o json ++ jq '.data["tls.crt"]' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nmtY2PPyKg +++ mktemp ++ local LAST_ERR=/tmp/tmp.qj9DJ7URCi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get secrets/some-name-tls-issueref-ssl -o json ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.nmtY2PPyKg ++ cat /tmp/tmp.qj9DJ7URCi ++ rm /tmp/tmp.nmtY2PPyKg /tmp/tmp.qj9DJ7URCi ++ return 0 + local 'secretData="LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURsVENDQW4yZ0F3SUJBZ0lSQUp0UEV2M3ljUWFkRkUrM1VIeXFmV0V3RFFZSktvWklodmNOQVFFTEJRQXcKS2pFb01DWUdBMVVFQXhNZmMyOXRaUzF1WVcxbExYUnNjeTFwYzNOMVpYSmxaaTF3Y205NGVYTnhiREFlRncweQpNakEwTVRJeE16UXdOVEJhRncweU1qQTNNVEV4TXpRd05UQmFNQ294S0RBbUJnTlZCQU1USDNOdmJXVXRibUZ0ClpTMTBiSE10YVhOemRXVnlaV1l0Y0hKdmVIbHpjV3d3Z2dFaU1BMEdDU3FHU0liM0RRRUJBUVVBQTRJQkR3QXcKZ2dFS0FvSUJBUURLbHhXYjVyS0ZyKytkWUJCcFl5VzZJOG8zbml2QXpIZVVCeDhuengxT1RPOFNNcEVvUTNqMApxdDVsVVdRaEhQaTcxc3hXRi9JcWZaZmZmN1BsTHd4emdkT0dlUTBQWk1BRVQzUmpIcUx1OW5DMmlBcXJ1NlZPCmFhaGh2NisyWkJ3MjY4ZWRVazhHZGtRWGgzUW9NNXFMZ3NiT2NMbHBnbUFGeXByYU82bHhaOEIzR1lzQzI5YzYKZFBIOUJGL2svajRvWE5wT3VFOGZueXhyVW8zY0xqTE0rT0IwaEx3NHFpUmFzZXRLL3F3bkRoNlVibTFEMDA1NApCQytUTzZYVGhHait4MVhYVWl2eTlpLzNWaFUvWEZ1RHhPR2h1YjgyWG1WZ0QvSitVSElZRkphZnB2a3NaRWU3CjBJVytRRGNZeDA0aHdQVDQrOHh1cEhYRE1vZHcvb1IzQWdNQkFBR2pnYlV3Z2JJd0RnWURWUjBQQVFIL0JBUUQKQWdLa01BOEdBMVVkRXdFQi93UUZNQU1CQWY4d0hRWURWUjBPQkJZRUZINUN2V3RxTlBxc0FObXhyTTRRRjFzdwplczhJTUhBR0ExVWRFUVJwTUdlQ0duTnZiV1V0Ym1GdFpTMTBiSE10YVhOemRXVnlaV1l0Y0hoamdod3FMbk52CmJXVXRibUZ0WlMxMGJITXRhWE56ZFdWeVpXWXRjSGhqZ2lFcUxuTnZiV1V0Ym1GdFpTMTBiSE10YVhOemRXVnkKWldZdGNISnZlSGx6Y1d5Q0NIUmxjM1F1WTI5dE1BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQTdCUjE5M2V1VwpPQ1NUTzBRU3dJREEzdGltdEc3THBvUUJXSk5jYXl2RnEySkZGcGV0MnNpS21Qc1N3K3dOMWxvUUFHMTNpTzhjCkVwYkdEN2MvVklENnRMemhhODBuRkJ4SGsyZ3EvaVBlVDZiQU5EZWE4Y2pZU005TVRwVnhSYitIS2o2SmxsSnQKdFQzTmtSUzhiZEZBMTRjUlhNaUJRRHRiTGRONjh3VkEwYXhwMFNjV25aMWpMUjNKdEo2V2tOOXovcjhuOTNQdQpIT0gvTmpXRzVLOFlhZmFRVXhJUnNIcHRFWEppWVYvSHZjeVlBS1Y1S1djNGFvWnpxUWVjMkp5VGZ0MzdVanVwClBLeXhQZHQ0NG9YUDlwWWRoWlpqL0l3YjRHTTlJTXJWWExHTmxqNnNRcWJoNy8rcnMxUVpQd2xLanZ3V2xSL3EKdzFZZGhOb3FFZFA3Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K"' + '[' -z '"LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURsVENDQW4yZ0F3SUJBZ0lSQUp0UEV2M3ljUWFkRkUrM1VIeXFmV0V3RFFZSktvWklodmNOQVFFTEJRQXcKS2pFb01DWUdBMVVFQXhNZmMyOXRaUzF1WVcxbExYUnNjeTFwYzNOMVpYSmxaaTF3Y205NGVYTnhiREFlRncweQpNakEwTVRJeE16UXdOVEJhRncweU1qQTNNVEV4TXpRd05UQmFNQ294S0RBbUJnTlZCQU1USDNOdmJXVXRibUZ0ClpTMTBiSE10YVhOemRXVnlaV1l0Y0hKdmVIbHpjV3d3Z2dFaU1BMEdDU3FHU0liM0RRRUJBUVVBQTRJQkR3QXcKZ2dFS0FvSUJBUURLbHhXYjVyS0ZyKytkWUJCcFl5VzZJOG8zbml2QXpIZVVCeDhuengxT1RPOFNNcEVvUTNqMApxdDVsVVdRaEhQaTcxc3hXRi9JcWZaZmZmN1BsTHd4emdkT0dlUTBQWk1BRVQzUmpIcUx1OW5DMmlBcXJ1NlZPCmFhaGh2NisyWkJ3MjY4ZWRVazhHZGtRWGgzUW9NNXFMZ3NiT2NMbHBnbUFGeXByYU82bHhaOEIzR1lzQzI5YzYKZFBIOUJGL2svajRvWE5wT3VFOGZueXhyVW8zY0xqTE0rT0IwaEx3NHFpUmFzZXRLL3F3bkRoNlVibTFEMDA1NApCQytUTzZYVGhHait4MVhYVWl2eTlpLzNWaFUvWEZ1RHhPR2h1YjgyWG1WZ0QvSitVSElZRkphZnB2a3NaRWU3CjBJVytRRGNZeDA0aHdQVDQrOHh1cEhYRE1vZHcvb1IzQWdNQkFBR2pnYlV3Z2JJd0RnWURWUjBQQVFIL0JBUUQKQWdLa01BOEdBMVVkRXdFQi93UUZNQU1CQWY4d0hRWURWUjBPQkJZRUZINUN2V3RxTlBxc0FObXhyTTRRRjFzdwplczhJTUhBR0ExVWRFUVJwTUdlQ0duTnZiV1V0Ym1GdFpTMTBiSE10YVhOemRXVnlaV1l0Y0hoamdod3FMbk52CmJXVXRibUZ0WlMxMGJITXRhWE56ZFdWeVpXWXRjSGhqZ2lFcUxuTnZiV1V0Ym1GdFpTMTBiSE10YVhOemRXVnkKWldZdGNISnZlSGx6Y1d5Q0NIUmxjM1F1WTI5dE1BMEdDU3FHU0liM0RRRUJDd1VBQTRJQkFRQTdCUjE5M2V1VwpPQ1NUTzBRU3dJREEzdGltdEc3THBvUUJXSk5jYXl2RnEySkZGcGV0MnNpS21Qc1N3K3dOMWxvUUFHMTNpTzhjCkVwYkdEN2MvVklENnRMemhhODBuRkJ4SGsyZ3EvaVBlVDZiQU5EZWE4Y2pZU005TVRwVnhSYitIS2o2SmxsSnQKdFQzTmtSUzhiZEZBMTRjUlhNaUJRRHRiTGRONjh3VkEwYXhwMFNjV25aMWpMUjNKdEo2V2tOOXovcjhuOTNQdQpIT0gvTmpXRzVLOFlhZmFRVXhJUnNIcHRFWEppWVYvSHZjeVlBS1Y1S1djNGFvWnpxUWVjMkp5VGZ0MzdVanVwClBLeXhQZHQ0NG9YUDlwWWRoWlpqL0l3YjRHTTlJTXJWWExHTmxqNnNRcWJoNy8rcnMxUVpQd2xLanZ3V2xSL3EKdzFZZGhOb3FFZFA3Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K"' ']' + checkTLSSecret some-name-tls-issueref-ssl tls.key + local secretName=some-name-tls-issueref-ssl + local dataKey=tls.key ++ kubectl_bin get secrets/some-name-tls-issueref-ssl -o json ++ jq '.data["tls.key"]' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hywXdHK5l6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.0LfY8Nxitf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get secrets/some-name-tls-issueref-ssl -o json ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.hywXdHK5l6 ++ cat /tmp/tmp.0LfY8Nxitf ++ rm /tmp/tmp.hywXdHK5l6 /tmp/tmp.0LfY8Nxitf ++ return 0 + local 'secretData="LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBeXBjVm0rYXloYS92bldBUWFXTWx1aVBLTjU0cndNeDNsQWNmSjg4ZFRrenZFaktSCktFTjQ5S3JlWlZGa0lSejR1OWJNVmhmeUtuMlgzMyt6NVM4TWM0SFRobmtORDJUQUJFOTBZeDZpN3Zad3RvZ0sKcTd1bFRtbW9ZYit2dG1RY051dkhuVkpQQm5aRUY0ZDBLRE9haTRMR3puQzVhWUpnQmNxYTJqdXBjV2ZBZHhtTApBdHZYT25UeC9RUmY1UDQrS0Z6YVRyaFBINThzYTFLTjNDNHl6UGpnZElTOE9Lb2tXckhyU3Y2c0p3NGVsRzV0ClE5Tk9lQVF2a3p1bDA0Um8vc2RWMTFJcjh2WXY5MVlWUDF4Ymc4VGhvYm0vTmw1bFlBL3lmbEJ5R0JTV242YjUKTEdSSHU5Q0Z2a0EzR01kT0ljRDArUHZNYnFSMXd6S0hjUDZFZHdJREFRQUJBb0lCQVFDK3BINkpQd2N3RnVmOQp5TENvKzh3ODVnRWRaWkwyMDNLVHNqLzNUa2ZkcnpncWY3dHBrRlQ2cVBDdmwzQUtXU0VZWW9IWFlxRllxdWVJClFyd2ZKNjhuNHp3U2gwZHEvUHhWRWhBTVZvQXRMMnZSMlNOeGlCdC9OcDVsbmVUVVZsRFd2Zm9TYWFwRUFuY0IKVFU1WnBuWUFZd2lDckVGT3NJbDc2YlRmT05XekNLcS9MdUcrNEFwUFdFdU5DK3hLZ3lOTnlBVEpiUjFKdUlDMQpxbFpJOXpRV2VWa1B5ODNKYUUyZDdRd0VBbFN1Tno3YzNmNVI3U1ZmTTZpbUJIMnpmN1I3eitGNEx2OElHa0lpCklRRGJSN1luYmhIRU41N21JazZoblpxU3hGdk0zT2lMM2k0MFROWVpXYW80b0NOSld2UHl3N280c2J6ay9kN1EKRW5XRmxmT0JBb0dCQU9mQm5mZVliajJSNHV5bFpoYTI2b0U5VFZ6Q1FSRzlUbWFsUWZYMU5uWkEzKzRiK2hSYgpYSWd4M1pmdEZnVmVyaDFzdCs3RHhTcDR4c0FnU1h1aytqSDJNczdjT081TDlUdFZwWk9KQlhGRllXR1JlRTZSClZEK0VIa285ZDV0WHJFUjRVSG9wZWRRNlp0elExQ2ZCWmovVXAvVTZUV3lwTE9mNUJGOFBJd3pwQW9HQkFOL0kKWjR3dml3UjlzQ2xpUjcyTUR0VU9lc1pOSVY2UVNOTmxoRVgvYTAvM2lTaUExZnM2WEJlN3kvTCtOZHRxUDFjTQpCODZoblRRTk4xdmg4RktwallrczVEelR4MGx4NWgxQ2lTU1VWUXltaytFV2JUSDFHVlpjRTJtT0lGZHRZbnl3CmwzRTRSTE5PTlFLdDBiRjB6d1MvS3hPZU9IWjFLS0xScFpWZTJxcGZBb0dBRkNPWUZZN3liWGxDN2V1RmpvM0EKblJCMXE0eloySy9KbzFQVEFvdVhldmt5UEYvRUFJYmI2SjNrc3JJUWl0RHFjbllqR2FWN2ViQmJIUkR4TWNScApWWFFlZ0NsVTdTZytuRFJhR2hMN0hSZTVXYUFwZ3pKUHp1RXRQS2hRdjNnRDZVTkhRd1pkcHRxWVc3SHQ3Yko1CnJ5eE1tQnZENUtCaTNabWZEcCtqSFJrQ2dZQWRSdHkxTEdzbGxVUDh4Z3pZc1lBblE1T3dhZHJad2txa3pYMDIKb3lubzlsdU5YdjJmdG96aDFnSGlzbnRkVDFNbUVqN2JPNjJUWjc4OFdCelcwN1BlcEc3ZnZsMHVNS2FJWVhJTAp4UlF6YVgralRRblFKVjdCTzU2V2lxK1VnSTBIL3hweEI0bUZZOC92bjhuMVMzRGlFQmlVb25pRWx1TlFHZGx0CjUxRDRSUUtCZ0Z5QlY2czBsTFo0REFZZVlybXpTcllRM2hiUVZiMVpVQnZtUVczVEJ2OFNxRGcrajVMNit4VkYKVW8vc1NrZUFnL212RzZsdURyUlFIWHpqTDJ5cTVpQ0VUR0pyaEJ0em9DVnRXTmJ5aHNMeFZIS0FTOGh0emYvcAoxRnF1Nk5GYnZwbHNwdWZZN1hEcjd0M05OOGRKcFoyVFBmMm9SWFVMNlFJVXJBVmZJdWFUCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="' + '[' -z '"LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFb3dJQkFBS0NBUUVBeXBjVm0rYXloYS92bldBUWFXTWx1aVBLTjU0cndNeDNsQWNmSjg4ZFRrenZFaktSCktFTjQ5S3JlWlZGa0lSejR1OWJNVmhmeUtuMlgzMyt6NVM4TWM0SFRobmtORDJUQUJFOTBZeDZpN3Zad3RvZ0sKcTd1bFRtbW9ZYit2dG1RY051dkhuVkpQQm5aRUY0ZDBLRE9haTRMR3puQzVhWUpnQmNxYTJqdXBjV2ZBZHhtTApBdHZYT25UeC9RUmY1UDQrS0Z6YVRyaFBINThzYTFLTjNDNHl6UGpnZElTOE9Lb2tXckhyU3Y2c0p3NGVsRzV0ClE5Tk9lQVF2a3p1bDA0Um8vc2RWMTFJcjh2WXY5MVlWUDF4Ymc4VGhvYm0vTmw1bFlBL3lmbEJ5R0JTV242YjUKTEdSSHU5Q0Z2a0EzR01kT0ljRDArUHZNYnFSMXd6S0hjUDZFZHdJREFRQUJBb0lCQVFDK3BINkpQd2N3RnVmOQp5TENvKzh3ODVnRWRaWkwyMDNLVHNqLzNUa2ZkcnpncWY3dHBrRlQ2cVBDdmwzQUtXU0VZWW9IWFlxRllxdWVJClFyd2ZKNjhuNHp3U2gwZHEvUHhWRWhBTVZvQXRMMnZSMlNOeGlCdC9OcDVsbmVUVVZsRFd2Zm9TYWFwRUFuY0IKVFU1WnBuWUFZd2lDckVGT3NJbDc2YlRmT05XekNLcS9MdUcrNEFwUFdFdU5DK3hLZ3lOTnlBVEpiUjFKdUlDMQpxbFpJOXpRV2VWa1B5ODNKYUUyZDdRd0VBbFN1Tno3YzNmNVI3U1ZmTTZpbUJIMnpmN1I3eitGNEx2OElHa0lpCklRRGJSN1luYmhIRU41N21JazZoblpxU3hGdk0zT2lMM2k0MFROWVpXYW80b0NOSld2UHl3N280c2J6ay9kN1EKRW5XRmxmT0JBb0dCQU9mQm5mZVliajJSNHV5bFpoYTI2b0U5VFZ6Q1FSRzlUbWFsUWZYMU5uWkEzKzRiK2hSYgpYSWd4M1pmdEZnVmVyaDFzdCs3RHhTcDR4c0FnU1h1aytqSDJNczdjT081TDlUdFZwWk9KQlhGRllXR1JlRTZSClZEK0VIa285ZDV0WHJFUjRVSG9wZWRRNlp0elExQ2ZCWmovVXAvVTZUV3lwTE9mNUJGOFBJd3pwQW9HQkFOL0kKWjR3dml3UjlzQ2xpUjcyTUR0VU9lc1pOSVY2UVNOTmxoRVgvYTAvM2lTaUExZnM2WEJlN3kvTCtOZHRxUDFjTQpCODZoblRRTk4xdmg4RktwallrczVEelR4MGx4NWgxQ2lTU1VWUXltaytFV2JUSDFHVlpjRTJtT0lGZHRZbnl3CmwzRTRSTE5PTlFLdDBiRjB6d1MvS3hPZU9IWjFLS0xScFpWZTJxcGZBb0dBRkNPWUZZN3liWGxDN2V1RmpvM0EKblJCMXE0eloySy9KbzFQVEFvdVhldmt5UEYvRUFJYmI2SjNrc3JJUWl0RHFjbllqR2FWN2ViQmJIUkR4TWNScApWWFFlZ0NsVTdTZytuRFJhR2hMN0hSZTVXYUFwZ3pKUHp1RXRQS2hRdjNnRDZVTkhRd1pkcHRxWVc3SHQ3Yko1CnJ5eE1tQnZENUtCaTNabWZEcCtqSFJrQ2dZQWRSdHkxTEdzbGxVUDh4Z3pZc1lBblE1T3dhZHJad2txa3pYMDIKb3lubzlsdU5YdjJmdG96aDFnSGlzbnRkVDFNbUVqN2JPNjJUWjc4OFdCelcwN1BlcEc3ZnZsMHVNS2FJWVhJTAp4UlF6YVgralRRblFKVjdCTzU2V2lxK1VnSTBIL3hweEI0bUZZOC92bjhuMVMzRGlFQmlVb25pRWx1TlFHZGx0CjUxRDRSUUtCZ0Z5QlY2czBsTFo0REFZZVlybXpTcllRM2hiUVZiMVpVQnZtUVczVEJ2OFNxRGcrajVMNit4VkYKVW8vc1NrZUFnL212RzZsdURyUlFIWHpqTDJ5cTVpQ0VUR0pyaEJ0em9DVnRXTmJ5aHNMeFZIS0FTOGh0emYvcAoxRnF1Nk5GYnZwbHNwdWZZN1hEcjd0M05OOGRKcFoyVFBmMm9SWFVMNlFJVXJBVmZJdWFUCi0tLS0tRU5EIFJTQSBQUklWQVRFIEtFWS0tLS0tCg=="' ']' + desc 'check if issuer created' + set +o xtrace ----------------------------------------------------------------------------------- check if issuer created ----------------------------------------------------------------------------------- + compare_kubectl clusterissuer/special-selfsigned-issuer + local resource=clusterissuer/special-selfsigned-issuer + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/clusterissuer_special-selfsigned-issuer.yml + local new_result=/tmp/tmp.Dipb9FFYMI/clusterissuer_special-selfsigned-issuer.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/clusterissuer_special-selfsigned-issuer-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/clusterissuer_special-selfsigned-issuer-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/clusterissuer_special-selfsigned-issuer-80.yml ']' + version_gt 1.22 ++ echo '1.20 >= 1.22' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.21 ++ echo '1.20 >= 1.21' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + kubectl_bin get -o yaml clusterissuer/special-selfsigned-issuer + yq d - '**.uid' + yq d - metadata.resourceVersion + yq d - '**.env.(name==CLUSTER_HASH)' + yq d - metadata.selfLink + yq d - metadata.deletionTimestamp + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - spec.nodeName + yq d - '**.volumeName' + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**.(volumeMode==Filesystem).volumeMode' + yq d - spec.volumeMode + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - '**.procMount' + yq d - '**.imagePullSecrets' + yq d - '**.storageClassName' + yq d - '**.enableServiceLinks' + yq d - '**."kubernetes.io/pvc-protection"' + yq d - status + yq d - '**."percona.com/*"' + yq d - '**.(name==NAMESPACE)' + yq d - '**.healthCheckNodePort' + yq d - '**.(name==percona-xtradb-cluster-operator-workload-token*)' + yq d - '**.finalizers' + yq d - '**.nodePort' + yq d - '**.creationTimestamp' + yq d - '**.image' + yq d - '**.(name==S3_BUCKET_PATH)' + yq d - '**.clusterIP' + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - '**.(name==suffix)' + yq d - '**.(name==S3_BUCKET_URL)' + yq d - '**.dataSource' + yq d - spec.ipFamilies + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - 'metadata.ownerReferences.*.apiVersion' + yq d - '**.controller-uid' + yq d - '**.clusterIPs' + yq d - '**.preemptionPolicy' + yq d - '**.namespace' + yq d - spec.ipFamilyPolicy + /usr/bin/sed 's#^apiVersion: policy/v1beta1#apiVersion: policy/v1#' + /usr/bin/sed 's/name: kube-api-access-.*$/name: kube-api-access/' + /usr/bin/sed 's/namespace\:.*name/name/' + /usr/bin/sed s/tls-issue-cert-manager-ref-21277/namespace/g + yq d - '**.creationTimestamp' + yq d - metadata.managedFields ++ mktemp + local LAST_OUT=/tmp/tmp.0wo3GZ5KAq ++ mktemp + local LAST_ERR=/tmp/tmp.9u0J7gFzkq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl get -o yaml clusterissuer/special-selfsigned-issuer + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.0wo3GZ5KAq + cat /tmp/tmp.9u0J7gFzkq + rm /tmp/tmp.0wo3GZ5KAq /tmp/tmp.9u0J7gFzkq + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/clusterissuer_special-selfsigned-issuer.yml /tmp/tmp.Dipb9FFYMI/clusterissuer_special-selfsigned-issuer.yml + desc 'check if issuer used during certificate creation' + set +o xtrace ----------------------------------------------------------------------------------- check if issuer used during certificate creation ----------------------------------------------------------------------------------- + compare_kubectl certificate/some-name-tls-issueref-ssl + local resource=certificate/some-name-tls-issueref-ssl + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/certificate_some-name-tls-issueref-ssl.yml + local new_result=/tmp/tmp.Dipb9FFYMI/certificate_some-name-tls-issueref-ssl.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/certificate_some-name-tls-issueref-ssl-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/certificate_some-name-tls-issueref-ssl-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/certificate_some-name-tls-issueref-ssl-80.yml ']' + version_gt 1.22 ++ echo '1.20 >= 1.22' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.21 ++ echo '1.20 >= 1.21' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + kubectl_bin get -o yaml certificate/some-name-tls-issueref-ssl ++ mktemp + local LAST_OUT=/tmp/tmp.YoAPENkUNZ + yq d - '**.namespace' + yq d - '**.uid' + yq d - metadata.resourceVersion + yq d - 'metadata.annotations."kubernetes.io/psp"' + yq d - 'metadata.annotations."cloud.google.com/neg"' + yq d - '**.(name==percona-xtradb-cluster-operator-workload-token*)' + yq d - '**.creationTimestamp' + yq d - '**.image' + yq d - '**.clusterIP' + yq d - '**.(volumeMode==Filesystem).volumeMode' ++ mktemp + yq d - '**.nodePort' + yq d - '**.healthCheckNodePort' + yq d - '**.finalizers' + yq d - '**."kubernetes.io/pvc-protection"' + yq d - '**.clusterIPs' + yq d - '**.dataSource' + yq d - '**.procMount' + yq d - '**.storageClassName' + yq d - '**.volumeName' + yq d - '**.imagePullSecrets' + yq d - '**.enableServiceLinks' + yq d - '**."volume.beta.kubernetes.io/storage-provisioner"' + yq d - '**."volume.kubernetes.io/storage-provisioner"' + yq d - 'metadata.annotations."k8s.v1.cni.cncf.io*"' + local LAST_ERR=/tmp/tmp.IODGrzZB9I + local exit_status=0 + yq d - spec.volumeMode + yq d - metadata.deletionTimestamp + yq d - 'spec.volumeClaimTemplates.*.apiVersion' + yq d - spec.nodeName + yq d - '**."volume.kubernetes.io/selected-node"' + yq d - '**."percona.com/*"' + yq d - status + yq d - metadata.selfLink + yq d - '**.(name==NAMESPACE)' + yq d - '**.(name==suffix)' + yq d - '**.(name==S3_BUCKET_PATH)' + yq d - '**.env.(name==CLUSTER_HASH)' + yq d - '**.(name==S3_BUCKET_URL)' + yq d - '**.preemptionPolicy' ++ seq 0 2 + /usr/bin/sed 's#^apiVersion: policy/v1beta1#apiVersion: policy/v1#' + yq d - spec.ipFamilyPolicy + yq d - 'spec.volumeClaimTemplates.*.kind' + yq d - '**.controller-uid' + yq d - 'metadata.ownerReferences.*.apiVersion' + /usr/bin/sed 's/name: kube-api-access-.*$/name: kube-api-access/' + yq d - spec.ipFamilies + yq d - '**.creationTimestamp' + /usr/bin/sed 's/namespace\:.*name/name/' + for i in '$(seq 0 2)' + kubectl get -o yaml certificate/some-name-tls-issueref-ssl + yq d - metadata.managedFields + /usr/bin/sed s/tls-issue-cert-manager-ref-21277/namespace/g + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.YoAPENkUNZ + cat /tmp/tmp.IODGrzZB9I + rm /tmp/tmp.YoAPENkUNZ /tmp/tmp.IODGrzZB9I + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1125/e2e-tests/tls-issue-cert-manager-ref/compare/certificate_some-name-tls-issueref-ssl.yml /tmp/tmp.Dipb9FFYMI/certificate_some-name-tls-issueref-ssl.yml + destroy tls-issue-cert-manager-ref-21277 + local namespace=tls-issue-cert-manager-ref-21277 + local ignore_logs=false + [[ false == \f\a\l\s\e ]] + grep -v 'get backup status: Job.batch' + grep -v 'the object has been modified' + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + sort -u + tee /tmp/tmp.Dipb9FFYMI/operator.log + grep -v level=info ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ grep -c percona-xtradb-cluster-operator +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.0FIz963KlX +++ mktemp ++ local LAST_ERR=/tmp/tmp.zlL3qkxyvn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ [[ 0 != 0 ]] ++ break ++ cat /tmp/tmp.0FIz963KlX ++ cat /tmp/tmp.zlL3qkxyvn ++ rm /tmp/tmp.0FIz963KlX /tmp/tmp.zlL3qkxyvn ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-5699d7755d-bc2ck ++ mktemp + local LAST_OUT=/tmp/tmp.TCckztBRzv ++ mktemp + local LAST_ERR=/tmp/tmp.PMYdBzGf5R + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-5699d7755d-bc2ck + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.TCckztBRzv + cat /tmp/tmp.PMYdBzGf5R + rm /tmp/tmp.TCckztBRzv /tmp/tmp.PMYdBzGf5R + return 0 I0412 13:38:49.464100 1 request.go:645] Throttling request took 1.026345859s, request: GET:https://10.7.240.1:443/apis/apiextensions.k8s.io/v1beta1?timeout=32s I0412 13:40:48.840150 1 request.go:645] Throttling request took 1.04518626s, request: GET:https://10.7.240.1:443/apis/storage.k8s.io/v1?timeout=32s {"level":"error",,"caller":"pxc/controller.go:1142","msg":"sync users","error":"exec syncusers: command terminated with exit code 1 / / ERROR (line:1812) : Could not find any nodes belonging to the cluster with writer hostgroup:11\nERROR (line:1921) : Could not find a primary cluster node\n","errorVerbose":"exec syncusers: command terminated with exit code 1 / / ERROR (line:1812) : Could not find any nodes belonging to the cluster with writer hostgroup:11\nERROR (line:1921) : Could not find a primary cluster node\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:491\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1140\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1581","stacktrace":"github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1142"} {"level":"info",,"caller":"pxc/controller.go:468","msg":"reconcile replication error","err":"get primary pxc pod: not found"} {"level":"info",,"caller":"pxc/version.go:328","msg":"update PXC version (fetched from db)","new version":"8.0.27-18.1"} {"level":"info",,"caller":"v1/pxc_types.go:874","msg":"ProxySQL size will be changed from 1 to 2 due to safe config"} {"level":"info",,"caller":"v1/pxc_types.go:875","msg":"Set allowUnsafeConfigurations=true to disable safe configuration"} {"level":"info",,"logger":"cmd","msg":"Git commit: 706f792ae47c369cb3556faff186b6873a8a247f Git branch: PR-1125-706f792a Build time: 2022-04-12T09:09:41Z"} {"level":"info",,"logger":"cmd","msg":"Go OS/Arch: linux/amd64"} {"level":"info",,"logger":"cmd","msg":"Go Version: go1.17.8"} {"level":"info",,"logger":"cmd","msg":"operator-sdk Version: v0.19.4"} {"level":"info",,"logger":"cmd","msg":"Registering Components."} {"level":"info",,"logger":"cmd","msg":"Runs on","platform":"kubernetes","version":"v1.20.15-gke.4100"} {"level":"info",,"logger":"cmd","msg":"Starting the Cmd."} {"level":"info",,"logger":"controller-runtime.certwatcher","msg":"Starting certificate watcher"} {"level":"info",,"logger":"controller-runtime.certwatcher","msg":"Updated current TLS certificate"} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterbackup-controller","msg":"Starting Controller"} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterbackup-controller","msg":"Starting EventSource","source":"kind source: /, Kind="} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterbackup-controller","msg":"Starting workers","worker count":1} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbcluster-controller","msg":"Starting Controller"} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbcluster-controller","msg":"Starting EventSource","source":"kind source: /, Kind="} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbcluster-controller","msg":"Starting workers","worker count":1} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterrestore-controller","msg":"Starting Controller"} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterrestore-controller","msg":"Starting EventSource","source":"kind source: /, Kind="} {"level":"info",,"logger":"controller-runtime.manager.controller.perconaxtradbclusterrestore-controller","msg":"Starting workers","worker count":1} {"level":"info",,"logger":"controller-runtime.manager","msg":"starting metrics server","path":"/metrics"} {"level":"info",,"logger":"controller-runtime.metrics","msg":"metrics server is starting to listen","addr":":8080"} {"level":"info",,"logger":"controller-runtime.webhook","msg":"registering webhook","path":"/validate-percona-xtradbcluster"} {"level":"info",,"logger":"controller-runtime.webhook","msg":"serving webhook server","host":"","port":9443} {"level":"info",,"logger":"controller-runtime.webhook.webhooks","msg":"starting webhook server"} {"level":"info",,"logger":"leader","msg":"Became the leader."} {"level":"info",,"logger":"leader","msg":"No pre-existing lock was found."} {"level":"info",,"logger":"leader","msg":"Trying to become the leader."} + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl patch pxc -n tls-issue-cert-manager-ref-21277 some-name-tls-issueref --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name-tls-issueref patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.oJsxStDq0n ++ mktemp + local LAST_ERR=/tmp/tmp.IiRro3cVAS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.oJsxStDq0n perconaxtradbcluster.pxc.percona.com "some-name-tls-issueref" deleted + cat /tmp/tmp.IiRro3cVAS + rm /tmp/tmp.oJsxStDq0n /tmp/tmp.IiRro3cVAS + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.NinZaebybP ++ mktemp + local LAST_ERR=/tmp/tmp.yTb62IPFQx + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.NinZaebybP No resources found + cat /tmp/tmp.yTb62IPFQx + rm /tmp/tmp.NinZaebybP /tmp/tmp.yTb62IPFQx + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.eTNBi70n6W ++ mktemp + local LAST_ERR=/tmp/tmp.FFDMlBxbHU + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.eTNBi70n6W No resources found + cat /tmp/tmp.FFDMlBxbHU + rm /tmp/tmp.eTNBi70n6W /tmp/tmp.FFDMlBxbHU + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.2m1fFj35Jt ++ mktemp + local LAST_ERR=/tmp/tmp.eKincNRGDR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + [[ 0 != 0 ]] + break + cat /tmp/tmp.2m1fFj35Jt validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.eKincNRGDR + rm /tmp/tmp.2m1fFj35Jt /tmp/tmp.eKincNRGDR + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.5.4/cert-manager.yaml namespace "cert-manager" deleted + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace tls-issue-cert-manager-ref-21277 + rm -rf /tmp/tmp.Dipb9FFYMI + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.SCdI3kVNfb ++ mktemp + local LAST_OUT=/tmp/tmp.PrclKSPvGg ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.xrdgQYoCXF + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.eqe4UAQcKv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + kubectl delete --grace-period=0 --force=true namespace pxc-operator + for i in '$(seq 0 2)' + kubectl delete --grace-period=0 --force=true namespace tls-issue-cert-manager-ref-21277