Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/logs/tls-issue-self-8-0.log Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.36) and server (1.33) exceeds the supported minor version skew of +/-1 + main + create_infra tls-issue-self-3272 + local ns=tls-issue-self-3272 + '[' -n pxc-operator ']' + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl get pxc --all-namespaces -o wide + kubectl patch pxc -n tls-issue-self-4433 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.NCgQodm5Lk ++ mktemp + local LAST_ERR=/tmp/tmp.xaK2rezw08 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NCgQodm5Lk perconaxtradbcluster.pxc.percona.com "some-name" deleted from tls-issue-self-4433 namespace + cat /tmp/tmp.xaK2rezw08 + rm /tmp/tmp.NCgQodm5Lk /tmp/tmp.xaK2rezw08 + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.KaHz2zgWwg ++ mktemp + local LAST_ERR=/tmp/tmp.5kPj3xlND5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KaHz2zgWwg No resources found + cat /tmp/tmp.5kPj3xlND5 + rm /tmp/tmp.KaHz2zgWwg /tmp/tmp.5kPj3xlND5 + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.XhjSXOZJfL ++ mktemp + local LAST_ERR=/tmp/tmp.fAeItgTlyc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.XhjSXOZJfL No resources found + cat /tmp/tmp.fAeItgTlyc + rm /tmp/tmp.XhjSXOZJfL /tmp/tmp.fAeItgTlyc + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// ++ tail -n1 + local chaos_mesh_ns= + '[' -n '' ']' ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl api-resources ++ awk '{print $1}' ++ awk '{print $1}' ++ grep chaos-mesh.org ++ kubectl get crd + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get clusterrole + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator+ kubectl_bin get ns ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + xargs kubectl delete ns + awk '{print$1}' ++ mktemp + local LAST_OUT=/tmp/tmp.4PkMpHbytC ++ mktemp + local LAST_OUT=/tmp/tmp.qgZYBmMbNR ++ mktemp + local LAST_ERR=/tmp/tmp.qNIUSOpci9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator ++ mktemp + local LAST_ERR=/tmp/tmp.gzEzOXuoSH + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qgZYBmMbNR + cat /tmp/tmp.gzEzOXuoSH + rm /tmp/tmp.qgZYBmMbNR /tmp/tmp.gzEzOXuoSH + return 0 namespace "tls-issue-self-4433" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4PkMpHbytC namespace "pxc-operator" deleted + cat /tmp/tmp.qNIUSOpci9 + rm /tmp/tmp.4PkMpHbytC /tmp/tmp.qNIUSOpci9 + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.DJYdd2QodS ++ mktemp + local LAST_ERR=/tmp/tmp.59VTh7dWlD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.DJYdd2QodS namespace/pxc-operator created + cat /tmp/tmp.59VTh7dWlD + rm /tmp/tmp.DJYdd2QodS /tmp/tmp.59VTh7dWlD + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.gVoKHNcEPH +++ mktemp ++ local LAST_ERR=/tmp/tmp.LPOgHQfFQy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gVoKHNcEPH ++ cat /tmp/tmp.LPOgHQfFQy ++ rm /tmp/tmp.gVoKHNcEPH /tmp/tmp.LPOgHQfFQy ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster12 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.1PogZeUErl ++ mktemp + local LAST_ERR=/tmp/tmp.20T5r4mbV5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster12 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.1PogZeUErl Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster12" modified. + cat /tmp/tmp.20T5r4mbV5 + rm /tmp/tmp.1PogZeUErl /tmp/tmp.20T5r4mbV5 + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.ponElErFTa ++ mktemp + local LAST_ERR=/tmp/tmp.NTw7ftMG4O + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ponElErFTa customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.NTw7ftMG4O + rm /tmp/tmp.ponElErFTa /tmp/tmp.NTw7ftMG4O + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.jWQQl4LSg8 + sed -e 's^namespace: .*^namespace: pxc-operator^' ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/deploy/cw-rbac.yaml + local LAST_ERR=/tmp/tmp.plXEN5G0Pp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jWQQl4LSg8 clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.plXEN5G0Pp + rm /tmp/tmp.jWQQl4LSg8 /tmp/tmp.plXEN5G0Pp + return 0 + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "PXCO_FEATURE_GATES").value) = ""' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + kubectl_bin apply -f - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2467-3dc7f023^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/deploy/cw-operator.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.fa7XxtWcpO ++ mktemp + local LAST_ERR=/tmp/tmp.POqR9jnbBW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.fa7XxtWcpO deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.POqR9jnbBW + rm /tmp/tmp.fa7XxtWcpO /tmp/tmp.POqR9jnbBW + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.rVb2FdTNlf ++ mktemp + local LAST_ERR=/tmp/tmp.LMTSoXoVeI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rVb2FdTNlf pod/percona-xtradb-cluster-operator-9d9fbdb5-4hk6m condition met + cat /tmp/tmp.LMTSoXoVeI E0516 21:55:45.128871 12821 reflector.go:227] "Failed to watch" err="Get \"https://34.121.115.52/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-9d9fbdb5-4hk6m&resourceVersion=1778968544767736000&timeoutSeconds=371&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" + rm /tmp/tmp.rVb2FdTNlf /tmp/tmp.LMTSoXoVeI + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ head -1 ++ jq -r '.items[] | select(.metadata.deletionTimestamp == null) | .metadata.name' ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.kpKpvLA24o +++ mktemp ++ local LAST_ERR=/tmp/tmp.8GsSlxAYnS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator --field-selector=status.phase=Running -o json -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kpKpvLA24o ++ cat /tmp/tmp.8GsSlxAYnS ++ rm /tmp/tmp.kpKpvLA24o /tmp/tmp.8GsSlxAYnS ++ return 0 + wait_pod percona-xtradb-cluster-operator-9d9fbdb5-4hk6m 480 pxc-operator + local pod=percona-xtradb-cluster-operator-9d9fbdb5-4hk6m + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-9d9fbdb5-4hk6m ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-9d9fbdb5-4hk6m condition met E0516 21:55:48.478690 13361 reflector.go:227] "Failed to watch" err="Get \"https://34.121.115.52/api/v1/namespaces/pxc-operator/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpercona-xtradb-cluster-operator-9d9fbdb5-4hk6m&resourceVersion=1778968547683607000&timeoutSeconds=519&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/percona-xtradb-cluster-operator-9d9fbdb5-4hk6m to become Ready.Ok + sleep 3 + create_namespace tls-issue-self-3272 + local namespace=tls-issue-self-3272 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ sed s/NAMESPACE// ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ grep chaos-mesh ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces tls-issue-self-3272' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces tls-issue-self-3272 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace tls-issue-self-3272 + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + xargs kubectl delete ns + kubectl_bin get ns ++ mktemp + awk '{print$1}' + local LAST_OUT=/tmp/tmp.ZtRSGBulMg ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.4lqNtpxJRZ + local LAST_ERR=/tmp/tmp.NukvDVbtGZ + local exit_status=0 ++ mktemp ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace tls-issue-self-3272 + local LAST_ERR=/tmp/tmp.TJU3vKtcp0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4lqNtpxJRZ + cat /tmp/tmp.TJU3vKtcp0 + rm /tmp/tmp.4lqNtpxJRZ /tmp/tmp.TJU3vKtcp0 + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace tls-issue-self-3272 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace tls-issue-self-3272 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.ZtRSGBulMg + cat /tmp/tmp.NukvDVbtGZ Error from server (NotFound): namespaces "tls-issue-self-3272" not found + rm /tmp/tmp.ZtRSGBulMg /tmp/tmp.NukvDVbtGZ + return 1 + : + wait_for_delete namespace/tls-issue-self-3272 + local res=namespace/tls-issue-self-3272 + echo -n 'waiting for namespace/tls-issue-self-3272 to be deleted' waiting for namespace/tls-issue-self-3272 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "tls-issue-self-3272" not found + desc 'create namespace tls-issue-self-3272' + set +o xtrace ----------------------------------------------------------------------------------- create namespace tls-issue-self-3272 ----------------------------------------------------------------------------------- + kubectl_bin create namespace tls-issue-self-3272 ++ mktemp + local LAST_OUT=/tmp/tmp.FsB2C2d7zk ++ mktemp + local LAST_ERR=/tmp/tmp.Es5ie4taYW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace tls-issue-self-3272 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FsB2C2d7zk namespace/tls-issue-self-3272 created + cat /tmp/tmp.Es5ie4taYW + rm /tmp/tmp.FsB2C2d7zk /tmp/tmp.Es5ie4taYW + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.0KSRtt1F3V +++ mktemp ++ local LAST_ERR=/tmp/tmp.prR8nd7SmA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0KSRtt1F3V ++ cat /tmp/tmp.prR8nd7SmA ++ rm /tmp/tmp.0KSRtt1F3V /tmp/tmp.prR8nd7SmA ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster12 --namespace=tls-issue-self-3272 ++ mktemp + local LAST_OUT=/tmp/tmp.QYUoNZMBqT ++ mktemp + local LAST_ERR=/tmp/tmp.xFQSrvUtgA + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster12 --namespace=tls-issue-self-3272 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QYUoNZMBqT Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2467-3dc7f023-16-cluster12" modified. + cat /tmp/tmp.xFQSrvUtgA + rm /tmp/tmp.QYUoNZMBqT /tmp/tmp.xFQSrvUtgA + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.9kTR5KXxps ++ mktemp + local LAST_ERR=/tmp/tmp.qG2QfD3ENG + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9kTR5KXxps secret/minio-secret created secret/aws-s3-secret created secret/do-spaces-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.qG2QfD3ENG + rm /tmp/tmp.9kTR5KXxps /tmp/tmp.qG2QfD3ENG + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.19.2/cert-manager.yaml + : + sleep 10 + desc 'create PXC cluster without cert manager' + set +o xtrace ----------------------------------------------------------------------------------- create PXC cluster without cert manager ----------------------------------------------------------------------------------- + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/some-name.yml 3 10 /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/secrets_without_tls.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/secrets_without_tls.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/secrets_without_tls.yml ++ mktemp + local LAST_OUT=/tmp/tmp.68jmmGagf2 ++ mktemp + local LAST_ERR=/tmp/tmp.524yocD54B + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/secrets_without_tls.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.68jmmGagf2 secret/my-cluster-secrets created + cat /tmp/tmp.524yocD54B + rm /tmp/tmp.68jmmGagf2 /tmp/tmp.524yocD54B + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/client.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/client.yml + local pvc_name= + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2467-3dc7f023#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.tls-issue-self-3272~ + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' ++ mktemp + local LAST_OUT=/tmp/tmp.diqUlEVvZ1 ++ mktemp + local LAST_ERR=/tmp/tmp.64ORGAkk4g + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.diqUlEVvZ1 deployment.apps/pxc-client created + cat /tmp/tmp.64ORGAkk4g + rm /tmp/tmp.diqUlEVvZ1 /tmp/tmp.64ORGAkk4g + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/some-name.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/some-name.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/some-name.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/some-name.yml + local pvc_name= + kubectl_bin apply -f - ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/conf/some-name.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + local LAST_OUT=/tmp/tmp.YZqt8rc87A + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2467-3dc7f023#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' ++ mktemp + /usr/bin/sed -e s~minio-service.#namespace~minio-service.tls-issue-self-3272~ + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + local LAST_ERR=/tmp/tmp.QaabqOB5e3 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YZqt8rc87A perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.QaabqOB5e3 + rm /tmp/tmp.YZqt8rc87A /tmp/tmp.QaabqOB5e3 + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.RuXtbaDko9 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.5j4OQ4Mc6r +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.RuXtbaDko9 +++ cat /tmp/tmp.5j4OQ4Mc6r +++ rm /tmp/tmp.RuXtbaDko9 /tmp/tmp.5j4OQ4Mc6r +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.V8hcWNGman ++++ mktemp +++ local LAST_ERR=/tmp/tmp.bpt0px46oH +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.V8hcWNGman +++ cat /tmp/tmp.bpt0px46oH +++ rm /tmp/tmp.V8hcWNGman /tmp/tmp.bpt0px46oH +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n tls-issue-self-3272 ++ mktemp + local LAST_OUT=/tmp/tmp.vSJmXScFKH ++ mktemp + local LAST_ERR=/tmp/tmp.efIkHl1nRR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n tls-issue-self-3272 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n tls-issue-self-3272 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n tls-issue-self-3272 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.vSJmXScFKH + cat /tmp/tmp.efIkHl1nRR error: no matching resources found + rm /tmp/tmp.vSJmXScFKH /tmp/tmp.efIkHl1nRR + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ax8eOpe5PC +++ mktemp ++ local LAST_ERR=/tmp/tmp.jg424nLtsX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ax8eOpe5PC ++ cat /tmp/tmp.jg424nLtsX ++ rm /tmp/tmp.Ax8eOpe5PC /tmp/tmp.jg424nLtsX ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TMy6L2xICe +++ mktemp ++ local LAST_ERR=/tmp/tmp.52Cyg1P7xc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TMy6L2xICe ++ cat /tmp/tmp.52Cyg1P7xc ++ rm /tmp/tmp.TMy6L2xICe /tmp/tmp.52Cyg1P7xc ++ return 0 + client_pod=pxc-client-67fc4995bb-mfp9p + wait_pod pxc-client-67fc4995bb-mfp9p + local pod=pxc-client-67fc4995bb-mfp9p + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-mfp9p ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-mfp9p condition met E0516 22:01:12.734709 18101 reflector.go:227] "Failed to watch" err="Get \"https://34.121.115.52/api/v1/namespaces/tls-issue-self-3272/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-mfp9p&resourceVersion=1778968871264642000&timeoutSeconds=510&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-mfp9p to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sK5leYr6bQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.heCtKs8IWb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sK5leYr6bQ ++ cat /tmp/tmp.heCtKs8IWb ++ rm /tmp/tmp.sK5leYr6bQ /tmp/tmp.heCtKs8IWb ++ return 0 + client_pod=pxc-client-67fc4995bb-mfp9p + wait_pod pxc-client-67fc4995bb-mfp9p + local pod=pxc-client-67fc4995bb-mfp9p + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-mfp9p ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-mfp9p condition met E0516 22:01:20.194099 18932 reflector.go:227] "Failed to watch" err="Get \"https://34.121.115.52/api/v1/namespaces/tls-issue-self-3272/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-mfp9p&resourceVersion=1778968878518224000&timeoutSeconds=532&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-mfp9p to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.spuuDhDDgT +++ mktemp ++ local LAST_ERR=/tmp/tmp.25SCaCufRQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.spuuDhDDgT ++ cat /tmp/tmp.25SCaCufRQ ++ rm /tmp/tmp.spuuDhDDgT /tmp/tmp.25SCaCufRQ ++ return 0 + client_pod=pxc-client-67fc4995bb-mfp9p + wait_pod pxc-client-67fc4995bb-mfp9p + local pod=pxc-client-67fc4995bb-mfp9p + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-67fc4995bb-mfp9p + local container= + set +o xtrace pod/pxc-client-67fc4995bb-mfp9p condition met E0516 22:01:57.260440 22936 reflector.go:227] "Failed to watch" err="Get \"https://34.121.115.52/api/v1/namespaces/tls-issue-self-3272/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-mfp9p&resourceVersion=1778968916489534000&timeoutSeconds=356&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-mfp9p to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.nt78ipZG5a/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.nt78ipZG5a/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/select-1.sql /tmp/tmp.nt78ipZG5a/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sANMxmAjIb +++ mktemp ++ local LAST_ERR=/tmp/tmp.4YBwLPhKcr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sANMxmAjIb ++ cat /tmp/tmp.4YBwLPhKcr ++ rm /tmp/tmp.sANMxmAjIb /tmp/tmp.4YBwLPhKcr ++ return 0 + client_pod=pxc-client-67fc4995bb-mfp9p + wait_pod pxc-client-67fc4995bb-mfp9p + local pod=pxc-client-67fc4995bb-mfp9p + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-mfp9p ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-mfp9p condition met E0516 22:02:03.253911 23489 reflector.go:227] "Failed to watch" err="Get \"https://34.121.115.52/api/v1/namespaces/tls-issue-self-3272/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-mfp9p&resourceVersion=1778968922505522000&timeoutSeconds=577&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-mfp9p to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.nt78ipZG5a/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.nt78ipZG5a/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/select-1.sql /tmp/tmp.nt78ipZG5a/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DG3ENppOg5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Hxqw7YviY8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DG3ENppOg5 ++ cat /tmp/tmp.Hxqw7YviY8 ++ rm /tmp/tmp.DG3ENppOg5 /tmp/tmp.Hxqw7YviY8 ++ return 0 + client_pod=pxc-client-67fc4995bb-mfp9p + wait_pod pxc-client-67fc4995bb-mfp9p + local pod=pxc-client-67fc4995bb-mfp9p + local max_retry=480 + local ns= ++ echo pxc-client-67fc4995bb-mfp9p ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-67fc4995bb-mfp9p condition met E0516 22:02:08.646836 23935 reflector.go:227] "Failed to watch" err="Get \"https://34.121.115.52/api/v1/namespaces/tls-issue-self-3272/pods?allowWatchBookmarks=true&fieldSelector=metadata.name%3Dpxc-client-67fc4995bb-mfp9p&resourceVersion=1778968926996467000&timeoutSeconds=543&watch=true\": context canceled" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*unstructured.Unstructured" waiting for pod/pxc-client-67fc4995bb-mfp9p to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ ! -s /tmp/tmp.nt78ipZG5a/select-1.sql ]] ++ grep 'Unknown MySQL server host' /tmp/tmp.nt78ipZG5a/select-1.sql + [[ -n '' ]] + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/select-1.sql /tmp/tmp.nt78ipZG5a/select-1.sql + is_keyring_plugin_in_use some-name + local cluster=some-name + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + kubectl exec some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + grep -E -o 'early-plugin-load=keyring_\w+.so' + return 1 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.mHmUE1cIxD +++ mktemp ++ local LAST_ERR=/tmp/tmp.hQBBtuY0S7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.mHmUE1cIxD ++ cat /tmp/tmp.hQBBtuY0S7 ++ rm /tmp/tmp.mHmUE1cIxD /tmp/tmp.hQBBtuY0S7 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.p5FeExKMtJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.TaDR7v4PQo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.p5FeExKMtJ ++ cat /tmp/tmp.TaDR7v4PQo ++ rm /tmp/tmp.p5FeExKMtJ /tmp/tmp.TaDR7v4PQo ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.muOOH4fbFO ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.8niyLeLv8j +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.muOOH4fbFO +++++ cat /tmp/tmp.8niyLeLv8j +++++ rm /tmp/tmp.muOOH4fbFO /tmp/tmp.8niyLeLv8j +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.gqf7O5cRIR ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.FNLxdaXmyU +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.gqf7O5cRIR +++++ cat /tmp/tmp.FNLxdaXmyU +++++ rm /tmp/tmp.gqf7O5cRIR /tmp/tmp.FNLxdaXmyU +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1wasPMD7ik +++ mktemp ++ local LAST_ERR=/tmp/tmp.HHyGuDHePq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1wasPMD7ik ++ cat /tmp/tmp.HHyGuDHePq ++ rm /tmp/tmp.1wasPMD7ik /tmp/tmp.HHyGuDHePq ++ return 0 + [[ 2 == \2 ]] + echo + desc 'check if certificate issued manually' + set +o xtrace ----------------------------------------------------------------------------------- check if certificate issued manually ----------------------------------------------------------------------------------- + tlsSecretsShouldExist some-name-ssl + local secretName=some-name-ssl + checkTLSSecret some-name-ssl ca.crt + local secretName=some-name-ssl + local dataKey=ca.crt ++ kubectl_bin get secrets/some-name-ssl -o json +++ mktemp ++ jq '.data["ca.crt"]' ++ local LAST_OUT=/tmp/tmp.1gATLEpS7f +++ mktemp ++ local LAST_ERR=/tmp/tmp.kiHuGQx82M ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/some-name-ssl -o json ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1gATLEpS7f ++ cat /tmp/tmp.kiHuGQx82M ++ rm /tmp/tmp.1gATLEpS7f /tmp/tmp.kiHuGQx82M ++ return 0 + local 'secretData="LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURFRENDQWZpZ0F3SUJBZ0lSQUlMeXZvb3pXWm1jYzRBdFk4S0UzV0V3RFFZSktvWklodmNOQVFFTEJRQXcKRWpFUU1BNEdBMVVFQ2hNSFVtOXZkQ0JEUVRBZUZ3MHlOakExTVRZeU1UVTJOVFZhRncweU9UQTFNVFV5TVRVMgpOVFZhTUJJeEVEQU9CZ05WQkFvVEIxSnZiM1FnUTBFd2dnRWlNQTBHQ1NxR1NJYjNEUUVCQVFVQUE0SUJEd0F3CmdnRUtBb0lCQVFDenk2Zk4wMDdPd3FPUXpsbCtwYmVHdm5jR1h1UEJzcXNXTHZTZG5xdCtMQlBVOUlieWNxWjEKUEpnVWJVMHBTQlNRZXFZQXBSVm5TSjNWQjFsVk9oaUZaOTNmajNsWmR0WVR1dVFKc3hyV2xUR1RkU3ZPbzc4Uwp4QXp6Z3VnSDl6ZXIybnBIY0M0YVpDQW5NU1BxNjFqSFRBWW1CLzlST3lLNFFYZm9PTTBsZ05tOHBydDRyYVhECkpocUs3M0E2a0xkYzBOQ0N1OEF3L1lsMWlhNGQreG5PM0hwQ1hrS2RIaURwd2xtdElYVndYbEs0dWF2RVlmMkEKUnRvZXQwL2JwY0d2d1lObE92RG53cHF0MG0wd2RSMm9PZ1VrZ1NBbmw2OGsyMlQrTktSd0JXQUJTYUltT01seApCVXUwbklZaGZscC9TZ1h0dVRDNEc5TldRR29DSVBqdEFnTUJBQUdqWVRCZk1BNEdBMVVkRHdFQi93UUVBd0lDCkJEQWRCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3RHdZRFZSMFRBUUgvQkFVd0F3RUIKL3pBZEJnTlZIUTRFRmdRVUtrTVBaK1laSTdRcTRwUU8veDhySFV0Sm5sb3dEUVlKS29aSWh2Y05BUUVMQlFBRApnZ0VCQUE3NTNxVnpIOWZrM0pObFZGQ2RoNlp5RFRuYThTR29HdWdCeDlqbDFKOXEvVmJmWjJjb3NhQ0UwMUp4Cm5nejZTSUtDbmpFbzN6WnU0dHBjdGo5MkFzNm90dTFmS1pkYVpNQktPT29wN0U1N1FMT2lDUk8xZWlkc1Q2NnUKN3pzRENLZXVITE1yc3FwcmZNSCtBcVRDNnlFRDZFTXdHT2JUUlVWRGQ0cHg0Q3lDTmtXMHRtU3ovaExzUGxsZQowaHk3WlN4dU9Vakl4YlBwMlowdG9yVEN6YkZ5TmEvalQyOFJKakZpMHRwQlB0bzNnazZONWh4Q1hlbkE3aE9YClZEbExKMXB1QWZwbGhZK1E1UHZoVEsxcWd5WHY2SUQreGgxOHpjYjR1UncrUDFCQnJNS0ZCVUM3WEg2c211NEQKMmtNS2YzUjRqZjdnMXVoeWVaMEJ0Vll5TGRnPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg=="' + '[' -z '"LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURFRENDQWZpZ0F3SUJBZ0lSQUlMeXZvb3pXWm1jYzRBdFk4S0UzV0V3RFFZSktvWklodmNOQVFFTEJRQXcKRWpFUU1BNEdBMVVFQ2hNSFVtOXZkQ0JEUVRBZUZ3MHlOakExTVRZeU1UVTJOVFZhRncweU9UQTFNVFV5TVRVMgpOVFZhTUJJeEVEQU9CZ05WQkFvVEIxSnZiM1FnUTBFd2dnRWlNQTBHQ1NxR1NJYjNEUUVCQVFVQUE0SUJEd0F3CmdnRUtBb0lCQVFDenk2Zk4wMDdPd3FPUXpsbCtwYmVHdm5jR1h1UEJzcXNXTHZTZG5xdCtMQlBVOUlieWNxWjEKUEpnVWJVMHBTQlNRZXFZQXBSVm5TSjNWQjFsVk9oaUZaOTNmajNsWmR0WVR1dVFKc3hyV2xUR1RkU3ZPbzc4Uwp4QXp6Z3VnSDl6ZXIybnBIY0M0YVpDQW5NU1BxNjFqSFRBWW1CLzlST3lLNFFYZm9PTTBsZ05tOHBydDRyYVhECkpocUs3M0E2a0xkYzBOQ0N1OEF3L1lsMWlhNGQreG5PM0hwQ1hrS2RIaURwd2xtdElYVndYbEs0dWF2RVlmMkEKUnRvZXQwL2JwY0d2d1lObE92RG53cHF0MG0wd2RSMm9PZ1VrZ1NBbmw2OGsyMlQrTktSd0JXQUJTYUltT01seApCVXUwbklZaGZscC9TZ1h0dVRDNEc5TldRR29DSVBqdEFnTUJBQUdqWVRCZk1BNEdBMVVkRHdFQi93UUVBd0lDCkJEQWRCZ05WSFNVRUZqQVVCZ2dyQmdFRkJRY0RBUVlJS3dZQkJRVUhBd0l3RHdZRFZSMFRBUUgvQkFVd0F3RUIKL3pBZEJnTlZIUTRFRmdRVUtrTVBaK1laSTdRcTRwUU8veDhySFV0Sm5sb3dEUVlKS29aSWh2Y05BUUVMQlFBRApnZ0VCQUE3NTNxVnpIOWZrM0pObFZGQ2RoNlp5RFRuYThTR29HdWdCeDlqbDFKOXEvVmJmWjJjb3NhQ0UwMUp4Cm5nejZTSUtDbmpFbzN6WnU0dHBjdGo5MkFzNm90dTFmS1pkYVpNQktPT29wN0U1N1FMT2lDUk8xZWlkc1Q2NnUKN3pzRENLZXVITE1yc3FwcmZNSCtBcVRDNnlFRDZFTXdHT2JUUlVWRGQ0cHg0Q3lDTmtXMHRtU3ovaExzUGxsZQowaHk3WlN4dU9Vakl4YlBwMlowdG9yVEN6YkZ5TmEvalQyOFJKakZpMHRwQlB0bzNnazZONWh4Q1hlbkE3aE9YClZEbExKMXB1QWZwbGhZK1E1UHZoVEsxcWd5WHY2SUQreGgxOHpjYjR1UncrUDFCQnJNS0ZCVUM3WEg2c211NEQKMmtNS2YzUjRqZjdnMXVoeWVaMEJ0Vll5TGRnPQotLS0tLUVORCBDRVJUSUZJQ0FURS0tLS0tCg=="' ']' + checkTLSSecret some-name-ssl tls.crt + local secretName=some-name-ssl + local dataKey=tls.crt ++ kubectl_bin get secrets/some-name-ssl -o json ++ jq '.data["tls.crt"]' +++ mktemp ++ local LAST_OUT=/tmp/tmp.x5OVjIumCR +++ mktemp ++ local LAST_ERR=/tmp/tmp.oIOjbshquw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/some-name-ssl -o json ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.x5OVjIumCR ++ cat /tmp/tmp.oIOjbshquw ++ rm /tmp/tmp.x5OVjIumCR /tmp/tmp.oIOjbshquw ++ return 0 + local 'secretData="LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURRRENDQWlpZ0F3SUJBZ0lRR3VvVnVwcGJhT2h4Y0JUK0VEcWpsekFOQmdrcWhraUc5dzBCQVFzRkFEQVMKTVJBd0RnWURWUVFLRXdkU2IyOTBJRU5CTUI0WERUSTJNRFV4TmpJeE5UWTFOVm9YRFRJMk1EZ3hOREl4TlRZMQpOVm93RGpFTU1Bb0dBMVVFQ2hNRFVGaERNSUlCSWpBTkJna3Foa2lHOXcwQkFRRUZBQU9DQVE4QU1JSUJDZ0tDCkFRRUF2RmtvbFkxQ1hGaWdxZVpMQnFWN3JIOFBYSXRodzFvUG1lZXdBYkJGYU1ZVUkxQlRFQkI2clQxM1l6ZDQKZG1aRFltaXczN0RxRVYwanZVaUNNdHRONDhKUE1pazRsOVVZT0pUVWcvMWZtNmtkTE5hams3NVZ0K2hjR0RGbAoxN3B2LytPY3JVWTFpV0I0ZGdFb29rT3BUZlltMGQ2Q0J0dzdoRXg4OFE1YU9kTllOVk5kT1hPMk95MHZ6M0l4CmNGTnIxUEI3T3E2bnNyV3cyYkx6Qm94bU1zb05CekVkaUVEYVhYZVFmam5QdFdwclN4RGFlNEEwNDU4dERoMXEKQXZ3QlJ3cWFzRU5uNmxaZ2ZYOWFCcEZiUUhzL3YreTFzN0c1WG9lSTNjU3lZOFozR2V3cGdoaEVnakViQkNCKwpoRUhBL2tydlhCWnl2UHFCdDB1Rmswb1o4UUlEQVFBQm80R1ZNSUdTTUE0R0ExVWREd0VCL3dRRUF3SUZvREFkCkJnTlZIU1VFRmpBVUJnZ3JCZ0VGQlFjREFRWUlLd1lCQlFVSEF3SXdEQVlEVlIwVEFRSC9CQUl3QURCVEJnTlYKSFJFRVREQktnZzF6YjIxbExXNWhiV1V0Y0hoamdoSnpiMjFsTFc1aGJXVXRjSEp2ZUhsemNXeUNEeW91YzI5dApaUzF1WVcxbExYQjRZNElVS2k1emIyMWxMVzVoYldVdGNISnZlSGx6Y1d3d0RRWUpLb1pJaHZjTkFRRUxCUUFECmdnRUJBSFo1ekhGRjZ6bVJaYnVLVW5yWDM4RUU0UndqVExUaGhzSXhjb0F0MWY3b05KMW13RVdxdzZ1UHE3dE0KQlRwd2F5VytwdUUxS0JmZWJXTUhlNmdHa3NOdE0vTDI0bGM1S1NybFZCUDQvMytMdWViTVNHNVlvdDNONTU1QgpabHlwaFM1bWJJbHJaZUFUUVhXUUpkYW12dk1aOTM4SEk2WXJjOW9UZlg3blpsQUhoVUdkWUhwdjFGU0paZFlhCndSWCtjME0walZPSTk2dHRjaWtFSmhUK0F5b3NRV2VDYXljUG9qNHhBanVtZzJHM3VaT1dNdjR3RllkK09TSnIKc085U3VrVDN6SDg4dXdhbjhsTlBHcVhVc2dFS1R6U2c2Q0ZjSzBldUNCSjAvMG52c01JNS9RemRXbmI5cmZiUQpXdEJMMFJveEt0Um9sNnVHM0JMMi8rU3krLzg9Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K"' + '[' -z '"LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURRRENDQWlpZ0F3SUJBZ0lRR3VvVnVwcGJhT2h4Y0JUK0VEcWpsekFOQmdrcWhraUc5dzBCQVFzRkFEQVMKTVJBd0RnWURWUVFLRXdkU2IyOTBJRU5CTUI0WERUSTJNRFV4TmpJeE5UWTFOVm9YRFRJMk1EZ3hOREl4TlRZMQpOVm93RGpFTU1Bb0dBMVVFQ2hNRFVGaERNSUlCSWpBTkJna3Foa2lHOXcwQkFRRUZBQU9DQVE4QU1JSUJDZ0tDCkFRRUF2RmtvbFkxQ1hGaWdxZVpMQnFWN3JIOFBYSXRodzFvUG1lZXdBYkJGYU1ZVUkxQlRFQkI2clQxM1l6ZDQKZG1aRFltaXczN0RxRVYwanZVaUNNdHRONDhKUE1pazRsOVVZT0pUVWcvMWZtNmtkTE5hams3NVZ0K2hjR0RGbAoxN3B2LytPY3JVWTFpV0I0ZGdFb29rT3BUZlltMGQ2Q0J0dzdoRXg4OFE1YU9kTllOVk5kT1hPMk95MHZ6M0l4CmNGTnIxUEI3T3E2bnNyV3cyYkx6Qm94bU1zb05CekVkaUVEYVhYZVFmam5QdFdwclN4RGFlNEEwNDU4dERoMXEKQXZ3QlJ3cWFzRU5uNmxaZ2ZYOWFCcEZiUUhzL3YreTFzN0c1WG9lSTNjU3lZOFozR2V3cGdoaEVnakViQkNCKwpoRUhBL2tydlhCWnl2UHFCdDB1Rmswb1o4UUlEQVFBQm80R1ZNSUdTTUE0R0ExVWREd0VCL3dRRUF3SUZvREFkCkJnTlZIU1VFRmpBVUJnZ3JCZ0VGQlFjREFRWUlLd1lCQlFVSEF3SXdEQVlEVlIwVEFRSC9CQUl3QURCVEJnTlYKSFJFRVREQktnZzF6YjIxbExXNWhiV1V0Y0hoamdoSnpiMjFsTFc1aGJXVXRjSEp2ZUhsemNXeUNEeW91YzI5dApaUzF1WVcxbExYQjRZNElVS2k1emIyMWxMVzVoYldVdGNISnZlSGx6Y1d3d0RRWUpLb1pJaHZjTkFRRUxCUUFECmdnRUJBSFo1ekhGRjZ6bVJaYnVLVW5yWDM4RUU0UndqVExUaGhzSXhjb0F0MWY3b05KMW13RVdxdzZ1UHE3dE0KQlRwd2F5VytwdUUxS0JmZWJXTUhlNmdHa3NOdE0vTDI0bGM1S1NybFZCUDQvMytMdWViTVNHNVlvdDNONTU1QgpabHlwaFM1bWJJbHJaZUFUUVhXUUpkYW12dk1aOTM4SEk2WXJjOW9UZlg3blpsQUhoVUdkWUhwdjFGU0paZFlhCndSWCtjME0walZPSTk2dHRjaWtFSmhUK0F5b3NRV2VDYXljUG9qNHhBanVtZzJHM3VaT1dNdjR3RllkK09TSnIKc085U3VrVDN6SDg4dXdhbjhsTlBHcVhVc2dFS1R6U2c2Q0ZjSzBldUNCSjAvMG52c01JNS9RemRXbmI5cmZiUQpXdEJMMFJveEt0Um9sNnVHM0JMMi8rU3krLzg9Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K"' ']' + checkTLSSecret some-name-ssl tls.key + local secretName=some-name-ssl + local dataKey=tls.key ++ kubectl_bin get secrets/some-name-ssl -o json ++ jq '.data["tls.key"]' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MKjVvD6Pdy +++ mktemp ++ local LAST_ERR=/tmp/tmp.WLJ8NJnj7U ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/some-name-ssl -o json ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MKjVvD6Pdy ++ cat /tmp/tmp.WLJ8NJnj7U ++ rm /tmp/tmp.MKjVvD6Pdy /tmp/tmp.WLJ8NJnj7U ++ return 0 + local 'secretData="LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFcEFJQkFBS0NBUUVBdkZrb2xZMUNYRmlncWVaTEJxVjdySDhQWEl0aHcxb1BtZWV3QWJCRmFNWVVJMUJUCkVCQjZyVDEzWXpkNGRtWkRZbWl3MzdEcUVWMGp2VWlDTXR0TjQ4SlBNaWs0bDlVWU9KVFVnLzFmbTZrZExOYWoKazc1VnQraGNHREZsMTdwdi8rT2NyVVkxaVdCNGRnRW9va09wVGZZbTBkNkNCdHc3aEV4ODhRNWFPZE5ZTlZOZApPWE8yT3kwdnozSXhjRk5yMVBCN09xNm5zcld3MmJMekJveG1Nc29OQnpFZGlFRGFYWGVRZmpuUHRXcHJTeERhCmU0QTA0NTh0RGgxcUF2d0JSd3Fhc0VObjZsWmdmWDlhQnBGYlFIcy92K3kxczdHNVhvZUkzY1N5WThaM0dld3AKZ2hoRWdqRWJCQ0IraEVIQS9rcnZYQlp5dlBxQnQwdUZrMG9aOFFJREFRQUJBb0lCQUJLZDM5ZXFBSXZJa3RjKwpycXRlSEdhNE9zdDU4UWVsaUtXakpzTWFYWEgxOEpwdHhlVFd6Q1l6NXUrcmJ2TzlCRklTYVFSalBhTEtJd0diCi83MjhWTGY3ZVdkSzNjTDF5cERsS2VPQ3hHMTlaTUVOTzhDMDcwN2dFdThUcEhzczR3V0tMQ0lwKzJhMEsyREQKWmNITUtlWWFPMzdjTnR2NHU1SVRHdmpPU2NSZzMvakhwbFBIQmxlMXNwMHF0OU1sRDFpdkxJYzRYNk5aUmdzdApTdHZzT1BPTHhZaldHdk9PazU4Qnh1aDMra2RGZVpQQ2pGV25YbWNFK3Q2UnhKTzNnR0ZzTVF1RjM2MHQ2dGtrCnFJVnVpWGNTMCtwajA4eWR2bGVqUE5oZWFZcWlyaDZZVWZ2UXJ6ZmNxdXQxVWd6STlIK0ZOU0MvZ1MwOGIrYXIKVHoyWnlLVUNnWUVBdzkwVGVBTTdRRGRpb0ZCT3liVkVma0N2blJJN1ZvbVFKYndFQnl6OXF2MGpIbzJ1T0pVLwpOQkg0MVJWNW5RdjZUN1ZTSzBpZzVVZldDdURwWDdrbkwyZFpWeGd4SXBBNlpNZURRV1RMcHJUY2ZzRUtxY0hxCnNrTnFQaTR0SFJHS1pEZk1Rb25JdE5FK0kxbWozMFA4UmNDbS9DZmhud3M2WFBaZEc1T0sxamNDZ1lFQTlpMWcKdldRS25SQVFmMWFtTXhhUEFlYjhvZjVkc0lwV00wTU1NYW9mZHdRTTFXVXNScG1GYUtVZUhqREhuU1pBSXcrSQpDRzhTaHJFQUlCMmJmU0VGTWlOOFJnUVk4Undra05qMm1iY1VOZ2JtMnBaN2ZtWThiU0ZrRTNIMmgxaGdDZEtMClgyV0lSMHNlbVFrSUtWWVQvK1dzYmRPNjEwRGh6eEFrMmhhSWZSY0NnWUVBZ1FMOTcrOHc5L3VEVzFXS2tCSjIKS3pNd2VhV29zRGVzV3k2YytKMEorYmNEaXZ5SDVYMXBSRU1TUkVrQThoRVJyNjhma0tNbVYrQVpmTkRvNThSdApxeXZJRFRaQ0p5SVpWSWlXVEZ5NFIvYVp4bVFoUWZBWlI1SmdTTlRob21RTFpEZmNzM2RPWE12MndoSXBmc2FwClJNakJOUmFQNXRMZWZwMUlkd0dZVWNFQ2dZQjE3ZWtROFRMUWFOaXVWN0VJZWo4SnQwREF6WnpPTEliOHlacTIKMC90OHhUanNwU25mckQybmtxbnA1bzVKYndCOGl6MHNaU0VuMnZUbmJhZUYrdUtGVVZDQUxnaFF3SjJYc2dmNApQU2IyZHJueWJ2STc2L2Y0ZzIxd2owQjlOODNTQnNWemJxMHBFZ2hYWVdkQ2ZoVFltb3JPbGdaRWhiVm12eUlZClE0RWhBd0tCZ1FDV0pzbHBUK3NPdkcvZTRxZ0VTREZHTk5jNE83cDgwcWFFYWxJeU50NUs1bXluaHhsaGZVUnMKKzJkVmhKSzBSZHA2MUFjTGxoNXhyelR2ZmdPM0c0Q3NELzJtcm50akFmU2RWc05ZQWkzSkVuSFZDYWVoaVc1dwpxdTNOM1VLVXBwVzBhYXNpbXoxaUVlWTVzMGFqOE1qSER5MXF5MmFhQ0h0dkczZ1cvRVFjZGc9PQotLS0tLUVORCBSU0EgUFJJVkFURSBLRVktLS0tLQo="' + '[' -z '"LS0tLS1CRUdJTiBSU0EgUFJJVkFURSBLRVktLS0tLQpNSUlFcEFJQkFBS0NBUUVBdkZrb2xZMUNYRmlncWVaTEJxVjdySDhQWEl0aHcxb1BtZWV3QWJCRmFNWVVJMUJUCkVCQjZyVDEzWXpkNGRtWkRZbWl3MzdEcUVWMGp2VWlDTXR0TjQ4SlBNaWs0bDlVWU9KVFVnLzFmbTZrZExOYWoKazc1VnQraGNHREZsMTdwdi8rT2NyVVkxaVdCNGRnRW9va09wVGZZbTBkNkNCdHc3aEV4ODhRNWFPZE5ZTlZOZApPWE8yT3kwdnozSXhjRk5yMVBCN09xNm5zcld3MmJMekJveG1Nc29OQnpFZGlFRGFYWGVRZmpuUHRXcHJTeERhCmU0QTA0NTh0RGgxcUF2d0JSd3Fhc0VObjZsWmdmWDlhQnBGYlFIcy92K3kxczdHNVhvZUkzY1N5WThaM0dld3AKZ2hoRWdqRWJCQ0IraEVIQS9rcnZYQlp5dlBxQnQwdUZrMG9aOFFJREFRQUJBb0lCQUJLZDM5ZXFBSXZJa3RjKwpycXRlSEdhNE9zdDU4UWVsaUtXakpzTWFYWEgxOEpwdHhlVFd6Q1l6NXUrcmJ2TzlCRklTYVFSalBhTEtJd0diCi83MjhWTGY3ZVdkSzNjTDF5cERsS2VPQ3hHMTlaTUVOTzhDMDcwN2dFdThUcEhzczR3V0tMQ0lwKzJhMEsyREQKWmNITUtlWWFPMzdjTnR2NHU1SVRHdmpPU2NSZzMvakhwbFBIQmxlMXNwMHF0OU1sRDFpdkxJYzRYNk5aUmdzdApTdHZzT1BPTHhZaldHdk9PazU4Qnh1aDMra2RGZVpQQ2pGV25YbWNFK3Q2UnhKTzNnR0ZzTVF1RjM2MHQ2dGtrCnFJVnVpWGNTMCtwajA4eWR2bGVqUE5oZWFZcWlyaDZZVWZ2UXJ6ZmNxdXQxVWd6STlIK0ZOU0MvZ1MwOGIrYXIKVHoyWnlLVUNnWUVBdzkwVGVBTTdRRGRpb0ZCT3liVkVma0N2blJJN1ZvbVFKYndFQnl6OXF2MGpIbzJ1T0pVLwpOQkg0MVJWNW5RdjZUN1ZTSzBpZzVVZldDdURwWDdrbkwyZFpWeGd4SXBBNlpNZURRV1RMcHJUY2ZzRUtxY0hxCnNrTnFQaTR0SFJHS1pEZk1Rb25JdE5FK0kxbWozMFA4UmNDbS9DZmhud3M2WFBaZEc1T0sxamNDZ1lFQTlpMWcKdldRS25SQVFmMWFtTXhhUEFlYjhvZjVkc0lwV00wTU1NYW9mZHdRTTFXVXNScG1GYUtVZUhqREhuU1pBSXcrSQpDRzhTaHJFQUlCMmJmU0VGTWlOOFJnUVk4Undra05qMm1iY1VOZ2JtMnBaN2ZtWThiU0ZrRTNIMmgxaGdDZEtMClgyV0lSMHNlbVFrSUtWWVQvK1dzYmRPNjEwRGh6eEFrMmhhSWZSY0NnWUVBZ1FMOTcrOHc5L3VEVzFXS2tCSjIKS3pNd2VhV29zRGVzV3k2YytKMEorYmNEaXZ5SDVYMXBSRU1TUkVrQThoRVJyNjhma0tNbVYrQVpmTkRvNThSdApxeXZJRFRaQ0p5SVpWSWlXVEZ5NFIvYVp4bVFoUWZBWlI1SmdTTlRob21RTFpEZmNzM2RPWE12MndoSXBmc2FwClJNakJOUmFQNXRMZWZwMUlkd0dZVWNFQ2dZQjE3ZWtROFRMUWFOaXVWN0VJZWo4SnQwREF6WnpPTEliOHlacTIKMC90OHhUanNwU25mckQybmtxbnA1bzVKYndCOGl6MHNaU0VuMnZUbmJhZUYrdUtGVVZDQUxnaFF3SjJYc2dmNApQU2IyZHJueWJ2STc2L2Y0ZzIxd2owQjlOODNTQnNWemJxMHBFZ2hYWVdkQ2ZoVFltb3JPbGdaRWhiVm12eUlZClE0RWhBd0tCZ1FDV0pzbHBUK3NPdkcvZTRxZ0VTREZHTk5jNE83cDgwcWFFYWxJeU50NUs1bXluaHhsaGZVUnMKKzJkVmhKSzBSZHA2MUFjTGxoNXhyelR2ZmdPM0c0Q3NELzJtcm50akFmU2RWc05ZQWkzSkVuSFZDYWVoaVc1dwpxdTNOM1VLVXBwVzBhYXNpbXoxaUVlWTVzMGFqOE1qSER5MXF5MmFhQ0h0dkczZ1cvRVFjZGc9PQotLS0tLUVORCBSU0EgUFJJVkFURSBLRVktLS0tLQo="' ']' + desc 'check disabling tls' + set +o xtrace ----------------------------------------------------------------------------------- check disabling tls ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{"spec": {"tls":{"enabled": false}, "unsafeFlags": {"tls": true}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.mNmVJyx2DW ++ mktemp + local LAST_ERR=/tmp/tmp.LAtKoAxQBD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{"spec": {"tls":{"enabled": false}, "unsafeFlags": {"tls": true}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mNmVJyx2DW perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.LAtKoAxQBD + rm /tmp/tmp.mNmVJyx2DW /tmp/tmp.LAtKoAxQBD + return 0 + sleep 10 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TSE2e6MWp4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Zwkdaa5zwA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TSE2e6MWp4 ++ cat /tmp/tmp.Zwkdaa5zwA ++ rm /tmp/tmp.TSE2e6MWp4 /tmp/tmp.Zwkdaa5zwA ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PVNfsv6Fkx +++ mktemp ++ local LAST_ERR=/tmp/tmp.OJYkHU6e3G ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PVNfsv6Fkx ++ cat /tmp/tmp.OJYkHU6e3G ++ rm /tmp/tmp.PVNfsv6Fkx /tmp/tmp.OJYkHU6e3G ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TqAR03zPl6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ziVPNZPgPK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TqAR03zPl6 ++ cat /tmp/tmp.ziVPNZPgPK ++ rm /tmp/tmp.TqAR03zPl6 /tmp/tmp.ziVPNZPgPK ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Zs2WPlBmQe +++ mktemp ++ local LAST_ERR=/tmp/tmp.qW5TyVVuUj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Zs2WPlBmQe ++ cat /tmp/tmp.qW5TyVVuUj ++ rm /tmp/tmp.Zs2WPlBmQe /tmp/tmp.qW5TyVVuUj ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2yGzA7V1ka +++ mktemp ++ local LAST_ERR=/tmp/tmp.gAMAel8PJH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2yGzA7V1ka ++ cat /tmp/tmp.gAMAel8PJH ++ rm /tmp/tmp.2yGzA7V1ka /tmp/tmp.gAMAel8PJH ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rrVtKIbBnf +++ mktemp ++ local LAST_ERR=/tmp/tmp.goiay9zgUH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rrVtKIbBnf ++ cat /tmp/tmp.goiay9zgUH ++ rm /tmp/tmp.rrVtKIbBnf /tmp/tmp.goiay9zgUH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bMsNvimDOE +++ mktemp ++ local LAST_ERR=/tmp/tmp.sp4WDTvkwE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bMsNvimDOE ++ cat /tmp/tmp.sp4WDTvkwE ++ rm /tmp/tmp.bMsNvimDOE /tmp/tmp.sp4WDTvkwE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lUfevfx6jp +++ mktemp ++ local LAST_ERR=/tmp/tmp.TghZPwelce ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lUfevfx6jp ++ cat /tmp/tmp.TghZPwelce ++ rm /tmp/tmp.lUfevfx6jp /tmp/tmp.TghZPwelce ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vc1rh8MPLG +++ mktemp ++ local LAST_ERR=/tmp/tmp.NPHtpr2fKc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vc1rh8MPLG ++ cat /tmp/tmp.NPHtpr2fKc ++ rm /tmp/tmp.vc1rh8MPLG /tmp/tmp.NPHtpr2fKc ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rcKwmPr5LI +++ mktemp ++ local LAST_ERR=/tmp/tmp.dCtm1nIUqk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rcKwmPr5LI ++ cat /tmp/tmp.dCtm1nIUqk ++ rm /tmp/tmp.rcKwmPr5LI /tmp/tmp.dCtm1nIUqk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XVMzgYgWjt +++ mktemp ++ local LAST_ERR=/tmp/tmp.FkucPuIJTS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XVMzgYgWjt ++ cat /tmp/tmp.FkucPuIJTS ++ rm /tmp/tmp.XVMzgYgWjt /tmp/tmp.FkucPuIJTS ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NPsEolrME3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.rO8nHvQXZo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NPsEolrME3 ++ cat /tmp/tmp.rO8nHvQXZo ++ rm /tmp/tmp.NPsEolrME3 /tmp/tmp.rO8nHvQXZo ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NJsh95GxzM +++ mktemp ++ local LAST_ERR=/tmp/tmp.9iYYRDEfnA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NJsh95GxzM ++ cat /tmp/tmp.9iYYRDEfnA ++ rm /tmp/tmp.NJsh95GxzM /tmp/tmp.9iYYRDEfnA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.T3mHxSawER +++ mktemp ++ local LAST_ERR=/tmp/tmp.RXvfS976OH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.T3mHxSawER ++ cat /tmp/tmp.RXvfS976OH ++ rm /tmp/tmp.T3mHxSawER /tmp/tmp.RXvfS976OH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7lNvbHQlAi +++ mktemp ++ local LAST_ERR=/tmp/tmp.PKbYXWDfzE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7lNvbHQlAi ++ cat /tmp/tmp.PKbYXWDfzE ++ rm /tmp/tmp.7lNvbHQlAi /tmp/tmp.PKbYXWDfzE ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IWf3fYGj8K +++ mktemp ++ local LAST_ERR=/tmp/tmp.jx38vnzKAz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IWf3fYGj8K ++ cat /tmp/tmp.jx38vnzKAz ++ rm /tmp/tmp.IWf3fYGj8K /tmp/tmp.jx38vnzKAz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ba77zqDU0Q +++ mktemp ++ local LAST_ERR=/tmp/tmp.LCjHjx9NuQ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ba77zqDU0Q ++ cat /tmp/tmp.LCjHjx9NuQ ++ rm /tmp/tmp.ba77zqDU0Q /tmp/tmp.LCjHjx9NuQ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yBR2JtGfOj +++ mktemp ++ local LAST_ERR=/tmp/tmp.8lBfb7YRDi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yBR2JtGfOj ++ cat /tmp/tmp.8lBfb7YRDi ++ rm /tmp/tmp.yBR2JtGfOj /tmp/tmp.8lBfb7YRDi ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KUBYEEM7SX +++ mktemp ++ local LAST_ERR=/tmp/tmp.m6gi7z29Ax ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KUBYEEM7SX ++ cat /tmp/tmp.m6gi7z29Ax ++ rm /tmp/tmp.KUBYEEM7SX /tmp/tmp.m6gi7z29Ax ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A8kV5tCUSz +++ mktemp ++ local LAST_ERR=/tmp/tmp.DIgC4uzfVd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.A8kV5tCUSz ++ cat /tmp/tmp.DIgC4uzfVd ++ rm /tmp/tmp.A8kV5tCUSz /tmp/tmp.DIgC4uzfVd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Zm2u17EPKd +++ mktemp ++ local LAST_ERR=/tmp/tmp.zaGzX5Bh8r ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Zm2u17EPKd ++ cat /tmp/tmp.zaGzX5Bh8r ++ rm /tmp/tmp.Zm2u17EPKd /tmp/tmp.zaGzX5Bh8r ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bYZo0xw4nZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.ZytKJoN3bI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bYZo0xw4nZ ++ cat /tmp/tmp.ZytKJoN3bI ++ rm /tmp/tmp.bYZo0xw4nZ /tmp/tmp.ZytKJoN3bI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yMsPrJ2dDd +++ mktemp ++ local LAST_ERR=/tmp/tmp.DeLXRwy7Lj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yMsPrJ2dDd ++ cat /tmp/tmp.DeLXRwy7Lj ++ rm /tmp/tmp.yMsPrJ2dDd /tmp/tmp.DeLXRwy7Lj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gODvSnvHmL +++ mktemp ++ local LAST_ERR=/tmp/tmp.BoryXP4l06 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gODvSnvHmL ++ cat /tmp/tmp.BoryXP4l06 ++ rm /tmp/tmp.gODvSnvHmL /tmp/tmp.BoryXP4l06 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yRA7DVVtP4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.zZBoYInOZ8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yRA7DVVtP4 ++ cat /tmp/tmp.zZBoYInOZ8 ++ rm /tmp/tmp.yRA7DVVtP4 /tmp/tmp.zZBoYInOZ8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QTJOmhKl9C +++ mktemp ++ local LAST_ERR=/tmp/tmp.0QXtVVNq2h ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QTJOmhKl9C ++ cat /tmp/tmp.0QXtVVNq2h ++ rm /tmp/tmp.QTJOmhKl9C /tmp/tmp.0QXtVVNq2h ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.26So0gaX9j +++ mktemp ++ local LAST_ERR=/tmp/tmp.RrO1i8muVe ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.26So0gaX9j ++ cat /tmp/tmp.RrO1i8muVe ++ rm /tmp/tmp.26So0gaX9j /tmp/tmp.RrO1i8muVe ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9PDERp784R +++ mktemp ++ local LAST_ERR=/tmp/tmp.KVg4qpKK7I ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9PDERp784R ++ cat /tmp/tmp.KVg4qpKK7I ++ rm /tmp/tmp.9PDERp784R /tmp/tmp.KVg4qpKK7I ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.LQpJ9xYrWN ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.7yVqZczpsH +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.LQpJ9xYrWN +++++ cat /tmp/tmp.7yVqZczpsH +++++ rm /tmp/tmp.LQpJ9xYrWN /tmp/tmp.7yVqZczpsH +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.e9wSrfMdBJ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.HFHWXmJZht +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.e9wSrfMdBJ +++++ cat /tmp/tmp.HFHWXmJZht +++++ rm /tmp/tmp.e9wSrfMdBJ /tmp/tmp.HFHWXmJZht +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hssylfP50W +++ mktemp ++ local LAST_ERR=/tmp/tmp.nqk6kazeHf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hssylfP50W ++ cat /tmp/tmp.nqk6kazeHf ++ rm /tmp/tmp.hssylfP50W /tmp/tmp.nqk6kazeHf ++ return 0 + [[ 2 == \2 ]] + echo + desc 'secrets should be deleted' + set +o xtrace ----------------------------------------------------------------------------------- secrets should be deleted ----------------------------------------------------------------------------------- + kubectl get secret some-name-ssl + kubectl get secret some-name-ssl-internal + compare_kubectl pxc/some-name + local resource=pxc/some-name + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name.yml + local new_result=/tmp/tmp.nt78ipZG5a/pxc_some-name.yml + desc 'compare pxc/some-name-' + set +o xtrace ----------------------------------------------------------------------------------- compare pxc/some-name- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.33 >= 1.33' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-k133.yml ']' + version_gt 1.29 ++ echo '1.33 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-k129.yml ']' + version_gt 1.27 ++ echo '1.33 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-k127.yml ']' + version_gt 1.24 ++ echo '1.33 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-k124.yml ']' + version_gt 1.22 ++ echo '1.33 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-k122.yml ']' + version_gt 1.21 ++ bc -l ++ echo '1.33 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-oc.yml ']' + version_gt 1.29 ++ echo '1.33 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.spec.template.spec.containers[].env[] | select(.name == "XTRABACKUP_ENABLED")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.metadata.annotations."kubernetes.digitalocean.com/load-balancer-id") | del(.metadata.annotations."service.beta.kubernetes.io/do-loadbalancer-type") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | del(.spec.updateStrategy.rollingUpdate.maxUnavailable) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("tls-issue-self-3272", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml pxc/some-name ++ mktemp + local LAST_OUT=/tmp/tmp.uVtaOKcju6 ++ mktemp + local LAST_ERR=/tmp/tmp.l7AdbBbIAv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml pxc/some-name + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uVtaOKcju6 + cat /tmp/tmp.l7AdbBbIAv + rm /tmp/tmp.uVtaOKcju6 /tmp/tmp.l7AdbBbIAv + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2467/e2e-tests/tls-issue-self/compare/pxc_some-name.yml /tmp/tmp.nt78ipZG5a/pxc_some-name.yml + log 'compare_kubectl: pxc/some-name OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2026-05-16T22:05:32+0000]' compare_kubectl: pxc/some-name OK [2026-05-16T22:05:32+0000] compare_kubectl: pxc/some-name OK + desc 'check enabling tls' + set +o xtrace ----------------------------------------------------------------------------------- check enabling tls ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{"spec": {"tls":{"enabled": true}}}' ++ mktemp + local LAST_OUT=/tmp/tmp.zAjfskpb9T ++ mktemp + local LAST_ERR=/tmp/tmp.fXOiy0R8Xo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{"spec": {"tls":{"enabled": true}}}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zAjfskpb9T perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.fXOiy0R8Xo + rm /tmp/tmp.zAjfskpb9T /tmp/tmp.fXOiy0R8Xo + return 0 + sleep 10 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.afcc2WjbTU +++ mktemp ++ local LAST_ERR=/tmp/tmp.jGgj9RqBqs ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.afcc2WjbTU ++ cat /tmp/tmp.jGgj9RqBqs ++ rm /tmp/tmp.afcc2WjbTU /tmp/tmp.jGgj9RqBqs ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kUcY3ZJCQ8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.UWZLd2YpJi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kUcY3ZJCQ8 ++ cat /tmp/tmp.UWZLd2YpJi ++ rm /tmp/tmp.kUcY3ZJCQ8 /tmp/tmp.UWZLd2YpJi ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ly9163j0Eo +++ mktemp ++ local LAST_ERR=/tmp/tmp.CQMdFPqqja ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ly9163j0Eo ++ cat /tmp/tmp.CQMdFPqqja ++ rm /tmp/tmp.ly9163j0Eo /tmp/tmp.CQMdFPqqja ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RayteSuKvq +++ mktemp ++ local LAST_ERR=/tmp/tmp.L2Lnb60iXZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RayteSuKvq ++ cat /tmp/tmp.L2Lnb60iXZ ++ rm /tmp/tmp.RayteSuKvq /tmp/tmp.L2Lnb60iXZ ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6BuUfloGf4 +++ mktemp ++ local LAST_ERR=/tmp/tmp.fkatfHMtCL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6BuUfloGf4 ++ cat /tmp/tmp.fkatfHMtCL ++ rm /tmp/tmp.6BuUfloGf4 /tmp/tmp.fkatfHMtCL ++ return 0 + [[ paused == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jZ73HGCLWc +++ mktemp ++ local LAST_ERR=/tmp/tmp.zADvJpqRo9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jZ73HGCLWc ++ cat /tmp/tmp.zADvJpqRo9 ++ rm /tmp/tmp.jZ73HGCLWc /tmp/tmp.zADvJpqRo9 ++ return 0 + [[ error == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xQHdtJFyhq +++ mktemp ++ local LAST_ERR=/tmp/tmp.SGq59pULTq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xQHdtJFyhq ++ cat /tmp/tmp.SGq59pULTq ++ rm /tmp/tmp.xQHdtJFyhq /tmp/tmp.SGq59pULTq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hK8yKljJWS +++ mktemp ++ local LAST_ERR=/tmp/tmp.tj7F5yDXeH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hK8yKljJWS ++ cat /tmp/tmp.tj7F5yDXeH ++ rm /tmp/tmp.hK8yKljJWS /tmp/tmp.tj7F5yDXeH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RSNCa1a55P +++ mktemp ++ local LAST_ERR=/tmp/tmp.6qKTouBGdr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}'