Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/logs/upgrade-consistency-8-0.log WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 + cluster=some-name + main + create_infra upgrade-consistency-15810 + local ns=upgrade-consistency-15810 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n upgrade-consistency-5726 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.LEXB9BtVEf ++ mktemp + local LAST_ERR=/tmp/tmp.CG4Una61LL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.LEXB9BtVEf perconaxtradbcluster.pxc.percona.com "some-name" deleted + cat /tmp/tmp.CG4Una61LL + rm /tmp/tmp.LEXB9BtVEf /tmp/tmp.CG4Una61LL + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.yBAT7erUPa ++ mktemp + local LAST_ERR=/tmp/tmp.Gj5HvqlUzC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.yBAT7erUPa No resources found + cat /tmp/tmp.Gj5HvqlUzC + rm /tmp/tmp.yBAT7erUPa /tmp/tmp.Gj5HvqlUzC + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.qNYUnZkMnd ++ mktemp + local LAST_ERR=/tmp/tmp.au4pkzwGSh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qNYUnZkMnd No resources found + cat /tmp/tmp.au4pkzwGSh + rm /tmp/tmp.qNYUnZkMnd /tmp/tmp.au4pkzwGSh + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// ++ tail -n1 ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ grep chaos-mesh ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep validate-auth + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl api-resources ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.DLZubxCyJ2 ++ mktemp + local LAST_ERR=/tmp/tmp.l7TnSg3WYq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + awk '{print$1}' + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.TknGdFU3xu ++ mktemp + local LAST_ERR=/tmp/tmp.pU0zPOaFWY + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.DLZubxCyJ2 + cat /tmp/tmp.l7TnSg3WYq + rm /tmp/tmp.DLZubxCyJ2 /tmp/tmp.l7TnSg3WYq + return 0 namespace "upgrade-consistency-5726" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.TknGdFU3xu namespace "pxc-operator" deleted + cat /tmp/tmp.pU0zPOaFWY + rm /tmp/tmp.TknGdFU3xu /tmp/tmp.pU0zPOaFWY + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'namespace/pxc-operator - ' namespace/pxc-operator - + set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.1wF9EyJb3T ++ mktemp + local LAST_ERR=/tmp/tmp.4mchSD0yKR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.1wF9EyJb3T namespace/pxc-operator created + cat /tmp/tmp.4mchSD0yKR + rm /tmp/tmp.1wF9EyJb3T /tmp/tmp.4mchSD0yKR + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.yn2LKEmIyi +++ mktemp ++ local LAST_ERR=/tmp/tmp.dwKfNMxlMf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yn2LKEmIyi ++ cat /tmp/tmp.dwKfNMxlMf ++ rm /tmp/tmp.yn2LKEmIyi /tmp/tmp.dwKfNMxlMf ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1715-225b38be-1-cluster3 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.HrZlwgI4Al ++ mktemp + local LAST_ERR=/tmp/tmp.dzcbRiWIXh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1715-225b38be-1-cluster3 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HrZlwgI4Al Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1715-225b38be-1-cluster3" modified. + cat /tmp/tmp.dzcbRiWIXh + rm /tmp/tmp.HrZlwgI4Al /tmp/tmp.dzcbRiWIXh + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.OuSLAPi7OM ++ mktemp + local LAST_ERR=/tmp/tmp.SBWSKFfUNv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.OuSLAPi7OM customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.SBWSKFfUNv + rm /tmp/tmp.OuSLAPi7OM /tmp/tmp.SBWSKFfUNv + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.oMDP0MY5Mo ++ mktemp + local LAST_ERR=/tmp/tmp.ElWRxx0WFJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oMDP0MY5Mo clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.ElWRxx0WFJ + rm /tmp/tmp.oMDP0MY5Mo /tmp/tmp.ElWRxx0WFJ + return 0 + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "DEBUG"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-1715-225b38be^' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/deploy/cw-operator.yaml + kubectl_bin apply -f - ++ mktemp + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + local LAST_OUT=/tmp/tmp.VOPC1lZgkT ++ mktemp + local LAST_ERR=/tmp/tmp.YQBIkmVhhu + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VOPC1lZgkT deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.YQBIkmVhhu + rm /tmp/tmp.VOPC1lZgkT /tmp/tmp.YQBIkmVhhu + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.RrAojhHlrM ++ mktemp + local LAST_ERR=/tmp/tmp.wz5TahmSGG + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.RrAojhHlrM pod/percona-xtradb-cluster-operator-59b7fbbc57-7298j condition met + cat /tmp/tmp.wz5TahmSGG + rm /tmp/tmp.RrAojhHlrM /tmp/tmp.wz5TahmSGG + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.dvbphG0U8l +++ mktemp ++ local LAST_ERR=/tmp/tmp.mQ7BUQbc19 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dvbphG0U8l ++ cat /tmp/tmp.mQ7BUQbc19 ++ rm /tmp/tmp.dvbphG0U8l /tmp/tmp.mQ7BUQbc19 ++ return 0 + wait_pod percona-xtradb-cluster-operator-59b7fbbc57-7298j 480 pxc-operator + local pod=percona-xtradb-cluster-operator-59b7fbbc57-7298j + local max_retry=480 + local ns=pxc-operator ++ egrep '^(pxc|proxysql)$' ++ echo percona-xtradb-cluster-operator-59b7fbbc57-7298j ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-59b7fbbc57-7298j condition met percona-xtradb-cluster-operator-59b7fbbc57-7298j.Ok + sleep 3 + create_namespace upgrade-consistency-15810 + local namespace=upgrade-consistency-15810 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces upgrade-consistency-15810' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces upgrade-consistency-15810 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace upgrade-consistency-15810 + awk '{print$1}' ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.4t43CCLqSy + local LAST_OUT=/tmp/tmp.PuDlBoo4wz ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.1iyNhMb6qJ + local LAST_ERR=/tmp/tmp.I3F7raM5Dg + local exit_status=0 + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-consistency-15810 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-consistency-15810 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4t43CCLqSy + cat /tmp/tmp.I3F7raM5Dg + rm /tmp/tmp.4t43CCLqSy /tmp/tmp.I3F7raM5Dg + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-consistency-15810 Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.PuDlBoo4wz + cat /tmp/tmp.1iyNhMb6qJ Error from server (NotFound): namespaces "upgrade-consistency-15810" not found + rm /tmp/tmp.PuDlBoo4wz /tmp/tmp.1iyNhMb6qJ + return 1 + : + wait_for_delete namespace/upgrade-consistency-15810 + local res=namespace/upgrade-consistency-15810 + echo -n 'namespace/upgrade-consistency-15810 - ' namespace/upgrade-consistency-15810 - + set +o xtrace Error from server (NotFound): namespaces "upgrade-consistency-15810" not found + desc 'create namespace upgrade-consistency-15810' + set +o xtrace ----------------------------------------------------------------------------------- create namespace upgrade-consistency-15810 ----------------------------------------------------------------------------------- + kubectl_bin create namespace upgrade-consistency-15810 ++ mktemp + local LAST_OUT=/tmp/tmp.a175u0ZO4t ++ mktemp + local LAST_ERR=/tmp/tmp.9LFqrkNPsa + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace upgrade-consistency-15810 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.a175u0ZO4t namespace/upgrade-consistency-15810 created + cat /tmp/tmp.9LFqrkNPsa + rm /tmp/tmp.a175u0ZO4t /tmp/tmp.9LFqrkNPsa + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.DG6ePtsqrt +++ mktemp ++ local LAST_ERR=/tmp/tmp.IN8ROJNC6u ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DG6ePtsqrt ++ cat /tmp/tmp.IN8ROJNC6u ++ rm /tmp/tmp.DG6ePtsqrt /tmp/tmp.IN8ROJNC6u ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1715-225b38be-1-cluster3 --namespace=upgrade-consistency-15810 ++ mktemp + local LAST_OUT=/tmp/tmp.Ua85P3PN7l ++ mktemp + local LAST_ERR=/tmp/tmp.kIO64FZVCR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1715-225b38be-1-cluster3 --namespace=upgrade-consistency-15810 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Ua85P3PN7l Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1715-225b38be-1-cluster3" modified. + cat /tmp/tmp.kIO64FZVCR + rm /tmp/tmp.Ua85P3PN7l /tmp/tmp.kIO64FZVCR + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.igt3FinvVN ++ mktemp + local LAST_ERR=/tmp/tmp.MWv0TdNwUF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.igt3FinvVN secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.MWv0TdNwUF + rm /tmp/tmp.igt3FinvVN /tmp/tmp.MWv0TdNwUF + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/conf/client.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1715-225b38be#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.upgrade-consistency-15810~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + local LAST_OUT=/tmp/tmp.WrqQUzpjLw ++ mktemp + local LAST_ERR=/tmp/tmp.ibLCn3fxxu + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WrqQUzpjLw deployment.apps/pxc-client created + cat /tmp/tmp.ibLCn3fxxu + rm /tmp/tmp.WrqQUzpjLw /tmp/tmp.ibLCn3fxxu + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.Fa65Kw5Sbg ++ mktemp + local LAST_ERR=/tmp/tmp.S4apViuMPt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Fa65Kw5Sbg secret/my-cluster-secrets created + cat /tmp/tmp.S4apViuMPt + rm /tmp/tmp.Fa65Kw5Sbg /tmp/tmp.S4apViuMPt + return 0 + yq eval '.spec.crVersion = "1.12.0"' - + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.yBdbEfwxEA ++ mktemp + local LAST_ERR=/tmp/tmp.5e3egG7daA + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/conf/some-name.yml + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1715-225b38be#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.upgrade-consistency-15810~ + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/conf/some-name.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.yBdbEfwxEA perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.5e3egG7daA + rm /tmp/tmp.yBdbEfwxEA /tmp/tmp.5e3egG7daA + return 0 + desc 'test 1.12.0' + set +o xtrace ----------------------------------------------------------------------------------- test 1.12.0 ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.12.0"} }' ++ mktemp + local LAST_OUT=/tmp/tmp.CjSa6D6z7E ++ mktemp + local LAST_ERR=/tmp/tmp.KZjOHlTtTt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.12.0"} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CjSa6D6z7E perconaxtradbcluster.pxc.percona.com/some-name patched (no change) + cat /tmp/tmp.KZjOHlTtTt + rm /tmp/tmp.CjSa6D6z7E /tmp/tmp.KZjOHlTtTt + return 0 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DlzvuzYk9e +++ mktemp ++ local LAST_ERR=/tmp/tmp.ddtYSL3K9q ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DlzvuzYk9e ++ cat /tmp/tmp.ddtYSL3K9q ++ rm /tmp/tmp.DlzvuzYk9e /tmp/tmp.ddtYSL3K9q ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8gX0EnJjcS +++ mktemp ++ local LAST_ERR=/tmp/tmp.7riQ7qyH3m ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8gX0EnJjcS ++ cat /tmp/tmp.7riQ7qyH3m ++ rm /tmp/tmp.8gX0EnJjcS /tmp/tmp.7riQ7qyH3m ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.obmVXgdijl +++ mktemp ++ local LAST_ERR=/tmp/tmp.OWxunB9SFl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.obmVXgdijl ++ cat /tmp/tmp.OWxunB9SFl ++ rm /tmp/tmp.obmVXgdijl /tmp/tmp.OWxunB9SFl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QlDqtD1pBQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.GjBrm8fpjG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QlDqtD1pBQ ++ cat /tmp/tmp.GjBrm8fpjG ++ rm /tmp/tmp.QlDqtD1pBQ /tmp/tmp.GjBrm8fpjG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y41RrF2NdQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.dBu6Cd4BgP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y41RrF2NdQ ++ cat /tmp/tmp.dBu6Cd4BgP ++ rm /tmp/tmp.Y41RrF2NdQ /tmp/tmp.dBu6Cd4BgP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DA1Pwn6BUN +++ mktemp ++ local LAST_ERR=/tmp/tmp.zCo0CUKhih ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DA1Pwn6BUN ++ cat /tmp/tmp.zCo0CUKhih ++ rm /tmp/tmp.DA1Pwn6BUN /tmp/tmp.zCo0CUKhih ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5K48k1ws4G +++ mktemp ++ local LAST_ERR=/tmp/tmp.LnvphIBFKl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5K48k1ws4G ++ cat /tmp/tmp.LnvphIBFKl ++ rm /tmp/tmp.5K48k1ws4G /tmp/tmp.LnvphIBFKl ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 6 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jKKQuQwSTr +++ mktemp ++ local LAST_ERR=/tmp/tmp.EpjiOynPwC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jKKQuQwSTr ++ cat /tmp/tmp.EpjiOynPwC ++ rm /tmp/tmp.jKKQuQwSTr /tmp/tmp.EpjiOynPwC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 7 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.LnqCqftWWk +++ mktemp ++ local LAST_ERR=/tmp/tmp.ErlUVzQlZD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.LnqCqftWWk ++ cat /tmp/tmp.ErlUVzQlZD ++ rm /tmp/tmp.LnqCqftWWk /tmp/tmp.ErlUVzQlZD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 8 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jI0SAp3xV5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.MBnp5HUpJC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jI0SAp3xV5 ++ cat /tmp/tmp.MBnp5HUpJC ++ rm /tmp/tmp.jI0SAp3xV5 /tmp/tmp.MBnp5HUpJC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 9 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9oBwW7UXAl +++ mktemp ++ local LAST_ERR=/tmp/tmp.9Jx0pHG0W5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9oBwW7UXAl ++ cat /tmp/tmp.9Jx0pHG0W5 ++ rm /tmp/tmp.9oBwW7UXAl /tmp/tmp.9Jx0pHG0W5 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 10 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TzGreoy0WN +++ mktemp ++ local LAST_ERR=/tmp/tmp.g3y9JWlmJW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TzGreoy0WN ++ cat /tmp/tmp.g3y9JWlmJW ++ rm /tmp/tmp.TzGreoy0WN /tmp/tmp.g3y9JWlmJW ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EuM0pe7zw6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.WelRVjiEUZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EuM0pe7zw6 ++ cat /tmp/tmp.WelRVjiEUZ ++ rm /tmp/tmp.EuM0pe7zw6 /tmp/tmp.WelRVjiEUZ ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.7fVkmB5vbI ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.jIo2CwdI3v +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.7fVkmB5vbI +++++ cat /tmp/tmp.jIo2CwdI3v +++++ rm /tmp/tmp.7fVkmB5vbI /tmp/tmp.jIo2CwdI3v +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.vZQZnb9WlD ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.JwdlsfjJJ4 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.vZQZnb9WlD +++++ cat /tmp/tmp.JwdlsfjJJ4 +++++ rm /tmp/tmp.vZQZnb9WlD /tmp/tmp.JwdlsfjJJ4 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.a6wo50wCIm +++ mktemp ++ local LAST_ERR=/tmp/tmp.MiqpqtSjNd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.a6wo50wCIm ++ cat /tmp/tmp.MiqpqtSjNd ++ rm /tmp/tmp.a6wo50wCIm /tmp/tmp.MiqpqtSjNd ++ return 0 + [[ 2 == \2 ]] + wait_for_sts_generation some-name-pxc 1 1 + local name=some-name-pxc + local desired_generation=1 + local size=1 + set +o xtrace Waiting for statefulset readyness + desc 'compare k8s objects' + set +o xtrace ----------------------------------------------------------------------------------- compare k8s objects ----------------------------------------------------------------------------------- + compare_kubectl service/some-name-pxc -1120 + local resource=service/some-name-pxc + local postfix=-1120 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/service_some-name-pxc.yml + desc 'compare service/some-name-pxc--1120' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc--1120 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-eks.yml ']' + kubectl_bin get -o yaml service/some-name-pxc + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.bvR5cEPRqe ++ mktemp + local LAST_ERR=/tmp/tmp.xeNC5xJ3St + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.bvR5cEPRqe + cat /tmp/tmp.xeNC5xJ3St + rm /tmp/tmp.bvR5cEPRqe /tmp/tmp.xeNC5xJ3St + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120.yml /tmp/tmp.Vx5fyn6AHh/service_some-name-pxc.yml + compare_kubectl service/some-name-proxysql -1120 + local resource=service/some-name-proxysql + local postfix=-1120 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql--1120' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql--1120 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-eks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.rLO9zxqdfL ++ mktemp + local LAST_ERR=/tmp/tmp.9hjaIbypk2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rLO9zxqdfL + cat /tmp/tmp.9hjaIbypk2 + rm /tmp/tmp.rLO9zxqdfL /tmp/tmp.9hjaIbypk2 + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120.yml /tmp/tmp.Vx5fyn6AHh/service_some-name-proxysql.yml + compare_kubectl statefulset/some-name-pxc -1120 + local resource=statefulset/some-name-pxc + local postfix=-1120 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc--1120' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc--1120 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ bc -l ++ echo '1.26 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc ++ mktemp + local LAST_OUT=/tmp/tmp.egDcQEo0Ga + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_ERR=/tmp/tmp.IokAu96sWy + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.egDcQEo0Ga + cat /tmp/tmp.IokAu96sWy + rm /tmp/tmp.egDcQEo0Ga /tmp/tmp.IokAu96sWy + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120.yml /tmp/tmp.Vx5fyn6AHh/statefulset_some-name-pxc.yml + compare_kubectl statefulset/some-name-proxysql -1120 + local resource=statefulset/some-name-proxysql + local postfix=-1120 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql--1120' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql--1120 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ bc -l ++ echo '1.26 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-proxysql ++ mktemp + local LAST_OUT=/tmp/tmp.u9Qfyw8WXw ++ mktemp + local LAST_ERR=/tmp/tmp.YAaarQoPhG + local exit_status=0 + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.u9Qfyw8WXw + cat /tmp/tmp.YAaarQoPhG + rm /tmp/tmp.u9Qfyw8WXw /tmp/tmp.YAaarQoPhG + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120.yml /tmp/tmp.Vx5fyn6AHh/statefulset_some-name-proxysql.yml + desc 'test 1.13.0' + set +o xtrace ----------------------------------------------------------------------------------- test 1.13.0 ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.13.0"} }' ++ mktemp + local LAST_OUT=/tmp/tmp.oVPCr8ipfF ++ mktemp + local LAST_ERR=/tmp/tmp.qshX4uChL2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.13.0"} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oVPCr8ipfF perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.qshX4uChL2 + rm /tmp/tmp.oVPCr8ipfF /tmp/tmp.qshX4uChL2 + return 0 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OjRz0tvAIj +++ mktemp ++ local LAST_ERR=/tmp/tmp.dXA92EVSIP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OjRz0tvAIj ++ cat /tmp/tmp.dXA92EVSIP ++ rm /tmp/tmp.OjRz0tvAIj /tmp/tmp.dXA92EVSIP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XHU5FvPfBn +++ mktemp ++ local LAST_ERR=/tmp/tmp.pvM9So3My7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XHU5FvPfBn ++ cat /tmp/tmp.pvM9So3My7 ++ rm /tmp/tmp.XHU5FvPfBn /tmp/tmp.pvM9So3My7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IRmRJ4mqwM +++ mktemp ++ local LAST_ERR=/tmp/tmp.v69QlfJSks ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IRmRJ4mqwM ++ cat /tmp/tmp.v69QlfJSks ++ rm /tmp/tmp.IRmRJ4mqwM /tmp/tmp.v69QlfJSks ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.20Zaz2lIyp +++ mktemp ++ local LAST_ERR=/tmp/tmp.A4jhphPAA3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.20Zaz2lIyp ++ cat /tmp/tmp.A4jhphPAA3 ++ rm /tmp/tmp.20Zaz2lIyp /tmp/tmp.A4jhphPAA3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.d75ShN73v6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.GA1T0PenTU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.d75ShN73v6 ++ cat /tmp/tmp.GA1T0PenTU ++ rm /tmp/tmp.d75ShN73v6 /tmp/tmp.GA1T0PenTU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.RFmxBxByHW +++ mktemp ++ local LAST_ERR=/tmp/tmp.ftfswwRGYI ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RFmxBxByHW ++ cat /tmp/tmp.ftfswwRGYI ++ rm /tmp/tmp.RFmxBxByHW /tmp/tmp.ftfswwRGYI ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iwY4DIUX8L +++ mktemp ++ local LAST_ERR=/tmp/tmp.3QBZYeuzT1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iwY4DIUX8L ++ cat /tmp/tmp.3QBZYeuzT1 ++ rm /tmp/tmp.iwY4DIUX8L /tmp/tmp.3QBZYeuzT1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 6 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BD8GhGTQF9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.zkoy0Wxk8P ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BD8GhGTQF9 ++ cat /tmp/tmp.zkoy0Wxk8P ++ rm /tmp/tmp.BD8GhGTQF9 /tmp/tmp.zkoy0Wxk8P ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 7 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ynCeETRFsy +++ mktemp ++ local LAST_ERR=/tmp/tmp.87DpPq3K9f ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ynCeETRFsy ++ cat /tmp/tmp.87DpPq3K9f ++ rm /tmp/tmp.ynCeETRFsy /tmp/tmp.87DpPq3K9f ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.de4GrfSF3M +++ mktemp ++ local LAST_ERR=/tmp/tmp.5FOtu1Ud9Z ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.de4GrfSF3M ++ cat /tmp/tmp.5FOtu1Ud9Z ++ rm /tmp/tmp.de4GrfSF3M /tmp/tmp.5FOtu1Ud9Z ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.K2WoqnVV6D ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.b8xikSTapx +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.K2WoqnVV6D +++++ cat /tmp/tmp.b8xikSTapx +++++ rm /tmp/tmp.K2WoqnVV6D /tmp/tmp.b8xikSTapx +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.G1TBFykbDu ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.25Yd5iL0Vs +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.G1TBFykbDu +++++ cat /tmp/tmp.25Yd5iL0Vs +++++ rm /tmp/tmp.G1TBFykbDu /tmp/tmp.25Yd5iL0Vs +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7sZwElokJs +++ mktemp ++ local LAST_ERR=/tmp/tmp.17J5SHfgfp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7sZwElokJs ++ cat /tmp/tmp.17J5SHfgfp ++ rm /tmp/tmp.7sZwElokJs /tmp/tmp.17J5SHfgfp ++ return 0 + [[ 2 == \2 ]] + wait_for_sts_generation some-name-pxc 2 1 + local name=some-name-pxc + local desired_generation=2 + local size=1 + set +o xtrace Waiting for statefulset readyness + desc 'compare k8s objects' + set +o xtrace ----------------------------------------------------------------------------------- compare k8s objects ----------------------------------------------------------------------------------- + compare_kubectl service/some-name-pxc -1130 + local resource=service/some-name-pxc + local postfix=-1130 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/service_some-name-pxc.yml + desc 'compare service/some-name-pxc--1130' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc--1130 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-eks.yml ']' + kubectl_bin get -o yaml service/some-name-pxc + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | ++ mktemp del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.mB2z2xCA9D ++ mktemp + local LAST_ERR=/tmp/tmp.p2392T00UM + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mB2z2xCA9D + cat /tmp/tmp.p2392T00UM + rm /tmp/tmp.mB2z2xCA9D /tmp/tmp.p2392T00UM + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130.yml /tmp/tmp.Vx5fyn6AHh/service_some-name-pxc.yml + compare_kubectl service/some-name-proxysql -1130 + local resource=service/some-name-proxysql + local postfix=-1130 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql--1130' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql--1130 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-eks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.VsXwPB9X41 ++ mktemp + local LAST_ERR=/tmp/tmp.QRJ2YqOXcZ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VsXwPB9X41 + cat /tmp/tmp.QRJ2YqOXcZ + rm /tmp/tmp.VsXwPB9X41 /tmp/tmp.QRJ2YqOXcZ + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130.yml /tmp/tmp.Vx5fyn6AHh/service_some-name-proxysql.yml + compare_kubectl statefulset/some-name-pxc -1130 + local resource=statefulset/some-name-pxc + local postfix=-1130 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc--1130' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc--1130 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.eJdAr4vSc5 ++ mktemp + local LAST_ERR=/tmp/tmp.I6YmgGZzKM + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.eJdAr4vSc5 + cat /tmp/tmp.I6YmgGZzKM + rm /tmp/tmp.eJdAr4vSc5 /tmp/tmp.I6YmgGZzKM + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130.yml /tmp/tmp.Vx5fyn6AHh/statefulset_some-name-pxc.yml + compare_kubectl statefulset/some-name-proxysql -1130 + local resource=statefulset/some-name-proxysql + local postfix=-1130 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql--1130' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql--1130 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.mgSeetHKj6 ++ mktemp + local LAST_ERR=/tmp/tmp.JaZohoFAve + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mgSeetHKj6 + cat /tmp/tmp.JaZohoFAve + rm /tmp/tmp.mgSeetHKj6 /tmp/tmp.JaZohoFAve + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130.yml /tmp/tmp.Vx5fyn6AHh/statefulset_some-name-proxysql.yml + desc 'test 1.14.0' + set +o xtrace ----------------------------------------------------------------------------------- test 1.14.0 ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.14.0"} }' ++ mktemp + local LAST_OUT=/tmp/tmp.mNwIyBN4F5 ++ mktemp + local LAST_ERR=/tmp/tmp.7K8WtIz4V8 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.14.0"} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mNwIyBN4F5 perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.7K8WtIz4V8 + rm /tmp/tmp.mNwIyBN4F5 /tmp/tmp.7K8WtIz4V8 + return 0 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ki45VHrsZG +++ mktemp ++ local LAST_ERR=/tmp/tmp.e1CfnWe2Fj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ki45VHrsZG ++ cat /tmp/tmp.e1CfnWe2Fj ++ rm /tmp/tmp.Ki45VHrsZG /tmp/tmp.e1CfnWe2Fj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8CX0KQgepe +++ mktemp ++ local LAST_ERR=/tmp/tmp.qIeUzPnWed ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8CX0KQgepe ++ cat /tmp/tmp.qIeUzPnWed ++ rm /tmp/tmp.8CX0KQgepe /tmp/tmp.qIeUzPnWed ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6r0cgnScmt +++ mktemp ++ local LAST_ERR=/tmp/tmp.lDdb8jFhv4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6r0cgnScmt ++ cat /tmp/tmp.lDdb8jFhv4 ++ rm /tmp/tmp.6r0cgnScmt /tmp/tmp.lDdb8jFhv4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o3h0ajkvTd +++ mktemp ++ local LAST_ERR=/tmp/tmp.RDQqf1CpGA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.o3h0ajkvTd ++ cat /tmp/tmp.RDQqf1CpGA ++ rm /tmp/tmp.o3h0ajkvTd /tmp/tmp.RDQqf1CpGA ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eyfT29ZCbX +++ mktemp ++ local LAST_ERR=/tmp/tmp.qkfHElILpC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eyfT29ZCbX ++ cat /tmp/tmp.qkfHElILpC ++ rm /tmp/tmp.eyfT29ZCbX /tmp/tmp.qkfHElILpC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.a1nWlVoQbm +++ mktemp ++ local LAST_ERR=/tmp/tmp.M96bRZnS0V ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.a1nWlVoQbm ++ cat /tmp/tmp.M96bRZnS0V ++ rm /tmp/tmp.a1nWlVoQbm /tmp/tmp.M96bRZnS0V ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.l7lJNoN6Ck +++ mktemp ++ local LAST_ERR=/tmp/tmp.Qz6KkxURss ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.l7lJNoN6Ck ++ cat /tmp/tmp.Qz6KkxURss ++ rm /tmp/tmp.l7lJNoN6Ck /tmp/tmp.Qz6KkxURss ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 6 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bikKSIK5Hc +++ mktemp ++ local LAST_ERR=/tmp/tmp.sP4O8JJlZX ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bikKSIK5Hc ++ cat /tmp/tmp.sP4O8JJlZX ++ rm /tmp/tmp.bikKSIK5Hc /tmp/tmp.sP4O8JJlZX ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GHnDOrgmmO +++ mktemp ++ local LAST_ERR=/tmp/tmp.GTKRURljqP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GHnDOrgmmO ++ cat /tmp/tmp.GTKRURljqP ++ rm /tmp/tmp.GHnDOrgmmO /tmp/tmp.GTKRURljqP ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.snQ1R7wK9I ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.CRzQjq4MDj +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.snQ1R7wK9I +++++ cat /tmp/tmp.CRzQjq4MDj +++++ rm /tmp/tmp.snQ1R7wK9I /tmp/tmp.CRzQjq4MDj +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.jckd4CDEAA ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.FtmnqKauvB +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.jckd4CDEAA +++++ cat /tmp/tmp.FtmnqKauvB +++++ rm /tmp/tmp.jckd4CDEAA /tmp/tmp.FtmnqKauvB +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MlJijX6Cok +++ mktemp ++ local LAST_ERR=/tmp/tmp.52oEYVVG1h ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MlJijX6Cok ++ cat /tmp/tmp.52oEYVVG1h ++ rm /tmp/tmp.MlJijX6Cok /tmp/tmp.52oEYVVG1h ++ return 0 + [[ 2 == \2 ]] + wait_for_sts_generation some-name-pxc 3 1 + local name=some-name-pxc + local desired_generation=3 + local size=1 + set +o xtrace Waiting for statefulset readyness + desc 'compare k8s objects' + set +o xtrace ----------------------------------------------------------------------------------- compare k8s objects ----------------------------------------------------------------------------------- + compare_kubectl service/some-name-pxc -1140 + local resource=service/some-name-pxc + local postfix=-1140 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/service_some-name-pxc.yml + desc 'compare service/some-name-pxc--1140' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc--1140 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ bc -l ++ echo '1.26 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ bc -l ++ echo '1.26 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ bc -l ++ echo '1.26 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-eks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml service/some-name-pxc ++ mktemp + local LAST_OUT=/tmp/tmp.QbFBZRk09a ++ mktemp + local LAST_ERR=/tmp/tmp.2Ry5H4i1YD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QbFBZRk09a + cat /tmp/tmp.2Ry5H4i1YD + rm /tmp/tmp.QbFBZRk09a /tmp/tmp.2Ry5H4i1YD + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140.yml /tmp/tmp.Vx5fyn6AHh/service_some-name-pxc.yml + compare_kubectl service/some-name-proxysql -1140 + local resource=service/some-name-proxysql + local postfix=-1140 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql--1140' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql--1140 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ bc -l ++ echo '1.26 >= 1.29' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-eks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | ++ mktemp del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.QtnbeQ8sPM ++ mktemp + local LAST_ERR=/tmp/tmp.vWvkjcVKbj + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QtnbeQ8sPM + cat /tmp/tmp.vWvkjcVKbj + rm /tmp/tmp.QtnbeQ8sPM /tmp/tmp.vWvkjcVKbj + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140.yml /tmp/tmp.Vx5fyn6AHh/service_some-name-proxysql.yml + compare_kubectl statefulset/some-name-pxc -1140 + local resource=statefulset/some-name-pxc + local postfix=-1140 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc--1140' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc--1140 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.zCJG2eMaa6 ++ mktemp + local LAST_ERR=/tmp/tmp.WG9rO1x5n5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zCJG2eMaa6 + cat /tmp/tmp.WG9rO1x5n5 + rm /tmp/tmp.zCJG2eMaa6 /tmp/tmp.WG9rO1x5n5 + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140.yml /tmp/tmp.Vx5fyn6AHh/statefulset_some-name-pxc.yml + compare_kubectl statefulset/some-name-proxysql -1140 + local resource=statefulset/some-name-proxysql + local postfix=-1140 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql--1140' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql--1140 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ bc -l ++ echo '1.26 >= 1.29' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.WOQNfiZgGE ++ mktemp + local LAST_ERR=/tmp/tmp.78yKZADoFe + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WOQNfiZgGE + cat /tmp/tmp.78yKZADoFe + rm /tmp/tmp.WOQNfiZgGE /tmp/tmp.78yKZADoFe + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140.yml /tmp/tmp.Vx5fyn6AHh/statefulset_some-name-proxysql.yml + desc 'test 1.15.0' + set +o xtrace ----------------------------------------------------------------------------------- test 1.15.0 ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.15.0"} }' ++ mktemp + local LAST_OUT=/tmp/tmp.sumxs5BV5V ++ mktemp + local LAST_ERR=/tmp/tmp.lA4K4O2krL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.15.0"} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sumxs5BV5V perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.lA4K4O2krL + rm /tmp/tmp.sumxs5BV5V /tmp/tmp.lA4K4O2krL + return 0 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pB6o2eVILp +++ mktemp ++ local LAST_ERR=/tmp/tmp.sxcC9Eceio ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pB6o2eVILp ++ cat /tmp/tmp.sxcC9Eceio ++ rm /tmp/tmp.pB6o2eVILp /tmp/tmp.sxcC9Eceio ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6u5oZj8KCS +++ mktemp ++ local LAST_ERR=/tmp/tmp.xxP4TSOGFZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6u5oZj8KCS ++ cat /tmp/tmp.xxP4TSOGFZ ++ rm /tmp/tmp.6u5oZj8KCS /tmp/tmp.xxP4TSOGFZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UeqTINaXp1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.4DFv90caz2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UeqTINaXp1 ++ cat /tmp/tmp.4DFv90caz2 ++ rm /tmp/tmp.UeqTINaXp1 /tmp/tmp.4DFv90caz2 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lupabNmuK0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.ap1EYWLCnz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lupabNmuK0 ++ cat /tmp/tmp.ap1EYWLCnz ++ rm /tmp/tmp.lupabNmuK0 /tmp/tmp.ap1EYWLCnz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Kk730V2Vrt +++ mktemp ++ local LAST_ERR=/tmp/tmp.EJCr5oF8Ng ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Kk730V2Vrt ++ cat /tmp/tmp.EJCr5oF8Ng ++ rm /tmp/tmp.Kk730V2Vrt /tmp/tmp.EJCr5oF8Ng ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xmQdfyEi5T +++ mktemp ++ local LAST_ERR=/tmp/tmp.6s3PECkqu7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xmQdfyEi5T ++ cat /tmp/tmp.6s3PECkqu7 ++ rm /tmp/tmp.xmQdfyEi5T /tmp/tmp.6s3PECkqu7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0EpUnEE3Q3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.HpFMVHvgOg ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0EpUnEE3Q3 ++ cat /tmp/tmp.HpFMVHvgOg ++ rm /tmp/tmp.0EpUnEE3Q3 /tmp/tmp.HpFMVHvgOg ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 6 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4VBqQJlI8h +++ mktemp ++ local LAST_ERR=/tmp/tmp.qjVbp4HEMd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4VBqQJlI8h ++ cat /tmp/tmp.qjVbp4HEMd ++ rm /tmp/tmp.4VBqQJlI8h /tmp/tmp.qjVbp4HEMd ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 7 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ws41v4fHdc +++ mktemp ++ local LAST_ERR=/tmp/tmp.rQsjnUCYsG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ws41v4fHdc ++ cat /tmp/tmp.rQsjnUCYsG ++ rm /tmp/tmp.ws41v4fHdc /tmp/tmp.rQsjnUCYsG ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PMc8Jofn0f +++ mktemp ++ local LAST_ERR=/tmp/tmp.z1yvAvMTw1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PMc8Jofn0f ++ cat /tmp/tmp.z1yvAvMTw1 ++ rm /tmp/tmp.PMc8Jofn0f /tmp/tmp.z1yvAvMTw1 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.PVESm3tQ95 ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.hnbJWYrkty +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.PVESm3tQ95 +++++ cat /tmp/tmp.hnbJWYrkty +++++ rm /tmp/tmp.PVESm3tQ95 /tmp/tmp.hnbJWYrkty +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.dxDEBg1UEA ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.87Mk2wdsdL +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.dxDEBg1UEA +++++ cat /tmp/tmp.87Mk2wdsdL +++++ rm /tmp/tmp.dxDEBg1UEA /tmp/tmp.87Mk2wdsdL +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sjYSFejpDl +++ mktemp ++ local LAST_ERR=/tmp/tmp.34Uc0GRQrr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sjYSFejpDl ++ cat /tmp/tmp.34Uc0GRQrr ++ rm /tmp/tmp.sjYSFejpDl /tmp/tmp.34Uc0GRQrr ++ return 0 + [[ 2 == \2 ]] + wait_for_sts_generation some-name-pxc 4 1 + local name=some-name-pxc + local desired_generation=4 + local size=1 + set +o xtrace Waiting for statefulset readyness + desc 'compare k8s objects' + set +o xtrace ----------------------------------------------------------------------------------- compare k8s objects ----------------------------------------------------------------------------------- + compare_kubectl service/some-name-pxc -1150 + local resource=service/some-name-pxc + local postfix=-1150 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/service_some-name-pxc.yml + desc 'compare service/some-name-pxc--1150' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc--1150 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-eks.yml ']' + kubectl_bin get -o yaml service/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.J2aWxh3718 ++ mktemp + local LAST_ERR=/tmp/tmp.p4gzw546aq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.J2aWxh3718 + cat /tmp/tmp.p4gzw546aq + rm /tmp/tmp.J2aWxh3718 /tmp/tmp.p4gzw546aq + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150.yml /tmp/tmp.Vx5fyn6AHh/service_some-name-pxc.yml + compare_kubectl service/some-name-proxysql -1150 + local resource=service/some-name-proxysql + local postfix=-1150 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql--1150' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql--1150 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-eks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.yI3xtkjZrg ++ mktemp + local LAST_ERR=/tmp/tmp.cDHq2T5044 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.yI3xtkjZrg + cat /tmp/tmp.cDHq2T5044 + rm /tmp/tmp.yI3xtkjZrg /tmp/tmp.cDHq2T5044 + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150.yml /tmp/tmp.Vx5fyn6AHh/service_some-name-proxysql.yml + compare_kubectl statefulset/some-name-pxc -1150 + local resource=statefulset/some-name-pxc + local postfix=-1150 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc--1150' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc--1150 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.FTrUvsTUac ++ mktemp + local LAST_ERR=/tmp/tmp.pmFLfO2jFl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FTrUvsTUac + cat /tmp/tmp.pmFLfO2jFl + rm /tmp/tmp.FTrUvsTUac /tmp/tmp.pmFLfO2jFl + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150.yml /tmp/tmp.Vx5fyn6AHh/statefulset_some-name-pxc.yml + compare_kubectl statefulset/some-name-proxysql -1150 + local resource=statefulset/some-name-proxysql + local postfix=-1150 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150.yml + local new_result=/tmp/tmp.Vx5fyn6AHh/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql--1150' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql--1150 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-proxysql + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-15810", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.xt2eEsLqn9 ++ mktemp + local LAST_ERR=/tmp/tmp.Vn7MYt5N6x + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xt2eEsLqn9 + cat /tmp/tmp.Vn7MYt5N6x + rm /tmp/tmp.xt2eEsLqn9 /tmp/tmp.Vn7MYt5N6x + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1715/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150.yml /tmp/tmp.Vx5fyn6AHh/statefulset_some-name-proxysql.yml + destroy upgrade-consistency-15810 + local namespace=upgrade-consistency-15810 + local ignore_logs=true + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' + grep -v 'the object has been modified' ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ + tee /tmp/tmp.Vx5fyn6AHh/operator.log + grep -v 'get backup status: Job.batch' + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator + sort -u + grep -v level=info +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.2sq3OgFWqN +++ mktemp ++ local LAST_ERR=/tmp/tmp.nnpWTkKbIo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2sq3OgFWqN ++ cat /tmp/tmp.nnpWTkKbIo ++ rm /tmp/tmp.2sq3OgFWqN /tmp/tmp.nnpWTkKbIo ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-59b7fbbc57-7298j ++ mktemp + local LAST_OUT=/tmp/tmp.Wf2yDZ2Jx0 ++ mktemp + local LAST_ERR=/tmp/tmp.kiZiIWeO7Y + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-59b7fbbc57-7298j + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Wf2yDZ2Jx0 + cat /tmp/tmp.kiZiIWeO7Y + rm /tmp/tmp.Wf2yDZ2Jx0 /tmp/tmp.kiZiIWeO7Y + return 0 2024-05-16T03:20:19.156Z INFO setup Manager starting up {"gitCommit": "225b38be01f1d85ab3ba5404a1190d44219127c8", "gitBranch": "PR-1715-225b38be", "buildTime": "2024-05-16T01:35:00Z", "goVersion": "go1.22.3", "os": "linux", "arch": "amd64"} 2024-05-16T03:20:19.156Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.26.15-gke.1300000"} 2024-05-16T03:20:19.229Z INFO setup Registering Components. 2024-05-16T03:20:21.039Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2024-05-16T03:20:21.042Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2024-05-16T03:20:21.042Z INFO controller-runtime.metrics Starting metrics server 2024-05-16T03:20:21.042Z INFO controller-runtime.webhook Starting webhook server 2024-05-16T03:20:21.042Z INFO setup Starting the Cmd. 2024-05-16T03:20:21.042Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2024-05-16T03:20:21.043Z INFO controller-runtime.certwatcher Starting certificate watcher 2024-05-16T03:20:21.043Z INFO controller-runtime.certwatcher Updated current TLS certificate 2024-05-16T03:20:21.043Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2024-05-16T03:20:21.144Z INFO attempting to acquire leader lease pxc-operator/08db1feb.percona.com... 2024-05-16T03:20:21.160Z INFO successfully acquired lease pxc-operator/08db1feb.percona.com 2024-05-16T03:20:21.161Z DEBUG events percona-xtradb-cluster-operator-59b7fbbc57-7298j_843c1a39-cf84-463a-b770-6ae45699dab1 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"3e723c4e-a869-46a9-8825-7f71d0e63801","apiVersion":"coordination.k8s.io/v1","resourceVersion":"57571"}, "reason": "LeaderElection"} 2024-05-16T03:20:21.161Z INFO Starting Controller {"controller": "pxcbackup-controller"} 2024-05-16T03:20:21.161Z INFO Starting Controller {"controller": "pxc-controller"} 2024-05-16T03:20:21.161Z INFO Starting Controller {"controller": "pxcrestore-controller"} 2024-05-16T03:20:21.161Z INFO Starting EventSource {"controller": "pxcbackup-controller", "source": "kind source: unknown type"} 2024-05-16T03:20:21.161Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: unknown type"} 2024-05-16T03:20:21.161Z INFO Starting EventSource {"controller": "pxcrestore-controller", "source": "kind source: unknown type"} 2024-05-16T03:20:21.329Z INFO Starting workers {"controller": "pxcrestore-controller", "worker count": 1} 2024-05-16T03:20:21.330Z INFO Starting workers {"controller": "pxcbackup-controller", "worker count": 1} 2024-05-16T03:20:21.371Z INFO Starting workers {"controller": "pxc-controller", "worker count": 1} 2024-05-16T03:20:58.531Z INFO User secrets updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "a758fb99-0312-4a4c-a207-c72723f8df97", "secrets": "my-cluster-secrets"} 2024-05-16T03:22:18.587Z INFO User monitor: granted privileges {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "2a1de214-5851-4ea0-b96a-de4374aa3789"} 2024-05-16T03:22:18.616Z INFO monitor user privileges granted {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "2a1de214-5851-4ea0-b96a-de4374aa3789"} 2024-05-16T03:22:18.651Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "2a1de214-5851-4ea0-b96a-de4374aa3789"} 2024-05-16T03:22:18.738Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "2a1de214-5851-4ea0-b96a-de4374aa3789", "err": "get primary pxc pod: not found"} 2024-05-16T03:22:23.697Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "b1eaa96d-60e0-4388-b87b-6e28b9289bce", "err": "get primary pxc pod: not found"} 2024-05-16T03:22:28.896Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "b141eece-74b5-4db6-b288-fae19adcd942", "err": "get primary pxc pod: not found"} 2024-05-16T03:22:34.131Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "ee75c8ab-d0e5-40eb-aff2-74e53b1666b4", "err": "get primary pxc pod: not found"} 2024-05-16T03:22:39.339Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "38916180-86ce-4a0d-887f-a9cd6b34a538", "err": "get primary pxc pod: not found"} 2024-05-16T03:24:52.570Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "efbd227e-461b-46d5-9a33-d3356f587c15", "new version": "8.0.36-28.1"} 2024-05-16T03:24:55.720Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "efbd227e-461b-46d5-9a33-d3356f587c15"} 2024-05-16T03:25:00.558Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "12a93286-96c7-418d-91f7-2f63a5781d42"} 2024-05-16T03:25:06.129Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "ccd789c5-b02c-4db5-ab61-9f152e777088"} 2024-05-16T03:25:10.238Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "365e84d5-6abe-422d-8bcc-174d433f0e4a", "user": "root"} 2024-05-16T03:25:10.301Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "365e84d5-6abe-422d-8bcc-174d433f0e4a", "user": "operator"} 2024-05-16T03:25:10.356Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "365e84d5-6abe-422d-8bcc-174d433f0e4a", "user": "monitor"} 2024-05-16T03:25:10.407Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "365e84d5-6abe-422d-8bcc-174d433f0e4a", "user": "xtrabackup"} 2024-05-16T03:25:10.460Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "365e84d5-6abe-422d-8bcc-174d433f0e4a", "user": "replication"} 2024-05-16T03:25:11.344Z ERROR sync users {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "ac9ca4ef-ba23-497e-a8b9-88d728d14730", "error": "exec syncusers: command terminated with exit code 137 / \nSyncing user accounts from PXC(some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local:3306) to ProxySQL\n / ", "errorVerbose": "exec syncusers: command terminated with exit code 137 / \nSyncing user accounts from PXC(some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local:3306) to ProxySQL\n / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:920\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1246\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1695"} 2024-05-16T03:26:12.293Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "9dc9bbb0-744e-4361-99c9-16adc9d9d6ea", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:26:49.465Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "37cfc1f2-6b64-4532-ae83-cdb82df1391a", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:26:49.911Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "5ef91d85-e25b-4c01-925f-a0d53f96deed", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:26:54.809Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "343a7214-3093-4dfa-b7fc-ae28320a2b01", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:27:00.107Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "aece079e-9b38-416e-9bbd-0107e9514992", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:27:00.383Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "3f56db3a-2b19-4e59-880e-09d76dfce942", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:27:01.718Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "4b318bce-1597-4e90-b425-0ed69588979c", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:27:05.397Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "f643bec4-813d-46af-9088-f1b6ea6f4557", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:27:10.572Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "5f906cb1-2a6e-401a-b7d5-95db12cb2f38", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:27:15.777Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "aa9d7f10-e4d2-4d6c-9497-6634ac45e33a", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:27:20.976Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "20ed3b43-4fe1-44d9-b100-1c916ab3be80", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:27:26.143Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "b84b9f9f-421c-4e1d-a287-74afdaa058f5", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:27:31.570Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "34fd75d7-26da-4acb-a3c0-2c28abd2fc49", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:27:36.780Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "5cd0796f-e155-4764-a966-ee80691ff6cb", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:27:51.179Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "9da5dc3b-abec-4532-b3fa-6da75211a8f0"} 2024-05-16T03:27:55.986Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "81d7738e-53ab-4655-ab81-9da56fac8406"} 2024-05-16T03:28:01.394Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "6bda902b-1522-424a-a046-9be608d9ce4b"} 2024-05-16T03:28:06.839Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "bfdfc1e1-d91a-4dd6-ab72-c1eb71ea44ab"} 2024-05-16T03:28:12.472Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "5b1bb288-4c09-41dd-98c9-a9130a1a2519"} 2024-05-16T03:28:17.444Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "4f4ffbcd-7d70-4b58-80fe-aab6fea72b78"} 2024-05-16T03:28:18.463Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "3b3edbbd-af99-42c6-a74f-5d8725d60b63", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-2: dial tcp: lookup some-name-pxc-2.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:28:19.450Z ERROR sync users {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "1afbdaf6-f2b7-4c4c-be32-3ec4fca2578c", "error": "exec syncusers: unable to upgrade connection: container not found (\"proxysql\") / / ", "errorVerbose": "exec syncusers: unable to upgrade connection: container not found (\"proxysql\") / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:920\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1246\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1695"} 2024-05-16T03:29:09.387Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "f68c4f99-89c1-4276-ab6e-6e6450668b97", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:29:14.402Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "f97fa6f4-ddb1-46fb-b57e-0f4aaefcd46c", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:29:19.864Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "677967cd-076d-4c19-ab6c-38cb7c5887ea", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:30:07.140Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "3da397d3-6710-498d-9dbf-c1a461e99b9f", "err": "failed to connect to pod some-name-pxc-0: dial tcp 10.196.192.57:33062: connect: connection refused"} 2024-05-16T03:30:12.343Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "040e5cc4-3582-48d2-bbae-92ac8b26e7c6", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:30:17.523Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "90d0ed03-4a29-4625-8073-0bb1581af133", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:30:22.852Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "55c180fd-0081-45f0-bf5e-572cc889b64a", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:30:28.055Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "0b81fe59-6bf1-4ded-80ff-64ed96a91936", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:30:33.275Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "6229a3aa-2bc9-40cf-bc26-54ed728e35e3", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:30:38.458Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "9ee9797f-673c-413c-8f06-b5cbf3254d96", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:30:47.280Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "c6b17b39-5427-4cb4-b9a6-57d80ab969e2"} 2024-05-16T03:30:52.748Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "a1254019-e652-4bb1-93dd-d69ea36b4eed"} 2024-05-16T03:30:57.848Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "0f2b6fed-03af-447a-b3ae-36eba80a3d72"} 2024-05-16T03:31:03.266Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "d80d531e-03e6-468c-bdab-5a93fec7aec2"} 2024-05-16T03:31:06.965Z ERROR sync users {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "eb2bcd2c-c9e9-4850-9a21-7b7997addd05", "error": "exec syncusers: unable to upgrade connection: container not found (\"proxysql\") / / ", "errorVerbose": "exec syncusers: unable to upgrade connection: container not found (\"proxysql\") / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:920\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1246\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1695"} 2024-05-16T03:31:59.138Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "fde6db85-335d-48a6-8483-45ee10140801", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:32:04.087Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "343f4949-2108-44c3-a5e2-df3d09ae288f", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-15810 on 10.163.128.10:53: no such host"} 2024-05-16T03:32:57.588Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "edfd335e-3967-4dbe-95f5-57cadae937ab", "err": "failed to connect to pod some-name-pxc-0: dial tcp 10.196.192.59:33062: connect: connection refused"} 2024-05-16T03:33:03.007Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "9dea2ea4-ba67-4cc1-a96f-251b64cea35a", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:33:08.189Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "b919285f-3c92-4a83-9447-b95f595ead01", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:33:13.358Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "b4102035-513d-449d-8386-16623d45d347", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-15810.svc.cluster.local"} 2024-05-16T03:33:49.044Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "ad762943-8ba0-48ee-808f-6721c0950a31"} 2024-05-16T03:33:53.766Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "20102ea6-0c01-4cc3-9548-5a9d63c9b10d"} 2024-05-16T03:33:59.157Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "19ba7988-bbd1-4add-84a2-3903c5031173"} 2024-05-16T03:34:05.063Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-15810", "name": "some-name", "reconcileID": "92178e33-a033-4842-b462-7cfd253725b2"} github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.18.0/pkg/internal/controller/controller.go:222 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.18.0/pkg/internal/controller/controller.go:261 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.18.0/pkg/internal/controller/controller.go:324 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1248 [mysql] 2024/05/16 03:29:19 packets.go:37: read tcp 10.196.192.53:52688->10.196.194.20:33062: read: connection reset by peer sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2 + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n upgrade-consistency-15810 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.BzTJuhnYOQ ++ mktemp + local LAST_ERR=/tmp/tmp.dHV49V6w0E + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.BzTJuhnYOQ perconaxtradbcluster.pxc.percona.com "some-name" deleted + cat /tmp/tmp.dHV49V6w0E + rm /tmp/tmp.BzTJuhnYOQ /tmp/tmp.dHV49V6w0E + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.AleHNQNGHp ++ mktemp + local LAST_ERR=/tmp/tmp.iTxTAGHYaA + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.AleHNQNGHp No resources found + cat /tmp/tmp.iTxTAGHYaA + rm /tmp/tmp.AleHNQNGHp /tmp/tmp.iTxTAGHYaA + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.FgrSbxdB9A ++ mktemp + local LAST_ERR=/tmp/tmp.uWVhNzzAkF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FgrSbxdB9A No resources found + cat /tmp/tmp.uWVhNzzAkF + rm /tmp/tmp.FgrSbxdB9A /tmp/tmp.uWVhNzzAkF + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.o3J7xDWQ5S ++ mktemp + local LAST_ERR=/tmp/tmp.XOzx1GRwOR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.o3J7xDWQ5S validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.XOzx1GRwOR + rm /tmp/tmp.o3J7xDWQ5S /tmp/tmp.XOzx1GRwOR + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.14.2/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace upgrade-consistency-15810 + rm -rf /tmp/tmp.Vx5fyn6AHh ++ mktemp + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.V6bZ3WC4Cg + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- ++ mktemp + local LAST_OUT=/tmp/tmp.oRyEQaqYCu ++ mktemp + local LAST_ERR=/tmp/tmp.BzKB8RKTx8 + local exit_status=0 + local LAST_ERR=/tmp/tmp.ttbv6fvoi3 + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace upgrade-consistency-15810 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator