Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/logs/upgrade-consistency-8-0.log WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.30) and server (1.26) exceeds the supported minor version skew of +/-1 + cluster=some-name + main + create_infra upgrade-consistency-26343 + local ns=upgrade-consistency-26343 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n upgrade-consistency-8254 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.94Eg6KPi1b ++ mktemp + local LAST_ERR=/tmp/tmp.RpQgyyBQBx + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.94Eg6KPi1b perconaxtradbcluster.pxc.percona.com "some-name" deleted + cat /tmp/tmp.RpQgyyBQBx + rm /tmp/tmp.94Eg6KPi1b /tmp/tmp.RpQgyyBQBx + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.uggBAEEDKa ++ mktemp + local LAST_ERR=/tmp/tmp.c5ksDPcBhR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uggBAEEDKa No resources found + cat /tmp/tmp.c5ksDPcBhR + rm /tmp/tmp.uggBAEEDKa /tmp/tmp.c5ksDPcBhR + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.oi7EdREXmZ ++ mktemp + local LAST_ERR=/tmp/tmp.TkLFUbpiUb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oi7EdREXmZ No resources found + cat /tmp/tmp.TkLFUbpiUb + rm /tmp/tmp.oi7EdREXmZ /tmp/tmp.TkLFUbpiUb + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ grep chaos-mesh.org ++ awk '{print $1}' ++ kubectl get crd + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.6wAhJ3nlbk ++ mktemp + local LAST_ERR=/tmp/tmp.Ypw14O9PwF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + awk '{print$1}' + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.EvWom8FTXm ++ mktemp + local LAST_ERR=/tmp/tmp.U4EzKl3fyi + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6wAhJ3nlbk + cat /tmp/tmp.Ypw14O9PwF + rm /tmp/tmp.6wAhJ3nlbk /tmp/tmp.Ypw14O9PwF + return 0 namespace "upgrade-consistency-8254" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EvWom8FTXm namespace "pxc-operator" deleted + cat /tmp/tmp.U4EzKl3fyi + rm /tmp/tmp.EvWom8FTXm /tmp/tmp.U4EzKl3fyi + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'namespace/pxc-operator - ' namespace/pxc-operator - + set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.fZR9D4mVmU ++ mktemp + local LAST_ERR=/tmp/tmp.dq34j8i6IB + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.fZR9D4mVmU namespace/pxc-operator created + cat /tmp/tmp.dq34j8i6IB + rm /tmp/tmp.fZR9D4mVmU /tmp/tmp.dq34j8i6IB + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.ReFocMaO65 +++ mktemp ++ local LAST_ERR=/tmp/tmp.S9v81aTzWd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ReFocMaO65 ++ cat /tmp/tmp.S9v81aTzWd ++ rm /tmp/tmp.ReFocMaO65 /tmp/tmp.S9v81aTzWd ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1739-a91d0247-1-cluster8 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.ADSkTUE5jC ++ mktemp + local LAST_ERR=/tmp/tmp.JAdSF4CiUq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1739-a91d0247-1-cluster8 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ADSkTUE5jC Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1739-a91d0247-1-cluster8" modified. + cat /tmp/tmp.JAdSF4CiUq + rm /tmp/tmp.ADSkTUE5jC /tmp/tmp.JAdSF4CiUq + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.eUIbTFDbfy ++ mktemp + local LAST_ERR=/tmp/tmp.4qZLtv6O6H + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.eUIbTFDbfy customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.4qZLtv6O6H + rm /tmp/tmp.eUIbTFDbfy /tmp/tmp.4qZLtv6O6H + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.JBG8FoL3m8 ++ mktemp + local LAST_ERR=/tmp/tmp.YFqEZJsSXt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JBG8FoL3m8 clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.YFqEZJsSXt + rm /tmp/tmp.JBG8FoL3m8 /tmp/tmp.YFqEZJsSXt + return 0 + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-1739-a91d0247^' + kubectl_bin apply -f - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "DEBUG"' - ++ mktemp + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/deploy/cw-operator.yaml + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + local LAST_OUT=/tmp/tmp.vKyGrqr3ff ++ mktemp + local LAST_ERR=/tmp/tmp.K1Uz0bndoS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vKyGrqr3ff deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.K1Uz0bndoS + rm /tmp/tmp.vKyGrqr3ff /tmp/tmp.K1Uz0bndoS + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.p3oNET3tmb ++ mktemp + local LAST_ERR=/tmp/tmp.g4yT802qRO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.p3oNET3tmb pod/percona-xtradb-cluster-operator-86d767f4fd-2ccjm condition met + cat /tmp/tmp.g4yT802qRO + rm /tmp/tmp.p3oNET3tmb /tmp/tmp.g4yT802qRO + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.L3NdbJyL4d +++ mktemp ++ local LAST_ERR=/tmp/tmp.1LQcM0B8F1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.L3NdbJyL4d ++ cat /tmp/tmp.1LQcM0B8F1 ++ rm /tmp/tmp.L3NdbJyL4d /tmp/tmp.1LQcM0B8F1 ++ return 0 + wait_pod percona-xtradb-cluster-operator-86d767f4fd-2ccjm 480 pxc-operator + local pod=percona-xtradb-cluster-operator-86d767f4fd-2ccjm + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-86d767f4fd-2ccjm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-86d767f4fd-2ccjm condition met percona-xtradb-cluster-operator-86d767f4fd-2ccjm.Ok + sleep 3 + create_namespace upgrade-consistency-26343 + local namespace=upgrade-consistency-26343 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ helm list --all-namespaces --filter chaos-mesh ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl api-resources ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' ++ mktemp + '[' -n '' ']' + desc 'cleaned up old namespaces upgrade-consistency-26343' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces upgrade-consistency-26343 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace upgrade-consistency-26343 + awk '{print$1}' ++ mktemp + local LAST_OUT=/tmp/tmp.rpFjqfjUW4 ++ mktemp + local LAST_OUT=/tmp/tmp.BlQjwE2Go0 ++ mktemp + local LAST_ERR=/tmp/tmp.1oT1iL3KF6 + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.fdcLutZ0Ip + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-consistency-26343 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-consistency-26343 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rpFjqfjUW4 + cat /tmp/tmp.1oT1iL3KF6 + rm /tmp/tmp.rpFjqfjUW4 /tmp/tmp.1oT1iL3KF6 + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + for i in '$(seq 0 2)' + set +e + kubectl delete namespace upgrade-consistency-26343 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.BlQjwE2Go0 + cat /tmp/tmp.fdcLutZ0Ip Error from server (NotFound): namespaces "upgrade-consistency-26343" not found + rm /tmp/tmp.BlQjwE2Go0 /tmp/tmp.fdcLutZ0Ip + return 1 + : + wait_for_delete namespace/upgrade-consistency-26343 + local res=namespace/upgrade-consistency-26343 + echo -n 'namespace/upgrade-consistency-26343 - ' namespace/upgrade-consistency-26343 - + set +o xtrace Error from server (NotFound): namespaces "upgrade-consistency-26343" not found + desc 'create namespace upgrade-consistency-26343' + set +o xtrace ----------------------------------------------------------------------------------- create namespace upgrade-consistency-26343 ----------------------------------------------------------------------------------- + kubectl_bin create namespace upgrade-consistency-26343 ++ mktemp + local LAST_OUT=/tmp/tmp.4EiYBVrFng ++ mktemp + local LAST_ERR=/tmp/tmp.fAJEYSdovr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace upgrade-consistency-26343 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4EiYBVrFng namespace/upgrade-consistency-26343 created + cat /tmp/tmp.fAJEYSdovr + rm /tmp/tmp.4EiYBVrFng /tmp/tmp.fAJEYSdovr + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.kpMgpUCH4a +++ mktemp ++ local LAST_ERR=/tmp/tmp.RQ0LWQV6Lw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kpMgpUCH4a ++ cat /tmp/tmp.RQ0LWQV6Lw ++ rm /tmp/tmp.kpMgpUCH4a /tmp/tmp.RQ0LWQV6Lw ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1739-a91d0247-1-cluster8 --namespace=upgrade-consistency-26343 ++ mktemp + local LAST_OUT=/tmp/tmp.VNvsOE0bhx ++ mktemp + local LAST_ERR=/tmp/tmp.dqONbPDIwL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1739-a91d0247-1-cluster8 --namespace=upgrade-consistency-26343 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VNvsOE0bhx Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1739-a91d0247-1-cluster8" modified. + cat /tmp/tmp.dqONbPDIwL + rm /tmp/tmp.VNvsOE0bhx /tmp/tmp.dqONbPDIwL + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.0L4NNV1AED ++ mktemp + local LAST_ERR=/tmp/tmp.KgwDXwgO8y + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0L4NNV1AED secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.KgwDXwgO8y + rm /tmp/tmp.0L4NNV1AED /tmp/tmp.KgwDXwgO8y + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/conf/client.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/conf/client.yml + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/conf/client.yml ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + local LAST_OUT=/tmp/tmp.Xu2H6oFGJT + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1739-a91d0247#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' ++ mktemp + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + local LAST_ERR=/tmp/tmp.66vZS07dIe + local exit_status=0 + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' ++ seq 0 2 + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.upgrade-consistency-26343~ + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Xu2H6oFGJT deployment.apps/pxc-client created + cat /tmp/tmp.66vZS07dIe + rm /tmp/tmp.Xu2H6oFGJT /tmp/tmp.66vZS07dIe + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.1TOi9f4Kqs ++ mktemp + local LAST_ERR=/tmp/tmp.FiyPl8sBfC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.1TOi9f4Kqs secret/my-cluster-secrets created + cat /tmp/tmp.FiyPl8sBfC + rm /tmp/tmp.1TOi9f4Kqs /tmp/tmp.FiyPl8sBfC + return 0 + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/conf/some-name.yml + yq eval '.spec.crVersion = "1.12.0"' - + kubectl_bin apply -f - + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' ++ mktemp + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.upgrade-consistency-26343~ + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1739-a91d0247#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/conf/some-name.yml + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + local LAST_OUT=/tmp/tmp.W5SrTK3FTg ++ mktemp + local LAST_ERR=/tmp/tmp.kXjcYKgbiV + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.W5SrTK3FTg perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.kXjcYKgbiV + rm /tmp/tmp.W5SrTK3FTg /tmp/tmp.kXjcYKgbiV + return 0 + desc 'test 1.12.0' + set +o xtrace ----------------------------------------------------------------------------------- test 1.12.0 ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.12.0"} }' ++ mktemp + local LAST_OUT=/tmp/tmp.sv2hOjjbiz ++ mktemp + local LAST_ERR=/tmp/tmp.CG4RzcrWoE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.12.0"} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sv2hOjjbiz perconaxtradbcluster.pxc.percona.com/some-name patched (no change) + cat /tmp/tmp.CG4RzcrWoE + rm /tmp/tmp.sv2hOjjbiz /tmp/tmp.CG4RzcrWoE + return 0 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.iCCb9DGuuS +++ mktemp ++ local LAST_ERR=/tmp/tmp.EPbzGCH5IZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.iCCb9DGuuS ++ cat /tmp/tmp.EPbzGCH5IZ ++ rm /tmp/tmp.iCCb9DGuuS /tmp/tmp.EPbzGCH5IZ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cPQek4wmtM +++ mktemp ++ local LAST_ERR=/tmp/tmp.1UNpCn2MGH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cPQek4wmtM ++ cat /tmp/tmp.1UNpCn2MGH ++ rm /tmp/tmp.cPQek4wmtM /tmp/tmp.1UNpCn2MGH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OF23wE2hUA +++ mktemp ++ local LAST_ERR=/tmp/tmp.FtIydWYGYw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OF23wE2hUA ++ cat /tmp/tmp.FtIydWYGYw ++ rm /tmp/tmp.OF23wE2hUA /tmp/tmp.FtIydWYGYw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.V91V6Arev8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.inPKpcBxNj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.V91V6Arev8 ++ cat /tmp/tmp.inPKpcBxNj ++ rm /tmp/tmp.V91V6Arev8 /tmp/tmp.inPKpcBxNj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6tZJuYRgli +++ mktemp ++ local LAST_ERR=/tmp/tmp.snBGBJl9bW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6tZJuYRgli ++ cat /tmp/tmp.snBGBJl9bW ++ rm /tmp/tmp.6tZJuYRgli /tmp/tmp.snBGBJl9bW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pRe6Z97w6Z +++ mktemp ++ local LAST_ERR=/tmp/tmp.XpHaCXcmqG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pRe6Z97w6Z ++ cat /tmp/tmp.XpHaCXcmqG ++ rm /tmp/tmp.pRe6Z97w6Z /tmp/tmp.XpHaCXcmqG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dV8MQ8BXob +++ mktemp ++ local LAST_ERR=/tmp/tmp.LuCOtJHB6j ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dV8MQ8BXob ++ cat /tmp/tmp.LuCOtJHB6j ++ rm /tmp/tmp.dV8MQ8BXob /tmp/tmp.LuCOtJHB6j ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 6 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ofUOQ0EXXM +++ mktemp ++ local LAST_ERR=/tmp/tmp.O1xTKOOqOz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ofUOQ0EXXM ++ cat /tmp/tmp.O1xTKOOqOz ++ rm /tmp/tmp.ofUOQ0EXXM /tmp/tmp.O1xTKOOqOz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 7 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wMxNsDX7je +++ mktemp ++ local LAST_ERR=/tmp/tmp.zEiPTleR6l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wMxNsDX7je ++ cat /tmp/tmp.zEiPTleR6l ++ rm /tmp/tmp.wMxNsDX7je /tmp/tmp.zEiPTleR6l ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 8 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.SuXmWutckr +++ mktemp ++ local LAST_ERR=/tmp/tmp.tTtexfDxwz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.SuXmWutckr ++ cat /tmp/tmp.tTtexfDxwz ++ rm /tmp/tmp.SuXmWutckr /tmp/tmp.tTtexfDxwz ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 9 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.3V7RHuEw20 +++ mktemp ++ local LAST_ERR=/tmp/tmp.bC3qIWIqgH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.3V7RHuEw20 ++ cat /tmp/tmp.bC3qIWIqgH ++ rm /tmp/tmp.3V7RHuEw20 /tmp/tmp.bC3qIWIqgH ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 10 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MgRCioXi7A +++ mktemp ++ local LAST_ERR=/tmp/tmp.1b0FyWG4uk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MgRCioXi7A ++ cat /tmp/tmp.1b0FyWG4uk ++ rm /tmp/tmp.MgRCioXi7A /tmp/tmp.1b0FyWG4uk ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kNNcrRuBXt +++ mktemp ++ local LAST_ERR=/tmp/tmp.KCgRJrgv8R ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kNNcrRuBXt ++ cat /tmp/tmp.KCgRJrgv8R ++ rm /tmp/tmp.kNNcrRuBXt /tmp/tmp.KCgRJrgv8R ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.4PkuppZbiQ ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.MlcmgHedqf +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.4PkuppZbiQ +++++ cat /tmp/tmp.MlcmgHedqf +++++ rm /tmp/tmp.4PkuppZbiQ /tmp/tmp.MlcmgHedqf +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.zpy7Jmfaip ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.LW44a90W7t +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.zpy7Jmfaip +++++ cat /tmp/tmp.LW44a90W7t +++++ rm /tmp/tmp.zpy7Jmfaip /tmp/tmp.LW44a90W7t +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9AccwNRaVR +++ mktemp ++ local LAST_ERR=/tmp/tmp.dF6KTdN9sR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9AccwNRaVR ++ cat /tmp/tmp.dF6KTdN9sR ++ rm /tmp/tmp.9AccwNRaVR /tmp/tmp.dF6KTdN9sR ++ return 0 + [[ 2 == \2 ]] + wait_for_sts_generation some-name-pxc 1 1 + local name=some-name-pxc + local desired_generation=1 + local size=1 + set +o xtrace Waiting for statefulset readyness + desc 'compare k8s objects' + set +o xtrace ----------------------------------------------------------------------------------- compare k8s objects ----------------------------------------------------------------------------------- + compare_kubectl service/some-name-pxc -1120 + local resource=service/some-name-pxc + local postfix=-1120 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120.yml + local new_result=/tmp/tmp.iLd3TvQh90/service_some-name-pxc.yml + desc 'compare service/some-name-pxc--1120' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc--1120 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120-eks.yml ']' + kubectl_bin get -o yaml service/some-name-pxc + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.hkEY8Oz1Vz ++ mktemp + local LAST_ERR=/tmp/tmp.YYnSRUheyJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hkEY8Oz1Vz + cat /tmp/tmp.YYnSRUheyJ + rm /tmp/tmp.hkEY8Oz1Vz /tmp/tmp.YYnSRUheyJ + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1120.yml /tmp/tmp.iLd3TvQh90/service_some-name-pxc.yml + compare_kubectl service/some-name-proxysql -1120 + local resource=service/some-name-proxysql + local postfix=-1120 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120.yml + local new_result=/tmp/tmp.iLd3TvQh90/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql--1120' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql--1120 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120-eks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.OdI2h0S1Go ++ mktemp + local LAST_ERR=/tmp/tmp.qUJlGmUK30 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.OdI2h0S1Go + cat /tmp/tmp.qUJlGmUK30 + rm /tmp/tmp.OdI2h0S1Go /tmp/tmp.qUJlGmUK30 + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1120.yml /tmp/tmp.iLd3TvQh90/service_some-name-proxysql.yml + compare_kubectl statefulset/some-name-pxc -1120 + local resource=statefulset/some-name-pxc + local postfix=-1120 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120.yml + local new_result=/tmp/tmp.iLd3TvQh90/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc--1120' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc--1120 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.SoUEGwxmxI ++ mktemp + local LAST_ERR=/tmp/tmp.AU8Ody7hjv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.SoUEGwxmxI + cat /tmp/tmp.AU8Ody7hjv + rm /tmp/tmp.SoUEGwxmxI /tmp/tmp.AU8Ody7hjv + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1120.yml /tmp/tmp.iLd3TvQh90/statefulset_some-name-pxc.yml + compare_kubectl statefulset/some-name-proxysql -1120 + local resource=statefulset/some-name-proxysql + local postfix=-1120 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120.yml + local new_result=/tmp/tmp.iLd3TvQh90/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql--1120' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql--1120 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.nVinvMZECM ++ mktemp + local LAST_ERR=/tmp/tmp.mBvkipuyXN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.nVinvMZECM + cat /tmp/tmp.mBvkipuyXN + rm /tmp/tmp.nVinvMZECM /tmp/tmp.mBvkipuyXN + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1120.yml /tmp/tmp.iLd3TvQh90/statefulset_some-name-proxysql.yml + desc 'test 1.13.0' + set +o xtrace ----------------------------------------------------------------------------------- test 1.13.0 ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.13.0"} }' ++ mktemp + local LAST_OUT=/tmp/tmp.2XHvYPIEEI ++ mktemp + local LAST_ERR=/tmp/tmp.4IdngSwwRa + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.13.0"} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.2XHvYPIEEI perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.4IdngSwwRa + rm /tmp/tmp.2XHvYPIEEI /tmp/tmp.4IdngSwwRa + return 0 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ELGcyfLsp7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.LsvKzm9rfR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ELGcyfLsp7 ++ cat /tmp/tmp.LsvKzm9rfR ++ rm /tmp/tmp.ELGcyfLsp7 /tmp/tmp.LsvKzm9rfR ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.lIlGO1G0jC +++ mktemp ++ local LAST_ERR=/tmp/tmp.fxVA8WHHfD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.lIlGO1G0jC ++ cat /tmp/tmp.fxVA8WHHfD ++ rm /tmp/tmp.lIlGO1G0jC /tmp/tmp.fxVA8WHHfD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.y8pVuMCypX +++ mktemp ++ local LAST_ERR=/tmp/tmp.54AGXJooSY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.y8pVuMCypX ++ cat /tmp/tmp.54AGXJooSY ++ rm /tmp/tmp.y8pVuMCypX /tmp/tmp.54AGXJooSY ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PE6GVAsEdg +++ mktemp ++ local LAST_ERR=/tmp/tmp.64fmTQKGi4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PE6GVAsEdg ++ cat /tmp/tmp.64fmTQKGi4 ++ rm /tmp/tmp.PE6GVAsEdg /tmp/tmp.64fmTQKGi4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.L7buemVtoK +++ mktemp ++ local LAST_ERR=/tmp/tmp.k8bSMRLATC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.L7buemVtoK ++ cat /tmp/tmp.k8bSMRLATC ++ rm /tmp/tmp.L7buemVtoK /tmp/tmp.k8bSMRLATC ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TypAuFZqKC +++ mktemp ++ local LAST_ERR=/tmp/tmp.opELtv5AWf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TypAuFZqKC ++ cat /tmp/tmp.opELtv5AWf ++ rm /tmp/tmp.TypAuFZqKC /tmp/tmp.opELtv5AWf ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Kud8VKaPqu +++ mktemp ++ local LAST_ERR=/tmp/tmp.5EF67w5d0k ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Kud8VKaPqu ++ cat /tmp/tmp.5EF67w5d0k ++ rm /tmp/tmp.Kud8VKaPqu /tmp/tmp.5EF67w5d0k ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 6 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2zLzhUYoVM +++ mktemp ++ local LAST_ERR=/tmp/tmp.0i3h6nMAot ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2zLzhUYoVM ++ cat /tmp/tmp.0i3h6nMAot ++ rm /tmp/tmp.2zLzhUYoVM /tmp/tmp.0i3h6nMAot ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nrZ5y4gKsU +++ mktemp ++ local LAST_ERR=/tmp/tmp.um1kEXPSuV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nrZ5y4gKsU ++ cat /tmp/tmp.um1kEXPSuV ++ rm /tmp/tmp.nrZ5y4gKsU /tmp/tmp.um1kEXPSuV ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.sZ2RohcAOq ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.q01PkeH7z9 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.sZ2RohcAOq +++++ cat /tmp/tmp.q01PkeH7z9 +++++ rm /tmp/tmp.sZ2RohcAOq /tmp/tmp.q01PkeH7z9 +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.qy0Buk9rFq ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.5lZ1hOxL9z +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.qy0Buk9rFq +++++ cat /tmp/tmp.5lZ1hOxL9z +++++ rm /tmp/tmp.qy0Buk9rFq /tmp/tmp.5lZ1hOxL9z +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pmp4MfOfQz +++ mktemp ++ local LAST_ERR=/tmp/tmp.yeBpUqGgjU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pmp4MfOfQz ++ cat /tmp/tmp.yeBpUqGgjU ++ rm /tmp/tmp.pmp4MfOfQz /tmp/tmp.yeBpUqGgjU ++ return 0 + [[ 2 == \2 ]] + wait_for_sts_generation some-name-pxc 2 1 + local name=some-name-pxc + local desired_generation=2 + local size=1 + set +o xtrace Waiting for statefulset readyness + desc 'compare k8s objects' + set +o xtrace ----------------------------------------------------------------------------------- compare k8s objects ----------------------------------------------------------------------------------- + compare_kubectl service/some-name-pxc -1130 + local resource=service/some-name-pxc + local postfix=-1130 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130.yml + local new_result=/tmp/tmp.iLd3TvQh90/service_some-name-pxc.yml + desc 'compare service/some-name-pxc--1130' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc--1130 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130-eks.yml ']' + kubectl_bin get -o yaml service/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.wxizp6XnVL ++ mktemp + local LAST_ERR=/tmp/tmp.nwmBr0v4og + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.wxizp6XnVL + cat /tmp/tmp.nwmBr0v4og + rm /tmp/tmp.wxizp6XnVL /tmp/tmp.nwmBr0v4og + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1130.yml /tmp/tmp.iLd3TvQh90/service_some-name-pxc.yml + compare_kubectl service/some-name-proxysql -1130 + local resource=service/some-name-proxysql + local postfix=-1130 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130.yml + local new_result=/tmp/tmp.iLd3TvQh90/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql--1130' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql--1130 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130-eks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.Va2hpdHX2U ++ mktemp + local LAST_ERR=/tmp/tmp.lzhSBy61uy + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Va2hpdHX2U + cat /tmp/tmp.lzhSBy61uy + rm /tmp/tmp.Va2hpdHX2U /tmp/tmp.lzhSBy61uy + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1130.yml /tmp/tmp.iLd3TvQh90/service_some-name-proxysql.yml + compare_kubectl statefulset/some-name-pxc -1130 + local resource=statefulset/some-name-pxc + local postfix=-1130 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130.yml + local new_result=/tmp/tmp.iLd3TvQh90/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc--1130' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc--1130 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.bselbMOLAv ++ mktemp + local LAST_ERR=/tmp/tmp.DRpzyZKic5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.bselbMOLAv + cat /tmp/tmp.DRpzyZKic5 + rm /tmp/tmp.bselbMOLAv /tmp/tmp.DRpzyZKic5 + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1130.yml /tmp/tmp.iLd3TvQh90/statefulset_some-name-pxc.yml + compare_kubectl statefulset/some-name-proxysql -1130 + local resource=statefulset/some-name-proxysql + local postfix=-1130 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130.yml + local new_result=/tmp/tmp.iLd3TvQh90/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql--1130' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql--1130 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.iu2GBghHqo ++ mktemp + local LAST_ERR=/tmp/tmp.1nODuVXknu + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.iu2GBghHqo + cat /tmp/tmp.1nODuVXknu + rm /tmp/tmp.iu2GBghHqo /tmp/tmp.1nODuVXknu + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1130.yml /tmp/tmp.iLd3TvQh90/statefulset_some-name-proxysql.yml + desc 'test 1.14.0' + set +o xtrace ----------------------------------------------------------------------------------- test 1.14.0 ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.14.0"} }' ++ mktemp + local LAST_OUT=/tmp/tmp.PJzSw6lUp9 ++ mktemp + local LAST_ERR=/tmp/tmp.GK4vd7IsJq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.14.0"} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.PJzSw6lUp9 perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.GK4vd7IsJq + rm /tmp/tmp.PJzSw6lUp9 /tmp/tmp.GK4vd7IsJq + return 0 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.c2zc7jNeQM +++ mktemp ++ local LAST_ERR=/tmp/tmp.4MHbgwbCCK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.c2zc7jNeQM ++ cat /tmp/tmp.4MHbgwbCCK ++ rm /tmp/tmp.c2zc7jNeQM /tmp/tmp.4MHbgwbCCK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7wcPE2hMmY +++ mktemp ++ local LAST_ERR=/tmp/tmp.tgMsIDS7HB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7wcPE2hMmY ++ cat /tmp/tmp.tgMsIDS7HB ++ rm /tmp/tmp.7wcPE2hMmY /tmp/tmp.tgMsIDS7HB ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jMOx6Av82K +++ mktemp ++ local LAST_ERR=/tmp/tmp.n47uEzWR7L ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jMOx6Av82K ++ cat /tmp/tmp.n47uEzWR7L ++ rm /tmp/tmp.jMOx6Av82K /tmp/tmp.n47uEzWR7L ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6gfiVXDufP +++ mktemp ++ local LAST_ERR=/tmp/tmp.LeYx25JimM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6gfiVXDufP ++ cat /tmp/tmp.LeYx25JimM ++ rm /tmp/tmp.6gfiVXDufP /tmp/tmp.LeYx25JimM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yEwrbKltMT +++ mktemp ++ local LAST_ERR=/tmp/tmp.2mKgPoT9SL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yEwrbKltMT ++ cat /tmp/tmp.2mKgPoT9SL ++ rm /tmp/tmp.yEwrbKltMT /tmp/tmp.2mKgPoT9SL ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1xsK7gjP1e +++ mktemp ++ local LAST_ERR=/tmp/tmp.shUvu4ffVj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1xsK7gjP1e ++ cat /tmp/tmp.shUvu4ffVj ++ rm /tmp/tmp.1xsK7gjP1e /tmp/tmp.shUvu4ffVj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cVdeEov0U7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.QlwzOL4yFU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cVdeEov0U7 ++ cat /tmp/tmp.QlwzOL4yFU ++ rm /tmp/tmp.cVdeEov0U7 /tmp/tmp.QlwzOL4yFU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 6 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rJ9CPujz1K +++ mktemp ++ local LAST_ERR=/tmp/tmp.8HuAN9sK21 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rJ9CPujz1K ++ cat /tmp/tmp.8HuAN9sK21 ++ rm /tmp/tmp.rJ9CPujz1K /tmp/tmp.8HuAN9sK21 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pibDTLkiGH +++ mktemp ++ local LAST_ERR=/tmp/tmp.EgX1fDlvL3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pibDTLkiGH ++ cat /tmp/tmp.EgX1fDlvL3 ++ rm /tmp/tmp.pibDTLkiGH /tmp/tmp.EgX1fDlvL3 ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.qjMWoaVpzI ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.TgyvpLOHPj +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.qjMWoaVpzI +++++ cat /tmp/tmp.TgyvpLOHPj +++++ rm /tmp/tmp.qjMWoaVpzI /tmp/tmp.TgyvpLOHPj +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.Umyntu4Xgc ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.WbOIALiQuv +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.Umyntu4Xgc +++++ cat /tmp/tmp.WbOIALiQuv +++++ rm /tmp/tmp.Umyntu4Xgc /tmp/tmp.WbOIALiQuv +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KUhIX3knia +++ mktemp ++ local LAST_ERR=/tmp/tmp.DYUF59MCpF ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KUhIX3knia ++ cat /tmp/tmp.DYUF59MCpF ++ rm /tmp/tmp.KUhIX3knia /tmp/tmp.DYUF59MCpF ++ return 0 + [[ 2 == \2 ]] + wait_for_sts_generation some-name-pxc 3 1 + local name=some-name-pxc + local desired_generation=3 + local size=1 + set +o xtrace Waiting for statefulset readyness + desc 'compare k8s objects' + set +o xtrace ----------------------------------------------------------------------------------- compare k8s objects ----------------------------------------------------------------------------------- + compare_kubectl service/some-name-pxc -1140 + local resource=service/some-name-pxc + local postfix=-1140 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140.yml + local new_result=/tmp/tmp.iLd3TvQh90/service_some-name-pxc.yml + desc 'compare service/some-name-pxc--1140' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc--1140 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140-eks.yml ']' + kubectl_bin get -o yaml service/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.Q8v3X0Pc5w ++ mktemp + local LAST_ERR=/tmp/tmp.S8COu6e1JK + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Q8v3X0Pc5w + cat /tmp/tmp.S8COu6e1JK + rm /tmp/tmp.Q8v3X0Pc5w /tmp/tmp.S8COu6e1JK + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1140.yml /tmp/tmp.iLd3TvQh90/service_some-name-pxc.yml + compare_kubectl service/some-name-proxysql -1140 + local resource=service/some-name-proxysql + local postfix=-1140 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140.yml + local new_result=/tmp/tmp.iLd3TvQh90/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql--1140' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql--1140 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140-eks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.fGEYoeQfFa ++ mktemp + local LAST_ERR=/tmp/tmp.TCHC9ZRgni + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.fGEYoeQfFa + cat /tmp/tmp.TCHC9ZRgni + rm /tmp/tmp.fGEYoeQfFa /tmp/tmp.TCHC9ZRgni + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1140.yml /tmp/tmp.iLd3TvQh90/service_some-name-proxysql.yml + compare_kubectl statefulset/some-name-pxc -1140 + local resource=statefulset/some-name-pxc + local postfix=-1140 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140.yml + local new_result=/tmp/tmp.iLd3TvQh90/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc--1140' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc--1140 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ bc -l ++ echo '1.26 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.JWZV6fQL10 ++ mktemp + local LAST_ERR=/tmp/tmp.28O4cuyn9Z + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JWZV6fQL10 + cat /tmp/tmp.28O4cuyn9Z + rm /tmp/tmp.JWZV6fQL10 /tmp/tmp.28O4cuyn9Z + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1140.yml /tmp/tmp.iLd3TvQh90/statefulset_some-name-pxc.yml + compare_kubectl statefulset/some-name-proxysql -1140 + local resource=statefulset/some-name-proxysql + local postfix=-1140 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140.yml + local new_result=/tmp/tmp.iLd3TvQh90/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql--1140' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql--1140 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.IYuJ1Fkrij ++ mktemp + local LAST_ERR=/tmp/tmp.01A7T9X2dJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.IYuJ1Fkrij + cat /tmp/tmp.01A7T9X2dJ + rm /tmp/tmp.IYuJ1Fkrij /tmp/tmp.01A7T9X2dJ + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1140.yml /tmp/tmp.iLd3TvQh90/statefulset_some-name-proxysql.yml + desc 'test 1.15.0' + set +o xtrace ----------------------------------------------------------------------------------- test 1.15.0 ----------------------------------------------------------------------------------- + kubectl_bin patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.15.0"} }' ++ mktemp + local LAST_OUT=/tmp/tmp.6gnUp4hkJa ++ mktemp + local LAST_ERR=/tmp/tmp.NjmFeKPfKJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl patch pxc some-name --type=merge --patch '{ "spec": {"crVersion":"1.15.0"} }' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6gnUp4hkJa perconaxtradbcluster.pxc.percona.com/some-name patched + cat /tmp/tmp.NjmFeKPfKJ + rm /tmp/tmp.6gnUp4hkJa /tmp/tmp.NjmFeKPfKJ + return 0 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=36 + sleep 7 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.8pMP8rwxYi +++ mktemp ++ local LAST_ERR=/tmp/tmp.xERmEVgOVq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.8pMP8rwxYi ++ cat /tmp/tmp.xERmEVgOVq ++ rm /tmp/tmp.8pMP8rwxYi /tmp/tmp.xERmEVgOVq ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 0 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.rhzMFjdHek +++ mktemp ++ local LAST_ERR=/tmp/tmp.IAdwjH6Exn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.rhzMFjdHek ++ cat /tmp/tmp.IAdwjH6Exn ++ rm /tmp/tmp.rhzMFjdHek /tmp/tmp.IAdwjH6Exn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 1 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oBvIPgbSuy +++ mktemp ++ local LAST_ERR=/tmp/tmp.Vra11BpK99 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oBvIPgbSuy ++ cat /tmp/tmp.Vra11BpK99 ++ rm /tmp/tmp.oBvIPgbSuy /tmp/tmp.Vra11BpK99 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 2 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.md3F6LlCuo +++ mktemp ++ local LAST_ERR=/tmp/tmp.2RDyaSiHtk ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.md3F6LlCuo ++ cat /tmp/tmp.2RDyaSiHtk ++ rm /tmp/tmp.md3F6LlCuo /tmp/tmp.2RDyaSiHtk ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 3 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bx9oGJ9Dcy +++ mktemp ++ local LAST_ERR=/tmp/tmp.YvmVF0Mquw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bx9oGJ9Dcy ++ cat /tmp/tmp.YvmVF0Mquw ++ rm /tmp/tmp.bx9oGJ9Dcy /tmp/tmp.YvmVF0Mquw ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 4 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.N5hEu39mVX +++ mktemp ++ local LAST_ERR=/tmp/tmp.KymEfSsvNa ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.N5hEu39mVX ++ cat /tmp/tmp.KymEfSsvNa ++ rm /tmp/tmp.N5hEu39mVX /tmp/tmp.KymEfSsvNa ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 5 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.DTjicNGmfy +++ mktemp ++ local LAST_ERR=/tmp/tmp.PQq2PTn6xj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.DTjicNGmfy ++ cat /tmp/tmp.PQq2PTn6xj ++ rm /tmp/tmp.DTjicNGmfy /tmp/tmp.PQq2PTn6xj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 6 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.saUcowgl8p +++ mktemp ++ local LAST_ERR=/tmp/tmp.GfCqagJFO0 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.saUcowgl8p ++ cat /tmp/tmp.GfCqagJFO0 ++ rm /tmp/tmp.saUcowgl8p /tmp/tmp.GfCqagJFO0 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo 'waiting for cluster readyness' waiting for cluster readyness + sleep 20 + [[ 7 -ge 36 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KTWiIWdiip +++ mktemp ++ local LAST_ERR=/tmp/tmp.lkxEQ9gXh1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KTWiIWdiip ++ cat /tmp/tmp.lkxEQ9gXh1 ++ rm /tmp/tmp.KTWiIWdiip /tmp/tmp.lkxEQ9gXh1 ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NtR3FJvUO0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.HmZlPSF2KH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NtR3FJvUO0 ++ cat /tmp/tmp.HmZlPSF2KH ++ rm /tmp/tmp.NtR3FJvUO0 /tmp/tmp.HmZlPSF2KH ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.284ZBr0Pgc ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.6MQOkiI2pI +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.284ZBr0Pgc +++++ cat /tmp/tmp.6MQOkiI2pI +++++ rm /tmp/tmp.284ZBr0Pgc /tmp/tmp.6MQOkiI2pI +++++ return 0 ++++ [[ '' == \t\r\u\e ]] +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.k4wMoAMoJX ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.RKX1scV1l7 +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.k4wMoAMoJX +++++ cat /tmp/tmp.RKX1scV1l7 +++++ rm /tmp/tmp.k4wMoAMoJX /tmp/tmp.RKX1scV1l7 +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-proxysql ++++ return +++ local cluster_proxy=some-name-proxysql +++ echo proxysql ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.proxysql.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hOQCuif5Ry +++ mktemp ++ local LAST_ERR=/tmp/tmp.UMs8eZmqhZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.proxysql.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hOQCuif5Ry ++ cat /tmp/tmp.UMs8eZmqhZ ++ rm /tmp/tmp.hOQCuif5Ry /tmp/tmp.UMs8eZmqhZ ++ return 0 + [[ 2 == \2 ]] + wait_for_sts_generation some-name-pxc 4 1 + local name=some-name-pxc + local desired_generation=4 + local size=1 + set +o xtrace Waiting for statefulset readyness + desc 'compare k8s objects' + set +o xtrace ----------------------------------------------------------------------------------- compare k8s objects ----------------------------------------------------------------------------------- + compare_kubectl service/some-name-pxc -1150 + local resource=service/some-name-pxc + local postfix=-1150 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150.yml + local new_result=/tmp/tmp.iLd3TvQh90/service_some-name-pxc.yml + desc 'compare service/some-name-pxc--1150' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc--1150 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150-eks.yml ']' + kubectl_bin get -o yaml service/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.GzFvUrPx45 ++ mktemp + local LAST_ERR=/tmp/tmp.tQwsixOkTO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.GzFvUrPx45 + cat /tmp/tmp.tQwsixOkTO + rm /tmp/tmp.GzFvUrPx45 /tmp/tmp.tQwsixOkTO + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-pxc-1150.yml /tmp/tmp.iLd3TvQh90/service_some-name-pxc.yml + compare_kubectl service/some-name-proxysql -1150 + local resource=service/some-name-proxysql + local postfix=-1150 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150.yml + local new_result=/tmp/tmp.iLd3TvQh90/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql--1150' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql--1150 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150-eks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.q3eVqP1TyV ++ mktemp + local LAST_ERR=/tmp/tmp.P76eAEWFNO + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.q3eVqP1TyV + cat /tmp/tmp.P76eAEWFNO + rm /tmp/tmp.q3eVqP1TyV /tmp/tmp.P76eAEWFNO + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/service_some-name-proxysql-1150.yml /tmp/tmp.iLd3TvQh90/service_some-name-proxysql.yml + compare_kubectl statefulset/some-name-pxc -1150 + local resource=statefulset/some-name-pxc + local postfix=-1150 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150.yml + local new_result=/tmp/tmp.iLd3TvQh90/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc--1150' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc--1150 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.oV1ZpJJ4ZD ++ mktemp + local LAST_ERR=/tmp/tmp.jdSB75Jsh6 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oV1ZpJJ4ZD + cat /tmp/tmp.jdSB75Jsh6 + rm /tmp/tmp.oV1ZpJJ4ZD /tmp/tmp.jdSB75Jsh6 + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-pxc-1150.yml /tmp/tmp.iLd3TvQh90/statefulset_some-name-pxc.yml + compare_kubectl statefulset/some-name-proxysql -1150 + local resource=statefulset/some-name-proxysql + local postfix=-1150 + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150.yml + local new_result=/tmp/tmp.iLd3TvQh90/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql--1150' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql--1150 ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-80.yml ']' + version_gt 1.29 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.29' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.27 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.27' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.24 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-k124.yml ']' + version_gt 1.22 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-k122.yml ']' + version_gt 1.21 + desc 'return true if kubernetes version equal or greater than desired' + set +o xtrace ----------------------------------------------------------------------------------- return true if kubernetes version equal or greater than desired ----------------------------------------------------------------------------------- ++ echo '1.26 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150-eks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-proxysql + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("upgrade-consistency-26343", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.NmMCOvp1rs ++ mktemp + local LAST_ERR=/tmp/tmp.TOzaIZiNEo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.NmMCOvp1rs + cat /tmp/tmp.TOzaIZiNEo + rm /tmp/tmp.NmMCOvp1rs /tmp/tmp.TOzaIZiNEo + return 0 + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1739/e2e-tests/upgrade-consistency/compare/statefulset_some-name-proxysql-1150.yml /tmp/tmp.iLd3TvQh90/statefulset_some-name-proxysql.yml + destroy upgrade-consistency-26343 + local namespace=upgrade-consistency-26343 + local ignore_logs=true + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' + grep -v level=info + grep -v 'the object has been modified' + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + tee /tmp/tmp.iLd3TvQh90/operator.log ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ + sort -u + grep -v 'get backup status: Job.batch' +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.bJITvP3JAJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.gTFZuxHppf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bJITvP3JAJ ++ cat /tmp/tmp.gTFZuxHppf ++ rm /tmp/tmp.bJITvP3JAJ /tmp/tmp.gTFZuxHppf ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-86d767f4fd-2ccjm ++ mktemp + local LAST_OUT=/tmp/tmp.hdWhnVuNvd ++ mktemp + local LAST_ERR=/tmp/tmp.NM4eC3LxgV + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-86d767f4fd-2ccjm + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hdWhnVuNvd + cat /tmp/tmp.NM4eC3LxgV + rm /tmp/tmp.hdWhnVuNvd /tmp/tmp.NM4eC3LxgV + return 0 2024-06-27T03:36:43.686Z INFO setup Manager starting up {"gitCommit": "a91d02478d7c4c292126db7185bdb6b6009b2b80", "gitBranch": "PR-1739-a91d0247", "buildTime": "2024-06-27T01:53:16Z", "goVersion": "go1.22.4", "os": "linux", "arch": "amd64"} 2024-06-27T03:36:43.686Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.26.15-gke.1436000"} 2024-06-27T03:36:43.687Z INFO setup Registering Components. 2024-06-27T03:36:47.966Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2024-06-27T03:36:47.970Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2024-06-27T03:36:47.970Z INFO controller-runtime.metrics Starting metrics server 2024-06-27T03:36:47.970Z INFO controller-runtime.webhook Starting webhook server 2024-06-27T03:36:47.970Z INFO setup Starting the Cmd. 2024-06-27T03:36:47.970Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2024-06-27T03:36:47.971Z INFO controller-runtime.certwatcher Starting certificate watcher 2024-06-27T03:36:47.971Z INFO controller-runtime.certwatcher Updated current TLS certificate 2024-06-27T03:36:47.971Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2024-06-27T03:36:48.071Z INFO attempting to acquire leader lease pxc-operator/08db1feb.percona.com... 2024-06-27T03:36:48.086Z DEBUG events percona-xtradb-cluster-operator-86d767f4fd-2ccjm_fdbebda1-66ee-48e4-bfa6-0341ac614981 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"64a22dd7-2e53-4b43-9996-f87cf47ce86a","apiVersion":"coordination.k8s.io/v1","resourceVersion":"53492"}, "reason": "LeaderElection"} 2024-06-27T03:36:48.086Z INFO successfully acquired lease pxc-operator/08db1feb.percona.com 2024-06-27T03:36:48.087Z INFO Starting Controller {"controller": "pxcbackup-controller"} 2024-06-27T03:36:48.087Z INFO Starting Controller {"controller": "pxc-controller"} 2024-06-27T03:36:48.087Z INFO Starting Controller {"controller": "pxcrestore-controller"} 2024-06-27T03:36:48.087Z INFO Starting EventSource {"controller": "pxcbackup-controller", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2024-06-27T03:36:48.087Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.PerconaXtraDBCluster"} 2024-06-27T03:36:48.087Z INFO Starting EventSource {"controller": "pxcrestore-controller", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2024-06-27T03:36:48.192Z INFO Starting workers {"controller": "pxc-controller", "worker count": 1} 2024-06-27T03:36:48.199Z INFO Starting workers {"controller": "pxcrestore-controller", "worker count": 1} 2024-06-27T03:36:48.257Z INFO Starting workers {"controller": "pxcbackup-controller", "worker count": 1} 2024-06-27T03:37:26.471Z INFO User secrets updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "2b124bb2-cbe1-4b98-9d47-bbad9f87a35f", "secrets": "my-cluster-secrets"} 2024-06-27T03:38:49.597Z INFO User monitor: granted privileges {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "5abacc08-c9c5-4bf8-bb40-8b86580bd4f5"} 2024-06-27T03:38:49.702Z INFO monitor user privileges granted {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "5abacc08-c9c5-4bf8-bb40-8b86580bd4f5"} 2024-06-27T03:38:49.802Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "5abacc08-c9c5-4bf8-bb40-8b86580bd4f5"} 2024-06-27T03:38:50.044Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "5abacc08-c9c5-4bf8-bb40-8b86580bd4f5", "err": "get primary pxc pod: not found"} 2024-06-27T03:38:54.117Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "1af7a65b-2640-4cb4-8811-1978a7026398", "err": "get primary pxc pod: not found"} 2024-06-27T03:38:59.365Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "7e5fdee6-c777-42ca-9e10-f50f82512bc5", "err": "get primary pxc pod: not found"} 2024-06-27T03:39:04.773Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "43acbf8c-6fd7-4b1e-b07c-25f6e1f63769", "err": "get primary pxc pod: not found"} 2024-06-27T03:41:11.857Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "9a1880ee-8e71-4324-bf14-4955fcca8470", "new version": "8.0.36-28.1"} 2024-06-27T03:41:15.095Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "9a1880ee-8e71-4324-bf14-4955fcca8470"} 2024-06-27T03:41:20.099Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "7280c640-6340-4ec8-ad10-a222f99ed103"} 2024-06-27T03:41:25.513Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "caa6f00c-3dd8-43c5-ae88-4d9b6b15ab41"} 2024-06-27T03:41:30.915Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "512a6f58-7ad5-4f67-9ecf-45e4b2b708de"} 2024-06-27T03:41:36.385Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "a37ec357-6a60-40f9-8fa6-efcdcfa6d3bd"} 2024-06-27T03:41:40.048Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "b8fbe36c-084b-4c4d-bbfe-8bff28ce4e1e", "user": "root"} 2024-06-27T03:41:40.096Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "b8fbe36c-084b-4c4d-bbfe-8bff28ce4e1e", "user": "operator"} 2024-06-27T03:41:40.140Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "b8fbe36c-084b-4c4d-bbfe-8bff28ce4e1e", "user": "monitor"} 2024-06-27T03:41:40.185Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "b8fbe36c-084b-4c4d-bbfe-8bff28ce4e1e", "user": "xtrabackup"} 2024-06-27T03:41:40.229Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "b8fbe36c-084b-4c4d-bbfe-8bff28ce4e1e", "user": "replication"} 2024-06-27T03:41:40.826Z ERROR sync users {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "b5082fa3-a32a-4ea8-9a59-b33d4ed66c68", "error": "exec syncusers: command terminated with exit code 137 / / ", "errorVerbose": "exec syncusers: command terminated with exit code 137 / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:920\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1246\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1695"} 2024-06-27T03:42:34.081Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "44966e54-6cfe-4e2d-9192-a12da1a56bc4", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:42:39.020Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "e67cc1a7-a136-4035-86ba-0c1782a4f939", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:42:44.308Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "7072b222-0712-42c5-8ad2-412907ea8053", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp 10.164.104.31:33062: connect: connection refused"} 2024-06-27T03:43:32.195Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "374c2027-aeb7-47f2-8561-025cb5f5f8c8", "err": "failed to connect to pod some-name-pxc-0: dial tcp 10.164.106.38:33062: connect: connection refused"} 2024-06-27T03:43:37.419Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "10f4197b-8c9d-4e38-9043-237345fdef8d", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:43:42.597Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "3e965255-00d9-48ec-b2b8-1785dcfbc49f", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:43:47.778Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "81f8e324-1848-469c-b934-103c93991f74", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:43:52.979Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "4ad8bb58-cc8c-480e-bf5e-6227e30259a4", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:43:58.157Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "5c5cd7a6-d270-4b07-b568-ce66c9f44a6c", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:44:03.360Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "8cd269e3-db90-4c0c-bd9f-224347b2e59c", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:44:08.537Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "536066e1-b473-4731-a261-595b32818b84", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:44:17.515Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "250cdeb2-c2d6-498c-a1e1-b3242fa48461"} 2024-06-27T03:44:23.171Z ERROR sync users {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "069a5513-4e7d-428a-9511-b8aeb0df9e7a", "error": "exec syncusers: command terminated with exit code 137 / \nSyncing user accounts from PXC(some-name-pxc-1.some-name-pxc.upgrade-consistency-26343.svc.cluster.local:3306) to ProxySQL\n / ERROR 2013 (HY000): Lost connection to MySQL server at 'reading initial communication packet', system error: 104\n", "errorVerbose": "exec syncusers: command terminated with exit code 137 / \nSyncing user accounts from PXC(some-name-pxc-1.some-name-pxc.upgrade-consistency-26343.svc.cluster.local:3306) to ProxySQL\n / ERROR 2013 (HY000): Lost connection to MySQL server at 'reading initial communication packet', system error: 104\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:920\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1246\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1695"} 2024-06-27T03:45:21.316Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "2501cc96-85ec-472f-85b5-493e3586dd32", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:45:26.676Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "c898e5a6-7191-4de3-a299-ba973a21c47b", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp 10.164.104.32:33062: connect: connection refused"} 2024-06-27T03:46:08.636Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "a853cba3-4c77-4f10-b122-51f0ffcd4553", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:46:19.698Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "8f55ca32-e15e-490f-8e78-9e696dd2b409", "err": "failed to connect to pod some-name-pxc-0: dial tcp: lookup some-name-pxc-0.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:46:25.028Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "e5a7f391-f7af-40f1-bcbb-9514bc8e2365", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:46:30.209Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "d4d0ec8e-fa55-41c7-b86c-8b8744b1f91e", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:46:35.445Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "e21caa7c-c1f0-4e4e-a232-866b54aec9cc", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:47:00.212Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "c8b1386d-cc85-4c0d-ac4f-f6a5439b590a"} 2024-06-27T03:47:05.200Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "36053c92-c3a2-4b73-8f53-2bc8ca349b81"} 2024-06-27T03:47:05.974Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "2e1e647e-576d-4be3-98a8-1ebcd3b87fe4", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-2: dial tcp: lookup some-name-pxc-2.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:47:06.619Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "c41f45c8-d524-4685-9574-5746861b5d27", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-2: dial tcp: lookup some-name-pxc-2.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:47:07.298Z ERROR sync users {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "9cf92b78-764b-4b60-b854-80a7b83807a0", "error": "syncusers: ERROR 2005 (HY000): Unknown MySQL server host 'some-name-pxc-2.some-name-pxc.upgrade-consistency-26343.svc.cluster.local' (-2)\n", "errorVerbose": "syncusers: ERROR 2005 (HY000): Unknown MySQL server host 'some-name-pxc-2.some-name-pxc.upgrade-consistency-26343.svc.cluster.local' (-2)\n\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:923\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1246\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1695"} 2024-06-27T03:47:58.842Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "36a4b264-7874-4ec3-8ec0-bd98d8e29fc3", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:48:04.446Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "72cc461d-b37c-4dc4-a088-db33e53c2d38", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:48:10.025Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "a47bde81-92d3-442e-ba31-b6fa8e766c9f", "err": "failed to ensure cluster readonly status: connect to pod some-name-pxc-1: dial tcp: lookup some-name-pxc-1.some-name-pxc.upgrade-consistency-26343 on 10.164.112.10:53: no such host"} 2024-06-27T03:48:58.247Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "458b54e6-6c13-4cd0-8f27-e56edffaa026", "err": "failed to connect to pod some-name-pxc-0: dial tcp 10.164.106.43:33062: connect: connection refused"} 2024-06-27T03:49:03.432Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "d54304cf-3e8d-4b16-897f-f4a599b2cf21", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:49:08.656Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "298624a6-db1d-4e4e-881b-6d866817491d", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:49:13.874Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "649f8770-edd6-4061-8288-066878cefa35", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:49:19.622Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "c0cd6b11-d36e-4dd5-bf01-c2a2f982efc4", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:49:24.919Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "8d9d8707-3c5b-4f2a-adef-b1d21f1dcf66", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:49:30.140Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "95d54bf8-5936-48a0-b33d-8a6bab1fdeec", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:49:35.372Z INFO Unable to find primary pod for replication. No pod with name or ip like this {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "aa0f8ed2-412b-417e-aedc-717d4582035f", "primary name": "some-name-pxc-0.some-name-pxc.upgrade-consistency-26343.svc.cluster.local"} 2024-06-27T03:49:44.413Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "a8016e8c-8e8b-43d1-8ac1-c719fcb398d1"} 2024-06-27T03:49:49.322Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "8439b5e6-a41b-430e-9b46-28b6ca745b35"} 2024-06-27T03:49:54.696Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "deb5232c-b10d-4e12-b0c2-c21e98732402"} 2024-06-27T03:50:00.093Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "e2e26658-9069-41d6-9e0c-0a206a911463"} 2024-06-27T03:50:06.485Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "upgrade-consistency-26343", "name": "some-name", "reconcileID": "5a9dfc9a-1d2c-492e-91b7-22429d1fb637"} github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.18.4/pkg/internal/controller/controller.go:222 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.18.4/pkg/internal/controller/controller.go:261 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.18.4/pkg/internal/controller/controller.go:324 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1248 [mysql] 2024/06/27 03:43:32 connection.go:49: read tcp 10.164.106.36:56498->10.164.106.38:33062: read: connection reset by peer [mysql] 2024/06/27 03:48:58 connection.go:49: read tcp 10.164.106.36:33084->10.164.106.43:33062: read: connection reset by peer sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2 + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n upgrade-consistency-26343 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.4rWdp5Smme ++ mktemp + local LAST_ERR=/tmp/tmp.viFa8wgPqW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4rWdp5Smme perconaxtradbcluster.pxc.percona.com "some-name" deleted + cat /tmp/tmp.viFa8wgPqW + rm /tmp/tmp.4rWdp5Smme /tmp/tmp.viFa8wgPqW + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.7i8eSZsonD ++ mktemp + local LAST_ERR=/tmp/tmp.kf8a8552qE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.7i8eSZsonD No resources found + cat /tmp/tmp.kf8a8552qE + rm /tmp/tmp.7i8eSZsonD /tmp/tmp.kf8a8552qE + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.STBayZbiU0 ++ mktemp + local LAST_ERR=/tmp/tmp.gghWOcJxl2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.STBayZbiU0 No resources found + cat /tmp/tmp.gghWOcJxl2 + rm /tmp/tmp.STBayZbiU0 /tmp/tmp.gghWOcJxl2 + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.Rer70pROew ++ mktemp + local LAST_ERR=/tmp/tmp.BUxAkJhYTp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Rer70pROew validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.BUxAkJhYTp + rm /tmp/tmp.Rer70pROew /tmp/tmp.BUxAkJhYTp + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.14.2/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace upgrade-consistency-26343 + rm -rf /tmp/tmp.iLd3TvQh90 + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- + local LAST_OUT=/tmp/tmp.rJ8SDLppiC + local LAST_OUT=/tmp/tmp.Rp0B4FP732 test passed ----------------------------------------------------------------------------------- ++ mktemp ++ mktemp + local LAST_ERR=/tmp/tmp.VOz1vzo6iC + local exit_status=0 + local LAST_ERR=/tmp/tmp.hnfbLnKCsF + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace upgrade-consistency-26343 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator