Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/logs/init-deploy-5-7.log Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + create_infra init-deploy-5676 + local ns=init-deploy-5676 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n init-deploy-14865 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.N9zRL7DCCG ++ mktemp + local LAST_ERR=/tmp/tmp.l5CBCz6b86 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.N9zRL7DCCG perconaxtradbcluster.pxc.percona.com "some-name" deleted from init-deploy-14865 namespace + cat /tmp/tmp.l5CBCz6b86 + rm /tmp/tmp.N9zRL7DCCG /tmp/tmp.l5CBCz6b86 + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.vfR0OKUFUX ++ mktemp + local LAST_ERR=/tmp/tmp.BjojLQYNCn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vfR0OKUFUX No resources found + cat /tmp/tmp.BjojLQYNCn + rm /tmp/tmp.vfR0OKUFUX /tmp/tmp.BjojLQYNCn + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.3Alfd8LsmK ++ mktemp + local LAST_ERR=/tmp/tmp.i9PCj7B9cr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3Alfd8LsmK No resources found + cat /tmp/tmp.i9PCj7B9cr + rm /tmp/tmp.3Alfd8LsmK /tmp/tmp.i9PCj7B9cr + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' ++ mktemp + xargs kubectl delete ns + awk '{print$1}' + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.sGuxgzjiJj ++ mktemp + local LAST_OUT=/tmp/tmp.p9PLxUZ3ic + local LAST_ERR=/tmp/tmp.TUT0odz7T8 + local exit_status=0 ++ mktemp ++ seq 0 2 + local LAST_ERR=/tmp/tmp.DDJzhksyss + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.p9PLxUZ3ic + cat /tmp/tmp.DDJzhksyss + rm /tmp/tmp.p9PLxUZ3ic /tmp/tmp.DDJzhksyss + return 0 namespace "cert-manager" deleted namespace "init-deploy-14865" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sGuxgzjiJj namespace "pxc-operator" deleted + cat /tmp/tmp.TUT0odz7T8 + rm /tmp/tmp.sGuxgzjiJj /tmp/tmp.TUT0odz7T8 + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.hI8KRVTiNM ++ mktemp + local LAST_ERR=/tmp/tmp.9EZpzFod2m + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hI8KRVTiNM namespace/pxc-operator created + cat /tmp/tmp.9EZpzFod2m + rm /tmp/tmp.hI8KRVTiNM /tmp/tmp.9EZpzFod2m + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.w7pYImOran +++ mktemp ++ local LAST_ERR=/tmp/tmp.13qMXPCT60 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.w7pYImOran ++ cat /tmp/tmp.13qMXPCT60 ++ rm /tmp/tmp.w7pYImOran /tmp/tmp.13qMXPCT60 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-ff96f032-11-cluster2 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.610kQNbNT8 ++ mktemp + local LAST_ERR=/tmp/tmp.sH8sgP8bV7 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-ff96f032-11-cluster2 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.610kQNbNT8 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-ff96f032-11-cluster2" modified. + cat /tmp/tmp.sH8sgP8bV7 + rm /tmp/tmp.610kQNbNT8 /tmp/tmp.sH8sgP8bV7 + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.EBcBV1WnQb ++ mktemp + local LAST_ERR=/tmp/tmp.Fj2hOWEaO7 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EBcBV1WnQb customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.Fj2hOWEaO7 + rm /tmp/tmp.EBcBV1WnQb /tmp/tmp.Fj2hOWEaO7 + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.6BXuYNxnb6 ++ mktemp + local LAST_ERR=/tmp/tmp.n4pCpX1qZK + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6BXuYNxnb6 clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.n4pCpX1qZK + rm /tmp/tmp.6BXuYNxnb6 /tmp/tmp.n4pCpX1qZK + return 0 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/cw-operator.yaml + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2234-ff96f032^' + kubectl_bin apply -f - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - ++ mktemp + local LAST_OUT=/tmp/tmp.xw4GQVPBnl + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - ++ mktemp + local LAST_ERR=/tmp/tmp.5d6zqFgK4a + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.xw4GQVPBnl deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.5d6zqFgK4a + rm /tmp/tmp.xw4GQVPBnl /tmp/tmp.5d6zqFgK4a + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.rWSBDmeYiP ++ mktemp + local LAST_ERR=/tmp/tmp.Rzv1xzY26R + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rWSBDmeYiP pod/percona-xtradb-cluster-operator-5d8c4fb844-8qvwn condition met + cat /tmp/tmp.Rzv1xzY26R + rm /tmp/tmp.rWSBDmeYiP /tmp/tmp.Rzv1xzY26R + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.TotwqIOrt9 +++ mktemp ++ local LAST_ERR=/tmp/tmp.aKGW6KXfZC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TotwqIOrt9 ++ cat /tmp/tmp.aKGW6KXfZC ++ rm /tmp/tmp.TotwqIOrt9 /tmp/tmp.aKGW6KXfZC ++ return 0 + wait_pod percona-xtradb-cluster-operator-5d8c4fb844-8qvwn 480 pxc-operator + local pod=percona-xtradb-cluster-operator-5d8c4fb844-8qvwn + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-5d8c4fb844-8qvwn ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-5d8c4fb844-8qvwn condition met waiting for pod/percona-xtradb-cluster-operator-5d8c4fb844-8qvwn to become Ready.Ok + sleep 3 + create_namespace init-deploy-5676 + local namespace=init-deploy-5676 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl api-resources ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces init-deploy-5676' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces init-deploy-5676 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace init-deploy-5676 ++ mktemp + xargs kubectl delete ns + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.vYSUVguW7d ++ mktemp + local LAST_ERR=/tmp/tmp.ukf95KL1A6 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + local LAST_OUT=/tmp/tmp.reJtUJ6IOr ++ mktemp + local LAST_ERR=/tmp/tmp.ix6IfuDbJG + local exit_status=0 + awk '{print$1}' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace init-deploy-5676 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vYSUVguW7d + cat /tmp/tmp.ukf95KL1A6 + rm /tmp/tmp.vYSUVguW7d /tmp/tmp.ukf95KL1A6 + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace init-deploy-5676 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace init-deploy-5676 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.reJtUJ6IOr + cat /tmp/tmp.ix6IfuDbJG Error from server (NotFound): namespaces "init-deploy-5676" not found + rm /tmp/tmp.reJtUJ6IOr /tmp/tmp.ix6IfuDbJG + return 1 + : + wait_for_delete namespace/init-deploy-5676 + local res=namespace/init-deploy-5676 + echo -n 'waiting for namespace/init-deploy-5676 to be deleted' waiting for namespace/init-deploy-5676 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "init-deploy-5676" not found + desc 'create namespace init-deploy-5676' + set +o xtrace ----------------------------------------------------------------------------------- create namespace init-deploy-5676 ----------------------------------------------------------------------------------- + kubectl_bin create namespace init-deploy-5676 ++ mktemp + local LAST_OUT=/tmp/tmp.KYounOr5jC ++ mktemp + local LAST_ERR=/tmp/tmp.6qaZkZfzoR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace init-deploy-5676 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KYounOr5jC namespace/init-deploy-5676 created + cat /tmp/tmp.6qaZkZfzoR + rm /tmp/tmp.KYounOr5jC /tmp/tmp.6qaZkZfzoR + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.5d74C66aSY +++ mktemp ++ local LAST_ERR=/tmp/tmp.Aq4JP1lhCL ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5d74C66aSY ++ cat /tmp/tmp.Aq4JP1lhCL ++ rm /tmp/tmp.5d74C66aSY /tmp/tmp.Aq4JP1lhCL ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-ff96f032-11-cluster2 --namespace=init-deploy-5676 ++ mktemp + local LAST_OUT=/tmp/tmp.yHK7Vzom3N ++ mktemp + local LAST_ERR=/tmp/tmp.mFHEQeCavC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-ff96f032-11-cluster2 --namespace=init-deploy-5676 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.yHK7Vzom3N Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-ff96f032-11-cluster2" modified. + cat /tmp/tmp.mFHEQeCavC + rm /tmp/tmp.yHK7Vzom3N /tmp/tmp.mFHEQeCavC + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.HUkDN4BEf0 ++ mktemp + local LAST_ERR=/tmp/tmp.Y7Tdx6XP4E + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HUkDN4BEf0 secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.Y7Tdx6XP4E + rm /tmp/tmp.HUkDN4BEf0 /tmp/tmp.Y7Tdx6XP4E + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.i2z9Jyydn8 ++ mktemp + local LAST_ERR=/tmp/tmp.qZHFw95Z1I + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.i2z9Jyydn8 namespace/cert-manager created + cat /tmp/tmp.qZHFw95Z1I + rm /tmp/tmp.i2z9Jyydn8 /tmp/tmp.qZHFw95Z1I + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.DqWVq8igqQ ++ mktemp + local LAST_ERR=/tmp/tmp.piR6Y6CapL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.DqWVq8igqQ namespace/cert-manager labeled + cat /tmp/tmp.piR6Y6CapL + rm /tmp/tmp.DqWVq8igqQ /tmp/tmp.piR6Y6CapL + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.4Gt3hPVn84 ++ mktemp + local LAST_ERR=/tmp/tmp.w57ai97dUs + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4Gt3hPVn84 namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.w57ai97dUs Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.4Gt3hPVn84 /tmp/tmp.w57ai97dUs + return 0 + '[' '' == 4.10 ']' + sleep 70 + desc 'apply my-env-var-secrets to add PEER_LIST_SRV_PROTOCOL' + set +o xtrace ----------------------------------------------------------------------------------- apply my-env-var-secrets to add PEER_LIST_SRV_PROTOCOL ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/conf/envsecrets.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.8LuiBjbEg9 ++ mktemp + local LAST_ERR=/tmp/tmp.PQmwHi8gcY + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/conf/envsecrets.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.8LuiBjbEg9 secret/my-env-var-secrets created + cat /tmp/tmp.PQmwHi8gcY + rm /tmp/tmp.8LuiBjbEg9 /tmp/tmp.PQmwHi8gcY + return 0 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/some-name.yml 3 10 /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/secrets_without_tls.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/secrets_without_tls.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/secrets_without_tls.yml ++ mktemp + local LAST_OUT=/tmp/tmp.MhGUuZESZ5 ++ mktemp + local LAST_ERR=/tmp/tmp.RjL49UmEKp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/secrets_without_tls.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.MhGUuZESZ5 secret/my-cluster-secrets created + cat /tmp/tmp.RjL49UmEKp + rm /tmp/tmp.MhGUuZESZ5 /tmp/tmp.RjL49UmEKp + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.init-deploy-5676~ + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2234-ff96f032#' ++ mktemp + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + local LAST_OUT=/tmp/tmp.tqjejjiRnL + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' ++ mktemp + local LAST_ERR=/tmp/tmp.VkKonDy9lS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.tqjejjiRnL deployment.apps/pxc-client created + cat /tmp/tmp.VkKonDy9lS + rm /tmp/tmp.tqjejjiRnL /tmp/tmp.VkKonDy9lS + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ some-name == \d\e\m\a\n\d\-\b\a\c\k\u\p ]] + [[ some-name == \d\e\m\a\n\d\-\b\a\c\k\u\p\-\c\l\o\u\d ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/some-name.yml + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2234-ff96f032#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.init-deploy-5676~ + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/some-name.yml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.oiKZ1dXM5j ++ mktemp + local LAST_ERR=/tmp/tmp.RQn93GJzOf + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.oiKZ1dXM5j perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.RQn93GJzOf + rm /tmp/tmp.oiKZ1dXM5j /tmp/tmp.RQn93GJzOf + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.OJkKfyWXR0 ++++ mktemp +++ local LAST_ERR=/tmp/tmp.c7DJluFvod +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.OJkKfyWXR0 +++ cat /tmp/tmp.c7DJluFvod +++ rm /tmp/tmp.OJkKfyWXR0 /tmp/tmp.c7DJluFvod +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.t581JoPmIB ++++ mktemp +++ local LAST_ERR=/tmp/tmp.9bfOhIGdaX +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.t581JoPmIB +++ cat /tmp/tmp.9bfOhIGdaX +++ rm /tmp/tmp.t581JoPmIB /tmp/tmp.9bfOhIGdaX +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n init-deploy-5676 ++ mktemp + local LAST_OUT=/tmp/tmp.aM7l6aOQpE ++ mktemp + local LAST_ERR=/tmp/tmp.ebDJoChR6N + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n init-deploy-5676 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n init-deploy-5676 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n init-deploy-5676 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.aM7l6aOQpE + cat /tmp/tmp.ebDJoChR6N error: no matching resources found + rm /tmp/tmp.aM7l6aOQpE /tmp/tmp.ebDJoChR6N + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo some-name-pxc-1 ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.uSwK0sUtFZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.g2nOj8A1GD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.uSwK0sUtFZ ++ cat /tmp/tmp.g2nOj8A1GD ++ rm /tmp/tmp.uSwK0sUtFZ /tmp/tmp.g2nOj8A1GD ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ grep -E -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KE2QsumhIR +++ mktemp ++ local LAST_ERR=/tmp/tmp.IAWO3QEm83 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KE2QsumhIR ++ cat /tmp/tmp.IAWO3QEm83 Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.KE2QsumhIR /tmp/tmp.IAWO3QEm83 ++ return 0 + [[ -n '' ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6XIxxtgxgk +++ mktemp ++ local LAST_ERR=/tmp/tmp.edHvCnpdRR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6XIxxtgxgk ++ cat /tmp/tmp.edHvCnpdRR ++ rm /tmp/tmp.6XIxxtgxgk /tmp/tmp.edHvCnpdRR ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-857d976497-mwvqg + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.NSp6ZgjNjW +++ mktemp ++ local LAST_ERR=/tmp/tmp.FHngEGuwKc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NSp6ZgjNjW ++ cat /tmp/tmp.FHngEGuwKc ++ rm /tmp/tmp.NSp6ZgjNjW /tmp/tmp.FHngEGuwKc ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-857d976497-mwvqg + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.evid81l1K3 +++ mktemp ++ local LAST_ERR=/tmp/tmp.lYnobNuD72 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.evid81l1K3 ++ cat /tmp/tmp.lYnobNuD72 ++ rm /tmp/tmp.evid81l1K3 /tmp/tmp.lYnobNuD72 ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-857d976497-mwvqg + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.0ENVfr81ik/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bsF3JuTkE6 +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ak1Isb438F ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bsF3JuTkE6 ++ cat /tmp/tmp.Ak1Isb438F ++ rm /tmp/tmp.bsF3JuTkE6 /tmp/tmp.Ak1Isb438F ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.0ENVfr81ik/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.oTpnBoCGkE +++ mktemp ++ local LAST_ERR=/tmp/tmp.4KNmAhH9RP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oTpnBoCGkE ++ cat /tmp/tmp.4KNmAhH9RP ++ rm /tmp/tmp.oTpnBoCGkE /tmp/tmp.4KNmAhH9RP ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.0ENVfr81ik/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ grep -E -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ATi0Dcr9Gn +++ mktemp ++ local LAST_ERR=/tmp/tmp.2er0Qcke1L ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ATi0Dcr9Gn ++ cat /tmp/tmp.2er0Qcke1L Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.ATi0Dcr9Gn /tmp/tmp.2er0Qcke1L ++ return 0 + '[' '' ']' + desc 'check if cert-manager issued ssl certificates' + set +o xtrace ----------------------------------------------------------------------------------- check if cert-manager issued ssl certificates ----------------------------------------------------------------------------------- ++ kubectl get secrets some-name-ssl -o 'jsonpath={.metadata.annotations.cert-manager\.io/issuer-kind}' + '[' Issuer '!=' Issuer ']' + desc 'check if service and statefulset created with expected config' + set +o xtrace ----------------------------------------------------------------------------------- check if service and statefulset created with expected config ----------------------------------------------------------------------------------- + compare_kubectl statefulset/some-name-pxc + local resource=statefulset/some-name-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc.yml + local new_result=/tmp/tmp.0ENVfr81ik/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127.yml ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127-aks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.d9ntY30Svf ++ mktemp + local LAST_ERR=/tmp/tmp.DDoIVICkuC + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.d9ntY30Svf + cat /tmp/tmp.DDoIVICkuC + rm /tmp/tmp.d9ntY30Svf /tmp/tmp.DDoIVICkuC + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127.yml /tmp/tmp.0ENVfr81ik/statefulset_some-name-pxc.yml + log 'compare_kubectl: statefulset/some-name-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:17:24+0000]' compare_kubectl: statefulset/some-name-pxc OK [2025-11-24T18:17:24+0000] compare_kubectl: statefulset/some-name-pxc OK + compare_kubectl statefulset/some-name-proxysql + local resource=statefulset/some-name-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql.yml + local new_result=/tmp/tmp.0ENVfr81ik/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127.yml ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127-aks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-proxysql + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.D4z1Xm7iSl ++ mktemp + local LAST_ERR=/tmp/tmp.lLkxIsdLzU + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.D4z1Xm7iSl + cat /tmp/tmp.lLkxIsdLzU + rm /tmp/tmp.D4z1Xm7iSl /tmp/tmp.lLkxIsdLzU + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127.yml /tmp/tmp.0ENVfr81ik/statefulset_some-name-proxysql.yml + log 'compare_kubectl: statefulset/some-name-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:17:26+0000]' compare_kubectl: statefulset/some-name-proxysql OK [2025-11-24T18:17:26+0000] compare_kubectl: statefulset/some-name-proxysql OK + compare_kubectl service/some-name-pxc + local resource=service/some-name-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc.yml + local new_result=/tmp/tmp.0ENVfr81ik/service_some-name-pxc.yml + desc 'compare service/some-name-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc-aks.yml ']' + kubectl_bin get -o yaml service/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.CJQG9dKM99 ++ mktemp + local LAST_ERR=/tmp/tmp.z3tZNmpoaK + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.CJQG9dKM99 + cat /tmp/tmp.z3tZNmpoaK + rm /tmp/tmp.CJQG9dKM99 /tmp/tmp.z3tZNmpoaK + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-pxc.yml /tmp/tmp.0ENVfr81ik/service_some-name-pxc.yml + log 'compare_kubectl: service/some-name-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:17:27+0000]' compare_kubectl: service/some-name-pxc OK [2025-11-24T18:17:27+0000] compare_kubectl: service/some-name-pxc OK + compare_kubectl service/some-name-proxysql + local resource=service/some-name-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql.yml + local new_result=/tmp/tmp.0ENVfr81ik/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-k122.yml ']' + version_gt 1.21 ++ bc -l ++ echo '1.31 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml service/some-name-proxysql ++ mktemp + local LAST_OUT=/tmp/tmp.zaBuiv4V5h ++ mktemp + local LAST_ERR=/tmp/tmp.Ny5bRGlRIh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zaBuiv4V5h + cat /tmp/tmp.Ny5bRGlRIh + rm /tmp/tmp.zaBuiv4V5h /tmp/tmp.Ny5bRGlRIh + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql.yml /tmp/tmp.0ENVfr81ik/service_some-name-proxysql.yml + log 'compare_kubectl: service/some-name-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:17:29+0000]' compare_kubectl: service/some-name-proxysql OK [2025-11-24T18:17:29+0000] compare_kubectl: service/some-name-proxysql OK + compare_kubectl service/some-name-proxysql-unready + local resource=service/some-name-proxysql-unready + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready.yml + local new_result=/tmp/tmp.0ENVfr81ik/service_some-name-proxysql-unready.yml + desc 'compare service/some-name-proxysql-unready-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql-unready- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-aks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql-unready ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.aLYCoF1D5O ++ mktemp + local LAST_ERR=/tmp/tmp.I4uFZ2Cele + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql-unready + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.aLYCoF1D5O + cat /tmp/tmp.I4uFZ2Cele + rm /tmp/tmp.aLYCoF1D5O /tmp/tmp.I4uFZ2Cele + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready.yml /tmp/tmp.0ENVfr81ik/service_some-name-proxysql-unready.yml + log 'compare_kubectl: service/some-name-proxysql-unready OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:17:30+0000]' compare_kubectl: service/some-name-proxysql-unready OK [2025-11-24T18:17:30+0000] compare_kubectl: service/some-name-proxysql-unready OK + [[ ! perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + desc 'check PodDisruptionBudget' + set +o xtrace ----------------------------------------------------------------------------------- check PodDisruptionBudget ----------------------------------------------------------------------------------- + compare_kubectl pdb/some-name-pxc + local resource=pdb/some-name-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc.yml + local new_result=/tmp/tmp.0ENVfr81ik/pdb_some-name-pxc.yml + desc 'compare pdb/some-name-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare pdb/some-name-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k129.yml ']' + version_gt 1.27 ++ bc -l ++ echo '1.31 >= 1.27' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.31 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc-aks.yml ']' + kubectl_bin get -o yaml pdb/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.DwXAaZr7qw ++ mktemp + local LAST_ERR=/tmp/tmp.i3LmgMtvLz + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml pdb/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.DwXAaZr7qw + cat /tmp/tmp.i3LmgMtvLz + rm /tmp/tmp.DwXAaZr7qw /tmp/tmp.i3LmgMtvLz + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-pxc.yml /tmp/tmp.0ENVfr81ik/pdb_some-name-pxc.yml + log 'compare_kubectl: pdb/some-name-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:17:32+0000]' compare_kubectl: pdb/some-name-pxc OK [2025-11-24T18:17:32+0000] compare_kubectl: pdb/some-name-pxc OK + compare_kubectl pdb/some-name-proxysql + local resource=pdb/some-name-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql.yml + local new_result=/tmp/tmp.0ENVfr81ik/pdb_some-name-proxysql.yml + desc 'compare pdb/some-name-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare pdb/some-name-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.31 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-aks.yml ']' + kubectl_bin get -o yaml pdb/some-name-proxysql ++ mktemp + local LAST_OUT=/tmp/tmp.fPopVU7Zgl + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_ERR=/tmp/tmp.ACoByGBOgE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml pdb/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.fPopVU7Zgl + cat /tmp/tmp.ACoByGBOgE + rm /tmp/tmp.fPopVU7Zgl /tmp/tmp.ACoByGBOgE + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pdb_some-name-proxysql.yml /tmp/tmp.0ENVfr81ik/pdb_some-name-proxysql.yml + log 'compare_kubectl: pdb/some-name-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:17:33+0000]' compare_kubectl: pdb/some-name-proxysql OK [2025-11-24T18:17:33+0000] compare_kubectl: pdb/some-name-proxysql OK + desc 'check if MySQL users created' + set +o xtrace ----------------------------------------------------------------------------------- check if MySQL users created ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-pxc -uroot -proot_password' + local 'uri=-h some-name-pxc -uroot -proot_password' + local postfix= ++ echo -h some-name-pxc -uroot -proot_password ++ sed -e 's/.*-u//; s/ .*//' + local user=root + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/root.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/root-57.sql ]] + run_mysql 'SHOW GRANTS;' '-h some-name-pxc -uroot -proot_password' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qSFrpVVthf +++ mktemp ++ local LAST_ERR=/tmp/tmp.xe3qsvEHuW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qSFrpVVthf ++ cat /tmp/tmp.xe3qsvEHuW ++ rm /tmp/tmp.qSFrpVVthf /tmp/tmp.xe3qsvEHuW ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/root.sql /tmp/tmp.0ENVfr81ik/root.sql + compare_mysql_user '-h some-name-pxc -umonitor -pmonitor_password' + local 'uri=-h some-name-pxc -umonitor -pmonitor_password' + local postfix= ++ sed -e 's/.*-u//; s/ .*//' ++ echo -h some-name-pxc -umonitor -pmonitor_password + local user=monitor + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/monitor.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/monitor-57.sql ]] + run_mysql 'SHOW GRANTS;' '-h some-name-pxc -umonitor -pmonitor_password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-pxc -umonitor -pmonitor_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0cS6KTxhOg +++ mktemp ++ local LAST_ERR=/tmp/tmp.jafDs8qvUB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0cS6KTxhOg ++ cat /tmp/tmp.jafDs8qvUB ++ rm /tmp/tmp.0cS6KTxhOg /tmp/tmp.jafDs8qvUB ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/monitor.sql /tmp/tmp.0ENVfr81ik/monitor.sql + compare_mysql_user '-h some-name-pxc -uproxyuser -ps3cret' + local 'uri=-h some-name-pxc -uproxyuser -ps3cret' + local postfix= ++ echo -h some-name-pxc -uproxyuser -ps3cret ++ sed -e 's/.*-u//; s/ .*//' + local user=proxyuser + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/proxyuser.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/proxyuser-57.sql ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-pxc -uproxyuser -ps3cret' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-pxc -uproxyuser -ps3cret' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.WDpy9CiU8B +++ mktemp ++ local LAST_ERR=/tmp/tmp.ohCDcX0mf7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WDpy9CiU8B ++ cat /tmp/tmp.ohCDcX0mf7 ++ rm /tmp/tmp.WDpy9CiU8B /tmp/tmp.ohCDcX0mf7 ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/proxyuser.sql /tmp/tmp.0ENVfr81ik/proxyuser.sql + compare_mysql_user_local '-uxtrabackup -pbackup_password' some-name-pxc-0 '' pxc + local 'uri=-uxtrabackup -pbackup_password' + local pod=some-name-pxc-0 + local postfix= + local container_name=pxc ++ echo -uxtrabackup -pbackup_password ++ sed -e 's/.*-u//; s/ .*//' + local user=xtrabackup + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/xtrabackup.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql_local 'SHOW GRANTS;' '-uxtrabackup -pbackup_password' some-name-pxc-0 pxc + local 'command=SHOW GRANTS;' + local 'uri=-uxtrabackup -pbackup_password' + local pod=some-name-pxc-0 + local container_name=pxc + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/xtrabackup.sql /tmp/tmp.0ENVfr81ik/xtrabackup.sql + desc 'check that pmm server user don'\''t have access' + set +o xtrace ----------------------------------------------------------------------------------- check that pmm server user don't have access ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-pxc -upmmserver -ppmmserver_password' + local 'uri=-h some-name-pxc -upmmserver -ppmmserver_password' + local postfix= ++ echo -h some-name-pxc -upmmserver -ppmmserver_password ++ sed -e 's/.*-u//; s/ .*//' + local user=pmmserver + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pmmserver.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pmmserver-57.sql ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-pxc -upmmserver -ppmmserver_password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-pxc -upmmserver -ppmmserver_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hiEcY3cwVV +++ mktemp ++ local LAST_ERR=/tmp/tmp.4dEHf7B41B ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hiEcY3cwVV ++ cat /tmp/tmp.4dEHf7B41B ++ rm /tmp/tmp.hiEcY3cwVV /tmp/tmp.4dEHf7B41B ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-857d976497-mwvqg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pmmserver.sql /tmp/tmp.0ENVfr81ik/pmmserver.sql + desc 'check if ProxySQL users created' + set +o xtrace ----------------------------------------------------------------------------------- check if ProxySQL users created ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-proxysql -uroot -proot_password' + local 'uri=-h some-name-proxysql -uroot -proot_password' + local postfix= ++ echo -h some-name-proxysql -uroot -proot_password ++ sed -e 's/.*-u//; s/ .*//' + local user=root + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/root.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/root-57.sql ]] + run_mysql 'SHOW GRANTS;' '-h some-name-proxysql -uroot -proot_password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ local LAST_OUT=/tmp/tmp.jh3NKjuWiv +++ mktemp ++ local LAST_ERR=/tmp/tmp.7SaIXDbEvC ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jh3NKjuWiv ++ cat /tmp/tmp.7SaIXDbEvC ++ rm /tmp/tmp.jh3NKjuWiv /tmp/tmp.7SaIXDbEvC ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/root.sql /tmp/tmp.0ENVfr81ik/root.sql + compare_mysql_user '-h some-name-proxysql -umonitor -pmonitor_password' + local 'uri=-h some-name-proxysql -umonitor -pmonitor_password' + local postfix= ++ sed -e 's/.*-u//; s/ .*//' ++ echo -h some-name-proxysql -umonitor -pmonitor_password + local user=monitor + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/monitor.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/monitor-57.sql ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-proxysql -umonitor -pmonitor_password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-proxysql -umonitor -pmonitor_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Kz5DxYPBOV +++ mktemp ++ local LAST_ERR=/tmp/tmp.0qZXWlJ0EY ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Kz5DxYPBOV ++ cat /tmp/tmp.0qZXWlJ0EY ++ rm /tmp/tmp.Kz5DxYPBOV /tmp/tmp.0qZXWlJ0EY ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-857d976497-mwvqg + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/monitor.sql /tmp/tmp.0ENVfr81ik/monitor.sql + desc 'check that pmm server user don'\''t have access' + set +o xtrace ----------------------------------------------------------------------------------- check that pmm server user don't have access ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-proxysql -uproxyuser -ps3cret' -proxysql + local 'uri=-h some-name-proxysql -uproxyuser -ps3cret' + local postfix=-proxysql ++ echo -h some-name-proxysql -uproxyuser -ps3cret ++ sed -e 's/.*-u//; s/ .*//' + local user=proxyuser + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/proxyuser-proxysql.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/proxyuser-proxysql-57.sql ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-proxysql -uproxyuser -ps3cret' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-proxysql -uproxyuser -ps3cret' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.5I3Zw54mG5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.lx4HVwAyP4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.5I3Zw54mG5 ++ cat /tmp/tmp.lx4HVwAyP4 ++ rm /tmp/tmp.5I3Zw54mG5 /tmp/tmp.lx4HVwAyP4 ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/proxyuser-proxysql.sql /tmp/tmp.0ENVfr81ik/proxyuser.sql + compare_mysql_user '-h some-name-proxysql -upmmserver -ppmmserver_password' -proxysql + local 'uri=-h some-name-proxysql -upmmserver -ppmmserver_password' + local postfix=-proxysql ++ echo -h some-name-proxysql -upmmserver -ppmmserver_password ++ sed -e 's/.*-u//; s/ .*//' + local user=pmmserver + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pmmserver-proxysql.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pmmserver-proxysql-57.sql ]] + run_mysql 'SHOW GRANTS;' '-h some-name-proxysql -upmmserver -ppmmserver_password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-proxysql -upmmserver -ppmmserver_password' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.v4OyEVMI8i +++ mktemp ++ local LAST_ERR=/tmp/tmp.bNwIZcvbPW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.v4OyEVMI8i ++ cat /tmp/tmp.bNwIZcvbPW ++ rm /tmp/tmp.v4OyEVMI8i /tmp/tmp.bNwIZcvbPW ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-857d976497-mwvqg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/pmmserver-proxysql.sql /tmp/tmp.0ENVfr81ik/pmmserver.sql + desc 'write data directly, read from all' + set +o xtrace ----------------------------------------------------------------------------------- write data directly, read from all ----------------------------------------------------------------------------------- + run_mysql 'INSERT myApp.myApp (id) VALUES (100501)' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local 'command=INSERT myApp.myApp (id) VALUES (100501)' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7lYssPbzf0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.4x2owOmvyR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7lYssPbzf0 ++ cat /tmp/tmp.4x2owOmvyR ++ rm /tmp/tmp.7lYssPbzf0 /tmp/tmp.4x2owOmvyR ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + compare_mysql_cmd select-2 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-2-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.tcADHQcWhM +++ mktemp ++ local LAST_ERR=/tmp/tmp.KcmnMww4JV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.tcADHQcWhM ++ cat /tmp/tmp.KcmnMww4JV ++ rm /tmp/tmp.tcADHQcWhM /tmp/tmp.KcmnMww4JV ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-857d976497-mwvqg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-2.sql /tmp/tmp.0ENVfr81ik/select-2.sql + compare_mysql_cmd select-2 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-2-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.fKUypSkE1I +++ mktemp ++ local LAST_ERR=/tmp/tmp.aHrJvLTpeG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.fKUypSkE1I ++ cat /tmp/tmp.aHrJvLTpeG ++ rm /tmp/tmp.fKUypSkE1I /tmp/tmp.aHrJvLTpeG ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-2.sql /tmp/tmp.0ENVfr81ik/select-2.sql + compare_mysql_cmd select-2 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-2-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.CRqcpO9x1n +++ mktemp ++ local LAST_ERR=/tmp/tmp.Aeli887pfJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CRqcpO9x1n ++ cat /tmp/tmp.Aeli887pfJ ++ rm /tmp/tmp.CRqcpO9x1n /tmp/tmp.Aeli887pfJ ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-2.sql /tmp/tmp.0ENVfr81ik/select-2.sql + desc 'kill Primary Pod, check reelection, check data' + set +o xtrace ----------------------------------------------------------------------------------- kill Primary Pod, check reelection, check data ----------------------------------------------------------------------------------- ++ get_proxy_primary '-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' some-name-proxysql-0 ++ local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' ++ local pod=some-name-proxysql-0 +++ run_mysql_local 'SELECT hostname FROM runtime_mysql_servers WHERE hostgroup_id=11 AND status='\''ONLINE'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' some-name-proxysql-0 proxysql +++ local 'command=SELECT hostname FROM runtime_mysql_servers WHERE hostgroup_id=11 AND status='\''ONLINE'\'';' +++ local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' +++ local pod=some-name-proxysql-0 +++ local container_name=proxysql +++ set +o xtrace ++ local ip=some-name-pxc-2.some-name-pxc.init-deploy-5676.svc.cluster.local +++ echo some-name-pxc-2.some-name-pxc.init-deploy-5676.svc.cluster.local +++ wc -l ++ '[' 1 '!=' 1 ']' ++ echo some-name-pxc-2.some-name-pxc.init-deploy-5676.svc.cluster.local ++ cut -d. -f1 + initial_primary=some-name-pxc-2 + kubectl_bin delete pods --grace-period=0 --force some-name-pxc-2 ++ mktemp + local LAST_OUT=/tmp/tmp.mYd19DlTQY ++ mktemp + local LAST_ERR=/tmp/tmp.0un9ezVWtI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pods --grace-period=0 --force some-name-pxc-2 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mYd19DlTQY pod "some-name-pxc-2" force deleted from init-deploy-5676 namespace + cat /tmp/tmp.0un9ezVWtI Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. + rm /tmp/tmp.mYd19DlTQY /tmp/tmp.0un9ezVWtI + return 0 + sleep 10 ++ get_proxy_primary '-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' some-name-proxysql-0 ++ local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' ++ local pod=some-name-proxysql-0 +++ run_mysql_local 'SELECT hostname FROM runtime_mysql_servers WHERE hostgroup_id=11 AND status='\''ONLINE'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' some-name-proxysql-0 proxysql +++ local 'command=SELECT hostname FROM runtime_mysql_servers WHERE hostgroup_id=11 AND status='\''ONLINE'\'';' +++ local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' +++ local pod=some-name-proxysql-0 +++ local container_name=proxysql +++ set +o xtrace ++ local ip=some-name-pxc-1.some-name-pxc.init-deploy-5676.svc.cluster.local +++ echo some-name-pxc-1.some-name-pxc.init-deploy-5676.svc.cluster.local +++ wc -l ++ '[' 1 '!=' 1 ']' ++ echo some-name-pxc-1.some-name-pxc.init-deploy-5676.svc.cluster.local ++ cut -d. -f1 + changed_primary=some-name-pxc-1 + '[' -n some-name-pxc-1 ']' + '[' some-name-pxc-2 '!=' some-name-pxc-1 ']' + run_mysql 'INSERT myApp.myApp (id) VALUES (100502)' '-h some-name-proxysql -uroot -proot_password' + local 'command=INSERT myApp.myApp (id) VALUES (100502)' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.90wKWlz8zM +++ mktemp ++ local LAST_ERR=/tmp/tmp.U9tLNq6NGm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.90wKWlz8zM ++ cat /tmp/tmp.U9tLNq6NGm ++ rm /tmp/tmp.90wKWlz8zM /tmp/tmp.U9tLNq6NGm ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo some-name-pxc-1 + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + compare_mysql_cmd select-3 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local command_id=select-3 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-3-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.KMnWGmuIpw +++ mktemp ++ local LAST_ERR=/tmp/tmp.PGLABBJnTw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.KMnWGmuIpw ++ cat /tmp/tmp.PGLABBJnTw ++ rm /tmp/tmp.KMnWGmuIpw /tmp/tmp.PGLABBJnTw ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-3.sql /tmp/tmp.0ENVfr81ik/select-3.sql + compare_mysql_cmd select-3 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local command_id=select-3 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-3-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.b1JZZE5F1k +++ mktemp ++ local LAST_ERR=/tmp/tmp.D9nxk1WUX4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.b1JZZE5F1k ++ cat /tmp/tmp.D9nxk1WUX4 ++ rm /tmp/tmp.b1JZZE5F1k /tmp/tmp.D9nxk1WUX4 ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-3.sql /tmp/tmp.0ENVfr81ik/select-3.sql + compare_mysql_cmd select-3 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local command_id=select-3 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-3-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.B1n9ICQL2B +++ mktemp ++ local LAST_ERR=/tmp/tmp.JwI520I9US ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.B1n9ICQL2B ++ cat /tmp/tmp.JwI520I9US ++ rm /tmp/tmp.B1n9ICQL2B /tmp/tmp.JwI520I9US ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-3.sql /tmp/tmp.0ENVfr81ik/select-3.sql + compare_mysql_cmd max_allowed_packet-1 'SELECT @@max_allowed_packet;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local command_id=max_allowed_packet-1 + local 'command=SELECT @@max_allowed_packet;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/max_allowed_packet-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/max_allowed_packet-1-57.sql ]] + run_mysql 'SELECT @@max_allowed_packet;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local 'command=SELECT @@max_allowed_packet;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bQTj9XLD2V +++ mktemp ++ local LAST_ERR=/tmp/tmp.bdG79GtBGf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bQTj9XLD2V ++ cat /tmp/tmp.bdG79GtBGf ++ rm /tmp/tmp.bQTj9XLD2V /tmp/tmp.bdG79GtBGf ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-857d976497-mwvqg + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/max_allowed_packet-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/max_allowed_packet-1.sql /tmp/tmp.0ENVfr81ik/max_allowed_packet-1.sql + desc 'check if possible to create no-proxysql cluster' + set +o xtrace ----------------------------------------------------------------------------------- check if possible to create no-proxysql cluster ----------------------------------------------------------------------------------- + cluster3=no-proxysql + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/conf/no-proxysql.yml + '[' -z '' ']' + kubectl_bin apply -f - ++ mktemp + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/conf/no-proxysql.yml + local LAST_OUT=/tmp/tmp.ZiiZxZKDWz + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7-backup#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.init-deploy-5676~ + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/conf/no-proxysql.yml + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc5.7#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2234-ff96f032#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' ++ mktemp + local LAST_ERR=/tmp/tmp.MuUkx9SQuP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZiiZxZKDWz perconaxtradbcluster.pxc.percona.com/no-proxysql created + cat /tmp/tmp.MuUkx9SQuP + rm /tmp/tmp.ZiiZxZKDWz /tmp/tmp.MuUkx9SQuP + return 0 + wait_for_running no-proxysql-pxc 3 + local name=no-proxysql-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod no-proxysql-pxc-0 480 + local pod=no-proxysql-pxc-0 + local max_retry=480 + local ns= ++ echo no-proxysql-pxc-0 ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace Error from server (NotFound): pods "no-proxysql-pxc-0" not found waiting for pod/no-proxysql-pxc-0 to become Ready.................................Ok + for i in '$(seq 0 $last_pod)' + wait_pod no-proxysql-pxc-1 480 + local pod=no-proxysql-pxc-1 + local max_retry=480 + local ns= ++ echo no-proxysql-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/no-proxysql-pxc-1 condition met waiting for pod/no-proxysql-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod no-proxysql-pxc-2 480 + local pod=no-proxysql-pxc-2 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo no-proxysql-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/no-proxysql-pxc-2 condition met waiting for pod/no-proxysql-pxc-2 to become Ready.Ok + desc 'check if service and statefulset created' + set +o xtrace ----------------------------------------------------------------------------------- check if service and statefulset created ----------------------------------------------------------------------------------- + compare_kubectl service/no-proxysql-pxc + local resource=service/no-proxysql-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc.yml + local new_result=/tmp/tmp.0ENVfr81ik/service_no-proxysql-pxc.yml + desc 'compare service/no-proxysql-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/no-proxysql-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k124.yml ']' + version_gt 1.22 ++ bc -l ++ echo '1.31 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml service/no-proxysql-pxc ++ mktemp + local LAST_OUT=/tmp/tmp.wajDRZvrQc ++ mktemp + local LAST_ERR=/tmp/tmp.HE9bVQtsMJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/no-proxysql-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.wajDRZvrQc + cat /tmp/tmp.HE9bVQtsMJ + rm /tmp/tmp.wajDRZvrQc /tmp/tmp.HE9bVQtsMJ + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-pxc.yml /tmp/tmp.0ENVfr81ik/service_no-proxysql-pxc.yml + log 'compare_kubectl: service/no-proxysql-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:25:27+0000]' compare_kubectl: service/no-proxysql-pxc OK [2025-11-24T18:25:27+0000] compare_kubectl: service/no-proxysql-pxc OK + compare_kubectl service/no-proxysql-proxysql + local resource=service/no-proxysql-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql.yml + local new_result=/tmp/tmp.0ENVfr81ik/service_no-proxysql-proxysql.yml + desc 'compare service/no-proxysql-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/no-proxysql-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k129.yml ']' + version_gt 1.27 ++ bc -l ++ echo '1.31 >= 1.27' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.31 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k124.yml ']' + version_gt 1.22 ++ bc -l ++ echo '1.31 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k122.yml ']' + version_gt 1.21 ++ bc -l ++ echo '1.31 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-aks.yml ']' + kubectl_bin get -o yaml service/no-proxysql-proxysql + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.IKFJ5Jyiiw ++ mktemp + local LAST_ERR=/tmp/tmp.tIMBJ1sw2s + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.IKFJ5Jyiiw + cat /tmp/tmp.tIMBJ1sw2s Error from server (NotFound): services "no-proxysql-proxysql" not found + rm /tmp/tmp.IKFJ5Jyiiw /tmp/tmp.tIMBJ1sw2s + return 1 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql.yml /tmp/tmp.0ENVfr81ik/service_no-proxysql-proxysql.yml + log 'compare_kubectl: service/no-proxysql-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:25:31+0000]' compare_kubectl: service/no-proxysql-proxysql OK [2025-11-24T18:25:31+0000] compare_kubectl: service/no-proxysql-proxysql OK + compare_kubectl configmap/no-proxysql-pxc + local resource=configmap/no-proxysql-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc.yml + local new_result=/tmp/tmp.0ENVfr81ik/configmap_no-proxysql-pxc.yml + desc 'compare configmap/no-proxysql-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare configmap/no-proxysql-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-aks.yml ']' + kubectl_bin get -o yaml configmap/no-proxysql-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.4CTxXSXNSf ++ mktemp + local LAST_ERR=/tmp/tmp.cDW2mLgbhI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml configmap/no-proxysql-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4CTxXSXNSf + cat /tmp/tmp.cDW2mLgbhI + rm /tmp/tmp.4CTxXSXNSf /tmp/tmp.cDW2mLgbhI + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc.yml /tmp/tmp.0ENVfr81ik/configmap_no-proxysql-pxc.yml + log 'compare_kubectl: configmap/no-proxysql-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:25:32+0000]' compare_kubectl: configmap/no-proxysql-pxc OK [2025-11-24T18:25:32+0000] compare_kubectl: configmap/no-proxysql-pxc OK + desc 'check if storageClass changed' + set +o xtrace ----------------------------------------------------------------------------------- check if storageClass changed ----------------------------------------------------------------------------------- + compare_kubectl statefulset/no-proxysql-pxc + local resource=statefulset/no-proxysql-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc.yml + local new_result=/tmp/tmp.0ENVfr81ik/statefulset_no-proxysql-pxc.yml + desc 'compare statefulset/no-proxysql-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/no-proxysql-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127.yml ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml statefulset/no-proxysql-pxc ++ mktemp + local LAST_OUT=/tmp/tmp.IpP277dM3V ++ mktemp + local LAST_ERR=/tmp/tmp.WKbHAE5H3B + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/no-proxysql-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.IpP277dM3V + cat /tmp/tmp.WKbHAE5H3B + rm /tmp/tmp.IpP277dM3V /tmp/tmp.WKbHAE5H3B + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127.yml /tmp/tmp.0ENVfr81ik/statefulset_no-proxysql-pxc.yml + log 'compare_kubectl: statefulset/no-proxysql-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:25:34+0000]' compare_kubectl: statefulset/no-proxysql-pxc OK [2025-11-24T18:25:34+0000] compare_kubectl: statefulset/no-proxysql-pxc OK + compare_kubectl statefulset/no-proxysql-proxysql + local resource=statefulset/no-proxysql-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql.yml + local new_result=/tmp/tmp.0ENVfr81ik/statefulset_no-proxysql-proxysql.yml + desc 'compare statefulset/no-proxysql-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/no-proxysql-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.31 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k122.yml ']' + version_gt 1.21 ++ bc -l ++ echo '1.31 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-aks.yml ']' + kubectl_bin get -o yaml statefulset/no-proxysql-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.bG7M51O9CB ++ mktemp + local LAST_ERR=/tmp/tmp.xcpaXXu83n + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.bG7M51O9CB + cat /tmp/tmp.xcpaXXu83n Error from server (NotFound): statefulsets.apps "no-proxysql-proxysql" not found + rm /tmp/tmp.bG7M51O9CB /tmp/tmp.xcpaXXu83n + return 1 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql.yml /tmp/tmp.0ENVfr81ik/statefulset_no-proxysql-proxysql.yml + log 'compare_kubectl: statefulset/no-proxysql-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:25:39+0000]' compare_kubectl: statefulset/no-proxysql-proxysql OK [2025-11-24T18:25:39+0000] compare_kubectl: statefulset/no-proxysql-proxysql OK + compare_kubectl issuer/no-proxysql-pxc-issuer + local resource=issuer/no-proxysql-pxc-issuer + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer.yml + local new_result=/tmp/tmp.0ENVfr81ik/issuer_no-proxysql-pxc-issuer.yml + desc 'compare issuer/no-proxysql-pxc-issuer-' + set +o xtrace ----------------------------------------------------------------------------------- compare issuer/no-proxysql-pxc-issuer- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k129.yml ']' + version_gt 1.27 ++ bc -l ++ echo '1.31 >= 1.27' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.31 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml issuer/no-proxysql-pxc-issuer ++ mktemp + local LAST_OUT=/tmp/tmp.Z8BX8mVUh6 ++ mktemp + local LAST_ERR=/tmp/tmp.SSXYsEgU3W + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml issuer/no-proxysql-pxc-issuer + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Z8BX8mVUh6 + cat /tmp/tmp.SSXYsEgU3W + rm /tmp/tmp.Z8BX8mVUh6 /tmp/tmp.SSXYsEgU3W + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer.yml /tmp/tmp.0ENVfr81ik/issuer_no-proxysql-pxc-issuer.yml + log 'compare_kubectl: issuer/no-proxysql-pxc-issuer OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:25:41+0000]' compare_kubectl: issuer/no-proxysql-pxc-issuer OK [2025-11-24T18:25:41+0000] compare_kubectl: issuer/no-proxysql-pxc-issuer OK + compare_kubectl certificate/no-proxysql-ssl + local resource=certificate/no-proxysql-ssl + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl.yml + local new_result=/tmp/tmp.0ENVfr81ik/certificate_no-proxysql-ssl.yml + desc 'compare certificate/no-proxysql-ssl-' + set +o xtrace ----------------------------------------------------------------------------------- compare certificate/no-proxysql-ssl- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.31 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k122.yml ']' + version_gt 1.21 ++ bc -l ++ echo '1.31 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml certificate/no-proxysql-ssl ++ mktemp + local LAST_OUT=/tmp/tmp.6oAmLtkgD5 ++ mktemp + local LAST_ERR=/tmp/tmp.4jSWFtgCJ5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml certificate/no-proxysql-ssl + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6oAmLtkgD5 + cat /tmp/tmp.4jSWFtgCJ5 + rm /tmp/tmp.6oAmLtkgD5 /tmp/tmp.4jSWFtgCJ5 + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl.yml /tmp/tmp.0ENVfr81ik/certificate_no-proxysql-ssl.yml + log 'compare_kubectl: certificate/no-proxysql-ssl OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:25:43+0000]' compare_kubectl: certificate/no-proxysql-ssl OK [2025-11-24T18:25:43+0000] compare_kubectl: certificate/no-proxysql-ssl OK + compare_kubectl certificate/no-proxysql-ssl-internal + local resource=certificate/no-proxysql-ssl-internal + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal.yml + local new_result=/tmp/tmp.0ENVfr81ik/certificate_no-proxysql-ssl-internal.yml + desc 'compare certificate/no-proxysql-ssl-internal-' + set +o xtrace ----------------------------------------------------------------------------------- compare certificate/no-proxysql-ssl-internal- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k124.yml ']' + version_gt 1.22 ++ bc -l ++ echo '1.31 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-5676", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml certificate/no-proxysql-ssl-internal ++ mktemp + local LAST_OUT=/tmp/tmp.hADhjbBNhp ++ mktemp + local LAST_ERR=/tmp/tmp.IeRoQanh74 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml certificate/no-proxysql-ssl-internal + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hADhjbBNhp + cat /tmp/tmp.IeRoQanh74 + rm /tmp/tmp.hADhjbBNhp /tmp/tmp.IeRoQanh74 + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal.yml /tmp/tmp.0ENVfr81ik/certificate_no-proxysql-ssl-internal.yml + log 'compare_kubectl: certificate/no-proxysql-ssl-internal OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-24T18:25:44+0000]' compare_kubectl: certificate/no-proxysql-ssl-internal OK [2025-11-24T18:25:44+0000] compare_kubectl: certificate/no-proxysql-ssl-internal OK + desc 'write data, read from all' + set +o xtrace ----------------------------------------------------------------------------------- write data, read from all ----------------------------------------------------------------------------------- + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' '-h no-proxysql-pxc -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local 'uri=-h no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2yCexeBYWl +++ mktemp ++ local LAST_ERR=/tmp/tmp.7y7RPFb6yA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2yCexeBYWl ++ cat /tmp/tmp.7y7RPFb6yA ++ rm /tmp/tmp.2yCexeBYWl /tmp/tmp.7y7RPFb6yA ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-857d976497-mwvqg + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h no-proxysql-pxc -uroot -proot_password' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0s8mR8jSZE +++ mktemp ++ local LAST_ERR=/tmp/tmp.xscvB2NUGZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0s8mR8jSZE ++ cat /tmp/tmp.xscvB2NUGZ ++ rm /tmp/tmp.0s8mR8jSZE /tmp/tmp.xscvB2NUGZ ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 10 + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vEKLWMd71R +++ mktemp ++ local LAST_ERR=/tmp/tmp.8VibKOkwmo ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vEKLWMd71R ++ cat /tmp/tmp.8VibKOkwmo ++ rm /tmp/tmp.vEKLWMd71R /tmp/tmp.8VibKOkwmo ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-857d976497-mwvqg + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.0ENVfr81ik/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-1.no-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-1.no-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-1.no-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-1.no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.s5fnT0eJLN +++ mktemp ++ local LAST_ERR=/tmp/tmp.4iwddpCfsE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.s5fnT0eJLN ++ cat /tmp/tmp.4iwddpCfsE ++ rm /tmp/tmp.s5fnT0eJLN /tmp/tmp.4iwddpCfsE ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.0ENVfr81ik/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-2.no-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-2.no-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1-57.sql ]] + run_mysql 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-2.no-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-2.no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IBDQa8DYhK +++ mktemp ++ local LAST_ERR=/tmp/tmp.QOg4lGTwHn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IBDQa8DYhK ++ cat /tmp/tmp.QOg4lGTwHn ++ rm /tmp/tmp.IBDQa8DYhK /tmp/tmp.QOg4lGTwHn ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-857d976497-mwvqg + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.0ENVfr81ik/select-1.sql + compare_mysql_cmd max_allowed_packet-2 'SELECT @@max_allowed_packet;' '-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local command_id=max_allowed_packet-2 + local 'command=SELECT @@max_allowed_packet;' + local 'uri=-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/max_allowed_packet-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 8\.0 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc5.7 =~ 5\.7 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/max_allowed_packet-2-57.sql ]] + run_mysql 'SELECT @@max_allowed_packet;' '-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local 'command=SELECT @@max_allowed_packet;' + local 'uri=-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.A6sisJgJxZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.i2n3JeO1O5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.A6sisJgJxZ ++ cat /tmp/tmp.i2n3JeO1O5 ++ rm /tmp/tmp.A6sisJgJxZ /tmp/tmp.i2n3JeO1O5 ++ return 0 + client_pod=pxc-client-857d976497-mwvqg + wait_pod pxc-client-857d976497-mwvqg + local pod=pxc-client-857d976497-mwvqg + local max_retry=480 + local ns= ++ echo pxc-client-857d976497-mwvqg ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-857d976497-mwvqg condition met waiting for pod/pxc-client-857d976497-mwvqg to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.0ENVfr81ik/max_allowed_packet-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/compare/max_allowed_packet-2.sql /tmp/tmp.0ENVfr81ik/max_allowed_packet-2.sql + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/conf/no-proxysql.yml ++ mktemp + local LAST_OUT=/tmp/tmp.JvJCce0GbF ++ mktemp + local LAST_ERR=/tmp/tmp.Dt5FkusSrd + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/init-deploy/conf/no-proxysql.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JvJCce0GbF perconaxtradbcluster.pxc.percona.com "no-proxysql" deleted from init-deploy-5676 namespace + cat /tmp/tmp.Dt5FkusSrd + rm /tmp/tmp.JvJCce0GbF /tmp/tmp.Dt5FkusSrd + return 0 + destroy init-deploy-5676 + local namespace=init-deploy-5676 + local ignore_logs=true + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' + grep -v level=info + sort -u + tee /tmp/tmp.0ENVfr81ik/operator.log + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + grep -v 'the object has been modified' + grep -v 'get backup status: Job.batch' ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.S0EtB5Fh50 +++ mktemp ++ local LAST_ERR=/tmp/tmp.DaOnaTzKgE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.S0EtB5Fh50 ++ cat /tmp/tmp.DaOnaTzKgE ++ rm /tmp/tmp.S0EtB5Fh50 /tmp/tmp.DaOnaTzKgE ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-5d8c4fb844-8qvwn ++ mktemp + local LAST_OUT=/tmp/tmp.mIsd2ho1Vq ++ mktemp + local LAST_ERR=/tmp/tmp.7rVD2GPt5k + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-5d8c4fb844-8qvwn + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mIsd2ho1Vq + cat /tmp/tmp.7rVD2GPt5k + rm /tmp/tmp.mIsd2ho1Vq /tmp/tmp.7rVD2GPt5k + return 0 2025-11-24T18:09:45.669Z INFO setup Manager starting up {"gitCommit": "ff96f032c4f13c9e8ce3df749a12fd530376ef2e", "gitBranch": "PR-2234-ff96f032", "buildTime": "2025-11-24T17:27:09Z", "goVersion": "go1.25.4", "os": "linux", "arch": "amd64"} 2025-11-24T18:09:45.669Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.31.13-gke.1454000"} 2025-11-24T18:09:45.672Z INFO setup Registering Components. 2025-11-24T18:09:46.113Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2025-11-24T18:09:46.113Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2025-11-24T18:09:46.113Z INFO controller-runtime.metrics Starting metrics server 2025-11-24T18:09:46.113Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2025-11-24T18:09:46.113Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2025-11-24T18:09:46.113Z INFO controller-runtime.webhook Starting webhook server 2025-11-24T18:09:46.113Z INFO setup Starting the Cmd. 2025-11-24T18:09:46.113Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2025-11-24T18:09:46.114Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2025-11-24T18:09:46.214Z INFO attempting to acquire leader lease pxc-operator/08db1feb.percona.com... 2025-11-24T18:09:46.243Z DEBUG events percona-xtradb-cluster-operator-5d8c4fb844-8qvwn_d8b359db-8c6e-4018-8017-08a845b0cd2a became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"c8074f9a-cda1-41cf-bea0-2065dab66c4f","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1764007786237439009"}, "reason": "LeaderElection"} 2025-11-24T18:09:46.243Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.Secret"} 2025-11-24T18:09:46.243Z INFO successfully acquired lease pxc-operator/08db1feb.percona.com 2025-11-24T18:09:46.244Z INFO Starting EventSource {"controller": "pxcbackup-controller", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2025-11-24T18:09:46.244Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.PerconaXtraDBCluster"} 2025-11-24T18:09:46.244Z INFO Starting EventSource {"controller": "pxcrestore-controller", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2025-11-24T18:09:46.344Z INFO Starting Controller {"controller": "pxcbackup-controller"} 2025-11-24T18:09:46.345Z INFO Starting Controller {"controller": "pxc-controller"} 2025-11-24T18:09:46.345Z INFO Starting Controller {"controller": "pxcrestore-controller"} 2025-11-24T18:09:46.345Z INFO Starting workers {"controller": "pxcbackup-controller", "worker count": 1} 2025-11-24T18:09:46.345Z INFO Starting workers {"controller": "pxc-controller", "worker count": 1} 2025-11-24T18:09:46.345Z INFO Starting workers {"controller": "pxcrestore-controller", "worker count": 1} 2025-11-24T18:12:00.953Z INFO Set CR version {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6e816078-7939-4766-bf87-a55e7d809c13", "version": "1.19.0"} 2025-11-24T18:12:01.393Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6e816078-7939-4766-bf87-a55e7d809c13"} 2025-11-24T18:12:04.477Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6e816078-7939-4766-bf87-a55e7d809c13"} 2025-11-24T18:12:04.521Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6e816078-7939-4766-bf87-a55e7d809c13"} 2025-11-24T18:12:07.623Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6e816078-7939-4766-bf87-a55e7d809c13", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-11-24T18:12:07.695Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6e816078-7939-4766-bf87-a55e7d809c13", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-11-24T18:12:08.248Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6e816078-7939-4766-bf87-a55e7d809c13", "error": "failed to deploy proxysql: updatePod for proxysql: reconcile config: reconcile autotune config: create or update configmap: configmaps \"auto-some-name-pxc\" already exists", "errorVerbose": "configmaps \"auto-some-name-pxc\" already exists\ncreate or update configmap\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileAutotuneConfigMap\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:92\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:25\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:52\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:577\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:595\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:379\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile autotune config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:27\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:52\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:577\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:595\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:379\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:54\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:577\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:595\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:379\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nupdatePod for proxysql\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:578\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:595\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:379\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nfailed to deploy proxysql\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).deploy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:596\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:379\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-24T18:12:08.351Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "334c2bf1-35b9-40d0-a49b-a71f02acecfc", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-24T18:12:08.441Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "334c2bf1-35b9-40d0-a49b-a71f02acecfc", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-24T18:12:08.540Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "334c2bf1-35b9-40d0-a49b-a71f02acecfc", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-24T18:12:08.574Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "334c2bf1-35b9-40d0-a49b-a71f02acecfc", "object": "some-name-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-24T18:12:08.631Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "334c2bf1-35b9-40d0-a49b-a71f02acecfc", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-24T18:12:08.747Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "334c2bf1-35b9-40d0-a49b-a71f02acecfc", "object": "some-name-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-24T18:12:09.243Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "02226b1b-d85b-40ab-bded-3c40e92a55d2", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-24T18:12:09.314Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "02226b1b-d85b-40ab-bded-3c40e92a55d2", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-24T18:13:26.071Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "362981c2-a9f9-4abe-846c-58676cb5102c", "user": "operator"} 2025-11-24T18:13:26.135Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "362981c2-a9f9-4abe-846c-58676cb5102c", "user": "monitor"} 2025-11-24T18:13:26.189Z INFO User monitor: granted privileges {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "362981c2-a9f9-4abe-846c-58676cb5102c"} 2025-11-24T18:13:26.250Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "362981c2-a9f9-4abe-846c-58676cb5102c", "user": "xtrabackup"} 2025-11-24T18:13:26.308Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "362981c2-a9f9-4abe-846c-58676cb5102c"} 2025-11-24T18:13:26.316Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "362981c2-a9f9-4abe-846c-58676cb5102c", "err": "get primary pxc pod: not found"} 2025-11-24T18:13:30.926Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "4c6e57e8-23c3-4931-8453-a567158372de", "err": "get primary pxc pod: not found"} 2025-11-24T18:13:36.121Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "dcc4f5d2-86e8-4070-90a4-26087c7f579e", "err": "get primary pxc pod: not found"} 2025-11-24T18:13:41.253Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "556feca1-9174-4bb9-b344-0497e125a5fe", "err": "get primary pxc pod: not found"} 2025-11-24T18:15:52.055Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "d1f8f325-044c-4aa6-81ab-78ae4f47a822", "user": "root"} 2025-11-24T18:15:52.122Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "d1f8f325-044c-4aa6-81ab-78ae4f47a822", "user": "replication"} 2025-11-24T18:15:52.170Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "d1f8f325-044c-4aa6-81ab-78ae4f47a822", "new version": "5.7.44-48-57"} 2025-11-24T18:15:53.972Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "d1f8f325-044c-4aa6-81ab-78ae4f47a822"} 2025-11-24T18:15:58.701Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "959b0130-f484-46c0-89b2-a73687baf8ff"} 2025-11-24T18:16:04.049Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "3bd9aafa-827f-4918-b736-99ec1feed3d7"} 2025-11-24T18:16:09.158Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "37ce2013-a914-463f-9413-b281752a49ad"} 2025-11-24T18:16:14.369Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "7b2abc96-4417-4e46-bd38-9f415a9c1a41"} 2025-11-24T18:16:19.872Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "98eb2db5-e01e-4828-aa75-d9c8c34a9522"} 2025-11-24T18:16:25.003Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "2bb90971-c714-4828-87cb-23bf94ee3f98"} 2025-11-24T18:16:30.497Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "1898da0a-1847-464a-b29c-cbeaf0d49758"} 2025-11-24T18:16:35.552Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "3ff876cb-e17a-44a7-b132-b21f849b2b36"} 2025-11-24T18:16:40.646Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "bdbd5824-09ec-41b2-bd7f-8cf54ed90cd5"} 2025-11-24T18:16:46.051Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "a4ff5685-be4a-46b2-b160-3de25c8eb35a"} 2025-11-24T18:16:51.146Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "0f29b29a-1bb6-4687-9578-81ef262da3df"} 2025-11-24T18:16:56.759Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "5212dced-d18f-40ae-b201-3dac26d6cb2b"} 2025-11-24T18:17:02.092Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "0809a4c0-b9f7-406c-a7bd-d661c470322c"} 2025-11-24T18:17:07.753Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "03d5a2a9-71fe-4a58-8256-066e3321ac7b"} 2025-11-24T18:17:12.151Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "9b1477dc-91be-443d-a44e-a2bfe9c8b6f1"} 2025-11-24T18:17:17.804Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "b6b8d591-b9d0-46a5-9cc2-9281fe362bc8"} 2025-11-24T18:17:23.052Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "a3e07510-b0ba-4525-a7ba-203a93481252"} 2025-11-24T18:17:28.082Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "0b141812-cedb-4abb-a3ef-c8665c7d8417"} 2025-11-24T18:17:33.353Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "ec76cd82-ed03-4187-a18a-74ce3fcdab62"} 2025-11-24T18:17:38.854Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "611e1619-8d15-4304-929e-f5ad39034a99"} 2025-11-24T18:17:43.901Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "4675fd54-a942-4ae1-8d81-77148f7e0e1c"} 2025-11-24T18:17:49.200Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "0acec04c-07b1-4014-b26a-7a9ba997e5fc"} 2025-11-24T18:17:54.761Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "ffe8139d-63f4-403e-a0eb-59e1e3942a11"} 2025-11-24T18:18:00.014Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "44911347-b4c3-4ea7-9b16-800141fe3572"} 2025-11-24T18:18:05.051Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "a0e29c68-e1d6-4d3f-8a03-3dbdbaa2cee2"} 2025-11-24T18:18:10.269Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "0cad6e9f-a163-4c13-81fb-d8ea750d2533"} 2025-11-24T18:18:16.051Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "7bdbe5a3-0fb9-44d4-93a8-fe6879b38a6e"} 2025-11-24T18:18:20.760Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "3b78232c-8f55-495e-a22d-28c0855c73a2"} 2025-11-24T18:18:26.353Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "9bd66605-a43b-4b79-8bab-fab99c5ec132"} 2025-11-24T18:18:32.053Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "1b20fea9-1d01-4681-bb22-36174c33c72a"} 2025-11-24T18:18:37.753Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "ae56605d-8873-4d41-85c5-a2c50b91dd7b"} 2025-11-24T18:18:42.759Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "2e723ff0-6ecd-4803-a847-82fddefe4cf4"} 2025-11-24T18:18:48.049Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "3d5dc27c-0927-4af9-b61d-70a9b95bf0bf"} 2025-11-24T18:18:53.598Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "96658b30-4d59-4738-ab2b-d8579a99cfd8"} 2025-11-24T18:18:58.382Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "c0df339b-f0f6-423d-b68f-f3d9db7f2d2f"} 2025-11-24T18:19:03.650Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "fa78de27-1b7b-4791-95fe-9d4809aa9f18"} 2025-11-24T18:19:08.746Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "f3aebfa6-bc07-4562-a528-95aad5d70b3a"} 2025-11-24T18:19:14.095Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "8ecf02a2-537d-4f7e-92af-5e468e011928"} 2025-11-24T18:19:19.350Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "75b432c9-dcfb-4fe3-bb51-bb278d7e6ae4"} 2025-11-24T18:19:24.561Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "d4dadb34-b266-423f-92d0-ef72d4fa47be"} 2025-11-24T18:19:30.126Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "2a9243d0-7da0-43f8-8a59-1649987a78e7"} 2025-11-24T18:19:35.172Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "2a3701fc-361e-4ff0-a2a7-cd9a7aca1e17"} 2025-11-24T18:19:40.382Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "e77fa9ba-c6e9-4301-a8ac-7cb7966d5875"} 2025-11-24T18:19:45.673Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "0714aabc-c66c-43a9-aadf-234e7af7ec57"} 2025-11-24T18:19:50.644Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "f1151c2a-2beb-42e2-8008-94791bfee8be"} 2025-11-24T18:19:55.869Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "5a5ba913-5390-4067-a9de-bf7ad5591231"} 2025-11-24T18:21:00.028Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "b58546fc-8bf8-4813-a929-9571ba1f912e"} 2025-11-24T18:21:05.206Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "697582e8-b930-49eb-be30-a5eabf96b015"} 2025-11-24T18:21:10.452Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "a35deedb-cdf1-4c4e-8473-9ce93a6629ae"} 2025-11-24T18:21:16.054Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "2256d27c-c629-4eee-a2f3-c4ec6424d9b6"} 2025-11-24T18:21:20.853Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "3347da9c-c292-4516-a591-a2ba1a0496a2"} 2025-11-24T18:21:26.048Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "71662df2-1f74-4f3d-8b4c-8395d1df7a87"} 2025-11-24T18:21:31.350Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "9daeeef4-a1f4-4a62-b512-32f1c8fe1443"} 2025-11-24T18:21:36.951Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "cfbad96d-6fd5-4db7-a8a9-13221f1ac6af"} 2025-11-24T18:21:37.057Z INFO Set CR version {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "affdeaca-f5ac-46e2-b613-2a3fbddbae0f", "version": "1.19.0"} 2025-11-24T18:21:37.298Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "affdeaca-f5ac-46e2-b613-2a3fbddbae0f"} 2025-11-24T18:21:40.346Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "affdeaca-f5ac-46e2-b613-2a3fbddbae0f"} 2025-11-24T18:21:40.385Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "affdeaca-f5ac-46e2-b613-2a3fbddbae0f"} 2025-11-24T18:21:43.386Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "affdeaca-f5ac-46e2-b613-2a3fbddbae0f", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-11-24T18:21:43.409Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "affdeaca-f5ac-46e2-b613-2a3fbddbae0f", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-11-24T18:21:44.003Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "affdeaca-f5ac-46e2-b613-2a3fbddbae0f", "error": "pxc upgrade error: reconcile config: reconcile custom config: create or update config map: configmaps \"no-proxysql-pxc\" already exists", "errorVerbose": "configmaps \"no-proxysql-pxc\" already exists\ncreate or update config map\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileCustomConfigMap\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:115\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:31\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:52\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:385\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile custom config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:33\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:52\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:385\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:54\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:385\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\npxc upgrade error\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:387\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-11-24T18:21:44.417Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "5b00c8dc-8576-4320-b04d-37bdab62df0e", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-24T18:21:44.488Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "5b00c8dc-8576-4320-b04d-37bdab62df0e", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-24T18:21:44.556Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "5b00c8dc-8576-4320-b04d-37bdab62df0e", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-24T18:21:44.586Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "5b00c8dc-8576-4320-b04d-37bdab62df0e", "object": "no-proxysql-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-24T18:21:45.233Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "691edba8-0337-4387-9cf4-9dcb3c98b304", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-24T18:21:46.058Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "0281dfd6-748a-414e-9cfc-ae0f5348703c"} 2025-11-24T18:21:50.853Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "1919153f-8804-410f-9c61-846b3b3b018b"} 2025-11-24T18:21:56.349Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "c4cb3666-e9ac-4020-b649-d4cf970752b1"} 2025-11-24T18:22:01.800Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6c840d47-1149-4e0f-a46c-32e8fcc70ead"} 2025-11-24T18:22:07.549Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "a0e0920e-0d75-492c-ad8e-fee461fb08d3"} 2025-11-24T18:22:12.668Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "c7c66441-6fd9-4b26-9f59-b8703e834493"} 2025-11-24T18:22:17.846Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "a7fed8c8-6434-4227-afc5-79249c49d5e9"} 2025-11-24T18:22:23.155Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6deeb141-0d33-4a94-9e89-c52f31bc0787"} 2025-11-24T18:22:28.286Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "d0a8561c-6bb2-4159-bd6e-5a7dbab5520b"} 2025-11-24T18:22:33.350Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "323153b5-d6c9-4d90-aaa1-f32253a7dd96"} 2025-11-24T18:22:38.866Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "def6cec8-724d-411f-b8a5-9c4044836074"} 2025-11-24T18:22:43.950Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "feaa72a8-af9c-481f-8771-b7507357942d"} 2025-11-24T18:22:49.300Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "f0302216-3a3d-41fb-bf6b-dd1e87706d1d"} 2025-11-24T18:22:54.377Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "21cab636-b5d0-41b0-9a40-ff011c4c09f3"} 2025-11-24T18:22:58.431Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "74ad320e-7225-4a35-ac35-473bc15c969b", "user": "operator"} 2025-11-24T18:22:58.473Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "74ad320e-7225-4a35-ac35-473bc15c969b", "user": "monitor"} 2025-11-24T18:22:58.523Z INFO User monitor: granted privileges {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "74ad320e-7225-4a35-ac35-473bc15c969b"} 2025-11-24T18:22:58.567Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "74ad320e-7225-4a35-ac35-473bc15c969b", "user": "xtrabackup"} 2025-11-24T18:22:58.615Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "74ad320e-7225-4a35-ac35-473bc15c969b"} 2025-11-24T18:22:58.617Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "74ad320e-7225-4a35-ac35-473bc15c969b", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:22:59.600Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "d6ff5dec-b288-4f5c-959a-88c26651eb7c"} 2025-11-24T18:23:03.538Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "b8ba5b45-b667-44c4-8220-4c62654e099c", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:04.996Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "f8806ee2-7533-4433-aab3-ddf75531ebab"} 2025-11-24T18:23:08.803Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "8ce11ac5-b68d-4324-ab46-a74876e72310", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:10.295Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "230bfb01-1ba4-4de1-b873-baeb02c73bbb"} 2025-11-24T18:23:14.037Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "23c499de-4274-4d03-b058-29a179eee434", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:16.161Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "bf7c407a-1f56-4f04-af87-6221caafe304"} 2025-11-24T18:23:19.317Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "4cb9ded6-cc93-4550-8957-c6cb7d433edf", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:20.704Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "c356c499-1740-4f25-a9ef-0ac86537bb31"} 2025-11-24T18:23:24.551Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "ac8b8b79-eb81-45cc-bc4f-7cfc6c78f153", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:26.247Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "b71710ca-29cc-4b60-beba-de87be0cad38"} 2025-11-24T18:23:29.830Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "579a12e1-051a-44af-957f-08b23f7da6e0", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:31.285Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "7861f8b1-c4ae-4bf2-ad88-50bbebd94db1"} 2025-11-24T18:23:35.344Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "0deddcef-2d8e-4f3f-a2b5-14298697b3ea", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:36.753Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "adfa50d6-58dc-456e-9915-e22f613d02a3"} 2025-11-24T18:23:40.439Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "424013b4-9289-476c-b592-e3407406d2ff", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:41.746Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "f803da6d-09a3-4231-b3cd-3709347c6f4a"} 2025-11-24T18:23:45.695Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "71741d77-6fb9-4344-afa0-203c2fba61f0", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:47.299Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "ca8ffb57-d2e4-4ac5-a0d0-8369c00bbf6c"} 2025-11-24T18:23:50.950Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "e16b1c48-8807-4519-9798-36b6218966f5", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:52.148Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "11f533dc-7e04-47b9-a3fa-86296c204347"} 2025-11-24T18:23:56.215Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "89731662-5734-4e04-bc02-e100d9323c96", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:23:57.847Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "e9d97804-3bfb-4580-acd5-01048c55785a"} 2025-11-24T18:24:01.474Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "9c2024b8-c421-48c0-84c1-aff0b1992938", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:03.153Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "164ef146-f254-4986-b7cc-491762f6de1b"} 2025-11-24T18:24:06.889Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "f6874740-bcaf-4581-a4eb-62f37ba20a74", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:08.149Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "e6f1e50d-67c4-4d32-a948-6b94d94f6087"} 2025-11-24T18:24:12.092Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "0de1a299-5ece-4c35-a155-0e55b848c4b1", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:12.818Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "62b8ab54-91da-48fc-8041-1f48fbdef05b", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:13.576Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "e3ad2a24-1517-424d-8348-df8d2d4c5f55"} 2025-11-24T18:24:17.499Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "60d71c09-de35-4396-80d4-ead3ad5a92cc", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:18.753Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "572dd626-15eb-4de6-9962-5a75c3c67b08"} 2025-11-24T18:24:22.739Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "1bd5403b-155a-4a79-a921-0db6bc9a1939", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:24.589Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "cdee152e-d7ee-4a3f-ab9b-edb8b2037dac"} 2025-11-24T18:24:28.010Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "58d468e6-7550-46a1-9d0c-f3f9c3bc0597", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:29.504Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "77d7c242-61e9-4bdd-8983-d813adcb0fd6"} 2025-11-24T18:24:33.265Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "4543a78f-096e-477c-b745-24c7668ce902", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:34.802Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "49fe9621-8101-4b2a-b442-40419ceb97a2"} 2025-11-24T18:24:38.536Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "b31b7768-df16-4c48-a4d4-7e318de9d22e", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:40.092Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "5ee22e49-c9d5-4b0e-b80c-0999c3ad7d20"} 2025-11-24T18:24:43.776Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "6478de3a-3f13-475f-83ae-47c96382e6d1", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:45.349Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "8e51d50b-4441-4a5b-ad8f-c71f71a3ac0c"} 2025-11-24T18:24:49.004Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "b9d6a9c9-970f-4924-a04a-2dad2792e670", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:50.461Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "b088155c-9b69-4da1-9675-d8ba0f69cdca"} 2025-11-24T18:24:54.268Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "bf398e3b-ff82-4887-8dee-06a28f2f46bf", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:24:56.056Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "1d4ea33c-e496-41d1-b9c7-982f53f82681"} 2025-11-24T18:24:59.525Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "b56cded2-fe49-4056-a694-af49dd1a104a", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:01.277Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "af082cad-7aa4-479c-ab50-a9d36c758769"} 2025-11-24T18:25:04.736Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "b318d5d3-3f04-495f-a80d-ec7aedfcc805", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:06.503Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "fdcdafae-4cf9-49f5-9387-0877c47d2478"} 2025-11-24T18:25:10.040Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "34f24866-2724-4173-8df1-5226fbc99410", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:11.778Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "83e2a2e0-c2a9-4fb2-81be-f572ce1e7f60"} 2025-11-24T18:25:15.844Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "5576cd24-1d6d-4ec1-ace4-49700138cb3e", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:17.449Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "6c010b34-9a2c-4116-9802-469b59776865"} 2025-11-24T18:25:21.125Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "8e7eceb0-6c0e-4eea-bb21-d7211181effa", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:22.460Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "f45230a8-e877-4a16-b02a-6cb3478ac4b5", "user": "root"} 2025-11-24T18:25:22.510Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "f45230a8-e877-4a16-b02a-6cb3478ac4b5", "user": "replication"} 2025-11-24T18:25:22.512Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "f45230a8-e877-4a16-b02a-6cb3478ac4b5", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:22.521Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "f45230a8-e877-4a16-b02a-6cb3478ac4b5", "new version": "5.7.44-48-57"} 2025-11-24T18:25:22.978Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "411980f3-a5f6-4b41-89fd-c4622e68c347"} 2025-11-24T18:25:23.804Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "a0ecce83-b916-4056-bc23-e667d91c7118", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:26.901Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "a86ae81a-3beb-416a-bc85-5fb98109b490", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:28.152Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "140cb6c9-b4bb-4e10-ac67-52d830becd33"} 2025-11-24T18:25:32.318Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "fb40a297-98ae-4e62-a864-271ecfe2ec35", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:33.373Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "454e7d18-066e-4630-a4c6-0aa033451f1a"} 2025-11-24T18:25:37.514Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "2462c45b-af6e-44a5-bec6-1f7383b288bd", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:38.193Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "63736ae6-e661-41ad-ad47-17cd09f57b22"} 2025-11-24T18:25:42.721Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "f0a599b0-7489-47f5-8d0b-29437f207687", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:43.752Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "b3e6b27a-9f10-4d00-8c26-6eba131661f3"} 2025-11-24T18:25:47.936Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "5be20c46-d2f6-4b8e-b849-e896be968f46", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:49.156Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "7ecb55a3-f09b-4903-9bfb-f53dc6dbae6c"} 2025-11-24T18:25:53.125Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "532a3f38-a664-4a3a-a131-76f2ec169fcb", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:54.168Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "acbbd5fa-f0d7-4540-9562-1fefb6ea0a65"} 2025-11-24T18:25:58.327Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "1a5c7ebf-3b63-41a0-8115-de685e4e587e", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:25:59.798Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "cf384910-b211-46fe-96da-ba79354496c1"} 2025-11-24T18:26:03.544Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "6bf55e58-97d1-45e8-9db7-eb5bb15ab78a", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:26:04.982Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "50717e9c-18e8-45cc-98e1-21943dd09893"} 2025-11-24T18:26:08.770Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "b11c550b-3607-4efc-93e7-a7285f76cb18", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:26:10.089Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "76302cb5-1d2d-4efc-86a1-b73abe2088ff"} 2025-11-24T18:26:14.050Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "e980eb90-c0a3-42a6-abc6-3ddfecf19531", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:26:15.476Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "51924196-6762-4f7e-a978-f7ba6d0431d5"} 2025-11-24T18:26:19.290Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "df74ae2a-c649-48b0-9903-eb4c8fb1863c", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:26:20.458Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "19d943ec-fdb9-46a4-8d57-d9a2ddced482"} 2025-11-24T18:26:24.542Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "d12ad3ff-f045-452a-82aa-c4b1c675081e", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:26:25.610Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "0c857960-6ba8-40b7-9588-44b0161cf146"} 2025-11-24T18:26:29.727Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "09d46b42-d6a8-409f-a11c-03a66c4b68f1", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:26:31.308Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "ff5623fd-e591-41b6-9f2c-e58b43196cea"} 2025-11-24T18:26:35.404Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "c128ea7c-de0e-4823-a58f-366218ab458c", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:26:36.398Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "f0e538ff-4571-49c6-a662-a4e57666c810"} 2025-11-24T18:26:40.582Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "341af94c-fb6e-4af2-b863-db946c8c924d", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:26:41.457Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "1920f7a3-e5b0-43b1-bd10-bf7051feffb9"} 2025-11-24T18:26:45.763Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "no-proxysql", "reconcileID": "89b0fd20-2310-409f-b14f-c023131e10a6", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-11-24T18:26:46.755Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "ee7f355b-7d3b-42e9-a666-0e24f7ecff42"} 2025-11-24T18:26:51.904Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "init-deploy-5676", "name": "some-name", "reconcileID": "0b5ce462-f849-451a-8808-02a44c9cf298"} /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:474 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 + grep -v NAMESPACE + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n init-deploy-5676 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.eYzZZc0vb7 ++ mktemp + local LAST_ERR=/tmp/tmp.PuFUNUjXrF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.eYzZZc0vb7 perconaxtradbcluster.pxc.percona.com "some-name" deleted from init-deploy-5676 namespace + cat /tmp/tmp.PuFUNUjXrF + rm /tmp/tmp.eYzZZc0vb7 /tmp/tmp.PuFUNUjXrF + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.vTynzKVIud ++ mktemp + local LAST_ERR=/tmp/tmp.IOjSv3UmfG + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vTynzKVIud No resources found + cat /tmp/tmp.IOjSv3UmfG + rm /tmp/tmp.vTynzKVIud /tmp/tmp.IOjSv3UmfG + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.JEyJY04uPK ++ mktemp + local LAST_ERR=/tmp/tmp.T0TUKW4hpz + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.JEyJY04uPK No resources found + cat /tmp/tmp.T0TUKW4hpz + rm /tmp/tmp.JEyJY04uPK /tmp/tmp.T0TUKW4hpz + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.i7NI0w9ql2 ++ mktemp + local LAST_ERR=/tmp/tmp.QcmYyKhzW3 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.i7NI0w9ql2 validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.QcmYyKhzW3 + rm /tmp/tmp.i7NI0w9ql2 /tmp/tmp.QcmYyKhzW3 + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + rm -rf /tmp/tmp.0ENVfr81ik + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator + kubectl_bin delete --grace-period=0 --force=true namespace init-deploy-5676 ++ mktemp + local LAST_OUT=/tmp/tmp.DJ6NflUNhs ++ mktemp + local LAST_ERR=/tmp/tmp.ku6LPZAcHu + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace init-deploy-5676 + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- ++ mktemp + local LAST_OUT=/tmp/tmp.2UYkTmMsIH ++ mktemp + local LAST_ERR=/tmp/tmp.NvDbfhBsfz + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator