Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/logs/init-deploy-8-0.log Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + create_infra init-deploy-9470 + local ns=init-deploy-9470 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n init-deploy-26962 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.i0exryKj3p ++ mktemp + local LAST_ERR=/tmp/tmp.kWk0EcngDN + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.i0exryKj3p perconaxtradbcluster.pxc.percona.com "some-name" deleted from init-deploy-26962 namespace + cat /tmp/tmp.kWk0EcngDN + rm /tmp/tmp.i0exryKj3p /tmp/tmp.kWk0EcngDN + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.n65DWwMeGH ++ mktemp + local LAST_ERR=/tmp/tmp.jLfHgF4kMh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.n65DWwMeGH No resources found + cat /tmp/tmp.jLfHgF4kMh + rm /tmp/tmp.n65DWwMeGH /tmp/tmp.jLfHgF4kMh + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.fEEH4EMbSL ++ mktemp + local LAST_ERR=/tmp/tmp.XFsWh79Bs7 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.fEEH4EMbSL No resources found + cat /tmp/tmp.XFsWh79Bs7 + rm /tmp/tmp.fEEH4EMbSL /tmp/tmp.XFsWh79Bs7 + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ sed s/NAMESPACE// ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep validate-auth ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl api-resources ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl get clusterrole ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator ++ mktemp + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + local LAST_OUT=/tmp/tmp.HW4u4P7GUd ++ mktemp + local LAST_ERR=/tmp/tmp.H6u9kV9YFA + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator + awk '{print$1}' + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.UnG1E20Z4Y ++ mktemp + local LAST_ERR=/tmp/tmp.BBiIS6dzmp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.UnG1E20Z4Y + cat /tmp/tmp.BBiIS6dzmp + rm /tmp/tmp.UnG1E20Z4Y /tmp/tmp.BBiIS6dzmp + return 0 namespace "cert-manager" deleted namespace "init-deploy-26962" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HW4u4P7GUd namespace "pxc-operator" deleted + cat /tmp/tmp.H6u9kV9YFA + rm /tmp/tmp.HW4u4P7GUd /tmp/tmp.H6u9kV9YFA + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.0Yfg5M2R05 ++ mktemp + local LAST_ERR=/tmp/tmp.eT4YIIV9ww + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0Yfg5M2R05 namespace/pxc-operator created + cat /tmp/tmp.eT4YIIV9ww + rm /tmp/tmp.0Yfg5M2R05 /tmp/tmp.eT4YIIV9ww + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Sa2m763M0Z +++ mktemp ++ local LAST_ERR=/tmp/tmp.TWboIhTqLA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Sa2m763M0Z ++ cat /tmp/tmp.TWboIhTqLA ++ rm /tmp/tmp.Sa2m763M0Z /tmp/tmp.TWboIhTqLA ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2221-abd79966-3-cluster3 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.rQnYtFcw77 ++ mktemp + local LAST_ERR=/tmp/tmp.8O1UrFUWaW + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2221-abd79966-3-cluster3 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rQnYtFcw77 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2221-abd79966-3-cluster3" modified. + cat /tmp/tmp.8O1UrFUWaW + rm /tmp/tmp.rQnYtFcw77 /tmp/tmp.8O1UrFUWaW + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.znWHKdispG ++ mktemp + local LAST_ERR=/tmp/tmp.sg2Vkiaq2o + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.znWHKdispG customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.sg2Vkiaq2o + rm /tmp/tmp.znWHKdispG /tmp/tmp.sg2Vkiaq2o + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.QME4j0fVaN ++ mktemp + local LAST_ERR=/tmp/tmp.VL2nP0aFpl + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QME4j0fVaN clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.VL2nP0aFpl + rm /tmp/tmp.QME4j0fVaN /tmp/tmp.VL2nP0aFpl + return 0 + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2221-abd79966^' + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + kubectl_bin apply -f - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' ++ mktemp + local LAST_OUT=/tmp/tmp.6dpXGRLQ5z + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/deploy/cw-operator.yaml ++ mktemp + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + local LAST_ERR=/tmp/tmp.qCfAxlQqgI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.6dpXGRLQ5z deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.qCfAxlQqgI + rm /tmp/tmp.6dpXGRLQ5z /tmp/tmp.qCfAxlQqgI + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.sFcXDt5q9J ++ mktemp + local LAST_ERR=/tmp/tmp.JJsliy5mBe + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sFcXDt5q9J pod/percona-xtradb-cluster-operator-bbcb56d74-drjwh condition met + cat /tmp/tmp.JJsliy5mBe + rm /tmp/tmp.sFcXDt5q9J /tmp/tmp.JJsliy5mBe + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.neos3Rc5Gu +++ mktemp ++ local LAST_ERR=/tmp/tmp.Wl7QUAbGHx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.neos3Rc5Gu ++ cat /tmp/tmp.Wl7QUAbGHx ++ rm /tmp/tmp.neos3Rc5Gu /tmp/tmp.Wl7QUAbGHx ++ return 0 + wait_pod percona-xtradb-cluster-operator-bbcb56d74-drjwh 480 pxc-operator + local pod=percona-xtradb-cluster-operator-bbcb56d74-drjwh + local max_retry=480 + local ns=pxc-operator ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo percona-xtradb-cluster-operator-bbcb56d74-drjwh ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-bbcb56d74-drjwh condition met waiting for pod/percona-xtradb-cluster-operator-bbcb56d74-drjwh to become Ready.Ok + sleep 3 + create_namespace init-deploy-9470 + local namespace=init-deploy-9470 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ grep chaos-mesh ++ kubectl get MutatingWebhookConfiguration ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl api-resources ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + grep -E -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + '[' -n '' ']' + desc 'cleaned up old namespaces init-deploy-9470' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces init-deploy-9470 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace init-deploy-9470 ++ mktemp + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.L4PLxokuKN ++ mktemp + local LAST_OUT=/tmp/tmp.sS3V7O7U1z ++ mktemp + local LAST_ERR=/tmp/tmp.bWbKAavAft + local exit_status=0 ++ seq 0 2 + local LAST_ERR=/tmp/tmp.o27iaKuQIF + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace init-deploy-9470 + for i in '$(seq 0 2)' + set +e + kubectl get ns + xargs kubectl delete ns + awk '{print$1}' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sS3V7O7U1z + cat /tmp/tmp.o27iaKuQIF + rm /tmp/tmp.sS3V7O7U1z /tmp/tmp.o27iaKuQIF + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace init-deploy-9470 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace init-deploy-9470 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.L4PLxokuKN + cat /tmp/tmp.bWbKAavAft Error from server (NotFound): namespaces "init-deploy-9470" not found + rm /tmp/tmp.L4PLxokuKN /tmp/tmp.bWbKAavAft + return 1 + : + wait_for_delete namespace/init-deploy-9470 + local res=namespace/init-deploy-9470 + echo -n 'waiting for namespace/init-deploy-9470 to be deleted' waiting for namespace/init-deploy-9470 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "init-deploy-9470" not found + desc 'create namespace init-deploy-9470' + set +o xtrace ----------------------------------------------------------------------------------- create namespace init-deploy-9470 ----------------------------------------------------------------------------------- + kubectl_bin create namespace init-deploy-9470 ++ mktemp + local LAST_OUT=/tmp/tmp.K4SSm3VElO ++ mktemp + local LAST_ERR=/tmp/tmp.9IYnSDgweB + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace init-deploy-9470 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.K4SSm3VElO namespace/init-deploy-9470 created + cat /tmp/tmp.9IYnSDgweB + rm /tmp/tmp.K4SSm3VElO /tmp/tmp.9IYnSDgweB + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.CXrK5pIxdT +++ mktemp ++ local LAST_ERR=/tmp/tmp.1m1OtSLMNm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.CXrK5pIxdT ++ cat /tmp/tmp.1m1OtSLMNm ++ rm /tmp/tmp.CXrK5pIxdT /tmp/tmp.1m1OtSLMNm ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2221-abd79966-3-cluster3 --namespace=init-deploy-9470 ++ mktemp + local LAST_OUT=/tmp/tmp.uOfOvPMpbY ++ mktemp + local LAST_ERR=/tmp/tmp.IOvPSeW62g + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2221-abd79966-3-cluster3 --namespace=init-deploy-9470 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uOfOvPMpbY Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2221-abd79966-3-cluster3" modified. + cat /tmp/tmp.IOvPSeW62g + rm /tmp/tmp.uOfOvPMpbY /tmp/tmp.IOvPSeW62g + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.UFBZuZR1Xe ++ mktemp + local LAST_ERR=/tmp/tmp.HngU9n0AYt + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.UFBZuZR1Xe secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.HngU9n0AYt + rm /tmp/tmp.UFBZuZR1Xe /tmp/tmp.HngU9n0AYt + return 0 + deploy_cert_manager + desc 'deploy cert manager' + set +o xtrace ----------------------------------------------------------------------------------- deploy cert manager ----------------------------------------------------------------------------------- + kubectl_bin create namespace cert-manager ++ mktemp + local LAST_OUT=/tmp/tmp.l049Eb5XCQ ++ mktemp + local LAST_ERR=/tmp/tmp.LdHMkdbUxI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace cert-manager + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.l049Eb5XCQ namespace/cert-manager created + cat /tmp/tmp.LdHMkdbUxI + rm /tmp/tmp.l049Eb5XCQ /tmp/tmp.LdHMkdbUxI + return 0 + kubectl_bin label namespace cert-manager certmanager.k8s.io/disable-validation=true ++ mktemp + local LAST_OUT=/tmp/tmp.uH076Uv7gk ++ mktemp + local LAST_ERR=/tmp/tmp.iQaMM6rjwE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl label namespace cert-manager certmanager.k8s.io/disable-validation=true + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.uH076Uv7gk namespace/cert-manager labeled + cat /tmp/tmp.iQaMM6rjwE + rm /tmp/tmp.uH076Uv7gk /tmp/tmp.iQaMM6rjwE + return 0 + kubectl_bin apply -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml --validate=false ++ mktemp + local LAST_OUT=/tmp/tmp.YZPnvNAFvO ++ mktemp + local LAST_ERR=/tmp/tmp.0FUKGlFi1h + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml --validate=false + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YZPnvNAFvO namespace/cert-manager configured customresourcedefinition.apiextensions.k8s.io/certificaterequests.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/certificates.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/challenges.acme.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterissuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/issuers.cert-manager.io unchanged customresourcedefinition.apiextensions.k8s.io/orders.acme.cert-manager.io unchanged serviceaccount/cert-manager-cainjector created serviceaccount/cert-manager created serviceaccount/cert-manager-webhook created clusterrole.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-cluster-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-view unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-edit unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrole.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-cainjector unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-issuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-clusterissuers unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificates unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-orders unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-challenges unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-ingress-shim unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-approve:cert-manager-io unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-controller-certificatesigningrequests unchanged clusterrolebinding.rbac.authorization.k8s.io/cert-manager-webhook:subjectaccessreviews unchanged role.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged role.rbac.authorization.k8s.io/cert-manager-tokenrequest created role.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created rolebinding.rbac.authorization.k8s.io/cert-manager-cainjector:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager:leaderelection unchanged rolebinding.rbac.authorization.k8s.io/cert-manager-cert-manager-tokenrequest created rolebinding.rbac.authorization.k8s.io/cert-manager-webhook:dynamic-serving created service/cert-manager-cainjector created service/cert-manager created service/cert-manager-webhook created deployment.apps/cert-manager-cainjector created deployment.apps/cert-manager created deployment.apps/cert-manager-webhook created mutatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured validatingwebhookconfiguration.admissionregistration.k8s.io/cert-manager-webhook configured + cat /tmp/tmp.0FUKGlFi1h Warning: resource namespaces/cert-manager is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. + rm /tmp/tmp.YZPnvNAFvO /tmp/tmp.0FUKGlFi1h + return 0 + '[' '' == 4.10 ']' + sleep 70 + desc 'apply my-env-var-secrets to add PEER_LIST_SRV_PROTOCOL' + set +o xtrace ----------------------------------------------------------------------------------- apply my-env-var-secrets to add PEER_LIST_SRV_PROTOCOL ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/conf/envsecrets.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.nzThtoKTt6 ++ mktemp + local LAST_ERR=/tmp/tmp.nfOHk1KFfS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/conf/envsecrets.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.nzThtoKTt6 secret/my-env-var-secrets created + cat /tmp/tmp.nfOHk1KFfS + rm /tmp/tmp.nzThtoKTt6 /tmp/tmp.nfOHk1KFfS + return 0 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/some-name.yml 3 10 /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/secrets_without_tls.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/secrets_without_tls.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/secrets_without_tls.yml ++ mktemp + local LAST_OUT=/tmp/tmp.WoWF7ZFrdk ++ mktemp + local LAST_ERR=/tmp/tmp.qoBf6QA4mc + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/secrets_without_tls.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WoWF7ZFrdk secret/my-cluster-secrets created + cat /tmp/tmp.qoBf6QA4mc + rm /tmp/tmp.WoWF7ZFrdk /tmp/tmp.qoBf6QA4mc + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/client.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/client.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/client.yml '' + kubectl_bin apply -f - + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/client.yml + local pvc_name= + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2221-abd79966#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.init-deploy-9470~ + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + local LAST_OUT=/tmp/tmp.jHBljsQXU5 ++ mktemp + local LAST_ERR=/tmp/tmp.oIhY8rNVfu + local exit_status=0 + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jHBljsQXU5 deployment.apps/pxc-client created + cat /tmp/tmp.oIhY8rNVfu + rm /tmp/tmp.jHBljsQXU5 /tmp/tmp.oIhY8rNVfu + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/some-name.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/some-name.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/some-name.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/some-name.yml + local pvc_name= + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/conf/some-name.yml ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_OUT=/tmp/tmp.IXkYL8MUQt + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.init-deploy-9470~ + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' ++ mktemp + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2221-abd79966#' + local LAST_ERR=/tmp/tmp.qgheeLXP3q + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.IXkYL8MUQt perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.qgheeLXP3q + rm /tmp/tmp.IXkYL8MUQt /tmp/tmp.qgheeLXP3q + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.qFDlzPva6B ++++ mktemp +++ local LAST_ERR=/tmp/tmp.Gn3ISHg0f6 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.qFDlzPva6B +++ cat /tmp/tmp.Gn3ISHg0f6 +++ rm /tmp/tmp.qFDlzPva6B /tmp/tmp.Gn3ISHg0f6 +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.RWeLoadpxr ++++ mktemp +++ local LAST_ERR=/tmp/tmp.IQPv7mryil +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.RWeLoadpxr +++ cat /tmp/tmp.IQPv7mryil +++ rm /tmp/tmp.RWeLoadpxr /tmp/tmp.IQPv7mryil +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n init-deploy-9470 ++ mktemp + local LAST_OUT=/tmp/tmp.hvARcZ2Ivu ++ mktemp + local LAST_ERR=/tmp/tmp.hzfQ5k4UG9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n init-deploy-9470 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n init-deploy-9470 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n init-deploy-9470 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.hvARcZ2Ivu + cat /tmp/tmp.hzfQ5k4UG9 error: no matching resources found + rm /tmp/tmp.hvARcZ2Ivu /tmp/tmp.hzfQ5k4UG9 + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met waiting for pod/some-name-proxysql-0 to become Ready.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo some-name-pxc-1 + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + sleep 10 ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.1L7QRNyQOK +++ mktemp ++ local LAST_ERR=/tmp/tmp.zzbU2VRh6D ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1L7QRNyQOK ++ cat /tmp/tmp.zzbU2VRh6D ++ rm /tmp/tmp.1L7QRNyQOK /tmp/tmp.zzbU2VRh6D ++ return 0 + local root_pass=root_password + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zlSCC2r9hJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.aXElZz7Kf2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zlSCC2r9hJ ++ cat /tmp/tmp.aXElZz7Kf2 ++ rm /tmp/tmp.zlSCC2r9hJ /tmp/tmp.aXElZz7Kf2 ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qagfzM2YCV +++ mktemp ++ local LAST_ERR=/tmp/tmp.MdMvNtBAJd ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qagfzM2YCV ++ cat /tmp/tmp.MdMvNtBAJd ++ rm /tmp/tmp.qagfzM2YCV /tmp/tmp.MdMvNtBAJd ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Dr4ru7auQL +++ mktemp ++ local LAST_ERR=/tmp/tmp.HZbrxQrL2A ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Dr4ru7auQL ++ cat /tmp/tmp.HZbrxQrL2A ++ rm /tmp/tmp.Dr4ru7auQL /tmp/tmp.HZbrxQrL2A ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-4bjr6 + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.RXZu460Y3x/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.wREG2vvFxq +++ mktemp ++ local LAST_ERR=/tmp/tmp.eEBcbrzUzG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.wREG2vvFxq ++ cat /tmp/tmp.eEBcbrzUzG ++ rm /tmp/tmp.wREG2vvFxq /tmp/tmp.eEBcbrzUzG ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.RXZu460Y3x/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -p'\''root_password'\'' -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ie1tJbxsCm +++ mktemp ++ local LAST_ERR=/tmp/tmp.YYuOd2WsKH ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ie1tJbxsCm ++ cat /tmp/tmp.YYuOd2WsKH ++ rm /tmp/tmp.ie1tJbxsCm /tmp/tmp.YYuOd2WsKH ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.RXZu460Y3x/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ grep -E -o 'early-plugin-load=keyring_\w+.so' ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2yOXxkQZcL +++ mktemp ++ local LAST_ERR=/tmp/tmp.Sf875AdNKU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2yOXxkQZcL ++ cat /tmp/tmp.Sf875AdNKU Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.2yOXxkQZcL /tmp/tmp.Sf875AdNKU ++ return 0 + '[' '' ']' + desc 'check if cert-manager issued ssl certificates' + set +o xtrace ----------------------------------------------------------------------------------- check if cert-manager issued ssl certificates ----------------------------------------------------------------------------------- ++ kubectl get secrets some-name-ssl -o 'jsonpath={.metadata.annotations.cert-manager\.io/issuer-kind}' + '[' Issuer '!=' Issuer ']' + desc 'check if service and statefulset created with expected config' + set +o xtrace ----------------------------------------------------------------------------------- check if service and statefulset created with expected config ----------------------------------------------------------------------------------- + compare_kubectl statefulset/some-name-pxc + local resource=statefulset/some-name-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc.yml + local new_result=/tmp/tmp.RXZu460Y3x/statefulset_some-name-pxc.yml + desc 'compare statefulset/some-name-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127.yml ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127-aks.yml ']' + kubectl_bin get -o yaml statefulset/some-name-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.cYYJaouHI4 ++ mktemp + local LAST_ERR=/tmp/tmp.5XYahf7LGL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.cYYJaouHI4 + cat /tmp/tmp.5XYahf7LGL + rm /tmp/tmp.cYYJaouHI4 /tmp/tmp.5XYahf7LGL + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-pxc-k127.yml /tmp/tmp.RXZu460Y3x/statefulset_some-name-pxc.yml + log 'compare_kubectl: statefulset/some-name-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:49:40+0000]' compare_kubectl: statefulset/some-name-pxc OK [2025-12-01T16:49:40+0000] compare_kubectl: statefulset/some-name-pxc OK + compare_kubectl statefulset/some-name-proxysql + local resource=statefulset/some-name-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql.yml + local new_result=/tmp/tmp.RXZu460Y3x/statefulset_some-name-proxysql.yml + desc 'compare statefulset/some-name-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/some-name-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127.yml ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml statefulset/some-name-proxysql ++ mktemp + local LAST_OUT=/tmp/tmp.KavTvQRx7C ++ mktemp + local LAST_ERR=/tmp/tmp.d0bHQpJDMP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KavTvQRx7C + cat /tmp/tmp.d0bHQpJDMP + rm /tmp/tmp.KavTvQRx7C /tmp/tmp.d0bHQpJDMP + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_some-name-proxysql-k127.yml /tmp/tmp.RXZu460Y3x/statefulset_some-name-proxysql.yml + log 'compare_kubectl: statefulset/some-name-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:49:42+0000]' compare_kubectl: statefulset/some-name-proxysql OK [2025-12-01T16:49:42+0000] compare_kubectl: statefulset/some-name-proxysql OK + compare_kubectl service/some-name-pxc + local resource=service/some-name-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc.yml + local new_result=/tmp/tmp.RXZu460Y3x/service_some-name-pxc.yml + desc 'compare service/some-name-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc-aks.yml ']' + kubectl_bin get -o yaml service/some-name-pxc + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.EVWR2sTF75 ++ mktemp + local LAST_ERR=/tmp/tmp.CPIpckZxrs + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EVWR2sTF75 + cat /tmp/tmp.CPIpckZxrs + rm /tmp/tmp.EVWR2sTF75 /tmp/tmp.CPIpckZxrs + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-pxc.yml /tmp/tmp.RXZu460Y3x/service_some-name-pxc.yml + log 'compare_kubectl: service/some-name-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:49:43+0000]' compare_kubectl: service/some-name-pxc OK [2025-12-01T16:49:43+0000] compare_kubectl: service/some-name-pxc OK + compare_kubectl service/some-name-proxysql + local resource=service/some-name-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql.yml + local new_result=/tmp/tmp.RXZu460Y3x/service_some-name-proxysql.yml + desc 'compare service/some-name-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-k129.yml ']' + version_gt 1.27 ++ bc -l ++ echo '1.31 >= 1.27' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.31 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml service/some-name-proxysql ++ mktemp + local LAST_OUT=/tmp/tmp.w6q7rvxqpT ++ mktemp + local LAST_ERR=/tmp/tmp.iwcZECsHIg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.w6q7rvxqpT + cat /tmp/tmp.iwcZECsHIg + rm /tmp/tmp.w6q7rvxqpT /tmp/tmp.iwcZECsHIg + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql.yml /tmp/tmp.RXZu460Y3x/service_some-name-proxysql.yml + log 'compare_kubectl: service/some-name-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:49:45+0000]' compare_kubectl: service/some-name-proxysql OK [2025-12-01T16:49:45+0000] compare_kubectl: service/some-name-proxysql OK + compare_kubectl service/some-name-proxysql-unready + local resource=service/some-name-proxysql-unready + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready.yml + local new_result=/tmp/tmp.RXZu460Y3x/service_some-name-proxysql-unready.yml + desc 'compare service/some-name-proxysql-unready-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/some-name-proxysql-unready- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready-aks.yml ']' + kubectl_bin get -o yaml service/some-name-proxysql-unready + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.9I6NUGFlAC ++ mktemp + local LAST_ERR=/tmp/tmp.QyUb5LxJxg + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/some-name-proxysql-unready + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9I6NUGFlAC + cat /tmp/tmp.QyUb5LxJxg + rm /tmp/tmp.9I6NUGFlAC /tmp/tmp.QyUb5LxJxg + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_some-name-proxysql-unready.yml /tmp/tmp.RXZu460Y3x/service_some-name-proxysql-unready.yml + log 'compare_kubectl: service/some-name-proxysql-unready OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:49:46+0000]' compare_kubectl: service/some-name-proxysql-unready OK [2025-12-01T16:49:46+0000] compare_kubectl: service/some-name-proxysql-unready OK + [[ ! perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + desc 'check if mysql-state-monitor works as expected' + set +o xtrace ----------------------------------------------------------------------------------- check if mysql-state-monitor works as expected ----------------------------------------------------------------------------------- ++ awk '{print $1}' ++ kubectl_bin get pod -l app.kubernetes.io/component=pxc --no-headers +++ mktemp ++ local LAST_OUT=/tmp/tmp.YeQIKbCVUB +++ mktemp ++ local LAST_ERR=/tmp/tmp.QUNg9a77s5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pod -l app.kubernetes.io/component=pxc --no-headers ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.YeQIKbCVUB ++ cat /tmp/tmp.QUNg9a77s5 ++ rm /tmp/tmp.YeQIKbCVUB /tmp/tmp.QUNg9a77s5 ++ return 0 + for pod in '$(kubectl_bin get pod -l app.kubernetes.io/component=pxc --no-headers | awk '\''{print $1}'\'')' ++ kubectl_bin exec some-name-pxc-0 -- cat /var/lib/mysql/mysql.state +++ mktemp ++ local LAST_OUT=/tmp/tmp.ylGaIc5vAP +++ mktemp ++ local LAST_ERR=/tmp/tmp.2SDfHTELfy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec some-name-pxc-0 -- cat /var/lib/mysql/mysql.state ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ylGaIc5vAP ++ cat /tmp/tmp.2SDfHTELfy ++ rm /tmp/tmp.ylGaIc5vAP /tmp/tmp.2SDfHTELfy ++ return 0 + state=ready + [[ ready != \r\e\a\d\y ]] + for pod in '$(kubectl_bin get pod -l app.kubernetes.io/component=pxc --no-headers | awk '\''{print $1}'\'')' ++ kubectl_bin exec some-name-pxc-1 -- cat /var/lib/mysql/mysql.state +++ mktemp ++ local LAST_OUT=/tmp/tmp.7TSf8gwke7 +++ mktemp ++ local LAST_ERR=/tmp/tmp.CglrSjyQha ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec some-name-pxc-1 -- cat /var/lib/mysql/mysql.state ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7TSf8gwke7 ++ cat /tmp/tmp.CglrSjyQha ++ rm /tmp/tmp.7TSf8gwke7 /tmp/tmp.CglrSjyQha ++ return 0 + state=ready + [[ ready != \r\e\a\d\y ]] + for pod in '$(kubectl_bin get pod -l app.kubernetes.io/component=pxc --no-headers | awk '\''{print $1}'\'')' ++ kubectl_bin exec some-name-pxc-2 -- cat /var/lib/mysql/mysql.state +++ mktemp ++ local LAST_OUT=/tmp/tmp.9bK8U5TPQq +++ mktemp ++ local LAST_ERR=/tmp/tmp.9r5O5uWcIn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec some-name-pxc-2 -- cat /var/lib/mysql/mysql.state ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9bK8U5TPQq ++ cat /tmp/tmp.9r5O5uWcIn ++ rm /tmp/tmp.9bK8U5TPQq /tmp/tmp.9r5O5uWcIn ++ return 0 + state=ready + [[ ready != \r\e\a\d\y ]] + desc 'check PodDisruptionBudget' + set +o xtrace ----------------------------------------------------------------------------------- check PodDisruptionBudget ----------------------------------------------------------------------------------- + compare_kubectl pdb/some-name-pxc + local resource=pdb/some-name-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc.yml + local new_result=/tmp/tmp.RXZu460Y3x/pdb_some-name-pxc.yml + desc 'compare pdb/some-name-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare pdb/some-name-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k124.yml ']' + version_gt 1.22 ++ bc -l ++ echo '1.31 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc-aks.yml ']' + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + kubectl_bin get -o yaml pdb/some-name-pxc ++ mktemp + local LAST_OUT=/tmp/tmp.2KZ5ketKhR ++ mktemp + local LAST_ERR=/tmp/tmp.Y2TmE6BRfZ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml pdb/some-name-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.2KZ5ketKhR + cat /tmp/tmp.Y2TmE6BRfZ + rm /tmp/tmp.2KZ5ketKhR /tmp/tmp.Y2TmE6BRfZ + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-pxc.yml /tmp/tmp.RXZu460Y3x/pdb_some-name-pxc.yml + log 'compare_kubectl: pdb/some-name-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:49:54+0000]' compare_kubectl: pdb/some-name-pxc OK [2025-12-01T16:49:54+0000] compare_kubectl: pdb/some-name-pxc OK + compare_kubectl pdb/some-name-proxysql + local resource=pdb/some-name-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql.yml + local new_result=/tmp/tmp.RXZu460Y3x/pdb_some-name-proxysql.yml + desc 'compare pdb/some-name-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare pdb/some-name-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql-aks.yml ']' + kubectl_bin get -o yaml pdb/some-name-proxysql ++ mktemp + local LAST_OUT=/tmp/tmp.hWPoL4Ks5s ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_ERR=/tmp/tmp.kkVEqQoVMr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml pdb/some-name-proxysql + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hWPoL4Ks5s + cat /tmp/tmp.kkVEqQoVMr + rm /tmp/tmp.hWPoL4Ks5s /tmp/tmp.kkVEqQoVMr + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pdb_some-name-proxysql.yml /tmp/tmp.RXZu460Y3x/pdb_some-name-proxysql.yml + log 'compare_kubectl: pdb/some-name-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:49:55+0000]' compare_kubectl: pdb/some-name-proxysql OK [2025-12-01T16:49:55+0000] compare_kubectl: pdb/some-name-proxysql OK + desc 'check if MySQL users created' + set +o xtrace ----------------------------------------------------------------------------------- check if MySQL users created ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-pxc -uroot -proot_password' + local 'uri=-h some-name-pxc -uroot -proot_password' + local postfix= ++ echo -h some-name-pxc -uroot -proot_password ++ sed -e 's/.*-u//; s/ .*//' + local user=root + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/root.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/root-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/root-80.sql + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-pxc -uroot -proot_password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.sVaKnAKFVh +++ mktemp ++ local LAST_ERR=/tmp/tmp.xjf6QvFw3J ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sVaKnAKFVh ++ cat /tmp/tmp.xjf6QvFw3J ++ rm /tmp/tmp.sVaKnAKFVh /tmp/tmp.xjf6QvFw3J ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/root-80.sql /tmp/tmp.RXZu460Y3x/root.sql + compare_mysql_user '-h some-name-pxc -umonitor -pmonitor_password' + local 'uri=-h some-name-pxc -umonitor -pmonitor_password' + local postfix= ++ echo -h some-name-pxc -umonitor -pmonitor_password ++ sed -e 's/.*-u//; s/ .*//' + local user=monitor + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/monitor.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/monitor-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/monitor-80.sql + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-pxc -umonitor -pmonitor_password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-pxc -umonitor -pmonitor_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UxShNlo2gq +++ mktemp ++ local LAST_ERR=/tmp/tmp.hm1NSxhBVt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UxShNlo2gq ++ cat /tmp/tmp.hm1NSxhBVt ++ rm /tmp/tmp.UxShNlo2gq /tmp/tmp.hm1NSxhBVt ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/monitor-80.sql /tmp/tmp.RXZu460Y3x/monitor.sql + compare_mysql_user '-h some-name-pxc -uproxyuser -ps3cret' + local 'uri=-h some-name-pxc -uproxyuser -ps3cret' + local postfix= ++ echo -h some-name-pxc -uproxyuser -ps3cret ++ sed -e 's/.*-u//; s/ .*//' + local user=proxyuser + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/proxyuser.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/proxyuser-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-pxc -uproxyuser -ps3cret' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-pxc -uproxyuser -ps3cret' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.EnWfJ1Yp0a +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ys4GWm3MrT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.EnWfJ1Yp0a ++ cat /tmp/tmp.Ys4GWm3MrT ++ rm /tmp/tmp.EnWfJ1Yp0a /tmp/tmp.Ys4GWm3MrT ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/proxyuser.sql /tmp/tmp.RXZu460Y3x/proxyuser.sql + compare_mysql_user_local '-uxtrabackup -pbackup_password' some-name-pxc-0 '' pxc + local 'uri=-uxtrabackup -pbackup_password' + local pod=some-name-pxc-0 + local postfix= + local container_name=pxc ++ echo -uxtrabackup -pbackup_password ++ sed -e 's/.*-u//; s/ .*//' + local user=xtrabackup + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/xtrabackup.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/xtrabackup-80.sql ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/xtrabackup-80.sql + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql_local 'SHOW GRANTS;' '-uxtrabackup -pbackup_password' some-name-pxc-0 pxc + local 'command=SHOW GRANTS;' + local 'uri=-uxtrabackup -pbackup_password' + local pod=some-name-pxc-0 + local container_name=pxc + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/xtrabackup-80.sql /tmp/tmp.RXZu460Y3x/xtrabackup.sql + desc 'check that pmm server user don'\''t have access' + set +o xtrace ----------------------------------------------------------------------------------- check that pmm server user don't have access ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-pxc -upmmserver -ppmmserver_password' + local 'uri=-h some-name-pxc -upmmserver -ppmmserver_password' + local postfix= ++ echo -h some-name-pxc -upmmserver -ppmmserver_password ++ sed -e 's/.*-u//; s/ .*//' + local user=pmmserver + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pmmserver.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pmmserver-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-pxc -upmmserver -ppmmserver_password' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-pxc -upmmserver -ppmmserver_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6u2osgwQTi +++ mktemp ++ local LAST_ERR=/tmp/tmp.8WF48FejsZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6u2osgwQTi ++ cat /tmp/tmp.8WF48FejsZ ++ rm /tmp/tmp.6u2osgwQTi /tmp/tmp.8WF48FejsZ ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-4bjr6 + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pmmserver.sql /tmp/tmp.RXZu460Y3x/pmmserver.sql + desc 'check if ProxySQL users created' + set +o xtrace ----------------------------------------------------------------------------------- check if ProxySQL users created ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-proxysql -uroot -proot_password' + local 'uri=-h some-name-proxysql -uroot -proot_password' + local postfix= ++ echo -h some-name-proxysql -uroot -proot_password ++ sed -e 's/.*-u//; s/ .*//' + local user=root + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/root.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/root-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/root-80.sql + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-proxysql -uroot -proot_password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cLeDC3G8DD +++ mktemp ++ local LAST_ERR=/tmp/tmp.R4HuGaHJPt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cLeDC3G8DD ++ cat /tmp/tmp.R4HuGaHJPt ++ rm /tmp/tmp.cLeDC3G8DD /tmp/tmp.R4HuGaHJPt ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-4bjr6 ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/root-80.sql /tmp/tmp.RXZu460Y3x/root.sql + compare_mysql_user '-h some-name-proxysql -umonitor -pmonitor_password' + local 'uri=-h some-name-proxysql -umonitor -pmonitor_password' + local postfix= ++ echo -h some-name-proxysql -umonitor -pmonitor_password ++ sed -e 's/.*-u//; s/ .*//' + local user=monitor + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/monitor.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/monitor-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/monitor-80.sql + run_mysql 'SHOW GRANTS;' '-h some-name-proxysql -umonitor -pmonitor_password' + local 'command=SHOW GRANTS;' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + local 'uri=-h some-name-proxysql -umonitor -pmonitor_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BHoTv83bOT +++ mktemp ++ local LAST_ERR=/tmp/tmp.rVNZWe5CYl ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BHoTv83bOT ++ cat /tmp/tmp.rVNZWe5CYl ++ rm /tmp/tmp.BHoTv83bOT /tmp/tmp.rVNZWe5CYl ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/monitor-80.sql /tmp/tmp.RXZu460Y3x/monitor.sql + desc 'check that pmm server user don'\''t have access' + set +o xtrace ----------------------------------------------------------------------------------- check that pmm server user don't have access ----------------------------------------------------------------------------------- + compare_mysql_user '-h some-name-proxysql -uproxyuser -ps3cret' -proxysql + local 'uri=-h some-name-proxysql -uproxyuser -ps3cret' + local postfix=-proxysql ++ echo -h some-name-proxysql -uproxyuser -ps3cret ++ sed -e 's/.*-u//; s/ .*//' + local user=proxyuser + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/proxyuser-proxysql.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/proxyuser-proxysql-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' + run_mysql 'SHOW GRANTS;' '-h some-name-proxysql -uproxyuser -ps3cret' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-proxysql -uproxyuser -ps3cret' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.qvZ3MHmsOL +++ mktemp ++ local LAST_ERR=/tmp/tmp.2XDEmOX2fb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.qvZ3MHmsOL ++ cat /tmp/tmp.2XDEmOX2fb ++ rm /tmp/tmp.qvZ3MHmsOL /tmp/tmp.2XDEmOX2fb ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/proxyuser-proxysql.sql /tmp/tmp.RXZu460Y3x/proxyuser.sql + compare_mysql_user '-h some-name-proxysql -upmmserver -ppmmserver_password' -proxysql + local 'uri=-h some-name-proxysql -upmmserver -ppmmserver_password' + local postfix=-proxysql ++ echo -h some-name-proxysql -upmmserver -ppmmserver_password ++ sed -e 's/.*-u//; s/ .*//' + local user=pmmserver + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pmmserver-proxysql.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pmmserver-proxysql-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SHOW GRANTS;' '-h some-name-proxysql -upmmserver -ppmmserver_password' + local 'command=SHOW GRANTS;' + local 'uri=-h some-name-proxysql -upmmserver -ppmmserver_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' + /usr/bin/sed -E 's/'\''(10|192)[.][0-9][^'\'']*'\''//; s/'\''[^'\'']*[.]internal'\''//' +++ mktemp ++ local LAST_OUT=/tmp/tmp.0MFKHe7Pvm +++ mktemp ++ local LAST_ERR=/tmp/tmp.Nk8M6n3Y85 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.0MFKHe7Pvm ++ cat /tmp/tmp.Nk8M6n3Y85 ++ rm /tmp/tmp.0MFKHe7Pvm /tmp/tmp.Nk8M6n3Y85 ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-4bjr6 ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/pmmserver-proxysql.sql /tmp/tmp.RXZu460Y3x/pmmserver.sql + desc 'write data directly, read from all' + set +o xtrace ----------------------------------------------------------------------------------- write data directly, read from all ----------------------------------------------------------------------------------- + run_mysql 'INSERT myApp.myApp (id) VALUES (100501)' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local 'command=INSERT myApp.myApp (id) VALUES (100501)' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.JyMIstUB39 +++ mktemp ++ local LAST_ERR=/tmp/tmp.wBcpcynJWx ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JyMIstUB39 ++ cat /tmp/tmp.wBcpcynJWx ++ rm /tmp/tmp.JyMIstUB39 /tmp/tmp.wBcpcynJWx ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + compare_mysql_cmd select-2 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.TuDa6bj9Fz +++ mktemp ++ local LAST_ERR=/tmp/tmp.NxqEMwKV8w ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.TuDa6bj9Fz ++ cat /tmp/tmp.NxqEMwKV8w ++ rm /tmp/tmp.TuDa6bj9Fz /tmp/tmp.NxqEMwKV8w ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-2.sql /tmp/tmp.RXZu460Y3x/select-2.sql + compare_mysql_cmd select-2 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.95GwSd9T5z +++ mktemp ++ local LAST_ERR=/tmp/tmp.rFZpEqPYHr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.95GwSd9T5z ++ cat /tmp/tmp.rFZpEqPYHr ++ rm /tmp/tmp.95GwSd9T5z /tmp/tmp.rFZpEqPYHr ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-4bjr6 ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-2.sql /tmp/tmp.RXZu460Y3x/select-2.sql + compare_mysql_cmd select-2 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local command_id=select-2 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Ztzo3ZQN49 +++ mktemp ++ local LAST_ERR=/tmp/tmp.3kWiKl7zcc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Ztzo3ZQN49 ++ cat /tmp/tmp.3kWiKl7zcc ++ rm /tmp/tmp.Ztzo3ZQN49 /tmp/tmp.3kWiKl7zcc ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-2.sql /tmp/tmp.RXZu460Y3x/select-2.sql + desc 'kill Primary Pod, check reelection, check data' + set +o xtrace ----------------------------------------------------------------------------------- kill Primary Pod, check reelection, check data ----------------------------------------------------------------------------------- ++ get_proxy_primary '-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' some-name-proxysql-0 ++ local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' ++ local pod=some-name-proxysql-0 +++ run_mysql_local 'SELECT hostname FROM runtime_mysql_servers WHERE hostgroup_id=11 AND status='\''ONLINE'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' some-name-proxysql-0 proxysql +++ local 'command=SELECT hostname FROM runtime_mysql_servers WHERE hostgroup_id=11 AND status='\''ONLINE'\'';' +++ local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' +++ local pod=some-name-proxysql-0 +++ local container_name=proxysql +++ set +o xtrace ++ local ip=some-name-pxc-2.some-name-pxc.init-deploy-9470.svc.cluster.local +++ echo some-name-pxc-2.some-name-pxc.init-deploy-9470.svc.cluster.local +++ wc -l ++ '[' 1 '!=' 1 ']' ++ echo some-name-pxc-2.some-name-pxc.init-deploy-9470.svc.cluster.local ++ cut -d. -f1 + initial_primary=some-name-pxc-2 + kubectl_bin delete pods --grace-period=0 --force some-name-pxc-2 ++ mktemp + local LAST_OUT=/tmp/tmp.WCt7m5B0Yt ++ mktemp + local LAST_ERR=/tmp/tmp.HggbpdiQvy + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pods --grace-period=0 --force some-name-pxc-2 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WCt7m5B0Yt pod "some-name-pxc-2" force deleted from init-deploy-9470 namespace + cat /tmp/tmp.HggbpdiQvy Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. + rm /tmp/tmp.WCt7m5B0Yt /tmp/tmp.HggbpdiQvy + return 0 + sleep 10 ++ get_proxy_primary '-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' some-name-proxysql-0 ++ local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' ++ local pod=some-name-proxysql-0 +++ run_mysql_local 'SELECT hostname FROM runtime_mysql_servers WHERE hostgroup_id=11 AND status='\''ONLINE'\'';' '-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' some-name-proxysql-0 proxysql +++ local 'command=SELECT hostname FROM runtime_mysql_servers WHERE hostgroup_id=11 AND status='\''ONLINE'\'';' +++ local 'uri=-h127.0.0.1 -P6032 -uproxyadmin -padmin_password' +++ local pod=some-name-proxysql-0 +++ local container_name=proxysql +++ set +o xtrace ++ local ip=some-name-pxc-1.some-name-pxc.init-deploy-9470.svc.cluster.local +++ wc -l +++ echo some-name-pxc-1.some-name-pxc.init-deploy-9470.svc.cluster.local ++ '[' 1 '!=' 1 ']' ++ echo some-name-pxc-1.some-name-pxc.init-deploy-9470.svc.cluster.local ++ cut -d. -f1 + changed_primary=some-name-pxc-1 + '[' -n some-name-pxc-1 ']' + '[' some-name-pxc-2 '!=' some-name-pxc-1 ']' + run_mysql 'INSERT myApp.myApp (id) VALUES (100502)' '-h some-name-proxysql -uroot -proot_password' + local 'command=INSERT myApp.myApp (id) VALUES (100502)' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.cFzNAQMVdG +++ mktemp ++ local LAST_ERR=/tmp/tmp.QOh1D3tl3R ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cFzNAQMVdG ++ cat /tmp/tmp.QOh1D3tl3R ++ rm /tmp/tmp.cFzNAQMVdG /tmp/tmp.QOh1D3tl3R ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-59944c5bbf-4bjr6 ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ grep -E '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo some-name-pxc-0 + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met waiting for pod/some-name-pxc-0 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met waiting for pod/some-name-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met waiting for pod/some-name-pxc-2 to become Ready.Ok + compare_mysql_cmd select-3 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local command_id=select-3 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-3-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zSq8qRvdjR +++ mktemp ++ local LAST_ERR=/tmp/tmp.A6zSiCZTm9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zSq8qRvdjR ++ cat /tmp/tmp.A6zSiCZTm9 ++ rm /tmp/tmp.zSq8qRvdjR /tmp/tmp.A6zSiCZTm9 ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-3.sql /tmp/tmp.RXZu460Y3x/select-3.sql + compare_mysql_cmd select-3 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local command_id=select-3 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-3-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bfFBShTVTW +++ mktemp ++ local LAST_ERR=/tmp/tmp.LRhhuWWVEU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bfFBShTVTW ++ cat /tmp/tmp.LRhhuWWVEU ++ rm /tmp/tmp.bfFBShTVTW /tmp/tmp.LRhhuWWVEU ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' ++ echo pxc-client-59944c5bbf-4bjr6 + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-3.sql /tmp/tmp.RXZu460Y3x/select-3.sql + compare_mysql_cmd select-3 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local command_id=select-3 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-3.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-3-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.eHVeAmsuAX +++ mktemp ++ local LAST_ERR=/tmp/tmp.pRQQ7Yo7ru ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.eHVeAmsuAX ++ cat /tmp/tmp.pRQQ7Yo7ru ++ rm /tmp/tmp.eHVeAmsuAX /tmp/tmp.pRQQ7Yo7ru ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-3.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-3.sql /tmp/tmp.RXZu460Y3x/select-3.sql + compare_mysql_cmd max_allowed_packet-1 'SELECT @@max_allowed_packet;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local command_id=max_allowed_packet-1 + local 'command=SELECT @@max_allowed_packet;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/max_allowed_packet-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/max_allowed_packet-1-80.sql ]] + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/max_allowed_packet-1-80.sql + run_mysql 'SELECT @@max_allowed_packet;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local 'command=SELECT @@max_allowed_packet;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.hUuL9BVRpZ +++ mktemp ++ local LAST_ERR=/tmp/tmp.k5JCOl1Nnp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.hUuL9BVRpZ ++ cat /tmp/tmp.k5JCOl1Nnp ++ rm /tmp/tmp.hUuL9BVRpZ /tmp/tmp.k5JCOl1Nnp ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/max_allowed_packet-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/max_allowed_packet-1-80.sql /tmp/tmp.RXZu460Y3x/max_allowed_packet-1.sql + desc 'check if possible to create no-proxysql cluster' + set +o xtrace ----------------------------------------------------------------------------------- check if possible to create no-proxysql cluster ----------------------------------------------------------------------------------- + cluster3=no-proxysql + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/conf/no-proxysql.yml + local config_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/conf/no-proxysql.yml + local pvc_name= + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/conf/no-proxysql.yml '' + local input_file=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/conf/no-proxysql.yml + local pvc_name= + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/conf/no-proxysql.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' ++ mktemp + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2221-abd79966#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + local LAST_OUT=/tmp/tmp.X5LRxLDlra + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/fluentbit:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.init-deploy-9470~ + /usr/bin/sed -e 's#claimName:..*-backup-pvc$#claimName: #' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + local LAST_ERR=/tmp/tmp.qY5IedgbRD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.X5LRxLDlra perconaxtradbcluster.pxc.percona.com/no-proxysql created + cat /tmp/tmp.qY5IedgbRD + rm /tmp/tmp.X5LRxLDlra /tmp/tmp.qY5IedgbRD + return 0 + wait_for_running no-proxysql-pxc 3 + local name=no-proxysql-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod no-proxysql-pxc-0 480 + local pod=no-proxysql-pxc-0 + local max_retry=480 + local ns= ++ echo no-proxysql-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace Error from server (NotFound): pods "no-proxysql-pxc-0" not found waiting for pod/no-proxysql-pxc-0 to become Ready................................Ok + for i in '$(seq 0 $last_pod)' + wait_pod no-proxysql-pxc-1 480 + local pod=no-proxysql-pxc-1 + local max_retry=480 + local ns= ++ echo no-proxysql-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/no-proxysql-pxc-1 condition met waiting for pod/no-proxysql-pxc-1 to become Ready.Ok + for i in '$(seq 0 $last_pod)' + wait_pod no-proxysql-pxc-2 480 + local pod=no-proxysql-pxc-2 + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo no-proxysql-pxc-2 ++ grep -E '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/no-proxysql-pxc-2 condition met waiting for pod/no-proxysql-pxc-2 to become Ready.Ok + desc 'check if service and statefulset created' + set +o xtrace ----------------------------------------------------------------------------------- check if service and statefulset created ----------------------------------------------------------------------------------- + compare_kubectl service/no-proxysql-pxc + local resource=service/no-proxysql-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc.yml + local new_result=/tmp/tmp.RXZu460Y3x/service_no-proxysql-pxc.yml + desc 'compare service/no-proxysql-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/no-proxysql-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc-aks.yml ']' + kubectl_bin get -o yaml service/no-proxysql-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.jLQ4fO5cVc ++ mktemp + local LAST_ERR=/tmp/tmp.op9Wxo4w3b + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/no-proxysql-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jLQ4fO5cVc + cat /tmp/tmp.op9Wxo4w3b + rm /tmp/tmp.jLQ4fO5cVc /tmp/tmp.op9Wxo4w3b + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-pxc.yml /tmp/tmp.RXZu460Y3x/service_no-proxysql-pxc.yml + log 'compare_kubectl: service/no-proxysql-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:57:25+0000]' compare_kubectl: service/no-proxysql-pxc OK [2025-12-01T16:57:25+0000] compare_kubectl: service/no-proxysql-pxc OK + compare_kubectl service/no-proxysql-proxysql + local resource=service/no-proxysql-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql.yml + local new_result=/tmp/tmp.RXZu460Y3x/service_no-proxysql-proxysql.yml + desc 'compare service/no-proxysql-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare service/no-proxysql-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql-aks.yml ']' + kubectl_bin get -o yaml service/no-proxysql-proxysql ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.J7cYfos3Yx ++ mktemp + local LAST_ERR=/tmp/tmp.Z67Bek9WR9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml service/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.J7cYfos3Yx + cat /tmp/tmp.Z67Bek9WR9 Error from server (NotFound): services "no-proxysql-proxysql" not found + rm /tmp/tmp.J7cYfos3Yx /tmp/tmp.Z67Bek9WR9 + return 1 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/service_no-proxysql-proxysql.yml /tmp/tmp.RXZu460Y3x/service_no-proxysql-proxysql.yml + log 'compare_kubectl: service/no-proxysql-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:57:30+0000]' compare_kubectl: service/no-proxysql-proxysql OK [2025-12-01T16:57:30+0000] compare_kubectl: service/no-proxysql-proxysql OK + compare_kubectl configmap/no-proxysql-pxc + local resource=configmap/no-proxysql-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc.yml + local new_result=/tmp/tmp.RXZu460Y3x/configmap_no-proxysql-pxc.yml + desc 'compare configmap/no-proxysql-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare configmap/no-proxysql-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k129.yml ']' + version_gt 1.27 ++ bc -l ++ echo '1.31 >= 1.27' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k124.yml ']' + version_gt 1.22 ++ bc -l ++ echo '1.31 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k122.yml ']' + version_gt 1.21 ++ bc -l ++ echo '1.31 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc-aks.yml ']' + kubectl_bin get -o yaml configmap/no-proxysql-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.icYS2aDvEJ ++ mktemp + local LAST_ERR=/tmp/tmp.EX8X5LRBxX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml configmap/no-proxysql-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.icYS2aDvEJ + cat /tmp/tmp.EX8X5LRBxX + rm /tmp/tmp.icYS2aDvEJ /tmp/tmp.EX8X5LRBxX + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/configmap_no-proxysql-pxc.yml /tmp/tmp.RXZu460Y3x/configmap_no-proxysql-pxc.yml + log 'compare_kubectl: configmap/no-proxysql-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:57:31+0000]' compare_kubectl: configmap/no-proxysql-pxc OK [2025-12-01T16:57:31+0000] compare_kubectl: configmap/no-proxysql-pxc OK + desc 'check if storageClass changed' + set +o xtrace ----------------------------------------------------------------------------------- check if storageClass changed ----------------------------------------------------------------------------------- + compare_kubectl statefulset/no-proxysql-pxc + local resource=statefulset/no-proxysql-pxc + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc.yml + local new_result=/tmp/tmp.RXZu460Y3x/statefulset_no-proxysql-pxc.yml + desc 'compare statefulset/no-proxysql-pxc-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/no-proxysql-pxc- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127.yml ']' + expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127.yml + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127-aks.yml ']' + kubectl_bin get -o yaml statefulset/no-proxysql-pxc ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.2ELWUddPD1 ++ mktemp + local LAST_ERR=/tmp/tmp.KKCxqviFEx + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/no-proxysql-pxc + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.2ELWUddPD1 + cat /tmp/tmp.KKCxqviFEx + rm /tmp/tmp.2ELWUddPD1 /tmp/tmp.KKCxqviFEx + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-pxc-k127.yml /tmp/tmp.RXZu460Y3x/statefulset_no-proxysql-pxc.yml + log 'compare_kubectl: statefulset/no-proxysql-pxc OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:57:32+0000]' compare_kubectl: statefulset/no-proxysql-pxc OK [2025-12-01T16:57:32+0000] compare_kubectl: statefulset/no-proxysql-pxc OK + compare_kubectl statefulset/no-proxysql-proxysql + local resource=statefulset/no-proxysql-proxysql + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql.yml + local new_result=/tmp/tmp.RXZu460Y3x/statefulset_no-proxysql-proxysql.yml + desc 'compare statefulset/no-proxysql-proxysql-' + set +o xtrace ----------------------------------------------------------------------------------- compare statefulset/no-proxysql-proxysql- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ bc -l ++ echo '1.31 >= 1.33' + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.31 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k122.yml ']' + version_gt 1.21 ++ bc -l ++ echo '1.31 >= 1.21' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql-aks.yml ']' + kubectl_bin get -o yaml statefulset/no-proxysql-proxysql + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.VmBbmVFk37 ++ mktemp + local LAST_ERR=/tmp/tmp.qJUkOHuz5o + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml statefulset/no-proxysql-proxysql + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.VmBbmVFk37 + cat /tmp/tmp.qJUkOHuz5o Error from server (NotFound): statefulsets.apps "no-proxysql-proxysql" not found + rm /tmp/tmp.VmBbmVFk37 /tmp/tmp.qJUkOHuz5o + return 1 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/statefulset_no-proxysql-proxysql.yml /tmp/tmp.RXZu460Y3x/statefulset_no-proxysql-proxysql.yml + log 'compare_kubectl: statefulset/no-proxysql-proxysql OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:57:37+0000]' compare_kubectl: statefulset/no-proxysql-proxysql OK [2025-12-01T16:57:37+0000] compare_kubectl: statefulset/no-proxysql-proxysql OK + compare_kubectl issuer/no-proxysql-pxc-issuer + local resource=issuer/no-proxysql-pxc-issuer + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer.yml + local new_result=/tmp/tmp.RXZu460Y3x/issuer_no-proxysql-pxc-issuer.yml + desc 'compare issuer/no-proxysql-pxc-issuer-' + set +o xtrace ----------------------------------------------------------------------------------- compare issuer/no-proxysql-pxc-issuer- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k127.yml ']' + version_gt 1.24 ++ bc -l ++ echo '1.31 >= 1.24' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer-aks.yml ']' + kubectl_bin get -o yaml issuer/no-proxysql-pxc-issuer + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - ++ mktemp + local LAST_OUT=/tmp/tmp.HhtpXI9SkP ++ mktemp + local LAST_ERR=/tmp/tmp.3NlMuFFEi6 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml issuer/no-proxysql-pxc-issuer + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HhtpXI9SkP + cat /tmp/tmp.3NlMuFFEi6 + rm /tmp/tmp.HhtpXI9SkP /tmp/tmp.3NlMuFFEi6 + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/issuer_no-proxysql-pxc-issuer.yml /tmp/tmp.RXZu460Y3x/issuer_no-proxysql-pxc-issuer.yml + log 'compare_kubectl: issuer/no-proxysql-pxc-issuer OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:57:38+0000]' compare_kubectl: issuer/no-proxysql-pxc-issuer OK [2025-12-01T16:57:38+0000] compare_kubectl: issuer/no-proxysql-pxc-issuer OK + compare_kubectl certificate/no-proxysql-ssl + local resource=certificate/no-proxysql-ssl + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl.yml + local new_result=/tmp/tmp.RXZu460Y3x/certificate_no-proxysql-ssl.yml + desc 'compare certificate/no-proxysql-ssl-' + set +o xtrace ----------------------------------------------------------------------------------- compare certificate/no-proxysql-ssl- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k129.yml ']' + version_gt 1.27 ++ echo '1.31 >= 1.27' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k124.yml ']' + version_gt 1.22 ++ echo '1.31 >= 1.22' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-oc.yml ']' + version_gt 1.29 ++ echo '1.31 >= 1.29' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-aks.yml ']' + kubectl_bin get -o yaml certificate/no-proxysql-ssl ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_OUT=/tmp/tmp.pst6cTjhDk ++ mktemp + local LAST_ERR=/tmp/tmp.OJxKAbZHrJ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml certificate/no-proxysql-ssl + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.pst6cTjhDk + cat /tmp/tmp.OJxKAbZHrJ + rm /tmp/tmp.pst6cTjhDk /tmp/tmp.OJxKAbZHrJ + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl.yml /tmp/tmp.RXZu460Y3x/certificate_no-proxysql-ssl.yml + log 'compare_kubectl: certificate/no-proxysql-ssl OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:57:40+0000]' compare_kubectl: certificate/no-proxysql-ssl OK [2025-12-01T16:57:40+0000] compare_kubectl: certificate/no-proxysql-ssl OK + compare_kubectl certificate/no-proxysql-ssl-internal + local resource=certificate/no-proxysql-ssl-internal + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal.yml + local new_result=/tmp/tmp.RXZu460Y3x/certificate_no-proxysql-ssl-internal.yml + desc 'compare certificate/no-proxysql-ssl-internal-' + set +o xtrace ----------------------------------------------------------------------------------- compare certificate/no-proxysql-ssl-internal- ----------------------------------------------------------------------------------- + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-eks.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-80.yml ']' + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + version_gt 1.33 ++ echo '1.31 >= 1.33' ++ bc -l + '[' 0 -eq 1 ']' + return 1 + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k129.yml ']' + version_gt 1.27 ++ bc -l ++ echo '1.31 >= 1.27' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k127.yml ']' + version_gt 1.24 ++ echo '1.31 >= 1.24' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k124.yml ']' + version_gt 1.22 ++ bc -l ++ echo '1.31 >= 1.22' + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k122.yml ']' + version_gt 1.21 ++ echo '1.31 >= 1.21' ++ bc -l + '[' 1 -eq 1 ']' + return 0 + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k121.yml ']' + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-oc.yml ']' + version_gt 1.29 ++ bc -l ++ echo '1.31 >= 1.29' + '[' 1 -eq 1 ']' + return 0 + '[' '!' -z '' -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-k129-oc.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-eks.yml ']' + '[' 0 = 1 -a -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal-aks.yml ']' + kubectl_bin get -o yaml certificate/no-proxysql-ssl-internal ++ mktemp + local LAST_OUT=/tmp/tmp.MhIIkTAzJU ++ mktemp + yq eval ' del(.metadata.managedFields) | del(.. | select(has("creationTimestamp")).creationTimestamp) | del(.. | select(has("namespace")).namespace) | del(.. | select(has("uid")).uid) | del(.metadata.resourceVersion) | del(.spec.template.spec.containers[].env[] | select(.name == "CLUSTER_HASH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "BACKUP_PATH")) | del(.spec.template.spec.containers[].env[] | select(.name == "S3_BUCKET_URL")) | del(.spec.template.spec.containers[].env[] | select(.name == "AZURE_CONTAINER_NAME")) | del(.metadata.selfLink) | del(.metadata.deletionTimestamp) | del(.metadata.annotations."kubectl.kubernetes.io/last-applied-configuration") | del(.metadata.annotations."kubernetes.io/psp") | del(.metadata.annotations."batch.kubernetes.io/job-tracking") | del(.metadata.labels."batch.kubernetes.io/job-name") | del(.metadata.labels."job-name") | del(.metadata.annotations."cloud.google.com/neg") | del(.metadata.annotations."k8s.v1.cni.cncf.io*") | del(.metadata.annotations."k8s.ovn.org/pod-networks") | del(.spec.template.metadata.annotations."last-applied-secret") | del(.spec.template.metadata.labels."batch.kubernetes.io/job-name") | del(.spec.template.metadata.labels."job-name") | del(.. | select(has("batch.kubernetes.io/controller-uid"))."batch.kubernetes.io/controller-uid") | del(.. | select(has("image")).image) | del(.. | select(has("clusterIP")).clusterIP) | del(.. | select(has("clusterIPs")).clusterIPs) | del(.. | select(has("dataSource")).dataSource) | del(.. | select(has("procMount")).procMount) | del(.. | select(has("storageClassName")).storageClassName) | del(.. | select(has("finalizers")).finalizers) | del(.. | select(has("kubernetes.io/pvc-protection"))."kubernetes.io/pvc-protection") | del(.. | select(has("volumeName")).volumeName) | del(.. | select(has("volume.beta.kubernetes.io/storage-provisioner"))."volume.beta.kubernetes.io/storage-provisioner") | del(.. | select(has("volume.kubernetes.io/storage-provisioner"))."volume.kubernetes.io/storage-provisioner") | del(.spec.volumeMode) | del(.spec.nodeName) | del(.. | select(has("volume.kubernetes.io/selected-node"))."volume.kubernetes.io/selected-node") | del(.. | select(has("percona.com/last-config-hash"))."percona.com/last-config-hash") | del(.. | select(has("percona.com/configuration-hash"))."percona.com/configuration-hash") | del(.. | select(has("percona.com/env-secret-config-hash"))."percona.com/env-secret-config-hash") | del(.. | select(has("percona.com/ssl-hash"))."percona.com/ssl-hash") | del(.. | select(has("percona.com/ssl-internal-hash"))."percona.com/ssl-internal-hash") | del(.. | select(has("kubectl.kubernetes.io/default-container"))."kubectl.kubernetes.io/default-container") | del(.spec.volumeClaimTemplates[].spec.volumeMode | select(. == "Filesystem")) | del(.. | select(has("healthCheckNodePort")).healthCheckNodePort) | del(.. | select(has("nodePort")).nodePort) | del(.. | select(has("imagePullSecrets")).imagePullSecrets) | del(.. | select(has("enableServiceLinks")).enableServiceLinks) | del(.status) | del(.spec.volumeClaimTemplates[].apiVersion) | del(.spec.volumeClaimTemplates[].kind) | del(.metadata.ownerReferences[].apiVersion) | del(.. | select(has("controller-uid")).controller-uid) | del(.. | select(has("preemptionPolicy")).preemptionPolicy) | del(.spec.ipFamilies) | del(.spec.ipFamilyPolicy) | (.. | select(. == "policy/v1beta1")) = "policy/v1" | del(.. | select(has("kubernetes.io/hostname"))."kubernetes.io/hostname") | (.. | select(tag == "!!str")) |= sub("init-deploy-9470", "namespace") | (.. | select(tag == "!!str")) |= sub("kube-api-access-.*", "kube-api-access") | del(.. | select(has("annotations")).annotations | select(length==0)) | del(.spec.crVersion) | del(.. | select(.[] == "percona-xtradb-cluster-operator-workload-token*"))' - + local LAST_ERR=/tmp/tmp.QXND4ty0Ge + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get -o yaml certificate/no-proxysql-ssl-internal + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.MhIIkTAzJU + cat /tmp/tmp.QXND4ty0Ge + rm /tmp/tmp.MhIIkTAzJU /tmp/tmp.QXND4ty0Ge + return 0 + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/certificate_no-proxysql-ssl-internal.yml /tmp/tmp.RXZu460Y3x/certificate_no-proxysql-ssl-internal.yml + log 'compare_kubectl: certificate/no-proxysql-ssl-internal OK' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-12-01T16:57:41+0000]' compare_kubectl: certificate/no-proxysql-ssl-internal OK [2025-12-01T16:57:41+0000] compare_kubectl: certificate/no-proxysql-ssl-internal OK + desc 'write data, read from all' + set +o xtrace ----------------------------------------------------------------------------------- write data, read from all ----------------------------------------------------------------------------------- + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' '-h no-proxysql-pxc -uroot -proot_password' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local 'uri=-h no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kMZdgUXLAF +++ mktemp ++ local LAST_ERR=/tmp/tmp.yg23tX5hSB ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kMZdgUXLAF ++ cat /tmp/tmp.yg23tX5hSB ++ rm /tmp/tmp.kMZdgUXLAF /tmp/tmp.yg23tX5hSB ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h no-proxysql-pxc -uroot -proot_password' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ytPO1tsZs2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.CQi8gnPctA ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ytPO1tsZs2 ++ cat /tmp/tmp.CQi8gnPctA ++ rm /tmp/tmp.ytPO1tsZs2 /tmp/tmp.CQi8gnPctA ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + sleep 10 + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IZKjIgLNMc +++ mktemp ++ local LAST_ERR=/tmp/tmp.Ku016YQnel ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IZKjIgLNMc ++ cat /tmp/tmp.Ku016YQnel ++ rm /tmp/tmp.IZKjIgLNMc /tmp/tmp.Ku016YQnel ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.RXZu460Y3x/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-1.no-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-1.no-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-1.no-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-1.no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.m74f1gOOvr +++ mktemp ++ local LAST_ERR=/tmp/tmp.8mq2W1Lc55 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.m74f1gOOvr ++ cat /tmp/tmp.8mq2W1Lc55 ++ rm /tmp/tmp.m74f1gOOvr /tmp/tmp.8mq2W1Lc55 ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.RXZu460Y3x/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-2.no-proxysql-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-2.no-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT * from myApp.myApp;' '-h no-proxysql-pxc-2.no-proxysql-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h no-proxysql-pxc-2.no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.1f4UfQN8XR +++ mktemp ++ local LAST_ERR=/tmp/tmp.SD3eXRQcxf ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.1f4UfQN8XR ++ cat /tmp/tmp.SD3eXRQcxf ++ rm /tmp/tmp.1f4UfQN8XR /tmp/tmp.SD3eXRQcxf ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/select-1.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/select-1.sql /tmp/tmp.RXZu460Y3x/select-1.sql + compare_mysql_cmd max_allowed_packet-2 'SELECT @@max_allowed_packet;' '-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local command_id=max_allowed_packet-2 + local 'command=SELECT @@max_allowed_packet;' + local 'uri=-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/max_allowed_packet-2.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.4 ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + [[ -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/max_allowed_packet-2-80.sql ]] + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'SELECT @@max_allowed_packet;' '-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' + local 'command=SELECT @@max_allowed_packet;' + local 'uri=-h no-proxysql-pxc-0.no-proxysql-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6eI8BRGqmT +++ mktemp ++ local LAST_ERR=/tmp/tmp.MpUlHWPigt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6eI8BRGqmT ++ cat /tmp/tmp.MpUlHWPigt ++ rm /tmp/tmp.6eI8BRGqmT /tmp/tmp.MpUlHWPigt ++ return 0 + client_pod=pxc-client-59944c5bbf-4bjr6 + wait_pod pxc-client-59944c5bbf-4bjr6 + local pod=pxc-client-59944c5bbf-4bjr6 + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-4bjr6 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ grep -E '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-4bjr6 condition met waiting for pod/pxc-client-59944c5bbf-4bjr6 to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + '[' '!' -s /tmp/tmp.RXZu460Y3x/max_allowed_packet-2.sql ']' + [[ 0 -eq 0 ]] + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/compare/max_allowed_packet-2.sql /tmp/tmp.RXZu460Y3x/max_allowed_packet-2.sql + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/conf/no-proxysql.yml ++ mktemp + local LAST_OUT=/tmp/tmp.kyFbkmq0Ga ++ mktemp + local LAST_ERR=/tmp/tmp.cerBFvhZFT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2221/e2e-tests/init-deploy/conf/no-proxysql.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.kyFbkmq0Ga perconaxtradbcluster.pxc.percona.com "no-proxysql" deleted from init-deploy-9470 namespace + cat /tmp/tmp.cerBFvhZFT + rm /tmp/tmp.kyFbkmq0Ga /tmp/tmp.cerBFvhZFT + return 0 + destroy init-deploy-9470 + local namespace=init-deploy-9470 + local ignore_logs=true + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' + grep -v level=info + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + sort -u + tee /tmp/tmp.RXZu460Y3x/operator.log + grep -v 'the object has been modified' + grep -v 'get backup status: Job.batch' ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.oVPx0eVS8m +++ mktemp ++ local LAST_ERR=/tmp/tmp.3EQ86tSa9N ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.oVPx0eVS8m ++ cat /tmp/tmp.3EQ86tSa9N ++ rm /tmp/tmp.oVPx0eVS8m /tmp/tmp.3EQ86tSa9N ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-bbcb56d74-drjwh ++ mktemp + local LAST_OUT=/tmp/tmp.Ogl0TuNzql ++ mktemp + local LAST_ERR=/tmp/tmp.6dRPgQOsm5 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-bbcb56d74-drjwh + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Ogl0TuNzql + cat /tmp/tmp.6dRPgQOsm5 + rm /tmp/tmp.Ogl0TuNzql /tmp/tmp.6dRPgQOsm5 + return 0 2025-12-01T16:42:14.079Z INFO setup Manager starting up {"gitCommit": "abd799666834bf5ec898319addec65137c9e717f", "gitBranch": "PR-2221-abd79966", "buildTime": "2025-12-01T15:59:26Z", "goVersion": "go1.25.4", "os": "linux", "arch": "amd64"} 2025-12-01T16:42:14.079Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.31.14-gke.1033000"} 2025-12-01T16:42:14.082Z INFO setup Registering Components. 2025-12-01T16:42:14.766Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2025-12-01T16:42:14.766Z INFO controller-runtime.metrics Starting metrics server 2025-12-01T16:42:14.766Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2025-12-01T16:42:14.766Z INFO controller-runtime.webhook Starting webhook server 2025-12-01T16:42:14.766Z INFO setup Starting the Cmd. 2025-12-01T16:42:14.766Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2025-12-01T16:42:14.767Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2025-12-01T16:42:14.767Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2025-12-01T16:42:14.767Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2025-12-01T16:42:14.867Z INFO attempting to acquire leader lease pxc-operator/08db1feb.percona.com... 2025-12-01T16:42:14.893Z DEBUG events percona-xtradb-cluster-operator-bbcb56d74-drjwh_eb84f8c9-ec6a-4d54-a9db-f1d1557b1899 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"1942b9f0-e364-432d-9db7-078ef664d268","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1764607334886911009"}, "reason": "LeaderElection"} 2025-12-01T16:42:14.893Z INFO Starting EventSource {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2025-12-01T16:42:14.893Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.PerconaXtraDBCluster"} 2025-12-01T16:42:14.893Z INFO Starting EventSource {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "source": "kind source: *v1.Secret"} 2025-12-01T16:42:14.893Z INFO Starting EventSource {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2025-12-01T16:42:14.893Z INFO successfully acquired lease pxc-operator/08db1feb.percona.com 2025-12-01T16:42:14.994Z INFO Starting Controller {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup"} 2025-12-01T16:42:14.994Z INFO Starting Controller {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster"} 2025-12-01T16:42:14.994Z INFO Starting Controller {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore"} 2025-12-01T16:42:14.994Z INFO Starting workers {"controller": "pxcbackup-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterBackup", "worker count": 1} 2025-12-01T16:42:14.994Z INFO Starting workers {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "worker count": 1} 2025-12-01T16:42:14.994Z INFO Starting workers {"controller": "pxcrestore-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBClusterRestore", "worker count": 1} 2025-12-01T16:44:26.436Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296", "version": "1.19.0"} 2025-12-01T16:44:26.762Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296"} 2025-12-01T16:44:29.813Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296"} 2025-12-01T16:44:29.845Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296"} 2025-12-01T16:44:32.947Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-12-01T16:44:33.098Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-12-01T16:44:33.149Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-12-01T16:44:33.229Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-12-01T16:44:33.289Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296", "object": "some-name-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-12-01T16:44:33.330Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-12-01T16:44:33.440Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2f63d9c7-45dc-4de7-a0a6-d7830c2a5296", "object": "some-name-proxysql-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-12-01T16:44:34.093Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "3293611d-0060-4050-b095-e4283fefe33d", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-12-01T16:44:34.142Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "3293611d-0060-4050-b095-e4283fefe33d", "object": "some-name-proxysql", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-12-01T16:45:51.308Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "67245dfe-054b-460e-9bc5-d0036304c0f9", "user": "operator"} 2025-12-01T16:45:51.374Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "67245dfe-054b-460e-9bc5-d0036304c0f9", "user": "monitor"} 2025-12-01T16:45:51.430Z INFO User monitor: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "67245dfe-054b-460e-9bc5-d0036304c0f9"} 2025-12-01T16:45:51.463Z INFO monitor user privileges granted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "67245dfe-054b-460e-9bc5-d0036304c0f9"} 2025-12-01T16:45:51.495Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "67245dfe-054b-460e-9bc5-d0036304c0f9", "user": "xtrabackup"} 2025-12-01T16:45:51.538Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "67245dfe-054b-460e-9bc5-d0036304c0f9"} 2025-12-01T16:45:51.581Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "67245dfe-054b-460e-9bc5-d0036304c0f9", "user": "replication"} 2025-12-01T16:45:51.591Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "67245dfe-054b-460e-9bc5-d0036304c0f9", "err": "get primary pxc pod: not found"} 2025-12-01T16:45:56.082Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "fed8d106-692f-43fb-84a4-df2deb5de51d", "err": "get primary pxc pod: not found"} 2025-12-01T16:46:01.225Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "694147a0-99a4-4992-8335-f7ebe4b20883", "err": "get primary pxc pod: not found"} 2025-12-01T16:46:06.374Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "34125db0-ab95-490b-b3c6-54487ca23ae2", "err": "get primary pxc pod: not found"} 2025-12-01T16:48:12.435Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "5be21be8-54e5-4d9d-94a8-1d0ec62dc860", "user": "root"} 2025-12-01T16:48:12.596Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "5be21be8-54e5-4d9d-94a8-1d0ec62dc860", "new version": "8.0.43-34.1"} 2025-12-01T16:48:14.081Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "5be21be8-54e5-4d9d-94a8-1d0ec62dc860"} 2025-12-01T16:48:19.187Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2b655e43-18f5-4fff-9508-e2ad1fa4de76"} 2025-12-01T16:48:24.487Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "e165beef-1844-4a60-9e67-2de490c6fb81"} 2025-12-01T16:48:29.681Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "89c639e8-d413-4606-8473-4ae1455aead2"} 2025-12-01T16:48:35.105Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "81453d20-fe45-4d39-aeba-3176948912ba"} 2025-12-01T16:48:40.502Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "996c57af-f5d9-4332-bf75-0640b2de15f4"} 2025-12-01T16:48:45.691Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "88b9bc9a-c64e-49f4-b797-1a306d7e9f10"} 2025-12-01T16:48:51.684Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "e99582d6-82c6-42e7-b267-73ccf0015c2a"} 2025-12-01T16:48:56.363Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "912f638e-5e00-4ce6-9e2e-bb5c66fb85ab"} 2025-12-01T16:49:01.676Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "d73dcb51-74d9-4947-a840-e75a292e909a"} 2025-12-01T16:49:07.985Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2a0fecb5-030c-448a-a679-8f33e55366e4"} 2025-12-01T16:49:13.077Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "fe51bcf9-65c5-4051-af4f-90fe5eb78cac"} 2025-12-01T16:49:18.082Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "1ee30aaf-3693-41d2-bf31-0c625900b0e3"} 2025-12-01T16:49:23.576Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "1e517fed-ea40-489a-82eb-46f28b19aaea"} 2025-12-01T16:49:28.383Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "c968815b-fc5f-491d-a963-a13e9538f46a"} 2025-12-01T16:49:33.889Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "e4d280e7-08a1-4224-a52f-221617fe17c1"} 2025-12-01T16:49:39.179Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "417295a6-3ca2-42e4-b3fb-8ceb2c5aee30"} 2025-12-01T16:49:44.485Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2cc139a9-e040-436b-9813-5a306efcc090"} 2025-12-01T16:49:49.882Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "30e362cc-772b-448a-b349-5007ea98dbaa"} 2025-12-01T16:49:55.184Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "f53e6269-2254-40fd-b1b5-4a6464b6be81"} 2025-12-01T16:50:00.585Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "106cca02-afc5-4da0-926a-eeae31419718"} 2025-12-01T16:50:06.045Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "9f834442-e7cc-4c20-9a0a-3569aa13459f"} 2025-12-01T16:50:11.087Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "8df0eadc-e99a-49f5-bd94-8ad6319f6b11"} 2025-12-01T16:50:16.509Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "754fe5f2-f47e-4f37-b60e-16c66dabbf4a"} 2025-12-01T16:50:21.812Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "7b80a2ba-2edc-403b-bfb9-446d6d5481c8"} 2025-12-01T16:50:27.014Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "164a3ee2-a191-47da-81e2-bf812939f59d"} 2025-12-01T16:50:32.285Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "774c4f3d-bd18-4079-8ff2-98ce82a008a1"} 2025-12-01T16:50:37.783Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "ad18e4f5-8481-449a-8f37-cdbfc62849fb"} 2025-12-01T16:50:43.284Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "7e3790d0-27d5-44bf-8695-c59e9ff0d20d"} 2025-12-01T16:50:48.594Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "f046e53f-2fcf-4b71-83d9-abff0e76cf0e"} 2025-12-01T16:50:53.893Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "97431b99-f701-4658-bedd-de102458dadd"} 2025-12-01T16:50:59.263Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "c6784180-7736-49c0-b535-f83b2c94f0f0"} 2025-12-01T16:51:04.520Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "78d28354-ed50-42ad-ba3a-58a19a42eaab"} 2025-12-01T16:51:09.895Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "8fc27768-8fa7-48e2-838d-9ea0b0a30fb5"} 2025-12-01T16:51:15.183Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "9456625e-3d3e-4abb-aff5-c78e04565883"} 2025-12-01T16:51:20.380Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "9325c778-c035-49c2-9123-c731364f2fd8"} 2025-12-01T16:51:25.814Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "d09e0f47-c467-4db9-baa2-cd39d8b860e3"} 2025-12-01T16:51:31.116Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "a70a407b-28f3-4bcd-9014-49463bb79b2b"} 2025-12-01T16:51:36.907Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "51b83f9a-3240-43da-b78c-7be7257995f6"} 2025-12-01T16:51:41.700Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "cd8c1b74-8dcf-4b5c-8ae9-aaaadae90c33"} 2025-12-01T16:51:47.111Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "f2013161-d3c1-4a9a-8362-df17a380a258"} 2025-12-01T16:52:02.772Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "54be4c85-3c3a-4851-a001-ce5e51150f1d"} 2025-12-01T16:52:56.387Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "37dec713-4cb9-441a-a6f6-cf82a7bf207b"} 2025-12-01T16:53:01.475Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "585ece5c-4115-4d02-93d3-6affe09bdd5b"} 2025-12-01T16:53:07.208Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "8e0cf346-640b-488c-b96f-530e58dac968"} 2025-12-01T16:53:11.886Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "37a1d245-15fc-4357-89d6-84577683c068"} 2025-12-01T16:53:17.084Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "727ef0fc-78cb-4a2f-b37a-354f00012db0"} 2025-12-01T16:53:22.868Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "4978140b-5f84-4811-a307-539edc4448dc"} 2025-12-01T16:53:28.117Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "b938b115-1c8a-419e-afc3-cbf9b1b16ed8"} 2025-12-01T16:53:32.498Z INFO Set CR version {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "2e8dc10e-3e3f-474c-ad8f-0122d7af3c91", "version": "1.19.0"} 2025-12-01T16:53:32.693Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "2e8dc10e-3e3f-474c-ad8f-0122d7af3c91"} 2025-12-01T16:53:33.320Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "ad2f5222-15a5-4263-9350-fa2fc9720729"} 2025-12-01T16:53:36.357Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "2e8dc10e-3e3f-474c-ad8f-0122d7af3c91"} 2025-12-01T16:53:36.521Z INFO spec.privateKey.rotationPolicy: In cert-manager >= v1.18.0, the default value changed from `Never` to `Always`. {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "2e8dc10e-3e3f-474c-ad8f-0122d7af3c91"} 2025-12-01T16:53:39.522Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "2e8dc10e-3e3f-474c-ad8f-0122d7af3c91", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-12-01T16:53:39.540Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "2e8dc10e-3e3f-474c-ad8f-0122d7af3c91", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-12-01T16:53:40.087Z ERROR Reconciler error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "2e8dc10e-3e3f-474c-ad8f-0122d7af3c91", "error": "pxc upgrade error: reconcile config: reconcile custom config: create or update config map: configmaps \"no-proxysql-pxc\" already exists", "errorVerbose": "configmaps \"no-proxysql-pxc\" already exists\ncreate or update config map\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileCustomConfigMap\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:115\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:31\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:52\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:373\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile custom config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileConfigMaps\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/config.go:33\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:52\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:373\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\nreconcile config\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updatePod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:54\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:373\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\npxc upgrade error\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:375\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} 2025-12-01T16:53:40.500Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "78746fd6-ae04-4c7e-9543-a81053f8a4fb", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-12-01T16:53:40.537Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "78746fd6-ae04-4c7e-9543-a81053f8a4fb", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-12-01T16:53:40.597Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "78746fd6-ae04-4c7e-9543-a81053f8a4fb", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-12-01T16:53:40.653Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "78746fd6-ae04-4c7e-9543-a81053f8a4fb", "object": "no-proxysql-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-12-01T16:53:41.351Z DEBUG Creating object {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "7b975f9c-0ed0-4eed-8bbb-7e5786d5c5c5", "object": "no-proxysql-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-12-01T16:53:41.790Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "df3f1bcd-fec6-4dc7-9d17-50db70f7fa3f"} 2025-12-01T16:53:47.086Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "20de95d9-c5cf-4d62-856a-b7710064807a"} 2025-12-01T16:53:53.290Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "d95eeb3d-fa39-45cd-ae70-0854b954a0c1"} 2025-12-01T16:53:58.682Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "4aba73c9-7fc3-4c29-9386-1ce087e726c6"} 2025-12-01T16:54:04.101Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "44586c07-78bc-4749-9323-bec79e5f0f94"} 2025-12-01T16:54:09.183Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "fb065e8c-b55c-4f88-a898-d4ed3c3b2671"} 2025-12-01T16:54:14.684Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "fda668f7-9d1d-493b-b4ea-ad4432b17f2a"} 2025-12-01T16:54:19.888Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "a1344570-1855-4193-9c96-9effb1b73f5d"} 2025-12-01T16:54:25.185Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "a9d67c01-f3cf-4746-b203-b61685fa23a3"} 2025-12-01T16:54:30.687Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2be32ead-0f36-4711-9cc2-f472aaceb536"} 2025-12-01T16:54:36.688Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "3cdd55c0-55ee-4eaa-9268-4d354b78617d"} 2025-12-01T16:54:41.185Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "3d5edc02-365d-460f-8c78-198677dd6944"} 2025-12-01T16:54:46.326Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "b40963b5-ee1c-4418-96b4-fd6557052bf6"} 2025-12-01T16:54:52.690Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "33fa9cf9-0fc1-475a-9543-5551c4745fdd"} 2025-12-01T16:54:55.844Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "09b6e57d-fedd-4b87-a008-557552fbedef", "user": "operator"} 2025-12-01T16:54:55.892Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "09b6e57d-fedd-4b87-a008-557552fbedef", "user": "monitor"} 2025-12-01T16:54:55.945Z INFO User monitor: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "09b6e57d-fedd-4b87-a008-557552fbedef"} 2025-12-01T16:54:55.979Z INFO monitor user privileges granted {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "09b6e57d-fedd-4b87-a008-557552fbedef"} 2025-12-01T16:54:56.019Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "09b6e57d-fedd-4b87-a008-557552fbedef", "user": "xtrabackup"} 2025-12-01T16:54:56.064Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "09b6e57d-fedd-4b87-a008-557552fbedef"} 2025-12-01T16:54:56.106Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "09b6e57d-fedd-4b87-a008-557552fbedef", "user": "replication"} 2025-12-01T16:54:56.108Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "09b6e57d-fedd-4b87-a008-557552fbedef", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:54:57.186Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "86002b55-680c-4ced-bcd5-dbcf8fc2badb"} 2025-12-01T16:55:00.935Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "e8037349-e36b-49bf-a44a-6c29add7a0e7", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:02.508Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "3ab751b6-e46b-4764-a932-91c71bca9d5f"} 2025-12-01T16:55:06.625Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "be11fbc6-bddf-4eeb-ad8a-e83ae780ae25", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:07.887Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "9fa5f33e-367a-4a9d-9ea1-3c0881e4f952"} 2025-12-01T16:55:11.940Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "130d01c1-17f5-4add-b29f-78cbdf68a8b4", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:13.400Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "0020bc26-0505-45a7-9959-fc6614205bde"} 2025-12-01T16:55:17.268Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "f8f638cf-8e13-44c9-be7f-bed3f0da586e", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:18.977Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "acd3fd74-0c5b-4dce-ae64-94241c1f3bfb"} 2025-12-01T16:55:22.562Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "98fe4ab1-4b63-4e31-b310-65714dc4aaa2", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:24.296Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "2b7f775d-7631-44b8-a43d-9475ee8576a6"} 2025-12-01T16:55:27.855Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "44358cab-872f-4cd5-8ae4-2e9eacf9a750", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:29.500Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "de960380-8991-433c-bf8e-ec381f1ff170"} 2025-12-01T16:55:33.232Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "82f358de-4776-4cb6-8825-30d4b81cd6b3", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:34.686Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "0c12aaf1-f330-47c7-bfa8-1bdc9c26f341"} 2025-12-01T16:55:38.568Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "2c9b4398-ddf4-426b-8259-4f2af1616395", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:40.194Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "f72ac437-9316-4b40-9619-0f112292ebae"} 2025-12-01T16:55:43.835Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "e82f3ba8-a20d-4767-8855-1f94181f7c23", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:45.690Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "40dd0158-504a-4881-9e64-b172c7dc2b30"} 2025-12-01T16:55:49.145Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "71504443-e58d-4ea8-bff1-80bc58dd4466", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:50.681Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "62e3dff4-59f4-451d-89d1-0ea210573888"} 2025-12-01T16:55:54.432Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "b316154a-d217-4efe-9cfc-edc39bc95c8b", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:55:56.170Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "c1d76fd8-40c8-4ae5-abfa-6dff1e1df4e7"} 2025-12-01T16:55:59.701Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "4badf45b-1a63-4f05-a009-149ced256a4e", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:01.401Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "3fedc942-3d30-48e3-b56d-050fb29d91b7"} 2025-12-01T16:56:05.039Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "0ec7f8b8-0254-469d-a097-067b384e8487", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:06.991Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "d30a212a-9759-4737-a03e-c8ed617503df"} 2025-12-01T16:56:10.330Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "652b3936-9289-43c8-8a5c-bf7d3250f5ef", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:11.037Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "d8047b84-39c5-45a9-8c73-689bae838f33", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:12.008Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "a31011d7-0aa6-4578-8c0c-633d182d9c15"} 2025-12-01T16:56:15.616Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "b847ad9e-e496-4418-964d-8d42df4e463f", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:17.289Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "6325f204-99c3-47e3-9f43-89b471433d5c"} 2025-12-01T16:56:21.158Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "07056cc0-bff9-400a-b4f0-e3093e9f1754", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:22.861Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "f29b9608-ac7f-4ad4-9532-23b429eae4df"} 2025-12-01T16:56:26.421Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "12dd5a7c-a594-46c0-8329-959cfb28b3ac", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:28.001Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "a4f3d9d4-e355-45db-831f-c6789cb94abc"} 2025-12-01T16:56:31.765Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "65c09d56-2146-4e5a-a0c8-78f5d48eb6ef", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:33.592Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "eef1f111-acec-4093-a34b-c8e405bc7475"} 2025-12-01T16:56:37.112Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "b87359b9-6b5b-44c9-a40b-acd98ac46e8e", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:38.995Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "70430ae1-bfad-4a7e-84d0-a72342ba580b"} 2025-12-01T16:56:42.354Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "dcd034d8-4fe6-413b-b4a2-e0bf869e3d44", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:43.911Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "e785c5b2-6807-4000-8cf1-b242eb10dc09"} 2025-12-01T16:56:47.653Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "02382941-2b47-4a30-83c7-d789ac17e9dc", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:49.417Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "4d1473fd-7b9c-42f3-9ba7-ed23883c57c7"} 2025-12-01T16:56:52.961Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "6a96b295-08f3-4fa1-b30f-8f25e6814d72", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:56:54.766Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "e6ed1b8d-c0c0-464a-a442-c3213a3b3239"} 2025-12-01T16:56:58.265Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "d2b15348-b203-4d75-9ac9-1fbed0eb45f3", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:00.065Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "d118d57c-93bb-46f8-b7bb-0da564361ef5"} 2025-12-01T16:57:03.564Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "3ae582c5-1df7-4505-b4cb-8e1fd4a0d275", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:05.397Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "c01be87b-aea5-4edc-af6e-0296b43e1136"} 2025-12-01T16:57:08.874Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "1049e91b-39f8-47ae-8680-abddfde20c81", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:10.677Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "cf91ca5c-07ca-48e3-ab5c-72a8fafe74cf"} 2025-12-01T16:57:14.173Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "9d2924d3-067f-447a-982a-734025999efd", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:15.790Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "deddc07b-eef5-4ac7-ad91-a05e87e0f857"} 2025-12-01T16:57:19.465Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "c835cc56-4302-47e6-8c9c-e91037d771a5", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:21.095Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "da1f36c4-8cbf-49ea-b9c8-10ce53f11f27"} 2025-12-01T16:57:24.836Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "73e48420-fe1f-4e1e-8cbf-0035a61a866f", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:26.106Z INFO Password expiration policy updated {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "5d7d0ba1-fd8a-4f7c-9197-d5080eef972d", "user": "root"} 2025-12-01T16:57:26.151Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "5d7d0ba1-fd8a-4f7c-9197-d5080eef972d", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:26.157Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "5d7d0ba1-fd8a-4f7c-9197-d5080eef972d", "new version": "8.0.43-34.1"} 2025-12-01T16:57:26.298Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "894ba87b-f16a-4900-9a2c-08f32788e6d6"} 2025-12-01T16:57:27.477Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "f09896d1-9258-483c-84d6-586a5de8a905", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:30.955Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "a6883ed0-6c73-4925-9744-602b40d80444", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:31.889Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "95a627c6-c5db-4d53-904d-1d2ce89179fa"} 2025-12-01T16:57:36.704Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "15767bec-236d-4d45-9400-4dbff65a4986", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:37.982Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "1dd8b3de-4a01-4a16-bc12-e57d952d2176"} 2025-12-01T16:57:41.919Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "b2c928ae-bfc5-4de9-ba29-bf93b8047f43", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:43.385Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "9aa2c0c3-1157-49d5-afea-b932d403a0c3"} 2025-12-01T16:57:47.482Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "8e535637-2a31-46f3-8b8d-6173949883f6", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:49.105Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "4cbb4c21-7f33-45b5-a2ee-7612389e8632"} 2025-12-01T16:57:52.752Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "fb4541c0-a71c-4ca8-96c0-2143ef9b0f99", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:54.189Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "dd7640e8-4044-4f00-8040-8dc8edad6a33"} 2025-12-01T16:57:58.057Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "10967ecb-3dfb-41bb-bc14-163dfb216e78", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:57:59.306Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "9ebdab37-c16d-4852-adf8-bbe9ebb17802"} 2025-12-01T16:58:03.353Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "6364f843-5d76-46ba-a564-b1e9239ac44e", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:58:04.575Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "1dd404ec-e648-4bee-8371-2d9d199d7d17"} 2025-12-01T16:58:08.657Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "85ec05d7-0c20-40e8-b1eb-9f4326bd6310", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:58:10.078Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "c2ce61ed-9c9d-46eb-acf0-ae57a61d46d8"} 2025-12-01T16:58:13.860Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "441fdcb7-190b-4501-89c0-1e42b689bbf5", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:58:15.295Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "91c37ee1-c416-43a7-8c65-ff1d1fcfa33c"} 2025-12-01T16:58:19.185Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "990230d0-8fb7-4156-a03b-24a892c45bfd", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:58:20.380Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "50064a8f-da76-4da1-a38b-bcd9ce9cff2b"} 2025-12-01T16:58:24.466Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "326e7f89-0255-4316-b9d7-f3ea20fbb727", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:58:25.567Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "a001d9e4-7e3d-4a46-b7b5-4106859467a6"} 2025-12-01T16:58:29.761Z INFO reconcile replication error {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"no-proxysql","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "no-proxysql", "reconcileID": "ced47f04-8523-426e-8f92-8003570a9026", "err": "get primary pxc pod: failed to get proxy connection: can't detect enabled proxy, please enable HAProxy or ProxySQL"} 2025-12-01T16:58:31.275Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "controllerGroup": "pxc.percona.com", "controllerKind": "PerconaXtraDBCluster", "PerconaXtraDBCluster": {"name":"some-name","namespace":"init-deploy-9470"}, "namespace": "init-deploy-9470", "name": "some-name", "reconcileID": "8655a409-bce1-4072-85ea-cf6bdc58fbb5"} /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:474 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 + kubectl get pxc --all-namespaces -o wide + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + grep -v NAMESPACE + kubectl patch pxc -n init-deploy-9470 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.VxWjSXXkHG ++ mktemp + local LAST_ERR=/tmp/tmp.1HghXgpJpy + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.VxWjSXXkHG perconaxtradbcluster.pxc.percona.com "some-name" deleted from init-deploy-9470 namespace + cat /tmp/tmp.1HghXgpJpy + rm /tmp/tmp.VxWjSXXkHG /tmp/tmp.1HghXgpJpy + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.wSGlqKpvhP ++ mktemp + local LAST_ERR=/tmp/tmp.dugurznwDk + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.wSGlqKpvhP No resources found + cat /tmp/tmp.dugurznwDk + rm /tmp/tmp.wSGlqKpvhP /tmp/tmp.dugurznwDk + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.7lFdfqCX3V ++ mktemp + local LAST_ERR=/tmp/tmp.UL2pDK76uT + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.7lFdfqCX3V No resources found + cat /tmp/tmp.UL2pDK76uT + rm /tmp/tmp.7lFdfqCX3V /tmp/tmp.UL2pDK76uT + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.ZBqQ70Hsod ++ mktemp + local LAST_ERR=/tmp/tmp.ugxYdMTlBa + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ZBqQ70Hsod validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.ugxYdMTlBa + rm /tmp/tmp.ZBqQ70Hsod /tmp/tmp.ugxYdMTlBa + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.18.2/cert-manager.yaml namespace "cert-manager" deleted + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace init-deploy-9470 + rm -rf /tmp/tmp.RXZu460Y3x ++ mktemp + local LAST_OUT=/tmp/tmp.Pxz3MA6w2R ++ mktemp + local LAST_ERR=/tmp/tmp.J0JPX9VSl5 + local exit_status=0 ++ seq 0 2 + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace init-deploy-9470 + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.1bPqnNG5Gm ++ mktemp + local LAST_ERR=/tmp/tmp.jxW4XDWfJ6 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator