Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/logs/demand-backup-without-passwords-8-0.log Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 Warning: version difference between client (1.34) and server (1.31) exceeds the supported minor version skew of +/-1 + cluster=some-name + create_infra demand-backup-without-passwords-25983 + local ns=demand-backup-without-passwords-25983 + '[' -n pxc-operator ']' + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + kubectl patch pxc -n demand-backup-without-passwords-16328 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.EUBqoehF6k ++ mktemp + local LAST_ERR=/tmp/tmp.L8kYVP9atB + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.EUBqoehF6k perconaxtradbcluster.pxc.percona.com "some-name" deleted from demand-backup-without-passwords-16328 namespace + cat /tmp/tmp.L8kYVP9atB + rm /tmp/tmp.EUBqoehF6k /tmp/tmp.L8kYVP9atB + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.TzmJu9hWaH ++ mktemp + local LAST_ERR=/tmp/tmp.9dZvUJThyh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.TzmJu9hWaH perconaxtradbclusterbackup.pxc.percona.com "backup1" deleted from demand-backup-without-passwords-16328 namespace + cat /tmp/tmp.9dZvUJThyh + rm /tmp/tmp.TzmJu9hWaH /tmp/tmp.9dZvUJThyh + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.YwEBce0Tv8 ++ mktemp + local LAST_ERR=/tmp/tmp.TshRpr6e4t + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YwEBce0Tv8 No resources found + cat /tmp/tmp.TshRpr6e4t + rm /tmp/tmp.YwEBce0Tv8 /tmp/tmp.TshRpr6e4t + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ tail -n1 ++ helm list --all-namespaces --filter chaos-mesh ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + '[' -n '' ']' ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get MutatingWebhookConfiguration + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ kubectl api-resources ++ awk '{print $1}' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ grep chaos-mesh ++ kubectl get clusterrolebinding + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrole ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator ++ mktemp + xargs kubectl delete ns + local LAST_OUT=/tmp/tmp.dPsuGCLWk9 ++ mktemp + awk '{print$1}' + local LAST_ERR=/tmp/tmp.4tjpRP3Uny + local exit_status=0 + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.jWH3Jxueuk ++ mktemp + local LAST_ERR=/tmp/tmp.YDKqLq9gPI + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.jWH3Jxueuk + cat /tmp/tmp.YDKqLq9gPI + rm /tmp/tmp.jWH3Jxueuk /tmp/tmp.YDKqLq9gPI + return 0 namespace "demand-backup-without-passwords-16328" deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.dPsuGCLWk9 namespace "pxc-operator" deleted + cat /tmp/tmp.4tjpRP3Uny + rm /tmp/tmp.dPsuGCLWk9 /tmp/tmp.4tjpRP3Uny + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'waiting for namespace/pxc-operator to be deleted' waiting for namespace/pxc-operator to be deleted+ set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.swydLiCkL7 ++ mktemp + local LAST_ERR=/tmp/tmp.tlo8R3GFmY + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.swydLiCkL7 namespace/pxc-operator created + cat /tmp/tmp.tlo8R3GFmY + rm /tmp/tmp.swydLiCkL7 /tmp/tmp.tlo8R3GFmY + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.cyisOEyL2d +++ mktemp ++ local LAST_ERR=/tmp/tmp.VQ1IZOPpwi ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.cyisOEyL2d ++ cat /tmp/tmp.VQ1IZOPpwi ++ rm /tmp/tmp.cyisOEyL2d /tmp/tmp.VQ1IZOPpwi ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.x1lDjICRMC ++ mktemp + local LAST_ERR=/tmp/tmp.abV4MmV5Xn + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.x1lDjICRMC Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7" modified. + cat /tmp/tmp.abV4MmV5Xn + rm /tmp/tmp.x1lDjICRMC /tmp/tmp.abV4MmV5Xn + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.HItyEHPeq4 ++ mktemp + local LAST_ERR=/tmp/tmp.bikK65I7MB + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HItyEHPeq4 customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.bikK65I7MB + rm /tmp/tmp.HItyEHPeq4 /tmp/tmp.bikK65I7MB + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' ++ mktemp + local LAST_OUT=/tmp/tmp.H32iC2ubGX ++ mktemp + local LAST_ERR=/tmp/tmp.UrdJvPW1PL + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.H32iC2ubGX clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.UrdJvPW1PL + rm /tmp/tmp.H32iC2ubGX /tmp/tmp.UrdJvPW1PL + return 0 + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-2234-269f3694^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.HIqCbS6F2d ++ mktemp + local LAST_ERR=/tmp/tmp.MxbXKQbypr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/deploy/cw-operator.yaml + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "VERBOSE"' - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HIqCbS6F2d deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.MxbXKQbypr + rm /tmp/tmp.HIqCbS6F2d /tmp/tmp.MxbXKQbypr + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.QFvpMVCFSq ++ mktemp + local LAST_ERR=/tmp/tmp.tqGTKJFsal + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.QFvpMVCFSq pod/percona-xtradb-cluster-operator-b5f9c4897-hns8f condition met + cat /tmp/tmp.tqGTKJFsal + rm /tmp/tmp.QFvpMVCFSq /tmp/tmp.tqGTKJFsal + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ grep -c percona-xtradb-cluster-operator +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.sYleuNjx7V +++ mktemp ++ local LAST_ERR=/tmp/tmp.Q1O2kCT30u ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.sYleuNjx7V ++ cat /tmp/tmp.Q1O2kCT30u ++ rm /tmp/tmp.sYleuNjx7V /tmp/tmp.Q1O2kCT30u ++ return 0 + wait_pod percona-xtradb-cluster-operator-b5f9c4897-hns8f 480 pxc-operator + local pod=percona-xtradb-cluster-operator-b5f9c4897-hns8f + local max_retry=480 + local ns=pxc-operator ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo percona-xtradb-cluster-operator-b5f9c4897-hns8f + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-b5f9c4897-hns8f condition met waiting for pod/percona-xtradb-cluster-operator-b5f9c4897-hns8f to become Ready.Ok + sleep 3 + create_namespace demand-backup-without-passwords-25983 + local namespace=demand-backup-without-passwords-25983 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ sed s/NAMESPACE// ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' + local chaos_mesh_ns= + '[' -n '' ']' ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep chaos-mesh ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ grep validate-auth ++ awk '{print $1}' ++ kubectl get ValidatingWebhookConfiguration + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl api-resources ++ grep chaos-mesh ++ grep chaos-mesh.org ++ kubectl get crd ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrole ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + '[' -n '' ']' + desc 'cleaned up old namespaces demand-backup-without-passwords-25983' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces demand-backup-without-passwords-25983 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace demand-backup-without-passwords-25983 + awk '{print$1}' + kubectl_bin get ns ++ mktemp + local LAST_OUT=/tmp/tmp.mjfPsG8Xx3 ++ mktemp + local LAST_ERR=/tmp/tmp.bF6DWn7mcq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + egrep -v '^kube-|^default|Terminating|pxc-operator|openshift|^gke-|^gmp-|^NAME' + xargs kubectl delete ns ++ mktemp + local LAST_OUT=/tmp/tmp.CdykFEVutP ++ mktemp + local LAST_ERR=/tmp/tmp.A31X6kS2kP + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-without-passwords-25983 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-without-passwords-25983 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.mjfPsG8Xx3 + cat /tmp/tmp.bF6DWn7mcq + rm /tmp/tmp.mjfPsG8Xx3 /tmp/tmp.bF6DWn7mcq + return 0 error: resource(s) were provided, but no name was specified + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace demand-backup-without-passwords-25983 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.CdykFEVutP + cat /tmp/tmp.A31X6kS2kP Error from server (NotFound): namespaces "demand-backup-without-passwords-25983" not found + rm /tmp/tmp.CdykFEVutP /tmp/tmp.A31X6kS2kP + return 1 + : + wait_for_delete namespace/demand-backup-without-passwords-25983 + local res=namespace/demand-backup-without-passwords-25983 + echo -n 'waiting for namespace/demand-backup-without-passwords-25983 to be deleted' waiting for namespace/demand-backup-without-passwords-25983 to be deleted+ set +o xtrace Error from server (NotFound): namespaces "demand-backup-without-passwords-25983" not found + desc 'create namespace demand-backup-without-passwords-25983' + set +o xtrace ----------------------------------------------------------------------------------- create namespace demand-backup-without-passwords-25983 ----------------------------------------------------------------------------------- + kubectl_bin create namespace demand-backup-without-passwords-25983 ++ mktemp + local LAST_OUT=/tmp/tmp.WhHnvI5Ur6 ++ mktemp + local LAST_ERR=/tmp/tmp.CTG6e71WIV + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace demand-backup-without-passwords-25983 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.WhHnvI5Ur6 namespace/demand-backup-without-passwords-25983 created + cat /tmp/tmp.CTG6e71WIV + rm /tmp/tmp.WhHnvI5Ur6 /tmp/tmp.CTG6e71WIV + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.NjPwuwUcqt +++ mktemp ++ local LAST_ERR=/tmp/tmp.tUTZ2mm3G2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.NjPwuwUcqt ++ cat /tmp/tmp.tUTZ2mm3G2 ++ rm /tmp/tmp.NjPwuwUcqt /tmp/tmp.tUTZ2mm3G2 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7 --namespace=demand-backup-without-passwords-25983 ++ mktemp + local LAST_OUT=/tmp/tmp.S51TovUbCh ++ mktemp + local LAST_ERR=/tmp/tmp.IKOffYxnJo + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7 --namespace=demand-backup-without-passwords-25983 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.S51TovUbCh Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-2234-269f3694-3-cluster7" modified. + cat /tmp/tmp.IKOffYxnJo + rm /tmp/tmp.S51TovUbCh /tmp/tmp.IKOffYxnJo + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.ne0Lt7wxg3 ++ mktemp + local LAST_ERR=/tmp/tmp.zLogBPtXO9 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.ne0Lt7wxg3 secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.zLogBPtXO9 + rm /tmp/tmp.ne0Lt7wxg3 /tmp/tmp.zLogBPtXO9 + return 0 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.9pfXVxmmuT ++ mktemp + local LAST_ERR=/tmp/tmp.qjxDqND9TD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9pfXVxmmuT deployment.apps/pxc-client created + cat /tmp/tmp.qjxDqND9TD + rm /tmp/tmp.9pfXVxmmuT /tmp/tmp.qjxDqND9TD + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/demand-backup-without-passwords/conf/some-name.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/demand-backup-without-passwords/conf/some-name.yml + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/demand-backup-without-passwords/conf/some-name.yml + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.demand-backup-without-passwords-25983~ + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-2234-269f3694#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0-backup#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' ++ mktemp + local LAST_OUT=/tmp/tmp.HDpoQ3TfaX ++ mktemp + local LAST_ERR=/tmp/tmp.bxzVKHZjsS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HDpoQ3TfaX perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.bxzVKHZjsS + rm /tmp/tmp.HDpoQ3TfaX /tmp/tmp.bxzVKHZjsS + return 0 + wait_cluster_consistency some-name 3 2 + local cluster_name=some-name + local cluster_size=3 + local proxy_size=2 + '[' -z 2 ']' + desc 'wait cluster consistency' + set +o xtrace ----------------------------------------------------------------------------------- wait cluster consistency ----------------------------------------------------------------------------------- + local i=0 + local max=300 + sleep 7 + echo -n 'waiting for pxc/some-name to be ready' waiting for pxc/some-name to be ready++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zXKzUS8REP +++ mktemp ++ local LAST_ERR=/tmp/tmp.iMB2eODbG1 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zXKzUS8REP ++ cat /tmp/tmp.iMB2eODbG1 ++ rm /tmp/tmp.zXKzUS8REP /tmp/tmp.iMB2eODbG1 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 0 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.yvWzHRiZje +++ mktemp ++ local LAST_ERR=/tmp/tmp.PYiIwEHydU ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.yvWzHRiZje ++ cat /tmp/tmp.PYiIwEHydU ++ rm /tmp/tmp.yvWzHRiZje /tmp/tmp.PYiIwEHydU ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 1 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.OcmtycVKia +++ mktemp ++ local LAST_ERR=/tmp/tmp.yf4awl5LDO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.OcmtycVKia ++ cat /tmp/tmp.yf4awl5LDO ++ rm /tmp/tmp.OcmtycVKia /tmp/tmp.yf4awl5LDO ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 2 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zJBmRCa9yH +++ mktemp ++ local LAST_ERR=/tmp/tmp.8oTOb6ANyv ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zJBmRCa9yH ++ cat /tmp/tmp.8oTOb6ANyv ++ rm /tmp/tmp.zJBmRCa9yH /tmp/tmp.8oTOb6ANyv ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 3 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.o9u4kR4jAM +++ mktemp ++ local LAST_ERR=/tmp/tmp.3FTvD1S7iD ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.o9u4kR4jAM ++ cat /tmp/tmp.3FTvD1S7iD ++ rm /tmp/tmp.o9u4kR4jAM /tmp/tmp.3FTvD1S7iD ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 4 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.QibuGsiDuu +++ mktemp ++ local LAST_ERR=/tmp/tmp.hYykRv6kv4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.QibuGsiDuu ++ cat /tmp/tmp.hYykRv6kv4 ++ rm /tmp/tmp.QibuGsiDuu /tmp/tmp.hYykRv6kv4 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 5 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9FsIiJb1ak +++ mktemp ++ local LAST_ERR=/tmp/tmp.4N0g7n8Cls ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9FsIiJb1ak ++ cat /tmp/tmp.4N0g7n8Cls ++ rm /tmp/tmp.9FsIiJb1ak /tmp/tmp.4N0g7n8Cls ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 6 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.p9nyEs1L3E +++ mktemp ++ local LAST_ERR=/tmp/tmp.fdgGvPB13l ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.p9nyEs1L3E ++ cat /tmp/tmp.fdgGvPB13l ++ rm /tmp/tmp.p9nyEs1L3E /tmp/tmp.fdgGvPB13l ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 7 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.caP27SXdta +++ mktemp ++ local LAST_ERR=/tmp/tmp.ezGzH7Pheb ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.caP27SXdta ++ cat /tmp/tmp.ezGzH7Pheb ++ rm /tmp/tmp.caP27SXdta /tmp/tmp.ezGzH7Pheb ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 8 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.P9NRM1Z1Hl +++ mktemp ++ local LAST_ERR=/tmp/tmp.yBucDTixCT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.P9NRM1Z1Hl ++ cat /tmp/tmp.yBucDTixCT ++ rm /tmp/tmp.P9NRM1Z1Hl /tmp/tmp.yBucDTixCT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 9 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UmN5v6r0eJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.Fh2ghb0hG3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UmN5v6r0eJ ++ cat /tmp/tmp.Fh2ghb0hG3 ++ rm /tmp/tmp.UmN5v6r0eJ /tmp/tmp.Fh2ghb0hG3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 10 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.PWA090jONH +++ mktemp ++ local LAST_ERR=/tmp/tmp.KNTTQzHOTp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.PWA090jONH ++ cat /tmp/tmp.KNTTQzHOTp ++ rm /tmp/tmp.PWA090jONH /tmp/tmp.KNTTQzHOTp ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 11 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.94Pc9iA4rb +++ mktemp ++ local LAST_ERR=/tmp/tmp.h6apnozZK8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.94Pc9iA4rb ++ cat /tmp/tmp.h6apnozZK8 ++ rm /tmp/tmp.94Pc9iA4rb /tmp/tmp.h6apnozZK8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 12 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.pWHbbf7wap +++ mktemp ++ local LAST_ERR=/tmp/tmp.Jva41iMlLT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.pWHbbf7wap ++ cat /tmp/tmp.Jva41iMlLT ++ rm /tmp/tmp.pWHbbf7wap /tmp/tmp.Jva41iMlLT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 13 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.k6lB01qpSg +++ mktemp ++ local LAST_ERR=/tmp/tmp.XTzoAKhMt8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.k6lB01qpSg ++ cat /tmp/tmp.XTzoAKhMt8 ++ rm /tmp/tmp.k6lB01qpSg /tmp/tmp.XTzoAKhMt8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 14 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.t0kwkl4zXe +++ mktemp ++ local LAST_ERR=/tmp/tmp.vPTlZ7Z3j8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.t0kwkl4zXe ++ cat /tmp/tmp.vPTlZ7Z3j8 ++ rm /tmp/tmp.t0kwkl4zXe /tmp/tmp.vPTlZ7Z3j8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 15 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.M3ezh74NiS +++ mktemp ++ local LAST_ERR=/tmp/tmp.auU4hEWLQn ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.M3ezh74NiS ++ cat /tmp/tmp.auU4hEWLQn ++ rm /tmp/tmp.M3ezh74NiS /tmp/tmp.auU4hEWLQn ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 16 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.bvFboOOgcd +++ mktemp ++ local LAST_ERR=/tmp/tmp.R3dkziyvT7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.bvFboOOgcd ++ cat /tmp/tmp.R3dkziyvT7 ++ rm /tmp/tmp.bvFboOOgcd /tmp/tmp.R3dkziyvT7 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 17 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.UUqbDjMUYh +++ mktemp ++ local LAST_ERR=/tmp/tmp.8atdvIktLP ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.UUqbDjMUYh ++ cat /tmp/tmp.8atdvIktLP ++ rm /tmp/tmp.UUqbDjMUYh /tmp/tmp.8atdvIktLP ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 18 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.ra5lRLAOTm +++ mktemp ++ local LAST_ERR=/tmp/tmp.Q9ORhrj8cy ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.ra5lRLAOTm ++ cat /tmp/tmp.Q9ORhrj8cy ++ rm /tmp/tmp.ra5lRLAOTm /tmp/tmp.Q9ORhrj8cy ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 19 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XsFBuzTENq +++ mktemp ++ local LAST_ERR=/tmp/tmp.qpk9Y4YzC8 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XsFBuzTENq ++ cat /tmp/tmp.qpk9Y4YzC8 ++ rm /tmp/tmp.XsFBuzTENq /tmp/tmp.qpk9Y4YzC8 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 20 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.nmgoSrL3tf +++ mktemp ++ local LAST_ERR=/tmp/tmp.CgQDhDUomK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.nmgoSrL3tf ++ cat /tmp/tmp.CgQDhDUomK ++ rm /tmp/tmp.nmgoSrL3tf /tmp/tmp.CgQDhDUomK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 21 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.xfB5fJvr1x +++ mktemp ++ local LAST_ERR=/tmp/tmp.CUoLViyQSM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.xfB5fJvr1x ++ cat /tmp/tmp.CUoLViyQSM ++ rm /tmp/tmp.xfB5fJvr1x /tmp/tmp.CUoLViyQSM ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 22 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.vstekElPGE +++ mktemp ++ local LAST_ERR=/tmp/tmp.snyuMeLFsW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.vstekElPGE ++ cat /tmp/tmp.snyuMeLFsW ++ rm /tmp/tmp.vstekElPGE /tmp/tmp.snyuMeLFsW ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 23 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.97XZRFwoLM +++ mktemp ++ local LAST_ERR=/tmp/tmp.ANdGjpJ1sT ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.97XZRFwoLM ++ cat /tmp/tmp.ANdGjpJ1sT ++ rm /tmp/tmp.97XZRFwoLM /tmp/tmp.ANdGjpJ1sT ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 24 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.20ywUvoMWQ +++ mktemp ++ local LAST_ERR=/tmp/tmp.46ukWFIQhm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.20ywUvoMWQ ++ cat /tmp/tmp.46ukWFIQhm ++ rm /tmp/tmp.20ywUvoMWQ /tmp/tmp.46ukWFIQhm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 25 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.gTxWyt0kGV +++ mktemp ++ local LAST_ERR=/tmp/tmp.RAgnzVmQoG ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.gTxWyt0kGV ++ cat /tmp/tmp.RAgnzVmQoG ++ rm /tmp/tmp.gTxWyt0kGV /tmp/tmp.RAgnzVmQoG ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 26 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GATYozMXpb +++ mktemp ++ local LAST_ERR=/tmp/tmp.cak9CXoGwK ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GATYozMXpb ++ cat /tmp/tmp.cak9CXoGwK ++ rm /tmp/tmp.GATYozMXpb /tmp/tmp.cak9CXoGwK ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 27 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Gf7ibG8Fgs +++ mktemp ++ local LAST_ERR=/tmp/tmp.OOdDlxE9lj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Gf7ibG8Fgs ++ cat /tmp/tmp.OOdDlxE9lj ++ rm /tmp/tmp.Gf7ibG8Fgs /tmp/tmp.OOdDlxE9lj ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 28 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.q2BgmDmpm0 +++ mktemp ++ local LAST_ERR=/tmp/tmp.o6fU34CHnm ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.q2BgmDmpm0 ++ cat /tmp/tmp.o6fU34CHnm ++ rm /tmp/tmp.q2BgmDmpm0 /tmp/tmp.o6fU34CHnm ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 29 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.kU97P6TdD8 +++ mktemp ++ local LAST_ERR=/tmp/tmp.HZrMdnNOJJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.kU97P6TdD8 ++ cat /tmp/tmp.HZrMdnNOJJ ++ rm /tmp/tmp.kU97P6TdD8 /tmp/tmp.HZrMdnNOJJ ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 30 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.MMgP20Bq4i +++ mktemp ++ local LAST_ERR=/tmp/tmp.nLU3KW17f3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.MMgP20Bq4i ++ cat /tmp/tmp.nLU3KW17f3 ++ rm /tmp/tmp.MMgP20Bq4i /tmp/tmp.nLU3KW17f3 ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 31 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Rv9xZCMibW +++ mktemp ++ local LAST_ERR=/tmp/tmp.frKw5IIz8s ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Rv9xZCMibW ++ cat /tmp/tmp.frKw5IIz8s ++ rm /tmp/tmp.Rv9xZCMibW /tmp/tmp.frKw5IIz8s ++ return 0 + [[ initializing == \r\e\a\d\y ]] + echo -n . .+ sleep 5 + [[ 32 -ge 300 ]] + let i+=1 ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.state}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.4oQf7rdpyJ +++ mktemp ++ local LAST_ERR=/tmp/tmp.pqqLW8EqcR ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.state}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.4oQf7rdpyJ ++ cat /tmp/tmp.pqqLW8EqcR ++ rm /tmp/tmp.4oQf7rdpyJ /tmp/tmp.pqqLW8EqcR ++ return 0 + [[ ready == \r\e\a\d\y ]] ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.pxc.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.zMPGBsetsD +++ mktemp ++ local LAST_ERR=/tmp/tmp.TJ05NeDErj ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.pxc.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.zMPGBsetsD ++ cat /tmp/tmp.TJ05NeDErj ++ rm /tmp/tmp.zMPGBsetsD /tmp/tmp.TJ05NeDErj ++ return 0 + [[ 3 == \3 ]] +++ get_proxy_engine some-name +++ local cluster_name=some-name ++++ get_proxy some-name ++++ local target_cluster=some-name +++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++++ mktemp +++++ local LAST_OUT=/tmp/tmp.juVYNpXm7P ++++++ mktemp +++++ local LAST_ERR=/tmp/tmp.b46IZicqHj +++++ local exit_status=0 ++++++ seq 0 2 +++++ for i in '$(seq 0 2)' +++++ set +e +++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ exit_status=0 +++++ set -e +++++ '[' 0 '!=' 0 ']' +++++ break +++++ cat /tmp/tmp.juVYNpXm7P +++++ cat /tmp/tmp.b46IZicqHj +++++ rm /tmp/tmp.juVYNpXm7P /tmp/tmp.b46IZicqHj +++++ return 0 ++++ [[ true == \t\r\u\e ]] ++++ echo some-name-haproxy ++++ return +++ local cluster_proxy=some-name-haproxy +++ echo haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.status.haproxy.ready}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.S1oQhcQfWD +++ mktemp ++ local LAST_ERR=/tmp/tmp.hz7NqTNC6i ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.status.haproxy.ready}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.S1oQhcQfWD ++ cat /tmp/tmp.hz7NqTNC6i ++ rm /tmp/tmp.S1oQhcQfWD /tmp/tmp.hz7NqTNC6i ++ return 0 + [[ 2 == \2 ]] + echo + write_data ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.yyzNtKFPcO ++++ mktemp +++ local LAST_ERR=/tmp/tmp.MjBAUJiItG +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.yyzNtKFPcO +++ cat /tmp/tmp.MjBAUJiItG +++ rm /tmp/tmp.yyzNtKFPcO /tmp/tmp.MjBAUJiItG +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-haproxy ++ return + local proxy=some-name-haproxy ++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.secretsName}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.2fCYiX1cO5 +++ mktemp ++ local LAST_ERR=/tmp/tmp.gxNw9PEOoE ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pxc some-name -o 'jsonpath={.spec.secretsName}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.2fCYiX1cO5 ++ cat /tmp/tmp.gxNw9PEOoE ++ rm /tmp/tmp.2fCYiX1cO5 /tmp/tmp.gxNw9PEOoE ++ return 0 + local secret_name=my-cluster-secrets ++ getSecretData my-cluster-secrets root ++ local secretName=my-cluster-secrets ++ local dataKey=root ++ kubectl_bin get secrets/my-cluster-secrets '--template={{.data.root}}' ++ base64 --decode +++ mktemp ++ local LAST_OUT=/tmp/tmp.JsHyRGiYqG +++ mktemp ++ local LAST_ERR=/tmp/tmp.nEOdo2idy2 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get secrets/my-cluster-secrets '--template={{.data.root}}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.JsHyRGiYqG ++ cat /tmp/tmp.nEOdo2idy2 ++ rm /tmp/tmp.JsHyRGiYqG /tmp/tmp.nEOdo2idy2 ++ return 0 + local 'root_pass=K4IaPwid-C0jNT$wDF' + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' '-h some-name-haproxy -uroot -p'\''K4IaPwid-C0jNT$wDF'\''' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY);' + local 'uri=-h some-name-haproxy -uroot -p'\''K4IaPwid-C0jNT$wDF'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Pv8ZyfW2Ua +++ mktemp ++ local LAST_ERR=/tmp/tmp.4twIdQa6uJ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Pv8ZyfW2Ua ++ cat /tmp/tmp.4twIdQa6uJ ++ rm /tmp/tmp.Pv8ZyfW2Ua /tmp/tmp.4twIdQa6uJ ++ return 0 + client_pod=pxc-client-59944c5bbf-ffzlm + wait_pod pxc-client-59944c5bbf-ffzlm + local pod=pxc-client-59944c5bbf-ffzlm + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-ffzlm ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-ffzlm condition met waiting for pod/pxc-client-59944c5bbf-ffzlm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-haproxy -uroot -p'\''K4IaPwid-C0jNT$wDF'\''' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-haproxy -uroot -p'\''K4IaPwid-C0jNT$wDF'\''' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.IGKoGFCm6D +++ mktemp ++ local LAST_ERR=/tmp/tmp.JYbYpUGfES ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IGKoGFCm6D ++ cat /tmp/tmp.JYbYpUGfES ++ rm /tmp/tmp.IGKoGFCm6D /tmp/tmp.JYbYpUGfES ++ return 0 + client_pod=pxc-client-59944c5bbf-ffzlm + wait_pod pxc-client-59944c5bbf-ffzlm + local pod=pxc-client-59944c5bbf-ffzlm + local max_retry=480 + local ns= ++ echo pxc-client-59944c5bbf-ffzlm ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-59944c5bbf-ffzlm condition met waiting for pod/pxc-client-59944c5bbf-ffzlm to become ReadyDefaulted container "pxc-client" out of: pxc-client, backup .Ok + set +o xtrace + run_backup some-name backup1 + local cluster=some-name + local backup=backup1 + log 'run pxc-backup/backup1' ++ date +%Y-%m-%dT%H:%M:%S%z + echo '[2025-11-12T10:30:21+0000]' run pxc-backup/backup1 [2025-11-12T10:30:21+0000] run pxc-backup/backup1 + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/demand-backup-without-passwords/conf/backup1.yml ++ mktemp + local LAST_OUT=/tmp/tmp.4VWeWJpZE3 ++ mktemp + local LAST_ERR=/tmp/tmp.sP3j862OhS + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-2234/e2e-tests/demand-backup-without-passwords/conf/backup1.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.4VWeWJpZE3 perconaxtradbclusterbackup.pxc.percona.com/backup1 created + cat /tmp/tmp.sP3j862OhS + rm /tmp/tmp.4VWeWJpZE3 /tmp/tmp.sP3j862OhS + return 0 + wait_backup backup1 + local backup=backup1 + local status=Succeeded + set +o xtrace waiting for pxc-backup/backup1 to reach Succeeded state........................................................................................................................................................................................................................................................................................................................................................................2025-11-12T10:25:20.671Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.31.13-gke.1231000"} 2025-11-12T10:25:20.671Z INFO setup Manager starting up {"gitCommit": "269f3694d9306fde7f2ffd7864a3f22b0ef247c5", "gitBranch": "PR-2234-269f3694", "buildTime": "2025-11-12T09:55:34Z", "goVersion": "go1.25.4", "os": "linux", "arch": "amd64"} 2025-11-12T10:25:20.674Z INFO setup Registering Components. 2025-11-12T10:25:21.354Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2025-11-12T10:25:21.354Z INFO setup Starting the Cmd. 2025-11-12T10:25:21.354Z INFO controller-runtime.metrics Starting metrics server 2025-11-12T10:25:21.354Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2025-11-12T10:25:21.354Z INFO starting server {"name": "health probe", "addr": "[::]:8081"} 2025-11-12T10:25:21.354Z INFO controller-runtime.webhook Starting webhook server 2025-11-12T10:25:21.354Z INFO controller-runtime.certwatcher Updated current TLS certificate {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key"} 2025-11-12T10:25:21.355Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2025-11-12T10:25:21.355Z INFO controller-runtime.certwatcher Starting certificate poll+watcher {"cert": "/tmp/k8s-webhook-server/serving-certs/tls.crt", "key": "/tmp/k8s-webhook-server/serving-certs/tls.key", "interval": "10s"} 2025-11-12T10:25:21.455Z INFO attempting to acquire leader lease pxc-operator/08db1feb.percona.com... 2025-11-12T10:25:21.493Z INFO successfully acquired lease pxc-operator/08db1feb.percona.com 2025-11-12T10:25:21.493Z DEBUG events percona-xtradb-cluster-operator-b5f9c4897-hns8f_e24d0d96-5adc-4a76-a811-490ec9237989 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db1feb.percona.com","uid":"504a834c-aca5-459d-9469-9db8082280aa","apiVersion":"coordination.k8s.io/v1","resourceVersion":"1762943121486607009"}, "reason": "LeaderElection"} 2025-11-12T10:25:21.493Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.Secret"} 2025-11-12T10:25:21.493Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.PerconaXtraDBCluster"} 2025-11-12T10:25:21.493Z INFO Starting EventSource {"controller": "pxcrestore-controller", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2025-11-12T10:25:21.494Z INFO Starting EventSource {"controller": "pxcbackup-controller", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2025-11-12T10:25:21.594Z INFO Starting Controller {"controller": "pxc-controller"} 2025-11-12T10:25:21.594Z INFO Starting workers {"controller": "pxc-controller", "worker count": 1} 2025-11-12T10:25:21.594Z INFO Starting Controller {"controller": "pxcrestore-controller"} 2025-11-12T10:25:21.594Z INFO Starting workers {"controller": "pxcrestore-controller", "worker count": 1} 2025-11-12T10:25:21.594Z INFO Starting Controller {"controller": "pxcbackup-controller"} 2025-11-12T10:25:21.594Z INFO Starting workers {"controller": "pxcbackup-controller", "worker count": 1} 2025-11-12T10:26:13.641Z INFO Set CR version {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "version": "1.19.0"} 2025-11-12T10:26:13.772Z INFO Created user secrets {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "secrets": "my-cluster-secrets"} 2025-11-12T10:26:16.081Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "object": "auto-some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-11-12T10:26:16.099Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:ConfigMap,APIVersion:v1,}"} 2025-11-12T10:26:16.218Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-12T10:26:16.260Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:StatefulSet,APIVersion:apps/v1,}"} 2025-11-12T10:26:16.293Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-12T10:26:16.330Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "object": "some-name-pxc-unready", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-12T10:26:16.431Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-12T10:26:16.539Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "0d9e1777-f7b9-4201-8ea3-e2bd7e0ff5b6", "object": "some-name-haproxy-replicas", "kind": "&TypeMeta{Kind:Service,APIVersion:v1,}"} 2025-11-12T10:26:17.488Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "3f7227d2-afbf-4dcf-b34d-fe98995e0b9f", "object": "some-name-pxc", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-12T10:26:17.528Z DEBUG Creating object {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "3f7227d2-afbf-4dcf-b34d-fe98995e0b9f", "object": "some-name-haproxy", "kind": "&TypeMeta{Kind:PodDisruptionBudget,APIVersion:policy/v1,}"} 2025-11-12T10:27:34.096Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "bb8271ed-68a1-4d9f-af27-0ad96ba05702", "user": "operator"} 2025-11-12T10:27:34.139Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "bb8271ed-68a1-4d9f-af27-0ad96ba05702", "user": "monitor"} 2025-11-12T10:27:34.193Z INFO User monitor: granted privileges {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "bb8271ed-68a1-4d9f-af27-0ad96ba05702"} 2025-11-12T10:27:34.242Z INFO monitor user privileges granted {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "bb8271ed-68a1-4d9f-af27-0ad96ba05702"} 2025-11-12T10:27:34.274Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "bb8271ed-68a1-4d9f-af27-0ad96ba05702", "user": "xtrabackup"} 2025-11-12T10:27:34.313Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "bb8271ed-68a1-4d9f-af27-0ad96ba05702"} 2025-11-12T10:27:34.343Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "bb8271ed-68a1-4d9f-af27-0ad96ba05702", "user": "replication"} 2025-11-12T10:29:54.651Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "ad13868f-45df-44be-a3a4-9d926f926bf8", "user": "root"} 2025-11-12T10:29:54.748Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "namespace": "demand-backup-without-passwords-25983", "name": "some-name", "reconcileID": "ad13868f-45df-44be-a3a4-9d926f926bf8", "new version": "8.0.43-34.1"} 2025-11-12T10:30:23.982Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "namespace": "demand-backup-without-passwords-25983", "name": "backup1", "reconcileID": "e250fa40-bda6-478c-801b-8f0b250cff9c", "cluster": "some-name", "storage": "pvc", "allowed": false} 2025-11-12T10:30:24.100Z INFO Creating a new volume for backup {"controller": "pxcbackup-controller", "namespace": "demand-backup-without-passwords-25983", "name": "backup1", "reconcileID": "e250fa40-bda6-478c-801b-8f0b250cff9c", "Namespace": "demand-backup-without-passwords-25983", "Name": "xb-backup1"} 2025-11-12T10:30:24.182Z INFO Created a new backup job {"controller": "pxcbackup-controller", "namespace": "demand-backup-without-passwords-25983", "name": "backup1", "reconcileID": "e250fa40-bda6-478c-801b-8f0b250cff9c", "namespace": "demand-backup-without-passwords-25983", "name": "xb-backup1"} 2025-11-12T10:30:24.182Z DEBUG Check if parallel backups are allowed {"controller": "pxcbackup-controller", "namespace": "demand-backup-without-passwords-25983", "name": "backup1", "reconcileID": "376a9cf4-8894-4a98-b811-4133095a2221", "cluster": "some-name", "storage": "pvc", "allowed": false} 2025-11-12T10:30:24.281Z INFO Releasing backup lock {"controller": "pxcbackup-controller", "namespace": "demand-backup-without-passwords-25983", "name": "backup1", "reconcileID": "376a9cf4-8894-4a98-b811-4133095a2221", "cluster": "some-name", "storage": "pvc", "lease": "pxc-some-name-backup-lock"} 2025-11-12T10:30:24.312Z ERROR Reconciler error {"controller": "pxcbackup-controller", "namespace": "demand-backup-without-passwords-25983", "name": "backup1", "reconcileID": "376a9cf4-8894-4a98-b811-4133095a2221", "error": "create backup job: create backup job: jobs.batch \"xb-backup1\" already exists", "errorVerbose": "jobs.batch \"xb-backup1\" already exists\ncreate backup job\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup.(*ReconcilePerconaXtraDBClusterBackup).createBackupJob\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup/controller.go:383\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup.(*ReconcilePerconaXtraDBClusterBackup).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup/controller.go:291\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693\ncreate backup job\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup.(*ReconcilePerconaXtraDBClusterBackup).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup/controller.go:293\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:216\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:461\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1693"} sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).reconcileHandler /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:474 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).processNextWorkItem /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:421 sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller[...]).Start.func1.1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.22.4/pkg/internal/controller/controller.go:296 2025-11-12T10:34:24.473Z INFO controller-runtime.cache Warning: watch ended with error {"reflector": "pkg/mod/k8s.io/client-go@v0.34.1/tools/cache/reflector.go:290", "type": "*v1.Event", "err": "very short watch: pkg/mod/k8s.io/client-go@v0.34.1/tools/cache/reflector.go:290: Unexpected watch close - watch lasted less than a second and no items received"} max retry count 360 reached. something went wrong with operator or kubernetes cluster