Log: /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/logs/restore-to-encrypted-cluster-8-0.log WARNING: version difference between client (1.29) and server (1.25) exceeds the supported minor version skew of +/-1 WARNING: version difference between client (1.29) and server (1.25) exceeds the supported minor version skew of +/-1 + main + create_infra restore-to-encrypted-cluster-4709 + local ns=restore-to-encrypted-cluster-4709 + '[' -n pxc-operator ']' + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n restore-to-encrypted-cluster-9707 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.1A4fMbrFwj ++ mktemp + local LAST_ERR=/tmp/tmp.rkpFaDcHWj + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.1A4fMbrFwj perconaxtradbcluster.pxc.percona.com "some-name" deleted + cat /tmp/tmp.rkpFaDcHWj + rm /tmp/tmp.1A4fMbrFwj /tmp/tmp.rkpFaDcHWj + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.HCogXZcui7 ++ mktemp + local LAST_ERR=/tmp/tmp.9Ol1rVRk8X + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HCogXZcui7 perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-aws-s3" deleted perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-pvc" deleted + cat /tmp/tmp.9Ol1rVRk8X + rm /tmp/tmp.HCogXZcui7 /tmp/tmp.9Ol1rVRk8X + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.o9BzCXq4XG ++ mktemp + local LAST_ERR=/tmp/tmp.0NVkfrz77Q + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.o9BzCXq4XG perconaxtradbclusterrestore.pxc.percona.com "on-demand-backup-pvc" deleted + cat /tmp/tmp.0NVkfrz77Q + rm /tmp/tmp.o9BzCXq4XG /tmp/tmp.0NVkfrz77Q + return 0 + create_namespace pxc-operator + local namespace=pxc-operator + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ sed s/NAMESPACE// ++ tail -n1 ++ awk '-F ' '{print $2}' ++ helm list --all-namespaces --filter chaos-mesh + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' + '[' -n '' ']' ++ kubectl get crd ++ grep chaos-mesh.org ++ awk '{print $1}' + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ awk '{print $1}' ++ kubectl get clusterrole ++ grep chaos-mesh + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + kubectl_bin get ns + '[' -n '' ']' + desc 'cleaned up old namespaces pxc-operator' + xargs kubectl delete ns + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin delete namespace pxc-operator + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' ++ mktemp + local LAST_OUT=/tmp/tmp.0ntfx65bUB ++ mktemp + local LAST_ERR=/tmp/tmp.3QawBOqfSH + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.caGLsjSFc6 + awk '{print$1}' ++ mktemp + local LAST_ERR=/tmp/tmp.DqzFDqRKEb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl get ns + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.caGLsjSFc6 + cat /tmp/tmp.DqzFDqRKEb + rm /tmp/tmp.caGLsjSFc6 /tmp/tmp.DqzFDqRKEb + return 0 namespace "restore-to-encrypted-cluster-9707" deleted namespace "vault-service-1-20359" deleted Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0ntfx65bUB namespace "pxc-operator" deleted + cat /tmp/tmp.3QawBOqfSH + rm /tmp/tmp.0ntfx65bUB /tmp/tmp.3QawBOqfSH + return 0 + wait_for_delete namespace/pxc-operator + local res=namespace/pxc-operator + echo -n 'namespace/pxc-operator - ' namespace/pxc-operator - + set +o xtrace Error from server (NotFound): namespaces "pxc-operator" not found + desc 'create namespace pxc-operator' + set +o xtrace ----------------------------------------------------------------------------------- create namespace pxc-operator ----------------------------------------------------------------------------------- + kubectl_bin create namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.qDtQuKbr9T ++ mktemp + local LAST_ERR=/tmp/tmp.1eG89PPa8J + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.qDtQuKbr9T namespace/pxc-operator created + cat /tmp/tmp.1eG89PPa8J + rm /tmp/tmp.qDtQuKbr9T /tmp/tmp.1eG89PPa8J + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.BMtNnats4J +++ mktemp ++ local LAST_ERR=/tmp/tmp.GgmYS1HqHp ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BMtNnats4J ++ cat /tmp/tmp.GgmYS1HqHp ++ rm /tmp/tmp.BMtNnats4J /tmp/tmp.GgmYS1HqHp ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9 --namespace=pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.RYOvJs5M2B ++ mktemp + local LAST_ERR=/tmp/tmp.tMB2QkkfQd + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9 --namespace=pxc-operator + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.RYOvJs5M2B Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9" modified. + cat /tmp/tmp.tMB2QkkfQd + rm /tmp/tmp.RYOvJs5M2B /tmp/tmp.tMB2QkkfQd + return 0 + deploy_operator + desc 'start PXC operator' + set +o xtrace ----------------------------------------------------------------------------------- start PXC operator ----------------------------------------------------------------------------------- + kubectl_bin apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/deploy/crd.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.0LLtoTCSme ++ mktemp + local LAST_ERR=/tmp/tmp.J2sFZpHnDQ + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --server-side --force-conflicts -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/deploy/crd.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0LLtoTCSme customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterbackups.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusterrestores.pxc.percona.com serverside-applied customresourcedefinition.apiextensions.k8s.io/perconaxtradbclusters.pxc.percona.com serverside-applied + cat /tmp/tmp.J2sFZpHnDQ + rm /tmp/tmp.0LLtoTCSme /tmp/tmp.J2sFZpHnDQ + return 0 + '[' -n pxc-operator ']' + apply_rbac cw-rbac + local operator_namespace=pxc-operator + local rbac=cw-rbac + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/deploy/cw-rbac.yaml + sed -e 's^namespace: .*^namespace: pxc-operator^' + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.FKZ5oMIk7R ++ mktemp + local LAST_ERR=/tmp/tmp.ZiHCiS3wcq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FKZ5oMIk7R clusterrole.rbac.authorization.k8s.io/percona-xtradb-cluster-operator unchanged serviceaccount/percona-xtradb-cluster-operator created clusterrolebinding.rbac.authorization.k8s.io/service-account-percona-xtradb-cluster-operator unchanged + cat /tmp/tmp.ZiHCiS3wcq + rm /tmp/tmp.FKZ5oMIk7R /tmp/tmp.ZiHCiS3wcq + return 0 + sed -e 's^image: .*^image: perconalab/percona-xtradb-cluster-operator:PR-1522-1ed4605b^' + kubectl_bin apply -f - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "DISABLE_TELEMETRY").value) = "true"' - + yq eval '(select(.kind == "Deployment").spec.template.spec.containers[] | select(.name == "percona-xtradb-cluster-operator").env[] | select(.name == "LOG_LEVEL").value) = "DEBUG"' - + sed -e 's^failureThreshold: .*^failureThreshold: 10^' ++ mktemp + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/deploy/cw-operator.yaml + local LAST_OUT=/tmp/tmp.HLEgFrm1bS ++ mktemp + local LAST_ERR=/tmp/tmp.f1xzHRWfpR + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.HLEgFrm1bS deployment.apps/percona-xtradb-cluster-operator created service/percona-xtradb-cluster-operator created + cat /tmp/tmp.f1xzHRWfpR + rm /tmp/tmp.HLEgFrm1bS /tmp/tmp.f1xzHRWfpR + return 0 + sleep 10 + kubectl_bin wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s ++ mktemp + local LAST_OUT=/tmp/tmp.0bZSpyrI56 ++ mktemp + local LAST_ERR=/tmp/tmp.OvZkgiJvqU + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pods -l app.kubernetes.io/component=operator,app.kubernetes.io/instance=percona-xtradb-cluster-operator,app.kubernetes.io/name=percona-xtradb-cluster-operator --timeout=30s + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0bZSpyrI56 pod/percona-xtradb-cluster-operator-6dbd475d79-7244c condition met + cat /tmp/tmp.OvZkgiJvqU + rm /tmp/tmp.0bZSpyrI56 /tmp/tmp.OvZkgiJvqU + return 0 ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.WxlwwvN9aa +++ mktemp ++ local LAST_ERR=/tmp/tmp.gzCBxPtwZt ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.WxlwwvN9aa ++ cat /tmp/tmp.gzCBxPtwZt ++ rm /tmp/tmp.WxlwwvN9aa /tmp/tmp.gzCBxPtwZt ++ return 0 + wait_pod percona-xtradb-cluster-operator-6dbd475d79-7244c 480 pxc-operator + local pod=percona-xtradb-cluster-operator-6dbd475d79-7244c + local max_retry=480 + local ns=pxc-operator ++ echo percona-xtradb-cluster-operator-6dbd475d79-7244c ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/percona-xtradb-cluster-operator-6dbd475d79-7244c condition met percona-xtradb-cluster-operator-6dbd475d79-7244c.Ok + sleep 3 + create_namespace restore-to-encrypted-cluster-4709 + local namespace=restore-to-encrypted-cluster-4709 + local skip_clean_namespace= + [[ 1 == 1 ]] + [[ -z '' ]] + destroy_chaos_mesh ++ helm list --all-namespaces --filter chaos-mesh ++ tail -n1 ++ awk '-F ' '{print $2}' ++ sed s/NAMESPACE// + local chaos_mesh_ns= + desc 'destroy chaos-mesh' + set +o xtrace ----------------------------------------------------------------------------------- destroy chaos-mesh ----------------------------------------------------------------------------------- ++ kubectl api-resources ++ grep chaos-mesh ++ awk '{print $1}' + '[' -n '' ']' ++ kubectl get crd ++ awk '{print $1}' ++ grep chaos-mesh.org + timeout 30 kubectl delete crd error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrolebinding ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrolebinding error: resource(s) were provided, but no name was specified + : ++ kubectl get clusterrole ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete clusterrole error: resource(s) were provided, but no name was specified + : ++ kubectl get MutatingWebhookConfiguration ++ grep chaos-mesh ++ awk '{print $1}' + timeout 30 kubectl delete MutatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ awk '{print $1}' ++ grep chaos-mesh + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : ++ kubectl get ValidatingWebhookConfiguration ++ grep validate-auth ++ awk '{print $1}' + timeout 30 kubectl delete ValidatingWebhookConfiguration error: resource(s) were provided, but no name was specified + : + desc 'cleaned up all old namespaces' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up all old namespaces ----------------------------------------------------------------------------------- + egrep -v '^kube-|^default$|Terminating|pxc-operator|openshift|^NAME' + awk '{print$1}' + '[' -n '' ']' + desc 'cleaned up old namespaces restore-to-encrypted-cluster-4709' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces restore-to-encrypted-cluster-4709 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace restore-to-encrypted-cluster-4709 + kubectl_bin get ns ++ mktemp ++ mktemp + local LAST_OUT=/tmp/tmp.2DXDth5Txp + local LAST_OUT=/tmp/tmp.8oNpbYzXx7 ++ mktemp ++ mktemp + xargs kubectl delete ns + local LAST_ERR=/tmp/tmp.DGoBqKZZ5v + local exit_status=0 + local LAST_ERR=/tmp/tmp.Wwz4sevPHB + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + for i in '$(seq 0 2)' + set +e + set +e + kubectl get ns + kubectl delete namespace restore-to-encrypted-cluster-4709 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace restore-to-encrypted-cluster-4709 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.2DXDth5Txp + cat /tmp/tmp.DGoBqKZZ5v + rm /tmp/tmp.2DXDth5Txp /tmp/tmp.DGoBqKZZ5v + return 0 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace restore-to-encrypted-cluster-4709 Error from server (Forbidden): namespaces "default" is forbidden: this namespace may not be deleted + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.8oNpbYzXx7 + cat /tmp/tmp.Wwz4sevPHB Error from server (NotFound): namespaces "restore-to-encrypted-cluster-4709" not found + rm /tmp/tmp.8oNpbYzXx7 /tmp/tmp.Wwz4sevPHB + return 1 + : + wait_for_delete namespace/restore-to-encrypted-cluster-4709 + local res=namespace/restore-to-encrypted-cluster-4709 + echo -n 'namespace/restore-to-encrypted-cluster-4709 - ' namespace/restore-to-encrypted-cluster-4709 - + set +o xtrace Error from server (NotFound): namespaces "restore-to-encrypted-cluster-4709" not found + desc 'create namespace restore-to-encrypted-cluster-4709' + set +o xtrace ----------------------------------------------------------------------------------- create namespace restore-to-encrypted-cluster-4709 ----------------------------------------------------------------------------------- + kubectl_bin create namespace restore-to-encrypted-cluster-4709 ++ mktemp + local LAST_OUT=/tmp/tmp.YHaW99Rv1y ++ mktemp + local LAST_ERR=/tmp/tmp.6iTI3rC40O + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace restore-to-encrypted-cluster-4709 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.YHaW99Rv1y namespace/restore-to-encrypted-cluster-4709 created + cat /tmp/tmp.6iTI3rC40O + rm /tmp/tmp.YHaW99Rv1y /tmp/tmp.6iTI3rC40O + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.IShxw9Ci7e +++ mktemp ++ local LAST_ERR=/tmp/tmp.sJdziwyBiV ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.IShxw9Ci7e ++ cat /tmp/tmp.sJdziwyBiV ++ rm /tmp/tmp.IShxw9Ci7e /tmp/tmp.sJdziwyBiV ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9 --namespace=restore-to-encrypted-cluster-4709 ++ mktemp + local LAST_OUT=/tmp/tmp.cH664V2TN6 ++ mktemp + local LAST_ERR=/tmp/tmp.8HEvvJgX6q + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9 --namespace=restore-to-encrypted-cluster-4709 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.cH664V2TN6 Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9" modified. + cat /tmp/tmp.8HEvvJgX6q + rm /tmp/tmp.cH664V2TN6 /tmp/tmp.8HEvvJgX6q + return 0 + apply_secrets + desc 'create secrets for cloud storages' + set +o xtrace ----------------------------------------------------------------------------------- create secrets for cloud storages ----------------------------------------------------------------------------------- + '[' -z '' ']' + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/cloud-secret.yml ++ mktemp + local LAST_OUT=/tmp/tmp.vTvZq9CE6t ++ mktemp + local LAST_ERR=/tmp/tmp.E4uFlyqtUs + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/minio-secret.yml -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/cloud-secret.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vTvZq9CE6t secret/minio-secret created secret/aws-s3-secret created secret/gcp-cs-secret created secret/azure-secret created + cat /tmp/tmp.E4uFlyqtUs + rm /tmp/tmp.vTvZq9CE6t /tmp/tmp.E4uFlyqtUs + return 0 + cluster=some-name + spinup_pxc some-name /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/some-name.yml + local cluster=some-name + local config=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/some-name.yml + local size=3 + local sleep=10 + local secretsFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/secrets.yml + local pxcClientFile=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/client.yml + local port=3306 + desc 'create first PXC cluster' + set +o xtrace ----------------------------------------------------------------------------------- create first PXC cluster ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/secrets.yml ++ mktemp + local LAST_OUT=/tmp/tmp.lTclR9ZFe3 ++ mktemp + local LAST_ERR=/tmp/tmp.lpRlONVror + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/secrets.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.lTclR9ZFe3 secret/my-cluster-secrets created secret/some-name-ssl created secret/some-name-ssl-internal created + cat /tmp/tmp.lpRlONVror + rm /tmp/tmp.lTclR9ZFe3 /tmp/tmp.lpRlONVror + return 0 + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/client.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/client.yml + kubectl_bin apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/client.yml + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' ++ mktemp + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + local LAST_OUT=/tmp/tmp.TOetB1Yri8 + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1522-1ed4605b#' + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' ++ mktemp + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.restore-to-encrypted-cluster-4709~ + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:k8spxc-1313-8.0-backup#' + local LAST_ERR=/tmp/tmp.hxCBjm7nBX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.TOetB1Yri8 deployment.apps/pxc-client created + cat /tmp/tmp.hxCBjm7nBX + rm /tmp/tmp.TOetB1Yri8 /tmp/tmp.hxCBjm7nBX + return 0 + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/some-name.yml + '[' -z '' ']' + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/some-name.yml + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:k8spxc-1313-8.0-backup#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + kubectl_bin apply -f - ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + local LAST_OUT=/tmp/tmp.FbYex9IO7T ++ mktemp + local LAST_ERR=/tmp/tmp.aHdhJk0znk + local exit_status=0 ++ seq 0 2 + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/some-name.yml + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.restore-to-encrypted-cluster-4709~ + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1522-1ed4605b#' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.FbYex9IO7T perconaxtradbcluster.pxc.percona.com/some-name created + cat /tmp/tmp.aHdhJk0znk + rm /tmp/tmp.FbYex9IO7T /tmp/tmp.aHdhJk0znk + return 0 + desc 'check if all 3 Pods started' + set +o xtrace ----------------------------------------------------------------------------------- check if all 3 Pods started ----------------------------------------------------------------------------------- ++ get_proxy some-name ++ local target_cluster=some-name +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.H2URwYnBcb ++++ mktemp +++ local LAST_ERR=/tmp/tmp.TLGcZyRk2H +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.H2URwYnBcb +++ cat /tmp/tmp.TLGcZyRk2H +++ rm /tmp/tmp.H2URwYnBcb /tmp/tmp.TLGcZyRk2H +++ return 0 ++ [[ '' == \t\r\u\e ]] +++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ mktemp +++ local LAST_OUT=/tmp/tmp.CvNZWfk9fR ++++ mktemp +++ local LAST_ERR=/tmp/tmp.cO5grWXIJ6 +++ local exit_status=0 ++++ seq 0 2 +++ for i in '$(seq 0 2)' +++ set +e +++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++ exit_status=0 +++ set -e +++ '[' 0 '!=' 0 ']' +++ break +++ cat /tmp/tmp.CvNZWfk9fR +++ cat /tmp/tmp.cO5grWXIJ6 +++ rm /tmp/tmp.CvNZWfk9fR /tmp/tmp.cO5grWXIJ6 +++ return 0 ++ [[ true == \t\r\u\e ]] ++ echo some-name-proxysql ++ return + local proxy=some-name-proxysql + kubectl_bin wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n restore-to-encrypted-cluster-4709 ++ mktemp + local LAST_OUT=/tmp/tmp.1AqIPp86Pm ++ mktemp + local LAST_ERR=/tmp/tmp.5wJzmO2iMh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n restore-to-encrypted-cluster-4709 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n restore-to-encrypted-cluster-4709 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl wait --for=condition=Ready pod -l app.kubernetes.io/instance=monitoring,app.kubernetes.io/managed-by=percona-xtradb-cluster-operator --timeout=300s -n restore-to-encrypted-cluster-4709 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.1AqIPp86Pm + cat /tmp/tmp.5wJzmO2iMh error: no matching resources found + rm /tmp/tmp.1AqIPp86Pm /tmp/tmp.5wJzmO2iMh + return 1 + true + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met some-name-proxysql-0.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met some-name-pxc-0.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met some-name-pxc-1.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met some-name-pxc-2.Ok + sleep 10 + desc 'write data' + set +o xtrace ----------------------------------------------------------------------------------- write data ----------------------------------------------------------------------------------- + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 5\.7 ]] + run_mysql 'CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' '-h some-name-proxysql -uroot -proot_password -P3306' + local 'command=CREATE DATABASE IF NOT EXISTS myApp; use myApp; CREATE TABLE IF NOT EXISTS myApp (id int PRIMARY KEY) ;' + local 'uri=-h some-name-proxysql -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9DQ4MmP9eS +++ mktemp ++ local LAST_ERR=/tmp/tmp.EoXTksAZZ5 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9DQ4MmP9eS ++ cat /tmp/tmp.EoXTksAZZ5 ++ rm /tmp/tmp.9DQ4MmP9eS /tmp/tmp.EoXTksAZZ5 ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + run_mysql 'INSERT myApp.myApp (id) VALUES (100500)' '-h some-name-proxysql -uroot -proot_password -P3306' + local 'command=INSERT myApp.myApp (id) VALUES (100500)' + local 'uri=-h some-name-proxysql -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Idru4k9XIY +++ mktemp ++ local LAST_ERR=/tmp/tmp.Bs1Q7qhle9 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Idru4k9XIY ++ cat /tmp/tmp.Bs1Q7qhle9 ++ rm /tmp/tmp.Idru4k9XIY /tmp/tmp.Bs1Q7qhle9 ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + sleep 30 ++ seq 0 2 + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XU0W9N6f74 +++ mktemp ++ local LAST_ERR=/tmp/tmp.e0yOK9bCTS ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XU0W9N6f74 ++ cat /tmp/tmp.e0yOK9bCTS ++ rm /tmp/tmp.XU0W9N6f74 /tmp/tmp.e0yOK9bCTS ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.LW7qwu72v2/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.LW7qwu72v2/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.Y27Ge3AMmo +++ mktemp ++ local LAST_ERR=/tmp/tmp.34i1NrrYyz ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Y27Ge3AMmo ++ cat /tmp/tmp.34i1NrrYyz ++ rm /tmp/tmp.Y27Ge3AMmo /tmp/tmp.34i1NrrYyz ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.LW7qwu72v2/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.LW7qwu72v2/select-1.sql + for i in '$(seq 0 $((size - 1)))' + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password -P3306' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.s6ySPJJ61q +++ mktemp ++ local LAST_ERR=/tmp/tmp.UZl6eASKM7 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.s6ySPJJ61q ++ cat /tmp/tmp.UZl6eASKM7 ++ rm /tmp/tmp.s6ySPJJ61q /tmp/tmp.UZl6eASKM7 ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.LW7qwu72v2/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.LW7qwu72v2/select-1.sql ++ is_keyring_plugin_in_use some-name ++ local cluster=some-name ++ kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ egrep -o 'early-plugin-load=keyring_\w+.so' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dnBDnPoJz2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.feiWpaFJVW ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dnBDnPoJz2 ++ cat /tmp/tmp.feiWpaFJVW Unable to use a TTY - input is not a terminal or the right kind of file ++ rm /tmp/tmp.dnBDnPoJz2 /tmp/tmp.feiWpaFJVW ++ return 0 + '[' '' ']' + keyring_plugin_must_not_be_in_use some-name + local cluster=some-name + is_keyring_plugin_in_use some-name + local cluster=some-name + kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ mktemp + local LAST_OUT=/tmp/tmp.e0XvN1CBIR + egrep -o 'early-plugin-load=keyring_\w+.so' ++ mktemp + local LAST_ERR=/tmp/tmp.mmVY7wi4KH + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.e0XvN1CBIR + cat /tmp/tmp.mmVY7wi4KH Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.e0XvN1CBIR /tmp/tmp.mmVY7wi4KH + return 0 + table_must_not_be_encrypted some-name myApp + local cluster=some-name + local table=myApp + is_table_encrypted some-name myApp + local cluster=some-name + local table=myApp + egrep -o 'ENCRYPTION=('\''Y'\''|"Y")' + run_mysql 'SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' '-h some-name-proxysql -uroot -proot_password' + local 'command=SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.BJ7na0qLno +++ mktemp ++ local LAST_ERR=/tmp/tmp.cy8MH1Uw7I ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.BJ7na0qLno ++ cat /tmp/tmp.cy8MH1Uw7I ++ rm /tmp/tmp.BJ7na0qLno /tmp/tmp.cy8MH1Uw7I ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-599db5b9bc-hllss ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + run_backup some-name on-demand-backup-pvc + local cluster=some-name + local backup1=on-demand-backup-pvc + desc 'make backup on-demand-backup-pvc' + set +o xtrace ----------------------------------------------------------------------------------- make backup on-demand-backup-pvc ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/on-demand-backup-pvc.yml ++ mktemp + local LAST_OUT=/tmp/tmp.kimSzLPzil ++ mktemp + local LAST_ERR=/tmp/tmp.jrJVzqOUuh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/on-demand-backup-pvc.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.kimSzLPzil perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.jrJVzqOUuh + rm /tmp/tmp.kimSzLPzil /tmp/tmp.jrJVzqOUuh + return 0 + wait_backup on-demand-backup-pvc + local backup=on-demand-backup-pvc + local status=Succeeded + set +o xtrace on-demand-backup-pvc.................Succeeded + '[' -z '' ']' + run_backup some-name on-demand-backup-aws-s3 + local cluster=some-name + local backup1=on-demand-backup-aws-s3 + desc 'make backup on-demand-backup-aws-s3' + set +o xtrace ----------------------------------------------------------------------------------- make backup on-demand-backup-aws-s3 ----------------------------------------------------------------------------------- + kubectl_bin apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/on-demand-backup-aws-s3.yml ++ mktemp + local LAST_OUT=/tmp/tmp.hKHAHPQPU2 ++ mktemp + local LAST_ERR=/tmp/tmp.ErkqySbJRK + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/on-demand-backup-aws-s3.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hKHAHPQPU2 perconaxtradbclusterbackup.pxc.percona.com/on-demand-backup-aws-s3 created + cat /tmp/tmp.ErkqySbJRK + rm /tmp/tmp.hKHAHPQPU2 /tmp/tmp.ErkqySbJRK + return 0 + wait_backup on-demand-backup-aws-s3 + local backup=on-demand-backup-aws-s3 + local status=Succeeded + set +o xtrace on-demand-backup-aws-s3................Succeeded + vault1=vault-service-1-7243 + start_vault vault-service-1-7243 + name=vault-service-1-7243 + protocol=http + local platform=kubernetes + [[ -n '' ]] + create_namespace vault-service-1-7243 skip_clean + local namespace=vault-service-1-7243 + local skip_clean_namespace=skip_clean + [[ 1 == 1 ]] + [[ -z skip_clean ]] + '[' -n '' ']' + desc 'cleaned up old namespaces vault-service-1-7243' + set +o xtrace ----------------------------------------------------------------------------------- cleaned up old namespaces vault-service-1-7243 ----------------------------------------------------------------------------------- + kubectl_bin delete namespace vault-service-1-7243 ++ mktemp + local LAST_OUT=/tmp/tmp.PEkyDNWa8P ++ mktemp + local LAST_ERR=/tmp/tmp.8Eo9zgDGKb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace vault-service-1-7243 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace vault-service-1-7243 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + for i in '$(seq 0 2)' + set +e + kubectl delete namespace vault-service-1-7243 + exit_status=1 + set -e + '[' 1 '!=' 0 ']' + '[' 1 == 1 ']' + sleep 0 + cat /tmp/tmp.PEkyDNWa8P + cat /tmp/tmp.8Eo9zgDGKb Error from server (NotFound): namespaces "vault-service-1-7243" not found + rm /tmp/tmp.PEkyDNWa8P /tmp/tmp.8Eo9zgDGKb + return 1 + : + wait_for_delete namespace/vault-service-1-7243 + local res=namespace/vault-service-1-7243 + echo -n 'namespace/vault-service-1-7243 - ' namespace/vault-service-1-7243 - + set +o xtrace Error from server (NotFound): namespaces "vault-service-1-7243" not found + desc 'create namespace vault-service-1-7243' + set +o xtrace ----------------------------------------------------------------------------------- create namespace vault-service-1-7243 ----------------------------------------------------------------------------------- + kubectl_bin create namespace vault-service-1-7243 ++ mktemp + local LAST_OUT=/tmp/tmp.zTYabxOgUZ ++ mktemp + local LAST_ERR=/tmp/tmp.b0rc8NdPI1 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl create namespace vault-service-1-7243 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.zTYabxOgUZ namespace/vault-service-1-7243 created + cat /tmp/tmp.b0rc8NdPI1 + rm /tmp/tmp.zTYabxOgUZ /tmp/tmp.b0rc8NdPI1 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.FkyOAwotks +++ mktemp ++ local LAST_ERR=/tmp/tmp.cocECYFO1v ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.FkyOAwotks ++ cat /tmp/tmp.cocECYFO1v ++ rm /tmp/tmp.FkyOAwotks /tmp/tmp.cocECYFO1v ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9 --namespace=vault-service-1-7243 ++ mktemp + local LAST_OUT=/tmp/tmp.S3v7lQpVXx ++ mktemp + local LAST_ERR=/tmp/tmp.VET8c4INOV + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9 --namespace=vault-service-1-7243 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.S3v7lQpVXx Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9" modified. + cat /tmp/tmp.VET8c4INOV + rm /tmp/tmp.S3v7lQpVXx /tmp/tmp.VET8c4INOV + return 0 + deploy_helm vault-service-1-7243 + helm repo add hashicorp https://helm.releases.hashicorp.com "hashicorp" already exists with the same configuration, skipping + helm repo add minio https://charts.min.io/ "minio" already exists with the same configuration, skipping + helm repo update Hang tight while we grab the latest from your chart repositories... ...Successfully got an update from the "hashicorp" chart repository ...Successfully got an update from the "chaos-mesh" chart repository ...Successfully got an update from the "minio" chart repository ...Successfully got an update from the "percona" chart repository Update Complete. ⎈Happy Helming!⎈ + helm uninstall vault-service-1-7243 Error: uninstall: Release not loaded: vault-service-1-7243: release: not found + : + desc 'install Vault vault-service-1-7243' + set +o xtrace ----------------------------------------------------------------------------------- install Vault vault-service-1-7243 ----------------------------------------------------------------------------------- + '[' http == https ']' + helm install vault-service-1-7243 hashicorp/vault --disable-openapi-validation --version 0.16.1 --namespace vault-service-1-7243 --set dataStorage.enabled=false --set global.platform=kubernetes NAME: vault-service-1-7243 LAST DEPLOYED: Fri Dec 29 11:39:48 2023 NAMESPACE: vault-service-1-7243 STATUS: deployed REVISION: 1 NOTES: Thank you for installing HashiCorp Vault! Now that you have deployed Vault, you should look over the docs on using Vault with Kubernetes available here: https://www.vaultproject.io/docs/ Your release is named vault-service-1-7243. To learn more about the release, try: $ helm status vault-service-1-7243 $ helm get manifest vault-service-1-7243 + [[ -n '' ]] + set +o xtrace pod/vault-service-1-7243-0.....{"running":{"startedAt":"2023-12-29T11:40:05Z"}} + kubectl_bin exec -it vault-service-1-7243-0 -- vault operator init -tls-skip-verify -key-shares=1 -key-threshold=1 -format=json ++ mktemp + local LAST_OUT=/tmp/tmp.vJ02QhiNAP ++ mktemp + local LAST_ERR=/tmp/tmp.Y8V2h0hiDb + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec -it vault-service-1-7243-0 -- vault operator init -tls-skip-verify -key-shares=1 -key-threshold=1 -format=json + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.vJ02QhiNAP + cat /tmp/tmp.Y8V2h0hiDb Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.vJ02QhiNAP /tmp/tmp.Y8V2h0hiDb + return 0 ++ jq -r '.unseal_keys_b64[]' + unsealKey=iZNb7C9P1+BkChWmQwVdgyR74OaKRadvdzZy7I62ZVY= ++ jq -r .root_token + token=s.Pon35VyeiFRtISlCTu6SC9fc + sleep 10 + kubectl_bin exec -it vault-service-1-7243-0 -- vault operator unseal -tls-skip-verify iZNb7C9P1+BkChWmQwVdgyR74OaKRadvdzZy7I62ZVY= ++ mktemp + local LAST_OUT=/tmp/tmp.hcJ5lywFZ2 ++ mktemp + local LAST_ERR=/tmp/tmp.lIbRQKiXbv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec -it vault-service-1-7243-0 -- vault operator unseal -tls-skip-verify iZNb7C9P1+BkChWmQwVdgyR74OaKRadvdzZy7I62ZVY= + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hcJ5lywFZ2 Key Value --- ----- Seal Type shamir Initialized true Sealed false Total Shares 1 Threshold 1 Version 1.8.3 Storage Type file Cluster Name vault-cluster-26ed29a9 Cluster ID 6864e3f6-680d-275e-cbbb-44d478c0e4c0 HA Enabled false + cat /tmp/tmp.lIbRQKiXbv Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.hcJ5lywFZ2 /tmp/tmp.lIbRQKiXbv + return 0 + kubectl_bin exec -it vault-service-1-7243-0 -- sh -c 'export VAULT_TOKEN=s.Pon35VyeiFRtISlCTu6SC9fc && export VAULT_LOG_LEVEL=trace && vault secrets enable --version=1 -tls-skip-verify -path=secret kv && vault audit enable file file_path=/vault/vault-audit.log' ++ mktemp + local LAST_OUT=/tmp/tmp.gteQfl8RIc ++ mktemp + local LAST_ERR=/tmp/tmp.uJrjUvmMKX + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec -it vault-service-1-7243-0 -- sh -c 'export VAULT_TOKEN=s.Pon35VyeiFRtISlCTu6SC9fc && export VAULT_LOG_LEVEL=trace && vault secrets enable --version=1 -tls-skip-verify -path=secret kv && vault audit enable file file_path=/vault/vault-audit.log' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.gteQfl8RIc Success! Enabled the kv secrets engine at: secret/ Success! Enabled the file audit device at: file/ + cat /tmp/tmp.uJrjUvmMKX Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.gteQfl8RIc /tmp/tmp.uJrjUvmMKX + return 0 + sleep 10 + sed -e s/#secret/secret/ + sed -e 's/#vault_url/http:\/\/vault-service-1-7243.vault-service-1-7243.svc.cluster.local:8200/' + sed -e s/#token/s.Pon35VyeiFRtISlCTu6SC9fc/ + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/conf/vault-secret.yaml + '[' http == https ']' + /usr/bin/sed -i /#vault_ca/d /tmp/tmp.LW7qwu72v2/vault-secret.yaml + kubectl_bin apply --namespace=restore-to-encrypted-cluster-4709 -f /tmp/tmp.LW7qwu72v2/vault-secret.yaml ++ mktemp + local LAST_OUT=/tmp/tmp.odsMjGVyg1 ++ mktemp + local LAST_ERR=/tmp/tmp.mww2nabLJ0 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply --namespace=restore-to-encrypted-cluster-4709 -f /tmp/tmp.LW7qwu72v2/vault-secret.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.odsMjGVyg1 secret/some-name-vault created + cat /tmp/tmp.mww2nabLJ0 + rm /tmp/tmp.odsMjGVyg1 /tmp/tmp.mww2nabLJ0 + return 0 ++ kubectl_bin config current-context +++ mktemp ++ local LAST_OUT=/tmp/tmp.Xfhu2aPd8J +++ mktemp ++ local LAST_ERR=/tmp/tmp.K4iT5OM055 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl config current-context ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.Xfhu2aPd8J ++ cat /tmp/tmp.K4iT5OM055 ++ rm /tmp/tmp.Xfhu2aPd8J /tmp/tmp.K4iT5OM055 ++ return 0 + kubectl_bin config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9 --namespace=restore-to-encrypted-cluster-4709 ++ mktemp + local LAST_OUT=/tmp/tmp.hjHaS9FzUO ++ mktemp + local LAST_ERR=/tmp/tmp.QuaTpKvFsv + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl config set-context gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9 --namespace=restore-to-encrypted-cluster-4709 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.hjHaS9FzUO Context "gke_cloud-dev-112233_us-central1-a_jen-pxc-1522-1ed4605b-21-cluster9" modified. + cat /tmp/tmp.QuaTpKvFsv + rm /tmp/tmp.hjHaS9FzUO /tmp/tmp.QuaTpKvFsv + return 0 + run_recovery_check some-name on-demand-backup-pvc + local cluster=some-name + local backup1=on-demand-backup-pvc ++ get_proxy_engine some-name ++ local cluster_name=some-name +++ get_proxy some-name +++ local target_cluster=some-name ++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.0Z1RmY5O73 +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.sd0bEktqn9 ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.0Z1RmY5O73 ++++ cat /tmp/tmp.sd0bEktqn9 ++++ rm /tmp/tmp.0Z1RmY5O73 /tmp/tmp.sd0bEktqn9 ++++ return 0 +++ [[ '' == \t\r\u\e ]] ++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.KGKXhj1yLn +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.F9gmOipww0 ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.KGKXhj1yLn ++++ cat /tmp/tmp.F9gmOipww0 ++++ rm /tmp/tmp.KGKXhj1yLn /tmp/tmp.F9gmOipww0 ++++ return 0 +++ [[ true == \t\r\u\e ]] +++ echo some-name-proxysql +++ return ++ local cluster_proxy=some-name-proxysql ++ echo proxysql + local proxy=proxysql + desc 'recover backup on-demand-backup-pvc' + set +o xtrace ----------------------------------------------------------------------------------- recover backup on-demand-backup-pvc ----------------------------------------------------------------------------------- + kubectl_bin apply -f - + /usr/bin/sed -e s~minio-service.#namespace~minio-service.restore-to-encrypted-cluster-4709~ ++ mktemp + local LAST_OUT=/tmp/tmp.5NQiNLlt4Y ++ mktemp + local LAST_ERR=/tmp/tmp.E0p5gGSSs2 + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/restore-on-demand-backup-pvc.yaml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.5NQiNLlt4Y perconaxtradbclusterrestore.pxc.percona.com/on-demand-backup-pvc created + cat /tmp/tmp.E0p5gGSSs2 + rm /tmp/tmp.5NQiNLlt4Y /tmp/tmp.E0p5gGSSs2 + return 0 + wait_backup_restore on-demand-backup-pvc + local backup_name=on-demand-backup-pvc + local status=Succeeded + local wait_time=720 + desc 'wait backup restore' + set +o xtrace ----------------------------------------------------------------------------------- wait backup restore ----------------------------------------------------------------------------------- + set +o xtrace on-demand-backup-pvc....................................................................................................................Succeeded + kubectl_bin logs job/restore-job-on-demand-backup-pvc-some-name ++ mktemp + local LAST_OUT=/tmp/tmp.1XjaNQvsKb ++ mktemp + local LAST_ERR=/tmp/tmp.eiuWjiSyhd + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs job/restore-job-on-demand-backup-pvc-some-name + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.1XjaNQvsKb + LIB_PATH=/usr/lib/pxc + . /usr/lib/pxc/check-version.sh + . /usr/lib/pxc/vault.sh ++ set -o errexit ++ keyring_vault=/etc/mysql/vault-keyring-secret/keyring_vault.conf + SOCAT_OPTS=TCP:restore-src-on-demand-backup-pvc-some-name:3307,retry=30 + check_ssl + CA=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt + '[' -f /var/run/secrets/kubernetes.io/serviceaccount/service-ca.crt ']' + SSL_DIR=/etc/mysql/ssl + '[' -f /etc/mysql/ssl/ca.crt ']' + CA=/etc/mysql/ssl/ca.crt + SSL_INTERNAL_DIR=/etc/mysql/ssl-internal + '[' -f /etc/mysql/ssl-internal/ca.crt ']' + CA=/etc/mysql/ssl-internal/ca.crt + KEY=/etc/mysql/ssl/tls.key + CERT=/etc/mysql/ssl/tls.crt + '[' -f /etc/mysql/ssl-internal/tls.key -a -f /etc/mysql/ssl-internal/tls.crt ']' + KEY=/etc/mysql/ssl-internal/tls.key + CERT=/etc/mysql/ssl-internal/tls.crt + '[' -f /etc/mysql/ssl-internal/ca.crt -a -f /etc/mysql/ssl-internal/tls.key -a -f /etc/mysql/ssl-internal/tls.crt ']' + SOCAT_OPTS='openssl-connect:restore-src-on-demand-backup-pvc-some-name:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' + ping -c1 restore-src-on-demand-backup-pvc-some-name + : + rm -rf /datadir/#ib_16384_0.dblwr /datadir/#ib_16384_1.dblwr /datadir/#innodb_redo /datadir/#innodb_temp /datadir/auth_plugin /datadir/auto.cnf /datadir/binlog.000001 /datadir/binlog.000002 /datadir/binlog.000003 /datadir/binlog.index /datadir/galera.cache /datadir/get-pxc-state /datadir/grastate.dat /datadir/ib_buffer_pool /datadir/ibdata1 /datadir/innobackup.backup.full.log /datadir/innobackup.backup.log /datadir/liveness-check.sh /datadir/myApp /datadir/mysql /datadir/mysql.ibd /datadir/mysqld-error.log /datadir/peer-list /datadir/performance_schema /datadir/pmm-prerun.sh /datadir/private_key.pem /datadir/public_key.pem /datadir/pxc-configure-pxc.sh /datadir/pxc-entrypoint.sh /datadir/readiness-check.sh /datadir/sys /datadir/undo_001 /datadir/undo_002 /datadir/unsafe-bootstrap.sh /datadir/version_info ++ mktemp --directory /datadir/pxc_sst_XXXX + tmp=/datadir/pxc_sst_aGZn + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-some-name:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio + socat -u 'openssl-connect:restore-src-on-demand-backup-pvc-some-name:3307,reuseaddr,cert=/etc/mysql/ssl-internal/tls.crt,key=/etc/mysql/ssl-internal/tls.key,cafile=/etc/mysql/ssl-internal/ca.crt,verify=1,commonname='\'''\'',retry=30,no-sni=1' stdio ++ grep -c processor /proc/cpuinfo + xbstream --decompress -x -C /datadir/pxc_sst_aGZn --parallel=4 + set +o xtrace % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 14 100 14 0 0 823 0 --:--:-- --:--:-- --:--:-- 823 + xtrabackup --use-memory=100MB --prepare --rollback-prepared-trx --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_aGZn 2023-12-29T11:41:51.811174-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --innodb_checksum_algorithm=crc32 --innodb_log_checksums=1 --innodb_data_file_path=ibdata1:12M:autoextend --innodb_log_file_size=50331648 --innodb_page_size=16384 --innodb_undo_directory=./ --innodb_undo_tablespaces=2 --server-id=24694912 --innodb_log_checksums=ON --innodb_redo_log_encrypt=0 --innodb_undo_log_encrypt=0 2023-12-29T11:41:51.811304-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --use-memory=100MB --prepare=1 --rollback-prepared-trx=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_aGZn xtrabackup version 8.0.34-29 based on MySQL server 8.0.34 Linux (x86_64) (revision id: 5ba706ee) 2023-12-29T11:41:51.811355-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_aGZn/ 2023-12-29T11:41:51.811430-00:00 0 [Note] [MY-011825] [Xtrabackup] This target seems to be not prepared yet. 2023-12-29T11:41:51.824398-00:00 0 [Note] [MY-011825] [Xtrabackup] xtrabackup_logfile detected: size=8388608, start_lsn=(32578649) 2023-12-29T11:41:51.825343-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2023-12-29T11:41:51.825367-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2023-12-29T11:41:51.825374-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2023-12-29T11:41:51.825404-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2023-12-29T11:41:51.825419-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2023-12-29T11:41:51.825430-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2023-12-29T11:41:51.825714-00:00 0 [Note] [MY-011825] [Xtrabackup] inititialize_service_handles suceeded 2023-12-29T11:41:51.825999-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2023-12-29T11:41:51.826020-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2023-12-29T11:41:51.826026-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2023-12-29T11:41:51.826036-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2023-12-29T11:41:51.826041-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2023-12-29T11:41:51.826047-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2023-12-29T11:41:51.826058-00:00 0 [Note] [MY-011825] [Xtrabackup] Starting InnoDB instance for recovery. 2023-12-29T11:41:51.826068-00:00 0 [Note] [MY-011825] [Xtrabackup] Using 104857600 bytes for buffer pool (set by --use-memory parameter) 2023-12-29T11:41:51.826102-00:00 0 [Note] [MY-012932] [InnoDB] PUNCH HOLE support available 2023-12-29T11:41:51.826129-00:00 0 [Note] [MY-012944] [InnoDB] Uses event mutexes 2023-12-29T11:41:51.826144-00:00 0 [Note] [MY-012945] [InnoDB] GCC builtin __atomic_thread_fence() is used for memory barrier 2023-12-29T11:41:51.826154-00:00 0 [Note] [MY-012948] [InnoDB] Compressed tables use zlib 1.2.13 2023-12-29T11:41:51.826401-00:00 0 [Note] [MY-012951] [InnoDB] Using hardware accelerated crc32 and polynomial multiplication. 2023-12-29T11:41:51.826876-00:00 0 [Note] [MY-012203] [InnoDB] Directories to scan './' 2023-12-29T11:41:51.826920-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2023-12-29T11:41:51.836566-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 8 files. 2023-12-29T11:41:51.837621-00:00 0 [Note] [MY-012955] [InnoDB] Initializing buffer pool, total size = 128.000000M, instances = 1, chunk size =128.000000M 2023-12-29T11:41:51.852183-00:00 0 [Note] [MY-012957] [InnoDB] Completed initialization of buffer pool 2023-12-29T11:41:51.858473-00:00 0 [Note] [MY-011952] [InnoDB] If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). 2023-12-29T11:41:51.930685-00:00 0 [Note] [MY-013883] [InnoDB] The latest found checkpoint is at lsn = 32578649 in redo log file ./#innodb_redo/#ib_redo0. 2023-12-29T11:41:51.930751-00:00 0 [Note] [MY-012560] [InnoDB] The log sequence number 32456618 in the system tablespace does not match the log sequence number 32578649 in the redo log files! 2023-12-29T11:41:51.930764-00:00 0 [Note] [MY-012551] [InnoDB] Database was not shutdown normally! 2023-12-29T11:41:51.930775-00:00 0 [Note] [MY-012552] [InnoDB] Starting crash recovery. 2023-12-29T11:41:51.933239-00:00 0 [Note] [MY-013086] [InnoDB] Starting to parse redo log at lsn = 32578639, whereas checkpoint_lsn = 32578649 and start_lsn = 32578560 2023-12-29T11:41:51.933265-00:00 0 [Note] [MY-012550] [InnoDB] Doing recovery: scanned up to log sequence number 32578659 2023-12-29T11:41:51.951072-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2023-12-29T11:41:51.973540-00:00 0 [Note] [MY-012532] [InnoDB] Applying a batch of 1 redo log records ... 2023-12-29T11:41:51.973656-00:00 0 [Note] [MY-012533] [InnoDB] 100% 2023-12-29T11:41:51.973672-00:00 0 [Note] [MY-012535] [InnoDB] Apply batch completed! 2023-12-29T11:41:52.075508-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2023-12-29T11:41:52.076409-00:00 0 [Note] [MY-013888] [InnoDB] Upgrading redo log: 1032M, LSN=32578659. 2023-12-29T11:41:52.076441-00:00 0 [Note] [MY-012968] [InnoDB] Starting to delete and rewrite redo log files. 2023-12-29T11:41:52.076609-00:00 0 [Note] [MY-011825] [InnoDB] Removing redo log file: ./#innodb_redo/#ib_redo0 2023-12-29T11:41:52.137753-00:00 0 [Note] [MY-011825] [InnoDB] Creating redo log file at ./#innodb_redo/#ib_redo0_tmp with file_id 0 with size 33554432 bytes 2023-12-29T11:41:52.140766-00:00 0 [Note] [MY-011825] [InnoDB] Renaming redo log file from ./#innodb_redo/#ib_redo0_tmp to ./#innodb_redo/#ib_redo0 2023-12-29T11:41:52.143156-00:00 0 [Note] [MY-012893] [InnoDB] New redo log files created, LSN=32579084 2023-12-29T11:41:52.143275-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2023-12-29T11:41:52.164893-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2023-12-29T11:41:52.167633-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2023-12-29T11:41:52.170415-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2023-12-29T11:41:52.170552-00:00 0 [Note] [MY-011980] [InnoDB] GTID recovery trx_no: 5192 2023-12-29T11:41:52.215905-00:00 0 [Note] [MY-013776] [InnoDB] Parallel initialization of rseg complete 2023-12-29T11:41:52.215970-00:00 0 [Note] [MY-013777] [InnoDB] Time taken to initialize rseg using 4 thread: 45425 ms. 2023-12-29T11:41:52.216066-00:00 0 [Note] [MY-012923] [InnoDB] Creating shared tablespace for temporary tables 2023-12-29T11:41:52.216143-00:00 0 [Note] [MY-012265] [InnoDB] Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... 2023-12-29T11:41:52.246590-00:00 0 [Note] [MY-012266] [InnoDB] File './ibtmp1' size is now 12 MB. 2023-12-29T11:41:52.246814-00:00 0 [Note] [MY-013627] [InnoDB] Scanning temp tablespace dir:'./#innodb_temp/' 2023-12-29T11:41:52.264521-00:00 0 [Note] [MY-013018] [InnoDB] Created 128 and tracked 128 new rollback segment(s) in the temporary tablespace. 128 are now active. 2023-12-29T11:41:52.264817-00:00 0 [Note] [MY-012976] [InnoDB] 8.0.34 started; log sequence number 32579094 2023-12-29T11:41:52.265320-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2023-12-29T11:41:52.266990-00:00 0 [Note] [MY-011825] [Xtrabackup] Completed loading of 6 tablespaces into cache in 0.00210736 seconds 2023-12-29T11:41:52.287620-00:00 0 [Note] [MY-011825] [Xtrabackup] Time taken to build dictionary: 0.020594 seconds 2023-12-29T11:41:52.288305-00:00 0 [Note] [MY-011825] [Xtrabackup] Recovered WSREP position: 1410f97c-a63e-11ee-810f-1a4fefb65588:38 2023-12-29T11:41:52.288349-00:00 0 [Note] [MY-011825] [Xtrabackup] starting shutdown with innodb_fast_shutdown = 1 2023-12-29T11:41:52.288477-00:00 0 [Note] [MY-012330] [InnoDB] FTS optimize thread exiting. 2023-12-29T11:41:53.288155-00:00 0 [Note] [MY-013072] [InnoDB] Starting shutdown... 2023-12-29T11:41:53.294063-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2023-12-29T11:41:53.315030-00:00 0 [Note] [MY-012980] [InnoDB] Shutdown completed; log sequence number 32579094 2023-12-29T11:41:53.319340-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + xtrabackup --defaults-group=mysqld --datadir=/datadir --move-back --force-non-empty-directories --keyring-vault-config=/etc/mysql/vault-keyring-secret/keyring_vault.conf --early-plugin-load=keyring_vault.so --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_aGZn 2023-12-29T11:41:53.340185-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --defaults_group=mysqld --datadir=/datadir 2023-12-29T11:41:53.340292-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --move-back=1 --force-non-empty-directories=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_aGZn xtrabackup version 8.0.34-29 based on MySQL server 8.0.34 Linux (x86_64) (revision id: 5ba706ee) 2023-12-29T11:41:53.340330-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_aGZn/ 2023-12-29T11:41:53.341156-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_001 to /datadir/undo_001 2023-12-29T11:41:53.341214-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_001 to /datadir/undo_001 2023-12-29T11:41:53.341246-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_002 to /datadir/undo_002 2023-12-29T11:41:53.341281-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_002 to /datadir/undo_002 2023-12-29T11:41:53.341484-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving ibdata1 to /datadir/ibdata1 2023-12-29T11:41:53.341518-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file ibdata1 to /datadir/ibdata1 2023-12-29T11:41:53.341802-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.000009 to /datadir//binlog.000009 2023-12-29T11:41:53.341837-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.000009 to /datadir//binlog.000009 2023-12-29T11:41:53.341910-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.index to /datadir//binlog.index 2023-12-29T11:41:53.341961-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.index to /datadir//binlog.index 2023-12-29T11:41:53.342541-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2023-12-29T11:41:53.342633-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2023-12-29T11:41:53.342688-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2023-12-29T11:41:53.342731-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2023-12-29T11:41:53.342760-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2023-12-29T11:41:53.342803-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2023-12-29T11:41:53.342832-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2023-12-29T11:41:53.342863-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2023-12-29T11:41:53.342891-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2023-12-29T11:41:53.342921-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2023-12-29T11:41:53.342970-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2023-12-29T11:41:53.343011-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2023-12-29T11:41:53.343047-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2023-12-29T11:41:53.343086-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2023-12-29T11:41:53.343134-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2023-12-29T11:41:53.343174-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2023-12-29T11:41:53.343214-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2023-12-29T11:41:53.343261-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2023-12-29T11:41:53.343293-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2023-12-29T11:41:53.343330-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2023-12-29T11:41:53.343361-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2023-12-29T11:41:53.343398-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2023-12-29T11:41:53.343431-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2023-12-29T11:41:53.343467-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2023-12-29T11:41:53.343501-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2023-12-29T11:41:53.343534-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2023-12-29T11:41:53.343568-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2023-12-29T11:41:53.343603-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2023-12-29T11:41:53.343636-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2023-12-29T11:41:53.343676-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2023-12-29T11:41:53.343712-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2023-12-29T11:41:53.343749-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2023-12-29T11:41:53.343785-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2023-12-29T11:41:53.343817-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2023-12-29T11:41:53.343858-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2023-12-29T11:41:53.343896-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2023-12-29T11:41:53.343926-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2023-12-29T11:41:53.343976-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2023-12-29T11:41:53.344014-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2023-12-29T11:41:53.344050-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2023-12-29T11:41:53.344087-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2023-12-29T11:41:53.344128-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2023-12-29T11:41:53.344158-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2023-12-29T11:41:53.344188-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2023-12-29T11:41:53.344224-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2023-12-29T11:41:53.344268-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2023-12-29T11:41:53.344305-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2023-12-29T11:41:53.344342-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2023-12-29T11:41:53.344373-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2023-12-29T11:41:53.344406-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2023-12-29T11:41:53.344444-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2023-12-29T11:41:53.344482-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2023-12-29T11:41:53.344513-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2023-12-29T11:41:53.344552-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2023-12-29T11:41:53.344586-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2023-12-29T11:41:53.344625-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2023-12-29T11:41:53.344659-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2023-12-29T11:41:53.344693-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2023-12-29T11:41:53.344728-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2023-12-29T11:41:53.344759-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2023-12-29T11:41:53.344793-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2023-12-29T11:41:53.344825-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2023-12-29T11:41:53.344860-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2023-12-29T11:41:53.344899-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2023-12-29T11:41:53.344961-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2023-12-29T11:41:53.344998-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2023-12-29T11:41:53.345033-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2023-12-29T11:41:53.345069-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2023-12-29T11:41:53.345125-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2023-12-29T11:41:53.345163-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2023-12-29T11:41:53.345194-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2023-12-29T11:41:53.345233-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2023-12-29T11:41:53.345277-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2023-12-29T11:41:53.345315-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2023-12-29T11:41:53.345389-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2023-12-29T11:41:53.345427-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2023-12-29T11:41:53.345456-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2023-12-29T11:41:53.345491-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2023-12-29T11:41:53.345524-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2023-12-29T11:41:53.345558-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2023-12-29T11:41:53.345594-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2023-12-29T11:41:53.345666-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2023-12-29T11:41:53.345701-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2023-12-29T11:41:53.345734-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2023-12-29T11:41:53.345802-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2023-12-29T11:41:53.345871-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2023-12-29T11:41:53.345909-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2023-12-29T11:41:53.345952-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2023-12-29T11:41:53.345994-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2023-12-29T11:41:53.346031-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2023-12-29T11:41:53.346065-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2023-12-29T11:41:53.346102-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2023-12-29T11:41:53.346135-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2023-12-29T11:41:53.346172-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2023-12-29T11:41:53.346203-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2023-12-29T11:41:53.346278-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2023-12-29T11:41:53.346315-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2023-12-29T11:41:53.346374-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2023-12-29T11:41:53.346412-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2023-12-29T11:41:53.346445-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2023-12-29T11:41:53.346482-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2023-12-29T11:41:53.346521-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2023-12-29T11:41:53.346557-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2023-12-29T11:41:53.346589-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2023-12-29T11:41:53.346626-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2023-12-29T11:41:53.346663-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2023-12-29T11:41:53.346696-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2023-12-29T11:41:53.346732-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2023-12-29T11:41:53.346762-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2023-12-29T11:41:53.346796-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2023-12-29T11:41:53.346829-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2023-12-29T11:41:53.346868-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2023-12-29T11:41:53.346903-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2023-12-29T11:41:53.346947-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2023-12-29T11:41:53.346983-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2023-12-29T11:41:53.347021-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2023-12-29T11:41:53.347055-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2023-12-29T11:41:53.347089-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2023-12-29T11:41:53.347121-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2023-12-29T11:41:53.347160-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2023-12-29T11:41:53.347196-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2023-12-29T11:41:53.347233-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2023-12-29T11:41:53.347279-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2023-12-29T11:41:53.347320-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2023-12-29T11:41:53.347357-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2023-12-29T11:41:53.347399-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2023-12-29T11:41:53.347439-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2023-12-29T11:41:53.347472-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2023-12-29T11:41:53.347513-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2023-12-29T11:41:53.347545-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2023-12-29T11:41:53.347576-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2023-12-29T11:41:53.347613-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2023-12-29T11:41:53.347648-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2023-12-29T11:41:53.347684-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2023-12-29T11:41:53.347719-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2023-12-29T11:41:53.347755-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2023-12-29T11:41:53.347792-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2023-12-29T11:41:53.347827-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2023-12-29T11:41:53.347857-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2023-12-29T11:41:53.347897-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2023-12-29T11:41:53.347931-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2023-12-29T11:41:53.347980-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2023-12-29T11:41:53.348017-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2023-12-29T11:41:53.348050-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2023-12-29T11:41:53.348087-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2023-12-29T11:41:53.348118-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2023-12-29T11:41:53.348155-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2023-12-29T11:41:53.348194-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2023-12-29T11:41:53.348232-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2023-12-29T11:41:53.348274-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2023-12-29T11:41:53.348310-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2023-12-29T11:41:53.348348-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2023-12-29T11:41:53.348380-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2023-12-29T11:41:53.348413-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2023-12-29T11:41:53.348444-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2023-12-29T11:41:53.348476-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2023-12-29T11:41:53.348511-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2023-12-29T11:41:53.348547-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2023-12-29T11:41:53.348581-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2023-12-29T11:41:53.348619-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2023-12-29T11:41:53.348663-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2023-12-29T11:41:53.348705-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2023-12-29T11:41:53.348742-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2023-12-29T11:41:53.348775-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2023-12-29T11:41:53.348811-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2023-12-29T11:41:53.348847-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2023-12-29T11:41:53.348882-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2023-12-29T11:41:53.348912-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2023-12-29T11:41:53.348959-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2023-12-29T11:41:53.348991-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2023-12-29T11:41:53.349030-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2023-12-29T11:41:53.349066-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2023-12-29T11:41:53.349103-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2023-12-29T11:41:53.349140-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2023-12-29T11:41:53.349177-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2023-12-29T11:41:53.349209-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2023-12-29T11:41:53.349238-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2023-12-29T11:41:53.349282-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2023-12-29T11:41:53.349316-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2023-12-29T11:41:53.349350-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2023-12-29T11:41:53.349383-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2023-12-29T11:41:53.349422-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2023-12-29T11:41:53.349455-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2023-12-29T11:41:53.349502-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2023-12-29T11:41:53.349540-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2023-12-29T11:41:53.349573-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2023-12-29T11:41:53.349610-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2023-12-29T11:41:53.349649-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2023-12-29T11:41:53.349679-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2023-12-29T11:41:53.349716-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2023-12-29T11:41:53.349752-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2023-12-29T11:41:53.349787-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2023-12-29T11:41:53.349820-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2023-12-29T11:41:53.349852-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2023-12-29T11:41:53.349890-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2023-12-29T11:41:53.349927-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2023-12-29T11:41:53.349973-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2023-12-29T11:41:53.350006-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2023-12-29T11:41:53.350041-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2023-12-29T11:41:53.350074-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2023-12-29T11:41:53.350107-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2023-12-29T11:41:53.350142-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2023-12-29T11:41:53.350178-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2023-12-29T11:41:53.350211-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2023-12-29T11:41:53.350241-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2023-12-29T11:41:53.350289-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2023-12-29T11:41:53.350326-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2023-12-29T11:41:53.350366-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2023-12-29T11:41:53.350412-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2023-12-29T11:41:53.350451-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2023-12-29T11:41:53.350486-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2023-12-29T11:41:53.350525-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2023-12-29T11:41:53.350558-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2023-12-29T11:41:53.350593-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2023-12-29T11:41:53.350628-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2023-12-29T11:41:53.350667-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2023-12-29T11:41:53.350705-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2023-12-29T11:41:53.350742-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2023-12-29T11:41:53.350781-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2023-12-29T11:41:53.350822-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2023-12-29T11:41:53.350859-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2023-12-29T11:41:53.350899-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2023-12-29T11:41:53.350947-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2023-12-29T11:41:53.350983-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2023-12-29T11:41:53.351021-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2023-12-29T11:41:53.351062-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2023-12-29T11:41:53.351160-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2023-12-29T11:41:53.351197-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2023-12-29T11:41:53.351232-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2023-12-29T11:41:53.351272-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2023-12-29T11:41:53.351307-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2023-12-29T11:41:53.351336-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2023-12-29T11:41:53.351369-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2023-12-29T11:41:53.351413-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2023-12-29T11:41:53.351451-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2023-12-29T11:41:53.351483-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2023-12-29T11:41:53.351518-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2023-12-29T11:41:53.351546-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2023-12-29T11:41:53.351578-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2023-12-29T11:41:53.351606-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2023-12-29T11:41:53.351640-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2023-12-29T11:41:53.351669-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2023-12-29T11:41:53.351702-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2023-12-29T11:41:53.351729-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2023-12-29T11:41:53.351821-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2023-12-29T11:41:53.351860-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2023-12-29T11:41:53.351876-00:00 1 [Note] [MY-011825] [Xtrabackup] Creating directory ./#innodb_redo 2023-12-29T11:41:53.351891-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: creating directory ./#innodb_redo 2023-12-29T11:41:53.351924-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sst_info to /datadir/sst_info 2023-12-29T11:41:53.351970-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sst_info to /datadir/sst_info 2023-12-29T11:41:53.352001-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2023-12-29T11:41:53.352036-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2023-12-29T11:41:53.352072-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql.ibd to /datadir/mysql.ibd 2023-12-29T11:41:53.352103-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql.ibd to /datadir/mysql.ibd 2023-12-29T11:41:53.352139-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ibtmp1 to /datadir/ibtmp1 2023-12-29T11:41:53.352168-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ibtmp1 to /datadir/ibtmp1 2023-12-29T11:41:53.352272-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2023-12-29T11:41:53.352311-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2023-12-29T11:41:53.352344-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_info to /datadir/xtrabackup_info 2023-12-29T11:41:53.352372-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_info to /datadir/xtrabackup_info 2023-12-29T11:41:53.352399-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ib_buffer_pool to /datadir/ib_buffer_pool 2023-12-29T11:41:53.352428-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ib_buffer_pool to /datadir/ib_buffer_pool 2023-12-29T11:41:53.443530-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + cat /tmp/tmp.eiuWjiSyhd + rm /tmp/tmp.1XjaNQvsKb /tmp/tmp.eiuWjiSyhd + return 0 + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met some-name-proxysql-0.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met some-name-pxc-0.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met some-name-pxc-1.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met some-name-pxc-2.Ok + sleep 35 + desc 'check data after backup' + set +o xtrace ----------------------------------------------------------------------------------- check data after backup ----------------------------------------------------------------------------------- + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.dVGjc5fHxe +++ mktemp ++ local LAST_ERR=/tmp/tmp.fr3BvCyt1L ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.dVGjc5fHxe ++ cat /tmp/tmp.fr3BvCyt1L ++ rm /tmp/tmp.dVGjc5fHxe /tmp/tmp.fr3BvCyt1L ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.LW7qwu72v2/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.LW7qwu72v2/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.GPWiRV7ymt +++ mktemp ++ local LAST_ERR=/tmp/tmp.LufkB4e1aO ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.GPWiRV7ymt ++ cat /tmp/tmp.LufkB4e1aO ++ rm /tmp/tmp.GPWiRV7ymt /tmp/tmp.LufkB4e1aO ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.LW7qwu72v2/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.LW7qwu72v2/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.6OenruCSay +++ mktemp ++ local LAST_ERR=/tmp/tmp.lWpj6a6qmZ ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.6OenruCSay ++ cat /tmp/tmp.lWpj6a6qmZ ++ rm /tmp/tmp.6OenruCSay /tmp/tmp.lWpj6a6qmZ ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ echo pxc-client-599db5b9bc-hllss ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.LW7qwu72v2/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.LW7qwu72v2/select-1.sql + '[' on-demand-backup-pvc '!=' on-demand-backup-minio ']' + desc 'copy backup' + set +o xtrace ----------------------------------------------------------------------------------- copy backup ----------------------------------------------------------------------------------- + bash /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/deploy/backup/copy-backup.sh on-demand-backup-pvc /tmp/tmp.LW7qwu72v2/backup Log: /tmp/tmp.hbaymTQvwX/log pvc/xb-on-demand-backup-pvcpod/backup-access created Starting pod.........[done] Downloading started tar: Removing leading `/' from member names Downloading finished pod "backup-access" deleted You can recover data locally with following commands: $ service mysqld stop $ rm -rf /var/lib/mysql/* $ cat /tmp/tmp.LW7qwu72v2/backup/xtrabackup.stream | xbstream --decompress -x -C /var/lib/mysql $ xtrabackup --prepare --target-dir=/var/lib/mysql $ chown -R mysql:mysql /var/lib/mysql $ service mysqld start + check_pvc_md5 + desc 'check backup file md5sum' + set +o xtrace ----------------------------------------------------------------------------------- check backup file md5sum ----------------------------------------------------------------------------------- + apply_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/client.yml + '[' -z '' ']' + kubectl_bin apply -f - + cat_config /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/client.yml + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/client.yml ++ mktemp + /usr/bin/sed -e 's#apiVersion: pxc.percona.com/v.*$#apiVersion: pxc.percona.com/v1#' + local LAST_OUT=/tmp/tmp.9zsoYFAZZI + /usr/bin/sed -e 's#image:.*\/percona-xtradb-cluster:.*$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-pxc\([0-9]*.[0-9]*\)\{0,1\}$#image: perconalab/percona-xtradb-cluster-operator:main-pxc8.0#' + /usr/bin/sed -e 's#image:.*-backup$#image: perconalab/percona-xtradb-cluster-operator:k8spxc-1313-8.0-backup#' ++ mktemp + /usr/bin/sed -e 's#image:.*-pmm$#image: perconalab/pmm-client:dev-latest#' + /usr/bin/sed -e 's#apply:.*#apply: Never#' + /usr/bin/sed -e 's#image:.*-haproxy$#image: perconalab/percona-xtradb-cluster-operator:main-haproxy#' + /usr/bin/sed -e 's#image:.*-logcollector$#image: perconalab/percona-xtradb-cluster-operator:main-logcollector#' + /usr/bin/sed -e s~minio-service.#namespace~minio-service.restore-to-encrypted-cluster-4709~ + local LAST_ERR=/tmp/tmp.InxLvT8Kcd + local exit_status=0 + /usr/bin/sed -e 's#image:.*-init$#image: perconalab/percona-xtradb-cluster-operator:PR-1522-1ed4605b#' + /usr/bin/sed -e 's#image:.*-proxysql$#image: perconalab/percona-xtradb-cluster-operator:main-proxysql#' ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.9zsoYFAZZI deployment.apps/backup-client created + cat /tmp/tmp.InxLvT8Kcd + rm /tmp/tmp.9zsoYFAZZI /tmp/tmp.InxLvT8Kcd + return 0 + sleep 10 ++ kubectl_bin get pods --selector=name=backup-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.9eim42bI0l +++ mktemp ++ local LAST_ERR=/tmp/tmp.cUafrujBgc ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=backup-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.9eim42bI0l ++ cat /tmp/tmp.cUafrujBgc ++ rm /tmp/tmp.9eim42bI0l /tmp/tmp.cUafrujBgc ++ return 0 + bak_client_pod=backup-client-6bc75f79d6-nhncm + wait_pod backup-client-6bc75f79d6-nhncm + local pod=backup-client-6bc75f79d6-nhncm + local max_retry=480 + local ns= ++ echo backup-client-6bc75f79d6-nhncm ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/backup-client-6bc75f79d6-nhncm condition met backup-client-6bc75f79d6-nhncm.Ok + kubectl_bin exec backup-client-6bc75f79d6-nhncm -- bash -c 'cd /backup; md5sum -c md5sum.txt' ++ mktemp + local LAST_OUT=/tmp/tmp.T8YXPtyccE ++ mktemp + local LAST_ERR=/tmp/tmp.cL9MUyxVah + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec backup-client-6bc75f79d6-nhncm -- bash -c 'cd /backup; md5sum -c md5sum.txt' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.T8YXPtyccE xtrabackup.stream: OK + cat /tmp/tmp.cL9MUyxVah + rm /tmp/tmp.T8YXPtyccE /tmp/tmp.cL9MUyxVah + return 0 + kubectl_bin delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/client.yml ++ mktemp + local LAST_OUT=/tmp/tmp.aVcL4lGWVj ++ mktemp + local LAST_ERR=/tmp/tmp.4pf5rkLecD + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/client.yml + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.aVcL4lGWVj deployment.apps "backup-client" deleted + cat /tmp/tmp.4pf5rkLecD + rm /tmp/tmp.aVcL4lGWVj /tmp/tmp.4pf5rkLecD + return 0 + keyring_plugin_must_be_in_use some-name + local cluster=some-name + desc 'check keyring plugin usage' + set +o xtrace ----------------------------------------------------------------------------------- check keyring plugin usage ----------------------------------------------------------------------------------- + is_keyring_plugin_in_use some-name + local cluster=some-name + kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ mktemp + local LAST_OUT=/tmp/tmp.rsdwCpbHeR + egrep -o 'early-plugin-load=keyring_\w+.so' ++ mktemp + local LAST_ERR=/tmp/tmp.slHc97Y9ax + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.rsdwCpbHeR + cat /tmp/tmp.slHc97Y9ax Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.rsdwCpbHeR /tmp/tmp.slHc97Y9ax + return 0 early-plugin-load=keyring_vault.so + table_must_not_be_encrypted some-name myApp + local cluster=some-name + local table=myApp + is_table_encrypted some-name myApp + local cluster=some-name + local table=myApp + run_mysql 'SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' '-h some-name-proxysql -uroot -proot_password' + local 'command=SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' + local 'uri=-h some-name-proxysql -uroot -proot_password' + egrep -o 'ENCRYPTION=('\''Y'\''|"Y")' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.XuoVoTWf6K +++ mktemp ++ local LAST_ERR=/tmp/tmp.eT2UploBpq ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.XuoVoTWf6K ++ cat /tmp/tmp.eT2UploBpq ++ rm /tmp/tmp.XuoVoTWf6K /tmp/tmp.eT2UploBpq ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' -z '' ']' + run_recovery_check some-name on-demand-backup-aws-s3 + local cluster=some-name + local backup1=on-demand-backup-aws-s3 ++ get_proxy_engine some-name ++ local cluster_name=some-name +++ get_proxy some-name +++ local target_cluster=some-name ++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.kybHVXsO4T +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.nqew5koEdQ ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc some-name -o 'jsonpath={.spec.haproxy.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.kybHVXsO4T ++++ cat /tmp/tmp.nqew5koEdQ ++++ rm /tmp/tmp.kybHVXsO4T /tmp/tmp.nqew5koEdQ ++++ return 0 +++ [[ '' == \t\r\u\e ]] ++++ kubectl_bin get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' +++++ mktemp ++++ local LAST_OUT=/tmp/tmp.r6H2cp4RIl +++++ mktemp ++++ local LAST_ERR=/tmp/tmp.BsOobNiT34 ++++ local exit_status=0 +++++ seq 0 2 ++++ for i in '$(seq 0 2)' ++++ set +e ++++ kubectl get pxc some-name -o 'jsonpath={.spec.proxysql.enabled}' ++++ exit_status=0 ++++ set -e ++++ '[' 0 '!=' 0 ']' ++++ break ++++ cat /tmp/tmp.r6H2cp4RIl ++++ cat /tmp/tmp.BsOobNiT34 ++++ rm /tmp/tmp.r6H2cp4RIl /tmp/tmp.BsOobNiT34 ++++ return 0 +++ [[ true == \t\r\u\e ]] +++ echo some-name-proxysql +++ return ++ local cluster_proxy=some-name-proxysql ++ echo proxysql + local proxy=proxysql + desc 'recover backup on-demand-backup-aws-s3' + set +o xtrace ----------------------------------------------------------------------------------- recover backup on-demand-backup-aws-s3 ----------------------------------------------------------------------------------- + /usr/bin/sed -e s~minio-service.#namespace~minio-service.restore-to-encrypted-cluster-4709~ + cat /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/conf/restore-on-demand-backup-aws-s3.yaml + kubectl_bin apply -f - ++ mktemp + local LAST_OUT=/tmp/tmp.wP5AwSdLqZ ++ mktemp + local LAST_ERR=/tmp/tmp.4WlkVcaxDG + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl apply -f - + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.wP5AwSdLqZ perconaxtradbclusterrestore.pxc.percona.com/on-demand-backup-aws-s3 created + cat /tmp/tmp.4WlkVcaxDG + rm /tmp/tmp.wP5AwSdLqZ /tmp/tmp.4WlkVcaxDG + return 0 + wait_backup_restore on-demand-backup-aws-s3 + local backup_name=on-demand-backup-aws-s3 + local status=Succeeded + local wait_time=720 + desc 'wait backup restore' + set +o xtrace ----------------------------------------------------------------------------------- wait backup restore ----------------------------------------------------------------------------------- + set +o xtrace on-demand-backup-aws-s3......................................................................................................................Succeeded + kubectl_bin logs job/restore-job-on-demand-backup-aws-s3-some-name ++ mktemp + local LAST_OUT=/tmp/tmp.KzyxaFGucm ++ mktemp + local LAST_ERR=/tmp/tmp.GWj3FVLASp + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs job/restore-job-on-demand-backup-aws-s3-some-name + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.KzyxaFGucm + LIB_PATH=/usr/lib/pxc + . /usr/lib/pxc/check-version.sh + . /usr/lib/pxc/vault.sh ++ set -o errexit ++ keyring_vault=/etc/mysql/vault-keyring-secret/keyring_vault.conf + XBCLOUD_ARGS='--curl-retriable-errors=7 ' + MC_ARGS='-C /tmp/mc' + '[' -n true ']' + [[ true == \f\a\l\s\e ]] + '[' -n operator-testing/some-name-2023-12-29-11:38:59-full ']' + mc -C /tmp/mc config host add dest https://s3.amazonaws.com ACCESS_KEY_ID SECRET_ACCESS_KEY Added `dest` successfully. + mc -C /tmp/mc ls dest/operator-testing/some-name-2023-12-29-11:38:59-full [2023-12-29 11:39:27 UTC] 501B STANDARD backup-my.cnf.00000000000000000000 [2023-12-29 11:39:27 UTC] 27B STANDARD backup-my.cnf.00000000000000000001 [2023-12-29 11:39:27 UTC] 244B STANDARD binlog.000010.00000000000000000000 [2023-12-29 11:39:27 UTC] 27B STANDARD binlog.000010.00000000000000000001 [2023-12-29 11:39:27 UTC] 62B STANDARD binlog.index.00000000000000000000 [2023-12-29 11:39:27 UTC] 26B STANDARD binlog.index.00000000000000000001 [2023-12-29 11:39:27 UTC] 3.7KiB STANDARD ib_buffer_pool.00000000000000000000 [2023-12-29 11:39:27 UTC] 28B STANDARD ib_buffer_pool.00000000000000000001 [2023-12-29 11:39:21 UTC] 10MiB STANDARD ibdata1.00000000000000000000 [2023-12-29 11:39:22 UTC] 2.0MiB STANDARD ibdata1.00000000000000000001 [2023-12-29 11:39:22 UTC] 21B STANDARD ibdata1.00000000000000000002 [2023-12-29 11:39:21 UTC] 10MiB STANDARD mysql.ibd.00000000000000000000 [2023-12-29 11:39:22 UTC] 10MiB STANDARD mysql.ibd.00000000000000000001 [2023-12-29 11:39:22 UTC] 10MiB STANDARD mysql.ibd.00000000000000000002 [2023-12-29 11:39:22 UTC] 23B STANDARD mysql.ibd.00000000000000000003 [2023-12-29 11:39:21 UTC] 10MiB STANDARD undo_001.00000000000000000000 [2023-12-29 11:39:22 UTC] 6.0MiB STANDARD undo_001.00000000000000000001 [2023-12-29 11:39:22 UTC] 22B STANDARD undo_001.00000000000000000002 [2023-12-29 11:39:21 UTC] 10MiB STANDARD undo_002.00000000000000000000 [2023-12-29 11:39:22 UTC] 6.0MiB STANDARD undo_002.00000000000000000001 [2023-12-29 11:39:22 UTC] 22B STANDARD undo_002.00000000000000000002 [2023-12-29 11:39:27 UTC] 116B STANDARD xtrabackup_binlog_info.00000000000000000000 [2023-12-29 11:39:27 UTC] 36B STANDARD xtrabackup_binlog_info.00000000000000000001 [2023-12-29 11:39:27 UTC] 190B STANDARD xtrabackup_checkpoints.00000000000000000000 [2023-12-29 11:39:27 UTC] 36B STANDARD xtrabackup_checkpoints.00000000000000000001 [2023-12-29 11:39:27 UTC] 852B STANDARD xtrabackup_info.00000000000000000000 [2023-12-29 11:39:27 UTC] 29B STANDARD xtrabackup_info.00000000000000000001 [2023-12-29 11:39:27 UTC] 3.1KiB STANDARD xtrabackup_logfile.00000000000000000000 [2023-12-29 11:39:27 UTC] 32B STANDARD xtrabackup_logfile.00000000000000000001 [2023-12-29 11:39:27 UTC] 95B STANDARD xtrabackup_tablespaces.00000000000000000000 [2023-12-29 11:39:27 UTC] 36B STANDARD xtrabackup_tablespaces.00000000000000000001 [2023-12-29 11:49:01 UTC] 0B myApp/ [2023-12-29 11:49:01 UTC] 0B mysql/ [2023-12-29 11:49:01 UTC] 0B performance_schema/ [2023-12-29 11:49:01 UTC] 0B sys/ + '[' -n '' ']' + rm -rf /datadir/#ib_16384_0.dblwr /datadir/#ib_16384_1.dblwr /datadir/#innodb_redo /datadir/#innodb_temp /datadir/auth_plugin /datadir/auto.cnf /datadir/binlog.000009 /datadir/binlog.000010 /datadir/binlog.000011 /datadir/binlog.index /datadir/galera.cache /datadir/get-pxc-state /datadir/grastate.dat /datadir/ib_buffer_pool /datadir/ibdata1 /datadir/innobackup.backup.full.log /datadir/innobackup.backup.log /datadir/liveness-check.sh /datadir/myApp /datadir/mysql /datadir/mysql.ibd /datadir/mysqld-error.log /datadir/peer-list /datadir/performance_schema /datadir/pmm-prerun.sh /datadir/private_key.pem /datadir/public_key.pem /datadir/pxc-configure-pxc.sh /datadir/pxc-entrypoint.sh /datadir/readiness-check.sh /datadir/sst_info /datadir/sys /datadir/undo_001 /datadir/undo_002 /datadir/unsafe-bootstrap.sh /datadir/version_info /datadir/wsrep_recovery_verbose_history.log /datadir/xtrabackup_info ++ mktemp --directory /datadir/pxc_sst_XXXX + tmp=/datadir/pxc_sst_1JAQ ++ grep -c processor /proc/cpuinfo ++ grep -c processor /proc/cpuinfo + xbstream -x -C /datadir/pxc_sst_1JAQ --parallel=4 ++ destination ++ '[' -n operator-testing/some-name-2023-12-29-11:38:59-full ']' ++ echo -n s3://operator-testing/some-name-2023-12-29-11:38:59-full + xbcloud get --parallel=4 --curl-retriable-errors=7 s3://operator-testing/some-name-2023-12-29-11:38:59-full.sst_info 231229 11:49:02 xbcloud: Successfully connected. 231229 11:49:02 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full.sst_info/sst_info.00000000000000000000. 231229 11:49:02 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full.sst_info/sst_info.00000000000000000000, size 145 231229 11:49:02 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full.sst_info/sst_info.00000000000000000001. 231229 11:49:02 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full.sst_info/sst_info.00000000000000000001, size 22 231229 11:49:02 xbcloud: Download completed. ++ grep -c processor /proc/cpuinfo ++ grep -c processor /proc/cpuinfo ++ destination + xbstream --decompress -x -C /datadir/pxc_sst_1JAQ --parallel=4 ++ '[' -n operator-testing/some-name-2023-12-29-11:38:59-full ']' ++ echo -n s3://operator-testing/some-name-2023-12-29-11:38:59-full + xbcloud get --parallel=4 --curl-retriable-errors=7 s3://operator-testing/some-name-2023-12-29-11:38:59-full 231229 11:49:02 xbcloud: Successfully connected. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_checkpoints.00000000000000000000. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/undo_002.00000000000000000000. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/sys/sys_config.ibd.00000000000000000000. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/variables_by_thr_192.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_checkpoints.00000000000000000000, size 190 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/variables_by_thr_192.sdi.00000000000000000000, size 4195 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/variables_by_thr_192.sdi.00000000000000000001. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_checkpoints.00000000000000000001. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/variables_by_thr_192.sdi.00000000000000000001, size 57 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_checkpoints.00000000000000000001, size 36 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_147.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/hosts_155.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/sys/sys_config.ibd.00000000000000000000, size 114740 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/sys/sys_config.ibd.00000000000000000001. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/hosts_155.sdi.00000000000000000000, size 5368 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_147.sdi.00000000000000000000, size 15009 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/sys/sys_config.ibd.00000000000000000001, size 32 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_147.sdi.00000000000000000001. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/hosts_155.sdi.00000000000000000001. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_145.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_147.sdi.00000000000000000001, size 57 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/hosts_155.sdi.00000000000000000001, size 46 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_145.sdi.00000000000000000000, size 15856 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_145.sdi.00000000000000000001. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_144.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_127.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_145.sdi.00000000000000000001, size 57 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_info.00000000000000000000. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_127.sdi.00000000000000000000, size 6919 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_144.sdi.00000000000000000000, size 16712 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_127.sdi.00000000000000000001. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_144.sdi.00000000000000000001. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_info.00000000000000000000, size 852 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_144.sdi.00000000000000000001, size 57 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_127.sdi.00000000000000000001, size 57 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_info.00000000000000000001. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_139.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_his_94.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_info.00000000000000000001, size 29 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/undo_001.00000000000000000000. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/undo_002.00000000000000000000, size 10485802 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_139.sdi.00000000000000000000, size 7946 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_his_94.sdi.00000000000000000000, size 17482 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_his_94.sdi.00000000000000000001. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_139.sdi.00000000000000000001. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/undo_002.00000000000000000001. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_his_94.sdi.00000000000000000001, size 56 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_139.sdi.00000000000000000001, size 57 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_131.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_135.sdi.00000000000000000000. 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_131.sdi.00000000000000000000, size 26360 231229 11:49:03 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_135.sdi.00000000000000000000, size 25484 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_135.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_131.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_131.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_135.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_134.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_177.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/undo_001.00000000000000000000, size 10485802 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_177.sdi.00000000000000000000, size 27459 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_177.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/undo_001.00000000000000000001. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_134.sdi.00000000000000000000, size 26332 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/undo_002.00000000000000000001, size 6291498 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_177.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_134.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/undo_002.00000000000000000002. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/variables_info_195.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_134.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/undo_002.00000000000000000002, size 22 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/global_status_190.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_141.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/variables_info_195.sdi.00000000000000000000, size 8057 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_141.sdi.00000000000000000000, size 21546 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/variables_info_195.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_141.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/global_status_190.sdi.00000000000000000000, size 3314 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/global_status_190.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_141.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/variables_info_195.sdi.00000000000000000001, size 55 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/global_status_190.sdi.00000000000000000001, size 54 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/users_153.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_180.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/persisted_variab_196.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/undo_001.00000000000000000001, size 6291498 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_180.sdi.00000000000000000000, size 4700 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/persisted_variab_196.sdi.00000000000000000000, size 3324 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/persisted_variab_196.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_180.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/undo_001.00000000000000000002. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/users_153.sdi.00000000000000000000, size 5367 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/undo_001.00000000000000000002, size 22 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_180.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/persisted_variab_196.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/users_153.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_binlog_info.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_128.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_126.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/users_153.sdi.00000000000000000001, size 46 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_binlog_info.00000000000000000000, size 116 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_128.sdi.00000000000000000000, size 39008 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_128.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_binlog_info.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_130.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_126.sdi.00000000000000000000, size 7766 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_128.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_binlog_info.00000000000000000001, size 36 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_126.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/socket_summary_b_158.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_cur_93.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_130.sdi.00000000000000000000, size 38567 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_126.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_130.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_137.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_cur_93.sdi.00000000000000000000, size 17482 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/socket_summary_b_158.sdi.00000000000000000000, size 20584 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_130.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_cur_93.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/socket_summary_b_158.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/table_io_waits_s_117.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_cur_93.sdi.00000000000000000001, size 56 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/socket_summary_b_158.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_124.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_123.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/table_io_waits_s_117.sdi.00000000000000000000, size 32057 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_124.sdi.00000000000000000000, size 8620 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_123.sdi.00000000000000000000, size 7794 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_123.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_124.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/table_io_waits_s_117.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_123.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_124.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/table_io_waits_s_117.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_137.sdi.00000000000000000000, size 31445 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_137.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_hi_122.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_132.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/slow_log.CSM.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/slow_log.CSM.00000000000000000000, size 87 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_137.sdi.00000000000000000001, size 57 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/slow_log.CSM.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/accounts_154.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_hi_122.sdi.00000000000000000000, size 11369 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_132.sdi.00000000000000000000, size 27187 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/slow_log.CSM.00000000000000000001, size 32 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_132.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_hi_122.sdi.00000000000000000001. 231229 11:49:04 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/general_log_225.sdi.00000000000000000000. 231229 11:49:04 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/accounts_154.sdi.00000000000000000000, size 6225 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_132.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_hi_122.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/accounts_154.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_his_95.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_143.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/general_log_225.sdi.00000000000000000000, size 6053 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/accounts_154.sdi.00000000000000000001, size 49 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_143.sdi.00000000000000000000, size 15885 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_143.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/general_log_225.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/binary_log_trans_198.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_his_95.sdi.00000000000000000000, size 17041 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_143.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/binary_log_trans_198.sdi.00000000000000000000, size 13298 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/general_log_225.sdi.00000000000000000001, size 39 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/binary_log_trans_198.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_his_95.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/general_log.CSM.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/slow_log_226.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/binary_log_trans_198.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/general_log.CSM.00000000000000000000, size 90 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_his_95.sdi.00000000000000000001, size 56 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/general_log.CSM.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_129.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/cond_instances_91.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/slow_log_226.sdi.00000000000000000000, size 12583 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/general_log.CSM.00000000000000000001, size 35 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/cond_instances_91.sdi.00000000000000000000, size 4051 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_129.sdi.00000000000000000000, size 39008 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/cond_instances_91.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_129.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/slow_log_226.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_149.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/cond_instances_91.sdi.00000000000000000001, size 54 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_129.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/slow_log_226.sdi.00000000000000000001, size 36 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/table_lock_waits_118.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_142.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_148.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_149.sdi.00000000000000000000, size 7763 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_149.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_142.sdi.00000000000000000000, size 21105 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_148.sdi.00000000000000000000, size 7762 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/table_lock_waits_118.sdi.00000000000000000000, size 56730 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_149.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_148.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_142.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/table_lock_waits_118.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_hi_121.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_148.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_142.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/table_lock_waits_118.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/general_log.CSV.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_100.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/file_summary_by__104.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_hi_121.sdi.00000000000000000000, size 11810 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/general_log.CSV.00000000000000000000, size 35 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_hi_121.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/ibdata1.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_hi_121.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/myApp/myApp.ibd.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/file_summary_by__104.sdi.00000000000000000000, size 25721 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_100.sdi.00000000000000000000, size 8619 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/file_summary_by__104.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_100.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_100.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/file_summary_by__104.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_grou_172.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/slow_log.CSV.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/myApp/myApp.ibd.00000000000000000000, size 114737 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/slow_log.CSV.00000000000000000000, size 32 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/myApp/myApp.ibd.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql.ibd.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_grou_172.sdi.00000000000000000000, size 9587 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/ibdata1.00000000000000000000, size 10485801 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/myApp/myApp.ibd.00000000000000000001, size 29 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_grou_172.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/ibdata1.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/wsrep_cluster.ibd.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_grou_172.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/pxc_cluster_view_203.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/ibdata1.00000000000000000001, size 2097193 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/pxc_cluster_view_203.sdi.00000000000000000000, size 5831 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/wsrep_cluster.ibd.00000000000000000000, size 114745 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/pxc_cluster_view_203.sdi.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/wsrep_cluster.ibd.00000000000000000001. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/ibdata1.00000000000000000002. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/pxc_cluster_view_203.sdi.00000000000000000001, size 57 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/ibdata1.00000000000000000002, size 21 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/wsrep_cluster.ibd.00000000000000000001, size 37 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/binlog.index.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/session_account__160.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/error_log_92.sdi.00000000000000000000. 231229 11:49:05 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/error_log_92.sdi.00000000000000000000, size 7730 231229 11:49:05 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/error_log_92.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql.ibd.00000000000000000000, size 10485803 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/session_account__160.sdi.00000000000000000000, size 4802 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/error_log_92.sdi.00000000000000000001, size 49 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/binlog.index.00000000000000000000, size 62 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/session_account__160.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/binlog.index.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql.ibd.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_152.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/session_account__160.sdi.00000000000000000001, size 57 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/binlog.index.00000000000000000001, size 26 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/wsrep_streaming_log.ibd.00000000000000000000. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/table_handles_167.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_152.sdi.00000000000000000000, size 6921 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_152.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/wsrep_streaming_log.ibd.00000000000000000000, size 114751 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_152.sdi.00000000000000000001, size 57 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/wsrep_streaming_log.ibd.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/ib_buffer_pool.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/table_handles_167.sdi.00000000000000000000, size 9745 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/wsrep_streaming_log.ibd.00000000000000000001, size 43 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/ib_buffer_pool.00000000000000000000, size 3808 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/ib_buffer_pool.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/table_handles_167.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_99.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/ib_buffer_pool.00000000000000000001, size 28 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/backup-my.cnf.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_99.sdi.00000000000000000000, size 7763 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql.ibd.00000000000000000001, size 10485803 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/table_handles_167.sdi.00000000000000000001, size 54 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/backup-my.cnf.00000000000000000000, size 501 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/backup-my.cnf.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_99.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql.ibd.00000000000000000002. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_151.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/backup-my.cnf.00000000000000000001, size 27 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_99.sdi.00000000000000000001, size 56 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_151.sdi.00000000000000000000, size 7793 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_151.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/wsrep_cluster_members.ibd.00000000000000000000. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/data_locks_169.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_151.sdi.00000000000000000001, size 57 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/data_locks_169.sdi.00000000000000000000, size 16203 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/data_locks_169.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_101.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/wsrep_cluster_members.ibd.00000000000000000000, size 114753 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/data_locks_169.sdi.00000000000000000001, size 51 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_101.sdi.00000000000000000000, size 6918 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_101.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql/wsrep_cluster_members.ibd.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_logfile.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_101.sdi.00000000000000000001, size 57 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql/wsrep_cluster_members.ibd.00000000000000000001, size 45 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_96.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/keyring_componen_202.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_logfile.00000000000000000000, size 3124 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_logfile.00000000000000000001. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql.ibd.00000000000000000002, size 10485803 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_96.sdi.00000000000000000000, size 7764 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/keyring_componen_202.sdi.00000000000000000000, size 9608 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_logfile.00000000000000000001, size 32 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/keyring_componen_202.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_96.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/mysql.ibd.00000000000000000003. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_97.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/keyring_componen_202.sdi.00000000000000000001, size 57 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/mysql.ibd.00000000000000000003, size 23 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_96.sdi.00000000000000000001, size 56 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_98.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_166.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_threads_115.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_97.sdi.00000000000000000000, size 8094 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_166.sdi.00000000000000000000, size 11786 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_166.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_97.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_98.sdi.00000000000000000000, size 7792 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_threads_115.sdi.00000000000000000000, size 6986 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_166.sdi.00000000000000000001, size 57 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_97.sdi.00000000000000000001, size 56 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_threads_115.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_98.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/file_summary_by__103.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/malloc_stats_tot_200.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_threads_115.sdi.00000000000000000001, size 54 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/global_variables_193.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_waits_sum_98.sdi.00000000000000000001, size 56 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/malloc_stats_tot_200.sdi.00000000000000000000, size 5799 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/file_summary_by__103.sdi.00000000000000000000, size 20566 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/malloc_stats_tot_200.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/file_summary_by__103.sdi.00000000000000000001. 231229 11:49:06 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_146.sdi.00000000000000000000. 231229 11:49:06 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/global_variables_193.sdi.00000000000000000000, size 3321 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/malloc_stats_tot_200.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/global_variables_193.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/metadata_locks_168.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/file_summary_by__103.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_146.sdi.00000000000000000000, size 15857 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/global_variables_193.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_146.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/host_cache_105.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/keyring_keys_161.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/metadata_locks_168.sdi.00000000000000000000, size 11994 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_146.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/keyring_keys_161.sdi.00000000000000000000, size 4758 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/keyring_keys_161.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/metadata_locks_168.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_163.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/host_cache_105.sdi.00000000000000000000, size 25403 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/keyring_keys_161.sdi.00000000000000000001, size 53 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/metadata_locks_168.sdi.00000000000000000001, size 55 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/host_cache_105.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_133.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/log_status_183.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_163.sdi.00000000000000000000, size 12642 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/host_cache_105.sdi.00000000000000000001, size 51 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/log_status_183.sdi.00000000000000000000, size 4376 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/log_status_183.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_163.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_136.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_133.sdi.00000000000000000000, size 26331 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/log_status_183.sdi.00000000000000000001, size 51 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_163.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_133.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_grou_178.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/malloc_stats_201.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_136.sdi.00000000000000000000, size 32818 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_133.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_grou_178.sdi.00000000000000000000, size 12455 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_grou_178.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_136.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/status_by_host_187.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/malloc_stats_201.sdi.00000000000000000000, size 5014 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_grou_178.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_136.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/status_by_host_187.sdi.00000000000000000000, size 4162 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/status_by_host_187.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/malloc_stats_201.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_164.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_165.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/malloc_stats_201.sdi.00000000000000000001, size 53 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_asyn_182.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_164.sdi.00000000000000000000, size 11787 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/status_by_host_187.sdi.00000000000000000001, size 55 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_165.sdi.00000000000000000000, size 11815 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_asyn_182.sdi.00000000000000000000, size 5416 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_164.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_asyn_182.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_165.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_tablespaces.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_164.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_b_165.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_asyn_182.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_tablespaces.00000000000000000000, size 95 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/xtrabackup_tablespaces.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_cu_120.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_consumers_112.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/objects_summary__107.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/xtrabackup_tablespaces.00000000000000000001, size 36 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_g_162.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_cu_120.sdi.00000000000000000000, size 11810 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/objects_summary__107.sdi.00000000000000000000, size 7921 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_consumers_112.sdi.00000000000000000000, size 3359 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/objects_summary__107.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_consumers_112.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_cu_120.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_g_162.sdi.00000000000000000000, size 10939 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/objects_summary__107.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_consumers_112.sdi.00000000000000000001, size 56 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_cu_120.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_g_162.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/prepared_stateme_184.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/binlog.000010.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/performance_time_108.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/memory_summary_g_162.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/binlog.000010.00000000000000000000, size 244 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/binlog.000010.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/processlist_109.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/prepared_stateme_184.sdi.00000000000000000000, size 35609 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/performance_time_108.sdi.00000000000000000000, size 4392 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/binlog.000010.00000000000000000001, size 27 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/prepared_stateme_184.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/performance_time_108.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_asyn_181.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/processlist_109.sdi.00000000000000000000, size 10878 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/prepared_stateme_184.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/performance_time_108.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/processlist_109.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_174.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_175.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_asyn_181.sdi.00000000000000000000, size 7342 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/processlist_109.sdi.00000000000000000001, size 52 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_175.sdi.00000000000000000000, size 4985 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_175.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_asyn_181.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/threads_119.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_174.sdi.00000000000000000000, size 8324 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_asyn_181.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_175.sdi.00000000000000000001, size 57 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_174.sdi.00000000000000000001. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_176.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_179.sdi.00000000000000000000. 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/threads_119.sdi.00000000000000000000, size 23733 231229 11:49:07 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_174.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/threads_119.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_conn_171.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_179.sdi.00000000000000000000, size 6829 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/threads_119.sdi.00000000000000000001, size 48 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_176.sdi.00000000000000000000, size 19978 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_176.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_179.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/mutex_instances_106.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_conn_171.sdi.00000000000000000000, size 46320 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_176.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/mutex_instances_106.sdi.00000000000000000000, size 5234 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_appl_179.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/mutex_instances_106.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_conn_171.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/file_instances_102.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/rwlock_instances_110.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/mutex_instances_106.sdi.00000000000000000001, size 56 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_conn_171.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/file_instances_102.sdi.00000000000000000000, size 7640 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/rwlock_instances_110.sdi.00000000000000000000, size 6050 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/rwlock_instances_110.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/file_instances_102.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_conn_173.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/session_connect__159.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/file_instances_102.sdi.00000000000000000001, size 55 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/rwlock_instances_110.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_138.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_actors_111.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_conn_173.sdi.00000000000000000000, size 24447 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/session_connect__159.sdi.00000000000000000000, size 4794 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_actors_111.sdi.00000000000000000000, size 6254 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_actors_111.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/session_connect__159.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/replication_conn_173.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_138.sdi.00000000000000000000, size 6224 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/replication_conn_173.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_actors_111.sdi.00000000000000000001, size 53 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/session_connect__159.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_138.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/session_status_191.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_140.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/session_variable_194.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_statement_138.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/session_status_191.sdi.00000000000000000000, size 3316 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/session_status_191.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_instrument_113.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_140.sdi.00000000000000000000, size 21546 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/session_variable_194.sdi.00000000000000000000, size 3322 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/session_status_191.sdi.00000000000000000001, size 55 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/session_variable_194.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_140.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/socket_instances_156.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_instrument_113.sdi.00000000000000000000, size 8036 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_transacti_140.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/session_variable_194.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_instrument_113.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_objects_114.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/socket_summary_b_157.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/socket_instances_156.sdi.00000000000000000000, size 9209 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_instrument_113.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_objects_114.sdi.00000000000000000000, size 6502 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/setup_objects_114.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/socket_instances_156.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/status_by_accoun_186.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/socket_summary_b_157.sdi.00000000000000000000, size 21769 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/setup_objects_114.sdi.00000000000000000001, size 54 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/socket_instances_156.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/socket_summary_b_157.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/data_lock_waits_170.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/status_by_thread_188.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/status_by_accoun_186.sdi.00000000000000000000, size 5018 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/status_by_thread_188.sdi.00000000000000000000, size 4192 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/socket_summary_b_157.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/data_lock_waits_170.sdi.00000000000000000000, size 14214 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/status_by_accoun_186.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/status_by_thread_188.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/data_lock_waits_170.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/status_by_user_189.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/status_by_accoun_186.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/data_lock_waits_170.sdi.00000000000000000001, size 56 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/status_by_thread_188.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/table_io_waits_s_116.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_150.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/tls_channel_stat_199.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/status_by_user_189.sdi.00000000000000000000, size 4161 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/status_by_user_189.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/table_io_waits_s_116.sdi.00000000000000000000, size 32923 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_150.sdi.00000000000000000000, size 8617 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/tls_channel_stat_199.sdi.00000000000000000000, size 15992 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/status_by_user_189.sdi.00000000000000000001, size 55 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/tls_channel_stat_199.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_150.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/table_io_waits_s_116.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_125.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/tls_channel_stat_199.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/table_io_waits_s_116.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_errors_su_150.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_125.sdi.00000000000000000000, size 7765 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_125.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/user_defined_fun_197.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/user_variables_b_185.sdi.00000000000000000000. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/events_stages_su_125.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/user_variables_b_185.sdi.00000000000000000000, size 4199 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/user_defined_fun_197.sdi.00000000000000000000, size 5890 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/user_variables_b_185.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Downloading some-name-2023-12-29-11:38:59-full/performance_schema/user_defined_fun_197.sdi.00000000000000000001. 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/user_variables_b_185.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: [0] Download successfull some-name-2023-12-29-11:38:59-full/performance_schema/user_defined_fun_197.sdi.00000000000000000001, size 57 231229 11:49:08 xbcloud: Download completed. + set +o xtrace % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 14 100 14 0 0 1166 0 --:--:-- --:--:-- --:--:-- 1166 + xtrabackup --use-memory=100MB --prepare --rollback-prepared-trx --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_1JAQ 2023-12-29T11:49:08.991924-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --innodb_checksum_algorithm=crc32 --innodb_log_checksums=1 --innodb_data_file_path=ibdata1:12M:autoextend --innodb_log_file_size=50331648 --innodb_page_size=16384 --innodb_undo_directory=./ --innodb_undo_tablespaces=2 --server-id=24694912 --innodb_log_checksums=ON --innodb_redo_log_encrypt=0 --innodb_undo_log_encrypt=0 2023-12-29T11:49:08.992082-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --use-memory=100MB --prepare=1 --rollback-prepared-trx=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_1JAQ xtrabackup version 8.0.34-29 based on MySQL server 8.0.34 Linux (x86_64) (revision id: 5ba706ee) 2023-12-29T11:49:08.992130-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_1JAQ/ 2023-12-29T11:49:08.992226-00:00 0 [Note] [MY-011825] [Xtrabackup] This target seems to be not prepared yet. 2023-12-29T11:49:09.005723-00:00 0 [Note] [MY-011825] [Xtrabackup] xtrabackup_logfile detected: size=8388608, start_lsn=(32600564) 2023-12-29T11:49:09.006764-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2023-12-29T11:49:09.006793-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2023-12-29T11:49:09.006803-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2023-12-29T11:49:09.006848-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2023-12-29T11:49:09.006858-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2023-12-29T11:49:09.006871-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2023-12-29T11:49:09.007202-00:00 0 [Note] [MY-011825] [Xtrabackup] inititialize_service_handles suceeded 2023-12-29T11:49:09.007529-00:00 0 [Note] [MY-011825] [Xtrabackup] using the following InnoDB configuration for recovery: 2023-12-29T11:49:09.007553-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_home_dir = . 2023-12-29T11:49:09.007559-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_data_file_path = ibdata1:12M:autoextend 2023-12-29T11:49:09.007570-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_group_home_dir = . 2023-12-29T11:49:09.007576-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_files_in_group = 1 2023-12-29T11:49:09.007581-00:00 0 [Note] [MY-011825] [Xtrabackup] innodb_log_file_size = 8388608 2023-12-29T11:49:09.007595-00:00 0 [Note] [MY-011825] [Xtrabackup] Starting InnoDB instance for recovery. 2023-12-29T11:49:09.007607-00:00 0 [Note] [MY-011825] [Xtrabackup] Using 104857600 bytes for buffer pool (set by --use-memory parameter) 2023-12-29T11:49:09.007645-00:00 0 [Note] [MY-012932] [InnoDB] PUNCH HOLE support available 2023-12-29T11:49:09.007676-00:00 0 [Note] [MY-012944] [InnoDB] Uses event mutexes 2023-12-29T11:49:09.007693-00:00 0 [Note] [MY-012945] [InnoDB] GCC builtin __atomic_thread_fence() is used for memory barrier 2023-12-29T11:49:09.007703-00:00 0 [Note] [MY-012948] [InnoDB] Compressed tables use zlib 1.2.13 2023-12-29T11:49:09.008032-00:00 0 [Note] [MY-012951] [InnoDB] Using hardware accelerated crc32 and polynomial multiplication. 2023-12-29T11:49:09.008637-00:00 0 [Note] [MY-012203] [InnoDB] Directories to scan './' 2023-12-29T11:49:09.008692-00:00 0 [Note] [MY-012204] [InnoDB] Scanning './' 2023-12-29T11:49:09.017518-00:00 0 [Note] [MY-012208] [InnoDB] Completed space ID check of 8 files. 2023-12-29T11:49:09.018355-00:00 0 [Note] [MY-012955] [InnoDB] Initializing buffer pool, total size = 128.000000M, instances = 1, chunk size =128.000000M 2023-12-29T11:49:09.029669-00:00 0 [Note] [MY-012957] [InnoDB] Completed initialization of buffer pool 2023-12-29T11:49:09.033492-00:00 0 [Note] [MY-011952] [InnoDB] If the mysqld execution user is authorized, page cleaner thread priority can be changed. See the man page of setpriority(). 2023-12-29T11:49:09.089958-00:00 0 [Note] [MY-013883] [InnoDB] The latest found checkpoint is at lsn = 32600564 in redo log file ./#innodb_redo/#ib_redo0. 2023-12-29T11:49:09.090020-00:00 0 [Note] [MY-012560] [InnoDB] The log sequence number 32456618 in the system tablespace does not match the log sequence number 32600564 in the redo log files! 2023-12-29T11:49:09.090031-00:00 0 [Note] [MY-012551] [InnoDB] Database was not shutdown normally! 2023-12-29T11:49:09.090040-00:00 0 [Note] [MY-012552] [InnoDB] Starting crash recovery. 2023-12-29T11:49:09.090240-00:00 0 [Note] [MY-013086] [InnoDB] Starting to parse redo log at lsn = 32600097, whereas checkpoint_lsn = 32600564 and start_lsn = 32600064 2023-12-29T11:49:09.090265-00:00 0 [Note] [MY-012550] [InnoDB] Doing recovery: scanned up to log sequence number 32600590 2023-12-29T11:49:09.106990-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2023-12-29T11:49:09.131300-00:00 0 [Note] [MY-012532] [InnoDB] Applying a batch of 1 redo log records ... 2023-12-29T11:49:09.131464-00:00 0 [Note] [MY-012533] [InnoDB] 100% 2023-12-29T11:49:09.131489-00:00 0 [Note] [MY-012535] [InnoDB] Apply batch completed! 2023-12-29T11:49:09.231729-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2023-12-29T11:49:09.233363-00:00 0 [Note] [MY-013888] [InnoDB] Upgrading redo log: 1032M, LSN=32600590. 2023-12-29T11:49:09.233414-00:00 0 [Note] [MY-012968] [InnoDB] Starting to delete and rewrite redo log files. 2023-12-29T11:49:09.233488-00:00 0 [Note] [MY-011825] [InnoDB] Removing redo log file: ./#innodb_redo/#ib_redo0 2023-12-29T11:49:09.292275-00:00 0 [Note] [MY-011825] [InnoDB] Creating redo log file at ./#innodb_redo/#ib_redo0_tmp with file_id 0 with size 33554432 bytes 2023-12-29T11:49:09.295554-00:00 0 [Note] [MY-011825] [InnoDB] Renaming redo log file from ./#innodb_redo/#ib_redo0_tmp to ./#innodb_redo/#ib_redo0 2023-12-29T11:49:09.298223-00:00 0 [Note] [MY-012893] [InnoDB] New redo log files created, LSN=32601100 2023-12-29T11:49:09.298352-00:00 0 [Note] [MY-013083] [InnoDB] Log background threads are being started... 2023-12-29T11:49:09.320841-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_001'. 2023-12-29T11:49:09.322886-00:00 0 [Note] [MY-013252] [InnoDB] Using undo tablespace './undo_002'. 2023-12-29T11:49:09.325571-00:00 0 [Note] [MY-012910] [InnoDB] Opened 2 existing undo tablespaces. 2023-12-29T11:49:09.325684-00:00 0 [Note] [MY-011980] [InnoDB] GTID recovery trx_no: 5225 2023-12-29T11:49:09.386855-00:00 0 [Note] [MY-013776] [InnoDB] Parallel initialization of rseg complete 2023-12-29T11:49:09.386909-00:00 0 [Note] [MY-013777] [InnoDB] Time taken to initialize rseg using 4 thread: 61237 ms. 2023-12-29T11:49:09.387022-00:00 0 [Note] [MY-012923] [InnoDB] Creating shared tablespace for temporary tables 2023-12-29T11:49:09.387120-00:00 0 [Note] [MY-012265] [InnoDB] Setting file './ibtmp1' size to 12 MB. Physically writing the file full; Please wait ... 2023-12-29T11:49:09.419632-00:00 0 [Note] [MY-012266] [InnoDB] File './ibtmp1' size is now 12 MB. 2023-12-29T11:49:09.419913-00:00 0 [Note] [MY-013627] [InnoDB] Scanning temp tablespace dir:'./#innodb_temp/' 2023-12-29T11:49:09.441176-00:00 0 [Note] [MY-013018] [InnoDB] Created 128 and tracked 128 new rollback segment(s) in the temporary tablespace. 128 are now active. 2023-12-29T11:49:09.441611-00:00 0 [Note] [MY-012976] [InnoDB] 8.0.34 started; log sequence number 32601110 2023-12-29T11:49:09.442185-00:00 0 [Warning] [MY-012091] [InnoDB] Allocated tablespace ID 1 for sys/sys_config, old maximum was 0 2023-12-29T11:49:09.444511-00:00 0 [Note] [MY-011825] [Xtrabackup] Completed loading of 6 tablespaces into cache in 0.00282663 seconds 2023-12-29T11:49:09.466919-00:00 0 [Note] [MY-011825] [Xtrabackup] Time taken to build dictionary: 0.022352 seconds 2023-12-29T11:49:09.467718-00:00 0 [Note] [MY-011825] [Xtrabackup] Recovered WSREP position: 1410f97c-a63e-11ee-810f-1a4fefb65588:40 2023-12-29T11:49:09.467788-00:00 0 [Note] [MY-011825] [Xtrabackup] starting shutdown with innodb_fast_shutdown = 1 2023-12-29T11:49:09.467863-00:00 0 [Note] [MY-012330] [InnoDB] FTS optimize thread exiting. 2023-12-29T11:49:10.467677-00:00 0 [Note] [MY-013072] [InnoDB] Starting shutdown... 2023-12-29T11:49:10.474592-00:00 0 [Note] [MY-013084] [InnoDB] Log background threads are being closed... 2023-12-29T11:49:10.496235-00:00 0 [Note] [MY-012980] [InnoDB] Shutdown completed; log sequence number 32601110 2023-12-29T11:49:10.501048-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + xtrabackup --defaults-group=mysqld --datadir=/datadir --move-back --force-non-empty-directories --keyring-vault-config=/etc/mysql/vault-keyring-secret/keyring_vault.conf --early-plugin-load=keyring_vault.so --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_1JAQ 2023-12-29T11:49:10.521994-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized server arguments: --defaults_group=mysqld --datadir=/datadir 2023-12-29T11:49:10.522133-00:00 0 [Note] [MY-011825] [Xtrabackup] recognized client arguments: --move-back=1 --force-non-empty-directories=1 --xtrabackup-plugin-dir=/usr/lib64/xtrabackup/plugin --target-dir=/datadir/pxc_sst_1JAQ xtrabackup version 8.0.34-29 based on MySQL server 8.0.34 Linux (x86_64) (revision id: 5ba706ee) 2023-12-29T11:49:10.522178-00:00 0 [Note] [MY-011825] [Xtrabackup] cd to /datadir/pxc_sst_1JAQ/ 2023-12-29T11:49:10.523063-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_001 to /datadir/undo_001 2023-12-29T11:49:10.523133-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_001 to /datadir/undo_001 2023-12-29T11:49:10.523167-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving undo_002 to /datadir/undo_002 2023-12-29T11:49:10.523201-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file undo_002 to /datadir/undo_002 2023-12-29T11:49:10.523407-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving ibdata1 to /datadir/ibdata1 2023-12-29T11:49:10.523451-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file ibdata1 to /datadir/ibdata1 2023-12-29T11:49:10.523770-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.000010 to /datadir//binlog.000010 2023-12-29T11:49:10.523812-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.000010 to /datadir//binlog.000010 2023-12-29T11:49:10.523902-00:00 0 [Note] [MY-011825] [Xtrabackup] Moving binlog.index to /datadir//binlog.index 2023-12-29T11:49:10.523938-00:00 0 [Note] [MY-011825] [Xtrabackup] Done: Moving file binlog.index to /datadir//binlog.index 2023-12-29T11:49:10.524589-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2023-12-29T11:49:10.524659-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_keys_161.sdi to /datadir/performance_schema/keyring_keys_161.sdi 2023-12-29T11:49:10.524698-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2023-12-29T11:49:10.524729-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_131.sdi to /datadir/performance_schema/events_statement_131.sdi 2023-12-29T11:49:10.524771-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2023-12-29T11:49:10.524803-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/error_log_92.sdi to /datadir/performance_schema/error_log_92.sdi 2023-12-29T11:49:10.524857-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2023-12-29T11:49:10.524912-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_181.sdi to /datadir/performance_schema/replication_asyn_181.sdi 2023-12-29T11:49:10.524944-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2023-12-29T11:49:10.524973-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_consumers_112.sdi to /datadir/performance_schema/setup_consumers_112.sdi 2023-12-29T11:49:10.525012-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2023-12-29T11:49:10.525051-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_123.sdi to /datadir/performance_schema/events_stages_su_123.sdi 2023-12-29T11:49:10.525090-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2023-12-29T11:49:10.525130-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_176.sdi to /datadir/performance_schema/replication_appl_176.sdi 2023-12-29T11:49:10.525167-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2023-12-29T11:49:10.525217-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_variable_194.sdi to /datadir/performance_schema/session_variable_194.sdi 2023-12-29T11:49:10.525258-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2023-12-29T11:49:10.525288-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_124.sdi to /datadir/performance_schema/events_stages_su_124.sdi 2023-12-29T11:49:10.525316-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2023-12-29T11:49:10.525346-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_134.sdi to /datadir/performance_schema/events_statement_134.sdi 2023-12-29T11:49:10.525373-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2023-12-29T11:49:10.525423-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_locks_169.sdi to /datadir/performance_schema/data_locks_169.sdi 2023-12-29T11:49:10.525466-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2023-12-29T11:49:10.525504-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_handles_167.sdi to /datadir/performance_schema/table_handles_167.sdi 2023-12-29T11:49:10.525542-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2023-12-29T11:49:10.525578-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_101.sdi to /datadir/performance_schema/events_waits_sum_101.sdi 2023-12-29T11:49:10.525609-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2023-12-29T11:49:10.525644-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_info_195.sdi to /datadir/performance_schema/variables_info_195.sdi 2023-12-29T11:49:10.525680-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2023-12-29T11:49:10.525713-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__104.sdi to /datadir/performance_schema/file_summary_by__104.sdi 2023-12-29T11:49:10.525750-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2023-12-29T11:49:10.525789-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_asyn_182.sdi to /datadir/performance_schema/replication_asyn_182.sdi 2023-12-29T11:49:10.525828-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2023-12-29T11:49:10.525865-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_97.sdi to /datadir/performance_schema/events_waits_sum_97.sdi 2023-12-29T11:49:10.525897-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2023-12-29T11:49:10.525933-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_summary_by__103.sdi to /datadir/performance_schema/file_summary_by__103.sdi 2023-12-29T11:49:10.525973-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2023-12-29T11:49:10.526013-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_152.sdi to /datadir/performance_schema/events_errors_su_152.sdi 2023-12-29T11:49:10.526049-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2023-12-29T11:49:10.526087-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/tls_channel_stat_199.sdi to /datadir/performance_schema/tls_channel_stat_199.sdi 2023-12-29T11:49:10.526132-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2023-12-29T11:49:10.526171-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_139.sdi to /datadir/performance_schema/events_statement_139.sdi 2023-12-29T11:49:10.526223-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2023-12-29T11:49:10.526263-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_instrument_113.sdi to /datadir/performance_schema/setup_instrument_113.sdi 2023-12-29T11:49:10.526297-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2023-12-29T11:49:10.526338-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/binary_log_trans_198.sdi to /datadir/performance_schema/binary_log_trans_198.sdi 2023-12-29T11:49:10.526374-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2023-12-29T11:49:10.526423-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_variables_b_185.sdi to /datadir/performance_schema/user_variables_b_185.sdi 2023-12-29T11:49:10.526457-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2023-12-29T11:49:10.526496-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_150.sdi to /datadir/performance_schema/events_errors_su_150.sdi 2023-12-29T11:49:10.526532-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2023-12-29T11:49:10.526568-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_158.sdi to /datadir/performance_schema/socket_summary_b_158.sdi 2023-12-29T11:49:10.526599-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2023-12-29T11:49:10.526638-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_165.sdi to /datadir/performance_schema/memory_summary_b_165.sdi 2023-12-29T11:49:10.526674-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2023-12-29T11:49:10.526711-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/cond_instances_91.sdi to /datadir/performance_schema/cond_instances_91.sdi 2023-12-29T11:49:10.526751-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2023-12-29T11:49:10.526787-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_128.sdi to /datadir/performance_schema/events_statement_128.sdi 2023-12-29T11:49:10.526826-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2023-12-29T11:49:10.526862-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_137.sdi to /datadir/performance_schema/events_statement_137.sdi 2023-12-29T11:49:10.526898-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2023-12-29T11:49:10.526935-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_164.sdi to /datadir/performance_schema/memory_summary_b_164.sdi 2023-12-29T11:49:10.526970-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2023-12-29T11:49:10.527002-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/accounts_154.sdi to /datadir/performance_schema/accounts_154.sdi 2023-12-29T11:49:10.527046-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2023-12-29T11:49:10.527084-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_163.sdi to /datadir/performance_schema/memory_summary_b_163.sdi 2023-12-29T11:49:10.527118-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2023-12-29T11:49:10.527155-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/persisted_variab_196.sdi to /datadir/performance_schema/persisted_variab_196.sdi 2023-12-29T11:49:10.527200-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2023-12-29T11:49:10.527239-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_145.sdi to /datadir/performance_schema/events_transacti_145.sdi 2023-12-29T11:49:10.527274-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2023-12-29T11:49:10.527312-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_summary_b_157.sdi to /datadir/performance_schema/socket_summary_b_157.sdi 2023-12-29T11:49:10.527352-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2023-12-29T11:49:10.527388-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_172.sdi to /datadir/performance_schema/replication_grou_172.sdi 2023-12-29T11:49:10.527453-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2023-12-29T11:49:10.527491-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_201.sdi to /datadir/performance_schema/malloc_stats_201.sdi 2023-12-29T11:49:10.527531-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2023-12-29T11:49:10.527570-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_126.sdi to /datadir/performance_schema/events_stages_su_126.sdi 2023-12-29T11:49:10.527609-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2023-12-29T11:49:10.527649-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_objects_114.sdi to /datadir/performance_schema/setup_objects_114.sdi 2023-12-29T11:49:10.527691-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2023-12-29T11:49:10.527730-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_143.sdi to /datadir/performance_schema/events_transacti_143.sdi 2023-12-29T11:49:10.527767-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2023-12-29T11:49:10.527806-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/keyring_componen_202.sdi to /datadir/performance_schema/keyring_componen_202.sdi 2023-12-29T11:49:10.527843-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2023-12-29T11:49:10.527877-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/user_defined_fun_197.sdi to /datadir/performance_schema/user_defined_fun_197.sdi 2023-12-29T11:49:10.527912-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2023-12-29T11:49:10.527953-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/metadata_locks_168.sdi to /datadir/performance_schema/metadata_locks_168.sdi 2023-12-29T11:49:10.527995-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2023-12-29T11:49:10.528032-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_account__160.sdi to /datadir/performance_schema/session_account__160.sdi 2023-12-29T11:49:10.528067-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2023-12-29T11:49:10.528102-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_136.sdi to /datadir/performance_schema/events_statement_136.sdi 2023-12-29T11:49:10.528137-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2023-12-29T11:49:10.528181-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_125.sdi to /datadir/performance_schema/events_stages_su_125.sdi 2023-12-29T11:49:10.528224-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2023-12-29T11:49:10.528265-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_130.sdi to /datadir/performance_schema/events_statement_130.sdi 2023-12-29T11:49:10.528304-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2023-12-29T11:49:10.528341-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_thread_188.sdi to /datadir/performance_schema/status_by_thread_188.sdi 2023-12-29T11:49:10.528379-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2023-12-29T11:49:10.528429-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_94.sdi to /datadir/performance_schema/events_waits_his_94.sdi 2023-12-29T11:49:10.528467-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2023-12-29T11:49:10.528506-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_threads_115.sdi to /datadir/performance_schema/setup_threads_115.sdi 2023-12-29T11:49:10.528541-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2023-12-29T11:49:10.528581-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_146.sdi to /datadir/performance_schema/events_transacti_146.sdi 2023-12-29T11:49:10.528617-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2023-12-29T11:49:10.528652-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/file_instances_102.sdi to /datadir/performance_schema/file_instances_102.sdi 2023-12-29T11:49:10.528689-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2023-12-29T11:49:10.528727-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_user_189.sdi to /datadir/performance_schema/status_by_user_189.sdi 2023-12-29T11:49:10.528766-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2023-12-29T11:49:10.528804-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_144.sdi to /datadir/performance_schema/events_transacti_144.sdi 2023-12-29T11:49:10.528840-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2023-12-29T11:49:10.528881-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_149.sdi to /datadir/performance_schema/events_errors_su_149.sdi 2023-12-29T11:49:10.528918-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2023-12-29T11:49:10.528957-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/mutex_instances_106.sdi to /datadir/performance_schema/mutex_instances_106.sdi 2023-12-29T11:49:10.528996-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2023-12-29T11:49:10.529036-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_174.sdi to /datadir/performance_schema/replication_appl_174.sdi 2023-12-29T11:49:10.529072-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2023-12-29T11:49:10.529105-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/processlist_109.sdi to /datadir/performance_schema/processlist_109.sdi 2023-12-29T11:49:10.529143-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2023-12-29T11:49:10.529187-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/hosts_155.sdi to /datadir/performance_schema/hosts_155.sdi 2023-12-29T11:49:10.529224-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2023-12-29T11:49:10.529263-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/socket_instances_156.sdi to /datadir/performance_schema/socket_instances_156.sdi 2023-12-29T11:49:10.529297-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2023-12-29T11:49:10.529337-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_98.sdi to /datadir/performance_schema/events_waits_sum_98.sdi 2023-12-29T11:49:10.529376-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2023-12-29T11:49:10.529458-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_142.sdi to /datadir/performance_schema/events_transacti_142.sdi 2023-12-29T11:49:10.529501-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2023-12-29T11:49:10.529535-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_122.sdi to /datadir/performance_schema/events_stages_hi_122.sdi 2023-12-29T11:49:10.529573-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2023-12-29T11:49:10.529613-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/performance_time_108.sdi to /datadir/performance_schema/performance_time_108.sdi 2023-12-29T11:49:10.529651-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2023-12-29T11:49:10.529690-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/prepared_stateme_184.sdi to /datadir/performance_schema/prepared_stateme_184.sdi 2023-12-29T11:49:10.529731-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2023-12-29T11:49:10.529768-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_135.sdi to /datadir/performance_schema/events_statement_135.sdi 2023-12-29T11:49:10.529808-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2023-12-29T11:49:10.529845-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/threads_119.sdi to /datadir/performance_schema/threads_119.sdi 2023-12-29T11:49:10.529888-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2023-12-29T11:49:10.529930-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_grou_178.sdi to /datadir/performance_schema/replication_grou_178.sdi 2023-12-29T11:49:10.529969-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2023-12-29T11:49:10.530010-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/rwlock_instances_110.sdi to /datadir/performance_schema/rwlock_instances_110.sdi 2023-12-29T11:49:10.530048-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2023-12-29T11:49:10.530086-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_179.sdi to /datadir/performance_schema/replication_appl_179.sdi 2023-12-29T11:49:10.530122-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2023-12-29T11:49:10.530162-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_180.sdi to /datadir/performance_schema/replication_appl_180.sdi 2023-12-29T11:49:10.530205-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2023-12-29T11:49:10.530241-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_status_190.sdi to /datadir/performance_schema/global_status_190.sdi 2023-12-29T11:49:10.530278-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2023-12-29T11:49:10.530316-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_hi_121.sdi to /datadir/performance_schema/events_stages_hi_121.sdi 2023-12-29T11:49:10.530352-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2023-12-29T11:49:10.530440-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_141.sdi to /datadir/performance_schema/events_transacti_141.sdi 2023-12-29T11:49:10.530484-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2023-12-29T11:49:10.530528-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/objects_summary__107.sdi to /datadir/performance_schema/objects_summary__107.sdi 2023-12-29T11:49:10.530564-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2023-12-29T11:49:10.530603-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_lock_waits_118.sdi to /datadir/performance_schema/table_lock_waits_118.sdi 2023-12-29T11:49:10.530638-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2023-12-29T11:49:10.530676-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_116.sdi to /datadir/performance_schema/table_io_waits_s_116.sdi 2023-12-29T11:49:10.530716-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2023-12-29T11:49:10.530753-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_138.sdi to /datadir/performance_schema/events_statement_138.sdi 2023-12-29T11:49:10.530790-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2023-12-29T11:49:10.530829-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_g_162.sdi to /datadir/performance_schema/memory_summary_g_162.sdi 2023-12-29T11:49:10.530872-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2023-12-29T11:49:10.530913-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_129.sdi to /datadir/performance_schema/events_statement_129.sdi 2023-12-29T11:49:10.530946-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2023-12-29T11:49:10.530988-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_connect__159.sdi to /datadir/performance_schema/session_connect__159.sdi 2023-12-29T11:49:10.531026-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2023-12-29T11:49:10.531060-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_96.sdi to /datadir/performance_schema/events_waits_sum_96.sdi 2023-12-29T11:49:10.531099-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2023-12-29T11:49:10.531138-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_host_187.sdi to /datadir/performance_schema/status_by_host_187.sdi 2023-12-29T11:49:10.531183-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2023-12-29T11:49:10.531223-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/global_variables_193.sdi to /datadir/performance_schema/global_variables_193.sdi 2023-12-29T11:49:10.531259-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2023-12-29T11:49:10.531300-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_177.sdi to /datadir/performance_schema/replication_appl_177.sdi 2023-12-29T11:49:10.531332-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2023-12-29T11:49:10.531374-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/memory_summary_b_166.sdi to /datadir/performance_schema/memory_summary_b_166.sdi 2023-12-29T11:49:10.531429-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2023-12-29T11:49:10.531475-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/session_status_191.sdi to /datadir/performance_schema/session_status_191.sdi 2023-12-29T11:49:10.531512-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2023-12-29T11:49:10.531553-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_99.sdi to /datadir/performance_schema/events_waits_sum_99.sdi 2023-12-29T11:49:10.531590-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2023-12-29T11:49:10.531627-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_cur_93.sdi to /datadir/performance_schema/events_waits_cur_93.sdi 2023-12-29T11:49:10.531666-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2023-12-29T11:49:10.531843-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_sum_100.sdi to /datadir/performance_schema/events_waits_sum_100.sdi 2023-12-29T11:49:10.531923-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2023-12-29T11:49:10.531984-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_cu_120.sdi to /datadir/performance_schema/events_stages_cu_120.sdi 2023-12-29T11:49:10.532027-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2023-12-29T11:49:10.532069-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/malloc_stats_tot_200.sdi to /datadir/performance_schema/malloc_stats_tot_200.sdi 2023-12-29T11:49:10.532108-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2023-12-29T11:49:10.532147-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/setup_actors_111.sdi to /datadir/performance_schema/setup_actors_111.sdi 2023-12-29T11:49:10.532188-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2023-12-29T11:49:10.532248-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/log_status_183.sdi to /datadir/performance_schema/log_status_183.sdi 2023-12-29T11:49:10.532293-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2023-12-29T11:49:10.532336-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_140.sdi to /datadir/performance_schema/events_transacti_140.sdi 2023-12-29T11:49:10.532375-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2023-12-29T11:49:10.532453-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/pxc_cluster_view_203.sdi to /datadir/performance_schema/pxc_cluster_view_203.sdi 2023-12-29T11:49:10.532494-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2023-12-29T11:49:10.532533-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/table_io_waits_s_117.sdi to /datadir/performance_schema/table_io_waits_s_117.sdi 2023-12-29T11:49:10.532574-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2023-12-29T11:49:10.532619-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_171.sdi to /datadir/performance_schema/replication_conn_171.sdi 2023-12-29T11:49:10.532656-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2023-12-29T11:49:10.532698-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_appl_175.sdi to /datadir/performance_schema/replication_appl_175.sdi 2023-12-29T11:49:10.532737-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2023-12-29T11:49:10.532782-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_transacti_147.sdi to /datadir/performance_schema/events_transacti_147.sdi 2023-12-29T11:49:10.532822-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2023-12-29T11:49:10.532864-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/users_153.sdi to /datadir/performance_schema/users_153.sdi 2023-12-29T11:49:10.532904-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2023-12-29T11:49:10.532945-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/status_by_accoun_186.sdi to /datadir/performance_schema/status_by_accoun_186.sdi 2023-12-29T11:49:10.532979-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2023-12-29T11:49:10.533020-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_133.sdi to /datadir/performance_schema/events_statement_133.sdi 2023-12-29T11:49:10.533067-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2023-12-29T11:49:10.533108-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_151.sdi to /datadir/performance_schema/events_errors_su_151.sdi 2023-12-29T11:49:10.533145-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2023-12-29T11:49:10.533189-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_statement_132.sdi to /datadir/performance_schema/events_statement_132.sdi 2023-12-29T11:49:10.533237-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2023-12-29T11:49:10.533275-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_stages_su_127.sdi to /datadir/performance_schema/events_stages_su_127.sdi 2023-12-29T11:49:10.533313-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2023-12-29T11:49:10.533354-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/host_cache_105.sdi to /datadir/performance_schema/host_cache_105.sdi 2023-12-29T11:49:10.533389-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2023-12-29T11:49:10.533439-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/data_lock_waits_170.sdi to /datadir/performance_schema/data_lock_waits_170.sdi 2023-12-29T11:49:10.533479-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2023-12-29T11:49:10.533521-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_errors_su_148.sdi to /datadir/performance_schema/events_errors_su_148.sdi 2023-12-29T11:49:10.533558-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2023-12-29T11:49:10.533600-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/events_waits_his_95.sdi to /datadir/performance_schema/events_waits_his_95.sdi 2023-12-29T11:49:10.533640-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2023-12-29T11:49:10.533681-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/variables_by_thr_192.sdi to /datadir/performance_schema/variables_by_thr_192.sdi 2023-12-29T11:49:10.533721-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2023-12-29T11:49:10.533763-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./performance_schema/replication_conn_173.sdi to /datadir/performance_schema/replication_conn_173.sdi 2023-12-29T11:49:10.534005-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2023-12-29T11:49:10.534059-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster_members.ibd to /datadir/mysql/wsrep_cluster_members.ibd 2023-12-29T11:49:10.534109-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2023-12-29T11:49:10.534153-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSV to /datadir/mysql/general_log.CSV 2023-12-29T11:49:10.534215-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2023-12-29T11:49:10.534250-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_streaming_log.ibd to /datadir/mysql/wsrep_streaming_log.ibd 2023-12-29T11:49:10.534284-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2023-12-29T11:49:10.534320-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log_225.sdi to /datadir/mysql/general_log_225.sdi 2023-12-29T11:49:10.534357-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2023-12-29T11:49:10.534389-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log_226.sdi to /datadir/mysql/slow_log_226.sdi 2023-12-29T11:49:10.534457-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2023-12-29T11:49:10.534494-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/wsrep_cluster.ibd to /datadir/mysql/wsrep_cluster.ibd 2023-12-29T11:49:10.534526-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2023-12-29T11:49:10.534563-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSV to /datadir/mysql/slow_log.CSV 2023-12-29T11:49:10.534593-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2023-12-29T11:49:10.534628-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/slow_log.CSM to /datadir/mysql/slow_log.CSM 2023-12-29T11:49:10.534664-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2023-12-29T11:49:10.534696-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql/general_log.CSM to /datadir/mysql/general_log.CSM 2023-12-29T11:49:10.534797-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2023-12-29T11:49:10.534841-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sys/sys_config.ibd to /datadir/sys/sys_config.ibd 2023-12-29T11:49:10.534851-00:00 1 [Note] [MY-011825] [Xtrabackup] Creating directory ./#innodb_redo 2023-12-29T11:49:10.534867-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: creating directory ./#innodb_redo 2023-12-29T11:49:10.534903-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./sst_info to /datadir/sst_info 2023-12-29T11:49:10.534946-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./sst_info to /datadir/sst_info 2023-12-29T11:49:10.534986-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2023-12-29T11:49:10.535024-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_galera_info to /datadir/xtrabackup_galera_info 2023-12-29T11:49:10.535057-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./mysql.ibd to /datadir/mysql.ibd 2023-12-29T11:49:10.535091-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./mysql.ibd to /datadir/mysql.ibd 2023-12-29T11:49:10.535123-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ibtmp1 to /datadir/ibtmp1 2023-12-29T11:49:10.535152-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ibtmp1 to /datadir/ibtmp1 2023-12-29T11:49:10.535252-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2023-12-29T11:49:10.535295-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./myApp/myApp.ibd to /datadir/myApp/myApp.ibd 2023-12-29T11:49:10.535322-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./xtrabackup_info to /datadir/xtrabackup_info 2023-12-29T11:49:10.535352-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./xtrabackup_info to /datadir/xtrabackup_info 2023-12-29T11:49:10.535381-00:00 1 [Note] [MY-011825] [Xtrabackup] Moving ./ib_buffer_pool to /datadir/ib_buffer_pool 2023-12-29T11:49:10.535450-00:00 1 [Note] [MY-011825] [Xtrabackup] Done: Moving file ./ib_buffer_pool to /datadir/ib_buffer_pool 2023-12-29T11:49:10.625642-00:00 0 [Note] [MY-011825] [Xtrabackup] completed OK! + cat /tmp/tmp.GWj3FVLASp + rm /tmp/tmp.KzyxaFGucm /tmp/tmp.GWj3FVLASp + return 0 + wait_for_running some-name-proxysql 1 + local name=some-name-proxysql + let last_pod=0 + : + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 0 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-proxysql-0 480 + local pod=some-name-proxysql-0 + local max_retry=480 + local ns= ++ echo some-name-proxysql-0 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=proxysql + set +o xtrace pod/some-name-proxysql-0 condition met some-name-proxysql-0.Ok + wait_for_running some-name-pxc 3 + local name=some-name-pxc + let last_pod=2 + local max_retry=480 + desc 'wait for running cluster' + set +o xtrace ----------------------------------------------------------------------------------- wait for running cluster ----------------------------------------------------------------------------------- ++ seq 0 2 + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-0 480 + local pod=some-name-pxc-0 + local max_retry=480 + local ns= ++ echo some-name-pxc-0 ++ egrep '^(pxc|proxysql)$' ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' + local container=pxc + set +o xtrace pod/some-name-pxc-0 condition met some-name-pxc-0.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-1 480 + local pod=some-name-pxc-1 + local max_retry=480 + local ns= ++ echo some-name-pxc-1 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-1 condition met some-name-pxc-1.Ok + for i in '$(seq 0 $last_pod)' + wait_pod some-name-pxc-2 480 + local pod=some-name-pxc-2 + local max_retry=480 + local ns= ++ echo some-name-pxc-2 ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container=pxc + set +o xtrace pod/some-name-pxc-2 condition met some-name-pxc-2.Ok + sleep 35 + desc 'check data after backup' + set +o xtrace ----------------------------------------------------------------------------------- check data after backup ----------------------------------------------------------------------------------- + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-0.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-0.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.URVMUm22B2 +++ mktemp ++ local LAST_ERR=/tmp/tmp.l3UU8PipP3 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.URVMUm22B2 ++ cat /tmp/tmp.l3UU8PipP3 ++ rm /tmp/tmp.URVMUm22B2 /tmp/tmp.l3UU8PipP3 ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.LW7qwu72v2/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.LW7qwu72v2/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-1.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-1.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.7yr09tdOEU +++ mktemp ++ local LAST_ERR=/tmp/tmp.EGcb7vHKdw ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.7yr09tdOEU ++ cat /tmp/tmp.EGcb7vHKdw ++ rm /tmp/tmp.7yr09tdOEU /tmp/tmp.EGcb7vHKdw ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.LW7qwu72v2/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.LW7qwu72v2/select-1.sql + compare_mysql_cmd select-1 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local command_id=select-1 + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local postfix= + local expected_result=/mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql + [[ perconalab/percona-xtradb-cluster-operator:main-pxc8.0 =~ 8\.0 ]] + '[' -f /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1-80.sql ']' + run_mysql 'SELECT * from myApp.myApp;' '-h some-name-pxc-2.some-name-pxc -uroot -proot_password' + local 'command=SELECT * from myApp.myApp;' + local 'uri=-h some-name-pxc-2.some-name-pxc -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.upoc4WCcj1 +++ mktemp ++ local LAST_ERR=/tmp/tmp.D1rNdOnVH4 ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.upoc4WCcj1 ++ cat /tmp/tmp.D1rNdOnVH4 ++ rm /tmp/tmp.upoc4WCcj1 /tmp/tmp.D1rNdOnVH4 ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' ++ echo pxc-client-599db5b9bc-hllss + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + '[' '!' -s /tmp/tmp.LW7qwu72v2/select-1.sql ']' + diff -u /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/e2e-tests/restore-to-encrypted-cluster/compare/select-1.sql /tmp/tmp.LW7qwu72v2/select-1.sql + '[' on-demand-backup-aws-s3 '!=' on-demand-backup-minio ']' + desc 'copy backup' + set +o xtrace ----------------------------------------------------------------------------------- copy backup ----------------------------------------------------------------------------------- + bash /mnt/jenkins/workspace/cloud-pxc-operator_PR-1522/deploy/backup/copy-backup.sh on-demand-backup-aws-s3 /tmp/tmp.LW7qwu72v2/backup Log: /tmp/tmp.4I8Pj5I6LF/log s3://operator-testing/some-name-2023-12-29-11:38:59-full Downloading started Downloading finished You can recover data locally with following commands: $ service mysqld stop $ rm -rf /var/lib/mysql/* $ cat /tmp/tmp.LW7qwu72v2/backup/xtrabackup.stream | xbstream --decompress -x -C /var/lib/mysql $ xtrabackup --prepare --target-dir=/var/lib/mysql $ chown -R mysql:mysql /var/lib/mysql $ service mysqld start + keyring_plugin_must_be_in_use some-name + local cluster=some-name + desc 'check keyring plugin usage' + set +o xtrace ----------------------------------------------------------------------------------- check keyring plugin usage ----------------------------------------------------------------------------------- + is_keyring_plugin_in_use some-name + local cluster=some-name + kubectl_bin exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' ++ mktemp + local LAST_OUT=/tmp/tmp.3hqF7UMxeg ++ mktemp + local LAST_ERR=/tmp/tmp.xgjZQz3mNh + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + egrep -o 'early-plugin-load=keyring_\w+.so' + kubectl exec -it some-name-pxc-0 -c pxc -- bash -c 'cat /etc/mysql/node.cnf' + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.3hqF7UMxeg + cat /tmp/tmp.xgjZQz3mNh Unable to use a TTY - input is not a terminal or the right kind of file + rm /tmp/tmp.3hqF7UMxeg /tmp/tmp.xgjZQz3mNh + return 0 early-plugin-load=keyring_vault.so + table_must_not_be_encrypted some-name myApp + local cluster=some-name + local table=myApp + is_table_encrypted some-name myApp + local cluster=some-name + local table=myApp + egrep -o 'ENCRYPTION=('\''Y'\''|"Y")' + run_mysql 'SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' '-h some-name-proxysql -uroot -proot_password' + local 'command=SELECT CREATE_OPTIONS FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_NAME=\"myApp\";' + local 'uri=-h some-name-proxysql -uroot -proot_password' ++ get_client_pod ++ kubectl_bin get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' +++ mktemp ++ local LAST_OUT=/tmp/tmp.jMxALmq4kt +++ mktemp ++ local LAST_ERR=/tmp/tmp.lOsQnRc0jM ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=name=pxc-client -o 'jsonpath={.items[].metadata.name}' ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.jMxALmq4kt ++ cat /tmp/tmp.lOsQnRc0jM ++ rm /tmp/tmp.jMxALmq4kt /tmp/tmp.lOsQnRc0jM ++ return 0 + client_pod=pxc-client-599db5b9bc-hllss + wait_pod pxc-client-599db5b9bc-hllss + local pod=pxc-client-599db5b9bc-hllss + local max_retry=480 + local ns= ++ echo pxc-client-599db5b9bc-hllss ++ /usr/bin/sed -E 's/.*-(pxc|proxysql)-[0-9]/\1/' ++ egrep '^(pxc|proxysql)$' + local container= + set +o xtrace pod/pxc-client-599db5b9bc-hllss condition met pxc-client-599db5b9bc-hllss.Ok + set +o xtrace + helm uninstall vault-service-1-7243 Error: uninstall: Release not loaded: vault-service-1-7243: release: not found + : + destroy restore-to-encrypted-cluster-4709 + kubectl_bin delete --grace-period=0 --force=true namespace vault-service-1-7243 + local namespace=restore-to-encrypted-cluster-4709 + local ignore_logs=true + desc 'destroy cluster/operator and all other resources' + set +o xtrace ----------------------------------------------------------------------------------- destroy cluster/operator and all other resources ----------------------------------------------------------------------------------- + '[' true == false -o 1 == 1 ']' ++ mktemp + grep -v level=info + /usr/bin/sed -r 's/"ts":[0-9.]+//; s^limits-[0-9.]+/^^g' + local LAST_OUT=/tmp/tmp.Lw2mv2UNIf + tee /tmp/tmp.LW7qwu72v2/operator.log ++ get_operator_pod ++ local label_prefix=app.kubernetes.io/ + sort -u + grep -v 'get backup status: Job.batch' ++ mktemp + grep -v 'the object has been modified' + local LAST_ERR=/tmp/tmp.IoElcph4DM + local exit_status=0 +++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -n pxc-operator +++ grep -c percona-xtradb-cluster-operator ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace vault-service-1-7243 ++ local check_label=1 ++ [[ 1 -eq 0 ]] ++ kubectl_bin get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator +++ mktemp ++ local LAST_OUT=/tmp/tmp.RGQ0l2O8kC +++ mktemp ++ local LAST_ERR=/tmp/tmp.bHIWhIXSCr ++ local exit_status=0 +++ seq 0 2 ++ for i in '$(seq 0 2)' ++ set +e ++ kubectl get pods --selector=app.kubernetes.io/name=percona-xtradb-cluster-operator -o 'jsonpath={.items[].metadata.name}' -n pxc-operator ++ exit_status=0 ++ set -e ++ '[' 0 '!=' 0 ']' ++ break ++ cat /tmp/tmp.RGQ0l2O8kC ++ cat /tmp/tmp.bHIWhIXSCr ++ rm /tmp/tmp.RGQ0l2O8kC /tmp/tmp.bHIWhIXSCr ++ return 0 + kubectl_bin logs -n pxc-operator percona-xtradb-cluster-operator-6dbd475d79-7244c ++ mktemp + local LAST_OUT=/tmp/tmp.paIUps7M5r ++ mktemp + local LAST_ERR=/tmp/tmp.DvvaoviA0J + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl logs -n pxc-operator percona-xtradb-cluster-operator-6dbd475d79-7244c + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.paIUps7M5r + cat /tmp/tmp.DvvaoviA0J + rm /tmp/tmp.paIUps7M5r /tmp/tmp.DvvaoviA0J + return 0 2023-12-29T11:32:15.933Z INFO setup Manager starting up {"gitCommit": "1ed4605b677654d7ec76c0662b204abb3cbe9a67", "gitBranch": "PR-1522-1ed4605b", "buildTime": "2023-12-29T10:12:44Z", "goVersion": "go1.21.5", "os": "linux", "arch": "amd64"} 2023-12-29T11:32:15.933Z INFO setup Runs on {"platform": "kubernetes", "version": "v1.25.16-gke.1041000"} 2023-12-29T11:32:15.934Z INFO setup Registering Components. 2023-12-29T11:32:17.411Z INFO controller-runtime.metrics Starting metrics server 2023-12-29T11:32:17.411Z INFO controller-runtime.webhook Registering webhook {"path": "/validate-percona-xtradbcluster"} 2023-12-29T11:32:17.411Z INFO setup Starting the Cmd. 2023-12-29T11:32:17.412Z INFO attempting to acquire leader lease pxc-operator/08db0feb.percona.com... 2023-12-29T11:32:17.412Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":8080", "secure": false} 2023-12-29T11:32:17.412Z INFO controller-runtime.webhook Starting webhook server 2023-12-29T11:32:17.412Z INFO starting server {"kind": "health probe", "addr": "[::]:8081"} 2023-12-29T11:32:17.413Z INFO controller-runtime.certwatcher Starting certificate watcher 2023-12-29T11:32:17.413Z INFO controller-runtime.certwatcher Updated current TLS certificate 2023-12-29T11:32:17.413Z INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443} 2023-12-29T11:32:17.423Z DEBUG events percona-xtradb-cluster-operator-6dbd475d79-7244c_244d18bd-9b5e-4114-9c64-2bc6e8074688 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"pxc-operator","name":"08db0feb.percona.com","uid":"f9cf9d79-0452-4bf2-96ed-1989523a7a59","apiVersion":"coordination.k8s.io/v1","resourceVersion":"35690"}, "reason": "LeaderElection"} 2023-12-29T11:32:17.423Z INFO Starting Controller {"controller": "pxcrestore-controller"} 2023-12-29T11:32:17.423Z INFO Starting EventSource {"controller": "pxc-controller", "source": "kind source: *v1.PerconaXtraDBCluster"} 2023-12-29T11:32:17.423Z INFO Starting EventSource {"controller": "pxcrestore-controller", "source": "kind source: *v1.PerconaXtraDBClusterRestore"} 2023-12-29T11:32:17.423Z INFO successfully acquired lease pxc-operator/08db0feb.percona.com 2023-12-29T11:32:17.424Z INFO Starting Controller {"controller": "pxcbackup-controller"} 2023-12-29T11:32:17.424Z INFO Starting Controller {"controller": "pxc-controller"} 2023-12-29T11:32:17.424Z INFO Starting EventSource {"controller": "pxcbackup-controller", "source": "kind source: *v1.PerconaXtraDBClusterBackup"} 2023-12-29T11:32:17.526Z INFO Starting workers {"controller": "pxc-controller", "worker count": 1} 2023-12-29T11:32:17.526Z INFO Starting workers {"controller": "pxcrestore-controller", "worker count": 1} 2023-12-29T11:32:17.531Z INFO Starting workers {"controller": "pxcbackup-controller", "worker count": 1} 2023-12-29T11:32:58.010Z INFO Set CR version {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "f88d96d1-2dd6-459b-ae95-33630d1928e2", "version": "1.14.0"} 2023-12-29T11:34:17.849Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ce8ae8ce-1075-4d17-b164-8707c90bb96c", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:17.849Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ce8ae8ce-1075-4d17-b164-8707c90bb96c"} 2023-12-29T11:34:18.061Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8837f23c-2f43-4820-9798-c464b987e10b", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:18.061Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8837f23c-2f43-4820-9798-c464b987e10b"} 2023-12-29T11:34:18.260Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0d75b58b-116b-4eb5-8c2f-8db5f4ab3c11", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:18.260Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0d75b58b-116b-4eb5-8c2f-8db5f4ab3c11"} 2023-12-29T11:34:18.457Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c7cf1dba-b733-4c0a-8678-b99c76719834", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:18.457Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c7cf1dba-b733-4c0a-8678-b99c76719834"} 2023-12-29T11:34:18.650Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1af3277d-fe7b-42da-85d3-1051920d1741", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:18.650Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1af3277d-fe7b-42da-85d3-1051920d1741"} 2023-12-29T11:34:18.871Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "73ffd3d9-3d01-4585-ad64-dcf8c407dc0c", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:18.871Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "73ffd3d9-3d01-4585-ad64-dcf8c407dc0c"} 2023-12-29T11:34:19.155Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c39ee5f4-1872-44f2-aa4e-f3c52e30ec9c", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:19.155Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c39ee5f4-1872-44f2-aa4e-f3c52e30ec9c"} 2023-12-29T11:34:19.399Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4e4bbe55-f642-4213-afee-de5c8c4d41d8", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:19.399Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4e4bbe55-f642-4213-afee-de5c8c4d41d8"} 2023-12-29T11:34:19.678Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "cd9401f1-e500-4298-8987-06246f6356c7", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:19.678Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "cd9401f1-e500-4298-8987-06246f6356c7"} 2023-12-29T11:34:21.161Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "823a36c2-2288-442d-9366-fd7d921e9ca7", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:21.161Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "823a36c2-2288-442d-9366-fd7d921e9ca7"} 2023-12-29T11:34:23.986Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "27dd4ac7-99b7-469a-9364-51ab86ee9ba9", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:23.986Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "27dd4ac7-99b7-469a-9364-51ab86ee9ba9"} 2023-12-29T11:34:29.311Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "dd1659d5-57f9-49a8-99db-d04b963a888b", "error": "reconcile users: manage sys users: get primary pod: failed to get primary pod: not found", "errorVerbose": "not found\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries.init\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/pxc/queries/queries.go:52\nruntime.doInit1\n\t/usr/local/go/src/runtime/proc.go:6740\nruntime.doInit\n\t/usr/local/go/src/runtime/proc.go:6707\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:249\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to get primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:544\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUserPassExpirationPolicy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1084\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleOperatorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:274\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:161\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:34:29.311Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "dd1659d5-57f9-49a8-99db-d04b963a888b"} 2023-12-29T11:34:39.844Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "67f17ca8-e4db-4e64-9a84-f7cd09521391", "user": "operator"} 2023-12-29T11:34:40.036Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "67f17ca8-e4db-4e64-9a84-f7cd09521391", "user": "monitor"} 2023-12-29T11:34:40.296Z INFO User monitor: granted privileges {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "67f17ca8-e4db-4e64-9a84-f7cd09521391"} 2023-12-29T11:34:40.485Z INFO monitor user privileges granted {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "67f17ca8-e4db-4e64-9a84-f7cd09521391"} 2023-12-29T11:34:40.697Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "67f17ca8-e4db-4e64-9a84-f7cd09521391", "user": "xtrabackup"} 2023-12-29T11:34:40.916Z INFO User xtrabackup: granted privileges {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "67f17ca8-e4db-4e64-9a84-f7cd09521391"} 2023-12-29T11:34:41.112Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "67f17ca8-e4db-4e64-9a84-f7cd09521391", "user": "replication"} 2023-12-29T11:34:41.573Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "67f17ca8-e4db-4e64-9a84-f7cd09521391", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:34:42.363Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e1881984-b350-4930-87e9-d8f74e7b7337", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:34:47.762Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0bfbcfba-1fda-4cb9-b0bc-98e31de42e78", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:34:54.108Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "99635f87-2520-427e-8244-b297f936ba6b", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:00.138Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "95262140-bcc2-420d-bdb0-b36981e4e4b2", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:05.969Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "f1000e96-cb9b-4041-b698-d8a14d780ebd", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:11.717Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "71806aca-8e90-4f32-8f51-27b6dc8b8746", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:17.399Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0c12b2b0-127d-4166-86b5-77e706b03fef", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:23.639Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "9bd71977-351d-4549-8f96-8ff0862f91b2", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:29.460Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4ac329e0-9768-49d4-ad20-34789f013ce4", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:35:29.563Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4ac329e0-9768-49d4-ad20-34789f013ce4", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:30.827Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "33d3a345-acb0-495c-be5b-5a3ccc4e6e2b", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:30.904Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "33d3a345-acb0-495c-be5b-5a3ccc4e6e2b", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:35:35.304Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "091064d8-62eb-4086-b677-ea3f03b98db0", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:35.379Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "091064d8-62eb-4086-b677-ea3f03b98db0", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:35:41.439Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a588a943-9eee-4733-b702-958595889a15", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:35:41.524Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a588a943-9eee-4733-b702-958595889a15", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:47.293Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e06655ed-5a28-4dbc-9308-60c041939e28", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:47.370Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e06655ed-5a28-4dbc-9308-60c041939e28", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:35:53.446Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ee1cea14-e48d-4c94-ad60-79bb280247d1", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:53.524Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ee1cea14-e48d-4c94-ad60-79bb280247d1", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:35:59.283Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "29b3225c-3eb4-40d3-9292-aabea3e1e300", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:35:59.487Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "29b3225c-3eb4-40d3-9292-aabea3e1e300", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:05.221Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "47303c99-7ff5-44c1-863e-7e4e33c1a241", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:05.299Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "47303c99-7ff5-44c1-863e-7e4e33c1a241", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:11.141Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "6c140eb7-4aad-4833-931d-6bc9eb45a24c", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:11.221Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "6c140eb7-4aad-4833-931d-6bc9eb45a24c", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:17.029Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "73c0c34b-9629-41b9-bdee-3fb1fb1fa57c", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:17.112Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "73c0c34b-9629-41b9-bdee-3fb1fb1fa57c", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:22.936Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b7039be3-5b57-4455-b364-c99b181bf892", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:23.208Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b7039be3-5b57-4455-b364-c99b181bf892", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:28.965Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8f5bbdf8-ba5e-4d08-98a5-3d90522e085b", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:29.036Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8f5bbdf8-ba5e-4d08-98a5-3d90522e085b", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:34.934Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "bb5d554c-d21e-4714-b3d4-b366cc06af1c", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:35.028Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "bb5d554c-d21e-4714-b3d4-b366cc06af1c", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:40.834Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "9bed27f6-95d9-4950-8857-bebf9f148e18", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:40.918Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "9bed27f6-95d9-4950-8857-bebf9f148e18", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:41.154Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "9bed27f6-95d9-4950-8857-bebf9f148e18", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:36:41.782Z INFO Password expiration policy updated {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eb09fa49-56ad-41b3-a645-662d5865511a", "user": "root"} 2023-12-29T11:36:43.272Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eb09fa49-56ad-41b3-a645-662d5865511a", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:43.355Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eb09fa49-56ad-41b3-a645-662d5865511a", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:43.446Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eb09fa49-56ad-41b3-a645-662d5865511a", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:36:43.546Z INFO update PXC version (fetched from db) {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eb09fa49-56ad-41b3-a645-662d5865511a", "new version": "8.0.34-26.1"} 2023-12-29T11:36:44.696Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eb09fa49-56ad-41b3-a645-662d5865511a"} 2023-12-29T11:36:45.417Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "103d2f9f-4694-4cd3-a565-cd7a69e28d50", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:45.503Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "103d2f9f-4694-4cd3-a565-cd7a69e28d50", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:45.606Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "103d2f9f-4694-4cd3-a565-cd7a69e28d50", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:36:47.006Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "103d2f9f-4694-4cd3-a565-cd7a69e28d50"} 2023-12-29T11:36:47.870Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0d87a6ad-5dfe-4529-9cdc-aa764a39d4ea", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:47.945Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0d87a6ad-5dfe-4529-9cdc-aa764a39d4ea", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:48.033Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0d87a6ad-5dfe-4529-9cdc-aa764a39d4ea", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:36:49.384Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0d87a6ad-5dfe-4529-9cdc-aa764a39d4ea"} 2023-12-29T11:36:54.688Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "706316f4-9bf0-4319-b66b-30b4d053af41", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:36:54.763Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "706316f4-9bf0-4319-b66b-30b4d053af41", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:36:54.847Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "706316f4-9bf0-4319-b66b-30b4d053af41", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:36:56.087Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "706316f4-9bf0-4319-b66b-30b4d053af41"} 2023-12-29T11:37:01.641Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5c0530d9-62d7-408e-8e02-e94ede70c6d3", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:37:01.716Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5c0530d9-62d7-408e-8e02-e94ede70c6d3", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:37:01.805Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5c0530d9-62d7-408e-8e02-e94ede70c6d3", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:37:03.097Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5c0530d9-62d7-408e-8e02-e94ede70c6d3"} 2023-12-29T11:37:08.660Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "3aaa3646-92c2-410e-8ae4-634df0d83a59", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:37:08.752Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "3aaa3646-92c2-410e-8ae4-634df0d83a59", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:37:08.867Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "3aaa3646-92c2-410e-8ae4-634df0d83a59", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:37:09.887Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "3aaa3646-92c2-410e-8ae4-634df0d83a59"} 2023-12-29T11:37:16.481Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8084b548-4250-4fc2-b986-f6a5bcb1d618", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:37:16.600Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8084b548-4250-4fc2-b986-f6a5bcb1d618", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:37:16.687Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8084b548-4250-4fc2-b986-f6a5bcb1d618", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:37:17.097Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8084b548-4250-4fc2-b986-f6a5bcb1d618"} 2023-12-29T11:37:23.679Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "71b28552-d432-48e0-9684-78cff62dc10e", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:37:23.767Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "71b28552-d432-48e0-9684-78cff62dc10e", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:37:23.927Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "71b28552-d432-48e0-9684-78cff62dc10e", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:37:25.014Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "71b28552-d432-48e0-9684-78cff62dc10e"} 2023-12-29T11:37:30.754Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4f1414ab-d64f-42db-964f-6b3396b3a84f", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:37:30.829Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4f1414ab-d64f-42db-964f-6b3396b3a84f", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:37:30.912Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4f1414ab-d64f-42db-964f-6b3396b3a84f", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:37:31.827Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4f1414ab-d64f-42db-964f-6b3396b3a84f"} 2023-12-29T11:37:37.574Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b3f42791-2896-420e-881a-85560e5ba5a7", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:37:37.689Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b3f42791-2896-420e-881a-85560e5ba5a7", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:37:37.790Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b3f42791-2896-420e-881a-85560e5ba5a7", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:37:39.021Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b3f42791-2896-420e-881a-85560e5ba5a7"} 2023-12-29T11:37:44.609Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d4b8c481-86d2-41e1-bc66-30262bb70838", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:37:44.698Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d4b8c481-86d2-41e1-bc66-30262bb70838", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:37:44.772Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d4b8c481-86d2-41e1-bc66-30262bb70838", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:37:46.106Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d4b8c481-86d2-41e1-bc66-30262bb70838"} 2023-12-29T11:37:51.310Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ebdb64c9-b69d-472e-a53d-4c12198a8d27", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:37:51.403Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ebdb64c9-b69d-472e-a53d-4c12198a8d27", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:37:51.490Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ebdb64c9-b69d-472e-a53d-4c12198a8d27", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:37:52.816Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ebdb64c9-b69d-472e-a53d-4c12198a8d27"} 2023-12-29T11:37:58.182Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "37420c00-cb1a-4828-8ce8-5acc7459b1c8", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:37:58.268Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "37420c00-cb1a-4828-8ce8-5acc7459b1c8", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:37:58.344Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "37420c00-cb1a-4828-8ce8-5acc7459b1c8", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:37:59.535Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "37420c00-cb1a-4828-8ce8-5acc7459b1c8"} 2023-12-29T11:38:04.880Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "651ad537-b099-43d1-9182-6070ded3f76e", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:38:04.955Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "651ad537-b099-43d1-9182-6070ded3f76e", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:38:05.044Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "651ad537-b099-43d1-9182-6070ded3f76e", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:38:06.294Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "651ad537-b099-43d1-9182-6070ded3f76e"} 2023-12-29T11:38:11.618Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "722a96ad-52cf-445e-ac78-69df832a91fa", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:38:11.711Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "722a96ad-52cf-445e-ac78-69df832a91fa", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:38:11.799Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "722a96ad-52cf-445e-ac78-69df832a91fa", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:38:13.017Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "722a96ad-52cf-445e-ac78-69df832a91fa"} 2023-12-29T11:38:17.085Z INFO Creating a new volume for backup {"controller": "pxcbackup-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-pvc", "reconcileID": "275e6c69-63af-4bc0-aeba-c9c2146e9fe7", "Namespace": "restore-to-encrypted-cluster-4709", "Name": "xb-on-demand-backup-pvc"} 2023-12-29T11:38:17.119Z INFO Created a new backup job {"controller": "pxcbackup-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-pvc", "reconcileID": "275e6c69-63af-4bc0-aeba-c9c2146e9fe7", "Namespace": "restore-to-encrypted-cluster-4709", "Name": "xb-on-demand-backup-pvc"} 2023-12-29T11:38:18.274Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "86d4ef6b-f0b9-486b-99ec-e20615ddebe4", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:38:18.359Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "86d4ef6b-f0b9-486b-99ec-e20615ddebe4", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:38:18.464Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "86d4ef6b-f0b9-486b-99ec-e20615ddebe4", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:38:19.705Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "86d4ef6b-f0b9-486b-99ec-e20615ddebe4"} 2023-12-29T11:38:25.085Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "16d49332-6084-4c69-a621-48505fbb75a4", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:38:25.163Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "16d49332-6084-4c69-a621-48505fbb75a4", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:38:25.248Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "16d49332-6084-4c69-a621-48505fbb75a4", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:38:26.510Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "16d49332-6084-4c69-a621-48505fbb75a4"} 2023-12-29T11:38:32.026Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b4da22c3-9d22-42a4-a820-947ee8d02b64", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:38:32.106Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b4da22c3-9d22-42a4-a820-947ee8d02b64", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:38:32.204Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b4da22c3-9d22-42a4-a820-947ee8d02b64", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:38:33.486Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b4da22c3-9d22-42a4-a820-947ee8d02b64"} 2023-12-29T11:38:39.055Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "675baddd-b7d4-4064-9163-816a8b049167", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:38:39.149Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "675baddd-b7d4-4064-9163-816a8b049167", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:38:39.238Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "675baddd-b7d4-4064-9163-816a8b049167", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:38:40.510Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "675baddd-b7d4-4064-9163-816a8b049167"} 2023-12-29T11:38:45.969Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eec28d7a-ba3b-4440-acb8-6f4726c5007a", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:38:46.119Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eec28d7a-ba3b-4440-acb8-6f4726c5007a", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:38:46.225Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eec28d7a-ba3b-4440-acb8-6f4726c5007a", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:38:47.202Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "eec28d7a-ba3b-4440-acb8-6f4726c5007a"} 2023-12-29T11:38:52.822Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2167ee4b-2ff2-43f6-94f3-73f09d88e42b", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:38:53.128Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2167ee4b-2ff2-43f6-94f3-73f09d88e42b", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:38:53.394Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2167ee4b-2ff2-43f6-94f3-73f09d88e42b", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:38:54.222Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2167ee4b-2ff2-43f6-94f3-73f09d88e42b"} 2023-12-29T11:38:59.366Z INFO Created a new backup job {"controller": "pxcbackup-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-aws-s3", "reconcileID": "0328f48a-5e1d-41a8-9039-709c1ebc9b49", "Namespace": "restore-to-encrypted-cluster-4709", "Name": "xb-on-demand-backup-aws-s3"} 2023-12-29T11:39:00.065Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d0a5df46-a53d-4d6a-8c9f-d8d29a76044e", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:39:00.166Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d0a5df46-a53d-4d6a-8c9f-d8d29a76044e", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:39:00.712Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d0a5df46-a53d-4d6a-8c9f-d8d29a76044e", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:39:01.511Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d0a5df46-a53d-4d6a-8c9f-d8d29a76044e"} 2023-12-29T11:39:07.280Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "11cd945f-8099-4cd3-8986-878e2aea3612", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:39:07.394Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "11cd945f-8099-4cd3-8986-878e2aea3612", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:39:07.491Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "11cd945f-8099-4cd3-8986-878e2aea3612", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:39:08.797Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "11cd945f-8099-4cd3-8986-878e2aea3612"} 2023-12-29T11:39:14.058Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "93c7a423-e5c6-453d-9b80-cd75de676e61", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:39:14.131Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "93c7a423-e5c6-453d-9b80-cd75de676e61", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:39:14.219Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "93c7a423-e5c6-453d-9b80-cd75de676e61", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:39:15.442Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "93c7a423-e5c6-453d-9b80-cd75de676e61"} 2023-12-29T11:39:21.184Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "754a8a1c-b5b2-4bec-bc07-bc303be7b0b1", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:39:21.267Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "754a8a1c-b5b2-4bec-bc07-bc303be7b0b1", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:39:21.445Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "754a8a1c-b5b2-4bec-bc07-bc303be7b0b1", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:39:22.538Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "754a8a1c-b5b2-4bec-bc07-bc303be7b0b1"} 2023-12-29T11:39:28.128Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e3f7d349-09ea-47f7-bfec-241ee19bca5b", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:39:28.220Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e3f7d349-09ea-47f7-bfec-241ee19bca5b", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:39:28.306Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e3f7d349-09ea-47f7-bfec-241ee19bca5b", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:39:29.508Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e3f7d349-09ea-47f7-bfec-241ee19bca5b"} 2023-12-29T11:39:34.891Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "74eb7b4c-cbe2-47de-8332-d13267100609", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:39:34.982Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "74eb7b4c-cbe2-47de-8332-d13267100609", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:39:35.067Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "74eb7b4c-cbe2-47de-8332-d13267100609", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:39:36.415Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "74eb7b4c-cbe2-47de-8332-d13267100609"} 2023-12-29T11:39:41.505Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0055eb56-4b21-4808-b68d-e3ce9d4ec0e6", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:39:41.597Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0055eb56-4b21-4808-b68d-e3ce9d4ec0e6", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:39:41.689Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0055eb56-4b21-4808-b68d-e3ce9d4ec0e6", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:39:42.999Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0055eb56-4b21-4808-b68d-e3ce9d4ec0e6"} 2023-12-29T11:39:48.316Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b3385c19-7923-4008-a912-60c4694feb22", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:39:48.407Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b3385c19-7923-4008-a912-60c4694feb22", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:39:48.489Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b3385c19-7923-4008-a912-60c4694feb22", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:39:49.711Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b3385c19-7923-4008-a912-60c4694feb22"} 2023-12-29T11:39:55.019Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2cf57063-6142-4eb3-8fe7-75aee782a95f", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:39:55.111Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2cf57063-6142-4eb3-8fe7-75aee782a95f", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:39:55.203Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2cf57063-6142-4eb3-8fe7-75aee782a95f", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:39:56.408Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2cf57063-6142-4eb3-8fe7-75aee782a95f"} 2023-12-29T11:40:02.109Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "f0b958c5-c4f1-4a6f-8a47-28922bac13f8", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:40:02.191Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "f0b958c5-c4f1-4a6f-8a47-28922bac13f8", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:40:02.291Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "f0b958c5-c4f1-4a6f-8a47-28922bac13f8", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:40:03.529Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "f0b958c5-c4f1-4a6f-8a47-28922bac13f8"} 2023-12-29T11:40:09.029Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ff6dfaa0-8cb2-49e0-b6cd-f7ea3481125c", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:40:09.132Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ff6dfaa0-8cb2-49e0-b6cd-f7ea3481125c", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:40:09.229Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ff6dfaa0-8cb2-49e0-b6cd-f7ea3481125c", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:40:10.426Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ff6dfaa0-8cb2-49e0-b6cd-f7ea3481125c"} 2023-12-29T11:40:16.316Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "96598107-8273-49e1-b540-69f0a6dc25a9", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:40:16.413Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "96598107-8273-49e1-b540-69f0a6dc25a9", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:40:16.499Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "96598107-8273-49e1-b540-69f0a6dc25a9", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:40:17.729Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "96598107-8273-49e1-b540-69f0a6dc25a9"} 2023-12-29T11:40:23.563Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "345ec92c-dab3-47d3-8ff1-93ffbff00d59", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:40:23.649Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "345ec92c-dab3-47d3-8ff1-93ffbff00d59", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:40:23.742Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "345ec92c-dab3-47d3-8ff1-93ffbff00d59", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:40:24.532Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "345ec92c-dab3-47d3-8ff1-93ffbff00d59"} 2023-12-29T11:40:30.639Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "18772a2e-5e9d-4b41-9ac2-b34f0a2ee0ed", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:40:30.721Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "18772a2e-5e9d-4b41-9ac2-b34f0a2ee0ed", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:40:30.816Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "18772a2e-5e9d-4b41-9ac2-b34f0a2ee0ed", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:40:31.801Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "18772a2e-5e9d-4b41-9ac2-b34f0a2ee0ed"} 2023-12-29T11:40:38.195Z ERROR sync users {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c7fa5987-2f82-4b24-8d01-d088d19d515d", "error": "exec syncusers: unable to upgrade connection: container not found (\"proxysql\") / / ", "errorVerbose": "exec syncusers: unable to upgrade connection: container not found (\"proxysql\") / / \ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).syncPXCUsersWithProxySQL\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:896\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1229\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:40:38.458Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c7fa5987-2f82-4b24-8d01-d088d19d515d", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:40:38.538Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c7fa5987-2f82-4b24-8d01-d088d19d515d", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:40:38.669Z INFO reconcile replication error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c7fa5987-2f82-4b24-8d01-d088d19d515d", "err": "failed to ensure cluster readonly status: check readonly status: select global read_only param: stdout: , stderr: mysql: [Warning] Using a password on the command line interface can be insecure.\nERROR 2005 (HY000): Unknown MySQL server host 'some-name-pxc-2.some-name-pxc.restore-to-encrypted-cluster-4709' (-2)\n: command terminated with exit code 1"} 2023-12-29T11:40:42.556Z INFO backup restore request {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-pvc", "reconcileID": "5825267a-6276-42d5-947c-cec71fc99c41"} 2023-12-29T11:40:54.602Z INFO stopping cluster {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-pvc", "reconcileID": "5825267a-6276-42d5-947c-cec71fc99c41", "cluster": "some-name"} 2023-12-29T11:41:29.855Z INFO starting restore {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-pvc", "reconcileID": "5825267a-6276-42d5-947c-cec71fc99c41", "cluster": "some-name", "backup": "on-demand-backup-pvc"} 2023-12-29T11:41:57.130Z INFO starting cluster {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-pvc", "reconcileID": "5825267a-6276-42d5-947c-cec71fc99c41", "cluster": "some-name"} 2023-12-29T11:42:34.069Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "700b9b9d-761b-44b4-905d-828948d93e19", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:34.069Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "700b9b9d-761b-44b4-905d-828948d93e19"} 2023-12-29T11:42:34.107Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "3d716490-5123-4ad9-949e-f90fd48b95c6", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:34.107Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "3d716490-5123-4ad9-949e-f90fd48b95c6"} 2023-12-29T11:42:34.144Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1c801378-53ad-437b-b554-5aaa0376a7f1", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:34.144Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1c801378-53ad-437b-b554-5aaa0376a7f1"} 2023-12-29T11:42:34.180Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ec5a8375-0c1d-4b2f-a8e8-cd4c7654d085", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:34.180Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ec5a8375-0c1d-4b2f-a8e8-cd4c7654d085"} 2023-12-29T11:42:34.493Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "348c278c-6830-4154-8938-6516247455ce", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:34.493Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "348c278c-6830-4154-8938-6516247455ce"} 2023-12-29T11:42:34.897Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0e57db6f-d023-4edb-bf10-50b24f913909", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:34.897Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0e57db6f-d023-4edb-bf10-50b24f913909"} 2023-12-29T11:42:35.291Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "f3c1c851-2d41-4b91-9ed1-062a8dc225fe", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:35.291Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "f3c1c851-2d41-4b91-9ed1-062a8dc225fe"} 2023-12-29T11:42:35.688Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "9b3b840c-124a-4ccd-b6cf-df9b8b829e73", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:35.688Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "9b3b840c-124a-4ccd-b6cf-df9b8b829e73"} 2023-12-29T11:42:36.097Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ea30c27b-77a4-44f8-892f-e63aa1331645", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:36.097Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "ea30c27b-77a4-44f8-892f-e63aa1331645"} 2023-12-29T11:42:36.490Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "324b24cb-6dee-4c5b-90f7-e683df13113f", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:36.490Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "324b24cb-6dee-4c5b-90f7-e683df13113f"} 2023-12-29T11:42:39.089Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c65fcf42-80ca-4f15-be6e-98a03cc516fc", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:39.089Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c65fcf42-80ca-4f15-be6e-98a03cc516fc"} 2023-12-29T11:42:44.245Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "51ab80ec-e13d-406e-8c82-38c2460097f0", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:44.245Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "51ab80ec-e13d-406e-8c82-38c2460097f0"} 2023-12-29T11:42:54.521Z ERROR Reconciler error {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8e0d6dfd-7ccc-4628-9423-7148f5ab360c", "error": "reconcile users: manage sys users: get primary pod: failed to connect to proxy: failed connect to proxy", "errorVerbose": "failed connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).connectProxy\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:524\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:528\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nfailed to connect to proxy\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getPrimaryPod\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/upgrade.go:530\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1066\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nget primary pod\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).getUserManager\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:1068\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).handleMonitorUser\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:394\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).updateUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:165\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:110\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nmanage sys users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).reconcileUsers\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/users.go:112\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:281\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650\nreconcile users\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:283\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:119\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:316\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1650"} 2023-12-29T11:42:54.521Z INFO Warning: Reconciler returned both a non-zero result and a non-nil error. The result will always be ignored if the error is non-nil and the non-nil error causes reqeueuing with exponential backoff. For more details, see: https://pkg.go.dev/sigs.k8s.io/controller-runtime/pkg/reconcile#Reconciler {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8e0d6dfd-7ccc-4628-9423-7148f5ab360c"} 2023-12-29T11:43:16.005Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b1643589-cc0e-4756-ab6c-04dc309263c3", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:16.765Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "54471734-7dc9-4af7-b2a3-94e5d9021fed", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:21.642Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "6a77a10f-9782-4ae3-b099-0b190588007c", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:27.192Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "3b0fd428-82c8-4a0c-abb7-7cfca7d1b1b8", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:27.847Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "79708fab-e98f-4e1e-b28f-3564166ece6a", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:32.804Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "838bf67d-7ca3-4da0-81df-139c332972a0", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:38.632Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "50416528-7a78-42a9-ba17-cde98877976b", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:44.447Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "97e9e3d5-81fe-4b8c-aa91-6965482a2a3d", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:44.548Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "97e9e3d5-81fe-4b8c-aa91-6965482a2a3d", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:43:45.165Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0f8f9be0-4723-4132-ac81-04deddc12c02", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:43:45.265Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0f8f9be0-4723-4132-ac81-04deddc12c02", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:50.149Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "241619fb-040e-4dbe-a8d0-ed2195c711e7", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:43:50.233Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "241619fb-040e-4dbe-a8d0-ed2195c711e7", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:43:55.886Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e2d43b96-77b8-4ac8-a607-72429f5ece18", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:43:55.994Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e2d43b96-77b8-4ac8-a607-72429f5ece18", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:01.531Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0e2e06cc-88a2-46a7-96d0-82cd127127da", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:01.610Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0e2e06cc-88a2-46a7-96d0-82cd127127da", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:07.164Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "efdbcd1f-dae6-4427-82ec-78d86afedacc", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:07.239Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "efdbcd1f-dae6-4427-82ec-78d86afedacc", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:12.897Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "de23f453-1e14-4fda-816d-33c5d35a325c", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:12.974Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "de23f453-1e14-4fda-816d-33c5d35a325c", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:18.548Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b4fcb82e-69d9-4c15-991c-221ff7bac723", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:18.622Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b4fcb82e-69d9-4c15-991c-221ff7bac723", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:24.476Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7369078e-9a11-4a4e-bf13-18de1072ce62", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:24.591Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7369078e-9a11-4a4e-bf13-18de1072ce62", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:30.208Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a8a63f86-42dd-4e5d-923e-3a0d5d6d2df3", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:30.382Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a8a63f86-42dd-4e5d-923e-3a0d5d6d2df3", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:36.027Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4fb715d3-a109-4e54-8243-cfca103d7de3", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:36.112Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "4fb715d3-a109-4e54-8243-cfca103d7de3", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:41.673Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "171733e5-eff9-4bc3-a80d-b60f4ecaa754", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:41.750Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "171733e5-eff9-4bc3-a80d-b60f4ecaa754", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:47.323Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1a396d2e-22f2-42e5-91b8-59842ab7e151", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:47.398Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1a396d2e-22f2-42e5-91b8-59842ab7e151", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:53.559Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a9c05ae4-5949-4530-aa3b-369ce6d4e9cd", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:53.635Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a9c05ae4-5949-4530-aa3b-369ce6d4e9cd", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:59.268Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c205b852-fa4e-4d68-846d-a413f27b4620", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:44:59.343Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c205b852-fa4e-4d68-846d-a413f27b4620", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:44:59.426Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c205b852-fa4e-4d68-846d-a413f27b4620", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:00.320Z INFO You can view xtrabackup log: 2023-12-29T11:45:01.394Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "204caf36-deac-407d-b52f-57862c5895eb", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:45:01.490Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "204caf36-deac-407d-b52f-57862c5895eb", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:45:01.582Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "204caf36-deac-407d-b52f-57862c5895eb", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:02.905Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "204caf36-deac-407d-b52f-57862c5895eb"} 2023-12-29T11:45:05.894Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "afc721d2-a6ff-457a-9be3-1f8da84e3494", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:45:05.971Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "afc721d2-a6ff-457a-9be3-1f8da84e3494", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:45:06.068Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "afc721d2-a6ff-457a-9be3-1f8da84e3494", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:07.310Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "afc721d2-a6ff-457a-9be3-1f8da84e3494"} 2023-12-29T11:45:12.491Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "98ddf94c-a3b4-4e55-8a44-f4c8cffdf098", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:45:12.568Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "98ddf94c-a3b4-4e55-8a44-f4c8cffdf098", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:45:12.652Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "98ddf94c-a3b4-4e55-8a44-f4c8cffdf098", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:14.012Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "98ddf94c-a3b4-4e55-8a44-f4c8cffdf098"} 2023-12-29T11:45:19.089Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "50b28c09-d079-4400-be33-9d5f977f9249", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:19.174Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "50b28c09-d079-4400-be33-9d5f977f9249", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:45:19.256Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "50b28c09-d079-4400-be33-9d5f977f9249", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:45:20.511Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "50b28c09-d079-4400-be33-9d5f977f9249"} 2023-12-29T11:45:25.758Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b1f296bf-199e-43ff-b772-fdf5ed34b2b2", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:25.846Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b1f296bf-199e-43ff-b772-fdf5ed34b2b2", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:45:25.940Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b1f296bf-199e-43ff-b772-fdf5ed34b2b2", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:45:27.109Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b1f296bf-199e-43ff-b772-fdf5ed34b2b2"} 2023-12-29T11:45:32.382Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e3b66a44-05d0-4835-a9f7-0143b5b1a536", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:45:32.469Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e3b66a44-05d0-4835-a9f7-0143b5b1a536", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:45:32.550Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e3b66a44-05d0-4835-a9f7-0143b5b1a536", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:33.926Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e3b66a44-05d0-4835-a9f7-0143b5b1a536"} 2023-12-29T11:45:39.268Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0c83cfc7-d42b-4631-9d00-0b48d183f68b", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:39.351Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0c83cfc7-d42b-4631-9d00-0b48d183f68b", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:45:39.447Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0c83cfc7-d42b-4631-9d00-0b48d183f68b", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:45:40.788Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0c83cfc7-d42b-4631-9d00-0b48d183f68b"} 2023-12-29T11:45:46.634Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "36cd50f9-4909-486a-a219-5142140bae94", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:45:46.731Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "36cd50f9-4909-486a-a219-5142140bae94", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:45:46.823Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "36cd50f9-4909-486a-a219-5142140bae94", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:48.294Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "36cd50f9-4909-486a-a219-5142140bae94"} 2023-12-29T11:45:53.780Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1d4466d0-17ed-4711-a6fb-40c033356ef2", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:45:53.904Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1d4466d0-17ed-4711-a6fb-40c033356ef2", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:45:54.003Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1d4466d0-17ed-4711-a6fb-40c033356ef2", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:45:55.292Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1d4466d0-17ed-4711-a6fb-40c033356ef2"} 2023-12-29T11:46:00.899Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "27cbf471-47b5-4b91-b035-1a5a99dc243e", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:46:00.976Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "27cbf471-47b5-4b91-b035-1a5a99dc243e", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:46:01.066Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "27cbf471-47b5-4b91-b035-1a5a99dc243e", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:46:01.911Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "27cbf471-47b5-4b91-b035-1a5a99dc243e"} 2023-12-29T11:46:07.609Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "cafe564a-25a5-45cf-90f1-0de4d034175d", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:46:07.709Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "cafe564a-25a5-45cf-90f1-0de4d034175d", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:46:07.796Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "cafe564a-25a5-45cf-90f1-0de4d034175d", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:46:09.094Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "cafe564a-25a5-45cf-90f1-0de4d034175d"} 2023-12-29T11:46:14.386Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "545b6d90-8b7d-4e66-8068-1162cee10f22", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:46:14.487Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "545b6d90-8b7d-4e66-8068-1162cee10f22", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:46:14.579Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "545b6d90-8b7d-4e66-8068-1162cee10f22", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:46:15.807Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "545b6d90-8b7d-4e66-8068-1162cee10f22"} 2023-12-29T11:46:21.157Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a5dc30f7-b43e-4c0c-854a-a4f203784205", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:46:21.240Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a5dc30f7-b43e-4c0c-854a-a4f203784205", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:46:21.340Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a5dc30f7-b43e-4c0c-854a-a4f203784205", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:46:22.588Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a5dc30f7-b43e-4c0c-854a-a4f203784205"} 2023-12-29T11:46:27.898Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7e0aaeae-fa8f-453f-ae4a-5ffc76fc2c46", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:46:27.971Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7e0aaeae-fa8f-453f-ae4a-5ffc76fc2c46", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:46:28.053Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7e0aaeae-fa8f-453f-ae4a-5ffc76fc2c46", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:46:29.398Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7e0aaeae-fa8f-453f-ae4a-5ffc76fc2c46"} 2023-12-29T11:46:34.608Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "82b06fbe-f51e-445e-ad35-264284307bac", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:46:34.717Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "82b06fbe-f51e-445e-ad35-264284307bac", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:46:34.837Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "82b06fbe-f51e-445e-ad35-264284307bac", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:46:36.113Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "82b06fbe-f51e-445e-ad35-264284307bac"} 2023-12-29T11:46:41.372Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "dc898baf-7012-4351-9b46-17af3884b899", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:46:41.456Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "dc898baf-7012-4351-9b46-17af3884b899", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:46:41.540Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "dc898baf-7012-4351-9b46-17af3884b899", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:46:42.812Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "dc898baf-7012-4351-9b46-17af3884b899"} 2023-12-29T11:46:49.310Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2103e86e-1a7a-4acf-91e9-f18e541fd663", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:46:49.436Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2103e86e-1a7a-4acf-91e9-f18e541fd663", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:46:50.075Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2103e86e-1a7a-4acf-91e9-f18e541fd663", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:46:50.706Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "2103e86e-1a7a-4acf-91e9-f18e541fd663"} 2023-12-29T11:46:56.584Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "38526446-4ef3-44f8-a1d2-725bc8db3a8d", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:46:56.660Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "38526446-4ef3-44f8-a1d2-725bc8db3a8d", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:46:56.751Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "38526446-4ef3-44f8-a1d2-725bc8db3a8d", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:46:58.101Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "38526446-4ef3-44f8-a1d2-725bc8db3a8d"} 2023-12-29T11:47:03.298Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "23385465-00c2-4d5f-81c8-e3dbfccc781e", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:47:03.376Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "23385465-00c2-4d5f-81c8-e3dbfccc781e", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:47:03.472Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "23385465-00c2-4d5f-81c8-e3dbfccc781e", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:47:04.739Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "23385465-00c2-4d5f-81c8-e3dbfccc781e"} 2023-12-29T11:47:10.004Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8f381c2f-45a9-49ff-8a21-74a581ec26e6", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:47:10.081Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8f381c2f-45a9-49ff-8a21-74a581ec26e6", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:47:10.158Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8f381c2f-45a9-49ff-8a21-74a581ec26e6", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:47:11.389Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8f381c2f-45a9-49ff-8a21-74a581ec26e6"} 2023-12-29T11:47:17.108Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "807f084f-c683-4692-8247-b158d7c491cc", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:47:17.193Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "807f084f-c683-4692-8247-b158d7c491cc", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:47:17.279Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "807f084f-c683-4692-8247-b158d7c491cc", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:47:18.601Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "807f084f-c683-4692-8247-b158d7c491cc"} 2023-12-29T11:47:24.471Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "92c17ab6-f270-4173-99ca-7393312db6c0", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:47:24.796Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "92c17ab6-f270-4173-99ca-7393312db6c0", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:47:24.869Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "92c17ab6-f270-4173-99ca-7393312db6c0", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:47:30.824Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "92c17ab6-f270-4173-99ca-7393312db6c0"} 2023-12-29T11:47:31.681Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "38a0eea2-22fd-4d6d-bb1c-b45a852392a6", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:47:31.759Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "38a0eea2-22fd-4d6d-bb1c-b45a852392a6", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:47:31.847Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "38a0eea2-22fd-4d6d-bb1c-b45a852392a6", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:47:33.199Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "38a0eea2-22fd-4d6d-bb1c-b45a852392a6"} 2023-12-29T11:47:38.573Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "646881f7-4885-41e0-a6ec-8608a75f7234", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:47:38.646Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "646881f7-4885-41e0-a6ec-8608a75f7234", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:47:38.731Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "646881f7-4885-41e0-a6ec-8608a75f7234", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:47:40.003Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "646881f7-4885-41e0-a6ec-8608a75f7234"} 2023-12-29T11:47:45.444Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c6a60d6b-2726-4380-9345-7865bea64cd4", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:47:45.568Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c6a60d6b-2726-4380-9345-7865bea64cd4", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:47:45.723Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c6a60d6b-2726-4380-9345-7865bea64cd4", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:47:46.918Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c6a60d6b-2726-4380-9345-7865bea64cd4"} 2023-12-29T11:47:52.271Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fb806c4a-449e-48df-b71d-4264e2125487", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:47:52.370Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fb806c4a-449e-48df-b71d-4264e2125487", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:47:52.492Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fb806c4a-449e-48df-b71d-4264e2125487", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:47:54.033Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fb806c4a-449e-48df-b71d-4264e2125487"} 2023-12-29T11:47:58.879Z INFO backup restore request {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-aws-s3", "reconcileID": "995dae7a-36d9-489f-98f2-72da8ffa4dbf"} 2023-12-29T11:47:58.982Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "493b5016-d41c-44a2-8f97-3e905584524d", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:47:59.063Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "493b5016-d41c-44a2-8f97-3e905584524d", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:47:59.184Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "493b5016-d41c-44a2-8f97-3e905584524d", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:47:59.281Z INFO stopping cluster {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-aws-s3", "reconcileID": "995dae7a-36d9-489f-98f2-72da8ffa4dbf", "cluster": "some-name"} 2023-12-29T11:48:00.591Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "493b5016-d41c-44a2-8f97-3e905584524d"} 2023-12-29T11:48:00.822Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5fb1c58b-44c6-4911-b760-f1f80f9c131f"} 2023-12-29T11:48:42.406Z INFO starting restore {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-aws-s3", "reconcileID": "995dae7a-36d9-489f-98f2-72da8ffa4dbf", "cluster": "some-name", "backup": "on-demand-backup-aws-s3"} 2023-12-29T11:49:14.455Z INFO starting cluster {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-aws-s3", "reconcileID": "995dae7a-36d9-489f-98f2-72da8ffa4dbf", "cluster": "some-name"} 2023-12-29T11:50:07.758Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "93072099-17a5-4d8b-b26e-9b79abb82f80", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:50:12.643Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b2fc11cf-4639-4703-bec2-16c74f77c319", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:50:18.229Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "dd4b2d7c-2871-4ee2-9eee-c8c87c13ad08", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:50:23.736Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "25f55422-17e0-4d1d-8ee7-834f138d4be3", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:50:29.347Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "04e066cf-e989-4252-a897-5b5406a4f0a8", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:50:34.890Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5f59c4ff-01ad-4f54-8623-384c304fc7a2", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:50:40.438Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "77a75b92-5e87-443b-a09a-57753b6bc442", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:50:46.411Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5899939c-22b5-47a9-867d-5b15a2b78fca", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:50:52.093Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "66201d3b-bdb4-4784-8a78-ef1d4f36c805", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:50:57.636Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1fc61727-94c4-4386-a72a-236441e882aa", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:03.141Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "9990d177-2cd3-4846-a424-d9b43e4e3f7d", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:08.719Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "8caa76e7-4957-4e8e-bb51-a62b5525df20", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:14.372Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "29dd6506-0a39-4a42-b360-456525d73606", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:20.036Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e19b7622-da54-45d6-8032-c0af0e4f7b9f", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:20.120Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e19b7622-da54-45d6-8032-c0af0e4f7b9f", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:51:20.805Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1590c16c-3942-47b7-80bf-e4aa0349e07c", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:20.888Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1590c16c-3942-47b7-80bf-e4aa0349e07c", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:51:25.677Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "cc439293-56c4-4191-98ed-10c5944383b9", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:25.759Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "cc439293-56c4-4191-98ed-10c5944383b9", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:51:31.450Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0fddf84c-0ef6-446f-b595-bbc34e3e0d9c", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:31.532Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "0fddf84c-0ef6-446f-b595-bbc34e3e0d9c", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:51:37.111Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "12479faa-cfda-4f76-9835-871b73a4ece6", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:37.186Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "12479faa-cfda-4f76-9835-871b73a4ece6", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:51:42.716Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d942cba3-9ce5-41c7-97ef-88e1281215bc", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:42.797Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d942cba3-9ce5-41c7-97ef-88e1281215bc", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:51:48.408Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a882e866-7699-4132-afc3-cb649c0da283", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:48.620Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "a882e866-7699-4132-afc3-cb649c0da283", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:51:54.293Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7a96ff44-0285-40aa-88ba-d5a9b6924924", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:51:54.387Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7a96ff44-0285-40aa-88ba-d5a9b6924924", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:51:59.939Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "6ed530f4-7106-4392-a82a-b4bfd46b287e", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:00.092Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "6ed530f4-7106-4392-a82a-b4bfd46b287e", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:06.195Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d3bb4ba2-96a8-4191-a91c-1ac4bcf31024", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:06.312Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "d3bb4ba2-96a8-4191-a91c-1ac4bcf31024", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:11.951Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "36aa4d9b-0061-44af-9854-f259a6baa295", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:12.042Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "36aa4d9b-0061-44af-9854-f259a6baa295", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:17.598Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e4bb05e6-3e7b-43ab-881f-ee160bfa53a6", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:17.704Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e4bb05e6-3e7b-43ab-881f-ee160bfa53a6", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:23.488Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c564b678-c91c-43a2-b960-2c9162aa6e05", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:23.589Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c564b678-c91c-43a2-b960-2c9162aa6e05", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:29.235Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e1b9fa8a-472a-46ef-91bf-f693805284d9", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:29.314Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e1b9fa8a-472a-46ef-91bf-f693805284d9", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:29.396Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "e1b9fa8a-472a-46ef-91bf-f693805284d9", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:52:29.648Z INFO You can view xtrabackup log: 2023-12-29T11:52:31.298Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "12a949c0-30be-447a-b7c4-06120139c4e6", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:52:31.371Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "12a949c0-30be-447a-b7c4-06120139c4e6", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:31.475Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "12a949c0-30be-447a-b7c4-06120139c4e6", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:32.776Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "12a949c0-30be-447a-b7c4-06120139c4e6"} 2023-12-29T11:52:36.091Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1a4d8304-e5bb-4e5b-b267-ac963abd2ea8", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:36.167Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1a4d8304-e5bb-4e5b-b267-ac963abd2ea8", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:36.247Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1a4d8304-e5bb-4e5b-b267-ac963abd2ea8", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:52:37.459Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "1a4d8304-e5bb-4e5b-b267-ac963abd2ea8"} 2023-12-29T11:52:42.756Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "6117b1f2-c059-4b46-b9a4-b4428a9299eb", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:52:42.829Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "6117b1f2-c059-4b46-b9a4-b4428a9299eb", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:42.917Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "6117b1f2-c059-4b46-b9a4-b4428a9299eb", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:44.283Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "6117b1f2-c059-4b46-b9a4-b4428a9299eb"} 2023-12-29T11:52:49.329Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b6136e34-3b28-45cc-b7b3-06d35fd5207d", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:52:49.409Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b6136e34-3b28-45cc-b7b3-06d35fd5207d", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:49.491Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b6136e34-3b28-45cc-b7b3-06d35fd5207d", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:50.941Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b6136e34-3b28-45cc-b7b3-06d35fd5207d"} 2023-12-29T11:52:55.856Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "42a167a3-fcc8-4b67-beea-35fc87201300", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:52:55.939Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "42a167a3-fcc8-4b67-beea-35fc87201300", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:52:56.033Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "42a167a3-fcc8-4b67-beea-35fc87201300", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:52:57.464Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "42a167a3-fcc8-4b67-beea-35fc87201300"} 2023-12-29T11:53:02.577Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5c21dabd-7995-474e-a252-ce578673e459", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:53:02.651Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5c21dabd-7995-474e-a252-ce578673e459", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:53:02.729Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5c21dabd-7995-474e-a252-ce578673e459", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:53:04.088Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5c21dabd-7995-474e-a252-ce578673e459"} 2023-12-29T11:53:09.564Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fb5c9991-af88-4db8-9bd3-46a5839d9e13", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:53:09.645Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fb5c9991-af88-4db8-9bd3-46a5839d9e13", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:53:09.723Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fb5c9991-af88-4db8-9bd3-46a5839d9e13", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:53:11.143Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fb5c9991-af88-4db8-9bd3-46a5839d9e13"} 2023-12-29T11:53:16.927Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5a8f8555-7c4e-4677-b89e-dd4af5965268", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:53:17.008Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5a8f8555-7c4e-4677-b89e-dd4af5965268", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:53:17.095Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5a8f8555-7c4e-4677-b89e-dd4af5965268", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:53:18.281Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5a8f8555-7c4e-4677-b89e-dd4af5965268"} 2023-12-29T11:53:23.875Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7b3d741e-02f9-403b-8e92-342852ecab04", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:53:23.954Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7b3d741e-02f9-403b-8e92-342852ecab04", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:53:24.039Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7b3d741e-02f9-403b-8e92-342852ecab04", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:53:25.379Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "7b3d741e-02f9-403b-8e92-342852ecab04"} 2023-12-29T11:53:30.886Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "49b06487-f605-4011-a317-452709455e5e", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:53:30.970Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "49b06487-f605-4011-a317-452709455e5e", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:53:31.052Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "49b06487-f605-4011-a317-452709455e5e", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:53:32.062Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "49b06487-f605-4011-a317-452709455e5e"} 2023-12-29T11:53:37.576Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b364cfe2-89d7-43d0-b78b-87511e35ca76", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:53:37.666Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b364cfe2-89d7-43d0-b78b-87511e35ca76", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:53:37.765Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b364cfe2-89d7-43d0-b78b-87511e35ca76", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:53:39.263Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "b364cfe2-89d7-43d0-b78b-87511e35ca76"} 2023-12-29T11:53:44.546Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "44c4c7a6-47e9-4d91-a478-9b73219a8506", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:53:44.632Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "44c4c7a6-47e9-4d91-a478-9b73219a8506", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:53:44.721Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "44c4c7a6-47e9-4d91-a478-9b73219a8506", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:53:46.178Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "44c4c7a6-47e9-4d91-a478-9b73219a8506"} 2023-12-29T11:53:51.377Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "829d5bf7-e3cb-4621-8a8c-81d5744bb2e0", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:53:51.456Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "829d5bf7-e3cb-4621-8a8c-81d5744bb2e0", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:53:51.535Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "829d5bf7-e3cb-4621-8a8c-81d5744bb2e0", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:53:52.984Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "829d5bf7-e3cb-4621-8a8c-81d5744bb2e0"} 2023-12-29T11:53:58.067Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fbd59765-d0e1-4efa-9a6a-7bd831e351ab", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:53:58.145Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fbd59765-d0e1-4efa-9a6a-7bd831e351ab", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:53:58.242Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fbd59765-d0e1-4efa-9a6a-7bd831e351ab", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:53:59.642Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "fbd59765-d0e1-4efa-9a6a-7bd831e351ab"} 2023-12-29T11:54:04.873Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "96091b75-0e25-49b7-9443-4239ba9f4956", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:54:04.962Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "96091b75-0e25-49b7-9443-4239ba9f4956", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:54:05.062Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "96091b75-0e25-49b7-9443-4239ba9f4956", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:54:06.436Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "96091b75-0e25-49b7-9443-4239ba9f4956"} 2023-12-29T11:54:11.573Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "49bbb7c6-3ac0-4b70-a178-bb02236c4a2e", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:54:11.648Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "49bbb7c6-3ac0-4b70-a178-bb02236c4a2e", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:54:11.731Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "49bbb7c6-3ac0-4b70-a178-bb02236c4a2e", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:54:13.146Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "49bbb7c6-3ac0-4b70-a178-bb02236c4a2e"} 2023-12-29T11:54:18.251Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5e798f4b-f89f-4e7a-b8b7-f441c500aecc", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:54:18.336Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5e798f4b-f89f-4e7a-b8b7-f441c500aecc", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:54:18.409Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5e798f4b-f89f-4e7a-b8b7-f441c500aecc", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:54:19.791Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "5e798f4b-f89f-4e7a-b8b7-f441c500aecc"} 2023-12-29T11:54:24.960Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c3f2de9e-0f1c-4f4b-8847-c2a1e25dadad", "pod": "some-name-pxc-2", "isReplica": false, "readonly": false} 2023-12-29T11:54:25.049Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c3f2de9e-0f1c-4f4b-8847-c2a1e25dadad", "pod": "some-name-pxc-1", "isReplica": false, "readonly": false} 2023-12-29T11:54:25.140Z DEBUG Read only status is correct {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c3f2de9e-0f1c-4f4b-8847-c2a1e25dadad", "pod": "some-name-pxc-0", "isReplica": false, "readonly": false} 2023-12-29T11:54:26.491Z DEBUG PXC users synced with ProxySQL {"controller": "pxc-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "some-name", "reconcileID": "c3f2de9e-0f1c-4f4b-8847-c2a1e25dadad"} {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-aws-s3", "reconcileID": "995dae7a-36d9-489f-98f2-72da8ffa4dbf"} {"controller": "pxcrestore-controller", "namespace": "restore-to-encrypted-cluster-4709", "name": "on-demand-backup-pvc", "reconcileID": "5825267a-6276-42d5-947c-cec71fc99c41"} github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc.(*ReconcilePerconaXtraDBCluster).resyncPXCUsersWithProxySQL.func1 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:227 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:266 /go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.1/pkg/internal/controller/controller.go:329 /go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxc/controller.go:1231 If everything is fine, you can cleanup the job: $ kubectl delete pxc-restore/on-demand-backup-aws-s3 $ kubectl delete pxc-restore/on-demand-backup-pvc $ kubectl logs job/restore-job-on-demand-backup-aws-s3-some-name $ kubectl logs job/restore-job-on-demand-backup-pvc-some-name sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2 + kubectl get pxc --all-namespaces -o wide + grep -v NAMESPACE + xargs -L 1 sh -xc 'kubectl patch pxc -n $0 $1 --type=merge -p "{\"metadata\":{\"finalizers\":[]}}"' + kubectl patch pxc -n restore-to-encrypted-cluster-4709 some-name --type=merge -p '{"metadata":{"finalizers":[]}}' perconaxtradbcluster.pxc.percona.com/some-name patched + kubectl_bin delete pxc --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.tDsF6dhKax ++ mktemp + local LAST_ERR=/tmp/tmp.56BaC7szyq + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.tDsF6dhKax perconaxtradbcluster.pxc.percona.com "some-name" deleted + cat /tmp/tmp.56BaC7szyq + rm /tmp/tmp.tDsF6dhKax /tmp/tmp.56BaC7szyq + return 0 + kubectl_bin delete pxc-backup --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.sQHmZWuePQ ++ mktemp + local LAST_ERR=/tmp/tmp.myz3cuxSaA + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-backup --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.sQHmZWuePQ perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-aws-s3" deleted perconaxtradbclusterbackup.pxc.percona.com "on-demand-backup-pvc" deleted + cat /tmp/tmp.myz3cuxSaA + rm /tmp/tmp.sQHmZWuePQ /tmp/tmp.myz3cuxSaA + return 0 + kubectl_bin delete pxc-restore --all --all-namespaces ++ mktemp + local LAST_OUT=/tmp/tmp.0lgGwb6Okx ++ mktemp + local LAST_ERR=/tmp/tmp.v0aTeAfaQE + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete pxc-restore --all --all-namespaces + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.0lgGwb6Okx perconaxtradbclusterrestore.pxc.percona.com "on-demand-backup-aws-s3" deleted perconaxtradbclusterrestore.pxc.percona.com "on-demand-backup-pvc" deleted + cat /tmp/tmp.v0aTeAfaQE + rm /tmp/tmp.0lgGwb6Okx /tmp/tmp.v0aTeAfaQE + return 0 + kubectl_bin delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook ++ mktemp + local LAST_OUT=/tmp/tmp.euT7WVsIpy ++ mktemp + local LAST_ERR=/tmp/tmp.AtIrkKiWhr + local exit_status=0 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete ValidatingWebhookConfiguration percona-xtradbcluster-webhook + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.Lw2mv2UNIf namespace "vault-service-1-7243" force deleted + cat /tmp/tmp.IoElcph4DM Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. + rm /tmp/tmp.Lw2mv2UNIf /tmp/tmp.IoElcph4DM + return 0 + exit_status=0 + set -e + '[' 0 '!=' 0 ']' + break + cat /tmp/tmp.euT7WVsIpy validatingwebhookconfiguration.admissionregistration.k8s.io "percona-xtradbcluster-webhook" deleted + cat /tmp/tmp.AtIrkKiWhr + rm /tmp/tmp.euT7WVsIpy /tmp/tmp.AtIrkKiWhr + return 0 + kubectl_bin delete -f https://github.com/jetstack/cert-manager/releases/download/v1.12.1/cert-manager.yaml + : + '[' '!' -z '' ']' + '[' -n pxc-operator ']' + kubectl_bin delete --grace-period=0 --force=true namespace restore-to-encrypted-cluster-4709 + rm -rf /tmp/tmp.LW7qwu72v2 ++ mktemp + kubectl_bin delete --grace-period=0 --force=true namespace pxc-operator ++ mktemp + local LAST_OUT=/tmp/tmp.0d6J7tMfvv + local LAST_OUT=/tmp/tmp.Hf7sTXMBrl ++ mktemp + desc 'test passed' + set +o xtrace ----------------------------------------------------------------------------------- test passed ----------------------------------------------------------------------------------- ++ mktemp + local LAST_ERR=/tmp/tmp.f4TpxpcIor + local exit_status=0 + local LAST_ERR=/tmp/tmp.66lWxA3iXm + local exit_status=0 ++ seq 0 2 ++ seq 0 2 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace restore-to-encrypted-cluster-4709 + for i in '$(seq 0 2)' + set +e + kubectl delete --grace-period=0 --force=true namespace pxc-operator