text
stringlengths 9
7.94M
| subset
stringclasses 1
value | meta
dict | file_path
stringclasses 1
value | question
dict | answers
listlengths |
|---|---|---|---|---|---|
\begin{document}
\title{\bf A Generalized Heckman Model With Varying Sample Selection Bias and Dispersion Parameters}
\author[1,3]{Fernando de S. Bastos\thanks{E-mail: \texttt{[email protected]}}} \affil[1]{\it Instituto de Ci\^encias Exatas e Tecnol\'ogicas, Universidade Federal de Vi\c cosa, Florestal, Brazil}
\author[2,3]{Wagner Barreto-Souza\thanks{E-mail: \texttt{[email protected]}}} \affil[2]{\it Statistics Program, King Abdullah University of Science and Technology, Thuwal, Saudi Arabia} \affil[3]{\it Departamento de Estat\'istica, Universidade Federal de Minas Gerais, Belo Horizonte, Brazil}
\author[2]{Marc G. Genton\thanks{E-mail: \texttt{[email protected]}}}
\date{}
\maketitle
\begin{abstract} Many proposals have emerged as alternatives to the Heckman selection model, mainly to address the non-robustness of its normal assumption. The 2001 Medical Expenditure Panel Survey data is often used to illustrate this non-robustness of the Heckman model. In this paper, we propose a generalization of the Heckman sample selection model by allowing the sample selection bias and dispersion parameters to depend on covariates. We show that the non-robustness of the Heckman model may be due to the assumption of the constant sample selection bias parameter rather than the normality assumption. Our proposed methodology allows us to understand which covariates are important to explain the sample selection bias phenomenon rather than to only form conclusions about its presence. We explore the inferential aspects of the maximum likelihood estimators (MLEs) for our proposed generalized Heckman model. More specifically, we show that this model satisfies some regularity conditions such that it ensures consistency and asymptotic normality of the MLEs. Proper score residuals for sample selection models are provided, and model adequacy is addressed. Simulated results are presented to check the finite-sample behavior of the estimators and to verify the consequences of not considering varying sample selection bias and dispersion parameters. We show that the normal assumption for analyzing medical expenditure data is suitable and that the conclusions drawn using our approach are coherent with findings from prior literature. Moreover, we identify which covariates are relevant to explain the presence of sample selection bias in this important dataset. \end{abstract}
\noindent {\it \textbf{Keywords}:} Asymptotics; Heteroscedasticity; Regularity conditions; Score residuals; Varying sample selection bias.
\doublespacing
\section{Introduction} \label{sec:intro}
\noindent \citet{heckman1974,heckman1976} introduced a model for dealing with the sample selection bias problem with the aid of a bivariate normal distribution to relate the outcome of interest and a selection rule. A semiparametric alternative to this model, known as Heckman’s two-step method, was proposed by \cite{heckman1979}, so that the non-robustness of the normal distribution in the presence of outliers could be handled.
The most discussed problem regarding the Heckman model is its sensitivity to the normal assumption of errors. Misspecification of the error distribution leads to inconsistent maximum likelihood estimators, yielding biased estimates \citep{TsayPin}. On the other hand, when the error terms are correctly specified, the estimation by maximum likelihood or by procedures based on likelihood produces consistent and efficient estimators \citep{leung1996choice,enders2010applied}.
However, even when the shape of the error density is correctly specified, the heteroskedasticity of the error terms can cause inconsistencies in the parameter estimates, as shown by \cite{hurd1979estimation} and \cite{arabmazar1981further}. In response to this concern, \cite{donald1995two} discussed how heteroskedasticity in sample selection models is relatively neglected and provided two reasons to motivate the importance of taking this into account in practice. The first reason is that typically, the data used to fit sample selection models comprises large databases, in which heterogeneity is commonly found. The second reason is that the estimates of the parameters obtained by fitting the usual selection models may in some cases be more severely affected by heteroscedasticity than by incorrect distribution of the error terms \citep{powell1986symmetrically}.
Nevertheless, even though there is a large body of recent research and studies on sample selection models, few studies have been carried out to correct or minimize the impact of heteroscedasticity. This is one of the aims of our paper. \cite{Siddhartha} proposed a semiparametric model for data with sample selection bias. They considered nonparametric functions in their model, which allowed great flexibility in the way covariates affect response variables. They still presented a Bayesian method for the analysis of such models. Subsequently, \cite{Wiesenfarth} introduced another general estimation method based on Markov Chain Monte Carlo simulation techniques and used a simultaneous equation system that incorporates Bayesian versions of penalized smoothing splines.
Recent works on sample selection models have aimed to address robust alternatives to the Heckman model. In this direction, \cite{marchenkoGenton} proposed a Student-$t$ sample selection model for dealing with the robustness to the normal assumption in the Heckman model. \cite{zhelonkin2013robustness} proposed a modified robust semiparametric alternative based on Heckman's two-step estimation method. They proved the asymptotic normality of the proposed estimators and provided the asymptotic covariance matrix. To deal with departures from normality due to skewness, \cite{ogundimu2016sample} introduced the skew-normal sample selection model to mitigate the remaining effect of skewness after applying a logarithmic transformation to the outcome variable.
Another direction that has been explored in the last few years is the modeling of discrete data with sample selection. For instance, \cite{MARRA2016110} introduced sample selection models for count data, potentially allowing for the use of any discrete distribution, non-Gaussian dependencies between the selection and outcome equations, and flexible covariate effects. The modeling of zero-inflated count data with sample selection bias is discussed by \cite{Wysmar2018}. \cite{Beili} considered the semiparametric identification and estimation of a heteroscedastic binary choice model with endogenous dummy regressors, and no parametric restriction on the distribution of the error term was assumed. This yields general multiplicative heteroscedasticity in both selection and outcome equations and multiple discrete endogenous regressors. A class of sample selection models for discrete and other non-Gaussian outcomes was recently proposed by \cite{azzalinietal}.
\cite{Giampiero} introduced a generalized additive model for location, scale and shape, which accounts for non-random sample selection, and \cite{Taeyoung} proposed a Bayesian methodology to correct the bias of estimation of the sample selection models based on a semiparametric Bernstein polynomial regression model that incorporates the sample selection scheme into a stochastic monotone trend constraint, variable selection, and robustness against departures from the normality assumption.
In the aforementioned papers, the solution to deal with departures from normality for continuous outcomes is to assume robust alternatives like Student-$t$ or skew-normal distributions. Another common approach is to consider non-parametric structures for the density of the error terms. Our proposal goes in a different direction, one which has not yet been explored in the literature.
In this paper, we propose a generalization of the Heckman sample selection model by allowing the sample selection bias and dispersion parameters to depend on covariates. We show that the non-robustness of the Heckman model may be due to the assumption of the constant sample selection bias parameter rather than the normality assumption. Our proposed methodology allows us to understand what covariates are important to explain the sample selection bias phenomenon rather than to solely make conclusions about its presence. It is worthy to mention that our methodology can be straightforwardly adapted for existing sample selection like those proposed by \cite{marchenkoGenton}, \cite{ogundimu2016sample}, and \cite{Wysmar2018}. We now highlight other contributions of our paper:
\begin{itemize}
\item We explore the inferential aspects of the maximum likelihood estimators (MLEs) for our proposed generalized Heckman model. More specifically, we show that this model satisfies regularity conditions so that it ensures consistency and asymptotic normality of the MLEs. In particular, we show that the Heckman model satisfies the regularity conditions, which is a new finding.
\item Proper residual for sample selection models is proposed as a byproduct of our asymptotic analysis. This is another relevant contribution of our paper, since this point has not yet been thoroughly addressed.
\item We develop an \texttt{R} package for fitting our proposed generalized Heckman model. It also includes the Student-$t$ and skew-normal sample selection models, which have not been implemented in \texttt{R} \citep{SoftwareR} before. This makes the paper replicable and facilitates the use of our generalized Heckman model by practitioners.
\item We show that the normal assumption for analyzing medical expenditure data is suitable and that the conclusions drawn using our approach are consistent with findings from prior literature. Moreover, we identify which covariates are relevant for explaining the presence of sample selection bias in this important dataset. \end{itemize}
This paper is organized as follows. In Section \ref{HeckmanGen} we define the generalized Heckman (GH) sample selection model and discuss estimation of the parameters through the maximum likelihood method. Further, diagnostics tools and residual analysis are discussed. Section \ref{asymptotics} is devoted to showing that the GH model satisfies regularity conditions that ensure consistency and asymptotic normality of the maximum likelihood estimators. In Section \ref{simulations}, we present Monte Carlo simulation results for evaluating the performance of the maximum likelihood estimators of our proposed model and for checking the behavior of other existing methodologies under misspecification. In Section \ref{3:application} we apply our generalized Heckman model to the data on ambulatory expenditures from the 2001 Medical Expenditure Panel Survey and show that our methodology overcomes an existing problem in a simple way. Concluding remarks are addressed in Section \ref{concluding_remarks}. This paper contains a Supplementary Material which can be obtained from the authors upon request.
\section{Generalized Heckman Model}\label{HeckmanGen}
Assume that $\{(Y_{1i}^{*},Y_{2i}^{*})\}_{i=1}^n$ are linearly related to covariates $\pmb{x}_{i}\in \mathbb{R}^{p}$ and $\pmb{w}_{i}\in \mathbb{R}^{q}$ through the following regression structures: \begin{align} Y_{1i}^{*}&=\mu_{1i}+\epsilon_{1i}, \label{eq_reg_heckman} \\ Y_{2i}^{*}&=\mu_{2i}+\epsilon_{2i}, \label{eq_sel_heckman} \end{align} where $\mu_{1i}=\pmb{x}_{i}^\top\pmb\beta$, $\mu_{2i}=\pmb w_{i}^\top\pmb\gamma$, $\pmb{\beta}=(\beta_1,\ldots,\beta_p)^\top\in \mathbb{R}^{p}$ and $\pmb{\gamma}=(\gamma_1,\ldots,\gamma_q)^\top\in \mathbb{R}^{q}$ are vectors of unknown parameters with associated covariate vectors $\pmb x_i$ and $\pmb w_i$, for $i=1,\ldots,n$, and $\{(\epsilon_{1i},\epsilon_{2i})\}_{i=1}^n$ is a sequence of independent bivariate normal random vectors. More specifically, we suppose that \begin{align}\label{2:disterro2} \begin{pmatrix} \epsilon_{1i}\\ \epsilon_{2i} \end{pmatrix} &\sim \mathcal{N}_2 \begin{bmatrix} \begin{pmatrix} 0\\ 0 \end{pmatrix}\!\!,& \begin{pmatrix} \sigma_{i}^{2} & \rho_{i}\sigma_{i} \\ \rho_{i}\sigma_{i} & 1 \end{pmatrix} \end{bmatrix}, \end{align} with the following regression structures for the sample selection bias and dispersion parameters: \begin{eqnarray}\label{addreg} \mbox{arctanh}\,\rho_i=\pmb v_i^\top\pmb\kappa \quad \mbox{and}\quad \log\sigma_i=\pmb e_i^\top\pmb\lambda, \end{eqnarray} where $\pmb{\lambda}=(\lambda_{1},\ldots,\lambda_{r})^\top\in \mathbb{R}^{r}$ and $\pmb{\kappa}=(\kappa_{1},\ldots,\kappa_{s})^\top\in \mathbb{R}^{s}$ are parameter vectors, with associated covariate vectors $\pmb v_i$ and $\pmb e_i$, for $ i=1,\ldots,n$. The $\mbox{arctanh}$ (inverse hyperbolic tangent) link function used for the sample selection bias parameter ensures that it belongs to the interval $(-1,1)$. The variable $Y_{1i}^{*}$ is observed only if $Y_{2i}^{*}>0$ while the variable $Y_{2i}^{*}$ is latent. We only know if $Y_{2i}^{*}$ is greater or less than 0. The equation (\ref{eq_reg_heckman}) is the primary interest equation and the equation (\ref{eq_sel_heckman}) represents the selection equation. In practice, we observe the variables \begin{align} U_{i}&=I\{Y_{2i}^{*}>0\},\\ Y_{i}&=Y_{1i}^{*}U_{i},\label{mod_heckman_ind} \end{align} for $ i=1,\ldots,n$, where $I\{Y_{2i}^{*}>0\}=1$ if $Y_{2i}^{*}>0$ and equals 0 otherwise.
Our generalized Heckman sample selection model is defined by (\ref{eq_reg_heckman})-(\ref{mod_heckman_ind}). The classic Heckman model is obtained by assuming constant sample selection bias and dispersion parameters in (\ref{addreg}).
The mixed distribution of $Y_{i}$ is composed by the discrete component \begin{align} P(U_i=u_i)=\Phi(\pmb{w}_{i}^\top\boldsymbol{\gamma})^{u_i}\Phi(-\pmb{w}_{i}^\top\boldsymbol{\gamma})^{1-u_i},\quad u_i=0,1, \end{align} and by a continuous part given by the conditional density function \begin{align} \label{1:dens_heckman}
f(y_{i}|U_i=1)&=\dfrac{1}{\sigma_{i}}\phi\left(\dfrac{y_{i}-\pmb{x}_{i}^\top\boldsymbol{\beta}}{\sigma_{i}}\right) \dfrac{\Phi\left(\dfrac{\rho_{i}}{\sqrt{1-\rho_{i}^{2}}}\left(\dfrac{y_{i}-\pmb{x}_{i}^\top\boldsymbol{\beta}}{\sigma_{i}}\right)+ \dfrac{\pmb{w}_{i}^\top\boldsymbol{\gamma}}{\sqrt{1-\rho_{i}^{2}}}\right)}{\Phi(\pmb{w}_{i}^\top\boldsymbol{\gamma})}, \end{align} where $\phi(\cdot)$ and $\Phi(\cdot)$ denote the density and cumulative distribution functions of the standard normal, respectively. More details about this are provided in the Supplementary Material.
Let $\boldsymbol{\theta}=(\pmb{\beta}^{\top},\pmb{\gamma}^{\top},\pmb{\lambda}^{\top},\pmb{\kappa}^{\top})^{\top}$ be the parameter vector. The log-likelihood function is given by \begin{align}\label{likelihoodGEN}
\mathcal{L}(\boldsymbol{\theta})&=\sum_{i=1}^{n} \left\{u_{i}\log{f(y_{i}|u_{i}=1)}+u_{i}\log{\Phi(\mu_{2i})}+ (1-u_{i})\log{\Phi(-\mu_{2i})}\right\}\nonumber\\ \displaystyle &=\sum_{i=1}^{n}u_{i}\left\{\log{\Phi(\zeta_{i})}+ \log{\phi(z_{i})}-\log{\sigma_{i}}\right\}+\sum_{i=1}^{n}(1-u_{i})\log{\Phi(-\mu_{2i})}, \end{align} where $u_{i}=1$ if $y_{i}$ is observed and $u_{i}=0$ otherwise, $ z_{i}\equiv\dfrac{y_{i}-\mu_{1i}}{\sigma_{i}}$, and $\zeta_{i}\equiv\dfrac{\mu_{2i}+\rho_iz_{i}}{\sqrt{1-\rho_i^2}}$, for $i=1,\ldots,n$.
Expressions for the score function ${\bf S}_{\boldsymbol{\theta}}=\dfrac{\partial \mathcal{L}(\boldsymbol{\boldsymbol{\theta}})}{\partial\boldsymbol{\theta}}$ and the respective Hessian matrix are presented in the Supplementary Material. The maximum likelihood estimators (MLEs) are obtained as solution of the non-linear system of equations ${\bf S}_{\boldsymbol{\theta}}={\bf 0}$, which does not have an explicit analytic form. We use the quasi-Newton algorithm of Broyden-Fletcher-Goldfarb-Shanno (BFGS) via the package \texttt{optim} of the software \texttt{R} \citep{SoftwareR} to maximize the log-likelihood function.
It is important to emphasize that the maximum likelihood method may suffer from possible multicollinearity problems when the selection equation has the same covariates as the regression equation (for example, see \cite{marchenkoGenton}). To reduce the impact of this problem in parameter estimation, the exclusion restriction is suggested in the literature. According to this approach, at least one significant covariate included in the selection equation should not be included in the primary regression. The interested reader can find more details on the exclusion restriction procedure for the Heckman sample selection model in \cite{heckman1976}, \cite{Leung2000} and \cite{Newey2009}.
We now discuss diagnostic techniques, which have been proposed to detect observations that could exercise some influence on the parameter estimates or inference in general. Next, for the generalized Heckman model, we describe the generalized Cook distance (GCD), and in the next section, we propose the score residual.
Cook's distance is a method commonly used in statistical modeling to evaluate changes in the estimated vector of parameters when observations are deleted. It allows us to assess the effect of each observation on the estimated parameters. The methodology proposed by \cite{RDCook} suggests the deletion of each observation and the evaluation of the log-likelihood function without such a case. According to \cite{XIE20074692}, the generalized Cook distance (GCD) is defined by \begin{align*} \mathrm{GCD}_{i}(\boldsymbol{\theta})=\left(\widehat{\boldsymbol{\theta}}-\widehat{\boldsymbol{\theta}}_{(i)}\right)^\top\boldsymbol{M} \left(\widehat{\boldsymbol{\theta}}-\widehat{\boldsymbol{\theta}}_{(i)}\right), \quad i=1, \ldots, n, \end{align*} where $\boldsymbol{M}$ is a nonnegative definite matrix, which measures the weighted combination of the elements for the difference $\widehat{\boldsymbol{\theta}}-\widehat{\boldsymbol{\theta}}_{(i)}$, and $\widehat{\boldsymbol{\theta}}_{(i)}$ is the MLE of $\boldsymbol\theta$ when removing the $i$th observation. Many choices for $\boldsymbol{M}$ were considered by \cite{cook1982residuals}. We use the inverse variance-covariance matrix $\boldsymbol{M}=-\ddot{\mathcal{L}}(\widehat{\boldsymbol{\theta}})^{-1}.$ To determine whether the $i$th case is potentially influential on inference about $\boldsymbol\theta$, we check if its associated GCD value is greater than $2p/n$. In this case, this point would be a possible influential observation.
We illustrate the usage of the GCD in the analysis of the medical expenditure data in Section~\ref{3:application}. Regarding the residual analysis, in the next section, we propose a proper residual for sample selection models, which is one of the aims of this paper.
\section{Asymptotic Properties and Score Residuals}\label{asymptotics}
Our aim in this section is to show that under some conditions, our proposed generalized Heckman sample selection model satisfies the regularity conditions stated by \cite{cox1979theoretical}. As a consequence, the maximum likelihood estimators discussed in the previous section are consistent and asymptotically normal distributed. As a by-product of our findings here, we propose a score residual that is well-known to be approximately normal distributed. Proofs of the theorems stated in this section can be found in the Appendix.
Let $\boldsymbol\Theta$ be the parameter space and $\ell_i(\boldsymbol\theta)= u_{i}\left\{\log{\Phi(\zeta_{i})}+\log{\phi(z_{i})}-\log{\sigma_{i}}\right\}+(1-u_{i})\log{\Phi(-\mu_{2i})}$ be the contribution of the $i$th observation to the log-likelhood function, where $\zeta_{i}$ retains its definition from previous section, for $i=1,\ldots,n$.
\begin{theorem}\label{regcond} The score function associated to the generalized Heckman model has mean zero and satisfies the identity $E\left({\bf S}_{\boldsymbol{\theta}}{\bf S}_{\boldsymbol{\theta}}^\top\right)=-E\left(\partial {\bf S}^\top_{\boldsymbol{\theta}}/\partial\boldsymbol\theta\right)$. \end{theorem}
We now propose a new residual for sample selection models inspired from Theorem \ref{regcond}. From (\ref{scorebeta}), we define the ordinary score residual by $s_i=z_i-\dfrac{\rho_i}{\sqrt{1-\rho_i^2}}\dfrac{\phi(\zeta_{i})}{\Phi(\zeta_{i})}$ for the non-censored observations (where $u_i=1$) and the standardized score residual by \begin{eqnarray*}
S_i=\dfrac{s_i-E(s_i|U_i=1)}{\sqrt{\mbox{Var}(s_i|U_i=1)}}=\dfrac{z_i-\dfrac{\rho_i}{\sqrt{1-\rho_i^2}}\dfrac{\phi(\zeta_{i})}{\Phi(\zeta_{i})}}{\sqrt{1+\mu_{2i}\rho_i^2\dfrac{\phi(\mu_{2i})}{\Phi(\mu_{2i})}+\dfrac{\rho_i^2}{1-\rho_i^2}\Psi_i}}, \end{eqnarray*}
where $\Psi_i=E\left(\dfrac{\phi^2(\zeta_i)}{\Phi^2(\zeta_i)}\big|U_i=1\right)=\dfrac{1}{\sigma_i\Phi(\mu_{2i})}\displaystyle\int_{-\infty}^\infty\phi(z_i)\phi(\zeta_i)/\Phi(\zeta_i)dy_i$, for $i=1,\ldots,n$ such that $u_i=1$. Alternatively, a score residual based on all observations (including the censored ones) can be defined by \begin{eqnarray}\label{score_residual} S^*_i=\dfrac{s_i-E(s_i)}{\sqrt{\mbox{Var}(s_i)}}=\dfrac{z_i-\dfrac{\rho_i}{\sqrt{1-\rho_i^2}}\dfrac{\phi(\zeta_{i})}{\Phi(\zeta_{i})}}{\sqrt{\Phi(\mu_{2i})\left(1+\mu_{2i}\rho_i^2\dfrac{\phi(\mu_{2i})}{\Phi(\mu_{2i})}+\dfrac{\rho_i^2}{1-\rho_i^2}\Psi_i\right)}}, \end{eqnarray} for $i=1,\ldots,n$. In practice, we replace the unknown parameters by their maximum likelihood estimates. The evaluation of the goodness-of-fit of our proposed generalized Heckman model will be performed through the score residual analysis. Based on this approach, discrepant observations are identified, besides being possible to evaluate the existence of serious departures from the assumptions inherent to the model. If the model is appropriate, plots of residuals versus predicted values should have a random behavior around zero. Alternatively, a common approach is to build residual graphics with simulated envelopes \citep{atkinson1985}. In this case, it is not necessary to know about the distribution of the residuals, they just need to be within the region formed by the envelopes so indicating a good fit. Otherwise, residuals outside the envelopes are possible outliers or indicate that the model is not properly specified. We will apply the proposed score residual (\ref{score_residual}) to the MEPS data analysis. As will be shown, the residual analysis indicates that the normal assumption for the data is suitable in contrast with the non-robustness of the Heckman model mentioned in existing papers in the literature on sample selection models.
We now establish the consistency and asymptotic normality of the MLEs for our proposed generalized Heckman model. For this, we need to assume some usual regularity conditions.
\noindent (C1) The parameter space $\boldsymbol\Theta$ is closed and compact, and the true parameter value, say $\boldsymbol\theta_0$, is an interior point of $\boldsymbol\Theta$.
\noindent (C2) The covariates are a sequence of iid random vectors, and ${\bf F}_n$ is the information matrix conditional on the covariates.
\noindent (C3) $E({\bf F}_n)$ is well-defined and positive definite, and $E\left(\max_{\boldsymbol\theta\in\boldsymbol\Theta}\|{\bf F}_n\|\right)<\infty$, where $\|\cdot\|$ is the Frobenius norm.
\begin{remark} Conditions (C2) and (C3) enable us to apply a multivariate Central Limit Theorem for iid random vectors for establishing the asymptotic normality of the MLEs. These conditions are discussed for instance in \cite{fahkau1985}. \end{remark}
\begin{theorem}\label{asympt_results} Under Conditions (C1)--(C3), the maximum likelihood estimator $\widehat{\boldsymbol\theta}$ of $\boldsymbol\theta$ for the generalized Heckman model is consistent and satisfies the weak convergence: ${\bf F}_n^{1/2}\left(\widehat{\boldsymbol\theta}-\boldsymbol\theta_0\right)\stackrel{d}{\longrightarrow}{\cal N}({\bf0},\bf{I})$, where $\bf{I}$ is the identity matrix, ${\bf F}_n$ is the conditional information matrix, and $\boldsymbol\theta_0$ denotes the true parameter vector value. \end{theorem}
\begin{remark} An important consequence of Theorem \ref{asympt_results} is that the classic Heckman model is regular under Conditions (C1)--(C3). Therefore, the MLEs for this model are consistent and asymptotically normal distributed, which is a new finding of this paper. \end{remark}
\section{Monte Carlo Simulations}\label{simulations}
\subsection{Simulation design} In this section, we develop Monte Carlo simulation studies to evaluate and compare the performance of the maximum likelihood estimators under the generalized Heckman, classic Heckman, Heckman-Skew, and Heckman-$t$ models when the assumption of either constant sample selection bias parameter or constant dispersion is not satisfied. To do this, six different scenarios with relevant characteristics for a more detailed evaluation were considered. In Scenarios 1 and 2, we use models with both varying dispersion and correlation (sample selection bias parameter) and with (I) exclusion restriction and (II) without the exclusion restriction.
For Scenarios 3-6, the exclusion restriction is considered. More specifically, in Scenarios 3, 4, and 5, we have specified the following (III) constant dispersion and varying correlation; (IV) varying dispersion and constant correlation; (V) both constant dispersion and correlation. To evaluate the sensitivity in the parameter estimation of the selection models at high censoring, in Scenario 6 we simulated from the generalized Heckman model with (VI) varying both sample selection bias and dispersion parameters and an average censorship of $50\%.$
Scenario 1 aims to evaluate the performance of the generalized Heckman model and compare it with its competitors when the assumption of constant sample selection bias parameter and dispersion is not satisfied. Scenario 2 is devoted to demonstrating that despite the absence of exclusion restriction, our model can yield satisfactory parameter estimates. Scenarios 3 and 4 aim to justify the importance of modeling through covariates the correlation and dispersion parameters, respectively. Scenario 5 illustrates some problems that the generalized Heckman model can face as with the classic Heckman model. Finally, Scenario 6 was included to demonstrate the sensitivity of selection models to high correlation and high censoring. We here present the results from Scenario~1. The remaining results are presented in the Supplementary Material.
All scenarios were based on the following regression structures: \begin{align} \mu_{1i}&=\beta_{0}+\beta_{1}x_{1i}+\beta_{2}x_{2i},\label{eq_sel_heckman_psim1} \\ \mu_{2i}&=\gamma_{0}+\gamma_{1}x_{1i}+\gamma_{2}x_{2i}+\gamma_{3}x_{3i},\label{eq_sel_heckman_psim2}\\ \log\sigma_{i}&=\lambda_{0}+\lambda_{1}x_{1i},\label{eq_sel_heckman_psim3}\\ \hspace{2pt}\mathrm{arctanh}\,\rho_{i}&=\kappa_{0}+\kappa_{1}x_{1i}, \label{eq_sel_heckman_psim4} \end{align} for $i=1,\ldots,n$. All covariates were generated from a standard normal distribution and were kept constant throughout the experiment. The responses were generated from the generalized Heckman model according to each of the six configurations. We set the sample sizes $n=500,1000,2000$ and $N=1000$ Monte Carlo replicates. Pilot simulations showed that the choice of parameters used in the simulations does not affect the results, as long as they maintain the same average percentage of censorship.
We would like to highlight that there is no \texttt{R} package for fitting the Heckman-$t$ and Heckman skew-normal models. Therefore, we developed an \texttt{R} package (to be submitted) able to fit our proposed generalized Heckman model and also the sample selection models by \cite{marchenkoGenton} and \cite{ogundimu2016sample}.
In the maximization procedure to estimate parameters based on the $\texttt{optim}$ package, we consider as initial values for $\pmb{\beta}$, $\pmb{\gamma}$, $\lambda_{0}$, and $\kappa_{0}$ the maximum likelihood estimates by the classic Heckman model. For the remaining parameters, we set $\lambda_{i}=0$ and $\kappa_{j}=0$ for $i=1,\ldots,r$ and $j=1,\ldots,s$. The initial values for the degrees of freedom of the Heckman-$t$ model and the skewness parameter for the Heckman-Skew model were set to $1$ and $2$, respectively. These values were chosen after some pilot simulations.
\subsection{Scenario 1: Varying sample selection bias and dispersion parameters}
Here, we consider (\ref{eq_sel_heckman_psim1})-(\ref{eq_sel_heckman_psim4}) with $\pmb\beta=(1.1,0.7,0.1)^\top$, $\pmb\gamma=(0.9,0.5,1.1,0.6)^\top$, and also $\pmb\lambda=(-0.4,0.7)^\top$ and $\pmb\kappa=(0.3,0.5)^\top$.
The regressors are kept fixed throughout the simulations with $x_{1i}, x_{2i}, x_{3i}\overset{iid}{\sim} \mathcal{N}(0,1)$ for all $i=1,\ldots,n.$
In Table \ref{2:res_cen3_tab1}, we present the empirical mean and root mean square error (RMSE) of the maximum likelihood estimates of the parameters based on the generalized Heckman, classic Heckman, Student-$t$, and skew-normal sample selection models under Scenario 1. From this table, we observe a good performance of the MLEs based on the generalized Heckman model, even for estimating the parameters related to the sample selection bias and dispersion. The bias and the RMSE under this model decrease for all the estimates as the sample size increases; therefore, suggesting the consistency of the MLEs, which is in line with our Theorem \ref{asympt_results}. On the other hand, even the regression structures for $\pmb\beta$ and $\pmb\gamma$ being correctly specified, we see that the MLEs for these parameters, based on the classic Heckman, skew-normal, and Student-$t$, do not provide satisfactory estimates even for a large sample. This illustrates the importance of considering covariates for the sample selection bias and dispersion parameters. The mean estimates of the degrees of freedom and skewness for the Student-$t$ and skew-normal sample selection models were $2.4$ and $0.8$, respectively.
The above comments are also supported by Figures \ref{2:cen3bp1}, \ref{2:cen3bp2}, and \ref{2:cen3bp3}, where boxplots of the parameter estimates are presented for sample sizes $n=500$, $n=1000$, and $n=2000$, respectively. We did not present the boxplots of the estimates of $\gamma_{1}$, $\gamma_{2}$, and $\beta_{1}$ since they behaved similarly to other boxplots.
\begin{table} \centering \caption{Empirical mean and root mean square error (RMSE) of the maximum likelihood estimates of the parameters based on the generalized Heckman, classic Heckman, Student-$t$, and skew-normal sample selection models under Scenario 1.}
\begin{tabular}{cccccc@{\hspace{-0.15ex}}ccc@{\hspace{-0.15ex}}ccc@{\hspace{-0.15ex}}cc}
\hline \hline
&&&\multicolumn{2}{c}{Generalized}&&\multicolumn{2}{c}{Classic}&&\multicolumn{2}{c}{Heckman}&&\multicolumn{2}{c}{Heckman} \\
&&&\multicolumn{2}{c}{Heckman}&&\multicolumn{2}{c}{Heckman}&&\multicolumn{2}{c}{Skew-Normal}&&\multicolumn{2}{c}{Student-$t$} \\
\cline{4-5} \cline{7-8} \cline{10-11}\cline{13-14}
\multicolumn{2}{c}{Parameters}& $n$ & mean & RMSE& & mean & RMSE & & mean & RMSE & & mean & RMSE\\ \hline
\hline
\multirow{3}{*}{$\gamma_{0}$}& \multirow{3}{*}{0.9}&500 & 0.912 & 0.093 & & 0.880 & 0.095 & & 0.854 & 0.364 & & 1.300 & 0.435 \\
&&1000 & 0.905 & 0.063 & & 0.878 & 0.069 & & 0.820 & 0.364 & & 1.293 & 0.409 \\
&&2000 & 0.903 & 0.042 & & 0.877 & 0.048 & & 0.804 & 0.348 & & 1.286 & 0.395 \\
\hline
\multirow{3}{*}{$\gamma_{1}$}& \multirow{3}{*}{0.5} &500 & 0.507 & 0.081 & & 0.555 & 0.104 & & 0.527 & 0.095 & & 0.652 & 0.200 \\
&&1000 & 0.506 & 0.056 & & 0.558 & 0.085 & & 0.530 & 0.075 & & 0.663 & 0.185 \\
&&2000 & 0.504 & 0.040 & & 0.555 & 0.070 & & 0.529 & 0.057 & & 0.662 & 0.174 \\
\hline
\multirow{3}{*}{$\gamma_{2}$}& \multirow{3}{*}{1.1} &500 & 1.118 & 0.104 & & 1.053 & 0.125 & & 1.001 & 0.169 & & 1.540 & 0.480 \\
&&1000 & 1.109 & 0.074 & & 1.046 & 0.099 & & 0.978 & 0.162 & & 1.522 & 0.441 \\
&&2000 & 1.107 & 0.053 & & 1.033 & 0.090 & & 0.980 & 0.146 & & 1.515 & 0.425 \\
\hline
\multirow{3}{*}{$\gamma_{3}$}& \multirow{3}{*}{0.6}&500 & 0.606 & 0.084 & & 0.556 & 0.106 & & 0.531 & 0.127 & & 0.848 & 0.288 \\
&&1000 & 0.604 & 0.065 & & 0.539 & 0.098 & & 0.504 & 0.130 & & 0.828 & 0.253 \\
&&2000 & 0.602 & 0.040 & & 0.543 & 0.074 & & 0.513 & 0.106 & & 0.833 & 0.242 \\
\hline
\multirow{3}{*}{$\beta_{0}$} & \multirow{3}{*}{1.1} &500 & 1.105 & 0.068 & & 0.998 & 0.261 & & 0.496 & 0.734 & & 1.112 & 0.103 \\
&&1000 & 1.102 & 0.044 & & 0.954 & 0.273 & & 0.553 & 0.701 & & 1.108 & 0.078 \\
&&2000 & 1.099 & 0.029 & & 0.892 & 0.230 & & 0.498 & 0.697 & & 1.096 & 0.051 \\
\hline
\multirow{3}{*}{$\beta_{1}$} & \multirow{3}{*}{0.7} &500 & 0.702 & 0.036 & & 0.882 & 0.219 & & 0.739 & 0.156 & & 0.787 & 0.106 \\
&&1000 & 0.701 & 0.020 & & 0.860 & 0.191 & & 0.753 & 0.158 & & 0.777 & 0.087 \\
&&2000 & 0.700 & 0.014 & & 0.881 & 0.191 & & 0.770 & 0.154 & & 0.774 & 0.079 \\
\hline
\multirow{3}{*}{$\beta_{2}$} & \multirow{3}{*}{0.1} &500 & 0.098 & 0.048 & & 0.140 & 0.196 & & 0.040 & 0.225 & & 0.101 & 0.078 \\
&&1000 & 0.099 & 0.029 & & 0.174 & 0.196 & & 0.069 & 0.237 & & 0.104 & 0.057 \\
&&2000 & 0.100 & 0.019 & & 0.221 & 0.147 & & 0.105 & 0.193 & & 0.113 & 0.042 \\
\hline
\multirow{3}{*}{$\lambda_{0}$}& \multirow{3}{*}{ $-$0.4} &500 & $-$0.400 & 0.042 & & 0.174 & 0.577 & & 0.366 & 0.771 & & $-$0.476 & 0.113 \\
&&1000 & $-$0.402 & 0.028 & & 0.182 & 0.584 & & 0.339 & 0.743 & & $-$0.470 & 0.092 \\
&&2000 & $-$0.401 & 0.019 & & 0.182 & 0.582 & & 0.331 & 0.734 & & $-$0.463 & 0.075 \\
\hline
\multirow{3}{*}{$\lambda_{1}$}& \multirow{3}{*}{0.7} &500 & 0.699 & 0.042 & & $-$ & $-$ & & $-$ & $-$ & & $-$ & $-$ \\
&&1000 & 0.700 & 0.031 & & $-$ & $-$ & & $-$ & $-$ & & $-$ & $-$ \\
&&2000 & 0.701 & 0.020 & & $-$ & $-$ & & $-$ & $-$ & & $-$ & $-$ \\
\hline
\multirow{3}{*}{$\kappa_{0}$} & \multirow{3}{*}{0.3 } &500 & 0.314 & 0.248 & & 0.507 & 0.688 & & 0.133 & 0.846 & & 0.307 & 0.300 \\
&&1000 & 0.311 & 0.154 & & 0.588 & 0.714 & & 0.223 & 0.918 & & 0.302 & 0.235 \\
&&2000 & 0.309 & 0.106 & & 0.811 & 0.608 & & 0.307 & 0.814 & & 0.336 & 0.158 \\
\hline
\multirow{3}{*}{$\kappa_{1}$} & \multirow{3}{*}{0.5} &500 & 0.573 & 0.240 & & $-$ & $-$ & & $-$ & $-$ & & $-$ & $-$\\
&&1000 & 0.531 & 0.135 & & $-$ & $-$ & & $-$ & $-$ & & $-$ & $-$ \\
&&2000 & 0.510 & 0.092 & & $-$ & $-$ & & $-$ & $-$ & & $-$ & $-$ \\
\hline
\end{tabular}
\label{2:res_cen3_tab1} \end{table}
\begin{sidewaysfigure} \includegraphics[width=\textwidth]{plot500.pdf} \caption{Boxplots of the maximum likelihood estimates of the parameters (a) $\gamma_{0},$ (b) $\gamma_{1},$ (c) $\beta_{0},$ (d) $\beta_{1},$ (e) $\lambda_{0},$ (f) $\lambda_{1}$ and (g) $\kappa_{0}$ and (h) $\kappa_{1}$ based on the (i) generalized Heckman, (ii) classic Heckman, (iii) Heckman-Skew, and (iv) Heckman-$t$ sample selection models. Sample size $n=500.$} \label{2:cen3bp1} \end{sidewaysfigure}
\begin{sidewaysfigure} \includegraphics[width=\textwidth]{plot1000.pdf} \caption{Boxplots of the maximum likelihood estimates of the parameters (a) $\gamma_{0},$ (b) $\gamma_{1},$ (c) $\beta_{0},$ (d) $\beta_{1},$ (e) $\lambda_{0},$ (f) $\lambda_{1}$ and (g) $\kappa_{0}$ and (h) $\kappa_{1}$ based on the (i) generalized Heckman, (ii) classic Heckman, (iii) Heckman-Skew, and (iv) Heckman-$t$ sample selection models. Sample size $n=1000.$} \label{2:cen3bp2} \end{sidewaysfigure}
\begin{sidewaysfigure} \includegraphics[width=\textwidth]{plot2000.pdf} \caption{Boxplots of the maximum likelihood estimates of the parameters (a) $\gamma_{0},$ (b) $\gamma_{1},$ (c) $\beta_{0},$ (d) $\beta_{1},$ (e) $\lambda_{0},$ (f) $\lambda_{1}$ and (g) $\kappa_{0}$ and (h) $\kappa_{1}$ based on the (i) generalized Heckman, (ii) classic Heckman, (iii) Heckman-Skew, and (iv) Heckman-$t$ sample selection models. Sample size $n=2000.$} \label{2:cen3bp3} \end{sidewaysfigure}
We now provide some simulations to evaluate the size and power of likelihood ratio, gradient, and Wald tests. We consider Scenario 1 and present the empirical significance level of the tests in Table \ref{2:tamanhoTesteCen1} for nominal significance levels at 1\%, 5\%, and 10\%.
Under the null hypothesis of absence of sample selection bias ($\rho_i=0$ for all $i$, that is $\kappa_0=\kappa_1=0$), the likelihood ratio, gradient, and Wald tests presented empirical values close to the nominal values only under the generalized Heckman model. For the other models, the type-I error was inflated and indicated the presence of selection bias. This suggests that the tests should be used with caution to test parameters of the sample selection models and that some confounding can be caused due to either varying sample selection bias or heteroskedasticity. It is important to point out that, even for the generalized Heckman model, the Wald test presents a considerable inflated type-I error for $n=500$.
\begin{table} \centering \caption{Empirical significance level of the likelihood ratio (LR), gradient (G) and Wald (W) tests for $H_{0}:\rho=0$.}
\begin{tabular}{c|ccccccccccccccccc} \hline\hline
&\multicolumn{3}{c}{Generalized} &&\multicolumn{3}{c}{Classic}&& \multicolumn{3}{c}{Heckman}&& \multicolumn{3}{c}{Heckman} \\
&\multicolumn{3}{c}{Heckman} &&\multicolumn{3}{c}{Heckman} && \multicolumn{3}{c}{Skew-normal} && \multicolumn{3}{c}{Student-$t$} \\ \cline{2-4} \cline{6-8} \cline{10-12} \cline{14-16} $n$ & LR & G & W && LR & G & W && LR & G & W && LR & G & W \\ \hline \hline
\cline{1-16}
\multicolumn{16}{c}{$\alpha=1\%$} \\ \hline \hline $500 $ & 1.5& 1.1& 3.5&& 30.3& 1.6& 61.4&& 41.9& 6.4& 71.3&& 2.8& 1.5& 4.3\\ $1000$ & 0.9& 0.7& 2.6&& 72.4& 7.7& 92.7&& 82.7& 26.2& 95.6&& 3.4& 2.5& 4.0\\ $2000$ & 1.1& 0.7& 1.9&& 85.9& 16.5& 96.8&& 91.0& 52.4& 99.0&& 3.2& 2.6& 3.9\\
\hline \hline \multicolumn{16}{c}{$\alpha=5\%$} \\ \hline \hline $500 $ & 6.2& 5.6& 12.0 && 50.5& 12.5& 69.9&& 61.7& 26.9& 78.3&& 9.9& 7.7& 11.6\\ $1000$ & 4.9& 3.7& 6.9&& 86.9& 27.1& 95.5&& 89.9& 49.7& 97.8&& 10.1& 8.6& 11.8\\ $2000$ & 5.9& 5.1& 7.4&& 94.2& 40.8& 97.7&& 94.5& 68.3& 99.5&& 10.2& 9.4& 10.6\\
\hline \hline \multicolumn{16}{c}{$\alpha=10\%$} \\ \hline \hline $500 $ & 13.0& 10.5& 19.3&& 59.8& 25.2& 74.4&& 69.8& 40.8& 82.2&& 16.2& 14.8& 18.4\\ $1000$ & 9.4& 8.2& 13.7&& 91.5& 41.4& 96.1&& 93.3& 61.1& 98.2&& 16.9& 15.4& 18.0\\ $2000$ &11.2& 10.8& 12.6&& 96.3& 55.7& 98.3&& 95.5& 74.8& 99.6&& 16.1& 15.8& 16.4\\
\hline \hline \end{tabular} \label{2:tamanhoTesteCen1} \end{table}
In Table \ref{2:Poder1TesteCen1}, we present the empirical power of the likelihood ratio, gradient and Wald tests (in percentage) for simulated data according to Scenario 1 under generalized Heckman, classic Heckman, Heckman-Skew, and Heckman-$t$ models, with significance level at $1\%, 5\%$ and $10\%$. From these results, we can observe that the tests, under the generalized Heckman model, provide high power, mainly when the sample size increases. On the other hand, since tests based on the other models do not provide the correct nominal significance level, the power of tests in these cases are not really comparable.
\begin{table} \centering \caption{Empirical power of the likelihood ratio (LR), gradient (G) and Wald (W) tests (in percentage) for simulated data according to Scenario 1 under generalized Heckman, classic Heckman, Heckman-Skew, and Heckman-$t$ models, with significance level at $1\%, 5\%$ and $10\%$.}
\begin{tabular}{c|ccccccccccccccccc} \hline\hline
&\multicolumn{3}{c}{Generalized} &&\multicolumn{3}{c}{Classic}&& \multicolumn{3}{c}{Heckman}&& \multicolumn{3}{c}{Heckman} \\
&\multicolumn{3}{c}{Heckman} &&\multicolumn{3}{c}{Heckman} && \multicolumn{3}{c}{Skew-normal} && \multicolumn{3}{c}{Student-$t$} \\ \cline{2-4} \cline{6-8} \cline{10-12} \cline{14-16} $n$ & LR & G & W && LR & G & W && LR & G & W && LR & G & W \\ \hline \hline
\cline{1-16}
\multicolumn{16}{c}{$\alpha=1\%$} \\ \hline \hline $500 $ &71.7&68.4&75.5&&58.4& 8.9&81.9&&45.7& 4.5&79.7&&14.6& 9.3&18.4\\ $1000$ &95.3&95.2&96.9&&93.7&30.8&99.2&& 88.0& 14.0&98.4&&20.9& 16.0&26.6\\ $2000$ &99.9&99.9&99.9&&99.6&74.4& 100&&98.2&28.4&99.4&&48.3& 46.0&52.1\\
\hline \hline \multicolumn{16}{c}{$\alpha=5\%$} \\ \hline \hline $500 $ &88.8&87.8&91.6&&74.1&32.6&89.3&&69.2&20.1&86.3&&29.6&25.1&33.6\\ $1000$ & 99&98.9&99.4&&98.1&57.4&99.9&&94.5&33.5& 99&&41.9&37.9&46.2\\ $2000$ & 100& 100& 100&&99.9& 87.0& 100&&98.9&46.8&99.7&&68.8&66.9&70.9\\
\hline \hline \multicolumn{16}{c}{$\alpha=10\%$} \\ \hline \hline $500 $ &94.3&93.1&95.5&&82.4&48.4&91.9&&78.2&34.9&89.8&&39.8&37.1&43.1\\ $1000$ &99.8&99.8&99.8&&99.1&69.6& 100&& 96&45.5& 99&&52.4&49.4&54.7\\ $2000$ & 100& 100& 100&& 100&91.6& 100&&99.2&55.9&99.7&&78.4&77.8&79.4\\
\hline \hline \end{tabular} \label{2:Poder1TesteCen1}
\end{table}
\section{MEPS Data Analysis}\label{3:application}
We present an application of the proposed model to a set of real data. Consider the outpatient expense data of the 2001 Medical Expenditure Panel Survey (MEPS) available in the \texttt{R} software in the package \texttt{ssmrob} \citep{ssmrob}. These data were also used by \cite{cameron2009microeconometrics}, \cite{marchenkoGenton}, and \cite{zhelonkin2013robustness} to fit the classic Heckman model, Heckman-$t$ model, and the robust version of the two-step method, respectively. The MEPS is a set of large-scale surveys of families, individuals, and their medical providers (doctors, hospitals, pharmacies, etc.) in the United States. It has data on the health services Americans use, how often they use them, the cost of these services, and how they are paid, as well as data on the cost and reach of health insurance available to American workers.
The sample is restricted to persons aged between 21 and 64 years and contains a variable response with $3,328$ observations of outpatient costs, of which 526 (15.8 \%) correspond to unobserved expenditure values identified as zero expenditure. It also includes the following explanatory variables: \texttt{Age} represents age measured in tens of years; \texttt{Fem} is an indicator variable for gender (female receives value 1); \texttt{Educ} represents years of schooling; \texttt{Blhisp} is an indicator for ethnicity (black or Hispanic receive a value of 1); \texttt{Totcr} is the total number of chronic diseases; \texttt{Ins} is the insurance status; and \texttt{Income} denotes the individual income.
The variable of interest $Y_{1i}^{*}$ represents the log-expenditure on medical services of the $i$th individual. We consider the logarithm of the expenditure since it is highly skewed (i.e., see Figure~\ref{histogram}, where plots of the expenditure and log-expenditure are presented). The variable $Y_{2i}^{*}$ denoting the willingness of the $i$th individual to spend is not observed. We only observe $U_i=I\{Y_{2i}^{*}>0\}$, which represents the decision or not of the $i$th individual to spend on medical care.
\begin{figure}
\caption{Histogram of medical expense data (a) and medical expense log (b).}
\label{histogram}
\end{figure}
According to \cite{cameron2009microeconometrics} and \cite{zhelonkin2013robustness}, it is natural to fit a sample selection model to such data, since the willingness to spend $(Y_{2}^{*})$ is likely to be related to the expense amount $(Y_{1}^{*}).$ However, after fitting the classic Heckman model and using the Wald statistics for testing $H_{0}:\rho=0$ against $H_{1}:\rho\neq 0,$ the conclusion is that there is no statistical evidence $(p\textrm{-value}>0.1)$ to reject $H_{0},$ that is, there is no sample selection bias. \cite{cameron2009microeconometrics} suspected this conclusion on the absence of sample selection bias, and \cite{marchenkoGenton} argued that a more robust model could evidence the presence of sample selection bias in the data; these authors proposed a Student-$t$ sample selection model to deal with this problem. However, as will be illustrated in this application, this problem of the classic Heckman model can be due to the assumption of constant sample selection bias and constant dispersion parameters rather than the normal assumption itself.
After a preliminary analysis, we consider the following regression structures for our proposed generalized Heckman model: \begin{align*} \mu_{1i}&= \beta_{0}+ \beta_{1}\texttt{Age}_{i}+ \beta_{2}\texttt{Fem}_{i}+ \beta_{3}\texttt{Educ}_{i}+ \beta_{4}\texttt{Blhisp}_{i}+\beta_{5}\texttt{Totchr}_{i}+\beta_{6}\texttt{Ins}_{i}, \\ \mu_{2i}&=\gamma_{0}+\gamma_{1}\texttt{Age}_{i}+\gamma_{2}\texttt{Fem}_{i}+\gamma_{3}\texttt{Educ}_{i}+\gamma_{4}\texttt{Blhisp}_{i}+\gamma_{5}\texttt{Totchr}_{i}+\gamma_{6}\texttt{Ins}_{i}+\gamma_{7}\texttt{Income}_{i}, \\ \log{\sigma_{i}} &=\lambda_{0}+\lambda_{1}\texttt{Age}_{i}+\lambda_{2}\texttt{Totchr}_{i}+\lambda_{3}\texttt{Ins}_{i},\\ \mbox{arctanh}\,\rho_{i}&=\kappa_{0}+\kappa_{1}\texttt{Fem}_{i}+\kappa_{2}\texttt{Totchr}_{i}, \end{align*} for $i=1,\ldots, 3328$. The primary equation has the same covariates of the selection equation with the additional covariate \texttt{Income}, so that exclusion restriction is in force. In Table \ref{2:dados_reais_tab1}, we present the summary of the fits of the classic Heckman and generalized Heckman models. From this table, we can observe that the covariates \texttt{Fem} and \texttt{Totchr} are significant to explain the sample selection bias, by using any significance level. We perform a likelihood (LR) test for checking the absence ($H_0:\,\kappa_0=\kappa_1=\kappa_2=0$) or presence of sample selection bias. The LR statistic was $28.16$ with corresponding $p$-value equal to $3\times10^{-6}$. Therefore, our proposed generalized Heckman model is able to detect the presence of sample selection bias even under normal assumption. We also performed the gradient and Wald tests, which confirmed this conclusion.
Further, the covariates \texttt{age}, \texttt{totcr}, and \texttt{ins} were significant for the dispersion parameter. For the selection equation, the covariate \texttt{Income} is not significant (significance level at 5\%) based on the generalized Heckman model in contrast with the classic Heckman model. Anyway, it is important to keep it in order to satisfy the exclusion restriction. Regarding the primary equation, we observe that the covariate \texttt{Ins} is only significant under the classic Heckman model. Another interesting point is that \texttt{Educ} is strongly significant for the primary equation under our proposed generalized Heckman model.
\begin{table} \centering \caption{Summary fits of the classic Heckman model (HM) and generalized Heckman model (GHM). The GHM summary fit contains estimates with their respective standard errors, z-value, $p\textrm{-value}$, and inferior and superior bounds of the $95\%$ confidence interval.} \begin{tabular}{lcccccccc} \hline
\multicolumn{9}{c}{\texttt{Selection Equation}}\\
\hline covariates & HM-est. & p-value & GHM-est. & stand. error & z-value & p-value & Inf. & Sup. \\ \hline \texttt{Intercept} & $-$0.676 & 0.000 & $-$0.590 & 0.187 & $-$3.162 & 0.002 & $-$0.956 & $-$0.224 \\ \texttt{Age} & 0.088 & 0.001 & 0.086 & 0.026 & 3.260 & 0.001 & 0.034 & 0.138 \\ \texttt{Fem} & 0.663 & 0.000 & 0.630 & 0.060 & 10.544 & 0.000 & 0.513 & 0.747 \\ \texttt{Educ} & 0.062 & 0.000 & 0.057 & 0.011 & 4.984 & 0.000 & 0.035 & 0.079 \\ \texttt{Blhisp} & $-$0.364 & 0.000 & $-$0.337 & 0.060 & $-$5.644 & 0.000 & $-$0.454 & $-$0.220 \\ \texttt{Totchr} & 0.797 & 0.000 & 0.758 & 0.069 & 11.043 & 0.000 & 0.624 & 0.893 \\ \texttt{Ins} & 0.170 & 0.007 & 0.173 & 0.061 & 2.825 & 0.005 & 0.053 & 0.293 \\ \texttt{Income} & 0.003 & 0.040 & 0.002 & 0.001 & 1.837 & 0.066 & 0.000 & 0.005 \\
\hline
\multicolumn{9}{c}{\texttt{Primary Equation}}\\
\hline
covariates & HM-est. & p-value & GHM-est. & stand. error & z-value & p-value & Inf. & Sup. \\
\hline \texttt{Intercept} & 5.044 & 0.000 & 5.704 & 0.193 & 29.553 & 0.000 & 5.326 & 6.082 \\ \texttt{Age} & 0.212 & 0.000 & 0.184 & 0.023 & 7.848 & 0.000 & 0.138 & 0.230 \\ \texttt{Fem} & 0.348 & 0.000 & 0.250 & 0.059 & 4.252 & 0.000 & 0.135 & 0.365 \\ \texttt{Educ} & 0.019 & 0.076 & 0.001 & 0.010 & 0.129 & 0.897 & $-$0.019 & 0.021 \\ \texttt{Blhisp} & $-$0.219 & 0.000 & $-$0.128 & 0.058 & $-$2.221 & 0.026 & $-$0.242 & $-$0.015 \\ \texttt{Totchr} & 0.540 & 0.000 & 0.431 & 0.031 & 14.113 & 0.000 & 0.371 & 0.490 \\ \texttt{Ins} & $-$0.030 & 0.557 & $-$0.103 & 0.051 & $-$1.999 & 0.046 & $-$0.203 & $-$0.002 \\
\hline
\multicolumn{9}{c}{\texttt{Dispersion Parameter}}\\
\hline
covariates & HM-est. & p-value & GHM-est. & stand. error & z-value & p-value & Inf. & Sup. \\
\hline \texttt{Intercept} & 1.271 & $-$ & 0.508 & 0.057 & 8.853 & 0.000 & 0.396 & 0.621 \\ \texttt{Age} &$-$ &$-$ & $-$0.025 & 0.013 & $-$1.986 & 0.047 & $-$0.049 & 0.000 \\ \texttt{Totchr} &$-$ &$-$ & $-$0.105 & 0.019 & $-$5.475 & 0.000 & $-$0.142 & $-$0.067 \\ \texttt{Ins} &$-$ &$-$ & $-$0.107 & 0.028 & $-$3.864 & 0.000 & $-$0.161 & $-$0.053 \\
\hline
\multicolumn{9}{c}{\texttt{Sample Selection Bias Parameter}}\\
\hline
covariates & HM-est. & p-value & GHM-est. & stand. error & z-value & p-value & Inf. & Sup. \\
\hline \texttt{Intercept} & $-$0.131 & 0.375 & $-$0.648 & 0.114 & $-$5.668 & 0.000 & $-$0.872 & $-$0.424 \\ \texttt{Fem} & $-$ &$-$ & $-$0.403 & 0.136 & $-$2.973 & 0.003 & $-$0.669 & $-$0.137 \\ \texttt{Totchr} & $-$ & $-$ & $-$0.438 & 0.186 & $-$2.353 & 0.019 & $-$0.803 & $-$0.073 \\
\hline \end{tabular} \label{2:dados_reais_tab1} \end{table}
We conclude this application by checking the goodness-of-fit of the fitted generalized sample selection Heckman model. In Figure \ref{QQplot}, we provide the QQ-plot of the score residuals given in (\ref{score_residual}) with simulated envelopes, and also a Cook distance plot for detecting global influence. Based on this last plot, we do not detect any outlier observation, since all points are below the reference line $2p/n=0.013$. Anyway, we investigate if the highlighted point \#2602 (above the line $8\times 10^{-4}$) is influential. We fitted our model by removing this observation and no changes either on the parameter estimates or different conclusions about the significance of covariates were obtained. Regarding the QQ-plot of the score residuals, we observe a great performance of our model, since 96\% of the points are inside of the envelope. This confirms that the normal assumption for this particular dataset is adequate and that our generalized Heckman model is suitable for the MEPS data analysis.
\begin{figure}
\caption{QQ-plot and its simulated envelope for the score residuals (left) and index plot of the GCD (right) for the generalized Heckman model for the medical expenditure panel survey data.}
\label{QQplot}
\end{figure}
\section{Concluding Remarks}\label{concluding_remarks}
In this paper, a generalization of the Heckman model was proposed by allowing both sample selection bias and dispersion parameters to vary across covariates. We showed that the proposed model satisfies certain regularity conditions that ensure consistency and asymptotic normality of the maximum likelihood estimators. Furthermore, a proper score residual for sample selection models was proposed. These finding are new contributions on this topic. The MEPS data analysis based on the generalized Heckman model showed that the normal assumption for the data is suitable in contrast with existing findings in the literature. Future research should address (i) generalization of other sample selection models such as Student-$t$ and skew-normal to allow varying sample selection bias and dispersion parameters; (ii) proposal of proper residuals for other sample selection models; and (iii) deeper study of influence analysis. An \texttt{R} package for fitting our proposed generalized Heckman, Student-$t$, and skew-normal models has been developed and will be available soon.
\section*{Appendix}
\noindent {\it Proof of Theorem \ref{regcond}}. We here show the results for the derivatives with respect to $\boldsymbol\beta$. The results involving the other derivatives follow similarly and therefore they are omitted.
For $i=1,\ldots,n$ and $j=1,\ldots,p$, we have that \begin{eqnarray}\label{scorebeta} \dfrac{\partial\ell_i}{\partial\beta_j}=\left\{-\dfrac{\rho_i}{\sqrt{1-\rho_i^2}}\dfrac{\phi(\zeta_{i})}{\Phi(\zeta_{i})}+z_i\right\}x_{ij}u_i/\sigma_i. \end{eqnarray}
By using basic properties of conditional expectation, it follows that $E\left(\dfrac{\partial\ell_i}{\partial\beta_j}\right)=E\left[E\left(\dfrac{\partial\ell_i}{\partial\beta_j}\big|U_i\right)\right]$ and it is immediate that $E\left(\dfrac{\partial\ell_i}{\partial\beta_j}\big|U_i=0\right)=0$. Let us now compute the conditional expectations involved in $E\left(\dfrac{\partial\ell_i}{\partial\beta_j}\big|U_i=1\right)$.
Here it is worth to remember the notations $z_i=\dfrac{y_i-\mu_{1i}}{\sigma_i}$ and $\zeta_i=\dfrac{\mu_{2i}+z_i\rho_i}{\sqrt{1-\rho_i^2}}$, for $i=1,\ldots,n$. We now use the conditional density function given in (\ref{1:dens_heckman}) to obtain that \begin{eqnarray*}
E\left(\dfrac{\phi(\zeta_{i})}{\Phi(\zeta_{i})}\big|U_i=1\right)&=&\dfrac{1}{\sigma_i\Phi(\mu_{2i})}\int_{-\infty}^\infty\phi(\zeta_i)\phi(z_i)dy_i=\dfrac{1}{2\pi\sigma_i\Phi(\mu_{2i})}\int_{-\infty}^\infty\exp\{-(\zeta_i^2+z_i^2)/2\}dy_i\\ &=&\dfrac{e^{-\mu_{2i}^2/2}}{2\pi\sigma_i\Phi(\mu_{2i})}\int_{-\infty}^\infty\exp\left\{-\dfrac{(y_i-\mu_{1i}+\sigma_i\mu_{2i}\rho_i)^2}{2\sigma_i^2(1-\rho_i^2)}\right\}dy_i= \sqrt{1-\rho_i^2}\dfrac{\phi(\mu_{2i})}{\Phi(\mu_{2i})}, \end{eqnarray*} where the last equality follows by identifying a normal kernel in the integral. On the other hand, we use the fact that $Z_i$ given $U_i=1$ has mean equal to $\mu_{1i}+\rho_i\sigma_i\dfrac{\phi(\mu_{2i})}{\Phi(\mu_{2i})}$ (more details are given in the Supplementary Material) and get \begin{eqnarray*}
E\left(Z_i\big|U_i=1\right)=-\dfrac{\mu_{1i}}{\sigma_i}+\dfrac{1}{\sigma_i}E(Y_i|U_i=1)=-\dfrac{\mu_{1i}}{\sigma_i}+\dfrac{1}{\sigma_i}\left(\mu_{1i}+\rho_i\sigma_i\dfrac{\phi(\mu_{2i})}{\Phi(\mu_{2i})}\right)=\rho_i\dfrac{\phi(\mu_{2i})}{\Phi(\mu_{2i})}. \end{eqnarray*}
With the results above, we obtain that $E\left(\dfrac{\phi(\zeta_{i})}{\Phi(\zeta_{i})}\big|U_i\right)=0$ almost surely and therefore $E\left(\dfrac{\partial\ell_i}{\partial\beta_j}\right)=0$ for $j=1,\ldots,p$.
We now concentrate our attention to prove the identity stated in the theorem. It follows that \begin{eqnarray*} \dfrac{\partial^2\ell_i}{\partial\beta_j\beta_l}=-\left\{\dfrac{\rho_i^2}{1-\rho_i^2}\left[\zeta_i\dfrac{\phi(\zeta_i)}{\Phi(\zeta_i)}+\dfrac{\phi^2(\zeta_i)}{\Phi^2(\zeta_i)}\right]+1\right\}\dfrac{x_{ij}x_{il}}{\sigma_i^2}u_i \end{eqnarray*} and \begin{eqnarray*} \dfrac{\partial\ell_i}{\partial\beta_j}\dfrac{\partial\ell_i}{\partial\beta_l}=\left\{z_i^2-2z_i\dfrac{\rho_i}{\sqrt{1-\rho_i^2}}\dfrac{\phi(\zeta_i)}{\Phi(\zeta_i)}+\dfrac{\rho_i^2}{1-\rho_i^2}\dfrac{\phi^2(\zeta_i)}{\Phi^2(\zeta_i)}\right\}\dfrac{x_{ij}x_{il}}{\sigma_i^2}u_i, \end{eqnarray*}
where we have used that $u_i^2=u_i$ (since $u_i\in\{0,1\}$) in the last equality. It is immediate that $E\left(\dfrac{\partial^2\ell_i}{\partial\beta_j\beta_l}\big|U_i=0\right)=-E\left(\dfrac{\partial\ell_i}{\partial\beta_j}\dfrac{\partial\ell_i}{\partial\beta_l}\big|U_i=0\right)=0$.
Following in a similar way as before, after some algebra we obtain that \begin{eqnarray*}
E\left(\zeta_i\dfrac{\phi(\zeta_i)}{\Phi(\zeta_i)}\big|U_i=1\right)=-\mu_{2i}(1-\rho_i^2)\dfrac{\phi(\mu_{2i})}{\Phi(\mu_{2i})}\quad\mbox{and}\quad E\left(Z_i^2\big|U_i=1\right)=1-\mu_{2i}\rho_i^2\dfrac{\phi(\mu_{2i})}{\Phi(\mu_{2i})}. \end{eqnarray*}
By combining these results, we have that \begin{eqnarray*}
-E\left(\dfrac{\partial^2\ell_i}{\partial\beta_j\beta_l}\big|U_i=1\right)&=&E\left(\dfrac{\partial\ell_i}{\partial\beta_j}\dfrac{\partial\ell_i}{\partial\beta_l}\big|U_i=1\right)\\
&=&\left\{1+\mu_{2i}\rho_i^2\dfrac{\phi(\mu_{2i})}{\Phi(\mu_{2i})}+\dfrac{\rho_i^2}{1-\rho_i^2}E\left(\dfrac{\phi^2(\zeta_i)}{\Phi^2(\zeta_i)}\big|U_i=1\right)\right\}\dfrac{x_{ij}x_{il}}{\sigma_i^2}. \end{eqnarray*}
Since the conditional expectations coincide, the marginal expectations also coincide so giving the desired result. $\square$
\noindent {\it Proof of Theorem \ref{asympt_results}}. Conditions (C1)--(C3) and Theorem \ref{regcond} give us the consistency of the MLEs. To establish the asymptotic normality of the estimators, we need to show that the third derivatives of the log-likelihood function are bounded by integrable functions not depending on the parameters.
We will show here that this is possible for the derivatives involving the $\boldsymbol\beta$'s. The other cases follow in a similar way as discussed in the proof of Theorem \ref{regcond} and therefore they are omitted.
By computing the third derivatives with respect to the $\boldsymbol\beta$'s and using the triangular inequality, we have that \begin{eqnarray*}
\left|\dfrac{\partial^3\ell_i}{\partial\beta_j\partial\beta_l\partial\beta_k}\right|&\leq&\dfrac{\rho_i^2}{\sigma_i^3(1-\rho_i^2)^{3/2}}\left\{(1+z_i^2)\dfrac{\phi(\zeta_i)}{\Phi(\zeta_i)}+\zeta_i^2\dfrac{\phi^2(\zeta_i)}{\Phi^2(\zeta_i)}+2\zeta_i\dfrac{\phi^2(\zeta_i)}{\Phi^2(\zeta_i)}+2\dfrac{\phi^2(\zeta_i)}{\Phi^3(\zeta_i)}\right\}x_{ij}x_{il}x_{ik}\\ &\equiv& g_i(\boldsymbol{\theta})\leq g_i(\boldsymbol{\theta}^*), \end{eqnarray*} for $j,l,k=1,\ldots,p$, where $\boldsymbol{\theta}^*=\mbox{argmax}_{\boldsymbol{\theta}\in\boldsymbol\Theta}g_i(\boldsymbol{\theta})$, which is well-defined due to Assumption (C1).
We now need to show that the expectations of the terms in $g_i(\boldsymbol{\theta}^*)$ are finite. Let us show that $E_{\boldsymbol\theta_0}\left(\zeta_i^{*2}\dfrac{\phi^2(\zeta_i^*)}{\Phi^2(\zeta_i^*)}\right)<\infty$, where $E_{\boldsymbol\theta_0}(\cdot)$ denotes the expectation with respect to the true parameter vector value $\boldsymbol\theta_0$ and $\zeta_i^*$ is defined as $\zeta_i$ by replacing $\boldsymbol{\theta}$ by $\boldsymbol{\theta}^*$. The proofs for the remaining terms follow from this one or in a similar way.
For $\zeta_i^*\leq\sqrt2$, it follows that $\phi^2(\zeta_i^*)/\Phi^2(\zeta_i^*)\leq \left\{2\pi\Phi^2(\sqrt2)\right\}^{-1}$. Now, consider $\zeta_i^*>\sqrt2$. Theorem 1.2.6 from \cite{durrett} gives us the following inequality for $x>0$: \begin{eqnarray*} \left(\dfrac{1}{x}-\dfrac{1}{x^3}\right)e^{-x^2/2}\leq \int_x^\infty e^{-y^2/2}dy. \end{eqnarray*}
Using this inequality and under $\zeta_i^*>\sqrt2$, we obtain that $\dfrac{\phi^2(\zeta_i^*)}{\Phi^2(\zeta_i^*)}\leq \dfrac{\zeta_i^{*3}}{\zeta_i^{*2}-1}\leq \zeta_i^{*}$. These results imply that \begin{eqnarray*} E_{\boldsymbol\theta_0}\left(\zeta_i^{*2}\dfrac{\phi^2(\zeta_i^*)}{\Phi^2(\zeta_i^*)}\right)&=&E_{\boldsymbol\theta_0}\left(\zeta_i^{*2}\dfrac{\phi^2(\zeta_i^*)}{\Phi^2(\zeta_i^*)}I\left\{\zeta_i^*\leq\sqrt2\right\}\right)+E_{\boldsymbol\theta_0}\left(\zeta_i^{*2}\dfrac{\phi^2(\zeta_i^*)}{\Phi^2(\zeta_i^*)}I\left\{\zeta_i^*>\sqrt2\right\}\right)\\
&\leq& \sqrt{2}\left\{2\pi\Phi^2(\sqrt2)\right\}^{-1}+E_{\boldsymbol\theta_0}\left(\left|\zeta_i^{*}\right|^3\right)<\infty, \end{eqnarray*}
with $E_{\boldsymbol\theta_0}\left(\left|\zeta_i^{*}\right|^3\right)<\infty$ being proved in the same lines that the first two moments presented in the Supplementary Material, which completes the proof of the desired result. $\square$
\end{document}
|
arXiv
|
{
"id": "2012.01807.tex",
"language_detection_score": 0.737532377243042,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\definecolor{light-gray}{gray}{0.95} \lstset{columns=fullflexible, basicstyle=\ttfamily,
backgroundcolor=\color{white},xleftmargin=0.5cm,frame=lr,framesep=8pt,framerule=0pt}
\begin{frontmatter}
\title{Pseudo Bayesian Mixed Models under Informative Sampling} \runtitle{Pseudo Bayesian Mixed Models}
\author{\fnms{Terrance D.} \snm{Savitsky}\thanksref{t1}\ead[label=e1]{[email protected]}} \and \author{\fnms{Matthew R.} \snm{Williams}\thanksref{t2,t3}\ead[label=e2]{[email protected]}}
\thankstext{t1}{U.S. Bureau of Labor Statistics, Office of Survey Methods Research,
Washington, DC, USA,
[email protected]
} \thankstext{t2}{National Center for Science and Engineering Statistics,
National Science Foundation,
Alexandria, VA, USA,
[email protected]
} \thankstext{t3}{Corresponding Author}
\runauthor{Savitsky and Williams}
\begin{abstract} When random effects are correlated with survey sample design variables, the usual approach of employing individual survey weights (constructed to be inversely proportional to the unit survey inclusion probabilities) to form a pseudo-likelihood no longer produces asymptotically unbiased inference. We construct a weight-exponentiated formulation for the random effects distribution that achieves approximately unbiased inference for generating hyperparameters of the random effects. We contrast our approach with frequentist methods that rely on numerical integration to reveal that only the Bayesian method achieves both unbiased estimation with respect to the sampling design distribution and consistency with respect to the population generating distribution. Our simulations and real data example for a survey of business establishments demonstrate the utility of our approach across different modeling formulations and sampling designs. This work serves as a capstone for recent developmental efforts that combine traditional survey estimation approaches with the Bayesian modeling paradigm and provides a bridge across the two rich but disparate sub-fields. \end{abstract}
\begin{keyword} \kwd{Labor force dynamics} \kwd{Markov chain Monte Carlo} \kwd{Pseudo-Posterior distribution} \kwd{Survey sampling} \kwd{Weighted likelihood} \end{keyword}
\end{frontmatter}
\section{Introduction} Hierarchical Bayesian models provide a flexible and powerful framework for social science and economic data, which often include nested units of analysis such as industry, geography, and individual. Yet, social science and economic data are commonly acquired from a survey sampling procedure. It is typically the case that the underlying survey sampling design distribution governing the procedure induces a correlation between the response variable(s) of interest and the survey sampling inclusion probabilities assigned to units in the target population from which the sample was taken. Survey sampling designs where there is a correlation between the response variable and the sampling inclusion probabilities are referred to as informative and will result in the distribution of the response variable in observed samples being different from that from the underlying population about which we seek to perform model-based inference. Sample designs may also be informative when the inclusion probabilities for groups are correlated with the corresponding latent random effects. The resulting distribution of random effects in the sample is different from that of the population of random effects.
The current literature for Bayesian methods has partially addressed population model estimation of survey data under informative sampling designs through the use of survey sampling weights to obtain consistent estimates of fixed effects or top level global parameters. Yet the survey statistics literature suggests that parameters related to random effects, or local parameters are still potentially estimated with bias. The possibility for survey-induced bias in estimation of random effects severely limits the applicability of the full suite of Bayesian models to complex social and economic data.
This paper proposes a Bayesian survey sample-weighted, plug-in framework for the simultaneous estimation of fixed effects and generating hyperparameters (e.g., variance) of random effects that is unbiased with the respect to the distribution over samples and asymptotically consistent with respect to the population distribution.
\subsection{Informative Sampling Designs} Survey sampling designs that induce a correlation between the response variable of interest, on the one hand, and the survey sample inclusion probabilities, on the other hand, are deemed ``informative" and produce samples that express a different balance of information than that of the underlying population, thus estimation methods that do not incorporate sample design information lead to incorrect inferences. For example, the U.S. Bureau of Labor Statistics (BLS) administers the Job Openings and Labor Turnover Survey (JOLTS) to business establishments for the purpose of estimating labor force dynamics, such as the total number of hires, separations and job openings for area-indexed domains. The units are business establishments and their inclusion probabilities are set to be proportional to their total employment (as obtained on a lagged basis from a census instrument). Since the number of hires, separations and job openings for establishments are expected to be correlated to the number of employees, this survey sampling design induces informativeness, so that hiring, separations and openings would be expected to be larger in the samples than in the underlying population.
\subsection{Bayesian Models for Survey Data} There is growing and rich literature on employment of survey sampling weights (constructed to be inversely proportional to unit inclusion probabilities) for correction of the population model estimated on the observed survey sample to produce asymptotically unbiased estimation. Some recent papers focus on the use of Bayesian modeling for the specific purpose of producing mean and total statistics under either empirical or nonparametric likelihoods, but these methods don't allow the data analyst to specify their own population models for the purpose of parameter estimation and prediction \citep{kunihama:2014,wu:2010,si2015}. In particular, the set-up for our paper is one where the data analyst has specified a particular Bayesian hierarchical model for the population (from which the sample was taken) under which they wish to perform inference from data taken from a complex sampling design. So, having to specify a model that is specific to the realized sample, but unrelated to the population model constructed by the data analyst does not allow them to conduct the inference they seek. \citet{2015arXiv150707050S} and \citet{2018dep} complement these efforts by employing a pseudo-posterior to allow the data analyst to estimate a population model of their choice on an informative sample taken from that population. The pseudo-likelihood exponentiates each unit likelihood contribution by its sampling weight, which re-balances information in the observed sample to approximate that in the population. The use of the pseudo-posterior may be situated in the more general class of approximate or composite likelihoods \citep{2009arXiv0911.5357R}. All of the above Bayesian approaches that allow analyst specification of the underlying population generating model to be estimated on the observed informative sample \emph{only} address models with fixed or global effects, not random effects. Yet, it is routine in Bayesian modeling to employ one or more sets of random effects under prior formulations designed to capture complex covariance structures. Hierarchical specifications make such population models readily estimable.
\subsection{Extending the Pseudo-Posterior to Mixed Effects Models} There are two survey designs considered in this paper: 1. Clusters or groups of units are sampled in a first stage, followed by the sampling of nested units in a second stage. We refer to this procedure as the ``direct" sampling of clusters or groups; 2. Units are sampled in a single stage without directly sampling the clusters or groups in which they naturally nest (e.g., geography). We refer to the case where groups used in the population model are not included in the sampling design as ``indirect" sampling of groups, since a group is included in the sample to the extent that a nested unit is directly sampled.
This paper extends the approaches of \citet{2015arXiv150707050S} and \citet{2018dep} from global-only parameters to mixed effects (global and local parameter) models by exponentiating both the data likelihood contributions \emph{and} the group-indexed random effects prior distributions by sampling weights - an approach that we label, ``double-weighting" - that is multiplied, in turn, by the joint prior distribution for the population model parameters to form a joint pseudo-posterior distribution with respect to the observed data and random effects for the sampled groups.
Our augmented (by sample-weighting the prior for random effects) pseudo-posterior method introduced in the next section is motivated by a data analyst who specifies a population generating (super population) model that includes group-indexed random effects for which they desire to perform inference. The observed data are generated under an informative sampling design such that simply estimating the population model on the observed sample will produce biased parameter estimates. Our augmented pseudo-posterior model estimator uses survey sampling weights to perform a relatively minor adjustment to the model augmented likelihood such that parameter draws taken on the observed informative sample approximates inference with respect to the population generating distribution (and the resulting parameter estimates are consistent with respect to the population generating distribution).
We demonstrate that our pseudo-posterior formulation achieves both unbiasedness with respect to the two-stage survey sampling design distribution \emph{and} consistency with respect to the population generating (super population) model for the observed response variable and the latent cluster/group random effects under both direct and indirect sampling of groups. Our weighted pseudo-posterior is relatively simple in construction because we co-sample the random effects and model parameters in our numerical Markov chain Monte Carlo (MCMC) scheme and marginalize out the random effects \emph{after} estimation.
\begin{comment}
, $P^{\pi}(\bm{\delta})$ for sampled inclusion indicators $\delta_{h}, \delta_{\ell | h},\delta_{h\ell }\in \{0,1\}$ with $\pi_{h} = Pr(\delta_{i} = 1)$, $\pi_{\ell | h} = Pr(\delta_{h|\ell} = 1)$, and $\pi_{h\ell} = \pi_{\ell | h} \pi_{h}$, \emph{and} consistency with respect to the population generating (super population) model, $P_{\theta_{0}, \phi_{0}}(\bm{y}, \bm{u})$ for response $y_{h\ell}$ and cluster random effects $u_{h}$ under both direct and indirect sampling of groups. \end{comment}
\begin{comment} We specify and evaluate alternative forms for induced group inclusion probabilities used to formulate weights for group-indexed random effects distributions in the case of indirect group sampling. We both formalize and extend \citet{doi:10.1081/STA-120017802} that discusses, but does not implement, the possibility for inducing group-indexed weights under indirect sampling of groups. We discuss the extensions of frequentist consistency for our Bayesian survey sampling-weighted mixed effects models under each of direct and indirect sampling of groups by generalizing \citet{2018dep}. \end{comment} The case of indirect sampling of groups is particularly important in Bayesian estimation as it is common to specify multiple random effects terms that parameterize a complex covariance structure because the random effects terms are readily estimable in an MCMC scheme.
\begin{comment} A recent publication of \citet{doi:10.1111/rssa.12362} addresses frequentist estimation of a population mean under informative non-response from a population model that includes unit-indexed random effects. The score function for non-responding units is replaced with it’s expectation with respect to the distribution for non-responding units and then uses a Bayes' rule formula to replace that expectation with respect a ratio of expectations with respect to distribution for responding units. The approach next uses expectation conditioning to compute a double integral where the first conditions on observation indexed random effects and the second integrates the random effects with respect to the distribution for the observed sample units. Because the approach focuses on estimation of the population model and produces a complex expression (composed of a ratio of expectations), they simply replace the population estimates for random effects with those computed from the sample as a plug-in. Porting this method for Bayesian estimation would require numerical computation of complicated integrals on every iteration of the posterior sampler and it would not address more general inference beyond the population mean. A major advantage of our pseudo-posterior construction under double-weighting is that the data analyst is only required to make minor changes to the posterior sampler designed for their population model to insert weights into the computations of the full conditional distributions. \end{comment}
The remainder of the paper proceeds to develop our proposed double-weighting methods for estimation of mixed effects models under both direct and indirect sampling of groups on data acquired under informative sampling in Section~\ref{doublepseudo}. Simulation studies are conducted in Section~\ref{simulation} that compare our proposed method to the usual case of likelihood weighting under direct sampling of groups. Section~\ref{application} applies our double-weighting method to estimation of the number of hires for business establishments under employment of industry-indexed random effects in the population model where we reveal that our double-weighting approach produces more heterogeneous random effects estimates to better approximate the population from the observed sample than does the usual practice. We offer a concluding discussion in Section~\ref{discussion}.
\section{Mixed Effects Pseudo Posterior Distribution}\label{doublepseudo} \begin{comment} \subsection{Fixed Effects Pseudo Posterior} \citet{2015arXiv150707050S} and \citet{2017pair} construct a sampling-weighted pseudo-likelihood that provides a noisy approximation to the population likelihood that when convolved with a prior distribution induces a pseudo-posterior distribution for \emph{fixed} effects models. We begin by briefly reviewing this construction before we proceed to extend their formulation to a linear, mixed effects population model construction that employs random effects.
Suppose there exists a Lebesgue measurable population-generating density, $\pi\left(y\vert\theta\right)$, indexed by parameters, $\theta \in \Theta$, where values for a response variable of interest, $y_{1},\ldots,y_{N} \in P_{\theta_{0}},~\theta_{0}\in\Theta$. Let $\delta_{i} \in \{0,1\}$ denote the sample inclusion indicator for unit, $i = 1,\ldots,N$, from a population, $U$. The density for the \emph{observed} sample, $S = (1,\ldots,n)$ (where the index of units drawn from the population, $U$, are re-labeled sequentially without loss of generality) is denoted by, $f\left(y\vert\theta\right) := \pi\left(y\vert \delta = 1,\theta\right)$.
The plug-in estimator for posterior density under the analyst-specified model for $\theta \in \Theta$ is \begin{equation} f^{\pi}\left(\theta\vert \mathbf{y}\right) \propto \left[\mathop{\prod}_{j = 1}^{n}f\left(y_{j}\vert \theta\right)^{w_{j}}\right]f\left(\theta\right), \label{pseudolike} \end{equation} where $\mathop{\prod}_{j = 1}^{n}f\left(y_{j}\vert \theta\right)^{w_{j}}$ denotes the pseudo-likelihood for observed sample responses, $\mathbf{y}$. The joint prior density on model space assigned by the analyst is denoted by $f\left(\theta\right)$. The sampling weights, $\{w_{j} \propto 1/\pi_{j}\}$, are inversely proportional to unit inclusion probabilities and normalized to sum to the sample size, $n$. Let $f^{\pi}(\cdot\vert \mathbf{y})$ denote the noisy approximation to (the population) posterior distribution, $f(\cdot\vert\mathbf{y})$, based on the data, $\mathbf{y}$, and sampling weights, $\{\mathbf{w}\}$, confined to those units \emph{observed} in the sample, $S$. \citet{2017pair} demonstrate asymptotic consistency of the pseudo-posterior distribution with respect to the joint distribution, $(P_{\theta_{0}},P^{\pi})$, that governs population generation and the taking of an informative sample, $S$, from the underlying population, $U$. \end{comment}
The focus of this paper addresses sampling units naturally collected into a population of groups; for example, defined by geography. There is typically a dependence among the response values for units within each group such that units are more similar within than between groups. Sampling designs are typically constructed as multi-stage where the collection of groups in the population are first sampled, followed by the sequential taking of a sub-sample of units from the population of units within \emph{selected} or sampled groups.
By contrast, an alternative set of sampling designs may proceed to draw a sample from the population of units in a \emph{single} stage such that the groups are included the sample, indirectly, when one or more of their member units are selected. These two sampling designs - sampling groups, followed by sampling units within groups in a multi-stage sampling design, on the one hand, as compared to a single-stage sampling of units without directly sampling their member groups, on the other hand - will lead us to design two formulations for extending the pseudo-posterior distribution of \citet{2018dep}.
The pseudo-posterior exponentiates the likelihood contributions of the \emph{single level} fixed effects model (that does not utilize random effects) by the survey weights for the observed sample units, which are inversely proportional to their probabilities of being selected into the sample $w_{i} \propto 1/\pi_{i}$: \begin{equation}\label{eq:pseudopost}
f^{\pi}\left(\theta\vert \mathbf{y},\tilde{\mathbf{w}}\right) \propto \left[\textcolor{red}{\mathop{\prod}_{i = 1}^{n}f\left(y_{i}\vert \theta\right)^{\tilde{w}_{i}}}\right]f\left(\theta\right) \end{equation} where the normalized weights sum to the sample size $\tilde{w}_{i} = w_{i}/\frac{\sum w_{i}}{n}$. The sum of the weights directly affects the amount of posterior uncertainty estimated in the posterior distribution for $\bm{\theta}$, so normalizing it to sum to the effective sample size regulates that uncertainty. Equation \ref{eq:pseudopost} is a noisy approximation to the true (but not fully known) joint distribution of the population model $P_{\theta_{0}}(\mathbf{y})$ and the sampling process $P^{\pi}(\bm{\delta})$, where $\bm{\delta}$ denotes a vector of sample design inclusion indicators for units and groups (that are governed by $P^{\pi}$) and formally defined under the 2-stage class of sampling designs considered in this paper in the upcoming section.
The noisy approximation for the population likelihood obtained by constructing the sample-weighted pseudo-posterior estimator for the observed (informative) sample leads to consistent estimation of population generating (super-population) parameters $\theta$ (as the sample size, $n$, grows) .
The pseudo-posterior construction requires only a minor change to the population model specified by the data analyst on which they wish to perform inference (by weighting each unit-indexed likelihood contribution by its associated marginal sampling weight). In particular, the data analyst may specify population distributions for $f\left(y_{i}\vert \theta\right)$ and priors $f(\bm{\theta})$; for example, if the data are count data that we work with in the sequel, they may specify a Poisson likelihood with mean, $\bm{\mu}$, for which they define a latent regression model formulation. The data analyst is interested to perform inference for the generating parameters under the population generation and not the distribution of the observed sample. Under informative sampling the two distributions are different and the pseudo-posterior corrects the distribution of the observed sample back to the population of interest.
We demonstrate in the sequel, that the formulation in Equation \ref{eq:pseudopost} can be extended to multi-level models by exponentiating \emph{both} the likelihood (conditioned on the random effects) \emph{and} the prior distribution for random effects by sampling weights.
\subsection{Mixed Effects Posterior Under \emph{Direct} Sampling of Population Groups}\label{sec:direct} Assign units, $i \in (1,\ldots,N)$, that index a population, $U$, to groups, $h \in (1,\ldots,G_{U})$, where each population group, $h$, nests $U_{h} = 1,\ldots,N_{h}$ units, such that $N = \lvert U \rvert = \mathop{\sum}_{h = 1}^{G_{U}}N_{h}$, with $N_{h} = \lvert U_{h} \rvert$. Construct a 2-stage informative sampling design whose first stage takes a direct sample of the $G_{U}$ groups, where $\pi_{h}\in(0,1]$ denotes the marginal sample inclusion probability for group, $h \in (1,\ldots,G_{U})$. Let $g \in (1,\ldots,G_{S})$, index the \emph{sampled} groups, where $G_{S}$ denotes the number of observed groups from the population of groups, $G_{U} \supset G_{S}$.
Our first result defines a pseudo-posterior estimator on the observed sample for our population model that includes group-indexed random effects in the case where we \emph{directly} sample groups, followed by the sampling of units nested within selected groups, in a multistage survey sampling design. Our goal is to achieve unbiased inference for $(\theta,\phi)$ (where $\theta$ denotes fixed effects for generating population responses, $y$, and $\phi$ denotes the generating parameters of random effects, $u$, for the population), estimated on our observed sample taken under an informative survey sampling design. We assume that random effects $u$ are indexed by group and are independent conditional on the generating parameter $\phi$. Multistage designs that sample groups or clusters, followed by the further sampling of nested units, are commonly used for convenience to mitigate costs of administration where in-person interviews are required and also in the case where a sampling frame of end-stage units is constructed after sampling groups in the first stage. The second stage of the survey sampling design takes a sample from the $N_{g}$ (second stage) units $\forall g \in S_{g}$, where $S_{g} \subset U_{g}$. The second stage units are sampled with conditional inclusion probabilities, $\pi_{\ell\mid g} \in (0,1]$ for $\ell = 1,\ldots,N_{g}$, conditioned on inclusion of group, $g\in (1,\ldots,G_{S})$. Let $j \in (1,\ldots,n_{g})$ index the sampled or observed second stage units linked to or nested within sampled group, $g \in (1,\ldots,G_{S})$. Denote the marginal group survey sampling weight, $w_{g} \propto 1/\pi_{g}$ for $\pi_{g} \in (0,1]$. Denote the marginal unit survey sampling weight, $w_{g j} \propto 1/\pi_{g j}$, for $\pi_{g j} \in (0,1]$, the joint inclusion probability for unit, $j$, nested in group, $g$, both selected into the sample. The group marginal inclusion probabilities and conditional unit inclusion probabilities under our 2-stage survey sampling design are governed by distribution, $P^{\pi}$.
\begin{theorem}\label{th:direct}
Under a proper prior specification, $f(\bm{\theta})f(\bm{\phi})$, the following pseudo-posterior estimator achieves approximately unbiased inference for super-population (population generating) model, $f(\bm{\theta},\bm{\phi}|\mathbf{y})$, with respect to the distribution governing the taking of samples from an underlying finite population, $P^{\pi}$, \begin{equation}\label{eq:sampdirmodel} f^{\pi}\left(\theta, \phi \vert \mathbf{y} \right) \propto \left[\mathop{\int}_{\mathbf{u} \in \mathcal{U}} \left\{ \textcolor{blue}{\mathop{\prod}_{g\in S}} \left( \textcolor{red}{\mathop{\prod}_{j \in S_{g}} f\left(y_{g j}\mid u_{g}, \theta \right)^{w_{g j}}} \right) \textcolor{blue}{f\left(u_{g}\mid \phi \right)^{w_g}} \right\} d\mathbf{u} \right]f(\theta)f(\phi). \end{equation}
where $f^{\pi}(\cdot)$ denotes a sampling-weighted pseudo-distribution, $j \in S_{g}$ denotes the subset of units, $j \in (1,\ldots,n_{g} = \lvert S_{g}\rvert)$, linked to group, $g\in (1,\ldots,G_{S})$. Parameters, $(\bm{\theta},\bm{\phi})$, index the super-population model posterior distribution, $f(\bm{\theta},\bm{\phi}|\mathbf{y})$, that is the target for estimation. The integral for the vector $\mathbf{u} = (u_{1},\ldots,u_{n_{g}})$ is taken over its support, $\mathcal{U}$, for each component, $u_{g} \in \mathbf{u}$. \end{theorem}
We employ a pseudo-likelihood for the first level of the model for sampled observations $y_{gj}$ within sampled clusters $g$ by exponentiating by the sample weight $w_{gj}$. This provides a noisy approximation to the first stage likelihood. For the second level model (or prior) for the random effects $u_{g}$, we exponentiate this distribution by its corresponding sampling weights $w_{g}$. This provides a noisy approximation to the population distribution of random effects. Both approximations are needed because the distributions of both the responses and the random effects in the sample can differ substantially from those in the corresponding population due to the informative sampling design at both the cluster $g$ and the within cluster $j|g$ stages, where the latter notation denotes the sampling of unit $j$ conditioned on / within sampled group $g$.
Under our augmented (by weighting the prior of the group-indexed random effects) pseudo-likelihood of Equation~\ref{eq:sampindirmodel}, $f\left(y_{g j}\mid u_{g}, \theta \right)$ and $f\left(u_{g}\mid \phi \right)$ are not restricted; for example, we select a Poisson distribution for the observed data likelihood, $f\left(y_{g j}\mid u_{g}, \theta \right)$, for our simulation study and application in the sequel. Similarly, the form of the distribution for the random effects prior distribution is not restricted under our construction, though it is most commonly defined as Gaussian under a GLM specification. Replacing the single Gaussian with a mixture of Gaussian distributions would also fit our set-up. Our approach also readily incorporates additional levels of random effects with no conceptual changes.
\begin{proof} We first construct the complete joint model for the finite population, $U$, as if the random effects, $(u_{h})$, were directly observed, \begin{equation}\label{eq:popmodel} f_{U}\left(\theta, \phi \vert \mathbf{y},\mathbf{u} \right) \propto \left[\mathop{\prod}_{h = 1}^{G_{U}} \left(\mathop{\prod}_{\ell= 1}^{N_{h}} f\left(y_{h \ell}\mid u_{h}, \theta \right)\right) f\left(u_{h}\mid \phi \right) \right]f(\theta)f(\phi). \end{equation}
Under a complex sampling design, we specify random sample inclusion indicators for groups, $\delta_h$, with marginal probabilities $\pi_{h} = P(\delta_h= 1)$ for $h \in (1,\ldots,G_{U})$, governed by $P^{\pi}$. We further specify random sample inclusion indicator, $\delta_{\ell \mid h} = (\delta_{\ell} \mid \delta_h = 1) \in \{0,1\}$, with probability $\pi_{\ell\mid h} = P(\delta_{\ell \mid h}= 1)$, for unit $\ell \in (1,\ldots,N_{h})$, conditioned on the inclusion of group, $h$, such that the indicator for the joint sampling of unit $\ell$ nested within group $h$ is denoted as $\delta_{h\ell} = \delta_{\ell \mid h} \times \delta_{h}$, with the associated marginal inclusion probability, $\pi_{h\ell} = P(\delta_{h\ell}=1)$.
The taking of an observed sample is governed by the survey sampling distribution, $P^{\pi}$ (as contrasted with $P_{\theta,\phi}$, the population generation distribution for $(y,u)$). The pseudo-likelihood with respect to the joint distribution, $(P^{\pi},P_{\theta,\phi})$, is then constructed by exponentiating components of the likelihood in the population such that the \emph{expected value} of the survey sample pseudo log-likelihood function with respect to $P^{\pi}$ equals that of the log-likelihood for the entire population (and thus the score functions also match in expectation). Let $\ell_{U}\left(\mathbf{y}, \mathbf{u} \vert \theta, \phi \right) \equiv \log f_{U}\left(\theta, \phi \vert \mathbf{y},\mathbf{u} \right)$ denote the population model log-likelihood. Applying this approach to the log-likelihood of the joint model, above, leads to the following pseudo- likelihood formulation: \begin{align} \ell ^{\pi}_{U}\left(\mathbf{y}, \mathbf{u} \vert \theta, \phi \right) &\propto \mathop{\sum}_{h=1}^{G_{U}} \left(\mathop{\sum}_{\ell = 1}^{N_{h}} \left(\frac{\delta_{\ell\mid h}}{\pi_{\ell\mid h}}\right)\left(\frac{\delta_h}{\pi_h}\right) \ell \left(y_{h \ell}\mid u_{h}, \theta \right)\right) +
\left(\frac{\delta_h}{\pi_h}\right) \ell \left(u_{h}\mid \phi \right)\\
& = \mathop{\sum}_{h=1}^{G_{U}} \left(\mathop{\sum}_{\ell = 1}^{N_{h}} \left(\frac{\delta_{h \ell}}{\pi_{h \ell}}\right) \ell \left(y_{h \ell}\mid u_{h}, \theta \right)\right) +
\left(\frac{\delta_h}{\pi_h}\right) \ell \left(u_{h}\mid \phi \right)\label{estimator} \end{align} where $P^{\pi}$ governs all possible samples, $(\delta_{h},\delta_{\ell\mid h})_{\ell\in U_{h}, h=1,\ldots,G_{U}}$, taken from population, $U$. Let joint group-unit inclusion indicator, $\delta_{h \ell} = \delta_{h}\times\delta_{\ell\mid h}$ with $\pi_{h \ell} = P(\delta_{h \ell}= 1) = P(\delta_{h}=1,\delta_{\ell\mid h} = 1)$. For each \emph{observed} sample $\ell ^{\pi}_{U}\left(\mathbf{y}, \mathbf{u} \vert \theta, \phi \right) = \ell ^{\pi}_{S}\left(\mathbf{y}, \mathbf{u} \vert \theta, \phi \right)$ where
\begin{equation}\label{estimatorsample} \ell ^{\pi}_{S}\left(\mathbf{y}, \mathbf{u} \vert \theta, \phi \right) = \mathop{\sum}_{g=1}^{G_{S}} \left(\mathop{\sum}_{j \in S_{g}} w_{g j} \ell \left(y_{g j}\mid u_{g}, \theta \right)\right) +
w_g \ell \left(u_{g}\mid \phi \right) \end{equation} and $w_{g j} \propto \pi_{g,j}^{-1}$ and $w_g \propto \pi_{g}^{-1}$. The expectation of our estimator in Equation~\ref{estimator} is unbiased with respect to $P^{\pi}$, \begin{align} \mathbb{E}^{\pi}\left[\ell ^{\pi}_{U}\left(\mathbf{y}, \mathbf{u} \vert \theta, \phi \right)\bigg\vert P_{\theta,\phi}\right] &\equiv \\ \mathbb{E}^{\pi}\left[\ell ^{\pi}_{U}\left(\mathbf{y}, \mathbf{u} \vert \theta, \phi \right)\right] &= \ell_{U}\left(\mathbf{y}, \mathbf{u} \vert \theta, \phi \right)\label{unbiased}, \end{align} where the expectation, $\mathbb{E}^{\pi}(\cdot)$, is taken with respect to the survey sampling distribution, $P^{\pi}$, that governs the survey sampling inclusion indicators, $\{\delta_{h \ell}, \delta_h\}$, conditional on the data $\{\mathbf{y}, \mathbf{u}\}$ generated by $P_{\theta,\phi}$. The final equality in Equation~\ref{unbiased} is achieved since $\mathbb{E}^{\pi}(\delta_{h \ell}) = \pi_{h \ell}$ and $\mathbb{E}^{\pi}(\delta_{h}) = \pi_{h}$.
Thus, we use the following sampling-weighted model approximation to the complete population model of Equation~\ref{eq:popmodel}: \begin{equation}\label{eq:sampcompletemodel} f^{\pi}\left(\theta, \phi \vert \mathbf{y}, \mathbf{u}\right) \propto \left[\mathop{\prod}_{g \in S}\left(\mathop{\prod}_{j \in S_{g}} f\left(y_{g j}\mid u_{g}, \theta \right)^{w_{g j}}\right) f\left(u_{g}\mid \phi \right)^{w_g} \right]f(\theta)f(\phi). \end{equation} We can then construct a sampling-weighted version of the observed model: \begin{equation}\label{eq:sampobsmodel} f^{\pi}\left(\theta, \phi \vert \mathbf{y} \right) \propto \left[\mathop{\int}_{\mathbf{u} \in \mathcal{U}} \left\{ \mathop{\prod}_{g \in S} \left(\mathop{\prod}_{j \in S_{g}} f\left(y_{g j}\mid u_{g}, \theta \right)^{w_{g j}}\right) f\left(u_{g}\mid \phi \right)^{w_g} \right\} d \mathbf{u} \right]f(\theta)f(\phi). \end{equation}
The walk from Equation~\ref{eq:sampcompletemodel} to Equation~\ref{eq:sampobsmodel} is possible because we co-estimate the $(\mathbf{u})$ with $(\theta,\phi)$ and then perform the integration step to marginalize over the $(\mathbf{u})$ \emph{after} estimation. \end{proof}
Theorem~\ref{th:direct} requires the exponentiation of the prior contributions for the sampled random effects, $(u_{g})$, by a sampling weight, $w_{g} \propto 1/\pi_{g}$ in order to achieve approximately unbiased inference for $\phi$; it is not enough to exponentiate each data likelihood contribution, $f\left(y_{g j}\mid u_{g}, \theta \right)$, by a unit (marginal) sampling weight, $w_{g j}$. This formulation is generally specified for \emph{any} population generating model, $P_{\theta,\phi}$. Our result may be readily generalized to survey sampling designs of more than two stages where each collection of later stage groups are nested in earlier stage groups (such as households of units nested within geographic PSUs).
The proposed method under direct sampling of Equation~\ref{eq:sampdirmodel} is categorized as a plug-in estimator that exponentiates the likelihood contributions for nested units by the unit-level marginal sampling weights \emph{and}, in turn, exponentiates the prior distribution for cluster-indexed random effects by the cluster (or PSU) marginal sampling weights. Samples from the joint pseudo-posterior distribution over parameters are interpreted as samples from the underlying (latent) population generating model since the augmented (by the weighted random effects prior distribution) pseudo-likelihood estimated on the observed sample provides a noisy approximation for the population generating likelihood.
Although the augmented pseudo-likelihood is unbiased with respect to the distribution over samples, it will not, generally, produce correct uncertainty quantification. In particular, the credibility intervals will be too optimistic relative to valid frequentist confidence intervals because the plug-in method is not fully Bayesian in that it doesn't model uncertainty in the sampling weights, themselves.
Although the employment of random effects captures dependence among nested units, the warping and scaling induced by the sampling weights will result in failure of Bartlett's second identity such that the asymptotic hyperparameter covariance matrix for our plug-in mixed effects model will not be equal to the sandwich form of the asymptotic covariance matrix for the MLE. The result of the lack of equality is that the model credibility intervals, without adjustment, will not contract on valid frequentist confidence intervals.
A recent work of \citet{leonnovelo2021fully} jointly models the unit level marginal sampling weights and the response variable and includes group-indexed random effects parameters in their joint model. They demonstrate correct uncertainty quantification because the asymptotic covariance matrix of their fully Bayesian model (that also co-models the sampling weights) is equal to that for the MLE. Their method specifies an exact likelihood for the observed sample that is complicated and requires a closed-form solution for an integral that restricts the class of models that may be considered. This approach requires a different model formulation than that specified for the population and of interest to the data analyst.
By contrast, our plug-in augmented pseudo-posterior distribution requires only minor change to the underlying population model specified by the data analyst and may be easily adapted to complicated population models.
Correct uncertainty quantification for $(\bm{\theta},\bm{\phi})$ may be achieved by using the method of \citet{2018arXiv180711796W} to perform a post-processing of the MCMC parameter draws that replaces the pseudo-posterior covariance with the sandwich form of the MLE. This paper, by contrast, focuses on providing unbiased point estimation for mixed effects models as an extension of \citet{2015arXiv150707050S} because \citet{2018arXiv180711796W} may be readily applied, post sampling.
\begin{comment} There are two related papers \citep{pfeffmix:1998, rh:1998} that address unbiased inference for linear mixed models under a continuous and dichotomous response, respectively, in the frequentist setting where the random effects utilized in the population model index groups used to draw nested units in a multistage sampling design where, in the case of a 2-stage sampling design, each group in the population is assigned an inclusion probability. In a set-up where the inclusion probabilities for the random effects are informative, both papers accomplish estimation by multiplying the logarithm of the distribution for each group-indexed random effect by a weight set to be inversely proportional to the group inclusion probability. The log-likelihood contribution for each unit, which is nested in a group, is multiplied by a weight set to be inversely proportional to the conditional inclusion probability for that unit, given that its group was sampled in an earlier stage. Each paper evaluates alternatives for normalizing the collection of conditional units weights in each group in order to reduce the bias for estimation of the generating random effects variance in the case of a small number of units linking to each group, though they both recommend normalizing to the sum of within group sample size. \end{comment}
Our pseudo-likelihood in Equation~\ref{estimator} is jointly conditioned on $(\mathbf{y},\mathbf{u})$, such that the random weights, $\left(\delta_{h \ell}/\pi_{h \ell}\right)$, are specified in linear summations. This linear combination of weights times log-likelihoods ensures (design) unbiasedness with respect to $P^{\pi}$ because the weight term is separable from the population likelihood term. We may jointly condition on $(\mathbf{y},\mathbf{u})$ in our Bayesian set-up because we co-sample $(\mathbf{u},\theta,\phi)$, numerically, in our MCMC such that the integration step over $\mathbf{u}$ is applied \emph{after} co-estimation. In other words, we accomplish estimation in our MCMC by sampling $\mathbf{u}$ jointly with $(\theta,\phi)$ on each MCMC iteration and then ignoring $\mathbf{u}$ to perform marginal inferences on $\theta$ and $\phi$, which is a common approach with Bayesian hierarchical models. By contrast, \citet{pfeffmix:1998}, \citet{rh:1998}, and others \citep[for example see][]{kim2017statistical} specify the following integrated likelihood under frequentist estimation for an observed sample where units are nested within groups, \begin{equation} \ell^{\pi}(\theta,\phi) = \mathop{\sum}_{g=1}^{G_{S}} w_{g} \ell_{i}^{\pi}(\theta,\phi), \end{equation} for $\ell_{i}^{\pi}(\theta,\phi) = \log L_{i}^{\pi}(\theta,\phi)$ for, \begin{equation}\label{eq:frequnitlike} L_{i}^{\pi}(\theta,\phi) = \mathop{\int}_{u_{g} \in \mathcal{U}}\exp\left[\mathop{\sum}_{j \in S_{g}} w_{j\mid g}\ell \left(y_{g j}\mid u_{g}, \theta \right)\right]f\left(u_{g}\mid \phi \right)du_{g}, \end{equation} which will \emph{not}, generally, be unbiased with respect to the distribution governing the taking of samples for the population likelihood because the unit level conditional weights, $(w_{j\mid g})_{j}$, are nested inside an exponential function (such that replacing $w_{j\mid g}$ with $\delta_{\ell\mid h}/\pi_{\ell \mid h}$ inside the exponential and summing over the population groups and nested units will not produce separable sampling design terms that each integrate to $1$ with respect to $P^{\pi}$, conditioned on the generated population) \citep{yi:2016}. The non-linear specification in Equation~\ref{eq:frequnitlike} results from an estimation procedure that integrates out $\mathbf{u}$ \emph{before} pseudo-maximum likelihood point estimation of $(\theta,\phi)$.
This design biasedness (with respect to $P^{\pi}$) is remedied for pseudo-maximum likelihood estimation by \citet{yi:2016} with their alternative formulation, \begin{align} \ell^{\pi}(\theta,\phi) &=\mathop{\sum}_{g=1}^{G_{S}} w_{g} \mathop{\sum}_{j < k; j,k\in S_{g}} w_{j,k\mid g} \ell_{gj,k}(\theta,\phi)\label{composite}\\ \ell_{gj,k}(\theta,\phi) &= \log \left\{\mathop{\int}_{u_{g} \in \mathcal{U}}f(y_{g j}\mid u_{g},\theta)f(y_{g k}\mid u_{g},\theta)f(u_{g}\mid\phi)du_{g} \right\}, \end{align} where $w_{j,k\mid g} \propto 1/\pi_{j,k\mid g}$ denotes the joint inclusion probability for units $(j,k)$, both nested in group, $g$, conditioned on the inclusion of group, $g$, in the observed sample. Equation~\ref{composite} specifies an integration over $u_{g}$ for \emph{each} $f(y_{g j}\mid u_{g},\theta)f(y_{g k}\mid u_{g},\theta)$ pair, which allows the design weights to enter in a linear construction outside of each integral. This set-up establishes linearity for inclusion of design weights, resulting in unbiasedness with respect to the distribution governing the taking of samples for computation of the pseudo-maximum likelihood estimate, though under the requirement that pairwise unit sampling weights be published to the data analyst or estimated by them.
Yet, the marginalization of the random effects \emph{before} applying the group weight, $w_{g}$, fails to fully correct for the prior distribution for $u_{g}$. We show in the sequel that $\bm{\phi}$ is estimated with bias by \citet{yi:2016} due to this integration of the random effects being performed on the unweighted prior of $u_g$. Our method, by contrast, weights the prior for $u_{g}$ and performs the integration of $u_{g}$ numerically in our MCMC (after sampling $u_{g}$).
\subsection{Mixed Effects Posterior Under \emph{Indirect} Sampling of Population Groups}\label{sec:indirect} Bayesian model specifications commonly employ group-level random effects (often for multiple simultaneous groupings) to parameterize a complex marginal covariance structure. Those groups are often not \emph{directly} sampled by the survey sampling design. We, next, demonstrate that weighting the prior contributions for the group-indexed random effects is \emph{still} required, even when the groups are not directly sampled, in order to achieve unbiased inference for the generating parameters of the random effects, $\phi$. Again, as throughout, we assume the group-indexed random effects are conditionally independent given generating parameter $\phi$. We focus our result on a simple, single-stage sampling design, that may be readily generalized, where we reveal that the group-indexed survey sampling weights are constructed from unit marginal inclusion probabilities. Constructing sampled group weights from those of member units appeals to intuition because groups are included in the observed sample only if any member unit is selected under our single-stage survey sampling design.
Suppose the same population set-up as for Theorem~\ref{th:direct}, with population units, $\ell \in U_{h}$, linked to groups, $h \in (1,\ldots,G_{U})$, where each unit, $(h,\ell)$, maps to $i \in (1,\ldots,N)$. We now construct a \emph{single} stage sampling design that directly samples each ($h,\ell$) unit with marginal inclusion probability, $\pi_{h \ell}$, governed by $P^{\pi}$. Group, $g \in G_{S}$, is \emph{indirectly} sampled based on whether there is \emph{any} linked unit, ($g j$), observed in the sample.
\begin{theorem}\label{th:indirect} The following pseudo-posterior estimator achieves approximately unbiased inference with respect to $P^{\pi}$, \begin{comment} \begin{dmath}\label{eq:sampindirmodel} f^{\pi}\left(\theta, \phi \vert \mathbf{y} \right) \propto \left[\mathop{\prod}_{g\in S}\mathop{\int}_{u_{g} \in \mathcal{U}} \left\{\left(\mathop{\prod}_{j \in S_{g}} f\left(y_{g j}\mid u_{g}, \theta \right)^{w_{g j}}\right) f\left(u_{g}\mid \phi \right)^{w_{g} = \frac{1}{N_{g}}\mathop{\sum}_{j\in S_{g}} w_{g j}} \right\} du_{g} \right]f(\theta)f(\phi), \end{dmath} \end{comment}
\begin{equation}\label{eq:sampindirmodel} \begin{split} f^{\pi}\left(\theta, \phi \vert \mathbf{y} \right) \propto &\left[\mathop{\int}_{\mathbf{u} \in \mathcal{U}} \left\{ \textcolor{blue}{ \mathop{\prod}_{g\in S}} \left( \textcolor{red}{\mathop{\prod}_{j \in S_{g}} f\left(y_{g j}\mid u_{g}, \theta \right)^{w_{g j}}} \right)\right.\right.\\ &\left.\left.\vphantom{
\mathop{\int}_{u_{g} \in \mathcal{U}}\left(\mathop{\prod}_{j \in S_{g}} f\left(y_{g j}\mid u_{g}, \theta \right)^{w_{g j}}\right)
}
\textcolor{blue}{ f\left(u_{g}\mid \phi \right)^{w_g = \frac{1}{N_{g}}\displaystyle\mathop{\sum}_{j\in S_{g}}w_{g j}}}
\right\} d \mathbf{u} \right]f(\theta)f(\phi), \end{split} \end{equation}
where $w_{g j} \propto 1/\pi_{gj}$. \end{theorem} \begin{proof} We proceed as in Theorem~\ref{th:direct} by supposing the population $U$ of units and associated group-indexed random effects, $(u_{h})$, were fully observed. We first construct the likelihood for the fully observed population. \begin{align} f_{U}\left(\theta, \phi \vert \mathbf{y},\mathbf{u} \right) &\propto \left[\mathop{\prod}_{h = 1}^{G_{U}} \left(\mathop{\prod}_{\ell= 1}^{N_{h}} f\left(y_{h \ell}\mid u_{h}, \theta \right)\right) f\left(u_{h}\mid \phi \right) \right]f(\theta)f(\phi)\label{eq:popunitindirect}\\ &=\left[\mathop{\prod}_{h = 1}^{G_{U}} \mathop{\prod}_{\ell= 1}^{N_{h}} \left\{f\left(y_{h \ell}\mid u_{h}, \theta \right) f\left(u_{h}\mid \phi \right)^{\frac{1}{N_{h}}} \right\} \right]f(\theta)f(\phi)\label{eq:popindirect}. \end{align} We proceed to formulate the pseudo-likelihood for all possible random samples taken from $U$, $f^{\pi}_{U}(\cdot)$, governed jointly by $(P^{\pi},P_{\theta,\phi})$, from which we render the pseudo-likelihood for any sample, $f^{\pi}(\cdot)$, which is constructed to be unbiased with respect to the distribution governing the taking of samples for the population model of Equation~\ref{eq:popindirect} under $P^{\pi}$, \begin{align}\label{eq:popindmodel} f^{\pi}_{U}\left(\theta, \phi \vert \mathbf{y}, \mathbf{u}\right) &\propto \left[\mathop{\prod}_{h=1}^{G_{U}}\mathop{\prod}_{\ell \in U_{h}} \left\{f\left(y_{h \ell}\mid u_{h}, \theta \right)f\left(u_{h}\mid \phi \right)^{\frac{1}{N_{h}}}\right\}^{\frac{\delta_{h \ell}}{\pi_{h \ell}}} \right]f(\theta)f(\phi)\\ &=\left[\mathop{\prod}_{h=1}^{G_{U}}f\left(u_{h}\mid \phi \right)^{\frac{1}{N_{h}}\mathop{\sum}_{\ell\in U_{h}}\frac{\delta_{h \ell}}{\pi_{h \ell}}}\mathop{\prod}_{\ell \in U_{h}} f\left(y_{h \ell}\mid u_{\ell}, \theta \right)^{\frac{\delta_{h \ell}}{\pi_{h \ell}}}\right]f(\theta)f(\phi). \end{align} This pseudo-posterior reduces to the following expression for the observed sample, \begin{equation} f^{\pi}\left(\theta, \phi \vert \mathbf{y}, \mathbf{u}\right) \propto \left[\mathop{\prod}_{g=1}^{G_{S}}f\left(u_{g}\mid \phi \right)^{\frac{1}{N_{g}}\mathop{\sum}_{j\in S_{g}}w_{g j}}\mathop{\prod}_{j \in S_{g}} f\left(y_{g j}\mid u_{j}, \theta \right)^{w_{g j}}\right]f(\theta)f(\phi), \end{equation} where $\pi_{g j} = P(\delta_{g j} = 1)$ (under $P^{\pi}$), $w_{g j} \propto 1/\pi_{g,j}$ and $N_{g}$ denotes the number of units in the population linked to observed group, $g \in(1,\ldots,G_{S})$ observed in the sample. We set $w_{g} := 1/N_{g}\times \mathop{\sum}_{j\in S_{g}}w_{g j}$ and the result is achieved. \end{proof} This result derives from eliciting group-indexed weights from unit inclusion probabilities for units linked to the groups. While the resulting pseudo-posterior estimators look very similar across the two theorems, the sampling designs are very different from one another in that groups are not directly sampled in this latter case, which is revealed in their differing formulations for $w_{g}$.
The averaging of unit weights formulation for $w_{g}$ naturally arises under the derivation of Equation~\ref{eq:popindirect} when sampling units, rather than groups under a model that utilizes group-indexed random effects to capture within group dependence that naturally arises among units in the population. Exponentiating the augmented pseudo-likelihood of Equation~\ref{eq:popunitindirect} by survey variables anticipates the integration of the random effects to produce an observed data pseudo-likelihood. We may intuit this result by interpreting this form for $w_{g}$ proportional to the average importance of units nested in group each group, $g$. It bears mention that in the indirect sampling case, there is no probability of group selection defined for a single stage design.
In practice, it is not common for the data analyst to know the population group sizes, $(N_{g})$, for the groups, $g\in (1,\ldots,G_{S})$ observed in the sample, so one estimates an $\hat{N}_{g}$ to replace $N_{g}$ in Equation~\ref{eq:sampindirmodel}. Under a single stage sampling design where the groups are indirectly sampled through inclusions of nested units into the observed sample, we assume that we only have availability of the marginal unit inclusion sampling weights, $(w_{g j})$. The group population size, $N_{g}$, needed for the sum-weights method of Equation~\ref{eq:sampindirmodel}, may be estimated by $\hat{N}_{g} = \mathop{\sum}_{j = 1}^{n_{g}}w_{j\mid g}$. To approximate $w_{j\mid g}$, we first utilize the sum-probabilities result to estimate, $\hat{w}_{g} = 1/\hat{\tilde{\pi}}_{g}$, and proceed to extract $(w_{j\mid g})$ from $w_{gj} \approx w_{g}w_{j\mid g}$. If we invert the resultant group-indexed weight, $w_{g}= 1/N_{g}\times\mathop{\sum}_{j \in S_{g}}w_{g j}$, for the case where groups are not directly sampled, we may view the inverse of the group $g$ weight, $\tilde{\pi}_{g} = 1/w_{g}$, as a ``pseudo" group inclusion probability, since we don't directly sample groups. One may envision other formulations for the pseudo group inclusion probabilities, $\tilde{\pi}_{g}$, that we may, in turn, invert to formulate alternative group-indexed survey sampling weights, $(w_{g})$. Please see Appendix \ref{sec:pseudoprobs} where we develop other methods, in addition to sum-weights, for computing $\tilde{\pi}_{g}$.
In application, we normalize the by-group, survey sampling weights, $\displaystyle (w_{g})_g = 1,\ldots,G_{S}$, to sum to the number of observed groups in the sample, $G_{S}$, and normalize unit weights, $(w_{gj})_{j = 1,\ldots,n_{g}}$ to sum to the overall sample size, $n$. These normalizations regulate uncertainty quantification for posterior estimation of $(u_{g})$ and global parameters, $(\phi,\theta)$ by encoding an effective number of observed groups and units. So, we normalize them to sum to the number of groups and units observed in the sample to regulate the estimated pseudo-posterior variance of $(\phi,\theta)$. (In practice, these normalizations often produce somewhat optimistic credibility intervals due to dependence induced by the survey sampling design. \citet{2018arXiv180711796W} provide an algorithm that adjusts pseudo-posterior draws to incorporate this dependence).
We refer to our proposed procedure for weight exponentiating both the data likelihood contributions and the prior distributions of the $(u_{g})$ as ``double-weighting", as mentioned in the introduction, to be contrasted with the usual approach of ``single-weighting" of \citet{2018dep} developed for models with global effects parameters.
\begin{comment} \subsection{Frequentist Consistency of the Pseudo-Posterior Estimators} The frequentist consistency of our estimators of Theorems~\ref{th:direct} and \ref{th:indirect} with respect to the joint distribution of population generation and the taking of a sample, $(P_{\theta_{0},\phi_{0}},P^{\pi})$, is readily shown under the same conditions and contraction rate as specified in \citet{2018dep}. Let $\nu \in \mathbb{Z}^{+}$ index a collection of populations, $U_{\nu}$, such that $N_{\nu^{'}} > N_{\nu}$, for $\nu^{'} > \nu$. $\nu$ controls the asymptotics such that for each increment of $\nu$ we formulate new units and generate $(y_{\nu},\pi_{\nu})$. Let $p_{\lambda}$ denote a population model density with respect to parameters, $\lambda$. Under direct sampling of groups, we construct the estimator, \begin{equation}\label{eq:theory1} \mathop{\prod}_{h=1}^{G_{\nu}}p_{\phi}\left(u_{\nu h}\right)^{\frac{\delta_{\nu h}}{\pi_{\nu h}}} \times \prod_{\ell = 1}^{N_{\nu g}}p_{\theta}\left(y_{\nu h\ell}\mid u_{\nu h}\right)^{\frac{\delta_{\nu h \ell}}{\pi_{\nu h \ell}}} , \end{equation} which may be readily shown achieves the contraction rate specified in Theorem 1 of \citet{2018dep}. We note that Condition (A4) of \citet{2018dep}, which imposes an upper bound on the inverse of the marginal unit inclusion probability, $1/\pi_{\nu h\ell} = 1/(\pi_{\nu\ell\mid h}\pi_{\nu h}) < \gamma$ serves to bound $(\pi_{\nu h},\pi_{\nu\ell \mid h}) > 0 $ (away from $0$). Condition (A5.2) constructs blocks in which unit pairwise dependencies induced by the survey sampling design are allowed to never attenuate with increasing $\nu$, so long as the blocks grow at less than $\mathcal{O}(N_{\nu})$ and the dependence among units induced by the sampling design \emph{between} the blocks attenuates to $0$. We meet this condition by defining the population groups, $h \in (1,\ldots,G_{U})$, as the blocks, so that asymptotic dependence within the groups is allowed, but asymptotic independence must be achieved between groups. This is a very weak condition that is met by nearly all sampling designs used, in practice, under direct sampling of groups.
Under the indirect sampling of groups, we formulate the estimator, \begin{equation}\label{eq:theory2} \mathop{\prod}_{h=1}^{G_{\nu}}\prod_{\ell = 1}^{N_{\nu g}}\left[p_{\phi}\left(u_{\nu h}\right)^{\frac{1}{N_{\nu h}}} \times p_{\theta}\left(y_{\nu h\ell}\mid u_{\nu h}\right)\right]^{\frac{\delta_{\nu h \ell}}{\pi_{\nu h \ell}}}. \end{equation} This estimator is only a slight generalization from that of \citet{2018dep} and achieves consistency under the specified conditions. We note that the prior contributions, $p_{\phi}(\cdot)$, in both Equations~\ref{eq:theory1} and \ref{eq:theory2} are treated as likelihoods under a population generating model governed by $\phi$.
\end{comment}
\section{Simulation study}\label{simulation} Our simulation study in the sequel focuses on a count data response rather than the usual continuous response, both because count data are the most common data type for the employment data collected by BLS and because our Bayesian construction is readily estimable under any response data type.
We generate a count data response variable, $y$, for a population of size, $N = 5000$ units, where the logarithm of the generating mean parameter, $\mu$ is constructed to depend on a size predictor, $x_{2}$, in both fixed and random effects terms; in this way, we construct both fixed and random effects to be informative, since our proportion-to-size survey sampling design sets unit inclusion probabilities to be proportional to $x_{2}$. We generate a population of responses using, \begin{align}\label{eq:margmodel} y_{i} &\sim \mathcal{P}\left(\mu_{i}\right) \nonumber\\ \log\mu_{i} &= \beta_{0} + x_{1i}\beta_{1} + x_{2i}\beta_{2} + \left[1,x_{2i}\right]\bm{\gamma}_{h\{i\}}, \end{align} where $\mathcal{P}(\cdot)$ denotes the Poisson distribution, $x_{1i} \sim \mathcal{N}\left(0,1\right)$ is the inferential predictor of interest to the data analyst and $x_{2i} \sim \mathcal{E}\left(1/2.5\right)$ (where $\mathcal{E}(\cdot)$ denotes the Exponential distribution) is the size variable, which is generated from a skewed distribution to reflect real-world survey data, particularly for employment counts. The expression, $h\{i\}$, denotes the group $h \in (1,\ldots,G_{U})$ linked to unit $i\in (1,\ldots,N)$. We generate $\displaystyle\mathop{\bm{\gamma}_{h}}^{2\times 1} \sim \mathcal{N}_{2}\left(\mathbf{0},\mbox{diag}(\bm{\sigma})\times \mathop{R}^{2\times 2} \times \mbox{diag}(\bm{\sigma})\right)$, where $\bm{\sigma} = (1.0,0.5)^{'}$. We set $R = \mathbb{I}_{2}$, where $\mathbb{I}_{2}$ denotes the identity matrix of size $2$. Finally, we set $\bm{\beta} = \left(\beta_{0}, \beta_{1},\beta_{2}\right)^{'} = (0, 1.0,0.5)$, where we choose the coefficient of $x_{2}$ to be lower than that for $x_{1}$ to be moderately informative, which is conservative.
The allocation of units, $i = 1,\ldots,N$ to groups, $h = 1,\ldots,G_{U}$ is performed by sorting the units, $i$, based on size variable, $x_{2}$. This allocation procedure constructs sized-based groups that accord well with survey designs that define groups as geographic regions, for convenience, where there is expected more homogeneity within groups than between groups.
The population size for each group, $N_{h}$, is fixed under direct sampling of groups; for example, $N_{h} = 4$ in the case of $G_{U} = 1250$, which produces $N = 5000$ units, so the number of population units per group is constructed as $(4, 10, 25, 50, 100)$ for population group sizes, $G_{U} = (1250,500,200,100,50)$, respectively. \begin{comment} In the case where we conduct an indirect sampling of groups by directly sampling units in a single-stage pps design, the number of population units per group, $N_{h}$, is set to randomly vary among the $G_{U}$ population groups using a log-normal distribution centered on the $(4, 10, 25, 50, 100)$ units per group used in the case of direct sampling, with a variance of $0.5$. In the case of $G_{U} = 1250$, this produces a right-skewed distribution of the number of units in each group, ranging from approximately $1$ to $40$ units per group and the total number of population units per group is restricted to sum to $N = 5000$.
We sort the groups such that groups with larger-sized units are assigned relatively fewer units and groups with smaller-sized units are assigned relatively more units. This set-up of assigning more units to smaller-sized groups mimics the estimation of employment counts among business establishments analyzed in our application in the sequel, where there are relatively few establishments with a large number of employees (e.g., $> 50$) (which is the size variable), while there are, by contrast, many more establishments (small businesses) that have a small number of employees (e.g., $< 10$). \end{comment}
Although the population response $y$ is generated with $\mu = f(x_1,x_2)$, we estimate the marginal model for the \emph{population}, $\mu = f(x_1)$ to which we will compare estimated results on samples taken from the population to assess bias and mean-squared error (MSE). We use $x_{2}$ in the generation of the population values for $y$ because the survey sampling inclusion probabilities are set proportionally to $x_{2}$, which instantiates the informativeness of the sampling design. In practice, however, the analyst does not have access to $x_{2}$ for the population units or, more generally, to all the information used to construct the survey sampling distribution that sets inclusion probabilities for all units in the population. The marginal estimation model under exclusion of size variable, $x_{2}$, is specified as:
\begin{align} \label{eq:estmodel} y_{i} &\sim \mathcal{P}\left(\mu_{i}\right) \nonumber\\ \log\mu_{i} &= \beta_{0} + x_{1i}\beta_{1} + u_{h\{i\}}\\ u_{h} & \sim \mathcal{N}\left(0,\sigma_{u}^{2}\right) \nonumber \end{align}
where now $u_{h}$ is an intercept random effect, $h = 1,\ldots,G_{U}$.
Our goal is to estimate the global parameters, $(\beta_{0},\beta_{1},\sigma_{u}^{2})$, from informative samples of size, $n = 500$, taken from the population (of size, $N = 5000$). We utilize the following simulation algorithm: \begin{enumerate}
\item Each Monte Carlo iteration of our simulator (that we run for $B = 300$ iterations) generates the population $(y_{i},x_{1i},x_{2i})_{i=1}^{N}$ from Equation~\ref{eq:margmodel} on which we estimate the marginal population model of Equation~\ref{eq:estmodel} to determine the population true values for $(\mu_{i},\sigma_{u}^{2})$. \item Our simulation study focuses on the direct sampling of groups, followed by a sub-sampling of units within the selected groups. We use a proportion-to-size design to directly sample from the $G_{U}$ groups in the first stage, where the group inclusion probabilities, $\pi_{h} \propto
\frac{1}{N_{h}}\mathop{\sum}_{i\in U_{h}}x_{2i}$.
We draw a sample of groups in the first stage and observe $G_{S} < G_{U}$ groups. In particular, fixed sample of total size $n = 500$ is taken where the number of groups sampled, $G_{S} = n/(Nf) \times G_{U}$.
\item The second stage size-based sampling of units is accomplished with inclusion probabilities, $\pi_{\ell\mid g} \propto x_{2\ell}$ for $\ell \in
(1,\ldots,N_{g})$.
We perform a further sub-sampling of $f\%$ of population units in the selected $G_{S}$ groups.
\item Estimation is performed for $(\beta_{0},\beta_{1},\sigma_{u}^{2})$ from the observed sample of $n = 500$ under three alternatives:
\begin{enumerate}
\item Single-weighting, where we solely exponentiate the likelihood contributions for $(y_{g j})$ by sampling weights, $(w_{g j} \propto 1/\pi_{g j})$ (and don't weight the prior for the random effects, $(u_{g})$);
\item Double-weighting, where we exponentiate \emph{both} the likelihood for the $(y_{g j})$ by sampling weights, $(w_{g j})$, and also exponentiate the prior distribution for $u_{g}$ by weight, $w_{g} \propto 1/\pi_{g}$ (for each of $g = 1,\ldots,G_{S}$). We compute the marginal unit weights used in both single- and double-weighting as $w_{g j} \propto 1/\pi_{g j}$, where $\pi_{g j}$ is the marginal inclusion probability, formulated as, $\pi_{g j} = \pi_{g}\pi_{j\mid g}$ for $ j = 1,\ldots,n_{g}$ for each group, $g \in 1,\ldots, G_{S}$ in the case of direct sampling of groups.
\item SRS, where in the case of direct sampling of groups, we take a simple random (equal probability) sample of groups in a first stage, followed by a simple random sample of units within selected groups.
We take the SRS sample from the same population as is used to take the two-stage, pps informative sample.
The inclusion of model estimation under (a non-informative) SRS is intended to serve as a gold standard against which we may judge the bias and MSE performance of single- and double-weighting under informative sampling.
\end{enumerate} \end{enumerate}
\begin{comment} Our second simulation study conducts an indirect sampling of groups by directly sampling units in a single stage design such that groups are included in the sample to the extent that at least one of their member units is sampled. As in the case of the direct sampling of groups, a proportion to size design is used with the marginal inclusion probability, $\pi_{i} \propto x_{2i},~i = 1,\ldots,N$. Group inclusion probabilities for double-weighting are composed using the sum-weights method as described in Section~\ref{sec:indirect}. Please see Appendix \ref{sec:simstudy} for description for comparative simulation results for other methods (in addition to sum-weights) for computing pseudo inclusion probabilities and associated weights for estimation. \end{comment}
We use Stan \citep{stan:2015} to estimate the double-weighted mixed effects model of Equation~\ref{eq:sampcompletemodel} for the specific case of the Poisson likelihood that we use in our simulations and application that next follows. We fully specify our Stan probability model for the Poisson likelihood under double-weighting in Appendix~\ref{sec:stanscript}. In particular, we specify a multivariate Gaussian joint prior distribution for the $K\times 1$, $\bm{\beta}$ coefficients with a vector of standard deviation parameters, $\sigma_{\beta}$ drawn from a truncated Cauchy prior. The associated correlation matrix for the multivariate Gaussian prior for $\bm{\beta}$ is drawn from a prior distribution that is uniform over the space of $K \times K$ correlation matrices. The prior for the standard deviation parameter of the random effects, $\sigma_{u}$, is also specified as a truncated Cauchy distribution. These prior distributions are designed as weakly informative by placing large probability mass over a large support region, while expressing a mode to promote regularity and a stable posterior geometry that is readily explored under Stan's Hamiltonian Monte Carlo scheme.
The single-weighting case is achieved as a special / simplified case of the double-weighting model.
Please see Appendix \ref{sec:simstudy} for results of a second simulation study under the \emph{indirect} sampling of groups.
\subsection{Informative Random Effects Under Direct Sampling of Groups} \begin{figure}
\caption{Each plot panel compares distributions of $\overline{(y-\exp(x\beta))}_{g}$ for each of a synthetic Population and a single sample taken from that population, faceted by a sequence for the number of population groups, $G_{U}$. The resulting violin plots present each distribution within $95\%$ quantiles.}
\label{fig:biasgrp}
\end{figure} \afterpage{\FloatBarrier} To make concrete the notion of informative random effects, we generate a single population and subsequently take a single, informative sample of groups from that population of groups under a proportion-to-size design, using the procedures for population generation and the direct sampling of groups, described above. The size for each population is $N = 5000$ and the sample size is $n = 500$. We, next, average the item responses, $y$ in each group after centering by removing the fixed effects observed in the sample (excluding $x_{2}$). For illustration, the computed $\left(\overline{(y-\exp(x\beta))}_{g}\right)$ will be used as a naive indicator of the distribution of the random effects, $\exp(u_{g})$. Each plot panel in Figure~\ref{fig:biasgrp} compares the distributions of this group-indexed centered mean statistic between the generated population and resulting informative sample for a population. A collection of plot panels for a sequence of populations with $G_{U} = (1250, 500, 100)$ number of population groups is presented, from left-to-right. Fixing a plot panel, each violin distribution plot includes horizontal lines for the $(0.25,0.5,0.75)$ quantiles. We see that under a proportion-to-size design that the distributions for the centered, group mean statistic in the sample are different from the underlying populations and skew larger than those for the populations. This upward skewness in each sample indicates that performing population estimation on the observed sample will induce bias for random effects variance, $\phi$, without correction of the group indexed random effects distribution in the sample, which we accomplish by weighting the distribution over random effects back to that for the population. \begin{comment} We begin our simulation study by assessing estimation bias for $\phi$, the generating hyperparameter of the random effects from Equation~\ref{eq:sampdirmodel}, that arises from informative random effects under the canonical case of unit-indexed random effects, where $G_{U} = N$ (the number of units in the population, $U$), such that the number of units per group is $N_{h} = 1$ for all groups, which reduces to setting $\pi_{i} \propto x_{2i}$, a single stage, proportion-to-size design, with size variable, $x_{2} \sim \mathcal{E}(1)$. We compare the use of single- and double-weighting utilizing bias and MSE statistics (based on the true population values from the marginal model) for the generating random effects variance, $\phi \equiv\sigma_{u}^{2}$ and fixed effects, $\left(\beta_{0}, \beta_{1}\right)$.
Our results show in Figure~\ref{fig:unitre} and associated Table~\ref{tab:unitre} reveal a strong bias under single-weighting for the generating random effects variance, $\sigma_{u}^{2}$, that is substantially removed under double-weighting. This is a consequential result because, as with fixed effects, we lack \emph{a priori} knowledge of the informativeness of the sampling design for any data set acquired under a complex survey sampling design, so this result suggests the necessity to exponentiate the prior contribution for each random effect, $u_{i}$. We also note a bias reduction in the intercept parameter, $\beta_{0}$, for double-weighting. Bias may arise in $\beta_{0}$ due to a location non-identifiability when the mean of the random effects, $(u_{i})$, are biased away from $0$.
\begin{figure}
\caption{Distributions and quantiles (25\%, 50\%, 75\%) of Posterior Means for random effects generating variance, $\sigma_{u}^{2}$ for Unit-Indexed Random Effects model. Double- and Single-weighting under informative sampling across B = \# simulations compared to simple random sampling (SRS).}
\label{fig:unitre}
\end{figure} \afterpage{\FloatBarrier}
\rowcolors{2}{gray!6}{white} \begin{table}[!h]
\caption{\label{tab:unitre}Bias and Mean Square Error of Posterior Means for Unit-Indexed Random Effects model. Double- and Single-weighting under informative sampling across B = 300 iterations compared to simple random sampling (SRS).} \centering
\begin{tabular}[t]{r|r|r|l|l} \hiderowcolors \toprule $\beta_{0}$ & $\beta_{1}$ & $\sigma_{u}^{2}$ & Statistic & Method\\ \midrule \showrowcolors 0.15 & 0.02 & 0.98 & bias & Single-weighting\\ 0.03 & 0.01 & 1.15 & MSE & Single-weighting\\ \hline 0.01 & -0.03 & -0.12 & bias & Double-weighting\\ 0.01 & 0.01 & 0.05 & MSE & Double-weighting\\ \hline -0.04 & -0.01 & 0.06 & bias & SRS\\ 0.01 & 0.01 & 0.13 & MSE & SRS\\ \bottomrule \end{tabular} \end{table} \rowcolors{2}{white}{white} \afterpage{
}
\end{comment}
\subsection{Varying Number of Population Groups, $G_{U}$}\label{sim:varygroups} We assess bias for a population model constructed using group-indexed random effects, where each group links to multiple units. Our results presented in Figure~\ref{fig:groupre} compare our double-weighting method to single-weighting in the case we conduct a proportion-to-size \emph{direct} sampling of groups and, subsequently, sub-sample $f = 50\%$ of member units within groups. We include an SRS sample of groups and units within selected groups taken from the same population. The results reveal that bias is most pronounced in the case of a relatively larger number of groups e.g., $G_{U}=(1250,500)$ for $N=5000$, where each group links relatively few units. By contrast, as the number of groups decreases, fixing the population size, $N$, the number of units linking to each group increases, which will have the effect of reducing variation in resulting sampling weights among the groups until, in the limit, there is a single group (with $\pi_{g} = 1$). The relative bias of single-weighting, therefore, declines as the number of groups declines (and units per group increases), such that residual bias in $\sigma_{u}^{2}$ for $G_{U}=100$ is dominated by increasing variability (because we sample fewer groups) for all methods. We, nevertheless, detect a small decrease in bias when we use double-weighting. We include Table~\ref{tab:groupre} that presents the bias in the estimation for the posterior mean values of $(\beta_{0},\beta_{1},\sigma_{u}^{2})$ that confirms the reduction in bias for $\sigma_{u}^{2}$ under double-weighting for $G_{U} = 100$. Our set-up may be viewed as more likely to induce bias because we assign units to groups by sorting units on the values of the size variable, $x_{2} \sim \mathcal{E}(1/(2.5))$ for allocation to groups. Our proportion-to-size sampling design selects groups based on the mean size variable for each group, $\bar{x}_{2}$. This set-up will tend to accentuate the variance in the resulting group-indexed size variable (and, hence, the resulting survey sample inclusion probabilities) as compared to a random allocation of units to groups. Our simulation set-up is, nevertheless, realistic because many surveys are characterized by relatively homogenous clusters; for example, the geographically-indexed metropolitan statistical areas (MSAs) (which may be viewed as clusters) used by the Current Employment Statistics survey (administered by the Bureau of Labor Statistics) tends to express larger (higher number of employees) establishments in more highly populated MSAs. \begin{figure}
\caption{Direct sampling of groups: Monte Carlo distributions and quantiles (0.5\%, 25\%, 50\%, 75\%, 99.5\%) for $B = 300$ iterations of differences between Posterior Means and truth under Single- and Double-weighting schema as compared to SRS for varying number of random effect groups, $G_{U}$, under $x_{2}\sim \mathcal{E}(1/2.5)$ for $N = 5000$ and $n=500$. Parameters $(\beta_{0},\beta_{1},\sigma_{u}^{2})$ along columns and number of population groups, $G_{U}$, in descending order along the rows.}
\label{fig:groupre}
\end{figure} \afterpage{
}
\rowcolors{2}{gray!6}{white} \begin{table}[!h]
\caption{\label{tab:groupre}Difference of Single- and Double-weighting with Increasing Number of Groups, $G$, under $x_{2}\sim \mathcal{E}(1/2.5)$} \centering
\begin{tabular}[t]{r|r|r|l|l|r} \hiderowcolors \toprule $\beta_{0}$ & $\beta_{1}$ & $\sigma_{u}^{2}$ & Statistic & Method\\ \midrule 0.03 & 0.03 & 0.25 & bias & Single-weighting & 1250\\ -0.06 & 0.03 & 0.06 & bias & Double-weighting & 1250\\ -0.05 & 0.01 & 0.02 & bias & sRS & 1250\\ \hline 0.15 & -0.13 & 0.19 & bias & Single-weighting & 500\\ 0.04 & -0.13 & 0.02 & bias & Double-weighting & 500\\ 0.04 & -0.12 & 0.04 & bias & sRS & 500\\ \hline 0.10 & -0.02 & 0.26 & bias & Single-weighting & 200\\ -0.01 & -0.02 & 0.09 & bias & Double-weighting & 200\\ -0.02 & 0.00 & 0.05 & bias & sRS & 200\\ \hline 0.09 & -0.09 & 0.31 & bias & Single-weighting & 100\\ -0.01 & -0.10 & 0.11 & bias & Double-weighting & 100\\ -0.03 & -0.10 & 0.07 & bias & sRS & 100\\ \hline -0.06 & -0.03 & 0.44 & bias & Single-weighting & 50\\ -0.14 & -0.03 & 0.29 & bias & Double-weighting & 50\\ -0.13 & -0.02 & 0.27 & bias & sRS & 50\\ \bottomrule \end{tabular} \end{table} \rowcolors{2}{white}{white} \afterpage{\FloatBarrier}
We next compare our double-weighting approach to the best available method in the literature, the pairwise composite likelihood method of \citet{yi:2016}, specified in Equation~\ref{composite} , which we refer to as ``pair-integrated''. We compare both methods in the case of relatively few units linked to each group (e.g., $G = 500, 1250$) because \citet{yi:2016} demonstrate superior bias removal properties as compared to \citet{rh:1998} in this setup. We exclude smaller values of $G$ because as the number of individuals within each group grows, the number of pairwise terms to include in the pair-integrated method grows quadratically. Our simulation set-up conducts a first-stage proportion to size sampling of groups in exactly the same manner as the previous simulation study. We additionally include an SRS of groups and, in turn, units within groups, as a benchmark. The custom R code to implement the ``pair-integrated" point estimation can be found in Appendix \ref{sec:pairwisecode}.
Figure \ref{fig:groupre_pair} presents the Monte Carlo distributions for parameter estimates, where the columns denote parameters, $(\beta_{0},\beta_{1},\sigma_{u}^{2})$, and the rows denote number of population groups, $G_{U}$. The results demonstrate that double-weighting leads to unbiased estimation of both the fixed effects parameters and the random effects variance relative to using a two-stage SRS sample. By contrast, the pair-integrated method demonstrates both bias and variability for the random effects variance, $\sigma_{u}^{2}$, which is exactly the set-up where it is hoped to perform relatively well. This bias for pair-integrated in the random effects variance also induces bias for the fixed effects intercept, $\beta_{0}$. As mentioned in Section~\ref{sec:direct} the pair-integrated method integrates out the random effects (from the unweighted prior distribution) \emph{before} applying the group weights, which fails to fully correct for the informative sampling of groups. Our method, by contrast, weights the prior the random effects and integrates them out numerically through our MCMC (after sampling the random effects). It bears mention that \cite{yi:2016} only evaluate informative sampling of units within groups, but not the informative sampling of the groups themselves, which may be why the estimation bias for $\sigma_{u}$ was not discovered.
\begin{figure}
\caption{Direct sampling of groups: Monte Carlo distributions and quantiles (0.5\%, 25\%, 50\%, 75\%, 99.5\%) for $B = 300$ iterations within a $99\%$ interval for difference of Posterior Means and truth under Double-weighting, Pair-integrated estimation and Simple Random Sampling (SRS). Parameters $(\beta_{0},\beta_{1},\sigma_{u}^{2})$ (columns); Varying number of random effects, $G_{U}$ (rows), under $x_{2}\sim \mathcal{E}(1/2.5)$, for a population of $N = 5000$.}
\label{fig:groupre_pair}
\end{figure} \afterpage{
}
\begin{comment}
\subsection{Sampling Units Rather than Groups (under Indirect Sampling of Groups)}\label{sim:indirectsamp} This next Monte Carlo simulation study implements case $3$ from Section~\ref{doublepseudo} that addresses sampling designs where the population group identifiers used to generate the response variable of interest are \emph{not} directly sampled by the survey sampling design. The synthetic population (for each Monte Carlo iteration) utilizes group-indexed random effects under size-based assignment of population units to groups under each alternative for total number of groups, $G_{U}$, as in Section~\ref{sim:varygroups}. In this study, however, we randomly vary the number of population units assigned to each group with the mean values for each $G_{U}$ set to be equal to the fixed number of units per group used in Section~\ref{sim:varygroups}. We allocate a relatively higher number of units to those groups with smaller-sized units under each group size, $G_{U}$, to mimic our application. The survey sampling design employed here is a \emph{single-stage}, proportion-to-size design that directly samples the units (not the groups) with unit inclusion probabilities proportional to the size variable, $x_{2} \sim \mathcal{E}(1/3.5)$.
Each plot panel in Figure~\ref{fig:inducegrpre} shows the distributions over Monte Carlo simulations for estimates of the generating variance, $\sigma_{u}^{2}$, of the random effects, $(u_{g})$, under each of the following weighting methods: single-weighting, sum-weights double-weighting (Equation~\ref{eq:sampindirmodel}), and SRS (no weighting under simple random sampling of the same population from which the pps sample was taken). The panels are ordered from left-to-right for a sequence of $G_{U} = \left(1250,500,200,100,50\right)$. As earlier mentioned, the number of population units per group, $N_{h}$, is set to randomly vary under a lognormal distribution, though there will on average be more units sampled per group from synthetic populations with a smaller number of population groups, $G_{U}$, than there will be units per group sampled under a larger number of population groups. The sum-weights method for accomplishing double-weighting generally performs better than single-weighting for all group sizes. When the number of population groups, $G_{U}$, is small, however, noise induced by sampling error results in double-weighting under-performing compared to SRS. Yet, as the number of units per group increases with $G_{U} = 500$, we see that sum-weights outperforms SRS, which is expected because the pps design is generally more efficient such that the contraction rate of the estimator on the truth will be faster for pps (occur at a lower sample size).
Table~\ref{tab:inducegrpre} presents the relative bias, defined as the bias divided by the true value, and the normalized root MSE, defined as the square root of MSE divided by the true value, for the regression coefficients, $(\beta_{0},\beta_{1})$, to accompany Figure~\ref{fig:inducegrpre}. We show the relative bias and normalized RMSE quantities in this study because the true values of the marginal model, $\sigma_{u}^{2} = \left(0.578,0.349,0.216,0.169,0.136\right)$, varies over the sequence of sizes for $G_{U}$. As in the case of direct sampling of groups, there is an association between the amount bias in estimation of $\sigma_{u}^{2}$ and in the intercept coefficient, $\beta_{0}$. \begin{figure}
\caption{Indirect sampling of groups: Monte Carlo distributions and quantiles (0.5\%, 25\%, 50\%, 75\%, 99.5\%) for $B = 300$ iterations for $\sigma_{u}^{2}$ for difference of posterior means and truth under alternative weighting schema for varying number of groups, $G$. Population of $N = 5000$ and sample of $n = 500$. In each plot panel, from left-to-right is the Single-weighting method, Sum-weights double-weighting method of Equation~\ref{eq:sampindirmodel} and Simple-random sampling (SRS).}
\label{fig:inducegrpre}
\end{figure} \afterpage{\FloatBarrier}
\rowcolors{2}{gray!6}{white} \begin{table}[!h]
\caption{\label{tab:inducegrpre}Normalized Bias and RMSE for Double-weighting as compared to Single-weighting and SRS for Increasing Units Per Random Effect Under \emph{Indirect} Sampling of Groups for $B = 300$ iterations with population of $N = 5000$ and sample of $n = 500$.} \centering \begin{adjustbox}{width=0.8\textwidth} \small
\begin{tabular}{l|r|r|r|r|r} \toprule \multicolumn{2}{c}{ } & \multicolumn{2}{c}{Relative Bias} & \multicolumn{2}{c}{Normalized RMSE} \\ \cmidrule(l{3pt}r{3pt}){3-4} \cmidrule(l{3pt}r{3pt}){5-6} Model & G & beta\_0 & beta\_1 & beta\_0 & beta\_1\\ \midrule Single-weighting & 1250 & -0.55 & 0.00 & 0.58 & 0.11\\ \rowcolor{gray!6} SRS & 1250 & -0.58 & 0.00 & 0.61 & 0.11\\ \rowcolor{gray!6} Sum-weights & 1250 & -0.52 & -0.01 & 0.55 & 0.11\\ \addlinespace \rowcolor{gray!6} Single-weighting & 500 & -0.45 & 0.01 & 0.48 & 0.14\\ SRS & 500 & -0.46 & 0.00 & 0.49 & 0.11\\ Sum-weights & 500 & -0.41 & 0.00 & 0.45 & 0.13\\ \addlinespace Single-weighting & 200 & -0.26 & 0.00 & 0.34 & 0.12\\ \rowcolor{gray!6} SRS & 200 & -0.29 & 0.00 & 0.36 & 0.11\\ \rowcolor{gray!6} Sum-weights & 200 & -0.23 & 0.00 & 0.33 & 0.12\\ \addlinespace \rowcolor{gray!6} Single-weighting & 100 & -0.13 & -0.02 & 0.30 & 0.13\\ SRS & 100 & -0.19 & 0.00 & 0.33 & 0.12\\ Sum-weights & 100 & -0.13 & -0.02 & 0.30 & 0.13\\ \addlinespace Single-weighting & 50 & -0.07 & 0.00 & 0.39 & 0.11\\ \rowcolor{gray!6} SRS & 50 & -0.08 & -0.01 & 0.39 & 0.11\\ \rowcolor{gray!6} Sum-weights & 50 & -0.07 & 0.00 & 0.39 & 0.11\\ \bottomrule \end{tabular} \end{adjustbox} \end{table} \rowcolors{2}{white}{white} \afterpage{\FloatBarrier}
We note that under $G_{U} = 50$ the results in Figure~\ref{fig:inducegrpre} show that the performance for the sum-weights double-weighting methods collapses onto those for SRS and single-weighting. This result is explained by the de facto inclusion of all $G_{U} = 50$ groups in \emph{every sample} across the Monte Carlo iterations. The inclusion of all population groups in the observed sample under $G_{U} = 50$ is not surprising as there are many units per group, so it is likely that at least one unit in \emph{every} group will be sampled in each Monte Carlo iteration. In the case that every sample contains a member of every group, there is no weighting of the random effects distributions needed and we see in Figure~\ref{fig:inducegrpre} that all of the double-weighting methods perform nearly the same as single-weighting and SRS.
\begin{figure}
\caption{Monte Carlo distribution for number of sampled groups, $G_{S}$, over $B=300$ iterations.}
\label{fig:numgrps}
\end{figure} \afterpage{\FloatBarrier} \end{comment}
\begin{comment} Lastly, it is often the case that while the data analyst may know aspects of the survey sampling design, such as the groups sampled in a multistage design, they do \emph{not} have access to marginal stagewise group inclusion probabilities or conditional last stage unit inclusion probabilities; rather, they may only have last stage marginal weights. So we next compare the performances under use of our \emph{induced} group weighting methods (Product-complement, Sum-probabilities, Sum-weights) based on pseudo group inclusion probabilities formulated when the groups are not directed sampled. In this case, however, the groups \emph{are} directly sampled and we label with ``Double-weighting", the usual method that employs both group-level weights and unit-level conditional weights as outlined in Theorem 1. Our goal is to determine whether our weighting approach formulated in the case of indirect sampling of groups is able to perform better than single-weighting when the groups are directly sampled, but stagewise sampling weights are not available. We follow the same simulation procedure as outlined in Section~\ref{sim:varygroups} to directly sample groups (and to take all units), but we generate our response using $x_{2}\sim \mathcal{E}(1/3.5)$. Figure~\ref{fig:sampgrpsindest} illustrates that, while the resulting double-weighting Sum-weights indirect estimator performs about as well as Double-weighting (formulated using group-indexed weights under direct sampling of groups), both of which handily outperform the other two indirect estimators (Product-complement, Sum-probabilities) and Single-weighting.
\begin{figure}
\caption{Monte Carlo distributions for $B = 300$ iterations within a $99\%$ interval for difference of Posterior Means and truth under direct sampling of groups to compare Single-weighting with two cases of Double-weighting: 1. Direct estimator (``Double-weighting"); 2. Indirect estimator (``Product-complement",``Sum-probabilities",``Sum-weights").}
\label{fig:sampgrpsindest}
\end{figure} \afterpage{\FloatBarrier}
\end{comment}
We briefly comment on the simulation study for the indirect sampling of of groups detailed in Appendix \ref{sec:simstudy}. The results accord with the direct sampling of groups where double-weighting outperforms single-weighting. When the number of population groups, $G_{U}$, is small, however, noise induced by sampling error results in double-weighting under-performing compared to SRS. Yet, as the number of units per group increases with $G_{U} = 500$, the sum-weights approach outperforms SRS, which is expected because the pps design is generally more efficient such that the contraction rate of the estimator on the truth will be faster for pps (occur at a lower sample size).
Lastly, we note that while we have focused on a simple Poisson random effects formulation, our survey-weighted pseudo Bayesian posterior method readily extends to any number of levels and simultaneous employment of multiple sets of random effects without any modification to the approach. Competitor methods, by contrast, are not readily estimable.
\section{Application}\label{application} We compare single- and double-weighting under a linear mixed effects model estimated on a dataset published by the Job Openings and Labor Turnover Survey (JOLTS), which is administered by BLS on a monthly basis to a randomly-selected sample from a frame composed of non-agricultural U.S. private (business) and public establishments. JOLTS focuses on the demand side of U.S. labor force dynamics and measures job hires, separations (e.g. quits, layoffs and discharges) and openings. We construct a univariate count data population estimation model with our response, $y$, defined to be the number of hires. We formulate the associated log mean with, \begin{equation} \log~\mu_{i} = \mathbf{x}_{i}^{'}\bm{\beta} + u_{g\{i\}}, \end{equation} where groups, $g = 1,\ldots,(G=892)$, denote industry groupings (defined as $6-$ digit North American Industry Classification (NAICS) codes) that collect the participating business establishments. We expect a within-industry dependence among the hiring levels for business establishments since there are common, industry-driven economic factors that impact member establishments. We construct the fixed effects predictors, $\mathbf{x} = \left[1,\text{ownership status},\text{region}\right]$, which are categorical predictors where ownership status holds four levels, $1.$ Private; $2.$ Federal government; $3.$ State government; $4.$ Local government. The region predictor holds four levels, $1.$ Northeast; $2.$ South; $3.$ Midwest; $4.$ West. Private and Northeast are designated as the reference levels.
The JOLTS sampling design assigns inclusion probabilities (under sampling without replacement) to establishments to be proportional to the number of employees for each establishment (as obtained from the Quarterly Census of Employment and Wages (QCEW)). This design is informative in that the number of employees for an establishment will generally be correlated with the number of hires, separations and openings. We perform our modeling analysis on a May, $2012$ data set of $n = 9743$ responding establishments. We \emph{a priori} expect the random effects, $(u_{g})$, to be informative since larger-sized establishments would be expected to express larger variances in their hiring levels. We choose the sum-weights method for inducing industry-level weights (from Equation~\ref{eq:sampindirmodel}) to construct our double-weighted estimation model on the observed sample. \begin{figure}
\caption{Distributions and quantiles (25\%, 50\%, 75\%) of estimated posterior mean values for random effects, $u_{g},~ g = 1,\ldots,(G=892)$, for the JOLTS application under Single- and Double-weighting.}
\label{fig:joltsu}
\end{figure} \afterpage{\FloatBarrier}
\begin{figure}
\caption{Distributions and quantiles (25\%, 50\%, 75\%) of posterior samples for $\sigma_{u}^{2}$, the generating variance for random effects, and a single random effect parameter, $u_{i}$, for the JOLTS application, under Single- and Double-weighting.}
\label{fig:joltsig}
\end{figure} \afterpage{\FloatBarrier}
The more diffuse distribution over the $G = 892$ posterior mean values for random effects, $(u_{g})$, under double-weighting than single-weighting shown in Figure~\ref{fig:joltsu} demonstrates that co-weighting the likelihood and random effects distribution produces notably different inference for the group-indexed random effects; in particular, the observed sample is more homogenous in the number of hires by setting inclusion probabilities to concentrate or over-sample large-sized establishments relative to the more size-diverse population of establishments. So the weighting of the random effects distributions in the observed sample produces a distribution over the posterior mean values for the random effects that better reflects the size-diversity of establishments in the underlying population from which the sample was taken. Figure~\ref{fig:joltsig} presents the estimated pseudo-posterior distributions for the generating random effects variance, $\sigma_{u}^{2}$ and also a single random effect parameter, $u_{i}$, under both single- and double-weighting. This figure reinforces the observed result for the random effects where the observed hiring levels in the survey data are more homogenous than those in the underlying population, which induces a larger posterior variation in the estimated random effects parameters for double-weighting.
\section{Discussion}\label{discussion}
In this work, we demonstrate the existence of biased estimation of both fixed \emph{and} random effects parameters when performing inference on data generated from a complex survey sample. This risk is largely unrecognized in the Bayesian literature. The current remedies come from the survey literature and are motivated from a frequentist perspective. They provide an incomplete and somewhat ad-hoc approach to the solution. We present a principled development of the ``double-weighting'' approach based on the joint distribution of the population generating model of inferential interest and the complex sampling design represented by sample inclusion indicators. We exploit the latent variable formulation of mixed models and their related posterior sampling techniques to avoid awkward numerical integration required for frequentist solutions. We show that this simplicity also leads to reductions in bias.
\begin{comment}
Through Monte Carlo simulations, we demonstrate the effectiveness of double-weighting for a variety of modeling and sampling situations such as unit-indexed random effects and group-indexed random effects models. These scenarios were chosen because they correspond to common situations in practice, as demonstrated by the JOLTS example. However, it should be clear that this approach also applies to more complex hierarchical models in general. For example, we could consider more levels of sampling and modeling and utilize more complex distributions for random effects, such as non-parametric process mixtures. In such cases, we might replace the term ``double-weighting'' with ``hierarchical-weighting''. \end{comment}
This work culminates recent developmental work in combining traditional survey estimation approaches with the Bayesian modeling paradigm. The pseudo-posterior framework simultaneously offers complex survey data analysis to Bayesian modelers and the full suite of hierarchical Bayesian methods to those well-versed in traditional fixed effect analysis of survey data.
\appendix \section{Alternative Pseudo Group Inclusion Probabilities Under Indirect Sampling}\label{sec:pseudoprobs} If we invert the resultant group-indexed weight, $w_{g}= 1/N_{g}\times\mathop{\sum}_{j \in S_{g}}w_{g j}$, from Equation~\ref{eq:sampindirmodel}, where groups are not directly sampled, we may view the inverse of the group $g$ weight, $\tilde{\pi}_{g} = 1/w_{g}$, as a ``pseudo" group inclusion probability, since we don't directly sample groups. The construction for one form of $\tilde{\pi}_{g}$ motivates our consideration of other formulations for the pseudo group inclusion probabilities that we may, in turn, invert to formulate alternative group-indexed survey sampling weights, $(w_{g})$.
The resulting $w_{g}$ of Equation~\ref{eq:sampindirmodel} requires either knowledge of $N_{g}$ or a method for its approximation. The sum of nested unit weights is further composed as a harmonic sum of inverse inclusion probabilities of member units in each group, which may be overly dominated by units with small unit inclusion probabilities. Our first alternative more directly constructs a pseudo group inclusion probability as the union of probabilities for inclusion of any member unit in the observed sample (in which case the group will be represented in the sample) and does not require estimation of population quantities, such as $N_{g}$. Under a weak assumption of nearly independent sampling within groups, this alternative is constructed as, \begin{align} \tilde{\pi}_{g} &= \mathop{\sum}_{\ell = 1}^{N_{g}}\pi_{\ell}\label{eq:sprpop}\\ \hat{\tilde{\pi}}_{g} &= \mathop{\sum}_{\ell = 1}^{N_{g}}\frac{\delta_{\ell}}{\pi_{\ell}} \times \pi_{\ell}\\
&= \mathop{\sum}_{j=1}^{n_{g}} w_{j} \times \pi_{j}\label{eq:sprsamp} \end{align} where $\pi_{\ell}$ denotes the marginal inclusion probability for unit, $\ell \in (1,\ldots,N_{g})$, where we recall that $N_{g}$ denotes the number of units linked to group, $g \in(1,\ldots,G_{U})$, in the population of groups. We may estimate the pseudo group inclusion probabilities in the observed sample by making the same walk from population-to-observed-sample as is done in Equation~\ref{estimator} to Equation~\ref{estimatorsample}; by including unit sampling weights, $(w_{j})_{j \in S_{g}}$ ($S_{g} = \{1,\ldots,n_{g}\}$). We normalize the $(w_{j})_{j \in S_{g}}$ to sum to $1$ as our intent is to re-balance the information (among sampled units) within a group to approximate that of the population of units within the group. While this estimator has the undesirable property of computing $\tilde{\pi}_{g} > 1$, we utilize this quantity to weight the random effects prior density contributions with, $w_{g} \propto 1/\tilde{\pi}_{g}$, so we focus on the effectiveness of estimation bias removal for generating hyperparameters of the $(u_{g})_{g \in G_{U}}$. We label this method as the ``sum-probabilities" method in contrast to the ``sum-weights" methods with which we label the result of Equation~\ref{eq:sampindirmodel}.
\begin{comment} Under a single stage sampling design where the groups are indirectly sampled through inclusions of nested units into the observed sample, we assume that we only have availability of the marginal unit inclusion sampling weights, $(w_{g j})$. The group population size, $N_{g}$, needed for the sum-weights method of Equation~\ref{eq:sampindirmodel}, may be estimated by $\hat{N}_{g} = \mathop{\sum}_{j = 1}^{n_{g}}w_{j\mid g}$. To approximate $w_{j\mid g}$, we first utilize the sum-probabilities result to estimate, $\hat{w}_{g} = 1/\hat{\tilde{\pi}}_{g}$, and proceed to extract $(w_{j\mid g})$ from $w_{gj} \approx w_{g}w_{j\mid g}$. \end{comment}
Our second alternative for estimation of a pseudo group inclusion probability is designed to ensure $\tilde{\pi}_{g} \leq 1$ by using a product complement approach that computes the union of member unit probabilities for a group, indirectly, by first computing its complement and subtracting that from $1$. To construct this estimator, we assume that units, $j \in S$ are sampled independently \emph{with} replacement, which is a tenable assumption when drawing a small sample from a large population of units. Let $\pi^{(1)}_{j}$ denote the probability of selecting unit, $j$, in a sample of size $1$ (e.g., a single draw). Then we may construct the marginal inclusion probability of unit, $\pi_{j}$, for a sample of size, $n = \vert S\vert$, as the complement that unit $j$ does not appear in any of the $n$ draws, \begin{equation}\label{eq:pwr} \pi_{j} = 1- \left(1-\pi^{(1)}_{j}\right)^{n}, \end{equation} where $\mathop{\sum}_{j\in U}\pi^{(1)}_{j} = 1$. By extension, $0 < \tilde{\pi}^{(1)}_{g} = \mathop{\sum}_{j \in U_{g}}\pi^{(1)}_{j} \leq 1$, where $\tilde{\pi}^{(1)}_{g}$ denotes the pseudo group, $g \in (1,\ldots,G_{U})$ inclusion probability for a sample of size $1$ and is composed as the union of size $1$ probabilities for member units. The expression for the pseudo group inclusion probability derives from the underlying sampling of members with replacement, \begin{equation}\label{eq:gpwr} \tilde{\pi}_{g} = 1 - \left(1-\tilde{\pi}^{(1)}_{g}\right)^{n} = 1 - \left(1-\mathop{\sum}_{j=1}^{N_{g}}\pi^{(1)}_{j}\right)^{n}, \end{equation} where we exponentiate the complement term, $\left(1-\tilde{\pi}^{(1)}_{g}\right)$, by the number of draws of units, $n$, (rather than $G_{S}$, the number of groups represented in the observed sample) because we don't directly sample groups. We solve for $\pi^{(1)}_{j}$ using Equation~\ref{eq:pwr}, $\pi^{(1)}_{j} = 1 - \left(1-\pi_{j}\right)^{(1/n)}$, and plug into Equation~\ref{eq:gpwr} to achieve, \begin{align} \tilde{\pi}_{g} &= 1 - \left(1-\mathop{\sum}_{j=1}^{N_{g}}\left(1-\left(1-\pi_{j}\right)^{(1/n)}\right)\right)^{n}\\ \hat{\tilde{\pi}}_{g} &= 1 - \left(1-\mathop{\sum}_{j=1}^{N_{g}}\frac{\delta_{j}}{\pi_{j}}\left(1-\left(1-\pi_{j}\right)^{(1/n)}\right)\right)^{n}\\ &= 1 - \left(1-\mathop{\sum}_{\ell=1}^{n_{g}}w_{\ell}\left(1-\left(1-\pi_{\ell}\right)^{(1/n)}\right)\right)^{n}\label{eq:pcprsamp}, \end{align} where, as with the sum-probabilities formulation, we normalize the unit weights within each group, $(w_{\ell})_{\ell \in S_{g}}$, to sum to $1$. We label this method as ``product-complement". \begin{comment} Where did we conclude on normalizing these weights vs. not. For some reason I was thinking PC did better with normalizing but sum prod did better without? \end{comment}
\section{Simulation Study Results for Alternative Pseudo Group Inclusion Probabilities}\label{sec:simstudy} We present the results for the simulation study that samples units, rather than groups, for the expanded set of methods developed in Appendix \ref{sec:pseudoprobs} for the pseudo group inclusion probabilities. We recall that under this single stage sampling of units, groups are not directly sampled under the survey sampling and are included to the extent that one or more member units are sampled.
The synthetic population (for each Monte Carlo iteration) utilizes group-indexed random effects under size-based assignment of population units to groups under each alternative for total number of groups, $G_{U}$. In this study, we randomly vary the number of population units assigned to each group with the mean values for each $G_{U}$ set to be equal to the fixed number of units per group. We allocate a relatively higher number of units to those groups with smaller-sized units under each group size, $G_{U}$, to mimic our application. The number of population units per group, $N_{h}$, is set to randomly vary among the $G_{U}$ population groups using a log-normal distribution centered on the $(4, 10, 25, 50, 100)$ units per group used in the case of direct sampling, with a variance of $0.5$. In the case of $G_{U} = 1250$, this produces a right-skewed distribution of the number of units in each group, ranging from approximately $1$ to $40$ units per group and the total number of population units per group is restricted to sum to $N = 5000$.
We sort the groups such that groups with larger-sized units are assigned relatively fewer units and groups with smaller-sized units are assigned relatively more units. This set-up of assigning more units to smaller-sized groups mimics the estimation of employment counts among business establishments analyzed in our application in the sequel, where there are relatively few establishments with a large number of employees (e.g., $> 50$) (which is the size variable), while there are, by contrast, many more establishments (small businesses) that have a small number of employees (e.g., $< 10$).
The survey sampling design employed here is a \emph{single-stage}, proportion-to-size design that directly samples the units (not the groups) with unit inclusion probabilities proportional to the size variable, $x_{2} \sim \mathcal{E}(1/(m_{2}=3.5))$. Each Monte Carlo iteration of our simulator (that we run for $B = 300$ iterations) generates the population $(y_{i},x_{1i},x_{2i})_{i=1}^{N}$, assigns group and unit inclusion probabilities for the population in the case of direct sampling of groups or assigns unit inclusion probabilities in the case of indirect sampling. A sample of $n = 500$ is then taken and estimation is performed for $(\beta_{0},\beta_{1},\sigma_{u}^{2})$ from the observed sample under three alternatives: \begin{enumerate} \item Single-weighting, where we solely exponentiate the likelihood contributions for $(y_{g j})$ by sampling weights, $(w_{g j} \propto 1/\pi_{g j})$ (and don't weight the prior for the random effects, $(u_{g})$); \item Double-weighting, where we exponentiate \emph{both} the likelihood for the $(y_{g j})$ by sampling weights, $(w_{g j})$, and also exponentiate the prior distribution for $u_{g}$ by weight, $w_{g} \propto 1/\pi_{g}$ (for each of $g = 1,\ldots,G_{S}$). We estimate $\pi_{g}$ using each of the three methods presented in Appendix \ref{sec:pseudoprobs}: ``sum-weights'', ``sum-probabilities'', and ``product-complement''. \item SRS, where we take a single-stage simple random sample of units. \end{enumerate} The inclusion of model estimation under (a non-informative) SRS is intended to serve as a gold standard against which we may judge the bias and MSE performance of single- and double-weighting under informative sampling.
Each plot panel in Figure~\ref{fig:inducegrprealts} shows the distributions over Monte Carlo simulations for estimates of the generating variance, $\sigma_{u}^{2}$, of the random effects, $(u_{g})$, under each of the following weighting methods: single-weighting, product-complement double-weighting (Equation~\ref{eq:pcprsamp}), sum-probabilities double-weighting (Equation~\ref{eq:sprsamp}), sum-weights double-weighting (Equation~\ref{eq:sampindirmodel}), and SRS (no weighting under simple random sampling of the same population from which the pps sample was taken). The panels are ordered from left-to-right for a sequence of $G_{U} = \left(1250,500,200,100,50\right)$. As earlier mentioned, the number of population units per group, $N_{h}$, is set to randomly vary under a lognormal distribution, though there will on average be more units sampled per group from synthetic populations with a smaller number of population groups, $G_{U}$, than there will be units per group sampled under a larger number of population groups. The sum-probabilities and sum-weights methods for accomplishing double-weighting generally perform nearly identically to one another and better than single-weighting for all group sizes. Since sum-probabilities and sum-weights perform nearly identically, one may choose to prefer use of the former because it does not require our estimation of $\hat{N}_{g}$, as does the latter.
Table~\ref{tab:inducegrprealts} presents the relative bias, defined as the bias divided by the true value, and the normalized root MSE, defined as the square root of MSE divided by the true value, for the regression coefficients, $(\beta_{0},\beta_{1})$, to accompany Figure~\ref{fig:inducegrprealts}. We show the relative bias and normalized RMSE quantities in this study because the true values of the marginal model, $\sigma_{u}^{2} = \left(0.578,0.349,0.216,0.169,0.136\right)$, varies over the sequence of sizes for $G_{U}$. As in the case of direct sampling of groups, there is an association between the amount bias in estimation of $\sigma_{u}^{2}$ and in the intercept coefficient, $\beta_{0}$. \begin{figure}
\caption{Indirect sampling of groups: Monte Carlo distributions and quantiles (0.5\%, 25\%, 50\%, 75\%, 99.5\%) for $B = 300$ iterations for $\sigma_{u}^{2}$ for difference of posterior means and truth under alternative weighting schema for varying number of groups, $G$. Population of $N = 5000$ and sample of $n = 500$. In each plot panel, from left-to-right is the Single-weighting method, Product Complement double-weighting method of Equation~\ref{eq:pcprsamp}, Sum-probabilities double-weighting method of Equation~\ref{eq:sprsamp}, Sum-weights double-weighting method of Equation~\ref{eq:sampindirmodel} and Simple-random sampling (SRS).}
\label{fig:inducegrprealts}
\end{figure} \afterpage{\FloatBarrier}
\rowcolors{2}{gray!6}{white} \begin{table}[!h] \caption{\label{tab:inducegrprealts} Normalized Bias and RMSE for Double-weighting methods as compared to Single-weighting and SRS for Increasing Units Per Random Effect Under \emph{Indirect} Sampling of Groups for $B = 300$ iterations with population of $N = 5000$ and sample of $n = 500$.} \centering \begin{adjustbox}{width=0.6\textwidth}
\begin{tabular}{l|r|r|r|r|r} \toprule \multicolumn{2}{c}{ } & \multicolumn{2}{c}{Relative Bias} & \multicolumn{2}{c}{Normalized RMSE} \\ \cmidrule(l{3pt}r{3pt}){3-4} \cmidrule(l{3pt}r{3pt}){5-6} Model & G & beta\_0 & beta\_1 & beta\_0 & beta\_1\\ \midrule \rowcolor{gray!6} Product-complement & 1250 & -0.54 & -0.01 & 0.58 & 0.11\\ Single-weighting & 1250 & -0.55 & 0.00 & 0.58 & 0.11\\ \rowcolor{gray!6} SRS & 1250 & -0.58 & 0.00 & 0.61 & 0.11\\ Sum-probabilities & 1250 & -0.52 & -0.01 & 0.55 & 0.11\\ \rowcolor{gray!6} Sum-weights & 1250 & -0.52 & -0.01 & 0.55 & 0.11\\ \addlinespace Product-complement & 500 & -0.45 & 0.01 & 0.48 & 0.13\\ \rowcolor{gray!6} Single-weighting & 500 & -0.45 & 0.01 & 0.48 & 0.14\\ SRS & 500 & -0.46 & 0.00 & 0.49 & 0.11\\ \rowcolor{gray!6} Sum-probabilities & 500 & -0.41 & 0.00 & 0.45 & 0.13\\ Sum-weights & 500 & -0.41 & 0.00 & 0.45 & 0.13\\ \addlinespace \rowcolor{gray!6} Product-complement & 200 & -0.26 & 0.00 & 0.34 & 0.12\\ Single-weighting & 200 & -0.26 & 0.00 & 0.34 & 0.12\\ \rowcolor{gray!6} SRS & 200 & -0.29 & 0.00 & 0.36 & 0.11\\ Sum-probabilities & 200 & -0.23 & 0.00 & 0.33 & 0.12\\ \rowcolor{gray!6} Sum-weights & 200 & -0.23 & 0.00 & 0.33 & 0.12\\ \addlinespace Product-complement & 100 & -0.13 & -0.02 & 0.30 & 0.13\\ \rowcolor{gray!6} Single-weighting & 100 & -0.13 & -0.02 & 0.30 & 0.13\\ SRS & 100 & -0.19 & 0.00 & 0.33 & 0.12\\ \rowcolor{gray!6} Sum-probabilities & 100 & -0.13 & -0.02 & 0.30 & 0.13\\ Sum-weights & 100 & -0.13 & -0.02 & 0.30 & 0.13\\ \addlinespace \rowcolor{gray!6} Product-complement & 50 & -0.07 & 0.00 & 0.39 & 0.11\\ Single-weighting & 50 & -0.07 & 0.00 & 0.39 & 0.11\\ \rowcolor{gray!6} SRS & 50 & -0.08 & -0.01 & 0.39 & 0.11\\ Sum-probabilities & 50 & -0.07 & 0.00 & 0.39 & 0.11\\ \rowcolor{gray!6} Sum-weights & 50 & -0.07 & 0.00 & 0.39 & 0.11\\ \bottomrule \end{tabular} \end{adjustbox} \end{table} \rowcolors{2}{white}{white} \afterpage{\FloatBarrier}
\section{Stan Code for Estimating Poisson Mixed Effects Model }\label{sec:stanscript}
We next present the Stan \citep{stan:2015} script that enumerates the probability model specification of the Poisson likelihood and associated prior distributions. We utilized Stan to estimate the mixed effects Poisson model implemented in Sections~\ref{simulation} and \ref{application}.
{\small \begin{lstlisting}[language=R] functions{
real wt_pois_lpmf(int[] y, vector mu, vector weights, int n){
real check_term;
check_term = 0.0;
for( i in 1:n )
{
check_term = check_term + weights[i] * poisson_log_lpmf(y[i] | mu[i]);
}
return check_term;
}
real normalwt_lpdf(vector y, vector mu, real sigma, vector weights, int n)
{
real check_term;
check_term = 0.0;
for( i in 1:n )
{
check_term = check_term + weights[i] * normal_lpdf(y[i] | mu[i], sigma);
}
return check_term;
}
} /* end function{} block */
data {
int<lower=1> n; // number of observations
int<lower=1> K; // number of linear predictors
int<lower=1> G; // number of random effects groups
int<lower=0> s_re[n]; // assignment of groups to items
int<lower=0> y[n]; // Response variable
vector<lower=0>[n] weights; // observation-indexed (sampling) weights
vector<lower=0>[G] weights_re; // group-indexed (sampling) weights
matrix[n, K] X; // coefficient matrix }
transformed data{
vector<lower=0>[G] zeros_g;
vector<lower=0>[K] zeros_beta;
zeros_beta = rep_vector(0,K);
zeros_g = rep_vector(0,G); } /* end transformed data block */
parameters{
vector[K] beta; /* regression coefficients from linear predictor */
vector<lower=0>[K] sigma_beta;
/* cholesky of correlation matrix for Sigma_beta */
cholesky_factor_corr[K] L_beta;
vector[G] u;
real<lower=0> sigma_u; }
transformed parameters{
real<lower=0> sigma_u2;
vector[n] mu;
vector[n] fixed_effects;
fixed_effects = X * beta;
mu = fixed_effects + u[s_re];
sigma_u2 = pow(sigma_u,2); } /* end transformed parameters block */
model{
L_beta ~ lkj_corr_cholesky(6);
sigma_u ~ student_t(3,0,1);
sigma_beta ~ student_t(3,0,1);
/* Implement a beta ~ N_{T}(0,Q^{-1}) */
/* K x 1, vector */
beta ~ multi_normal_cholesky( zeros_beta, diag_pre_multiply(sigma_beta,L_beta) );
/* directly update the log-probability for sampling */
u ~ normalwt(zeros_g, sigma_u, weights_re, G); // weighting the random effects
target += wt_pois_lpmf(y | mu, weights, n); } /* end model{} block */
\end{lstlisting} }
\section{R Code for Pairwise Integrated Poisson Mixed Effects Model }\label{sec:pairwisecode}
We next present the R \citep{R} scripts to implement the pairwise integrated likelihood approach of \citep{yi:2016} for Poison models. We utilized this model in Sections~\ref{simulation}.
{\small \begin{lstlisting}[language=R] library(fastGHQuad)
##Poison model integrated likelihood
#joint density - does not include exp(-u^2) which is included in ghQuad
#change of variables z = sqrt(2)*u*sig.z,
#where z ~N(0, sig.z) random effect
pwpois.GH <- function(u,x1,x2, mu1, mu2, sig.z){
#first argument is what is integrated out
d <- dpois(x1, mu1*exp(sqrt(2)*u*sig.z))*
dpois(x2, mu2*exp(sqrt(2)*u*sig.z))*1/sqrt(pi)
return(d) }
#see qhQuad help for more details intpois.ghfs <- function(x1, x2, mu1, mu2, sig.z){
#scalar arguments for each
intz <- ghQuad(pwpois.GH, rule = gaussHermiteData(5),
x1 = x1, x2 = x2, mu1 = mu1, mu2 = mu2, sig.z = sig.z)
return(intz) }
##Weighted Logliklihood
#assume y1 and y2 are vectors of pairs
#with corresponding matrix predictors X1 and X2 and vector
#of pairwise weights pwt
wt_comp_pois_ll <- function(par, y1,y2, X1,X2, pwt){
#need vector of parameters par as input for optim
k <- length(par)
beta <- par[1:(k-1)]
#have log(sig) as parameter to keep bounds -Inf to Inf in optim
sig <- exp(par[k])
mu1 <- exp(X1
mu2 <- exp(X2
ll <- 0
for(i in seq_along(y1)){
#print(i)
inti <- intpois.ghfs(y1[i],y2[i],mu1[i], mu2[i], sig)
ll <- ll + pwt[i]*max(log(inti), -743) #truncate to keep above -INF
#print(ll)
}
return(-ll) #need to minimize -LL }
## Input into optim
#Example with 3 parameters (Beta1, Beta2, Sigma_u)
## starting values testpar <- c(rnorm(1,0,.5), rnorm(1,0,.5), log(rgamma(1,3,rate=3))) est1 <- optim(par = testpar, fn = wt_comp_pois_ll, method = "BFGS",
y1 = y1, y2=y2, X1= X1, X2= X2, pwt = pwt)
## extract parameter estimates pars <- est1$par pars[3] <- exp(pars[3]) #back-transform to sigma in [0,Inf] beta_hat <- pars[1:2] sigmau2_hat <- pars[3]^2 \end{lstlisting} }
\end{document}
|
arXiv
|
{
"id": "1904.07680.tex",
"language_detection_score": 0.8068677186965942,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[]{Edgeworth expansions for independent bounded integer valued random variables.}
\vskip 0.1cm \author{Dmitry Dolgopyat and Yeor Hafouta}
\dedicatory{ }
\maketitle
\begin{abstract} We obtain asymptotic expansions for local probabilities of partial sums for uniformly bounded independent but not necessarily identically distributed integer-valued random variables. The expansions involve products of polynomials and trigonometric polynomials. Our results do not require any additional assumptions. As an application of our expansions we find necessary and sufficient conditions for the classical Edgeworth expansion. It turns out that there are two possible obstructions for the validity of the Edgeworth expansion of order $r$. First, the distance between the distribution of the underlying partial sums modulo some $h\in {\mathbb N}$ and the uniform distribution could fail to be $o(\sigma_N^{1-r})$, where $\sigma_N$ is the standard deviation of the partial sum. Second, this distribution could have the required closeness but this closeness is unstable, in the sense that it could be destroyed by removing finitely many terms. In the first case, the expansion of order $r$ fails. In the second case it may or may not hold depending on the behavior of the derivatives of the characteristic functions of the summands whose removal causes the break-up of the uniform distribution. We also show that a quantitative version of the classical Prokhorov condition (for the strong local central limit theorem) is sufficient for Edgeworth expansions, and moreover this condition is, in some sense, optimal. \end{abstract}
\tableofcontents
\section{Introduction.}\label{SecInt} Let $X_1,X_2,...$ be a uniformly bounded sequence of independent integer-valued random variables. Set $S_N=X_1+X_2+...+X_N$, $V_N=V(S_N)=\text{Var}(S_N)$ and ${\sigma}_N=\sqrt{V_N}$. Assume also that $V_N\to\infty$ as $N\to\infty$. Then the central limit theorem (CLT) holds true, namely the distribution of $(S_N-{\mathbb E}(S_N))/{\sigma}_N$ converges to the standard normal distribution as $N\to\infty$.
Recall that the local central limit theorem (LLT) states that, uniformly in $k$ we have \[ {\mathbb P}(S_N=k)=\frac1{\sqrt{2\pi}{\sigma}_N}e^{-\left(k-{\mathbb E}(S_N)\right)^2/2V_N}+o({\sigma}_N^{-1}). \] This theorem is also a classical result, and it has origins in De Moivre-Laplace theorem. The stable local central limit theorem (SLLT) states that the LLT holds true for any integer-valued square integrable independent sequence $X_1',X_2',...$ which differs from $X_1,X_2,...$ by a finite number of elements. We recall a classical result due to Prokhorov.
\begin{theorem} \cite{Prok} \label{ThProkhorov} The SLLT holds iff for each integer $h>1$, \begin{equation}\label{Prokhorov} \sum_n {\mathbb P}(X_n\neq m_n \text{ mod } h)=\infty \end{equation} where $m_n=m_n(h)$ is the most likely residue of $X_n$ modulo $h$. \end{theorem} We refer the readers' to \cite{Rozanov, VMT} for extensions of this result to the case when $X_n$'s are not necessarily bounded (for instance, the result holds true when
$\displaystyle \sup_n\|X_n\|_{L^3}<\infty$). Related results for local convergence to more general limit laws are discussed in \cite{D-MD, MS74}.
The above result provides a necessary and sufficient condition for the SLLT. It turns out that the difference between LLT and SLLT is not that big.
\begin{proposition} \label{PrLLT-SLLT} Suppose $S_N$ obeys LLT. Then for each integer $h\geq2$ at least one of the following conditions occur:
either (a) $\displaystyle \sum_n {\mathbb P}(X_n\neq m_n(h) \text{ mod } h)=\infty$.
or (b) $\exists j_1, j_2, \dots, j_k$ with $k<h$ such that $\displaystyle \sum_{s=1}^k X_{j_s}$ mod $h$ is uniformly distributed. In that case for all $N\geq \max(j_1, \dots, j_k)$ we have that $S_N$ mod $h$ is uniformly distributed. \end{proposition}
Since we could not find this result in the literature we include the proof in Section \ref{FirstOrder}.
Next, we provide necessary and sufficient conditions for the regular LLT.
We need an additional notation. Let $\displaystyle K=\sup_n\|X_n\|_{L^\infty}$. Call $t$ \textit{resonant} if $t=\frac{2\pi l}{m}$ with $0<m\leq 2K$ and $0\leq l<m.$
\begin{theorem} \label{ThLLT} The following conditions are equivalent:
(a) $S_N$ satisfies LLT;
(b) For each $\xi\in {\mathbb R}\setminus {\mathbb Z}$, $\displaystyle \lim_{N\to\infty} {\mathbb E}\left(e^{2\pi i \xi S_N}\right)=0$;
(c) For each non-zero resonant point $\xi$, $\displaystyle \lim_{N\to\infty} {\mathbb E}\left(e^{2\pi i \xi S_N}\right)=0$;
(d) For each integer $h$ the distribution of $S_N$ mod $h$ converges to uniform. \end{theorem}
The proof of this result is also given in Section \ref{FirstOrder}. We refer the readers to \cite{Do, DS} for related results in more general settings.
The local limit theorem deals with approximation of $P(S_N=k)$ up to an error term of order $o({\sigma}_N^{-1})$. Given $r\geq1$, the Edgeworth expansion of order $r$ holds true if there are polynomials $P_{b, N}$, whose coefficients are uniformly bounded in $N$ and their degrees do no depend on $N,$ so that
uniformly in $k\in{\mathbb Z}$ we have that \begin{equation}\label{EdgeDef} {\mathbb P}(S_N=k)=\sum_{b=1}^r \frac{P_{b, N} (k_N)}{\sigma_N^b}\mathfrak{g}(k_N)+o(\sigma_N^{-r}) \end{equation} where $k_N=\left(k-{\mathbb E}(S_N)\right)/{\sigma}_N$ and $\mathfrak{g}(u)=\frac{1}{\sqrt{2\pi}} e^{-u^2/2}. $ In Section \ref{Sec5} we will show, in particular, that Edgeworth expansions of any order $r$ are unique up to terms of order $o({\sigma}_N^{-r}),$ and so the case $r=1$ coincides with the LLT. Edgeworth expansions for discrete (lattice-valued) random variables have been studied in literature for iid random variables \cite[Theorem 4.5.4]{IL} \cite[Chapter VII]{Pet75}, (see also \cite[Theorem 5]{Ess}),
homogeneous Markov chains \cite[Theorems 2-4]{Nag2},
decomposable statistics \cite{MRM}, or { dynamical systems \cite{FL} with good spectral properties such as expanding maps. Papers \cite{Bor16, GW17} discuss the rate of convergence in the LLT. Results for non-lattice variables were obtained in \cite{Feller, BR76, CP, Br} (which considered random vectors) and \cite{FL} (see also \cite{Ha} for corresponding results for random expanding dynamical systems).
In this paper we obtain analogues of Theorems \ref{ThProkhorov} and \ref{ThLLT} for higher order Edgeworth expansions for independent but not identically distributed integer-valued uniformly bounded random variables. We begin with the following result.
\begin{theorem} \label{ThEdgeMN}
Let $\displaystyle K=\sup_j\|X_j\|_{L^\infty}.$ For each $r\in{\mathbb N}$ there is a constant $R\!\!=\!\!R(r, K)$ such that the Edgeworth expansion of order $r$ holds~if \[ M_N:=\min_{2\leq h\leq 2K}\sum_{n=1}^N{\mathbb P}(X_n\neq m_n(h) \text{ mod } h)\geq R\ln V_N. \] In particular, $S_N$ obeys Edgeworth expansions of all orders if $$ \lim_{N\to\infty} \frac{M_N}{\ln V_N}=\infty. $$ \end{theorem} The number $R(r,K)$ can be chosen according to Remark \ref{R choice}. This theorem is a quantitative version of Prokhorov's Theorem \ref{ThProkhorov}. We observe that logarithmic in $V_N$ growth of various non-perioidicity characteristics of individual summands are often used in the theory of local limit theorems (see e.g. \cite{Mal78, MS70, MPP}). We will see from the examples of Section \ref{ScExamples} that this result is close to optimal. However, to justify the optimality we need to understand the conditions necessary for the validity of the Edgeworth expansion.
\begin{theorem}\label{r Char} For any $r\geq1$, the Edgeworth expansion of order $r$ holds if and only if for any nonzero resonant point $t$ and $0\leq\ell<r$ we have \[ \bar \Phi_{N}^{(\ell)}(t)=o\left({\sigma}_N^{\ell+1-r}\right). \] where $\bar\Phi_{N}(x)={\mathbb E}[e^{ix (S_N-{\mathbb E}(S_N))}]$ and $\bar\Phi_{N}^{(\ell)}(\cdot)$ is its $\ell$-th derivative. \end{theorem}
This result\sout{s} generalizes Theorem \ref{ThLLT}, however in contrast with that theorem, in the case $r>1$ we also need to take into account the behavior of the derivatives of the characteristic function at nonzero resonant points. The values of the characteristic function at the resonant points $2\pi l/m$ have clear probabilistic meaning. Namely, they control the rate
equidistribution modulo $m$ (see part (d) of Theorem \ref{ThLLT} or Lemma \ref{LmUnifFourier}).
Unfortunately, the probabilistic meaning of the derivatives is less clear, so it is desirable to characterize the validity of the Edgeworth expansions of orders higher than 1 without
considering the derivatives. Example \ref{ExUniform} shows that this is impossible without additional assumptions. Some of the reasonable additional conditions are presented below.
We start with the expansion of order 2.
\begin{theorem}\label{Thm SLLT VS Ege} Suppose $S_N$ obeys the SLLT. Then the following are equivalent:
(a) Edgeworth expansion of order 2 holds;
(b) $|\Phi_N(t)|=o(\sigma_N^{-1})$ for each nonzero resonant point $t$;
(c) For each $h\leq 2K$ the distribution of $S_N$ mod $h$ is $o(\sigma_N^{-1})$ close to uniform. \end{theorem}
Corollary \ref{CorNoDer} provides an extension
of Theorem \ref{Thm SLLT VS Ege} for expansions of an arbitrary order $r$ under an additional
assumption that
$\displaystyle \varphi:=\min_{t\in{\mathcal{R}}}\inf_{n}|\phi_n(t)|>0$, where ${\mathcal{R}}$ is the set of all nonzero resonant points. The latter condition implies in particular,
that for each $\ell$ there is a uniform lower bound on the distance between the distribution of $X_{n_1}+X_{n_2}+\dots +X_{n_\ell}$
\text{mod }$m$ and the uniform distribution, when $\{n_1, n_2,\dots ,n_\ell\}
\in{\mathbb N}^\ell$ and $m\geq2$.
Next we discuss an analogue of Theorem \ref{ThProkhorov} for expansions of order higher than 2. It requires
a stronger condition which uses an additional notation. Given $j_1, j_2,\dots,j_s$ with $j_l\in [1, N]$ we write $$ S_{N; j_1, j_2, \dots,j_s}=S_N-\sum_{l=1}^s X_{j_l}. $$ Thus $S_{N; j_1, j_2, \dots ,j_s}$ is a partial sum of our sequence with $s$ terms removed. We will say that $\{X_n\}$ {\em admits an Edgeworth expansion of order $r$ in a superstable way} (which will be denoted by $\{X_n\}\in EeSs(r)$) if for each ${\bar s}$ and each sequence $j_1^N, j_2^N,\dots ,j_{s_N}^N$ with $s_N\leq {\bar s}$ there are polynomials $P_{b, N}$ whose coefficients are $O(1)$ in $N$ and their degrees do not depend on $N$ so that
uniformly in $k\in{\mathbb Z}$ we have that \begin{equation}\label{EdgeDefSS} {\mathbb P}(S_{N; j_1^N, j_2^N, \dots, j_{s_N}^N}=k)=\sum_{b=1}^r \frac{P_{b, N} (k_N)}{\sigma_N^b} \mathfrak{g}(k_N)+o(\sigma_N^{-r}) \end{equation} and the estimates in $O(1)$ and $o(\sigma_N^{-r})$ are uniform in the choice of the tuples $j_1^N, \dots ,j_{s_N}^N.$ That is, by removing a finite number of terms we can not destroy the validity of the Edgeworth expansion (even though the coefficients of the underlying polynomials will of course depend on the choice of the removed terms). Let $\Phi_{N; j_1, j_2,\dots, j_s}(t)$ be the characteristic function of $S_{N; j_1, j_2,\dots, j_s}.$
\begin{remark} We note that in contrast with SLLT, in the definition of the superstrong Edgeworth expansion one is only allowed to remove old terms, but not to add new ones. This difference in the definition is not essential, since adding terms with sufficiently many moments (in particular, adding bounded terms) does not destroy the validity of the Edgeworth expansion. See the proof of Theorem \ref{Thm Stable Cond} (i) or the second part of Example \ref{ExNonAr}, starting with equation \eqref{Convolve}, for details. \end{remark}
\begin{theorem}\label{Thm Stable Cond} (1) $S_N\in EeSs(1)$ (that is, $S_N$ satisfies the LLT in a superstable way) if and if it satisfies the SLLT.
(2) For arbitrary $r\geq 1$ the following conditions are equivalent:
(a) $\{X_n\}\in EeSs(r)$;
(b) For each $j_1^N, j_2^N,\dots ,j_{s_N}^N$ and each nonzero resonant point $t$ we have $\Phi_{N; j_1^N, j_2^N,\dots, j_{s_N}^N}(t)=o(\sigma_N^{1-r});$
(c) For each $j_1^N, j_2^N,\dots ,j_{s_N}^N$, and each $h \leq 2K$ the distribution of $S_{N; j_1^N, j_2^N,\dots, j_{s_N}^N}$ mod $h$ is $o(\sigma_N^{1-r})$ close to uniform. \end{theorem}
To prove the above results we will show that for any order $r$, we can always approximate $\mathbb{P}(S_N=k)$ up to an error $o(\sigma_N^{-r})$ provided that instead of polynomials we use products of regular and
the trigonometric polynomials. Those products allow us} to take into account possible oscillatory behavior of $P(S_N=k)$ when $k$ belongs to different residues mod $h$, where $h$ is denominator of a {\em resonant frequency.} When $M_N\geq R V_N$ for $R$ large enough, the new expansion coincides with the usual Edgeworth expansions. We thus derive that the condition $M_N\geq R\ln V_N$ is
in a certain sense optimal.
\section{Main result}\label{Main R} Let $X_1,X_2,...$ be a sequence of independent integer-valued random variables. For each $N\in{\mathbb N}$ we set
$\displaystyle S_N=\sum_{n=1}^N X_n$ and $V_N=\text{Var}(S_N)$. We assume in this paper that
$\displaystyle \lim_{N\to\infty}V_N=\infty$ and that
there is a constant $K$ such that
$$\sup_n \|X_n\|_{L^\infty}\leq K. $$ Denote $\sigma_N=\sqrt{V_N}.$ For each positive integer $m$, let $ q_n(m)$ be the second largest among $$\displaystyle \sum_{l\equiv j \text{ mod }m} {\mathbb P}(X_n=l)={\mathbb P}(X_n\equiv j \text{ mod }m),\,j=1,2,...,m$$ and $j_n(m)$ be the corresponding residue class. Set $$ M_N(m)=\sum_{n=1}^N q_n(m)\quad\text{and}\quad M_N=\min_m M_N(m).$$
\begin{theorem}\label{IntIndThm}
There $\exists J=J(K)<\infty$ and polynomials $P_{a, b, N}$, where $a\in 0, \dots ,J-1,$ $b\in \mathbb{N}$,
with degrees depending only on $b$ but not on $a, K$ or on any other characteristic of $\{X_n\}$, such that the coefficients of $P_{a,b, N}$ are uniformly bounded in $N$, and, for any $r\geq1$ uniformly in $k\in{\mathbb Z}$ we have $${\mathbb P}(S_N=k)-\sum_{a=0}^{J-1} \sum_{b=1}^r \frac{P_{a, b, N} ((k-a_N)/\sigma_N)}{\sigma_N^b} \mathfrak{g}((k-a_N)/\sigma_N) e^{2\pi i a k/J} =o(\sigma_N^{-r}) $$ where $a_N={\mathbb E}(S_N)$ and $\mathfrak{g}(u)=\frac{1}{\sqrt{2\pi}} e^{-u^2/2}. $
Moreover, $P_{0,1,N}\equiv1$, and given $K, r$, there exists $R=R(K,r)$ such that if $M_N\geq R \ln V_N$ then we can choose $P_{a, b, N}=0$ for $a\neq 0.$ \end{theorem} We refer the readers to (\ref{r=1}) for more details on these expansions in the case $r=1$, and to Section \ref{FirstOrder} for a discussion about the relations with local limit theorems. The resulting expansions in the case $r=2$ are given in (\ref{r=2'}). We note that the constants $J(K)$ and $R(K,r)$ can be recovered from the proof of Theorem \ref{IntIndThm}.
\begin{remark} Since the coefficients of the polynomials $P_{a,b,N}$ are uniformly bounded, the terms corresponding to $b=r+1$ are of order $O({\sigma}_N^{-(r+1)})$ uniformly in $k$. Therefore, in the $r$-th order expansion we actually get that the error term is $O({\sigma}_N^{-(r+1)})$. \end{remark}
\begin{remark} In fact, the coefficients of the polynomials $P_{a,b,N}$ for $a>0$ are bounded by a constant times $(1+M_N^{q})e^{-c_0 M_N}$, where $c_0>0$ depends only on $K$ and $q\geq 0$ depends only on $r$ and $K$. Therefore, these coefficient are small when $M_N$ is large. When $M_N\geq R(r,K)\ln V_N$ these coefficients become of order $o({\sigma}^{-r}_N).$ Therefore, they only contribute to the error term, and so we can replace them by $0$, as stated in Theorem \ref{IntIndThm}. \end{remark}
\begin{remark} As in the derivation of the classical Edgeworth expansion, the main idea of the proof of Theorem \ref{IntIndThm} is the stationary phase analysis of the characteristic function. However, in contrast with the iid case
there may be
resonances other than 0 which contribute to the oscillatory terms in the expansion. Another interesting case where the classical Edgeworth analysis fails is the case of iid terms where the summands are non-arithmetic but take only finitely many values. It is shown in \cite{DF} that in that case, the leading correction to the Edgeworth expansion also comes from resonances. However, in the case studied in \cite{DF} the geometry of resonances is more complicated, so in contrast to our Theorem \ref{IntIndThm}, \cite{DF} does not get the expansion of all orders. \end{remark}
\section{Edgeworth expansions under quantitative Prokhorov condition.} \label{ScEdgeLogProkh} In this section we prove Theorem \ref{ThEdgeMN}. In the course of the proof we obtain the estimates of the characteristic function on intervals not containing resonant points which will also play an important role in the proof of Theorem \ref{IntIndThm}. The proof of Theorem \ref{IntIndThm} will be completed in Section \ref{ScGEE} where we analyze additional contribution coming from nonzero resonant points which appear in the case $M_N\leq R \ln {\sigma}_N.$ Those contributions constitute the source of the trigonometric polynomials in the generalized Edgeworth expansions.
\subsection{Characterstic function near 0.} Here we recall some facts about the behavior of the characteristic function near 0, which will be useful in the proofs of Theorems \ref{ThEdgeMN} and \ref{IntIndThm}. The first result holds for general uniformly bounded sequences $\{X_n\}$ (which are not necessarily integer-valued). \begin{proposition}\label{PropEdg} Suppose that $\displaystyle \lim_{N\to\infty} {\sigma}_N=\infty$, where ${\sigma}_N\!=\!\sqrt{V_N}= \sqrt{V(S_N)}$. Then for $k=1,2,3,...$ there exists a sequence of polynomials $(A_{k,N})_N$ whose degree $d_k$ depends only on $k$ so that for any $r\geq 1$ there is ${\delta}_r>0$ such that for all $N\geq1$ and $t\in[-{\delta}_r{\sigma}_N,{\delta}_r{\sigma}_N]$, \begin{equation}\label{FinStep.0} {\mathbb E}\left(e^{it(S_N-{\mathbb E}(S_N))/{\sigma}_N}\right)=e^{-t^2/2}\left(1+\sum_{k=1}^r \frac{A_{k,N}(t)}{{\sigma}_N^k}+\frac{t^{r+1}}{{\sigma}_N^{r+1}} O(1)\right). \end{equation} Moreover, the coefficients of $A_{k,N}$ are algebraic combinations of moments of the $X_m$'s and they are uniformly bounded in $N$. Furthermore \begin{equation}\label{A 1,n .1} A_{1,N}(t)= -\frac{i}{6}{\gamma}_N t^3\,\,\text{ and }\,\,A_{2,N}(t)=\Lambda_{4}(\bar S_N){\sigma}_N^{-2}\frac{t^4}{4!}-\frac{1}{36}{\gamma}_N^2 t^6 \end{equation} where $\bar S_N=S_N-{\mathbb E}(S_N)$, ${\gamma}_N={\mathbb E}[(\bar S_N)^3]/{\sigma}_N^2$ and $\Lambda_{4}(\bar S_N)$ is the fourth comulant of $\bar S_N$. \end{proposition} The proof is quite standard, so we just sketch the argument.
The idea is to fix some $B_2>B_1>0$, and to partition $\{1,...,N\}$ into intervals $I_1,...,I_{m_N}$ so that $\displaystyle B_1\leq \text{Var}(S_{I_l})\leq B_2 $ where for each $l$ we set $\displaystyle S_{I_l}=\sum_{j\in I_l}X_i$. It is clear that $m_N/{\sigma}_N^2$ is bounded away from $0$ and $\infty$ uniformly in $N$. Recall next that there are constants $C_p$, $p\geq2$ so that for any $n\geq1$ and $m\geq 0$ we have \begin{equation} \label{CenterMoments}
\left\|\sum_{j=n}^{n+m}\big(X_j-{\mathbb E}(X_j)\big)\right\|_{L^p}\leq C_p\left(1+\left\|\sum_{j=n}^{n+m}\big(X_j-{\mathbb E}(X_j)\big)\right\|_{L^2}\right). \end{equation} This is a consequence of the multinomial theorem and some elementary estimates, and we refer the readers to either Lemma 2.7 in \cite{DS}, or Theorem 6.17 in \cite{Pel} for such a result in a much more general settings. Using the latter estimates we get that the $L^p$-norms of $S_{I_l}$ are uniformly bounded in $l$. This reduces the problem to the case when the variance of $X_n$ is uniformly bounded from below, and all the moments of $X_n-{\mathbb E}(X_n)$ are uniformly bounded. In this case, the proposition follows by considering the Taylor expansion of the function $\ln {\mathbb E}\big(e^{it(S_N-{\mathbb E}(S_N))/{\sigma}_N}\big)+\frac12t^2$, see \cite[\S XVI.6]{Feller}.
\begin{proposition} \label{PrHalf} Given a square integrable random variable $X$, let $\bar{X}=X-{\mathbb E}(X).$ Then for each $h\in\mathbb{R}$ we have
$$\left|{\mathbb E}(e^{ih \bar X})-1\right|\leq \frac12 h^2V(X).$$ \end{proposition}
\begin{proof} Set $\varphi(h)={\mathbb E}(e^{ih \bar X})$. Then by the integral form of the second order Taylor reminder we have $$
|\varphi(h)-\varphi(0)-h\varphi'(0)|=|\varphi(h)-\varphi(0)|=\left|\int_0^h(t-h)\varphi''(t)dt\right|$$ $\hskip4.7cm \displaystyle \leq
V(X)\int_{0}^{|h|}(|h|-t)dt= \frac12 h^2V(X). $ \end{proof}
\subsection{ Non resonant intervals.} \label{SSNonRes} As in almost all the proofs of the LLT, the starting point in the proof of Theorem \ref{ThEdgeMN} (and Theorem \ref{IntIndThm}) is that for $k, N\in{\mathbb N}$ we have \begin{equation} \label{EqDual} 2\pi {\mathbb P}(S_N=k)=\int_{0}^{2\pi} e^{-itk}{\mathbb E}(e^{it S_N})dt. \end{equation}
Denote ${\mathbb T}={\mathbb R}/2\pi {\mathbb Z}.$ Let $$ \Phi_N(t)={\mathbb E}(e^{it S_N})=\prod_{n=1}^N \phi_n(t) \quad \text{where}\quad \phi_n(t)={\mathbb E}(e^{it X_n}).$$
Divide ${\mathbb T}$ into intervals $I_j$ of small size $\delta$ such that each interval contains at most one resonant point and this point is strictly inside $I_j.$ We call an interval resonant if it contains a resonant point inside. Then \begin{equation}\label{SplitInt} 2\pi {\mathbb P}(S_N=k)=\sum_{j}\int_{I_j} e^{-itk}{\mathbb E}(e^{it S_N})dt. \end{equation} We will consider the integrals appearing in the above sum individually.
\begin{lemma}\label{Step1} There are constants $C,c>0$ which depend only on ${\delta}$ and $K$ so that for any non-resonant interval $I_j$ and $N\geq1$ we have
$$ \int_{I_j} |\Phi_N(t)| dt\leq C e^{-c V_N}. $$ \end{lemma} \begin{proof} Let ${\hat q}_n, {\bar q}_n$ be the largest and the second largest values of ${\mathbb P}(X_n=j)$ and let ${\hat j}_n, {\bar j}_n$ be the corresponding values. Note that \begin{equation} \label{phiNSum}
\phi_n(t)={\hat q}_n e^{it {\hat j}_n}+{\bar q}_n e^{it {\bar j}_n}+\sum_{l\neq {\hat j}_n, {\bar j}_n} {\mathbb P}(X_n=l) e^{i t l}. \end{equation} Since $I_j$ is non resonant, the angle between $e^{it {\hat j}_n}$ and $e^{it {\bar j}_n}$ is uniformly bounded from below. Indeed if this was not the case we would have $t {\bar j}_n-t{\hat j}_n\approx 2\pi l_n$ for some $l_n\in{\mathbb Z}.$ Then $t\approx \frac{2\pi l_n}{m_n}$ where $m_n={\bar j}_n-{\hat j}_n$ contradicting the assumption that $I_j$ is non-resonant. Accordingly $\exists c_1>0$ such that
$\displaystyle \left|e^{it {\hat j}_n}+e^{it {\bar j}_n}\right|\leq 2-c_1. $ Therefore
$$ \left|{\hat q}_n e^{it {\hat j}_n}+{\bar q}_n e^{it {\bar j}_n}\right|
\leq { ({\hat q}_n-{\bar q}_n)+{\bar q}_n \left| e^{it {\hat j}_n}+e^{it {\bar j}_n} \right|}
\leq {\hat q}_n+{\bar q}_n-2c_1 {\bar q}_n. $$
Plugging this into \eqref{phiNSum}, we conclude that
$|\phi_n(t)|\leq 1-2c_1 {\bar q}_n$ for $t\in I_j.$ Multiplying these estimates over $n$ and using that $1-x\leq e^{-x},\, x>0$, we get
$$ |\Phi_N(t)|\leq e^{-2c_1\sum_n {\bar q}_n}. $$ Since $V(X_n)\leq c_2 {\bar q}_n$ for a suitable constant $c_2$ we can rewrite the preceding as \begin{equation}\label{NonResDec}
|\Phi_N(t)|\leq e^{-c_3V_N},\, c_3>0. \end{equation} Integrating over $I_j$ we obtain the result. \end{proof}
\subsection{Prokhorov estimates} Next we consider the case where $I_j$ contains a nonzero resonant point $t_j=\frac{2\pi l}{m}.$
\begin{lemma}\label{Step2} There is a constant $c_0$ which depends only on $K$ so that for any nonzero resonant point $t_j=2\pi l/m$ we have \begin{equation}\label{Roz0}
\sup_{t\in I_j}|{\mathbb E}(e^{it S_N})|\leq e^{-c_0 M_N(m)}. \end{equation} Thus, for any $r\geq1$ there is a constant $R=R(r,K)$ such that if $M_N(m)\geq R \ln V_N$, then the integral $\int_{I_j} e^{-itk}{\mathbb E}(e^{it S_N})dt$ is $o(\sigma_N^{-r})$ uniformly in $k$, and so it only contributes to the error term. \end{lemma}
\begin{proof} The estimate (\ref{Roz0}) follows from the arguments in \cite{Rozanov}, but
for readers' convenience we recall its proof. Let $X$ be an integer-valued random variable so that $\|X\|_{L^\infty}\leq K$. Let $t_0=2\pi l/m$ be a nonzero resonant point, where $\gcd(l,m)=1$. Let $t\in{\mathbb T}$ be so that \begin{equation} \label{Neart0}
|t-t_0|\leq{\delta}, \end{equation} where ${\delta}$ is a small positive number. Let $\phi_X(\cdot)$ denote the characteristic function of $X$. Since $x\leq e^{x-1}$ for any real $x$ we have \[
|\phi_X(t)|^2\leq e^{|\phi(t)|^2-1}. \] Next, we have \[
|\phi_X(t)|^2-1=\phi(t)\phi(-t)-1=\sum_{j=-2K}^{2K}\sum_s\tilde P_j
\left[\cos(t_j)-1\right] \] where \[ \tilde P_j=\sum_{s}{\mathbb P}(X=s){\mathbb P}(X=j+s). \] Fix some $-2K\leq j\leq 2K$.
We claim that if ${\delta}$ in \eqref{Neart0} is small enough and $j\not\equiv 0\text{ mod }m$ then for each integer $w$ we have $|t-2\pi w/j|\geq {\varepsilon}_0$ for some ${\varepsilon}_0>0$ which depends only on $K$. This follows from the fact that $-2K\leq j\leq 2K$ and that $2\pi w/j\not=t_0$ (and there is a finite number of resonant points). Therefore, \[ \cos(tj)-1\leq -{\delta}_0 \] for some ${\delta}_0>0$. On the other hand, if $j=km$ for some integer $k$ then with $w=lk$ we have \begin{eqnarray*} \cos(tj)-1=-2\sin^2(tj/2)=-2\sin^2\left((tj-2\pi w)/2\right)\\=-2\sin^2\left(j(t-t_0)/2\right)\leq -{\delta}_1(t-t_0)^2 \end{eqnarray*}
for some ${\delta}_1>0$ (assuming that $|t-t_0|$ is small enough). We conclude that \[
|\phi_X(t)|^2-1\leq -{\delta}_0\sum_{j\in A}\tilde P_j-{\delta}_1(t-t_0)^2\sum_{j\in B}\tilde P_j \] where $A=A(X)$ is the set of $j$'s between $-2K$ and $2K$ so that $j\not\equiv 0\text{ mod }m$ and $B=B(X)$ is its complement in ${\mathbb Z}\cap[-2K,2K]$. Let $s_0$ be the most likely
residue of $X$ mod $m$ and $s_1$ be the second most likely residue class. Since $$ {\mathbb P}(X\equiv s_0 \text{ mod }m)\geq \frac{1}{m} \quad\text{and}\quad {\mathbb P}(X\equiv s_1 \text{ mod }m)=q_m(X) $$ it follows that $\displaystyle \sum_{j\in A} \tilde P_j\geq \frac{q_m(X)}{m}.$
Combining this with the trivial bound $\displaystyle \sum_{j\in B} \tilde P_j\geq {\mathbb P}^2(X\equiv s_0)\geq \frac{1}{m^2}$ we obtain $$
|\phi_X(t)|\leq \exp-\left[\frac12\left(\frac{{\delta}_0 q_m(X)}{m} + \frac{{\delta}_1 (t-t_0)^2}{m^2}\right)\right]. $$ Applying the above with $t_0=t_j$ and $X=X_n$, $1\leq n\leq N$ we get that \begin{equation}\label{CharEstRoz}
|\Phi_N(t)|\leq e^{-c_0M_N(m)-\bar{c}_0 N(t-t_j)^2 }\leq e^{-c_0M_N(m)} \end{equation} where $c_0$ is some constant. \end{proof}
\begin{remark} Using the first inequality in (\ref{CharEstRoz}) and arguing as in \cite[page 264]{Rozanov}, we can deduce that there are positive constants $C, c_1, c_2$ such that \begin{equation}\label{RozArg}
\int_{I_j}|{\mathbb E}(e^{it S_N})|dt\leq C\left(e^{-c_1{\sigma}_N}+\frac{e^{-c_2M_N(m)}}{{\sigma}_N}\right). \end{equation} This estimate plays an important role in the proof of the SLLT in \cite{Rozanov}, but for our purposes a weaker bound \eqref{Roz0} is enough. Note also that in order to prove (\ref{Roz0}) we could have just used the trivial inequality $\cos(t_j)-1\leq 0$ when $j\equiv 0\text{ mod }m$, but we have decided to present this part from \cite{Rozanov} in full. \end{remark}
\begin{remark}\label{R choice} Let $d_{{\mathcal{R}}}$ be the minimal distance between two different resonant points. Then, when ${\delta}<2d_{{\mathcal{R}}}$, we can take ${\delta}_0=1-\cos(d_{{\mathcal{R}}})$ in the proof of Lemma \ref{Step2}. Therefore, we can take $c_0=\frac{1-\cos(d_{{\mathcal{R}}})}{4K}$ in \eqref{Roz0}. Hence Lemma \ref{Step2} holds with
$R(r,K)=\frac{r+1}{2c_0}.$ \end{remark} \vskip0.2cm
\subsection{ Proof of Theorem \ref{ThEdgeMN}} \label{Cmplt1} Fix some $r\geq1$. Lemmas \ref{Step1} and \ref{Step2} show that if $M_N\geq R(r,K)\ln V_N$, then all the integrals in the right hand side of (\ref{SplitInt}) are of order $o({\sigma}_N^{-r})$, except for the one corresponding to the resonant point $t_j=0$. That is, for any ${\delta}>0$ small enough, uniformly in $k$ we have $$2\pi {\mathbb P}(S_N=k)=\int_{-{\delta}}^{\delta} e^{-ih k}\Phi_N(h)dh+o({\sigma}_N^{-r}). $$ In order to complete the proof of Theorem \ref{ThEdgeMN}, we need to expand the above integral. Making a change of variables $h\to h/{\sigma}_N$ and using Proposition \ref{PropEdg}, we conclude that if ${\delta}$ is small enough then $$ \int_{-{\delta}}^{\delta} e^{-ih k}\Phi_N(h)dh=$$ $${\sigma}_N^{-1}\int_{-{\delta}{\sigma}_N}^{{\delta}{\sigma}_N}e^{-ihk_N}e^{-h^2/2}\left(1+\sum_{u=1}^r \frac{A_{u,N}(h)}{{\sigma}_N^k}+\frac{h^{r+1}}{{\sigma}_N^{r+1}} O(1)\right)dh $$ where $k_N=\left(k-{\mathbb E}(S_N)\right)/{\sigma}_N$. Since the coefficients of the polynomials $A_{u,N}$ are uniformly bounded in $N$, we can just replace the above integral with the corresponding integral over all ${\mathbb R}$ (i.e. replace $\pm{\delta}{\sigma}_N$ with $\pm\infty$). Now the Edgeworth expansions are achieved using that for any nonnegative integer $q$ we have that $(it)^qe^{-t^2/2}$ is the Fourier transform of the $q$-th derivative of $\textbf{n}(t)=\frac{1}{\sqrt{2\pi}}e^{-t^2/2}$ and that for any real $a$, \begin{equation}\label{Fourir} \int_{-\infty}^\infty e^{-iat}\widehat{\textbf{n}^{(q)}}(t) dt=\textbf{n}^{(q)}(a)=\frac{1}{\sqrt{2\pi}}(-1)^{q}H_q(a)e^{-a^2/2} \end{equation} where $H_q(a)$ is the $q$-th Hermite polynomial.
\section{Generalized Edgeworth expansions: Proof of Theorem Theorem~ \ref{IntIndThm}} \label{ScGEE} \subsection{Contributions of resonant intervals.} Let $r\geq1$. As in the proof of Theorem \ref{ThEdgeMN}, our starting point is the equality \begin{equation} \label{EqDual} 2\pi {\mathbb P}(S_N=k)=\int_{0}^{2\pi} e^{-itk}{\mathbb E}(e^{it S_N})dt=\sum_{j}\int_{I_j} e^{-itk}{\mathbb E}(e^{it S_N})dt \end{equation} which holds for any $k\in{\mathbb N}$. We will consider the integrals appearing in the above sum individually. By Lemma \ref{Step1} the integrals over non-resonant intervals are of order $o({\sigma}_N^{-r})$, and so they can be disregarded. Moreover, in \S \ref{Cmplt1} we have expanded the integral over the resonant interval containing $0$.
Now we will see that in the case $M_N< R(r,K) \ln V_N$ the contribution of nonzero resonant points need not be negligible.
Let $t_j=\frac{2\pi l}{m}$ be a nonzero resonant point so that $M_N(m)<R(r,K) \ln V_N$ and let $I_j$ be the resonant interval containing it. Theorem~\ref{IntIndThm} will follow from an appropriate expansion of the integral $$\int_{I_j} e^{-itk}{\mathbb E}(e^{it S_N})dt.$$ We need the following simple result, which for readers' convenience is formulated as a lemma. \begin{lemma}\label{EpsLem}
There exists ${\bar\varepsilon}>0$ so that for each $n\geq1$ with $q_n(m)\leq {\bar\varepsilon}$ we have $|\phi_n(t_j)|\geq\frac12$. In fact, we can take ${\bar\varepsilon}=\frac1{4m}$. \end{lemma}
\begin{proof} Recall that $t_j=2\pi l/m$. The lemma follows since for any random variable $X$ we have
$\displaystyle |{\mathbb E}(e^{i t_j X})|=$
$$\left|e^{it_js(m,X)}-\sum_{u\not\equiv s(m,X)\text{ mod m}}\big(e^{it_j s(m,X)}-e^{it_j u}\big)P(X\equiv u\text{ mod } m)\right| $$ $$ \geq 1-2mq(m,X) $$ where $s(m,X)$ is the most likely value of $X\text{ mod }m$ and $q(m,X)$ is the second largest value among $P(X\equiv u\text{ mod } m)$, $u=0,1,2,...,m-1$. Therefore, we can take ${\bar\varepsilon}=\frac1{4m}$. \end{proof} Next, set ${\bar\varepsilon}=\frac{1}{8K}$ and let $N_0=N_0(N,t_j,{\bar\varepsilon})$ be the number of all $n$'s between $1$ to $N$ so that $q_n(m)>{\bar\varepsilon}$. Then $N_0\leq \frac{R \ln V_N}{{\bar\varepsilon}}$ because $M_N(m)\leq R \ln V_N.$ By permuting the indexes $n=1,2,...,N$ if necessary we can assume that $q_n(m)$ is non increasing. Let $N_0$ be the largest number such that $q_{N_0}\geq {\bar\varepsilon}.$
Decompose \begin{equation} \label{NPert-Pert} \Phi_N(t)=\Phi_{N_0}(t) \Phi_{N_0, N}(t) \end{equation} where $\displaystyle \Phi_{N_0, N}(t)=\prod_{n=N_0+1}^N \phi_n(t).$
\begin{lemma}\label{Step3} If the length ${\delta}$ of $I_j$ is small enough then for any $t=t_j+h\in I_j$ and $N\geq1$ we have \[ \Phi_{N_0,N}(t)=\Phi_{N_0,N}(t_j)\Phi_{N_0,N}(h) \Psi_{N_0,N}(h) \]
where $$ \Psi_{N_0,N}(h)=\exp\left[O(M_N(m))\sum_{u=1}^\infty (O(1))^u h^u\right]. $$ \end{lemma} \begin{proof}
Denote $$ \mu_n={\mathbb E}(X_n), \quad {\bar X}_n=X_n-\mu_n, \quad {\bar\phi}_n(t)={\mathbb E}(e^{it {\bar X}_n}).$$ Let $j_n(m)$ be the most likely residue mod $m$ for $X_n.$ Decompose $$ {\bar X}_n=s_n+Y_n+Z_n$$ where $Z_n\in m{\mathbb Z}$, $s_n=j_n(m)-\mu_n$, so that ${\mathbb P}(Y_n\neq 0)\leq m q_n(m).$ Then for $t=t_j+h$, \begin{equation} \label{IndCharTaylor}
{\bar\phi}_n(t)=e^{i t_j s_n} {\mathbb E}\left(e^{i t_j Y_n} e^{ih {\bar X}_n}\right)={\bar\phi}_n(t_j) \psi_n(h)
\end{equation}
where $$ \psi_n(h)= \left(1+ \frac{i h {\mathbb E}(e^{i t_j Y} {\bar X}_n)-\frac{h^2}{2} {\mathbb E}(e^{i t_j Y} ({\bar X}_n)^2)+\dots}{{\mathbb E}(e^{i t_j Y_n})}\right) . $$ Next, using that for any $x\in(-1,1)$ we have $$1+x=e^{\ln (1+x)}=e^{x-x^2/2+x^3/3-...}$$ we obtain that for $h$ close enough to $0$, \begin{equation}\label{Expand}
\psi_n(h)= \exp\left(\sum_{k=1}^\infty\frac{(-1)^{k+1}}{k}\left(\frac1{{\mathbb E}(e^{it_jY_n})}\sum_{q=1}^\infty \frac{(ih)^q}{q!}{\mathbb E}(e^{it_j Y_n}(\bar X_n)^q)\right)^k\right) \end{equation} \begin{eqnarray} =\exp\left(\sum_{k=1}^\infty\frac{(-1)^{k+1}}{k}\sum_{1\leq j_1,...,j_k}\frac1{({\mathbb E}(e^{it_jY_n}))^k}\prod_{r=1}^{k}\frac{(ih)^{j_r}}{j_r!}{\mathbb E}(e^{it_j Y_n}(\bar X_n)^{j_r})\right)\nonumber\\= \exp\left(\sum_{u=1}^\infty\left(\sum_{k=1}^{u}\frac{(-1)^{k+1}}{k}\sum_{j_1+...+j_k=u}\,\prod_{r=1}^{k}\frac{{\mathbb E}(e^{it_j Y_n}(\bar X_n)^{j_r})}{{\mathbb E}(e^{it_j Y_n})j_r!}\right)(ih)^u\right).\nonumber \end{eqnarray} Observe next that \[ {\mathbb E}[e^{it_j Y_n}(\bar X_n)^{j_r}]={\mathbb E}\left[(e^{it_j Y_n}-1)\big((\bar X_n)^{j_r}-{\mathbb E}[(\bar X_n)^{j_r}]\big)\right]+ {\mathbb E}[(\bar X_n)^{j_r}]{\mathbb E}(e^{it_j Y_n}) \] and so with $C=2K$, we have \[ \frac{{\mathbb E}[e^{it_j Y_n}(\bar X_n)^{j_r}]}{{\mathbb E}(e^{it_j Y_n})}=O(q_n(m))O(C^{j_r})+{\mathbb E}[(\bar X_n)^{j_r}]. \] Plugging this into (\ref{Expand}) and using
that for
$h$ small enough, \begin{equation*} \exp\left[ \sum_{u=1}^\infty\left(\sum_{k=1}^u\frac{(-1)^{k+1}}k\sum_{j_1+...+j_k=u}\prod_{r=1}^k\frac{{\mathbb E}({\bar X}_n^{j_r})}{j_r!}\right)(ih)^u\right]={\mathbb E}\left(e^{ih {\bar X}_n}\right) \end{equation*} we conclude that $$
\psi_n(h)={\mathbb E}(e^{ih\bar X_n}) \exp\left[\sum_{u=1}^\infty(O(1))^uO(q_n(m))h^u\right].$$ Therefore, \[ \Phi_{N_0,N}(t)=\Phi_{N_0,N}(t_j)\Phi_{N_0,N}(h) \Psi_{N_0,N}(h) \]
where $\displaystyle \Psi_{ N_0, N}(h)=\exp\left[O(M_N(m))\sum_{u=1}^\infty (O(1))^u h^u\right]. $ \end{proof} \begin{remark}\label{CoeffRem}
We will see in \S\ref{Fin} that the coefficients of the polynomials appearing in Theorem~\ref{IntIndThm} depend on the coefficients of the power series $\Psi_{N_0, N}(h)$ (see, in particular, \eqref{Cj(k)}). The first term in this series is $\displaystyle ih \sum_{n={N_0+1}}^N a_{n,j}$, where \begin{equation}\label{a n,j} a_{n,j}=\frac{{\mathbb E}[(e^{it_j Y_n}-1)\bar X_n]}{{\mathbb E}(e^{i t_j Y_n})}=\frac{{\mathbb E}(e^{it_j X_n}\bar X_n)}{{\mathbb E}(e^{it_j X_n})} \end{equation} while the second term is $\displaystyle \frac{h^2}{2} \sum_{n={N_0+1}}^N b_{n,j}$, where \begin{equation} \label{SecondTerm}
b_{n,j}=\frac{{\mathbb E}[(e^{it_j Y_n}-1)\bar X_n]^2}{{\mathbb E}(e^{i t_j Y_n})^2}-\frac{{\mathbb E}[(e^{it_j Y_n}-1)((\bar X_n)^2-V(X_n))]}{{\mathbb E}(e^{i t_j Y_n})} \end{equation} $$ =a_{n,j}^2-\frac{{\mathbb E}\big(e^{it_j X_n}(\bar X_n)^2\big)}{{\mathbb E}(e^{it_j X_n})}. $$ In Section \ref{Sec2nd} we will use (\ref{a n,j}) to compute the coefficients of the polynomials from Theorem \ref{IntIndThm} in the case $r=2$, and \eqref{SecondTerm}
is one of the main ingredients for the computation in the case $r=3$
(which will not be explicitly discussed in this manuscript). \end{remark}
The next step in the proof of Theorem \ref{IntIndThm} is the following. \begin{lemma}\label{Step4} For $t=t_j+h\in I_j$ we can decompose \begin{equation}\label{S4} \Phi_{N_0}(t)=\Phi_{N_0}(t_j+h)=\sum_{l=0}^L \frac{\Phi_{N_0}^{(l)}(t_j)}{l!} h^l+O\left((h \ln V_N)^{L+1}\right). \end{equation} \end{lemma} \begin{proof} The lemma follows from the observation that
the derivatives of $\Phi_{N_0}$ satisfy
$|\Phi_{N_0}^{(k)}(t)|\leq O(N_0^k)\leq (C \ln V_N)^k$. \end{proof}
\subsection{Completing the proof}\label{Fin}
Recall \eqref{EqDual} and consider a resonant interval $I_j$ which does not contain $0$ such that $M_N(m)\leq R\ln{\sigma}_N$. Set $U_j=[-u_j,v_j]=I_j-t_j$. Let $N_0$ be as described below Lemma \ref{EpsLem}. Denote \begin{equation} \label{SNN0} S_{N_0,N}=S_N-S_{N_0},\, S_0=0, \end{equation} $$V_{N_0,N}=\text{Var}(S_N-S_{N_0})=V_N-V_{N_0}\quad\text{and} \quad {\sigma}_{N_0,N}=\sqrt{V_{N_0,N}}.$$ Then \begin{equation}\label{Vars} V_{N_0,N}=V_N+O(\ln V_N)=V_N(1+o(1)). \end{equation}
Denote $h_{N_0,N}=h/{\sigma}_{N_0,N}.$
By \eqref{FinStep.0}, if $|h_{N, N_0}|$ is small enough then \begin{equation}\label{FinStep} {\mathbb E}(e^{ih_{N_0,N} S_{N_0,N}})= \end{equation} $$e^{ih_{N_0,N}{\mathbb E}(S_{N_0,N})}e^{-h^2/2}\left(1+\sum_{k=1}^r \frac{A_{k,N_0,N}(h)}{{\sigma}_{N_0,N}^k}+\frac{h^{r+1}}{{\sigma}_{N_0,N}^{r+1}} O(1)\right)$$ where $A_{k,N_0,N}$ are polynomials with bounded coefficients (the degree of $A_{k,N_0,N}$ depends only on $k$). Let us now evaluate $\displaystyle \int_{I_j}e^{-itk}{\mathbb E}(e^{it S_N})dt. $ By Lemma \ref{Step3}, \begin{equation} \label{ResInt} \int_{I_j}e^{-itk}\Phi_N(t)dt= \end{equation} $$ e^{-it_j k}\Phi_{N_0,N}(t_j)\int_{U_j}e^{-ihk}\Phi_{N_0}(t_j+h)\Phi_{N_0,N}(h) \Psi_{N_0, N}(h)\; dh.
$$
Therefore, it is enough to expand the integral on the RHS of \eqref{ResInt}. Fix a large positive integer $L$ and plug \eqref{S4} into \eqref{ResInt}.
Note that for $N$ is large enough, $h_0$ small enough and $|h|\leq h_0$, Proposition \ref{PrHalf} and \eqref{Vars} show that there exist positive constants $c_0, c$ such that \begin{equation}\label{expo}
|\Phi_{N_0,N}(h)|=|{\mathbb E}(e^{ih S_{N_0,N}})|\leq e^{-c_0(V_N-V_{N_0})h^2}\leq
e^{-cV_N h^2}. \end{equation}
Thus, the contribution coming from the term $O\left((h \ln V_N)^{L+1}\right)$ in the right hand side of (\ref{S4}) is at most of order \[ V_N^{R{\delta}}(\ln V_n)^{L+1}\int_{-\infty}^\infty h^{L+1}e^{-c V_N h^2}dh \] where ${\delta}$ is the diameter of $I_j$. Changing variables $x={\sigma}_N h$, where ${\sigma}_N=\sqrt{V_N}$ we get that the latter term is of order $(\ln V_n)^{L+1}{\sigma}_{N}^{-(L+1-2R{\delta})}$ and so when $L$ is large enough we get that this term is $o({\sigma}_N^{-r-1})$ (alternatively, we can take $L=r$ and ${\delta}$ to be sufficiently small). This means that it is enough to expand each integral of the form \begin{equation} \label{HLInt} \int_{U_j}e^{-ih k}h^l\Phi_{N_0,N}(h) \Psi_{N_0, N}(h) dh
\end{equation} where $l=0,1,...,L$ (after changing variables the above integral is divided by ${\sigma}_{N_0,N}^{l+1}$). Next,
Lemma \ref{Step3} shows that for any $d\in {\mathbb Z}$ we have \begin{equation}
\label{EqOrderD}
\Psi_{N_0, N}(h)
=1+\sum_{u=1}^{d}C_{w,N}h^u+h^{d+1}O(1+M_N(m)^{d+1})|V_N|^{O(|h|)}, \end{equation} where $C_{w,N}=C_{w,N, t_j}$ are $O(M_{N}^u(m))=O((\ln V_N)^{u})$. Note that, with $a_{n,j}$ and $b_{n,j}$ defined in Remark \ref{CoeffRem}, we have \begin{equation}\label{C 1 N} C_{1,N}=i\sum_{n=N_0+1}^{N}a_{n,j} \end{equation} and \[ C_{2,N}=\frac12\sum_{n=N_0+1}^{N}b_{n,j}-\frac 12\left(\sum_{n=N_0+1}^{N}a_{n,j}\right)^2. \]
Take $d$ large enough and plug \eqref{EqOrderD} into \eqref{HLInt}. Using (\ref{expo}), we get again that the contribution of the term
$$h^{d+1}O(1+M_N(m)^{d+1})|V_N|^{O(|h|)}h^l\Phi_{N_0,N}(h)$$ to the above integral is $o({\sigma}_N^{-r})$. Thus, it is enough to expand each term of the form \[ \int_{U_j}e^{-ih k}h^{q}\Phi_{N_0,N}(h)dh \] where $0\leq q\leq L+d$. Using (\ref{FinStep}) and making the change of variables $h\to h/{\sigma}_{N_0,N}$ it is enough to expand \begin{equation}\label{MainInt} \int_{-\infty}^\infty e^{-ih (k-{\mathbb E}[S_{N_0,N}])/{\sigma}_{N_0,N}}h^qe^{-h^2/2}\left(1+\sum_{w=1}^r \frac{A_{w,N_0,N}(h)}{{\sigma}_{N_0,N}^w}+\frac{h^{r+1}}{{\sigma}_{N_0,N}^{r+1}} O(1)\right)
\frac{dh}{{\sigma}_{N_0, N}}. \end{equation} This is achieved by using that $(it)^qe^{-t^2/2}$ is the Fourier transform of the $q$-th derivative of $\textbf{n}(t)=\frac{1}{\sqrt{2\pi}}e^{-t^2/2}$ and that for any real $a$, \begin{equation}\label{Fourir} \int_{-\infty}^\infty e^{-iat}\widehat{\textbf{n}^{(q)}}(t) dt=\textbf{n}^{(q)}(a)=\frac{1}{\sqrt{2\pi}}(-1)^{q}H_q(a)e^{-a^2/2} \end{equation} where $H_q(a)$ is the $q$-th Hermite polynomial.
Note that in the above expansion we get polynomials in the variable $\displaystyle k_{N_0,N}=\frac{k-{\mathbb E}[S_N-S_{N_0}]}{{\sigma}_{N,N_0}}$, not in the variable $k_N=\frac{k-{\mathbb E}(S_N)}{{\sigma}_N}$. Since $k_{N_0,N}=k_N{\alpha}_{N_0,N}+O(\ln {\sigma}_N/{\sigma}_N)$, where ${\alpha}_{N_0,N}={\sigma}_{N}/{\sigma}_{N_0,N}=O(1)$, the binomial theorem shows that such polynomials can be rewritten as polynomials in the variable $k_N$ whose coefficients are uniformly bounded in $N$. We also remark that in the above expansions we get the exponential terms \[ e^{-\frac{(k-a_{N_0,N})^2}{2(V_N-V_{N_0})}} \quad\text{where}\quad a_{N_0,N}={\mathbb E}[S_N-S_{N_0}] \] and not $e^{-(k-a_N)^2/2V_N}$ (as claimed in Theorem \ref{IntIndThm}). In order to address this
fix some ${\varepsilon}<1/2$. Note that for
$|k-a_{N_0,N}|\geq V_N^{\frac12+{\varepsilon}}$ we have
\[ e^{-\frac{(k-a_{N_0,N})^2}{2(V_N-V_{N_0})}}=o(e^{-cV_N^{2{\varepsilon}}})
\quad \text{and} \quad
e^{-\frac{(k-a_{N_0,N})^2}{2V_N}}=o(e^{-cV_N^{2{\varepsilon}}})
\text{ for some }c>0.
\]
Since both terms are $o({\sigma}_N^{-s})$ for any $s$,
it is enough to explain how to replace
$\displaystyle
e^{-\frac{(k-a_{N_0,N})^2}{2(V_N-V_{N_0})}}
$
with $\displaystyle e^{-\frac{(k-a_{N})^2}{2V_N}}$ when
$|k-a_{N_0,N}|\leq V_N^{\frac12+{\varepsilon}}$
(in which case $\displaystyle |k-a_N|=O(V_N^{\frac12+{\varepsilon}})$).
For such $k$'s we can write \begin{equation}\label{ExpTrans1} \exp\left[-\frac{(k-a_{N_0,N})^2}{2(V_N-V_{N_0})}\right]= \end{equation} $$ \exp\left[-\frac{(k-a_{N_0,N})^2}{2V_N}\right]\; \exp\left[-\frac{(k-a_{N_0,N})^2 V_{N_0}}{2V_N(V_N-V_{N_0})}\right]. $$ Since $\displaystyle \frac{(k-a_{N_0,N})^2 V_{N_0}}{2V_N(V_N-V_{N_0})} =O\left(V_N^{-(1-3{\varepsilon})}\right)$, for any $d_1$ we have \begin{equation}\label{ExTrans1.1} \exp\left[-\frac{(k-a_{N_0,N})^2 V_{N_0}}{2V_N(V_N-V_{N_0})}\right]= \end{equation} $$\sum_{j=0}^{d_1}\frac{V_{N_0}^j}{2^j(V_N-V_{N_0})^j j!} \left(\frac{(k-a_{N_0,N})^2}{{\sigma}_N^2}\right)^j+O(V_N^{-(d_1+1)(2-3{\varepsilon})}).$$ Note that (using the binomial formula) the first term on the above right hand side
is a polynomial of the variable $(k-a_{N})/{\sigma}_N$ whose coefficients are uniformly bounded in $N$.
Next we analyze the first factor in the RHS of \eqref{ExpTrans1}. As before, it is enough to consider $k$'s such that
$|k-a_N|\leq V_N^{\frac12+{\varepsilon}}$ for a sufficiently small ${\varepsilon}.$ We have \begin{equation}\label{Centring} \exp\left[-\frac{(k-a_{N,N_0})^2}{2V_N}\right]=
\end{equation} $$ \exp\left[-\frac{(k-a_N)^2}{2V_N}\right] \exp\left[-\frac{2(k-a_N)a_{N_0}+a_{N_0}^2}{2V_N}\right]. $$
Note that $\frac{(k-a_N)a_{N_0}+a_{N_0}^2}{2V_N} =k_N\beta_{N_0,N}+\theta_{N_0,N}$, where $$\beta_{N_0,N}=\frac{a_{N_0}}{2{\sigma}_N}=O\left(\frac{\ln {\sigma}_N}{{\sigma}_N}\right) \;\text{ and }\; \theta_{N_0,N}=\frac{a_{N_0}^2}{2V_N}=O\left(\frac{\ln^2{\sigma}_N}{V_N}\right).$$
Approximating $e^{\frac{(k-a_N)a_{N_0}+a_{N_0}^2}{2V_N}}$ by a polynomial of a sufficiently large degree $d_2$ in the variable $\frac{(k-a_N)a_{N_0}+a_{N_0}^2}{2V_N}$ completes the proof of existence of polynomials $P_{a,b,N}$ claimed in the theorem (the Taylor reminder in the last approximation is of order $\displaystyle O\left(V_N^{-d_2(\frac12-{\varepsilon})}\right)$, so we can take $d_2=4(r+1)$ assuming that ${\varepsilon}$ is small enough).
Finally, let us show that the coefficients of the polynomials $P_{a,b,N}$
constructed above are uniformly bounded in $N$. In fact, we will show that for each nonzero resonant point $t_j=2\pi l/m$, the coefficients of the polynomials coming from integration over $I_j$ are of order $$O\left((1+M_N^{q_0}(m))e^{-M_N(m)}\right),$$ where $q_0=q_0(r)$ depends only on $r$.
Observe that the additional contribution to the coefficients of the polynomials coming from the transition between the variables $k_N$ and $k_{N_0,N}$ is uniformly bounded in $N$. Hence we only need to show that the coefficients of the (original) polynomials in the variable $k_{N_0,N}$ are uniformly bounded in $N$. The possible largeness of these coefficient can only come from the terms $C_{u,N,t_j}$, for $u=0,1,2,...,d$ which are of order $M_N^u(m)$, respectively. However, the corresponding terms are multiplied by terms of the form $\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(\ell)}(t_j)$ for certain $\ell$'s which are uniformly bounded in $N$ (see also \eqref{Cj(k)}). We conclude that there are constants $W_j\in{\mathbb N}$ and $a_j\in{\mathbb N}$ which depend only on $t_j$ and $r$ so that the coefficients of the resulting polynomials are composed of a sum of at most $W_j$ terms of order $(M_N(m))^{a_j}\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(\ell)}(t_j)$, where $\ell\leq E(r)$ for some $E(r)\in{\mathbb N}$. Next, we have \begin{equation} \label{DerComb} \Phi_{N_0}^{(\ell)}(t_j)\Phi_{N_0, N}(t_j)= \end{equation} $$\sum_{n_1,\dots, n_k\leq N_0; \atop \ell_1+\dots+\ell_k=\ell} \gamma_{\ell_1,\dots, \ell_k} \left(\prod_{q=1}^k \phi_{n_q}^{(\ell_q)}(t_j) \right) \left[\prod_{n\leq N, \; n\neq n_k} \phi_n(t_j)\right]$$ where $\gamma_{\ell_1,\dots, \ell_k}$ are bounded coefficients of combinatorial nature.
Using (\ref{Roz0}) we see that
for each $n_1,\dots, n_k$ the product in the square brackets
is at most $C e^{-c_0 M_N(m)+O(1)}$ for some $C, c_0>0$. Hence
$$|\Phi_{N_0}^{(\ell)}(t_j)\Phi_{N_0, N}(t_j)|\leq \hat{C} N_0^{\ell} \; e^{-c_0 M_N(m)},\,\,\hat C>0.$$ Now, observe that the definition of $N_0$ gives $M_N(m)\geq {\varepsilon}_0 N_0$, ${\varepsilon}_0>0$. Therefore
$\displaystyle |\Phi_{N_0}^{(\ell)}(t_j)\Phi_{N_0, N}(t_j)|\leq C_0 M_N^\ell (m)e^{-c_0 M_N(m)}$, and so each one of the above coefficients is of order $M_N^{\ell'}(m)e^{-c_0 M_N(m)}$ for some $\ell'$ which does not depend on $N$. \qed
\begin{remark} The transition between the variables $k_{N_0,N}$ and $k_{N}$ changes the monomials of the polynomials $P_{a,b,N}$, $a\not=0$ coming from integration over $I_j$, for $t_j\not=0$ into monomials of the form $\frac{c_N a_{N_0}^{j_1}{\sigma}_{N_0}^{j_2}k_N^{j_3}}{{\sigma}_N^{u}}$ for some bounded sequence $(c_N)$, $j_1,j_2,j_3\geq0$ and $u\in{\mathbb N}$. As we have explained, the coefficients of these monomials are uniformly bounded. Still, it seems more natural to consider such monomials as part of the polynomial $P_{a,b+u,N}$. In this case we still get polynomials with bounded coefficients since $a_{N_0}$ and ${\sigma}_{N_0}$ are both $O(N_0)$, $N_0=O(M_{N}(m))$ and $c_N$ contains a term of the form $\Phi_{N_0}^{(\ell)}(t_j)\Phi_{N_0, N}(t_j)$. \end{remark}
\begin{remark} As can be seen from the proof, the resulting expansions might contain terms corresponding to ${\sigma}_N^{-s}$ for $s>r$.
Such terms can be disregarded. For $\frac{|k-a_N|}{{\sigma}_N}\leq V_N^{\varepsilon}$ this follows because the coefficients of our exapansions are $O(1)$ and for $\frac{|k-a_N|}{{\sigma}_N}\geq V_N^{\varepsilon}$ this follows from \eqref{expo}. In practice, some of the polynomials $P_{a,b, N}$ with $b\leq r$ might
have coefficients which are $\displaystyle o({\sigma}_N^{b-r})$ (e.g. when $b+u>r$ in the last remark) so they also can be
disregarded. The question when the terms $P_{a,b, N}$ may be disregarded
is in the heart of the proof of
Theorem \ref{r Char} given in the next section.
\end{remark}
\subsection{A summary}
The proofs of Proposition \ref{PrLLT-SLLT}, Theorem \ref{ThLLT} and Theorem \ref{r Char} will be based on careful analysis of the formulas of the polynomials from Theorem \ref{IntIndThm}. For this purpose, it will be helpful to summarize the main conclusions
from the proof of Theorem \ref{IntIndThm}. Let $r\geq1$ and $t_j=2\pi l/m$ be a nonzero resonant point. Then the arguments in the proof of Theorem \ref{IntIndThm} yield that the contribution to the expansion coming from $t_j$ is \begin{equation}\label{Cj(k)} \textbf{C}_j(k):= \end{equation} $$ e^{-it_j k}\Phi_{N_0,N}(t_j)\sum_{s\leq r-1}\left(\sum_{u+l=s}\frac{\Phi_{N_0}^{(l)}(t_j)C_{u,N}}{l!}\right)\int_{U_j}e^{-ihk}h^{s}\Phi_{N_0,N}(h)dh $$ where $U_j=I_j-t_j$, $C_{u, N}$ are given by \eqref{EqOrderD} and $C_{0,N}=1$. When $t_j=0$ then it is sufficient to consider only $s=0$, $N_0=0$ and the contribution is just the integral \[ \int_{-{\delta}}^{\delta} e^{-ihk}\Phi_{N}(h)dh \] where ${\delta}$ is small enough. As in (\ref{MainInt}), changing variables we can replace the integral corresponding to $h^s$ with \begin{eqnarray}\label{s} {\sigma}_{N_0,N}^{-s-1}\int_{-\infty}^\infty e^{-ih (k-{\mathbb E}[S_{N_0,N}])/{\sigma}_{N_0,N}}h^s e^{-h^2/2}\\\times\left(1+\sum_{w=1}^r \frac{A_{w,N_0,N}(h)}{{\sigma}_{N_0,N}^w}+\frac{h^{r+1}}{{\sigma}_{N_0,N}^{r+1}} O(1)\right)dh.\nonumber \end{eqnarray} After that was established, the proof was completed using \eqref{Fourir} and some estimates whose whose purpose was to make the transition between the variables $k_{N_0,N}$ and $k_N$.
\section{Uniqueness of trigonometric expansions.}\label{Sec5} In several proofs we will need the following result. \begin{lemma}\label{Lemma} Let $r\geq1$ and $d\geq0$. Set ${\mathcal{R}}_0={\mathcal{R}}\cup\{0\}$ where ${\mathcal{R}}$ is the set of nonzero resonant points. For any $t_j\in{\mathcal{R}}_0$, let $A_{0,N}(t_j)$,...,$A_{d,N}(t_j)$ be sequences so that, uniformly in $k$ such that $\displaystyle k_N=\frac{k-{\mathbb E}(S_N)}{{\sigma}_N}=O(1)$ we have \[ \sum_{t_j\in{\mathcal{R}}_0}e^{-it_j k}\left(\sum_{m=0}^{d}k_N^m A_{m,N}(t_j)\right) =o({\sigma}_N^{-r}). \] Then for all $m$ and $t_j$ \begin{equation}\label{A def} A_{m,N}(t_j)=o({\sigma}_N^{-r}). \end{equation} In particular the polynomials from the definition of the (generalized) Edgeworth expansions are unique up to terms of order $o({\sigma}_N^{-r})$. \end{lemma}
\begin{proof} The proof is by induction on $d$. Let us first set $d=0$. Then, for any $k\in{\mathbb N}$ we have \begin{equation}\label{d=0} \sum_{t_j\in{\mathcal{R}}_0}e^{-it_j k}A_{0,N}(t_j)=o({\sigma}_N^{-r}). \end{equation}
Let $T$ be the number of nonzero resonant points, and let us relabel them as $\{x_1,...,x_T\}$. Consider the vector $$\mathfrak{A}_N=(A_{0,N}(0), A_{0,N}(x_1),...,A_{0,N}(x_T)).$$ Let ${\mathcal V}$ be the transpose of the Vandermonde matrix of the distinct numbers ${\alpha}_j=e^{-ix_j}, j=0,1,2,...,T$ where $x_0:=0$. Then ${\mathcal V}$ is invertible and by considering $k=0,1,2,...,T$ in (\ref{d=0}) we see that (\ref{d=0}) holds true if and only if \[ \mathfrak{A}_N={\mathcal V}^{-1}o({\sigma}_N^{-r})=o({\sigma}_N^{-r}). \] Alternatively, let $Q$ be the least common multiple of the denominators of $t_j\in{\mathcal{R}}$. Let $a_N(p)=A_{0,N}(2\pi p/Q)$ if $2\pi p/Q$ is a resonant point and $0$ otherwise. Then for $m=0,1,...,Q-1$ we have \[ \hat a_N(m):=\sum_{p=0}^{Q-1}a_N(p)e^{-2\pi pm/Q}=o({\sigma}_N^{-r}). \] Therefore, by the inversion formula of the discrete Fourier transform, \[ a_N(p)=Q^{-1}\sum_{m=0}^{Q-1} \hat a_N(m) e^{2\pi i m p/Q}=o({\sigma}_N^{-r}). \]
Assume now that the theorem is true for some $d\geq 0$ and any sequences functions $A_{0,N}(t_j),...,A_{d,N}(t_j)$. Let $A_{0,N}(t_j),...,A_{d+1,N}(t_j)$ be sequences so that uniformly in $k$ such that $\displaystyle k_N:=\frac{k-{\mathbb E}(S_N)}{{\sigma}_N}=O(1)$ we have \begin{equation}\label{I} \sum_{t_j\in{\mathcal{R}}_0}e^{-it_j k}\left(\sum_{m=0}^{d+1}k_N^m A_{m,N}(t_j)\right)=o({\sigma}_N^{-r}). \end{equation} Let us replace $k$ with $k'=k+[{\sigma}_N]Q$, where $Q$ is the least common multiply of all the denominators of the nonzero $t_j$'s. Then $e^{-it_jk}=e^{-it_j k'}$. Thus, \[ \sum_{t_j\in{\mathcal{R}}_0}e^{-it_j k}\left(\sum_{m=0}^{d+1}(k_N'^m-k_N^{m})A_{m,N}(t_j)\right)=o({\sigma}_N^{-r}). \] Set $L_N=[{\sigma}_N]Q/{\sigma}_N\thickapprox Q$. Then the LHS above equals \[ L_N\sum_{t_j\in{\mathcal{R}}_0}e^{-it_j k}\left(\sum_{s=0}^{d}k_N^s{\mathcal A}_{s,N}(t_j)\right) \] where \[ {\mathcal A}_{s,N}(t_j)=\sum_{m=s+1}^{d+1}A_{m,N}(t_j)L_N^{m-s-1}. \] By the induction hypothesis we get that \[ {\mathcal A}_{s,N}(t_j)=o({\sigma}_N^{-r}) \] for any $s=0,1,...,d$. In particular \[ {\mathcal A}_{d,N}(t_j)=A_{d+1,N}(t_j)=o({\sigma}_N^{-r}). \] Substituting this into \eqref{I} we can disregard the last term $A_{d+1,N}(t_j)$. Using the induction hypothesis with $A_{0,N}(t_j),A_{1,N}(t_j),...,A_{d,N}(t_j)$ we
obtain \eqref{A def}. \end{proof}
\section{First order expansions}\label{FirstOrder} In this section we will consider the case $r=1$. By \eqref{Cj(k)} and \eqref{s}, we see that the contribution coming from the integral over $I_j$ is \[ {\sigma}_{N_0,N}^{-1}e^{-it_j k}\Phi_{N}(t_j)\sqrt {2\pi} e^{-k_{N_0,N}^2/2}+o({\sigma}_N^{-1}) \] where $k_{N_0,N}=(k-{\mathbb E}(S_{N_0,N}))/{\sigma}_{N_0,N}$. Now, using the arguments at the end of the proof of Theorem \ref{IntIndThm} when $r=1$ we can just replace $e^{-k_{N_0,N}^2/2}$ with $e^{-(k-{\mathbb E}(S_N))^2/2V_N}$ (since it is enough to consider the case when $k_{N_0,N}$ and $k_{0,N}$ are of order $V_N^{\varepsilon}$). Therefore, taking into account that ${\sigma}_{N_0,N}^{-1}-{\sigma}_{N}^{-1}=O({\sigma}_N^{-2}N_0)$ we get that \begin{equation}\label{r=1} \sqrt 2\pi{\mathbb P}(S_N=k)= \end{equation} $$\left(1+\sum_{t_j\in {\mathcal{R}}}e^{-it_j k}\Phi_{N}(t_j)\right){\sigma}_{N}^{-1}e^{-(k-{\mathbb E}[S_N])^2/2V_N}+o(\sigma_N^{-1}). $$
Here ${\mathcal{R}}$ is the set of all nonzero resonant points $t_j=2\pi l_j/m_j$.
Indeed \sout{for} the contribution of the resonant points satisfying $M_N(m_j)\leq R(r,K)\ln V_N$ is analyzed in \S \ref{Fin}.
The contribution of the other nonzero resonant points $t$ is $o({\sigma}_N^{-1})$ due to \eqref{Roz0} in Section \ref{ScEdgeLogProkh}. In particular, \eqref{Roz0} implies that $\Phi_N(t)=o({\sigma}_N^{-1})$ so adding the points with $M_N(m_j)\geq R(r,K)\ln V_N$ only changes the sum in the RHS of \eqref{r=1} by $o(\sigma_N^{-1}). $
\begin{corollary}\label{FirstOrCor}
The local limit theorem holds if and only if $\displaystyle \max_{t\in R}|\Phi_N(t)|=o(1)$. \end{corollary}
\begin{proof} It follows from (\ref{r=1}) that the LLT holds true if and only if for any $k$ we have $$ \sum_{t_j\in {\mathcal{R}}}e^{-it_j k}\Phi_{N}(t_j)=o(1). $$ Now, the corollary follows from Lemma \ref{Lemma}. \end{proof}
Before proving Theorem \ref{ThLLT} we recall a standard fact which will also be useful in the proofs of Theorems \ref{Thm SLLT VS Ege} and \ref{Thm Stable Cond}.
\begin{lemma} \label{LmUnifFourier} Let $\{\mu_N\}$ be a sequence of measures probability measures on ${\mathbb Z}/m{\mathbb Z}$ and $\{\gamma_N\}$ be a positive sequence. Then $\mu_N(a)=\frac{1}{m}+O(\gamma_N)$ for all $a\in {\mathbb Z}/m{\mathbb Z}$ if and only iff $\hat\mu_N(b)=O(\gamma_N)$ for all $b\in \left({\mathbb Z}/m{\mathbb Z}\right)\setminus \{0\}$ where $\hat\mu$ is the Fourier transform of $\mu.$ \end{lemma}
\begin{proof} If $\mu_N(a)=\frac{1}{m}+O(\gamma_n)$ then $$ \hat\mu_N(b)=\sum_{a=0}^{m-1} \mu_N(a) e^{2\pi i ab/m}= \sum_{a=0}^{m-1} \frac{1}{m} e^{2\pi iab/m}+O(\gamma_N)=O(\gamma_N).$$ Next $\hat\mu_N(0)=1$ since $\mu_N$ are probabilities. Hence if $\hat\mu_N(b)=O(\gamma_N)$ for all $b\in \left({\mathbb Z}/m{\mathbb Z}\right)\setminus \{0\}$ then $$ \mu_N(a)=\frac{1}{m} \sum_{b=0}^{m-1} \hat\mu_N(b) e^{-2\pi i ba/m}= \frac{1}{m} \left[1+\sum_{b=1}^{m-1} \hat\mu_N(b) e^{-2\pi i ba/m}\right]=\frac{1}{m}+O(\gamma_N)$$ as claimed. \end{proof}
\begin{proof}[Proof of Theorem \ref{ThLLT}] The equivalence of conditions (b) and (c) comes from the fact that for non-resonant points the characteristic function decays faster than any power of $\sigma_N$ (see \eqref{NonResDec}).
The equivalence of (a) and (c) is due to Corollary \ref{FirstOrCor}. Finally, the equivalence between (c) and (d) comes from Lemma \ref{LmUnifFourier}. \end{proof}
\begin{remark} Theorem \ref{ThLLT} can also be deduced from \cite[Corollary 1.4]{Do}. Indeed the corollary says that either the LLT holds or there is an integer $h\in (0, 2K)$ and a bounded sequence $\{a_N\}$ such that the limit $$ {\mathbf{p}}(j)=\lim_{N\to\infty} {\mathbb P}(S_N-a_N=j \text{ mod } h) $$ exists and moreover if $k-a_n\equiv j \text{ mod }h$ then $$ \sigma_N {\mathbb P}(S_N=k)={\mathbf{p}}(j) h \mathfrak{g}\left(\frac{k-{\mathbb E}(S_N)}{\sigma_N}\right)+o\left(\sigma_N^{-1}\right).$$ Thus in the second case the LLT holds iff ${\mathbf{p}}(j)=\frac{1}{h}$ for all $j$ which is equivalent to $S_N$ being asymptotically uniformly distributed mod $h$ and also to the Fourier transform of ${\mathbf{p}}(j)$ regarded as the measure on ${\mathbb Z}/(h{\mathbb Z})$ being the $\delta$ measure at 0. Thus the conditions (a), (c) and (d) of the theorem are equivalent. Also by the results of \cite[Section 2]{Do} (see also [\S 3.3.2]\cite{DS}) if ${\mathbb E}\left(e^{i\xi S_N}\right)$ does not converge to 0 for some non zero $\xi$ then $\displaystyle \left(\frac{2\pi}{\xi} \right) {\mathbb Z}\bigcap 2\pi {\mathbb Z}$ is a lattice in ${\mathbb R}$ which implies that $\xi$ is resonant, so condition (b) of the theorem is also equivalent to the other conditions. \end{remark}
\begin{proof}[Proof of Proposition \ref{PrLLT-SLLT}.] Let $S_N$ satisfy LLT. Fix $m\in \mathbb{N}$ and suppose that $\displaystyle \sum_n q_n(m)<\infty.$ Let $s_n$ be the most likely residue of $X_n$ mod $m$. Then for $t=\frac{2\pi l}{m}$ we have $$\phi_n(t)=e^{i t s_n}-\sum_{j\not\equiv s_n\; \text{mod}\; m} {\mathbb P}(X_n\equiv j\; \text{mod m})\left(e^{its_n}-e^{itj}\right),$$
so that $1\geq |\phi_n(t)|\geq 1-2 m q_n(m).$ It follows that for each ${\varepsilon}>0$ there is
$N({\varepsilon})$ such that $\displaystyle \left|\prod_{n=N({\varepsilon})+1}^\infty \phi_n(t)\right|>1-{\varepsilon}. $ Applying this for ${\varepsilon}=\frac{1}{2}$ we have \begin{equation} \label{PhiNHalf}
\frac{1}{2}\leq\liminf_{N\to\infty} \left|\Phi_{N(1/2), N}(t)\right|\neq 0. \end{equation} On the other hand the LLT implies that \begin{equation} \label{PhiLim} \lim_{N\to\infty} \Phi_N(t)=0. \end{equation} Since $\Phi_N=\Phi_{N(1/2)} \Phi_{N(1/2), N}$, \eqref{PhiNHalf} and \eqref{PhiLim} imply that $\Phi_{N(1/2)}(t)=0.$ Since $\displaystyle \Phi_{N(1/2)}\left(\frac{2\pi l}{m}\right)=\prod_{n=1}^{N(1/2)} \phi_n\left(\frac{2\pi l}{m}\right)$ we conclude that there exists $n_l\leq N(1/2)$ such that $\phi_{n_l}(\frac{2\pi l}{m})=0.$ Hence $Y=X_{n_1}+X_{n_2}+\dots X_{n_{m-1}}$ satisfies ${\mathbb E}\left(e^{2\pi i (k/m)Y}\right)=0$ for $k=1,\dots m-1.$ By Lemma \ref{LmUnifFourier} both $Y$ and $S_N$ for $N\geq N(1/2)$ are uniformly distributed. This proves the proposition. \end{proof}
\section{Characterizations of Edgeworth expansions of all orders.} \label{ScCharacterization}
\subsection{Derivatives of the non-perturbative factor.}
Next we prove the following result. \begin{proposition}\label{Thm} Fix $r\geq1,$ and assume that $M_N\leq R(r, K)\ln{\sigma}_N$ (possibly along a subsequence). Then Edgeworth expansions of order $r$ hold true (i.e. (\ref{EdgeDef}) holds for such $N$'s) iff for each $t_j\in{\mathcal{R}}$ and $0\leq\ell<r$ (along the underlying subsequence) we have \begin{equation}\label{Cond} {\sigma}_N^{r-\ell-1}\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(\ell)}(t_j)=o(1). \end{equation} \end{proposition}
\begin{proof} First, in view of \eqref{Cj(k)} and \eqref{expo}, it is clear
that the condition (\ref{Cond}) is sufficient for expansions of order $r$.
Let us now prove that the condition (\ref{Cond}) is necessary for the expansion of order $r.$
We will use induction on $r$. For $r=1$ (see \eqref{r=1}) our expansions read \[ {\mathbb P}(S_N=k)={\sigma}_N^{-1}e^{-k_N^2/2} \left[1+\sum_{t_j\in {\mathcal{R}}}e^{-it_j k}\Phi_N(t_j)\right]+o({\sigma}_N^{-1}). \] Therefore if \[ {\mathbb P}(S_N=k)={\sigma}_N^{-1}e^{-k_N^2/2}P_N(k_N)+o({\sigma}_N^{-1}) \]
for some polynomial $P_N$
then Lemma \ref{Lemma} tells us that,
in particular $\Phi_{N}(t_j)=o(1)$ for each $t_j\in{\mathcal{R}}$.
Let us assume now that the necessity part in Proposition \ref{Thm} holds for $r'=r-1$ and prove that it holds for $r$. We will use the following lemma.
\begin{lemma}\label{LemInd} Assume that for some $t_j\in{\mathcal{R}}$, \begin{equation}\label{Ind} {\sigma}_{N}^{r-2-l}\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(l)}(t_j)=o(1),\, l=0,1,...,r-2. \end{equation}
Then, up to an $o({\sigma}_N^{-r})$ error term, the contribution of $t_j$ to the generalized Edgeworh expansions of order $r$ is \begin{equation}\label{Val0} e^{-it_j k}e^{-k_{N}^2/2}\left(\frac{\Phi_{N_0}(t_j)}{{\sigma}_N}+\sum_{q=2}^{r}\frac{\mathscr H_{N,q}(k_{N})}{{\sigma}_N^{q}}\right) \end{equation} with \begin{equation} \label{DefCH} \mathscr H_{N,q}(x)=\mathscr H_{N,q}(x;t_j)= \end{equation} $$ {\mathcal H}_{N,q,1}(x)+{\mathcal H}_{N,q,2}(x)+{\mathcal H}_{N,q,3}(x)+{\mathcal H}_{N,q,4}(x) $$ where $$ {\mathcal H}_{N, q,1}(x)= \frac{(i)^{q-1}H_{q-1}(x)\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(q-1)}(t_j)}{(q-1)!},$$ $$ {\mathcal H}_{ N,q, 2}(x)= \frac{(i)^{q-1}H_{q-1}(x)\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(q-2)}(t_j)C_{1,N,t_j}}{(q-2)!}, $$ $${\mathcal H}_{N,q,3}(x)= \frac{a_{N_0}(i)^{q-2}H'_{q-2}(x)\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(q-2)}(t_j)}{(q-2)!},$$ $$ {\mathcal H}_{N,q,4}(x)=-\frac{xa_{N_0}(i)^{q-2}H_{q-2}(x)\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(q-2)}(t_j)}{(q-2)!}, $$
and $H_q$ are Hermite polynomials.
Here $C_{1,N,t_j}$ is given by \eqref{C 1 N} when $M_N(m)\leq R(K, r)\ln{\sigma}_N$, and $C_{1,N,t_j}=0$
when $M_N(m)>R(K,r)\ln{\sigma}_N.$ (Note that in either case
$C_{1,N,t_j}=O(M_N(m))=O(\ln{\sigma}_N)$).
As a consequence, when the Edgeworth expansions of order $r$ hold true and \eqref{Ind} holds, then uniformly in $k$ so that $k_N=O(1)$ we have \begin{equation}\label{Val} \frac{\Phi_{N}(t_j)}{{\sigma}_N}+\sum_{q=2}^{r}\frac{\mathscr H_{N,q}(k_{N};t_j)}{{\sigma}_N^{q}}=o({\sigma}_N^{-r}). \end{equation} \end{lemma}
The proof of the lemma will be given in \S \ref{SSSummingUp} after we finish the proof of Proposition \ref{Thm}.
By the induction hypothesis the condition \eqref{Ind} holds true. Let us prove now that for $\ell=0,1,2,...,r-1$ and $t_j\in{\mathcal{R}}$ we have \[ \Phi_{N_0,N}(t_j)\Phi_{N_0}^{(\ell)}(t_j)=o({\sigma}_N^{-r+1+\ell}). \] Let us write \[ \frac{\Phi_{N}(t_j)}{{\sigma}_N}+\sum_{q=2}^{r}\frac{\mathscr H_{N,q}(k_{N})}{{\sigma}_N^{q}}=\sum_{m=0}^{r-1}k_N^m A_{m,N}(t_j). \]
Applying Lemmas \ref{Lemma} and \ref{LemInd} we get that \[ A_{m,N}(t_j)=o({\sigma}_N^{-r}) \] for each $0\leq m\leq r-1$ and $t_j\in{\mathcal{R}}$.
Fix $t_j\in{\mathcal{R}}$. Using Lemma \ref{LemInd} and the fact that the Hermite polynomials $H_{u}$ have the same parity as $u$ and that their leading coefficient is $1$ we have \begin{eqnarray} \label{AEq1} A_{r-1,N}(t_j)={\sigma}_N^{-r}(i)^{r-2}\Bigg(i\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-1)}(t_j)/(r-1)\\+\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-2)}(t_j)(iC_{1,N}-a_{N_0})\Bigg)=o({\sigma}_N^{-r}) \notag \end{eqnarray} and \begin{eqnarray} \label{AEq2} \quad A_{r-2,N}(t_j)={\sigma}_N^{-r+1}(i)^{r-3}\Bigg(i\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-2)}(t_j)/(r-2)\\+\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-3)}(t_j)(iC_{1,N}-a_{N_0})\Bigg)=o({\sigma}_N^{-r}). \notag \end{eqnarray} Since $\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-3)}(t_j)=o({\sigma}_N^{-1})$,
\eqref{AEq2} yields \[ \Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-2)}(t_j)=o({\sigma}_N^{-1}\ln{\sigma}_N). \] Plugging this into \eqref{AEq1} we get \[ \Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-1)}(t_j)=o(1). \] Therefore we can just disregard $\mathscr H_{r,N}(k_N;t_j)$ since
its coefficients are of order $o({\sigma}_N^{-r})$. Since the term ${\mathcal H}_{r,N}(k_N;t_j)$ no longer appears, repeating the above arguments with $r-1$ in place of $r$ we have \begin{eqnarray*} A_{r-3,N}(t_j)={\sigma}_N^{-r+2}(i)^{r-4}\Bigg(i\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-3)}(t_j)/(r-3)\\+\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-4)}(t_j)(iC_{1,N}-a_{N_0})\Bigg)=o({\sigma}_N^{-r}). \end{eqnarray*} Since $\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-4)}(t_j)=o({\sigma}_N^{-2})$, the above asymptotic equality yields that \[ \Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-3)}(t_j)=o({\sigma}_N^{-2}\ln{\sigma}_N). \] Plugging this into \eqref{AEq2} we get \[ \Phi_{N_0,N}(t_j)\Phi_{N_0}^{(r-2)}(t_j)=o({\sigma}_N^{-1}). \] Hence, we can disregard also the term $\mathscr H_{r-1,N}(k_N;t_j)$. Proceeding this way we get that $\displaystyle \Phi_{N_0,N}(t_j)\Phi_{N_0}^{(\ell)}(t_j)=o({\sigma}_N^{\ell+1-r}) $ for any $0\leq \ell<r$. \qed
Before proving Lemma \ref{LemInd}, let us state the following result, which is a consequence of Proposition \ref{Thm} and \eqref{Roz0}. \begin{corollary}\label{CorNoDer}
Suppose that for each nonzero resonant point $t$ we have $\displaystyle \inf_{n}|\phi_n(t)|>0$. Then for any $r$, the sequence $S_N$ obeys Edgeworth expansions of order $r$ if and only if $\Phi_N(t)=o({\sigma}_N^{1-r})$ for each nonzero resonant point $t$. \end{corollary}
\subsection{Proof of Lemma \ref{LemInd}.} \label{SSSummingUp} \begin{proof} First, because of \eqref{Ind},
for each $0\leq s\leq r-1,$ the terms indexed by $l<s-1$ in \eqref{Cj(k)},
are of order $o({\sigma}_N^{-r})$
and so they can be disregarded. Therefore, we need only to consider the terms indexed by $l=s$ and $l=s-1$. For such $l$, using again \eqref{Ind} we can disregard all the terms in \eqref{s} indexed by $w\geq1$, since the resulting terms are of order $o({\sigma}_N^{-w-r}\ln{\sigma}_N)=o({\sigma}_N^{-r})$.
Now, since ${\sigma}_{N_0,N}^{-1}-{\sigma}_{N}^{-1}=O(V_{N_0}/{\sigma}_N^3)$ we can
replace ${\sigma}_{N_0,N}$ with ${\sigma}_{N}^{-1}$
in \eqref{Cj(k)}, as the remaining terms are of order $o({\sigma}_N^{-r-1})$. Therefore, using \eqref{Fourir} we get the following contributions from $t_j\in{\mathcal{R}}$, \begin{equation*} e^{-it_j k}\; e^{-k_{N_0,N}^2/2}\left(\frac{\Phi_{N}(t_j)}{{\sigma}_N}+\sum_{q=2}^{r}\frac{{\mathcal H}_{N,q}(k_{N_0,N})}{{\sigma}_N^{q}}\right) \end{equation*}
where ${\mathcal H}_{N,q}(x)={\mathcal H}_{N,q,1}(x)+{\mathcal H}_{N,q,2}(x)$ and ${\mathcal H}_{N,q,j}, j=1,2$
are defined after \eqref{DefCH}.
Note that when $x=O(1)$ and $q<r$, \begin{equation}\label{Order} \frac{{\mathcal H}_{N,q,1}(x)}{{\sigma}_N^q}=o({\sigma}_N^{-r+1})\,\,\text { and }\,\,\frac{{\mathcal H}_{N,q,2}(x)}{{\sigma}_N^q}=o({\sigma}_N^{-r}\ln{\sigma}_N). \end{equation} while when $q=r$, \begin{equation}\label{Order.1} \frac{{\mathcal H}_{N,r,1}(x)}{{\sigma}_N^r}=O({\sigma}_N^{-r}\ln{\sigma}_N)\,\,\text { and }\,\,\frac{{\mathcal H}_{N,r,2}(x)}{{\sigma}_N^r}=o({\sigma}_N^{-r}\ln{\sigma}_N). \end{equation} Next \[ k_{N_0,N}=(1+\rho_{N_0,N})k_N+\frac{a_{N_0}}{{\sigma}_N}+\theta_{N_0,N} \] where $\rho_{N_0,N}={\sigma}_{N}/{\sigma}_{N_0,N}-1=O(\ln{\sigma}_N/{\sigma}^2_N)$ and $$\theta_{N_0,N}=a_{N_0}\left(\frac{1}{{\sigma}_{N_0,N}}-\frac{1}{{\sigma}_N}\right)=O(\ln^2{\sigma}_N/{\sigma}_N^3).$$ Hence,
when $|k_{N_0,N}|\leq{\sigma}_N^{{\varepsilon}}$ (and so $\displaystyle k_N=O({\sigma}_N^{{\varepsilon}})$) for some ${\varepsilon}>0$ small enough then for each $m\geq 1$ we have \[ k_{N_0,N}^m=k_{N}^m+mk_{N}^{m-1}a_{N_0}/{\sigma}_N+o({\sigma}_N^{-1}). \] Therefore, \eqref{Order} and \eqref{Order.1} show that
upon replacing $H_{q-1}(k_{N_0,N})$ with $H_{q-1}(k_{N})$
the only additional term is \[ \frac{a_{N_0}(i)^{q-1}H'_{q-1}(k_N)\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(q-1)}(t_j)}{(q-1)!{\sigma}_N^{q+1}} \] for $q=2,3,...,r-1$. We thus get that the contribution of $t_j$ is \begin{equation*} e^{-it_j k} \;e^{-k_{N_0,N}^2/2}\left(\frac{\Phi_{N}(t_j)}{{\sigma}_N}+\sum_{q=2}^{r}\frac{{\mathcal C}_{N,q}(k_{N})}{{\sigma}_N^{q}}\right) \end{equation*} where \begin{equation*} {\mathcal C}_{N,q}(x)={\mathcal H}_{N,q}(x)+ \frac{a_{N_0}(i)^{q-2}H'_{q-2}(x)\Phi_{N_0,N}(t_j) \Phi_{N_0}^{(q-2)}(t_j)}{(q-2)!}. \end{equation*} Note that ${\mathcal C}_{N,2}(\cdot)={\mathcal H}_{N,2}(\cdot)$. Finally, we can replace $e^{-k_{N_0,N}^2/2}$ with \[ (1-k_Na_{N_0}/{\sigma}_N)e^{-k_{N}^2/2} \] since all other terms in the transition between $e^{-k_{N_0,N}^2/2}$ to $e^{-k_{N}^2/2}$ are of order $o({\sigma}_N^{-1})$ (see (\ref{ExpTrans1}) and (\ref{ExTrans1.1})). The term $-k_Na_{N_0}/{\sigma}_N$ shifts the $u$-th order term to the $u+1$-th term, $u=1,2,...,r-1$ multiplied by $-k_Na_{N_0}$.
Next, relying on \eqref{Order} and \eqref{Order.1} we see that after multiplied by $k_Na_{N_0}/{\sigma}_N$, the second term ${\mathcal H}_{N,q,2}(k_N)$ from the definition of ${\mathcal H}_{N,q}(k_N)$ is of order $o({\sigma}_N^{-r-1}\ln^2{\sigma}_N){\sigma}_N^q$ and so
this product can be disregarded.
Similarly, we can ignore the additional contribution coming from multiplying the second term from the definition of ${\mathcal C}_{N,q}(k_N)$ by $-k_Na_{N_0}/{\sigma}_N$ (since this term is of order $o({\sigma}_N^{-r}\ln{\sigma}_N){\sigma}_N^q$). We conclude that, up to a term of order $o({\sigma}_N^{-r})$, the total contribution of $t_j$ is \begin{equation*} e^{-it_j k} \; e^{-k_{N}^2/2}\left(\frac{\Phi_{N}(t_j)}{{\sigma}_N}+\sum_{q=2}^{r}\frac{\mathscr H_{N,q}(k_{N}; t_j)}{{\sigma}_N^{q}}\right) \end{equation*} where $\displaystyle
\mathscr H_{N,q}(x;t_j)={\mathcal C}_{N,q}(x)-\frac{x a_{N_0}(i)^{q-2}H_{q-2}(x) \Phi_{N_0,N}(t_j)\Phi_{N_0}^{(q-2)}(t_j)}{(q-2)!} $ which completes the proof of \eqref{Val0}.
Next we prove \eqref{Val}.
On the one hand, by assumption we have Edgeworth expansions or order $r$, and, on the other hand, we have the expansions from Theorem \ref{IntIndThm}. Therefore, the difference between the two must be $o({\sigma}_N^{-r})$. Since the usual Edgeworth expansions contain no terms corresponding to nonzero resonant points, applying Lemma \ref{Lemma} and \eqref{Val0} we obtain \eqref{Val}. \end{proof}
Note that the formulas of Lemma \ref{LemInd} together with already proven Proposition \ref{Thm} give the following result.
\begin{corollary} \label{CrFirstNonEdge} Suppose that $\mathbb{E}(S_N)$ is bounded, $S_N$ admits the Edgeworth expansion of order $r-1$, and, either
(a) for some ${\bar\varepsilon}\leq 1/(8K)$ we have $N_0=N_0(N,t_j,{\bar\varepsilon})=0$ for each each nonzero resonant point $t_j$,
or (b) $\displaystyle \varphi:=\min_{t\in{\mathcal{R}}}\inf_{n}|\phi_n(t)|>0$.
Then $$ \sqrt{2\pi} \mathbb{P}(S_N=k)$$ $$=e^{-k_N^2/2} \left[{\mathcal E}_r(k_N)+ \sum_{t_j\in{\mathcal{R}}}\left(\frac{\Phi_N(t_j)}{{\sigma}_N}+\frac{ik_N C_{1,N,t_j}\Phi_N(t_j)}{{\sigma}_N^2}\right) e^{-i t_j k}\right]+o(\sigma_N^{-r})$$ where ${\mathcal E}_r(\cdot)$ is the Edgeworth polynomial of order $r$ (i.e. the contribution of $t=0$), and we recall that $$iC_{1,N,t_j}=-\sum_{n=1}^N\frac{{\mathbb E}(e^{it_j X_n}\bar X_n)}{{\mathbb E}(e^{it_j X_n})}.$$ \end{corollary}
\begin{proof}
Part (a) holds since under the assumption that $N_0=0$ all terms ${\mathcal H}_{N,q,j}$ in \eqref{DefCH} except ${\mathcal H}_{N, 2, 2}$ vanish.
Part (b) holds since in this case the argument proceeds similarly to the proof of Theorem \ref{IntIndThm}
if we set $N_0=0$ for any $t_j$ (since we only needed $N_0$ to obtain a positive lower bound on $|\phi_n(t_j)|$ for $t_j\in{\mathcal{R}}$ and $N_0<n\leq N$). \end{proof}
\begin{remark}
Observe that $\displaystyle {\sigma}_N^{-1}\gg |C_{1, N, {t_j}}|{\sigma}_N^{-2},$
so if the conditions of the corollary are satisfied but $|\Phi_N(t_j)|\leq c {\sigma}_N^{1-r}$ (possibly along a subsequence), then the leading correction to the Edgeworth expansion comes from $$ e^{-k_N^2/2} \sum_{t_j\in{\mathcal{R}}}\left(\frac{\Phi_N(t_j)}{{\sigma}_N} \right).$$ Thus Corollary \ref{CrFirstNonEdge} strengthens Corollary \ref{CorNoDer} by computing the leading correction to the Edgeworth expansion when the expansion does not hold. \end{remark}
\subsection{Proof of Theorem \ref{r Char}} We will use the following. \begin{lemma}\label{Lem}
Let $t_j$ be a nonzero resonant point, $r>1$ and suppose that $M_N\leq R\ln{\sigma}_N$, $R=R(r,K)$ and that $|{\mathbb E}(S_N)|=O(\ln{\sigma}_N)$. Then (\ref{Cond}) holds for all $0\leq \ell<r$ if and only if \begin{equation}\label{CondDer}
|\Phi_{N}^{(\ell)}(t_j)|=o\left({\sigma}_N^{-r+\ell+1}\right) \end{equation} for all $0\leq \ell<r$.
\end{lemma} \begin{proof} Let us first assume that (\ref{Cond}) holds. Recall that \begin{equation} \label{PhiNProduct} \Phi_N(t)=\Phi_{N_0}(t) \Phi_{N_0, N}(t) \end{equation} with \begin{equation} \label{TripleProduct} \Phi_{N_0, N}(t)=\Phi_{N_0, N} (t_j)\Phi_{N_0, N}(h) \Psi_{N_0, N}(h) \end{equation}
where $t=t_j+h$ and \begin{equation} \label{TripleFactorization}
\Psi_{N_0, N}(h)=\exp\left[O(M_N(m))\sum_{u=1}^\infty (O(1))^u h^u\right]. \end{equation}
For $\ell=0$ the result reduces to \eqref{PhiNProduct}. For larger $\ell$'s we have \begin{equation}\label{Relation} \Phi_{N}^{(\ell)}(t_j)=\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(\ell)}(t_j)+\sum_{k=0}^{\ell-1}\binom{\ell}{k}\Phi_{N_0,N}^{(\ell-k)}(t_j)\Phi_{N_0}^{(k)}(t_j). \end{equation} Fix some $k<\ell$. Then by \eqref{TripleProduct}, \begin{eqnarray*} \Phi_{N_0, N}^{(\ell-k)}(t_j)=\Phi_{N_0,N}(t_j)\sum_{u=0}^{\ell-k}\binom{\ell-k}{u}\Phi_{N_0,N}^{(u)}(0)\Psi_{N_0,N}^{(\ell-k-u)}(0)\\=O(\ln^{\ell-k}{\sigma}_N)\Phi_{N_0,N}(t_j) \end{eqnarray*} where we have used that $\bar S_{N_0,N}=S_{N_0,N}-{\mathbb E}(S_{N_0,N})$ satisfies
$|{\mathbb E}[(\bar S_{N_0,N})^q]|\leq C_q{\sigma}_{N_0,N}^{2q}$, (see \eqref{CenterMoments}). Therefore \begin{equation}\label{Thereofre} \Phi_{N_0,N}^{(\ell-k)}(t_j)\Phi_{N_0}^{(k)}(t_j)=O(\ln^{\ell-k}{\sigma}_N)\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(k)}(t_j). \end{equation} Finally, by (\ref{Cond}) we have \[ \Phi_{N_0,N}(t_j)\Phi_{N_0}^{(k)}(t_j)=o({\sigma}_N^{k+1-r}) \] and so, since $k<\ell$, \[ \Phi_{N_0}^{(\ell-k)}(t_j)\Phi_{N_0}^{(k)}(t_j)=o({\sigma}_N^{\ell+1-r}). \] This completes the proof that \eqref{CondDer} holds.
Next, suppose that (\ref{CondDer}) holds for each $0\leq\ell<r$. Let use prove by induction on $\ell$ that \begin{equation}\label{Above}
|\Phi_{N_0,N}\Phi_{N_0}^{(\ell)}(t_j)|=o\left({\sigma}_N^{-r+\ell+1}\right). \end{equation} For $\ell=1$ this follows from \eqref{Relation}.
Now take $\ell>1$ and assume that \eqref{Above} holds with $k$ in place of $\ell$ for each $k<\ell$. Then by (\ref{Relation}), (\ref{Thereofre}) and the induction hypothesis we get that \[ \Phi_{N}^{(\ell)}(t_j)=\Phi_{N_0,N}(t_j)\Phi_{N_0}^{(\ell)}(t_j)+o({\sigma}_N^{\ell+1-r}). \] By assumption we have $\Phi_{N}^{(\ell)}(t_j)=o({\sigma}_N^{\ell+1-r})$ and hence \[ \Phi_{N_0,N}(t_j)\Phi_{N_0}^{(\ell)}(t_j)=o({\sigma}_N^{\ell+1-r}) \] as claimed. \end{proof}
Theorem \ref{r Char} in the case $M_N\leq R\ln \sigma_N$ follows now by first replacing $X_n$ with $X_n-c_n$, where $(c_n)$ is a bounded sequence of integers so that ${\mathbb E}[S_N-C_N]=O(1)$, where \begin{equation} \label{CNInteger}
C_N=\sum_{j=1}^n c_j \end{equation}
(see Lemma 3.4 in \cite{DS}), and then applying Lemma~\ref{Lem} and Proposition~\ref{Thm}.
It remains to consider the case when $M_N(m)\geq {\bar R} \ln \sigma_N$ where ${\bar R}$ is large enough. In that case, by Theorem \ref{ThEdgeMN}, the Edgeworth expansion of order $r$ hold true, and so, after the reduction to the case when ${\mathbb E}(S_N)$ is bounded, it is enough to show that $\displaystyle \Phi_N^{(\ell)}(t_j)=o\left(\sigma_N^{-r+\ell+1}\right)$ for all $0\leq \ell<r.$ By the arguments of Lemma \ref{Lem} it suffices to show that for each $0\leq \ell<r$ we have $\Phi_{N_0}^{(\ell)}(t_j)\Phi_{N_0, N}(t_j) =o(\sigma_N^{-r})$. To this end we write $$\Phi_{N_0}^{(\ell)}(t_j)\Phi_{N_0, N}(t_j) =\sum_{n_1,\dots, n_k\leq N_0; \atop \ell_1+\dots+\ell_k=\ell} \gamma_{\ell_1,\dots, \ell_k} \left(\prod_{q=1}^k \phi_{n_q}^{(\ell_q)}(t_j) \right) \left[\prod_{n\leq N, \; n\neq n_k} \phi_n(t_j)\right]$$ where $\gamma_{\ell_1,\dots, \ell_k}$ are bounded coefficients of combinatorial nature.
Using (\ref{Roz0}) we see that
for each $n_1,\dots, n_k$ the product in the square brackets
is at most $C e^{-c_0 M_N(m)+O(1)}$ for some $C, c_0>0$. Hence
$$|\Phi_{N_0}^{(\ell)}(t_j)\Phi_{N_0, N}(t_j)|\leq \hat{C} N_0^{\ell} \; e^{-c M_N(m)}.$$ It remains to observe that the definition of $N_0$ gives $M_N(m)\geq \hat{\varepsilon} N_0.$ Therefore
$\displaystyle |\Phi_{N_0}^{(\ell)}(t_j)\Phi_{N_0, N}(t_j)|\leq C^* M_N^\ell (m) \; e^{-c M_N(m)}=o(\sigma_N^{-r})$ provided that $M_N\geq {\bar R}\ln \sigma_N$ for ${\bar R}$ large enough. \qed
\section{Edgeworth expansions and uniform distribution.}
\subsection{Proof of Theorem \ref{Thm SLLT VS Ege}} \label{SSEdgeR=2} In view of Proposition \ref{Thm} with $r=2$, it is enough to show that if $\Phi_{N}(t_j)=o(\sigma_N^{-1})$ then the SLLT implies that \begin{equation} \label{C11-C02}
|\Phi_{N_0,N}(t_j)\Phi_{N_0}'(t_j)|=o(1) \end{equation}
for any non-zero resonant point $t_j$ (note that the equivalence of conditions (b) and (c)
of the theorem follows from Lemma \ref{LmUnifFourier}).
Denote $\displaystyle \Phi_{N; k}(t)=\prod_{l\neq k, l\leq N} \phi_l(t)$.
Let us first assume that $\phi_k(t_j)\not=0$ for all $1\leq k\leq N$. Then $\phi_k'(t_j) \Phi_{N; k}(t_j)=\phi_k'(t_j)\Phi_N(t_j)/\phi_k(t_j)$. Let ${\varepsilon}_N=\frac{\ln{\sigma}_N}{{\sigma}_N}$. If for all
$1\leq k\leq N_0$ we have $|\phi_k(t_j)|\geq {\varepsilon}_N$ then $$
\left|\Phi_{N_0,N}(t_j)\Phi_{N_0}'(t_j)\right|=\left|\sum_{k=1}^{N_0}\phi_k'(t_j) \Phi_{N; k}(t_j)\right|\leq|\Phi_N(t_j)|\sum_{k=1}^{N_0}|\phi'_k(t_j)/\phi_k(t_j)| $$
$$ \leq C{\varepsilon}_N^{-1} N_0 |\Phi_N(t_j)|
\leq C'{\sigma}_N|\Phi_N(t_j)|\to 0 \text{ as }N\to\infty $$ where we have used that $N_0=O(\ln V_N)$.
Next suppose there is at least one $1\leq k\leq N_0$ such that
$|\phi_k(t_j)|<{\varepsilon}_N$. Let us pick some $k=k_N$ with the latter property. Then for any $k\not=k_N$, $1\leq k\leq N_0$ we have \[
|\phi_k'(t_j) \Phi_{N; k}(t_j)|\leq C|\phi_{k_N}(t_j)|<C{\varepsilon}_N. \] Therefore, \[
\left|\sum_{k\not=k_N,\,1\leq k\leq N_0}\phi_k'(t_j) \Phi_{N; k}(t_j)\right|\leq \frac{C'\ln^2 {\sigma}_N}{{\sigma}_N}=o(1). \] It follows that \begin{equation} \label{SingleTerm}
\Phi_{N_0,N}(t_j)\Phi_{N_0}'(t_j)=\Phi_{N; k_N}(t_j) \phi_{k_N}'(t_j)+o(1).
\end{equation}
Next, in the case when $\phi_{k_0}(t_j)=0$ for some $1\leq k_0\leq N_0$, then \eqref{SingleTerm} clearly holds true with $k_N=k_0$ since all the other terms vanish.
In summary, either \eqref{C11-C02}
holds or we have \eqref{SingleTerm}. In the later case, using (\ref{Roz0}) we obtain \begin{equation} \label{PhiKBound}
\left|\mathbb{E}\left(e^{i t_j S_{N; k_N}}\right)\right|\leq e^{-c_2\sum_{s\not=k_N, 1\leq s\leq N}q_s(m)}=e^{-c_2M_{N}(m)-q_{k_N}(m)} \end{equation} where $S_{N; k}=S_N-X_k$, and $c_2>0$ depends only on $K$. Since the SLLT holds true, $M_{N}$ converges to $\infty$ as $N\to\infty$. Taking into account that $0\leq q_{k_N}(m)\leq1$ we get that the left hand side of \eqref{PhiKBound} converges to $0$, proving \eqref{C11-C02}. \end{proof}
\subsection{Proof of Theorem \ref{Thm Stable Cond}} We start with the proof of part (1). Assume that the LLT holds true in a superstable way. Let $X_1',X_2',...$ be a square integrable integer-valued independent sequence which differs from $X_1,X_2,...$ by a finite number of elements. Then there is $n_0\in{\mathbb N}$ so that $X_n=X'_n$ for any $n>n_0$. Set $\displaystyle S_N'=\sum_{n=1}^N X'_n$, $Y=S'_{n_0}$ and
$Y_N=Y{\mathbb I}(|Y|<{\sigma}_N^{1/2+{\varepsilon}})$, where ${\varepsilon}>0$ is a small constant. By the Markov inequality we have $$
{\mathbb P}(|Y|\geq {\sigma}_N^{1/2+{\varepsilon}})={\mathbb P}(|Y|^2\geq {\sigma}_N^{1+2{\varepsilon}})\leq\|Y\|_{L^2}^2{\sigma}_N^{-1-2{\varepsilon}}=o({\sigma}_N^{-1}). $$ Therefore, for any $k\in{\mathbb N}$ and $N>n_0$ we have \begin{eqnarray*}
{\mathbb P}(S'_N=k)={\mathbb P}(S_{N;1,2,...,n_0}+Y_N=k)+o({\sigma}_N^{-1})\\={\mathbb E}[{\mathbb P}(S_{N;1,2,...,n_0}=k-Y_N|X_1',...,X_{n_0}')]+o({\sigma}_N^{-1})\\= {\mathbb E}[P_{N:n_1,...,n_0}(k-Y_N)]+o({\sigma}_N^{-1}) \end{eqnarray*} where $P_{N:n_1,...,n_0}(s)={\mathbb P}(S_{N;1,2,...,n_0}=s)$ for any $s\in{\mathbb Z}$. Since the LLT holds true in a superstable way, we have, uniformly in $k$ and the realizations of $X_1',...,X_{n_0}'$ that $$ P_{N:n_1,...,n_0}(k-Y_N)=\frac{e^{-(k-Y_N-{\mathbb E}(S_N))^2/(2V_N)}}{\sqrt{2\pi}{\sigma}_N}+o({\sigma}_N^{-1}). $$ Therefore, \begin{equation}\label{Above1} {\mathbb P}(S'_N=k)= \end{equation} $$ \frac{e^{-(k-{\mathbb E}(S_N))^2/2V_N}}{\sqrt{2\pi}{\sigma}_N} {\mathbb E}\big(e^{-(k-{\mathbb E}(S_N))Y_N/V_N+Y_N^2/(2V_N)}\big)+o({\sigma}_N^{-1}). $$
Next, since $|Y_N|\leq {\sigma}_N^{1/2+{\varepsilon}}$ we have that $\|Y_N^2/(2V_N)\|_{L^\infty}\leq {\sigma}_{N}^{2{\varepsilon}-1}$, and so when ${\varepsilon}<1/2$ we have $\|Y_N^2/2V_N\|_{L^\infty}=o(1)$.
Recall that $k_N=(k-{\mathbb E}(S_N))/{\sigma}_N$. Suppose first that
$|k_N|\geq {\sigma}_N^{{\varepsilon}}$ with ${\varepsilon}<1/4.$
Since
$$\big|(k-{\mathbb E}(S_N))Y_N/V_N\big|\leq |k_N|{\sigma}_N^{{\varepsilon}-\frac12},$$ we get that the RHS of \eqref{Above1} is $o({\sigma}_N^{-1})$ (uniformly in such $k$'s).
On the other hand, if $|k_N|<{\sigma}_N^{{\varepsilon}}$ then $${\mathbb E}\big(e^{-(k-{\mathbb E}(S_N))Y_N/V_N+Y_N^2/2V_N}\big)=1+o(1)$$ (uniformly in that range of $k$'s).
We conclude that, uniformly in $k$, we have $$ {\mathbb P}(S'_N=k)=\frac{e^{-(k-{\mathbb E}(S_N))^2/(2V_N)}}{\sqrt{2\pi}{\sigma}_N}+o({\sigma}_N^{-1}). $$
Lastly, since $\displaystyle \sup_{N}|{\mathbb E}(S_N)-{\mathbb E}(S_N')|\!\!<\!\infty$ and
$\displaystyle \sup_{N}|\text{Var}(S_N)-\text{Var}(S_N')|\!<\!\!\infty,$ $$ {\mathbb P}(S'_N=k)=\frac{e^{-(k-{\mathbb E}(S'_N))^2/(2V'_N)}}{\sqrt{2\pi}{\sigma}'_N}+o(1/{\sigma}_N') $$ where $V_N'=\text{Var}(S_N')$ and ${\sigma}_N'=\sqrt{V_N'}$.
Conversely, if the SLLT holds then $M_N(h)\to \infty$ for each $h\geq2$. Now if $t$ is a nonzero resonant point with denominator $h$ then \eqref{Roz0} gives
$$ |\Phi_{N; j_1^{N}, j_2^{N},\dots ,j_{s_N}^{N}}(t)|\leq C e^{-c M_N(h)+\bar C {\bar s}},\, \,C,\bar C>0 $$ for any choice of $j_1^N,...,j_{s_N}^N$ and $\bar s$ with $s_N\leq \bar s$. Since the RHS tends to 0 as $N\to\infty$, $\{X_n\}\in EeSS(1)$ completing the proof of part (1).
For part (2) we only need to show that (a) is equivalent to (b) as the equivalence of (b) and (c) comes from Lemma \ref{LmUnifFourier}. By replacing again $X_n$ with $X_n-c_n$ it is enough to prove the equivalency in the case when ${\mathbb E}(S_N)=O(1)$. The proof that (a) and (b) are equivalent consists of two parts. The first part is the following statement whose proof is a straightforward adaptation of the proof of Theorem \ref{r Char} and is therefore omitted.
\begin{proposition} $\{X_n\}\in SsEe(r)$ if and only if for each ${\bar s}$, each sequence $j_1^N, j_2^N, \dots ,j_{s_N}^N$ with $s_N\leq {\bar s}$, each $\ell<r$ and each $t\in{\mathcal{R}}$ we have \begin{equation} \label{PhiDerFM} \Phi^{(\ell)}_{N; j_1^N, j_2^N, \dots, j_{s_N}^N}(t)= o(\sigma_N^{\ell+1-r}). \end{equation} \end{proposition}
Note that the above proposition shows that the condition $\displaystyle \Phi_{N; j_1^N, j_2^N, \dots ,j_{s_N}^N}(t)= o(\sigma^{1-r}_{N})$ is necessary.
The second part of the argument is to show that if $$ \Phi_{N; j_1^N, j_2^N, \dots ,j_{s_N}^N}(t)= o(\sigma^{1-r}_{N}) $$ holds for every finite modification of $S_N$ with $s_N\leq {\bar s}+\ell$ (uniformly) then \eqref{PhiDerFM} holds for every modifications with $s_N\leq {\bar s}$ so that the condition $\displaystyle \Phi_{N; j_1^N, j_2^N, \dots ,j_{s_N}^N}(t)= o(\sigma^{1-r}_{N})$ is also sufficient.
To this end we introduce some notation. Fix a nonzero resonant point $t=\frac{2\pi l}{m}.$ Let
$\check\Phi_{N}$ be the characteristic function of the sum $\check S_{N}$ of all $X_n$'s
such that $1\leq n\leq N$, $n\not\in\{j_1^{N},j_2^{N},...,j_{s_N}^N\}$ and $q_{n}(m)\geq\bar\epsilon$. Let $\check N$ be the number of terms in $\check S_N.$ Denote
$\tilde S_{N}=S_{N;j_1^N,j_2^N,\dots,j_{s_N}^{N}}-\check S_{N}$
and let $\tilde\Phi_N(t)$ be the characteristic function of $\tilde S_N.$
Similarly to the proof of Theorem \ref{r Char} it suffices to show
that for each $\ell<r$
$$ \left| \check \Phi_N^{(\ell)} \tilde \Phi_N(t)\right|=o(\sigma_N^{1+\ell-r}) $$ and, moreover, we can assume that $M_N(m)\leq {\bar R} \ln \sigma_N$ so that $\check N=O(\ln \sigma_N).$ We have (cf. \eqref{DerComb}) , $$ \check \Phi_N^{(\ell)} \tilde \Phi_N(t) =\sum_{\substack{n_1, \dots ,n_k; \;\\ \ell_1+\dots+\ell_k=\ell}} \prod_{w=1}^k \gamma_{\ell_1,\dots, \ell_k} \phi_{n_w}^{(\ell_w)}(t_j) \prod_{n\not\in\{n_1,n_2 \dots ,n_k, j_1^N,j_2^N \dots, j_{s_N}^N\}}
\phi_n(t_j)$$
where the summation is over all tuples $n_1,n_2, \dots ,n_k$ such that
$q_{n_w}(m)\geq\bar\epsilon$.
Note that the absolute value of each term in the above sum is bounded by
$C |\Phi_{N; n_1, \dots ,n_k, j_1\dots, j_{s_N}^N}(t_j)|=o(\sigma_N^{1-r}).$
It follows that the whole sum is $$ o\left(\sigma_N^{1-r} \check N^\ell\right) = o\left(\sigma_N^{1-r} \ln^\ell \sigma_N \right) $$ completing the proof. \qed
\begin{remark} Lemma \ref{LmUnifFourier} and Theorem \ref{r Char} show that the convergence to uniform distribution on any factor $\mathbb{Z}/h \mathbb{Z}$ with the speed $o({\sigma}_N^{1-r})$ is necessary for Edgeworth expansion of order $r.$ This is quite intuitive. Indeed calling $\mathscr E_r$
the Edgeworth function of order $r$, (i.e. the contribution from zero), then it is a standard result from numerical integration (see, for instance, \cite[Lemma A.2]{DNP}) that for each $s\in \mathbb{N}$ and each $j\in \mathbb{Z}$ $$ \sum_{k\in \mathbb{Z}} h \mathscr E_r\left(\frac{j+hk}{\sqrt{{\sigma}_N}}\right) =\int_{-\infty}^\infty \mathscr E_r(x) dx+o\left({\sigma}_N^{-s}\right)= 1+o\left({\sigma}_N^{-s}\right) $$ where in the last inequality we have used that the non-constant Hermite polynomials have zero mean with respect to the standard normal law (since they are orthogonal to the constant functions). However, using this result to show that $$\displaystyle \sum_{k\in \mathbb{Z}} \mathbb{P}(S_N=j+kh)=\frac{1}{h}+o\left({\sigma}_N^{1-r}\right) $$ requires a good control on large values of $k.$ While it appears possible to obtain such control using the large deviations theory it seems simpler to
estimate the convergence rate towards uniform distribution from our generalized Edgeworth expansion. \end{remark}
\section{Second order expansions}\label{Sec2nd} In this section we will compute the polynomials in the general expansions in the case $r=2$.
First, let us introduce some notations which depend on a resonant point $t_j.$ Let $t_j=2\pi l_j/m_j$ be a nonzero resonant point such that $M_N(m_j)\leq R(2,K)\ln V_N$ where $R(2,K)$ is specified in Remark \ref{R choice}.
Let
$\check\Phi_{j,N}$ be the characteristic function of the sum $\check S_{j,N}$ of all $X_n$'s
such that $1\leq n\leq N$ and $q_{n}(m_j)\geq\bar\epsilon=\frac1{8K}$.
Note that $\check S_{j,N}$ was previously denoted by $S_{N_0}$. Let $\tilde S_{N,j}=S_N-\check S_{N,j}$ and denote by $\tilde \Phi_{N,j}$ its characteristic function. (In previous sections we denoted the same expression by $S_{N_0,N}$, but here we want to emphasize the dependence on $t_j$.)
Let $\gamma_{N,j}$ be the ratio between the third moment of $\tilde S_{N,j}-{\mathbb E}(\tilde S_{N,j})$ and its variance. Recall that by \eqref{CenterMoments}
$|\gamma_{N,j}|\leq C$ for some $C$. Also, let $C_{1,N,t_j}$ be given by (\ref{C 1 N}), with the indexes rearranged so that the $n$'s with $q_n(m)\geq\bar{\varepsilon}$ are the first $N_0$ ones ($C_{1,N,t_j}$ is at most of order $M_N(m)=O(\ln V_{N})$). For the sake of convenience, when either $t_j=0$ or $M_N(m_j)\geq R(2,K)\ln V_N$ we set $C_{1,N,t_j}=0$, $\tilde S_{N,j}=S_N$ and $\check S_{N,j}=0$. In this case $\tilde\Phi_{N,j}=\tilde\Phi_{N}$ and $\check\Phi_{N,j}\equiv 1$. Also denote $k_N=(k-{\mathbb E}(S_N))/{\sigma}_N,$ $\bar S_N=S_N-{\mathbb E}(S_N)$, and $\gamma_N={\mathbb E}(\bar S_N^3)/V_N$, ($\gamma_N$ is bounded).
\begin{proposition}\label{2 Prop} Uniformly in $k$, we have \begin{eqnarray}\label{r=2'} \sqrt {2\pi}{\mathbb P}(S_N=k)=\Big(1+\sum_{t_j\in {\mathcal{R}}}e^{-it_j k}\Phi_{N}(t_j)\Big){\sigma}_{N}^{-1}e^{-k_N^2/2}\\-{\sigma}_{N}^{-2}e^{-k_N^2/2}\left(\gamma_N k_N^3/6+\sum_{t_j\in {\mathcal{R}}}e^{-it_j k}\tilde\Phi_{N,j}(t_j)P_{N,j}(k_N)\right)\nonumber\\+o(\sigma_N^{-2})\nonumber \end{eqnarray} where $$ P_{N,j}(x)=\big(\check\Phi_{N,j}(t_j)(iC_{1,N,t_j}-{\mathbb E}(\check S_{N,j}))+i\check\Phi_{N,j}'(t_j)\big)x+\check\Phi_{N,j}(t_j)\gamma_{N,j}x^3/6. $$ \end{proposition}
\begin{proof} Let $t_j=\frac{2\pi l}{m}$ be a resonant point with $M_N(m)\leq R(2,K) \ln V_N.$ Recall that $\textbf{C}_j(k)$ are given by \eqref{Cj(k)}.
First, in order to compute the term corresponding to ${\sigma}_{N_0,N}^{-2}$ we need only to consider the case $s\leq1$ in (\ref{s}). Using \eqref{A 1,n .1} we end up with
the following contribution
of the interval containing $t_j$, $$ \sqrt{(2\pi)^{-1}}e^{-it_j k}\tilde \Phi_{N,j}(t_j){\sigma}_{N_0,N}^{-1}\Bigg(\int_{-\infty}^{\infty}e^{-ih(k-{\mathbb E}[\tilde S_{N,j}])/{\sigma}_{N,j}}e^{-h^2/2}dh$$ $$+{\sigma}_{N,j}^{-1}\int_{-\infty}^\infty e^{-ih(k-{\mathbb E}[\tilde S_{N,j}])/{\sigma}_{N,j}}\left(\frac{ih^3}{6}{\mathbb E}\left[\big(\tilde S_{N,j}-{\mathbb E}(\tilde S_{N,j})\big)^3\right] {\sigma}_{N,j}^{-3}\right)dh$$ $$+{\sigma}_{N,j}^{-1}(C_{1,N}\check\Phi_{N,j}(t_j)+\check\Phi_{N,j}'(t_j))\int_{-\infty}^\infty e^{-ih(k-{\mathbb E}(\tilde S_{N,j}))/{\sigma}_{N,j}}he^{-h^2/2}dh\Bigg)$$ $$=e^{-it_j k}\tilde\Phi_{N,j}(t_j)\sqrt{(2\pi)^{-1}} e^{-k_{N,j}^2/2}{\sigma}_{N,j}^{-1}\Big(\check\Phi_{N,j}(t_j)+i\big(C_{1,N,t_j}\check\Phi_{N,j}(t_j)+\check\Phi_{N,j}'(t_j)\big)$$ $$\times k_{N,j}{\sigma}_{N,j}^{-1}+\check\Phi_{N,j}(t_j)(k_{N,j}^3-3k_{N,j})\gamma_{N,j}{\sigma}_{N,j}^{-1}/6\Big) $$ where ${\sigma}_{N,j}=\sqrt{V(\tilde S_{N,j})}$, $k_{N,j}=(k-{\mathbb E}(\tilde S_{N,j}))/{\sigma}_{N,j}$ and $\gamma_{N,j}= \frac{{\mathbb E}[(\tilde S_{N,j}-{\mathbb E}(\tilde S_{N,j}))^3]}{{\sigma}_{N,j}^2}$ (which is uniformly bounded).
As before we shall only consider the case where $|k_N|\leq V_N^{\varepsilon}$ with ${\varepsilon}=0.01$ since otherwise both the LHS and the RHS \eqref{r=2'} are $O({\sigma}_N^{-r})$ for all $r.$ Then, the last display can be rewritten as $I+I\!\!I$ where \begin{equation} I=\frac{e^{-i t_j k }}{\sqrt{2\pi} {\sigma}_{N,j}}\; e^{-k_{N,j}^2/2}\; \Phi_N(t_j); \end{equation} $$ I\!\!I =\frac{e^{-i t_j k }}{\sqrt{2\pi} {\sigma}^2_{N,j}}\; e^{-k_{N,j}^2/2}\times$$ $$ \left[\Phi_N(t_j) \left(i C_{1, N,t_j} k_{N,j}+\frac{\gamma_{N,j}}{6} \left(k_{N,j}^3-3 k_{N,j}\right)\right)+i\check\Phi_{N,j}'(t_j) \tilde\Phi_{N,j}(t_j) k_{N,j} \right]. $$
In the region $|k_N|\leq V_N^{\varepsilon}$ we have $$ I= \frac{e^{-i t_j k }}{\sqrt{2\pi} {\sigma}_{N}}\; e^{-k_{N}^2/2}\; \left[1-q_{N,j} k_N \right]\Phi_N(t_j)+ o\left({\sigma}_N^{-2}\right)$$ where $$q_{N,j}={\mathbb E}(\check S_{N,j})/{\sigma}_{N,j}={\mathbb E}(\check S_{N,j})/{\sigma}_N+O(\ln{\sigma}_N/{\sigma}_N^3)= O(\ln{\sigma}_N/{\sigma}_N) $$ while $$ I\!\!I=\frac{e^{-i t_j k }}{\sqrt{2\pi} {\sigma}^2_{N}}\; e^{-k_{N}^2/2}\times$$ $$ \left[\Phi_N(t_j) \left(i C_{1, N,t_j} k_{ N}+\frac{{\gamma}_{N,j} \left(k_{N}^3-3 k_{N}\right)}{6} \right)+i\check\Phi_{N,j}'(t_j) \tilde\Phi_{N,j}(t_j) k_{N} \right]$$ $$+o\left({\sigma}_N^{-2}\right).$$ This yields (\ref{r=2'}) with ${\mathcal{R}}_N$ in place of ${\mathcal{R}}$, where ${\mathcal{R}}_N$ is the set on nonzero resonant points $t_j=2\pi l/m$ such that $M_N(m)\leq R(2,K)\ln V_N$. Next, \eqref{Roz0} shows that if $M_N(m)\geq R(2,k)\ln V_N$ then \[
\sup_{t\in I_j}|\Phi_{N}(t)|\leq e^{-c_0M_N(m)}=o({\sigma}_N^{-2}) \] and so the contribution of $I_j$ to the right hand side of \eqref{r=2'} is $o({\sigma}_N^{-2}).$ Finally,
the contribution coming from $t_j=0$ is \[ e^{-k_N^2/2}\left({\sigma}_N^{-1}+{\sigma}_N^{-2}\gamma_N^3 k_N^3/6\right) \] and the proof of the proposition is complete. \end{proof} \begin{remark}\label{Alter 2nd Order}
Suppose that $M_N(m)\geq R(2,K)\ln V_N$ and let $N_0$ is the number of $n$'s between $1$ to $N$ so that $q_n(m)\geq \frac{1}{8K}$. Then using (\ref{Roz0}) we also have $$
|\check\Phi'_{N,j}(t_j)\tilde \Phi_{N,j}(t_j)|\leq\sum_{n\in{\mathcal B}_{{\bar\varepsilon}}(m)}|{\mathbb E}[X_ne^{it_j X_n}]|\cdot|\Phi_{N;n}(t_j)| $$ $$ \leq CN_0(N,t_j,{\bar\varepsilon})e^{-c_0 M_{N}(m)}\leq C'M_{N}(m)e^{-c_0 M_{N}(m)}, $$ where $${\mathcal B}_{N,{\bar\varepsilon}}(m)=\{1\leq n\leq N: q_n(m)>{\bar\varepsilon}\}.$$ Since $M_N(m)\geq R(2,K)\ln V_N$, for any $0<c_1<c_0$, when $N$ is large enough we have $$M_{N}(m)e^{-c_0 M_{N}(m)}\leq C_1e^{-c_1M_N(m)}=o({\sigma}_N^{-2}).$$
Similarly, $|{\mathbb E}(\check S_{N,j})\Phi_N(t_j)|=o({\sigma}_N^{-2})$ and $$C_{1,N,t_j}\Phi_{N}(t_j)=O(M_N(m))\Phi_{N}(t_j)=o({\sigma}_N^{-2}).$$ Therefore we get (\ref{r=2'}) when $\tilde S_{N,j}$ and $\check S_{N,j}$ are defined in the same way as in the case $M_N(m)\leq R(2,K)\ln V_N$. \end{remark}
Under additional assumptions the order 2 expansion can be simplified.
\begin{corollary} \label{CrR2-SLLT} If $S_N$ satisfies SLLT then $$ \sqrt {2\pi}{\mathbb P}(S_N=k)=\frac{e^{-k_N^2/2}}{{\sigma}_N} \left(1+\sum_{t_j\in {\mathcal{R}}}e^{-it_j k}\Phi_{N}(t_j) -\frac{\gamma_N k_N^3}{6{\sigma}_N}\right) +o(\sigma_N^{-2}). $$ \end{corollary}
\begin{proof} The estimates of \S \ref{SSEdgeR=2} together with \eqref{Roz0} show that if $S_N$ satisfies the SLLT then for all $j$ $$(1+M_N(m))\tilde\Phi_{N,j} (t_j) \check \Phi_{N,j}(t_j)=o(1) \text{ and } \tilde\Phi_{N,j} (t_j) \check \Phi'_{N,j}(t_j)=o(1).$$ Thus all terms in the second line of \eqref{r=2'} except the first one make a negligible contribution, and so they could be omitted. \end{proof}
Next, assume that $S_N$ satisfies the LLT but not SLLT. According to Proposition \ref{PrLLT-SLLT}, in this case there exists $m$ such that $M_N(m)$ is bounded and for $k=1,\dots ,m-1$ there exists $n=n(k)$ such that $\phi_{n}(k/m)=0$. Let ${\mathcal{R}}_s$ denote the set of nonzero resonant points $t_j=\frac{2\pi k}{m}$ so that $M_N(m)$ is bounded and $\phi_{\ell_j}(t_j)=0$ for unique $\ell_j.$
\begin{corollary}\label{CrR2Z} Uniformly in $k$, we have \begin{eqnarray*} \sqrt {2\pi}{\mathbb P}(S_N=k)=\Big(1+\sum_{t_j\in {\mathcal{R}}}e^{-it_j k}\Phi_{N}(t_j)\Big){\sigma}_{N}^{-1}e^{-k_N^2/2}\\-{\sigma}_{N}^{-2}e^{-k_N^2/2}\left(\gamma_N k_N^3/6+ \sum_{t_j\in {\mathcal{R}}_s}i e^{-it_j k}\Phi_{N; \ell_j}(t_j) \phi_{\ell_j}'(t_j) k_N \right)+o(\sigma_N^{-2}).\nonumber \end{eqnarray*} \end{corollary}
\begin{proof} As in the proof of Corollary \ref{CrR2-SLLT} we see that the contribution of the terms with $k/m$ with $M_N(m)\to\infty$ is negligible. Next, for terms in ${\mathcal{R}}_s$ the only non-zero term in the second line in \eqref{r=2'} corresponds to $\Phi_{N; \ell_j}(t_j) \phi_{\ell_j}'(t_j)$ while for the resonant points such that $\phi_{\ell}(t_j)=0$ for two different $\ell$s all terms vanish. \end{proof}
\section{Examples.} \label{ScExamples} \begin{example} \label{ExNonAr} Suppose $X_n$ are iid integer valued with step $h>1.$ That is there is $s\in {\mathbb Z}$ such that ${\mathbb P}(X_n\in s+h {\mathbb Z})=1$ and $h$ is the smallest number with this property. In this case \cite[Theorem 4.5.4]{IL} (see also \cite[Theorem 5]{Ess}) shows that there are polynomials $P_b$ such that \begin{equation} \label{EssEdge} {\mathbb P}(S_N=k)= \sum_{b=1}^r \frac{P_{b} ((k-{\mathbb E}[S_N])/\sigma_N)}{\sigma_N^b} \mathfrak{g}\left(\frac{k-{\mathbb E}(S_N)}{\sigma_N}\right) +o(\sigma_N^{-r}) \end{equation} for all $k\in sN+h{\mathbb Z}.$ Then $$ \sum_{a=0}^{h-1} \sum_{b=1}^r e^{2\pi i a(k-sN)/h} \frac{P_{b} ((k-{\mathbb E}[S_N])/\sigma_N)}{\sigma_N^b} \mathfrak{g}((k-{\mathbb E}(S_N))/\sigma_N)$$ provides $\displaystyle o(\sigma_N^{-r})$ approximation to ${\mathbb P}(S_N=k)$ which is valid for {\em all} $k\in {\mathbb Z}.$
Next let ${\bar S}_N=X_0+S_N$ where $X_0$ is bounded and arithmetic with step 1. Then using the identity \begin{equation} \label{Convolve}
{\mathbb P}({\bar S}_N=k)=\sum_{u\equiv k-s N\text{ mod } h} {\mathbb P}(X_0=u) {\mathbb P}(S_N=k-u), \end{equation}
\noindent invoking \eqref{EssEdge} and expanding $\displaystyle \mathfrak{g}\left(\frac{k-u-{\mathbb E}(S_N)}{\sigma_N}\right)$ in the Taylor series about $\frac{k-{ {\mathbb E}(S_N)}}{\sigma_N}$ we conclude that there are polynomials $P_{b,j}$ such that we have for $k\in j+h{\mathbb Z}$, $$ {\mathbb P}({\bar S}_N=k)= \sum_{b=1}^r \frac{P_{b, j} ((k-{\mathbb E}[S_N])/\sigma_N)}{\sigma_N^b} \mathfrak{g}\left(\frac{k-{\mathbb E}(S_N)}{\sigma_N}\right) +o(\sigma_N^{-r}). $$ Again $$ \sum_{a=0}^{h-1} \sum_{j=0}^{h-1} e^{2\pi i a (k-j)/h} \sum_{b=1}^r \frac{P_{b,j} ((k-{\mathbb E}[S_N])/\sigma_N)}{\sigma_N^b} \mathfrak{g}\left(\frac{k-{\mathbb E}(S_N)}{\sigma_N}\right) $$ provides the oscillatory expansion valid for all integers. \end{example}
\begin{example} \label{ExUniform}
Our next example is a small variation of the previous one. Fix a positive integer $m.$ Let $X'$ be a random variable such that $X'$ mod $m$ is uniformly distributed. Then its characteristic function satisfies $\phi_{X'}(\frac{2\pi a}{m})=0$ for $a=1, \dots, m-1.$ We also assume that $\phi_{X'}'(\frac{2\pi a}{m})\neq 0$ for $a$ as above
(for example one can suppose that $X'$ takes the values $Lm$, $1, 2, \dots ,m-1$ with equal probabilities where $L$ is a large integer). Let $X''$ take values
in $m{\mathbb Z}$ and have zero mean. We also assume that $X''$
does not take values at $m_0{\mathbb Z}$ for a larger $m_0$. Then $q(X'',m_0)>0$ for any $m_0\not=m$. Fix $r\in \mathbb{N}$ and let $$X_n=\begin{cases} X' & n\leq r, \\ X'' & n>r. \end{cases}$$ Then $M_N(m_0)$ grows linearly fast in $N$ if $m_0\not=m$ and $M_N(m)$ is bounded in $N$. We claim that $S_N$ admits the Edgeworth expansion of order $r$ but does not admit Edgeworth expansion of order $r+1.$ The first statement holds due to Theorem \ref{r Char}, since $\Phi^{(\ell)}_N(\frac{2\pi a}{m})=0$ for each $a\in {\mathbb Z}$ and each $\ell<r.$ On the other hand, since $\Phi_{N}^{(\ell)}(\frac{2\pi a}m)=0$ for any $\ell<r$, using Lemma \ref{Lem} we see that the conditions of Lemma \ref{LemInd} are satisfied with $r+1$ in place of $r$. Moreover, with $t_j=2\pi a/m$, $a\not=0$ we have ${\mathcal H}_{N,r+1,s}(x,t_j)\equiv0$ for any $q\leq r+1$ and $s=2,3,4$ while ${\mathcal H}_{N,q,w}(x,t_j)\equiv0$ for any $q\leq r$ and $w=1,2,3,4$. Furthermore, when $N\geq r$ we have \begin{eqnarray*} {\mathcal H}_{N,r+1,1}(x;t_j)=\frac{i^{r}H_{r}(x)\big(\phi_{X''}(2\pi a/m)\big)^{N-r}\Phi_{r}^{(r)}(2\pi a/m)}{r!}\\=(i)^{r}H_{r}(x)\big(\Phi_{X'}'(2\pi a/m)\big)^r. \end{eqnarray*} We conclude that $$ {\mathbb P}(S_N=k)$$ $$=\frac{e^{-k_N^2/2}}{\sqrt{2\pi}} \left[{\mathcal E}_{r+1}(k_N)+ \frac{i^r}{{\sigma}_N^{r+1}} \sum_{a=1}^{m-1} e^{-2\pi i ak/m} \left(\phi_{X'}'\left(\frac{2\pi a}{m}\right) \right)^r H_{r}(k_N) \right] $$ $$ +o({\sigma}_N^{-r-1}) $$ where ${\mathcal E}_{r+1}$ the Edgeworth polynomial (i.e. the contribution of 0) and $H_{r}(x)$ is the Hermite polynomial.
Observe that since the uniform distribution on ${\mathbb Z}/m{\mathbb Z}$ is shift invariant, $S_N$ are uniformly distributed mod $m$ for all $N\in {\mathbb N}.$ This shows that for $r\geq 1$, one can not characterize Edgeworth expansions just in term of the distributions of $S_N$ mod $m$, so the additional assumptions in Theorems \ref{Thm SLLT VS Ege} and \ref{Thm Stable Cond} are necessary.
Next, consider a more general case where for each $n$, $X_n$ equals in law to either $X'$ or $X''$, however, now we assume that $X'$ appears infinitely often. In this case $S_N$ obeys Edgeworth expansions of all orders since for large $N$, $\Phi_N(t)$ has zeroes of order greater $N$ at all points of the form $\frac{2\pi a}{m},$ $ a=1, \dots, m-1.$ In fact, the Edgeworth expansions hold in the superstable way since removing a finite number of terms does not make the order of zero to fall below $r.$
\end{example}
\begin{example}\label{Eg1} Let $\mathfrak{p}_n=\min(1, \frac{\theta}{n})$ and let $X_n$ take value $0$ with probability $\mathfrak{p}_n$ and values $\pm 1$ with probability $\frac{1-\mathfrak{p}_n}{2}.$ In this example the only non-zero resonant point is $\pi=2\pi\times \frac{1}{2}.$ Then for small $\theta$ the contributions of $P_{1, b, N}$ (the only non-zero $a$ is $1$) are significant and as a result $S_N$ does not admit the ordinary Edgeworth expansion. Increasing $\theta$ we can make $S_N$ to admit Edgeworth expansions of higher and higher orders.
Namely we get that for large $n,$ $\displaystyle \phi_n(\pi)=\frac{2\theta}{n}-1. $ Accordingly $$ \ln (-\phi_n(\pi))=-\frac{2\theta}{n}+O\left(\frac{1}{n^2}\right). $$
Now the asymptotic relation $$ \sum_{n=1}^N \frac{1}{n}=\ln N+{\mathfrak{c}}+O\left(\frac{1}{N}\right),$$ where ${\mathfrak{c}}$ is the Euler-Mascheroni constant,
implies that that there is a constant $\Gamma(\theta)$ such that $$ \Phi_N(\pi)=\frac{(-1)^N e^{\Gamma(\theta)}}{N^{2\theta}}\left(1+O(1/N)\right).$$ Therefore $S_N$ admits the Edgeworth expansions of order $r$ iff $\displaystyle \theta>\frac{r-1}{4}.$ Moreover, if $\displaystyle \theta\in \left(\frac{r-2}{4}, \frac{r-1}{4}\right],$ then Corollary \ref{CrFirstNonEdge} shows that $$ \mathbb{P}(S_N=k)=\frac{e^{-k_N^2/2}}{\sqrt{2\pi}} \left[{\mathcal E}_{r}(k_N)+ \frac{(-1)^{N+k} e^{\Gamma(\theta) }}{N^{2\theta+(1/2)}}+O\left(\frac{1}{N^{2\theta+1} }\right)\right] $$ where ${\mathcal E}_r$ is the Edgeworth polynomial of order $r.$ In particular if $\theta\in (0, 1/4)$ then using that \begin{equation} \label{VarNearN} V_N=N+O(\ln N)=N\left(1+O\left(\frac{\ln N}{N}\right)\right) \end{equation}
and hence
\begin{equation}
\label{SDNearN} \sigma_N=\sqrt{N}\left(1+O\left(\frac{\ln N}{N}\right)\right) \end{equation} we conclude that $$ \mathbb{P}(S_N=k)=\frac{e^{-k^2/(2N)}}{\sqrt{2\pi}} \left[\frac{1}{\sqrt{N}}+ \frac{(-1)^{N+k} e^{\Gamma(\theta) }}{N^{2\theta+(1/2)}}+O\left(\frac{1}{N^{2\theta+1}}\right)\right]. $$
Next, take $\mathfrak{p}_n=\min\left(1, \frac{\theta}{n^2}\right)$. Then the SLLT does not hold, since the Prokhorov condition fails. Instead we have (\ref{r=1}) with ${\mathcal{R}}=\{\pi\}$. Namely, uniformly in $k$ we have \begin{equation*} \sqrt 2\pi{\mathbb P}(S_N=k)=\left(1 +(-1)^k\prod_{u=1}^{N}\left(2\mathfrak{p}_u-1\right)\right){\sigma}_{N}^{-1}e^{-k^2/2V_N}+o(\sigma_N^{-1}). \end{equation*} Next, $\mathfrak{p}_u$ is summable and moreover \[ \prod_{u=1}^{N}(2\mathfrak{p}_u-1)=(-1)^N U(1+O(1/N)) \] where $\displaystyle U=\prod_{n=1}^{\infty}(1-2\mathfrak{p}_u)$. We conclude that \begin{equation} \label{A1Ex3} \sqrt 2\pi{\mathbb P}(S_N=k)=\left(1+(-1)^{k+N}U\right){\sigma}_{N}^{-1}e^{-k^2/2V_N}+ O\left(\sigma_N^{-2}\right) \end{equation} uniformly in $k$. In this case the usual LLT holds true if and only if $U=0$ in agreement with Proposition \ref{PrLLT-SLLT}.
In fact, in this case we have a faster rate of convergence. To see this we consider expansions of order $2$ for $\mathfrak{p}_n$ as above.
We observe that $q_m(2)=\mathfrak{p}_n$ for large $n.$ Thus
\[
|{\mathbb E}(e^{\pi iX_n})|=1-2 \mathfrak{p}_n \]
and so $|{\mathbb E}(e^{\pi X_n})|\geq \frac12$ when $n\geq N_{\theta}$ for some minimal $N_{\theta}$. Therefore, we can take $N_0=N_{\theta}$. Note also that we have $Y_n=X_n\text{ mod }2-1$. We conclude that for $n>N_0$ we have \[ a_{n}=a_{n,j}=\frac{{\mathbb E}[((-1)^{Y_n}-1)X_n]}{{\mathbb E}[(-1)^{Y_n}]}=0 \] and so the term $C_{1,N}$ vanishes. Next, we observe that \[ \gamma_{N,j}=\frac{\sum_{n=N_0+1}^{N}{\mathbb E}(X_n^3)}{\sum_{n=N_0+1}^{N}(1-\mathfrak{p}_n)}=0. \] Finally, we note that ${\mathbb E}[(-1)^{X_n}X_n]=0$, and hence $\Phi_{N_0}'(\pi)=0$. Therefore, the second term in (\ref{r=2'}) vanishes and we have \begin{equation*} \sqrt 2\pi{\mathbb P}(S_N=k)=\left(1 +(-1)^{k+N} U
\right){\sigma}_{N}^{-1}e^{-k^2/(2V_N)}+
O\left(\sigma_N^{-3}\right). \end{equation*} Taking into account \eqref{VarNearN} and \eqref{SDNearN} we obtain $$ \sqrt 2\pi{\mathbb P}(S_N=k)=\frac{1 +(-1)^{k+N} U}{\sqrt{N}} \; e^{-k^2/(2N)}+ O\left(N^{-3/2}\right). $$ In particular, \eqref{A1Ex3} holds with the stronger rate $O\left(\sigma_N^{-3}\right)$. \end{example}
\begin{example}\label{NonSym r=2 Eg}
The last example exhibited significant simplifications. Namely, there was only one resonant point, and, in addition, the second term vanished due to the symmetry. We now show how a similar analysis could be performed when the above simplifications are not present. Let us assume that $X_n$ takes the values $-1,0$ and $3$ with probabilities $a_n,b_n$ and $c_n$ so that $a_n+b_n+c_n=1$. Let us also assume that $ b_n<\frac18$ and that $a_n,c_n\geq \rho>0$ for some constant $\rho$. Then $$V(X_n)=9(c_n-c_n^2)+6a_nc_n+(a_n-a_n^2)\geq 6 \rho^2$$ \noindent and so $V_N$ grows linearly fast in $N$.
Next, since we can take $K=3$, the denominators $m$ of the nonzero resonant points can only be $2,3,4,5$ or $6$. An easy check shows that for $m=3,5,6$ we have $q_n(m)\geq \rho$, and that for $m=2,4$ we have $q_n(m)=b_n$. Therefore, for $m=3,5,6$ we have $M_N(m)\geq \rho N$, and so we can disregard all the nonzero resonant points except for $\pi/2,\pi$ and $3\pi/2$. For the latter points we have \begin{equation} \label{Res1} \phi_n\left(\frac{\pi}{2}\right)=b_n-i(1-b_n), \end{equation} \begin{equation} \label{Res2}
\phi_n(\pi)=2b_n-1,\quad \phi_n\left(\frac{3\pi}{2}\right)=b_n+i(1-b_n). \end{equation} Hence, denoting $\eta_n=b_n(1-b_n),$ we have \begin{equation*}
\left|\phi_n\left(\frac{\pi}{2}\right)\right|^2=\left|\phi_n\left(\frac{3\pi}{2}\right)\right|^2=1-2\eta_n, \quad
\big|\phi_n(\pi)\big|^2=1-4\eta_n. \end{equation*}
Since we suppose that $\eta_n\leq b_n<\frac{1}{8}$ it follows that $1-4\eta_n\geq \frac12$. Then for the above three resonant points we can take $N_0=0$.
Now Proposition~\ref{Thm} and a simple calculation show that for any $r$ we get the Edgeworth expansions of order $r$ if and only if \[ \prod_{n=1}^N(1-2\eta_n)= o\left(N^{1-r}\right). \] Let us focus for the moment on the case when $b_n=\gamma/n$ for $n$ large enough
where $\gamma>0$ is a constant.
Rewriting \eqref{Res1}, \eqref{Res2} as \begin{equation} \label{PhiNRatio}
\frac{\phi_n\left(\frac{\pi}{2}\right)}{-i}=(1-b_n)+ib_n, \quad \frac{\phi_n\left(\frac{3\pi}{2}\right)}{i}=(1-b_n)-ib_n, \quad
\end{equation}
$$ -\phi_n(\pi)=1-2 b_n $$
and, using that the condition $b_n<\frac{1}{8}$ implies that
that $\phi_n(t)\neq 0$ for
all $n\in \mathbb{N}$ and all $t\in \left\{\frac{\pi}{2}, \pi, \frac{3\pi}{2}\right\}$, we conclude similarly to Example \ref{Eg1} that there are non-zero complex numbers $\kappa_1, \kappa_3$ and a non-zero real number $\kappa_2$ such that $$ \Phi_N\left(\frac{\pi}{2}\right)=\frac{(-i)^N \kappa_1}{N^\gamma} e^{i \gamma \ln N
} \left(1+O\left(\frac1N\right)\right), $$ $$ \Phi_N\left(\frac{3\pi}{2}\right)=\frac{i^N \kappa_3}{N^\gamma} e^{-i \gamma \ln N } \left(1+O\left(\frac1N\right)\right), $$ $$ \Phi_N(\pi)=(-1)^N \frac{\kappa_2e^{2\gamma w_N}}{N^{2\gamma}} \left(1+O\left(\frac1N\right)\right). $$ It follows that $S_N$ admits Edgeworth expansion of order $r$ iff $\gamma>\frac{r-1}{2}.$ In fact if $\frac{r-1}{2}<\gamma\leq \frac{r}{2}$ then Corollary \ref{CrFirstNonEdge} shows that $$ \mathbb{P}(S_N=k)= \frac{e^{-k_N^2/2}}{\sqrt{2\pi}} \Bigg[{\mathcal E}_{r}(k_N)+ \frac{\kappa_1 e^{i\gamma \ln N}} {N^\gamma {\sigma}_N} + \frac{\kappa_3 e^{-i\gamma \ln N}}
{N^\gamma {\sigma}_N} +O\left(N^{-\eta}\right)\Bigg] $$ where ${\mathcal E}_r$ is the Edgeworth polynomial of order $r$ and $\eta\!\!=\!\!\min\left(2\gamma, \frac{r}{2}\right)+\frac{1}{2}.$
To give a specific example, let us suppose that $\frac{1}{2}\leq \gamma<1$ and that $E(X_n)=0$ which means that \begin{equation} \label{Ex4-A-C} a_n=\frac{3(1-b_n)}{4}, \quad c_n=\frac{1-b_n}{4}. \end{equation} Then \begin{equation} \label{Ex4M2M3}
V_N=3N-3\gamma \ln N+O(1), \quad E(S_N^3)=6N-6\gamma \ln N+O(1), \end{equation} so Proposition \ref{2 Prop} gives $$ \sqrt{2\pi} \mathbb{P}(S_N=k)=$$ $$e^{-k^2/6 N} \left[\frac{1}{\sqrt{3N}} \left(1+\frac{\kappa_1 i^{k-N} e^{i\gamma \ln N} +\kappa_3 i^{N-k} e^{-i\gamma \ln N}} {N^\gamma}\right)-
\frac{k^3}{81\sqrt{3 N^5}} \right] $$ $$ +O\left(N^{-3/2}\right) $$
Next, let us provide the second order trigonometric expansions under the sole assumption that $1-4\eta_n\geq \frac12$ and $a_n,c_n\geq \rho$. As we have mentioned, we only need to consider the nonzero resonant points $\pi/2,\pi,3\pi/2$ and for these points we have $N_0=0$. Therefore, the term involving the derivative in the right hand side of (\ref{r=2'}) vanishes. Now, a direct calculation shows that \[ C_{1,N,\pi}=\sum_{n=1}^{N}\frac{{\mathbb E}(e^{i\pi X_n}\bar X_n)}{{\mathbb E}(e^{i\pi X_n})}=2\sum_{n=1}^{N}\frac{(a_n-3c_n)b_n}{2b_n-1} \] and $$ C_{1,N,\pi/2}=\sum_{n=1}^N\frac{(a_n-3c_n)(1+i)b_n}{b_n-i(1-b_n)},\quad C_{1,N,3\pi/2}=\sum_{n=1}^N\frac{(a_n-3c_n)(1-i)b_n}{b_n+i(1-b_n)}. $$ Note that $3c_n-a_n={\mathbb E}(X_n)$. Set $$ \Gamma_{1,N}=\prod_{n=1}^{N}(b_n-i(1-b_n)), \quad \Gamma_{2,N}=\prod_{n=1}^N(2b_n-1), \quad
\Gamma_{3,N}=\prod_{n=1}^{N}(b_n+i(1-b_n)). $$ Then $\Gamma_{s,N}={\mathbb E}(e^{\frac{s\pi i}{2} S_N})$. We also set $$ \Theta_{s,N}=C_{1,N,s\pi/2}\Gamma_{s,N},\,s=1,2,3 $$ and $$\Gamma_{N}(k)=\sum_{j=1}^{3}e^{-j\pi i k/2}\Gamma_{j,N}, \quad \Theta_{N}(k)=\sum_{j=1}^{3}e^{-j\pi i k/2}\Theta_{j,N}.$$ Then by Proposition \ref{2 Prop} and Remark \ref{Alter 2nd Order}, uniformly in $k$ we have \begin{equation}\label{EgGen} \sqrt{2\pi}P(S_N=k)={\sigma}_N^{-1}\left(1+\Gamma_N(k)\right)e^{-k_N^2/2} \end{equation} $$ -{\sigma}_N^{-2}\left(k_N^3 T_N\big(1+\Gamma_{N}(k)\big)+ik_N\Theta_{N}(k)\right)e^{-k_N^2/2}+o({\sigma}_N^{-2}) $$ where $T_N=\frac{\displaystyle \sum_{n=1}^{N} {\mathbb E}(\bar X_n^3)}{6V_N}$, $\bar X_n=X_n-{\mathbb E}(X_n)$.
Let us now consider a more specific situation.
Namely we suppose that $b_n=\frac{\gamma}{n^{3/2}}$ for large $n$ and that $E(X_n)=0.$ Then \eqref{Ex4-A-C} shows that $C_{1, N, s\pi/2}=0.$ Next \eqref{PhiNRatio} gives $$ \frac{\Phi_N\left(\frac{\pi}{2}\right)}{(-i)^N}= \prod_{n=1}^N \left[(1-b_n)+ib_n\right]= \frac{{\bar\kappa}_1}{\displaystyle \prod_{n=N+1}^\infty\left[(1-b_n)+ib_n\right]}$$ $$= {\bar\kappa}_1 \left(1+\frac{2 \gamma(1-i)}{\sqrt{N}}+O\left(\frac{1}{N}\right)\right) $$ where $\displaystyle {\bar\kappa}_1=\prod_{n=1}^\infty \left[(1-b_n)+ib_n\right].$ Likewise $$ \frac{\Phi_N\left(\frac{3\pi}{2}\right)}{i^N}= {\bar\kappa}_3 \left(1+\frac{2 \gamma(1+i)}{\sqrt{N}}+O\left(\frac{1}{N}\right)\right) $$ and $$
\frac{\Phi_N\left(\pi\right)}{(-1)^N}= {\bar\kappa}_2 \left(1+\frac{4 \gamma}{\sqrt{N}}+O\left(\frac{1}{N}\right)\right). $$ Taking into account \eqref{Ex4M2M3} we can reduce \eqref{EgGen} to the following expansion $$ \sqrt{2\pi}{\mathbb P}(S_N=k)=e^{-k^2/6N} \left[ \frac{1}{\sqrt{3N}} \left(1+\sum_{s=1}^3 {\bar\kappa}_s i^{s(k-N)} \right)\right. $$ $$\left.+\frac{1}{N}\left( -\frac{\tilde k_N^3}{3} + \sum_{s=1}^3 {\bar\kappa}_s i^{s(k-N)} \left(\frac{2\gamma(1-i^{-s}) }{\sqrt{3}}-\frac{\tilde k_N^3}{3} \right)\right) \right]+O\left(\frac{1}{N^{3/2}}\right)
$$ where $\tilde k_N=k/\sqrt{3N}$. \end{example}
\begin{example} Let $X'$ take value $\pm 1$ with probability $\frac{1}{2}$, $X''$ take values $0$ and $1$ with probability $\frac{1}{2}$, and $X^{\delta}$, ${\delta}\in[0,1]$ be the mixture of $X'$ and $X''$ with weights $\delta$ and $1-\delta.$ Thus $X^\delta$ take value $-1$ with probability $\frac{\delta}{2},$ the value $0$ with probability $\frac{1-\delta}{2}$ and value $1$ with probability $\frac{1}{2}$. Therefore, ${\mathbb E}(e^{\pi i X^\delta})=-{\delta}$. We suppose that $X_{2m}$ and $X_{2m-1}$ have the same law which we call $Y_m.$ The distribution of $Y_m$ is defined as follows. Set $k_j=3^{3^j}$, and let $Y_{k_j}$ have the same distribution as $X^{{\delta}_j}$ where ${\delta}_j=\frac{1}{\sqrt{k_{j+1}}}$. When $m\not\in\{k_j\}$ we let $Y_m$ have the distribution of $X'$. It is clear that $V_N$ grows linearly fast in $N$. Note also that ${\mathbb E}(e^{\pi i Y_m})=-{\delta}_j$ when $m=k_j$ for some $j$, and otherwise ${\mathbb E}(e^{\pi i Y_m})=-1$. Now, take $N\in{\mathbb N}$ such that $N>2k_2$, and let $J_N$ be so that $2k_{J_N}\leq N<2k_{J_N+1}$. Then $$
|\Phi_{N}(\pi)|\leq \prod_{j=1}^{J_N}(k_{j+1})^{-1}. $$ Since $k_{J_N}\leq \frac{N}2<k_{J_N+1}$ and $k_j=(k_{j+1})^{1/3}$ we have $k_{J_N+1}^{-1}\leq 2N^{-1}$ and $k_{J_N+1-m}\leq 2^{3^{-m}} N^{-3^{-m}}$ for any $0<m\leq J_N$.
Denote $\displaystyle \alpha_N=\sum_{j=1}^{J_N-1}3^{-j}.$ Since $\alpha_N>1/3$ we get that \[
|\Phi_{N}(\pi)|\leq 2^{3/2}N^{-\alpha_N} =o(N^{-1-1/3}). \] Similarly, for each $j_1, j_2\leq N$, \begin{equation}\label{2nd}
|\Phi_{N: j_1}(\pi)|\leq 2^{3/2} N^{-1/2-\alpha_N}=o(N^{-1/2-1/3}) \end{equation} and \begin{equation}\label{3rd}
|\Phi_{N: j_1,j_2}(\pi)|\leq 2^{3/2} N^{-\alpha_N}=o(N^{-1/3}). \end{equation} Indeed, the largest possible values are obtained for $j_1=2k_{J_N}$ (or $j_1=2k_{J_{N+1}}-1$ if it is smaller than $N+1$) and $j_2=2k_{J_N}-1$ (or $j_2=2k_{J_{N}}$). Using the same estimates as in the proof of of Theorem \ref{Thm Stable Cond} we conclude from (\ref{2nd}) that $\displaystyle \Phi_N'(\pi)=o\left(1/\sqrt{N}\right)$ and we conclude from \eqref{3rd} that $\displaystyle \Phi_N''(\pi)=o(1).$ It follows from Lemma \ref{Lem} and Proposition \ref{Thm} that $S_N$ satisfies an Edgeworth expansion of order 3. The same conclusion holds if we remove a finite number of terms from the beginning of the sequence $\{X_n\}$ because the smallness of $\Phi_N(\pi)$ comes from the terms $X_{2 k_{j}-1} $ and $X_{2 k_j}$ for arbitrary large $j$'s.
On the other hand
$$ \left|\Phi_{2k_j; 2k_j, 2k_j-1, 2k_{j-1}, 2k_{j-1}-1}(\pi)\right|= \prod_{s=2}^{j-1} \left(3^{-3^{s}}\right)$$ $$=
3^{-(3^j-9)/2}=\frac{3^{9/2}}{\sqrt{k_j}} \gg \frac{1}{k_j}=3^{-3^j}. $$ It follows that $S_{2k_j; 2k_j, 2k_j-1, 2k_{j-1}, 2k_{j-1}-1}$ does not obey the Edgeworth expansion of order 3. Accordingly, stable Edgeworth expansions need not be superstable if $r=3.$ A similar argument allows to construct examples showing that those notions are different for all $r>2.$
\end{example}
\section{Extension for uniformly bounded integer-valued triangular arrays}
In this section we will describe our results for arrays of independent random variables. We refer to \cite{Feller}, \cite{Mu84, Mu91} and \cite{Dub}, \cite{VS}, \cite{Pel1} and \cite{DS} for results for triangular arrays of inhomogeneous Markov chains. Example where Markov arrays appear naturally include the theory of large deviations for inhomogeneous systems (see \cite{SaSt91, PR08, FH} and references wherein), random walks in random scenery \cite{CGPPS, GW17}, and statistical mechanics \cite{LPRS}.
Let $X_n^{(N)},\,1\leq n\leq L_N$ be a triangular array such that for each fixed $N$, the random variables $X_n^{(N)}$ are independent and integer valued. Moreover, we assume that
$$K:=\sup_{N}\sup_{n}\|X_n\|_{L^\infty}<\infty.$$ For each $N$ we set $\displaystyle S_N=\sum_{n=1}^{L_N}X_n^{(N)}$. Let $V_N=\text{Var}(S_N)$. We assume that $V_N\to\infty$, so that, by Lindenberg--Feller Theorem, the sequence $(S_N-{\mathbb E}(S_N))/{\sigma}_N$ obeys the CLT, where ${\sigma}_N=\sqrt{V_N}$.
We say that the array $X_n^{(N)}$ obeys the SLLT if for any $k$ the LLT holds true for any uniformly square integrable array $Y_n^{(N)},\,1\leq n\leq L_N$, so that $Y_n^{(N)}=X_n^{(N)}$ for all but $k$ indexes $n$. Set $$ M_N:=\min_{2\leq h\leq 2K}\sum_{n=1}^{L_N}P(X_n\neq m_n^{(N)}(h) \text{ mod } h)\geq R\ln V_N $$ where $m_n^{(N)}(h)$ is the most likely value of $X_n^{(N)}$ modulo $h$.
Observe now that the proofs of Proposition \ref{PropEdg} and Lemmas \ref{Step1}, \ref{Step2}, \ref{Step3} and \ref{Step4} proceed exactly the same for arrays. Therefore, all the arguments in the proof of Theorem \ref{IntIndThm} proceed the same for arrays instead of a fixed sequence $X_n$. That is, we have \begin{theorem}\label{IntIndThmAr} There $\exists J=J(K)<\infty$ and polynomials $P_{a, b, N}$ with degrees depending only on $a$ and $b$, whose coefficients are uniformly bounded in $N$ such that, for any $r\geq1$ uniformly in $k\in{\mathbb Z}$ we have $${\mathbb P}(S_N=k)-\sum_{a=0}^{J-1} \sum_{b=1}^r \frac{P_{a, b, N} ((k-a_N)/\sigma_N)}{\sigma_N^b} \mathfrak{g}((k-a_N)/\sigma_N) e^{2\pi i a k/J} =o(\sigma_N^{-r}) $$ where $a_N={\mathbb E}(S_N)$ and $\mathfrak{g}(u)=\frac{1}{\sqrt{2\pi}} e^{-u^2/2}. $
Moreover, $P_{0,1,N}\equiv1$ and given $K, r$, there exists $R=R(K,r)$ such that if $M_N\geq R \ln V_N$ then we can choose $P_{a, b, N}=0$ for $a\neq 0.$ \end{theorem} All the formulas for the coefficients of the polynomials $P_{a,b,N}$ remain the same in the arrays setup.
In particular, we get that, uniformly in $k$ we have \begin{equation}\label{r=1 Ar} {\mathbb P}(S_N=k)=\left(1+\sum_{t\in{\mathcal{R}}}e^{-itk}\Phi_N(t)\right)e^{-k_N^2/2}{\sigma}_N^{-1}+o({\sigma}_N^{-1}) \end{equation} where $\Phi_N(t)={\mathbb E}(e^{it S_N})$.
Next, our version for Proposition \ref{PrLLT-SLLT} for arrays is as follows. \begin{proposition} \label{PrLLT-SLLT Ar} Suppose $S_N$ obeys LLT. Then for each integer $h\geq2$, at least one of the following conditions occur: \vskip0.2cm either (a) $\displaystyle \lim_{N\to\infty}\sum_{n=1}^{L_{N}} {\mathbb P}(X_n\neq m_n^{(N)}(h) \text{ mod } h)=\infty$. \vskip0.2cm or (b) there exists a subsequence $N_k$, numbers $s\in{\mathbb N}$ and ${\varepsilon}_0>0$ and indexes $1\leq j_1^{k},...,j_{s_k}^{k}\leq L_{N_k}$, $s_k\leq s$ so that the distribution of $\displaystyle \sum_{u=1}^{s_k}X^{(N_k)}_{j_u}$ converges to uniform $\text{mod }h$, and the distance between the distribution of $\displaystyle S_{N_k}-\sum_{q=1}^{s_k}X^{(N_k)}_{j_q}$ and the uniform distribution $\text{mod }h$ is at least ${\varepsilon}_0$. \end{proposition}
\begin{proof} First, by \eqref{r=1 Ar} and Lemma \ref{Lemma} if the LLT holds
then for any nonzero resonant point $t$ we have $\displaystyle \lim_{N\to\infty}|\Phi_N(t)|=0$. Now, if (a) does not hold true then there is a subsequence $N_k$ so that $\displaystyle \sum_{n=1}^{L_{N_k}}q(X_n^{(N_k)},h)\leq C$, where $C$ is some constant. Set $\displaystyle q_n^{(N_k)}(h)=q\left(X_n^{(N_k)},h\right)$. Then there are at most $8hC$ $n$'s between $1$ and $L_{N_k}$ so that $q_n^{(N_k)}(h)>\frac1{8 h}$. Let us denote these $n$'s by $n_{1,k},...,n_{s_k,k}$, $s_k\leq 8h C$. Next, for any $n$ and a nonzero resonant point $t=2\pi l/h$ we have \begin{equation} \label{PhiQArr}
|\phi_n^{(N_k)}(t)|\geq 1-2hq_n^{(N_k)}(h)
\geq e^{-2\gamma h q_n^{(N_k)}} \end{equation} where $\phi_n^{(N_k)}$ is the characteristic function of $X_n^{(N_k)}$
and $\gamma$ is such that for $\theta\in [0, 1/4]$ we have $1-\theta\geq e^{-\gamma \theta}.$
We thus get that \begin{equation}\label{C1.}
\prod_{n\not\in\{n_{u,k}\}}|\phi_{n}^{(N_k)}(t)|\geq\prod_{n\not\in\{n_{u,k}\}}(1-2hq_n^{(N_k)}(h))\geq C_0 \end{equation} where $C_0>0$ is some constant. Therefore, $$
|\Phi_{N_k}(t)|\geq \prod_{u=1}^{s_k}|\phi_{n_{u,k}}^{(N_k)}(t)|\cdot C_0 $$ and so we must have \begin{equation}\label{C2.}
\lim_{k\to\infty}\prod_{u=1}^{s_k}|\phi_{n_{u,k}}^{(N_k)}(t)|=0. \end{equation} Now (b) follows from \eqref{C1.}, \eqref{C2.} and Lemma \ref{LmUnifFourier}. \end{proof}
Using (\ref{r=1 Ar}) we can now prove a version of Theorem \ref{ThProkhorov} for arrays. \begin{theorem}\label{ThProkhorovAr} The SLLT holds iff for each integer $h>1$, \begin{equation}\label{Prokhorov Ar} \lim_{N\to\infty}\sum_{n=1}^{L_N} {\mathbb P}(X_n^{(N)}\neq m_n \text{ mod } h)=\infty \end{equation} where $m_n=m_n^{(N)}(h)$ is the most likely residue of $X_n^{(N)}$ modulo $h$. \end{theorem}
\begin{proof} First, the arguments in the proof of (\ref{Roz0}) show that there are constants $c_0,C>0$ so that for any nonzero resonant point $t=2\pi l/h$ we have \begin{equation}\label{ROZ}
|\Phi_N(t)|\leq Ce^{-c_0M_N(h)},\quad \text{where}\quad M_N(h):=\sum_{n=1}^{L_N}q(X_n^{(N)},h). \end{equation} Let us assume that \eqref{Prokhorov Ar} holds for all integers $h>1$. Consider $s_N$--tuples
$1\leq j_1^N,...,j_{s_N}^N\leq L_N$, where $s_N\leq \bar s$ is bounded in $N$. Then by applying \eqref{ROZ} with $\displaystyle \tilde S_N=S_N-\sum_{l=1}^{s_N}X_{j_l^N}^{(N)}$ we have \begin{equation}\label{TILD}
\lim_{N\to\infty}|{\mathbb E}(e^{it\tilde S_N})|=0. \end{equation} Now, arguing as in the proof of Theorem \ref{Thm Stable Cond}(1), given a uniformly square integrable array $Y_n^{(N)}$ as in the definition of the SLLT, we still have \eqref{r=1 Ar}, even though the new array is not necessarily uniformly bounded. Applying (\ref{TILD}) we see that for any nonzero resonant point $t$ we have $$
\lim_{N\to\infty}\left|{\mathbb E}\left(\exp\left[it \sum_{n=1}^{L_N}Y_n^{(N)}\right]\right)\right|=0 $$ and so $\displaystyle S_N Y:=\sum_{n=1}^{L_N}Y_n^{(N)}$ satisfies the LLT.
Now let us assume that $M_N(h)\not\to\infty$ for some $2\leq h\leq 2K$ (it not difficult to see
that \eqref{Prokhorov Ar} holds for any $h>2K$).
In other words after taking a subsequence we have that $M_{N_k}(h)\leq L$ for some $L<\infty.$
The proof of Proposition \ref{PrLLT-SLLT Ar}
shows that
there $s<\infty$ such that
after possibly removing terms $n_{1, k}, n_{2, k},\dots ,n_{s_k, k}$ with $s_k\leq s$
we can obtain that $q_n^{(N_k)}(h) \leq \frac{1}{8h},$ $n\not\in\{n_{j, k}\}$. In this case \eqref{PhiQArr} shows that for each $\ell$
$$ |\Phi_{N_k; n_{1, k} , \dots, n_{s_k, k}}(2\pi\ell/h)|\geq e^{-2\gamma L}. $$ By Proposition \ref{PrLLT-SLLT Ar}, $S_{N_k; n_{1, k} , \dots, n_{s_k, k}}$ does not satisfy the LLT. \end{proof}
Next, all the other arguments in our paper proceed similarly for arrays since they essentially rely only on the specific structure of the polynomials from Theorem \ref{IntIndThm}.
For the sake of completeness, let us formulate the main (remaining) results here.
\begin{theorem}\label{ThLLT Ar} The following conditions are equivalent:
(a) $S_N$ satisfies LLT;
(b) For each $\xi\in {\mathbb R}\setminus {\mathbb Z}$, $\displaystyle \lim_{N\to\infty} {\mathbb E}\left(e^{2\pi i \xi S_N}\right)=0$;
(c) For each non-zero resonant point $\xi$, $\displaystyle \lim_{N\to\infty} {\mathbb E}\left(e^{2\pi i \xi S_N}\right)=0$;
(d) For each integer $h$ the distribution of $S_N$ mod $h$ converges to uniform. \end{theorem}
\begin{theorem} \label{ThEdgeMN Ar} For each $r$ there is $R=R(r, K)$ such that the Edgeworth expansion of order $r$ holds true if $M_N\geq R\ln V_N$. In particular, $S_N$ obeys Edgeworth expansions of all orders if $$ \lim_{N\to\infty} \frac{M_N}{\ln V_N}=\infty. $$ \end{theorem}
\begin{theorem}\label{r Char Ar}
For any $r\geq1$, the Edgeworth expansion of order $r$ holds if and only if for any nonzero resonant point $t$ and $0\leq\ell<r$ we have \[ \bar \Phi_{N}^{(\ell)}(t)=o\left({\sigma}_N^{\ell+1-r}\right) \] where $\bar\Phi_{N}(x)={\mathbb E}[e^{ix (S_N-{\mathbb E}(S_N))}]$. \end{theorem}
\begin{theorem}\label{Thm SLLT VS Ege Ar } Suppose $S_N$ obeys the SLLT. Then the following are equivalent:
(a) Edgeworth expansion of order 2 holds;
(b) $|\Phi_N(t)|=o(\sigma_N^{-1})$ for each nonzero resonant point $t$;
(c) For each $h\leq 2K$ the distribution of $S_N$ mod $h$ is $o(\sigma_N^{-1})$ close to uniform. \end{theorem}
Next, we say that an array $\{X_n^{(N)}\}$ {\em admits an Edgeworth expansion of order $r$ in a superstable way} (denoted by $\{X_n^{(N)}\}\in EeSs(r)$) if for each ${\bar s}$ and each sequence $j_1^N, j_2^N,\dots ,j_{s_N}^N$ with $s_N\leq {\bar s}$ and $j_i^N\leq L_N$ there are polynomials $P_{b, N}$ whose coefficients are $O(1)$ in $N$ and their degrees do not depend on $N$ so that
uniformly in $k\in{\mathbb Z}$ we have that \begin{equation}\label{EdgeDefSS Ar} {\mathbb P}(S_{N; j_1^N, j_2^N, \dots,j_{{s_N}^N}}=k)=\sum_{b=1}^r \frac{P_{b, N} (k_N)}{\sigma_N^b} \mathfrak{g}(k_N)+o(\sigma_N^{-r}) \end{equation} and the estimates in $O(1)$ and $o(\sigma_N^{-r})$ are uniform in the choice of the tuples $j_1^N, \dots ,j_{s_N}^N.$
Let $\Phi_{N; j_1, j_2,\dots, j_s}(t)$ be the characteristic function of $S_{N; j_1, j_2,\dots, j_s}.$
\begin{theorem}\label{Thm Stable Cond Ar} (1) $S_N\in EeSs(1)$ (that is, $S_N$ satisfies the LLT in a superstable way) if and if it satisfies the SLLT.
(2) For arbitrary $r\geq 1$ the following conditions are equivalent:
(a) $\{X_n^{(N)}\}\in EeSs(r)$;
(b) For each $j_1^N, j_2^N,\dots ,j_{s_N}^N$ and each nonzero resonant point $t$ we have $\Phi_{N; j_1^N, j_2^N,\dots, j_{s_N}^N}(t)=o(\sigma_N^{1-r});$
(c) For each $j_1^N, j_2^N,\dots ,j_{s_N}^N$, and each $h\leq 2K$ the distribution of $S_{N; j_1^N, j_2^N,\dots, j_{s_N}^N}$ mod $h$ is $o(\sigma_N^{1-r})$ close to uniform. \end{theorem}
\end{document}
|
arXiv
|
{
"id": "2011.14852.tex",
"language_detection_score": 0.6475480794906616,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Areas of triangles and $ ext{SL}
\begin{abstract} In Euclidean space, one can use the dot product to give a formula for the area of a triangle in terms of the coordinates of each vertex. Since this formula involves only addition, subtraction, and multiplication, it can be used as a definition of area in $R^2$, where $R$ is an arbitrary ring. The result is a quantity associated with triples of points which is still invariant under the action of $\text{SL}_2(R)$. One can then look at a configuration of points in $R^2$ in terms of the triangles determined by pairs of points and the origin, considering two such configurations to be of the same type if corresponding pairs of points determine the same areas. In this paper we consider the cases $R=\mathbb{F}_q$ and $R=\mathbb{Z}/p^\ell \mathbb{Z}$, and prove that sufficiently large subsets of $R^2$ must produce a positive proportion of all such types of configurations. \end{abstract}
\section{Introduction}
There are several interesting combinatorial problems asking whether a sufficiently large subset of a vector space over a finite field must generate many different objects of some type. The most well known example is the Erdos-Falconer problem, which asks whether such a set must contain all possible distances, or at least a positive proportion of distances. More precisely, given $E\subset\mathbb{F}_q^d$ we define the distance set
\[ \Delta(E)=\{(x_1-y_1)^2+\cdots +(x_d-y_d)^2:x,y\in E\}. \]
Obviously, $\Delta(E)\subset\mathbb{F}_q$. The Erdos-Falconer problem asks for an exponent $s$ such that $\Delta(E)=\mathbb{F}_q$, or more generally $|\Delta(E)|\gtrsim q$, whenever $|E|\gtrsim q^s$ (Throughout, the notation $X\lesssim Y$ means there is a constant $C$ such that $X\leq CY$, $X\approx Y$ means $X\lesssim Y$ and $Y\lesssim X$, and $O(X)$ denotes a quantity that is $\lesssim X$). In \cite{IR}, Iosevich and Rudnev proved that $\Delta(E)=\mathbb{F}_q$ if $|E|\gtrsim q^{\frac{d+1}{2}}$ In \cite{Sharpness} it is proved by Hart, Iosevich, Koh, and Rudnev that the exponent $\frac{d+1}{2}$ cannot be improved in odd dimensions, althought it has been improved to $4/3$ in the $d=2$ case (first in \cite{WolffExponent} in the case $q\equiv 3 \text{ (mod 4)}$ by Chapman, Erdogan, Hart, Iosevich, and Koh, then in general in \cite{GroupAction} by Bennett, Hart, Iosevich, Pakianathan, and Rudnev). Several interesting variants of the distance problem have been studied as well. A result of Pham, Phuong, Sang, Valculescu, and Vinh studies the problem when distances between pairs of points are replaced with distances between points and lines in $\mathbb{F}_q^2$; they prove that if sets $P$ and $L$ of points and lines, respectively, satisfy $|P||L|\gtrsim q^{8/3}$, then they determine a positive proportion of all distances \cite{Lines}. Birklbauer, Iosevich, and Pham proved an analogous result about distances determined by points and hyperplanes in $\mathbb{F}_q^d$ \cite{Planes}.\\
We can replace distances with dot products and ask the analogous question. Let
\[ \Pi(E)=\{x_1y_1+\cdots +x_dy_d:x,y\in E\}, \]
and again ask for an exponent $s$ such that $|E|\gtrsim q^s$ implies $\Pi(E)$ contains all distances (or at least a positive proportion of distances). Hart and Iosevich prove in \cite{HI} that the exponent $s=\frac{d+1}{2}$ works for this question as well. The proof is quite similar to the proof of the same exponent in the Erdos-Falconer problem; in each case, the authors consider a function which counts, for each $t\in\mathbb{F}_q$, the number of representations of $t$ as, respectively, a distance and a dot product determined by the set $E$. These representation functions are then studied using techniques from Fourier analysis. \\
Another interesting variant of this problem was studied in \cite{Angles}, where Lund, Pham, and Vinh defined the angle between two vectors in analogue with the usual geometric interpretation of the dot product. Namely, given vectors $x$ and $y$, they consider the quantity
\[
s(x,y)=1-\frac{(x\cdot y)^2}{\|x\|\|y\|}, \]
where $\|x\|=x_1^2+\cdots x_d^2$ is the finite field distance defined above. Note that since we cannot always take square roots in finite fields, the finite field distance corresponds to the square of the Euclidean distance; therefore, $s(x,y)$ above is the correct finite field analogue of $\sin^2\theta$, where $\theta$ is the angle between the vectors $x$ and $y$. This creates a variant of the dot product problem, since one can obtain different dot products from the same angle by varying length. The authors go on to prove that the exponent $\frac{d+2}{2}$ guarantees a positive proportion of angles. \\
It is of interest to generalize these types of results to point configurations. By a $(k+1)$-point configuration in $\mathbb{F}_q^d$, we simply mean an element of $(\mathbb{F}_q^d)^{k+1}$. Throughout, we will use superscripts to denote different vectors in a given configuration, and subscripts to denote the coordinates of each vector. For example, a $(k+1)$ point configuration $x$ is made up of vectors $x^1,...,x^{k+1}$, each of which has coordinates $x_1^i,x_2^i$. Given a set $E\subset\mathbb{F}_q^d$, we can consider $(k+1)$-point configurations in $E$ (i.e., elements of $E^{k+1}$) and ask whether $E$ must contain a positive proportion of all configurations, up to some notion of equivalence. For example, we may view $(k+1)$-point configurations as simplices, and our notion of equivalence is geometric congruence; any two simplices are congruent if there is a translation and a rotation that maps one onto the other. Since a $2$-simplex is simply a pair of points, and two such simplices are congruent if and only if the distance is the same, congruence classes simply correspond to distance. Hence, the Erdos-Falconer distance problem may be viewed as simply the $k=1$ case of the simplex congruence problem. In \cite{Ubiquity}, Hart and Iosevich prove that $E$ contains the vertices of a congruent copy of every non-degenerate simplex (non-degenerate here means the points are in general position) whenever $|E|\gtrsim q^{\frac{kd}{k+1}+\frac{k}{2}}$. However, in order for this result to be non-trivial the exponent must be $<d$, and that only happens when $\binom{k+1}{2}<d$. So, the result is limited to fairly small configurations. This result is improved in \cite{GroupAction} by Bennett, Hart, Iosevich, Pakianathan, and Rudnev, who prove that for any $k\leq d$ a set $E\subset\mathbb{F}_q^d$ determines a positive proportion of all congruence classes of $(k+1)$-point configurations provided $|E|\gtrsim q^{d-\frac{d-1}{k+1}}$. This exponent is clearly non-trivial for all $k$. In \cite{Me}, I extended this result to the case $k\geq d$. \\
In this paper, we consider a different notion of equivalence. We will consider the problem over both finite fields and rings of integers modulo powers of primes, so I will define the equivalence relation in an arbitrary ring.
\begin{dfn} Let $R$ be a ring, and let $E\subset R^2$. We define an equivalence relation $\sim$ on $E^{k+1}$ by $(x^1,...,x^{k+1})\sim (y^1,...,y^{k+1})$ (or more breifly $x\sim y$) if and only if for each pair $i,j$ we have $x^i\cdot x^{j\perp}=y^i\cdot y^{j\perp}$. Define $\mathcal{C}_{k+1}(E)$ to be the set of equivalence classes of $E$ under this relation. \end{dfn}
In the Euclidean setting, $\frac{1}{2} |x\cdot y^\perp|$ is the area of the triangle with vertices $0,x,y$. So, we may view each pair of points in a $(k+1)$-point configuration as determining a triangle with the origin, and we consider two such configurations to be equivalent if the triangles they determine all have the same areas. As we will prove in section $2$, this equivalence relation is closely related to the action of $\text{SL}_2(R)$ on tuples of points; except for some degenerate cases, two configurations are equivalent if and only if there is a unique $g$ mapping one to the other. This allows us to analyze the problem in terms of this action; in section 2, we define a counting function $f(g)$ and reduce matters to estimating the sum $\sum_g f(g)^{k+1}$. In section 3, we show how to turn an estimate for $\sum_g f(g)^2$ into an estimate for $\sum_g f(g)^{k+1}$. Since we already understand the $k=1$ case (it is essentially the same as the dot product problem discussed above), this reduction allows us to obtain a non-trivial result. Our first theorem is as follows.
\begin{thm} \label{MT1}
Let $q$ be a power of an odd prime, and let $E\subset\mathbb{F}_q^2$ satisfy $|E|\gtrsim q^s$, where $s={2-\frac{1}{k+1}}$. Then $\mathcal{C}_{k+1}(E)\gtrsim \mathcal{C}_{k+1}(\mathbb{F}_q^2)$. \end{thm}
In addition to proving this theorem, we will consider the case where the finite field $\mathbb{F}_q$ is replaced by the ring $\mathbb{Z}/p^\ell\mathbb{Z}$. The structure of the proof is largely the same; the dot product problem over such rings is studied in \cite{CIP}, giving us the $k=1$ case, and the machinery which lifts that case to arbitrary $k$ works the same way. However, many details in the proofs are considerably more complicated. The theorem is as follows.
\begin{thm} \label{MT2}
Let $p$ be an odd prime, let $\ell\geq 1$, and let $E\subset(\mathbb{Z}/p^\ell\mathbb{Z})^2$ satisfy $|E|\gtrsim \ell^{\frac{2}{k+1}}p^{\ell s}$, where $s=2-\frac{1}{\ell(k+1)}$. Then $\mathcal{C}_{k+1}(E)\gtrsim \mathcal{C}_{k+1}((\mathbb{Z}/p^\ell\mathbb{Z})^2).$ \end{thm}
We first note that, as we would expect, Theorem \ref{MT2} coincides with Theorem \ref{MT1} in the case $\ell=1$. We also note that, for fixed $p$ and $k$, the exponent in Theorem \ref{MT2} is always less than 2, but it tends to $2$ as $\ell\to\infty$. This does not happen in the finite field case, where the exponent depends on $k$ but not on the size of the field. \\
Finally, we want to state the extent to which these results are sharp. There are examples which show that the exponent must tend to $2$ as $k\to\infty$ in the finite field case, and as either $k\to\infty$ or $\ell\to\infty$ in the $\mathbb{Z}/p^\ell\mathbb{Z}$ case.
\begin{thm}[Sharpness] We have the following:
\begin{enumerate}[i]
\item For any $s<2-\frac{2}{k+1}$, there exists $E\subset\mathbb{F}_q^2$ such that $|E|\approx q^s$ and $\mathcal{C}_{k+1}(E)=o(\mathcal{C}_{k+1}(\mathbb{F}_q^2))$.
\item For any $s<2-\min\left(\frac{2}{k+1},\frac{1}{\ell}\right)$, there exists $E\subset(\mathbb{Z}/p^\ell\mathbb{Z})^2$ such that $|E|\approx p^{\ell s}$ and $\mathcal{C}_{k+1}(E)=o(\mathcal{C}_{k+1}((\mathbb{Z}/p^\ell\mathbb{Z})^2))$. \end{enumerate}
\end{thm}
\section{Characterization of the equivalence relation in terms of the $\text{SL}_2(R)$ action}
Our main tool in reducing the problem of $(k+1)$-point configurations to the $k=1$ case is the fact that we can express the equivalence relation in terms of the action of the special linear group; with some exceptions, tuples $x$ and $y$ are equivalent if and only if there exists a unique $g\in \text{SL}_2$ such that for each $i$, we have $y^i=gx^i$. In order to use this, we need to bound the number of exceptions to this rule. This is easy in the finite field case, and a little more tricky in the $\mathbb{Z}/p^\ell\mathbb{Z}$ case. The goal of this section is to describe and and bound the number of exceptional configurations in each case. We begin with a definition.
\begin{dfn} Let $R$ be a ring. A configuration $x=(x^1,...,x^{k+1})\in (R^2)^{k+1}$ is called \textbf{good} if there exist two indices $i,j$ such that $x^i\cdot x^{j\perp}$ is a unit. A configuration is \textbf{bad} if it is not good. \end{dfn}
As we will see, the good configurations are precisely those for which equivalence is determined by the action of $\text{SL}_2(R)$. To prove this, we will need the following theorems about determinants of matrices over rings, which can be found in \cite{DF}, section 11.4.
\begin{thm} \label{DF1} Let $R$ be a ring, let $A_1,...,A_n$ be the columns of an $n\times n$ matrix $A$ with entries in $R$. Fix an index $i$, and let $A'$ be the matrix obtained from $A$ by replacing column $A_i$ by $c_1A_1+\cdots+c_nA_n$, for some $c_1,...,c_n\in R$. Then $\det(A')=c_i\det(A)$. \end{thm}
\begin{thm} \label{DF2} Let $R$ be a ring, and let $A$ be an $n\times n$ matrix with entries in $R$. The matrix $A$ is invertible if and only if $\det(A)$ is a unit in $R$. \end{thm}
\begin{thm} \label{DF3} Let $R$ be a ring, and let $A$ and $B$ be $n\times n$ matrices with entries in $R$. Then $\det(AB)=\det(A)\det(B)$. \end{thm}
We are now ready to prove that equivalence of good configurations is given by the action of the special linear group.
\begin{lem} \label{SL2} Let $R$ be a ring, and let $x,y$ be good configurations such that $x^i\cdot x^{j\perp}=y^i\cdot y^{j\perp}$ for every pair of indices $i,j$. Then there exists a unique $g\in \text{SL}_2(R)$ such that $y^i=gx^i$ for each $i$. \end{lem}
\begin{proof} Because $x$ and $y$ are good, there exist indices $i$ and $j$ such that $x^i\cdot x^{j\perp}$ is a unit; equivalently, the determinant of the $2\times 2$ matrix with columns $x^i$ and $x^j$ is a unit. Denote this matrix by $(x^i\ x^j)$. By theorem \ref{DF2}, this matrix is invertible. Let
\[ g=(y^i\ y^j)(x^i\ x^j)^{-1}. \]
Since $g(x^i\ x^j)=(gx^i\ gx^j)$, it follows that $y^i=gx^i$ and $y^j=gx^j$. Also note that by Theorem \ref{DF3}, we have $\det(g)=1$. Let $n$ be any other index. We want to write $x^n=ax^i+bx^j$; this amounts to solving the matrix equation
\[ \begin{pmatrix} x_1^i & x_1^j \\ x_2^i & x_2^j \end{pmatrix} \begin{pmatrix} a \\ b \end{pmatrix} =x^n \]
Since we have already established the matrix $(x^i\ x^j)$ is invertible, we can solve for $a$ and $b$. Similarly, let $y^n=a'y^i+b'y^j$. By Theorem \ref{DF1}, we have $\det(x^i\ x^n)=b\det(x^i\ x^j)$ and $\det(y^i\ y^n)=b'\det(y^i\ y^j)$. It follows that $b=b'$, and an analogous argument yields $a=a'$. Therefore,
\[ gx^n=g(ax^i+bx^j)=agx^i+bgx^j=ay^i+by^j=y^n. \]
So, we have established existance. To prove uniqueness, note that $g$ must satisfy $g(x^i\ x^j)=(y^i\ y^j)$, and since $(x^i\ x^j)$ is invertible we can solve for $g$. \end{proof}
Now that we know that good tuples allow us to use the machinery we need, we must prove that the bad tuples are negligible.
\begin{lem} \label{countbad} Let $R$ be a ring and let $E\subset R^2$. We have the following: \begin{enumerate}[i]
\item If $R=\mathbb{F}_q$, then $E^{k+1}$ contains $\lesssim q^k|E|$ bad tuples. In particular, if $|E|\gtrsim q^{1+\e}$ for any constant $\e>0$, the number of bad tuples in $E^{k+1}$ is $o(|E|^{k+1})$.
\item If $R=\mathbb{Z}/p^\ell \mathbb{Z}$, the number of bad tuples in $R^{k+1}$ is $\lesssim p^{(2\ell-1)(k+1)+1}$. In particular, if $|E|\gtrsim p^{2\ell-1+\frac{1}{k+1}+\e}$ for any constant $\e>0$, then the number of bad tuples in $E^{k+1}$ is $o(|E|^{k+1})$. \end{enumerate} \end{lem}
\begin{proof} We first prove the first claim. Since the only non-unit of $\mathbb{F}_q$ is 0, a bad tuple must consist of $k+1$ points which all lie on a line through the origin. Therefore, we may choose $x^1$ to be anything in $E$, after which the next $k$ points must be chosen from the $q$ points on the line through the origin and $x^1$. \\
To prove the second claim, first observe that the number of tuples where at least one coordinate is a non-unit is $p^{2(\ell-1)(k+1)}$, which is less then the claimed bound. So, it suffices to bound the set of bad tuples where all coordinates are units. Let $B$ be this set. Define
\[ \psi(x_1^1,x_2^1,\cdots ,x_1^{k+1},x_2^{k+1})=(p^{\ell-1}x_1^1,x_2^1,\cdots ,p^{\ell-1}x_1^{k+1},x_2^{k+1}). \]
If $x\in B$, then $x^i\cdot x^{j\perp}$ is a non-unit, meaning it is divisible by $p$, and
\[ (p^{\ell-1}x_1^i,x_2^i)\cdot (p^{\ell-1}x_1^j,x_2^j)=p^{\ell-1}x^i\cdot x^{j\perp}=0. \]
Therefore, $\psi$ maps bad tuples $x$ to tuples $y$ with $y^i\cdot y^{j\perp}=0$, or $y_1^iy_2^j-y_1^jy_2^i=0$. Rearranging, using the fact that the second coordinate of each $y^i$ is a unit, we conclude that $\frac{y_1^i}{y_2^i}$ is a constant independent of $i$ which is divisible by $p^{\ell-1}$. In other words, each $y^i$ is on a common line through the origin and a point $(n,1)$ where $p^{\ell-1}|n$. There are $p$ such lines, and once we fix a line there are $p^{\ell(k+1)}$ choices of tuples $y$. Therefore, $|\psi(B)|\leq p\cdot p^{\ell(k+1)}$. Finally, we observe that the map $\psi$ is $p^{(\ell-1)(k+1)}$-to-1. This gives us the claimed bound on $|B|$. \end{proof}
\begin{lem} \label{flemma}
Let $R$ be either $\mathbb{F}_q$ or $\mathbb{Z}/p^\ell\mathbb{Z}$. Let $E\subset R^2$, and let $G\subset E^{k+1}$ be the set of good tuples. Suppose $|E|\gtrsim q^{1+\e}$ if $R=\mathbb{F}_q$ and $|E|\gtrsim p^{2\ell-1+\frac{1}{k+1}+\e}$ if $R=\mathbb{Z}/p^\ell\mathbb{Z}$. For $g\in \text{SL}_2(R)$, define $f(g)=\sum_x E(x)E(gx)$. Then
\[
|E|^{2(k+1)}\lesssim \mathcal{C}_{k+1}(E)\sum_{g\in\text{SL}_2(R)} f(g)^{k+1}. \] \end{lem}
\begin{proof} By Cauchy-Schwarz, we have
\[
|G|^2\leq |G/\sim|\cdot |\{(x,y)\in G\times G:x\sim y\}|. \]
By assumption and Lemma \ref{countbad}, $|E|^{k+1}\approx |G|$, and therefore the left hand side above is $\approx |E|^{2(k+1)}$. Since $G\subset E^{k+1}$ the right hand side above is $\leq \mathcal{C}_{k+1}(E)|\{(x,y)\in G\times G:x\sim y\}|$. It remains to prove $|\{(x,y)\in G\times G: x\sim y\}|\leq \sum_{g\in\text{SL}_2(R)} f(g)^{k+1}$. By lemma \ref{SL2},
\[
|\{(x,y)\in G\times G: x\sim y\}|=\sum_{x,y\in G}\sum_{\substack{g \\ y=gx}} 1. \]
By extending the sum over $G$ to one over all of $E^{k+1}$, we bound the above sum by
\begin{align*} &\sum_{x,y\in E^{k+1}}\sum_{\substack{g \\ y=gx}} 1 \\ =&\sum_x E(x^1)\cdots E(x^{k+1})\sum_g E(gx^1)\cdots E(gx^{k+1}) \\ =&\sum_g\left(\sum_{x^1} E(x^1)E(gx^1)\right)^{k+1} \\ =&\sum_g f(g)^{k+1} \end{align*}
\end{proof}
\section{Lifting $L^2$ estimates to $L^{k+1}$ estimates}
In both the case $R=\mathbb{F}_q$ and $R=\mathbb{Z}/p^\ell\mathbb{Z}$, results are known for pairs of points, which is essentially the $k=1$ case. The finite field version was studied in \cite{HI}, and the ring of integers modulo $p^\ell$ was studied in \cite{CIP}. In section 2, we defined a function $f$ on $\text{SL}_2(R)$ and related the number of equivalence classes determined by a set to the sum $\sum_g f(g)^{k+1}$. Since results are known for the $k=1$ case, we have information about the sum $\sum_g f(g)^2$. We wish to turn that into a bound for $\sum_g f(g)^{k+1}$. This is achieved with the following lemma.
\begin{lem} \label{induction} Let $S$ be a finite set, and let $F:S\to \mathbb{R}_{\geq 0}$. Let
\[
A=\frac{1}{|S|}\sum_{x\in S}F(x) \]
denote the average value of $F$, and
\[ M=\sup_{x\in S}F(x) \]
denote the maximum. Finally, suppose
\[
\sum_{x\in S}F(x)^2=A^2|S|+R. \]
Then there exist constants $c_k$, depending only on $k$, such that
\[
\sum_{x\in S}F(x)^{k+1}\leq c_k(M^{k-1}R+A^{k+1}|S|). \] \end{lem}
\begin{proof} We proceed by induction. For the base case, let $c_1=1$ and observe that the claimed bound is the one we assumed for $\sum_x F(x)^2$. Now, let $\{c_k\}$ be any sequence such that $k\binom{k}{j}c_j\leq c_k$ holds for all $j<k$; for example, $c_k=2^{k^2}$ works. Now, suppose the claimed bound holds for all $1\leq j<k$, and also observe that the bound is trivial for $j=0$. By direct computation, we have
\begin{align*} &\sum_{x\in S}(F(x)-A)^2 \\
=&\sum_{x\in S}F(x)^2-2A\sum_{x\in S}F(x)+A^2|S| \\
=&\sum_{x\in S}F(x)^2-A^2|S| \\ =&R. \end{align*}
We also have
\[ \sum_{x\in S}F(x)^{k+1}=\sum_{x\in S}(F(x)-A)^kF(x)+\sum_{j=0}^{k-1}\binom{k}{j}(-1)^{k-j+1}A^{k-j}\sum_{x\in S}F(x)^{j+1}. \]
To bound the first term, we simply use the trivial bound. Since $F(x)\leq M$ for all $x$, $A\leq M$, and $F(x),A\geq 0$, we conclude $|F(x)-A|\leq M$ for each $x$. Therefore,
\[ \sum_{x\in S}(F(x)-A)^kF(x)\leq M^{k-1}\sum_{x\in S}(F(x)-A)^2=M^{k-1}R. \]
To bound the second term, we use the inductive hypothesis and the triangle inequality. We have
\begin{align*}
&\left|\sum_{j=0}^{k-1}\binom{k}{j}(-1)^{k-j+1}A^{k-j}\sum_{x\in S}F(x)^{j+1}\right| \\ \leq &k\cdot\sup_{0\leq j<k} \binom{k}{j}A^{k-j}\sum_{x\in S}F(x)^{j+1} \\
\leq &k\cdot\sup_{0\leq j<k} \binom{k}{j}A^{k-j}c_j(M^{j-1}R+A^{j+1}|S|) \\
\leq & c_k \cdot\sup_{0\leq j<k}(A^{k-j}M^{j-1}R+A^{k+1}|S|) \end{align*}
Since $A\leq M$, it follows that $A^{k-j}M^{j-1}R\leq M^{k-1}R$ for any $j<k$, so the claimed bound holds. \end{proof}
\section{Some lemmas about the action of $\text{SL}_2(R)$}
\begin{lem} \label{action}
Let $G$ be a finite group acting transitively on a finite set $X$. Define $\ph:X\times X\to\mathbb{N}$ by $\ph(x,y)=|\{g\in G:gx=y\}|$. We have
\[
\ph(x,y)=\frac{|G|}{|X|} \]
for every pair $x,y$. If $h:X\to\mathbb{C}$ and $x_0\in X$, then
\[
\sum_{g\in G} h(gx_0)=\frac{|G|}{|X|}\sum_{x\in X}h(x). \]
\end{lem}
\begin{proof} The second statement follows from the first by a simple change of variables. To prove the first, we have
\[ \sum_{x,y\in X}\ph(x,y)=\sum_{g\in G}\sum_{\substack{x,y\in X \\ gx=y}}1. \]
On the right, for any fixed $g$, one can choose any $x$ and there is a unique corresponding $y$, so the inner sum is $|X|$ and the right hand side is therefore $|G||X|$. On the other hand, $\ph$ is constant. To prove this, let $x,y,z,w\in X$ and let $h_1,h_2\in G$ such that $h_1x=z$ and $h_2w=y$. This means for any $g$ with $gz=w$, we have $(g_2gh_1)x=y$, so $\ph(z,w)\leq \ph(x,y)$. By symmetry, equality holds. If $c$ is the constant value of $\ph(x,y)$, the left hand side above must be $c|X|^2$, and therefore $c=\frac{|G|}{|X|}$ as claimed. \end{proof}
\begin{lem} \label{SL2size}
We have $|\text{SL}_2(\mathbb{F}_q)|=q^3-q$ and $|\text{SL}_2(\mathbb{Z}/p^\ell\mathbb{Z})|=p^{3\ell}-p^{3\ell-2}$. \end{lem}
\begin{proof} We are counting solutions to the equation $ad-bc=1$ where $a,b,c,d\in\mathbb{F}_q$. We consider two cases. If $a$ is zero, then $d$ can be anything, and we must have $bc=1$. This means $b$ can be anything non-zero, and $c$ is determined. So, there are $q^2-q$ solutions with $a=0$. With $a\neq 0$, $b$ and $c$ can be anything, and $d$ is determined, giving $q^3-q^2$ solutions in this case. So, there are $(q^3-q^2)+(q^2-q)$ total solutions. \\
Next, we want to count solutions to $ad-bc=1$ with $a,b,c,d\in\mathbb{Z}/p^\ell\mathbb{Z}$. The arguments are essentially the same as in the proof of the finite field case, but slightly more complicated because there are non-zero elements which are still not units. We again consider separately two cases according to whether $a$ is a unit or not. If $a$ is a unit, then $b,c$ can be anything and then $d$ is determined, so there are $(p^{\ell}-p^{\ell-1})p^{2\ell}$ such solutions. If $a$ is not a unit, then $b$ and $c$ must be units, as otherwise $1$ would be divisible by $p$. So there are $p^{\ell-1}$ choices for $a$, $p^{\ell}$ choices for $d$, $p^{\ell}-p^{\ell-1}$ for $b$, and $c$ is determined. Putting this together, we get the claimed number of solutions. \end{proof}
\section{Proof of Theorem \ref{MT1}} We are now ready to prove theorem \ref{MT1}.
\begin{proof} First observe that good tuples are equivalent to $\approx q^{3}$ distinct tuples, so there are $\approx q^{2k-1}$ equivalence classes of good tuples. Since the only non-unit in the finite field case is 0, the bad tuples are all in the same equivalence class. So, our goal is to prove $\mathcal{C}_{k+1}(E)\gtrsim q^{2k-1}$. We first must prove the estimate
\[
\sum_g f(g)^2=\frac{|E|^4}{q}+O(q^2|E|^2). \]
We expand the sum on the left hand side and change variables to obtain
\[ \sum_g f(g)^2=\sum_{x^1,x^2,y^1,y^2}E(x^1)E(x^2)E(y^1)E(y^2)\left(\sum_{\substack{g \\ gx=y}}1\right). \]
We first observe we may ignore the pairs $x,y$ which are on a line through the origin. This is because if $x^2=tx^1$ and $y^2=sy^1$, there will exist $g$ with $gx=y$ if and only if $t=s$, in which case there are $\approx q$ choices for $g$. So, we have $|E|$ choices for $x^1$ and $y^1$, $q$ choices for $t$, and $\approx q$ choices for $g$ giving an error of $O(q^2|E|^2)$, as claimed. For all other pairs $x,y$, the inner sum in $g$ is 1 if $x\sim y$ and 0 otherwise. Therefore, if $\nu(t)=|\{(x,y)\in E\times E:x\cdot y^\perp=t\}|$, we have
\[
\sum_g f(g)^2=O(|E|^2q^2)+\sum_t\sum_{\substack{x^1, x^2, y^1, y^2 \\ x^1\cdot x^{2\perp}=t \\ y^1\cdot y^{2\perp}=t}}E(x^1)E(x^2)E(y^1)E(y^2)=O(|E|^2q^2) +\sum_t\nu(t)^2. \]
The proof of theorem 1.4 in \cite{HI} shows that $\nu(t)=\frac{|E|^2}{q}+O(|E|q^{1/2})$, so this gives
\[
\sum_t \nu(t)^2-\frac{|E|^4}{q}=\sum_t \left(\nu(t)-\frac{|E|^2}{q}\right)^2=O(|E|^2q^2), \]
which proves the equation above. We now apply lemma \ref{induction} with $F=f$. Lemmas \ref{SL2size} and \ref{action} imply
\[
A=\frac{1}{|\text{SL}_2(\mathbb{F}_q)|}\sum_xE(x)\sum_gE(gx)=\frac{1}{(q^2-1)}|E|^2=\frac{|E|^2}{q^2}+O\left(\frac{|E|^2}{q^4}\right) \]
and
\[
|S|=q^3+O(q^2). \]
Putting this together gives
\[
A^2|S|=\frac{|E|^4}{q}+O\left(\frac{|E|^4}{q^2}\right), \]
and therefore
\[
\sum_g f(g)^2=A^2|S|+R \]
with $R=O(q^2|E|^2)$. Finally, we observe that $f$ has maximum $M\leq |E|$. Therefore, lemma \ref{induction} gives
\[
\sum_g f(g)^{k+1}\lesssim q^2|E|^{k+1}+\frac{|E|^{2(k+1)}}{q^{2k-1}}. \]
Together with lemma \ref{flemma}, this gives
\[
|E|^{2(k+1)}\lesssim \mathcal{C}_{k+1}(E)\left(q^2|E|^{k+1}+\frac{|E|^{2(k+1)}}{q^{2k-1}}\right). \]
If the second term on the right is bigger, we get the result for free. If the first term is bigger, we get
\[
\mathcal{C}_{k+1}(E)\gtrsim \frac{|E|^{k+1}}{q^2}. \]
This will be $\gtrsim q^{2k-1}$ when $|E|\gtrsim q^{2-\frac{1}{k+1}}$, as claimed.
\end{proof}
\section{Size of $\mathcal{C}_{k+1}((\mathbb{Z}/p^\ell\mathbb{Z})^2)$}
Since $|\text{SL}_2(R)|\approx |R|^3$, we expect each tuple in $(R^2)^{k+1}$ to be equivalent to $\approx |R|^3$ other tuples, and therefore we expect the number of congruence classes to be $|R|^{2k-1}$. In the finite field case, this was proved as the first step of the proof of Theorem \ref{MT1}, but the proof in the $R=\mathbb{Z}/p^\ell\mathbb{Z}$ is more complicated so we will prove it here, separately from the proof of Theorem \ref{MT2} in the next section.
\begin{thm} \label{target} We have $\mathcal{C}_{k+1}((\mathbb{Z}/p^\ell\mathbb{Z})^2)\approx (p^\ell)^{2k-1}$. More precisely, the good $(k+1)$-point configurations of $(\mathbb{Z}/p^\ell\mathbb{Z})^2$ determine $\approx (p^\ell)^{2k-1}$ classes, and the bad configurations determine $o((p^\ell)^{2k-1})$ classes. \end{thm}
\begin{proof} We first establish that there are $\approx p^{\ell(2k-1)}$ classes of good tuples. This is easy; if $x$ is a good tuple, we have seen the map $g\mapsto gx$ is injective, so each class has size $\approx p^{3\ell}$ and there are $p^{2\ell (k+1)}$ tuples, meaning there are $p^{2\ell(k+1)-3\ell}$ classes. \\
It remains to bound the number of bad classes. We first establish the $k=1,2$ cases. When $k=1$, we want to prove there are $o(p^\ell)$ equivalence classes. This is clear, because in the $k=1$ case we are looking at pairs $(x^1,x^2)$ whose class is determined by the scalar $x^1\cdot x^{2\perp}$. The classes therefore correspond to the underlying set of scalars in $\mathbb{Z}/p^\ell\mathbb{Z}$, and the bad classes correspond to non-units. In the $k=2$ case, we are looking at triples $(x^1,x^2,x^3)$ whose class is determined by the three scalars $(x^1\cdot x^{2\perp},x^2\cdot x^{3\perp},x^3\cdot x^{1\perp})$. So, the space of equivalence classes can be identified with $((\mathbb{Z}/p^\ell\mathbb{Z})^2)^3$, and the bad classes correspond to triples of non-units. \\
For $k\geq 3$, we use the following theorem, which is really just a more specific version of Theorem \ref{DF2}, also found in \cite{DF}, chapter 11. \\
\begin{thm*}[\ref{DF2}'] For any $2\times 2$ matrix $A$, there exists a $2\times 2$ matrix $B$ with $AB=BA=(\det(A))I_2$, where $I_2$ is the $2\times 2$ identity matrix. \end{thm*}
We also make a more specific version of the definition of good and bad tuples. Namely, let$x$ be a $(k+1)$ point configuration in $\mathbb{Z}/p^\ell\mathbb{Z}$, and let $m\leq \ell$ be minimal with respect to the property that $p^m$ divides $x^i\cdot x^{j\perp}$ for every pair of indices $(i,j)$. We say that $x$ is \textbf{$m$-bad}. Observe that according to our previous definition, good tuples are $0$-bad and bad tuples are $m$-bad for some $m>0$. Also observe that $m$-badness is preserved by equivalence, so we may define $m$-bad equivalence classes analogously. An easy variant of the argument in Lemma \ref{countbad} shows that the number of $m$-bad tuples is $\lesssim p^{(2\ell-m)(k+1)+m}$; note that this bound can be rewritten as $p^{\ell(2k-1)+3\ell-km}$. We claim that every $m$-bad equivalence class has at least $p^{3\ell-2m}$ elements. It follows from the claim that there are $\lesssim p^{\ell(2k-1)+(2-k)m}$ $m$-bad classes, and since we may assume $k\geq 3$ the theorem follows from here. To prove the claim, note that the equivalence class containing $x$ also contains $gx$ for any $g\in\text{SL}_2(\mathbb{Z}/p^\ell\mathbb{Z})$, so for a lower bound on the size of a class we need to determine the size of the image of the map $g\mapsto gx$. First note that we may assume without loss of generality that each coordinate of $x^1$ is a unit. This is because given $x$ we can shift any factor of $p$ from $x^1$ onto each other vector $x^i$ and obtain another representative of the same equivalence class. Next, observe that if $x$ is $m$-bad and $gx=hx$, then by Theorem \ref{DF2}' we have $p^mg=p^mh$. It follows that $h=g+p^{\ell-m}A$ for some matrix $A$ with entries between $0$ and $p^m$. Using the fact that
\[ \det(A+B)=\det(A)+\det(B)+\mathcal{B}(A,B), \]
where $\mathcal{B}$ is bilinear, we conclude that if $h=g+p^{\ell-m}A$ and $\det(g)=\det(h)=1$, we must have
\[ 0=p^{2(\ell-m)}\det(A)+p^{\ell-m}\mathcal{B}(g,A). \]
Let $m'$ be the minimal power of $p$ which divides all entries of $A$. Since the entries of $g$ cannot all be divisible by $p$, it follows that $\ell-m+m'$ is the maximal power of $p$ which divides the second term above. Since $2(\ell-m+m')$ divides the first term, it follows that both terms must be 0 for the equation to hold. In particular, we must have $\mathcal{B}(g,A)=0$. Since at least one entry of $g$ must be a unit, we can solve for one entry of $A$ in terms of the others. Now observe that in order to have $gx=hx$, we must have $p^{\ell-m}Ax=0$. In particular, $p^{\ell-m}Ax^1=0$. Since each coordinate of $x^1$ is a unit, we may solve for another entry of the matrix $A$. This means there are at most $p^{2m}$ choices for $A$, and hence the map $g\mapsto gx$ is at most $p^{2m}$-to-one. It follows that $m$-bad classes have at least $p^{3\ell-2m}$ elements, as claimed.
\end{proof}
\section{Proof of Theorem \ref{MT2}}
\begin{proof} In keeping with the rest of this paper, the proof of the $\mathbb{Z}/p^\ell\mathbb{Z}$ case is essentially the same as the finite field case, but more complicated casework is required to deal with non-units. By our work in the previous section, our goal is to show $\mathcal{C}_{k+1}(E)\gtrsim (p^\ell)^{2k-1}$. Following the line of reasoning in the proof of Theorem \ref{MT1}, we want to establish the estimate
\[
\sum_g f(g)^2=\frac{|E|^4}{p^\ell}+O(\ell^2|E|^2(p^\ell)^{3-\frac{1}{\ell}}). \]
We have, after a change of variables,
\[ \tag{$*$} \sum_g f(g)^2=\sum_{x^1,x^2,y^1}E(x^1)E(x^2)E(y^1)\sum_{\substack{g \\ gx^1=y^1}}E(gx^2). \]
We first want to throw away terms where $x^1,y^1$ have non-units in their first coordinates. Note that there are $\approx p^{4\ell-2}$ such pairs. For each, there are $|E|$ many choices for $x^2$. We claim that there are $\leq p^\ell$ choices of $g$ which map $x^1$ to $y^1$ under this constraint. It follows from this claim that those terms contribute $\lesssim p^{5\ell-2}|E|$ to ($*$), which is less then the claimed error term. To prove the claim, observe that we are counting solutions to the system of equations
\begin{align*} ax_1^1+bx_2^1&=y_1^1 \\ cx_1^1+dx_1^1&=y_2^1\\ ad-bc&=1 \end{align*}
in $a,b,c,d$. Since $x_1^1$ is a unit, we can solve the first two equations for $a$ and $c$, respectively. Plugging these solutions into the third equation yields
\[ 1=\frac{y_1^1}{x_1^1}d-\frac{y_2^1}{y_1^1}b. \]
Since $y_1^1$ is a unit, for every $b$ there is a unique $d$ satisfying the equation. This proves the claim. Now, we want to remove all remaining terms from ($*$) corresponding to $x^1,x^2$ where $x^1\cdot x^{2\perp}$ is not a unit. To bound this contribution, we observe that for any such pair, we can write $x^2=tx^1+k$, where $0<t<p$ and $k$ is a vector where both entries are non-units. Therefore, there are $\leq |E|^2$ choices for $(x^1,y^1)$, there are $\leq p^{2\ell-1}$ choices for $x^2$, and there are $\leq p^\ell$ choices for $g$ as before. This gives the bound $|E|^2p^{3\ell-1}$, smaller than the claimed error term. This means, up to the error term, ($*$) can be written as
\[ \sum_{\substack{x^1,x^2,y^1,y^2 \\ x^1\cdot x^{2\perp}=y^1\cdot y^{2\perp}}}E(x^1)E(x^2)E(y^1)E(y^2)=\sum_t \nu(t)^2, \]
where $\nu(t)=|\{(x,y)\in E\times E:x\cdot y^\perp=t\}|$. This function was studied in \cite{CIP}; in that paper, it is proved that $\nu(t)=\frac{|E|^2}{q}+O(\ell|E|(p^\ell)^{\frac{1}{2}(2-\frac{1}{\ell})})$, leading to the claimed estimate for $\sum_g f(g)^2$, using the same reasoning as in the proof of Theorem \ref{MT1}. Applying Lemma \ref{induction} and Lemma \ref{flemma} with $A\approx \frac{|E|^2}{p^{2\ell}},|S|\approx p^{3\ell},M\leq |E|, R=O(\ell^2 |E|^2(p^\ell)^{3-\frac{1}{\ell}})$ gives
\[
|E|^{2(k+1)}\lesssim \mathcal{C}_{k+1}(E)\left(\ell^2|E|^{k+1}(p^\ell)^{3-\frac{1}{\ell}}+\frac{|E|^{2(k+1)}}{p^{\ell(2k-1)}}\right). \]
If the second term on the right is bigger, we get the result for free. If the first term is bigger, we have
\[
\mathcal{C}_{k+1}(E)\gtrsim \frac{|E|^{k+1}}{\ell^2p^{3\ell-1}}. \]
If $|E|\gtrsim \ell^{\frac{2}{k+1}}p^{\ell s}$, then this is $\gtrsim p^{\ell s(k+1)-3\ell+1}$, which is $\gtrsim p^{\ell(2k-1)}$ when $s\geq 2-\frac{1}{\ell(k+1)}$.
\end{proof}
\section{Proof of sharpness}
\begin{proof}
We first consider the finite field case. Let $1\leq s<2-\frac{2}{k+1}$, and let $E$ be a union of $q^{s-1}$ circles of distinct radii. Since each circle has size $\approx q$, this is a set of size $\approx q^s$. Observe that for any $x\in E^{k+1}$ and any $g$ in the orthogonal group $O_2(\mathbb{F}_q)$, we have $gx\in E^{k+1}$. Therefore, every configuration of points in $E$ is equivalent to at least $|O_2(\mathbb{F}_q)|\approx q$ other configurations. This means that
\[
\mathcal{C}_{k+1}(E)\lesssim q^{-1}|E|^{k+1}\approx q^{s(k+1)-1}=o(q^{2k-1}), \]
where in the last step we use the assumed bound on $s$. \\
Now, consider the $\mathbb{Z}/p^\ell\mathbb{Z}$ case. Let $1\leq s<2-\min\left(\frac{2}{k+1},\frac{1}{\ell}\right)$. We consider two different examples, according to which of $\frac{2}{k+1}$ or $\frac{1}{\ell}$ is smaller. In the first case, the example that works for finite fields also works here; circles still have size $\approx p^\ell$, so nothing is changed. In the second, let
\[ E=\{(t+pn,t+pm):0\leq t<p,0\leq m,n\leq p^{\ell-1}\}. \]
Clearly $|E|=p^{2\ell -1}=(p^\ell)^{2-\frac{1}{\ell}}$, but it is also easy to check that $x\cdot y^\perp$ is never a unit for any $x,y\in E$. Therefore, every configuration of points in $E$ is bad, and we have shown that this is $o(\mathcal{C}_{k+1}(p^{\ell(2k-1)}))$.
\end{proof}
\end{document}
|
arXiv
|
{
"id": "1906.04674.tex",
"language_detection_score": 0.7973494529724121,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\author{Olivia Dumitrescu} \title{Plane curves with prescribed triple points: a toric approach}
\maketitle
\begin{abstract} We will use toric degenerations of the projective plane ${{\mathbb{P}}^ 2}$ to give a new proof of the triple points interpolation problems in the projective plane. We also give a complete list of toric surfaces that are useful as components in this degeneration. \end{abstract}
\section{Introduction} Let ${\mathcal{L}}_{d}(m_{1},...,m_{r})$ denote the linear system of curves in ${{\mathbb{P}}^ 2}$ of degree $d$, that pass through $r$ points $P_{1},...,P_{r}$, with multiplicity at least $m_{i}$. A natural question would be to compute the projective dimension of the linear system ${\mathcal{L}}$. The virtual dimension of $\mathcal{L}$ is $$\ v ({\mathcal{L}}_{d}(m_1,...,m_r)):= \binom {d+2} {2}-\sum^{r}_{i=1} \binom {m_{i}+1} {2}-1 $$ and the expected dimension is $\ e ({\mathcal{L}}):=max\{v ({\mathcal{L}}),-1\}.$ There are some elementary cases for which $\ dim({\mathcal{L}}) \neq \ e ({\mathcal{L}}).$ A linear system for which $dim {\mathcal{L}}> e ({\mathcal{L}})$ is called a special linear system. However if we consider the homogeneous case when all multiplicities are equal $m_1=...=m_r=m$, the linear system (denoted by ${\mathcal{L}}_{d}(m^r)$) is expected to be non-special when $r$ is large enough. In this paper we will only consider the case $m=3$; the virtual dimension becomes $v({\mathcal{L}}_{d}(3^r))=\frac{d(d+3)}{2}-6m-1$. In \cite{CDM07} the authors used a toric degeneration of the Veronese surface into a union of projective planes for the double points interpolation problems i.e. $m=2$. This paper extends the degeneration used in \cite{CDM07} to the triple points interpolation problem. A triple point in the projective plane imposes six conditions, so in this paper we will classify the toric surfaces $(X, {\mathcal{L}})$ with $h^{0}(X, {\mathcal{L}})=6$ (see Theorem \ref{Classification}). In particular we will analyse the ones for which the linear system becomes empty when imposing a triple point, call them $Y_{i}$. We will then use a toric degeneration of the embedded projective space via a linear system ${\mathcal{L}}$ into a union of planes, quadircs and $r$ disjoint toric surfaces $Y_{i}$. On each surface $Y_{i}$ we will place one triple point and by a semicontinuity argument we will prove the non-speciality of ${\mathcal{L}}$, see Theorem \ref {triple plane}.\\\\ We remark that this result can be generelized to any dimension i.e. a list of toric varieties becoming empty when imposing a muliplicity $m$ point could be described in a similar way. However the combinatorical degeneration and the construction of the lifting function is not very well understood. In \cite{Len08} T. Lenarcik used an algebraic approach to study triple points interpolation in ${\mathbb{P}}^ {1}\times {\mathbb{P}}^ {1}$; however the list of the algebraic polygons is slightly different than ours and the connection with toric degenerations is not explicit. Toric degenerations of three dimensional projective space have been used by S. Brannetti to give a new proof of the Alexander-Hirschowitz theorem in dimension three in \cite{Brann08}. Degenerations of $n$-dimensional projective space have been used by E. Postinghel to give a new proof for the Alexander-Hirschowitz theorem in any dimension in \cite{Postinghel}.
\indent
\section{Toric varieties and toric degenerations.}\label{sec:toric}
We recall a few basic facts about toric degenerations of projective toric varieties. We refer to \cite {hu} and \cite{WF} for more information on the subject and to \cite {gath} for relations with tropical geometry.
The datum of a pair $(X,{\mathcal{L}})$, where $X$ is a projective, $n$--dimensional toric variety and ${\mathcal{L}}$ is a base point free, ample line bundle on $X$, is equivalent to the datum of an $n$ dimensional convex polytope ${\mathcal{P}}$ in ${\mathbb{R}}^ n$, determined up to translation and $SL_{n}^{\pm}({\mathbb{Z}})$. We consider a polytope ${\mathcal{P}}$ and a \emph{subdivision} ${\mathcal{D}}$ of ${\mathcal{P}}$ into convex subpolytopes; i.e. a finite family of $n$ dimensional convex polytopes whose union is ${\mathcal{P}}$ and such that any two of them intersect only along a face (which may be empty). Such a subdivision is called \emph{regular} if there is a convex, positive, linear on each face, function $F$ defined on ${\mathcal{P}}$. Such function $F$ will be called a lifting function. Regular subdivisions correspond to degeneration of toric varieties.
We will now prove a technical lemma that will enable us to easily demonstrate the existence of a lifting function when we need them. Let $X$ be a toric surface and ${\mathcal{P}}$ be its associated polygon. Consider $L$ a line that separates two disjoint polygons ${\mathcal{P}_1}$ and ${\mathcal{P}_2}$ in ${\mathcal{P}}$ and $X_{1}$ and $X_{2}$ their corresponding toric surfaces such that $L$ does not contain any integer point. \begin{lemma} The toric variety $X$ degenerates into a union of toric surfaces two of which are $X_{1}$ and $X_{2}$ which are skew. \end{lemma}
\begin{proof} We consider the convex piecewise linear function given by $$f(x,y,z)=\max \{z, L+z \}$$ Consider the image of the points on the boundary of the polygons $X_{1}$ and $X_{2}$ through $f$. Change the function $f$ by interposing the convex hull of the boundary points separated by $L$ between $l_{1}$ and $l_{2}$ (as in the Figure \ref{lift1}). The function will still be convex and piecewise linear, therefore we get a regular subdivision. We consider now the toric varieties associated to each polygon, and since ${\mathcal{P}_{1}}$ and ${\mathcal{P}_{2}}$ are disjoint, we obtain that two of the toric surfaces that appeared in the degeration, namely $X_1$ and $X_2$ are skew. \end{proof}
For example, in the picture below we have four polygons, two of which are disjoint. The corresponding degeneration will contain four toric varieties, two of them $X_{1}$ and $X_{2}$, being skew. \begin{figure}\label{lift1}
\end{figure}
It is easy to see how we could iterate this process. Let $M$ be a line cutting the polygon associated to $X_{2}$ and not containing any of its interior points. Then $X_{2}$ degenerates into a union of toric surfaces, two of which are skew, $Y_{2}$ and $Y_{3}$.
In this case, we conclude that $X$ degenerates into nine toric surfaces three of which $X_{1}$, $Y_{2}$ and $Y_{3}$ being skew, as the Figure \ref{lift2} indicates. \begin{figure}\label{lift2}
\end{figure}
Later on, we will ignore the varieties lying in between the disjoint ones; they are only important for the degeneration and not for the analysis itself.
\section{The Classification of polygons.}
We recall that the group $SL_2^{\pm}({\mathbb{Z}})$ acts on the column vectors of ${\mathbb{R}}^ 2$ by left multiplication. This induces an action of $SL_2^{\pm}({\mathbb{Z}})$ on the set of convex polygons ${\mathcal{P}}$ by acting on its enclosed points ($SL_2^{-}({\mathbb{Z}})$ corresponds to orientation reversing lattice equivalences). Obviously, ${\mathbb{Z}^{2}}$ acts on vectors of ${\mathbb{R}}^ 2$ by translation.
Next, we will classify all convex polygons enclosing six lattice integer points modulo the actions described above. We first start with a definition.
\begin{definition} We say the polygon ${\mathcal{P}}$ is equivalent to one in \it{standard position} if \begin{my_enumerate} \item It contains $O=(0,0)$ as a vertex \item $OS$ is a vertex where $S=(0,m)$ and $m$ is the largest edge length \item $OP$ is an edge where $P=(p,q)$ and $0\leq p <q$ \end{my_enumerate} \end{definition}
\begin{remark} Every polygon has a standard position. \end{remark}
Indeed, we first choose the longest edge and then we translate one of its vertices to the origin. We will now rotate the polygon to put the longest edge on the positive side of the $x$ axis and then we shift it such that the adjacent edge lies in the upper half of the first quadrant. Indeed, if $OP$ is an edge with $P=(s,q)$ and $s\geq q$; then $s=mq+p$ for $0\leq p<q$ so we shift left by $m$. We will call this procedure normalization.\\
It is easy to see that the standard position of the polygon may not be unique, it depends on the choice of the longest edge, and of the choice of the special vertex that becomes the origin.
We can now present the classification of the polygons in standard position according to $m$ (the maximum number of integral points lying on the edges of the polygon), and also according to their number of edges, $n$. Obviously, the polygons ${\mathcal{P}}$ will have at most six edges, and at most five points on an edge.
We leave the elementary details to the reader; we only remark that Pick's lemma is useful (for more details see \cite{Dum10}).
\begin{proposition}\label{Classification} Any polygon enclosing six lattice points is equivalent to exactly one from the following list
\end{proposition} We now recall that any rational convex polygon ${\mathcal{P}}$ in ${\mathbb{R}^{n}}$ enclosing a fixed number of integer lattice points defines an $n$ dimensional projective toric variety $X_{\mathcal{P}}$ endowed with an ample line bundle on $X_{\mathcal{P}}$ which has the integer points of the polygon as sections. We get the following result
\begin{corollary} Any toric surface endowed with an ample line bundle with six sections is completely described by exactly one of the polygons from the above list.
\end{corollary}
\section{Triple Point Analysis.}
We first observe that six, the number of integer points enclosed by the polygon, represents exactly the number of conditions imposed by a triple point. We will now classify all polygons from Proposition \ref{Classification} for which their corresponding linear system becomes empty when imposing a triple point. There are two methods for testing the emptiness of these linear systems: an algebraic method and a geometric method and we will briefly describe them below. For the algebraic approach, checking that a linear system is non-empty when imposing a triple point reduces to showing that the conditions imposed by a triple point in ${\mathbb{P}^{2}}$ are dependent. For this, one needs to look at the rank of the six by six matrix where the first column represents the sections of the line bundle and the other five columns represent all first and second derivatives in $x$ and $y$. We conclude that the six conditions are dependent if and only if the determinant of the matrix is identically zero.
The geometric method for testing when a planar linear system is empty is to explicitly find it and show that it contains no curve, using ${\mathbb{P}^{2}}$ as a minimal model for the surface $X$ and writing its resolution of singularities.
\begin{remark}\label{Elimination} The corresponding linear systems of the following polygons are non-empty when imposing a base point with multiplicity three.
\end{remark}
\begin{proof} It is easy to check that the algebraic conditions imposed by at least four sections on a line are always dependent. Indeed, we have two possible cases, if the line of sections is an edge, or if is enclosed by the polygon. For the first case, we can only have sections on two levels so the vanishing of the second derivative in $y$ gives a dependent condition (The same argument applies for case $m=3$ representing the embedded ${\mathbb{P}^{1}}\times {\mathbb{P}^{1}}$). For the second case we notice that the vanishing of the first derivative in $y$ and the second derivative in $x$ and $y$ give two linearly dependent conditions. \end{proof}
We will use the Remark \ref{Elimination} to eliminate the polygons that don't have the desired property and we obtain five polygons for which we will study the corresponding algebraic surfaces and linear systems using toric geometry methods.
For any polygon consider its fan by dualizing the polygon's angles and in the case that the toric variety obtained by gluing the cones is singular, take it's resolution of singularites. In this way we obtain the all the toric surfaces using ${\mathbb{P}^{2}}$ as a minimal model. The associated linear system may not have general points. In general, we will use the notations ${\mathcal{L}}_{d}([1,1]), {\mathcal{L}}_{d}([2,1]), {\mathcal{L}}_{d}([1,1,1])$ for linear systems of degree $d$ that pass through a base point with a defined tangent, a double point with a defined tangent or having a flex direction. For example, ${\mathcal{L}}_{4}([1,1,1]^3)$ represents quartics with three base points that are flex to the line joining any two of them. Since the base points are special, the linear systems will need a different analysis. We conclude the emptiness of each linear systems with a triple point by applying birational transformations and splitting off $-1$ curves. The last column of the table indicates the geometric conditions that corresponding to the infinitely near multiplicities. We obtain the following result:
\begin{lemma}\label {Empty poly} The linear systems corresponding to the following polygons become empty after imposing a triple point. \end{lemma}
All the linear systems from the table become empty when imposing a triple point. For example the fourth polygon in the above table describes a projective plane embedded by a linear system of conics, ${\mathcal{L}}_{2}$. By imposing a triple point ${\mathcal{L}}_{2}$ becomes empty. One can obtain more polygons with an empty linear system by rotating or by shifting the main ones by any integer numbers.
\section{Triple points in ${\mathbb{P}}^ 2$.}
We denote by $V_d$ the image of the Veronese embedding $v_d:\mathbb{P}^2 \to \mathbb{P}^{d(d+3)/2}$ that transforms the plane curves of degree $d$ to hyperplane sections of the Veronese variety $V_d$. We degenerate $V_d$ into a union of disjoint surfaces and ordinary planes and we place one point on each one of the disjoint surfaces. The surfaces are chosen such that the restriction of a hyperplane section to each one of them to be linear system that becomes empty when we impose a triple point. We conclude that any hyperplane section to $V_d$ needs to contain all disjoint surfaces, and in particular all of the coordinate points of the ambient projective space covered in this way. Therefore if $V_{d}$ degenerates exactly into a union of disjoint special surfaces and planes (or quadrics) with no points left over we conclude that the desired linear system is empty, and therefore it has the expected dimension. Using semicontinuity this argument can easily be extended to any degeneration as in \cite{CDM07}. In order to give an inductive proof for triple points in the projective plane we will first analyze triple points in ${\mathbb{P}}^ 1\times{\mathbb{P}}^ 1$.
We will only prove the most difficult case when the linear systems in ${\mathbb{P}}^ 1\times{\mathbb{P}}^ 1$ with virtual dimension $-1$ are empty. The general case will follow by induction, but it was already proved in a similar way using algebraic methods by T. Lenarcik in \cite{Len08}.
\begin{lemma}\label{bidegree $(5,n)$} Fix $n\geq 3$. Then linear systems of bidegree $(5,n)$ for $n\neq 4, (11,n), (2, 4n-9), (8,2n-3)$ and an arbitrary number of triple points have the expected dimension. \end{lemma}
\begin{proof} \begin{itemize} \item For any linear systems of bidegree $(5,n)$ we find a skew $n+1$ set of surfaces and we place each of the $n+1$ triple points in one of the surfaces. We denote the degenerations presented below as $C_{5}^{5}$, $C_{5}^{6}$, $C_{5}^{8}$ and $C_{5}^{3}$ For every $n>2$, $n\neq 4$ take $i\in \{3,5,6,8 \}$ such that $\frac{n-i}{4}$ is an integer, $k$. For any arbitrary $n$ we consider the degeneration $C_{5}^{n}=C_{5}^{i}+k C_{5}^{3}$ where the sum of two blocks means attaching the two disjoint blocks together along the edge of length $5$.
\item For linear systems of bidegree $(11,n)$ and 2$n+2$ triple points we find skew surfaces. We denote the degenerations presented below by $C_{11}^{2}$, $C_{11}^{3}$, and $C_{11}^{4}.$ For every $n>2$ take $i\in \{2,3,4\}$ such that $\frac{n-i}{3}$ is an integer, $k$. For any arbitrary $n$ we consider the degeneration $C_{11}^{n}=C_{11}^{i}+k C_{11}^{2}.$
\item For curves of bidegree $(2,4n-9)$ we consider the degeneration $C_{2}^{4n-9}$ given by $(n-3)C_{2}^{3}$ (in particular, $C_{2}^{11}=3C_{2}^{3}$) and for $C_{8}^{2n-3}$ we use combinations of $C_{8}^{3}$ and $C_{8}^{5}$
\end{itemize} \end{proof}
\begin{corollary}\label{main} Linear systems in ${\mathbb{P}}^ {1}\times {\mathbb{P}}^ {1}$ with triple points of virtual dimension $-1$ are empty. \end{corollary}
\begin{proof} We have to prove the statement for linear systems of bidegree $(6k-1, n)$ and $(3k-1, 2n-1).$ For the bidegree $(6k-1, n)$ we distinguish two cases. If $k$ is even $k=2k'$ we use the degeneration $C_{12k'-1}^{n}=k'C_{11}^{n}$; while if $k$ is odd of the form $2k'+1$ we use $C_{12k'+5}^{n}=C_{5}^{n}+k'C_{11}^{n}$, for $n\neq 4$. For $n=4$ we use the following degeneration for $C_{17}^{4}$ and we generalize this case by adding $C_{11}^{4}$ blocks For the bidegree $(3k-1, 2n-1)$ we reduce to the case when $k$ is odd of the form $2k'+1$ and depending on the parity of $k'$, if $6k'=6+12r$ we use the degeneration $C_{8}^{2n-1}+rC_{11}^{2n-1}$ while $6k'=12r$ we use $C_{5}^{2n-1}+C_{8}^{2n-1}+(r-1)C_{11}^{2n-1}$. \end{proof}
We can now obtain the following result
\begin{theorem}\label{triple plane} $\mathcal{L}_d(3^n)$ has the expected dimension whenever $d \geq 5$. \end{theorem}
\begin{proof} It is enough to prove the theorem for the number of triple points for which the virtual dimension is $-1$ so in that case we claim that the linear system is empty. Note that $\binom{d+2}{2}\equiv 0$ mod $6$ if $d\equiv \{2,7, 10, 11\}$ mod $6$; $\binom{d+2}{2}\equiv 1$ mod $6$ if $d\equiv \{0,9\}$ mod $6$; $\binom{d+2}{2}\equiv 3$ mod $6$ if $d\equiv \{1, 4, 5, 8\}$ mod $6$ and $\binom{d+2}{2}\equiv 4$ mod $6$ if $d\equiv \{3, 6\}$ mod $6$.
We will use the induction step $V_{12(k+1)+j}=V_{12k+j}+kC^{11}_{11}+C^{11}_{j+1}+V_{10}$ with $j=1,...,12$, $k\geq 0$, $(i,j)\neq (1,4)$ and to finish the proof we present the degenerations of $V_{j}$ if $j\leq 12.$
\end{proof}
\begin{remark} {\rm Notice that ${\mathcal{L}_4(3^2)}$ consists of quartics with two triple points and the expected dimension is $2$. This linear system has a fixed part, the double line through the two points and a movable part ${\mathcal{L}_2(1^2)}$ i.e. conics through two points, that has dimension $3$. A simple argument shows that if $d=4$, the linear system ${\mathcal{L}}$ is $-1$--special (we have a $-1$--curve, line connecting the 2 points, splitting off twice) and therefore special.\\ \indent One could mention that case $d=4$ is also a special case for the double points interpolation problem since ${\mathcal{L}_4(2^5)}$ is expected to be empty but it consists of the double conic determined by the $5$ general points. } \end{remark}
\end{document}
|
arXiv
|
{
"id": "1104.1755.tex",
"language_detection_score": 0.838331401348114,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{ A Unified Convergence Rate Analysis of The Accelerated Smoothed Gap Reduction Algorithm
}
\titlerunning{ A Unified Convergence Rate Analysis of The ASGARD Algorithm }
\author{Quoc Tran-Dinh$^{*}$}
\authorrunning{Q. Tran-Dinh}
\institute{Quoc Tran-Dinh \at
Department of Statistics and Operations Research\\
The University of North Carolina at Chapel Hill\\
333 Hanes Hall, UNC-CH, Chapel Hill, NC27599. \\
\email{[email protected]} }
\date{Received: date / Accepted: date}
\maketitle
\begin{abstract} In this paper, we develop a unified convergence analysis framework for the \textit{Accelerated Smoothed GAp ReDuction algorithm} (ASGARD) introduced in \cite[\textit{Tran-Dinh et al}]{TranDinh2015b}. Unlike \cite{TranDinh2015b}, the new analysis covers three settings in a single algorithm: general convexity, strong convexity, and strong convexity and smoothness. Moreover, we establish the convergence guarantees on three criteria: (i) gap function, (ii) primal objective residual, and (iii) dual objective residual. Our convergence rates are optimal (up to a constant factor) in all cases. While the convergence rate on the primal objective residual for the general convex case has been established in \cite{TranDinh2015b}, we prove additional convergence rates on the gap function and the dual objective residual. The analysis for the last two cases is completely new. Our results provide a complete picture on the convergence guarantees of ASGARD. Finally, we present four different numerical experiments on a representative optimization model to verify our algorithm and compare it with Nesterov's smoothing technique.
\keywords{ Accelerated smoothed gap reduction \and primal-dual algorithm \and Nesterov's smoothing technique \and convex-concave saddle-point problem. }
\subclass{90C25 \and 90C06 \and 90-08} \end{abstract}
\section{Introduction}\label{sec:intro}
We consider the following classical convex-concave saddle-point problem: \begin{equation}\label{eq:saddle_point_prob} \min_{x\in\mathbb{R}^p}\max_{y\in\mathbb{R}^n}\Big\{ \mathcal{L}(x, y) := f(x) + \iprods{Kx, y} - g^{*}(y) \Big\}, \end{equation} where $f : \mathbb{R}^p \to \mathbb{R}\cup\{+\infty\}$ and $g : \mathbb{R}^n\to\mathbb{R}\cup\{+\infty\}$ are proper, closed, and convex, $K : \mathbb{R}^p \to \mathbb{R}^n$ is a linear operator, and $g^{*}(y) := \sup_{x}\set{ \iprods{y, x} - g(x)}$ is the Fenchel conjugate of $g$.
The convex-concave saddle-point problem \eqref{eq:saddle_point_prob} can be written in the primal and dual forms. The primal problem is defined as \begin{equation}\label{eq:primal_prob} F^{\star} := \min_{x\in\mathbb{R}^p}\Big\{F(x) := f(x) + g(Kx) \Big\}. \end{equation} The corresponding dual problem is written as \begin{equation}\label{eq:dual_prob} D^{\star} := \min_{y\in\mathbb{R}^n}\Big\{ D(y) := f^{*}(-K^{\top}y) + g^{*}(y) \Big\}. \end{equation} Clearly, both the primal problem \eqref{eq:primal_prob} and its dual form \eqref{eq:dual_prob} are convex.
\noindent\textit{\textbf{Motivation.}} In \cite{TranDinh2015b}, two accelerated smoothed gap reduction algorithms are proposed to solve \eqref{eq:primal_prob} and its special constrained convex problem. Both algorithms achieve optimal sublinear convergence rates (up to a constant factor) in the sense of black-box first-order oracles \cite{Nemirovskii1983,Nesterov2004}, when $f$ and $g$ are convex, and when $f$ is strongly convex and $g$ is just convex, respectively. The first algorithm in \cite{TranDinh2015b} is called ASGARD (\textit{Accelerated Smoothed GAp ReDuction}). To the best of our knowledge, except for a special case \cite{TranDinh2012a}, ASGARD was the first primal-dual first-order algorithm that achieves a non-asymptotic optimal convergence rate on the last primal iterate. ASGARD is also different from the alternating direction method of multipliers (ADMM) and its variants, where it does not require solving complex subproblems but uses the proximal operators of $f$ and $g^{*}$. However, ASGARD (i.e. \cite[Algorithm 1]{TranDinh2015b}) only covers the general convex case, and it needs only one proximal operation of $f$ and of $g^{*}$ per iteration. To handle the strong convexity of $f$, a different variant is developed in \cite{TranDinh2015b}, called ADSGARD, but requires two proximal operations of $f$ per iteration. Therefore, the following natural question is arising: \begin{itemize}
\item[] ``\textit{Can we develop a unified variant of ASGARD that covers three settings: general convexity, strong convexity, and strong convexity and smoothness?}'' \end{itemize}
\noindent\textit{\textbf{Contribution.}} In this paper, we affirmatively answer this question by developing a unified variant of ASGARD that covers the following three settings: \begin{itemize} \item[] Case 1: Both $f$ and $g^{*}$ in \eqref{eq:saddle_point_prob} are only convex, but not strongly convex. \item[] Case 2: Either $f$ or $g^{*}$ is strongly convex, but not both $f$ and $g^{*}$. \item[] Case 3: Both $f$ and $g^{*}$ are strongly convex. \end{itemize} The new variant only requires one proximal operation of $f$ and of $g^{*}$ at each iteration as in the original ASGARD and existing primal-dual methods, e.g., in \cite{boct2020variable,Chambolle2011,Chen2013a,Condat2013,Esser2010,connor2014primal,Goldstein2013,vu2013variable}. Our algorithm reduces to ASGARD in Case 1, but uses a different update rule for $\eta_k$ (see Step~\ref{step:i2} of Algorithm~\ref{alg:A1}) compared to ASGARD. In Case 2 and Case 3, our algorithm is completely new by incorporating the strong convexity parameter $\mu_f$ of $f$ and/or $\mu_{g^{*}}$ of $g^{*}$ in the parameter update rules to achieve optimal $\BigO{1/k^2}$ sublinear and $(1 - \BigO{1/\sqrt{\kappa_F}})$ linear rates, respectively, where $k$ is the iteration counter and $\kappa_F := \norms{K}^2/(\mu_f\mu_{g^{*}})$.
In terms of convergence guarantees, we establish that, in all cases, our algorithm achieves optimal rates on the last primal sequence $\sets{x^k}$ and an averaging dual sequence $\set{\tilde{y}^k}$. Moreover, our convergence guarantees are on three different criteria: (i) the gap function for \eqref{eq:saddle_point_prob}, (ii) the primal objective residual $F(x^k) - F^{\star}$ for \eqref{eq:primal_prob}, and (iii) the dual objective residual $D(\tilde{y}^k) - D^{\star}$ for \eqref{eq:dual_prob}. Our paper therefore provides a unified and full analysis on convergence rates of ASGARD for solving three problems \eqref{eq:saddle_point_prob}, \eqref{eq:primal_prob}, and \eqref{eq:dual_prob} simultaneously.
We emphasize that primal-dual first-order methods for solving \eqref{eq:saddle_point_prob}, \eqref{eq:primal_prob}, and \eqref{eq:dual_prob}, and their convergence analysis have been well studied in the literature. To avoid overloading this paper, we refer to our recent works \cite{TranDinh2015b,tran2019non} for a more thorough discussion and comparison between existing methods. Hitherto, there have been numerous papers studying convergence rates of primal-dual first-order methods, including \cite{boct2020variable,Chambolle2011,chambolle2016ergodic,Chen2013a,Davis2014a,He2012,valkonen2019inertial}. However, the best known and optimal rates are only achieved via averaging or weighted averaging sequences, which are also known as ergodic rates. The convergence rates on the last iterate sequence are often slower and suboptimal. Recently, the optimal convergence rates of the last iterates have been studied for primal-dual first-order methods, including \cite{TranDinh2015b,tran2019non,valkonen2019inertial}. As pointed out in \cite{Chambolle2011,Davis2015,sabach2020faster,TranDinh2015b,tran2019non}, the last iterate convergence guarantee is very important in various applications to maintain some desirable structures of the final solutions such as sparsity, low-rankness, or sharp-edgedness in images. This also motivates us to develop ASGARD.
\noindent\textit{\textbf{Paper outline.}} The rest of this paper is organized as follows. Section~\ref{sec:preliminaries} recalls some basic concepts, states our assumptions, and characterizes the optimality condition of \eqref{eq:saddle_point_prob}. Section~\ref{sec:alg_scheme} presents our main results on the algorithm and its convergence analysis. Section~\ref{sec:num_exp} provides a set of experiments to verify our theoretical results and compare our method with Nesterov's smoothing scheme in \cite{Nesterov2005c}. Some technical proofs are deferred to the appendix.
\section{Basic Concepts, Assumptions, and Optimality Condition}\label{sec:preliminaries}
We are working with Euclidean spaces, $\mathbb{R}^p$ and $\mathbb{R}^n$, equipped with the standard inner product $\iprods{\cdot,\cdot}$ and the Euclidean norm $\norms{\cdot}$. We will use the Euclidean norm for the entire paper. Given a proper, closed, and convex function $f$, we use $\dom{f}$ and $\partial{f}(x)$ to denote its domain and its subdifferential at $x$, respectively. We also use $\nabla{f}(x)$ for a subgradient or the gradient (if $f$ is differentiable) of $f$ at $x$. We denote by $f^{*}(y) := \sup\set{\iprod{y, x} - f(x) : x\in\dom{f}}$ the Fenchel conjugate of $f$. We denote by $ \mathrm{ri}(\mathcal{X})$ the relative interior of $\mathcal{X}$.
A function $f$ is called $M_f$-Lipschitz continuous if $\vert f(x) - f(\tilde{x})\vert \leq M_f\norms{x - \tilde{x}}$ for all $x, \tilde{x} \in \dom{f}$, where $M_f \in [0, +\infty)$ is called a Lipschitz constant. A proper, closed, and convex function $f$ is $M_f$-Lipschitz continuous if and only if $\partial{f}(\cdot)$ is uniformly bounded by $M_f$ on $\dom{f}$. For a smooth function $f$, we say that $f$ is $L_f$-smooth (or Lipschitz gradient) if for any $x, \tilde{x}\in\dom{f}$, we have $\Vert\nabla{f}(x) - \nabla{f}(\tilde{x})\Vert \leq L_f\norms{x - \tilde{x}}$, where $L_f \in [0, +\infty)$ is a Lipschitz constant. A function $f$ is called $\mu_f$-strongly convex with a strong convexity parameter $\mu_f \geq 0$ if $f(\cdot) - \frac{\mu_f}{2}\norms{\cdot}^2$ remains convex. For a proper, closed, and convex function $f$, $\textrm{prox}_{\gamma f}(x) := \mathrm{arg}\!\min\big\{f(\tilde{x}) + \tfrac{1}{2\gamma}\norms{\tilde{x} - x}^2 : \tilde{x}\in\dom{f} \big\}$ is called the proximal operator of $\gamma f$, where $\gamma > 0$.
\subsection{\textbf{Basic assumptions and optimality condition}}\label{subsec:primal_dual_formulation}
In order to show the relationship between \eqref{eq:saddle_point_prob}, \eqref{eq:primal_prob} and its dual form \eqref{eq:dual_prob}, we require the following assumptions.
\begin{assumption}\label{as:A1} The following assumptions hold for \eqref{eq:saddle_point_prob}. \begin{itemize} \item[$\mathrm{(a)}$] Both functions $f$ and $g$ are proper, closed, and convex on their domain. \item[$\mathrm{(b)}$] There exists a saddle-point $z^{\star} := (x^{\star}, y^{\star})$ of $\mathcal{L}$ defined in \eqref{eq:saddle_point_prob}, i.e.: \begin{equation}\label{eq:saddle_point} \mathcal{L}(x^{\star}, y) \leq \mathcal{L}(x^{\star}, y^{\star}) \leq \mathcal{L}(x, y^{\star}), \quad \forall (x, y)\in\dom{f}\times\dom{g^{*}}. \end{equation} \item[$\mathrm{(c)}$] The Slater condition $0 \in \mathrm{ri}(\dom{g} - K\dom{f})$ holds. \end{itemize} \end{assumption} Assumption~\ref{as:A1} is standard in convex-concave saddle-point settings. Under Assumption~\ref{as:A1}, strong duality holds, i.e. $F^{\star} = \mathcal{L}(x^{\star}, y^{\star}) = -D^{\star}$.
To characterize a saddle-point of \eqref{eq:saddle_point_prob}, we define the following gap function: \begin{equation}\label{eq:gap_func} \mathcal{G}_{\mathcal{X}\times\mathcal{Y}}(x, y) := \sup\big\{ \mathcal{L}(x, \tilde{y}) - \mathcal{L}(\tilde{x}, y) : \tilde{x}\in\mathcal{X}, \ \tilde{y}\in\mathcal{Y} \big\}, \end{equation} where $\mathcal{X}\subseteq\dom{f}$ and $\mathcal{Y}\subseteq\dom{g^{*}}$ are nonempty, closed, and convex subsets such that $\mathcal{X}\times\mathcal{Y}$ contains a saddle-point $(x^{\star},y^{\star})$ of \eqref{eq:saddle_point_prob}. Clearly, we have $\mathcal{G}_{\mathcal{X}\times\mathcal{Y}}(x, y) \geq 0$ for all $(x, y)\in\mathcal{X}\times\mathcal{Y}$. Moreover, if $(x^{\star}, y^{\star})$ is a saddle-point of \eqref{eq:saddle_point_prob} in $\mathcal{X}\times \mathcal{Y}$, then $\mathcal{G}_{\mathcal{X}\times\mathcal{Y}}(x^{\star}, y^{\star}) = 0$.
\subsection{\textbf{Smoothing technique for $g$}}\label{subsec:proximal_smoothing}
We first smooth $g$ in \eqref{eq:primal_prob} using Nesterov's smoothing technique \cite{Nesterov2005c} as \begin{equation}\label{eq:smoothed_g} g_{\beta}(u, \dot{y}) := \max_{y\in\mathbb{R}^n}\set{\iprods{u, y} - g^{*}(y) - \tfrac{\beta}{2}\norms{y - \dot{y}}^2}, \end{equation} where $g^{*}$ is the Fenchel conjugate of $g$, $\beta > 0$ is a smoothness parameter, and $\dot{y}$ is a given proximal center. We denote by $\nabla_ug_{\beta}(u) = \textrm{prox}_{g^{*}/\beta}(\dot{y} + \frac{1}{\beta}u)$ the gradient of $g_{\beta}$ w.r.t. $u$.
Given $g_{\beta}$ defined by \eqref{eq:smoothed_g}, we can approximate $F$ in \eqref{eq:primal_prob} by \begin{equation}\label{eq:F_beta} F_{\beta}(x, \dot{y}) := f(x) + g_{\beta}(Kx, \dot{y}). \end{equation} The following lemma provides two key inequalities to link $F_{\beta}$ to $\mathcal{L}$ and $D$.
\begin{lemma}\label{le:basic_properties} Let $F_{\beta}$ be defined by \eqref{eq:F_beta} and $(x^{\star}, y^{\star})$ be a saddle-point of \eqref{eq:saddle_point_prob}. Then, for any $x, \bar{x}\in\dom{f}$ and $\dot{y}, \tilde{y}, y \in\dom{g^{*}}$, we have \begin{equation}\label{eq:key_est01} \begin{array}{lcl} \mathcal{L}(\bar{x}, y) & \leq & F_{\beta}(\bar{x}, \dot{y}) + \frac{\beta}{2}\norms{y - \dot{y}}^2,
\\ D(\tilde{y}) - D(y^{\star}) & \leq & F_{\beta}(\bar{x}, \dot{y}) - \mathcal{L}(\tilde{x}, \tilde{y}) + \frac{\beta}{2}\norms{y^{\star} - \dot{y}}^2, \quad \forall\tilde{x} \in \partial{f^{*}}(-K^{\top}\tilde{y}). \end{array} \end{equation} \end{lemma}
\begin{proof} Using the definition of $\mathcal{L}$ in \eqref{eq:saddle_point_prob} and of $F_{\beta}$ in \eqref{eq:F_beta}, we have $\mathcal{L}(\bar{x}, y) = f(\bar{x}) + \iprods{K\bar{x}, y} - g^{*}(y) \leq f(\bar{x}) + \sup_{y}\sets{\iprods{K\bar{x}, y} - g^{*}(y) - \frac{\beta}{2}\norms{y - \dot{y}}^2} + \frac{\beta}{2}\norms{y - \dot{y}}^2 = F_{\beta}(\bar{x}, \dot{y}) + \frac{\beta}{2}\norms{y - \dot{y}}^2$, which proves the first line of \eqref{eq:key_est01}.
Next, for any $\tilde{x} \in \partial{f^{*}}(-K^{\top}\tilde{y})$, we have $f^{*}(-K^{\top}\tilde{y}) = \iprods{-K^{\top}\tilde{y}, \tilde{x}} - f(\tilde{x})$ by Fenchel's equality \cite{Bauschke2011}. Hence, $D(\tilde{y}) = f^{*}(-K^{\top}\tilde{y}) + g^{*}(\tilde{y}) = -\mathcal{L}(\tilde{x}, \tilde{y})$. On the other hand, by \eqref{eq:saddle_point}, we have $\mathcal{L}(\bar{x}, y^{\star}) \geq \mathcal{L}(x^{\star}, y^{\star}) = -D(y^{\star})$. Combining these two expressions and the first line of \eqref{eq:key_est01}, we obtain $D(\tilde{y}) - D(y^{\star}) \leq \mathcal{L}(\bar{x}, y^{\star}) - \mathcal{L}(\tilde{x},\tilde{y}) \leq F_{\beta}(\bar{x}, \dot{y}) - \mathcal{L}(\tilde{x}, \tilde{y}) + \frac{\beta}{2}\norms{\dot{y} - y^{\star}}^2$.
$\square$ \end{proof}
\section{New ASGARD Variant and Its Convergence Guarantees}\label{sec:alg_scheme}
In this section, we derive a new and unified variant of ASGARD in \cite{TranDinh2015b} and analyze its convergence rate guarantees for three settings.
\subsection{\textbf{The derivation of algorithm and one-iteration analysis}}
Given $\dot{y} \in\mathbb{R}^n$ and $\hat{x}^k \in\mathbb{R}^p$, the main step of ASGARD consists of one primal and one dual updates as follows: \begin{equation}\label{eq:prox_grad_step0} \left\{\begin{array}{lcl} y^{k+1} & := & \textrm{prox}_{g^{*}/\beta_k}\big(\dot{y} + \tfrac{1}{\beta_k}K\hat{x}^k \big),
\\ x^{k+1} & := & \textrm{prox}_{f/L_k}\big(\hat{x}^k - \tfrac{1}{L_k}K^{\top}y^{k+1} \big), \end{array}\right. \end{equation} where $\beta_k > 0$ is the smoothness parameter of $g$, and $L_k > 0$ is an estimate of the Lipschitz constant of $\nabla{g_{\beta_k}}$. Here, \eqref{eq:prox_grad_step0} serves as basic steps of various primal-dual first-order methods in the literature, including \cite{Chambolle2011,He2012,TranDinh2015b}. However, instead of updating $\dot{y}$, ASGARD fixes it for all iterations.
The following lemma serves as a key step for our analysis in the sequel. Since its statement and proof are rather different from \cite[Lemma 2]{TranDinh2015b}, we provide its proof in Appendix~\ref{apdx:le:descent_pro}.
\begin{lemma}[\cite{TranDinh2015b}]\label{le:descent_pro} Let $(x^{k+1}, y^{k+1})$ be generated by \eqref{eq:prox_grad_step0}, $F_{\beta}$ be defined by \eqref{eq:F_beta}, and $\mathcal{L}$ be given by \eqref{eq:saddle_point_prob}. Then, for any $x\in\dom{f}$ and $\tau_k \in [0,1]$, we have \begin{equation}\label{eq:key_est00_a} \hspace{-2ex} \arraycolsep=0.1em \begin{array}{lcl} F_{\beta_k}(x^{k+1},\dot{y}) & \leq & (1 - \tau_k) F_{\beta_{k-1}}(x^k, \dot{y}) + \tau_k\mathcal{L}(x, y^{k+1})
\\ && + {~} \frac{L_k\tau_k^2}{2} \big\Vert \tfrac{1}{\tau_k}[\hat{x}^k - (1-\tau_k)x^k] - x \big\Vert^2 - \frac{\mu_f(1-\tau_k)\tau_k}{2}\norms{x - x^k}^2
\\ && - {~} \frac{\tau_k^2}{2}\left( L_k + \mu_f \right) \big\Vert \tfrac{1}{\tau_k}[x^{k+1} - (1-\tau_k)x^k] - x \big\Vert^2
\\ && - {~} \frac{L_k}{2}\norms{x^{k+1} - \hat{x}^k}^2 + \frac{1}{2(\mu_{g^{*}}+\beta_k)}\Vert K(x^{k+1} - \hat{x}^k)\Vert^2
\\ && - {~} \frac{1-\tau_k}{2}\left[ \tau_k\beta_k - (\beta_{k-1} - \beta_k)\right]\norms{ \nabla_ug_{\beta_k}(Kx^k,\dot{y}) - \dot{y}}^2. \end{array} \hspace{-8ex} \end{equation} \end{lemma}
\noindent Together with the primal-dual step \eqref{eq:prox_grad_step0}, we also apply Nesterov's accelerated step to $\hat{x}^k$ and an averaging step to $\tilde{y}^k$ as follows: \begin{equation}\label{eq:prox_grad_scheme} \left\{\begin{array}{lcl} \hat{x}^{k+1} & := & x^{k+1} + \eta_{k+1}(x^{k+1} - x^k),
\\ \tilde{y}^{k+1} &:= & (1-\tau_k)\tilde{y}^k + \tau_ky^{k+1}, \end{array}\right. \end{equation} where $\tau_k \in (0, 1)$ and $\eta_{k+1} > 0$ will be determined in the sequel.
To analyze the convergence of the new ASGARD variant, we define the following Lyapunov function (also called a potential function): \begin{equation}\label{eq:lyapunov_func1} \begin{array}{lcl} \mathcal{V}_k(x) & := & F_{\beta_{k-1}}(x^k, \dot{y}) - \mathcal{L}(x, \tilde{y}^k)
\\ && + {~} \tfrac{(L_{k-1} + \mu_f)\tau_{k-1}^2}{2}\norms{\tfrac{1}{\tau_{k-1}}[x^k - (1 - \tau_{k-1})x^{k-1}] - x}^2. \end{array} \end{equation} The following lemma provides a key recursive estimate to analyze the convergence of \eqref{eq:prox_grad_step0} and \eqref{eq:prox_grad_scheme}, whose proof is given in Appendix~\ref{apdx:le:maintain_gap_reduction1}.
\begin{lemma}\label{le:maintain_gap_reduction1} Let $(x^{k}, \hat{x}^k, y^k, \tilde{y}^{k})$ be updated by \eqref{eq:prox_grad_step0} and \eqref{eq:prox_grad_scheme}. Given $\beta_0 > 0$, $\tau_k, \tau_{k+1}\in (0, 1]$, let $\beta_k$, $L_k$, and $\eta_{k+1}$ be updated by
\begin{equation}\label{eq:Lk_m_k} \beta_k := \frac{\beta_{k-1}}{1+\tau_k}, \quad L_k := \frac{\norms{K}^2}{\beta_k + \mu_{g^{*}}}, \quad\text{and}\quad \eta_{k+1} := \frac{(1-\tau_{k})\tau_{k}}{\tau_{k}^2 + m_{k+1}\tau_{k+1}},
\end{equation} where $m_{k+1} := \frac{L_{k+1} + \mu_f}{L_{k} + \mu_f}$. Suppose further that $\tau_k \in (0, 1]$ satisfies \begin{equation}\label{eq:param_cond} \hspace{-1ex} \arraycolsep=0.2em \left\{\begin{array}{llcl} &(L_{k-1} + \mu_f)(1-\tau_k)\tau_{k-1}^2 + \mu_f(1-\tau_k)\tau_k &\geq & L_k\tau_k^2,
\\ &(L_{k-1} + \mu_f)(L_k + \mu_f)\tau_k\tau_{k-1}^2 + (L_k + \mu_f)^2\tau_k^2 & \geq & (L_{k-1} + \mu_f)L_k\tau_{k-1}^2. \end{array}\right. \hspace{-1ex} \end{equation} Then, for any $x\in\dom{f}$, the Lyapunov function $\mathcal{V}_k$ defined by \eqref{eq:lyapunov_func1} satisfies \begin{equation}\label{eq:key_est1_ncvx} \mathcal{V}_{k+1}(x) \leq (1-\tau_k)\mathcal{V}_{k}(x). \end{equation} \end{lemma}
\noindent\textbf{The unified ASGARD algorithm.} Our next step is to expand \eqref{eq:prox_grad_step0}, \eqref{eq:prox_grad_scheme}, and \eqref{eq:Lk_m_k} algorithmically to obtain a new ASGARD variant (called ASGARD+) as presented in Algorithm~\ref{alg:A1}.
\begin{algorithm}[ht!]\caption{(New Accelerated Smoothed GAp ReDuction (ASGARD+))}\label{alg:A1} \normalsize \begin{algorithmic}[1]
\State{\bfseries Initialization:} \label{step:i0a}
Choose $x^0 \in\dom{f}$, $\tilde{y}^0\in\dom{g^{*}}$, and $\dot{y}\in \mathbb{R}^n$.
\State\label{step:i0b}
Choose $\tau_0 \in (0, 1]$ and $\beta_0 > 0$.
Set $L_0 := \frac{\norms{K}^2}{\mu_{g^{*}} + \beta_0}$ and $\hat{x}^0 := x^0$.
\State{\bfseries For $k := 0, 1, \cdots, k_{\max}$ do}
\State\hspace{2ex}\label{step:i1}
Update $\tau_{k+1}$ as in Theorem~\ref{th:convergence1}, \ref{th:convergence_1a}, or \ref{th:convergence_1b}, and update $\beta_{k+1} := \frac{\beta_{k}}{1+\tau_{k+1}}$.
\State\hspace{2ex}\label{step:i2}
Let $L_{k+1} := \frac{\norms{K}^2}{\mu_{g^{*}} + \beta_{k+1}}$, $m_{k+1} := \frac{L_{k+1}+\mu_f}{L_{k} + \mu_f}$, and $\eta_{k+1} := \frac{(1-\tau_{k})\tau_{k}}{\tau_{k}^2 + m_{k+1}\tau_{k+1}}$.
\State\hspace{2ex}\label{step:i3} Update
\begin{equation*}
\left\{\begin{array}{lcl}
y^{k+1} & := & \textrm{prox}_{g^{*}/\beta_k}\big(\dot{y} + \tfrac{1}{\beta_k}K\hat{x}^k \big),
\\
x^{k+1} & := & \textrm{prox}_{f/L_k}\big(\hat{x}^k - \tfrac{1}{L_k}K^{\top}y^{k+1} \big),
\\
\hat{x}^{k+1} & := & x^{k+1} + \eta_{k+1}(x^{k+1} - x^k),
\\
\tilde{y}^{k+1} & := & (1-\tau_k)\tilde{y}^k + \tau_ky^{k+1}.
\end{array}\right.
\end{equation*}
\State{\bfseries EndFor} \end{algorithmic} \end{algorithm}
Compared to the original ASGARD in \cite{TranDinh2015b}, Algorithm~\ref{alg:A1} requires one additional averaging dual step on $\tilde{y}^k$ at Step~\ref{step:i3} to obtain the dual convergence. Note that Algorithm~\ref{alg:A1} also incorporates the strong convexity parameters $\mu_f$ of $f$ and $\mu_{g^{*}}$ of $g^{*}$ to cover three settings: general convexity ($\mu_f = \mu_{g^{*}} = 0$), strong convexity ($\mu_f > 0$ and $\mu_{g^{*}} = 0$), and strong convexity and smoothness ($\mu_f > 0$ and $\mu_{g^{*}} > 0$). More precisely, $L_k$ and the momentum step-size $\eta_{k+1}$ are also different from \cite{TranDinh2015b} by incorporating $\mu_{g^{*}}$ and $\mu_f$. The per-iteration complexity of ASGSARD+ remains the same as ASGARD except for the averaging dual update $\tilde{y}^k$. However, this step is not required if we only solve \eqref{eq:primal_prob}. We highlight that if we apply a new approach from \cite{tran2019non} to \eqref{eq:saddle_point_prob}, then we can also update the proximal center $\dot{y}$ at each iteration.
\subsection{\textbf{Case 1: Both $f$ and $g^{*}$ are just convex ($\mu_f = \mu_{g^{*}} = 0$)}}\label{subsec:general_cvx_ccv}
The following theorem establishes convergence rates of Algorithm~\ref{alg:A1} for the general convex case where both $f$ and $g^{*}$ are just convex.
\begin{theorem}\label{th:convergence1} Suppose that Assumption~\ref{as:A1} holds and both $f$ and $g^{*}$ are only convex, i.e. $\mu_f = \mu_{g^{*}} = 0$. Let $\sets{(x^k, \tilde{y}^k)}$ be generated by Algorithm~\ref{alg:A1} for solving \eqref{eq:saddle_point_prob}, where $\tau_0 := 1$ and $\tau_{k+1}$ is the unique solution of the cubic equation $\tau^3 + \tau^2 + \tau_k^2\tau - \tau_k^2 = 0$ in $\tau$, which always exists. Then, for all $k \geq 1$, we have: \begin{itemize} \item[$\mathrm{(a)}$] The gap function $\mathcal{G}_{\mathcal{X}\times\mathcal{Y}}$ defined by \eqref{eq:gap_func} satisfies \begin{equation}\label{eq:main_result1} \mathcal{G}_{\mathcal{X}\times\mathcal{Y}}(x^k, \tilde{y}^k) \leq \frac{\norms{K}^2}{2\beta_0k} \sup_{x\in\mathcal{X}}\norms{x^0 - x}^2 + \frac{\beta_0}{k+1}\sup_{y\in\mathcal{Y}}\norms{y - \dot{y}}^2. \end{equation} \item[$\mathrm{(b)}$] If $g$ is $M_g$-Lipschitz continuous on $\dom{g}$, then for \eqref{eq:primal_prob}, it holds that \begin{equation}\label{eq:main_result1b} F(x^k) - F^{\star} \leq \frac{\norms{K}^2}{2\beta_0 k}\norms{x^0 - x^{\star}}^2 + \frac{\beta_0}{k+1}(\norms{\dot{y}} + M_g)^2. \end{equation} \item[$\mathrm{(c)}$] If $f^{*}$ is $M_{f^{*}}$-Lipschitz continuous on $\dom{f^{*}}$, then for \eqref{eq:dual_prob}, it holds that \begin{equation}\label{eq:main_result1c} D(\tilde{y}^k) - D^{\star} \leq \frac{\norms{K}^2}{2\beta_0 k}(\norms{x^0} + M_{f^{*}})^2 + \frac{\beta_0}{k+1}\norms{\dot{y} - y^{\star}}^2. \end{equation} \end{itemize} \end{theorem}
\begin{proof} First, since $\mu_f = \mu_{g^{*}} = 0$, $L_k = \frac{\norms{K}^2}{\beta_k}$, and $(1+\tau_k)\beta_k = \beta_{k-1}$, the two conditions of \eqref{eq:param_cond} respectively reduce to \begin{equation*} (1-\tau_k)\tau_{k-1}^2 \geq (1+\tau_k)\tau_k^2 \quad\text{and}\quad (1 + \tau_k)\tau_k^2 \geq \tau_{k-1}^2(1-\tau_k). \end{equation*} These conditions hold if $\tau_k^3 + \tau_k^2 + \tau_{k-1}^2\tau_k - \tau_{k-1}^2 = 0$. We first choose $\tau_0 := 1$, and update $\tau_k$ by solving the cubic equation $\tau^3 + \tau + \tau_{k-1}^2\tau - \tau_{k-1}^2 = 0$ for $k\geq 1$. Note that this equation has a unique positive real solution $\tau_k \in (0, 1)$ due to Lemma~\ref{le:tech_lemma}(b). Moreover, we have $\prod_{i=1}^k(1-\tau_i) \leq \frac{1}{k+1}$ and $\beta_k \leq \frac{2\beta_0}{k+2}$.
Next, by induction, \eqref{eq:key_est1_ncvx} leads to $\mathcal{V}_k(x) \leq \left[\prod_{i=1}^{k-1}(1-\tau_i)\right]\mathcal{V}_1(x) \leq \frac{1}{k}\mathcal{V}_1(x)$, where we have used $\prod_{i=1}^{k-1}(1-\tau_i) \leq \frac{1}{k}$ from Lemma~\ref{le:tech_lemma}(b). However, from \eqref{eq:key_est00_proof1} in the proof of Lemma~\ref{le:maintain_gap_reduction1} and $\tau_0 = 1$, we have $\mathcal{V}_1(x) \leq (1-\tau_0)\mathcal{V}_0(x) + \frac{L_0\tau_0^2}{2}\norms{\hat{x}^0 - x}^2 = \frac{\norms{K}^2}{2\beta_0}\norms{x^0 - x}^2$. Hence, we eventually obtain \begin{equation}\label{eq:proof_th1_01} \mathcal{V}_k(x) \leq \frac{\norms{K}^2}{2\beta_0k}\norms{x^0 - x}^2. \end{equation} Using \eqref{eq:key_est01} from Lemma~\ref{le:basic_properties} and $\beta_{k-1} \leq \frac{2\beta_0}{k+1}$ from Lemma~\ref{le:tech_lemma}(b), we get \begin{equation*} \begin{array}{lcl} \mathcal{L}(x^k, y) - \mathcal{L}(x, \tilde{y}^k) & \overset{\tiny\eqref{eq:key_est01}}{\leq} & F_{\beta_{k-1}}(x^k, \dot{y}) - \mathcal{L}(x,\tilde{y}^k) + \frac{\beta_{k-1}}{2}\norms{y- \dot{y}}^2
\\ & \overset{\tiny\eqref{eq:lyapunov_func1}}{\leq} & \mathcal{V}_k(x) + \frac{\beta_{k-1}}{2}\norms{y - \dot{y}}^2
\\ & \overset{\tiny\eqref{eq:proof_th1_01}}{\leq} & \frac{\norms{K}^2}{2\beta_0k}\norms{x^0 - x}^2 + \frac{\beta_0}{k+1}\norms{\dot{y} - y}^2. \end{array} \end{equation*} Taking the supreme over $\mathcal{X}$ and $\mathcal{Y}$ both sides of the last estimate and using \eqref{eq:gap_func}, we obtain \eqref{eq:main_result1}.
Now, since $F_{\beta_{k-1}}(x^k,\dot{y}) - F^{\star} \overset{\tiny\eqref{eq:saddle_point}}{\leq} F_{\beta_{k-1}}(x^k,\dot{y}) - \mathcal{L}(x^{\star}, \tilde{y}^k) \overset{\tiny\eqref{eq:lyapunov_func1}}{\leq} \mathcal{V}_k(x^{\star})$, combining this inequality and \eqref{eq:proof_th1_01}, we get \begin{equation*} F_{\beta_{k-1}}(x^k,\dot{y}) - F^{\star} \leq \frac{\norms{K}^2}{2\beta_0k}\norms{x^0 - x^{\star}}^2. \end{equation*} On the other hand, since $g$ is $M_g$-Lipschitz continuous, we have \begin{equation*} \sup\set{ \norms{y-\dot{y}} : y\in\partial{g}(Kx^k)} \leq \norms{\dot{y}} + \sup\sets{\norms{y} : \norms{y} \leq M_g} = \norms{\dot{y}} + M_g. \end{equation*} Hence, by \eqref{eq:key_bound_apdx3} of Lemma~\ref{le:g_properties}, we have $F(x^k) \leq F_{\beta_{k-1}}(x^k, \dot{y}) + \frac{\beta_{k-1}}{2}(\norms{\dot{y}} + M_g)^2 \leq F_{\beta_{k-1}}(x^k, \dot{y}) + \frac{\beta_0}{k+1}(\norms{\dot{y}} + M_g)^2$. Combining both estimates, we obtain \eqref{eq:main_result1b}.
Finally, using \eqref{eq:key_est01}, we have \begin{equation*} \begin{array}{lcl} D(\tilde{y}^k) - D^{\star} & \leq & F_{\beta_{k-1}}(x^k, \dot{y}) - \mathcal{L}(\tilde{x}^k, \tilde{y}^k) + \frac{\beta_{k-1}}{2}\norms{\dot{y} - y^{\star}}^2
\\ &\overset{\tiny\eqref{eq:lyapunov_func1}}{\leq} & \mathcal{V}_k(\tilde{x}^k) + \frac{\beta_0}{k+1}\norms{ \dot{y} - y^{\star}}^2
\\ &\overset{\tiny\eqref{eq:proof_th1_01}}{\leq} & \frac{\norms{K}^2}{2\beta_0k} (\norms{x^0} + M_{f^{*}})^2 + \frac{\beta_0}{k+1}\norms{\dot{y} - y^{\star}}^2, \end{array} \end{equation*} which proves \eqref{eq:main_result1c}. Here, since $\tilde{x}^k \in \partial{f^{*}}(-K^{\top}\tilde{y}^k)$, we have $\norms{x^0 - \tilde{x}^k} \leq \norms{\tilde{x}^k} + \norms{x^0} \leq M_{f^{*}} + \norms{x^0}$, which has been used in the last inequality.
$\square$ \end{proof}
\subsection{\textbf{Case 2: $f$ is strongly convex and $g^{*}$ is convex ($\mu_f > 0$ and $\mu_{g^{*}} = 0$)}}\label{subsect:f_or_gstar_strongly_convex}
Next, we consider the case when only $f$ or $g^{*}$ is strongly convex. Without loss of generality, we assume that $f$ is strongly convex with a strong convexity parameter $\mu_f > 0$, but $g^{*}$ is only convex with $\mu_{g^{*}} = 0$. Otherwise, we switch the role of $f$ and $g^{*}$ in Algorithm~\ref{alg:A1}.
The following theorem establishes an optimal $\BigO{1/k^2}$ convergence rate (up to a constant factor) of Algorithm~\ref{alg:A1} in this case (i.e. $\mu_f > 0$ and $\mu_{g^{*}} = 0$).
\begin{theorem}\label{th:convergence_1a} Suppose that Assumption~\ref{as:A1} holds and that $f$ is strongly convex with a convexity parameter $\mu_f > 0$, but $g^{*}$ is just convex $($i.e. $\mu_{g^{*}} = 0$$)$. Let us choose $\tau_0 := 1$ and $\beta_0 \geq \frac{0.382\norms{K}^2}{\mu_f}$. Let $\sets{(x^k, \tilde{y}^k)}$ be generated by Algorithm~\ref{alg:A1} using the update rule $\tau_{k+1} := \frac{\tau_k}{2}\big(\sqrt{\tau_k^2 + 4} - \tau_k\big)$ for $\tau_k$. Then, we have: \begin{itemize} \item[$\mathrm{(a)}$] The gap function $\mathcal{G}_{\mathcal{X}\times\mathcal{Y}}$ be defined by \eqref{eq:gap_func} satisfies \begin{equation}\label{eq:main_est300} \mathcal{G}_{\mathcal{X}\times\mathcal{Y}}(x^k, \tilde{y}^k) \leq \frac{2\norms{K}^2}{\beta_0(k+1)^2}\sup_{x\in\mathcal{X}}\norms{x^0 -x}^2 + \frac{10\beta_0}{(k+3)^2} \sup_{y\in\mathcal{Y}}\norms{\dot{y} - y}^2. \end{equation}
\item[$\mathrm{(b)}$] If $g$ is $M_g$-Lipschitz continuous on $\dom{g}$, then for \eqref{eq:primal_prob}, it holds that \begin{equation}\label{eq:main_est300b} F(x^k) - F^{\star} \leq \frac{2\norms{K}^2\norms{x^0 - x^{\star}}^2}{\beta_0(k+1)^2} + \frac{10\beta_0(\norms{\dot{y}} + M_g)^2}{(k+3)^2}. \end{equation}
\item[$\mathrm{(c)}$] If $f^{*}$ is $M_{f^{*}}$-Lipschitz continuous on $\dom{f^{*}}$, then for \eqref{eq:dual_prob}, it holds that \begin{equation}\label{eq:main_est300c} D(\tilde{y}^k) - D^{\star} \leq \frac{2\norms{K}^2(\norms{x^0} + M_{f^{*}})^2}{\beta_0(k+1)^2} + \frac{10\beta_0\norms{\dot{y} - y^{\star}}^2}{(k+3)^2}. \end{equation} \end{itemize} \end{theorem}
\begin{proof} Since $\mu_{g^{*}} = 0$ and $\tau_k^2 = \tau_{k-1}^2(1 - \tau_k)$ by the update rule of $\tau_k$, $\beta_k = \frac{\beta_{k-1}}{1+\tau_k}$, and $L_k = \frac{\norms{K}^2}{\beta_k}$, the first condition of \eqref{eq:param_cond} is equivalent to $\beta_{k-1} \geq \frac{\norms{K}^2\tau_k^2}{\mu_f}$. However, since $\beta_{k-1} \geq \frac{\beta_0\tau_{k}^2}{\tau_1^2}$ due to Lemma~\ref{le:tech_lemma}(a), and $\tau_1 = 0.6180$, $\beta_{k-1} \geq \frac{\norms{K}^2\tau_k^2}{\mu_f}$ holds if $\beta_0 \geq \frac{\norms{K}^2\tau_1^2}{\mu_f} = \frac{0.382\norms{K}^2}{\mu_f}$. Thus we can choose $\beta_0 \geq \frac{0.382\norms{K}^2}{\mu_f}$ to guarantee the first condition of \eqref{eq:param_cond}.
Similarly, using $m_k = \frac{L_k + \mu_f}{L_{k-1} + \mu_f} \geq 1$, the second condition of \eqref{eq:param_cond} is equivalent to $m_k^2\tau_k^2 + m_k\tau_k\tau_{k-1}^2 \geq \frac{L_k\tau_{k-1}^2}{L_{k-1} + \mu_f}$. Since $\frac{L_k}{L_{k-1} + \mu_f} \leq m_k$, the last condition holds if $m_k^2\tau_k^2 + m_k\tau_k\tau_{k-1}^2 \geq m_k\tau_{k-1}^2$. Using again $\tau_k^2 = \tau_{k-1}^2(1-\tau_k)$, this condition becomes $m_k\tau_k^2 \geq \tau_{k-1}^2(1-\tau_k) = \tau_k^2$. This always holds true since $m_k \geq 1$. Therefore, the second condition of \eqref{eq:param_cond} is satisfied.
As a result, we have the recursive estimate \eqref{eq:key_est1_ncvx}, i.e.: \begin{equation}\label{eq:recursive_form} \mathcal{V}_{k+1}(x) \leq (1-\tau_k)\mathcal{V}_{k}(x). \end{equation} From \eqref{eq:recursive_form}, Lemma~\ref{le:tech_lemma}(a), \eqref{eq:key_est00_proof1}, and noting that $\tilde{x}^0 := x^0$ and $\tau_0 := 1$, we have \begin{equation*} \arraycolsep=0.2em \begin{array}{lcl} \mathcal{V}_k(x) & \leq & \big[\prod_{i=1}^{k-1}(1-\tau_i)\big]\mathcal{V}_1(x) = \Theta_{1,k-1}\mathcal{V}_1(x) \leq \frac{4}{(k+1)^2}\mathcal{R}_0(x). \end{array} \end{equation*} where $\mathcal{R}_0(x) := \frac{\Vert K\Vert^2}{2\beta_0}\Vert x^0 - x\Vert^2$. Similar to the proof of Theorem~\ref{th:convergence1}, using the last inequality and $\beta_{k-1} \leq \frac{4\beta_0\tau_0^2}{\tau_2^2[\tau_0(k+1) + 2]^2} \leq \frac{20\beta_0}{(k+3)^2}$ from Lemma~\ref{le:tech_lemma}(a), we obtain the bounds \eqref{eq:main_est300}, \eqref{eq:main_est300b}, and \eqref{eq:main_est300c} respectively.
$\square$ \end{proof}
\begin{remark}\label{re:compared} The variant of Algorithm~\ref{alg:A1} in Theorem~\ref{th:convergence_1a} is completely different from \cite[Algorithm 2]{TranDinh2015b} and \cite[Algorithm 2]{tran2019non}, where it requires only one $\textrm{prox}_{f/L_k}(\cdot)$ as opposed to two proximal operations of $f$ as in \cite{TranDinh2015b,tran2019non}. \end{remark}
\subsection{\textbf{Case 3: Both $f$ and $g^{*}$ are strongly convex ($\mu_f > 0$ and $\mu_{g^{*}} > 0$)}}\label{sec:both_f_gstar_scvx}
Finally, we assume that both $f$ and $g^{*}$ are strongly convex with strong convexity parameters $\mu_f > 0$ and $\mu_{g^{*}} > 0$, respectively. Then, the following theorem proves the optimal linear rate (up to a constant factor) of Algorithm~\ref{alg:A1}.
\begin{theorem}\label{th:convergence_1b} Suppose that Assumption~\ref{as:A1} holds and both $f$ and $g^{*}$ in \eqref{eq:saddle_point_prob} are strongly convex with $\mu_f > 0$ and $\mu_{g^{*}} > 0$, respectively. Let $\sets{(x^k, \tilde{y}^k)}$ be generated by Algorithm~\ref{alg:A1} using $\tau_k := \tau = \frac{1}{\sqrt{1 + \kappa_F}}\in (0, 1)$ and $\beta_0 > 0$, where $\kappa_F := \frac{\norms{K}^2}{\mu_f\mu_{g^{*}}}$. Then, the following statements hold: \begin{itemize} \item[$\mathrm{(a)}$] The gap function $\mathcal{G}_{\mathcal{X}\times\mathcal{Y}}$ defined by \eqref{eq:gap_func} satisfies \begin{equation}\label{eq:key_est40}
\mathcal{G}_{\mathcal{X}\times\mathcal{Y}}(x^k,\tilde{y}^k) \leq (1-\tau)^k\bar{\mathcal{R}}_p + \frac{1}{2(1+\tau)^k}\sup_{y\in\mathcal{Y}}\norms{\dot{y} - y}^2,
\end{equation} where $\bar{\mathcal{R}}_p := {\displaystyle\sup_{x\in\mathcal{X}}}\Big\{ (1-\tau)\big[\mathcal{F}_{\beta_0}(x^0) - \mathcal{L}(x, \tilde{y}^0)\big] + \frac{\norms{K}^2\tau^2}{2(\mu_{g^{*}} + \beta_{0})} \norms{x^0 - x}^2 \Big\}$.
\item[$\mathrm{(b)}$] If $g$ is $M_g$-Lipschitz continuous on $\dom{g}$, then for \eqref{eq:primal_prob}, it holds that \begin{equation}\label{eq:key_est40b} F(x^k) - F^{\star} \leq (1-\tau)^k\bar{\mathcal{R}}_p^{\star} \ + \ \frac{\beta_0}{2(1+\tau)^k} \left(\norms{\dot{y}} + M_g\right)^2, \end{equation} where $\bar{\mathcal{R}}_p^{\star} := (1-\tau)\big[ \mathcal{F}_{\beta_0}(x^0) - \mathcal{L}(x^{\star}, \tilde{y}^0) \big] + \frac{\norms{K}^2\tau^2}{2(\mu_{g^{*}} + \beta_{0})}\norms{x^0 - x^{\star}}^2$.
\item[$\mathrm{(d)}$] If $f^{*}$ is $M_{f^{*}}$-Lipschitz continuous on $\dom{f^{*}}$, then for \eqref{eq:dual_prob}, it holds that \begin{equation}\label{eq:key_est40c} D(\tilde{y}^k) - D^{\star} \leq (1-\tau)^k\bar{\mathcal{R}}_d^{\star} \ + \ \frac{\beta_0\norms{\dot{y} - y^{\star}}^2}{2(1+\tau)^k}, \end{equation} where $\bar{\mathcal{R}}_d^{\star} := (1-\tau)\big[\mathcal{F}_{\beta_0}(x^0) - D(\tilde{y}^0)\big] + \frac{\norms{K}^2\tau^2}{2(\mu_{g^{*}} + \beta_{0})}\left(\norms{x^0} + M_{f^{*}}\right)^2$. \end{itemize} \end{theorem}
\begin{proof} Since $\tau_k = \tau = \frac{1}{\sqrt{1+\kappa_F}} = \sqrt{\frac{\mu_f\mu_{g^{*}}}{\norms{K}^2 + \mu_f\mu_{g^{*}}}} \in (0, 1)$ and $\beta_{k-1} = (1+\tau)\beta_k$, after a few elementary calculations, we can show that the first condition of \eqref{eq:param_cond} automatically holds. The second condition of \eqref{eq:param_cond} is equivalent to $m_k\tau + m_k^2 \geq \frac{L_k}{L_{k-1} + \mu_f}$. Since $m_k \geq \frac{L_k}{L_{k-1} + \mu_f}$, this condition holds if $m_k\tau + m_k^2 \geq m_k$, which is equivalent to $\tau + m_k \geq 1$. This obviously holds true since $\tau > 0$ and $m_k \geq 1$.
From \eqref{eq:key_est1_ncvx} of Lemma~\ref{le:maintain_gap_reduction1}, we have $\mathcal{V}_{k+1}(x) \leq (1-\tau)\mathcal{V}_{k}(x)$. Therefore, by induction and using again \eqref{eq:key_est00_proof1}, we get \begin{equation}\label{eq:recursive_form} \mathcal{V}_k(x) \leq (1-\tau)^k\mathcal{V}_1(x) \leq (1-\tau)^k\bar{\mathcal{R}}_0(x). \end{equation} where $\bar{\mathcal{R}}_p(x) := (1-\tau)\big[ \mathcal{F}_{\beta_0}(x^0) - \mathcal{L}(x,\tilde{y}^0)\big] + \frac{\norms{K}^2\tau^2}{2(\mu_{g^{*}} + \beta_{0})}\norms{x^0 - x}^2$.
Now, since $\beta_{k-1} = \frac{\beta_0}{(1+\tau)^k}$ due to the update rule of $\beta_k$, by \eqref{eq:key_est01}, we have \begin{equation*} \mathcal{L}(x^k, y) - \mathcal{L}(x, \tilde{y}^k) \leq \mathcal{V}_k(x) + \frac{\beta_{k-1}}{2}\norms{\dot{y} - y}^2 \leq (1-\tau)^k\bar{\mathcal{R}}_p(x) + \frac{\norms{\dot{y} - y}^2}{2(1+\tau)^k}. \end{equation*} This implies \eqref{eq:key_est40}. The estimates \eqref{eq:key_est40b} and \eqref{eq:key_est40c} can be proved similarly as in Theorem~\ref{th:convergence1}, and we omit the details here.
$\square$ \end{proof}
\begin{remark}\label{re:cond_number} Since $g^{*}$ is $\mu_{g^{*}}$-strongly convex, it is well-known that $g\circ K$ is $\frac{\norms{K}^2}{\mu_{g^{*}}}$-smooth. Hence, $\kappa_F := \frac{\norms{K}^2}{\mu_f\mu_{g^{*}}}$ is the condition number of $F$ in \eqref{eq:primal_prob}. Theorem shows that Algorithm~\ref{alg:A1} can achieve a $\big(1 - \frac{1}{(1+\sqrt{2})\sqrt{\kappa_F}}\big)^k$-linear convergence rate. Consequently, it also achieves $\BigO{ \sqrt{\kappa_F}\log\left(\frac{1}{\varepsilon}\right)}$ oracle complexity to obtain an $\varepsilon$-primal-dual solution $(x^k,\tilde{y}^k)$. This linear rate and complexity are optimal (up to a constant factor) under the given assumptions in Theorem~\ref{th:convergence_1b}. However, Algorithm~\ref{alg:A1} is very different from existing accelerated proximal gradient methods, e.g., \cite{Beck2009,Nesterov2007,tseng2008accelerated} for solving \eqref{eq:primal_prob} since our method uses the proximal operator of $g^{*}$ (and therefore, the proximal operator of $g$) instead of the gradient of $g$ as in \cite{Beck2009,Nesterov2007,tseng2008accelerated}. \end{remark}
\begin{remark}\label{re:optimal_rate} The $\BigO{1/k}$, $\BigO{1/k^2}$, and linear convergence rates in Theorem~\ref{th:convergence1}, \ref{th:convergence_1a}, and \ref{th:convergence_1b}, respectively are already optimal (up to a constant factor) under given assumptions as discussed, e.g., in \cite{ouyang2018lower,tran2019non}. The primal convergence rate on $\sets{F(x^k) - F^{\star}}$ has been proved in \cite[Theorem~4]{TranDinh2015b}, but only for the case $\BigO{1/k}$. The convergence rates on $\sets{\mathcal{G}_{\mathcal{X}\times\mathcal{Y}}(x^k,\tilde{y}^k)}$ and $\sets{D(\tilde{y}^k) - D^{\star}}$ are new. Moreover, the convergence of the primal sequence is on the last iterate $x^k$, while the convergence of the dual sequence is on the averaging iterate $\tilde{y}^k$. \end{remark}
\section{Numerical Experiments}\label{sec:num_exp}
In this section, we provide four numerical experiments to verify the theoretical convergence aspects and the performance of Algorithm~\ref{alg:A1}. Our algorithm is implemented in Matlab R.2019b running on a MacBook Laptop with 2.8GHz Quad-Core Intel Core i7 and 16GB RAM. We also compare our method with Nesterov's smoothing technique in \cite{Nesterov2005c} as a baseline. We emphasize that our experiments bellow follow exactly the parameter update rules as stated in Theorem~\ref{th:convergence1} and Theorem~\ref{th:convergence_1a} without any parameter tuning trick. To further improve practical performance of Algorithm~\ref{alg:A1}, one can exploit the restarting strategy in \cite{TranDinh2015b}, where its theoretical guarantee is established in \cite{tran2018adaptive}.
The nonsmooth and convex optimization problem we use for our experiments is the following representative model: \begin{equation}\label{eq:LAD} \min_{x\in\mathbb{R}^p}\Big\{ F(x) := \Vert Kx - b\Vert_2 + \lambda\norms{x}_1 + \frac{\rho}{2}\norms{x}_2^2 \Big\}, \end{equation} where $K\in\mathbb{R}^{n\times p}$ is a given matrix, $b\in\mathbb{R}^n$ is also given, and $\lambda > 0$ and $\rho \geq 0$ are two given regularization parameters. The norm $\norms{\cdot}_2$ is the $\ell_2$-norm (or Euclidean norm). If $\rho = 0$, then \eqref{eq:LAD} reduces to the square-root LASSO model proposed in \cite{Belloni2011}. If $\rho > 0$, then \eqref{eq:LAD} becomes a square-root regression problem with elastic net regularizer similar to \cite{zou2005regularization}. Clearly, if we define $g(y) := \norms{y - b}_2$ and $f(x) := \lambda\norms{x}_1 + \frac{\rho}{2}\norms{x}_2^2$, then \eqref{eq:LAD} can be rewritten into \eqref{eq:primal_prob}.
To generate the input data for our experiments, we first generate $K$ from standard i.i.d. Gaussian distributions with either uncorrelated or $50\%$ correlated columns. Then, we generate an observed vector $b$ as $b := Kx^{\natural} + \mathcal{N}(0,\sigma)$, where $x^{\natural}$ is a predefined sparse vector and $\mathcal{N}(0, \sigma)$ stands for standard Gaussian noise with variance $\sigma = 0.05$. The regularization parameter $\lambda$ to promote sparsity is chosen as suggested in \cite{Belloni2011}, and the parameter $\rho$ is set to $\rho := 0.1$. We first fix the size of problem at $p := 1000$ and $n := 350$ and choose the number of nonzero entries of $x^{\natural}$ to be $s := 100$. Then, for each experiment, we generate $30$ instances of the same size but with different input data $(K, b)$.
For Nesterov's smoothing method, by following \cite{Nesterov2005c}, we smooth $g$ as \begin{equation*} g_{\gamma}(y) := \max_{v\in\mathbb{R}^n}\Big\{ \iprods{y - b, v} - \frac{\gamma}{2}\norms{v}^2: \norms{v}_2 \leq 1 \Big\}, \end{equation*} where $\gamma > 0$ is a smoothness parameter. In order to correctly choose $\gamma$ for Nesterov's smoothing method, we first solve \eqref{eq:LAD} with CVX \cite{Grant2004} using Mosek with high precision to get a high accurate solution $x^{\star}$ of \eqref{eq:LAD}. Then, we set $\gamma^{*} := \frac{\sqrt{2}\norms{K}\norms{x^0 - x^{\star}}}{k_{\max}\sqrt{D_{\mathcal{V}}}}$ by minimizing its theoretical bound from \cite{Nesterov2005c} w.r.t. $\gamma > 0$, where $D_{\mathcal{V}} := \frac{1}{2}$ is the prox-diameter of the unit $\ell_2$-norm ball, and $k_{\max}$ is the maximum number of iterations. For Algorithm~\ref{alg:A1}, using \eqref{eq:main_result1b} and $\dot{y} := 0$, we can set $\beta_0 = \beta^{*} := \frac{\norms{K}\norms{x^0 - x^{*}}}{M_g}$ by minimizing the right-hand side of \eqref{eq:main_result1b} w.r.t. $\beta_0 > 0$, where $M_g := 1$. We choose $k_{\max} := 5000$ for all experiments. To see the effect of the smoothness parameters $\gamma$ and $\beta_0$ on the performance of both algorithms, we also consider two variants by increasing or decreasing these parameters $10$ times, respectively. More specifically, we set them as follows. \begin{itemize} \itemsep=0.1em \item For Nesterov's smoothing scheme, we consider two additional variants by setting $\gamma := 10\gamma^{*}$ and $\gamma = 0.1\gamma^{*}$, respectively. \item For Algorithm~\ref{alg:A1}, we also consider two other variants with $\beta_0 := 10\beta^{*}$ and $\beta_0 := 0.1\beta^{*}$, respectively. \end{itemize} We first conduct two different experiments for the square-root LASSO model (i.e. setting $\rho := 0$ in \eqref{eq:LAD}). In this case, the underlying optimization problem is non-strongly convex and fully nonsmooth. \begin{itemize} \itemsep=0.1em \item \textit{Experiment 1:} We test Algorithm \ref{alg:A1} (abbreviated by \texttt{Alg. 1}) and Nesterov's smoothing method (abbreviated by \texttt{Nes. Alg.}) on $30$ problem instances with uncorrelated columns of $K$. Since both algorithms have essentially the same per-iteration complexity, we report the relative primal objective residual $\frac{F(x^k) - F(x^{\star})}{\max\sets{1, \vert F(x^{\star})\vert}}$ against the number of iterations. \item \textit{Experiment 2:} We conduct the same test on another set of $30$ problem instances, but using $50\%$ correlated columns in the input matrix $K$. \end{itemize} The results of both experiments are depicted in Figure~\ref{fig:sqrtLASSO_exp1}, where the left plot is for \textit{Experiment 1} and the right plot is for \textit{Experiment 2}. The solid line of each curve presents the mean over $30$ problem samples, and the corresponding shaded area represents the sample variance of $30$ problem samples (i.e. the area from the lowest and the highest deviation to the mean).
\begin{figure}
\caption{The convergence behavior of Algorithm~\ref{alg:A1} and Nesterov's smoothing scheme on $30$ problem instances of \eqref{eq:LAD} (the non-strongly convex case). Left plot: uncorrelated columns in $K$, and Right plot: $50\%$ correlated columns in $K$.}
\label{fig:sqrtLASSO_exp1}
\end{figure}
From Figure \ref{fig:sqrtLASSO_exp1}, we can observe that, with the choice $\beta_0 := \beta^{*}$ and $\gamma := \gamma^{*}$ as suggested by the theory, both algorithms perform best compared to other smaller or larger values of these parameters. We also see that Algorithm~\ref{alg:A1} outperforms Nesterov's smoothing scheme in both experiments. If $\beta_0$ (respectively, $\gamma$) is large, then both algorithms make good progress in early iterations, but become saturated at a given objective value in the last iterations. Alternatively, if $\beta_0$ (respectively, $\gamma$) is small, then both algorithms perform worse in early iterations, but further decrease the objective value when the number of iterations is increasing. This behavior also confirms the theoretical results stated in Theorem~\ref{th:convergence1} and in \cite{Nesterov2005c}. In fact, if $\beta_0$ (or $\gamma$) is small, then the algorithmic stepsize is small. Hence, the algorithm makes slow progress at early iterations, but it better approximates the nonsmooth function $g$, leading to more accurate approximation from $F(x^k)$ to $F(x^{\star})$. In contrast, if $\beta_0$ (or $\gamma$) is large, then we have a large stepsize and therefore a faster convergence rate in early iterations. However, the smoothed approximation is less accurate.
In order to test the strongly convex case in Theorem~\ref{th:convergence_1a}, we conduct two additional experiments on \eqref{eq:LAD} with $\rho := 0.1$. In this case, problem \eqref{eq:LAD} is strongly convex with $\mu_f := 0.1$. Since \cite{Nesterov2005c} does not directly handle the strongly convex case, we only compare two variants of Algorithm~\ref{alg:A1} stated in Theorem~\ref{th:convergence1} (\texttt{Alg. 1}) and Theorem~\ref{th:convergence_1a} (\texttt{Alg. 1b}), respectively. We set $\beta_0 := \frac{0.382\norms{K}^2}{\mu_f}$ in \texttt{Alg. 1b)} as suggested by Theorem~\ref{th:convergence_1a}. We consider two experiments as follows: \begin{itemize} \itemsep=0.1em \item \textit{Experiment 3:} Test two variants of Algorithm \ref{alg:A1} on a collection of $30$ problem instances with uncorrelated columns of $K$. \item \textit{Experiment 4:} Conduct the same test on another set of $30$ problem instances, but using $50\%$ correlated columns in $K$. \end{itemize} The results of both variants of Algorithm~\ref{alg:A1} are reported in Figure \ref{fig:sqrtLASSO_exp2}, where the left plot is for \textit{Experiment 3} and the right plot is for \textit{Experiment 4}.
\begin{figure}
\caption{The convergence behavior of the two variants of Algorithm~\ref{alg:A1} on a collection of $30$ problem instances of \eqref{eq:LAD} (the strongly convex case). Left plot: uncorrelated columns in $K$, and Right plot: $50\%$ correlated columns in $K$.}
\label{fig:sqrtLASSO_exp2}
\end{figure}
Clearly, as shown in Figure \ref{fig:sqrtLASSO_exp2}, \texttt{Alg. 1b} (i.e. corresponding to Theorem~\ref{th:convergence_1a}) highly outperforms \texttt{Alg. 1} (corresponding to Theorem~\ref{th:convergence1}). \texttt{Alg. 1} matches well the $\BigO{1/k}$ convergence rate as stated in Theorem~\ref{th:convergence1}, while \texttt{Alg. 1b} shows its $\BigO{1/k^2}$ convergence rate as indicated by Theorem~\ref{th:convergence_1a}. Note that since $g^{*}$ in \eqref{eq:LAD} is non-strongly convex, we omit testing the result of Theorem~\ref{th:convergence_1b}. This case is rather well studied in the literature, see, e.g., \cite{Chambolle2011}.
\section{Concluding Remarks}\label{sec:concl}
We have developed a new variant of ASGRD introduced in \cite[Algorithm 1]{TranDinh2015b}, Algorithm~\ref{alg:A1}, that unifies three different settings: general convexity, strong convexity, and strong convexity and smoothness. We have proved the convergence of Algorithm~\ref{alg:A1} for three settings on three convergence criteria: gap function, primal objective residual, and dual objective residual. Our convergence rates in all cases are optimal up to a constant factor and the convergence rates of the primal sequence is on the last iterate. Our preliminary numerical experiments have shown that the theoretical convergence rates of Algorithm~\ref{alg:A1} match well the actual rates observed in practice. The proposed algorithm can be easily extended to solve composite convex problems with three or multi-objective terms. It can also be customized to solve other models, including general linear and nonlinear constrained convex problems as discussed in \cite{sabach2020faster,TranDinh2015b,tran2019non}.
\vskip 2mm \noindent{\bf Acknowledgments.} This work is partly supported by the Office of Naval Research under Grant No. ONR-N00014-20-1-2088 (2020 - 2023), and the Nafosted Vietnam, Grant No. 101.01-2020.06 (2020 - 2022).
\appendix
\section{Appendix 1: Technical Lemmas}\label{apdx:basic}
We need the following technical lemmas for our convergence analysis in the main text.
\begin{lemma}\label{le:g_properties}$($\cite[Lemma 10]{TranDinh2015b}$)$ Given $\beta > 0$, $\dot{y}\in\mathbb{R}^n$, and a proper, closed, and convex function $g : \mathbb{R}^n \to\mathbb{R}\cup\{+\infty\}$ with its Fenchel conjugate $g^{*}$, we define \begin{equation}\label{eq:smoothed_varphi0} g_{\beta}(u, \dot{y}) := \max_{y\in\mathbb{R}^n}\set{\iprods{u, y} - g^{*}(y) - \tfrac{\beta}{2}\norms{y - \dot{y}}^2}. \end{equation} Let $y^{*}_{\beta}(u, \dot{y})$ be the unique solution of \eqref{eq:smoothed_varphi0}. Then, the following statements hold:
\begin{itemize} \item[$(a)$] $g_{\beta}(\cdot,\dot{y})$ is convex w.r.t. $u$ on $\dom{g}$ and $\frac{1}{\beta + \mu_{g^{*}}}$-smooth w.r.t. $u$ on $\dom{g}$, where $\nabla_u{g_{\beta}}(u,\dot{y}) = \textrm{prox}_{g^{*}/\beta}(\dot{y} + \frac{1}{\beta}u)$. Moreover, for any $u, \hat{u} \in\dom{g}$, we have \begin{equation}\label{eq:key_bound_apdx1} g_{\beta}(\hat{u},\dot{y}) + \iprods{\nabla{g}_{\beta}(\hat{u},\dot{y}), u - \hat{u}} \leq g_{\beta}(u,\dot{y}) - \frac{\beta + \mu_{g^{*}}}{2}\Vert \nabla_u{g_{\beta}}(\hat{u},\dot{y}) - \nabla_u{g_{\beta}}(u,\dot{y})\Vert^2. \end{equation}
\item[$(b)$] For any $\beta > 0$, $\dot{y}\in\mathbb{R}^n$, and $u\in\dom{g}$, we have \begin{equation}\label{eq:key_bound_apdx3} \hspace{-1ex} \begin{array}{lcl} g_{\beta}(u,\dot{y}) \leq g(u) \leq g_{\beta}(u,\dot{y}) + \frac{\beta}{2} [ D_{g}(\dot{y})]^2, \ \text{where} \ D_{g}(\dot{y}) := \sup_{y\in\partial{g(u)}} \norm{y - \dot{y}}. \end{array} \hspace{-1ex} \end{equation}
\item[$(c)$] For $u\in\dom{g}$ and $\dot{y}\in\mathbb{R}^n$, $g_{\beta}(u,\dot{y})$ is convex in $\beta$, and for all $\hat{\beta} \geq \beta > 0$, we have \begin{equation}\label{eq:key_bound_apdx4} g_{\beta}(u,\dot{y}) \leq g_{\hat{\beta}}(u,\dot{y}) + \big(\tfrac{\hat{\beta} - \beta}{2}\big)\norms{\nabla_u{g_{\beta}}(u,\dot{y}) - \dot{y} }^2. \hspace{-2.5ex} \end{equation}
\item[$(d)$] For any $\beta > 0$, and $u, \hat{u}\in\dom{g}$, we have \begin{equation}\label{eq:key_bound_apdx5} \hspace{5ex} \begin{array}{lcl} g_{\beta}(u,\dot{y}) + \iprods{\nabla_u{g_{\beta}}(u, \dot{y}), \hat{u} - u} & \leq & \ell_{\beta}(\hat{u}, \dot{y}) - \frac{\beta}{2}\norms{\nabla_u{g_{\beta}}(u,\dot{y}) - \dot{y}}^2, \end{array} \end{equation} where $\ell_{\beta}(\hat{u}, \dot{y}) := \iprods{\hat{u}, \nabla_u{g_{\beta}}(u,\dot{y}) } - g^{*}(\nabla_u{g_{\beta}}(u,\dot{y})) \leq g(\hat{u}) - \frac{\mu_{g^{*}}}{2}\norms{\nabla_u{g_{\beta}}(u,\dot{y}) - \nabla{g}(\hat{u})}^2$ for any $\nabla{g}(\hat{u}) \in \partial{g}(\hat{u})$. \end{itemize} \end{lemma}
\begin{lemma}\label{le:tech_lemma} The following statements hold. \begin{itemize} \item[$\mathrm{(a)}$] Let $\set{\tau_k}\subset (0, 1]$ be computed by $\tau_{k+1} := \frac{\tau_k}{2}\big[ (\tau_k^2 + 4)^{1/2} - \tau_k\big]$ for some $\tau_0 \in (0, 1]$. Then, we have \begin{equation*}
\tau_k^2 = (1-\tau_k)\tau_{k-1}^2, \quad \frac{1}{k + 1/\tau_0} \leq \tau_k < \frac{2}{k + 2/\tau_0}, \quad\text{and}\quad \frac{1}{1 + \tau_{k-2}} \leq 1 - \tau_k \leq \frac{1}{1+\tau_{k-1}}. \end{equation*} Moreover, we also have \begin{equation*} \begin{array}{ll} & \Theta_{l,k} := \displaystyle\prod_{i=l}^k(1-\tau_i) = \dfrac{\tau_k^2}{\tau_{l-1}^2} \quad\text{for}\ 0\leq l\leq k, \qquad \Theta_{0,k} = \dfrac{(1-\tau_0)\tau_k^2}{\tau_0^2} \leq \dfrac{4(1-\tau_0)}{(\tau_0k+2)^2},
\\ \text{and}\quad &\dfrac{\tau_{l+1}^2}{\tau_{k+2}^2} \leq \Gamma_{l,k} := \displaystyle\prod_{i=l}^k(1+\tau_i) \leq \dfrac{\tau_l^2}{\tau_{k+1}^2} \quad\text{for} \ 0 \leq l \leq k. \end{array} \end{equation*} If we update $\beta_k := \frac{\beta_{k-1}}{1+\tau_k}$ for a given $\beta_0 > 0$, then \begin{equation*} \frac{4\beta_0\tau_0^2}{\tau_1^2[\tau_0(k+1) + 2]^2} \leq \frac{\beta_0\tau_{k+1}^2}{\tau_1^2} \leq \beta_k = \frac{\beta_0}{\Gamma_{1,k}}\leq \frac{\beta_0\tau_{k+2}^2}{\tau_2^2} \leq \frac{4\beta_0\tau_0^2}{\tau_2^2[\tau_0(k+2) + 2]^2}. \end{equation*} \item[$\mathrm{(b)}$] Let $\set{\tau_k}\subset (0, 1]$ be computed by solving $\tau_k^3 + \tau_k^2 + \tau_{k-1}^2\tau_k - \tau_{k-1}^2 = 0$ for all $k\geq 1$ and $\tau_0 := 1$. Then, we have $\frac{1}{k+1} \leq \tau_k \leq \frac{2}{k+2}$ and $\Theta_{1,k} := \prod_{i=1}^k(1-\tau_i) \leq \frac{1}{k+1}$. Moreover, if we update $\beta_k := \frac{\beta_{k-1}}{1+\tau_k}$, then $\beta_k \leq \frac{2\beta_0}{k+2}$. \end{itemize} \end{lemma}
\begin{proof} The first two relations of (a) have been proved, e.g., in \cite{TranDinh2012a}. Let us prove the last inequality of (a). Since $\frac{1}{1+\tau_{k-2}} \leq 1-\tau_k$ is equivalent to $\tau_{k-2}(1-\tau_k) \geq \tau_k$. Using $1- \tau_k = \frac{\tau_k^2}{\tau_{k-1}^2}$, we have $\tau_k\tau_{k-2} \geq \tau_{k-1}^2$. Utilizing $\tau_k = \frac{\tau_{k-1}}{2}\big[(\tau_{k-1}^2 + 4)^2 - \tau_{k-1}\big]$, this condition is equivalent to $\tau_{k-2}^2 \geq \tau_{k-1}^2(1 + \tau_{k-2})$. However, since $\tau_{k-1}^2 = (1-\tau_{k-1})\tau_{k-2}^2$, the last condition becomes $1 \geq (1-\tau_{k-1})(1+\tau_{k-2})$, or equivalently, $\tau_{k-1} \leq \tau_{k-2}$, which automatically holds.
To prove $1-\tau_k \leq \frac{1}{1 + \tau_{k-1}}$, we write it as $\tau_{k-1}(1-\tau_k) \leq \tau_{k}$. Using again $\tau_k^2 = (1-\tau_k)\tau_{k-1}^2$, the last inequality is equivalent to $\tau_k \leq \tau_{k-1}$, which automatically holds. The last statements of (a) is a consequence of $1-\tau_k = \frac{\tau_k^2}{\tau_{k-1}^2}$ and the previous relations.
(b) We consider the function $\varphi(\tau) := \tau^3 + \tau^2 + \tau_{k-1}^2\tau - \tau_{k-1}^2$. Clearly, $\varphi(0) = -\tau_{k-1}^2 < 0$ and $\varphi(1) = 2 > 0$. Moreover, $\varphi'(\tau) = 3\tau^2 + 2\tau + \tau_{k-1}^2 > 0$ for $\tau \in [0, 1]$. Hence, the cubic equation $\varphi(\tau) = 0$ has a unique solution $\tau_k \in (0, 1)$. Therefore, $\sets{\tau_k}_{k\geq 0}$ is well-defined.
Next, since $\tau_k^3 + \tau_k^2 + \tau_k\tau_{k-1}^2 - \tau_{k-1}^2 = 0$ is equivalent to $\tau_{k-1}^2(1-\tau_k) = \tau_k^2(1+\tau_k)$, we have $\tau_{k-1}^2(1-\tau_k) = \tau_k^2(1+\tau_k) \leq \frac{\tau_k^2}{1-\tau_k}$. This inequality becomes $\tau_k \geq \frac{\tau_{k-1}}{1 + \tau_{k-1}}$. By induction and $\tau_0 = 1$, we can easily show that $\tau_k \geq \frac{1}{k+1}$. On the other hand, $\tau_{k-1}^2(1-\tau_k) = \tau_k^2(1+\tau_k) \geq \tau_k^2$. From this inequality, with a similar argument as in the proof of the statement (a), we can also easily show that $\tau_k \leq \frac{2}{k+2}$. Hence, we have $\frac{1}{k+1} \leq \tau_k \leq \frac{2}{k+2}$ for all $k\geq 0$.
Finally, since $\tau_k \geq \frac{1}{k+1}$, we have $\prod_{i=1}^k(1-\tau_i) \leq \prod_{i=1}^k\left(1 - \frac{1}{i+1}\right) = \frac{1}{k+1}$. Alternatively, $\prod_{i=1}^k(1+\tau_i) \geq \prod_{i=1}^k\left(1 + \frac{1}{i+1}\right) = \frac{k+2}{2}$. However, since $\beta_k = \frac{\beta_{k-1}}{1+\tau_k}$, we have $\beta_k = \beta_0\prod_{i=1}^k\frac{1}{1+\tau_i} \leq \frac{2\beta_0}{k+2}$.
$\square$ \end{proof}
\begin{lemma}\label{le:tech_lemma2}$($\cite[Lemma 4]{ZhuLiuTran2020} and \cite{TranDinh2015b}$)$ The following statements hold. \begin{itemize} \item[$\mathrm{(a)}$] For any $u, v, w\in\mathbb{R}^p$ and $t_1, t_2 \in \mathbb{R}$ such that $t_1 + t_2 \neq 0$, we have \begin{equation*} t_1\norms{u - w}^2 + t_2\norms{v - w}^2 = (t_1 + t_2)\norms{w - \tfrac{1}{t_1+t_2}(t_1u + t_2v)}^2 + \tfrac{t_1t_2}{t_1+t_2}\norms{u-v}^2. \end{equation*}
\item[$\mathrm{(b)}$] For any $\tau \in (0, 1)$, $\hat{\beta}, \beta > 0$, $w, z\in\mathbb{R}^p$, we have \begin{equation*} \begin{array}{lcl} \beta(1-\tau) \norms{w - z}^2 + \beta\tau\norms{w}^2 - (1-\tau)(\hat{\beta} - \beta)\norms{z}^2 & = & \beta\norms{w - (1-\tau)z}^2
\\ && + {~} (1-\tau)\big[\tau\beta - (\hat{\beta} - \beta) \big]\norms{z}^2. \end{array} \end{equation*} \end{itemize} \end{lemma}
The following lemma is a key step to address the strongly convex case of $f$ in \eqref{eq:saddle_point_prob}.
\begin{lemma}\label{le:element_est12} Given $L_k > 0$, $\mu_f > 0$, and $\tau_k \in (0, 1)$, let $m_k := \frac{L_k + \mu_f}{L_{k-1} + \mu_f}$ and $a_k := \frac{L_k}{L_{k-1} + \mu_f}$. Assume that the following two conditions hold: \begin{equation}\label{eq:element_cond2} \arraycolsep=0.2em \left\{\begin{array}{llcl} &(1-\tau_k) \big[ \tau_{k-1}^2 + m_k\tau_k \big]&\geq & a_k\tau_k
\\ &m_k\tau_k\tau_{k-1}^2 + m_k^2\tau_k^2 &\geq & a_k\tau_{k-1}^2. \end{array}\right. \end{equation} Let $\set{x^k}$ be a given sequence in $\mathbb{R}^p$. We define $\hat{x}^k := x^k + \frac{1}{\omega_k}(x^k - x^{k-1})$, where $\omega_k$ is chosen such that \begin{equation}\label{eq:x_extra_step} \max\set{\frac{\tau_{k-1} + \sqrt{\tau_{k-1}^2 + 4a_k}}{2(1-\tau_{k-1})}, \frac{a_k\tau_k}{(1-\tau_k)(1-\tau_{k-1})\tau_{k-1}}} \leq \omega_k \leq \frac{\tau_{k-1}^2 + m_k\tau_k}{\tau_{k-1}(1-\tau_{k-1})}. \end{equation} Then, $\omega_k$ is well-defined, and for any $x \in \mathbb{R}^p$, we have \begin{equation}\label{eq:element_est12} \hspace{-2ex} \arraycolsep=0.2em \begin{array}{ll} &L_k\tau_k^2\norms{\frac{1}{\tau_k}[\hat{x}^k - (1-\tau_k)x^k] - x}^2 - \mu_f \tau_k(1-\tau_k)\norms{x^k - x}^2
\\ &\qquad \leq {~} (1-\tau_k) \left(L_{k-1} + \mu_f\right)\tau_{k-1}^2\norms{\frac{1}{\tau_{k-1}}[x^k - (1-\tau_{k-1})x^{k-1}] - x}^2. \end{array} \hspace{-2ex} \end{equation} \end{lemma}
\begin{proof} Firstly, from the definition $\hat{x}^k := x^k + \frac{1}{\omega_k}(x^k - x^{k-1})$ of $\hat{x}^k$, we have $\omega_k(\hat{x}^k - x^k) = x^k - x^{k-1}$. Hence, we can show that \begin{equation*} \arraycolsep=0.2em \begin{array}{lcl} \tau_{k-1}^2\norms{\frac{1}{\tau_{k-1}}[x^k - (1-\tau_{k-1})x^{k-1}] - x}^2 &= & \norms{(1-\tau_{k-1})(x^k - x^{k-1}) + \tau_{k-1}(x^k - x)}^2
\\ &= & \norms{(1-\tau_{k-1})\omega_k(\hat{x}^k - x^{k}) + \tau_{k-1}(x^k - x)}^2
\\ &= & \omega_k^2(1-\tau_{k-1})^2\norms{\hat{x}^k - x^k}^2 + \tau_{k-1}^2\norms{x^k - x}^2
\\ && + {~} 2\omega_k(1-\tau_{k-1})\tau_{k-1}\iprods{\hat{x}^k - x^k, x^k - x}. \end{array} \end{equation*} Alternatively, we also have \begin{equation*} \arraycolsep=0.2em \begin{array}{lcl} \tau_k^2\norms{\frac{1}{\tau_k}[\hat{x}^k - (1-\tau_k)x^k] - x}^2 & = & \norms{\hat{x}^k - x^k}^2 + \tau_k^2\norms{x^k - x}^2 + 2\tau_k\iprods{\hat{x}^k - x^k, x^k - x}. \end{array} \end{equation*} Utilizing the two last expressions, \eqref{eq:element_est12} can be rewritten equivalently to \begin{equation*} \arraycolsep=0.2em \begin{array}{lcl} \mathcal{T}_{[1]} &:= & 2\left[ \left(L_{k-1} + \mu_f\right)(1-\tau_k)(1-\tau_{k-1})\tau_{k-1}\omega_k - L_k\tau_k \right] \iprods{\hat{x}^k - x^k, x - x^k}
\\ &\leq & \left[ \left(L_{k-1} + \mu_f \right)(1-\tau_k)(1-\tau_{k-1})^2\omega_k^2 - L_{k}\right] \norms{\hat{x}^k - x^k}^2
\\ && + {~} \left[ \left(L_{k-1} + \mu_f\right)(1-\tau_k)\tau_{k-1}^2 - L_k\tau_k^2 + \mu_f\tau_k(1-\tau_k)\right] \norms{x^k - x}^2. \end{array} \end{equation*} Now, let us denote \begin{equation*} \arraycolsep=0.2em \left\{\begin{array}{lcl} c_1 &:= & \left(L_{k-1} + \mu_f\right)(1-\tau_k)(1-\tau_{k-1})\tau_{k-1}\omega_k - L_k\tau_k
\\ c_2 &:= & \left(L_{k-1} + \mu_f\right)(1-\tau_k)(1-\tau_{k-1})^2\omega_k^2 - L_k
\\ c_3 &:= & \left(L_{k-1} + \mu_f\right)(1-\tau_k)\tau_{k-1}^2 - L_k\tau_k^2 + \mu_f(1-\tau_k)\tau_k. \end{array}\right. \end{equation*} Then, \eqref{eq:element_est12} is equivalent to \begin{equation}\label{eq:need_to_prove} 2c_1\iprods{\hat{x}^k - x^k, x - x^k} \leq c_2\norms{\hat{x}^k - x^k}^2 + c_3\norms{x - x^k}^2. \end{equation} Secondly, we need to guarantee that $c_1 \geq 0$. This condition holds if we choose $\omega_k$ such that \begin{equation}\label{eq:upper_bound_omegak2} \omega_k \geq \frac{a_k\tau_k}{(1-\tau_k)(1-\tau_{k-1})\tau_{k-1}}. \end{equation} Thirdly, we also need to guarantee $c_2 \geq c_1$, which is equivalent to \begin{equation*} c_2 - c_1 = \left(L_{k-1} + \mu_f\right)(1-\tau_k)(1-\tau_{k-1})\left[(1-\tau_{k-1})\omega_k^2 - \tau_{k-1}\omega_k \right] - L_k(1-\tau_k) \geq 0. \end{equation*} This condition holds if \begin{equation}\label{eq:upper_bound_omegak1} \omega_k \geq \frac{\tau_{k-1} + \sqrt{\tau_{k-1}^2 + 4a_k}}{2(1-\tau_{k-1})}. \end{equation} Alternatively, we also need to guarantee $c_3 \geq c_1$, which is equivalent to \begin{equation*} c_3 - c_1 = \left(L_{k-1} + \mu_f\right)(1-\tau_k)\left[\tau_{k-1}^2 - (1-\tau_{k-1})\tau_{k-1}\omega_k \right] + (L_k + \mu_f)\tau_k(1-\tau_k) \geq 0. \end{equation*} This condition holds if \begin{equation}\label{eq:lower_bound_omegak} \omega_k \leq \frac{\tau_{k-1}^2 + m_k\tau_k}{\tau_{k-1}(1-\tau_{k-1})}. \end{equation} Combining \eqref{eq:upper_bound_omegak2}, \eqref{eq:upper_bound_omegak1}, and \eqref{eq:lower_bound_omegak}, we obtain \begin{equation*} \max\set{\frac{\tau_{k-1} + \sqrt{\tau_{k-1}^2 + 4a_k}}{2(1-\tau_{k-1})}, \frac{a_k\tau_k}{(1-\tau_k)(1-\tau_{k-1})\tau_{k-1}}} \leq \omega_k \leq \frac{\tau_{k-1}^2 + m_k\tau_k}{\tau_{k-1}(1-\tau_{k-1})}, \end{equation*} which is exactly \eqref{eq:x_extra_step}. Here, under the condition \eqref{eq:element_cond2}, the left-hand side of the last expression is less than or equal to the right-hand side. Therefore, $\omega_k$ is well-defined.
Finally, under the choice of $\omega_k$ as in \eqref{eq:x_extra_step}, we have $c_2 \geq c_1 \geq 0$ and $c_3\geq c_1 \geq 0$. Hence, \eqref{eq:need_to_prove} holds, which is also equivalent to \eqref{eq:element_est12}.
$\square$ \end{proof}
\section{Appendix 2: Technical Proof of Lemmas \ref{le:descent_pro} and \ref{le:maintain_gap_reduction1} in Section~\ref{sec:alg_scheme}}\label{apdx:sec:A2}
This section provides the full proof of Lemma~\ref{le:descent_pro} and Lemma~\ref{le:maintain_gap_reduction1} in the main text.
\subsection{The proof of Lemma~\ref{le:descent_pro}: Key estimate of the primal-dual step~\eqref{eq:prox_grad_step0}}\label{apdx:le:descent_pro}
\begin{proof} From the first line of \eqref{eq:prox_grad_step0} and Lemma~\ref{le:g_properties}(a), we have $\nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}) = K^{\top}y^{k+1}$. Now, from the second line of \eqref{eq:prox_grad_step0}, we also have \begin{equation*} 0 \in \partial{f}(x^{k+1}) + L_k(x^{k+1} - \hat{x}^k) + K^{\top}\nabla_u{g_{\beta_k}}(K\hat{x}^k, \dot{y}). \end{equation*} Combining this inclusion and the $\mu_f$-convexity of $f$, for any $x\in\dom{f}$, we get \begin{equation*} \begin{array}{lcl} f(x^{k+1}) & \leq & f(x) + \iprods{\nabla_u{g_{\beta_k}}(K\hat{x}^k, \dot{y}), K(x - x^{k+1})} + L_k\iprods{x^{k+1} - \hat{x}^k, x - x^{k+1}}
\\ && - {~} \frac{\mu_f}{2}\norms{x^{k+1} - x}^2. \end{array} \end{equation*} Since $g_{\beta}(\cdot, \dot{y})$ is $\frac{1}{\beta + \mu_{g^{*}}}$-smooth by Lemma~\ref{le:g_properties}(a), for any $x\in\dom{f}$, we have \begin{equation*} \arraycolsep=0.2em \begin{array}{lcl} g_{\beta_k}(Kx^{k+1}, \dot{y}) & \leq & g_{\beta_k}(K\hat{x}^k, \dot{y}) + \iprods{\nabla_u{g}_{\beta_k}(K\hat{x}^k, \dot{y}), K(x^{k+1} - \hat{x}^k)}
\\ && + {~} \frac{1}{2(\beta_k + \mu_{g^{*}})}\Vert K(x^{k+1} - \hat{x}^k)\Vert^2
\\ & = & g_{\beta_k}(K\hat{x}^k, \dot{y}) + \iprods{\nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}), K(x - \hat{x}^k)}
\\ && - {~} \iprods{\nabla_u{g}_{\beta_k}(K\hat{x}^k, \dot{y}), K(x - x^{k+1})}
\\ && + {~} \frac{1}{2(\mu_{g^{*}} + \beta_k)}\Vert K(x^{k+1} - \hat{x}^k)\Vert^2. \end{array} \end{equation*} Now, combining the last two estimates, we get \begin{equation}\label{eq:proof_est1} \hspace{-1ex} \arraycolsep=0.2em \begin{array}{lcl} f(x^{k+1}) + g_{\beta_k}(Kx^{k+1}, \dot{y}) & \leq & f(x) + g_{\beta_k}(K\hat{x}^k, \dot{y}) + \iprods{\nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}), K(x - \hat{x}^k)}
\\ && + {~} L_k\iprods{x^{k+1} - \hat{x}^k, x - \hat{x}^k} - L_k\norms{x^{k+1} - \hat{x}^k}^2
\\ && + {~} \frac{1}{2(\mu_{g^{*}} + \beta_k)}\Vert K(x^{k+1} - \hat{x}^k)\Vert^2 - \frac{\mu_f}{2}\norms{x - x^{k+1}}^2. \end{array} \hspace{-5ex} \end{equation} Using Lemma~\ref{le:g_properties}(a) again, we have \begin{equation}\label{eq:proof_est2} \arraycolsep=0.2em \hspace{-2ex} \begin{array}{lcl} \ell_{\beta_k}(x^k, \dot{y}) &:= & g_{\beta_k}(K\hat{x}^k, \dot{y}) + \iprods{\nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}), K(x^k - \hat{x}^k)}
\\ &\leq & g_{\beta_k}(Kx^k, \dot{y}) - \frac{\beta_k+\mu_{g^{*}}}{2}\norms{\nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}) - \nabla_ug_{\beta_k}(Kx^k, \dot{y})}^2. \end{array} \hspace{-2ex} \end{equation} Substituting $x := x^k$ into \eqref{eq:proof_est1}, and multiplying the result by $1-\tau_k$ and adding the result to \eqref{eq:proof_est1} after multiplying it by $\tau_k$, then using \eqref{eq:proof_est2}, we can derive \begin{equation}\label{eq:proof_est3} \hspace{-1ex} \arraycolsep=0.2em \begin{array}{lcl} F_{\beta_k}(x^{k+1}, \dot{y}) &:= & f(x^{k+1}) + g_{\beta_k}(Kx^{k+1}, \dot{y})
\\ & \leq & (1 - \tau_k)[ f(x^k) + g_{\beta_k}(Kx^k, \dot{y})] + \tau_k\left[f(x) + \ell_{\beta_k}(x, \dot{y}) \right]
\\ && - {~} L_k\norms{x^{k+1} - \hat{x}^k}^2 + \frac{1}{2(\mu_{g^{*}}+\beta_k)}\Vert K(x^{k+1} - \hat{x}^k)\Vert^2
\\ && + {~} L_k\iprods{x^{k+1} - \hat{x}^k, \tau_kx - \hat{x}^k + (1-\tau_k)x^k}
\\ && - {~} \frac{\mu_f}{2}\left[(1-\tau_k)\norms{x^{k+1} - x^k}^2 + \tau_k\norms{x - x^{k+1}}^2\right]
\\ && - {~} \frac{(1-\tau_k)(\beta_k+\mu_{g^{*}})}{2}\norms{\nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}) - \nabla_ug_{\beta_k}(Kx^k, \dot{y})}^2. \end{array} \hspace{-5ex} \end{equation} From Lemma~\ref{le:tech_lemma2}(a), we can easily show that \begin{equation*} \begin{array}{lcl} (1-\tau_k)\norms{x^{k+1} - x^k}^2 + \tau_k\norms{x^{k+1} - x}^2 & = & \tau_k^2\norms{\tfrac{1}{\tau_k}[ x^{k+1} - (1-\tau_k)x^k] - x}^2
\\ && + {~} \tau_k(1-\tau_k)\norms{x - x^k}^2. \end{array} \end{equation*} We also have the following elementary relation \begin{equation*} \arraycolsep=0.1em \begin{array}{lcl} \iprods{x^{k+1} - \hat{x}^k, \tau_k x - [\hat{x}^k - (1-\tau_k)x^k]} &= & \frac{\tau_k^2}{2}\norms{\tfrac{1}{\tau_k}[\hat{x}^k - (1-\tau_k)x^k] - x}^2 + \frac{1}{2}\norms{x^{k+1} - \hat{x}^k}^2
\\ && - {~} \frac{\tau_k^2}{2}\norms{\tfrac{1}{\tau_k}[x^{k+1} - (1-\tau_k)x^k] - x}^2. \end{array} \end{equation*} Substituting the two last expressions into \eqref{eq:proof_est3}, we obtain \begin{equation}\label{eq:proof4} \arraycolsep=0.2em \begin{array}{lcl} F_{\beta_k}(x^{k+1}, \dot{y}) & \leq & (1 - \tau_k) F_{\beta_k}(x^k, \dot{y}) + \tau_k \left[ f(x) + \ell_{\beta_k}(x, \dot{y}) \right]
\\ && + {~} \frac{L_k\tau_k^2}{2} \norms{\tfrac{1}{\tau_k}[\hat{x}^k - (1-\tau_k)x^k] - x}^2
\\ && - {~} \frac{\tau_k^2}{2}\left( L_k + \mu_f \right) \norms{\tfrac{1}{\tau_k}[x^{k+1} - (1-\tau_k)x^k] - x}^2
\\ && - {~} \frac{(1-\tau_k)(\mu_{g^{*}} + \beta_k)}{2}\norms{\nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}) - \nabla_ug_{\beta_k}(Kx^k, \dot{y})}^2
\\ && - {~} \frac{L_k}{2}\norms{x^{k+1} - \hat{x}^k}^2 + \frac{1}{2(\mu_{g^{*}}+\beta_k)}\Vert K(x^{k+1} - \hat{x}^k)\Vert^2
\\ && - {~} \frac{\mu_f(1-\tau_k)\tau_k}{2}\norms{x - x^k}^2. \end{array} \hspace{-2ex} \end{equation} One the one hand, by \eqref{eq:key_bound_apdx4} of Lemma~\ref{le:g_properties}, we have \begin{equation*} F_{\beta_k}(x^k, \dot{y}) \leq F_{\beta_{k-1}}(x^k, \dot{y}) + \frac{(\beta_{k-1} - \beta_k)}{2}\norms{\nabla_ug_{\beta_{k}}(Kx^k, \dot{y}) - \dot{y}}^2. \end{equation*} On the other hand, by \eqref{eq:key_bound_apdx5} of Lemma~\ref{le:g_properties}, we get \begin{equation*} f(x) + \ell_{\beta_k}(x, \dot{y}) \leq \mathcal{L}(x, y^{k+1}) - \frac{\beta_k}{2}\norms{ \nabla_ug_{\beta_{k}}(K\hat{x}^k, \dot{y}) - \dot{y}}^2, \end{equation*} where $\mathcal{L}(x, y^{k+1}) := f(x) + \iprods{Kx, y^{k+1}} - g^{*}(y^{k+1})$ is the Lagrange function in \eqref{eq:saddle_point_prob}.
Now, substituting the last two inequalities into \eqref{eq:proof4}, and using Lemma~\ref{le:tech_lemma2}(b) with $w := \nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}) - \dot{y}$ and $z := \nabla_ug_{\beta_k}(Kx^k, \dot{y}) - \dot{y}$, we arrive at \begin{equation*} \arraycolsep=0.2em \begin{array}{lcl} F_{\beta_k}(x^{k+1}, \dot{y}) & \leq & (1 - \tau_k) F_{\beta_{k-1}}(x^k, \dot{y}) + \tau_k\mathcal{L}(x, y^{k+1}) + \frac{L_k\tau_k^2}{2} \norms{\tfrac{1}{\tau_k}[\hat{x}^k - (1-\tau_k)x^k] - x}^2
\\ && - {~} \frac{\tau_k^2}{2}\left( L_k + \mu_f \right) \norms{\tfrac{1}{\tau_k}[x^{k+1} - (1-\tau_k)x^k] - x}^2 - \frac{\mu_f(1-\tau_k)\tau_k}{2}\norms{x - x^k}^2
\\ && - {~} \frac{L_k}{2}\norms{x^{k+1} - \hat{x}^k}^2 + \frac{1}{2(\mu_{g^{*}}+\beta_k)}\Vert K(x^{k+1} - \hat{x}^k)\Vert^2
\\ && - {~} \frac{(1-\tau_k)}{2}\left[ \tau_k\beta_k - (\beta_{k-1} - \beta_k)\right]\norms{ \nabla_ug_{\beta_k}(Kx^k, \dot{y}) - \dot{y}}^2
\\ && - {~} \frac{\beta_k}{2}\norms{\nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}) - \dot{y} - (1-\tau_k)\left[ \nabla_ug_{\beta_k}(Kx^k, \dot{y}) - \dot{y} \right]}^2
\\ && - {~} \frac{(1-\tau_k)\mu_{g^{*}}}{2}\norms{\nabla_ug_{\beta_k}(K\hat{x}^k, \dot{y}) - \nabla_ug_{\beta_k}(Kx^k, \dot{y})}^2. \end{array} \end{equation*} By dropping the last two nonpositive terms in the last inequality, we obtain \eqref{eq:key_est00_a}.
$\square$ \end{proof}
\subsection{The proof of Lemma~\ref{le:maintain_gap_reduction1}: Recursive estimate of the Lyapunov function}\label{apdx:le:maintain_gap_reduction1}
\begin{proof} First, from the last line $\tilde{y}^{k+1} = (1-\tau_k)\tilde{y}^k + \tau_ky^{k+1}$ of \eqref{eq:prox_grad_scheme}, and the $\mu_{g^{*}}$-convexity of $g^{*}$, we have \begin{equation*} \begin{array}{lcl} \mathcal{L}(x, \tilde{y}^{k+1}) & := & f(x) + \iprods{Kx, \tilde{y}^{k+1}} - g^{*}(\tilde{y}^{k+1})
\\ & \geq & (1-\tau_k)\mathcal{L}(x, \tilde{y}^k) + \tau_k\mathcal{L}(x, y^{k+1}) + \frac{\mu_{g^{*}}\tau_k(1-\tau_k)}{2}\norms{y^{k+1} - \tilde{y}^k}^2. \end{array} \end{equation*} Hence, $\tau_k\mathcal{L}(x, y^{k+1}) \leq \mathcal{L}(x, \tilde{y}^{k+1}) - (1-\tau_k)\mathcal{L}(x, \tilde{y}^k) - \frac{\mu_{g^{*}}\tau_k(1-\tau_k)}{2}\norms{y^{k+1} - \tilde{y}^k}^2$. Substituting this estimate into \eqref{eq:key_est00_a} and dropping the term $- \frac{\mu_{g^{*}}\tau_k(1-\tau_k)}{2}\norms{y^{k+1} - \tilde{y}^k}^2$, we can derive \begin{equation}\label{eq:key_est00_proof1} \hspace{-2ex} \arraycolsep=0.1em \begin{array}{lcl} F_{\beta_k}(x^{k+1},\dot{y}) & \leq & (1 - \tau_k) F_{\beta_{k-1}}(x^k, \dot{y}) + \mathcal{L}(x, \tilde{y}^{k+1}) - (1-\tau_k)\mathcal{L}(x, \tilde{y}^k)
\\ && + {~} \frac{L_k\tau_k^2}{2} \big\Vert \tfrac{1}{\tau_k}[\hat{x}^k - (1-\tau_k)x^k] - x \big\Vert^2
\\ && - {~} \frac{\tau_k^2}{2}\left(L_k + \mu_f\right) \big\Vert \tfrac{1}{\tau_k}[x^{k+1} - (1-\tau_k)x^k] - x \big\Vert^2
\\ && - {~} \frac{L_k}{2}\norms{x^{k+1} - \hat{x}^k}^2 + \frac{1}{2(\mu_{g^{*}} + \beta_k)}\Vert K(x^{k+1} - \hat{x}^k)\Vert^2
\\ && - {~} \frac{\mu_f\tau_k(1-\tau_k)}{2}\norms{x^k - x}^2. \end{array} \hspace{-2ex} \end{equation} Now, it is obvious to show that the condition \eqref{eq:param_cond} is equivalent to the condition \eqref{eq:element_cond2} of Lemma~\ref{le:element_est12}. In addition, we choose $\eta_k = \frac{1}{\omega_k}$ in our update \eqref{eq:Lk_m_k}, where $\omega_k := \frac{\tau_{k-1}^2 + m_k\tau_k}{\tau_{k-1}(1-\tau_{k-1})}$, which is the upper bound of \eqref{eq:x_extra_step}. Hence, \eqref{eq:x_extra_step} automatically holds. Using \eqref{eq:element_est12}, we have \begin{equation*} \begin{array}{lcl} \mathcal{T}_{[2]} &:= & \frac{L_k\tau_k^2}{2} \big\Vert \tfrac{1}{\tau_k}[\hat{x}^k - (1-\tau_k)x^k] - x \big\Vert^2 - \frac{\mu_f\tau_k(1-\tau_k)}{2}\norms{x^k - x}^2
\\ &\leq & \frac{\tau_{k-1}^2}{2} (1-\tau_k)\left(L_{k-1} + \mu_f\right)\big\Vert \tfrac{1}{\tau_{k-1}}[x^{k} - (1-\tau_{k-1})x^{k-1}] - x \big\Vert^2. \end{array} \end{equation*} Moreover, $\frac{1}{2(\mu_{g^{*}} + \beta_k)}\Vert K(x^{k+1} - \hat{x}^k)\Vert^2 \leq \frac{\norms{K}^2}{2(\mu_{g^{*}} + \beta_k)}\Vert x^{k+1} - \hat{x}^k\Vert^2 = \frac{L_k}{2}\norms{x^{k+1} - \hat{x}^k}^2$ due to the definition of $L_k$ in \eqref{eq:Lk_m_k}. Substituting these two estimates into \eqref{eq:key_est00_proof1}, and utilizing the definition \eqref{eq:lyapunov_func1} of $\mathcal{V}_k$, we obtain \eqref{eq:key_est1_ncvx}.
$\square$ \end{proof}
\end{document}
|
arXiv
|
{
"id": "2104.01314.tex",
"language_detection_score": 0.5711344480514526,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Fixed points of $n$-valued maps, the fixed point property and the case of surfaces~--~a braid approach}
\author{DACIBERG~LIMA~GON\c{C}ALVES\\ Departamento de Matem\'atica - IME-USP,\\ Caixa Postal~66281~-~Ag.~Cidade de S\~ao Paulo,\\ CEP:~05314-970 - S\~ao Paulo - SP - Brazil.\\ e-mail:~\url{[email protected]}\vspace*{4mm}\\ JOHN~GUASCHI\\ Normandie Universit\'e, UNICAEN,\\ Laboratoire de Math\'ematiques Nicolas Oresme UMR CNRS~\textup{6139},\\ CS 14032, 14032 Caen Cedex 5, France.\\ e-mail:~\url{[email protected]}}
\maketitle
\begin{abstract} \noindent We study the fixed point theory of $n$-valued maps of a space $X$ using the fixed point theory of maps between $X$ and its configuration spaces. We give some general results to decide whether an $n$-valued map can be deformed to a fixed point free $n$-valued map. In the case of surfaces, we provide an algebraic criterion in terms of the braid groups of $X$ to study this problem. If $X$ is either the $k$-dimensional ball or an even-dimensional real or complex projective space, we show that the fixed point property holds for $n$-valued maps for all $n\geq 1$, and we prove the same result for even-dimensional spheres for all $n\geq 2$. If $X$ is the $2$-torus, we classify the homotopy classes of $2$-valued maps in terms of the braid groups of $X$. We do not currently have a complete characterisation of the homotopy classes of split $2$-valued maps of the $2$-torus that contain a fixed point free representative, but we give an infinite family of such homotopy classes.
\end{abstract}
\section{Introduction}\label{sec:intro}
Multifunctions and their fixed point theory have been studied for a number of years, see for example the books~\cite{Be,Gor}, where fairly general classes of multifunctions and spaces are considered. Continuous $n$-valued functions are of particular interest, and more information about their fixed point theory on finite complexes may be found in~\cite{Brr1,Brr2,Brr3,Brr4,Brr5,Brr6,Bet1,Bet2,Sch0,Sch1,Sch2}. In this paper, we will concentrate our attention on the case of metric spaces, and notably that of surfaces. In all of what follows, $X$ and $Y$ will be topological spaces, and $\phi\colon\thinspace X \multimap Y$ will be a multivalued function \emph{i.e.}\ a function that to each $x\in X$ associates a non-empty subset $\phi(x)$ of $Y$. Following the notation and the terminology of the above-mentioned papers, a multifunction $\phi\colon\thinspace X \multimap Y$ is \emph{upper semi-continuous} if for all $x\in X$, $\phi(x)$ is closed, and given an open set $V$ in $Y$, the set $\set{x\in X}{\phi(x)\subset V}$ is open in $X$, is \emph{lower semi-continuous} if the set $\set{x \in X}{\phi(x)\cap V \neq \varnothing}$ is open in $X$, and is \emph{continuous} if it is upper semi-continuous and lower semi-continuous. Let $I$ denote the unit interval $[0,1]$. We recall the definitions of (split) $n$-valued maps.
\begin{defns} Let $X$ and $Y$ be topological spaces, and let $n\in \ensuremath{\mathbb N}$. \begin{enumerate} \item An \emph{$n$-valued map} (or multimap) $\phi\colon\thinspace X \multimap Y$, is a continuous multifunction that to each $x\in X$ assigns an unordered subset of $Y$ of cardinal exactly $n$. \item A \emph{homotopy} between two $n$-valued maps $\phi_1,\phi_2\colon\thinspace X \multimap Y$ is an $n$-valued map $H\colon\thinspace X\times I \multimap Y$ such that $\phi_1=H ( \cdot , 0)$ and $\phi_2=H ( \cdot , 1)$. \end{enumerate} \end{defns}
\begin{defn}[\cite{Sch0}] An $n$-valued function $\phi\colon\thinspace X \multimap Y$ is said to be a \emph{split $n$-valued map} if there exist single-valued maps $f_1, f_2, \ldots, f_n\colon\thinspace X \ensuremath{\longrightarrow} Y$ such that $\phi(x)=\brak{f_1(x),\ldots,f_n(x)}$ for all $x\in X$. This being the case, we shall write $\phi=\brak{f_1,\ldots,f_n}$. Let $\splitmap{X}{Y}{n}$ denote the set of split $n$-valued maps between $X$ and $Y$.
\end{defn}
\emph{A priori}, $\phi\colon\thinspace X \multimap Y$ is just an $n$-valued function, but if it is split then it is continuous by \repr{multicont} in the Appendix, which justifies the use of the word `map' in the above definition. Partly for this reason, split $n$-valued maps play an important r\^ole in the theory.
We now recall the notion of coincidence of a pair $(\phi, f)$ where $\phi$ is an $n$-valued map and $f\colon\thinspace X \ensuremath{\longrightarrow} Y$ is a single-valued map (meaning continuous)~\emph{cf.}~\cite{Brr5}. Let $\ensuremath{\operatorname{\text{Id}}}_{X}\colon\thinspace X \ensuremath{\longrightarrow} X$ denote the identity map of $X$.
\begin{defns} Let $\phi\colon\thinspace X\multimap Y$ be an $n$-valued map, and let $f\colon\thinspace X \ensuremath{\longrightarrow} Y$ be a single-valued map. The set of coincidences of the pair $(\phi, f)$ is denoted by $\operatorname{\text{Coin}}(\phi, f)=\set{x\in X}{f(x)\in \phi(x)}$. If $X=Y$ and $f=\ensuremath{\operatorname{\text{Id}}}_X$ then $\operatorname{\text{Coin}}(\phi, \ensuremath{\operatorname{\text{Id}}}_X)=\set{x\in X}{x\in \phi(x)}$ is called the \emph{fixed point set} of $\phi$, and will be denoted by $\operatorname{\text{Fix}}(\phi)$. If $f$ is the constant map $c_{y_0}$ at a point $y_0\in Y$ then $\operatorname{\text{Coin}}(\phi, c_{y_0})=\set{x\in X}{y_0\in \phi(x)}$ is called the set of \emph{roots} of $\phi$ at $y_0$. \end{defns}
Recall that a space $X$ is said to have the \emph{fixed point property} if any self-map of $X$ has a fixed point. This notion may be generalised to $n$-valued maps as follows.
\begin{defn}\label{fppro} If $n\in \ensuremath{\mathbb N}$, a space $X$ is said to have the \emph{fixed point property} for $n$-valued maps if any $n$-valued map $\phi\colon\thinspace X \multimap X$ has a fixed point. \end{defn}
If $n=1$ then we obtain the classical notion of the fixed point property. It is well known that the fixed point theory of surfaces is more complicated than that of manifolds of higher dimension. This is also the case for $n$-valued maps. A number of results for singled-valued maps of manifolds of dimension at least three may be generalised to the setting of $n$-valued maps, see for example the results of Schirmer from the 1980's~\cite{Sch0,Sch1,Sch2}. In dimension one or two, the situation is more complex, and has only been analysed within the last ten years or so, see~\cite{Brr1} for the study of $n$-valued maps of the circle. The papers~\cite{Brr4,Brr6} illustrate some of the difficulties that occur when the manifold is the $2$-torus $\ensuremath{\mathbb{T}^{2}}$. Our expectation is that the case of surfaces of negative Euler characteristic will be much more involved.
In this paper, we explore the fixed point property for $n$-valued maps, and we extend the famous result of L.~E.~J.~Brouwer that every self-map of the disc has a fixed point to this setting \cite{Bru}. We will also develop some tools to decide whether an $n$-valued map can be deformed to a fixed point free $n$-valued map, and we give a partial classification of those split $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$ that can be deformed to fixed point free $2$-valued maps. Our approach to the study of fixed point theory of $n$-valued maps makes use of the homotopy theory of configuration spaces. It is probable that these ideas can also be adapted to coincidence theory. This viewpoint is fairly general. It helps us to understand the theory, and provides some means to perform (not necessarily easy) computations in general. Nevertheless, for some specific situations, such as for surfaces of non-negative Euler characteristic, these calculations are often tractable. To explain our approach, let $F_{n}(Y)$ denote the \emph{$n\ensuremath{\up{th}}$ (ordered) configuration space} of a space $Y$, defined by: \begin{equation*} F_n(Y)=\setr{(y_1,\ldots,y_n)}{\text{$y_i\in Y$, and $y_i\neq y_j$ if $i\neq j$}}. \end{equation*} Configuration spaces play an important r\^ole in several branches of mathematics and have been extensively studied, see~\cite{CG,FH} for example. The symmetric group $S_n$ on $n$ elements acts freely on $F_n(Y)$ by permuting coordinates. The corresponding quotient space, known as the \emph{$n\ensuremath{\up{th}}$ (unordered) configuration space of $Y$}, will be denoted by $D_n(Y)$, and the quotient map will be denoted by $\pi \colon\thinspace F_{n}(Y) \ensuremath{\longrightarrow} D_{n}(Y)$. The \emph{$n\ensuremath{\up{th}}$ pure braid group $P_n(Y)$} (respectively the \emph{$n\ensuremath{\up{th}}$ braid group $B_n(Y)$}) of $Y$ is defined to be the fundamental group of $F_n(Y)$ (resp.\ of $D_n(Y)$), and there is a short exact sequence: \begin{equation}\label{eq:sesbraid} 1\ensuremath{\longrightarrow} P_n(Y) \ensuremath{\longrightarrow} B_n(Y) \stackrel{\tau}{\ensuremath{\longrightarrow}} S_n \ensuremath{\longrightarrow} 1, \end{equation} where $\tau$ is the homomorphism that to a braid associates its induced permutation. For $i=1,\ldots,n$, let $p_{i}\colon\thinspace F_{n}(Y)\ensuremath{\longrightarrow} Y$ denote projection onto the $i\ensuremath{\up{th}}$ factor. The notion of intermediate configuration spaces was defined in~\cite{GG2,GG4}. More precisely, if $n, m\in \ensuremath{\mathbb N}$, the subgroup $S_n\times S_m \subset S_{n+m}$ acts freely on $F_{n+m}(Y)$ by restriction, and the corresponding orbit space $F_{n+m}(Y)/(S_n\times S_m)$ is denoted by $D_{n, m}(Y)$. Let $B_{n,m}=\pi_{1}(D_{n, m}(Y))$ denote the associated `mixed' braid group. The space $F_{n+m}(Y)$ is equipped with the topology induced by the inclusion $F_{n+m}(Y)\subset Y^{n+m}$, and $D_{n, m}(Y)$ is equipped with the quotient topology. If $Y$ is a manifold without boundary then the natural projections $\overline{p}_{m,n}\colon\thinspace D_{m,n}(Y) \ensuremath{\longrightarrow} D_m(Y)$ onto the first $m$ coordinates are fibrations. For maps whose target is a configuration space, we have the following notions.
\begin{defns} Let $X$ and $Y$ be topological spaces, and let $n\in \ensuremath{\mathbb N}$. A map $\Phi\colon\thinspace X \ensuremath{\longrightarrow} D_n(Y)$ will be called an \emph{$n$-unordered map}, and a map $\Psi\colon\thinspace X \ensuremath{\longrightarrow} F_n(Y)$ will be called an \emph{$n$-ordered map}. For such an $n$-ordered map, for $i=1,\ldots,n$, there exist maps $f_i\colon\thinspace X \ensuremath{\longrightarrow} Y$ such that $\Psi(x)=(f_1(x),\ldots, f_n(x))$ for all $x\in X$, and for which $f_i(x)\neq f_j(x)$ for all $1\leq i,j\leq n$, $i\neq j$, and all $x\in X$. In this case, we will often write $\Psi=(f_{1},\ldots,f_{n})$. \end{defns}
The fixed point-theoretic concepts that were defined earlier for $n$-valued maps carry over naturally to $n$-unordered and $n$-ordered maps as follows.
\begin{defns} Let $X$ and $Y$ be topological spaces, let $f\colon\thinspace X \ensuremath{\longrightarrow} Y$ be a single-valued map, let $y_0\in Y$, and let $n\in \ensuremath{\mathbb N}$. \begin{enumerate}[(a)] \item Given an $n$-unordered map $\Phi\colon\thinspace X \ensuremath{\longrightarrow} D_n(Y)$, $x\in X$ is said to be a \emph{coincidence} of the pair $(\Phi, f)$ if there exist $(x_1,\ldots,x_n)\in F_n(Y)$ and $j\in\brak{1,\ldots,n}$ such that $\Phi(x)= \pi(x_1,\ldots, x_n)$ and $f(x)=x_j$. The set of coincidences of the pair $(\Phi, f)$ will be denoted by $\operatorname{\text{Coin}}(\Phi, f)$.
If $X=Y$ and $f=\ensuremath{\operatorname{\text{Id}}}_X$ then $\operatorname{\text{Coin}}(\Phi, \ensuremath{\operatorname{\text{Id}}}_X)$
is called the \emph{fixed point set} of $\Phi$, and is denoted by $\operatorname{\text{Fix}}(\Phi)$. If $f$ is the constant map $c_{y_0}$ at $y_0$ then $\operatorname{\text{Coin}}(\Phi, c_{y_0})$
is called the set of \emph{roots} of $\Phi$ at $y_0$.
\item Given an $n$-ordered map $\Psi\colon\thinspace X\ensuremath{\longrightarrow} F_n(Y)$, the set of coincidences of the pair $(\Psi, f)$ is defined by $\operatorname{\text{Coin}}(\Psi, f)= \set{x\in X}{\text{$f(x)= p_j\circ \Psi(x)$ for some $1\leq j\leq n$}}$. If $X=Y$ and $f=\ensuremath{\operatorname{\text{Id}}}_X$ then $\operatorname{\text{Coin}}(\Psi, \ensuremath{\operatorname{\text{Id}}}_X)=\set{x\in X}{\text{$x=p_j\circ \Psi(x)$ for some $1\leq j\leq n$}}$ is called the \emph{fixed point set} of $\Psi$, and is denoted by $\operatorname{\text{Fix}}(\Psi)$. If $f$ is the constant map $c_{y_0}$ then $\operatorname{\text{Coin}}(\Psi, c_{y_0})=\set{x\in X}{\text{$y_0=p_j\circ \Psi(x)$ for some $1\leq j\leq n$}}$ is called the set of \emph{roots} of $\Psi$ at $y_0$. \end{enumerate} \end{defns}
In order to study $n$-valued maps via single-valued maps, we use the following natural relation between multifunctions and functions. First observe that there is an obvious bijection between the set of $n$-point subsets of a space $Y$ and the unordered configuration space $D_{n}(Y)$. This bijection induces a one-to-one correspondence between the set of $n$-valued functions from $X$ to $Y$ and the set of functions from $X$ to $D_n(Y)$. In what follows, given an $n$-valued function $\phi\colon\thinspace X \multimap Y$, we will denote the corresponding function whose target is the configuration space $D_{n}(Y)$ by $\Phi\colon\thinspace X \ensuremath{\longrightarrow} D_n(Y)$, and \emph{vice-versa}. Since we are concerned with the study of continuous multivalued functions, we wish to ensure that this correspondence restricts to a bijection between the set of (continuous) $n$-valued maps and the set of continuous single-valued maps whose target is $D_{n}(Y)$. It follows from \reth{metriccont} that this is indeed the case if $X$ and $Y$ are metric spaces. This hypothesis will clearly be satisfied throughout this paper. If the map $\Phi\colon\thinspace X \ensuremath{\longrightarrow} D_n(Y)$ associated to $\phi$ admits a lift $\widehat{\Phi}\colon\thinspace X \ensuremath{\longrightarrow} F_n(Y)$ via the covering map $\pi$ then we shall say that $\widehat{\Phi}$ is a \emph{lift} of $\phi$ (see \resec{pres} for a formal statement of this definition). We will make use of this notion to develop a correspondence between split $n$-valued maps and maps from $X$ into $F_n(Y)$. As we shall see, the problems that we are interested in for $n$-valued maps, such as coincidence, fixed point and root problems, may be expressed within the context of $n$-unordered maps, to which we may apply the classical theory of single-valued maps.
Our main aims in this paper are to explore the fixed point property of spaces for $n$-valued maps, and to study the problem of whether an $n$-valued map map can be deformed to a fixed point free $n$-valued map. We now give the statements of the main results of this paper. The first theorem shows that for simply-connected metric spaces, the usual fixed point property implies the fixed point property for $n$-valued maps.
\begin{thm}\label{th:sccfpp} Let $X$ be a simply-connected metric space that has the fixed point property, and let $n\in \ensuremath{\mathbb N}$. Then every $n$-valued map of $X$ has at least $n$ fixed points, so $X$ has the fixed point property for $n$-valued maps. In particular, for all $n,k\geq 1$, the $k$-dimensional disc $\dt[k]$ and the $2k$-dimensional complex projective space $\ensuremath{\mathbb C} P^{2k}$ have the fixed point property for $n$-valued maps. \end{thm}
It may happen that a space does not have the (usual) fixed point property but that it has the fixed point property for $n$-valued maps for $n>1$. This is indeed the case for the $2k$-dimensional sphere $\St[2k]$.
\begin{prop}\label{prop:S2fp} If $n\geq 2$ and $k\geq 1$, $\St[2k]$ has the fixed point property for $n$-valued maps. \end{prop}
\reth{sccfpp} and \repr{S2fp} will be proved in \resec{pres}. Although the $2k$-dimensional real projective space $\ensuremath{\mathbb R} P^{2k}$ is not simply connected, in \resec{rp2k} we will show that it has the fixed point property for $n$-valued maps for all $n\in \ensuremath{\mathbb N}$.
\begin{thm}\label{th:rp2Kfpp} Let $k,n\geq 1$. The real projective space $\ensuremath{\mathbb R} P^{2k}$ has the fixed point property for $n$-valued maps. Further, any $n$-valued map of $\ensuremath{\mathbb R} P^{2k}$ has at least $n$ fixed points. \end{thm}
We do not know of an example of a space that has the fixed point property, but that does not have the fixed point property for $n$-valued maps for some $n\geq 2$.
In \resec{fixfree}, we turn our attention to the question of deciding whether an $n$-valued map of a surface $X$ of non-negative Euler characteristic $\chi(X)$ can be deformed to a fixed point free $n$-valued map. In the following result, we give algebraic criteria involving the braid groups of $X$.
\begin{thm}\label{th:defchineg} Let $X$ be a compact surface without boundary such that $\chi(X)\leq 0$, let $n\geq 1$, and let $\phi\colon\thinspace X \multimap X$ be an $n$-valued map. \begin{enumerate}[(a)] \item\label{it:defchinega} The $n$-valued map $\phi$ can be deformed to a fixed point free $n$-valued map if and only if there is a homomorphism $\varphi\colon\thinspace \pi_1(X) \ensuremath{\longrightarrow} B_{1,n}(X)$ that makes the following diagram commute: \begin{equation}\label{eq:commdiag1} \begin{tikzcd}[ampersand replacement=\&] \&\& B_{1,n}(X) \ar{d}{(\iota_{1,n})_{\#}}\\ \pi_{1}(X) \ar[swap]{rr}{(\ensuremath{\operatorname{\text{Id}}}_{X}\times \Phi)_{\#}} \ar[dashrightarrow, end anchor=south west]{rru}{\varphi} \&\&\pi_{1}(X) \times B_{n}(X), \end{tikzcd} \end{equation} where $\iota_{1,n}\colon\thinspace D_{1,n}(X) \ensuremath{\longrightarrow} X \times D_{n}(X)$ is the inclusion map.
\item\label{it:defchinegb} If the $n$-valued map $\phi$ is split, it can be deformed to a fixed point free $n$-valued map if and only if there is a homomorphism $\widehat{\varphi}\colon\thinspace \pi_1(X) \ensuremath{\longrightarrow} P_{n+1}(X)$ that makes the following diagram commute: \begin{equation*} \begin{tikzcd}[ampersand replacement=\&] \&\& P_{n+1}(X) \ar{d}{(\widehat{\iota}_{n+1})_{\#}}\\ \pi_{1}(X) \ar[swap]{rr}{(\ensuremath{\operatorname{\text{Id}}}_{X}\times \widehat{\Phi})_{\#}} \ar[dashrightarrow, end anchor=south west]{rru}{\widehat{\varphi}} \&\& \pi_{1}(X) \times P_{n}(X). \end{tikzcd} \end{equation*} where $\widehat{\iota}_{n+1}\colon\thinspace F_{n+1}(X) \ensuremath{\longrightarrow} X \times F_{n}(X)$ is the inclusion map. \end{enumerate}\end{thm}
If $\phi\colon\thinspace X \multimap X$ is a split $n$-valued map given by $\phi=\{f_1, \cdots ,f_n\}$ that can be deformed to a fixed point free $n$-valued map, then certainly each of the single-valued maps $f_i$ can be deformed to a fixed point free map. The question of whether the converse of this statement holds for surfaces is open. We do not know the answer for any compact surface without boundary different from $\St$ or $\ensuremath{{\mathbb R}P^2}$, but it is likely that the converse does not hold. More generally, one would like to know if the homotopy class of $\phi$ contains a representative for which the number of fixed points is exactly the Nielsen number. Very little is known about this question, even for the $2$-torus. Recall that the Nielsen number of an $n$-valued map $\phi\colon\thinspace X\multimap X$, denoted $N(\phi)$, was defined by Schirmer~\cite{Sch1}, and generalises the usual Nielsen number in the single-valued case. She showed that $N(\phi)$ is a lower bound for the number of fixed points among all $n$-valued maps homotopic to $\phi$.
Within the framework of \reth{defchineg}, it is natural to study first the case of $2$-valued maps of the $2$-torus $\ensuremath{\mathbb{T}^{2}}$, which is the focus of \resec{toro}. In what follows, $\mu$ and $\lambda$ will denote the meridian and the longitude respectively of $\ensuremath{\mathbb{T}^{2}}$. Let $(e_1, e_2)$ be a basis of $\pi_1(\ensuremath{\mathbb{T}^{2}})$ such that $e_1=[\mu]$ and $e_2=[\lambda]$. For self-maps of $\ensuremath{\mathbb{T}^{2}}$, we will not be overly concerned with the choice of basepoints since the fundamental groups of $\ensuremath{\mathbb{T}^{2}}$ with respect to two different basepoints may be canonically identified. In \resec{toro2}, we will study the groups $P_{2}(\ensuremath{\mathbb{T}^{2}})$, $B_{2}(\ensuremath{\mathbb{T}^{2}})$ and $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$, and in \reco{compactpres}, we will see that $P_{2}(\ensuremath{\mathbb{T}^{2}})$ is isomorphic to the direct product of a free group $\mathbb{F}_2(u,v)$ of rank $2$ and $\ensuremath{\mathbb Z}^2$. In what follows, the elements of $P_2(\ensuremath{\mathbb{T}^{2}})$ will be written with respect to the decomposition $\mathbb{F}_2(u,v) \times \ensuremath{\mathbb Z}^2$, and $\operatorname{\text{Ab}}\colon\thinspace \mathbb{F}_2(u,v) \ensuremath{\longrightarrow} \ensuremath{\mathbb Z}^{2}$ will denote Abelianisation. \reth{helgath01}, which is a result of~\cite{Sch1} for the Nielsen number of split $n$-valued maps, will be used in part of the proof of the following proposition.
\begin{prop}\label{prop:exisfpf} Let $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ be a split $2$-valued map of the torus $\ensuremath{\mathbb{T}^{2}}$, and let $\widehat{\Phi}=(f_1,f_2) \colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$ be a lift of $\phi$ such that $\widehat{\Phi}_{\#}(e_{1})=(w^r,(a,b))$ and $\widehat{\Phi}_{\#}(e_{2})= (w^s, (c,d)))$, where $(r,s)\in \ensuremath{\mathbb Z}^{2}\setminus \brak{(0,0)}$, $a,b,c,d\in \ensuremath{\mathbb Z}$ and $w\in \mathbb{F}_2(u,v)$. Then the Nielsen number of $\phi$ is given by: \begin{equation*} N(\phi)=\left\lvert\det\begin{pmatrix}
a-1 & c \\
b & d-1 \end{pmatrix}\right\rvert + \left\lvert\det\begin{pmatrix}
rm+a-1 & sm+c \\
rn+b & sn+d-1 \end{pmatrix}\right\rvert, \end{equation*} where $\operatorname{\text{Ab}}(w)=(m,n)\in \ensuremath{\mathbb Z}^{2}$. If the map $\phi$ can be deformed to a fixed point free $2$-valued map, then both of the maps $f_1$ and $f_2$ can be deformed to fixed point free maps. Furthermore, $f_1$ and $f_2$ can be deformed to fixed point free maps if and only if either: \begin{enumerate}[(a)] \item\label{it:exisfpfa} the pairs of integers $(a-1, b),(c,d-1)$ and $(m,n)$ belong to a cyclic subgroup of $\ensuremath{\mathbb Z}^2$, or \item\label{it:exisfpfb} $s(a-1, b)=r(c,d-1)$. \end{enumerate} \end{prop}
Within the framework of \repr{exisfpf}, given a split $2$-valued map $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ for which $N(\phi)=0$, we would like to know whether $\phi$ can be deformed to a fixed point free $2$-valued map. If $N(\phi)=0$, then by this proposition, one of the conditions~(\ref{it:exisfpfa}) or~(\ref{it:exisfpfb}) must be satisfied. The following result shows that condition~(\ref{it:exisfpfb}) is also sufficient.
\begin{thm}\label{th:necrootfree3} Let $\widehat{\Phi}\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2({\ensuremath{\mathbb{T}^{2}}})$ be a lift of a split $2$-valued map $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ that satisfies $\widehat{\Phi}_{\#}(e_{1})=(w^{r},(a,b))$ and $\widehat{\Phi}_{\#}(e_{2})= (w^{s}, (c,d))$, where $w\in \mathbb{F}_2(u,v)$, $a,b,c,d\in \ensuremath{\mathbb Z}$ and $(r,s)\in \ensuremath{\mathbb Z}^{2}\setminus \brak{(0,0)}$ satisfy $s(a-1,b)=r(c,d-1)$. Then $\phi$ may be deformed to a fixed point free $2$-valued map.
\end{thm}
With respect to condition~(\ref{it:exisfpfa}), we obtain a partial converse for certain values of $a,b,c,d,m$ and $n$.
\begin{thm}\label{th:construct2val} Suppose that $(a-1, b),(c,d-1)$ and $(m,n)$ belong to a cyclic subgroup of $\ensuremath{\mathbb Z}^2$ generated by an element of the form $(0,q), (1,q), (p,0)$ or $(p,1)$, where $p,q\in \ensuremath{\mathbb Z}$, and let $r,s\in \ensuremath{\mathbb Z}$. Then there exist $w\in \mathbb{F}_2(u,v)$, a split fixed point free $2$-valued map $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ and a lift $\widehat{\Phi} \colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2(\ensuremath{\mathbb{T}^{2}})$ of $\phi$ such that such that $\operatorname{\text{Ab}}(w)=(m,n)$, $\widehat{\Phi}_{\#}(e_1)=((w^{r},(a,b))$ and $\widehat{\Phi}_{\#}(e_2)= (w^{s}, (c,d))$. \end{thm}
\repr{exisfpf} and Theorems~\ref{th:necrootfree3} and~\ref{th:construct2val} will be proved in \resec{toro}. Besides the introduction and an appendix, this paper is divided into 4 sections. In \resec{pres}, we give some basic definitions, we establish the connection between multimaps and maps whose target is a configuration space, and we show that simply-connected spaces have the fixed point property for $n$-valued maps if they have the usual fixed point property. In \resec{rp2k}, we show that even-dimensional real projective spaces have the fixed point property for $n$-valued maps. In \resec{fixfree}, we provide general criteria of a homotopic and algebraic nature, to decide whether an $n$-valued map can be deformed or not to a fixed point free $n$-valued map, and we give the corresponding statements for the case of roots. In \resec{toro}, we study the fixed point theory of $2$-valued maps of the $2$-torus. In \resec{toro2}, we give presentations of certain braid groups of $\ensuremath{\mathbb{T}^{2}}$, in \resec{descript}, we describe the set of homotopy classes of split $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$, and in \resec{fptsplit2}, we study the fixed point theory of split $2$-valued maps. In the Appendix, written with R.~F.~Brown, in \reth{metriccont}, we show that for the class of metric spaces that includes those considered in this paper, $n$-valued maps can be regarded as single-valued maps whose target is the associated unordered configuration space.
\section{Generalities and the $n$-valued fixed point property}\label{sec:pres}
In \resec{relnsnvm}, we begin by describing the relations between $n$-valued maps and $n$-unordered maps. We will assume throughout that $X$ and $Y$ are metric spaces, so that we can apply \reth{metriccont}. Making use of unordered configuration space, in \relem{split} and \reco{inject}, we prove some properties about the fixed points of $n$-valued maps. In \resec{disc}, we give an algebraic condition that enables us to decide whether an $n$-valued map is split. We also study the case where $X$ is simply connected (the $k$-dimensional disc for example, which has the usual fixed point property) and we prove \reth{sccfpp}, and in \resec{sph}, we analyse the case of the $2k$-dimensional sphere (which does not have the usual fixed point property), and we prove \repr{S2fp}.
\subsection{Relations between $n$-valued maps, $n$-(un)ordered maps and their fixed point sets}\label{sec:relnsnvm}
A proof of the following result may be found in the Appendix.
\begin{thm}\label{th:metriccont} Let $X$ and $Y$ be metric spaces, and let $n\in \ensuremath{\mathbb N}$. An $n$-valued function $\phi\colon\thinspace X \multimap Y$ is continuous if and only if the corresponding function $\Phi\colon\thinspace X \ensuremath{\longrightarrow} D_n(Y)$ is continuous. \end{thm}
It would be beneficial for the statement of \reth{metriccont} to hold under weaker hypotheses on $X$ and $Y$. See~\cite{Brr7} for some recent results in this direction.
\begin{defn} If $\phi\colon\thinspace X \multimap Y$ is an $n$-valued map and $\Phi\colon\thinspace X\ensuremath{\longrightarrow} D_{n}(Y)$ is the associated $n$-unordered map, an $n$-ordered map $\widehat{\Phi}\colon\thinspace X \ensuremath{\longrightarrow} F_n(Y)$ is said to be a \emph{lift} of $\phi$ if the composition $\pi\circ \widehat{\Phi}\colon\thinspace X \ensuremath{\longrightarrow} D_{n}(Y)$ of $\widehat{\Phi}$ with the covering map $\pi\colon\thinspace F_n(Y) \ensuremath{\longrightarrow} D_n(Y)$ is equal to $\Phi$. \end{defn}
If $\phi=\brak{f_1,\ldots,f_n}\colon\thinspace X \multimap Y$ is a split $n$-valued map, then it admits a lift $\widehat{\Phi}=(f_1,\ldots,f_n)\colon\thinspace X \ensuremath{\longrightarrow} F_n(Y)$. For any such lift, $\operatorname{\text{Fix}}(\widehat{\Phi})=\operatorname{\text{Fix}}(\phi)$, and the map $\widehat{\Phi}$ determines an ordered set of $n$ maps $(f_1=p_1\circ \widehat{\Phi},\ldots, f_n=p_n\circ \widehat{\Phi})$ from $X$ to $Y$ for which $f_i(x)\ne f_j(x)$ for all $x\in X$ and all $1\leq i<j\leq n$. Conversely, any ordered set of $n$ maps $(f_1,\ldots,f_n)$ from $X$ to $Y$ for which $f_i(x)\ne f_j(x)$ for all $x\in X$ and all $1\leq i<j\leq n$ determines an $n$-ordered map $\Psi\colon\thinspace X \ensuremath{\longrightarrow} F_n(Y)$ defined by $\Psi(x)=(f_1(x),\ldots,f_n(x))$ and a split $n$-valued map $\phi=\brak{f_1,\ldots,f_n}\colon\thinspace X \multimap Y$ of which $\Psi$ is a lift.
So the existence of such a split $n$-valued map $\phi$ is equivalent to that of an $n$-ordered map $\Psi\colon\thinspace X \ensuremath{\longrightarrow} F_n(Y)$, where $\Psi=(f_1,\ldots,f_n)$. This being the case, the composition $\pi\circ \Psi\colon\thinspace X \ensuremath{\longrightarrow} D_n(Y)$ is the map $\Phi\colon\thinspace X \ensuremath{\longrightarrow} D_n(Y)$ that corresponds (in the sense described in \resec{intro}) to the $n$-valued map $\phi$. Consequently, an $n$-valued map $\phi\colon\thinspace X \multimap Y$ admits a lift if and only if it is split. As we shall now see, \reth{metriccont} will be of help in the description of the relations between (split) $n$-valued maps and $n$-(un)ordered maps of metric spaces. As we have seen, to each $n$-valued map (resp.\ split $n$-valued map), we may associate an $n$-unordered map (resp.\ a lift), and \emph{vice-versa}. Note that the symmetric group $S_n$ not only acts (freely) on $F_n(Y)$ by permuting coordinates, but it also acts on the set of ordered $n$-tuples of maps between $X$ and $Y$. Further, the restriction of the latter action to the subset $F_{n}(Y)^{X}$ of $n$-ordered maps, \emph{i.e.}\ maps of the form $\Psi\colon\thinspace X \ensuremath{\longrightarrow} F_{n}(Y)$, where $\Psi(x)=(f_1(x),\ldots,f_n(x))$ for all $x\in X$ for which $f_i(x)\ne f_j(x)$ for all $x\in X$ and $1\leq i< j\leq n$, is also free. In what follows, $[X,Y]$ (resp.\ $[X,Y]_{0}$) will denote the set of homotopy classes (resp.\ based homotopy classes) of maps between $X$ and $Y$.
\begin{lem}\label{lem:split}\mbox{} Let $X$ and $Y$ be metric spaces, and let $n\in \ensuremath{\mathbb N}$. \begin{enumerate}[(a)] \item\label{it:splitII} The set $\splitmap{X}{Y}{n}$ of split $n$-valued maps from $X$ to $Y$ is in one-to-one correspondence with the orbits of the set of maps $F_{n}(Y)^{X}$ from $X$ to $F_n(Y)$ modulo the free action defined above of $S_n$ on $F_{n}(Y)^{X}$. \item\label{it:splitIII} If two $n$-valued maps from $X$ to $Y$ are homotopic and one is split, then the other is also split. Further, the set $\splitmap{X}{Y}{n}/\!\sim$ of homotopy classes of split $n$-valued maps from $X$ to $Y$ is in one-to-one correspondence with the orbits of the set $[X,F_{n}(Y)]$ of homotopy classes of maps from $X$ to $F_n(Y)$ under the action of $S_n$ induced by that of $S_n$ on $F_{n}(Y)^{X}$. \item\label{it:splitIV} Suppose that $X=Y$. If an $n$-valued map $\phi\colon\thinspace X \multimap X$ is split and deformable to a fixed point free map, then a lift $\widehat{\Phi}\colon\thinspace X \ensuremath{\longrightarrow} F_{n}(X)$ of $\phi$ may be written as ${\Psi}=(f_1,\ldots,f_n)$, where for all $i=1,\ldots,n$, the map $f_i\colon\thinspace X\ensuremath{\longrightarrow} X$ is a map that is deformable to a fixed point free map. \end{enumerate} \end{lem}
\begin{proof}\mbox{} \begin{enumerate}[(a)] \item Let $\phi\colon\thinspace X \multimap Y$ be a split $n$-valued map. From the definition, there exists an $n$-ordered map $\widehat{\Phi}\colon\thinspace X\ensuremath{\longrightarrow} F_n(Y)$ such that $\Phi=\pi\circ \widehat{\Phi}$, up to the identification given by \reth{metriccont}. If $\widehat{\Phi}=(f_1,\ldots,f_n)$, the other lifts of $\phi$ are obtained via the action of the group of deck transformations of the covering space, this group being $S_n$ in our case, and so are of the form $(f_{\sigma(1)},\ldots,f_{\sigma(n)})$, where $\sigma\in S_{n}$. This gives rise to the stated one-to-one correspondence between $\splitmap{X}{Y}{n}$ and the orbit space $F_{n}(Y)^{X}/S_{n}$.
\item By naturality, the map $\pi\colon\thinspace F_n(Y) \ensuremath{\longrightarrow} D_n(Y)$ induces a map $\widehat{\pi}\colon\thinspace [X, F_n(Y)]\ensuremath{\longrightarrow} [X, D_n(Y)]$ defined by $\widehat{\pi}([\Psi])=[\pi\circ \Psi]$ for any $n$-ordered map $\Psi\colon\thinspace X\ensuremath{\longrightarrow} F_{n}(Y)$. Given two homotopic $n$-valued maps between $X$ and $Y$, which we regard as maps from $X$ to $D_n(Y)$ using \reth{metriccont}, if the first has a lift to $F_n(Y)$, then the lifting property of a covering implies that the second also admits a lift to $F_n(Y)$, so if the first map is split then so is the second map. To prove the second part of the statement, first note that there is a surjective map $f\colon\thinspace F_{n}(Y)^{X} \ensuremath{\longrightarrow} \splitmap{X}{Y}{n}$ given by $f(g)=\pi\circ g$, where we identify $\splitmap{X}{Y}{n}$ with the set of maps $D_{n}(Y)^{X}$ from $X$ to $D_{n}(Y)$, that induces a surjective map $\overline{f}\colon\thinspace [X,F_{n}(Y)] \ensuremath{\longrightarrow} \splitmap{X}{Y}{n}/\!\sim$ on the corresponding sets of homotopy classes. Further, if ${\Psi}_1, {\Psi}_2\in F_{n}(Y)^{X}$ are two $n$-ordered maps that are homotopic via a homotopy $H$, and if $\alpha\in S_n$, then the maps $\alpha \circ\Psi_1,\alpha \circ \Psi_2 \in F_{n}(Y)^{X}$ are also homotopic via the homotopy $\alpha \circ H$, and so we obtain a quotient map $q\colon\thinspace [X,F_{n}(Y)] \ensuremath{\longrightarrow} [X,F_{n}(Y)]/S_{n}$. We claim that $\overline{f}$ factors through $q$ via the map $\overline{\overline{f}}\colon\thinspace [X,F_{n}(Y)]/S_{n}\ensuremath{\longrightarrow} \splitmap{X}{Y}{n}/\!\sim$ defined by $\overline{\overline{f}}([g])=[f(g)]$. To see this, let $g, h\in F_{n}(Y)^{X}$ be such that $q([g])=q([h])$. Then there exists $\alpha\in S_{n}$ such that $\alpha([g])=[h]$. Then $\overline{f}(\alpha[g])=[\alpha f(h)]= \overline{f}([h])=[f(h)]$. But from the definition of $\splitmap{X}{Y}{n}$, $[\alpha f(g)]=[f(g)]$, and so $[f(g)]=[f(h)]$, which proves the claim. By construction, the map $\overline{\overline{f}}$ is surjective. It remains to show that it is injective. Let $g,h\in F_{n}(Y)^{X}$ be such that $\overline{\overline{f}}([g])=\overline{\overline{f}}([h])$. Then $[f(g)]=[f(h)]$, and thus $f(g)$ and $f(h)$ are homotopic via a homotopy $H$ in $\splitmap{X}{Y}{n}$, where $H(0,f(g))=f(g)$ and $H(1,f(g))=f(h)$. Then $H$ lifts to a homotopy $\widetilde{H}$ such that $\widetilde{H}(0,g)=g$, and $\widetilde{H}(1,g)$ is a lift of $f(h)$. But $h$ is also a lift of $f(h)$, so there exists $\alpha\in S_{n}$ such that $f(h)=\alpha h$. Further, $g$ is homotopic to $f(h)$, so is homotopic to $\alpha(h)$, and hence $q([g])=q([\alpha(h)])= q([h])$ from the definition of $q$, which proves the injectivity of $\overline{\overline{f}}$.
\item Since $\phi$ is split, we may choose a lift $\widehat{\Phi}=(f_1,\ldots,f_n)\colon\thinspace X \ensuremath{\longrightarrow} F_{n}(X)$ of $\phi$. By hypothesis, there is a homotopy $H\colon\thinspace X\times I \ensuremath{\longrightarrow} D_n(X)$ such that $H(\cdot ,0)=\Phi$, and $H(\cdot ,1)$ is fixed point free. Since the initial part of the homotopy $H$ admits a lift, there exists a lift $\widehat{H}\colon\thinspace X\times I \ensuremath{\longrightarrow} F_n(X)$ of $H$ such that $\widehat{H}(\cdot, 0)=\widehat{\Phi}$, and $\widehat{H}(\cdot, 1)$ is fixed point free. So $\widehat{H}(\cdot, 1)$ is of the form $(f_1,\ldots,f_n)$, where $f_i$ is fixed point free for $1\leq i\leq n$, and the conclusion follows.\qedhere \end{enumerate} \end{proof}
\begin{rems}\mbox{} \begin{enumerate}[(a)] \item The action of $S_n$ on the set of homotopy classes $[X, F_n(Y)]$ is not necessarily free (see \repr{classif}(\ref{it:classifb})).
\item The question of whether the converse of \relem{split}(\ref{it:splitIV}) is valid for surfaces is open, see the introduction. \end{enumerate} \end{rems}
The following consequence of \relem{split}(\ref{it:splitIV}) will be useful in what follows, and implies that if a split $n$-valued map can be deformed to a fixed point free $n$-valued map (through $n$-valued maps), then the deformation
is through split $n$-valued maps.
\begin{cor}\label{cor:inject} Let $X$ be a metric space, and let $n\in \ensuremath{\mathbb N}$. A split $n$-valued map $\phi\colon\thinspace X\multimap X$ may be deformed within $\splitmap{X}{X}{n}$ to a fixed point free $n$-valued map if and only if any lift $\widehat{\Phi}=(f_1,\ldots,f_n)\colon\thinspace X \ensuremath{\longrightarrow} F_n(X)$ of $\phi$ may be deformed within $F_{n}(X)$ to a fixed point free map $\widehat{\Phi}'=(f_1',\ldots,f_n')\colon\thinspace X \ensuremath{\longrightarrow} F_n(X)$. In particular, for all $1\leq i\leq n$, there exists a homotopy $H_i\colon\thinspace X\times I\ensuremath{\longrightarrow} X$ between $f_i$ and $f_i'$, where $f_i'$ is a fixed point free map, and $H_j(x, t)\ne H_k(x, t)$ for all $1\leq j<k\leq n$, $x\in X$ and $t\in [0,1]$. \end{cor}
\begin{proof} The `if' part of the statement may be obtained by considering the composition of the deformation between ${\Psi}$ and ${\Psi}'$ by the projection $\pi$ and by applying \reth{metriccont}. The `only if' part follows in a manner similar to that of the proof of the first part of \relem{split}(\ref{it:splitIII}). \end{proof}
\subsection{The fixed point property of simply connected spaces and the $k$-disc $\dt[k]$ for $n$-valued maps} \label{sec:disc}
In this section, we analyse the case where $X$ is a simply-connected metric space that possesses the fixed point property, such as the closed $k$-dimensional disc $\dt[k]$. In \relem{split1}, we begin by proving a variant of the so-called `Splitting Lemma' that is more general than the versions that appear in the literature, such as that of Schirmer given in~\cite[Section~2, Lemma~1]{Sch0} for example. The hypotheses are expressed in terms of the homomorphism on the level of the fundamental group of the target $Y$, rather than that of the domain $X$, and the criterion is an algebraic condition, in terms of the fundamental group, for an $n$-valued map from $X$ to $Y$ to be split. This allows us to prove \reth{sccfpp}, which says that a simply-connected metric space that has the fixed point property also possesses the fixed point property for $n$-valued maps for all $n\geq 1$. In particular, $\dt[k]$ satisfies this property for all $k\geq 1$. The $2$-disc will be the only surface with boundary that will be considered in this paper. The cases of other surfaces with boundary, such as the annulus and the M\"obius band, will be studied elsewhere.
\begin{lem}\label{lem:split1} Let $n\geq 1$, let $\phi\colon\thinspace X \multimap Y$ be an $n$-valued map between metric spaces, where $X$ is connected and locally arcwise-connected, and let $\Phi\colon\thinspace X \ensuremath{\longrightarrow} D_n(Y)$ be the associated $n$-unordered map. Then $\phi$ is split if and only if the image of the induced homomorphism $\Phi_{\#}\colon\thinspace \pi_1(X) \ensuremath{\longrightarrow} B_n(Y)$ is contained in the image of the homomorphism $\pi_{\#}\colon\thinspace P_n(Y) \ensuremath{\longrightarrow} B_{n}(Y)$ induced by the covering map $\pi\colon\thinspace F_{n}(Y)\ensuremath{\longrightarrow} D_{n}(Y)$. In particular, if $X$ is simply connected then all $n$-valued maps from $X$ to $Y$ are split.
\end{lem}
\begin{proof} Since $X$ and $Y$ are metric spaces, using \reth{metriccont}, we may consider the $n$-unordered map $\Phi\colon\thinspace X \ensuremath{\longrightarrow} D_n(Y)$ that corresponds to $\phi$. The first part of the statement follows from standard results about the lifting property of a map to a covering space in terms of the fundamental group~\cite[Chapter~5, Section~5, Theorem~5.1]{Mas}. The second part is a consequence of the first part. \end{proof}
As a consequence of \relem{split1}, we are able to prove~\reth{sccfpp}.
\begin{proof}[Proof of \reth{sccfpp}] Let $X$ be a simply-connected metric space that has the fixed point property. By \relem{split1}, any $n$-valued map $\phi\colon\thinspace X \multimap X$ is split. Writing $\phi=\{f_1,\ldots,f_n\}$, each of the maps $f_i\colon\thinspace X \ensuremath{\longrightarrow} X$ is a self-map of $X$ that has at least one fixed point. So $\phi$ has at least $n$ fixed points, and in particular, $X$ has the fixed point property for $n$-valued maps. The last part of the statement then follows. \end{proof}
\subsection{$n$-valued maps of the sphere $\mathbb S^{2k}$}\label{sec:sph}
Let $k\geq 1$. Although $\mathbb S^{2k}$ does not have the fixed point property for self-maps, we shall show in this section that it has the fixed point property for $n$-valued maps for all $n>1$, which is the statement of \repr{S2fp}. We first prove a lemma.
\begin{lem}\label{lem:S2split} Let $n\geq 1$ and $k\geq 2$. Then any $n$-valued map of $\mathbb S^{k}$ is split. \end{lem}
\begin{proof} The result follows from \relem{split1} using the fact that $\mathbb S^k$ is simply connected. \end{proof}
\begin{proof}[Proof of \repr{S2fp}] Let $n\geq 2$, and let $\phi \colon\thinspace \mathbb S^{2k} \multimap \mathbb S^{2k}$ be an $n$-valued map. By \relem{S2split}, $\phi$ is split, so it admits a lift $\widehat{\Phi}\colon\thinspace \mathbb S^{2k} \ensuremath{\longrightarrow} F_{n}(\St[2k])$, where $\widehat{\Phi}=(f_1,f_2,\ldots,f_n)$. Since $f_1(x)\ne f_2(x)$ for all $x\in \St[2k]$, we have $f_2(x)\neq -(-f_1(x))$, it follows that $f_2$ is homotopic to $-f_1$ via a homotopy that for all $x\in \mathbb S^{2k}$, takes $-f_1(x)$ to $f_2(x)$ along the unique geodesic that joins them. Thus the degree of one of the maps $f_1$ and $f_2$ is different from $-1$, and so has a fixed point, which implies that $\phi$ has a fixed point. \end{proof}
\begin{rem} If $n>2$ and $k=1$ then the result of \repr{S2fp} is clearly true since by~\cite[pp.~43--44]{GG5}, the set $[\St, F_{n}(\St)]$ of homotopy classes of maps between $\St$ and $F_{n}(\St)$ contains only one class, which is that of the constant map. So any representative of this class is of the form $\phi=(f_1,\ldots,f_n)$, where all of the maps $f_i\colon\thinspace \St\ensuremath{\longrightarrow} \St$ are homotopic to the constant map. Such a map always has a fixed point, and hence $\phi$ has at least $n$ fixed points. \end{rem}
\section{$n$-valued maps of the projective space $\ensuremath{\mathbb R} P^{2k}$}\label{sec:rp2k}
In this section, we will show that the projective space $\ensuremath{\mathbb R} P^{2k}$ also has the fixed point property for $n$-valued maps, which is the statement of \reth{rp2Kfpp}. Since $\ensuremath{\mathbb R} P^{2k}$ is not simply connected, we will require more elaborate arguments than those used in Sections~\ref{sec:disc} and~\ref{sec:sph}. We separate the discussion into two cases, $k=1$, and $k>1$.
\subsection{$n$-valued maps of \ensuremath{{\mathbb R}P^2}}\label{sec:rp2ka}
The following result is the analogue of \relem{S2split} for $\ensuremath{{\mathbb R}P^2}$.
\begin{lem}\label{lem:rp2split}
Let $n\geq 1$. Then any $n$-valued map of the projective plane is split. \end{lem}
\begin{proof} Let $\phi\colon\thinspace \ensuremath{{\mathbb R}P^2} \multimap \ensuremath{{\mathbb R}P^2}$ be an $n$-valued map, let $\Phi\colon\thinspace \ensuremath{{\mathbb R}P^2} \ensuremath{\longrightarrow} D_n(\ensuremath{{\mathbb R}P^2})$ be the associated $n$-unordered map, and let $\Phi_{\#}\colon\thinspace \pi_{1}(\ensuremath{{\mathbb R}P^2}) \ensuremath{\longrightarrow} B_{n}(\ensuremath{{\mathbb R}P^2})$ be the homomorphism induced on the level of fundamental groups. Since $\pi_{1}(\ensuremath{{\mathbb R}P^2})$ is isomorphic to the cyclic group of order $2$, it follows that $\im{\Phi_{\#}}$ is contained in the subgroup $\ang{\ft}$ of $B_{n}(\ensuremath{{\mathbb R}P^2})$ generated by the full twist braid $\ft$ of $B_n(\ensuremath{{\mathbb R}P^2})$ because $\ft$ is the unique element of $B_n(\ensuremath{{\mathbb R}P^2})$ of order $2$ \cite[Proposition~23]{GG3}. But $\ft$ is a pure braid, so $\im{\Phi_{\#}}\subset P_{n}(\ensuremath{{\mathbb R}P^2})$. Thus $\Phi$ factors through $\pi$, and hence $\phi$ is split by \relem{split1} as required. \end{proof}
\begin{prop}\label{prop:RP2fp} Let $n\geq 1$. Then any $n$-valued map of $\ensuremath{{\mathbb R}P^2}$ has at least $n$ fixed points, in particular $\ensuremath{{\mathbb R}P^2}$ has the fixed point property for $n$-valued maps. \end{prop}
\begin{proof} Let $\phi\colon\thinspace \ensuremath{{\mathbb R}P^2} \multimap \ensuremath{{\mathbb R}P^2}$ be an $n$-valued map of $\ensuremath{{\mathbb R}P^2}$. Then $\phi$ is split by \relem{rp2split}, and so $\phi=\brak{f_1,\ldots,f_n}$, where $f_{1},\ldots, f_{n}\colon\thinspace \ensuremath{{\mathbb R}P^2} \ensuremath{\longrightarrow} \ensuremath{{\mathbb R}P^2}$ are pairwise coincidence-free self-maps of $\ensuremath{{\mathbb R}P^2}$. But $\ensuremath{{\mathbb R}P^2}$ has the fixed point property, and so for $i=1,\ldots,n$, $f_{i}$ has a fixed point. Hence $\phi$ has at least $n$ fixed points. \end{proof}
\subsection{$n$-valued maps of $\ensuremath{\mathbb R} P^{2k}$, $k>1$}\label{sec:proje}
The aim of this section is to prove that $\ensuremath{\mathbb R} P^{2k}$ has the fixed point property for $n$-valued maps for all $n\geq 1$ and $k>1$. Indeed, we will show that every such $n$-valued map has at least $n$ fixed points.
Given an $n$-valued map $\phi\colon\thinspace X \multimap X$ of a topological space $X$, we consider the corresponding map $\Phi\colon\thinspace X \ensuremath{\longrightarrow} D_n(X)$, and the induced homomorphism $\Phi_{\#}\colon\thinspace \pi_1(X) \ensuremath{\longrightarrow} \pi_1(D_n(X))$ on the level of fundamental groups, where $\pi_1(D_n(X))=B_n(X)$. By the short exact sequence~\reqref{sesbraid}, $P_n(X)$ is a normal subgroup of $B_n(X)$ of finite index $n!$, so the subgroup $H=\Phi_{\#}^{-1}(P_n(X))$ is a normal subgroup of $\pi_1(X)$ of finite index. Further, if $L=\pi_1(X)/H$, the composition $\pi_1(X) \stackrel{\Phi_{\#}}{\ensuremath{\longrightarrow}} B_n(X) \stackrel{\tau}{\ensuremath{\longrightarrow}} S_n$ is a homomorphism that induces a homomorphism between $L$ and $S_n$.
\begin{prop}\label{prop:nielsen} Let $n\in \ensuremath{\mathbb N}$. Suppose that $X$ is a connected, locally arcwise-connected metric space. With the above notation, there exists a covering $q\colon\thinspace \widehat{X} \ensuremath{\longrightarrow} X$ of $X$ that corresponds to the subgroup $H$, and the $n$-valued map $\phi_1=\phi \circ q\colon\thinspace \widehat{X}\multimap X$ admits exactly $n!$ lifts, which are $n$-ordered maps from $\widehat{X}$ to $F_n(X)$. If one such lift $\widehat{\Phi}_{1}\colon\thinspace \widehat{X}\ensuremath{\longrightarrow} F_n(X)$ is given by $\widehat{\Phi}_{1}=(f_1,\ldots, f_n)$, where for $i=1,\ldots,n$, $f_i$ is a map from $\widehat{X}$ to $X$, then the other lifts are of the form $(f_{\tau(1)},\ldots,f_{\tau(n)})$, where $\tau\in S_n$. \end{prop}
\begin{proof} The first part is a consequence of~\cite[Theorem~5.1, Chapter~V, Section~5]{Mas}, using the observation that $S_{n}$ is the deck transformation group corresponding to the covering $\pi\colon\thinspace F_n(X) \ensuremath{\longrightarrow} D_n(X)$. The second part follows from the fact that $S_n$ acts freely on the covering space $F_n(X)$ by permuting coordinates. \end{proof}
The fixed points of the $n$-valued map $\phi\colon\thinspace X\multimap X$ may be described in terms of the coincidences of the covering map $q\colon\thinspace \widehat{X} \ensuremath{\longrightarrow} X$ with the maps $f_{1},\ldots, f_{n}$ given in the statement of \repr{nielsen}.
\begin{prop}\label{prop:coinfix} Let $n\in \ensuremath{\mathbb N}$, let
$X$ be a connected, locally arcwise-connected, metric space, let $\phi\colon\thinspace X \multimap X$ be an $n$-valued map, and let ${\widehat \Phi_1}=(f_1,\ldots,f_n)\colon\thinspace \widehat{X} \ensuremath{\longrightarrow} F_n(X)$ be an $n$-ordered map that is a lift of $\phi_1=\phi\circ q\colon\thinspace \widehat{X} \multimap X$ as in \repr{nielsen}. Then the map $q$ restricts to a surjection $q\colon\thinspace \bigcup_{i=1}^{n} \operatorname{\text{Coin}}(q, f_i) \ensuremath{\longrightarrow} \operatorname{\text{Fix}}(\phi)$.
Furthermore, the pre-image of a point $x\in \operatorname{\text{Fix}}(\phi)$ by this map is precisely $q^{-1}(x)$, namely the fibre over $x\in X$ of the covering map $q\colon\thinspace \widehat{X} \ensuremath{\longrightarrow} X$. \end{prop}
\begin{proof} Let $\widehat{x}\in \operatorname{\text{Coin}}(q, f_i)$ for some $1\leq i\leq n$, and let $x=q(\widehat{x})$. Then $f_i(\widehat{x})=q(\widehat{x})$, and since $\Phi(x)= \pi\circ \widehat{\Phi}_1(\widehat{x})=\brak{f_1(\widehat{x}),\ldots,f_n(\widehat{x})}$, it follows that $x\in \phi(x)$, \emph{i.e.}\ $x\in \operatorname{\text{Fix}}(\phi)$, so the map is well defined. To prove surjectivity and the second part of the statement, it suffices to show that if $x \in \operatorname{\text{Fix}}(\phi)$, then any element $\widehat{x}$ of $q^{-1}(x)$ belongs to $\bigcup_{i=1}^{n} \operatorname{\text{Coin}}(q, f_i) $. So let $x\in \phi(x)$, and let $\widehat{x}\in \widehat{X}$ be such that $q(\widehat{x})=x$. By commutativity of the following diagram: \begin{equation*} \begin{tikzcd}[ampersand replacement=\&] \&\& F_{n}(X) \ar{d}{\pi}\\ \widehat{X} \ar[swap]{r}{q} \ar[dashrightarrow, end anchor=south west]{rru}{\widehat{\Phi}_1} \& X \ar[swap]{r}{\Phi} \& D_{n}(X), \end{tikzcd} \end{equation*} and the fact that $x\in \operatorname{\text{Fix}}(\phi)$, it follows that $x$ is one of the coordinates, the $j\up{th}$ coordinate say, of $\widehat{\Phi}_1(\widehat{x})$. This implies that $\widehat{x}\in \operatorname{\text{Coin}}(q, f_j)$, which completes the proof of the proposition. \end{proof}
\begin{prop}\label{prop:nullhomo} Let $n,k>1$. If $\phi\colon\thinspace \St[2k] \multimap \ensuremath{\mathbb R} P^{2k}$ is an $n$-valued map, then $\phi$ is split, and for $i=1,\ldots,n$, there exist maps $f_i\colon\thinspace \St[2k] \ensuremath{\longrightarrow} \ensuremath{\mathbb R} P^{2k}$ for which $\phi=\{f_1,\ldots,f_n\}$. Further, $f_i$ is null homotopic for all $i\in \brak{1,\ldots,n}$. \end{prop}
\begin{proof} The first part follows from \relem{split1}. It remains to prove the second part, \emph{i.e.}\ that each $f_i$ is null homotopic. Since $\St[2k]$ is simply connected, the set $[\mathbb{S}^{2k}, \mathbb{S}^{2k}]=[\mathbb{S}^{2k}, \mathbb{S}^{2k}]_{0}=\pi_{2k}(\mathbb{S}^{2k})$, where $[ \cdot , \cdot ]_{0}$ denotes basepoint-preserving homotopy classes of maps. Let $x_0\in \mathbb{S}^{2k}$ be a basepoint, let $p\colon\thinspace \St[2k] \ensuremath{\longrightarrow} \ensuremath{\mathbb R} P^{2k}$ be the two-fold covering, and let $\overline{x}_0=p(x_{0})\in \ensuremath{\mathbb R} P^{2k}$ be the basepoint of $\ensuremath{\mathbb R} P^{2k}$. Consider the natural map $[(\mathbb{S}^{2k}, x_0), (\mathbb{S}^{2k}, x_0)] \ensuremath{\longrightarrow} [(\mathbb{S}^{2k}, x_0), (\ensuremath{\mathbb R} P^{2k}, \overline{x}_0)]$ from the set of based homotopy classes of self-maps of $\St[2k]$ to the based homotopy classes of maps from $\St[2k]$ to $\ensuremath{\mathbb R} P^{2k}$, that to a homotopy class of a basepoint-preserving self-map of $\St[2k]$ associates the homotopy class of the composition of this self-map with $p$. This correspondence is an isomorphism. The covering map $p$ has topological degree $2$~\cite{Ep}, so the degree of a map $f\colon\thinspace \St[2k] \ensuremath{\longrightarrow} \ensuremath{\mathbb R} P^{2k}$ is an even integer (we use the system of local coefficients given by the orientation of $\ensuremath{\mathbb R} P^{2k}$~\cite{Ol}). Since $H^{l}( \ensuremath{\mathbb R} P^{2k} , \widetilde{\mathbb{Q}})=0$ for $\widetilde{\mathbb{Q}}$ twisted by the orientation and $l\ne 2k$, if $i\neq j$, it follows that the Lefschetz coincidence number $L(f_{i},f_{j})$ is equal to $\deg(f_i)$. But $f_i$ and $f_j$ are coincidence free, so their Lefschetz coincidence number must be zero, which implies that $\deg(f_i)=0$~\cite{GJ}. Since $n>1$, we conclude that $\deg(f_i)=0$ for all $1\leq i\leq n$, and the result follows. \end{proof}
We are now able to prove the main result of this section, that $\ensuremath{\mathbb R} P^{2k}$ has the fixed point property for $n$-valued maps for all $k,n\geq 1$.
\begin{proof}[Proof of \reth{rp2Kfpp}] The case $n=1$ is classical, so assume that $n\geq 2$. We use the notation introduced at the beginning of this section, taking $X=\ensuremath{\mathbb R} P^{2k}$. Since $\pi_1(\ensuremath{\mathbb R} P^{2k})\cong\ensuremath{\mathbb Z}_2$, $H$ is either $\pi_1(\ensuremath{\mathbb R} P^{2k})$ or the trivial group. In the former case, the $n$-valued map $\phi\colon\thinspace \ensuremath{\mathbb R} P^{2k} \multimap \ensuremath{\mathbb R} P^{2k}$ is split, $\operatorname{\text{Fix}}(\phi)=\bigcup_{i=1}^{n} \operatorname{\text{Fix}}(f_i)$, where $\phi=\brak{f_1,\ldots,f_n}$, and for all $i=1,\ldots,n$, $f_{i}$ is a self-map of $\ensuremath{\mathbb R} P^{2k}$. It follows that $\phi$ has at least $n$ fixed points. So suppose that $H$ is the trivial subgroup of $\pi_1(\ensuremath{\mathbb R} P^{2k})$. Then $\widehat{\ensuremath{\mathbb R} P^{2k}}=\St[2k]$, and $q$ is the covering map $p\colon\thinspace \St[2k] \ensuremath{\longrightarrow} \ensuremath{\mathbb R} P^{2k}$. We first consider the case $n=2$. Let $\phi\colon\thinspace \ensuremath{\mathbb R} P^{2k} \multimap \ensuremath{\mathbb R} P^{2k}$ be a $2$-valued map, and let $\widehat{\Phi}_1\colon\thinspace \mathbb{S}^{2k} \ensuremath{\longrightarrow} F_2(\ensuremath{\mathbb R} P^{2k})$ be a lift of the map $\Phi_1=\Phi\circ p\colon\thinspace \mathbb{S}^{2k} \ensuremath{\longrightarrow} D_2(\ensuremath{\mathbb R} P^{2k})$ that factors through the projection $\pi\colon\thinspace F_2(\ensuremath{\mathbb R} P^{2k}) \ensuremath{\longrightarrow} D_2(\ensuremath{\mathbb R} P^{2k})$. By \repr{nielsen}, $\widehat{\Phi}_1=(f_1, f_2)$, where for $i=1,2$, $f_i\colon\thinspace \mathbb{S}^{2k} \ensuremath{\longrightarrow} \ensuremath{\mathbb R} P^{2k}$ is a single-valued map, and it follows from \repr{nullhomo} that $f_1$ and $f_2$ are null homotopic. If $\operatorname{\text{Coin}}(f_i, p)= \ensuremath{\varnothing}$ for some $i\in \brak{1,2}$, then arguing as in the second part of the proof of \repr{nullhomo}, it follows that $L(p,f_i)=\deg(p)=2$, which yields a contradiction. So $\operatorname{\text{Coin}}(f_i, p)\ne \ensuremath{\varnothing}$ for all $i\in \brak{1,2}$. Using the fact that $f_{1}$ and $f_{2}$ are coincidence free, we conclude that $\phi$ has at least two fixed points, and the result follows in this case.
Finally suppose that $n>2$. Arguing as in the case $n=2$, we obtain a lift of $\phi\circ p$ of the form $(f_1, \ldots, f_n)$, where for $i=1,\ldots,n$, $f_i\colon\thinspace \mathbb{S}^{2k} \ensuremath{\longrightarrow} \ensuremath{\mathbb R} P^{2k}$ is a map. If $i=1,\ldots,n$ then for all $j\in \brak{1,\ldots,n}$, $j\neq i$, we may apply the above argument to $f_i$ and $f_j$ to obtain $\operatorname{\text{Coin}}(f_i, p)\ne \ensuremath{\varnothing}$. Hence $\phi$ has at least $n$ fixed points, and the result follows. \end{proof}
\begin{rem} If $n>1$, we do not know whether there exists a non-split $n$-valued map $\phi\colon\thinspace \ensuremath{\mathbb R} P^{2k} \multimap \ensuremath{\mathbb R} P^{2k}$. \end{rem}
\section{Deforming (split) $n$-valued maps to fixed point and root-free maps}\label{sec:fixfree}
In this section, we generalise a standard procedure for deciding whether a single-valued map may be deformed to a fixed point free map to the $n$-valued case. We start by giving a necessary and sufficient condition for an $n$-valued map (resp.\ a split $n$-valued map) $\phi\colon\thinspace X \ensuremath{\longrightarrow} X$ to be deformable to a fixed point free $n$-valued map (resp.\ split fixed point free $n$-valued map), at least in the case where $X$ is a manifold without boundary. This enables us to prove \reth{defchineg}. We then go on to give the analogous statements for roots. Recall from \resec{intro} that $D_{1,n}(X)$ is the quotient of $F_{n+1}(X)$ by the action of the subgroup $\brak{1} \times S_n$ of the symmetric group $S_{n+1}$, and that $B_{1,n}(X)=\pi_1(D_{1,n}(X))$.
\begin{prop}\label{prop:defor} Let $n\in \ensuremath{\mathbb N}$, let $X$ be a metric space, and let $\phi\colon\thinspace X \multimap X$ be an $n$-valued map. If $\phi$ can be deformed to a fixed point free $n$-valued map, then there exists a map $\Theta\colon\thinspace X\ensuremath{\longrightarrow} D_{1,n}(X)$ such that the following diagram is commutative up to homotopy: \begin{equation}\label{eq:commdiag3} \begin{tikzcd}[ampersand replacement=\&] \& \& D_{1,n}(X) \ar{d}{\iota_{1,n}}\\ X \ar[swap]{rr}{\ensuremath{\operatorname{\text{Id}}}_{X}\times \Phi} \ar[dashrightarrow, end anchor=south west]{rru}{\Theta} \& \& X \times D_{n}(X), \end{tikzcd} \end{equation} where $\iota_{1,n}\colon\thinspace D_{1,n}(X) \ensuremath{\longrightarrow} X \times D_n(X)$ is the inclusion map. Conversely, if $X$ is a manifold without boundary and there exists a map $\Theta\colon\thinspace X\ensuremath{\longrightarrow} D_{1,n}(X)$ such that diagram~\reqref{commdiag3} is commutative up to homotopy, then the $n$-valued map $\phi\colon\thinspace X \multimap X$ may be deformed to a fixed point free $n$-valued map. \end{prop}
\begin{proof} For the first part, if $\phi'$ is a fixed point free deformation of $\phi$, then we may take the factorisation map $\Theta\colon\thinspace X\ensuremath{\longrightarrow} D_{1,n}(X)$ to be that defined by $\Theta(x)= (x, \Phi'(x))$. For the converse, the argument is similar to the proof of the case of single-valued maps, and is as follows. Let $\Theta\colon\thinspace X \ensuremath{\longrightarrow} D_{1,n}(X)$ be a homotopy factorisation that satisfies the hypotheses. Composing $\Theta$ with the projection $\overline{p}_{1,n}\colon\thinspace D_{1,n}(X)\ensuremath{\longrightarrow} X$ onto the first coordinate, we obtain a self-map of $X$ that is homotopic to the identity. Let $H\colon\thinspace X\times I \ensuremath{\longrightarrow} X$ be a homotopy between $\overline{p}_{1,n}\circ \Theta$ and $\ensuremath{\operatorname{\text{Id}}}_{X}$. Since $\overline{p}_{1,n}\colon\thinspace D_{1,n}(X)\ensuremath{\longrightarrow} X$ is a fibration and there is a lift of the restriction of $H$ to $X\times \brak{0}$, $H$ lifts to a homotopy $\widetilde{H}\colon\thinspace X\times I \ensuremath{\longrightarrow} D_{1,n}(X)$. The restriction of $\widetilde{H}$ to $X\times \brak{1}$ yields the required deformation. \end{proof}
For split $n$-valued maps, the correspondence given by \relem{split}(\ref{it:splitII}) gives rise to a statement analogous to that of \repr{defor} in terms of $F_{n}(X)$.
\begin{prop}\label{prop:equivsplit} Let $n\in \ensuremath{\mathbb N}$, let $X$ be a metric space, and let $\phi\colon\thinspace X \multimap X$ be a split $n$-valued map. If $\phi$ can be deformed to a fixed point free $n$-valued map, and if $\widehat{\Phi}\colon\thinspace X \ensuremath{\longrightarrow} F_{n}(X)$ is a lift of $\phi$, then there exists a map $\widehat{\Theta}\colon\thinspace X \ensuremath{\longrightarrow} F_{n+1}(X)$ such that the following diagram is commutative up to homotopy: \begin{equation}\label{eq:commdiag4} \begin{tikzcd}[ampersand replacement=\&] \&\& F_{n+1}(X) \ar{d}{\widehat{\iota}_{n+1}}\\ X \ar[swap]{rr}{\ensuremath{\operatorname{\text{Id}}}_{X}\times \widehat{\Phi}} \ar[dashrightarrow, end anchor=south west]{rru}{\widehat{\Theta}} \&\& X \times F_{n}(X), \end{tikzcd} \end{equation} where $\widehat{\iota}_{n+1}\colon\thinspace F_{n+1}(X) \ensuremath{\longrightarrow} X \times F_n(X)$ is the inclusion map. Conversely, if $X$ is a manifold without boundary and there exists a map $\widehat{\Theta}\colon\thinspace X\ensuremath{\longrightarrow} F_{n+1}(X)$ such that diagram~\reqref{commdiag4} is commutative up to homotopy, then the split $n$-valued map $\phi\colon\thinspace X \multimap X$ may be deformed through split maps to a fixed point free split $n$-valued map. \end{prop}
\begin{proof} Similar to that of \repr{defor}. \end{proof}
We now apply Propositions~\ref{prop:defor} and~\ref{prop:equivsplit} to prove \reth{defchineg}, which treats the case where $X$ is a compact surface without boundary (orientable or not) of non-positive Euler characteristic.
\begin{proof}[Proof of \reth{defchineg}] The space $D_{1,n}(X)$ is a finite covering of $D_{1+n}(X)$, so it is a $K(\pi, 1)$ since
$D_{1+n}(X)$ is a $K(\pi, 1)$ by~\cite[Corollary~2.2]{FaN}. To prove the `only if' implication of part~(\ref{it:defchinega}), diagram~\reqref{commdiag1} implies the existence of diagram~\reqref{commdiag3} using the fact that the space $D_{1,n}(X)$ is a $K(\pi, 1)$, where $\varphi=\Theta_{\#}$ is the homomorphism induced by $\Theta$ on the level of the fundamental groups. Conversely, diagram~\reqref{commdiag3} implies that the two maps $\iota_{1,n}\circ \Theta$ and $\ensuremath{\operatorname{\text{Id}}}_X\times \Phi$ are homotopic, but not necessarily by a basepoint-preserving homotopy, so diagram~\reqref{commdiag1} is commutative up to conjugacy. Let $\delta\in\pi_{1}(X) \times B_{n}(X)$ be such that $(\iota_{1,n})_{\#}\circ \varphi(\alpha)=\delta (\ensuremath{\operatorname{\text{Id}}}_X\times \phi)_{\#}(\alpha) \delta^{-1}$ for all $\alpha\in \pi_1(X)$, and let $\widehat{\delta}\in B_{1,n}(X)$ be an element such that $(\iota_{1,n})_{\#}(\widehat{\delta})=\delta.$ Considering the homomorphism $\varphi'\colon\thinspace \pi_1(X) \ensuremath{\longrightarrow} B_{1,n}(X)$ defined by $\varphi'(\alpha)=\widehat{\delta}^{-1}\varphi (\alpha)\widehat{\delta}$ for all $\alpha\in \pi_1(X)$, we obtain the commutative diagram~\reqref{commdiag1}, where we replace $\varphi$ by $\varphi'$. The proof of part~(\ref{it:defchinegb}) is similar, and is left to the reader. \end{proof}
For the case of roots, we now give statements analogous to those of Propositions~\ref{prop:defor} and~\ref{prop:equivsplit} and of \reth{defchineg}. The proofs are similar to those of the corresponding statements for fixed points, and the details are left to the reader.
\begin{prop}\label{prop:rootI} Let $n\in \ensuremath{\mathbb N}$, let $X$ and $Y$ be metric spaces, let $y_0\in Y$ be a basepoint, and let $\phi\colon\thinspace X \multimap Y$ be an $n$-valued map. If $\phi$ can be deformed to a root-free $n$-valued map, then there exists a map $\Theta\colon\thinspace X\ensuremath{\longrightarrow} D_{n}(Y\backslash\{y_0\})$ such that the following diagram is commutative up to homotopy: \begin{equation}\label{eq:commdiag3I} \begin{tikzcd}[ampersand replacement=\&] \& \& D_{n}(Y\backslash\{y_0\}) \ar{d}{\iota_{n}}\\ X \ar[swap]{rr}{ \Phi} \ar[dashrightarrow, end anchor=south west]{rru}{\Theta} \& \& D_{n}(Y), \end{tikzcd} \end{equation} where the map $\iota_{n}\colon\thinspace D_{n}(Y\backslash\{y_0\}) \ensuremath{\longrightarrow} D_n(Y)$ is induced by the inclusion map $Y\backslash \{y_0\} \mathrel{\lhook\joinrel\ensuremath{\longrightarrow}} Y$. Conversely, if $Y$ is a manifold without boundary and there exists a map $\Theta\colon\thinspace X\ensuremath{\longrightarrow} D_{n}(Y)$ such that diagram~\reqref{commdiag3I} is commutative up to homotopy, then the $n$-valued map $\phi\colon\thinspace X \multimap Y$ may be deformed to a root-free $n$-valued map. \end{prop}
For split $n$-valued maps, the correspondence of \relem{split}(\ref{it:splitII}) gives rise to a statement analogous to that of Proposition \ref{prop:rootI} in terms of $F_{n}(Y)$.
\begin{prop}\label{prop:equivsplitI} Let $n\in \ensuremath{\mathbb N}$, let $X$ and $Y$ be metric spaces, let $y_0\in Y$ a basepoint, and let $\phi\colon\thinspace X \multimap Y$ be a split $n$-valued map. If $\phi$ can be deformed to a root-free $n$-valued map then there exists a map $\widehat{\Theta}\colon\thinspace X \ensuremath{\longrightarrow} F_{n}(Y\backslash \{y_0\})$ and a lift $\widehat{\Phi}$ of $\phi$ such that the following diagram is commutative up to homotopy: \begin{equation}\label{eq:commdiag4I} \begin{tikzcd}[ampersand replacement=\&] \&\& F_{n}(Y\backslash \{y_0\}) \ar{d}{\widehat{\iota}_{n}}\\ X \ar[swap]{rr}{ \widehat{\Phi}} \ar[dashrightarrow, end anchor=south west]{rru}{\widehat{\Theta}} \&\& F_{n}(Y), \end{tikzcd} \end{equation} where the map $\widehat{\iota}_{n}\colon\thinspace F_{n}(Y\backslash \{y_0\}) \ensuremath{\longrightarrow} F_n(Y)$ is induced by the inclusion map $Y\backslash \{y_0\} \mathrel{\lhook\joinrel\ensuremath{\longrightarrow}} Y$. Conversely, if $Y$ is a manifold without boundary, and there exists a map $\widehat{\Theta}\colon\thinspace X\ensuremath{\longrightarrow} F_{n}(Y\backslash \{y_0\})$ such that diagram~\reqref{commdiag4I} is commutative up to homotopy, then the split $n$-valued map $\phi\colon\thinspace X \multimap Y$ may be deformed through split maps to a root-free split $n$-valued map. \end{prop}
Propositions~\ref{prop:rootI} and~\ref{prop:equivsplitI} may be applied to the case where $X$ and $Y$ are compact surfaces without boundary of non-positive Euler characteristic to obtain the analogue of \reth{defchineg} for roots.
\begin{thm}\label{th:defchinegI} Let $n\in \ensuremath{\mathbb N}$, and let $X$ and $Y$ be compact surfaces without boundary of non-positive Euler characteristic. \begin{enumerate}[(a)] \item\label{it:defchinegal} An $n$-valued map $\phi\colon\thinspace X \multimap Y$ can be deformed to a root-free $n$-valued map if and only if there is a homomorphism $\varphi\colon\thinspace \pi_1(X) \ensuremath{\longrightarrow} B_{n}(Y\backslash\{y_0\})$ that makes the following diagram commute:
\begin{equation*} \begin{tikzcd}[ampersand replacement=\&] \&\& B_{n}(Y\backslash\{y_0\}) \ar{d}{(\iota_{n})_{\#}}\\ \pi_{1}(X) \ar[swap]{rr}{\Phi_{\#}} \ar[dashrightarrow, end anchor=south west]{rru}{\varphi} \&\& B_{n}(Y), \end{tikzcd} \end{equation*} where $\iota_{n}\colon\thinspace D_{n}(Y\backslash\{y_0\}) \ensuremath{\longrightarrow} D_{n}(Y)$ is induced by the inclusion map $Y\backslash \{y_0\} \mathrel{\lhook\joinrel\ensuremath{\longrightarrow}} Y$, and $\Phi\colon\thinspace X\ensuremath{\longrightarrow} D_{n}(Y)$ is the $n$-unordered map associated to $\phi$.
\item\label{it:defchinegbI} A split $n$-valued map $\phi\colon\thinspace X \multimap Y$ can be deformed to a root-free $n$-valued map if and only if there exist a lift $\widehat{\Phi}\colon\thinspace X \ensuremath{\longrightarrow} F_n(Y)$ of $\phi$ and a homomorphism $\widehat{\varphi}\colon\thinspace \pi_1(X) \ensuremath{\longrightarrow} P_{n}(Y\backslash\{y_0\})$ that make the following diagram commute: \begin{equation*} \begin{tikzcd}[ampersand replacement=\&] \&\& P_{n}(Y\backslash\{y_0\}) \ar{d}{(\widehat{\iota}_{n})_{\#}}\\ \pi_{1}(X) \ar[swap]{rr}{\widehat{\Phi}_{\#}} \ar[dashrightarrow, end anchor=south west]{rru}{\widehat{\varphi}} \&\& P_{n}(Y), \end{tikzcd} \end{equation*} where $\widehat{\iota}_{n}\colon\thinspace F_{n}(Y\backslash\{y_0\}) \ensuremath{\longrightarrow} F_{n}(Y)$ is induced by the inclusion map $Y\backslash \{y_0\} \mathrel{\lhook\joinrel\ensuremath{\longrightarrow}} Y$. \end{enumerate} \end{thm}
\section{An application to split $2$-valued maps of the $2$-torus}\label{sec:toro}
In this section, we will use some of the ideas and results of \resec{fixfree} to study the fixed point theory of $2$-valued maps of the $2$-torus $\ensuremath{\mathbb{T}^{2}}$. We restrict our attention to the case where the maps are split, \emph{i.e.}\ we consider $2$-valued maps of the form $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ that admit a lift $\widehat{\Phi}\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$, where $\widehat{\Phi}=(f_1, f_2)$, $f_1$ and $f_2$ being coincidence-free self-maps of $\ensuremath{\mathbb{T}^{2}}$. We classify the set of homotopy classes of split $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$, and we study the question of the characterisation of those split $2$-valued maps that can be deformed to fixed point free $2$-valued maps. The case of arbitrary $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$ will be treated in a forthcoming paper. In \resec{toro2}, we give presentations of the groups $P_{2}(\ensuremath{\mathbb{T}^{2}})$, $B_{2}(\ensuremath{\mathbb{T}^{2}})$ and $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ that will be used in the following sections, where $1$ denotes a basepoint of $\ensuremath{\mathbb{T}^{2}}$. In \resec{descript}, we describe the set of based and free homotopy classes of split $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$. In \resec{fptsplit2}, we give a formula for the Nielsen number, and we derive a necessary condition for such a split $2$-valued map to be deformable to a fixed point free $2$-valued map. We then give an infinite family of homotopy classes of split $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$ that satisfy this condition and that
may be deformed to fixed point free $2$-valued maps. To facilitate the calculations, in \resec{p2Tminus1}, we shall show that the fixed point problem is equivalent to a root problem.
\subsection{The groups $P_{2}(\ensuremath{\mathbb{T}^{2}})$, $B_{2}(\ensuremath{\mathbb{T}^{2}})$ and $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$}\label{sec:toro2}
In this section, we give presentations of $P_{2}(\ensuremath{\mathbb{T}^{2}})$, $B_{2}(\ensuremath{\mathbb{T}^{2}})$ and $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ that will be used in the following sections. Other presentations of these groups may be found in the literature, see~\cite{Bel,Bi,GM,Sco} for example. We start by considering the group $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$. If $u$ and $v$ are elements of a group $G$, we denote their commutator $uvu^{-1}v^{-1}$ by $[u,v]$, and the commutator subgroup of $G$ by $\Gamma_{2}(G)$. If $A$ is a subset of $G$ then $\ang{\!\ang{A}\!}_{G}$ will denote the normal closure of $A$ in $G$.
\begin{prop}\label{prop:presP2T1} The group $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ admits the following presentation: \begin{enumerate} \item[generators:] $\rho_{1,1}$, $\rho_{1,2}$, $\rho_{2,1}$, $\rho_{2,2}$, $B_{1,2}$, $B$ and $B'$. \item[relations:]\mbox{} \begin{enumerate}[(a)] \item\label{it:presP2T1a} $\rho_{2,1}\rho_{1,1}\rho_{2,1}^{-1}=B_{1,2}\rho_{1,1}B_{1,2}^{-1}$. \item $\rho_{2,1}\rho_{1,2}\rho_{2,1}^{-1}=B_{1,2}\rho_{1,2}\rho_{1,1}^{-1}B_{1,2}\rho_{1,1}B_{1,2}^{-1}$. \item $\rho_{2,2}\rho_{1,1}\rho_{2,2}^{-1}=\rho_{1,1}B_{1,2}^{-1}$. \item $\rho_{2,2}\rho_{1,2}\rho_{2,2}^{-1}=B_{1,2}\rho_{1,2}B_{1,2}^{-1}$. \item\label{it:presP2T1e} $\rho_{2,1}B\rho_{2,1}^{-1}=B$ and $\rho_{2,2}B\rho_{2,2}^{-1}=B$. \item\label{it:presP2T1f} $\rho_{2,1}B_{1,2}\rho_{2,1}^{-1}=B_{1,2} \rho_{1,1}^{-1}B_{1,2} \rho_{1,1}B_{1,2}^{-1}$ and $\rho_{2,2}B_{1,2}\rho_{2,2}^{-1}=B_{1,2} \rho_{1,2}^{-1}B_{1,2} \rho_{1,2}B_{1,2}^{-1}$. \item\label{it:relf} $B'\rho_{1,1}B'^{-1}=\rho_{1,1}$ and $B'\rho_{1,2}B'^{-1}=\rho_{1,2}$. \item\label{it:relg} $B'B_{1,2}B'^{-1}=B_{1,2}^{-1}B^{-1} B_{1,2}BB_{1,2}$ and $B'BB'^{-1}=B_{1,2}^{-1}BB_{1,2}$. \item\label{it:relh} $[\rho_{1,1},\rho_{1,2}^{-1}]=BB_{1,2}$ and $[\rho_{2,1},\rho_{2,2}^{-1}]=B_{1,2}B'$. \end{enumerate} \end{enumerate} In particular, $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ is a semi-direct product of the free group of rank three generated by $\brak{\rho_{1,1},\rho_{1,2},B_{1,2}}$ by the free group of rank two generated by $\brak{\rho_{2,1},\rho_{2,2}}$. \end{prop}
Geometric representatives of the generators of $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ are illustrated in Figure~\ref{fig:gens}. The torus is obtained from this figure by identifying the boundary to a point.
\begin{figure}\label{fig:gens}
\end{figure}
\begin{rem} The inclusion of $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\{1\})$ in $P_{2}(\ensuremath{\mathbb{T}^{2}})$ induces a surjective homomorphism $\map{\alpha}{P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\})}{P_{2}(\ensuremath{\mathbb{T}^{2}})}$ that sends $B$ and $B'$ to the trivial element, and sends each of the remaining generators (considered as an element of $P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\})$) to itself (considered as an element of $P_2(\ensuremath{\mathbb{T}^{2}})$). Applying this to the presentation of $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\{1\})$ given by \repr{presP2T1}, we obtain the presentation of $P_{2}(\ensuremath{\mathbb{T}^{2}})$ given in~\cite{FH}. \end{rem}
\begin{proof}[Proof of \repr{presP2T1}] Consider the following Fadell-Neuwirth short exact sequence: \begin{equation}\label{eq:fnses} 1\ensuremath{\longrightarrow} P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1,x_{2}},x_{1}) \ensuremath{\longrightarrow} P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, (x_{1},x_{2})) \xrightarrow{(p_{2})_{\#}} P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2}) \ensuremath{\longrightarrow} 1, \end{equation} where $(p_{2})_{\#}$ is the homomorphism given geometrically by forgetting the first string and induced by the projection $p_{2}\colon\thinspace F_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})\ensuremath{\longrightarrow} \ensuremath{\mathbb{T}^{2}}\setminus\brak{1}$ onto the second coordinate. The kernel $K=P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1,x_{2}},x_{1})$ of $(p_{2})_{\#}$ (resp.\ the quotient $Q=P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{1})$) is a free group of rank three (resp.\ two). It will be convenient to choose presentations for these two groups that have an extra generator. From Figure~\ref{fig:gens}, we take $K$ (resp.\ $Q$) to be generated by $X=\brak{\rho_{1,1},\rho_{1,2},B_{1,2},B}$ (resp.\ $Y=\brak{\rho_{2,1},\rho_{2,2},B'}$) subject to the single relation $[\rho_{1,1},\rho_{1,2}^{-1}]=BB_{1,2}$ (resp.\ $[\rho_{2,1},\rho_{2,2}^{-1}]=B'$). We apply standard methods to obtain a presentation of the group extension $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, (x_{1},x_{2}))$~\cite[Proposition~1, p.~139]{Jo}. This group is generated by the union of $X$ with coset representatives of $Y$, which we take to be the same elements geometrically, but considered as elements of $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, (x_{1},x_{2}))$. This yields the given generating set. There are three types of relation. The first is that of $K$. The second type of relation is obtained by lifting the relation of $Q$ to $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, (x_{1},x_{2}))$, which gives rise to the relation $[\rho_{2,1},\rho_{2,2}^{-1}]B'^{-1}=B_{1,2}$. The third type of relation is obtained by rewriting the conjugates of the elements of $X$ by the chosen coset representatives of the elements of $Y$ in terms of the elements of $X$ using the geometric representatives of $X$ and $Y$ illustrated in Figure~\ref{fig:gens}. We leave the details to the reader. The last part of the statement is a consequence of the fact that $K$ (resp.\ $Q$) is a free group of rank three (resp.\ two), so the short exact sequence~\reqref{fnses} splits. \end{proof}
\begin{rem} For future purposes, it will be convenient to have the following relations at our disposal: \begin{align*} \rho_{2,1}^{-1}\rho_{1,1}\rho_{2,1}&=\rho_{1,1}B_{1,2}^{-1}\rho_{1,1}B_{1,2}\rho_{1,1}^{-1} & \rho_{2,1}^{-1}\rho_{1,2}\rho_{2,1}&=\rho_{1,1}B_{1,2}^{-1}\rho_{1,1}^{-1} \rho_{1,2}B_{1,2}^{-1} \rho_{1,1}B_{1,2}\rho_{1,1}^{-1}\\ \rho_{2,2}^{-1}\rho_{1,1}\rho_{2,2}&=\rho_{1,1}\rho_{1,2}B_{1,2}\rho_{1,2}^{-1} & \rho_{2,2}^{-1}\rho_{1,2}\rho_{2,2}&=\rho_{1,2}B_{1,2}^{-1} \rho_{1,2}B_{1,2}\rho_{1,2}^{-1}\\ \rho_{2,1}^{-1}B_{1,2}\rho_{2,1}&=\rho_{1,1}B_{1,2}\rho_{1,1}^{-1} & \rho_{2,2}^{-1}B_{1,2}\rho_{2,2}&=\rho_{1,2}B_{1,2}\rho_{1,2}^{-1}. \end{align*} As in the proof of \repr{presP2T1}, these equalities may be derived geometrically. \end{rem}
The presentation of $P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\}, (x_1,x_2))$ given by \repr{presP2T1} may be modified to obtain another presentation that highlights its algebraic structure as a semi-direct product of free groups of finite rank.
\begin{prop}\label{prop:presTminus1alta} The group $P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\}, (x_1,x_2))$ admits the following presentation: \begin{enumerate} \item[generators:] $B$, $u$, $v$, $x$ and $y$. \item[relations:]\mbox{} \begin{enumerate}[(a)] \item\label{it:altpresa} $xux^{-1}=u$. \item\label{it:altpresb} $xvx^{-1}=v[v^{-1}, u]B^{-1}[u, v^{-1}]$. \item\label{it:altpresc} $xBx^{-1}=u[v^{-1}, u]B[u, v^{-1}]u^{-1}$. \item\label{it:altpresd} $yuy^{-1}=v[v^{-1}, u]Buv^{-1}$. \item\label{it:altprese} $yvy^{-1}=v$. \item\label{it:altpresf} $yBy^{-1}=v[v^{-1},u]B[u,v^{-1}]v^{-1}=uvu^{-1}Buv^{-1}u^{-1}$. \end{enumerate} \end{enumerate} In particular, $P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\}, (x_1,x_2))$ is a semi-direct product of the free group of rank three generated by $\brak{u,v,B}$ by the free group of rank two generated by $\brak{x,y}$. \end{prop}
\begin{proof} Using relation~(\ref{it:relh}) of \repr{presP2T1}, we define $B'$ as $B_{1,2}^{-1} [\rho_{2,1},\rho_{2,2}^{-1}]$ and $B_{1,2}$ as $B^{-1} [\rho_{1,1},\rho_{1,2}^{-1}]$. We then apply the following change of variables: \begin{equation}\label{eq:uvxy} \text{$u=\rho_{1,1}$, $v=\rho_{1,2}$, $x=\rho_{1,1}B_{1,2}^{-1}\rho_{2,1}$ and $y=\rho_{1,2}B_{1,2}^{-1}\rho_{2,2}$.} \end{equation} Relations~(\ref{it:presP2T1a})--(\ref{it:presP2T1e}) of \repr{presP2T1} may be seen to give rise to relations~(\ref{it:altpresa})--(\ref{it:altpresf}) of \repr{presTminus1alta}. Rewritten in terms of the generators of \repr{presTminus1alta}, relations~(\ref{it:presP2T1f})--(\ref{it:relg}) of \repr{presP2T1} are consequences of relations~(\ref{it:altpresa})--(\ref{it:altpresf}) of \repr{presTminus1alta}. To see this, using the relations of \repr{presTminus1alta}, first note that: \begin{align*} xB_{1,2}x^{-1}&= x B^{-1} [u,v^{-1}] x^{-1}=u[v^{-1}, u]B^{-1} [u, v^{-1}]u^{-1} \ldotp \bigl[u, [v^{-1}, u] B[u, v^{-1}] v^{-1}\bigr]\\ &= B^{-1} [u,v^{-1}] = B_{1,2}\;\text{and}\\
yB_{1,2}y^{-1} &= y B^{-1} [u,v^{-1}] y^{-1} = v[v^{-1},u]B^{-1}[u,v^{-1}]v^{-1}\ldotp \bigl[ v[v^{-1}, u]Buv^{-1}, v^{-1}\bigr]\\ &= B^{-1} [u,v^{-1}] = B_{1,2}.
\end{align*} In light of these relations, it is convenient to carry out the calculations using $B_{1,2}$ instead of $B$. In conjunction with the relations of the preceding remark, we obtain the following relations: \begin{equation}\label{eq:conjxyuv} \left\{ \begin{aligned} yuy^{-1}&=vB_{1,2}^{-1}uv^{-1},\; xvx^{-1}=uvu^{-1} B_{1,2}\\ y^{-1}uy&=B_{1,2} v^{-1}uv,\; x^{-1}vx=u^{-1}v B_{1,2}^{-1}u, \end{aligned} \right. \end{equation} from which it follows that: \begin{align*} \rho_{2,1}B_{1,2}\rho_{2,1}^{-1}&=B_{1,2}u^{-1}x B_{1,2} x^{-1}u B_{1,2}^{-1}= B_{1,2}u^{-1} B_{1,2}u B_{1,2}^{-1}=B_{1,2} \rho_{1,1}^{-1}B_{1,2} \rho_{1,1}B_{1,2}^{-1}\;\text{and}\\ \rho_{2,2}B_{1,2}\rho_{2,2}^{-1}&=B_{1,2}v^{-1}y B_{1,2} y^{-1}v B_{1,2}^{-1}=B_{1,2}v^{-1} B_{1,2} v B_{1,2}^{-1}= B_{1,2} \rho_{1,2}^{-1}B_{1,2} \rho_{1,2}B_{1,2}^{-1}. \end{align*} Thus relations~(\ref{it:altpresa})--(\ref{it:altpresf}) of \repr{presTminus1alta} imply relations~(\ref{it:presP2T1f}) of \repr{presP2T1}. Now $B'= B_{1,2}^{-1}[\rho_{2,1},\rho_{2,2}^{-1}]$, and a straightforward computation shows that: \begin{align*} B'=u^{-1}x y^{-1}vB_{1,2}^{-1} x^{-1}u v^{-1}y&=u^{-1}x y^{-1}vB_{1,2}^{-1} x^{-1}u v^{-1} \ldotp xyx^{-1}\ldotp xy^{-1}x^{-1}y\\ &=[v^{-1},u]B_{1,2}\ldotp xy^{-1}x^{-1}y. \end{align*} One may then check that: \begin{align*} xy^{-1}x^{-1}y u y^{-1}xyx^{-1}&=B_{1,2}^{-1}[u,v^{-1}] u[v^{-1}, u] B_{1,2}\;\text{and}\\ xy^{-1}x^{-1}y v y^{-1}xyx^{-1}&=B_{1,2}^{-1}[u,v^{-1}] v [v^{-1}, u] B_{1,2}, \end{align*} and that: \begin{align*} B' \rho_{1,1}B'^{-1}&=B' uB'^{-1}=[v^{-1},u] B_{1,2}\ldotp xy^{-1}x^{-1}y u y^{-1}xyx^{-1} B_{1,2}^{-1} [u,v^{-1}]=u=\rho_{1,1}\; \text{and}\\ B' \rho_{1,2} B'^{-1}&= B' vB'^{-1}=[v^{-1},u] B_{1,2}\ldotp xy^{-1}x^{-1}y v y^{-1}xyx^{-1} B_{1,2}^{-1} [u,v^{-1}]=v=\rho_{1,2}. \end{align*} Hence relations~(\ref{it:altpresa})--(\ref{it:altpresf}) of \repr{presTminus1alta} imply relations~(\ref{it:relf}) of \repr{presP2T1}. Furthermore, \begin{align*} B' B_{1,2}B'^{-1} &= v^{-1}uvu^{-1} B_{1,2} xy^{-1}x^{-1}y B_{1,2} y^{-1}xyx^{-1} B_{1,2}^{-1} uv^{-1}u^{-1}v\\ &= v^{-1}uvu^{-1} B_{1,2} uv^{-1}u^{-1}v=B_{1,2}^{-1}\ldotp B_{1,2} [v^{-1},u] B_{1,2} [u,v^{-1}] B_{1,2}^{-1} \ldotp B_{1,2}\\ &= B_{1,2}^{-1} B^{-1} B_{1,2} B B_{1,2}, \end{align*} and since \begin{align*} xy^{-1}x^{-1}y [u , v^{-1}] y^{-1}xyx^{-1} &= \left[B_{1,2}^{-1}[u,v^{-1}] u[v^{-1}, u] B_{1,2}, B_{1,2}^{-1}[u,v^{-1}] v^{-1} [v^{-1}, u] B_{1,2}\right]\\ &= B_{1,2}^{-1}[u,v^{-1}] B_{1,2}, \end{align*} we obtain: \begin{align*} B' B B'^{-1} &=[v^{-1},u]B_{1,2} B_{1,2}^{-1}[u,v^{-1}] B_{1,2}^{-1} B_{1,2} B_{1,2}^{-1}[u,v^{-1}]=B_{1,2}^{-1} [u,v^{-1}] B_{1,2}^{-1} \ldotp B_{1,2}\\ &=B_{1,2}^{-1} B B_{1,2}. \end{align*} Hence relations~(\ref{it:altpresa})--(\ref{it:altpresf}) of \repr{presTminus1alta} imply relations~(\ref{it:relg}) of \repr{presP2T1}. This proves the first part of the statement. The last part of the statement is a consequence of the nature of the presentation. \end{proof}
The homomorphism $\map{\alpha}{P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\})}{P_{2}(\ensuremath{\mathbb{T}^{2}})}$ mentioned in the remark that follows the statement of \repr{presP2T1} may be used to obtain a presentation of $P_{2}(T)$ in terms of the generators of \repr{presTminus1alta}. To do so, we first show that $\ker{\alpha}$ is the normal closure in $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ of $B$ and $B'$.
\begin{prop}\label{prop:presTminus2alta} $\ker{\alpha}=\ang{\!\ang{B,B'}\!}_{P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})}$. \end{prop}
\begin{proof} Consider the presentation of $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ given in \repr{presP2T1}, as well as the short exact sequence~\reqref{fnses} and the surjective homomorphism $\map{\alpha}{P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\})}{P_{2}(\ensuremath{\mathbb{T}^{2}})}$. In terms of the generators of \repr{presP2T1}, for $i=1,2$, $\alpha$ sends $\rho_{i,j}$ (resp.\ $B_{1,2}$) (considered as an element of $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, (x_{1},x_{2}))$) to $\rho_{i,j}$ (resp.\ to $B_{1,2}$) (considered as an element of $P_{2}(\ensuremath{\mathbb{T}^{2}})$), and it sends $B$ and $B'$ to the trivial element. Since $B,B'\in \ker{\alpha}$, it is clear that $\ang{\!\ang{B,B'}\!}_{P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})}\subset \ker{\alpha}$. We now proceed to prove the converse inclusion. Using the projection $(p_{2})_{\#}$ and $\alpha$, we obtain the following commutative diagram of short exact sequences: \begin{equation}\label{eq:keralpha} \begin{gathered} \begin{tikzcd}[ampersand replacement=\&] \& 1 \ar{d} \& 1 \ar{d} \& 1 \ar{d} \&\\ 1 \ar{r} \& \ker{\overline{\alpha}} \ar{r} \ar[dashed]{d}{\tau\left\lvert_{\ker{\overline{\alpha}}}\right.} \& P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1,x_{2}},x_{1}) \ar[dashed]{r}{\overline{\alpha}} \ar{d}{\tau} \& P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{x_{2}},x_{1}) \ar{r} \ar{d} \& 1\\ 1 \ar{r} \& \ker{\alpha} \ar{r} \ar[dashed, shift left=1]{d}{(p_{2})_{\#}\left\lvert_{\ker{\alpha}}\right.} \& P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, (x_{1},x_{2})) \ar{r}{\alpha} \ar{d}{(p_{2})_{\#}} \& P_{2}(\ensuremath{\mathbb{T}^{2}}, (x_{1},x_{2})) \ar{r} \ar{d}{(p_{2}')_{\#}} \& 1\\ 1 \ar{r} \& \ker{\alpha'} \ar{r} \ar{d} \ar[dashed, shift left=1]{u}{s} \& P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2}) \ar{r}{\alpha'} \ar{d} \& P_{1}(\ensuremath{\mathbb{T}^{2}},x_{2}) \ar{r} \ar{d} \& 1.\\ \& 1 \& 1 \& 1 \& \end{tikzcd} \end{gathered} \end{equation} Diagram~\reqref{keralpha} is constructed in the following manner: the surjective homomorphism $\map{\alpha'}{P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2})}{P_{1}(\ensuremath{\mathbb{T}^{2}},x_{2})}$ is induced by the inclusion $\map{\iota'}{\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}}{\ensuremath{\mathbb{T}^{2}}}$, and the right-hand column is a Fadell-Neuwirth short exact sequence, where the fibration $\map{p_{2}'}{F_{2}(\ensuremath{\mathbb{T}^{2}})}{\ensuremath{\mathbb{T}^{2}}}$ is given by projecting onto the second coordinate. The homomorphism $\map{\tau}{P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1,x_{2}},x_{1})}{P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, (x_{1},x_{2}))}$ is interpreted as inclusion. Taking $\brak{\rho_{2,1},\rho_{2,2},B'}$ to be a generating set of $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2})$ subject to the relation $[\rho_{2,1},\rho_{2,2}^{-1}]=B'$, as for $\alpha$, we see that $\alpha'(\rho_{2,j})=\rho_{2,j}$ and $\alpha'(B')=1$. Alternatively, we may consider $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2})$ to be the free group on $\brak{\rho_{2,1},\rho_{2,2}}$, and $P_{1}(\ensuremath{\mathbb{T}^{2}},x_{2})$ to be the free Abelian group on $\brak{\rho_{2,1},\rho_{2,2}}$, so $\alpha'$ is Abelianisation, and $\ker{\alpha'}=\Gamma_{2}(P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2}))$. Interpreting $\alpha'$ as the canonical projection of $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2})$ onto its quotient by the normal closure of the element $[\rho_{2,1},\rho_{2,2}^{-1}]$ in $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2})$, we see that $\ker{\alpha'}=\ang{\!\ang{\,[\rho_{2,1},\rho_{2,2}^{-1}]\,}\!}_{P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2})}$. But $[\rho_{2,1},\rho_{2,2}^{-1}]=B'$ in $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2})$, hence: \begin{equation}\label{eq:keralphaprime} \ker{\alpha'}=\ang{\!\ang{B'}\!}_{P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1},x_{2})}. \end{equation} The fact that $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2})$ is a free group implies that $\ker{\alpha'}$ is too (albeit of infinite rank). The commutativity of the lower right-hand square of~\reqref{keralpha} is a consequence of the following commutative square: \begin{equation*} \begin{tikzcd}[ampersand replacement=\&] F_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}) \ar[r, "\iota"] \ar[d, "p_{2}"] \& F_{2}(\ensuremath{\mathbb{T}^{2}}) \ar[d, "p_{2}'"]\\ F_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}) \ar[r, "\iota'"] \& F_{2}(\ensuremath{\mathbb{T}^{2}}), \end{tikzcd} \end{equation*} where $\map{\iota}{F_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})}{F_{2}(\ensuremath{\mathbb{T}^{2}})}$ is induced by the inclusion $\iota'$. Together with $\alpha$ and $\alpha'$, the second two columns of~\reqref{keralpha} give rise to a homomorphism $\map{\overline{\alpha}}{P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1,x_{2}}, x_{1})}{P_{1}(\ensuremath{\mathbb{T}^{2}} \setminus\brak{x_{2}},x_{1})}$. Taking $\brak{\rho_{1,1},\rho_{1,2},B_{1,2},B}$ to be a generating set of $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1,x_{2}}, x_{1})$, by commutativity of the diagram, we see that $\overline{\alpha}$ sends each of $\rho_{1,1}$, $\rho_{1,2}$ and $B_{1,2}$ (considered as an element of $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1,x_{2}}, (x_{1},x_{2}))$) to itself (considered as an element of $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{x_{2}}, x_{1})$), and sends $B'$ to the trivial element. In particular, $\overline{\alpha}$ is surjective. As for $\alpha'$, we may consider $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1,x_{2}}, x_{1})$ to be the free group of rank three generated by $\brak{\rho_{1,1},\rho_{1,2},B_{1,2}}$, and $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{x_{2}}, x_{1})$ to be the group generated by $\brak{\rho_{1,1},\rho_{1,2},B_{1,2}}$ subject to the relation $[\rho_{1,1},\rho_{1,2}^{-1}]=B_{1,2}$. The homomorphism $\alpha'$ may thus be interpreted as the canonical projection of $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{x_{2}}, x_{1})$ onto its quotient by $\ang{\!\ang{[\rho_{1,1},\rho_{1,2}^{-1}]B_{1,2}^{-1}}\!}_{P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{x_{2}}, x_{1})}$. But by relation~(\ref{it:relh}) of \repr{presP2T1}, $B'=[\rho_{1,1},\rho_{1,2}^{-1}]B_{1,2}^{-1}$, hence: \begin{equation*} \ker{\overline{\alpha}}=\ang{\!\ang{B'}\!}_{P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{x_{2}},x_{1})}. \end{equation*} By exactness, the first (resp.\ second) two rows of~\reqref{keralpha} give rise to an induced homomorphism $\map{\tau\left\lvert_{\ker{\overline{\alpha}}}\right.}{\ker{\overline{\alpha}}}{\ker{\alpha}}$ (resp.\ $\map{(p_{2})_{\#}\left\lvert_{\ker{\alpha}}\right.}{\ker{\alpha}}{\ker{\alpha'}}$), and $\tau\left\lvert_{\ker{\overline{\alpha}}}\right.$ is injective because $\tau$ is. The homomorphism $(p_{2})_{\#}\left\lvert_{\ker{\alpha}}\right.$ is surjective, because by~\reqref{keralphaprime}, any element $x$ of $\ker{\alpha'}$ may be written as a product of conjugates of $B'$ and its inverse by products of $\rho_{2,1}$, $\rho_{2,2}$ and $B'$. This expression, considered as an element of $P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, (x_{1},x_{2}))$, belongs to $\ker{\alpha}$, and its image under $(p_{2})_{\#}$ is equal to $x$.
The fact that $\im{\tau\left\lvert_{\ker{\overline{\alpha}}}\right.}\subset \ker{(p_{2})_{\#}\left\lvert_{\ker{\alpha}}\right.}$ follows from exactness of the second column of~\reqref{keralpha}. Conversely, if $z\in \ker{(p_{2})_{\#}\left\lvert_{\ker{\alpha}}\right.}$ then $z\in \ker{\tau}$ by exactness of the second column. So there exists $y\in P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1,x_{2}},x_{1})$ such that $\tau(y)=z$. But $\overline{\alpha}(y)=\alpha(\tau(y))=\alpha(z)=1$, and hence $y\in \ker{\overline{\alpha}}$. This proves that $\ker{(p_{2})_{\#}\left\lvert_{\ker{\alpha}}\right.}\subset \im{\tau\left\lvert_{\ker{\overline{\alpha}}}\right.}$, and we deduce that the first column is exact.
Finally, since $\ker{\alpha'}$ is free, we may pick an infinite basis consisting of conjugates of $B'$ by certain elements of $P_{1}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1}, x_{2})$. Further, there exists a section $\map{s}{\ker{\alpha'}}[\ker{\alpha}]$ for $(p_{2})_{\#}\left\lvert_{\ker{\alpha}}\right.$ that consists in sending each of these conjugates (considered as an element of $\ker{\alpha'}$) to itself (considered as an element of $\ker{\alpha}$). In particular, $\ker{\alpha}$ is an internal semi-direct product of $\tau(\ker{\overline{\alpha}})$, which is contained in $\ang{\!\ang{B}\!}_{P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})}$, by $s(\ker{\overline{\alpha}})$, which is contained in $\ang{\!\ang{B'}\!}_{P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})}$. Hence $\ker{\alpha}$ is contained in the subgroup $\ang{\!\ang{B,B'}\!}_{P_{2}(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})}$. This completes the proof of the proposition. \end{proof}
We may thus deduce the following useful presentation of $P_{2}(\ensuremath{\mathbb{T}^{2}})$.
\begin{cor}\label{cor:compactpres} The group $P_2(\ensuremath{\mathbb{T}^{2}}, (x_1,x_2))$ admits the following presentation: \begin{enumerate} \item[generators:] $u$, $v$, $x$ and $y$. \item[relations:]\mbox{} \begin{enumerate}[(a)] \item\label{it:altpresA} $xux^{-1}=u$ and $yuy^{-1}=u$. \item\label{it:altpresB} $xvx^{-1}=v$ and $yvy^{-1}=v$. \item\label{it:altpresC} $xyx^{-1}=y$. \end{enumerate} \end{enumerate} In particular, $P_2(\ensuremath{\mathbb{T}^{2}}, (x_1,x_2))$ is isomorphic to the direct product of the free group of rank two generated by $\brak{u,v}$ and the free Abelian group of rank two generated by $\brak{x,y}$. \end{cor}
\begin{rem} The decomposition of \reco{compactpres} is a special case of~\cite[Lemma~17]{BGG}. \end{rem}
\begin{proof}[Proof of \reco{compactpres}] By \repr{presTminus2alta}, a presentation of $P_2(\ensuremath{\mathbb{T}^{2}})$ may be obtained from the presentation of $P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\})$ given in \repr{presTminus1alta} by setting $B$ and $B'$ equal to $1$. Under this operation, relations~(\ref{it:altpresa}) and~(\ref{it:altpresd}) (resp.~(\ref{it:altpresb}) and~(\ref{it:altprese})) of \repr{presTminus1alta} are sent to relations~(\ref{it:altpresA}) (resp.~(\ref{it:altpresB})) of \reco{compactpres}, and relations~(\ref{it:altpresc}) and~(\ref{it:altpresf}) of \repr{presTminus1alta} become trivial. We must also take into account the fact that $B'=1$ in $P_2(\ensuremath{\mathbb{T}^{2}})$. In $P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\})$, we have: \begin{equation*} B' = B_{1,2}^{-1} [\rho_{2,1}, \rho_{2,2}^{-1}]= [v^{-1},u] B \bigl[B^{-1}[u,v^{-1}]\ldotp u^{-1}x, y^{-1}v [v^{-1},u] B\bigr], \end{equation*} and taking the image of this equation by $\alpha$, we obtain: \begin{align*} 1 &= [v^{-1},u] \bigl[[u,v^{-1}] u^{-1}x, y^{-1}v [v^{-1},u] \bigr]\\ & = [v^{-1},u] \ldotp [u,v^{-1}] u^{-1}x \ldotp y^{-1}v [v^{-1},u] \ldotp x^{-1}u [v^{-1},u] \ldotp [u,v^{-1}] v^{-1} y = u^{-1}x y^{-1}uvu^{-1} x^{-1}u v^{-1} y \end{align*} in $P_2(\ensuremath{\mathbb{T}^{2}})$. Using relations~(\ref{it:altpresA}) (resp.~(\ref{it:altpresB})) of \reco{compactpres}, it follows that $[x, y^{-1}]=1$, which yields relation~(\ref{it:altpresC}) of \reco{compactpres}. The last part of the statement is a consequence of the nature of the presentation.
\end{proof}
\begin{rem} As we saw in \repr{presTminus1alta}, $P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\}, (x_1, x_2))$ is a semi-direct product of the form $\mathbb{F}_3(u,v,B)\rtimes \mathbb{F}_2(x,y)$, the action being given by the relations of that proposition. Transposing the second two columns of the commutative diagram~\reqref{keralpha}, we obtain, up to isomorphism, the following commutative diagram: \begin{equation}\label{eq:f3f2} \begin{tikzcd}[ampersand replacement=\&] 1 \ar{r}\& \mathbb{F}_3(u,v,B) \ar{d}{\overline{\alpha}} \ar{r}{\tau} \& \mathbb{F}_3(u,v,B)\rtimes \mathbb{F}_2(x,y) \ar{d}{\alpha} \ar{r}{(p_{2})_{\#}} \ar{r} \& \mathbb{F}_2(x,y) \ar{d}{\alpha'} \ar{r} \& 1\\ 1 \ar{r}\& \mathbb{F}_2(u,v) \ar{r} \& \mathbb{F}_2(u,v) \times \ensuremath{\mathbb Z}^{2} \ar{r}{(p_{2}')_{\#}} \& \ensuremath{\mathbb Z}^{2} \ar{r} \& 1, \end{tikzcd} \end{equation} where $\alpha(u)=u$, $\alpha(v)=v$, $\alpha(B)=1$, $\alpha(x)=(1; (1,0))$ and $\alpha(y)=(1; (0,1))$. This is a convenient setting to study the question of whether a $2$-valued map may be deformed to a root-free $2$-valued map (which implies that the corresponding map may be deformed to a fixed point free map), since using~\reqref{f3f2} and \reth{defchinegI}(\ref{it:defchinegbI}), the question is equivalent to a lifting problem, to which we will refer in \resec{fptsplit2}, notably in Propositions~\ref{prop:necrootfree2} and~\ref{prop:construct2valprop}. \end{rem}
Using the short exact sequence~\reqref{sesbraid}, we obtain the following presentation of the full braid group $B_2(\ensuremath{\mathbb{T}^{2}}, (x_1,x_2))$ from that of $P_2(\ensuremath{\mathbb{T}^{2}}, (x_1,x_2))$ given by \reco{compactpres}.
\begin{prop}\label{prop:presful} The group $B_2(\ensuremath{\mathbb{T}^{2}}, (x_1,x_2))$ admits the following presentation: \begin{enumerate} \item[generators:] $u$, $v$, $x$, $y$ and $\sigma$. \item[relations:]\mbox{} \begin{enumerate}[(a)] \item\label{it:presB2Ta} $xux^{-1}=u$ and $yuy^{-1}=u$. \item $xvx^{-1}=v$ and $yvy^{-1}=v$. \item\label{it:presB2Tc} $xyx^{-1}=y$. \item\label{it:presB2Td} $\sigma^{2}=[u,v^{-1}]$. \item\label{it:presB2Te} $\sigma x \sigma^{-1}=x$ and $\sigma y \sigma^{-1}=y$. \item\label{it:presB2Tf} $\sigma u \sigma^{-1}=[u,v^{-1}]u^{-1}x$ and $\sigma v \sigma^{-1}=[u,v^{-1}]v^{-1}y$. \end{enumerate} \end{enumerate} \end{prop}
\begin{proof} Once more, we apply the methods of~\cite[Proposition~1, p.~139]{Jo}, this time to the short exact sequence~\reqref{sesbraid} for $X=\ensuremath{\mathbb{T}^{2}}$ and $n=2$, where we take $P_{2}(\ensuremath{\mathbb{T}^{2}})$ to have the presentation given by \reco{compactpres}. A coset representative of the generator of $\ensuremath{\mathbb Z}_{2}$ is given by the braid $\sigma=\sigma_{1}$ that swaps the two basepoints. Hence $\brak{u,v,x,y,\sigma}$ generates $B_{2}(\ensuremath{\mathbb{T}^{2}})$. Relations~(\ref{it:presB2Ta})--(\ref{it:presB2Tc}) emanate from the relations of $P_{2}(\ensuremath{\mathbb{T}^{2}})$. Relation~(\ref{it:presB2Td}) is obtained by lifting the relation of $\ensuremath{\mathbb Z}_{2}$ to $B_{2}(\ensuremath{\mathbb{T}^{2}})$ and using the fact that $\sigma^{2}=B_{1,2}=[u,v^{-1}]$. To obtain relations~(\ref{it:presB2Te}) and~(\ref{it:presB2Tf}), by geometric arguments, one may see that for $j\in \brak{1,2}$, $\sigma \rho_{1,j} \sigma^{-1}= \rho_{2,j}$ and $\sigma \rho_{2,j} \sigma^{-1}= B_{1,2}\rho_{2,j}B_{1,2}^{-1}$, and one then uses \req{uvxy} to express these relations in terms of $u,v,x$ and $y$. \end{proof}
\subsection{A description of the homotopy classes of $2$-ordered and split $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$, and the computation of the Nielsen number}\label{sec:descript}
In this section, we describe the homotopy classes of $2$-ordered (resp.\ split $2$-valued) maps of $\ensuremath{\mathbb{T}^{2}}$ using the group structure of $P_{2}(\ensuremath{\mathbb{T}^{2}})$ (resp.\ of $B_{2}(\ensuremath{\mathbb{T}^{2}})$) given in \resec{toro2}.
\begin{prop}\label{prop:baseT2}\mbox{} \begin{enumerate} \item\label{it:baseT2a} The set $[\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})]_0$ of based homotopy classes of $2$-ordered maps of $\ensuremath{\mathbb{T}^{2}}$ is in one-to-one correspondence with the set of commuting, ordered pairs of elements of $P_{2}(\ensuremath{\mathbb{T}^{2}})$. \item\label{it:baseT2b} The set $[\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})]$ of homotopy classes of $2$-ordered maps of $\ensuremath{\mathbb{T}^{2}}$ is in one-to-one correspondence with the set of commuting, conjugate, ordered pairs of $P_{2}(\ensuremath{\mathbb{T}^{2}})$, i.e.\ two commuting pairs $(\alpha_1, \beta_1)$ and $(\alpha_2, \beta_2)$ of $P_{2}(\ensuremath{\mathbb{T}^{2}})$ give rise to the same homotopy class of $2$-ordered maps of $\ensuremath{\mathbb{T}^{2}}$ if there exists $\delta\in P_{2}(\ensuremath{\mathbb{T}^{2}})$ such that $\delta\alpha_1\delta^{-1}=\alpha_2$ and $\delta\beta_1\delta^{-1}=\beta_2$. \item\label{it:baseT2c} Under the projection $\widehat{\pi}\colon\thinspace [\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})] \ensuremath{\longrightarrow} [\ensuremath{\mathbb{T}^{2}},D_{2}(\ensuremath{\mathbb{T}^{2}})]$ induced by the covering map $\pi\colon\thinspace F_2(\ensuremath{\mathbb{T}^{2}}) \ensuremath{\longrightarrow} D_2(\ensuremath{\mathbb{T}^{2}})$, two homotopy classes of $2$-ordered maps of $\ensuremath{\mathbb{T}^{2}}$ are sent to the same homotopy class of $2$-unordered maps if and only if any two pairs of braids that represent the maps are conjugate in $B_{2}(\ensuremath{\mathbb{T}^{2}})$. \end{enumerate} \end{prop}
\begin{proof} Let $x_0\in \ensuremath{\mathbb{T}^{2}}$ and $(y_0, z_0)\in F_2(\ensuremath{\mathbb{T}^{2}})$ be basepoints, and let $\Psi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$ be a basepoint-preserving $2$-ordered map. The restriction of $\Psi$ to the meridian $\mu$ and the longitude $\lambda$ of $\ensuremath{\mathbb{T}^{2}}$, which are geometric representatives of the elements of the basis $(e_{1},e_{2})$ of $\pi_1(\ensuremath{\mathbb{T}^{2}})$, gives rise to a pair of geometric braids. The resulting pair $(\Psi_{\#}(e_{1}),\Psi_{\#}(e_{2}))$ of elements of $P_{2}(\ensuremath{\mathbb{T}^{2}})$ obtained via the induced homomorphism $\Psi_{\#}\colon\thinspace \pi_{1}(\ensuremath{\mathbb{T}^{2}}) \ensuremath{\longrightarrow} P_{2}(\ensuremath{\mathbb{T}^{2}})$ is an invariant of the based homotopy class of the map $\Psi$, and the two braids $\Psi_{\#}(e_{1})$ and $\Psi_{\#}(e_{2})$ commute. Conversely, given a pair of braids $(\alpha, \beta)$ of $P_{2}(\ensuremath{\mathbb{T}^{2}})$, let $f_{1}\colon\thinspace \St[1] \ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$ and $f_{2}\colon\thinspace \St[1] \ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$ be geometric representatives of $\alpha$ and $\beta$ respectively, \emph{i.e.}\ $\alpha=[f_{1}]$ and $\beta=[f_{2}]$. Then we define a geometric map from the wedge of two circles into $F_{2}(\ensuremath{\mathbb{T}^{2}})$ by sending $x\in \St[1] \vee \St[1]$ to $f_{1}(x)$ (resp.\ to $f_{2}(x)$) if $x$ belongs to the first (resp.\ second) copy of $\St[1]$. By classical obstruction theory in low dimension, this map extends to $\ensuremath{\mathbb{T}^{2}}$ if and only if $\alpha$ and $\beta$ commute as elements of $P_{2}(\ensuremath{\mathbb{T}^{2}})$, and part~(\ref{it:baseT2a}) follows. Parts~(\ref{it:baseT2b}) and~(\ref{it:baseT2c}) are consequences of part~(\ref{it:baseT2a}) and classical general facts about maps between spaces of type $K(\pi, 1)$, see~\cite[Chapter~V, Theorem~4.3]{Wh} for example. \end{proof}
Applying \repr{equivsplit} to $\ensuremath{\mathbb{T}^{2}}$, we obtain the following consequence.
\begin{prop}\label{prop:lifttorus} If $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ is a split $2$-valued map and $\widehat{\Phi}\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$ is a lift of $\phi$, the map $\phi$ can be deformed to a fixed point free $2$-valued map if and only if there exist commuting elements $\alpha_1, \alpha_2 \in P_{3}(\ensuremath{\mathbb{T}^{2}})$ such that $\alpha_1$ (resp.\ $\alpha_2$) projects to $(e_1, \widehat{\Phi}_{\#}(e_1))$ (resp.\ $(e_2, \widehat{\Phi}_{\#}(e_2))$) under the homomorphism induced by the inclusion map $\widehat{\iota}_{3}\colon\thinspace F_{3}(\ensuremath{\mathbb{T}^{2}}) \ensuremath{\longrightarrow} \ensuremath{\mathbb{T}^{2}}\times F_{2}(\ensuremath{\mathbb{T}^{2}})$. \end{prop}
\begin{proof} Since $\ensuremath{\mathbb{T}^{2}}$ is a space of type $K(\pi, 1)$, the existence of diagram~\reqref{commdiag4} is equivalent to that of the corresponding induced diagram on the level of fundamental groups. It then suffices to take $\alpha_{1}=\Theta_{\#}(e_1)$ and $\alpha_{2}=\Theta_{\#}(e_2)$ in the statement of \repr{equivsplit}. \end{proof}
Proposition~\ref{prop:lifttorus} gives a criterion to decide whether a split $2$-valued map of $\ensuremath{\mathbb{T}^{2}}$ can be deformed to a fixed point free $2$-valued map. However, from a computational point of view, it seems better to use an alternative condition in terms of roots (see \resec{exrfm}).
In the following proposition, we make use of the identification of $P_{2}(\ensuremath{\mathbb{T}^{2}})$ with $\mathbb{F}_{2} \times \ensuremath{\mathbb Z}^{2}$ given in \reco{compactpres}.
\begin{prop}\label{prop:exismaps}\mbox{} \begin{enumerate} \item\label{it:exismapsa} The set $[\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})]_0$ of based homotopy classes of $2$-ordered maps of $\ensuremath{\mathbb{T}^{2}}$ is in one-to-one correspondence with the set of pairs $(\alpha, \beta)$ of elements of $\mathbb{F}_2 \times \ensuremath{\mathbb Z}^{2}$ of the form $\alpha=(w^r,(a,b))$, $\beta=(w^s, (c,d))$, where $(a,b), (c,d), (r,s)\in \ensuremath{\mathbb Z}^{2}$ and $w\in \mathbb{F}_2$. Further, up to taking a root of $w$ if necessary, we may assume that $w$ is either trivial or is a primitive element of $\mathbb{F}_2$ (i.e.\ $w$ is not a proper power of another element of $\mathbb{F}_2$).
\item\label{it:exismapsb} The set $[\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})]$ of homotopy classes of $2$-ordered maps of $\ensuremath{\mathbb{T}^{2}}$ is in one-to-one correspondence with the set of the equivalence classes of pairs $(\alpha, \beta)$ of elements of $\mathbb{F}_2 \times \ensuremath{\mathbb Z}^{2}$ of the form given in part~(\ref{it:exismapsa}), where the equivalence relation is defined as follows: the pairs of elements $( (w_1^{r_{1}},(a_1,b_1)), (w_1^{s_{1}}, (c_1,d_1)))$ and $( (w_2^{r_{2}},(a_2,b_2)), (w_2^{s_{2}}, (c_2,d_2)))$ of $\mathbb{F}_2 \times \ensuremath{\mathbb Z}^{2}$ are equivalent if and only if $(a_1,b_1,c_1,d_1) = (a_2,b_2,c_2,d_2)$, and either:
\begin{enumerate}[(i)] \item $w_1=w_2=1$, or \item $w_1$ and $w_2$ are primitive, and there exists $\ensuremath{\varepsilon}\in \brak{1,-1}$ such that $w_1$ and $w_2^{\ensuremath{\varepsilon}}$ are conjugate in $\mathbb{F}_2$, and $(r_1,s_{1})=\ensuremath{\varepsilon} (r_2,s_{2})\ne (0,0)$. \end{enumerate} \end{enumerate} \end{prop}
\begin{proof} Part~(\ref{it:exismapsa}) follows using \repr{baseT2}(\ref{it:baseT2a}), the identification of $P_{2}(\ensuremath{\mathbb{T}^{2}})$ with $\mathbb{F}_2\times \ensuremath{\mathbb Z}^2$ given by \reco{compactpres}, and the fact that two elements of $\mathbb{F}_2$ commute if and only if they are powers of some common element of $\mathbb{F}_2$. Part~(\ref{it:exismapsb}) is a consequence of \repr{baseT2}(\ref{it:baseT2b}), \reco{compactpres}, and the straightforward description of the conjugacy classes of the group $\mathbb{F}_2\times \ensuremath{\mathbb Z}^2$. \end{proof}
To describe the homotopy classes of split $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$, let us consider the set $[\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})]$ of homotopy classes of $2$-ordered maps and the action of $\ensuremath{\mathbb Z}_2$ on this set that is induced by the action of $\ensuremath{\mathbb Z}_2$ on $F_2(\ensuremath{\mathbb{T}^{2}})$. By \relem{split}(\ref{it:splitIII}), the corresponding set of orbits $[\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})]/\ensuremath{\mathbb Z}_{2}$ is in one-to-one correspondence with the set $\splitmap{\ensuremath{\mathbb{T}^{2}}}{\ensuremath{\mathbb{T}^{2}}}{2}/\!\sim$ of homotopy classes of split $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$. Given a homotopy class of a $2$-ordered map of $\ensuremath{\mathbb{T}^{2}}$, choose a based representative $f\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2(\ensuremath{\mathbb{T}^{2}})$. The based homotopy class of $f$ is determined by the element $f_{\#}$ of $\operatorname{\text{Hom}}(\ensuremath{\mathbb Z}^{2}, P_2(\ensuremath{\mathbb{T}^{2}}))$. In turn, by \repr{exismaps}(\ref{it:exismapsa}),
$f_{\#}$ is determined by a pair of elements of $P_{2}(\ensuremath{\mathbb{T}^{2}})$ of the form $( (w^r,(a,b)), (w^s, (c,d)))$, where $(a,b)$, $(c,d)$ and $(r,s)$ belong to $\ensuremath{\mathbb Z}^2$, and $w\in \mathbb{F}_2$. To characterise the equivalence class of $f$ in $[\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})]/\ensuremath{\mathbb Z}_{2}$, we first consider the set of conjugates of this pair by the elements of $P_2(\ensuremath{\mathbb{T}^{2}})$, which by \repr{baseT2}(\ref{it:baseT2b}) describes the homotopy class of $f$ in $[\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})]$, and secondly, we take into account the $\ensuremath{\mathbb Z}_2$-action by conjugating by the elements of $B_2(\ensuremath{\mathbb{T}^{2}})$. So the equivalence class of $f$ in $[\ensuremath{\mathbb{T}^{2}},F_{2}(\ensuremath{\mathbb{T}^{2}})]/\ensuremath{\mathbb Z}_{2}$ is characterised by the set of conjugates of the pair $( (w^r,(a,b)), (w^s, (c,d)))$ by elements of $B_2(\ensuremath{\mathbb{T}^{2}})$. The presentation of $B_{2}(\ensuremath{\mathbb{T}^{2}})$ given by~\repr{presful} contains the action by conjugation of $\sigma$ on $P_{2}(\ensuremath{\mathbb{T}^{2}})$. Consider the homomorphism involution of $\mathbb{F}_2(u,v)$ that is defined on the generators of $\mathbb{F}_2(u,v)$ by $u \longmapsto u^{-1}$ and $v \longmapsto v^{-1}$. The image of an element $w\in \mathbb{F}_2(u,v)$ by this automorphism will be denoted by $\widehat{w}$. With respect to the decomposition of $P_{2}(\ensuremath{\mathbb{T}^{2}})$ given by \reco{compactpres}, let $\gamma\colon\thinspace P_2(\ensuremath{\mathbb{T}^{2}}) \ensuremath{\longrightarrow} \mathbb{F}_2(u,v)$ denote the projection onto $\mathbb{F}_2(u,v)$. Let $\operatorname{\text{Ab}}\colon\thinspace \mathbb{F}_2(u,v) \ensuremath{\longrightarrow} \ensuremath{\mathbb Z}^{2}$ denote the Abelianisation homomorphism that sends $u$ to $(1,0)$ and $v$ to $(0,1)$. We write $\operatorname{\text{Ab}}(w)=(|w|_{u}, |w|_{v})$, where $|w|_{u}$ (resp.\ $|w|_{v}$) denotes the exponent sum of $u$ (resp.\ $v$) in the word $w$, and $\ell(w)$ will denote the word length of $w$ with respect to $u$ and $v$. One may check easily that $(\widehat{w})^{-1}=\widehat{w^{-1}}$ and $\ell(w)=\ell(\widehat{w})$ for all $w\in \mathbb{F}_2(u,v)$. Note also that if $\lambda\in \mathbb{F}_2(u,v)$ and $r\in \ensuremath{\mathbb Z}$ then: \begin{equation}\label{eq:lambdahat} \widehat{\lambda} (\lambda \widehat{\lambda})^{r} \lambda=\widehat{\lambda} (\lambda \widehat{\lambda})^{r} \widehat{\lambda}^{-1} \ldotp \widehat{\lambda} \lambda=(\widehat{\lambda} \lambda \widehat{\lambda}\widehat{\lambda}^{-1})^{r} \widehat{\lambda} \lambda=(\widehat{\lambda}\lambda)^{r+1}. \end{equation}
\begin{lem}\label{lem:conjhat} For all $w\in \mathbb{F}_{2}(u,v)$, $\widehat{w}=vu^{-1}\gamma(\sigma w\sigma^{-1})uv^{-1}$ and \begin{equation}\label{eq:conjsigma}
\sigma w \sigma^{-1}=(uv^{-1}\widehat{w} vu^{-1}, (|w|_{u}, |w|_{v})). \end{equation} In particular, $\widehat{w}$ is conjugate to $\gamma(\sigma w\sigma^{-1})$ in $\mathbb{F}_2(u,v)$. \end{lem}
\begin{proof} By \repr{presful},
we have: \begin{equation*}
\text{$\sigma u \sigma^{-1} = [u,v^{-1}]u^{-1} x=uv^{-1}\widehat{u} vu^{-1} x^{|u|_{u}}$ and
$\sigma v \sigma^{-1} = [u,v^{-1}]v^{-1} y=uv^{-1}\widehat{v} vu^{-1} y^{|v|_{v}}$.} \end{equation*} If $w\in \mathbb{F}_{2}(u,v)$ then~\reqref{conjsigma} follows because
$x$ and $y$ belong to the centre of $P_{2}(\ensuremath{\mathbb{T}^{2}})$. Thus $\gamma(\sigma w \sigma^{-1})=uv^{-1}\widehat{w} vu^{-1}$ as required. \end{proof}
\begin{lem}\label{lem:conjhat1}\mbox{} \begin{enumerate} \item\label{it:classifI} Let $a,b\in \mathbb{F}_{2}(u,v)$ be such that $ab$ is written in reduced form. Then $ab=\widehat{b}\,\widehat{a}$ if and only if there exist $\lambda\in \mathbb{F}_{2}(u,v)$ and $r,s\in \ensuremath{\mathbb Z}$ such that: \begin{equation}\label{eq:concl1} \text{$a=(\widehat{\lambda} \lambda)^s\widehat{\lambda}$ and $b=(\lambda\widehat{\lambda})^r\lambda$.} \end{equation}
\item\label{it:classifII} For all $w\in \mathbb{F}_{2}(u,v)$, $w$ and $\widehat{w}$ are conjugate in $\mathbb{F}_2(u,v)$ if and only if there exist $\lambda\in \mathbb{F}_2(u,v)$ and $l\in \ensuremath{\mathbb Z}$ such that: \begin{equation}\label{eq:concl2} w=(\lambda \widehat{\lambda})^l. \end{equation} \end{enumerate} \end{lem}
\begin{rem} By modifying the definition of the $\widehat{\cdot}$ homomorphism appropriately, \relem{conjhat1} and its proof may be generalised to any free group of finite rank on a given set. \end{rem}
\begin{proof}[Proof of \relem{conjhat1}] We first prove the `if' implications of~(\ref{it:classifI}) and~(\ref{it:classifII}). For part~(\ref{it:classifI}), let $a,b\in \mathbb{F}_{2}(u,v)$ be such that $ab$ is written in reduced form, and that~\reqref{concl1} holds. Using~\reqref{lambdahat}, we have: \begin{equation*} ab=(\widehat{\lambda} \lambda)^s\widehat{\lambda}(\lambda\widehat{\lambda})^r\lambda=(\widehat{\lambda} \lambda)^{r+s+1}=(\widehat{\lambda} \lambda)^{r}\widehat{\lambda}(\lambda \widehat{\lambda})^{s}\lambda= \widehat b\,\widehat a. \end{equation*} For part~(\ref{it:classifII}), if~\reqref{concl2} holds then $\widehat{w}=(\widehat{\lambda} \lambda)^l=\widehat{\lambda}(\lambda\widehat{\lambda})^l(\widehat{\lambda})^{-1}=\widehat{\lambda} w\widehat{\lambda}^{-1}$ by~\reqref{lambdahat}, so $w$ and $\widehat{w}$ are conjugate in $\mathbb{F}_2(u,v)$.
Finally, we prove the `only if' implications of~(\ref{it:classifI}) and~(\ref{it:classifII}) simultaneously by induction on the length $k$ of the words $ab$ and $w$, which we assume to be non trivial and written in reduced form. Let~(E1) denote the equation $ab=\widehat{b}\,\widehat{a}$, and let~(E2) denote the equation $\widehat{w}=\theta w \theta^{-1}$, where $\theta\in \mathbb{F}_2(u,v)$. Note that if $a$ and $b$ satisfy~(E1) then both $a$ and $b$ are non trivial, and that $\widehat{b}\,\widehat{a}$ is also in reduced form. Further, if $z\in \mathbb{F}_2(u,v)$, then since $|z|_{y}=-|\widehat{z}|_{y}$ for $y\in\brak{u,v}$, it follows from the form of~(E1) and~(E2) that $k$ must be even in both cases. We carry out the proof of the two implications by induction as follows.
\begin{enumerate}[(i)] \item If $k\leq 4$ then (E1) implies~\reqref{concl1}. One may check easily that $k$ cannot be equal to $2$. Suppose that $k=4$ and that~(E1) holds. Now $\ell(a)\neq 1$, for otherwise $b$ would start with $a^{-1}$, but then $ab$ would not be reduced. Similarly, $\ell(b)\neq 1$, so we must have $\ell(a)=\ell(b)=2$, in which case $b=\widehat{a}$, and it suffices to take $r=s=0$ and $\lambda=b$ in~\reqref{concl1}.
\item If $k\leq 4$ then (E2) implies~\reqref{concl2}. Once more, it is straightforward to see that $k$ cannot be equal to $2$. Suppose that $k=4$. Since $w$ and $\widehat{w}$ are conjugate, we have $|w|_{y}=|\widehat{w}|_{y}$ for $y\in\brak{u,v}$, and so $|w|_{y}=0$. Since $w$ is in reduced form, one may then check that~\reqref{concl2} holds, where $\ell(\lambda)=2$.
\item\label{it:induct} Suppose by induction that for some $k\geq 4$,~(E1) implies~\reqref{concl1} if $\ell(ab)<k$ and~(E2) implies~\reqref{concl2} if $\ell(w)<k$. Suppose that $a$ and $b$ satisfy~(E1) and that $\ell(ab)=k$. If $\ell(a)=\ell(b)$ then $b=\widehat{a}$, and as above, it suffices to take $r=s=0$ and $\lambda=b$ in~\reqref{concl1}. So assume that $\ell(a)\neq \ell(b)$. By applying the automorphism $\widehat{\cdot}$ and exchanging the r\^{o}les of $a$ and $b$ if necessary, we may suppose that $\ell(a)<\ell(b)$. We consider two subcases.
\begin{enumerate}[(A)] \item $\ell(a)\leq \ell(b)/2$: since both sides of~(E1) are in reduced form, $b$ starts and ends with $\widehat{a}$, and there exists $b_{1}\in \mathbb{F}_2(u,v)$ such that $b=\widehat{a}b_{1}\widehat{a}$, written in reduced form. Substituting this into~(E1), we see that $a\widehat{a}b_{1}\widehat{a}= a\widehat{b}_{1}a\widehat{a}$, and thus $\widehat{a}b_{1}=\widehat{b}_{1}a$, written in reduced form. This equation is of the form~(E1), and since $\ell(\widehat{a}b_{1})<\ell(b)<\ell(ab)$, we may apply the induction hypothesis. Thus there exist $\lambda\in \mathbb{F}_{2}(u,v)$ and $r,s\in \ensuremath{\mathbb Z}$ such that $\widehat{a}=(\widehat{\lambda} \lambda)^s\widehat{\lambda}$, and $b_1=(\lambda\widehat{\lambda})^{r}\lambda$. Therefore $a=(\lambda\widehat{\lambda} )^s \lambda$ and \begin{equation*} b=\widehat{a} b_1 \widehat{a}=(\widehat{\lambda} \lambda)^s\widehat{\lambda} (\lambda\widehat{\lambda})^{r}\lambda (\widehat{\lambda} \lambda)^s\widehat{\lambda}= (\widehat{\lambda} \lambda)^{2s+r+1}\widehat{\lambda}, \end{equation*} using~\reqref{lambdahat}, which proves the result in this case.
\item $\ell(b)/2< \ell(a) <\ell(b)$: since both sides of~(E1) are in reduced form, $b$ starts with $\widehat{a}$, and there exists $b_{1}\in \mathbb{F}_2(u,v)$, $b_{1}\neq 1$, such that $b=\widehat{a}b_{1}$. Substituting this into~(E1), we obtain $a\widehat{a}b_{1}=a\widehat{b}_{1}\widehat{a}$, which is equivalent to $\widehat{a}b_{1}\widehat{a}^{-1}=\widehat{b}_{1}$. This equation is of the form~(E2), and since $\ell(\widehat{b}_{1})<\ell(b)<\ell(ab)=k$, we may apply the induction hypothesis. Thus there exist $\lambda\in \mathbb{F}_{2}(u,v)$ and $l\in \ensuremath{\mathbb Z}$ such that $b_1=(\lambda\widehat{\lambda})^{l}$. The fact that $b_{1}\neq 1$ implies that $\lambda\widehat{\lambda}\neq 1$ and $l\neq 0$. We claim that $\lambda\widehat{\lambda}$ may be chosen to be primitive. To prove the claim, suppose that $\lambda\widehat{\lambda}$ is not primitive. Since $\mathbb{F}_{2}(u,v)$ is a free group of rank $2$, the centraliser of $\lambda\widehat{\lambda}$ in $\mathbb{F}_{2}(u,v)$ is infinite cyclic, generated by a primitive element $v$, and replacing $v$ by $v^{-1}$ if necessary, there exists $s\geq 2$ such that $v^s= \lambda\widehat{\lambda}$. Therefore $b_{1}=v^{sl}$, and substituting this into the relation $\widehat{b}_{1}=\widehat{a}b_{1}\widehat{a}^{-1}$, we obtain $\widehat{v}^{sl}=\widehat{a} v^{sl}\widehat{a}^{-1}=(\widehat{a} v\widehat{a}^{-1})^{sl}$ in the free group $\mathbb{F}_{2}(u,v)$, from which we conclude that $\widehat{v}=\widehat{a} v\widehat{a}^{-1}$. We may thus apply the induction hypothesis to this relation because $\ell(v)<\ell(b_1)<k$, and since $v$ is primitive, there exists $\gamma\in \mathbb{F}_{2}(u,v)$ for which $v=\gamma\widehat{\gamma}$. Hence $b_{1}=(\gamma\widehat{\gamma})^{sl}$, where $\gamma\widehat{\gamma}$ is primitive, which proves the claim. Substituting $b_1=(\lambda\widehat{\lambda})^{l}$ into the relation $\widehat{a} b_1\widehat{a}^{-1}=\widehat{b}_1$, we obtain: \begin{equation*} (\widehat{a} \lambda\widehat{\lambda} \widehat{a}^{-1})^{l}=\widehat{a} (\lambda\widehat{\lambda})^{l}\widehat{a}^{-1}=\widehat{(\lambda\widehat {\lambda})^{l}}=(\widehat{\lambda} \lambda)^{l}, \end{equation*} where we take $\lambda\widehat{\lambda}$ to be primitive. Once more, since $\mathbb{F}_{2}(u,v)$ is a free group of rank $2$ and $l\neq 0$, it follows that $\widehat{a} \lambda\widehat{\lambda} \widehat{a}^{-1}=\widehat{\lambda} \lambda=\widehat{\lambda} \lambda \widehat{\lambda} \widehat{\lambda}^{-1}$, from which we conclude that $\widehat{\lambda}^{-1}\widehat{a}$ belongs to the centraliser of $\lambda\widehat{\lambda}$. But $\lambda\widehat{\lambda}$ is primitive, so there exists $t\in \ensuremath{\mathbb Z}$ such that $\widehat{\lambda}^{-1}\widehat{a}=(\lambda\widehat{\lambda})^{t}$, and hence $\widehat{a}=\widehat{\lambda} (\lambda\widehat{\lambda})^{t}$. Hence $a=\lambda (\widehat\lambda \lambda)^{t}=(\lambda \widehat\lambda)^{t}\lambda$ and $b=\widehat a b_1=\widehat{\lambda} (\lambda\widehat{\lambda})^{t+l}= (\widehat{\lambda}\lambda)^{t+l}\widehat{\lambda}$ in a manner similar to that of~\reqref{lambdahat}, so~\reqref{concl1} holds. \end{enumerate}
\item By the induction hypothesis and~(\ref{it:induct}), we may suppose that for some $k\geq 4$,~(E1) implies~\reqref{concl1} if $\ell(ab)\leq k$ and~(E2) implies~\reqref{concl2} if $\ell(w)<k$. Suppose that $\ell(w)=k$ and that $w$ and $\widehat{w}$ are conjugate. Let $\widehat{w}=\theta w\theta^{-1}$, where $\theta\in \mathbb{F}_{2}(u,v)$. If $\theta=1$ then $w=\widehat{w}$, which is impossible. So $\theta\neq 1$, and since $\ell(w)=\ell(\widehat{w})$, there must be cancellation in the expression $\theta w\theta^{-1}$. Taking the inverse of the relation $\widehat{w}=\theta w\theta^{-1}$ if necessary, we may suppose that cancellation occurs between $\theta$ and $w$. So there exist $\theta_1,\theta_2\in \mathbb{F}_{2}(u,v)$ such that $\theta=\theta_1\theta_2$ written in reduced form, and such that the cancellation between $\theta$ and $w$ is maximal \emph{i.e.}\ if $w_{1}=\theta_{2}w$ is written in reduced form then $\theta_{1}w_{1}$ is also reduced. Let $\ell(\theta)=n$ and $\ell(\theta_2)=r$. We again consider two subcases. \begin{enumerate}[(A)] \item Suppose first that $r=n$. Then $\theta_{1}=1$, $\theta_{2}=\theta$ and $w=\theta^{-1}w_{1}$, so: \begin{equation}\label{eq:ellwinequ} \ell(w)=\ell(\theta^{-1}w_{1})\leq \ell(\theta^{-1})+\ell(w_{1})=n+\ell(w)-n=\ell(w). \end{equation} Hence $\ell(\theta^{-1}w_{1})= \ell(\theta^{-1})+\ell(w_{1})$, from which it follows that $w=\theta^{-1}w_1$ is written in reduced form. Therefore $\widehat{w}=\widehat{\theta}^{-1}\widehat{w}_{1}$ is also written in reduced form. Now $\widehat{w}=\theta w \theta^{-1}=w_{1}\theta^{-1}$, and applying an inequality similar to that of~\reqref{ellwinequ}, we see that $\widehat{w}=w_{1}\theta^{-1}$ is written in reduced form. Hence $\widehat{\theta}^{-1}\widehat{w}_{1}=w_{1}\theta^{-1}$, which is in the form of~(E1), both sides being written in reduced form. Thus $\ell(w_{1}\theta^{-1})= \ell(\theta^{-1}w_1)=\ell(w)=k$, and by the induction hypothesis, there exist $\lambda\in \mathbb{F}_{2}(u,v)$ and $r,s\in \ensuremath{\mathbb Z}$ such that $w_1=(\widehat{\lambda} \lambda)^s\widehat{\lambda} $ and $\theta^{-1}=(\lambda\widehat{\lambda})^r\lambda$. So by~\reqref{lambdahat}, $w=\theta^{-1}w_1= (\lambda\widehat{\lambda})^r\lambda(\widehat{\lambda} \lambda)^s\widehat{\lambda} =( \lambda \widehat{\lambda})^{r+s+1}$, which proves the result in the case $r=n$.
\item Now suppose that $r<n$. Then there must be cancellation on both sides of $w$. Taking the inverse of both sides of the equation $\widehat{w}=\theta w\theta^{-1}$ if necessary, we may suppose that the length of the cancellation on the left is less than or equal to that on the right. So there exist $\theta_1,\theta_2,\theta_3, w_2\in \mathbb{F}_{2}(u,v)$ such that $\theta=\theta_1\theta_2\theta_3$, $w=\theta_3^{-1}w_2 \theta_2\theta_3$ and $\widehat{w}=\theta_1\theta_2 w_2 \theta_1^{-1}$, all these expressions being written in reduced form. Since $\ell(w)=\ell(\widehat{w})$, it follows from the second two expressions that $\ell(\theta_1)=\ell(\theta_3)$, and that $w=\theta_3^{-1}w_2 \theta_2\theta_3=\widehat{\theta}_1\widehat{\theta}_2 \widehat{w}_2 \widehat{\theta}_1^{-1}$, written in reduced form, from which we conclude that $\widehat{\theta}_1=\theta_3^{-1}$, and that $w_2 \theta_2=\widehat{\theta}_2 \widehat{w}_2$, written in reduced form, which is in the form of~(E1). Now $\ell(w_2 \theta_2)<\ell(w)=k$, and applying the induction hypothesis, there exist $\lambda\in \mathbb{F}_{2}(u,v)$ and $r,s\in \ensuremath{\mathbb Z}$ such that $w_2=(\widehat{\lambda} \lambda)^s\widehat{\lambda} $ and $\theta_2=(\lambda\widehat{\lambda})^r\lambda$. Hence $w=\theta_3^{-1}w_2 \theta_2\theta_3= \theta_3^{-1} (\widehat{\lambda} \lambda)^{r+s+1} \theta_3= (\theta_3^{-1}\widehat{\lambda} \widehat{\theta}_3 \ldotp\widehat{\theta}_3^{-1} \lambda \theta_3)^{r+s+1}=(\gamma \widehat{\gamma})^{r+s+1}$, where $\gamma=\theta_3^{-1}\widehat{\lambda} \widehat{\theta}_3$. This completes the proof of the induction step, and hence that of the lemma.\qedhere \end{enumerate} \end{enumerate} \end{proof}
\begin{prop}\mbox{}\label{prop:classif} \begin{enumerate}
\item\label{it:classifa} Let $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ be a split $2$-valued map, and let $\widehat{\Phi}\colon\thinspace \ensuremath{\mathbb{T}^{2}}\ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$ be a lift of $\phi$ that is determined by the pair $((w^r,(a,b)), (w^s, (c,d)))$ as described in \repr{exismaps}. Let $\mathcal{O}_{\phi}$ denote the set of conjugates of this pair by elements of $B_2(\ensuremath{\mathbb{T}^{2}})$. Then $\mathcal{O}_{\phi}$ is the union of the sets $\mathcal{O}_{\phi}^{(1)}$ and $\mathcal{O}_{\phi}^{(2)}$, where $\mathcal{O}_{\phi}^{(1)}$ is the subset of pairs of the form $((w_1^r,(a,b)), (w_1^s, (c,d)))$, where $w_1$ runs over the set of conjugates of $w$ in $\mathbb{F}_2(u,v)$, and $\mathcal{O}_{\phi}^{(2)}$ is the subset of pairs of the form $((w_2^r,(a+r|w|_{u},b+r|w|_{v})), (w_2^s, (c+s|w|_{u},d+s|w|_{v})))$, where $w_{2}$ runs over the set of conjugates of $\widehat{w}$ in $\mathbb{F}_2(u,v)$. Further, the correspondence that to $\phi$ associates $\mathcal{O}_{\phi}$ induces a bijection between the set of homotopy classes of split $2$-valued maps and the set of conjugates of the pairs of the form given by \repr{exismaps}(\ref{it:exismapsa}) by elements of $B_2(\ensuremath{\mathbb{T}^{2}})$.
\item\label{it:classifb} Let $f=(f_1,f_2)\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2(\ensuremath{\mathbb{T}^{2}})$ be a $2$-ordered map of $\ensuremath{\mathbb{T}^{2}}$ determined by the pair $((w^r,(a,b)), (w^s, (c,d)))$, let $g=(g_1,g_2)\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2(\ensuremath{\mathbb{T}^{2}})$, and let $\widehat{\pi}\colon\thinspace [\ensuremath{\mathbb{T}^{2}}, F_2(\ensuremath{\mathbb{T}^{2}})] \ensuremath{\longrightarrow} [\ensuremath{\mathbb{T}^{2}}, D_2(\ensuremath{\mathbb{T}^{2}})]$ be the projection defined in the proof of \relem{split}(\ref{it:splitIII}). Then $\widehat{\pi}^{-1}(\widehat{\pi}([f]))= \brak{[f],[g]}$. Further, $[f]=[g]$ if and only if there exist $\lambda\in \mathbb{F}_2(u,v)$ and $l\in \ensuremath{\mathbb Z}$ such that $w=(\lambda \widehat{\lambda})^l$. \end{enumerate} \end{prop}
\begin{proof}\mbox{} \begin{enumerate}[(a)] \item To compute $\mathcal{O}_{\phi}$, we determine the conjugates of the pair $((w^r,(a,b)), (w^s, (c,d)))$ by elements of $B_2(\ensuremath{\mathbb{T}^{2}})$, namely by words of the form $\sigma^{\ensuremath{\varepsilon}}z$, where $\ensuremath{\varepsilon} \in \brak{0,1}$, and $z\in P_2(\ensuremath{\mathbb{T}^{2}})$. With respect to the decomposition of \reco{compactpres}, if $\ensuremath{\varepsilon}=0$, we obtain the elements of $\mathcal{O}_{\phi}^{(1)}$. If $\ensuremath{\varepsilon}=1$, using the computation for $\ensuremath{\varepsilon}=0$ and the fact that
$\sigma z^{m} \sigma^{-1}=(uv^{-1}\widehat{z}^{m} vu^{-1}, (m|z|_{u}, m|z|_{v}))$ for all $z\in \mathbb{F}_{2}(u,v)$ and $m\in \ensuremath{\mathbb Z}$ by \relem{conjhat},
we obtain the elements of $\mathcal{O}_{\phi}^{(2)}$. This proves the first part of the statement. The second part is a consequence of classical general facts about maps between spaces of type $K(\pi, 1)$, see~\cite[Chapter~V, Theorem~4.3]{Wh} for example.
\item Let $\alpha\in [\ensuremath{\mathbb{T}^{2}}, F_2(\ensuremath{\mathbb{T}^{2}})]$ and let $f=(f_1,f_2)\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2(\ensuremath{\mathbb{T}^{2}})$ be such that $f\in \alpha$. Taking $g=(f_2,f_1)\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2(\ensuremath{\mathbb{T}^{2}})$, under the projection $\widehat{\pi}\colon\thinspace [\ensuremath{\mathbb{T}^{2}}, F_2(\ensuremath{\mathbb{T}^{2}})] \ensuremath{\longrightarrow} [\ensuremath{\mathbb{T}^{2}}, D_2(\ensuremath{\mathbb{T}^{2}})]$, $\widehat{\pi}(\beta)=\widehat{\pi}(\alpha)$, where $\beta=[g]$. From \resec{relnsnvm}, $\alpha$ and $\beta$ are the only elements of $[\ensuremath{\mathbb{T}^{2}}, F_2(\ensuremath{\mathbb{T}^{2}})]$ that project under $\widehat{\pi}$ to $\widehat{\pi}(\alpha)$, which proves the first part. It remains to decide whether $\alpha=\beta$. Suppose that $f$ is determined by the pair $P_1=((w^r,(a,b)), (w^s, (c,d)))$. Since $\widehat{\pi}(\beta)=\widehat{\pi}(\alpha)$, $g$ is determined by a pair belonging to $\mathcal{O}_{\phi}$. Using the fact that $g$ is obtained from $f$ via the $\ensuremath{\mathbb Z}_2$-action on $F_2(\ensuremath{\mathbb{T}^{2}})^{\ensuremath{\mathbb{T}^{2}}}$ that arises from the covering map $\pi$ and applying covering space arguments, there exists $g'\in \beta$ that is determined by the pair $P_2=((\widehat{w}^r,(a+r|w|_{u},b+r|w|_{v})), (\widehat{w}^s, (c+s|w|_{u},d+s|w|_{v})))$. Then $\alpha=\beta$ if and only if $P_1$ and $P_2$ are conjugate by an element of $P_2(\ensuremath{\mathbb{T}^{2}})$, which is the case if and only if $w$ and $\widehat{w}$ are conjugate in the free group $\mathbb{F}_2(u,v)$ (recall from the proof of \relem{conjhat1} that if $w$ and $\widehat{w}$ are conjugate then $|w|_{u}=|w|_{v}=0$). By part~(\ref{it:classifI}) of that lemma, this is equivalent to the existence of $\lambda\in \mathbb{F}_2(u,v)$ and $l\in \ensuremath{\mathbb Z}$ such that $w=(\lambda \widehat{\lambda})^l$.\qedhere \end{enumerate} \end{proof}
\subsection{Fixed point theory of split $2$-valued maps}\label{sec:fptsplit2}
In this section, we give a sufficient condition for a split $2$-valued map of $\ensuremath{\mathbb{T}^{2}}$ to be deformable to a fixed point free $2$-valued map. \repr{lifttorus} already provides one such condition. We shall give an alternative condition in terms of roots, which seems to provide a more convenient framework from a computational point of view. To obtain fixed point free $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$, we have two possibilities at our disposal: we may either use \reth{defchineg}(\ref{it:defchinegb}), in which case we should determine the group $P_{3}(\ensuremath{\mathbb{T}^{2}})$, or \reth{defchinegI}(\ref{it:defchinegbI}), in which case we may make use of the results of \resec{toro2}, and notably \repr{presTminus1alta}. We choose the second possibility. We divide the discussion into three parts. In \resec{proofexisfpf}, we prove \repr{exisfpf}. In \resec{p2Tminus1}, we give the analogue for roots of the second part of \repr{exisfpf}, and in \resec{exrfm}, we will give some examples of split $2$-valued maps that may be deformed to root-free $2$-valued maps.
\subsubsection{The proof of \repr{exisfpf}}\label{sec:proofexisfpf}
Let $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ be a $2$-valued map of $\ensuremath{\mathbb{T}^{2}}$. As in \reco{compactpres}, we identify $P_{2}(\ensuremath{\mathbb{T}^{2}})$ with $\mathbb{F}_{2} \times \ensuremath{\mathbb Z}^{2}$. The Abelianisation homomorphism is denoted by $\operatorname{\text{Ab}}\colon\thinspace \mathbb{F}_2(u,v) \ensuremath{\longrightarrow} \ensuremath{\mathbb Z}^2$. Recall from the beginning of \resec{relnsnvm} that if $\phi$ is split and $\widehat{\Phi}\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$ is a lift of $\phi$, then $\operatorname{\text{Fix}}(\widehat{\Phi})=\operatorname{\text{Fix}}(\phi)$. In this section, we compute the Nielsen number $N(\phi)$ of $\phi$, and we give necessary conditions for $\phi$ to be homotopic to a fixed point free $2$-valued map. Although the Nielsen number is not the main subject of this paper, it is an important invariant in fixed point theory. The Nielsen number of $n$-valued maps was defined in~\cite{Sch1}. The following result of that paper will enable us to compute $N(\phi)$ in our setting.
\begin{thm}[{\cite[Corollary~7.2]{Sch1}}]\label{th:helgath01} Let $n\in \ensuremath{\mathbb N}$, let $K$ be a compact polyhedron, and let $\phi=\{f_1,\ldots,f_n\}\colon\thinspace K \multimap K$ be a split $n$-valued map. Then $N(\phi)=N(f_1)+\cdots+N(f_n)$. \end{thm}
\begin{proof}[Proof of \repr{exisfpf}] Let $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ be a split $2$-valued map of $\ensuremath{\mathbb{T}^{2}}$, and let $\widehat{\Phi}=(f_1,f_2) \colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$ be a lift of $\phi$ such that $\widehat{\Phi}_{\#}(e_{1})=(w^r,(a,b))$ and $\widehat{\Phi}_{\#}(e_{2})= (w^s, (c,d)))$, where $(r,s)\in \ensuremath{\mathbb Z}^{2}\setminus \brak{(0,0)}$, $a,b,c,d\in \ensuremath{\mathbb Z}$ and $w\in \mathbb{F}_2(u,v)$. For $i=1,2$, we shall compute the matrix $M_{i}$ of the homomorphism $f_{i\#}\colon\thinspace \ensuremath{\mathbb Z}^{2} \ensuremath{\longrightarrow} \ensuremath{\mathbb Z}^{2}$ induced by $f_i$ on the fundamental group of $\ensuremath{\mathbb{T}^{2}}$ with respect to the basis $(e_{1},e_{2})$ of $\pi_1(\ensuremath{\mathbb{T}^{2}})$ (up to the canonical identification of $\pi_1(\ensuremath{\mathbb{T}^{2}})$ for different basepoints if necessary). In practice, the bases in the target are the images of the elements $(\rho_{1,1},\rho_{1,2})$ and $(\rho_{2,1},\rho_{2,2})$ by the homomorphism $p_{i\#}\colon\thinspace P_2(\ensuremath{\mathbb{T}^{2}}) \ensuremath{\longrightarrow} \pi_{1}(\ensuremath{\mathbb{T}^{2}})$ induced by the projection $p_i\colon\thinspace F_2(\ensuremath{\mathbb{T}^{2}}) \ensuremath{\longrightarrow} \ensuremath{\mathbb{T}^{2}}$ onto the $i\up{th}$ coordinate, where $i=1,2$. Note that $f_{i\#}=p_{i\#}\circ \widehat{\Phi}_{\#}$. Setting $w=w(u,v)$, and using multiplicative notation and \req{uvxy}, we have: \begin{align*} \widehat{\Phi}_{\#}(e_{1})&=w^r x^{a} y^{b}=(w(\rho_{1,1},\rho_{1,2}))^r (\rho_{1,1}B_{1,2}^{-1}\rho_{2,1})^{a} (\rho_{1,2}B_{1,2}^{-1}\rho_{2,2})^{b}\\ \widehat{\Phi}_{\#}(e_{2})&=w^s x^{c} y^{d}=(w(\rho_{1,1},\rho_{1,2}))^s (\rho_{1,1}B_{1,2}^{-1}\rho_{2,1})^{c} (\rho_{1,2}B_{1,2}^{-1}\rho_{2,2})^{d}. \end{align*} Projecting onto the first (resp.\ second) coordinate, it follows that $f_{1\#}(e_{1})=\rho_{1,1}^{rm+a}\rho_{1,2}^{rn+b}$ and $f_{1\#}(e_{2})=\rho_{1,1}^{sm+c}\rho_{1,2}^{sn+d}$ (resp.\ $f_{2\#}(e_{1})=\rho_{2,1}^{a}\rho_{2,2}^{b}$ and $f_{2\#}(e_{2})=\rho_{2,1}^{c}\rho_{2,2}^{d}$), so $M_{1} =\left( \begin{smallmatrix}
rm+a & sm+c \\
rn+b & sn+d \end{smallmatrix}\right)$ and $M_{2} =\left( \begin{smallmatrix}
a & c \\
b & d \end{smallmatrix}\right)$. One then obtains the equation for $N(\phi)$ as a consequence of \reth{helgath01} and the usual formula for the Nielsen number of a self-map of $\ensuremath{\mathbb{T}^{2}}$~\cite{BrooBrPaTa}. The second part of the statement is clear, since if $\phi$ can be deformed to a fixed point free $2$-valued map then $f_1, f_2$ can both be deformed to fixed point free maps. To prove the last part, $f_1$ and $f_2$ can both be deformed to fixed point free maps if and only if $\det(M_{i}-I_{2})=0$ for $i=1,2$, which using linear algebra is equivalent to: \begin{gather} \text{$\det(M_{2}-I_{2})=0$, and}\label{eq:onedet}\\ \det\begin{pmatrix}
a-1 & sm \\
b & sn \end{pmatrix}+ \det\begin{pmatrix}
rm & c \\
rn & d-1 \end{pmatrix}=0.\label{eq:twodet} \end{gather} Equation~\reqref{onedet} is equivalent to the proportionality of $(c,d-1)$ and $(a-1,b)$. Suppose that \req{twodet} holds. If one of the determinants in that equation is zero, then so is the other, and it follows that $(a-1, b),(c,d-1)$ and $(m,n)$ generate a subgroup of $\ensuremath{\mathbb Z}^{2}$ isomorphic to $\ensuremath{\mathbb Z}$, which yields condition~(\ref{it:exisfpfa}) of the statement. If both of these determinants are non zero then $(m,n)$ is neither proportional to $(a-1,b)$ nor to $(c,d-1)$, and since $(c,d-1)$ and $(a-1,b)$ are proportional, $(m,n)$ is not proportional to any linear combination of the two. Further,~\reqref{twodet} may be written as: \begin{equation*} 0=s\det\begin{pmatrix}
a-1 & m \\
b & n \end{pmatrix}+ r\det\begin{pmatrix}
m & c \\
n & d-1 \end{pmatrix}= \det\begin{pmatrix} s(a-1)-rc & m \\ sb-r(d-1) & n \end{pmatrix}, \end{equation*} from which it follows that $s(a-1, b)=r(c,d-1)$, which is condition~(\ref{it:exisfpfb}) of the statement. The converse is straightforward. \end{proof}
\begin{rem} Within the framework of \repr{exisfpf}, the fact that $f_1$ and $f_2$ can be deformed to fixed point free maps does not necessarily imply that there exists a deformation of the pair $(f_1, f_2)$, regarded as a map from $\ensuremath{\mathbb{T}^{2}}$ to $F_2(\mathbb{T}^{2})$, to a pair $(f_1', f_2')$ where the maps $f_1'$ and $f_2'$ are fixed point free. To answer the question of whether the $2$-ordered map $(f_1, f_2)\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2(\ensuremath{\mathbb{T}^{2}})$ can be deformed or not to a fixed point free $2$-ordered map under the hypothesis that each map can be deformed to a fixed point free map would be a major step in understanding the fixed point theory of $n$-valued maps, and would help in deciding whether the Wecken property holds or not for $\ensuremath{\mathbb{T}^{2}}$ for the class of split $n$-valued maps. \end{rem}
\subsubsection{Deformations to root-free $2$-valued maps}\label{sec:p2Tminus1}
Recall from the introduction that a root of an $n$-valued map $\phi_0\colon\thinspace X \multimap Y$, with respect to a basepoint $y_0\in Y$, is a point $x$ such that $y_0\in \phi_0(x)$. If $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ is an $n$-valued map then we may construct another $n$-valued map $\phi_0\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ as follows. If $x\in \ensuremath{\mathbb{T}^{2}}$ and $\phi(x)=\{ x_1,\ldots,x_n\}$, then let $\phi_0(x)=\{x_1\ldotp x^{-1},\ldots,x_n \ldotp x^{-1}\}$. The correspondence that to $\phi$ associates $\phi_{0}$ is bijective. Moreover, if $\phi$ is split, so that $\phi=\{f_1, f_2,\ldots, f_n\}$, where the self-maps $f_{i}\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} \ensuremath{\mathbb{T}^{2}}$, $i=1,\ldots,n$, are coincidence-free, then $\phi_{0}$ is also split, and is given by $\phi_{0}(x)=\{f_1(x)\ldotp x^{-1},\ldots, f_n(x)\ldotp x^{-1}\}$ for all $x\in \ensuremath{\mathbb{T}^{2}}$. The restriction of the above-mentioned correspondence to the case where the $n$-valued maps are split is also a bijection. The following lemma implies that the question of deciding whether an $n$-valued map $\phi$ can be deformed to a fixed point free map is equivalent to deciding whether the associated map $\phi_{0}$ can be deformed to a root-free map. Let $1$ denote the basepoint of $\ensuremath{\mathbb{T}^{2}}$.
\begin{lem}\label{lem:equivroot} With the above notation, a point $x_0\in \ensuremath{\mathbb{T}^{2}}$ is a fixed point of an $n$-valued map $\phi$ of $\ensuremath{\mathbb{T}^{2}}$ if and only if it is a root of the $n$-valued map $\phi_0$ (i.e.\ $1\in \phi(x_0)$). Further, $\phi$ may be deformed to an $n$-valued map $\phi'$ such that $\phi'$ has $k$ fixed points if and only if $\phi_0$ may be deformed to an $n$-valued map $\phi_0'$ such that $\phi_0'$ has $k$ roots. \end{lem}
\begin{proof} Straightforward, and left to the reader. \end{proof}
The algebraic condition given by \reth{defchinegI}(\ref{it:defchinegbI}) is equivalent to the existence of a homomorphism $g_{\#}\colon\thinspace \pi_1(\ensuremath{\mathbb{T}^{2}}) \ensuremath{\longrightarrow} P_{2}(\ensuremath{\mathbb{T}^{2}})$ that factors through $P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\})$. The following result is the analogue for roots of the second part of \repr{exisfpf}.
\begin{prop}\label{prop:exisrf} Let $g\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ be a split $2$-valued map, and let $\widehat{g}=(g_1, g_2)\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_{2}(\ensuremath{\mathbb{T}^{2}})$ be a lift of $g$ such that $\widehat{g}_{\#}(e_{1})=(w^{r},(a',b))$ and $\widehat{g}_{\#}(e_{2})=(w^{s}, (c,d'))\in P_{2}(\ensuremath{\mathbb{T}^{2}})$, where $(r,s)\in \ensuremath{\mathbb Z}^{2}\setminus \brak{(0,0)}$, $a',b,c,d'\in \ensuremath{\mathbb Z}$, $w\in \mathbb{F}_2(u,v)$ and $\operatorname{\text{Ab}}(w)=(m,n)$.
If $g$ can be deformed to a root-free map, then each of the maps $g_1, g_2\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} \ensuremath{\mathbb{T}^{2}}$ can be deformed to a root-free map. Further, $g_1$ and $g_2$ can both be deformed to root-free maps if and only if either: \begin{enumerate} \item\label{it:exisrfa} the pairs $(a',b),(c,d')$ and $(m,n)$ belong to a cyclic subgroup of $\ensuremath{\mathbb Z}^2$, or \item\label{it:exisrfb} $s(a',b)=r(c,d')$. \end{enumerate} \end{prop}
\begin{proof} If $g$ can be deformed to a root-free map then clearly the maps $g_1, g_2\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} \ensuremath{\mathbb{T}^{2}}$ can be deformed to root-free maps. For the second part of the statement, for $i=1,2$, consider the maps $f_{i}\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} \ensuremath{\mathbb{T}^{2}}$, where $f_i(x)=g_i(x)\ldotp x$, and where $g_1$ and $g_2$ can both be deformed to root-free maps. Then $f_1$ and $f_2$ can be deformed to fixed point free maps, and the maps $f_1, f_2$ are determined by the elements $(w^r, (a,b))$, $(w^s, (c,d))$ of $P_2(\ensuremath{\mathbb{T}^{2}})$, where $a=a'+1$ and $d=d'+1$. By \repr{exisfpf}, either the elements $(a-1,b),(c,d-1)$ and $(m,n)$ belong to a cyclic subgroup of $\ensuremath{\mathbb Z}^2$, or $s(a-1,b)=r(c,d-1)$, which is the same as saying that either the elements $(a',b),(c,d')$ and $(m,n)$ belong to a cyclic subgroup of $\ensuremath{\mathbb Z}^2$, or $s(a',b)=r(c,d')$, and the result follows. \end{proof}
\subsubsection{Examples of split $2$-valued maps that may be deformed to root-free $2$-valued maps}\label{sec:exrfm}
We now give a family of examples of split $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$ that satisfy the necessary condition of \repr{exisrf} for such a map to be deformable to a root-free map. To do so, we exhibit a family of $2$-ordered maps that we compose with the projection $\pi\colon\thinspace F_2(\ensuremath{\mathbb{T}^{2}}) \ensuremath{\longrightarrow} D_2(\ensuremath{\mathbb{T}^{2}})$ to obtain a family of split $2$-valued maps. We begin by studying a $2$-ordered map of $\ensuremath{\mathbb{T}^{2}}$ determined by a pair of braids of the form $((w^{r},(a,b))$, $ (w^{s}, (c,d))$, where $s(a,b)=r(c,d)$ (we make use of the notation of \repr{exisrf}).
\begin{prop}\label{prop:necrootfree2} If $\widehat{\Phi}\colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2({\ensuremath{\mathbb{T}^{2}}})$ is a lift of a split $2$-valued map $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ that satisfies $\widehat{\Phi}_{\#}(e_{1})=(w^{r},(a,b))$ and $\widehat{\Phi}_{\#}(e_{2})= (w^{s}, (c,d))$, where $w\in \mathbb{F}_2(u,v)$, $a,b,c,d\in \ensuremath{\mathbb Z}$ and $(r,s)\in \ensuremath{\mathbb Z}^{2}\setminus \brak{(0,0)}$ satisfy $s(a,b)=r(c,d)$, then $\phi$ may be deformed to a root-free $2$-valued map. \end{prop}
\begin{proof} By hypothesis, the subgroup $\Gamma$ of $\ensuremath{\mathbb Z}^2$ generated by $(a,b)$ and $(c,d)$ is contained in a subgroup isomorphic to $\ensuremath{\mathbb Z}$. Let $\gamma$ be a generator of $\Gamma$. Suppose first that $r$ and $s$ are both non zero. Then we may take $\gamma=(a_0, b_0)=(\ell/r)(a,b)= (\ell/s)(c,d)$, where $\ell=\gcd(r,s)$, and the elements $(w^{r},(a,b))$ and $(w^{s}, (c,d))$ belong to the subgroup of $P_{2}(\ensuremath{\mathbb{T}^{2}})$ generated by $(w^{\ell},(a_0,b_0))$. Let $z\in P_2(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ be an element that projects to $(w^{\ell},(a_0,b_0))$ under the homomorphism $\alpha\colon\thinspace P_2(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})\ensuremath{\longrightarrow} P_2(\ensuremath{\mathbb{T}^{2}})$ induced by the inclusion $\ensuremath{\mathbb{T}^{2}}\setminus\brak{1} \ensuremath{\longrightarrow} \ensuremath{\mathbb{T}^{2}}$. The map $\varphi\colon\thinspace \pi_1(\ensuremath{\mathbb{T}^{2}})\ensuremath{\longrightarrow} P_2(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ defined by $\varphi(e_1)=z^{r/\ell}$ and $\varphi(e_2)=z^{s/\ell}$ extends to a homomorphism, and is a lift of $\widehat{\Phi}_{\#}$. The result in this case follows by \reth{defchinegI}(\ref{it:defchinegbI}). Now suppose thar $r=0$ (resp.\ $s=0$). Then $(a,b)=(0,0)$ (resp.\ $(c,d)=(0,0)$) and $(w^{r},(a,b))$ (resp.\ $(w^{s}, (c,d))$) is trivial in $P_{2}(\ensuremath{\mathbb{T}^{2}})$. Let $z\in P_2(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ be an element that projects to $(w^{s},(c,d))$ (resp.\ to $(w^{r},(a,b))$). Then we define $\varphi(e_1)=1$ and $\varphi(e_2)=z$ (resp.\ $\varphi(e_1)=z$ and $\varphi(e_2)=1$), and once more the result follows. \end{proof}
\begin{proof}[Proof of \reth{necrootfree3}] This follows directly from \repr{necrootfree2} and the relation between the fixed point and root problems described by \relem{equivroot}. \end{proof}
\begin{lem}\label{lem:exfreero} Let $k,l\in \ensuremath{\mathbb Z}$ and suppose that either $p\in \{0,1\}$ or $q\in \{0,1\}$. With the notation of \repr{presTminus1alta}, the elements $(x^py^q)^k$ and $(u^pv^q)^l$ of $P_2(\ensuremath{\mathbb{T}^{2}}\setminus\{1\}; (x_1,x_2))$ commute. \end{lem}
\begin{proof} We will make use of \repr{presTminus1alta} and some of the relations obtained in its proof. If $p$ or $q$ is zero then the result follows easily. So it suffices to consider the two cases $p=1$ and $q=1$. By \req{conjxyuv}, for $\ensuremath{\varepsilon}\in \brak{1,-1}$, we have $x^{\ensuremath{\varepsilon}}v x^{-\ensuremath{\varepsilon}}=u^{\ensuremath{\varepsilon}}v (u^{-1}B_{1,2})^{\ensuremath{\varepsilon}}$ and $y^{\ensuremath{\varepsilon}}u y^{-\ensuremath{\varepsilon}}=(vB_{1,2}^{-1})^{\ensuremath{\varepsilon}} uv^{-\ensuremath{\varepsilon}}$, and by induction on $r$, it follows that $x^{r} v x^{-r}=u^{r} v (u^{-1}B_{1,2})^{r}$ and $y^{r}u y^{-r}=(vB_{1,2}^{-1})^{r} uv^{-r}$ for all $r\in \ensuremath{\mathbb Z}$. So if $p=1$ or $q=1$ then we have respectively: \begin{align*} xy^{q} uv^{q} y^{-q} x^{-1} &= x (vB_{1,2}^{-1})^{q} uv^{-q} \ldotp v^{q}x^{-1}= uv^{q}u^{-1} u =uv^{q}, \;\text{and}\\ x^{p}y u^{p}v y^{-1} x^{-p} &= x^{p} (yvy^{-1})^{p} vx^{-p}= x^{p} v(B_{1,2}^{-1}u)^{p} v^{-1} \ldotp v x^{-p}= u^{p} v (u^{-1}B_{1,2})^{p}(B_{1,2}^{-1}u)^{p}=u^{p} v, \end{align*} as required. \end{proof}
This enables us to prove the following proposition and \reth{construct2val}.
\begin{prop}\label{prop:construct2valprop} Suppose that $(a, b),(c,d)$ and $(m,n)$ belong to a cyclic subgroup of $\ensuremath{\mathbb Z}^2$ generated by an element of the form $(0,q), (1,q), (p,0)$ or $(p,1)$, where $p,q\in \ensuremath{\mathbb Z}$, and let $r,s\in \ensuremath{\mathbb Z}$. Then there exist $w\in \mathbb{F}_2(u,v)$, a split $2$-valued map $\phi\colon\thinspace \ensuremath{\mathbb{T}^{2}} \multimap \ensuremath{\mathbb{T}^{2}}$ and a lift $\widehat{\Phi} \colon\thinspace \ensuremath{\mathbb{T}^{2}} \ensuremath{\longrightarrow} F_2(\ensuremath{\mathbb{T}^{2}})$ of $\phi$ for which $\operatorname{\text{Ab}}(w)=(m,n)$, $\widehat{\Phi}_{\#}(e_1)=((w^{r},(a,b))$ and $\widehat{\Phi}_{\#}(e_2)= (w^{s}, (c,d))$, and such that $\phi$ can be deformed to a root-free $2$-valued map. \end{prop}
\begin{proof} Once more we apply \reth{defchinegI}(\ref{it:defchinegbI}). Let $(p,q)$ be a generator of the cyclic subgroup given in the statement. So there exist $\lambda_{1}, \lambda_{2}, \lambda_{3}\in \ensuremath{\mathbb Z}$ such that $(a, b)=\lambda_1(p,q)$, $(c,d)=\lambda_2(p,q)$ and $(m,n)=\lambda_3(p,q)$. We define $\varphi\colon\thinspace \pi_1(\ensuremath{\mathbb{T}^{2}})\ensuremath{\longrightarrow} P_2(\ensuremath{\mathbb{T}^{2}}\setminus\brak{1})$ by $\varphi(e_1)=(u^pv^q)^{\lambda_3r}(x^py^q)^{\lambda_1}$ and $\varphi(e_2)=(u^pv^q)^{\lambda_3s}(x^py^q)^{\lambda_2}$. \relem{exfreero} implies that $\varphi$ extends to a well-defined homomorphism, and we may take $w=(u^pv^q)^{\lambda_3}$. \end{proof}
\begin{proof}[Proof of \reth{construct2val}] This follows directly from \repr{construct2valprop} and the relation between the fixed point and root problems described in \relem{equivroot}. \end{proof}
\begin{rem} \reth{construct2val} implies that there is an infinite family of homotopy classes of $2$-valued maps of $\ensuremath{\mathbb{T}^{2}}$ that satisfy the necessary condition of \repr{exisrf}(\ref{it:exisrfa}), and can be deformed to root-free maps. We do not know whether there exist examples of maps that satisfy this condition but that cannot be deformed to root-free, however it is likely that such examples exist. \end{rem}
\section*{Appendix: Equivalence between $n$-valued maps and maps into configuration spaces}
This appendix constitutes joint work with R.~F.~Brown. Let $n\in \ensuremath{\mathbb N}$. As observed in \resec{intro}, the set of $n$-valued functions from $X$ to $Y$ is in one-to-one correspondence with the set of functions from $X$ to $D_n(Y)$. As we have seen in the main part of this paper, this correspondence facilitates the study of $n$-valued maps, and more specifically of their fixed point theory. In this appendix, we prove \reth{metriccont} that clarifies the topological relationship preserved by the correspondence under some mild hypotheses on $X$ and $Y$. For the sake of completeness, we will include the proof of a simple fact (\repr{multicont}) mentioned in~\cite{Sch0} that relates the splitting of maps and the continuity of multifunctions.
Given a metric space $Y$, let $\mathcal K'$ be the family of non-empty compact sets of $Y$. We equip $\mathcal K'$ with the topology induced by the Hausdorff metric on $\mathcal K'$ defined in~\cite[Chapter~VI, Section~6]{Be}.
\begin{thm}[{\cite[Chapter~VI, Section~6, Theorem~1]{Be}}]\label{th:berge} Let $X$ and $Y$ be metric spaces, let $\mathcal K'$ denote the family of non-empty compact sets of $Y$, let $\Gamma\colon\thinspace X \multimap Y$ be a multifunction such that for all $x\in X$, $\Gamma(x)\in \mathcal K'$ and $\Gamma(x)\ne \varnothing$. Then $\Gamma$ is continuous if and only if it is a single-valued continuous mapping from $X$ to $\mathcal K'$. \end{thm}
As we mentioned in \resec{intro}, $F_n(Y)$ may be equipped with the topology induced by the inclusion of $F_{n}(Y)$ in $Y^n$, and $D_n(Y)$ may be equipped with the quotient topology using the quotient map $\pi\colon\thinspace F_n(Y) \ensuremath{\longrightarrow} D_n(Y)$, a subset $W$ of $D_n(Y)$ being open if and only if $\pi^{-1}(W)$ is open in $F_n(Y)$. If $Y$ is a metric space with metric $d$, the set $D_n(Y)$ is a subset of $\mathcal K'$, and the Hausdorff metric on $\mathcal K'$ mentioned above restricts to a Hausdorff metric $d_H$ on $D_n(Y)$ defined as follows. If $z,w \in D_n(Y)$ then there exist $(z_1, \ldots , z_n), (w_1, \ldots , w_n)\in F_n(Y)$ such that $z=\pi(z_1, \ldots , z_n)$ and $w=\pi(w_1, \ldots , w_n)$, and
we define $d_H$ by: \begin{equation*} d_H(z,w)=\max\Bigl(\max_{1\leq i\leq n} d(z_i,w), \max_{1\leq i\leq n} d(w_i,z) \Bigr), \end{equation*} where $\displaystyle d(z_i,w)=\min_{1\leq j\leq n} d(z_i,w_j)$ for all $1\leq i\leq n$. Notice that $d_H(z,w)$ does not depend on the choice of representatives in $F_n(Y)$. We now prove \reth{metriccont}.
\begin{proof}[Proof of \reth{metriccont}] By \reth{berge}, it suffices to show that the set $D_n(Y)$ equipped with the Hausdorff metric $d_{H}$ is homeomorphic to the unordered configuration space $D_n(Y)$ equipped with the quotient topology, or equivalently, to show that a subset of $D_n(Y)$ is open with respect to the Hausdorff metric topology if and only if it is open with respect to the quotient topology. Let $y\in D_n(Y)$, and let $(y_1,\ldots,y_n) \in F_n(Y)$ be such that $\pi(y_1,\ldots,y_n)=y$.
For the `if' part, let $U_1,\ldots, U_n$ be open balls in $Y$ whose centres are $y_1,\ldots,y_n$ respectively. Without loss of generality, we may assume that they have the same radius $\ensuremath{\varepsilon}>0$, and are pairwise disjoint. Consider the Hausdorff ball $U_{H}$ of radius $\ensuremath{\varepsilon}$ in $D_n(Y)$ whose centre is $y$.
Let $z$ be an element of $U_{H}$, and let $(z_1,\ldots,z_n) \in F_n(Y)$ be such that $\pi(z_1,\ldots,z_n)=z$. Suppose that $z \notin \pi(U_1 \times \cdots \times U_n)$. We argue for a contradiction. Then there exists a ball $U_{i}$ such that $z_j \notin U_i$ for all $j\in \brak{1,\ldots,n}$.
So $d(y_i,z)\geq \ensuremath{\varepsilon}$, and from the definition of $d_H$, it follows that $d_H(y,z)\geq \ensuremath{\varepsilon}$, which contradicts the choice of $z$. Hence $z\in \pi(U_1 \times \cdots \times U_n)$, and the `if' part follows.
For the `only if' part, let us consider an open ball $U_{H}$ of radius $\ensuremath{\varepsilon}>0$ in the Hausdorff metric $d_H$ whose centre is $y$.
We will show that there are open balls $U_1, \ldots ,U_n$ in $Y$ whose centres are $y_1, \ldots , y_n$ respectively, such that the subset of elements $z$ of $D_{n}(Y)$, where $z=\pi(z_1,\ldots,z_n)$ for some $(z_1,\ldots,z_n) \in F_n(Y)$, and where
each $U_i$ contains exactly one point of $\brak{z_1,\ldots,z_n}$, is a subset of $U_{H}$. Define $\delta>0$ to be the minimum of $\ensuremath{\varepsilon}$ and the distances $d(y_i, y_j)/2$ for all $i \ne j$. For $j = 1, \dots , n$ let $U_{j}$ be the open ball in $Y$ of radius $\delta$ with respect to $d$ and whose centre is $y_j$. Clearly, $U_i \cap U_j = \varnothing$ for all $i \ne j$. So for all $z=\brak{z_1, \dots , z_n}$ belonging to the set $\pi(U_1 \times \cdots \times U_n)$, each $U_j$ contains exactly one point of $\brak{z_1,\ldots,z_n}$, which up to permuting indices, we may suppose to be $z_j$. Further, for all $j = 1, \dots , n$, $d(z_j,y)=d(z_j,y_j)<\delta$ and $d(y_j,z)=d(y_j,z_j)<\delta$. So from the definition of $d_H$, $d_H(y,z)<\delta<\ensuremath{\varepsilon}$, hence $z\in U_H$, and the `only if' part follows. \end{proof}
Just above Lemma~1 of \cite[Section~2]{Sch0}, Schirmer wrote `Clearly a multifunction which splits into maps is continuous'. For the sake of completeness, we provide a short proof of this fact.
\begin{prop}\label{prop:multicont} Let $n\in \ensuremath{\mathbb N}$, let $X$ be a topological space, and let $Y$ be a Hausdorff topological space. For $i=1,\ldots,n$, let $f_i\colon\thinspace X \ensuremath{\longrightarrow} Y$ be continuous. Then the split $n$-valued map $\phi= \brak{f_1,\ldots,f_n}\colon\thinspace X \multimap Y$ is continuous. \end{prop}
\begin{proof} Let $x_0 \in X$, and let $V$ be an open subset of $Y$ such that $\phi(x_0) \cap V\ne \varnothing$. Then there exists $j\in \brak{1,\ldots,n}$ such that $f_j(x_0) \in V$. Since $f_j$ is continuous, there exists an open subset $U_j$ containing $x_0$ such that if $x \in U_j$ then $f_j(x) \in V$, so $\phi(x)\cap V\ne \varnothing$. Therefore $\phi$ is lower semi-continuous. To prove upper semi-continuity, first note that for all $x\in X$, $\phi(x)$ is closed in $Y$ because $\phi(x)$ is a finite set and $Y$ is Hausdorff. Now let $x_0 \in X$ be such that $\phi(x_0) \subset V$. We then use the continuity of the $f_j$ to define the $U_j$ as before, and we set $U = \bigcap_{j=1}^n U_j$. Since $\phi(U) \subset V$, we have proved that $\phi$ is also upper semi-continuous, and so it is continuous. \end{proof}
\end{document}
|
arXiv
|
{
"id": "1702.05016.tex",
"language_detection_score": 0.6290969848632812,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{abstract} We characterize one-dimensional compact repellers having non-concave Lyapunov spectra. For linear maps with two branches we give an explicit condition that characterizes non-concave Lyapunov spectra. \end{abstract}
\title{The Lyapunov spectrum is not always concave} \section{Introduction} The rigorous study of the Lyapunov spectrum finds its roots in the pioneering work of H. Weiss \cite{w1}. For the purpose of simplicity we restrict our discussion to a dynamical system $T : \Lambda \to \Lambda$ where $\Lambda$ is a compact subset of an interval. The Lyapunov exponent $\lambda (x)$ of $T$ at $x$ is
$$\lambda (x) := \lim_{n \to \infty} \frac{1}{n} \log |(T^n)'(x)|,$$ whenever this limit exists. The Lyapunov spectrum $L$ encodes the decomposition of the phase space $\Lambda$ into level sets $J_\alpha$ of the Lyapunov exponent (i.e. the set where $\lambda(x) = \alpha$). More precisely, $L$ is the function that assigns to each $\alpha$ (in an appropriate interval) the Hausdorff dimension of $J_\alpha$. Weiss proved that the Lyapunov spectrum is real analytic. A rather surprising result in light of the fact that this decomposition is fairly complicated. For instance, each level set turns out to be dense in $\Lambda$. Relying on his previous joint results with Pesin~\cite{pw1}, Weiss studied the Lyapunov spectrum via the dimension spectrum of the measure of maximal entropy. It is worth to mention that not only the analyticity of $L$ is obtained with this approach but it also gives, for each $\alpha$, an equilibrium measure such that $J_\alpha$ has full measure and the Hausdorff dimension of the measure coincides with that of $J_\alpha$.
In between the wealth of novel and correct results about Lyapunov spectra briefly summarized above, the aforementioned paper~\cite{w1} unfortunately contains a claim which is not fully correct. Namely, that compact conformal repellers have concave Lyapunov spectra~\cite[Theorem 2.4
(1)]{w1}. Recently, it has been shown that non-compact conformal repellers may have non-concave Lyapunov spectra (see \cite{ks} for the Gauss map and \cite{io} for the Renyi map). Puzzled by this phenomena, our aim here is to better understand the concavity properties of Lyapunov spectra.
On one hand we characterize interval maps with two linear full branches for which the Lyapunov spectra is not concave. In particular, we not only show that compact conformal repellers (see Example \ref{ex}) may have non-concave spectra, but that this already occurs in the simplest possible context. On the other hand, we establish some general conditions under which the Lyapunov spectra is non-concave. Roughly speaking, the asymptotic variance (of the $-t \log |T'|$ potential) should be sufficiently large (in a certain sense).
The class of maps that we consider is defined as follows. Given a pairwise disjoint finite family of closed intervals $I_1, \dots, I_n$ contained in $[0,1]$ we say that a map: $$T: \bigcup_{i=1}^n I_i \to [0,1]$$ is a {\sf cookie-cutter map with $n$ branches} if the following holds: \begin{enumerate} \item $T(I_i)= [0,1]$ for every $i \in \{1 , \dots, n\}$, \item The map $T$ is of class $C^{1 + \epsilon}$ for some $\epsilon >0$,
\item $|T' (x)| > 1$ for every $x \in I_1 \cup \cdots \cup I_n $. \end{enumerate} We say that $T$ is a {\sf linear cookie-cutter map} if $T$ restricted to each one of the intervals $I_i$ is an affine map. The {\sf repeller} $\Lambda \subset [0,1]$ of $T$ is \[ \Lambda:= \bigcap_{n=0}^{\infty} T^{-n} ([0,1]). \] The {\sf Lyapunov exponent} of the map $T$ at the point $x \in [0,1]$ is defined by
\[ \lambda(x) = \lim_{n \to \infty} \frac{1}{n} \log |(T^n)'(x)|, \] whenever the limit exists. Let us stress that the set of points for which the Lyapunov exponent does not exist has full Hausdorff dimension~\cite{bs}.
We will mainly be concerned with the Hausdorff dimension of the level sets of $\lambda$. More precisely, the range of $\lambda$ is an interval $[\alpha_{min}, \alpha_{\max}]$ and the {\sf multifractal spectrum of the Lyapunov exponent} is the function given by: $$\begin{array}{rccc}
L:& [\alpha_{min}, \alpha_{\max}] & \rightarrow & \mathbb{R} \\
& \alpha & \mapsto & \dim_H ( J_\alpha = \{ x \in \Lambda \mid \lambda(x) = \alpha \}), \end{array}$$ where $\dim_H(J)$ denotes the Hausdorff dimension of $J$. For short we say that $L$ is the {\sf Lyapunov spectrum of $T$}.
In our first result we consider linear cookie-cutters with two branches and obtain conditions on the slopes that ensure that the Lyapunov spectrum is concave. This result can also be used to construct examples of non-concave Lyapunov spectrum.
\begin{introtheorem}
\label{A}
Consider the linear cookie-cutter map with two branches $$T: \left[0, \frac{1}{a} \right] \cup \left[1-\frac{1}{b}, 1\right] \to \left[0,1\right] $$
defined by
$$T(x) = \begin{cases}
a x & \text{ if } x \le \frac{1}{a}, \\
b x + 1 - b & \text{ if } x \ge 1-\frac{1}{b}. \end{cases} $$
Then the Lyapunov spectrum $L:[\log a , \log b] \to \mathbb{R}$ of $T$ is concave if and only if $$\dfrac{\log b}{\log a} \leq \dfrac{\sqrt{2 \log 2} + 1}{\sqrt{2 \log 2} - 1} \approxeq 12.2733202...$$ \end{introtheorem}
The above relation gives the bifurcation point dividing the spectra with inflection points form the concave ones. For maps with two linear branches, the combination of Lemma~\ref{tran} with this Theorem implies that the bifurcation between concave spectra and non-concave one may only occur when the Lyapunov exponent, $\alpha_M$, corresponding to the measure of maximal entropy is an inflection point.
In order to describe the Lyapunov spectrum we will make use of the thermodynamic formalism. Let $T$ be a cookie-cutter map, denote by $ \mathcal{M}_T$ the set of
$T-$invariant probability measures. The {\sf topological pressure} of $-t \log |T'|$ with respect to $T$ is defined by \begin{equation*}
P(-t \log |T'|) = \sup \left\{ h(\mu) -t \int \log |T'| \ d\mu : \mu \in \mathcal{M}_T \right\}, \end{equation*}
where $ h(\mu)$ denotes the measure theoretic entropy of $T$ with respect to the measure $\mu$ (see \cite[Chapter 4]{wa} for a precise definition of entropy). A measure $\mu_{t} \in \mathcal{M}_T$ is called an {\sf equilibrium measure} for $-t \log |T'|$ if it satisfies:
\[ P(-t \log |T'|) = h(\mu_{t}) - t \int \log |T'| \ d\mu_{t}. \]
If the function $\log |T'|$ is not cohomologous to a constant then the function $t \mapsto P(-t \log |T'|)$ is strictly convex, strictly decreasing, real analytic and for every $t \in \mathbb{R}$ there exists a unique equilibrium measure $\mu_t$ corresponding to $-t \log |T'|$ (see \cite[Chapters 3 and 4]{pp}). Moreover, there are explicit expressions for the derivatives of the pressure. Indeed (see \cite[Chapter $4$]{pu}), the first derivative of the pressure is given by
\[\alpha(t_0):= -\frac{d}{dt} P(-t \log |T'|) \Big|_{t=t_0} = \int \log |T'| \ d \mu_{t_0}.\] The second derivative of the pressure is the {\sf asymptotic variance}
\[\frac{d^2}{dt^2} P(-t \log |T'|) \Big|_{t=t_0} = \sigma^2(t_0),\] where \[\sigma^2(t_0):=
\lim_{n \to \infty} \int \left( \sum_{i=0}^{n-1} - \log |T'(T^i x)| +n
\int \log |T'| d \mu_{t_0} \right) ^2 d\mu_{t_0}(x). \] There exists a close relation between the topological pressure and the Hausdorff dimension of the repeller. In fact, the number $t_d = \dim_H(\Lambda)$ is the unique zero of the Bowen equation (see \cite[Chapter $7$]{pe})
\[P(-t \log |T'|) =0.\]
Let $\mu_{t_d}$ be the unique equilibrium measure corresponding to the function $-t_d \log |T'|$ and, let
\[\alpha_d := \int \log |T'| \ d\mu_d.\] Our next theorem establishes general conditions for a cookie-cutter map to have (non-)concave Lyapunov spectrum. \begin{introtheorem}
\label{B}
Let $L$ be Lyapunov spectrum of a cookie-cutter map $T$. Then $L$ is always concave in $[\alpha_{min} , \alpha_d]$. Moreover, $L: [\alpha_{min}, \alpha_{max}] \to \mathbb{R}$ is concave if and only if
$${\sigma^2(t)} < \dfrac{\alpha(t)^2} {2 P(-t \log |T'|)} \quad \text{ for all } \quad t < t_d \cdot$$ \end{introtheorem} When considering linear cookie-cutter maps we obtain a simpler formula in terms of the slopes of the map.
\begin{introcorollary}
\label{C}
Consider a linear cookie-cutter map $T$ with $n$-branches of slopes $m_1, \dots, m_n$.
Then its Lypunov spectrum $L: [\min\{\log|m_i|\}, \max\{\log|m_i|\}] \to \mathbb{R}$ is concave if and only if, for all $t \in \mathbb{R}$, \begin{equation*}
2 \log \left( \sum_{i=1}^n |m_i|^t \right) \left(\frac{\left( \sum_{i=1}^n |m_i|^t (\log |m_i|)^2 \right) \left( \sum_{i=1}^n |m_i|^t \right)}{\left( \sum_{i=1}^n |m_i|^t \log |m_i| \right)^2} - 1 \right) \leq 1. \end{equation*} \end{introcorollary}
In terms of the $L^1(\mu_t)$ and $L^2(\mu_t)$ norms with respect to the corresponding equilibrium measure $\mu_t$ the above formula may be rewritten as: $$
2 P(-t \log |T'|) \left( \dfrac{ \| \log|T'| \|^2_{2,t}}{ \| \log|T'| \|^2_{1,t} } -1 \right) \le 1.$$
Although our results shed some light on the concavity properties of the Lyapunov spectrum, up to our knowledge, the occurrence of inflection points is
not well understood. In fact, given a map with Lyapunov spectrum having an inflection point at $\alpha_0$, it is natural to pose the general problem of understanding the geometric and ergodic properties of the equilibrium measure with exponent $\alpha_0$.
From our results it follows that the Lyapunov spectrum of a cookie-cutter map has an even (possibly zero) number of inflections points. Although Theorem~\ref{A} establishes the existence of maps with spectra having at least two inflection points, it does not give an exact count. We conjecture that the spectrum of a cookie-cutter map with two branches has at most two inflection points. Also one may ask: Is there an upper bound on the number of inflection points of the spectrum of a cookie-cutter map? of a compact conformal repeller? of a non-compact conformal repeller?
Our results are based on the following formula that ties up the topological pressure with the Lyapunov spectrum \begin{equation} \label{WeissFormula}
L(\alpha) = \frac{1}{\alpha} \inf_{t \in \mathbb{R}} (P(-t \log |T'|) +t \alpha). \end{equation} This formula follows form the work of Weiss \cite{w1} and can be found explicitly, for instance, in the work of Kesseb\"ohmer and Stratmann \cite{ks}. Actually, in this setting, the Lyapunov spectrum can be written as \begin{equation} \label{lyapunov}
L(\alpha)= \frac{1}{\alpha} (P(-t_{\alpha} \log |T'|) +t_{\alpha} \alpha) = \frac{h(\mu_{\alpha})}{\alpha}, \end{equation} where $t_{\alpha}$ is the unique real number such that
\[ -\dfrac{d}{d t} P(-t \log |T'|) \Big|_{t=t_{\alpha}} = \int \log |T'| \ d\mu_{\alpha} = \alpha, \]
and $\mu_{\alpha}$ is the unique equilibrium measure corresponding to the potential $-t_{\alpha} \log |T'|$. That is, $\alpha \mapsto t_\alpha$ is the inverse of $t \mapsto \alpha(t)$. Thus, after the substitution $\alpha = \alpha(t)$, equation~\eqref{WeissFormula} becomes: \begin{equation} \label{lt}
L(\alpha(t))= \frac{1}{\alpha(t)} \left( P(-t \log |T'|) +t \alpha(t) \right). \end{equation}
The structure of the paper is as follows. In Section~\ref{2} we prove Theorem~\ref{A} and construct explicit examples of maps for which the Lyapunov spectrum is not concave. In Section~\ref{nl} we prove Theorem~\ref{B} and finally is Section~\ref{lc} we prove Corollary~\ref{C}.
\section{Maps with two branches} \label{2} Our aim now is to prove Theorem~\ref{A}. Throughout this section we let $T$ be the cookie-cutter with two linear branches of slopes $b>a>1$ and Lyapunov spectrum $L$, as in the statement of Theorem~\ref{A}. The proof relies on explicit formulas for $L$ together with a characterization of the inflection points that persist under small changes of the slopes $a$ and $b$ (i.e. transversal). We show that unstable zeros of $\frac{d^2 L}{d \alpha^2} (\alpha)$ may only occur at the Lyapunov exponent of the measure of maximal entropy when the logarithmic ratio of the slopes is as in the statement of the theorem.
We start by obtaining an explicit formula for $L$, from equation~\eqref{WeissFormula}. \begin{lema} \label{2BranchFormula}
\begin{equation*} L(\alpha) =
\frac{1}{\alpha} \left[- \left( \frac{\log b - \alpha}{\log \frac{b}{a}} \right) \log \left( \frac{\log b - \alpha}{\log \frac{b}{a}} \right) - \left( \frac{\alpha - \log a }{\log \frac{b}{a}} \right) \log \left( \frac{\alpha - \log a}{\log \frac{b}{a}} \right) \right]. \end{equation*} \end{lema}
\begin{proof}
The equilibrium measure $\mu_{\alpha}$, in equation~\eqref{lyapunov}, can be explicitly determined. This is due to the fact that it is a Bernoulli measure (see \cite[Theorem 9.16]{wa}). Since the Lyapunov exponent corresponding to $\mu_{\alpha}$ is $\alpha$ we have that \begin{eqnarray*}
\alpha &=& \int \log |T'| \ d\mu_{\alpha} = \mu_{\alpha}(I_1) \log a + \mu_{\alpha}(I_2) \log b \\ &=& \mu_{\alpha}(I_1) \log a +(1- \mu_{\alpha}(I_1)) \log b \\
&=& \mu_{\alpha}(I_1) \left( \log a - \log b \right) + \log b. \end{eqnarray*} Hence, \begin{eqnarray*} \mu_{\alpha}(I_1)& =& \frac{\log b - \alpha}{\log b - \log a}, \\ \mu_{\alpha}(I_2) & = & \frac{\alpha - \log a}{\log b - \log a}. \end{eqnarray*} Therefore, $\mu_{\alpha}$ is the unique Bernoulli measure which satisfies the above conditions. Moreover, the entropy of this measure is \begin{equation} \label{entropy} h(\mu_{\alpha})=-\left( \frac{\log b - \alpha}{\log b - \log a} \right) \log \left( \frac{ \log b - \alpha}{\log b - \log a} \right) - \left( \frac{\alpha - \log a}{\log b - \log a} \right) \log \left( \frac{\alpha - \log a}{\log b - \log a} \right). \end{equation} Hence, from equation \eqref{lyapunov} we obtain that \begin{equation*} \label{dosramas} L(\alpha) = - \frac{1}{\alpha \log \frac{b}{a}} \left[ \left({\log b - \alpha} \right) \log \left( \frac{\log b - \alpha}{\log b - \log a} \right) + \left({\alpha - \log a } \right) \log \left( \frac{\alpha - \log a}{\log b - \log a} \right) \right]. \end{equation*} \end{proof}
As suggested by equation~\eqref{lyapunov} the behavior of the entropy function is closely related to the (in)existence of inflection points of the Lyapunov spectrum. In fact: \begin{lema}[Remark 8.1 \cite{io}] \label{et} A point $\alpha_0 \in (\log a , \log b)$ satisfies $ \frac{d^2L}{d \alpha^2} (\alpha_0)=0$ if and only if
\[ 2 \frac{d L}{d \alpha} (\alpha_0) = \frac{d^2}{d \alpha^2} h(\mu_{\alpha}) \Big{|}_{\alpha=\alpha_0}. \] \end{lema}
Thus, our aim now is to study the second derivative of the entropy:
\begin{prop} \label{segunda} Let $\alpha_M= (\log a+ \log b)/2$. Then the function $$\frac{d^2}{d \alpha^2} h(\mu_{\alpha}) :(\log a , \log b) \to \mathbb{R}$$ is concave, increasing in the interval $(\log a ,\alpha_M)$ and, decreasing in the interval $(\alpha_M , \log b)$. In particular, it has a unique maximum at $\alpha = \alpha_M$. Moreover, \begin{eqnarray*}
\frac{d^2}{d \alpha^2} h(\mu_{\alpha}) \Big|_{\alpha = \alpha_M} & = & - \Big(\frac{2}{\log \frac{b}{a}} \Big)^2, \\ \lim_{ \alpha \to \log a}\frac{d^2}{d \alpha^2} h(\mu_{\alpha}) &=& - \infty,\\ \lim_{ \alpha \to \log b} \frac{d^2}{d \alpha^2} h(\mu_{\alpha}) & = &- \infty. \end{eqnarray*} \end{prop}
\begin{proof} From equation \eqref{entropy} we have that the first derivative of the entropy with respect to $\alpha$ is given by \begin{equation} \label{1-derivative} \frac{d}{d \alpha} h(\mu_{\alpha})=\frac{1}{\log \frac{b}{a}} \left( \log \left( \frac{\alpha - \log a}{\log {b} - \log {a}} \right) - \log \left( \frac{\log b - \alpha}{\log {b}- \log{a}} \right) \right). \end{equation} We can also compute its second derivative, \begin{eqnarray} \label{2-derivative} \frac{d^2}{d \alpha^2} h(\mu_{\alpha})=\frac{-1}{\log \frac{b}{a}} \left( \frac{1}{\alpha - \log a} +\frac{1}{\log b - \alpha} \right).
\end{eqnarray} Note that this function has two asymptotes at $\log a$ and at $\log b$. Indeed \[ \lim_{ \alpha \to \log a} \frac{d^2}{d \alpha^2} h(\mu_{\alpha}) = - \infty \textrm{ and } \lim_{ \alpha \to \log b} \frac{d^2}{d \alpha^2} h(\mu_{\alpha})= - \infty. \] In order to obtain the maximum of $\frac{d^2}{d \alpha^2} h(\mu_{\alpha})$ we compute the third derivative of the entropy function,
\[ \frac{d^3}{d \alpha^3} h(\mu_{\alpha})= \frac{1}{\log \frac{b}{a}} \left( \frac{1}{(\alpha - \log a)^2} -\frac{1}{(\log b- \alpha)^2} \right), \] which is equal to zero if and only if \[ \alpha=\alpha_M= \frac{\log b + \log a}{2} . \]
Moreover, \[ \frac{d^4}{d \alpha^4} h(\mu_{\alpha})= \frac{-2}{\log \frac{b}{a}} \left( \frac{1}{(\alpha - \log a)^3} +\frac{1}{(\log b- \alpha)^3} \right), \] In particular \[ \frac{d^4}{d \alpha^4} h(\mu_{\alpha}) \leq 0. \] \end{proof}
Now we combine the Proposition \ref{segunda} with Lemma~\ref{et} in order to characterize inflection points of the spectrum which are stable under small changes of the slopes $a$ and $b$. Namely, transversal intersections between the graphs of $2 \frac{d L}{d \alpha} (\alpha)$ and
$\frac{d^2}{d \alpha^2} h(\mu_{\alpha})$.
\begin{lema} \label{tran} Assume that $\alpha_i \in (\log a, \log b)$ with $\alpha_i \neq \alpha_M$ is such that $$2 \frac{d L}{d \alpha} (\alpha_i) =
\frac{d^2}{d \alpha^2} h(\mu_{\alpha}) \Big|_{\alpha = \alpha_i}.$$ Then the intersection of the graphs of $ \frac{d L}{d \alpha} (\alpha)$ and $ \frac{d^2}{d \alpha^2} h(\mu_{\alpha}) $ at $\alpha = \alpha_i$ is transversal. \end{lema}
\begin{proof} By contradiction, suppose that the intersection is not transversal, that is \begin{equation} \label{trans}
2 \frac{d^2 L}{d \alpha^2} (\alpha_i) = \frac{d^3 h(\mu_\alpha)}{d \alpha^3} (\alpha_i).
\end{equation}
Since $$2 \frac{d}{d \alpha} L(\alpha_i) = \frac{d^2}{d \alpha^2}h (\alpha_i)$$ we have that $ \frac{d^2 L}{d \alpha^2} (\alpha_i)=0$. Therefore equation \eqref{trans} implies that $ \frac{d}{d \alpha}h (\alpha)\Big|_{\alpha =\alpha_i}=0$. Thus, by Proposition \ref{segunda}, $\alpha_i= \alpha_M$. \end{proof}
Finally, we are ready to prove the Theorem.
\begin{proof}[Proof of Theorem~\ref{A}] From equation \eqref{lyapunov} we obtain that \begin{eqnarray*} 2 \frac{d L}{d \alpha}(\alpha)= \frac{2}{\alpha^2 \log \frac{b}{a} } \left[ \log b \log
\left( \frac{\log b - \alpha}{\log {b} - \log {a}} \right) - \log a
\log \left( \frac{\alpha - \log a}{\log b - \log{a}} \right)
\right].
\end{eqnarray*} Therefore, \begin{eqnarray*}
2 \frac{d L}{d \alpha} (\alpha) \Big|_{\alpha=\alpha_M} &=& -\frac{8 \log 2}{ (\log a + \log b)^2}. \end{eqnarray*} Moreover, \begin{eqnarray*} \lim_{\alpha \to \log a} 2\frac{d L}{d \alpha}(\alpha) & = & \infty, \\ \lim_{\alpha \to \log b} 2\frac{d L'}{d \alpha} (\alpha) & = & -\infty = \lim_{\alpha \to \log b} \left( 2\frac{d L}{d \alpha}(\alpha) - \frac{d^2}{d \alpha^2} h(\mu_{\alpha}) \right). \end{eqnarray*} Hence, a sufficient condition to have two transversal intersections of the graphs of $2\frac{d L}{d \alpha}$ and $\frac{d^2}{d \alpha^2}h$ is $2\frac{d L}{d \alpha} (\alpha_M) > \frac{d^2}{d \alpha^2}h(\alpha_M)$. Now, $$2 \frac{d L}{d \alpha} (\alpha_M) \ge \frac{d^2}{d \alpha^2}h(\alpha_M) \iff \dfrac{\log b}{\log a} \geq \dfrac{\sqrt{2 \log 2} + 1}{\sqrt{2 \log 2} - 1},$$ where equality holds in one equation if and only if it holds in the other. To finish the proof of the theorem we must check that there exist values of $a$ and $b$ such that $$\dfrac{\log b}{\log a} < \dfrac{\sqrt{2 \log 2} + 1}{\sqrt{2 \log 2} - 1},$$ for which the corresponding graphs of $2\frac{d L}{d \alpha}$ and $\frac{d^2}{d \alpha}h$ do not intersect. For this purpose let $a >1$ and $b = \exp(\varepsilon) a$. To ease notation we introduce $$x = \dfrac{\alpha-\log a}{\log b - \log a},$$ and note that $0 \leq x \leq 1$. It follows that $$ \varepsilon^2 x (1 -x) 2 \frac{d L}{d \alpha} (\alpha) = \dfrac{2 \varepsilon}{\alpha^2} g(x),$$ where $g:[0,1] \rightarrow \mathbb{R}$ is the continuous function such that for all $x \in (0,1)$, $$g(x) = \varepsilon x (1-x) \log (1-x) + (\log a) x (1-x) \log (1-x) - (\log a) (1-x) x \log x.$$ Hence, uniformly for $\alpha \in [\log a, \log b]$, we have that $ \varepsilon^2 x (1 -x) 2 \frac{d}{d \alpha}L(\alpha) \to 0$, as $\varepsilon \searrow 0$. However, from equation~\eqref{2-derivative} $$ \varepsilon^2 x (1 -x) \frac{d^2}{d \alpha^2} h(\mu_{\alpha}) = -1.$$ Thus, for $\varepsilon >0$ sufficiently small, we have that $2\frac{d L}{d \alpha} (\alpha) > \frac{d^2}{d \alpha^2}h(\alpha)$ for all $\alpha \in [\log a, \log b]$. \end{proof}
\begin{eje}[Non-concave Lyapunov spectrum] \label{ex} {\em Let $T$ be a linear cookie-cutter map with two branches of slopes $a=\exp(1)$ and $b= \exp(45)$. The corresponding Lyapunov spectrum is not concave since \[45=\dfrac{\log b}{\log a} > \dfrac{\sqrt{2 \log 2} + 1}{\sqrt{2 \log 2} - 1}. \]
See Figure~\ref{2BranchFigure} (right).} \end{eje}
\begin{figure}
\caption{The Lyapunov spectra for maps with two linear branches of slopes $a$ and $b$. For all graphs $a=\exp(1)$. At the left a concave spectra corresponding to $b= \exp(10)$. At the center, also concave but $b \approx (\sqrt{2 \log 2} + 1) / (\sqrt{2 \log 2} - 1)$ is the bifurcation value of the slope. At the right, the non-concave graph corresponding to $b=\exp(45)$. }
\label{2BranchFigure}
\end{figure}
\section{The non-linear case} \label{nl} In this section we prove Theorem~\ref{B}, which gives a general condition that ensures the existence of inflection points for the Lyapunov spectrum. In the next section we will show that the mentioned condition is explicit for maps with linear branches.
\begin{proof}[Proof of Theorem~\ref{B}] From equation~\eqref{lt} we have
\[L(\alpha(t))= \frac{P(-t \log |T'|) +t \alpha(t)}{\alpha(t)}.\] Then, using $f(g(t))'$ to denote the derivative of $f \circ g$ with respect to $t$, we have
\[L(\alpha(t))'= \frac{1}{\alpha(t)^2} \left( - \alpha'(t) P(-t \log |T'|) + \alpha(t) P(-t\log |T'|)' +\alpha(t)^2 \right).\] Recall that \begin{equation} \label{rela}
\alpha(t) = -P(-t \log |T'|)'. \end{equation}
Hence
\[L(\alpha(t))' = -\frac{\alpha'(t) P(-t \log |T'|)}{ \alpha(t)^2}. \] Therefore, making use of equation \eqref{rela} we obtain that \begin{eqnarray*} \frac{d^2 L}{d \alpha^2} (\alpha(t)) & =& \frac{1}{\alpha'(t)} \cdot \left( \frac{L(\alpha(t))'}{\alpha'(t)} \right)' \\
& = & \frac{1}{\alpha'(t)} \cdot \left( \frac{-P(-t \log |T'|)}{\alpha(t)^2}\right)' \\
& =& \frac{1}{\alpha'(t)} \cdot \frac{2\alpha'(t) P(-t \log |T'|) - \alpha(t) P(-t \log |T'|)'}{\alpha(t)^3}\\
& =& \frac{1}{\alpha'(t)} \cdot \frac{-2 \alpha'(t) P(-t \log |T'|) + \alpha(t)^2}{\alpha(t)^3} \cdot\end{eqnarray*} Since
$$\alpha' (t) = \sigma^2 (t) = P(-t \log |T'|)'' < 0$$ for all $t \in \mathbb{R}$, we conclude that \begin{equation} \label{sindividir}
\frac{d^2 L}{d \alpha^2} (\alpha(t)) \le 0 \iff -2 \sigma^2(t) P(-t \log |T'|) + \alpha(t)^2 \ge 0. \end{equation}
The theorem follows since $P(-t \log |T'|)$ is positive if and only if $t < t_d$. \end{proof}
Let us stress that even though Theorem \ref{B} is very general, it is not easy to deduce explicit conditions on $T$ in order to guarantee the absence or presence of inflection points (in contrast with Theorem \ref{A}).
\section{The linear case} \label{lc} Throughout this section we consider a linear cookie-cutter map with $n$ branches of slopes $m_1, \dots, m_n$. We record a straightforward computation in the next lemma. This trivial calculation allow us to ''effectively'' draw the graphs of the corresponding Lyapunov spectra (e.g. see Figure~\ref{3BranchFigure} ).
\begin{lema}
\label{graph} Consider a linear cookie-cutter map $T$ with $n$ branches of slopes $m_1, \dots, m_n$.
Then,
\begin{eqnarray*}
\alpha(t) &= & \frac{\sum_{i=1}^n |m_i|^t \log |m_i|}{\sum_{i=1}^n |m_i| ^t},\\
L(\alpha(t)) &=& \frac{\left(\sum_{i=1}^n |m_i|^t \right) \log \left(\sum_{i=1}^n |m_i|^t \right)}{\sum_{i=1}^n | m_i|^t \log |m_i|} -t.
\end{eqnarray*} \end{lema}
Note that the graph of $L: [\min\{\log|m_i|\}, \max\{\log|m_i|\}] \to \mathbb{R}$ coincides with the graph of $\mathbb{R} \ni t \mapsto (\alpha(t), L(t))$.
\begin{figure}
\caption{The Lyapunov spectra for maps with three linear branches of slopes $a =\exp(1) < b=\exp(2) < c$. At the left a concave spectra corresponding to $c= \exp(4)$. At the center, a non-concave corresponding to $c= \exp(8)$. At the right, another non-concave spectrum corresponding to $b=\exp(16)$. }
\label{3BranchFigure}
\end{figure}
\begin{proof} The pressure function of a linear cookie-cutter map has a simple form (see, for instance, \cite[Example 1]{sa2}), \begin{equation*}
P(-t \log |T'|) =\log \sum_{i=1}^n |m_i|^t . \end{equation*} Hence, \begin{equation} \label{al}
\alpha(t) = -P(-t \log |T'|)' = \frac{\sum_{i=1}^n |m_i|^t \log |m_i|}{\sum_{i=1}^n |m_i|^t}. \end{equation} The formula for $L(\alpha(t))$ follows from Equation \eqref{lt}. \end{proof}
\begin{proof}[Proof of Corollary~\ref{C}] From the formula \eqref{al} for $\alpha(t)$, it follows that:
$$\sigma^2(t) = \frac{\sum_{i=1}^n |m_i|^t (\log |m_i|)^2}{\sum_{i=1}^n |m_i| ^t} -
\left(\frac{\sum_{i=1}^n |m_i|^t \log |m_i|}{\sum_{i=1}^n |m_i| ^t }\right)^2 . $$ We introduce the following notation:
\begin{eqnarray*}
\| \log|T'| \|^2_{2,t} & = & \frac{\sum_{i=1}^n |m_i|^t (\log |m_i|)^2}{\sum_{i=1}^n |m_i| ^t}\\
\| \log|T'| \|_{1,t} & = &\frac{\sum_{i=1}^n |m_i|^t \log |m_i|}{\sum_{i=1}^n |m_i| ^t } \end{eqnarray*} Now from Equation~\eqref{sindividir}, we have that $L$ is concave if and only if, for all $t \in \mathbb{R}$,
$$
2 P(-t \log |T'|) \left( \| \log|T'| \|^2_{2,t} - \| \log|T'| \|^2_{1,t} \right) \le \| \log|T'| \|^2_{1,t}.$$ \end{proof}
\end{document}
|
arXiv
|
{
"id": "0812.1936.tex",
"language_detection_score": 0.6629893183708191,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Measuring Phonon Dephasing with \\ Ultrafast Pulses}
\author{F. C. Waldermann} \author{Benjamin J. Sussman} \email{[email protected]} \altaffiliation[Also at ]{National Research Council of Canada, Ottawa, Ontario K1A 0R6, Canada} \author{J. Nunn} \author{V. O. Lorenz} \author{K. C. Lee} \author{K. Surmacz} \author{K. H. Lee} \author{D. Jaksch} \author{I. A. Walmsley} \affiliation{Clarendon Laboratory, University of Oxford, Parks Road, Oxford, OX1 3PU, UK} \author{P. Spizziri} \author{P. Olivero} \author{S. Prawer} \affiliation{Center for Quantum Computer Technology, School of Physics, The University of Melbourne, Parkville, Victoria 3010, Australia}
\begin{abstract} A technique to measure the decoherence time of optical phonons in a solid is presented. Phonons are excited with a pair of time delayed 80~fs, near infrared pulses via spontaneous, transient Raman scattering. The fringe visibility of the resulting Stokes pulse pair, as a function of time delay, is used to measure the phonon dephasing time. The method avoids the need to use either narrow band or few femtosecond pulses and is useful for low phonon excitations. The dephasing time of phonons created in bulk diamond is measured to be $\tau=$6.8~ps ($\Delta \nu=1.56$~cm$^{-1}$). \end{abstract}
\maketitle \section{Introduction}
Phonons are a fundamental excitation of solids that are responsible for numerous electric, thermal, and acoustic properties of matter. The lifetime of optical phonons plays an important role in determining these physical properties and has been the subject of extensive study. A technique to measure phonon dephasing times is presented here that utilizes ultrafast infrared laser pulses: Transient Coherent Ultrafast Phonon Spectroscopy (TCUPS) offers a number of conveniences for measuring phonon dephasing. TCUPS utilizes commercially available ultrafast pulses (80~fs) and hence does not require a narrow band or extremely short lasers to achieve high spectral or temporal resolution. As well, TCUPS is suitable for measurements in the single phonon excitation regime. The large sampling area and long sampling distance increase the generated Stokes power and avoid sample heating, which is a concern for low temperature studies. Diamonds are well known for their extraordinary physical properties \cite{Field1979} and, as well, offer interesting prospects for use in Quantum Information applications \cite{Wrachtrup2006,Childress2006, Neumann2008,Waldermann2007}. As such, diamond has been selected here as the material for demonstration of TCUPS.
Two methods have previously been utilized to measure phonon lifetimes: high-resolution Raman spectroscopy and differential reflectivity measurements. The first is the traditional technique, where the optical phonon lifetime is obtained from high-resolution linewidth measurements of the first-order Raman peak, usually conducted using narrowband excitation lasers and high-resolution spectrometers \cite{lbpb2000}. The alternative technique, working in the time-domain, can directly show the temporal evolution of the surface vibrations of solids \cite{ckk1990}. A femtosecond pump pulse is used to excite a phonon population. The reflectivity (or transmittivity) of a subsequent probe pulse displays a time dependence that follows the vibrational frequency and phonon population. This method was used to study the phonon decay in various solids \cite{cbkmddi1990}, their symmetry modes \cite{lyylkl2003}, and their interaction with charge carriers \cite{hkcp2003} and with other phonons \cite{bdk2000}. In these experiments, impulsive stimulated Raman scattering has been established as the coherent phonon generation mechanism \cite{sfigwn1985,lfgwfum1995}.
The time-domain experiments utilize the impulsive regime, \textit{i.e}.~laser pulse lengths much shorter than the phonon oscillation period (inverse phonon frequency). This requirement can be challenging for the application of the differential reflectivity technique to materials with high phonon energies, as laser systems with very short pulse lengths are required (\textit{e.g.}, for diamond, sub-10~fs pulses are required to resolve a phonon frequency of 40~THz). TCUPS operates in the transient Raman scattering regime, \textit{i.e.}, pulse lengths much shorter than the phonon decoherence time, which is usually orders of magnitude slower than the phonon oscillation period (about 25~fs for diamond) \cite{ihk2007}. Stimulated Raman scattering, which implies large phonon excitations, is often employed in dephasing measurements in order to achieve good signal-to-noise ratios. High phonon population numbers, often referred to as \emph{hot phonons}, can be subject to an increased decay rate, as previously observed \cite{ylokl2002} for {GaN}. By contrast, TCUPS investigates the properties of a coherent phonon excitation by direct analysis of the Stokes spectra generated in the Raman process. The use of single photon detectors extends the sensitivity of the experiment to low phonon populations, including the single phonon level.
\section{Experiment}
\begin{figure}
\caption{Experimental setup. An oscillator pulse is split into two time delayed pulses and focused through the diamond sample. Not shown, a bandpass filter cleans the oscillator pulse before the diamond and a longpass filter rejects the pump and transmits the Stokes before the spectrometer.}
\label{fig:bulk_raman_experiment}
\end{figure}
\begin{figure}
\caption{Diamond energy level schematic. Ground state phonons are excited with the incident 788~nm pump, via a Raman transition, to the optical phonon mode, emitting an 880~nm Stokes pulse}
\label{fig:diamondEnergyLevels}
\end{figure}
The diamond was classified as a type Ib high pressure high temperature (HPHT) sample with a nitrogen impurity concentration of less than 100~ppm. The Stokes shift of diamond \cite{zaitsev2001} is $\unit[1332]{cm}^{-1}$ and the Raman gain coefficient for diamond has been reported \cite{llk1971} as $g = \unit[7.4 \times 10^{-3}]{cm/MW}$ (corrected for $\lambda = \unit[800]{nm}$). With pump pulse energies ranging over $\unit[1.1 \ldots 380]{pJ}$, the collinear Stokes emission is calculated as $0.004 \ldots 1.3$ photons per pulse, in agreement with the count rates achieved experimentally. The Raman scatter is thus in the spontaneous regime, as verified by a linear pump power dependence ranging over three orders of magnitude (see inset of Fig.~\ref{fig:powerdep}). Therefore, the experiment is performed far below the hot phonon regime.
The experimental setup is depicted in fig.~\ref{fig:bulk_raman_experiment}. Phonons are excited, via a Raman transition, with a pair of time-delayed 80~fs, 788~nm pulses (fig.~\ref{fig:diamondEnergyLevels}) from a commercial Ti:Sapphire Oscillator (Coherent Mira). The pulses are focussed into a $2 \times 2 \times 1$~mm diamond with faces polished along [100] plane (Sumitomo). Stokes emission is detected collinearly. The pump laser is spectrally filtered using a bandpass filter to avoid extraneous light at the Stokes frequency, which might seed stimulated emission and decreases the signal-to-noise ratio when detecting single Stokes photons. The Stokes scatter is detected and spectrally analysed by means of a 30~cm spectrometer (Andor Shamrock 303i) and an electron multiplying charged coupled device (iXon DV887DCS-BV), which is capable of statistical single photon counting. The gratings were ruled at 150 lines/mm for data in figures~\ref{fig:two_pulse_spectra_principle}, and 1800 lines/mm for figures~\ref{fig:waterfall_etc} and~\ref{fig:powerdep}.
\begin{figure}
\caption{Example spectral interference for a delay $\tau = 0.39$~{ps}. Spectra of the broadband excitation laser (left) and the Stokes signal of diamond (right). The single pulse data in (a) and (c) show the pump laser spectrum and the corresponding broadband Raman spectrum. Spectral interference fringes appear for coherent pulse pairs in (b) and (d). }
\label{fig:two_pulse_spectra_principle}
\end{figure}
The spectral interference from the pump pair and Stokes pair is shown in fig.~\ref{fig:two_pulse_spectra_principle}. The fringe spacing $\Delta \lambda$ is as expected for two time-delayed coherent pulses: $\Delta \lambda={\lambda^2}/{c \tau}$ (see also (\ref{eq:IntensityFringes}), below). For the excitation pair, $\lambda$ is the centre wavelength of the pump (fig.~\ref{fig:two_pulse_spectra_principle}b) and for the generated output Raman pair, $\lambda$ is the centre wavelength of the Stokes (fig.~\ref{fig:two_pulse_spectra_principle}d). The fringe spacing of the Raman output corresponds to the Stokes peak wavelength, confirming that the process is a measure of the coherence of the Raman process. Figure \ref{fig:waterfall_etc}a shows the fringe visibility reduction as a function of time delay. The fringe visibility $V=\exp(-\Gamma|\tau|)$ is plotted in fig.~\ref{fig:waterfall_etc}b. The visibility has been renormalized using the laser visibility for each delay to account for beam walk-off and the spectrometer resolution which artificially reduces visibility, due to a sampling effect from the finite pixel size of the spectrometer CCD.
\begin{figure}
\caption{Decoherence measurement. (a) Stokes spectra for two pump pulses with various delays $\tau$, recorded with a 1800 lines/mm grating. The decrease of spectral interference visibility of the Stokes signal is due to decoherence of the optical phonons created. The respective visibilities are plotted in (b), obtained by curve-fitting the spectra. Asterisks denote data points, the continuous line an exponential decay fit. Part (c) shows a high-resolution Raman spectrum for the same diamond (dots) with a Lorentzian fit (line). }
\label{fig:waterfall_etc}
\end{figure}
\section{Theory} The observed interference visibility can be considered from two perspectives. In the first, the visibility decay arises due to fluctuations of the phase in the classical fields. Each input laser pulse excites optical phonons, via a Raman transition (fig.~\ref{fig:diamondEnergyLevels}), which in turn causes the emission of two Stokes pulses. That is, the Raman interaction maps the electric field of the two input pulses to two output Stokes shifted pulses \begin{equation} E_{Stokes}=E_1(t)+E_2(t). \end{equation} The phase of the first Stokes pulse $E_1$ is determined spontaneously, but the phase (and amplitude at stimulated intensities) of the second pulse $E_2$ is influenced by the coherence maintained in the phonon of the system following the first pulse, so that the output field may also be rewritten as \begin{equation} E_{Stokes}(t)=E_1(t)+e^{i \theta}E_1(t-\tau) \end{equation} where $\tau$ is the time delay between the input pulses and $\theta$ is the spontaneously fluctuating phase difference between the pulses. The spectral intensity of the Stokes pulse pair \begin{equation}
|E_{Stokes}(\omega)|^2=2|E_1(\omega)|^2 \left(1+\cos({\omega \tau +\theta}) \right ) \end{equation} contains interference fringes whose position depends on the relative phase $\theta$. Shot-to-shot, decoherence causes spontaneous fluctuations in $\theta$ and the fringe pattern loses visibility. At longer delays $\tau$, the fluctuations increase, eventually reducing the visibility of any integrated fringe pattern to zero. Assuming a Lorentzian lineshape with width $\Gamma$ for the distribution of the phase shift, the shot-to-shot averaged spectral intensity is broadened to \begin{equation} \label{eq:IntensityFringes}
\left<|E_{Stokes}(\omega)|^2 \right>_{shots}=2|E_1(\omega)|^2 \left(1+e^{-\Gamma|\tau|}\cos({\omega \tau}) \right ). \end{equation} The phase fluctuations cause a reduction of the fringe visibility.
Alternatively, the fluctuating phase perspective can be connected with the second, quantum field perspective. This formalism can also be made applicable in the stimulated regime. The observed spectral intensity expectation value is proportional to the number of Stokes photons: \begin{equation}
\left <|E_{Stokes}(\omega)|^2 \right>\propto \left < A^{\dagger} A\right > \end{equation} where the lowering operator $A(\omega)$ is a sum of the first $A_1$ and second $A_2$ pulse mode lowering operators: \begin{equation}
\left<|E_{Stokes}(\omega)|^2 \right>\propto\left < A_1^{\dagger} A_1\right > +\left < A_2^{\dagger} A_2\right > +2 \Re \left < A_1^{\dagger} A_2\right >. \end{equation} The final, correlated term (\textit{cf}. the decay term in (\ref{eq:IntensityFringes})) measures the phonon coherence that remains in the system between pulses. During the evolution of the system, the starting time phonon mode $B^\dagger(0)$ is `mixed' into the Stokes photon modes $A_i$ due to application of the laser field. The mixed-in term is then subsequently the source for spontaneous emission. The source term is the same for both pulses, but during the period between pulses the coherence is reduced due to crystal anharmonicity and impurities. For the correlation, the relevant terms to lowest perturbative order are (see appendix for the equations of motion) \begin{equation} A_1\approx A_1(0) -i g \tau_{pump} B^\dagger(0) \end{equation} \begin{equation} A_2\approx A_1(0)e^{i \omega \tau} -i g\tau_{pump} B^\dagger(0)e^{-\Gamma \tau} \end{equation} from which the correlation term can be evaluated as \begin{equation}
\left < A_1^{\dagger} A_2\right > \approx g^2\tau_{pump}^2 \left < B(0) B(0) ^\dagger \right > = g^2 \tau_{pump}^2\left < N_B(0) +1 \right > e^{-\Gamma \tau}
\end{equation} where $N_B(0)$ is the initial number of phonons, which in this case is the nearly zero thermal population. This result links the phonon decoherence rate $\Gamma$ with the fluctuating phase perspective linewidth $\Gamma$ from (\ref{eq:IntensityFringes}). Therefore, measuring the reduction of the fringe visibility is a direct measure of the phonon dephasing time.
\section{Discussion}
The TCUPS measurement indicates a phonon dephasing time of $1/\Gamma=6.8\pm0.9$~ps or a linewidth of $\Delta \nu=1.56$~cm$^{-1}$. The literature has reported a great deal of variation in linewidth measurements for diamond, varying from at least 1.1~cm$^{-1}$ to as high as $4.75$~cm$^{-1}$ \cite{lbpb2000,llk1971,Chen1995} . Here, the TCUPS (fig.~\ref{fig:waterfall_etc}b, $\Delta \nu=1.56$~cm$^{-1}$ ) and conventional Raman spectrum (fig.~\ref{fig:waterfall_etc}c, $\Delta \nu=1.95$~cm$^{-1}$) show comparatively good agreement. The lifetime measured here is slightly shorter than the decay rate calculated theoretically by Debernardi \emph{et al.} \cite{dbm1995} for an ideal crystal ($\Delta \nu= \unit[1.01]{cm}^{-1}$ or $1/\Gamma=\unit[10.5]{ps}$), as the decay process is enhanced by lattice imperfections, vacancies, and the high concentration of substitutional nitrogen atoms, as is typical for this sort of diamond. The decay model considering acoustic phonon modes suggests that this deviation from the theoretical optimum is due to inhomogeneous broadening rather than additional pure dephasing. Future work will reveal whether ultra-pure diamond with very low crystal defect density can achieve a longer phonon lifetime. The creation of coherent phonons in diamond is heralded by the emitted Stokes photon, which could be employed for quantum optical experiments operating at room temperature like, \textit{e.g.}, schemes that transfer optical entanglement to matter\cite{maku2004,crfpek2005}.
\begin{figure}
\caption{Power dependence. The power dependence of the Stokes interference visibility (at constant delay $\tau = \unit[0.51]{ps}$, reduced due to a limited alignment) showing that the experiment can be carried out at arbitrarily low phonon excitation levels (the green horizontal line is plotted to guide the eyes). The inset shows the dependence of the Stokes pulse energy on the average pump power (single pump pulses). The linear power dependence shows that the scattering is in the spontaneous Raman regime. The fraction of pump power converted into collinear Stokes light was measured to be less than $10^{-8}$.}
\label{fig:powerdep}
\end{figure}
The spectral interference pattern persists for low excitation levels, \textit{i.e.,}~a phonon excitation probability per mode of $p < 1$. (The case of $p \gg 1$ would correspond to a strongly stimulated regime, which has previously been studied in molecular hydrogen gas \cite{bswr1993}, although using spatial, not spectral interference.) A constant visibility with excitation power can be seen at low excitation level (fig.~\ref{fig:powerdep}). TCUPS can therefore be employed to measure the decoherence properties of single optical phonons, overcoming the need for large phonon populations for lifetime measurements of phonon decoherence.
The excitation probability per mode was much smaller than 1, ranging over $p \approx 10^{-7} \ldots 10^{-5}$ due to the large number of phonon modes in the Brillouin zone for which Stokes scatter is detected ($\sim 10^5$). This level is in fact smaller than the thermal population level of the optical phonons at room temperature, given by $p_\mathrm{thermal} = [\exp(E_\mathrm{vib} / k_B T) - 1]^{-1} \approx 0.0017$. However, a small thermal population of the optical phonon modes does not influence this measurement method, as only the phonons deliberately excited by the pump pulses lead to Stokes scatter, and only Stokes light and its interference are detected. As phonons are governed by bosonic statistics, any finite background excitation level does not inhibit a further excitation. The linewidth increase due to phonon-phonon interaction is negligible at ambient temperatures in diamond due to the low population level \cite{dbm1995}. An increase in the Raman linewidth of diamond due to temperature has been reported \cite{lbpb2000} to begin at around $T \approx \unit[300]{K}$. At $T \approx \unit[800]{K}$, it is more than twice the zero temperature linewidth. At room temperature, the phonon decay is only marginally enhanced by an acoustical phonon population. This insensitivity to a thermal background is in contrast to the differential reflectivity method, where thermal phonons lead to additional noise as both thermal and coherent phonons lead to a change of the material reflectivity.
TCUPS is a convenient approach to determining the quantum coherence properties of optical phonons in Raman active solids. The measurement technique relies solely on spontaneous Raman scattering and is therefore useful down to the single phonon levels. In particular, TCUPS enables the measurement of the decoherence time of phonons, which is of paramount importance in many Quantum Information Processing schemes. Spectral interference of the Stokes light from pump pulse pairs is used to measure the Raman linewidth of the material, while maintaining a coherent excitation due to ultrafast excitation. The phonon lifetime of diamond was measured as $\unit[6.8]{ps}$. This lifetime corresponds to a phonon $Q$-factor of $Q = \nu /\Gamma \sim 270$. Although the short lifetime of the excitation makes it unsuitable for long-distance quantum repeaters, such a high $Q$ and the low thermal population at room temperature make it feasible for proof-of-principle demonstrations of typical quantum optics schemes, such as collective-excitation entanglement in the solid state.
Acknowledgements. This work was supported by the {QIPIRC} and {EPSRC} (grant number GR/S\-82176\-/\-01), EU RTN project EMALI, and Toshiba Research Europe.
\section*{Appendix: Phonon-photon equations of motion}
Consider an incident pump laser that Raman scatters off a phonon field of the diamond to produce an output Stokes field. The equations of motion for Stokes field $A(t)$ and the phonon field $B(t)$ are linked by the pump coupling $g$ via \cite{Raymer1990}: \begin{equation} \dot{A}(t)=-ig B^\dagger(t) \end{equation} and \begin{equation} \dot{B}(t)=-igA^\dagger(t)-\Gamma B(t) + F^\dagger(t). \end{equation} The dephasing rate $\Gamma$ is due to crystal anharmonicity and impurities. The Langevin operator $F$ has been added to maintain the normalization of $B$ in the presence of decay, allowing the phonon to decohere, but keeping the operator norm, via the commutation relation $[B,B^\dagger]=1$, constant. The formal solutions are: \begin{equation} B=B(0)e^{-\Gamma t} +\int_0^t e^{-\Gamma (t-t^\prime)}\left [ -i g A^\dagger(t^\prime)+F^\dagger(t^\prime)\right ] dt^\prime \end{equation} and \begin{equation} A=A(0)-i\int_0^t g B^\dagger(t^\prime)dt^\prime. \end{equation} For brevity, the time argument has been dropped from the solutions. In the weak ($g \tau_{pump} \ll 1$) and transient ($\Gamma \tau_{pump}\ll 1$) pump pulse limit, the incident laser leaves the phonon operator approximately in the vacuum state $B(0)$ and the phonon operator solution at lowest order is \begin{equation} B\approx B(0)e^{-\Gamma t} +\int_0^t e^{-\Gamma (t-t^\prime)} F^\dagger(t^\prime) dt^\prime. \end{equation} The Stokes field to first order is then \begin{equation} A \approx A(0) -i g \tau_{pump}B^\dagger(0)e^{-\Gamma t} -i \int_0^t \int_0^{t^\prime} g e^{-\Gamma (\tau-t^{\prime \prime})} F(t^{\prime \prime}) dt^{\prime \prime} dt^\prime \end{equation} where the coupling $g$ in the second term has been taken as a constant step for the duration of the pump. The initial Stokes operator $A(0)$ annihilates the vacuum, but the solution for $A$ mixes in a component of the phonon raising operator $B^\dagger(0)$, which acts as a source for the spontaneous Raman scattering.
The decoherence rate $\Gamma$ represents the dephasing of the phonon raising operator $B^\dagger$. The phonon number $N_B=B^\dagger B$ therefore decays at a rate $2\Gamma$. The corresponding spectral frequency linewidth is $\Delta \nu={\Gamma}/{\pi}$.
\end{document}
|
arXiv
|
{
"id": "0805.1668.tex",
"language_detection_score": 0.8335253596305847,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[Compactifications of ${\mathcal M}_{0, \lowercase{n}}$ associated with Alexander self dual complexes]{Compactifications of ${\mathcal M}_{0, \lowercase{n}}$ associated with Alexander self dual complexes: Chow ring, $\psi$-classes, and intersection numbers} \author{Ilia Nekrasov$^1$} \email{[email protected]} \address{$^1$Chebyshev laboratory, Department of Math. and Mech., St. Petersburg State University} \author{Gaiane Panina$^2$} \email{[email protected]} \address{$^2$PDMI RAS, St. Petersburg State University}
\begin{abstract} An \textit{Alexander self-dual complex} gives rise to a compactification of $\mathcal{M}_{0,n}$, called \textit{ ASD compactification}, which is a smooth algebraic variety. ASD compactifications include (but are not exhausted by) the \textit{polygon spaces}, or the moduli spaces of flexible polygons. We present an explicit description of the Chow rings of ASD compactifications. We study the analogs of Kontsevich's tautological bundles, compute their Chern classes, compute top intersections of the Chern classes, and derive a recursion for the intersection numbers. \end{abstract}
\maketitle
\keywords{Alexander self dual complex, modular compactification, tautological ring, Chern class, Chow ring}
\section{Introduction}
The moduli space of $n$-punctured rational curves ${\mathcal M}_{0,n}$ and its various compactifications is a classical object, bringing together algebraic geometry, combinatorics, and topological robotics. Recently, D.I.Smyth \cite{Smyth13} classified all \textit{modular compactifications} of ${\mathcal M}_{0,n}$. We make use of an interplay between different compactifications, and:
\begin{itemize} \item describe the classification in terms of (what we call) \textit{preASD simplicial complexes};
\item describe the Chow rings of the compactifications arising from Alexander self-dual complexes (ASD compactifications);
\item compute for ASD compactifications the associated Kontsevich's \textit{$\psi$-classes}, their top monomials, and give a recurrence relation for the top monomials. \end{itemize}
Oversimplifying, the main approach is as follows. Some (but not all) compactifications are the well-studied \textit{polygon spaces}, that is, moduli spaces of flexible polygons. A polygon space corresponds to a \textit{threshold} Alexander self-dual complex. Its cohomology ring (which equals the Chow ring) is known due to J.-C. Hausmann and A. Knutson \cite{HKn}, and A. Klyachko\cite{Kl}. The paper \cite{NP} gives a computation-friendly presentation of the ring. Due to Smyth \cite{Smyth13}, all the modular compactifications correspond to \textit{preASD complexes}, that is, to those complexes that are contained in an ASD complex. A removal of a facet of a preASD complex amounts to a blow up of the associated compactification. Each ASD compactification is achievable from a threshold ASD compactification by a sequence of blow ups and blow downs. Since the changes in the Chow ring are controllable, one can start with a polygon space, and then (by elementary steps) reach any of the ASD compactifications and describe its Chow ring (Theorem \ref{ASDChow}).
M. Kontsevich's $\psi$-classes \cite{Kon} arise here in a standard way. Their computation of is a mere modification of the Chern number count for the tangent bundle over $\mathbb{S}^2$ (a classical exercise in a topology course). The recursion (Theorem \ref{ThmRecursion}) and the top monomial counts (Theorem \ref{main_theorem}) follow.
It is worthy mentioning that a disguised compactification by simple games, i.e., ASD complexes, is discussed from a combinatorial viewpoint in \cite{PanGal}.
Now let us give a very brief overview of moduli compactifications of ${\mathcal M}_{0,n}$. A compactification by a smooth variety is very desirable since it makes intersection theory applicable. We also expect that (1) a compactification is \textit{modular}, that is, itself is the moduli space of some curves and marked points lying on it, and (2)
the complement of ${\mathcal M}_{0,n}$ (the “boundary”) is a divisor.
The space ${\mathcal M}_{0,n}$ is viewed as the configuration space of $n$ distinct marked points (``particles'') living in the complex projective plane. The space ${\mathcal M}_{0,n}$ is non-compact due to forbidden collisions of the marked points. Therefore, each compactification should suggest an answer to the question: what happens when two (or more) marked points tend to each other? There exist two possible choices: either one allows some (not too many!) points to coincide, either one applies a blow up. It is important that the blow ups amount to adding points that correspond to $n$-punctured nodal curves of arithmetic genus zero.
A compactification obtained by blow ups only is the celebrated \textit{Deligne--Mumford compactification}. If one avoids blow ups and allows (some carefully chosen) collections of points to coincide, one gets an ASD-compactification; among them are the \textit{polygon spaces}. Diverse combinations of these two options (in certain cases one allows points to collide, in other cases one applies a blow up) are also possible; the complete classification is due to \cite{Smyth13}.
Now let us be more precise and look at the compactifications in more detail.
\subsection{Deligne--Mumford compactification}\label{DMcom}
\begin{definition}\cite{DM69} Let $B$ be a scheme. A \textit{family of rational nodal curves with $n$ marked points} over $B$ is \begin{itemize} \item a flat proper morphism $\pi: C \rightarrow B$ whose geometric fibers $E_{\bullet}$ are nodal connected curves of arithmetic genus zero, and \item a set of sections $(s_{1}, \dots, s_{n})$ that do not intersect nodal points of geometric fibers. \newline In this language, the sections correspond to marked points. The above condition means that a nodal point of a curve may not be marked. \end{itemize} A family $(\pi:C \rightarrow B; s_{1}, \dots, s_{n})$ is \textit{stable} if the divisor $K_{\pi}+s_{1}+\dots+s_{n}$ is $\pi$-relatively ample.
Let us rephrase this condition: a family $(\pi:C \rightarrow B; s_{1}, \dots, s_{n})$ is {stable} if each irreducible component of each geometric fiber has at least three \textit{special points} (nodal points and points of the sections $s_{i}$). \end{definition}
\begin{theorem} \cite{DM69} (1) There exists a smooth and proper over $\mathbb{Z}$ stack $\overline{\mathcal{M}}_{0,n}$, representing the moduli functor of stable rational curves. Corresponding moduli scheme is a smooth projective variety over $\mathbb{Z}$.
(2) The compactification equals the moduli space for $n$-punctured stable curves of arithmetic genus zero with $n$ marked points. A stable curve is a curve of arithmetic genus zero with at worst nodal singularities and finite automorphism group. This means that (i) every irreducible component has at least three marked or nodal points, and (ii) no marked point is a nodal point. \end{theorem}
The Deligne-Mumford compactification has a natural stratification by stable trees with $n$ leaves. A \textit{stable tree with $n$ leaves} is a tree with exactly $n$ leaves enumerated by elements of $[n]=\{1,...,n\}$ such that each vertice is at least trivalent.
Here and in the sequel, we use the following \textbf{notation}: by \textit{vertices} of a tree we mean all the vertices (in the usual graph-theoretical sense) excluding the leaves. A \textit{bold edge} is an edge connecting two vertices (see Figure \ref{Figtrees}).
\begin{figure}
\caption{Stable nodal curves (left) and the corresponding trees (right)}
\label{Figtrees}
\end{figure}
The initial space ${\mathcal M}_{0,n}$ is a stratum corresponding to the one-vertex tree. Two-vertex trees (Fig.\ref{Figtrees}(b)) are in a bijection with bipartitions of the set $[n]$: $T\sqcup T^{c} = [n]$ s.t. $|T|, |T^{c}| > 1$. We denote the closure of the corresponding stratum by $\D_{T}$. The latter are important since the (Poincaré duals of) closures of the strata generate the Chow ring $\textbf{A}^{*}(\overline{\mathcal{M}}_{0,n})$:
\begin{theorem}\cite[Theorem 1]{Keel92} \label{KeelPresentation} The Chow ring $\bf{A}^{*}$$(\overline{\mathcal{M}}_{0,n})$ is isomorphic to the polynomial ring
$$\mathbb{Z}\left[\D_{T}:\; T \subset [n]; |T|>1, |T^{c}| >1\right]$$ factorized by the relations: \begin{enumerate} \item $\D_{T} = \D_{T^{c}}$; \item $\D_{T}\D_{S} =0$ unless $S \subset T$ or $T \subset S$ or $S \subset T^{c}$ or $T \subset S^{c}$; \item For any distinct elements $i,j,k,l \in [n]$: $$\sum_{i,j \in T; k,l \not\in T} \D_{T} = \sum_{i,k \in T; j,l \not\in T} \D_{T} = \sum_{i,l \in T; j,k \not\in T} \D_{T}$$ \end{enumerate} \end{theorem}
\subsection{Weighted compactifications}\label{Weicom}
The next breakthrough step was done by B. Hassett in \cite{Has03}.
Define a \textit{weight data} as an element $\mathcal{A} = (a_{1}, \dots, a_{n}) \in \mathbb{R}^{n}$ such that \begin{itemize} \item $0 < a_{i} \leq 1$ for any $i \in [n]$,
\item $a_{1}+ \dots +a_{n} > 2$. \end{itemize}
\begin{definition} Let $B$ be a scheme. A family of nodal curves with $n$ marked points $(\pi: C \rightarrow B; s_{1}, \dots, s_{n})$ is \textit{$\mathcal{A}$--stable} if \begin{enumerate} \item $K_{\pi}+ a_{1}s_{1}+\dots + a_{n}s_{n}$ is $\pi$-relatively ample,
\item whenever the sections $\{s_{i}\}_{i \in I}$ intersect for some $I \subset [n]$, one has $\sum_{i \in I} a_{i} < 1$. \end{enumerate} The first condition can be rephrased as:
each irreducible component of any geometric fiber has at least three {distinct} special points.
\end{definition}
\begin{theorem}\cite[Theorem 2.1]{Has03} For any weight data $\mathcal{A}$ there exist a connected Deligne--Mumford stack $\overline{\mathcal{M}}_{0, \mathcal{A}}$ smooth and proper over $\mathbb{Z}$, representing the moduli functor of $\mathcal{A}$--stable rational curves. The corresponding moduli scheme is a smooth projective variety over $\mathbb{Z}$. \end{theorem}
The Deligne--Mumford compactification arises as a special case for the weight data $ (1,\dots, 1)$.
It is natural to ask: how much does a weighted compactification $\overline{\mathcal{M}}_{0, \mathcal{A}}$ depend on $\mathcal{A}$? Pursuing this question, let us consider the space of parameters: $${ \mathscr{A}}_{n} = \left\{ \mathcal{A} \in \mathbb{R}^{n}:\; 0< a_{i} \leq 1, \; \sum_{i} a_{i} >2 \right\} \subset \mathbb{R}^{n}.$$
The hyperplanes $\sum_{i \in I} a_{i} = 1$, $I \subset [n], |I| \geq 2$, (called \textit{walls}) cut the polytope ${ \mathscr{A}}_{n}$ into \textit{chambers}. The Hassett compactification depends on a chamber only \cite[Proposition 5.1]{Has03}.
Combinatorial stratification of the space $\overline{\mathcal{M}}_{0, \mathcal{A}}$ looks similarly to that of the Deligne--Mumford's with the only difference --- some of the marked points now can coincide \citep{Cey07}.
More precisely, a \textit{weighted tree} $(\gamma, I)$ is an ordered $k$-partition $I_{1}\sqcup \dots \sqcup I_{k} = [n]$ and a tree $\gamma$ with $k$ ordered leaves marked by elements of the partition such that (1) $\sum_{j \in I_{m}} a_{j} \leq 1$ for any $m$, and (2) for each vertex, the number of emanating bold edges plus the total weight is greater than $2$. Open strata are enumerated by weighted trees: the stratum of the space $\overline{\mathcal{M}}_{0,\mathcal{A}}$ corresponding to a weighted tree $({\gamma}, I)$ consists of curves whose irreducible components form the tree $\gamma$ and collisions of sections form the partition $I$. Closure of this stratum is denoted by $\D_{({\gamma}, I)}$.
\subsection{Polygon spaces as compactifications of ${\mathcal M}_{0,n}$}\label{FlePolcom}
Assume that an $n$-tuple of positive real numbers $\mathcal{L} = (l_{1},...,l_{n})$ is fixed. We associate with it a \textit{flexible polygon}, that is, $n$ rigid bars of lengths $l_{i}$ connected in a cyclic chain by revolving joints. A \textit{configuration} of $\mathcal{L}$ is an $n$-tuple of points $(q_{1},...,q_{n}), \; q_i \in \mathbb{R}^3,$ with $|q_i q_{i+1}|=l_{i}, \; \; |q_{n} q_{1}|=l_{n}$.
The following two definitions for the \textit{polygon space}, or the \textit{moduli space of the flexible polygon} are equivalent:
\begin{definition} \cite{HKn} \begin{itemize} \item[I.]The moduli space $M_{\mathcal{L}}$ is a set of all configurations of $\mathcal{L}$ modulo orientation preserving isometries of $\mathbb{R}^3$.
\item[II.] Alternatively, the space $M_{ \mathcal{L}}$ equals the quotient of the space $$\left\{(u_1,...,u_n) \in (\mathbb{S}^2)^n : \sum_{i=1}^n l_iu_i=0\right\}$$ by the diagonal action of the group $\SO_3(\mathbb{R})$. \end{itemize} \end{definition}
The second definition shows that the space $M_{\mathcal{L}}$ does not depend on the ordering of $\{l_1,...,l_n\}$; however, it does depend on the values of $l_i$.
Let us consider the \textit{parameter space} $$\left\{ (l_{1}, \dots, l_{n}) \in \mathbb{R}^{n}_{>0}:\ l_{i}< \sum_{j\neq i} l_{j} \text{ for }i=1, \dots, n\right\}.$$
This space is cut into open \textit{chambers} by \textit{walls}. The latter are hyperplanes with defining equations $$\sum_{i\in I} l_i = \sum_{j\notin I} l_j.$$
The diffeomorphic type of $M_{ \mathcal{L}}$ depends only on the chamber containing $\mathcal{L}$. For a point $\mathcal{L}$ lying strictly inside some chamber, the space $M_{\mathcal{L}}$ is a smooth $(2n-6)$-dimensional algebraic variety \cite{Kl}. In this case we say that the length vector is \textit{generic}.
\begin{definition} For a generic length vector $\mathcal{L}$, we call a subset $J\subset [n]$ {\it long} if $$\sum_{i \in J} l_i > \sum_{i\notin J} l_i.$$ Otherwise, $J$ is called \textit{short}. The set of all short sets we denote by $SHORT(\mathcal{L})$. \end{definition}
Each subset of a short set is also short, therefore $SHORT(\mathcal{L})$ is a (threshold Alexander self-dual) simplicial complex. Rephrasing the above, the diffeomorphic type of $M_{ \mathcal{L}}$ is defined by the simplicial complex $SHORT(\mathcal{L})$.
\section{ASD and preASD compactifications}\label{ASDpreASD}
\subsection{ASD and preASD simplicial complexes} Simplicial complexes provide a necessary combinatorial framework for the description of the category of smooth modular compactifications of ${\mathcal M}_{0,n}$.
A \textit{simplicial complex} (a \textit{complex,} for short) $K$ is a subset of $2^{[n]}$ with the hereditary property: $A \subset B\in K$ implies $A \in K$. Elements of $K$ are called \textit{faces} of the complex. Elements of $2^{[n]} \setminus K$ are called \textit{non-faces}. The maximal (by inclusion) faces are called \textit{facets}.
We assume that the set of $0$-faces (the set of vertices) of a complex is $[n]$. The complex $2^{[n]}$ is denoted by $\Delta_{n-1}$. Its $k$-skeleton is denoted by $\Delta_{n-1}^{k}$. {In particular, $\Delta_{n-1}^{n-2}$ is the boundary complex of the simplex $\Delta_{n-1}$.}
\begin{definition} For a complex $K \subset 2^{[n]}$, its \textit{Alexander dual} is the simplicial complex $$K^{\circ}:= \{A \subset [n]:\; A^{c} \not\in K\} = \{A^{c}:\; A \in 2^{[n]}\backslash K\}.$$ Here and in the sequel, $A^c=[n]\setminus A$ is the complement of $A$. A complex $K$ is \textit{Alexander self-dual (an ASD complex)} if $K^{\circ}=K$. A \textit{pre Alexander self-dual (a pre ASD)} complex is a complex contained in some ASD complex. \end{definition}
In other words, ASD complexes (pre ASD complexes, respectively) are characterized by the condition:
for any partition $[n]=A\sqcup B$, {exactly one} (at most one, respectively) of $A$, $B$ is a face.
Some ASD complexes are \textit{threshold complexes}: they equal $SHORT(\mathcal{L})$ for some generic weight vectors $\mathcal{L}$ (Section \ref{FlePolcom}). It is known that threshold ASD complexes exhaust all ASD complexes for $n \leq 5$. However, for bigger $n$ this is no longer true. Moreover, for $n \rightarrow \infty$ the percentage of threshold ASD complexes tends to zero.
To produce new examples of ASD complexes, we use \textit{flips}: \begin{definition}\cite{PanGal} For an ASD complex $K$ and a facet $A \in K$ we build a new ASD complex $$\flip_{A}(K):= (K\backslash A) \cup A^{c}.$$ \end{definition}
It is easy to see that \begin{proposition}\label{LemmaNotLonger2} (1) \cite{PanGal} Inverse of a flip is also some flip.
(2) \cite{PanGal} Any two ASD complexes are connected by a sequence of flips. (3) For any ASD complex $K$ there exists a threshold ASD complex $K'$ that can be obtained from $K$ by a sequence of flips with some $A_{i}\subset [n]$ such that $|A_{i}| > 2, |A_{i}^c| > 2$.
\end{proposition}
\begin{proof} We prove (3). It is sufficient to show that for any ASD complex, there exists a threshold ASD complex with the same collection of $2$-element non-faces. For this, let us observe that any two non-faces of an ASD complex necessarily intersect. Therefore, all possible collections of $2$-element non-faces of an ASD complex (up to renumbering) are: \begin{enumerate} \item empty set; \item $(12),\ (23),\ (31)$; \item $(12),(13),\dots ,(1k)$. \end{enumerate} It is easy to find appropriate threshold ASD complexes for all these cases. \end{proof}
ASD complexes appear in the game theory literature as ``simple games with constant sum'' (see \citep{vNMor}). One imagines $n$ players and all possible ways of partitioning them into two teams. The teams compete, and a team \textit{looses} if it belongs to $K$. In the language of flexible polygons, a short set is a loosing team.
\textbf{Contraction, or freezing operation.} Given an ASD complex $K$, let us build a new ASD complex $K_{(ij)}$ with $n-1$ vertices $\{1,...,\widehat{{i}},...,\widehat{{j}},...,n,(i,j)\}$ by \textit{contracting the edge} $\{i,j\}\in K$, or \textit{freezing} $i$ and $j$ together.
The formal definition is: for $A\subset \{1,...,\widehat{{i}},...,\widehat{{j}},...,n\}$, $A\in K_{(ij)}$ iff $A\in K$, and
$A\cup \{(ij)\}\in K_{(ij)}$ iff $A \cup \{i,j\}\in K$.
Contraction $K_I$ of any other face $I\in K$ is defined analogously.
Informally, in the language of simple game, contraction of an edge means making one player out of two. In the language of flexible polygons, ``freezing'' means producing one new edge out of two old ones (the lengths sum up).
\begin{figure}
\caption{Contraction of $\{1,2\}$ in a simplicial complex}
\label{Figcont}
\end{figure}
\subsection{Smooth extremal assignment compactifications}\label{Smycom}
Now we review the results of \cite{Smyth13} and \cite{Han15}, and indicate a relation with preASD complexes.
For a scheme $B$, consider the space $\mathcal{U}_{\,0, n}(B)$ of all flat, proper, finitely-presented morphisms $\pi: \;\mathcal{C} \rightarrow B$ with $n$ sections $\{s_{i}\}_{i \in [n]}$, and connected, reduced, one-dimensional geometric fibers of genus zero. Denote by $\mathcal{V}_{0, n}$ the irreducible component of $\mathcal{U}_{\,0,n}$ that contains ${\mathcal M}_{0,n}$. \begin{definition} A modular compactification of ${\mathcal M}_{0,n}$ is an open substack $\mathcal{X} \subset \mathcal{V}_{0,n}$ that is proper over $\mathbb{Z}$. A modular compactification is stable if every geometric point $(\pi:\;\mathcal{C} \rightarrow B; s_{1}, \dots, s_{n})$ is stable. We call a modular compactification {smooth} if it is a smooth algebraic variety. \end{definition}
\begin{definition} A smooth extremal assignment $\mathcal{Z}$ over $\overline{\mathcal{M}}_{0,n}$ is an assignment to each stable tree with $n$ leaves a subset of its vertices $$\gamma \mapsto \mathcal{Z}(\gamma) \subset Vert(\gamma)$$ such that: \begin{enumerate} \item for any tree $\gamma$, the assignment is a proper subset of vertices: $\mathcal{Z}(\gamma) \subsetneqq Vert(\gamma)$,
\item for any contraction $\gamma \rightsquigarrow \tau$ with $\{v_{i}\}_{i \in I} \subset Vert(\gamma)$ contracted to $v \in Vert(\tau)$, we have $v_{i}\in \mathcal{Z}(\gamma)$ for all $i \in I$ if and only if $v \in \mathcal{Z}(\tau)$. \item for any tree $\gamma$ and $v \in \mathcal{Z}(\gamma)$ there exists a two-vertex tree $\gamma'$ and $v'\in \mathcal{Z}(\gamma')$ such that $$\gamma' \rightsquigarrow \gamma \text{ and }v' \rightsquigarrow v.$$\end{enumerate}\end{definition} \begin{definition}\label{DefAss} Assume that $\mathcal{Z}$ is a smooth extremal assignment. A curve $(\pi:\;\mathcal{C} \rightarrow B; s_{1}, \dots, s_{n})$ is $\mathcal{Z}$--stable if it can be obtained from some Deligne--Mumford stable curve $(\pi':\;\mathcal{C}' \rightarrow B'; s_{1}', \dots, s_{n}')$ by (maximal) blowing down irreducible components of the curve $\mathcal{C}'$ corresponding to the vertices from the set $\mathcal{Z}(\gamma(\mathcal{C}'))$. \end{definition}
A smooth assignment is completely defined by its value on two-vertex stable trees with $n$ leaves. The latter bijectively correspond to unordered partitions $A\sqcup A^{c} = [n]$ with $|A|, |A^c| > 1$: sets $A$ and $A^{c}$ are affixed to two vertices of the tree. The first condition of Definition \ref{DefAss} implies that no more than one of $A$ and $A^{c}$ is ``assigned''. One concludes that preASD complexes are in bijection with smooth assignments.
All possible modular compactifications of ${\mathcal M}_{0, n}$ are parametrized by smooth extremal assignments:
\begin{theorem}\cite[Theorems 1.9 \& 1.21]{Smyth13} and \cite[Theorem 1.3]{Han15} \begin{itemize} \item For any smooth extremal assignment $\mathcal{Z}$ of ${\mathcal M}_{0,n}$, or equivalently, for any preASD complex $K$, there exists a stack $\overline{\mathcal{M}}_{0,\mathcal{Z}} = \overline{\mathcal{M}}_{0,K} \subset \mathcal{V}_{0,n}$ parameterizing all $\mathcal{Z}$--stable rational curves. \item For any smooth modular compactification $\mathcal{X} \subset \mathcal{V}_{0,n}$, there exist a smooth extremal assignment $\mathcal{Z}$ (a preASD complex $K$) such that $\mathcal{X} = \overline{\mathcal{M}}_{0,\mathcal{Z}} = \overline{\mathcal{M}}_{0,K}$. \end{itemize} \end{theorem}
There are two different ways to look at a moduli spaces. In the present paper we look at the moduli space as at a smooth algebraic variety equipped with $n$ sections (\textit{fine moduli space}). The other way is to look at it as at a smooth algebraic variety (\textit{coarse moduli space}). Different preASD complexes give rise to different fine moduli spaces. However, two different complexes can yield isomorphic coarse moduli spaces.
Indeed, consider two preASD complexes $K$ and $K\cup\{ij\}$ (we abbreviate the latter as $K+(ij)$), assuming that $\{ij\}\notin K$. The corresponding algebraic varieties $\overline{\mathcal{M}}_{0, K}$ and $\overline{\mathcal{M}}_{0, K + (ij)}$ are isomorphic. A vivid explanation is: to let a couple of marked points to collide is the same as to add a nodal curve with these two points sitting alone on an irreducible component. Indeed, this irreducible component would have exactly three special points, and $\PSL_{2}$ acts transitively on triples.
\begin{theorem}\cite[Statements 7.6--7.10]{Han15}\label{objects_fixedlevel} The set of smooth modular compactifications of ${\mathcal M}_{0,n}$ is in a bijection with objects of the ${\bf preASD}_{n}/ \thicksim$, where $K\thicksim L$ whenever $K\setminus L$ and $L\setminus K$ consist of two-element sets only. \end{theorem}
\begin{example}\label{example} PreASD complexes and corresponding compactifications. \begin{enumerate} \item the 0-skeleton $\Delta_{n-1}^{0} = [n]$ of the simplex $\Delta_{n-1}$ corresponds to the Deligne--Mumford compactification; \item {the complex $\mathcal{P}_{n} : = {\bf pt} \sqcup \Delta_{n-2}^{n-3}$ (disjoint union of a point and the boundary of a simplex $\Delta_{n-2}$)} is ASD. It corresponds to the Hassett weights $(1, \varepsilon, \dots, \varepsilon)$; this compactification is isomorphic to $\mathbb{P}^{n-3}$; \item the Losev--Manin compactification $\overline{\mathcal{M}}_{0, n}^{LM}$ \cite{LosMan00} corresponds to the weights $(1,1,\varepsilon, \dots, \varepsilon)$ and to the complex ${\bf pt}_{1} \sqcup {\bf pt}_{2} \sqcup \Delta_{n-3}$; \item the space $(\mathbb{P}^{1})^{n-3}$ corresponds to weights $(1,1,1,\varepsilon, \dots, \varepsilon)$, and to the complex ${\bf pt}_{1} \sqcup {\bf pt}_{2} \sqcup {\bf pt}_{3} \sqcup \Delta_{n-4}$. \end{enumerate} \end{example}
\subsection{ASD compactifications via stable point configurations}
ASD compactifications can be explained in a self-contained way, without referring to \cite{Smyth13}.
Fix an ASD complex $K$ and consider configurations of $n$ (not necessarily all distinct) points $p_1,...,p_n$ in the projective line. A configuration is called {\em stable} if the index set of each collection of coinciding points belongs to $K$. That is, whenever $p_{i_1}=...=p_{i_k}$, we have $\{i_1,...,i_k\}\in K$.
Denote by $STABLE(K)$ the space of stable configurations in the complex projective line. The group $\mathrm{PSL}_{2}(\mathbb{C})$ acts naturally on this space. Set $$\overline{\mathcal{M}}_{0, K}:=STABLE(K)/\mathrm{PSL}_{2}(\mathbb{C}).$$
If $K$ is a threshold complex, that is, arises from some flexible polygon $\mathcal{L}$, then the space $\overline{\mathcal{M}}_{0, K}$ is isomorphic to the polygon space $M_{\mathcal{L}}$ \cite{Kl}.
Although the next theorem fits in a broader context of \cite{Smyth13}, we give here its elementary proof. \begin{theorem}\label{ThmSmoothComp} The space $\overline{\mathcal{M}}_{0, K}$ is a compact smooth variety with a natural complex structure. \end{theorem} \begin{proof}$\;$ \textbf{Smoothness.} For a distinct triple of indices $i,j,k \in [n]$, denote by $U_{i,j,k}$ the subset of $\overline{\mathcal{M}}_{0, K}$ defined by $p_i\neq p_j,$ $p_j \neq p_k$, and $p_i \neq p_k$. For each of $U_{i,j,k}$, we get rid of the action of the group $\mathrm{PSL}_{2}(\mathbb{C})$, setting $$U_{i,j,k}=\big\{(p_1,...,p_n)\in \overline{\mathcal{M}}_{0, K}:\; p_i=0,p_j=1,\text{ and }p_k=\infty \big\}.$$
Clearly, each of the charts $U_{i,j,k}$ is an open smooth manifold. Since all the $U_{i,j,k}$ cover $\overline{\mathcal{M}}_{0, K}$, smoothness is proven.
\textbf{Compactness.} Let us show that each sequence of $n$-tuples has a converging subsequence.
Assume the contrary. Without loss of generality, we may assume that the sequence $(p_1^i=0,p_2^i=1,p_3^i=\infty,p_4^i,...,p_n^i)_{i=1}^{\infty}$ has no converging subsequence. We may assume that for some set $I \notin K$, all $p^i_j$ with $j\in I$ converge to a common point. We say that we have a \textit{collapsing long set} $I$. This notion depends on the choice of a chart. We may assume that our collapsing long set has the minimal cardinality among all long sets that can collapse without a limit (that is, violate compactness) for this complex $K$. We may assume that $I=\{3,4,5,...,k\}$.
This long set can contain at most one of the points $p_1,p_2,p_3$. We consider the case when it contains $p_3$; other cases are treated similarly.
That is, all the points $p^i_4,...,p^i_k$ tend to $\infty$. Denote by $C_i$ the circle with the minimal radius embracing the points $p^i_3=\infty ,p^i_4,p^i_5,...,p^i_k$. The circle contains at least two points of $p^i_4,...,p^i_k, p_3=\infty$. Apply a transform $\phi_i \in \mathrm{PSL}_{2}(\mathbb{C})$ which turns the radius of $C_i$ to $1$, and keeps at least two of the points $p^i_4,...,p^i_k, p_3=\infty$ away from each other. In this new chart the cardinality of the collapsing long set gets smaller. A contradiction to the minimality assumption. \end{proof}
A natural question is: what if one takes a simplicial complex (not a self-dual one), and cooks the analogous quotient space. Some heuristics are: if the complex contains simultaneously some set $A$ and its complement $[n]\setminus A$, we have a stable tuple with a non-trivial stabilizer in $\mathrm{PSL}_{2}(\mathbb{C})$, so the factor has a natural nontrivial orbifold structure. If a simplicial complex is smaller than some ASD complex $K'$, and therefore, we get a proper open subset of $\overline{\mathcal{M}}_{0, K'}$, that is, we lose compactness.
\subsection{Perfect cycles} Assume that we have an ASD complex $K$ and the associated compactification $\overline{\mathcal{M}}_{0, K}$. Let $K_I$ be the contraction of some face $I\in K$. Since the variety $\overline{\mathcal{M}}_{0, K_I}$ naturally embeds in $ \overline{\mathcal{M}}_{0, K}$, the contraction procedure gives rise to a number of subvarieties of $\overline{\mathcal{M}}_{0, K}$. These varieties (1) ``lie on the boundary'' \footnote{That is do not intersect the initial space ${\mathcal M}_{0, n}$.} and (2) generate the Chow ring (Theorem \ref{ASDChow} ). Let us look at them in more detail.
An \textit{elementary perfect cycle} $(ij)=(ij)_{K}\subset \overline{\mathcal{M}}_{0, K}$ is defined as
$$(ij)=(ij)_K=\{(p_1,...,p_n)\in \overline{\mathcal{M}}_{0, K}: p_i=p_j\}.$$
Let $[n]=A_1\sqcup...\sqcup A_k$ be an unordered partition of $[n]$. A \textit{perfect cycle} associated to the partition \begin{align*} (A_1)\cdot...\cdot(A_k)&=(A_1)_{K}\cdot...\cdot(A_k)_{K} = \\ &= \{(p_1,...,p_n)\in \overline{\mathcal{M}}_{0, K}: i,j \in A_m \Rightarrow p_i=p_j\}. \end{align*}
Each perfect cycle is isomorphic to $\overline{\mathcal{M}}_{0, K'}$ for some complex $K'$ obtained from $K$ by a series of contractions.
Singletons play no role, so we omit all one-element sets $A_i$ from our notation. Consequently, all the perfect cycles are labeled by partitions of some subset of $[n]$ such that all the $A_i$ have at least two elements.
Note that for arbitrary $A \in K$, the complex $K_A$ might be ill-defined. This happens if $A \notin K$. In this case the associated perfect cycle $(A)$ is empty.
For each perfect cycle there is an associated Poincaré dual element in the cohomology ring. These dual elements we denote by the same symbols as the perfect cycles.
The following rules allow to compute the cup-product of perfect cycles:
\begin{proposition}\label{ComputRules}$\;$ \begin{enumerate}
\item Let $A$ and $B$ be disjoint subsets of $[n]$. Then
\begin{enumerate}
\item $(A)\smile (B)=(A)\cdot (B).$
\item $(Ai)\smile (Bi)=(ABi).$
\end{enumerate}
\item For $A\notin K$, we have $(A)=0$. If one of $A_k$ is a non-face of $K$, then $(A_1)\cdot...\cdot(A_k)=0$.
\item The four-term relation:
$(ij)+(kl)=(jk)+(il)$ holds for any distinct $i,j,k,l$.
\end{enumerate} \end{proposition}
\begin{proof} In the cases (1) and (2) we have a transversal intersection of holomorphically embedded complex varieties. The item (3) will be proven in Theorem \ref{ASDChow}. \end{proof}
Examples: $$(123)\cdot (345)=(12345);\ \ (12)\cdot (34) \cdot (23)=(1234).$$ A more sophisticated computation: $$(12)\cdot(12) = (12)\cdot \big( (13) + (24) - (34) \big) = (123) + (124) -(12)\cdot(34).$$
\begin{proposition}\label{PropPerf}A cup product of perfect cycles is a perfect cycle.
\end{proposition} \textit{Proof.} Clearly, each perfect cycle is a product of elementary ones. Let us prove that the product of two perfect cycles is an integer linear combination of perfect cycles. We may assume that the second factor is an elementary perfect cycle, say, $(12)$. Let the first factor be $(A_1)\cdot(A_2)\cdot(A_3)\cdot\ldots\cdot(A_k)$.
We need the following case analysis: \begin{enumerate} \item If at least one of $1,2$ does not belong to $\bigcup A_i$, the product is a perfect cycle by Proposition \ref{ComputRules}, (1). \item If $1$ and $2$ belong to different $A_i$, we use the following:
for any perfect cycle $(A_1)\cdot(A_2)$ with $ i\in A_1, j\in A_2$, we have $$(A_1)\cdot(A_2)\smile (ij)=(A_1)\smile (ij)\smile (A_2)= (A_1j)\smile (A_2)= (A_1\cup A_2).$$
\item Finally, assume that $1,2 \in A_1$. Choose $i\notin A_1,\ \ j\notin A_1$ such that $i$ and $j$ do not belong to one and the same $A_k$. By Proposition \ref{ComputRules}, (3), $$(A_1)\cdot(A_2)\cdot(A_3)\cdot\dotso\cdot(A_k)\smile (12)=(A_1)\cdot(A_2)\cdot(A_3)\cdot\ldots\cdot(A_k)\smile \big((1i)+(2j)-(ij)\big).$$ After expanding the brackets, one reduces this to the above cases.\qed \end{enumerate}
\begin{lemma}\label{LemmaTriple} For an ASD complex $K$, let $A\sqcup B\sqcup C=[n]$ be a partition of $[n]$ into three faces. Then $(A)\cdot(B)\cdot(C)=1$ in the graded component ${\bf A}^{n-3}_K$ of the Chow ring, which is canonically identified with $\mathbb{Z}$. \end{lemma} \begin{proof} Indeed, the cycles $(A)$, $(B)$, and $(C)$ intersect transversally at a unique point. \end{proof}
Now we see that the set of perfect cycles is closed under cup-product. In the next section we show that the Chow ring equals the ring of perfect cycles.
\subsection{Flips and blow ups. }
Let $K$ be an ASD complex, and let $A \subset [n]$ be its facet.
\begin{lemma}\label{LemmaPerfCyclProj}
The perfect cycle $(A)$ is isomorphic to $\overline{\mathcal{M}}_{0, \mathcal{P}_{|A^c|+1}} \cong \mathbb{P}^{|A^c|-2}$. \end{lemma}
\begin{proof} Contraction of $A$ gives the complex ${\bf pt}\;\sqcup\;\Delta_{|A^c|}^{} = \mathcal{P}_{|A^c|+1}$ from the Example \ref{example}, (2). \end{proof}
\begin{lemma}\label{blowupVSflip} For an ASD complex $K$ and its facet $A$, there are two blow up morphisms $$\pi_{ A }: \overline{\mathcal{M}}_{0,K\backslash A} \rightarrow \overline{\mathcal{M}}_{0, K} \text{ and }\pi_{A^c}: \overline{\mathcal{M}}_{0,K\backslash A} \rightarrow \overline{\mathcal{M}}_{0, \flip_{A} (K)}.$$
The centers of these blow ups are the perfect cycles $(A)$ and $(A^c)$ respectively. The exceptional divisors are equal: $\D_{A} = \D_{A^c}$. Both are isomorphic to $\overline{\mathcal{M}}_{0, \mathcal{P}_{|A|+1}} \times \overline{\mathcal{M}}_{0, \mathcal{P}_{|A^c|+1}} \cong \mathbb{P}^{|A|} \times \mathbb{P}^{|A^c|}$. The maps $\pi_{A}|_{\D_{A}}$ and $\pi_{A^c}|_{\D_{A^c}}$ are projections to the first and the second components respectively. \end{lemma}
The \textit{proof} literally repeats \cite[Corollary 3.5]{Has03}: $K$--stable but not $K_{A}$($K_{A^c}$)--stable curves have two connected components. The marked points with indices from the set $A$ lie on one of the irreducible components, and marked points with indices from the set $A^c$ lie on the other. \qed
\begin{corollary} For an ASD complex $K$ and its facet $A$, the algebraic varieties $\overline{\mathcal{M}}_{0, K}$ and $\overline{\mathcal{M}}_{0, K \backslash A}$ are HI--schemes, i.e., the canonical map from the Chow ring to the cohomology ring is an isomorphism. \end{corollary} \textit{Proof.} This follows from Lemma \ref{blowupVSflip} and Theorem~\ref{ChowCoh}.\qed
\section{Chow rings of ASD compactifications}
As it was already mentioned, many examples of ASD compactifications are polygon spaces, that is, come from a threshold ASD complex. Their Chow rings were computed in \cite{HKn}. A more relevant to the present paper presentation of the ring is given in \cite{NP}. We recall it below.
\begin{definition} Let ${\bf A}^{*}_{univ} = {\bf A}^{*}_{univ, n}$ be the ring
$$\mathbb{Z}\big[(I):\; I \subset [n], 2 \leq |I| \leq n-2 \big]$$ factorized by relations: \begin{enumerate} \item ``\textit{The four-term relations}'': $(ij)+(kl)-(ik)-(jl)=0$ for any $i,j,k,l \in [n]$.
\item ``\textit{The multiplication rule}'': $(Ik)\cdot(Jk) = (IJk)$ for any disjoint $I, J \subset [n]$ not containing element $k$.
\end{enumerate} \end{definition}
There is a natural graded ring homomorphism from ${\bf A}^*_{univ}$ to the Chow ring of an ASD-compactification that sends each of the generators $(I)$ to the corresponding perfect cycle.
\begin{theorem}\cite{NP}\label{ChowForPolygons} The Chow ring (it equals the cohomology ring) of a polygon space equals the ring ${\bf A}^*_{univ}$ factorized by $$(I)=0 \ \ \ \hbox{whenever $I$ is a long set.}$$ \end{theorem}
The following generalization of Theorem \ref{ChowForPolygons} is the first main result of the paper:
\begin{theorem}\label{ASDChow} For an ASD complex $L$, the Chow ring ${\bf A}^{*}_{L}:= {\bf A}^{*}(\overline{\mathcal{M}}_{0, L})$ of the moduli space $\overline{\mathcal{M}}_{0, L}$ is isomorphic to the quotient ${\bf A}^{*}_{univ}$ by the ideal $\mathcal{I}_{L}:= \big\langle (I):\;I\not\in L \big\rangle$.
\end{theorem}
The idea of the proof is: the claim is true for threshold ASD complexes (i.e., for polygon spaces), and each ASD complex is achievable from a threshold ASD complex by a sequence of flips. Therefore it is sufficient to look at a unique flip. Let us consider an ASD complex $K + B$ where $B\notin K$ is a facet in $K + B$. Set $A:=[n]\setminus B$, and consider the ASD complex $K + A= \flip_{B}(K+B)$. \begin{center} \begin{tikzcd}[column sep=small] & K \arrow[dl, hook] \arrow[dr, hook] & \\ K+ B \arrow[rr, dashrightarrow, "\flip_{B}"] & & K+A \end{tikzcd} \end{center}
We are going to prove that if the claim of the theorem holds true for $K+B$, then it also holds for $K+A$.
By Lemma \ref{blowupVSflip},
the space $\overline{\mathcal{M}}_{0, K}$ is the blow up of $\overline{\mathcal{M}}_{0, K+B}$ along the subvariety $(B)$ and the blow up of $\overline{\mathcal{M}}_{0, K+A}$ along the subvariety $(A)$. The diagram of the blow ups looks as follows: \begin{center} \begin{tikzcd} (B)\arrow[d, hook, "i_{B}"] & \D \arrow[hook]{d}{j_{A} = j_{B}} \arrow{r}{g_{A}}\arrow{l}{g_{B}} & (A) \arrow[hook]{d}{i_{A}} \\
\overline{\mathcal{M}}_{0, K+B} & \overline{\mathcal{M}}_{0, K} \arrow[twoheadrightarrow]{r}{\pi_{A}}\arrow[l, two heads, "\pi_{B}"] & \overline{\mathcal{M}}_{0, K+A } \end{tikzcd} \end{center} The induced diagram of Chow rings is: \begin{center} \begin{tikzcd}
{\bf A}^{*}_{(B)} = {\bf A}^{*}_{\mathcal{P}_{|A|+1}} \arrow{r}{g_{B}^{*}} & {\bf A}^{*}_{\mathcal{P}_{|A|+1}}\times {\bf A}^{*}_{\mathcal{P}_{|B|+1}} & {\bf A}^{*}_{(A)} = {\bf A}^{*}_{\mathcal{P}_{|B|+1}} \arrow{l}{g_{A}^{*}} \\
{\bf A}^{*}_{K + B}\arrow[u, "i_{B}^{*}"]\arrow[r, hookrightarrow, "\pi_{B}^{*}" description] & {\bf A}^{*}_{K} \arrow[hook]{u}{j_{A}^{*} = j_{B}^{*}} & {\bf A}^{*}_{K + A} \arrow[l, hookrightarrow , "\pi_{A}^{*}" description]\arrow[u, "i_{A}^{*}" description] \end{tikzcd} \end{center}
Let ${\bf A}^{*}_{K+ A, comb}$ be the quotient of ${\bf A}^{*}_{univ}$ by the ideal $\mathcal{I}_{K+ A}$. We have a natural graded ring homomorphism $$\alpha = \alpha_{K+A}:{\bf A}^{*}_{K+ A, comb} \rightarrow {\bf A}^{*}_{K+ A}=:{\bf A}^{*}_{K+ A, alg},$$ where the map $\alpha$ sends each symbol $(I)$ to the associated perfect cycle.
A remark on notation: as a general rule, all objects related to ${\bf A}^{*}_{K+ A, comb}$ we mark with a subscript \textit{``comb''}, and objects related to ${\bf A}^{*}_{K+ A, alg}$ we mark with \textit{``alg''}.
We shall show that $\alpha$ is an isomorphism. The outline of the proof is:
\begin{enumerate} \item The ring ${\bf A}^{*}_{K+A,alg}$ is generated by the first graded component. (The ring ${\bf A}^{*}_{K+A,comb}$ is also generated by the first graded component; this is clear by construction.) \item The restriction of $\alpha$ to the first graded components is a group isomorphism. Therefore, $\alpha$ is surjective. \item The map $\alpha$ is injective. \end{enumerate}
\iffalse We make use of the general structure theory for Chow rings under blow ups (the main theorems we collect in Appendix). \fi
\begin{lemma} \label{OneGen} The ring ${\bf A}^{*}_{K+A, alg}$ is generated by the group ${\bf A}^{1}_{K+A, alg}$. \end{lemma}
\begin{proof} By Theorem \ref{NewChowRing} $${\bf A}^{*}_{K} \cong \frac{ {\bf A}^{*}_{K+A, alg}[T]}{\big( f_{A}(T), T\cdot \ker (i_{A}^{*})\big)}.$$
Observe that:
\begin{itemize} \item The zero graded components of ${\bf A}^{*}_{K+A, alg}, {\bf A}^{*}_{K+A, comb}$ equals $\mathbb{Z}$. \item The map $\pi_{A}^*:\;{\bf A}^{*}_{K+A, alg} \rightarrow {\bf A}^{*}_{K}$ is a homomorphism of graded rings. Moreover, the variable $T$ stands for the additive inverse of the class of the exceptional divisor $\D$. And so, $T$ a degree one homogeneous element.
\item Since $i_{A}^*$ is the multiplication by the cycle $(A)$, the kernel $\ker(i_{A}^{*})$ equals the annihilator $\Ann(A)_{alg}$ in the ring ${\bf A}^{*}_{K+A, alg}$. {Since the space $\overline{\mathcal{M}}_{0, K+A}$ is an HI-scheme}, the degree of the ideal $\Ann(A)_{alg}$ is strictly positive.
\item The polynomial $f_{A}(T)$ is a homogeneous element whose degree equals the degree $\deg_{T}(f_{A}(T))$. Besides, its coefficients are generated by elements from the first graded component since they all belong to the ring $\alpha ({\bf A}^{*}_{K+A, comb})$. \end{itemize}
Denote by $\langle{\bf A}^1_{K+A, alg}\rangle$ the subalgebra of ${\bf A}_{K+A, alg}$ generated by the first graded component.
First observe that the restriction of the map ${\bf A}^{*}_{K+A, alg}[T] \rightarrow {\bf A}^{*}_{K}$ to the first graded components is injective.
Assuming that the lemma is not true,
consider a homogeneous element $r$ of the ring ${\bf A}^{*}_{K+A, alg}$ with minimal degree among all not belonging to $\langle{\bf A}^{1}_{K+A, alg}\rangle$. There exist elements $b_{i}\in \langle{\bf A}^{1}_{K+A, alg}\rangle$ such that $b_{p}\cdot T^{p} + \dots + b_{1}\cdot T + b_{0} = r$ in the ring ${\bf A}^{*}_{K+A, alg}[T]$. The elements $b_{i}$ are necessarily homogeneous.
Equivalently, $b_{p}\cdot T^{p} + \dots + b_{1}\cdot T + b_{0} - r$ belongs to the ideal $\big(f_A(T), T\cdot \Ann(A_{alg})\big)$. Therefore $b_{p}\cdot T^{p} + \dots + b_{1}\cdot T + b_{0} - r=x\cdot f_A(T) + y \cdot T \cdot i$ with some $x, y \in R[T]$ and $i \in \Ann(A_{alg})$.
Setting $T=0$, we get $b_{0} - r = x_{0}\cdot f_{0}$. If the element $x_{0}$ belongs to $\langle{\bf A}^{1}_{K+A, alg}\rangle$, then we are done. Assume the contrary. Then from the minimality assumption we get the following inequalities: $\deg (b_0 - r) = \deg (x_{0}\cdot f_{0}) > \deg(x_{0}) \geq \deg(r)$. A contradiction. \end{proof}
\begin{lemma} \label{FirstGroup} For any ASD complex $L$ the groups ${\bf A}^1_{L, comb}$ and ${\bf A}^1_{L, alg}$ are isomorphic. The isomorphism is induced by the homomorphism $\alpha_{L}$. \end{lemma} The \textit{proof} analyses how do these groups change under flips.
We know that the claim is true for threshold complexes. Due to Lemma \ref{LemmaNotLonger2} we may consider flips only with $n-2>|A|>2$. Again, we suppose that the claim is true for the complex $K+B$ and will prove for the complex $K+A$ with $A \sqcup B = [n]$. Under such flips ${\bf A}^{1}_{comb}$ does not change. The group ${\bf A}^{1}$ does not change neither. This becomes clear with the following two short exact sequences (see Theorem \ref{SESBlow},e): \begin{align*}
0 &\rightarrow {\bf A}_{n-4}\big(\overline{\mathcal{M}}_{0, \mathcal{P}_{|A|+1}}\big) \rightarrow {\bf A}_{n-4}\big( \overline{\mathcal{M}}_{0, \mathcal{P}_{|A|+1} } \times \overline{\mathcal{M}}_{0, \mathcal{P}_{|B|+1} } \big) \oplus {\bf A}_{n-4}\big( \overline{\mathcal{M}}_{0, K + B} \big) \rightarrow {\bf A}_{n-4}\big(\overline{\mathcal{M}}_{0, K}\big) \rightarrow 0,\\
0 &\rightarrow {\bf A}_{n-4}\big(\overline{\mathcal{M}}_{0, \mathcal{P}_{|B|+1}}\big) \rightarrow {\bf A}_{n-4}\big( \overline{\mathcal{M}}_{0, \mathcal{P}_{|A|+1} } \times \overline{\mathcal{M}}_{0, \mathcal{P}_{|B|+1} } \big) \oplus {\bf A}_{n-4}\big( \overline{\mathcal{M}}_{0, K + A} \big) \rightarrow {\bf A}_{n-4}\big(\overline{\mathcal{M}}_{0, K}\big) \rightarrow 0. \end{align*} \qed
Now we know that $\alpha:\;{\bf A}^{*}_{K+A, comb} \rightarrow {\bf A}^{*}_{K+A, alg}$ is surjective.
\begin{proposition}\label{FirstUniv} Let $\Gamma$ be a graph $Vert(\Gamma)=[n]$ which equals a tree with one extra edge. Assume that the unique cycle in $\Gamma$ has the odd length. Then the set of perfect cycles $\{(ij)\}$ corresponding to the edges of $\Gamma$ is a basis of the (free abelian) group ${\bf A}^{1}_{univ}$. \end{proposition} \begin{proof} Any element of the group ${\bf A}^{1}_{univ}$ by definition has a form $\sum_{ij} a_{ij} \cdot (ij)$ with the sum ranges over all edges of the complete graph on the set $[n]$. The four-term relation can be viewed as an alternating relation for a four-edge cycle. One concludes that analogous alternating relation holds for each cycle of even length. Example: $(ij)-(jk)+(kl)-(lm)+(mp)-(pi)=0$. Such a cycle may have repeating vertices. Therefore, if a graph has an even cycle, the perfect cycles associated to its edges are dependant.
It remains to observe that the graph $\Gamma$ is a maximal graph without even cycles. \end{proof}
By Theorem \ref{NewChowRing}, the Chow rings of the compactifications corresponding to complexes $K$, $K+A$, and $K+B$ are related in the following way:
$${\bf A}^{*}_{K} \cong \frac{ {\bf A}^{*}_{K+A, alg}[T]}{\big( f_{A}(T), T\cdot \ker (i_{A}^{*})\big)} \cong \frac{ {\bf A}^{*}_{K+B}[S]}{\big( f_{B}(S), S\cdot \ker (i_{B}^{*})\big)}.$$
Now we need an explicit description of the polynomials $f_{A}$ and $f_{B}$.
Assuming that $A=\{x, x_{2}, \dots, x_{a}\}$ $B=\{y, y_{2}, \dots, y_{b}\}$, where $|A|= a$ and $|B| = b$, take the generators \begin{align*} &\big\{(xy); (xy_{i}), i \in \{2, \dots, b\}; (x_{j}y), j \in \{2, \dots, a\}; (yy_{2}) \big\} \text{ for } {\bf A}^{*}_{K+ B} \text{ , and }\\ &\big\{(xy); (xy_{i}), i \in \{2, \dots, b\}; (x_{j}y), j \in \{2, \dots, a\}; (xx_{2}) \big\} \text{ for } {\bf A}^{*}_{K+ A, comb}. \end{align*}
\begin{figure}
\caption{111}
\label{Figbases}
\end{figure}
Denote by ${\mathcal A}$ the subring of the Chow rings ${\bf A}^{*}_{K+ A, comb}$ and ${\bf A}^{*}_{K+ B}$ generated by the elements $\{ (xy); (xy_{i}), i \in \{2, \dots, b-1\}; (x_{j}y), j \in \{2, \dots, a-1\} \}$.
Then ${\bf A}^{*}_{K+ A, comb}$ is isomorphic to ${\mathcal A}[I]/F_{B}(I)$ where $I:=(xx_{2})$ and $F_{B}(I)$ is an incarnation of the expression $(B) = (yy_{2})\cdot\dots\cdot(yy_{b})=0$ via the generators. Analogously, ${\bf A}^{*}_{K+ B, comb} \cong {\mathcal A}[J]/F_{A}(J)$ with $V:= (yy_{2})$.
{The cycles $(A)$ and $(B)$ equal to the complete intersection of divisors $(xx_{2}), (xx_{3}), \dots, (xx_{a-1})$ and $(yy_{2}), (yy_{3}), \dots, (yy_{b-1})$ respectively. So the Chern polynomials are:} $$f_{A}(T) = \big(T + (xx_{2})\big)\cdot\dots\cdot\big(T + (xx_{a-1})\big) \text{ and }f_{B}(S) = \big(S + (yy_{2})\big)\cdot\dots\cdot\big(S + (yy_{b-1})\big).$$
Moreover, the new variables $T$ and $S$ correspond to one and the same exceptional divisor $\D_{A}=\D_{B}$.
Relation between polynomials $f_{\bullet}$ and $F_{\bullet}$ are clarified in the following lemma.
\begin{lemma}\label{PullingUp} The Chow class of the image of a divisor $(ab)_{K+A}$, $a,b \in [n]$ under the morphism $\pi_{A}^{*}$ equals \begin{enumerate} \item $(ab)_{K}$ for $a \in A, b\in B$, or vice versa;
\item ${\bf bl}_{(ab)(A)}\big((ab)_{K+A}\big)$ for $\{a,b\} \subset B$;
\item ${\bf bl}_{(A)}\big((ab)_{K+A}\big) + \D_{A}$ for $\{a,b\} \subset A$. \end{enumerate} \end{lemma} \begin{proof}
In case (1), the cycle $(ab)_{K+A}$ does not intersect $(A)_{K+A}$. It is by definition $\textbf{bl}_{(ab)\cap(A)}\big((ab)_{K+A}\big)$. Then (1) and (2) follow directly from Theorem \ref{CorCodim},(2) by dimension counts.
The claim (3) also follows from the blow-up formula Theorem \ref{CorCodim}: $$\pi_{A}^{*} (ab) = \textbf{bl}_{(ab)\cap (A)}\big((ab)_{K+A}\big) + j_{A, *}\big( g^{*}_{A} [(ab)\cap(A)] \cdot s(N_{\D_{A}}\overline{\mathcal{M}}_{0, K}) \big)_{n-4},$$ where $N_{\D_{A}}\overline{\mathcal{M}}_{0, K}$ is a normal bundle and $s(\;)$ is a total Segre class. This follows from the equalities by the functoriality of the total Chern and Segre classes and the equality $s(U)\cdot c(U) = 1$. Namely, we have \begin{align*} s((ab)\cap(A)) &= [(ab)\cap(A)] \cdot s(N_{(ab)\cap(A)}(ab)_{K+A}) = [(ab)\cap(A)]\cdot s\left(N_{(A)}\overline{\mathcal{M}}_{0, K+A}\right), \\ c\left( \frac{g^{*}_{A}N_{(A)}\overline{\mathcal{M}}_{0,K+A}}{N_{\D_{A}}\overline{\mathcal{M}}_{0,K}} \right) &\cdot g^{*}_{A} [(ab)\cap(A)]\cdot g^{*}_{A}s\left(N_{(A)}\overline{\mathcal{M}}_{0, K+A}\right) = g^{*}_{A} [(ab)\cap(A)] \cdot s(N_{\D_{A}}\overline{\mathcal{M}}_{0, K}). \end{align*}
Finally, we note that $g^{*}_{A} [(ab)\cap(A)] = \D_{A}$. \iffalse, and for the desired dimension $n-4$ we should take ``1'' from the total Segre class $s(N_{\D_{A}}\overline{\mathcal{M}}_{0, K})$. \textbf{GAIANE DOES NOT LIKE THIS SENTENCE} \fi \end{proof}
From Lemma \ref{PullingUp}, we have the equality $$f_{A}(T) = \pi_{A}^{*}\big( (xx_{2})\cdot\dots\cdot(xx_{a-1}) \big) = \pi_{A}^{*}(A)\text{ and }f_{B}(S) = \pi_{B}^{*}(B).$$
\iffalse The following lemma states that combinatorial annihilators of $(A)$ and $(B)$ coincide. \fi \begin{lemma}\label{kernels}$\;$ \begin{enumerate} \item The ideal $\Ann(A)_{comb}$ is generated by its first graded component.
\item More precisely, the generators of the ideal $\Ann(A)_{comb}$ are \begin{enumerate} \item the elements of type $(ab)$ with $a\in A, b\in B$, and
\item the elements of type $$(a_1a_2)-(b_1b_2),$$ where $a_1,a_2 \in A$; $b_1,b_2 \in B=A^c$. \end{enumerate}
\item The annihilators $\Ann(A)_{comb}$ and $\Ann(B)$ are canonically isomorphic. \end{enumerate} \end{lemma} \begin{proof}
First observe that the kernel $\ker(i_{A}^*)$ equal the annihilator of the cycle $(A)$. Set $\kappa$ be the ideal generated by $\ker (i_{A}^{*})\cap A^1$. Without loss of generality, we may assume that $A=\{1,2,...,m\}$. Observe that: \begin{enumerate} \item If $I\subset [n]$ has a nonempty intersection with both $A$ and $A^c$, then $(A)(I)=0$. In this case, $(I)$ can be expressed as $(ab)(I')$, where $a\in A,\ b \in A^c$.
\item \begin{enumerate} \item[(i)] If $I\subset A$, then
$(A)\smile (I)= (A)(m+1...m+|I|).$
\item[(ii)] In this case, the element $(I)-(m+1,...,m+|I|)$ is in $\kappa$. \end{enumerate}
Let us demonstrate this by giving an example with $A=\{1,2,3,4\}$, $(I)=(12)$: \newline $(1234)\smile (12)=(A)\smile ((15)+ (16)-(56))= 0+0-(1234)(56)$. We conclude that $(12)-(56)\in \kappa$.
Let us show that $(123)-(567)\in \kappa$.
Indeed, $(123)-(567)=(12) \smile (23)-(567)\in \kappa \Leftrightarrow (56)\smile(23)-(567) \in \kappa \Leftrightarrow (56)\smile ((23)-(67))\in \kappa$. Since $(23)-(67)\in \kappa$, the claim is proven.
\item If $I\subset A^c$, then
$(A)\smile (I)= (A)(m+1,...,m+|I|).$
The element $(I)-(m+1,...,m+|I|)$ is in $\kappa$.
This follows from (1) and (2). \end{enumerate}
Now let us prove the lemma.
Assume $x\smile (A)=0$. Let $x=\sum_i a_i (I_1^i)...(I^i_{k_i}).$
We may assume that $x$ is a homogeneous element. Modulo $\kappa$, each summand $(I_1)...(I_{k_i})$ can be reduced to some $ (m+1...m+r_1)(m+r_1+1,...,m+r_2)...(m+r_{k_i}+1...m')$. Modulo $\kappa$, $ (m+1...m+r_1)(m+r_1+1...m+r_2)...(m+r_{k_i}+1,...,m')$ can be reduced to a one-bracket element $ (m+1...m+r_1m+r_1+1...m+r_2...m+r_{k_i}+1,...,m'')$.
Indeed, for two brackets we have: $$ (m+1...m+r_1)(m+r_1+1,...,m+r_2) \equiv (m+1...m+r_1)(m+r_1,...,m+r_2-1) \equiv (m+1...m+r_2-1)\;\;(\mathrm{mod} \; \kappa).$$ For a bigger number of brackets, the statement follows by induction.
We conclude that a homogeneous $x\in \ker(i_{A}^*)$ modulo $\kappa$ reduces to some $a(m+1...m+m')$, where $a \in \mathbb{Z}$. Then $a=0$. Indeed, $(A)(m+1...m+m')\neq 0$ since by Lemma \ref{LemmaTriple} $(A)(m+1...m+m')(m+m'...n)\neq 0$. \end{proof} \begin{remark} Via the four-term relation any element from b) can be expressed as a linear combination of elements from a). So only a)--elements are sufficient to generate the annihilators. Actually, $${\mathcal A} \cong \mathbb{C} \oplus \Ann(A)_{comb}.$$ \end{remark}
We arrive at the following commutative diagram of graded rings: \begin{center} \begin{tikzcd}[column sep=small] & {\bf A}^{*}_{K}\cong \widetilde{{\bf A}^{*}_{K}}/\Ann(B) \cong \widetilde{{\bf A}^{*}_{K}}/\Ann(A)_{comb} & \\ & \widetilde{{\bf A}^{*}_{K}}:= {\bf A}^{*}_{K+A, comb} [T]/f_{(A)}(T) \cong {\bf A}^{*}_{K+B} [S]/f_{(B)}(S) \arrow[u] & \\ & &\\
{\bf A}^{*}_{K+B} \cong {\mathcal A}[J]/F_{A}(J)\arrow[uur, "f_{B} = \pi_{B}^{*} F_{B}"] & & {\bf A}^{*}_{K+A, comb}\cong {\mathcal A}[I]/F_{B}(I) \arrow[uul, "f_{A} = \pi_{A}^{*} F_{A}"] \\ & {\mathcal A} \arrow[ul, "F_{A}"] \arrow[ur, "F_{B}"] & \end{tikzcd} \end{center}
Therefore, the following diagram commutes:
\begin{center} \begin{tikzcd}[column sep=small] 0\arrow[r] & \Ann(A)_{comb} \arrow[r] \arrow[d]& {\bf A}^{*}_{K+A, comb} [f_{(A)}] \arrow[r]\arrow[d, two heads, "\alpha"] & \sfrac{{\bf A}^{*}_{K+A, comb}[f_{(A)}]}{\Ann(A)_{comb}} \arrow[r] \arrow[d, "\cong"] & 0\\ 0 \arrow[r]& \Ann(A)_{alg} \arrow[r] & {\bf A}^{*}_{K+A, alg} [f_{(A)}]\arrow[r] & \sfrac{{\bf A}^{*}_{K+A, alg} [f_{(A)}]}{\Ann(A)_{alg}} \arrow[r] & 0 \end{tikzcd} \end{center} Here $R[g]$ denotes the
extension of a ring $R$ by a polynomial $g(t)$. All three vertical maps are induced by the map $\alpha$; the last vertical map is an isomorphism since both rings are isomorphic to ${\bf A}^{*}_{K}$.
The ideals $\Ann(A)_{comb}$ and $\Ann(A)_{alg}$ coincide, so the homomorphism $$\alpha:\;{\bf A}^{*}_{K+A, comb} [f_{(A)}] \rightarrow {\bf A}^{*}_{K+A, alg} [f_{(A)}]$$ is {injective}, and the theorem is proven. \qed
\iffalse
\begin{lemma} The map $\alpha:\; {\bf A}^{*}_{K+A, comb} \rightarrow {\bf A}^{*}_{K+A, alg}$ is an isomorphism. \end{lemma}
The surjective homomorphism $\alpha:\; {\bf A}^{*}_{K+A, comb} \rightarrow {\bf A}^{*}_{K+A, alg}$ induces an isomorphism ${\bf A}^{*}_{K+A, comb} [f_{(A)}] \rightarrow {\bf A}^{*}_{K+A, alg} [f_{(A)}]$. We get the result from this applying the Snake Lemma . \fi
\section{Poincaré polynomials of ASD compactifications}
\begin{theorem}\label{PoiASD} Poincaré polynomial ${\mathscr P}_{q}\big(\overline{\mathcal{M}}_{0, L}\big)$ for an ASD complex $L$ equals
$${\mathscr P}_{q}\big(\overline{\mathcal{M}}_{0, L}\big) = \frac{1}{q(q-1)} \left( (1+q)^{n-1} - \sum\limits_{I \in L} q^{|I|} \right).$$ \end{theorem}
\begin{proof} This theorem is proven by Klyachko \cite[Theorem 2.2.4]{Kl} for polygon spaces, that is, for compactifications coming from a threshold ASD complex. Assume that $K + A$ be a threshold ASD complex. For the blow up of the space $\overline{\mathcal{M}}_{0, K+B}$ along the subvariety $(B)$ we have an exact sequence of Chow groups
$$0 \rightarrow {\bf A}_{p}\big(\overline{\mathcal{M}}_{0, \mathcal{P}_{|A|+1}}\big) \rightarrow {\bf A}_{p}\big( \overline{\mathcal{M}}_{0, \mathcal{P}_{|A|+1} } \times \overline{\mathcal{M}}_{0, \mathcal{P}_{|B|+1} } \big) \oplus {\bf A}_{p}\big( \overline{\mathcal{M}}_{0, K + B} \big) \rightarrow {\bf A}_{p}\big(\overline{\mathcal{M}}_{0, K}\big) \rightarrow 0,$$ where, as before, $A \sqcup B = [n]$ and $p$ is a natural number. We get the equality
$${\mathscr P}_{q}(K) = {\mathscr P}_{q}(K+B) + {\mathscr P}_{q}(\mathcal{P}_{|A|+1})\cdot{\mathscr P}_{q}(\mathcal{P}_{|B|+1})- {\mathscr P}_{q}(\mathcal{P}_{|A|+1}).$$
Also, we have the recurrent relations for the Poincaré polynomials: \begin{align*}
{\mathscr P}_{q}(K + B) &= {\mathscr P}_{q}(K + A) + {\mathscr P}_{q}(\mathcal{P}_{|A|+1}) - {\mathscr P}_{q}(\mathcal{P}_{|B|+1}) =\\
&=\frac{1}{q(q-1)}\left((1+q)^{n-1} - \sum\limits_{I\in K + A} q^{|I|} + q^{|A|} - q - q^{|B|} +q \right) = \\
&=\frac{1}{q(q-1)}\left((1+q)^{n-1} - \sum\limits_{I\in K} q^{|I|} -q^{|A|} + q^{|A|} - q - q^{|B|} +q \right) = \\
&= \frac{1}{q(q-1)} \left( (1+q)^{n-1} - \sum\limits_{I \in K + B} q^{|I|} \right). \end{align*}
We have used the following: if $U$ is a facet of an ASD complex, then there is an isomorphism $\overline{\mathcal{M}}_{0, \mathcal{P}_{|U|+1}} \cong \mathbb{P}^{|U| - 2}$ (see Example \ref{example}(2) ); the Poincaré polynomial of the projective space $\mathbb{P}^{|U| - 2}$ equals $\frac{q^{|U|} - q}{q(q-1)}$. \end{proof}
\section{The tautological line bundles over $\overline{\mathcal{M}}_{0,K}$ and the $\psi$--classes}\label{psi}
The \textit{tautological line bundles} $L_i, \ i=1,...,n$ were introduced by M. Kontsevich \cite{Kon} for the Deligne-Mumford compactification. The first Chern classes of $L_i$ are called the $\psi$-\textit{classes}.
We now mimic the Kontsevich's original definition for ASD compactifications. Let us fix an ASD complex $K$ and the corresponding compactification $\overline{\mathcal{M}}_{0,K}$. \begin{definition} The line bundle $E_i=E_{i}(L)$ is the complex line bundle over the space $\overline{\mathcal{M}}_{0,K}$ whose fiber over a point $(u_1,...,u_n)\in (\mathbb{P}^1)^n$ is the tangent line\footnote{In the original Kontsevich's definition, the fiber over a point is the cotangent line, whereas we have the tangent line. This replacement does not create much difference.} to the projective line $\mathbb{P}^1$ at the point $u_i$. The first Chern class of $E_{i}$ is called the $\psi$-class and is denoted by $\psi_i$. \end{definition}
\begin{proposition} \label{Ch2} \begin{enumerate} \item For any $i \neq j \neq k\in [n]$ we have $$\psi_i= (ij) +(ik)-(jk).$$ \item The four-term relation holds true:
$(ij)+(kl)=(ik)+(jl)$ for any distinct $i,j,k,l \in [n].$ \end{enumerate}
\end{proposition} \begin{proof} (1) Take a stable configuration $(x_1,...,x_n)\in \overline{\mathcal{M}}_{0,K}$. Take the circle passing through $x_i,x_j$, and $x_k$. It is oriented by the order $ijk$. Take the vector lying in the tangent complex line to $x_i$ which is tangent to the circle and points in the direction of $x_j$. It gives rise to a section of $E_{i}$ which is defined correctly whenever the points $x_i,x_j$, and $x_k$ are distinct. Therefore, $\psi_i = A(ij) +B(ik)+C(jk)$ for some integer $A,B,C$. Detailed analysis specifies their values.
Now (2) follows since the Chern class $\psi_i $ does not depend on the choice of $j$ and $k$. \end{proof}
Let us denote by $|d_{1}, \dots, d_{n}|_K$ the intersection number $\langle \psi_1^{ d_1} ... \psi_k^{ d_k}\rangle_K= \psi_1^{\smile d_1} \smile ... \smile \psi_k^{\smile d_k}$ related to the ASD complex $K$.
\begin{theorem}\label{ThmRecursion} Let $\overline{\mathcal{M}}_{0,K}$ be an ASD compactification. A recursion for the intersection numbers is \begin{align*}
|d_{1}, \dots, d_{n}|_{K} = &|d_{1}, \dots, d_{i}+d_{j}-1, \dots, \hat{d_{j}}, \dots, d_{n}|_{K_{(ij)}} + |d_{1}, \dots, d_{i}+d_{k}-1, \dots, \hat{d_{k}}, \dots, d_{n}|_{K_{(ik)}} \\
&-|d_{1}, \dots, d_{i}-1 ,\dots, d_{j}+d_{k}, \dots, \hat{d_{j}}, \hat{d_{k}},\dots, d_{n}|_{K_{(jk)}}, \end{align*} where $i, j, k \in [n]$ are distinct.
Remind that $K_{(ij)}$ denotes the complex $K$ with $i$ and $j$ frozen together.
{Might happen that $K_{(ij)}$ is ill-defined, that is, $(ij)\notin K$. Then we set the corresponding summand to be zero.} \end{theorem} \begin{proof} By Proposition \ref{Ch2}, $$\langle \psi_1^{d_1} \dots \psi_{n}^{d_n} \rangle_K = \langle \psi_1^{d_1-1} \dots \psi_{n}^{d_n} \rangle_K \smile \big((1i)+(1j)- (ij)\big).$$ It remains to observe that $\langle \psi_1^{d_1-1} \dots \psi_{n}^{d_n} \rangle_K\smile (ab)$ equals the $\langle \psi_1^{d_1-1} \dots \psi_{n}^{d_n} \rangle_{K_{(a,b)}}$.\end{proof}
\begin{theorem}\label{main_theorem} Let $\overline{\mathcal{M}}_{0,K}$ be an ASD compactification. Any top monomial in $\psi$-classes modulo renumbering has a form $$\psi_1^{d_1}\smile ...\smile \psi_m^{d_m}$$ with $\sum_{q=1}^m d_q=n-3$ and $d_q \neq 0$ for $q=1,...,m$. Its value equals the signed number of partitions $$[n-2]=I\cup J$$ with $m+1 \in I$ and $I,J \subset K$. Each partition is counted with the sign
$$(-1)^N \cdot \varepsilon ,$$
where $$N= |J|+\sum_{q \in J , q\leq m} d_q, \;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\varepsilon=
\left\{
\begin{array}{lll}
1, & \hbox{if \ \ } J\cup \{n\}\in K, \hbox{and\ \ }J\cup \{n-1\}\in K;\\
-1, & \hbox{if \ \ } I\cup \{n\}\in K, \hbox{and\ \ }I\cup \{n-1\}\in K;\\
0, & \hbox{otherwise.}
\end{array}
\right.$$
\end{theorem}
\textit{Proof} goes by induction.
Although \textbf{the base} is trivial, let us look at it. The smallest $n$ which makes sense is $n=4$. There exist two ASD complexes with four vertices, both are threshold. So there exist two types of fine moduli compactifications, both correspond to the configuration spaces of some flexible four-gon. The top monomials are the first powers of the $\psi$-classes.
\begin{enumerate}
\item For $l_1=1;\ l_2=1;\ l_3=1;\ l_4=0,1$ we have
$\psi_1=\psi_2=\psi_3=0$, and \ \ $\psi_4=2$.
Let us prove that the theorem holds for the monomial $\psi_1$. There are two partitions of \newline $[n-2]~=~[2]$: \begin{enumerate} \item $J=\{1\},\ I=\{2\}$. Here $\varepsilon =0$, so this partition contributes $0$. \item $J=\emptyset,\ I=\{1,2\}$. Here $I\notin K $, so this partition also contributes $0$. \end{enumerate}
\item For $l_1=2,9;\ l_2=1;\ l_3=1;\ l_4=1$, we have
$\psi_2=\psi_3=\psi_4=1$, and \ \ $\psi_1=-1$. Let us check that the theorem holds for the monomial $\psi_1$. (The other monomials are checked in a similar way.) There partitions of $[2]$ are the same: \begin{enumerate} \item $J=\{1\},\ I=\{2\}$. Here $\varepsilon =-1, \ N=1+1$, so this partition contributes $-1$. \item $J=\emptyset,\ I=\{1,2\}$. Here $I\notin K $, so it contributes $0$. \end{enumerate}
\end{enumerate}
For the \textbf{induction step}, let us use the recursion. We shall show that for any partition $[n-2]=I\cup J$, its contribution to the left hand side and the right hand side of the recursion are equal.
This is done through a case analysis. We present here three cases; the rest are analogous.
\begin{enumerate} \item Assume that $i,j,k \in I$, and $(I,J)$ contributes $1$ to the left hand side count. Then
\begin{itemize} \item[$\triangleright$] $(d_{1}, \dots, d_{i}+d_{j}-1, \dots, \hat{d_{j}}, \dots, d_{n})_{K_{(ij)}}$ contributes $1$ to the right hand side.
Indeed, neither $N$, nor $\varepsilon$ changes when we pass from $K$ to $K_{(ij)}$.
\item[$\triangleright$] $(d_{1}, \dots, d_{i}+d_{k}-1, \dots, \hat{d_{k}}, \dots, d_{n})_{K_{(ik)}} $ contributes $1$, and
\item[$\triangleright$] $-(d_{1}, \dots, d_{i}-1 ,\dots, d_{j}+d_{k}, \dots, \hat{d_{j}}, \hat{d_{k}},\dots, d_{n})_{K_{(jk)}}$ contributes $-1$. \end{itemize}
\item Assume that $i\in I,\ j,k \in J$, and $(I,J)$ contributes $1$ to the left hand side count. Then
\begin{itemize} \item[$\triangleright$] $(d_{1}, \dots, d_{i}+d_{j}-1, \dots, \hat{d_{j}}, \dots, d_{n})_{K_{(ij)}}$ contributes $0$ to the right hand side.
\item[$\triangleright$] $(d_{1}, \dots, d_{i}+d_{k}-1, \dots, \hat{d_{k}}, \dots, d_{n})_{K_{(ik)}} $ contributes $0$, and
\item[$\triangleright$] $-(d_{1}, \dots, d_{i}-1 ,\dots, d_{j}+d_{k}, \dots, \hat{d_{j}}, \hat{d_{k}},\dots, d_{n})_{K_{(jk)}}$ contributes $1$.
Indeed, $N$ turns to $N-1$, whereas $\varepsilon$ stays the same. \end{itemize}
\item Assume that $i\in J,\ j,k \in I$, and $(I,J)$ contributes $1$ to the left hand side count. Then
\begin{itemize} \item[$\triangleright$] $(d_{1}, \dots, d_{i}+d_{j}-1, \dots, \hat{d_{j}}, \dots, d_{n})_{K_{(ij)}}$ contributes $0$.
\item[$\triangleright$] $(d_{1}, \dots, d_{i}+d_{k}-1, \dots, \hat{d_{k}}, \dots, d_{n})_{K_{(ik)}} $ contributes $0$, and
\item[$\triangleright$] $-(d_{1}, \dots, d_{i}-1 ,\dots, d_{j}+d_{k}, \dots, \hat{d_{j}}, \hat{d_{k}},\dots, d_{n})_{K_{(jk)}}$ contributes $1$, since $N$ turns to $N-1$, and $\varepsilon$ stays the same.\qed \end{itemize}\end{enumerate}
This theorem was proven for polygon spaces (that is, for threshold ASD complexes) in \cite{Agapito}.
\section{Appendix. Chow rings and blow ups}
Assume we have a diagram of a blow up $\widetilde{Y}:={\bf bl}_{X}(Y)$. Here $X$ and $Y$ are smooth varieties, $\iota: X \hookrightarrow Y$ is a regular embedding, and $\widetilde{X}$ is the exceptional divisor. In this case, $\iota^{*}: A^*(Y) \rightarrow A^*(X)$ is surjective. \begin{center} \begin{tikzcd} \widetilde{X} \arrow{d}{\tau} \arrow[r, hookrightarrow, "\theta"] & \widetilde{Y} \arrow{d}{\pi} \\ X \arrow[r, hookrightarrow, "\iota"] & Y \end{tikzcd} \end{center} Denote by $E$ the relative normal bundle $$E:= \tau^* N_X Y/ N_{\widetilde{X}} \widetilde{Y}.$$
\begin{theorem}\cite[Appendix. Theorem 1]{Keel92}\label{NewChowRing} The Chow ring $A^{*}(\widetilde{Y})$ is isomorphic to $$\frac{A^{*}(Y)[T]}{(P(T), T\cdot \ker i^*)},$$ where $P(T) \in A^*(Y)[T]$ is the pullback from $A^*(X)[T]$ of Chern polynomial of the normal bundle $N_{X}Y$. This isomorphism is induced by $\pi^* : A^*(Y)[T] \rightarrow A^* (\widetilde{Y})$ which sends $-T$ to the class of the exceptional divisor $\widetilde{X}$. \end{theorem}
\begin{theorem}\cite[Proposition 6.7]{FulInter}\label{SESBlow} Let $k \in \mathbb{N}$. \begin{itemize} \item[a)](Key formula) For all $x \in A_{k}(X)$ $$\pi^* \iota_{*} (x) = \theta_* (c_{d-1}(E)\cap \tau^* x)$$ \item[e)] There are split exact sequences $$0 \rightarrow A_{k} X \xrightarrow{\upsilon} A_{k} \widetilde{X} \oplus A_{k} Y \xrightarrow{\eta} A_{k} \widetilde{Y} \rightarrow 0$$ with $\upsilon(x) = \big( c_{d-1}(E)\cap \tau^{*}x, -\iota_{*}x \big)$, and $\eta(\tilde{x}, y) = \theta_{*}\tilde{x}+ \pi^* y$. A left inverse for $\upsilon$ is given by $(\tilde{x}, y) \mapsto \tau_{*}(\tilde{x})$. \end{itemize} \end{theorem}
\begin{theorem}\cite[Theorem 6.7, Corollary 6.7.2]{FulInter}\label{CorCodim} $\;$ \begin{enumerate} \item (Blow-up Formula) Let $V$ be a $k$-dimensional subvariety of $Y$, and let $\tilde{V} \subset \tilde{Y}$ be the proper transform of $V$, i.e., the blow-up of $V$ along $V\cap X$. Then $$\pi^{*}[V] = [\tilde{V}] + j_{*}\{c(E)\cap \tau^{*} s(V\cap X, V)\}_{k} \text{ in }A_{k}\tilde{Y}.$$ \item If $\dim V\cap X \leq k-d$, then $\pi^* [V] = [\tilde{V}]$. \end{enumerate} \end{theorem}
An algebraic variety $Z$ is a \textit{HI--scheme} if the canonical map $\mathrm{cl}: A^{*}(Z) \rightarrow H^{*}(Z, \mathbb{Z})$ is an isomorphism.
\begin{theorem}\cite[Appendix. Theorem 2]{Keel92}\label{ChowCoh} \begin{itemize} \item If $X, \widetilde{X}$, and $Y$ are HI, then so is $\widetilde{Y}$.
\item If $X, \widetilde{X}$, and $\widetilde{Y}$ are HI, then so is $Y$. \end{itemize} \end{theorem}
\textbf{Acknowledgement.} This research is supported by the Russian Science Foundation under grant 16-11-10039.
\end{document}
|
arXiv
|
{
"id": "1808.08600.tex",
"language_detection_score": 0.6811769008636475,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Matrix random products with singular harmonic measure}
\author{Vadim A. Kaimanovich} \address{Mathematics, Jacobs University Bremen, Campus Ring 1, D-28759, Bremen, Germany} \email{[email protected]}
\author{Vincent Le Prince} \address{IRMAR, Campus de Beaulieu, 35042 Rennes, France} \email{[email protected]}
\subjclass[2000]{Primary 60J50; Secondary 28A78, 37D25, 53C35}
\keywords{Random walk, matrix random product, Lyapunov exponents, harmonic measure, Hausdorff dimension}
\begin{abstract} Any Zariski dense countable subgroup of $SL(d,\R)$ is shown to carry a non-degenerate finitely supported symmetric random walk such that its harmonic measure on the flag space is singular. The main ingredients of the proof are: (1) a new upper estimate for the Hausdorff dimension of the projections of the harmonic measure onto Grassmannians in $\R^d$ in terms of the associated differential entropies and differences between the Lyapunov exponents; (2) an explicit construction of random walks with uniformly bounded entropy and Lyapunov exponents going to infinity. \end{abstract}
\maketitle
\thispagestyle{empty}
\section*{Introduction}
The notion of \emph{harmonic measure} (historically first defined in the framework of the theory of potential) has an explicit probabilistic description in terms of the dynamical properties of the associated Markov processes as a \emph{hitting distribution}. Moreover, ``hitting'' can be interpreted both as attaining the target set in the usual sense (in finite time) and as converging to it at infinity (when the target is attached as a boundary to the state space).
In concrete situations the target set is usually endowed with additional structures giving rise to other ``natural'' measures (e.g., smooth, uniform, Haar, Hausdorff, maximal entropy, etc.), which leads to the question about \emph{coincidence} of the harmonic and these ``other'' measures (or, in a somewhat weaker form, about coincidence of the respective measure classes). As a general rule, such a coincidence either implies that the considered system has very special symmetry properties or is not possible at all. However, establishing it in a rigorous way is a notoriously difficult problem. See, for example, the cases of the Brownian motion on cocompact negatively curved Riemannian manifolds \cite{Katok88,Ledrappier95}, of Julia sets of endomorphisms of the Riemann sphere \cite{Przytycki-Urbanski-Zdunik89,Zdunik91} and of polynomial-like maps \cite{Lyubich-Volberg95, Balogh-Popovici-Volberg97, Zdunik97}, or of Cantor repellers in a Euclidean space \cite{Makarov-Volberg86,Volberg93}.
In the present paper we consider the singularity problem for \emph{random matrix products} $x_n=h_1h_2\dots h_n$ with Bernoulli increments, or, in other words, for \emph{random walks} on the group $SL(d,\R)$ (or its subgroups). Actually, our results are also valid for general non-compact semi-simple Lie groups, see \cite{LePrince04}, but for the sake of expositional simplicity we restrict ourselves just to the case of matrix groups. Another simplification is that we always assume that the considered subgroups of $SL(d,\R)$ are Zariski dense. It guarantees that the harmonic measure of the random walk is concentrated on the space of full flags in $\R^d$ (rather than on its quotient determined by the degeneracies of the Lyapunov spectrum). Modulo an appropriate technical modification our results remain valid without this assumption as well.
If the distribution $\mu$ of the increments $\{h_n\}$ has a finite \emph{first moment} $\int\log\|h\|\,d\mu(h)$, then by the famous \emph{Oseledets multiplicative ergodic theorem} \cite{Oseledec68} there exists the \emph{Lyapunov spectrum} $\l$ consisting of \emph{Lyapunov exponents} $\l_1\ge\l_2\ge\dots\l_d$ (they determine the growth of the random products in various directions), and, moreover, a.e.\ sample path $\{x_n\}$ gives rise to the associated \emph{Lyapunov flag} (filtration) of subspaces in $\R^d$. The distribution $\nu=\nu(\mu)$ of these Lyapunov flags is then naturally called the \emph{harmonic measure} of the random product. The geometric interpretation of the Oseledets theorem \cite{Kaimanovich89} is that the sequence $x_n o$ asymptotically follows a geodesic in the associated Riemannian symmetric space $S=SL(d,\R)/SO(d)$ (here $o=SO(d)\in S$); the Lyapunov spectrum determines the Cartan (``radial'') part of this geodesic, whereas the Lyapunov flag determines its direction.
If $\supp\mu$ generates a Zariski dense subgroup of $SL(d,\R)$, then the Lyapunov spectrum is \emph{simple} ($\equiv$ the vector $\l$ lies in the interior of the positive Weyl chamber), see \cite{Guivarch-Raugi85,Goldsheid-Margulis89}, so that the associated Lyapunov flags are full ($\equiv$ contain subspaces of all the intermediate dimensions). The space $\B=\B(d)$ of full flags in $\R^d$ is also known under the name of the \emph{Furstenberg boundary} of the associated symmetric space $S$, see \cite{Furstenberg63} for its definition and \cite{Kaimanovich89,Guivarch-Ji-Taylor98} for its relation with the boundaries of various compactifications of Riemannian symmetric spaces. In the case when the Lyapunov spectrum is simple, a.e.\ sequence $x_n o$ is convergent in all reasonable compactifications of the symmetric space $S$, and the corresponding hitting distributions can be identified with the harmonic measure $\nu$ on $\B$.
The flag space $\B$ is endowed with a natural smooth structure, therefore it makes sense to compare the harmonic measure class with the smooth measure class (the latter class contains the unique rotation invariant measure on the flag space). The harmonic measure is ergodic, so that it is either absolutely continuous or singular with respect to the smooth (or any other quasi-invariant) measure class. Accordingly, we shall call the jump distribution $\mu$ either \emph{absolutely continuous} or \emph{singular at infinity}.
If the measure $\mu$ is absolutely continuous with respect to the Haar measure on the group $SL(d,\R)$ (or even weaker: a certain convolution power of $\mu$ contains an absolutely continuous component) then it is absolutely continuous at infinity \cite{Furstenberg63,Azencott70}.
As it turns out, there are also measures $\mu$ which are absolutely continuous at infinity in spite of being supported by a \emph{discrete subgroup} of $SL(d,\R)$. Namely, Furstenberg \cite{Furstenberg71,Furstenberg73} showed that any lattice (for instance, $SL(d,\Z)$) carries a probability measure $\mu$ with a finite first moment which is absolutely continuous at infinity. It was used for proving one of the first results on the rigidity of lattices in semi-simple Lie groups.
Furstenberg's construction of measures absolutely continuous at infinity (based on discretization of the Brownian motion on the associated symmetric space) was further extended and generalized in \cite{Lyons-Sullivan84,Kaimanovich92a,Ballmann-Ledrappier96}. Another construction of random walks with a given harmonic measure was recently developed by Connell and Muchnik \cite{Connell-Muchnik07,Connell-Muchnik07a}. Note that the measures $\mu$ arising from all these constructions are inherently infinitely supported.
Let us now look at the singularity vs. absolute continuity dichotomy for the harmonic measure from the ``singularity end''. The first result of this kind was obtained by Chatterji \cite{Chatterji66} who established singularity of the distribution of infinite continuous fractions with independent digits. This distribution can indeed be viewed as the harmonic measure associated to a certain random walk on $SL(2,\Z)\subset SL(2,\R)$ \cite{Furstenberg63a}. See \cite{Chassaing-Letac-Mora83} for an explicit description of the harmonic measure in a similar situation and \cite{Kifer-Peres-Weiss01} for a recent very general result on singularity of distributions of infinite continuous fractions. Extending the continuous fractions setup Guivarc'h and Le Jan \cite{Guivarch-LeJan90,Guivarch-LeJan93} proved that the measures $\mu$ on non-compact lattices in $SL(2,\R)$ satisfying a certain moment condition (in particular, all finitely supported measures) are singular at infinity.
Together with some other circumstantial evidence (as, for instance, pairwise singularity of various natural boundary measure classes on the visibility boundary of the universal cover of compact negatively curved manidolds \cite{Ledrappier95} or on the hyperbolic boundary of free products \cite{Kaimanovich-LePrince08p}) the aforementioned results lead to the following
{\parindent 0pt \textsc{Conjecture.} Any finitely supported probability measure on $SL(d,\R)$ is singular at infinity. }
The principal result of the present paper is the following
\begin{thrm} \label{th:mn} Any countable Zariski dense subgroup $\G$ of $SL(d,\R)$ carries a non-degenerate symmetric probability measure $\mu$ which is singular at infinity. Moreover, if $\G$ is finitely generated then the measure $\mu$ can be chosen to be finitely supported. \end{thrm}
Note here an important difference between the case $d=2$ and the higher rank case $d\ge 3$. If $\G$ is discrete, then for $d=2$ the boundary circle can be endowed with different $\G$-invariant smooth structures (parameterized by the points from the Teichm\"uller space of the quotient surface), so that Furstenberg's discretization construction readily provides existence of measures $\mu$ (not finitely supported though!) which are singular at infinity. Obviously, this approach does not work in the higher rank case, where our work provides first examples of measures singular at infinity.
Our principal tool for establishing singularity is the notion of the \emph{dimension} of a measure. Namely, we show that in the setup of our Main Theorem the measure $\mu$ can be chosen in such a way that the Hausdorff dimension of the associated harmonic measure (more precisely, of its projection onto one of the Grassmanians in $\R^d$) is arbitrarily small, which obviously implies singularity. For doing this we establish an inequality connecting the \emph{Hausdorff dimension} with the \emph{asymptotic entropy} and the \emph{Lyapunov exponents} of the random walk.
There are several notions of dimension of a probability measure $m$ on a compact metric space $(Z,\r)$ (see \cite{Pesin97} and the discussion in \secref{sec:Hausd} below). These notions roughly fall into two categories. The \emph{global} ones are obtained by looking at the dimension of sets which ``almost'' (up to a piece of small measure $m$) coincide with $Z$; in particular, the \emph{Hausdorff dimension} $\dim_H m$ of the measure $m$ is $\inf\left\{ \dim_H A : m(A)=1 \right\}$, where $\dim_H A$ denotes the Hausdorff dimension of a subset $A\subset Z$. On the other hand, the \emph{local} ones are related to the asymptotic behavior of the measures of concentric balls $B(z,r)$ in $Z$ as the radius $r$ tends to $0$. More precisely, the lower $\un\dim_P m(z)$ (resp., the upper $\ov\dim_P m(z)$) \emph{pointwise dimension} of the measure $m$ at a point $z\in Z$ is defined as the $\liminf$ (resp., $\limsup$) of the ratio $\log m B(z,r)/\log r$ as $r\to 0$. In these terms $\dim_H m$ coincides with $\esssup_z \un\dim_P m(z)$. In particular, if $\un\dim_P m(z)=\ov\dim_P m(z)=D$ almost everywhere for a constant $D$, then $\dim_H m=D$. Moreover, in the latter case all the reasonable definitions of dimension of the measure $m$ coincide.
Numerous variants of the formula $\dim m = h/\l$ relating the dimension of an invariant measure $m$ of a differentiable map with its entropy $h=h(m)$ and the characteristic exponent(s) $\l=\l(m)$ have been known since the late 70s -- early 80s, see \cite{Ledrappier81,Young82} and the references therein. Ledrappier \cite{Ledrappier83} was the first to carry it over to the context of random walks by establishing the formula $\dim \nu =h/2\l$ for the dimension of the harmonic measure of random walks on discrete subgroups of $SL(2,\C)$. Here $h=h(\mu)$ is the asymptotic entropy of the random walk with the jump distribution $\mu$, and $\l=\l(\mu)$ is the Lyapunov exponent. However, the dimension appearing in this formula is somewhat different from the classical Hausdorff dimension, because convergence of the ratios $\log m B(z,r)/\log r$ to the value of dimension is only established in measure (rather than almost surely).
Le Prince has recently extended the approach of Ledrappier (also see \cite{Ledrappier01,Kaimanovich98}) to a general discrete group $G$ of isometries of a Gromov hyperbolic space $X$; in this case the Gromov product induces a natural metric (more rigorously, a gauge) on the hyperbolic boundary $\pt X$. He proved that if a probability measure $\mu$ on $G$ has a finite first moment with respect to the metric on $X$, then the associated harmonic measure $\nu$ on $\pt X$ has the property that $\ov\dim_P\nu(z)\le h/\ell$ for $\nu$-a.e. point $z\in\pt X$ (which implies that $\dim_H\nu\le h/\ell$) \cite{LePrince07}, and that the \emph{box dimension} of the measure $\nu$ is precisely $h/\ell$ \cite{LePrince08} (here, as before, $h=h(\mu)$ is the asymptotic entropy of the random walk $(G,\mu)$, and $\ell=\ell(\mu)$ is its linear rate of escape with respect to the hyperbolic metric). Note that the question about the pointwise dimension, i.e., about the asymptotic behaviour of the ratios $\log m B(z,r)/\log r$ for a.e. point $z\in\pt X$, is still open in this generality. It was recently proved that these ratios converge to $h/\ell$ almost surely for any symmetric finitely supported random walk on $G$ \cite{Blachere-Haissinsky-Mathieu08}. This should also be true for finitely supported random walks which are not necessarily symmetric. Namely, the results of Ancona \cite{Ancona88} on almost multiplicativity of the Green function imply that in this situation the harmonic measure has the \emph{doubling property }, which in turn can be used to prove existence of the pointwise dimension.
Yet another example is the inequality $\dim_H\nu\le h/\ell$ for the Hausdorff dimension of the harmonic measure of iterated function systems in the Euclidean space \cite{Nicol-Sidorov-Broomhead02} (when is the equality attained?).
In the context of random walks on general countable groups the asymptotic entropy $h(\mu)$, the rate of escape $\ell(\mu)$ and the exponential growth rate of the group $v$ satisfy the inequality $h(\mu)\le\ell(\mu)v$ (it was first established by Guivarc'h \cite{Guivarch79}; recently Vershik \cite{Vershik00} revitalized interest in it). As it was explained above, in the ``hyperbolic'' situations the ratio $h/\ell$ can (under various additional conditions) be interpreted as the dimension of the harmonic measure, whereas $v$ is the Hausdorff dimension of the boundary itself.
Unfortunately, the aforementioned formulas of the type $\dim\nu=h/\ell$ can not be directly carried over to the higher rank case. The reason for this is that in this case the boundary is ``assembled'' of several components which may have different dimension properties. The simplest illustration is provided by the product $\G_1\times\G_2$ of two discrete subgroups of $SL(2,\R)$ with a product jump distribution $\mu=\mu_1\otimes\mu_2$. The Furstenberg boundary of the bidisk associated with the group $SL(2,\R)\times SL(2,\R)$ is the product of the boundary circles of each hyperbolic disc. The harmonic measure of the jump distribution $\mu$ is then the product of the harmonic measures of the jump distributions $\mu_1$ and $\mu_2$ which may have different dimensions (it would be interesting to investigate the dimension properties of the harmonic measure on the boundary of the polydisc for groups and random walks which do not split into a direct product). In the case of $SL(d,\R)$ the role of elementary ``building blocks'' of the flag space $\B$ is played by the Grassmannians $\Gr_i$ in $\R^d$ (which are the minimal equivariant quotients of $\B$).
In order to deal with spaces with a $\mu$-stationary measure other than the Poisson boundary of the random walk $(\G,\mu)$ it is convenient to use the notion of the \emph{differential $\mu$-entropy} first introduced by Furstenberg \cite{Furstenberg71}. It measures the ``amount of information'' about the random walk contained in the action. The differential entropy does not exceed the asymptotic entropy of the random walk, coincides with it for the Poisson boundary, and is strictly smaller for any proper quotient of the Poisson boundary \cite{Kaimanovich-Vershik83}. Our point is that in the formulas of the type $\dim\nu=h/\ell$ (see above) for spaces with a $\mu$-stationary measure $\nu$ other than the Poisson boundary, the asymptotic entropy $h$ should be replaced with the differential entropy of the space.
In the setup of our Main Theorem let $\mu$ be a non-degenerate probability measure on $\G$ with a finite first moment, so that the associated Lyapunov spectrum $\l_1>\l_2>\dots>\l_d$ is simple. Denote by $\nu_i$ the image of the harmonic measure $\nu$ on the flag space $\B$ under the projection onto the rank $i$ Grassmannian $\Gr_i$, and let $E_i$ be the corresponding differential entropy. All the spaces $(\Gr_i,\nu_i)$ are quotients of the Poisson boundary of the random walk $(\G,\mu)$; if $\G$ is discrete then $(\B,\nu)$ is the Poisson boundary of the random walk $(\G,\mu)$ \cite{Kaimanovich85,Ledrappier85}, otherwise the Poisson boundary of the random walk $(\G,\mu)$ may be bigger than the flag space \cite{Kaimanovich-Vershik83,Bader-Shalom06,Brofferio06} (note that it is still unknown whether all the proper quotients of the flag space $\B$ are also always proper measure-theoretically with respect to the corresponding $\mu$-stationary measures).
\begin{thrm1} The Hausdorff dimensions of the measures $\nu_i$ satisfy the inequalities $$ \dim \nu_i \le \frac{E_i}{\l_i-\l_{i+1}} \;. $$ \end{thrm1}
In the case when $\G$ is a discrete subgroup of $SL(2,\R)$ the right-hand side of the above inequality precisely coincides with the ratio $h(\mu)/2\l(\mu)$ from Ledrappier's formula \cite{Ledrappier83}.
Our Main Theorem follows from a combination of this dimension estimate with the following construction:
\begin{thrm2} Let $\mu$ be a non-degenerate probability measure on $\G$ whose entropy and the first moment are finite, and let $\g\in\G$ be an $\R$-regular element (which always exists in a Zariski dense group). Then the measures $$ \mu^k = \frac12\mu+\frac14\left( \d_{\g^k}+\d_{\g^{-k}} \right) $$ have the property that the entropies $H(\mu^k)$ (and therefore the corresponding asymptotic entropies as well) are uniformly bounded, whereas the lengths of their Lyapunov vectors $\l(\mu^k)$ (equivalently, the top Lyapunov exponents $\l_1(\mu^k)$) go to infinity. \end{thrm2}
Let us mention here several issues naturally arising in connection with our results.
1. Our study of dimension in the present paper is subordinate to proving singularity at infinity, so that we only use rather rudimentary facts about the dimension of the harmonic measure. Under what conditions is the inequality from \thmref{th:dim} realized as equality? Is there an exact formula for the dimension of the harmonic measure on the flag space? Of course, in these questions one has to specify precisely which definition of the dimension is used, cf. \secref{sec:Hausd}.
2. For our purposes it was enough to establish in \thmref{th:estimate} that the lengths of the Lyapunov vectors $\l(\mu^k)$ go to infinity. Is this also true for all the spectral gaps $\l_i(\mu^k)-\l_{i+1}(\mu^k)$ (which would imply that the dimensions of the harmonic measures on $\B$ go to 0)? What happens with the harmonic measures $\nu^k$ (or with their projection); do they weakly converge, and if so what is the limit? Note that one can ask the latter questions in the hyperbolic (rank 1) situation as well, cf. \cite{LePrince07}.
3. Another interesting issue is the connection of the harmonic measure with the invariant measures of the geodesic flow (more generally, of Weyl chambers in the case of higher rank locally symmetric spaces). As it was proved by Katok and Spatzier \cite{Katok-Spatzier96}, the only invariant measure of the Weyl chamber flow associated to a lattice $\G$ in $SL(d,\R)$ with positive entropy along all one-dimensional directions is the Haar measure. In view of the correspondence between the Radon invariant measures of the Weyl chamber flow and $\G$-invariant Radon measures on the principal stratum of the product $\B\times\B$ (cf. \cite{Kaimanovich90} in the hyperbolic case), it implies that singular harmonic measures on $\B$ either can not be lifted to a Radon invariant measure on $\B\times\B$ (contrary to the hyperbolic case \cite{Kaimanovich90}) or, if they can be lifted to a Radon invariant measure on $\B\times\B$, then the associated invariant measure of the Weyl chamber flow has a vanishing directional entropy (in our setup the latter option most likely can be eliminated).
The paper has the following structure. In \secref{sec:RW} we set up the notations and introduce the necessary background information about random products (random walks) on groups. In \secref{sec:exponents}, \secref{sec:flags} and \secref{sec:harm} we specialize these notions to the case of matrix random products, remind the Oseledets multiplicative ergodic theorem, give several equivalent descriptions of the limit flags (Lyapunov filtrations) of random products, and introduce the harmonic measure $\nu$ on the flag space $\B$ and its projections $\nu_i$ to the Grassmannians $\Gr_i$ in $\R^d$. In \secref{sec:entr} we remind the definitions of the asymptotic entropy of a random walk and of the differential entropy of a $\mu$-stationary measure. In \lemref{lem:dentr} we establish an inequality relating the exponential rate of decay of translates of a stationary measure along the sample paths of the reversed random walk with the differential entropy. In \secref{sec:Hausd} we discuss the notion of dimension of a probability measure on a compact metric space. We introduce new \emph{lower} and \emph{upper mean dimensions} and establish in \thmref{th:dims} inequalities (used in the proof of \thmref{th:dim}) relating them to other more traditional definitions of dimension (including the classical Hausdorff dimension). After discussing the notion of the limit set in the higher rank case in \secref{sec:limit}, we finally pass to proving our Main Theorem in \secref{sec:link} and \secref{sec:const}. Namely, in \secref{sec:link} we establish the inequality of \thmref{th:dim}, and in \secref{sec:const} we prove \thmref{th:estimate} on existence of jump distributions with uniformly bounded entropy and arbitrarily big Lyapunov exponents.
\section{Random products} \label{sec:RW}
We begin by recalling the basic definitions from the theory of random walks on discrete groups, e.g., see \cite{Kaimanovich-Vershik83,Kaimanovich00a}. The \emph{random walk} determined by a probability measure $\mu$ on a countable group $\G$ is a Markov chain with the transition probabilities $$ p(g,gh)=\mu(h) \;, \qquad g,h\in\G \;. $$
\conv{Without loss of generality (passing, if necessary, to the subgroup generated by the support $\supp\mu$ of the measure $\mu$) we shall always assume that the measure $\mu$ is \emph{non-degenerate} in the sense that the smallest subgroup of $\G$ containing $\supp\mu$ is $\G$ itself.}
If the position of the random walk at time $0$ is a point $x_0\in\G$, then its position at time $n>0$ is the product $$ x_n = x_0 h_1 h_2 \dots h_n \;, $$ where $(h_n)_{n\ge 1}$ is the sequence of independent $\mu$-distributed \emph{increments} of the random walk. Therefore, provided that $x_0$ is the group identity $e$, the distribution of $x_n$ is the $n$-fold convolution $\mu^{*n}$ of the measure $\mu$.
Below it will be convenient to consider bilateral sequences of Bernoulli $\mu$-distributed increments $\h=(h_n)_{n\in\Z}$ and the associated \emph{bilateral sample paths} $\x=(x_n)_{n\in\Z}$ obtained by extending the formula $$ x_{n+1}=x_n h_{n+1} $$ to all $n\in\Z$ under the condition $x_0=e$, so that
\begin{equation} \label{eq:iso} x_n = \begin{cases}
h_0^{-1}h_{-1}^{-1} \dots h_{n+1}^{-1}\;, & n<0\;; \\
e\;, & n=0\;; \\
h_1 h_2\dots h_n\;, & n>0\;. \\ \end{cases} \end{equation}
Thus, the ``negative part'' $\ch x_n=x_{-n},\;n\ge 0,$ of the bilateral random walk is the random walk on $\G$ starting from the group identity at time 0 and governed by the \emph{reflected measure} $\ch\mu(g)=\mu(g^{-1})$.
We shall denote by $\P$ the probability measure in the space $\X=\G^\Z$ of bilateral sample paths $\x$ which is the image of the Bernoulli measure in the space of bilateral increments $\h$ under the isomorphism \eqref{eq:iso}. It is preserved by the transformation $T$ induced by the shift in the space of increments:
\begin{equation} \label{eq:T} (T\x)_n = h_1^{-1} x_{n+1}\;, \qquad n\in\Z\;. \end{equation}
\section{Lyapunov exponents} \label{sec:exponents}
We shall fix a standard basis $\E=(e_i)_{1\le i\le d}$ in $\R^d$ and identify elements of the Lie group $G=SL(d,\R)$ with their matrices in this basis. Throughout the paper we shall use the standard Euclidean norms associated with this basis both for vectors and matrices in $\R^d$.
\conv{From now on we shall assume that $\G$ is a countable subgroup of the group $G=SL(d,\R)$.}
Denote by $A$ the \emph{Cartan subgroup} of $G$ consisting of diagonal matrices $a=\diag(a_i)$ with positive entries $a_i$, and by $$ \af = \left\{ \a=(\a_1,\a_2,\dots,\a_d)\in\R^d : \sum\a_i=0 \right\} $$ the associated \emph{Cartan subalgebra}, so that $A=\exp\af$. Let $$ \af_+ = \{\a\in\af: \a_1>\a_2> \dots> \a_d \} \;, $$ be the standard \emph{positive Weyl chamber} in $\af$, and let $$ A_+ = \exp\af_+ = \{a\in A: a_1> a_2> \dots > a_d\} \;. $$ Denote the closures of $\af_+$ and $A_+$ by $\ov{\af_+}$ and $\ov{A_+}$, respectively.
Any element $g\in G$ can be presented as $g=k_1 a k_2$ with $k_{1,2}\in K=SO(d)$ (which is a maximal compact subgroup in $G$) and a uniquely determined $a=a(g)\in\ov{A_+}$ (\emph{Cartan} or \emph{polar} decomposition).
\conv{We shall always assume that the probability measure $\mu$ on $\G$ has a
\emph{finite first moment} in the ambient group $G$, i.e., $\sum \log \|g\| \mu(g) < \infty$. }
Then the asymptotic behaviour of the random walk $(\G,\mu)$ is described by the famous \emph{Oseledets multiplicative ergodic theorem} which we shall state in the form due to Kaimanovich \cite{Kaimanovich89} (and in the generality suitable for our purposes):
\begin{thm} \label{th:osel} There exists a vector $\l=\l(\mu)\in\ov\af_+$ (the \emph{Lyapunov spectrum} of the random walk) such that $$ \frac1n \log a(x_n) \mathop{\,\longrightarrow\,}_{n\to+\infty} \l $$ for $\P$-a.e.\ sample path $\x\in\X$. Moreover, for $\P$-a.e.\ $\x$ there exists a uniquely determined positive definite symmetric matrix $$ g=g(\x)=k(\exp\l) k^{-1} \;, \qquad k\in K \;, $$ such that $$
\log \|g^{-n}x_n\|=o(n) \quad \textrm{as} \quad n\to+\infty\;. $$ \end{thm}
\begin{thm}[\cite{Guivarch-Raugi85,Goldsheid-Margulis89}] \label{th:simple} If, in addition, the group $\G$ is Zariski dense in $G$, then the Lyapunov spectrum $\l(\mu)$ is \emph{simple}, i.e., it belongs to the Weyl chamber $\af_+$. \end{thm}
\begin{rem} \label{rem:inverse} The Lyapunov spectra $(\l_1,\dots,\l_d)$ and $(\ch\l_1,\dots,\ch\l_d)$ of the measure $\mu$ and of the reflected measure $\ch\mu$, respectively, are connected by the formula $$ \ch\l_i = -\l_{d+1-i} \;, \qquad 1\le i\le d\;. $$ \end{rem}
\section{Limit flags} \label{sec:flags}
Let $S=G/K$ be the \emph{Riemannian symmetric space} associated with the group $G$ (e.g., see \cite{Eberlein96} for the basic notions). We shall fix a reference point $o=K\in S$ (its choice is equivalent to choosing a Euclidean structure on $\R^d$ such that its rotation group is $K$). Being non-positively curved and simply connected, the space $S$ has a natural \emph{visibility compactification} $\ov S = S \cup \pt S$ whose boundary $\pt S$ consists of asymptotic equivalence classes of geodesic rays in $S$ and can be identified with the unit sphere of the tangent space at the point $o$ (since any equivalence class contains a unique ray issued from $o$). The action of the group $G$ extends from $S$ to $\pt S$, and the orbits of the latter action are naturally parameterized by unit length vectors $\a\in\ov{\af_+}$: the orbit $\pt S_\a$ consists of the equivalence classes of all the rays $\g(t)=k\exp(t\a)o,\; k\in K$. Algebraically the orbits $\pt S_\a$ corresponding to the interior vectors $\a\in\af_+$ are isomorphic to the space $\B$ of \emph{full flags} $$ \V=\{V_i\}\;,\quad V_0=\{0\}\subset V_1\subset\dots\subset V_{d-1}\subset V_d=\R^d $$ in $\R^d$ (also known as the \emph{Furstenberg boundary} of the symmetric space $S$), whereas the orbits corresponding to wall vectors are isomorphic to quotients of $\B$, i.e., to flag varieties for which certain intermediate dimensions are missing, see \cite{Kaimanovich89}.
\thmref{th:osel} implies that for $\P$-a.e.\ sample path $\x$ $$
d(x_n o, \g(t\|\l\|))=o(n) \quad \textrm{as} \quad n\to+\infty \;, $$
where $\g$ is the geodesic ray $\g(t)=k\exp\left(t\frac{\l}{\|\l\|}\right)o$, hence the sequence $x_n o$ converges in the visibility compactification to a limit point $\bnd\x\in\pt S_{\l/\|\l\|}$. Moreover, $\pt S_{\l/\|\l\|}\cong\B$ by \thmref{th:simple}, so that below we shall consider the aforementioned \emph{boundary map} $\bnd$ as a map from the path space to the flag space $\B$.
\thmref{th:osel} and \thmref{th:simple} easily imply the following descriptions of the limit flag $\bnd\x$.
\begin{prop} \label{pr:lyap}
\begin{itemize} \item[(i)] Denote by $\V_0$ the \emph{standard flag} $$ \qquad \{0\} \subset \spn\{e_1\} \subset \spn\{e_1,e_2\} \subset \dots \subset \spn\{e_1,e_2,\dots,e_{d-1}\} \subset \R^d $$ associated with the basis $\E$. Then $$ \bnd\x = k\V_0 $$ for $k=k(\x)\in K$ from \thmref{th:osel}.
\item[(ii)] The spaces $V_i$ from the flag $\bnd\x$ are increasing direct sums of the eigenspaces of the matrix $g=g(\x)$ from \thmref{th:osel} taken in the order of decreasing eigenvalues.
\item[(iii)] The flag $\bnd\x$ is the \emph{Lyapunov flag} of the sequence $x_n^{-1}$, i.e., $$
\lim_{n\to+\infty} \frac1n \log \|x_n^{-1} v\| = -\l_i \qquad \forall\, v\in V_i\setminus V_{i-1} \;. $$
\item[(iv)] For any smooth probability measure $\th$ on $\B$ and $\P$-a.e.\ sample path $\x$ $$ \lim_{n\to+\infty} x_n\th = \d_{\bnd\x} $$ in the weak$^*$ topology of the space of probability measures on $\B$. \end{itemize} \end{prop}
\begin{rem} Equivariance of the boundary map $\bnd$ implies that for the transformation $T$ \eqref{eq:T} $$ \bnd T^n\x = x_n^{-1}\bnd\x \qquad\forall\,n\in\Z \;. $$ In particular,
\begin{equation} \label{eq:Tc} \bnd T^{-n}\x = \ch x_n^{-1}\bnd\x \qquad\forall\,n\ge 0 \;. \end{equation}
\end{rem}
\section{Harmonic measure} \label{sec:harm}
\begin{defn} The image $\nu=\bnd(\P)$ of the probability measure $\P$ in the path space $\X$ under the map $\bnd:\X\to\B$ is called the \emph{harmonic measure} of the random walk $(\G,\mu)$. In other words, $\nu$ is the distribution of the limit flag $\bnd\x$ under the measure $\P$. \end{defn}
The harmonic measure is \emph{$\mu$-stationary} in the sense that it is invariant with respect to the convolution with $\mu$: $$ \mu*\nu = \sum_g \mu(g) g\nu = \nu \;. $$
\begin{thm}[\cite{Guivarch-Raugi85,Goldsheid-Margulis89}] \label{th:sub} Under conditions of \thmref{th:simple} $\nu$ is the unique $\mu$-stationary probability measure on $\B$, and any proper algebraic subvariety of $\B$ is $\nu$-negligible. \end{thm}
\begin{thm}[\cite{Kaimanovich85,Ledrappier85}] Under conditions of \thmref{th:simple}, if the subgroup $\G$ is discrete, then the measure space $(\B,\nu)$ is isomorphic to the Poisson boundary of the random walk $(\G,\mu)$. \end{thm}
\begin{rem} If $\G$ is not discrete, then the random walk may have limit behaviours other than described just by the limit flags, or, in other words, the Poisson boundary of the random walk $(\G,\mu)$ may be bigger than the flag space, see \cite{Kaimanovich-Vershik83} for the first example of this kind (the dyadic-rational affine group) and \cite{Bader-Shalom06,Brofferio06} for the recent developments. \end{rem}
Denote by $\Gr_i$ the dimension $i$ \emph{Grassmannian} (the space of all dimension $i$ subspaces) in $\R^d$. There is a natural projection $\pi_i:\B\to\Gr_i$ which consists in assigning to any flag in $\R^d$ its dimension $i$ subspace. Let $\nu_i=\pi_i(\nu),\;1\le i\le d-1,$ be the associated images of the measure $\nu$. Obviously, the measures $\nu_i$ are $\mu$-stationary (along with the measure $\nu$). We shall also use the notation $$ \bnd_i\x = \pi_i(\bnd\x) \in \Gr_i \;, $$ so that $\nu_i=\bnd_i(\P)$.
We shall embed each Grassmannian $\Gr_i$ into the projective space $P\bigwedge^i \R^d$ in the usual way and define a $K$-invariant metric $\r=\r_i$ on the latter as
\begin{equation} \label{eq:metric} \r (\xi,\z) = \sin\an(\xi,\z) \;, \end{equation}
where the angle (varying between $0$ and $\pi/2$) is measured with respect to the standard Euclidean structure on $\bigwedge^i\R^d$ determined by the basis $\E$ (so that $(e_{j_1}\wedge \cdots \wedge e_{j_i})_{1\le j_1 < \cdots < j_i \le d}$ is an orthonormal basis of $\bigwedge^i\R^d$).
The \emph{Furstenberg formula} (see \cite{Furstenberg63a,Bougerol-Lacroix85})
\begin{equation} \label{eq:furst} \l_1 + \l_2+\dots+\l_i =
\sum_g \mu(g) \int_{\Gr_i} \log\frac{\|gv\|}{\|v\|}\ d\nu_i (\ov{v}) \;,
\qquad 1\le i\le d-1 \;, \end{equation}
where $v\in\bigwedge^i\R^d$ is the vector presenting a point $\ov{v}\in\Gr_i\subset P\bigwedge^i\R^d$, relates Lyapunov exponents with the harmonic measure.
\section{Entropy} \label{sec:entr}
Recall that if the measure $\mu$ has a finite entropy $H(\mu)$, then the \emph{asymptotic entropy} of the random walk $(\G,\mu)$ is defined as
\begin{equation}\label{eq:entr} h(\G,\mu)=\lim_{n\to+\infty} \frac{H(\mu^{*n})}n \le H(\mu)\;, \end{equation}
where $H(\cdot)$ denotes the usual entropy of a discrete probability measure. The asymptotic entropy can also be defined ``pointwise'' along sample paths of the random walk as $$ h(\G,\mu) = \lim_{n\to+\infty} -\frac1n\log\mu^{*n}(x_n) \;, $$ where the convergence holds both $\P$-a.e.\ and in the space $L^1(\X,\P)$, see \cite{Kaimanovich-Vershik83,Derriennic86}.
The \emph{$\mu$-entropy} (\emph{Furstenberg entropy, differential entropy}) of a $\mu$-stationary measure $\th$ on a $\G$-space $X$ is defined as
\begin{equation}\label{eq:def diff}
E_\mu(X,\th) = -\sum_{g\in \G}\mu(g)\int\log \frac{dg^{-1}\th}{d\th}(b)d\th(b)
\;, \end{equation}
and it satisfies the inequality $E_\mu(X,\th)\le h(\G,\mu)$, see \cite{Furstenberg71,Kaimanovich-Vershik83,Nevo-Zimmer02}.
\conv{We shall always assume that the probability measure $\mu$ on $\G$ has a \emph{finite entropy} $H(\mu)<\infty$. }
In our context, if the subgroup $\G$ is discrete in $SL(d,\R)$, then finiteness of the first moment of the measure $\mu$ easily implies that $H(\mu)<\infty$ (e.g., see \cite{Derriennic86}). Therefore, $E_i=E_\mu(\Gr_i,\nu_i)<\infty$.
Below we shall need the following routine estimate (in fact valid for an arbitrary quotient of the Poisson boundary).
\begin{lem} \label{lem:dentr} For any index $i\in\{1,2,\dots,d-1\}$, any subset $A\subset\Gr_i$ with $\nu_i(A)>0$ and $\P$-a.e.\ sample path $\x$
\begin{equation}\label{eq:dentr} \limsup_{n\to+\infty}\left[-\frac{1}{n} \log \ch x_n\nu_i(A)\right] \le E_i \;. \end{equation}
\end{lem}
\begin{proof} Put $$ F_i(\x)=-\log\frac{d\ch x_1\nu_i}{d\nu_i} (\bnd_i\x)=-\log\frac{d\nu_i(\bnd_iT^{-1}\x)}{d\nu_i(\bnd_i\x)} $$ (see formula \eqref{eq:Tc}), so that $$ E_i = \int F_i(\x)\,d\P(\x) \;. $$ Then $$ \begin{aligned} -\log\frac{d\ch x_n\nu_i}{d\nu_i} (\bnd_i\x)
&= -\log\frac{d\nu_i(\bnd_iT^{-n}\x)}{d\nu_i(\bnd_i\x)} \\
&= F_i(\x) + F_i(T^{-1}\x) + \dots + F_i(T^{-n+1}\x) \;, \end{aligned} $$ whence by the ergodic theorem $$ -\frac1n \log\frac{d\ch x_n\nu_i}{d\nu_i} (\bnd_i\x) \mathop{\,\longrightarrow\,}_{n\to+\infty} E_i $$ in $L^1(\X,\P)$. It implies that $$ -\frac1n \int_A \log\frac{d\ch x_n\nu_i}{d\nu_i} (\xi) \frac{d\nu_i(\xi)}{\nu_i(A)} \mathop{\,\longrightarrow\,}_{n\to+\infty} E_i \;, $$ which, by a convexity argument, yields the claim. \end{proof}
\section{Dimension of measures} \label{sec:Hausd}
Let us recall several notions of the dimension of a probability measure $m$ on a compact metric space $(Z,\r)$ (all the details, unless otherwise specified, can be found in the book \cite{Pesin97}). These notions roughly fall into two categories: the \emph{global} ones are obtained by looking at the dimension of sets which ``almost'' coincide with $Z$ (up to a piece of small measure $m$), whereas the \emph{local} ones are related to the asymptotic behavior of the ratios $\log m B(z,r)/\log r$ as the radius $r$ tends to $0$.
\subsection{Global definitions}
The \emph{Hausdorff dimension} of the measure $m$ is $$ \dim_H m = \inf\left\{ \dim_H A : m(A)=1 \right\} \;, $$ where $\dim_H A$ denotes the Hausdorff dimension of a subset $A\subset Z$.
The \emph{lower} and the \emph{upper box dimensions} of a subset $A\subset Z$ are defined, respectively, as $$ \un\dim_B A = \liminf_{r\to 0}\frac{\log N(A,r)}{\log 1/r}
\quad\textrm{and}\quad
\ov{\dim}_B A = \limsup_{r\to 0}\frac{\log N(A,r)}{\log 1/r} \;, $$ where $N(A,r)$ is the minimal number of balls of radius $r$ needed to cover $A$. Ledrappier \cite{Ledrappier81} also considered the minimal number $N(r,\e,m)$ of balls of radius $r$ such that the measure $m$ of their union is at least $1-\e$ and defined the ``fractional dimension'' of the measure $m$ as $$ \ov\dim_L m = \sup_{\e\to 0} \limsup_{r\to 0} \frac{\log N(r,\e,m)}{\log 1/r} $$ (we use the notation from \cite{Young82} and below shall call it the \emph{upper Ledrappier dimension}). As it was noticed by Young \cite{Young82}, in the same way one can also define what we call the \emph{lower Ledrappier dimension} of the measure $m$ $$ \un\dim_L m = \sup_{\e\to 0} \liminf_{r\to 0} \frac{\log N(r,\e,m)}{\log 1/r} $$ as well as, in modern terminology, its \emph{lower} and the \emph{upper box dimensions}, respectively, $$ \un\dim_B m = \liminf_{m(A)\to 1} \left\{ \un\dim_B A \right\} \quad \textrm{and} \quad
\ov\dim_B m = \liminf_{m(A)\to 1} \left\{ \ov\dim_B A \right\} \;. $$ Obviously, $$ \un\dim_L m \le \ov\dim_L m \quad\textrm{and}\quad \un\dim_B m \le \ov\dim_B m \;. $$
The difference between the Ledrappier and the box dimensions is that in the definition of the box dimensions it is the same set $A$ which has to be covered by balls with varying radii $r$, unlike in the definition of Ledrappier, so that $$ \un\dim_L m \le \un\dim_B m \quad\textrm{and}\quad \ov\dim_L m \le \ov\dim_B m \;. $$
By \cite[Proposition 4.1]{Young82},
\begin{equation}\label{eq:dims} \dim_H m \le \un\dim_L m \;. \end{equation}
\subsection{Local definitions}
The \emph{lower} and the \emph{upper pointwise dimensions} of the measure $m$ at a point $z\in Z$ are $$ \un\dim_P m (z) = \liminf_{r\to 0}\frac{\log m B(z,r)}{\log r}
\quad
\textrm{and}
\quad
\ov\dim_P m (z) = \limsup_{r\to 0}\frac{\log m B(z,r)}{\log r} \;, $$ respectively. Then
\begin{equation} \label{eq:HD} \dim_H m = \esssup_z \un\dim_P m (z) \;. \end{equation}
In particular, if $m$-a.e.\
\begin{equation} \label{eq:D} \lim_{r\to 0}\frac{\log m B(z,r)}{\log r}=D \;, \end{equation}
then $\dim_H m=D$. Moreover, in this case all the reasonable definitions of dimension of the measure $m$ give the same result \cite[Theorem 4.4]{Young82}.
\begin{defn} \label{def:M} In the situation when the convergence in \eqref{eq:D} holds just in probability we shall say that $D$ is the \emph{mean dimension} $\dim_M m$ of the measure $m$. We shall also introduce the \emph{lower} and the \emph{upper mean dimensions} of the measure $m$ as, respectively, $$ \begin{aligned} \un\dim_M m &= \sup\left\{t: \left[\frac{\log m B(z,r)}{\log r} - t\right]_-\mathop{\,\longrightarrow\,}^m 0 \right\} \;, \\ \ov\dim_M m &= \inf\left\{t: \left[\frac{\log m B(z,r)}{\log r} - t\right]_+\mathop{\,\longrightarrow\,}^m 0 \right\} \;, \end{aligned} $$ where $[t]_+=\max\{0,t\}, [t]_-=\min\{0,t\}$, and $\displaystyle{\mathop{\,\longrightarrow\,}^m}$ denotes convergence in probability with respect to the measure $m$. \end{defn}
The definition of $\dim_M$ first appeared in Ledrappier's paper \cite{Ledrappier83} (also see \cite{Ledrappier84}), whereas $\un\dim_M$ and $\ov\dim_M$ (although obvious generalizations of $\dim_M$) are, apparently, new. Clearly, $\un\dim_M m \le \ov\dim_M m$. In slightly different terms, $[\un\dim_M m, \ov\dim_M m]$ is the minimal closed subinterval of $\R$ with the property that for any closed subset $I$ of its complement $$ m\left\{z\in Z: \frac{\log m B(z,r)}{\log r}\in I\right\} \mathop{\,\longrightarrow\,}_{r\to 0} 0 \;. $$ In particular, if $\dim_M m$ exists, then $\un\dim_M m=\ov\dim_M m=\dim_M m$.
\subsection{Mean, box and Ledrappier dimensions} \label{sec:bes}
We shall now establish simple inequalities between these dimensions.
\begin{prop} For any probability measure $m$ on a compact metric space $(Z,\r)$ $$ \un\dim_M m \le \un\dim_L m \;. $$ \end{prop}
\begin{proof} Fix a number $D<\un\dim_M m$. Then for any $\e>0$ there exist $r_0>0$ and a set $A\subset Z$ with $m(A)>1-\e$ and such that $m B(z,r)\le r^D$ for all $z\in A$ and $r\le r_0$. Suppose now that $$ m \left( \bigcup_i B(z_i,r) \right) \ge 1-\e $$ for a certain set of points $\{z_i\}$ of cardinality $N$ and a certain $r\le r_0/2$. If $B(z_i,r)$ intersects $A$, then $B(z_i,r)\subset B(z,2r)$ for some $z\in A$, so that $$ m\left( A \cap B(z_i,r) \right) \le m B(z_i,r) \le m B(z,2r) \le (2r)^D \;. $$ Thus, $$ 1-2\e \le m \left( A \cap \bigcup_i B(z_i,r) \right) \le N (2r)^D \;, $$ whence the claim. \end{proof}
For establishing the inverse inequality we shall additionally require that the metric space $(Z,\r)$ has the \emph{Besicovitch covering property}, i.e., that for any precompact subset $A\subset Z$ and any bounded function $r:A\to\R_+$ (important particular case: $r$ is constant) the cover $\{B(z,r(z)), z\in A\}$ of $A$ contains a countable subcover whose multiplicity is bounded from above by a universal constant $M=M(Z,\r)$ (recall that the \emph{multiplicity} of a cover is the maximal number of elements of this cover to which a single point may belong). The Besicovitch property is, for instance, satisfied for the Euclidean space, hence, for all its compact subsets endowed with a metric which is Lipschitz equivalent to the Euclidean one. Therefore, it is satisfied for each of the Grassmannians $\Gr_i$ endowed with the metric \eqref{eq:metric}.
\begin{prop} For any probability measure $m$ on a compact space $(Z,\r)$ satisfying the Besicovitch property $$ \ov\dim_B m \le \ov\dim_M m \;. $$ \end{prop}
\begin{proof} Take a number $D>\ov\dim_M m$, let $$ A_r = \left\{z\in Z : m B(z,r)\ge r^D \right\} \;, $$ and consider a cover of $A_r$ by the balls $B(z_i,r),\;z_i\in A_r$ obtained from applying the Besicovitch property. The cardinality of this cover is at least $N(A_r,r)$, whereas its multiplicity is at most $M$, whence $$ N(A,r)r^D \le \sum_i m B(z_i,r) \le M \;, $$ so that $$ \frac{\log N(A_r,r)}{\log 1/r}\le D+\frac{\log M}{\log 1/r} \;. $$ For $r\to0$ the right-hand side of the above inequality tends to $D$, whereas $m(A_r)\to 1$ by the choice of $D$, whence $\ov\dim_B m\le D$. \end{proof}
\subsection{Final conclusions}
Taking stock of the above discussion we obtain
\begin{thm}\label{th:dims} For any probability measure $m$ on a compact metric space $$ \left. \begin{aligned} &\dim_H m\\ &\,\un\dim_M m\\ \end{aligned} \right\} \le \un\dim_L m \le \un\dim_B m \;. $$ If, in addition, the space has the Besicovitch property, then $$ \ov\dim_L m \le \ov\dim_B m \le \ov\dim_M m \;. $$ \end{thm}
\begin{rem} There is no general inequality between $\dim_H m$ and $\un\dim_M m$. For instance, take two singular measures $m_1,m_2$ for which the dimensions \eqref{eq:D} exist and are different, say, $D_1<D_2$, and let $m$ be their convex combination. Then $\un\dim_M m=D_1$, whereas $\dim_H m=D_2>\un\dim_M m$ by \eqref{eq:HD}.
On the other hand, by exploiting the difference between the convergence in probability and the convergence almost everywhere one can also construct examples with $\dim_H m<\un\dim_M m$. We shall briefly describe one such example. Let $Z'$ be the space of unilateral binary sequences $\a=(\a_1,\a_2,\dots)$ with the uniform measure $m'$ and the usual metric $\r'$ for which $-\log\r'(\a,\b)$ is the length of the initial common segment of the sequences $\a$ and $\b$. Take a sequence of cylinder sets $A_n$ with the property that $m' A_n\to 0$, but any point of $Y$ belongs to infinitely many sets $A_n$. Also take a sequence of integers $s_n$ (to be specified later), and let $Z$ be the image of the space $Z'$ under the following map: given a sequence $\a\in Z'$ take the set $I=\{n:\a\in A_n\}$ and replace with 0 all the symbols $\a_k$ with $s_n\le k\le 2s_n$ for a certain $n\in I$. The space $Z$ is endowed with the quotient measure $m$ and the quotient metric $\r$. If the sequence $s_n$ is very rapidly increasing, then \eqref{eq:HD} can be used to show that $\dim_H m = \log 2/2$, whereas $\un\dim_M m=\dim_M m=\log 2$. \end{rem}
\section{Limit set} \label{sec:limit}
Denote by $\Pr(\B)$ the compact space of probability measures on the flag space $\B$ endowed with the weak$^*$ topology, and let $m$ be the unique $K$-invariant probability measure on $\B$. Then the map $go\mapsto gm$ determines an embedding of the symmetric space $S=G/K$ into $\Pr(\B)$, and gives rise to the \emph{Satake--Furstenberg compactification} of $S$. Its boundary $\ov S\setminus S$ contains the space $\B$ under the identification of its points with the associated delta-measures (but, unless the rank of $G=SL(d,\R)$ is 1, i.e., $d=2$, it also contains other limit measures, see \cite{Guivarch-Ji-Taylor98}). The \emph{limit set} $L_\G$ of a subgroup $\G\subset G$ in this compactification is then defined (see \cite{Guivarch90}) as $$ L_\G =\ov{\G o}\cap \B \subset \B \;. $$ The limit set is obviously $\G$-invariant and closed. Moreover,
\begin{thm}[\cite{Guivarch90}] \label{th:limset} If the group $\G$ is Zariski dense in $G$, then its action on the limit set $L_\G$ is minimal (i.e., $L_\G$ has no proper closed $\G$-invariant subsets). \end{thm}
Below we shall need the following elementary property.
\begin{prop} \label{pr:min} Under the conditions of \thmref{th:limset} let $U\subset\B$ be an open set with $U\cap L_\G \neq \emp$. Then there exists finitely many elements $\g_1,\dots,\g_r\in\G$ such that $$ L_\G \subset \bigcup_i \g_i U\ . $$ \end{prop}
\begin{proof} Minimality of $L_\G$ means that any closed $\G$-invariant subset of $\B$ either contains $L_\G$ or does not intersect it. Since the set $\B\setminus\bigcup_{\g\in\G}\g U$ does not contain $L_\G$ (as $U\cap L_\G \neq \emp$), it does not intersect $L_\G$, so that $L_\G \subset \bigcup_{\g\in\G}\g U$. Finally, since $L_\G$ is compact, the above cover contains a finite subcover. \end{proof}
\begin{rem} \thmref{th:limset} and \propref{pr:min} obviously carry over to the projections $L^i_\G=\pi_i(L_\G)\subset\Gr_i$ of the limit set $L_\G$ to the Grassmannians $\Gr_i$. \end{rem}
By \propref{pr:lyap}(iv) (see \cite{Guivarch-Raugi85} for a more general argument) a.e.\ sample path $\x$ converges as $n\to+\infty$ to the limit flag $\bnd\x$ in the Satake--Furstenberg compactification as well. Therefore, $\supp\nu\subset L_\G$. If $\supp\mu$ generates $\G$ as a semigroup, then $\mu$-stationarity of the harmonic measure $\nu$ implies its quasi-invariance, so that in this case $\supp\nu$ is $\G$-invariant, and, if $\G$ is Zariski dense, $\supp\nu=L_\G$ by \thmref{th:limset}.
\begin{rem} If $\G$ is a lattice, then $L_\G=\B$. On the other hand, if $\G$ is Zariski dense, then, as it is shown in \cite{Guivarch90}, $L_\G$ has positive Hausdorff dimension, which is deduced from the positivity of the dimension of the harmonic measure (under the assumption that $\mu$ has an exponential moment). Also see \cite{Link04} for recent results on the Hausdorff dimension of the \emph{radial limit set}. It would be interesting to investigate existence of random walks such that their harmonic measure has the maximal Hausdorff dimension (equal to the dimension of the limit set), cf. \cite{Connell-Muchnik07}. \end{rem}
\section{Dimension of the harmonic measure} \label{sec:link}
\subsection{Rate of contraction estimate}
Recall (see \secref{sec:RW}) that the negative part $(\ch x_n)_{n\ge 0}=(x_{-n})_{n\ge 0}$ of a bilateral sample path $\x$ performs the random walk on $\G$ governed by the reflected measure $\ch\mu$. Denote by $\bnd^-\x\in\B$ the corresponding limit flag, and for $i\in\{1,2,\dots,d-1\}$ let $$ \xi_\x=\left(\bnd_{d-i}^-\x\right)^\bot\in\Gr_i $$ be the orthogonal complement in $\R^d$ of the $(d-i)$-dimensional subspace of the flag $\bnd^-\x$ (for simplicity we omit the index $i$ in the notation for $\xi_\x$; the Grassmannian $\Gr_i$ to which it belongs should always be clear from the context).
\begin{thm} \label{th:contr} For any Grassmannian $\Gr_i$, any $r<1$, and $\P$-a.e.\ $\x\in\X$, $$ \liminf_n \left[ - \frac1n \log\diam \ch x_n^{-1} B(\xi_\x,r) \right] \ge \l_i - \l_{i+1} \;. $$ \end{thm}
\begin{proof} Let us consider first the case when $i=1$. Then by the definition of the metric $\r$ \eqref{eq:metric} on $\Gr_1\cong P\R^d$ the ball $B(\xi_\x,r)$ consists of the projective classes of all the vectors $v+w$, where $v\in\xi_\x\setminus\{0\}$ and $w\bot\xi_\x$ with
\begin{equation} \label{eq:C}
\frac{\|w\|}{\|v\|}\le C \quad \textrm{for a certain constant} \quad C=C(r) \;. \end{equation}
By \propref{pr:lyap}(iii) applied to the random walk $(\G,\ch\mu)$ we have that $$
\lim_n \frac1n \log\frac{\|\ch x_n^{-1} w\|}{\|\ch x_n^{-1} v\|} \le\l_2-\l_1 $$ uniformly under condition \eqref{eq:C}, whence the claim.
The general case argument follows the same lines with the only difference that now instead of the action on $P\R^d$ one has to consider the action on $P\bigwedge^i\R^d$. The sequence of matrices $\bigwedge^i \ch x_n^{-1}$ (which are the images of $\ch x_n^{-1}$ in the $i$-th external power representation) is also Lyapunov regular with the Lyapunov spectrum consisting of all the sums of the form $\l_{j_1}+\l_{j_2}+\dots+\l_{j_i}$ with $1\le j_1 < \cdots < j_i \le d$. In particular, the top of the spectrum (corresponding precisely to the eigenspace $\xi_\x\in P\bigwedge^i\R^d$ as orthogonal to the highest dimension proper subspace of the Lyapunov flag) is $\l_1+\l_2+\dots+\l_i$, whereas the second point in the spectrum is $\l_1+\l_2+\dots+\l_{i-1}+\l_{i+1}$, the spectral gap being $\l_i-\l_{i+1}$. \end{proof}
\begin{prop}\label{pr:balls} For any index $i\in\{1,2,\dots,d-1\}$, any $\e>0$ and $\P$-a.e.\ sample path $\x$ the sets $$ A_N = \bigcap_{n\ge N} \ch x_n B \left( \bnd_iT^{-n}\x, e^{-n(\l_i-\l_{i+1}-\e)}\right) \subset \Gr_i $$ have positive measure $\nu_i$ for all sufficiently large $N$. \end{prop}
\begin{proof} By definition of the metric $\r$ \eqref{eq:metric} its values do not exceed 1, and for any $\xi\in\Gr_i$ the radius 1 sphere $S(\xi,1)$ centered at $\xi$ consists precisely of those $\z\in\Gr_i$ for which the vectors from $\bigwedge^i\R^d$ associated with $\xi$ and $\z$ are orthogonal. The latter condition on $\z$ is algebraic, so that by \thmref{th:sub} $\nu_i S(\xi,1)=0$ for any $\xi\in\Gr_i$ (the measure $\nu_i$ being the projection of the measure $\nu$). Therefore, for a.e.\ sample path $\x$ there exists $r<1$ such that $\nu_i B(\xi_\x,r)>0$. On the other hand, by \thmref{th:contr} and \eqref{eq:Tc} $B(\xi_\x,r)\subset A_N$ for all sufficiently large $N$. \end{proof}
\subsection{Dimension estimate}
Recall that $\bnd_i T^{-n}\x = \ch x_n^{-1}\bnd_i\x$ \eqref{eq:Tc}. By \lemref{lem:dentr} and \propref{pr:balls}, for $\P$-a.e.\ sample path $\x$
\begin{equation}\label{eq:minentr} \begin{split} &\limsup_{n\to\infty} \left[ -\frac1n \log\nu_i B \left( \bnd_iT^{-n}\x, e^{-n(\l_{i}-\l_{i+1}-\e)} \right) \right] \\
&\le \limsup_{n\to\infty}\left[-\frac{1}{n} \log \ch x_n\nu_i(A_N)\right] \le E_i \;. \end{split} \end{equation}
The left-hand side of this inequality looks almost like in the definition of the pointwise dimension, with the only difference that the ball centers vary. We shall take care of this difference by switching to the mean dimension and using $\mu$-stationarity of the measures $\nu_i$.
\begin{thm}\label{th:dim} For any $i\in\{1,2,\dots,d-1\}$, $$ \dim_H(\nu_i)\le\ov\dim_B(\nu_i)\le\ov\dim_M\nu_i \le \frac{E_i}{\l_i-\l_{i+1}} \;. $$ \end{thm}
\begin{proof} Let $$ A = \bigcap_\eta\bigcup_N\bigcap_{n\ge N}\Big\{\x :
-\frac1n\log {\nu}^i B\big(\bnd_iT^{-n}\x,e^{-n(\l_i-\l_{i+1}-\e)}\big) \le
E_i+\eta\Big\} \;. $$ be the set of sample paths satisfying condition \eqref{eq:minentr}. Since $\P(A)=1$, for any $\eta,\chi>0$ $$ \P\Big\{\x :
-\frac{1}{n}\log{\nu}^i B\big(\bnd_iT^{-n}\x,e^{-n(\l_i-\l_{i+1}-\e)}\big)
\le E_i+\eta\Big\}\ge 1-\chi $$ for all sufficiently large $n$. The transformation $T$ preserves the measure $\P$, so that its image under the map $\x\mapsto\bnd_iT^{-n}\x$ is $\nu_i$, whence the rightmost inequality. The other inequalities follow from \thmref{th:dims} because the metric $\r$ has the Besicovitch property (see \secref{sec:bes}). \end{proof}
\section{The construction}\label{sec:const}
For the rest of this Section we shall fix a probability measure $\mu$ on a subgroup $\G\subset G$ satisfying our standing assumptions from \secref{sec:RW} and \secref{sec:exponents}.
\conv{In addition we shall assume in this Section that $\supp\mu$ generates $\G$ as a semigroup.}
Recall that an element of the group $G=SL(d,\R)$ is called \emph{$\R$-regular} if it is diagonalizable over $\R$ and the absolute values of its eigenvalues are pairwise distinct. By \cite{Benoist-Labourie93} any Zariski dense subgroup of $G$ contains such an element. Let us fix an $\R$-regular element $\g\in\G$ and consider the sequence of probability measures on $\G$
\begin{equation} \label{eq:mu_k} \mu^k = \frac12\mu+\frac14\left( \d_{\g^k}+\d_{\g^{-k}} \right) \;. \end{equation}
where $\d_g$ denotes the Dirac measure at $g$.
\begin{rem} Actually, we just need that $\g$ be diagonalizable over $\R$ with the absolute value of some of its eigenvalues being not $1$. \end{rem}
Denote by $\nu^k$ the harmonic measures on the flag space $\B$ of the random walks $\left(\G,\mu^k\right)$, and by $\nu^k_i$ their quotients on the Grassmannians $\Gr_i,\;1\le i \le d-1$. Our Main Theorem follows from
\begin{thm} \label{th:main} $$ \min_i\{\dim_H\nu_i^k\} \mathop{\,\longrightarrow\,}_{k\to\infty} 0 \;. $$ \end{thm}
In its turn, \thmref{th:main} is an immediate consequence of a combination of the inequality from \thmref{th:dim} and
\begin{thm} \label{th:estimate} The measures $\mu^k$ \eqref{eq:mu_k} have the property that the entropies $H(\mu^k)$ are uniformly bounded, whereas the lengths of their Lyapunov vectors $\l(\mu^k)$ (equivalently, the top Lyapunov exponents $\l_1(\mu^k)$) go to infinity. \end{thm}
The rest of this Section is devoted to a proof of \thmref{th:estimate} split into a number of separate claims. The first one is obvious:
\begin{claim}\label{pr:h is bdd} The measures $\mu^k$ have uniformly bounded entropies $H\left(\mu^k\right)$. \end{claim}
In view of the discussion from \secref{sec:entr} it immediately implies
\begin{cor}\label{cor:unif} The asymptotic entropies $h\left(\G,\mu^k\right)$ and therefore all the differential entropies $E^k_i=E_{\mu^k}\left(\Gr_i,\nu_i^k\right)$ are uniformly bounded. \end{cor}
For estimating the top Lyapunov exponent $\l_1\left(\mu^k\right)$ we shall use the Furstenberg formula \eqref{eq:furst}, by which
\begin{equation} \label{eq:l1} \begin{aligned} \l_1 \left(\mu^k\right)
&= \frac12 \sum_g \mu(g)\int_{P\R^d} \log\frac{\|gv\|}{\|v\|}\
d\nu^k_1(\ov{v}) \\
&\qquad + \frac14\int_{P\R^d}
\left( \log\frac{\|\g^k v\|}{\|v\|}+ \log\frac{\|\g^{-k} v\|}{\|v\|}\right)
d\nu^k_1(\ov{v}) \;. \end{aligned} \end{equation}
The absolute value of the first term of this sum is uniformly bounded because $\mu$ has a finite first moment. For dealing with the second term of the sum \eqref{eq:l1} we need the following claims.
\begin{claim}\label{claim:g^k} The sum $$
\left( \log\frac{\|\g^k v\|}{\|v\|}
+ \log\frac {\|\g^{-k} v\|} {\|v\|}\right) $$ is bounded from below uniformly on $v\in\R^d\setminus\{0\}$ and $k\ge 0$. \end{claim}
\begin{proof} If $\d$ is a diagonal matrix, then obviously, $$
\|\d^k v\| \|\d^{-k}v\| \ge \la \d^k v,\d^{-k} v\ra = \|v\|^2 \;. $$ Now, $\g=h^{-1}\d h$ is diagonalizable, whence $$ \begin{aligned}
\|\g^k v\| \|\g^{-k}v\| &\ge \|h\|^{-2} \|\d^k h v\| \|\d^{-k} h v \| \ge
\|h\|^{-2} \|hv\|^2\\ &\ge \|h\|^{-2}\|h^{-1}\|^{-2} \|v\|^2 \;. \end{aligned} $$ \end{proof}
\begin{claim} \label{claim:opens} For any open subset $U\subset P\R^d$ which intersects the limit set $L^1_\G$ (see \secref{sec:limit}), the measures $\nu^k_1(U)$ are bounded away from zero uniformly on $k$. \end{claim}
\begin{proof} By \propref{pr:min} there exists a finite set $A\subset\G$ such that $$ L^1_\G\subset \bigcup_{g\in A} g^{-1} U \;. $$ Since $\supp\mu$ generates $\G$ as a semigroup, for each $g\in A$ there exists an integer $s=s(g)$ such that $g\in\supp\mu^{*s}$. Then by $\mu^k$-stationarity of $\nu_1^k$, for each $g\in A$ $$ \nu^k_1(U) \ge \frac1{2^s} \mu^{*s}(g)g\nu_1^k(U) = \frac1{2^s} \mu^{*s}(g)\nu_1^k\left(g^{-1} U\right) \ge \e \nu_1^k\left(g^{-1} U\right) $$ for a certain $\e=\e(A)>0$. Summing up the above inequalities over all $g\in A$ we obtain $$
|A| \nu^k_1(U) \ge \e \sum_{g\in A} \nu_1^k\left(g^{-1} U\right) \ge \e\nu_1^k(L^1_\G) = \e \;, $$ whence the claim. \end{proof}
Now we are ready to prove
\begin{claim} \label{claim:l1} $$ \lim_{k\to\infty}\l_1\left(\mu^k\right)=+\infty \;. $$ \end{claim}
\begin{proof} We shall fix a diagonalization $\g=h^{-1}\d h$ with
$\d=\diag(\d_1,\dots,\d_d)$ in such a way that $|\d_1|>1$ and $|\d_d|<1$, and define the open set $U$ as $$ U = \left\{\ov{v}\in P\R^d :
\frac{\la h v,e_1\ra }{\|h v\|}>\b
\ \textrm{and}\ \frac{\la h v,e_d\ra }{\|h v\|}>\b \right\} \;, $$ i.e., by requiring that the first and the last coordinates (with respect to the standard basis $\E$) of the normalized vector $hv$ be greater than $\b$. The value of $\b$ is chosen to make sure that $U$ is non-empty (for instance, one can take $\b=1/2$).
If $\ov{v}\in U$, then $$
\|\g^k v\|= \|h^{-1}\d^k h v\|\ge \|h\|^{-1} |\d_1|^k \la h v,e_1\ra
\ge \|h\|^{-1} |\d_1|^k \b \|h v\| \;, $$ and thus $$
\frac{\|\g^k v\|}{\| v\|}\ge \|h^{-1}\|^{-1}~\|h\|^{-1}\b |\d_1|^k \;. $$ In the same way, $$
\frac{\|\g^{-k} v\|}{\| v\|}\ge \|h^{-1}\|^{-1}~\|h\|^{-1}\b |\d_d|^{-k} \;, $$ so that $$
\log\frac{\|\g^k v\|}{\| v\|} + \log\frac{\|\g^{-k} v\|}{\| v\|} \to \infty $$ as $k\to\infty$ uniformly on $\ov{v}\in U$, which in view of \eqref{eq:l1} in combination with \claimref{claim:g^k} and \claimref{claim:opens} finishes the argument. \end{proof}
\begin{rem} The measure $\mu$ in our construction can clearly be chosen symmetric and, if the group $\G$ is finitely generated, finitely supported. Obviously, the measures $\mu^k$ then also have these properties, so that \emph{singular harmonic measures can be produced by symmetric finitely supported measures on the group}. \end{rem}
\end{document}
|
arXiv
|
{
"id": "0807.1015.tex",
"language_detection_score": 0.7635774612426758,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{${}$ \vskip -1.2cm Half-arc-transitive graphs of arbitrary even valency greater than 2}
\author{
Marston D.E. Conder \\[+1pt] {\normalsize Department of Mathematics, University of Auckland,}\\[-4pt] {\normalsize Private Bag 92019, Auckland 1142, New Zealand} \\[-3pt] {\normalsize [email protected]}\\[+4pt] \and
Arjana \v{Z}itnik \\[+1pt] {\normalsize Faculty of Mathematics and Physics, University of Ljubljana,} \\[-4pt] {\normalsize Jadranska 19, 1000 Ljubljana, Slovenia} \\[-3pt] {\normalsize [email protected]} }
\date{}
\maketitle
\begin{abstract} A {\em half-arc-transitive\/} graph is a regular graph that is both vertex- and edge-transitive, but is not arc-transitive. If such a graph has finite valency, then its valency is even, and greater than $2$. In 1970, Bouwer proved that there exists a half-arc-transitive graph of every even valency greater than 2, by giving a construction for a family of graphs now known as $B(k,m,n)$, defined for every triple $(k,m,n)$ of integers greater than $1$ with $2^m \equiv 1 \mod n$. In each case, $B(k,m,n)$ is a $2k$-valent vertex- and edge-transitive graph of order $mn^{k-1}$, and Bouwer showed that $B(k,6,9)$ is half-arc-transitive for all $k > 1$.
For almost 45 years the question of exactly which of Bouwer's graphs are half-arc-transitive and which are arc-transitive has remained open, despite many attempts to answer it. In this paper, we use a cycle-counting argument to prove that almost all of the graphs constructed by Bouwer are half-arc-transitive. In fact, we prove that $B(k,m,n)$ is arc-transitive only when $n = 3$, or $(k,n) = (2,5)$, or $(k,m,n) = (2,3,7)$ or $(2,6,7)$ or $(2,6,21)$. In particular, $B(k,m,n)$ is half-arc-transitive whenever $m > 6$ and $n > 5$. This gives an easy way to prove that there are infinitely many half-arc-transitive graphs of each even valency $2k > 2$. \\
\noindent Keywords: graph, half-arc transitive, edge-transitive, vertex-transitive, arc-transitive,
automorphisms, cycles
\noindent Mathematics Subject Classification (2010): 05E18, 20B25.
\end{abstract}
\section{Introduction} \label{intro}
In the 1960s, W.T.~Tutte \cite{Tutte} proved that if a connected regular graph of odd valency is both vertex-transitive and edge-transitive, then it is also arc-transitive. At the same time, Tutte observed that it was not known whether the same was true for even valency. Shortly afterwards, I.Z.~Bouwer \cite{Bouwer} constructed a family of vertex- and edge-transitive graphs of any given even valency $2k > 2$, that are not arc-transitive.
Any graph that is vertex- and edge-transitive but not arc-transitive is now known as a {\em half-arc-transitive\/} graph. Every such graph has even valency, and since connected graphs of valency $2$ are cycles, which are arc-transitive, the valency must be at least $4$.
Quite a lot is now known about half-arc-transitive graphs, especially in the $4$-valent case --- see \cite{ConderMarusic,ConderPotocnikSparl,Marusic1998b,DM05} for example. Also a lot of attention has been paid recently to half-arc-transitive group actions on edge-transitive graphs --- see \cite {HujdurovicKutnarMarusic} for example.
In contrast, however, relatively little is known about half-arc-transitive graphs of higher valency. Bouwer's construction produced a vertex- and edge-transitive graph $B(k,m,n)$ of order $mn^{k-1}$ and valency $2k$ for every triple $(k,m,n)$ of integers greater than $1$ such that $2^{m} \equiv 1$ mod $n$, and Bouwer proved in \cite{Bouwer} that $B(k,6,9)$ is half-arc-transitive for every $k > 1$. Bouwer also showed that the latter is not true for every triple $(k,m,n)$; for example, $B(2,3,7)$, $B(2,6,7)$ and $B(2,4,5)$ are arc-transitive.
For the last $45$ years, the question of exactly which of Bouwer's graphs are half-arc-transitive and which are arc-transitive has remained open, despite a number of attempts to answer it. Three decades after Bouwer's paper, C.H.~Li and H.-S.~Sim \cite{LiSim} developed a quite different construction for a family of half-arc-transitive graphs, using Cayley graphs for metacyclic $p$-groups, and in doing this, they proved the existence of infinitely many half-arc-transitive graphs of each even valency $2k > 2$. Their approach, however, required a considerable amount of group-theoretic analysis.
In this paper, we use a cycle-counting argument to prove that almost all of the graphs constructed by Bouwer in \cite{Bouwer} are half-arc-transitive, and thereby give an easier proof of the fact that there exist infinitely many half-arc-transitive graphs of each even valency $2k > 2$.
Specifically, we prove the following:
\begin{theorem} \label{thm:main} The graph $B(k,m,n)$ is arc-transitive only if and only if $n = 3$, or $(k,n) = (2,5)$,
or $(k,m,n) = (2,3,7)$ or $(2,6,7)$ or $(2,6,21)$. In particular, $B(k,m,n)$ is half-arc-transitive whenever $m > 6$ and $n > 5$. \end{theorem}
By considering the $6$-cycles containing a given $2$-arc, we prove in Section~\ref{sec:main} that $B(k,m,n)$ is half-arc-transitive whenever $m > 6$ and $n > 7$, and then we adapt this for the other half-arc-transitive cases in Section~\ref{sec:other}. In between, we prove arc-transitivity in the cases given in the above theorem in Section~\ref{sec:AT}. But first we give some additional background about the Bouwer graphs in the following section.
\section{Further background} \label{further}
First we give the definition of the Bouwer graph $B(k,m,n)$, for every triple $(k,m,n)$ of integers greater than $1$ such that $2^{m} \equiv 1$ mod $n$.
The vertices of $B(k,m,n)$ are the $k$-tuples $(a,b_2,b_3,\dots,b_k)$ with $a \in \mathbb Z_m$ and $b_j \in \mathbb Z_n$ for $2 \le j \le k$. We will sometimes write a given vertex as $(a,{\bf b})$, where ${\bf b} = (b_2,b_3,\dots,b_k)$. Any two such vertices are adjacent if they can be written as
$(a,{\bf b})$ and $(a+1,{\bf c})$ where either ${\bf c} = {\bf b}$, or ${\bf c} = (c_2,c_3,\dots,c_k)$ differs from ${\bf b} = (b_2,b_3,\dots,b_k)$ in exactly one position, say the $(j\!-\!1)$st position, where $c_j = b_j +2^a$.
Note that the condition $2^{m} \equiv 1$ mod $n$ ensures that $2$ is a unit mod $n$, and hence that $n$ is odd. Also note that the graph is simple.
In what follows, we let ${\bf e}_j$ be the element of $(\mathbb Z_n)^{k-1}$ with $j\hskip 1pt$th term equal to $1$ and all other terms equal to $0$, for $1 \le j < k$. With this notation, we see that the neighbours of a vertex $(a,{\bf b})$ are precisely the vertices $(a+1,{\bf b})$, $(a-1,{\bf b})$, $(a+1,{\bf b}+2^{a}{\bf e}_j)$ and $(a-1,{\bf b}-2^{a-1}{\bf e}_j)$ for $1 \le j < k$, and in particular, this shows that the graph $B(k,m,n)$ is regular of valency $2k$.
Next, we recall that for all $(k,m,n)$, the graph $B(k,m,n)$ is
both vertex- and edge-transitive; see \cite[Proposition~1]{Bouwer}. Also $B(k,m,n)$ is bipartite if and only if $m$ is even.
Moreover, it is easy to see that $B(k,m,n)$ has the following three automorphisms:
\noindent (i) $\theta$, of order $k-1$, taking each vertex $(a,{\bf b}) = (a,b_2,b_3,\dots,b_{k-1},b_k)$ to the vertex $(a,{\bf b}') = (a,b_3,b_4,\dots,b_k,b_2)$, obtained by shifting its last $k-1$ entries,
\noindent (ii) $\tau$, of order $m$, taking each vertex $(a,{\bf b}) = (a,b_2,b_3,\dots,b_{k-1},b_k)$ to the vertex $(a,{\bf b}'') = (a+1,2b_2,2b_3,\dots,2b_{k-1},2b_k)$, obtained by increasing its first entry $a$ by $1$ and multiplying the others by $2$, and
\noindent (ii) $\psi$, of order $2$, taking each vertex $(a,{\bf b}) = (a,b_2,b_3,\dots,b_{k-1},b_k)$ to the vertex $(a,{\bf b}'') = (a,2^{a}-1-(b_2+b_3+\dots+b_k),b_3,\dots,b_{k-1},b_k)$, obtained by replacing its second entry $b_2$ by $2^{a}-1-(b_2+b_3+\dots+b_k)$.
\noindent (Note: in the notation of \cite{Bouwer}, the automorphism $\psi$ is $T_2 \circ S_2$.)
The automorphisms $\theta$ and $\psi$ both fix the `zero' vertex $(0,{\bf 0}) = (0,0,\dots,0)$, and $\theta$ induces a permutation of its $2k$ neighbours that fixes each of the two vertices $(1,{\bf 0}) = (1,0,0,\dots,0)$ and $(-1,{\bf 0}) = (-1,0,0,\dots,0)$ and induces two $(k\!-\!1)$-cycles on the others, while $\psi$ swaps $(1,{\bf 0})$ with $(1,{\bf e}_1) = (1,1,0,\dots,0)$, and swaps $(-1,{\bf 0})$ with $(-1,-2^{-1}{\bf e}_1) = (-1,-2^{-1},0,\dots,0)$, and fixes all the others.
It follows that the subgroup generated by $\theta$ and $\psi$ fixes the vertex $(0,{\bf 0})$ and has two orbits of length $k$ on its neighbours, with one orbit consisting of the vertices of the form $(1,{\bf b})$ where ${\bf b} = {\bf 0}$ or ${\bf e}_j$ for some $j$, and the other consisting of those of the form $(-1,{\bf c})$ where ${\bf c} = {\bf 0}$ or $-2^{-1}{\bf e}_j$ for some $j$.
By edge-transitivity, the graph $B(k,m,n)$ is arc-transitive if and only if it admits an automorphism that interchanges the `zero' vertex $(0,{\bf 0})$ with one of its neighbours, in which case the above two orbits of $\langle \theta, \psi \rangle$ on the neighbours of $(0,{\bf 0})$ are merged into one under the full automorphism group. We will use the automorphism $\tau$ in the next section.
We will also use the following, which is valid in all cases, not just those with $m > 6$ and $n > 7$ considered in the next section.
\begin{lemma} \label{threearcs} Every $3$-arc $\,v_0 \!\sim\! v_1 \!\sim\! v_2 \!\sim\! v_3$ in $X$ with first vertex $v_0 = (0,{\bf 0})$ is of one of the following forms, with $r,s,t \in \{1,\dots,k\!-\!1\}$ in each case$\,:$ \\[+10pt] \begin{tabular}{rl} \ {\rm (1)} & $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,{\bf 0}) \!\sim\! (3,{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,4{\bf e}_r$, \\[+4 pt] \ {\rm (2)} & $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,{\bf 0}) \!\sim\! (1,-2{\bf e}_r)$, \\[+4 pt] \ {\rm (3)} & $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,2{\bf e}_r) \!\sim\! (3,2{\bf e}_r\!+\!{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,4{\bf e}_s$, \\[+4 pt] \ {\rm (4)} & $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,2{\bf e}_r) \!\sim\! (1,2{\bf e}_r\!-\!{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,2{\bf e}_s$ with $s \ne r$, \\[+4 pt] \ {\rm (5)} & $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (0,-{\bf e}_r) \!\sim\! (1,-{\bf e}_r\!+\!{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,{\bf e}_s$ with $s \ne r$, \\[+4 pt] \ {\rm (6)} & $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (0,-{\bf e}_r) \!\sim\! (-1,-{\bf e}_r\!-\!{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,2^{-1}{\bf e}_s$, \\[+4 pt] \ {\rm (7)} & $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (2,{\bf e}_r) \!\sim\! (3,{\bf e}_r\!+\!{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,4{\bf e}_s$, \\[+4 pt] \ {\rm (8)} & $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (2,{\bf e}_r) \!\sim\! (1,{\bf e}_r\!-\!2{\bf e}_s)$, \\[+4 pt] \end{tabular} \begin{tabular}{rl} \ {\rm (9)} & $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (2,{\bf e}_r\!+\!2{\bf e}_s) \!\sim\! (3,{\bf e}_r\!+\!2{\bf e}_s\!+\!{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,4{\bf e}_t$, \\[+4 pt] {\rm (10)} & $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (2,{\bf e}_r\!+\!2{\bf e}_s) \!\sim\! (1,{\bf e}_r\!+\!2{\bf e}_s\!-\!{\bf d})$,
\\ & \quad where $\,{\bf d} = {\bf 0}$ or $\,2{\bf e}_t$ with $t \ne s$, \\[+4 pt] {\rm (11)} & $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r) \!\sim\! (1,{\bf e}_r+{\bf e}_s)$, \\[+4 pt] {\rm (12)} & $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r) \!\sim\! (-1,{\bf e}_r\!-\!{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,2^{-1}{\bf e}_s$, \\[+4 pt] {\rm (13)} & $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r-{\bf e}_s) \!\sim\! (1,{\bf e}_r-{\bf e}_s\!+\!{\bf d})$,
where $s \ne r$, \\ & \quad and $\,{\bf d} = {\bf 0}$ or $\,{\bf e}_t$ with $t \ne s$, \\[+4 pt] {\rm (14)} & $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r-{\bf e}_s) \!\sim\! (-1,{\bf e}_r-{\bf e}_s\!-\!{\bf d})$,
where $s \ne r$, \\ & \quad and $\,{\bf d} = {\bf 0}$ or $\,2^{-1}{\bf e}_t$, \\[+4 pt] {\rm (15)} & $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (0,2^{-1}{\bf e}_r) \!\sim\! (1,2^{-1}{\bf e}_r\!+\!{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,{\bf e}_s$, \\[+4 pt] {\rm (16)} & $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (0,2^{-1}{\bf e}_r) \!\sim\! (-1,2^{-1}{\bf e}_r\!-\!{\bf d})$,
\\ & \quad where $\,{\bf d} = {\bf 0}$ or $\,2^{-1}{\bf e}_s$ with $s \ne r$, \\[+4 pt] {\rm (17)} & $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (-2,{\bf 0}) \!\sim\! (-1,2^{-2}{\bf e}_r)$, \\[+4 pt] {\rm (18)} & $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (-2,{\bf 0}) \!\sim\! (-3,-{\bf d})$,
where $\,{\bf d} = {\bf 0}$ or $\,2^{-3}{\bf e}_r$, \\[+4 pt] {\rm (19)} & $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (-2,-2^{-2}{\bf e}_r) \!\sim\! (-1,-2^{-2}{\bf e}_r\!+\!{\bf d})$,
\\ & \quad where $\,{\bf d} = {\bf 0}$ or $\,2^{-2}{\bf e}_s$ with $s \ne r$, \\[+4 pt] {\rm (20)} & $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (-2,-2^{-2}{\bf e}_r) \!\sim\! (-3,-2^{-2}{\bf e}_r\!-\!{\bf d})$,
\\ & \quad where $\,{\bf d} = {\bf 0}$ or $\,2^{-3}{\bf e}_s$, \\[+4 pt] {\rm (21)} & $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r) \!\sim\! (1,-2^{-1}{\bf e}_r\!+\!{\bf d})$,
\\ & \quad where $\,{\bf d} = {\bf 0}$ or $\,{\bf e}_s$, \\[+4 pt] {\rm (22)} & $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r) \!\sim\! (-1,-2^{-1}{\bf e}_r\!-\!2^{-1}{\bf e}_s)$, \\[+4 pt] {\rm (23)} & $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r\!+\!2^{-1}{\bf e}_s) \!\sim\! (1,-2^{-1}{\bf e}_r\!+\!2^{-1}{\bf e}_s\!+\!{\bf d})$,
\\ & \quad where $s \ne r$, and $\,{\bf d} = {\bf 0}$ or $\,{\bf e}_t$, \\[+4 pt] {\rm (24)} & $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r\!+\!2^{-1}{\bf e}_s) \!\sim\! (-1,-2^{-1}{\bf e}_r\!+\!2^{-1}{\bf e}_s\!-\!{\bf d})$,
\\ & \quad where $s \ne r$, and $\,{\bf d} = {\bf 0}$ or $\,2^{-1}{\bf e}_t$ with $t \ne s$, \\[+4 pt] {\rm (25)} & $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (-2,-2^{-1}{\bf e}_r) \!\sim\! (-1,-2^{-1}{\bf e}_r\!+\!2^{-2}{\bf e}_s)$, \\[+4 pt] {\rm (26)} & $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (-2,-2^{-1}{\bf e}_r) \!\sim\! (-3,-2^{-1}{\bf e}_r\!-\!{\bf d})$,
\\ & \quad where $\,{\bf d} = {\bf 0}$ or $\,2^{-3}{\bf e}_s$, \\[+4 pt] {\rm (27)} & $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (-2,-2^{-1}{\bf e}_r\!-\!2^{-2}{\bf e}_s) \!\sim\! (-1,-2^{-1}{\bf e}_r\!-\!2^{-2}{\bf e}_s\!+\!{\bf d})$,
\\ & \quad where $\,{\bf d} = {\bf 0}$ or $\,2^{-2}{\bf e}_t$ with $t \ne s$, \\[+4 pt] {\rm (28)} & $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (-2,-2^{-1}{\bf e}_r\!-\!2^{-2}{\bf e}_s) \!\sim\! (-3,-2^{-1}{\bf e}_r\!-\!2^{-2}{\bf e}_s\!-\!{\bf d})$,
\\ & \quad where $\,{\bf d} = {\bf 0}$ or $\,2^{-3}{\bf e}_t$. \\
\end{tabular} \end{lemma}
\begin{proof} This follows directly from the definition of $X = B(k,m,n)$. \end{proof}
Note that the 28 cases given in Lemma~\ref{threearcs} fall naturally into $14$ pairs, with each pair determined by the form of the initial $2$-arc $\,v_0 \!\sim\! v_1 \!\sim\! v_2$.
Also is it easy to see that the number of $3$-arcs in each case is \\[-4pt] $$\begin{cases} \ k & \hbox{in cases 1 and 18,} \\[-1 pt] \ k\!-\!1 & \hbox{in cases 2 and 17,} \\[-1 pt] \ k(k\!-\!1) & \hbox{in cases 3, 6, 7, 12, 15, 20, 21 and 26,} \\[-1 pt] \ (k\!-\!1)^2 & \hbox{in cases 4, 5, 8, 11, 16, 19, 22 and 25,} \\[-1 pt] \ k(k\!-\!1)^2 & \hbox{in cases 9 and 28,} \\[-1 pt] \ (k\!-\!1)^3 & \hbox{in cases 10 and 27,} \\[-1 pt] \ (k\!-\!1)^{2}(k\!-\!2) & \hbox{in cases 13 and 24,} \\[-1 pt] \ k(k\!-\!1)(k\!-\!2) & \hbox{in cases 14 and 23,} \end{cases} $$ and the total of all these numbers is $2k(2k\!-\!1)^2$, as expected.
\section{The main approach} \label{sec:main}
Let $k$ be any integer greater than $1$, and suppose that $m > 6$ and $n > 7$.
We will prove that in every such case, the graph $X = B(k,m,n)$ is not arc-transitive, and is therefore half-arc-transitive.
We do this simply by considering the ways in which a given $2$-arc or $3$-arc lies in a cycle of length $6$ in $X$. (For any positive integer $s$, an $s$-arc in a simple graph is a walk $\,v_0 \!\sim\! v_1 \!\sim\! v_2 \!\sim\! \dots \!\sim\! v_s$ of length $s$ in which any three consecutive vertices are distinct.) By vertex-transitivity, we can consider what happens locally around the vertex $(0,{\bf 0})$.
\begin{lemma} \label{girth} The girth of $X$ is $6$. \end{lemma}
\begin{proof} First, $X = B(k,m,n)$ is simple, by definition. Also there are no cycles of length $3$, $4$ or $5$ in $X$, since in the list of cases for a $3$-arc $\,v_0 \!\sim\! v_1 \!\sim\! v_2 \!\sim\! v_3$ in $X$ with first vertex $v_0 = (0,{\bf 0})$ given by Lemma~\ref{threearcs}, the vertex $v_3$ is never equal to $v_0$, the vertex $v_1$ is uniquely determined by $v_2$, and every possibility for $v_2$ is different from every possibility for $v_3$.
On the other hand, there are certainly some cycles of length $6$ in $X$, such as $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,2{\bf e}_k) \!\sim\! (1,
2{\bf e}_k) \!\sim\! (0,{\bf e}_k) \!\sim\! (1,{\bf e}_k) \!\sim\! (0,{\bf 0})$. \end{proof}
Next, we can find all $6$-cycles based at the vertex $v_0 = (0,{\bf 0})$ in $X$.
\begin{lemma} \label{sixcycles} Up to reversal, every $6$-cycle based at the vertex $v_0 = (0,{\bf 0})$ has exactly one of the forms below, with $r$, $s,$ $t$ all different when they appear$\,:$ \\[+5pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,2{\bf e}_r) \!\sim\! (1,2{\bf e}_r) \!\sim\! (0,{\bf e}_r) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (0,-{\bf e}_r) \!\sim\! (1,-{\bf e}_r) \!\sim\! (2,{\bf e}_r) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (0,-{\bf e}_r) \!\sim\! (1,{\bf e}_s\!-\!{\bf e}_r) \!\sim\! (0,{\bf e}_s\!-\!{\bf e}_r) \!\sim\! (1,{\bf e}_s) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (0,-{\bf e}_r) \!\sim\! (-1,-{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (2,2{\bf e}_s\!+\!{\bf e}_r) \!\sim\! (1,2{\bf e}_s\!-\!{\bf e}_r) \!\sim\! (0,{\bf e}_s\!-\!{\bf e}_r) \!\sim\! (1,{\bf e}_s) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r) \!\sim\! (1,{\bf e}_s\!+\!{\bf e}_r) \!\sim\! (0,{\bf e}_s) \!\sim\! (1,{\bf e}_s) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r) \!\sim\! (-1,2^{-1}{\bf e}_r) \!\sim\! (0,2^{-1}{\bf e}_r) \!\sim\! (-1,{\bf 0}) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r\!-\!{\bf e}_s) \!\sim\! (1,{\bf e}_r\!-\!{\bf e}_s\!+\!{\bf e}_t) \!\sim\! (0,{\bf e}_t\!-\!{\bf e}_s) \!\sim\! (1,{\bf e}_t) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r\!-\!{\bf e}_s) \!\sim\! (-1,2^{-1}{\bf e}_r\!-\!{\bf e}_s) \!\sim\! (0,2^{-1}{\bf e}_r\!-\!2^{-1}{\bf e}_s)$ \\ ${}$ \qquad $\!\sim\! (-1,-\!2^{-1}{\bf e}_s) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (0,2^{-1}{\bf e}_r) \!\sim\! (1,2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r) \!\sim\! (-1,-2^{-1}{\bf e}_r) $ \\ ${}$ \qquad $\!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (0,2^{-1}{\bf e}_r) \!\sim\! (-1,2^{-1}{\bf e}_r\!-\!2^{-1}{\bf e}_s) \!\sim\! (0,2^{-1}{\bf e}_r\!-\!2^{-1}{\bf e}_s)$ \\ ${}$ \qquad $\!\sim\! (-1,-\!2^{-1}{\bf e}_s) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (-2,-2^{-2}{\bf e}_r) \!\sim\! (-1,-2^{-2}{\bf e}_r) \!\sim\! (-2,-2^{-1}{\bf e}_r) \!\sim\! (-1,-2^{-1}{\bf e}_r) $ \\ ${}$ \qquad $\!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r) \!\sim\! (-1,-2^{-1}{\bf e}_r\!-\!2^{-1}{\bf e}_s) \!\sim\! (0,-2^{-1}{\bf e}_s)$ \\ ${}$ \qquad $\!\sim\! (-1,-2^{-1}{\bf e}_s) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r\!+\!2^{-1}{\bf e}_s) \!\sim\! (1,2^{-1}{\bf e}_r\!+\!2^{-1}{\bf e}_s)$ \\ ${}$ \qquad $\!\sim\! (0,2^{-1}{\bf e}_r\!-\!2^{-1}{\bf e}_s) \!\sim\! (-1,-2^{-1}{\bf e}_s) \!\sim\! (0,{\bf 0})$, \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r\!+\!2^{-1}{\bf e}_s) \!\sim\! (-1,-2^{-1}{\bf e}_r\!+\!2^{-1}{\bf e}_s\!-\!2^{-1}{\bf e}_t)$ \\ ${}$ \qquad $\!\sim\! (0,2^{-1}{\bf e}_s\!-\!2^{-1}{\bf e}_t) \!\sim\! (-1,-2^{-1}{\bf e}_t) \!\sim\! (0,{\bf 0})$, or \\[+3 pt] $\bullet$ \ $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (-2,-2^{-1}{\bf e}_r\!-\!2^{-2}{\bf e}_s) \!\sim\! (-1,-2^{-2}{\bf e}_r\!-\!2^{-2}{\bf e}_s)$ \\ ${}$ \qquad $\!\sim\! (-2,-2^{-2}{\bf e}_r\!-\!2^{-1}{\bf e}_s) \!\sim\! (-1,-2^{-1}{\bf e}_s) \!\sim\! (0,{\bf 0})$.
\end{lemma}
\begin{proof} This is left as an exercise for the reader. It may be helpful to note that a $6$-cycle of the first form is obtainable as a $3$-arc of type 4 with final vertex $2{\bf e}_r$ followed by the reverse of a $3$-arc of type 11 with the same final vertex. The $6$-cycles of the other $15$ forms are similarly obtainable as the concatenation of a $3$-arc of type $i$ with the reverse of a $3$-arc of type $j$, for $(i,j) = (5,8)$, $(5,13)$, $(6,22)$, $(10,13)$, $(11,11)$, $(12,16)$, $(13,13)$, $(14,24)$, $(15,21)$, $(16,24)$, $(19,25)$, $(22,22)$, $(23,23)$, $(24,24)$ and $(27,27)$, respectively. Uniqueness follows from the assumptions about $m$ and $n$. \end{proof}
\begin{corollary} \label{count6cycles} The number of different $6$-cycles in $X$ that contain a given $2$-arc $\,v_0 \!\sim\! v_1 \!\sim\! v_2$ with first vertex $v_0 = (0,{\bf 0})$ is always $0$, $1$ or $k$. More precisely, this number is \\[-12pt] \begin{center} \begin{tabular}{cl} ${}$\hskip -18pt $0$
& \hskip -8pt
for the $2$-arcs $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,{\bf 0})$
and $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (-2,{\bf 0})$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (2,3{\bf e}_r)$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (-2,-(2^{-1}\!+\!2^{-2}){\bf e}_r)$, \\[+6pt] ${}$\hskip -18pt $1$
& \hskip -8pt
for the $2$-arcs of the form $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,2{\bf e}_r)$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (2,{\bf e}_r)$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (2,{\bf e}_r\!+\!2{\bf e}_s)$, when $k > 2$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (-2,-2^{-2}{\bf e}_r)$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (-2,-2^{-1}{\bf e}_r)$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (-2,-2^{-1}{\bf e}_r\!-\!2^{-2}{\bf e}_s)$,
\\ & \quad when $k > 2$, \ and \end{tabular} \begin{tabular}{cl} ${}$\hskip -18pt $k$
& \hskip -8pt
for the $2$-arcs of the form $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (0,-{\bf e}_r)$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r)$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r\!-\!{\bf e}_s)$, when $k > 2$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (-1,{\bf 0}) \!\sim\! (0,2^{-1}{\bf e}_r)$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r)$, \\
& and those of the form $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (0,-2^{-1}{\bf e}_r\!+\!2^{-1}{\bf e}_s)$,
\\ & \quad when $k > 2$. \\[-3pt] \end{tabular} \end{center} In particular, every $2$-arc of the form $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,2{\bf e}_r)$ lies in just one $6$-cycle, namely $(0,{\bf 0}) \!\sim\! (1,{\bf 0}) \!\sim\! (2,2{\bf e}_r) \!\sim\! (1,2{\bf e}_r) \!\sim\! (0,{\bf e}_r) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf 0})$, while every $2$-arc of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r)$ lies in $k$ $6$-cycles. \end{corollary}
\begin{proof} This follows easily from inspection of the list of $6$-cycles given in Lemma~\ref{sixcycles}, and their reverses. \end{proof}
At this stage, we could repeat the above calculations for $2$-arcs and $3$-arcs with first vertex $(1,{\bf 0}) = (1,0,0,\dots,0)$, but it is much easier to simply apply the automorphism $\tau$ defined in Section~\ref{further}.
Hence in particular, the $2$-arcs $\,v_0 \!\sim\! v_1 \!\sim\! v_2$ with first vertex $v_0 = (1,{\bf 0})$ that lie in a unique $6$-cycle are those of the form $(1,{\bf 0}) \!\sim\! (2,{\bf 0}) \!\sim\! (3,4{\bf e}_r)$, or $(1,{\bf 0}) \!\sim\! (2,2{\bf e}_r) \!\sim\! (3,2{\bf e}_r)$, or $(1,{\bf 0}) \!\sim\! (2,2{\bf e}_r) \!\sim\! (3,2{\bf e}_r\!+\!4{\bf e}_s)$ when $k > 2$, or $(1,{\bf 0}) \!\sim\! (0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r)$, or $(1,{\bf 0}) \!\sim\! (0,-{\bf e}_r) \!\sim\! (-1,-{\bf e}_r)$, or $(1,{\bf 0}) \!\sim\! (0,-{\bf e}_r) \!\sim\! (-1,-{\bf e}_r\!-\!2^{-1}{\bf e}_s)$ when $k > 2$.
Let $v$ and $w$ be the neighbouring vertices $(0,{\bf 0})$ and $(1,{\bf 0})$. We will show that there is no automorphism of $X$ that reverses the arc $(v,w)$.
Let $A = \{(2,2{\bf e}_r) : 1 \le r < k\}$, which is the set of all vertices $x$ in $X$ that extend the arc $(v,w)$ to a $2$-arc $(v,w,x)$ which lies in a unique $6$-cycle, and similarly, let $B = \{(0,-{\bf e}_r) : 1 \le r < k\}$, the set of all vertices that extend $(v,w)$ to a $2$-arc which lies in $k$ different $6$-cycles. Also let $C = \{(-1,-2^{-1}{\bf e}_r) : 1 \le r < k\}$ and $D = \{(1,{\bf e}_r) : 1 \le r < k\}$ be the analogous sets of vertices extending the arc $(w,v)$, as illustrated in Figure~\ref{ABCD}.
\begin{figure}
\caption{$6$-cycles containing the edge from $v = (0,{\bf 0})$ to $w = (1,{\bf 0})$}
\label{ABCD}
\end{figure}
Now suppose there exists an automorphism $\xi$ of $X$ that interchanges the two vertices of the edge $\{v,w\}$. Then by considering the numbers of $6$-cycles that contain a given $2$-arc, we see that $\xi$ must interchange the sets $A$ and $C$, and interchange the sets $B$ and $D$.
Next, observe that if $x = (2,2{\bf e}_r) \in A$, then the unique $6$-cycle containing the $2$-arc $v \!\sim\! w \!\sim\! x$ is $v \!\sim\! w \!\sim\! x \!\sim\! y \!\sim\! z \!\sim\! u \!\sim\! v$, where $(y,z,u) = ((1,2{\bf e}_r), (0,{\bf e}_r),(1,{\bf e}_r))$; in particular, the $6$th vertex $u = (1,{\bf e}_r)$ lies in $D$. Similarly, if $x' = (-1,-2^{-1}{\bf e}_r) \in C$, then the unique $6$-cycle containing the $2$-arc $w \!\sim\! v \!\sim\! x'$ is $w \!\sim\! v \!\sim\! x' \!\sim\! y' \!\sim\! z' \!\sim\! u' \!\sim\! w$, where $(y',z',u') = ((0,-2^{-1}{\bf e}_r), (-1,-{\bf e}_r), (0,-{\bf e}_r))$, and the $6$th vertex $u' = (0,-{\bf e}_r)$ lies in $B$.
The arc-reversing automorphism $\xi$ must take every $6$-cycle
of the first kind to a $6$-cycle
of the second kind, and hence must take each $2$-arc of the form $v \!\sim\! u \!\sim\! z$ ($= (0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r)$) to a $2$-arc of the form $w \!\sim\! u' \!\sim\! z'$ ($= (1,{\bf 0}) \!\sim\! (0,-{\bf e}_r) \!\sim\! (-1,-{\bf e}_r)$).
By Corollary~\ref{count6cycles}, however, each $2$-arc of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_r) \!\sim\! (0,{\bf e}_r)$ lies in $k$ different $6$-cycles, while each $2$-arc of the form $(1,{\bf 0}) \!\sim\! (0,-{\bf e}_r) \!\sim\! (-1,-{\bf e}_r)$ is the $\tau$-image of the $2$-arc $(0,{\bf 0}) \!\sim\! (-1,-2^{-1}{\bf e}_r) \!\sim\! (-2,-2^{-1}{\bf e}_r)$ and hence lies in only one $6$-cycle.
This is a contradiction, and shows that no such arc-reversing automorphism exists.
Hence the Bouwer graph $B(k,m,n)$ is half-arc-transitive whenever $m > 6$ and $n > 7$.
If $m \le 6$, then the order $m_2$ of $2$ as a unit mod $n$ is at most $6$, and so $n$ divides $2^{m_2} - 1 = 3, 7, 15, 31$ or $63$. In particular, if $m _2 = 2$, $3$, $4$ or $5$ then $n \in \{3\}$, $\{7\}$, $\{5,15\}$ or $\{31\}$ respectively, while if $m_2 = 6$, then $n$ divides $63$ but not $3$ or $7$, so $n \in \{9, 21, 63\}$.
Conversely, if $n = 3, 5, 7, 9,$ $ 15, 21, 31$ or $63$, then $m_2 = 2$, $4$, $3$, $6$, $4$, $6$, $5$ or $6$, respectively, and of course in each case, $m$ is a multiple of $m_2$.
We deal with these exceptional cases in the next two sections.
\section{Arc-transitive cases} \label{sec:AT}
The following observations are very easy to verify.
When $n = 3$ (and $m$ is even), the Bouwer graph $B(k,m,n)$ is always arc-transitive, for in that case there is an automorphism that takes each vertex $(a,{\bf b})$ to $(1\!-\!a,-{\bf b})$, and this
reverses the arc from $(0,{\bf 0})$ to $(1,{\bf 0})$.
Similarly, when $k = 2$ and $n = 5$ (and $m$ is divisible by $4$), the graph $B(k,m,n)$ is always arc-transitive, since it has an automorphism that takes $(a,b_2)$ to $(1\!-\!a,-b_2)$ for all $a \equiv 0$ or $1$ mod $4$, and to $(1\!-\!a,2-b_2)$ for all $a \equiv 2$ or $3$ mod $4$, and again this interchanges $(0,{\bf 0})$ with $(1,{\bf 0})$.
Next, $B(2,m,7)$ is arc-transitive when $m = 3$ or $6$, because in each of those two cases there is an automorphism that takes \\[-16pt] \begin{center} \begin{tabular}{lll} $\quad(a,0)$ to $(1\!-\!a,0)$, \quad \ & $(a,1)$ to $(a\!+\!1,2)$, \quad \ & $(a,2)$ to $(a\!-\!1,1)$, \\[+1 pt] $\quad (a,3)$ to $(\!-\!a,6)$, \quad \ & $(a,4)$ to $(a\!+\!3,4)$, \quad \ & $(a,5)$ to $(1\!-\!a,5)$, \\[+1 pt] $\quad (a,6)$ to $(1\!-\!a,3)$, & & \\[-6pt] \end{tabular} \end{center} for every $a \in \mathbb Z_m$.
Similarly, $B(2,6,21)$ is arc-transitive since it has an automorphism taking \\[-16pt] \begin{center} \begin{tabular}{lll} $\quad (a,0)$ to $(1\!-\!a,0)$, \quad \ & $(a,1)$ to $(a\!+\!1,2)$, \quad \ & $(a,2)$ to $(a\!-\!1,1)$, \\[+1 pt] $\quad (a,3)$ to $(5\!-\!a,6)$, \quad \ & $(a,4)$ to $(a\!+\!3,11)$, \quad \ & $(a,5)$ to $(5\!-\!a,19)$, \\[+1 pt] $\quad (a,6)$ to $(5\!-\!a,3)$, \quad & $(a,7)$ to $(1\!-\!a,14)$, \quad \ & $(a,8)$ to $(a\!+\!1,16)$, \\[+1 pt] $\quad (a,9)$ to $(a\!-\!1,15)$, \quad & $(a,10)$ to $(5\!-\!a,20)$, \quad \ & $(a,11)$ to $(a\!+\!3,4)$, \\[+1 pt] $\quad (a,12)$ to $(5\!-\!a,12)$, \quad & $(a,13)$ to $(5\!-\!a,17)$, \quad \ & $(a,14)$ to $(1\!-\!a,7)$, \\[+1 pt] $\quad (a,15)$ to $(a\!+\!1,9)$, \quad & $(a,16)$ to $(a\!-\!1,8)$, \quad \ & $(a,17)$ to $(5\!-\!a,13)$, \\[+1 pt] $\quad (a,18)$ to $(a\!+\!3,18)$, \quad & $(a,19)$ to $(5\!-\!a,5)$, \quad \ & $(a,20)$ to $(5\!-\!a,10)$, \\[-3pt] \end{tabular} \end{center} for every $a \in \mathbb Z_m$.
\section{Other half-arc-transitive cases} \label{sec:other}
In this final section, we give sketch proofs of half-arc-transitivity in all the remaining cases.
We first consider the three cases where $m$ and the multiplicative order of $2$ mod $n$ are both equal to $6$ (and $n = 9$, $21$ or $63$), and then deal with the cases where the multiplicative order of $2$ mod $n$ is less than $6$ (and $n = 5$, $7$, $15$ or $31$). In the cases $n = 15$ and $n = 31$, we can assume that $m < 6$ as well.
Note that one of these remaining cases is the one considered by Bouwer, namely where $(m,n) = (6,9)$. As this is one of the exceptional cases, it is not representative of the generic case --- which may help explain why previous attempts to generalise Bouwer's approach did not get very far.
To reduce unnecessary repetition, we will introduce some further notation that will be used in most of these cases. As in Section~\ref{sec:main}, we let $v$ and $w$ be the vertex $(0,{\bf 0})$ and its neighbour $(1,{\bf 0})$.
Next, for all $i \ge 0$, we define $V^{(i)}$ as the set of all vertices $x$ in $X$ for which $v \!\sim\! w \!\sim\! x $ is a $2$-arc that lies in exactly $i$ different $6$-cycles, and $W^{(i)}$ as the analogous set of vertices $x'$ in $X$ for which $w \!\sim\! v \!\sim\! x'$ is a $2$-arc that lies in exactly $i$ different $6$-cycles. (Hence, for example, the sets $A$, $B$, $C$ and $D$ used in Section~\ref{sec:main} are $V^{(1)}$, $V^{(k)}$, $W^{(1)}$ and $W^{(k)}$, respectively.)
Also for $i \ge 0$ and $j \ge 0$, we define $T_{v}^{\,(i,j)}$ as the set of all $2$-arcs $v \!\sim\! u \!\sim\! z$ that come from a $6$-cycle of the form $v \!\sim\! w \!\sim\! x \!\sim\! y \!\sim\! z \!\sim\! u \!\sim\! v$ with $x \in V^{(i)}$ and $u \in W^{(i)}$, and lie in exactly $j$ different $6$-cycles altogether, and define $T_{w}^{\,(i,j)}$ as the analogous set of all $2$-arcs $w \!\sim\! u' \!\sim\! z'$ that come from a $6$-cycle of the form $w \!\sim\! v \!\sim\! x' \!\sim\! y' \!\sim\! z' \!\sim\! u' \!\sim\! w$ with $x' \in W^{(i)}$ and $u' \in V^{(i)}$, and lie in exactly $j$ different $6$-cycles altogether.
Note that if the graph under consideration is arc-transitive, then it has an automorphism $\xi$ that reverses the arc $(v,w)$, and then clearly $\xi$ must interchange the two sets $V^{(i)}$ and $W^{(i)}$, for each $i$. Hence also $\xi$ interchanges the two sets $T_{v}^{\,(i,j)}$ and $T_{w}^{\,(i,j)}$ for all $i$ and $j$,
and therefore $|T_{v}^{\,(i,j)}| = |T_{w}^{\,(i,j)}|$ for all $i$ and $j$.
Equivalently, if $|T_{v}^{\,(i,j)}| \ne |T_{w}^{\,(i,j)}|$ for some pair $(i,j)$, then the graph cannot be arc-transitive, and therefore must be half-arc-transitive.
The approach taken in Section~\ref{sec:main} was similar, but compared the $2$-arcs $v \!\sim\! u \!\sim\! z$ that come from a $6$-cycle of the form $v \!\sim\! w \!\sim\! x \!\sim\! y \!\sim\! z \!\sim\! u \!\sim\! v$ where $x \in V^{(1)}$ and $u \in W^{(k)}$, with the $2$-arcs $w \!\sim\! u' \!\sim\! z'$ that come from a $6$-cycle of the form $w \!\sim\! v \!\sim\! x' \!\sim\! y' \!\sim\! z' \!\sim\! u' \!\sim\! w$ where $x' \in W^{(1)}$ and $u' \in V^{(k)}$.
We proceed by considering the first three cases below, in which the girth of the Bouwer graph $B(k,m,n)$ is $6$, but the numbers of $6$-cycles containing a given arc or $2$-arc are different from those found in Section~\ref{sec:main}.
\subsection{The graphs $B(k,6,9)$}
Suppose $m = 6$ and $n = 9$. This case was considered by Bouwer in \cite{Bouwer}, but it can also be dealt with in a similar way to the generic case in Section~\ref{sec:main}.
Every $2$-arc lies in either $k$ or $k+1$ cycles of length $6$, and each arc lies in exactly $k$ distinct $2$-arcs of the first kind, and $k-1$ of the second kind. Next, the set $V^{(k)}$ consists of $(2,{\bf 0})$ and $(0,-{\bf e}_r)$ for $1 \le r < k$, while $W^{(k)}$ consists of $(-1,{\bf 0})$ and $(1,{\bf e}_s)$ for $1 \le s < k$.
Now consider the $2$-arcs $v \!\sim\! u \!\sim\! z$ that come from a $6$-cycle of the form $v \!\sim\! w \!\sim\! x \!\sim\! y \!\sim\! z \!\sim\! u \!\sim\! v$ with $x \in V^{(k)}$ and $u \in W^{(k)}$. There are $k^{2}-2k+2$ such $2$-arcs, and of these, $k\!-\!1$ lie in $k\!+\!1$ different $6$-cycles altogether (namely the $2$-arcs of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_s) \!\sim\! (2,{\bf e}_s)$),
while the other $k^{2}-3k+3$ lie on only $k$ different 6-cycles. In particular, $|T_{v}^{\,(k,k+1)}| = k\!-\!1$.
On the other hand, among the $2$-arcs $w \!\sim\! u' \!\sim\! z'$ coming from a $6$-cycle of the form $w \!\sim\! v \!\sim\! x' \!\sim\! y' \!\sim\! z' \!\sim\! u' \!\sim\! w$ with $x' \in W^{(k)}$ and $u' \in V^{(k)}$,
none lies in $k+1$ different 6-cycles. Hence $|T_{w}^{\,(k,k+1)}| = 0$.
Thus $|T_{v}^{\,(k,k+1)}| \ne |T_{w}^{\,(k,k+1)}|$, and so the Bouwer graph $B(k,6,9)$ cannot be arc-transitive, and is therefore half-arc-transitive.
\subsection{The graphs $B(k,6,21)$ for $k > 2$}
Next, suppose $m = 6$ and $n = 21$, and $k > 2$. (The case $k = 2$ was dealt with in Section~\ref{sec:AT}.) Here every $2$-arc lies in one, two or $k$ cycles of length $6$, and each arc lies in exactly one $2$-arc of the first kind, and $k-1$ distinct $2$-arcs of each of the second and third kinds.
The set $V^{(k)}$ consists of the $k\!-\!1$ vertices of the form $(0,-{\bf e}_r)$, while $W^{(k)}$ consists of the $k\!-\!1$ vertices of the form $(1,{\bf e}_s)$. (Note that this does not hold when $k = 2$.) Also $T_{v}^{\,(k,2)}$ consists of the $2$-arcs of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_s) \!\sim\! (2,{\bf e}_s)$,
so $|T_{v}^{\,(k,2)}| = k\!-\!1$, but on the other hand, $T_{w}^{\,(k,2)}$ is empty.
Hence there can be no automorphism that reverses the arc $(v,w)$, and so the graph is half-arc-transitive.
\subsection{The graphs $B(k,6,63)$}
Suppose $m = 6$ and $n = 63$. Then every $2$-arc lies in either one or $k$ different cycles of length $6$, and each arc lies in exactly $k$ $2$-arcs of the first kind, and $k\!-\!1$ of the second kind.
The sets $V^{(k)}$ and $W^{(k)}$ are precisely as in the previous case, but for all $k \ge 2$, and in this case $T_{v}^{\,(k,1)}$ consists of the $2$-arcs of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_s) \!\sim\! (2,{\bf e}_s)$,
so $|T_{v}^{\,(k,1)}| = k\!-\!1$, but on the other hand, $T_{w}^{\,(k,1)}$ is empty.
Hence there can be no automorphism that reverses the arc $(v,w)$, and so the graph is half-arc-transitive.
Now we turn to the cases where $n = 5$, $7$, $15$ or $31$. In these cases, the order of $2$ as a unit mod $n$ is $4$, $3$, $4$ or $5$, respectively, and indeed when $n = 15$ or $31$ we may suppose that $m = 4$ or $m = 5$, while the cases $n = 5$ and $n = 7$ are much more tricky.
\subsection{The graphs $B(k,4,15)$}
Suppose $n = 15$ and $m = 4$. Then the girth is $4$, with $2k$ different $4$-cycles passing through the vertex $(0,{\bf 0})$. Apart from this difference, the approach taken in Section~\ref{sec:main} for counting $6$-cycles still works in the same way as before.
Hence the graph $B(k,4,15)$ is half-arc-transitive for all $k \ge 2$.
\subsection{The graphs $B(k,5,31)$}
Similarly, when $n = 31$ and $m = 5$, the girth is $5$, and the same approach as taken in Section~\ref{sec:main} using $6$-cycles
works, to show that the graph $B(k,5,31)$ is half-arc-transitive for all $k \ge 2$.
\subsection{The graphs $B(k,m,5)$ for $k > 2$}
Suppose $n = 5$ and $k > 2$. (The case $k = 2$ was dealt with in Section~\ref{sec:AT}.) Here we have $m \equiv 0$ mod $4$, and the number of $6$-cycles is much larger than in the generic case considered in Section~\ref{sec:main} and the cases with $m = 6$ above, but a similar argument works.
When $m > 4$, the girth of $B(k,m,n)$ is $6$, and every $2$-arc lies in either $2k$, $2k+3$ or $4k-4$ cycles of length 6. Also $V^{(2k+3)}$ consists of the $k\!-\!1$ vertices of the form $(0,-{\bf e}_r)$, while $W^{(2k+3)}$ consists of the $k\!-\!1$ vertices of the form $(1,{\bf e}_s)$, and then $T_{v}^{\,(2k+3,2k)}$ consists of the $2$-arcs of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_s) \!\sim\! (2,{\bf e}_s)$,
so $|T_{v}^{\,(2k+3,2k)}| = k\!-\!1$, but on the other hand, $T_{w}^{\,(2k+3,2k)}$ is empty.
When $m = 4$, the girth of $B(k,m,n)$ is $4$, and the situation is similar, but slightly different. In this case, every $2$-arc lies in either $2k+2$, $2k+5$ or $6k-6$ cycles of length 6,
and $|T_{v}^{\,(2k+5,2k+2)}| = k\!-\!1$ while $|T_{w}^{\,(2k+5,2k+2)}| = 0$.
Once again, it follows that no automorphism can reverse the arc $(v,w)$, and so the graph is half-arc-transitive, for all $k > 2$ and all $m \equiv 0$ mod $4$.
\subsection{The graphs $B(k,m,7)$ for $(k,m) \ne (2,3)$ or $(2,6)$}
Suppose finally that $n = 7$, with $m \equiv 0$ mod $3$, but $(k,m) \ne (2,3)$ or $(2,6)$. Here $m \equiv 0$ mod $3$, and we treat four sub-cases separately: (a) $k = 2$ and $m > 6$; (b) $k > 2$ and $m = 3$; (c) $k > 2$ and $m = 3$; and (d) $k > 2$ and $m > 6$.
In case (a), where $k = 2$ and $m > 6$, every $2$-arc lies in $1$, $2$ or $3$ cycles of length $6$. Also the set $V^{(3)}$ consists of the single vertex $(0,-1)$, while $W^{(3)}$ consists of the single vertex $(1,1)$, and then $T_{v}^{\,(3,1)}$ consists of the single $2$-arc $(0,0) \!\sim\! (1,1) \!\sim\! (2,1)$,
so $|T_{v}^{\,(3,1)}| = 1$, but on the other hand, $T_{w}^{\,(3,1)}$ is empty.
In case (b), where $k > 2$ and $m = 3$, every $2$-arc lies in $0$, $2$ or $k$ cycles of length $6$. In this case the set $V^{(k)}$ consists of the $k\!-\!1$ vertices of the form $(0,-{\bf e}_r)$, while $W^{(k)}$ consists of the $k\!-\!1$ vertices of the form $(1,{\bf e}_s)$, and then $T_{v}^{\,(k,2)}$ consists of the $2$-arcs of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_s) \!\sim\! (2,{\bf e}_s)$,
so $|T_{v}^{\,(k,2)}| = k-1$, but on the other hand, $T_{w}^{\,(k,2)}$ is empty.
In case (c), where $k > 2$ and $m = 6$, every $2$-arc lies in $3$, $k\!+\!1$ or $4k\!-\!3$ cycles of length $6$. Here $V^{(k+1)}$ consists of the $k\!-\!1$ vertices of the form $(0,-{\bf e}_r)$, while $W^{(k+1)}$ consists of the $k\!-\!1$ vertices of the form $(1,{\bf e}_s)$, and then $T_{v}^{\,(k+1,3)}$ consists of the $2$-arcs of the form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_s) \!\sim\! (2,{\bf e}_s)$,
and so $|T_{v}^{\,(k+1,3)}| = k-1$, but on the other hand, $T_{w}^{\,(k+1,3)}$ is empty.
In case (d), where $k > 2$ and $m > 6$, every 2-arc lies in $1$, $k\!+\!1$ or $2k\!-\!2$ cycles of length $6$. Next, if $k > 3$ then $V^{(k+1)}$ consists of the $k\!-\!1$ vertices of the form $(0,-{\bf e}_r)$, while $W^{(k+1)}$ consists of the $k\!-\!1$ vertices of the form $(1,{\bf e}_s)$, but if $k = 3$ then $k+1 = 2k-2$, and $V^{(k+1)}$ contains also $(2,{\bf 0})$ while $W^{(k+1)}$ contains also $(-1,{\bf 0})$. Whether $k = 3$ or $k > 3$, the set $T_{v}^{\,(k+1,1)}$ consists of the $2$-arcs of the
form $(0,{\bf 0}) \!\sim\! (1,{\bf e}_s) \!\sim\! (2,{\bf e}_s)$, and so $|T_{v}^{\,(k+1,1)}| = k-1$, but on the other hand, $T_{w}^{\,(k+1,1)}$ is empty.
Hence in all four of cases (a) to (d), no automorphism can reverse the arc $(v,w)$, and therefore the graph is half-arc-transitive.
This completes the proof of our Theorem.
\noindent {\Large\bf Acknowledgements}
This work was undertaken while the second author was visiting the University of Auckland in 2014, and was partially supported by the ARRS (via grant P1-0294), the European Science Foundation (via EuroGIGA/GReGAS grant N1-0011), and the N.Z. Marsden Fund (via grant UOA1323). The authors would also like to thank Toma{\v z} Pisanski and Primo{\v z} Poto\v{c}nik for discussions which encouraged them to work on this topic, and acknowledge considerable use of the {\sc Magma} system~\cite{Magma} for computational experiments and verification of some of the contents of the paper.
\end{document}
|
arXiv
|
{
"id": "1505.02299.tex",
"language_detection_score": 0.6081059575080872,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Finding All Useless Arcs in Directed Planar Graphs}
\begin{abstract}
We present a linear-time algorithm for simplifying flow networks on directed planar graphs: Given a directed planar graph on $n$ vertices, a source vertex $s$ and a sink vertex $t$, our algorithm removes all the arcs that do not participate in any simple $s,t$-path in linear-time.
The output graph produced by our algorithm satisfies the prerequisite needed by the $O(n\log n)$-time algorithm of Weihe [FOCS'94 \& JCSS'97] for computing maximum $s,t$-flow in directed planar graphs. Previously, Weihe's algorithm could not run in $O(n\log n)$-time due to the absence of the preprocessing step; all the preceding algorithms run in $\tilde{\Omega}(n^2)$-time [Misiolek-Chen, COCOON'05 \& IPL'06; Biedl, Brejov{\'{a}} and Vinar, MFCS'00].
Consequently, this provides an alternative $O(n\log n)$-time algorithm for computing maximum $s,t$-flow in directed planar graphs in addition to the known $O(n\log n)$-time algorithms [Borradaile-Klein, SODA'06 \& J.ACM'09;
Erickson, SODA'10].
Our algorithm can be seen as a (truly) linear-time $s,t$-flow sparsifier for directed planar graphs, which runs faster than any maximum $s,t$-flow algorithm (which can also be seen of as a sparsifier).
The simplified structures of the resulting graph might be useful in future developments of maximum $s,t$-flow algorithms in both directed and undirected planar graphs. \end{abstract}
\section{Introduction} \label{sec:intro}
The {\em maximum $s,t$-flow} problem is a fundamental problem in Combinatorial Optimization that has numerous applications in both theory and practice.
A basic instance of maximum flow where the underlying graph is planar has been considered as an important special case and has been studied since 50's in the early work of Ford and Fulkerson \cite{FordF56}.
Since then, there have been steady developments of maximum flow algorithms on undirected planar graphs. Itai and Shiloach \cite{ItaiS79} proposed an algorithm for the maximum $s,t$-flow problem on undirected planar graphs that runs in $O(n^2\log n)$ time, and in subsequent works \cite{Hassin81,Reif83,HassinJ85,Frederickson87,KleinRRS94-STOC94,HenzingerKRS97,ItalianoNSW11}, the running time have been improved to the current best $O(n\log\log{n})$-time algorithm by Italiano~et~al.~\cite{ItalianoNSW11}.
Another line of research is the study of the maximum $s,t$-flow problem in directed planar graphs. The fastest algorithm with the running time of $O(n\log n)$ is due to Borradaile and Klein~\cite{BorradaileK09}. Historically, in 1994, Weihe~\cite{Weihe97} presented a novel approach that would solve the maximum $s,t$-flow problem on directed planar graphs in $O(n\log n)$ time. However, Weihe's algorithm requires a preprocessing step that transforms an input graph into a particular form: (1) each vertex (except source and sink) has degree three, (2) the planar embedding has no clockwise cycle, and (3) every arc must participate in some {\em simple $s,t$-path}.
The first condition can be guaranteed by a simple (and thus linear-time) reduction, and the second condition can be obtained by an algorithm of Khuller, Naor and Klein in \cite{KhullerNK93} which runs in $O(n\log n)$-time (this was later improved to linear time \cite{HenzingerKRS97}).
Unfortunately, for the third condition, there was no known algorithm that could remove all such {\em useless} arcs in $O(n\log n)$-time. As this problem seems to be simple, this issue had not been noticed until it was pointed out much later by Biedl, Brejov{\'{a}} and Vinar \cite{BiedlBV00}.
Although an $O(n\log n)$-time algorithm for the maximum $s,t$-flow problem in directed planar graphs has been devised by Borradaile and Klein \cite{BorradaileK09}, the question of removing all the useless arcs in $O(n\log n)$-time remains unsolved.
In this paper, we study the {\em flow network simplification} problem, where we are given a directed planar graph $G=(V,E)$ on $n$ vertices, a source vertex $s$ and a sink vertex $t$, and the goal is to remove all the arcs $e$ that are not contained in any simple $s,t$-path.
One may observe that the problem of determining the usefulness of a single arc involves finding two vertex-disjoint paths, which is NP-complete in general directed graphs \cite{FortuneHW80}.
Thus, detecting all the useless arcs at once is non-trivial.
Here we present a linear-time algorithm that determines all the useless arcs, thus solving the problem left open in the work of Weihe~\cite{Weihe97} and settling the complexity of simplifying a flow network in directed planar graphs.
The main ingredient of our algorithm is a decomposition algorithm that slices a plane graph into small strips with simple structures. This allows us to design an appropriate algorithm and data structure to handle each strip separately. Our data structure is simple but requires a rigorous analysis of the properties of a planar embedding. We use information provided by the data structure to determine whether each arc $e$ is contained in a simple $s,t$-path $P$ in $O(1)$ time, thus yielding a linear-time algorithm.
The main difficulty is that we cannot afford to explicitly compute such $s,t$-path $P$ (if one exists) as it would lead to $O(n^2)$ running time. The existence of any path $P$ can only be determined implicitly by our structural lemmas.
Our main algorithm runs in linear time. However, it requires the planar embedding (which is assumed to be given together with the input graph) to contain no clockwise cycle and that every vertex except the source and the sink has degree exactly three. We provide a sequence of linear-time reductions that produces a plane graph that satisfies the above requirements with the guarantees that
the value of maximum $s,t$-flow is preserved.
In particular, we apply a standard degree-reduction to reduce the degree of the input graph and invoke the algorithm of Khuller, Naor, and Klein \cite{KhullerNK93} to modify the planar embedding using an application of shortest path computation in the dual graph, which can be done in linear time by the algorithm of Henzinger, Klein, Rao, and Subramanian~\cite{HenzingerKRS97}.
\subsection{Our Contributions} \label{sec:contributions}
Our contribution is two-fold.
Firstly, our algorithm removes all the useless arcs in a directed planar $s,t$-flow network in linear-time, which thus completes the framework of Weihe \cite{Weihe97} and yields an alternative $O(n\log n)$-time algorithm for computing maximum $s,t$-flow on directed planar graphs.
Secondly, our algorithm can be seen as a (truly) linear-time $s,t$-flow sparsifier, which runs faster than any known maximum $s,t$-flow algorithm on directed (and also undirected) planar graphs (which can be seen as an $s,t$-flow sparsifier as well).
The plane graph produced by our algorithm has simple structures that can be potentially exploited in the further developments of network flow algorithms; in particular, this may lead to a simpler or faster algorithm for computing maximum $s,t$-flow in directed planar graphs. In addition, our algorithm could be adapted to simplify a flow network in undirected planar graphs as well.
\subsection{Related Work} \label{sec:related-work}
Planar duality plays crucial roles in planar flow algorithm developments.
Many maximum flow algorithms in planar graphs exploit a shortest path algorithm as a subroutine.
However, the $O(n\log n)$-time algorithm for the maximum $s,t$-flow problem in directed planar graphs by Borradaile and Klein \cite{BorradaileK09}, which is essentially a left-most augmenting-path algorithm, is not based on shortest path algorithms. Erickson~\cite{Erickson10} reformulated the maximum flow problem as a parametric shortest path problem. The two $O(n\log n)$-algorithms were shown to be the same. \cite{Erickson10} but with different interpretation.
Roughly speaking, the first one by Borradaile and Klein runs mainly on the primal graph while the latter one by Erickson runs mainly on the dual graph.
Borradaile and Harutyunyan~\cite{BorradaileH13} explored this path-flow duality further and showed a correspondence between maximum flows and shortest paths in directed planar graphs with no restriction on the locations of the source and the sink.
For undirected planar unit capacity networks, Weihe~\cite{Weihe97-undirected} presented a linear time algorithm for the maximum $s,t$-flow. Brandes and Wagner~\cite{BrandesW00} and Eisenstat and Klein~\cite{EisenstatK13} presented linear time algorithms when the planar unit-capacity network is directed.
A seminal work of Miller and Naor~\cite{MillerN95} studied flow problems in planar graphs with multiple sources and multiple sinks. The problem is special in planar graphs as one cannot use standard reduction to the single-source single-sink problem without destroying the planarity. Using a wide range of planar graph techniques, Borradaile, Klein, Mozes, Nussbaum, Wulff-Nilsen~\cite{BorradaileKMNWN11} presented an $O(n\log^3 n)$ algorithm for this problem.
\section{Preliminaries and Background} \label{sec:prelim}
We use the standard terminology in graph theory. By the {\em degree} of a vertex in a directed graph, we mean the sum of its indegree and its outdegree.
A path $P\subseteq G$ is a {\em simple path} if each vertex appears in $P$ at most once.
A {\em cycle} in a directed graph is a simple directed walk (i.e., no internal vertices appear twice) that starts and ends at the same vertex.
A {\em strongly connected component} of $G$ is a maximal subgraph of $G$ that is strongly connected.
The {\em boundary} of a strongly connected component is the set of arcs that encloses the component (arcs on the unbounded face of the component). We call those arcs {\em boundary arcs}.
For any strongly connected component $C$ such that $s,t\notin V(C)$, a vertex $v\in V(C)$ is an {\em entrance} of $C$ if there exists an $s,v$-path $P\subseteq G$ such that $P$ contains no vertices of $C$ except $v$, i.e., $V(P)\cap V(C)=\{v\}$.
Similarly, a vertex $v\in V(C)$ is an {\em exit} of $C$ if there exists an $v,t$-path $P\subseteq G$ such that $P$ contains no vertices of $C$ except $v$, i.e., $V(P)\cap V(C)=\{v\}$.
Consider an $s,t$-flow network consisting of a directed planar graph $G$, a source vertex $s$ and a sink vertex $t$.
We say that an arc $uv$ is a {\em useful} arc (w.r.t. $s$ and $t$) if there is a simple $s,t$-path $P$ containing $uv$. Thus, the $s,u$ and the $v,t$ subpaths of $P$ have no common vertices. Otherwise, if there is no simple $s,t$-path containing $uv$, then we say that $uv$ is a {\em useless} arc (w.r.t. $s$ and $t$).
Similarly, a path $P$ is {\em useful} (w.r.t. $s$ and $t$) if there is a simple $s,t$-path $Q$ that contains $P$.
Note that if a path $P$ is useful, then all the arcs of $P$ are useful. However, the converse is not true, i.e., all the arcs of $P$ are useful does not imply that $P$ is a useful path.
Throughout the paper, we denote the input directed planar graph by $G$ and denote the number of vertices and arcs of $G$ by $n$ and $m=O(n)$, respectively. We also assume that a planar embedding of $G$ is given as input, and the sink vertex $t$ is on the unbounded face of the embedding.
Note that we assume that our flow networks have unit-capacities although the algorithm works on networks with arbitrary capacities as well.
\subsection{Planar Embedding and Basic Subroutines} \label{sec:prelim:planar-embedding}
We assume that the plane graph is given in a standard format in which the input graph is represented by an adjacency list whose arcs sorted in counterclockwise order. To be precise, our input is a table $\calT$ of adjacency lists of arcs incident to each vertex $v\in V(G)$ (i.e., $V(G)$ is the index of $\calT$) sorted in counterclockwise order. Each adjacency list $\calT(v)$ is a doubly linked-list of arcs having $v$ as either heads or tails (i.e., arcs of the forms $uv$ or $vw$).
Given an arc $vw$ one can find an arc $vw'$ next to $vw$ in the counterclockwise (or clockwise) order. This allows us to query in $O(1)$ time the {\em right-most arc} (resp., {\em left-most arc}) of $uv$, which is an arc $vw$ in the reverse direction that is nearest to $uv$ in the counterclockwise (resp., clockwise) order.
\subsection{Left and Right Directions} \label{sec:prelim:left-right}
Since we are given a planar embedding, we may compare two paths/arcs using the notion of {\em left to} and {\em right to}.
Let us consider a reference path $P$ and place it so that the start vertex of $P$ is below its end vertex.
We say that a path $Q$ is {\em right to} $P$ if we can cut the plane graph along the path $P$ in such a way that $Q$ is enclosed in the right side of the half-plane.
Our definition includes the case that $Q$ is a single arc.
We say that an arc $uu'$ {\em leaves $P$ to} (resp., {\em enters $P$ from}) the right if $u\in V(P)$ and $uu'$ lies on the right half-plane cutting along $P$.
Similarly, we say that a path $Q$ {\em leaves $P$ to} (resp., {\em enters $P$ from}) the right if the first arc of $Q$ leaves $P$ to (resp., enters $P$ from) the right half-plane cutting along $P$.
These terms are also used analogously for the case of the left direction.
As mentioned, the representation of a plane graph described in the previous section allow us to query in $O(1)$ time the {\em right-most} arc $vw$ of a given arc $uv$, which is the arc in the reverse direction nearest to $uv$ in the counterclockwise order. Consequently, we may define the {\em right-first-search} algorithm as a variant of the depth-first-search algorithm that chooses to traverse to the right-most (unvisited) arc in each step of the traversal.
We may assume that the source $s$ has a single arc $e$ leaving it. Thus, the right-first-search (resp., left-first-search) started from $s$ gives a unique ordering of vertices (because every path must start from the arc $e$).
\subsection{Forward and Backward Paths} \label{sec:prelim:forward-backward}
Forward and backward paths are crucial objects that we use to determine whether an edge is useful or useless.
Let $F$ be a reference path, which we call a {\em floor}. We order vertices on $F$ according to their appearance on $F$
and denote such order by $\pi:V(F)\rightarrow [|V(F)|]$. We may think that $F$ is a line that goes from left to right.
Consider any path $P$ that starts and ends at vertices on $F$. We say that $P$ is a {\em forward path $P$ w.r.t. $F$} if the order of vertices in $V(P)\cap V(F)$ agrees with that of $F$, i.e., for all vertices $u,v\in V(P)\cap V(F)$, $\pi(u) < \pi(v)$ implies that $u$ appears before $v$ on $P$; otherwise, we say that $P$ is a {\em backward path $P$ w.r.t. $F$}.
In general, we will use the term forward path to mean a minimal forward path, i.e., a forward path $P$ that intersects $F$ only at its start and end vertices and share no inner vertices with $F$.
We will use the term {\em extended forward path} to mean a non-minimal forward path.
In this sense, the ceiling $U$ is also an extended forward path.
Analogously, we use the same notions for backward paths.
A path $P$ is {\em flipped} w.r.t. $F$ if it starts from the left-side of $F$ and ends on the right-side of $F$. Similarly, $P$ is {\em reverse-flipped} w.r.t. $F$ if it starts from the right-side of $F$ and ends on the left-side of $F$.
A {\em non-flipped} path is a path that lies on the left-side of $F$, and a {\em hanging} path is a path that lies on the right-side of $F$.
\subsection{Strips and Links} \label{sec:prelim:strips}
We are ready to define our main structural components. A strip is formed by a {\em floor} $F$ and a forward path $U$ w.r.t. $F$, called a {\em ceiling}. A {\em strip} is a region enclosed by a floor $F$ and a ceiling $U$, and we denote the strip by $C_{U,F}$.
Observe that if we draw $F$ in such a way that $F$ is a line that goes from left to right, then $F$ lies beneath $U$ in the planar drawing.
The two paths $U$ and $F$ form the {\em boundary} of the strip $C_{U,F}$. A strip is {\em proper} if $U$ and $F$ are internally vertex disjoint, i.e., they have no vertices in common except the start and the end vertices.
Generally, we use the term strip to mean a proper strip except for the case of {\em primary strip}, which we will define later in Section~\ref{sec:s-inside}.
Consider a proper strip $C_{U,F}$. We call arcs in $U$ and $F$ {\em boundary arcs} and call other arcs (i.e., arcs in $E(C_{U,F})- (E(U)\cup E(F))$) {\em inner arcs}. Similarly, vertices in $U$ and $F$ are called {\em boundary vertices} and other vertices (i.e., vertices in $V(C_{U,F})- (V(U)\cup V(F))$) are called {\em inner vertices}.
A path $P\subseteq C_{U,F}$ which is not entirely contained in $F$ or $U$ is called a {\em link} in $C_{U,F}$ if its start and end vertices are boundary vertices and all other vertices are inner vertices, i.e., $V(P)\cap (V(U)\cup V(F))=\{u,v\}$ where $u,v$ are the first and the last vertices of $P$, respectively. Observe that a link has no boundary arcs, i.e., $E(P)\cap (E(U)\cup E(F))=\emptyset$.
A link in $C_{U,F}$ whose start and end vertices are in the floor $F$ can be classified as {\em forward} and {\em backward} in the same manner as forward and backward paths (w.r.t. the floor $F$). Specifically, {\em forward} (resp., {\em backward}) links are minimal forward (resp., backward) paths.
A link, however, can start at some vertex on the floor $F$ and end at an inner vertex of the ceiling $U$; in this case, we call it an {\em
up-cross} path. On the other hand, if it goes from an inner vertex of the ceiling $U$ to some vertex on the floor $F$, we call it a {\em
down-cross} path. A {\em hanging path} inside a strip is defined to be a hanging path w.r.t. the ceiling $U$. That is, a hanging path is a link that starts and ends at inner vertices of the ceiling $U$.
Note that a hanging path can be either forward or backward hanging path.
The classification of these links are shown in Figure~\ref{fig:paths-in-strip}.
\begin{figure}
\caption{The classification of links in a strip.}
\label{fig:paths-in-strip}
\end{figure}
A link inside a proper strip slices the strip into two pieces: one on the left and on the right of the link. If we slice a strip with a forward, up-cross, down-cross or forward hanging path, then we have two strips that are again proper strips.
In this sense, we say that a strip is {\em minimal} if we cannot slice it further, i.e., it has neither forward, up-cross, down-cross nor forward hanging paths (but, backward paths are allowed).
\subsection{The Usefulness of Forward Paths}
The usefulness of forward paths inside a strip $C_{U,F}$ can be determined by the usefulness of its floor as stated in the following lemma.
\begin{lemma} \label{lem:forward-path-is-useful} Consider a strip $C_{U,F}$ such that the source vertex $s$ and the sink vertex $t$ are not in $C_{U,F}$. If $F$ is useful, then so is any forward path w.r.t. $C_{U,F}$. \end{lemma} \begin{proof} Let $u$ and $v$ be the start and the end vertices of $F$, and let $P$ be a forward path connecting some vertices $q,w \in V(F)$. Since $F$ is useful, there exists a simple $s,t$-path $R$ that contains $F$. We may choose $R$ to be the shortest such path. Since $s,t$ are not in $C_{U,F}$, the minimality of $R$ implies that none of the vertices in $V(R)- V(F)$ are in $C_{U,F}$. Consequently, the path $P$ (which is a link) has no vertices in $V(R)- V(F)$. We will construct a new simple $s,t$-path by replacing the $q,w$-subpath of $R$ by $P$. To be formal, let $R_{s,q}$ and $R_{w,t}$ be the $s,q$ and $w,t$ subpaths of $R$, respectively. We then define a new path $R' = R_{s,q}\cdot P\cdot R_{w,t}$. It can be seen that $R'$ is a simple $s,t$-path. This proves that $P$ is useful. \end{proof}
The following corollary follows immediately from Lemma~\ref{lem:forward-path-is-useful}.
\begin{corollary} \label{cor:ext-forward-path-is-useful} Consider a strip $C_{U,F}$ such that the source vertex $s$ and the sink vertex $t$ are not in $C_{U,F}$. If $F$ is useful, then so is any extended forward path w.r.t. $C_{U,F}$. \end{corollary} \begin{proof} We proceed the proof in the same way as that of Lemma~\ref{lem:forward-path-is-useful}.
First consider any extended forward path $P$. Observe that $P$ can be partitioned into subpaths so that each subpath is either contained in $R$ or is a forward path w.r.t $C_{U,F}$. We replace subpaths of $F$ by these forward paths, which results in a new $u,v$-path $P'\supseteq P$, where $u$ and $v$ are the start and the end vertices of the floor $F$, respectively.
Since $F$ is useful, there exists a simple $s,t$-path $R$ that contains $F$. We may assume that $R$ is the shortest such path.
By the minimality of $R$, we know that $P'$ has no vertices in $V(R)- V(F)$ (because $P'\subseteq C_{U,F}$).
So, we can construct a new simple $s,t$-path $R'$ by replacing the path $F\subseteq R$ in $R$ by the path $P'$.
The simple $s,t$-path $R'\supseteq P'\supseteq P$ certifies that $P$ is useful. \end{proof}
\subsection{Other important facts}
Below are the useful facts that form the basis of our algorithm.
\begin{lemma} \label{lem:remain-useful}
Let $e'$ be any useless arc in $G$.
Then any useful arc $e\in E(G)$ is also useful in $G- \{e'\}$. \end{lemma} \begin{proof} By the definition of useful arc, there exists a simple $s,t$-path $P$ in $G$ containing the arc $e$, and such path $P$ cannot contain $e'$; otherwise, it would certify that $e'$ is useful. Therefore, $P\subseteq G-\{e'\}$, implying that $e$ is useful in $G-\{e'\}$. \end{proof}
\begin{lemma} \label{lem:boundary-is-ccw} Let $C$ be a strongly connected component, and let $B$ be the set of boundary arcs of $C$. Then $B$ forms a counter clockwise cycle. \end{lemma} \begin{proof}
Since we assume that the embedding has no clockwise cycle,
it suffices to show that $B$ is a cycle.
We prove by a contradiction.
Assume that $B$ is not a cycle.
Then we would have two consecutive arcs in $B$ that go
in opposite directions.
(Note that the underlying undirected graph of $B$ is always a
cycle.)
That is, there must exist a vertex $u$ with two leaving arcs,
say $uv$ and $uw$.
Since $u,v,w$ are in the same strongly connected component,
the component $C$ must have a $v,u$-path $P$ and and a $w,u$-path
$P'$.
We may assume minimality of $P$ and $P'$ and thus assume that
they are simple paths.
Now $P\cup\{uv\}$ and $P'\cup\{uw\}$ form a cycle,
and only one of them can be counterclockwise
(since $uv$ and $uw$ are in opposite direction),
a contradiction. \end{proof}
\begin{lemma} \label{lem:disjoint-entrance-exit-paths} Let $u$ and $v$ be an entrance and an exit of $C$, and let $P_u$ and $P_v$ be an $s,u$-path $P_u$ and a $v,t$-path $P_v$ that contains no vertices of $C$ except $u$ and $v$, respectively. Then $P_u$ and $P_v$ are vertex disjoint. \end{lemma} \begin{proof} Suppose $P_u$ and $P_v$ are not vertex-disjoint. Then the intersection of $P_u$ and $P_v$ induces a strongly connected component strictly containing $C$. This is a contradiction since $C$ is a maximal strongly connected subgraph of $G$. \end{proof}
\section{Overview of Our Algorithm} \label{sec:overview}
In this section, we give an overview of the algorithm.
First, we preprocess the plane graph so that it meets the following requirements. \begin{enumerate} \item There is {\em no clockwise cycle} in the planar embedding. \item The source $s$ is adjacent to only {\em one outgoing arc}. \item Every vertex except $s$ and $t$ has {\em degree three}. \end{enumerate} We provide in Section~\ref{sec:preprocessing} a sequence of reductions that outputs a graph satisfied the above conditions while guaranteeing that
the value of maximum $s,t$-flow is preserved.
It is worth noting that the degree-three condition is not a strict requirement for our main algorithm. We only need every vertex to have either indegree one or outdegree one; a network that meets this prerequisite is called a {\em unit network}.
After the preprocessing step, we apply the main algorithm, which first decomposes the plane graph into a collection of strongly connected components. The algorithm then processes each strongly connected component independently.
Notice that the usefulness of arcs that lie outside of the strongly connected components can be determined by basic graph search algorithms.
To see this, let us contract all the strongly connected components, which results in a {\em directed acyclic graph} (DAG). Since there is no cycle, an arc $uv$ is useful if and only if $u$ is reachable from $s$, and $t$ is reachable from $v$. So, it suffices to run any graph search algorithm to list all the vertices reachable from $s$ and those reachable from $t$ in the reverse graph. Hence, we are left to work on arcs inside each strongly connected component.
We classify each strongly connected component into an {\bf outside case} and an {\bf inside case}. The outside case is the case that the source $s$ lies outside of the component, and the inside case is the case that the source $s$ is enclosed inside the component. (Note that since $s$ has no incoming arc, it cannot be on the boundary of a strongly connected component.) We deal with the outside case in Section~\ref{sec:st-outside} and deal with the inside case in Section~\ref{sec:s-inside}. While the inside case is more complicated, we decompose a component of the inside case into subgraphs, which resemblance those in the outside cases.
In both cases, the main ingredient is the {\em strip decomposition}, which allows us to handle all the useless arcs in one go. We present the strip decomposition algorithm in Section~\ref{sec:strip-decomposition}.
Please see Figure~\ref{fig:outside-inside} for the example of outside and inside cases. It is possible to have many inside cases (see Figure~\ref{fig:component-example}).
\begin{figure}
\caption{The top figure illustrates the outside case, while the
bottom figure illustrates the inside case.}
\label{fig:outside-inside}
\end{figure}
\begin{figure}
\caption{An example showing 7 components. Note the nested inside cases.}
\label{fig:component-example}
\end{figure}
\section{Preprocessing The Plane Graph} \label{sec:preprocessing}
In this section, we present a sequence of reductions that provides a plane graph required by the main algorithm.
The input to this algorithm is a directed planar flow network $(G_0,s_0,t_0)$ consisting of a plane graph $G_0$ (i.e., a directed planar graph plus a planar embedding), a source $s_0$ and a sink $t_0$.
Our goal is to compute a directed planar flow network $(G,s,t)$ consisting of a plane graph $G$, a source $s$ and a sink $t$ with the following properties. \begin{itemize} \item[(1)] The value of the maximum flow in the network $(G,s,t)$
and that in the original network $(G_0,s_0,t_0)$ are the same. \item[(2)] The plane graph $G$ has no clockwise cycle. \item[(3)] Every vertex in $G$ except the source $s$ and the sink $t$ has
degree exactly three. \item[(4)] The source $s$ has no incoming arc and has outdegree one. \item[(5)] The resulting graph $G$ has
$|V(G)| \leq |E(G_0)|+2$ and $|E(G)| \leq 2|E(G_0)|+2$. \end{itemize}
The main subroutine in the preprocessing step is the algorithm for removing all the clockwise cycles of a given plane graph $H$ by Khuller, Naor and Klein \cite{KhullerNK93}, which guarantees to preserve the value of maximum $s,t$-flow. This subroutine can be implemented by computing single-source shortest path in the dual planar graph. The setting of this shortest path problem is by placing a source $s^*$ on the outer-face of the dual graph, which can be computed in linear-time using the algorithm of Henzinger, Klein, Rao, and Subramanian~\cite{HenzingerKRS97}. This subroutine introduces no new vertices but may reverse some arcs.
\subsection{Reduction} \label{sec:preprocessing:reduction}
Our preprocessing first removes all the clockwise cycles. Then it applies a standard reduction to reduce the degree of all the vertices:
\begin{itemize} \item {\bf Step 1:} We remove all the clockwise cycles from the plane graph $G_0$.
We denote the plane graph obtained in this step by $G_1$.
\item {\bf Step 2:} We apply the {\em degree reduction} to obtain a plane graph $G_2$
that satisfies the degree conditions in Properties (4) and (5):
We replace the source $s_0$ by an arc $ss_0$, i.e.,
we introduce the new source $s$ and add an arc $ss_0$ joining $s$ to $s_0$.
For notation convenience, we rename $t_0$ as $t$.
Then we replace each vertex $v$ of degree $d_v$ by a counterclockwise cycle
$\hat{C}_v$ of length $d_v$. We re-direct each arc going to $v$ and leaving $v$ to
exactly one vertex $\hat{C}_v$ in such a way that no arcs cross.
(So, each vertex of $\hat{C}_v$ corresponds to an arc incident to $v$ in $G_0$) \end{itemize}
As we discussed, it follows by the result in \cite{KhullerNK93} that removing clockwise cycles does not change the value of the maximum flow, and it is not hard to see that the degree reduction has no effect on the value of maximum flow on edge-capacitated networks. Thus, Property~(1) holds for $(G,s,t)$.
Property~(2) holds simply because we remove all the clockwise cycles at the end, and Property~(3)-(5) follows directly from the degree reduction.
\section{Strip-Slicing Decomposition} \label{sec:strip-decomposition}
The crucial part of our algorithm is the decomposition algorithm that decomposes a strongly connected component $C$ into (proper) {\em strips}, which are regions of $C$ in the planar embedding enclose by two arc-disjoint paths $U$ and $F$ that share start and end vertices.
We will present the {\em strip-slicing decomposition} algorithm that decomposes a given strip $C_{U,F}$ into a collection $\calS$ of {\em minimal} strips in the sense that any strip in $\calS$ cannot be sliced into smaller strips.
Moreover, any two strips obtained by the decomposition are either disjoint or share only their boundary vertices.
We claim that the above decomposition can be done in linear time.
If then certain prerequisites are met (we will discuss this in the later section), then our strip-slicing decomposition gives a collection of strips such that all the inner arcs of each strip are useless while all the boundary arcs are useful.
Before proceeding to the presentation of the algorithm, we advise the readers to recall definitions of links and paths in a strip in Section~\ref{sec:prelim:strips}.
\subsection{The Decomposition Algorithm} \label{sec:strip-decomposition:algorithm}
Now we describe our strip-slicing decomposition algorithm. We start our discussion by presenting an abstraction of the algorithm.
We denote the initial (input) strip by $C_{U^*,F^*}$. We call $U^*$ the {\em top-most} ceiling and call $F^*$ the {\em lowest floor}. The top-most ceiling $U^*$ can be a dummy ceiling that does not exist in the graph.
The decomposition algorithm taking as input a strip $C_{U,F}$. Then it finds the ``right-most'' path $P$ w.r.t. the floor $F$ that is either a forward path or an up-cross path, which we call a {\em slicing path}. This path $P$ slices the strips into two regions, the {\em up-strip} and the {\em down-strip}. Please see Figure~\ref{fig:slicing-strip} for illustration. Intuitively, we wish to slice the input strip into pieces so that the final collection consists of strips that have no forward, up-cross, down-cross nor forward hanging paths (each of these paths can slice a strip into smaller ones).
\begin{figure}
\caption{The path $P$ slices the strip into up-strip and down-strips.}
\label{fig:slicing-strip}
\end{figure}
The naive implementation yields an $O(n^2)$-time algorithm; however, with a careful implementation, we can speed up the running time to $O(n)$. We remark that our decomposition algorithms (both the quadratic and linear time algorithm) do not need the properties that the input plane graph has no clockwise cycles and that every vertex (except source and sink) has degree three. Thus, we are free to modify the graph as long as the graph remains planar.
To make the readers familiar with our notations, we start by presenting the quadratic-time algorithm in Section~\ref{sec:strip-decomposition:quadratic-time} and prove its correctness. Then we will present the linear-time algorithm in Section~\ref{sec:strip-decomposition:linear-time} and show that it has the same properties as that of the quadratic-time algorithm. The basic subroutines that are shared in both algorithms are presented in Section~\ref{sec:strip-decomposition:subroutines}
More precisely, we prove the following lemmas in the next two sections.
\begin{lemma} \label{lem:quadratic-time-decomposition} There is an $O(n^2)$-time algorithm that, given as input a strip $C_{U^*,F^*}$, outputs a collection $\calS$ of strips such that each strip $C_{U,F}\in\calS$ has neither forward, up-cross, down-cross nor forward hanging path (w.r.t. $U$).
\end{lemma}
\begin{lemma} \label{lem:linear-time-decomposition} There is a linear-time algorithm that, given as input a strip $C_{U^*,F^*}$, outputs a collection $\calS$ of strips such that each strip $C_{U,F}\in\calS$ has neither forward, up-cross, down-cross nor forward hanging path (w.r.t. $U$).
\end{lemma}
The important property that we need from the decomposition algorithm is that if the floor of the input strip $C_{U^*,F^*}$ is useful, then all the boundary arcs of every strip $C_{U,F}\in\mathcal{S}$ are also useful.
\begin{lemma} \label{lem:properties-strips} Let $\mathcal{S}$ be a collection of minimal strips obtained by running the strip-slicing decomposition on an input strip $C_{U^*,F^*}$. Suppose the source $s$ and the sink $t$ are not enclosed in $C_{U^*,F^*}$ and that the floor $F^*$ is useful. Then, for every strip $C_{U,F}\in\mathcal{S}$, all the boundary arcs of $C_{U,F}$ (i.e., arcs in $U\cup F$) are useful. \end{lemma} \begin{proof} We prove by induction that the decomposition algorithm maintains an invariant that the ceiling $U$ and the floor $F$ are useful in every recursive call.
For the base case, since $F^*$ is useful, Lemma~\ref{lem:forward-path-is-useful} implies that $U^*$ is useful because it is a forward path w.r.t. $F^*$. Inductively, assume that the claim holds prior to running the decomposition algorithm on a strip $C_{U,F}$.
Thus, $U$ and $F$ are useful. If the algorithm finds no slicing-path, then we are done. Otherwise, it finds a slicing-path $P$, which slices the strip $C_{U,F}$ into $C_{U_{up},F_{up}}$ and $C_{U_{down},F_{down}}$.
Let $s'$ and $t'$ be the start and end vertices of the strip $C_{U,F}$, respectively.
Observe that each of the path $U_{up},F_{up},U_{down}$ and $F_{down}$ is a simple $s',t'$-path, regardless of whether the slicing $P$ is a forward path or an up-cross path.
Since $s$ and $t$ are not enclosed in $C_{U,F}$, we can extend any simple $s',t'$-path $Q$ into a simple $s,t$-path, which certifies that $Q$ is a useful path.
Therefore, $U_{up},F_{up},U_{down}$ and $F_{down}$ are all useful, and the lemma follows. \end{proof}
\subsection{Basic Subroutines} \label{sec:strip-decomposition:subroutines}
Before proceeding, we formally describe the two basic procedures that will be used as subroutines in the main algorithms.
\paragraph*{PROCEDURE 1: Slice the strip $C_{U,F}$ by a path $P$.} This procedure outputs two strips $C_{U_{up},F_{up}},C_{U_{down},F_{down}}$. We require that $P$ is either a forward-path or an up-cross path. Let $s',t'$ be the start and end vertices of the strip $C_{U,F}$ (which are end vertices of both $U$ and $F$), and let $u,v$ be the start and the end vertices of $P$, respectively. We define the two new strips by defining the new floors and strips as below. \begin{itemize} \item {\bf Case 1}: $P$ is a {\bf forward path}.
Then we know that $u,v$ are on the floor $F$
and other vertices of $P$ are not on the boundary of $C_{U,F}$.
To define the up-strip, we let $U_{up} := U$ and
construct $F_{up}$ by replacing the $u,v$-subpath of $F$ by $P$.
To define the down-strip, we let $U_{down} := P$ and
let $F_{down}$ be the $u,v$-subpath of $F$.
\item {\bf Case 2}: $P$ is an {\bf up-cross path}.
Then we know that $u$ is on the floor $F$ while $v$ is on the ceiling $U$
and other vertices of $P$ has not on the boundary of $C_{U,F}$.
To define the up-strip, we let $U_{up}$ be the $s',u$-subpath of $U$
and construct $F_{up}$ by concatenating
the $s',u$-subpath of $F$ with $P$.
To define the down-strip, we construct $U_{down}$ by concatenating
the path $P$ with the $v,t'$-subpath of $U$,
and we let $F_{down}$ be the the $u,t'$-subpath of $F$. \end{itemize}
\paragraph*{PROCEDURE 2: Find a slicing path starting from $u\in F$.} We require that there are no arcs leaving the strip $C_{U,F}$ (we can temporary remove them). The algorithm runs recursively in the right first search manner (recall this is a variant of depth first search where we choose the traverse to the right-most arc). For simplicity, we assume that the first arc that will be scanned is the arc leaving $u$ that is next to the floor arc. Specifically, letting $uw\in F$ be the floor arc and $uw'$ be the arc next to $uw$ in counterclockwise order, the algorithm will ignore $uw$ and start by traversing $uw'$.
We terminate a recursive call when we reach a boundary vertex $v \in U\cup F$. The algorithm then decides whether the path $P$ is a slicing path. More precisely, if $v$ appears before $u$ (thus, $P$ is a forward path) or $v$ is an internal vertex of the ceiling (thus, $P$ is an up-cross path), then the algorithm returns that we found a slicing path $P$ plus reports that $P$ is a forward path or an up-cross path.
Otherwise, we continue the search until we visited all the inner vertices reachable from $u$; in this case, the algorithm returns that there is no slicing path.
Notice that any slicing path $P$ found by the algorithm is the {\em right most} slicing path starting from $u$.
\subsection{Quadratic-Time Strip-Slicing Decomposition (Proof of Lemma~\ref{lem:quadratic-time-decomposition})} \label{sec:strip-decomposition:quadratic-time}
\paragraph*{The Main Algorithm.}
Now we describe our $O(n^2)$-time algorithm in more details. The algorithm reads an input strip $C_{U^*,F^*}$ and then outputs a collection of minimal strips $\calS$. Our algorithm consists of initialization and recursive steps (we may think that these are two separated procedures).
Let $s',t'$ be the start and the end vertices of the input strip $C_{U^*,F^*}$ (i.e., the start and end vertices of the paths $U^*$ and $F^*$).
At the initialization step, we remove every arc leaving the input strip and then make sure that the {\em initial strip} has no hanging path by adding an auxiliary ceiling $(s',x,'t')$ to enclose the input strip, where $x'$ is a dummy vertex that does not exist in the graph.
Clearly, the modified strip has no hanging path.
Next we run the recursive procedure. Let us denote the input strip to this procedure by $C_{U,F}$. We first find a slicing path $P$. If there is no such path, then we add $C_{U,F}$ to the collection of strips $\calS$.
Otherwise, we found a slicing path $P$, and we use it to slice the strip $C_{U,F}$ into the up-strip $C_{U_{up},F_{up}}$ and the down-strip $C_{U_{down},F_{down}}$.
We then make recursive calls on $C_{U_{up},F_{up}}$, $C_{U_{down},F_{down}}$, and terminate the algorithm (or return to the parent recursive-call).
\paragraph*{Analysis}
The running-time of the algorithm can be analyzed by the following recurrent relation: \[ T(n) = T(n_1) + T(n - n_1) + O(n), \mbox{where $T(1)=1$}. \] This implies the quadratic running-time of the algorithm. It is clear from the construction that any two strips in $\calS$ can intersect only at their boundaries and that any strip in the collection $\calS$ contains no forward nor up-cross path. It remains to show that the any strip at the termination has no hanging nor down-cross path.
Observe that the assertion holds trivially at the first call to the recursive procedure because the ceiling $U=(s',x,t')$ is an auxiliary ceiling (we simply have no arc leaving the lone internal vertex of $U$). We will show that this property holds inductively, i.e., we claim that if $C_{U,F}$ has no forward hanging nor down-cross path, then both $C_{U_{up},F_{up}}$ and $C_{U_{down},F_{down}}$ also have no forward hanging nor down-cross path. We consider two cases of the slicing path $P$.
{\bf Case 1: $P$ is a forward path.} Then the up-strip $C_{U_{up},F_{up}}$ has no forward hanging path because $U_{up}=U$ (but, a backward hanging path may exist), and $C_{U_{up},F_{up}}$ has no down-cross path because, otherwise, any such path together with a subpath of $P$ would form a down-cross path in $C_{U,F}$.
For the down-strip $C_{U_{down},F_{down}}$, we know that $P=F_{U_{down}}$ and that $P$ is the right-most path w.r.t. to $F_{up}$. This means that there is no forward hanging path w.r.t. $U_{down}$ simply because the hanging path $Q$ by definition must lie on the right of $U_{down}=P$, which contradicts the fact that $P$ is the right-most forward path (we can replace a subpath of $P$ by $Q$ to form a forward path right to $P$). Similarly, any down-cross path w.r.t. $C_{U_{down},F_{down}}$ together with a subpath of $P$ would form a forward path right to $P$ in $C_{U,F}$, again a contradiction.
{\bf Case 2: $P$ is an up-cross path.} Then the up-strip $C_{U_{up},F_{up}}$ has no forward hanging path simply because $U_{up}$ is a subpath of $U$. It is also not hard to see that $C_{U_{up},F_{up}}$ has no down-cross path. Suppose not. Then we have a down-cross path $Q$ that starts from some internal vertex $x$ of $U_{up}$ and ends at some vertex $y$ in $F_{up}$. We have two cases: (1) if $y\in F$, then the path $Q$ is also a down-cross path in $C_{U,F}$ and (2) if $y\in P$, then the path $Q$ together with a subpath of $P$ form a forward hanging path in $C_{U,F}$. Both cases lead to contradiction.
Next consider the down-strip $C_{U_{down},F_{down}}$. The ceiling $U_{down}$ consists of the slicing path $P$ and a subpath of $U$. Suppose there is a down-cross path $Q$ in $C_{U_{down},F_{down}}$. Then we know that $Q$ must go from $P$ to $F_{down}=U$. But, this means that we can form a forward path right to $P$ by concatenating a subpath of $P$ with $Q$, which contradicts the fact that $P$ is the right-most slicing path.
Suppose there is a forward hanging path $Q$ in $C_{U_{down},F_{down}}$. Then it must start from some vertex $v$ in $P$ and ends on either $P$ or the subpath of $U$ in $U_{down}$. But, both cases imply that there exists a slicing path $P'$ that lie right to $P$, which contradicts our choice of $P$. \qed
\subsection{Linear Time Implementation} \label{sec:strip-decomposition:linear-time}
Now we present a linear-time implementation of our strip-slicing decomposition algorithm and thus prove Lemma~\ref{lem:linear-time-decomposition}.
The algorithm is similar to the one with quadratic running-time except that we keep the information of vertices that have already been scanned. Here the recursive procedure has an additional parameter $v^*$, which is a pointer to a vertex on the floor $F$ that we use as a starting point in finding the slicing path $P$.
To be precise, our algorithm again consists of two steps. Let us denote the input strip by $C_{U^*,F^*}$ and denote its start and end vertices by $s'$ and $t'$, respectively. At the initialization step, we remove all arc leaving $C_U^*,F^*$ and add an auxiliary ceiling $(s',x',t')$ to enclose the input strip ($x'$ is a dummy vertex). Then we set the starting point $v^*:=s'$ and call the recursive procedure with parameter $C_{U,F}$ (the modified strip) and $v^*$.
In the recursive procedure,
we first order vertices on the floor $F$ by $v_1,\ldots,v_{\ell}$ (note that $v_1=s'$ and $v_{\ell}=t'$). Let $i^*$ be the index such that $v_{i^*}=v^*$. For each $i=i^*$ to $\ell$ in this order, we find a slicing path starting from $v_i$ (using the right-first-search algorithm). If we cannot find a slicing path starting from $v_i$, then we remove all the inner vertices and arcs scanned by the search algorithm and iterate to the next vertex (i.e., to the vertex $v_{i+1}$). Otherwise, we found a slicing path $P$ and exit the loop.
Note that the loop terminates under two conditions. Either the search algorithm finds no slicing path or it reports that there exists a slicing path $P$ starting from a vertex $v_i$. For the former case, we add $C_{U,F}$ to the collection of strips $\calS$. For the latter case, we remove all the inner vertices and arcs scanned by the search algorithm except those on $P$, and we then slice the strip $C_{U,F}$ into the up-strip $C_{U_{up},F_{up}}$ and the down-strip $C_{U_{down},F_{down}}$. Finally, we make two calls to the recursive procedure with parameters $(C_{U_{up},F_{up}},v_i)$ and $(C_{U_{down},F_{down}},v_i)$, respectively. In particular, we recursively run the procedure on the up and down strips, and we pass the vertex $v_i$ as the starting point in finding slicing paths.
\subsubsection{Implementation Details}
Here we describe some implementation details that affect the running-time of our algorithm.
\paragraph*{Checking whether a vertex is in the floor or ceiling.} Notice that checking whether a vertex is in the the floor $F$ or the ceiling $U$ can be an issue since the up-strip and the down-strip share boundary vertices.
A trivial implementation would lead to storing the same vertex in $O(n)$ data structures, and the algorithm ends up running in $O(n^2)$-time.
We resolve this issue by using a coloring scheme. We color vertices on the floor $F$ by red and the vertices on the ceiling $U$ by green. (Note that we color only the internal vertices of $U$ since $U$ and $F$ share the start and end vertices). We color other vertices by white. Now we can easily check if a vertex is in the floor or ceiling by checking its color. When we make a recursive call to the up-strip and the down-strip. We have to maintain the colors on the slicing path $P$ to make sure that we have a right coloring. More precisely, when we make a recursive call to the up-strip we color internal vertices of $P$ by red (since they will be vertices on the floor $F_{up}$) and then flip their colors to green later when we make a recursive call to the down-strip (since they will be vertices on the ceiling $U_{down}$). Observe that internal vertices of the slicing path $P$ are internal vertices of the strip $C_{U,F}$, which will be boundary vertices of the up and down strips in the deeper recursive calls. As such, any vertices will change colors at most three times (from white to green and then to red).
\paragraph*{Checking if a path is a forward path.} Another issue is when we search for a slicing path $P$ and want to decide whether the path we found is a forward path or a backward path. We simply solve this issue by removing all the incoming arcs of vertices $v_{i'}$ for all $0 \leq i'< i$. This way any path we found must be either a forward path or an up-cross path.
\paragraph*{Slicing a Strip.} Observe that we may create $O(n)$ strips (and thus have $O(n)$ floors and ceilings) during the run of the decomposition algorithm. However, we cannot afford to store floors and ceilings of every strip separately as they are not arc disjoint. Otherwise, just storing them would require a running time of $O(n^2)$.
To avoid this issue, we keep paths (either floor or ceiling) as doubly linked lists.
Consequently, we can store $C_{U,F}$ by referencing to the first and last arcs of the floor (respectively, ceiling).
Moreover, since we store paths as doubly linked list, each {\em cut-and-join} operation can be done in $O(1)$ time.
\paragraph*{Maintaining the search to be inside $C_{U,F}$.} As we may have an arc going leaving $C_{U,F}$, we have to ensure that the right-first-search algorithm would not go outside the strip. (Note that our algorithm does not see the real drawing of $C_{U,F}$ on a plane and thus cannot decide whether an arc is in a strip or not.) To do so, we temporary remove the arcs leaving $C_{U,F}$ and put them back after the procedure terminates. Observe that the inner parts of up-strip $C_{U_{up},F_{up}}$ and $C_{U_{down},F_{down}}$ are disjoint. Thus, each arc leaving $C_{U,F}$ is removed and put back at most once.
\subsubsection{Correctness} \label{sec:strip-decomposition:linear-time:correctness}
For the correctness, it suffices the show that the slicing path $P$ that we found in every step is exactly the same as what we would have found by the quadratic-time algorithm. We prove this by reverse induction. The base case, where we call the procedure on the initial strip $C_{U^*,F^*}$, holds trivially.
For the inductive step, we assume the that both implementations found the same slicing paths so far.
Now if the recursive procedure found no slicing path, then we are done. Otherwise, we found a slicing path $P$. Let us order vertices on the floor $F$ by $v_1,\ldots,v_{\ell}$, and let $v_i$ be the start vertex of $P$. First, consider the up-strip $C_{U_{up},F_{up}}$. Clearly, $C_{U_{up},F_{up}}$ has no up-cross path starting from any vertex $v_{i'}$ with $0 \leq i' < i$. We also know that $C_{U_{up},F_{up}}$ has no forward path $Q$ starting from $v_{i'}$; otherwise, the path $Q$ is either a forward path in $C_{U,F}$ starting at $v_{i'}$ (if $Q$ starts and ends on $U$) or we can form another slicing path $P'$ starting from $v_{i'}$ by concatenate $Q$ with a subpath of $P$.
Moreover, all the vertices that scanned by running the right-first-search algorithm from $v_{i}$ must be enclosed in the down-strip. Thus, we conclude that removing all scanned vertices (those that are not on $P$) does not affect the choice of slicing paths in the up-strip.
Next consider the down-strip. In any case, we know that the slicing path $P$ becomes a subpath of the ceiling of $C_{U_{down},F_{down}}$. Clearly, any path $Q$ from $P$ to $F_{down}$ would either form a forward path $P'$ lying right to $P$ or would imply that $C_{U,F}$ has a down-cross path (this is the case when $Q$ is a subpath of a longer path that goes from $U$ to $F_{down}$), which are contradictions in both cases. Thus, we can remove every vertex that is scanned by the right-first-search algorithm and is not on the slicing path $P$ without any effect on the choices of slicing paths in deeper recursive calls.
To conclude, our linear-time and quadratic-time implementations have the same choices of slicing paths and must produce the same outputs. The correctness of our algorithm is thus implied by Lemma~\ref{lem:quadratic-time-decomposition}.
\subsubsection{Running Time Analysis} \label{sec:strip-decomposition:linear-time:running-time-analysis}
Clearly, the initialization step of the decomposition algorithm runs in linear-time. Thus, it suffices to consider the running time of the recursive procedure.
We claim that each arc is scanned by our algorithm at most $O(1)$ times, and each vertex is scanned only when we scan its arcs. This will immediately imply the linear running time. We prove this by induction.
First, note that we remove all the internal vertices and arcs that are scanned by the search algorithm but are not on the slicing path $P$ (if it exists), which means that these vertices and arcs are scanned only once throughout the algorithm.
Thus, if we found no slicing path, then the decomposition runs in linear time on the number of arcs in the strip.
Otherwise, we found a slicing path $P$ starting at some vertex $v_i\in F$.
The recursive procedure slices the strip $C_{U,F}$ in such a way that the two resulting strips $C_{U_{up},F_{up}}$ and $C_{U_{down},F_{down}}$ have no inner parts in common.
Thus, we can deduce inductively that the running time incurred by scanning inner arcs and vertices of the two strips are linear on the number of arcs in the strips. Moreover, since we pass $v_i$ as the starting point of iterations in deeper recursive calls, we can guarantee that any vertices $v_{i'}\in F$ with $i < i'$ (those that appear before $v_i$) will never be scanned again.
We note, however, that we may scan the vertex $v_i$ multiple times because some arcs leaving $v_i$ that lie on the left of $P$ have never been scanned. Nevertheless, we can guarantee that the number of times that we scan $v_i$ is at most the number of its outgoing arcs. To see this, first any arc that does not lead to a slicing path will be removed. For the arc that leads to a slicing path, we know that it will become an arc on the boundaries of up and down strips and thus will never be scanned again. In both cases, we can see that the number of times we scan $v_i$ is upper bounded by the number of its outgoing arcs.
Therefore, the running time of our algorithm is linear on the number of arcs in the input strip, completing the proof of Lemma~\ref{lem:linear-time-decomposition}.
\qed \section{The Outside Case: Determining useless arcs when $s$ is not
in the component} \label{sec:st-outside}
In this section, we describe an algorithm that determines all useful arcs of a strongly connected component $C$ when the source $s$ is in the infinite face.
We first outline our approach. Consider a strongly connected component $C$.
If the boundary of $C$ is not a cycle, then the strong connectivity implies that we can extend some path on the boundary to form a clockwise cycle, which contradicts to our initial assumption that the embedding has no clockwise cycle.
If all arcs in the boundary of $C$, denoted by $Q$, are useless, then no arcs of $C$ are useful. Otherwise, we claim that some arcs of $Q$ are useful and some are useless. Moreover, these arcs partition $Q$ into two paths $Q_1$ and $Q_2$ such that $Q_1$ is a useful path and all arcs of $Q_2$ are useless.
\begin{lemma} \label{lem:one-useful} Consider the boundary of a strongly connected component $C$. Let $Q$ be the cycle that forms the boundary of $C$. Then either all arcs of $C$ are useless or there are non-trivial paths $Q_1$ and $Q_2$ such that \begin{itemize}
\item $E(Q)=E(Q_1)\cup E(Q_2)$.
\item $Q_1$ is a useful path.
\item All arcs of $Q_2$ are useless. \end{itemize} Moreover, there is a linear-time algorithm that computes $Q_1$ and $Q_2$. \end{lemma} \begin{proof}
First, we claim that all arcs of $C$ are useless
if and only if $Q$ has no entrance or has no exit.
If $Q$ has no entrance or has no exit, then
it is clear that all arcs of $C$ are useless
because there can be no simple $s,t$-path using an edge of $C$.
Suppose $Q$ has at least one entrance $u$ and
at least one exit $v$.
We will show that $Q$ has both useful and useless arcs, which
can be partitioned into two paths as in the statement of the lemma.
Observe that no vertices of $Q$ can be both entrance and exit
because of the degree-three assumption.
(Any vertex that is both entrance and exit must have degree at least
four, two from $Q$ and two from the entrance and the exit paths.)
For any pair of entrance and exit $u,v$,
let $Q_{uv}$ be the $u,v$-subpath of $Q$.
We claim that $Q_{uv}$ is a useful path.
To see this, we apply
Lemma~\ref{lem:disjoint-entrance-exit-paths}.
Thus, we have an $s,u$-path $P_u$ and
a $v,t$-path $P_v$ that are vertex disjoint.
Moreover, $P_u$ and $P_v$ contain no vertices of $Q$ except
$u$ and $v$, respectively.
Thus, $R=P_u\cdot Q_{uv}\cdot P_v$ form a simple $s,t$-path,
meaning that $Q_{uv}$ is a useful path.
Notice that $Q$ has no useless arc only if
$Q$ has two distinct entrances $u_1,u_2$ and two
distinct exits $v_1,v_2$ that appear in $Q$ in interleaving
order $(u_1,v_1,u_2,v_2)$.
Now consider a shortest $s,u_1$-path $P_{u_1}$,
a shortest $s,u_2$-path $P_{u_2}$,
and the $u_1,u_2$-subpath $Q_{u_1,u_2}$ of $Q$.
These three paths together encloses any path that leaves
the component $C$ through an exit $v_1$.
Thus, any $v_1,t$-path must intersect either $P_{u_1}$ or
$P_{u_2}$, contradicting
Lemma~\ref{lem:disjoint-entrance-exit-paths}.
Consequently, a sequence of entrances and exits
appear in consecutive order on $Q$.
Let us take the first entrance $u^*$ and the last exit $v^*$.
We know from the previous arguments that any pair of entrance
and exit yields a useful subpath of $Q$.
More precisely, the $u^*,v^*$-subpath $Q_1$ of $Q$ must be
useful and must contain all the entrances and exits
because of the choices of $u^*$ and $v^*$.
Consider the other subpath -- the $v^*,u^*$-subpath $Q_2$ of $Q$.
The only entrance and exit on $Q_2$ are $u^*$ and $v^*$,
respectively.
Thus, any path $s,t$-path $P$ that contains an arc $e$ of $Q_2$ must
intersect with the path $Q_1$.
But, then the path $P$ together with the boundary $Q$
enclose the region of $C$ that has no exit.
Thus, $P$ cannot be a simple path.
It follows that no arcs of $Q_2$ are useful.
To distinguish the above two cases, it suffices to compute all the
entrances and exits of $Q$ using a standard graph searching
algorithm. The running time of the algorithm is linear, and we need
to apply it once for all the strongly connected components.
This completes the proof. \end{proof}
\subsection{Algorithm for The Outside Case} \label{sec:algo-outside-case}
Now we present our algorithm for the outside case. We assume that some arcs in the boundary of $C$ are useful and some are useless.
Our algorithm decides whether each arc in $C$ is useful or useless by decomposing $C$ into a collection of minimal strips using the algorithm in Section~\ref{sec:strip-decomposition}.
Then any arc enclosed inside a minimal strip is useless (only boundary arcs are useful). Thus, we can determine the usefulness of arcs in each strip.
Observe, however, that the component $C$ is not a strip because it is enclosed by a cycle instead of two paths that go in the same direction.
We transform $C$ into a strip as follows. First, we apply the algorithm as in Claim~\ref{lem:one-useful}. Then we know the boundary of $C$, which is a cycle consisting of two paths $Q_1$ and $Q_2$, where $Q_1$ is a useful path and $Q_2$ is a useless path. Let $F^*=Q_1$, and call it the {\em lowest-floor}. Note that $F^*$ starts at an entrance $s_C$ and ends at an exit $t_C$. We add to $C$ a dummy path $U^* = (s_C,u^*,t_C)$, where $u^*$ is dummy vertex, and call $U^*$ the {\em top-most ceiling}, which is a dummy ceiling. This transforms $C$ into a strip, denoted by $C_{U^*,F^*}$. (See Figure~\ref{fig:s-outside-init-strip})
\begin{figure}
\caption{The initial strip before the decomposition.}
\label{fig:s-outside-init-strip}
\end{figure}
Now we are able to decompose $C_{{U^*},{F^*}}$ into a collection $\calS$ of minimal strips by calling the strip-slicing decomposition on $C_{{U^*},{F^*}}$.
Since $F^*$ is a useful path, we deduce from Corollary~\ref{cor:ext-forward-path-is-useful} that all the strips $C_{U,F}\in \calS$ are such that both $U$ and $F$ are useful paths.
Next we show that no inner arcs of a strip $C_{U,F}\in\calS$ are useful, which then implies only arcs on the boundaries of strips computed by the strip-slicing decomposition are useful. Therefore, we can determine all the useful arcs in $C$ in linear-time.
\begin{lemma} \label{lem:outside-no-useful-backward} Consider a strip $C_{U,F}\in\calS$ computed by the strip-slicing decomposition algorithm. Then no inner arc $e$ of $C_{U,F}$ is useful. More precisely, all arcs in $e\in E(C_{U,F})- E(U\cup F)$ are useless. \end{lemma} \begin{proof} Suppose for a contradiction that there is an inner arc $e$ of $C_{U,F}$ that is useful. Then there is an $s,t$ path $P$ containing $e$. We may also assume that $P$ contains a link $Q$. That is, $Q$ starts and ends at some vertices $u$ and $v$ in $V(U\cup F)$, respectively, and $Q$ contains no arcs of $E(U\cup F)$. By Lemma~\ref{lem:linear-time-decomposition}, $Q$ must be a backward path. Hence, it suffices to show that any backward path inside the strip $C_{U,F}$ is useless.
By the definition of the backward path $Q$, $v$ must appear before $u$ on $F$, and by Lemma~\ref{lem:properties-strips}, $F$ is a useful path, meaning that there is an $s,t$-path $R$ containing $F$. Let $s'$ and $t'$ be the start and end vertices of the strip $C_{U,F}$ (i.e., $s'$ and $t'$ are the common start and end vertices of the paths $U$ and $F$, respectively). Then we have four cases.
\begin{itemize} \item {\bf Case 1: $v=s'$ and $u=t'$.}
In this case, the path $U$ and $Q$ form a clockwise cycle,
a contradiction.
\item {\bf Case 2: $v=s'$ and $u\neq t'$.}
In this case, $v$ has degree three in $Q\cup U\cup F$.
Thus, the path $R\supseteq F$ must start at the same vertex as $F$;
otherwise, $v$ would have degree at least four.
This means that $s=s'$, but then $P$ could not be a simple
$s,t$-path because $s$ must appear in $P$ at least twice,
a contradiction.
\item {\bf Case 3: $v\neq s'$ and $u=t'$.}
This case is similar to the former one.
The vertex $u$ has degree three in $Q\cup U\cup F$.
Thus, the path $R\supseteq F$ must end at the same vertex as $F$;
otherwise, $u$ would have degree at least four.
This means that $t=t'$, but then $P$ could not be a simple
$s,t$-path because $t$ must appear in $P$ at least twice,
a contradiction.
\item {\bf Case 4: $u\neq t'$ and $v\neq s'$.}
Observe that $P$ must enter and leave $C_{U,F}$ on
the right of $F$ (and thus $R$)
at some vertices $a$ and $b$, respectively.
We may assume wlog that the $a,u$-subpath
and the $v,b$-subpath of $P$ are contained in $R$.
Moreover, it can been seen that
$u$ and $v$ appear in different orders on
the path $R$ (that contains $F$) and
the path $P$ (that contains $Q$).
Thus, $P$ must intersect either
the $s,b$-subpath of $R$, say $R_{s,b}$, or
the $a,t$-subpath of $R$, say $R_{a,t}$.
(Note that $R$ contains no inner vertices of $Q$
by the definition of a backward path.)
Suppose $P$ intersects $R_{a,t}$ at some vertex $x$,
and assume that $x$ is the last vertex of $P$ in
the intersection of $P$ and $R_{a,t}$.
Since $P$ enters $R$ at the vertex $a$ from the right,
the $x,a$-subpath of $P$ must lie on the right of $R$.
But, then the union of the $x,a$-subpath of $P$ and
the $a,x$-subpath of $R$ forms a clockwise cycle,
a contradiction.
(See Figure ~\ref{fig:outside-no-usefull-backward-4-1}
for illustration.)
Now suppose $P$ intersects $R_{s,b}$ at some vertex $y$,
and assume that $y$ is the first vertex of $P$ in
the intersection of $P$ and $R_{s,b}$.
Since $P$ leaves $R$ at the vertex $b$ from the right,
the $b,y$-subpath of $P$ must lie on the right of $R$.
But, then the union of the $b,y$-subpath of $P$ and
the $y,b$-subpath of $R$ forms a clockwise cycle,
again a contradiction. \end{itemize} Therefore, all backward paths inside the strip $C_{U,F}$ are useless and so do inner arcs of $C_{U,F}$. \end{proof}
\begin{figure}
\caption{An illustration of Case~4 in the proof of
Lemma~\ref{lem:outside-no-useful-backward}}
\label{fig:outside-no-usefull-backward-4-1}
\end{figure}
\section{The Inside Case: Determining useful arcs of a strongly
connected component when $s$ is inside of the component} \label{sec:s-inside}
In this section, we describe an algorithm for the case that the source vertex $s$ is in the strongly connected component $C$.
In this case, the source vertex $s$ is enclosed in some face $f_s$ in the component $C$. It is possible that $s$ is also enclosed nestedly in other strongly connected components (see Figure~\ref{fig:component-example}). We will perform a simple reduction so that (1) the source $s$ is enclosed inside the component $C$, (2) $s$ has only one outgoing arc, and (3) every vertex except $s$ (and $t$) has degree three: First, we contract the maximal component containing $s$ inside $f_s$ into $s$. Now the source $s$ has degree more than one in a new graph. We thus add an arc $s's$ and declare $s'$ as a new source vertex. Then we replace $s$ by a binary tree on $d$ leaves to maintain the degree-three requirement. See Figure~\ref{fig:inside-reduction}.
\begin{figure}
\caption{A reduction for simplifying the source structure for the
inside case.}
\label{fig:inside-reduction}
\end{figure}
\subsection{Overview}
We decompose the problem of determining the usefulness of component $C$ into many subproblems and will deal with each subproblem (almost) independently.
The basis of our decomposition is in slicing the component with the {\em lowest-floor path} $F^*$ and the {\em top-most ceiling} $U^*$. These two paths then divide the component $C$ into two parts: (1) the primary strip and (2) the open strip. Figure~\ref{fig:floor-of-inside-case} illustrates the paths $F^*$ and $U^*$, and Figure~\ref{fig:inside-strip-types} shows the schematic view of the primary and the open strips. All types of strips and components will be defined later. The lowest-floor is constructed in such a way that it includes all the exits of a component $C$. We thus assume that the sink $t$ is a vertex on the boundary of $C$, and there is no other exit.
\begin{figure}
\caption{the lowest-floor path $F^*$ and the top-most ceiling $U^*$
in the inside case.}
\label{fig:floor-of-inside-case}
\end{figure}
\begin{figure}
\caption{The types of major strips in the inside case.}
\label{fig:inside-strip-types}
\end{figure}
\subsubsection{The lowest-floor path $F^*$.} \label{sec:s-inside:lowest-floor} We mainly apply the algorithm from Section~\ref{sec:st-outside} to the primary strip defined by two paths $F^*$ and $U^*$, which may not be arc-disjoint. The path $F^*$ is the {\em lowest-floor path} defined similarly to the outside case: We compute $F^*$ by finding the right-most path $P$ from $s$ to $t$ (hence, $P$ is an $s,t$-path).
We recall that the sink $t$ is placed on the unbounded face of the planar embedding. Thus, $t$ must lie outside of the region enclosed by $C$. This means that the right-most path $P$ has to go through some exit $t'$ on the boundary of $C$. Although we know that the right-most path goes from $t'$ to $t$ directly, we detour the path $P$ along the boundary to include all the exits, which is possible since the source $s$ is enclosed inside the component. See Figure~\ref{fig:finding-lowest-floor-path} for illustration.
\begin{figure}
\caption{How to find $F^*$: the path $P$ is shown in solid red. The
extension on the boundary is shown as a dashed line.}
\label{fig:finding-lowest-floor-path}
\end{figure}
\subsubsection{Flipped paths.} Unlike the outside case, the path $F^*$ in the inside case does not divide the component $C$ into two pieces because the source vertex $s$ is inside some inner face of $C$.
This leaves us more possibilities of paths that we have not encountered in the outside case, called ``flipped paths''.
These are paths that goes from the left-side of the floor $F^*$ to the right-side (or vice versa). (Please recall the terminology of left and right directions in Section~\ref{sec:prelim:left-right}, and recall the types of paths in Section~\ref{sec:prelim:forward-backward}.)
Fortunately, the reverse-flipped and hanging paths do not exist because of our choice of the lowest floor path $F^*$.
\begin{lemma} \label{lem:no-reserse-flip-nor-hanging} Consider the construction of the lowest floor path $F^*$ as in Section~\ref{sec:s-inside:lowest-floor}. Then there is no reverse-flipped nor hanging paths w.r.t. $F^*$. \end{lemma} \begin{proof} Consider a connected component $C$ that encloses the source $s$. We recall that $F^*$ is constructed by finding the right-most $s,t$-path and then detouring along the boundary of $C$.
Thus, $F^*$ is a union of two subpaths: (1) the path $F_1$ that is a subpath of the right-most $s,t$-path and (2) the path $F_2$ that is a subpath of the boundary of $C$.
First, we rule out the existence of a hanging path. Assume to the contrary that there is a path $P\subseteq C$ that leaves $F^*$ to the right and enters $F^*$ again from the right. Then clearly $P$ must start and end on the first subpath $F_1$ of $F^*$ (otherwise, $P$ would have left the boundary). But, this would contradict the fact that $F_1$ is a subpath of the right-most $s,t$-path.
Next, we rule out the existence of a reverse-flipped path. Assume to the contrary that there is a path $P$ that leaves $F^*$ to the right and enters $F^*$ again from the left. We assume the minimality of such path (thus, $P$ contains no internal vertices of $F^*$). Then we know that $P$ must start from $F_1$. (otherwise, $P$ would have left the boundary). If $P$ is a backward path, then the path $P$ together with a subpath of $F^*$ forms a clockwise cycle, a contradiction (we would have gotten ride of them in the preprocessing step).
If $P$ is a forward path, then it would contradict the fact that $F_1$ is a subpath of the right-most path since we can traverse along the path $P$ (which is right to $F_1$) to get an $s,t$-path. \end{proof}
\subsubsection{Primary Strip.} The {\em primary strip} is defined by two paths, which may not be arc-disjoint, namely $F^*$ and $U^*$. The path $F^*$ is the lowest-floor path. The path $U^*$ is the {\em top-most ceiling}, which is an $s,t$-path such that the strip $C_{U^*,F^*}$ encloses all (non-flipped) forward paths (w.r.t. $F^*$). The path $U^*$ is essentially the left-most path from $s$ to $t$. Section~\ref{sec:s-inside:init} describes how we find $U^*$ given $F^*$.
\subsubsection{Open Strip.} The {\em open strip} $\hat{C}$ is defined to be everything not in the inner parts of the primary strip. See Figure~\ref{fig:inside-path-types} for illustration.
\begin{figure}
\caption{The types of paths in the open strip.}
\label{fig:inside-path-types}
\end{figure}
To deal with arcs in $\hat{C}$, we first characterize the first set of useless arcs: ceiling arcs. An arc $e\in\hat{C}=(u,v)$ is a {\em
ceiling arc} if $u$ can be reachable only from paths leaving $F^*$ to the left and $v$ can reach $t$ only through paths entering $F^*$ from the left as well. These arcs form ``ceiling components''. (Note that paths that form ceiling components are all backward paths because all the forward paths w.r.t. $F^*$ have been enclosed in $C_{U^*,F^*}$.) We shall prove that these paths are useless in Lemma~\ref{lem:backward-comp-useless}. Therefore, as a preprocessing step, we delete all the ceiling arcs from $\hat{C}$.
To deal with the remaining arcs in $\hat{C}$, we first find strongly connected components $H_1,H_2,\ldots$ in $\hat{C}$.
The arcs outside strongly connected components then form a directed acyclic graph $\hat{D}$ that represents essentially the structures of all flipped paths. We can now process each component $H_i$ using the outside-case algorithm only if we know which adjacent arcs in $\hat{D}$ are valid entrances and exits. For arcs in $\hat{D}$ that belong to some flipped path, we can use a dynamic programming algorithm to compute all reachability information needed to determine their usefulness.
Our algorithm can be described shortly as follows. \begin{itemize} \item {\bf Initialization.}
We compute the lowest-floor path $F^*$ and the top-most ceiling
$U^*$, thus forming the primary strip $C_{U^*,F^*}$ and the open
strip $\hat{C}$. \item {\bf The open strip (1): structures of flipped paths.}
We find all strongly connected components in $\hat{C}$ and collapse
them to produce the directed acyclic graph $\hat{D}$.
We then compute the reachability information of arcs in $\hat{D}$ and
use it to determine the usefulness of arcs in $\hat{D}$.
\item {\bf The open strip (2): process each strongly
connected component.}
Then we apply the outside-case algorithm to each strongly connected
component in $\hat{C}$. \item {\bf The primary strip.}
We determine the usefulness of arcs in the primary strip. \end{itemize} In the following subsections, we describe how to deal with each part of $C$.
\subsection{Initialization: Finding the Primary and Open Strips} \label{sec:s-inside:init}
In this section, we discuss the initialization step of our algorithm, which computes the primary and the open strips. In this step, we also compute the reachability information of flipped paths, which will be used in the later part of our algorithm.
As discussed in the previous section, we compute the lowest-floor path $F^*$ by finding the right-most $s,t$-path, which is unique and well-defined because $s$ has a single arc leaving it. The path $F^*$ is contained in $C$ because we assume that $t$ is an exit vertex on the boundary of $C$.
The top-most ceiling $U^*$ can be computed by simply computing the left-most non-flipped path w.r.t. of $F^*$ from $s$ to $t$. This can be done in linear time by first (temporary) removing every arc on the right-side of $F^*$ and then computing the left-most path from $s$ to $t$.
Consequently, we have the primary strip $C_{U^*,F^*}$, we claim that it encloses all forward paths (w.r.t. $F^*$).
\begin{lemma} \label{lem:main-strip-encloses-all-forward} Every forward path w.r.t. $F^*$ is enclosed in $C_{U^*,F^*}$. \end{lemma} \begin{proof} Suppose to the contrary that there exists a forward path $P$ not enclosed by $C_{U^*,F^*}$. Then $P$ must intersect with $U^*$ or $F^*$. If it is the former case, then $P$ must have a subpath $Q$ that leaves and enters $U^*$ on the left (this is because $U^*$ is on the left of $F^*$). We choose a subpath $Q$ in which the start vertex of $Q$, say $u$, appears before its end vertex, say $v$, on $U^*$. Such path $Q$ must exist because, otherwise, $P$ would have a self-intersection inside $C_{U^*,F^*}$.
We may further assume that $Q$ is a minimal such path, which means that all the arcs of $Q$ lie entirely on the left of $U^*$. But, then $U^*$ would not be the left-most non-flipped path (w.r.t. $F^*$) because the left-first-search algorithm would have followed $Q$ instead of the $u,v$-subpath of $U^*$, a contradiction.
The case that $P$ intersects $F^*$ is similar, and such path would contradict the fact that $F^*$ is the right-most path. Therefore, all the forward paths (w.r.t. $F^*$) must be enclosed in $C_{U^*,F^*}$. \end{proof}
Next consider the remaining parts of the component, which forms the open strip $\hat{C}=C- C_{U^*,F^*}$. We have as a corollary of Lemma~\ref{lem:main-strip-encloses-all-forward} that there is no simple $s,t$-path in the open strip that leaves $F^*$ from the right.
\begin{corollary} \label{cor:no-right-path} No simple $s,t'$-path that lies on the right of $F^*$. \end{corollary} \begin{proof} Suppose there is an $s,t$-path $P$ in the open strip that leaves $F^*$ from the right. Then we have three cases.
First, if $P$ also enters $F^*$ from the right, then $P$ cannot be a forward-path and cannot have any forward subpath by Lemma~\ref{lem:main-strip-encloses-all-forward}. Thus, $P$ must have a self-intersection, contradicting to the fact that $P$ is a simple path.
Second, if $P$ enters $F^*$ from the left, then we can find a $u,v$-subpath $Q$ of $P$ such that $u$ appears before $v$ in $F^*$ and $Q$ is arc-disjoint from $F^*$. But, then the right-first-search algorithm would have followed $Q$ instead of the $u,v$-subpath of $F^*$, a contradiction.
Otherwise, if $P$ never enters $F^*$, then it must go directly to $t$. But, then the right-first-search algorithm would have followed the subpath of $P$ that leaves $F^*$, a contradiction. \end{proof}
\subsection{Working on the Open Strip (1): Dealing with Arcs in Flipped Paths} \label{sec:arcs-in-flipped-paths}
To determine the usefulness of arcs on the DAG $D$, which is formed by contracting strongly connected components $H_1,H_2,\ldots$ in the open strip, it suffices to check if it is contained in some useful flipped path w.r.t. $F^*$.
We first prove the characterization of useful flipped paths. Section~\ref{sec:s-inside:reachability} describes how we can check if an arc belongs to any useful path by using reachability information.
Our first observation is that every forward flipped path is useful.
\begin{lemma} \label{lem:flip-forward-is-useful} Every forward flipped path w.r.t. $F^*$ is useful. \end{lemma} \begin{proof} Consider any forward flipped path $P$. By the definition of the forward flipped path, $P$ shares no arcs with $F^*$. Moreover, $P$ starts at some vertex $v\in F^*$ and ends at some vertex $w\in F^*$ such that $v$ appears before $w$ in $F^*$. Note that $F^*$ is a path that goes from $s$ to the vertex $t'$ in the boundary of the component $C$. We extend $F^*$ to a simple $s,t$-path $R$ by adding a $t',t$-path that is not in the component $C$.
We then replace the $u,v$-subpath of $R$ by $P$, thus getting a new path $R'$, which is a simple path because $R$ shares no arcs with $P$ and contains no vertices of $V(P)-\{v,w\}$. Thus, $R'$ is a simple $s,t$-path, implying that $P$ is useful. \end{proof}
If an arc is not contained in any forward flipped path, it must be in some backward flipped path. In most cases, a backward flipped path is useless except when there exists an exit that provides an alternative route to $t$.
Consider the primary strip, which is enclosed by the lowest-floor $F^*$ and the top-most ceiling $U^*$. There are some arcs that $U^*$ and $F^*$ have in common. If we remove all of these arcs, then the remaining graph consists of weakly connected components, which we call humps. Formally, a {\em hump} is a strip obtained from the primary strip by removing $F^*\cap U^*$. We can also order humps by the ordering of their first vertices on the floor $F^*$. The structure of humps is important in determining the usefulness of a flipped path. See Figure~\ref{fig:useful-flipped-backward-path}.
\begin{figure}
\caption{An example of flipped backward paths $P_1$ (useless) and
$P_2$ (useful). $F^*$ is shown in red and $U^*$ is shown in blue.
There are 5 humps (shown in gray).}
\label{fig:useful-flipped-backward-path}
\end{figure}
The lemma below shows the conditions when arcs in a backward flipped path can be useful.
\begin{lemma} \label{lem:useful-backward-flip-paths} Let $P$ be any backward flipped-path such that $P$ starts at a vertex $u\in F^*$ and ends at a vertex $v\in F^*$.
Then $\hat{P}=P\cap\hat{C}$ (the subpath of $P$ in the open strip) is useful iff $P$ ends at some hump $H$ (thus, $H$ contains $v$) such that either \begin{enumerate} \item both $u$ and $v$ are in the same hump $H$, and
$v$ is not the first vertex in $H$. \item The vertex $u$ is not in the hump $H$, and
there is an exit after $v$ in the hump $H$. \end{enumerate} \end{lemma} \begin{proof}
First, we show that if the end vertex $v$ of $P$
is not in any hump, then $\hat{P}=P\cap\hat{C}$ is useless.
We assume wlog that the start vertex $u$ of $P$ is also
not contained in any hump;
otherwise, we contract all vertices in the hump
(that contains $u$) to the vertex $u$.
Thus, $P=\hat{P}$.
It suffices to show that there is no $s,u$-path $J$
that is vertex disjoint from $P$, which will imply that
$P$ cannot be a useful path.
To see this, assume a contradiction that there is an $s,u$-path $J$
that is vertex disjoint from $P$.
Then $J$ has to leave $F^*$ at some vertex $x$
and then enters $F^*$ again at some vertex $y$
(it is possible that $y=u$) in such a way that
$x,y,u,v$ appear in the order $(x,v,y,u)$ on $F^*$.
Let $J'$ be the $x,y$-subpath of $J$.
If $J'$ is contained in the primary strip,
then $J'$ would induce a hump containing $v$, a contradiction.
So, we assume that $J'$ is not contained in the primary strip.
If $J'$ leaves $F^*$ from the right,
then $J'$ cannot enter $F^*$;
otherwise, it would contradict the fact that $F^*$ is
the right-most path.
Hence, we are left with the case that $J'$ leaves $F^*$
from the left and then enters again from the right.
Since $J$ and $P$ are vertex-disjoint (and so are $J'$ and $P$),
$J'$ cannot enter $F^*$ at any vertex that appears after $v$
(including $u$ and $y$) because, otherwise, $J'$ has to cross the
path $P$.
Thus, we again have a contradiction.
Consequently, since there is no path from $s$ that can reach
$u$ or the hump containing $u$ without intersecting with $\hat{P}$,
the path $\hat{P}$ cannot be a useful path.
Next consider the case that $v$ is contained in some hump $H$.
Let $w$ be a vertex in $P\cap U^*$, i.e., $w$ be a vertex in the
ceiling intersecting with $P$.
Also, let $\hat{P}$ be $P\cap\hat{C}$.
For the backward direction, we construct
a useful path $Q$ containing $\hat{P}$ by taking the union of the
prefix of the top-most ceiling $U^*$ up to $w$, $\hat{P}$ and
the subpath $P_F$ of the floor $F^*$ after $v$.
In Case (1), $P_F$ is the suffix of $F^*$ after $v$, and
in Case (2), $P_F$ starts from $v$ and ends at the exit.
It can be seen that $Q$ is a simple path from $s$ to an exit,
meaning that $Q$ is a useful path and so is $\hat{P}$.
Now consider the forward direction.
We proceed by contraposition.
There are two subcases we have to consider:
(i) $u$ and $v$ are in the same hump $H$, but
$v$ is the first vertex in the hump and
(ii) $u$ is not in the hump $H$ and there is no exit after
$v$ in $H$.
In Subcase (i), any useful path $Q$ containing $\hat{P}$ whose
prefix up until $\hat{P}$ is entirely in the primary strip has to go
through $v$; therefore, it cannot be simple.
Consider the case when there is a useful path $Q$ containing
$\hat{P}$ outside the primary strip.
The prefix $Q'$ of this path must reach $w$ without touching $v$
(because $v\in \hat{P}$).
Observe that the union of $H$ and $\hat{P}$ contains a cycle
enclosing the source $s$,
and by the construction of $U^*$, there is no path from $s$ entering
$H\cap U^*$ (otherwise, it would have been included in $U^*$).
Consequently, because of planarity, the prefix $Q'$ must cross
$\hat{P}$, meaning that $Q$ cannot be simple.
Finally, we deal with Subcase (ii).
Let $x$ be the last floor vertex of $H$.
First note that, by the choice of $F^*$, any path from $s$ to $w$
not intersecting $\hat{P}$ cannot cross $F^*$ to the right;
therefore, it has to use $x$.
Because of the same reason, any path from $v\in F^*$ to $t$ cannot
leave $F^*$ to the right; thus, it must also go though $x$ as well.
Consequently, a path from $s$ to $t$ containing $\hat{P}$ in this case
cannot be simple. \end{proof}
\subsubsection{Reachability information} \label{sec:s-inside:reachability}
After we compute the primary strip and the open strip, and collapse strongly connected components in $\hat{C}$, we run a linear-time preprocessing to compute reachability information.
For each vertex $u$ in $\hat{D}$, we would like to compute $\mathsf{first}(u)\in F^*$ defined to be the first vertex $w$ in $F^*$ such that there is a path from $s$ to $u$ that uses $w$ but not other vertex in $F^*$ after $w$. We also want to compute $\mathsf{last}(u)\in F^*$ defined to be the last vertex $w'$ in $F^*$ such that there is a path from $u$ to $t$ that uses $w'$ as the first vertex in $F^*$.
To compute $\mathsf{first}(u)$, we find the left-first-search tree $T$ rooted at $s$. For $u\in\hat{C}$, we set $\mathsf{first}(u)$ to be the closest ancestor of $u$ on the floor $F^*$. We compute $\mathsf{last}(u)$ similarly by finding the left-first-search tree in the reverse graph rooted at $t$ and setting $\mathsf{last}(u)$ as the closest ancestor of $u$ on the floor $F^*$.
\subsubsection{Checking arcs in $\hat{D}$}
Now we describe our linear-time algorithm for determining the usefulness of arcs in $\hat{D}$
Consider an arc $e=(u,v)\in\hat{D}$. If $\phi(\mathsf{first}(u))>\phi(\mathsf{last}(v))$, then there exist a flipped forward path containing $e$. Thus, by Lemma~\ref{lem:flip-forward-is-useful}, $e$ is useful.
If $\phi(\mathsf{first}(u))\leq\phi(\mathsf{last}(v))$, then we need to check conditions in Lemma~\ref{lem:useful-backward-flip-paths}. To do so, we need to maintain additional information on every vertex $v$ in $F^*$, namely the hump $H(v)$ that contains it, a flag representing if $v$ is the first vertex in the hump, and a flag representing if the hump $H(v)$ contains an exit. Since these data can be preprocessed in linear-time, we can perform the check for each arc $e$ in constant time.
\subsection{Working on the Open Strip (2): Dealing with Arcs in Strongly Connected Components} \label{sec:arcs-in-scc}
Consider a strongly connected component $H_i$ in the open strip. If no arcs going into $H_i$ are useful or no arcs leaving $C$ are useful, clearly every arc in $H_i$ are useless. However, it is not obvious that applying the previous outside algorithm simply works because entering and leaving arcs determine the usefulness of the path. See Figure~\ref{fig:scc-in-open-strip-dependency}, for example.
\begin{figure}
\caption{The arc $e$ in the red path is useful when considering $C$
as an outside case instance. However, using the blue path as
an entering path (at the entrance $b_3$) and the green path as
a leaving path (at the exit $b'_2$) is not enough to show that.
One can, instead, use the orange path (at the entrance $b_1$)
as an entering path together with parts of the boundary to
construct a forward flipped path that contains $e$.}
\label{fig:scc-in-open-strip-dependency}
\end{figure}
The following lemma proves that we can take the strongly connected component $H_i$ and build an outside case instance by attaching, for each useful arc in the DAG entering $H_i$ or leaving $H_i$, corresponding to an entrance or an exit.
\begin{lemma} Every arc in an outside instance of $H_i$ is useful iff it is useful in the original graph. \end{lemma} \begin{proof}
The backward direction is straight-forward. Consider the forward
direction: we would like to show that if an arc $e$ is useful in the
instance, there exists a useful path containing it.
Suppose that $H_i$ has $n_1$ entrances and $n_2$ exits.
From Lemma~\ref{lem:boundary-is-ccw} and the observation
in the proof of Lemma~\ref{lem:one-useful} that
entrance-exit pairs do not appear interleaving,
we know that $H_i$ is enclosed by a counter-clockwise
cycle and we can name all entrances as $b_1,b_2,\ldots,b_{n_1}$ and
all exits as $b'_1,b'_2,\ldots,b'_{n_2}$ in such a way that they
appear in counter-clockwise order as
\[
b_1,b_2,\ldots,b_{n_1},b'_1,b'_2,\ldots,b'_{n_2}.
\]
Consider an arc $e$ which is useful in this instance. Since $e$ is
useful, there is a simple path $P$ from some entrance $b_i$
to some exit $b'_j$ containing $e$. Since the boundary arcs form a
counter-clockwise cycle, we can extend $P$ to start from $b_1$ by
adding boundary arcs from $b_1$ to $b_i$. We can also extend $P$ to
reach $b'_{n_2}$ by adding arcs from $b'_j$ to $b'_{n_2}$.
Since an arc entering $C$ at $b_1$ is useful and an arc leaving $C$
from $b'_{n_2}$ is also useful, we can construct a useful path
containing $e$ by joining a path from $s$ to $b_1$, the path $P$,
and a path from $b'_{n_2}$ to $t$. Thus, $e$ is useful. \end{proof}
\subsection{Working on the Primary Strip} \label{sec:s-inside:primary}
Given the primary strip $C_{U^*,F^*}$, we can use the algorithm from Section~\ref{sec:st-outside} to find all useful arcs inside $C_{U^*,F^*}$. The next lemma proves the correctness of this step.
\begin{lemma}
A useful arc $e$ in $C$ is useful iff it is useful in $C_{U^*,F^*}$. \end{lemma} \begin{proof}
The backward direction is obvious. We focus on the forward
direction. Assume for contradiction that there exist a useful arc
$e$ in $C$ but $e$ is not useful in $C_{U^*,F^*}$. Consider a
useful path $P$ from $s$ to $t$ containing $e$. Since $e$ is inside
$C_{U^*,F^*}$, $P$ must cross the lowest-floor $F^*$ or the top-most
ceiling $U^*$. However, $P$ cannot cross $U^*$ as it would create a
forward path outside the primary strip (contradicting
Lemma~\ref{lem:main-strip-encloses-all-forward}). Now suppose that
$P$ crosses $F^*$ at some vertex $w$ before $e$. If $P$ do not
leave the primary strip after $w$, then clearly $e$ must be useful inside
the primary strip. If $P$ leaves the primary strip at the ceiling
after $e$, the subpath of $P$ containing $e$ is again a forward
path. We also note that $P$ cannot leave $C_{U^*,F^*}$ at the floor
because it would cross itself. Since we reach contradiction in
every case, the lemma follows. \end{proof}
\subsection{The Ceiling Components are Useless} \label{sec:ceiling-components}
In this section, we show that every arc in the ceiling components is useless. The ceiling components are formed by arcs $e=(u,v)$ in $\hat{C}$ which are reachable only from paths leaving $F^*$ to the left and $v$ can reach $t$ only through paths entering $F^*$ from the left.
\begin{lemma}
Every ceiling arc $e$ is useless.
\label{lem:backward-comp-useless} \end{lemma} \begin{proof}
First, note that $e$ does not lie in a forward path; otherwise $e$
would be in the primary strip.
Let $P$ be a useful path from $s$ to $t$ containing $e$. Let $u'$
be the last vertices of $P\cap U^*$ before reaching $e$ and $v'$ be
the first vertices of $P\cap U^*$ after leaving $e$.
Because of the degree constraint, $P'$ must use the only incoming
arc of $u'$, say $e_u$, and the only outgoing arc of $v'$,
say $e_v$. Let $P_1$ be the prefix of $P'$ from $s$ to
the head of $e_u$ and $P_2$ be the suffix of $P'$ from the tail of $e_v$
to $t$. The only way $P_2$ can avoid crossing $P_1$ is to
cross $U^*$; however, this creates a clockwise cycle, a contradiction. \end{proof}
\section{Conclusions and Open Problems}
In this paper, we presented the algorithm that simplifies a directed planar network into a plane graph in which every vertex except the source $s$ and the sink $t$ has degree three, the graph has no clockwise cycle, and every arc is contained in some simple $s,t$-path.
Our algorithm can be applied as a preprocessing step for Weihe's algorithm \cite{Weihe97} and thus yields an $O(n\log n)$-time algorithm for computing maximum $s,t$-flow on directed planar graphs.
This gives an alternative approach that departs from the $O(n\log n)$-time algorithm of Borradaile and Klein \cite{BorradaileK09} and that of Erickson \cite{Erickson10}, which are essentially the same algorithm with different interpretations.
While other works mainly deal with maximum flow, the main concern in this paper is in simplifying the flow network. Henceforth, we believe that our algorithm will serve as supplementary tools for further developments of network flow algorithms.
Next let us briefly discuss open problems. A straightforward question is whether our approach can be generalized to a larger class of graphs on surfaces.
Another problem that might be interesting for readers is to remove the prerequisites from our main algorithm.
Specifically, prior to feeding a plane graph to the algorithm, we apply a sequence of reductions so that each vertex has degree three, and the plane graph contains no clockwise cycles. These are the prerequisites required by our main algorithm. Although the reduction does not change the value of the maximum flow, the usefulness of arcs in the modified network may differ from the original graph.
It would be interesting to simplify the flow network without changing the usefulness of arcs in the graph.
Lastly, we would like to note that if there exists a reduction that deals with clockwise cycles, then the degree requirement can be removed. Specifically, if there is a procedure that given a plane graph $G$ constructs a new plane graph $G'$ with no clockwise cycles together with another efficient procedure for identifying the usefulness of original arcs in $G$ based on the results in $G'$, then we can apply the following lemma.
\begin{lemma} \label{lem:no-cw-cycles}
If a plane graph $G$ contains no clockwise cycles, a new plane graph
$G'$ constructed by replacing every vertex of degree greater than three
in $G$ with a {\bf clockwise cycle} preserves the usefulness of
every arc from $G$ (those arcs that are not contained in any clockwise cycle). \end{lemma} \begin{proof} We first describe the reduction formally. Let $G$ be a plane graph with no clockwise cycle, and let $s$ and $t$ be the source and sink vertices. For each vertex $v\in V(G)$, let $d_v$ denote the degree of $v$ in $G$. We construct $G'$ by first adding copies of $s$ and $t$, namely $s'$ and $t'$, respectively. Then we add to $G'$ a clockwise cycle $C_v$ on $d_v$ vertices, for each vertex $v\in V(G)-\{s,t\}$. Each vertex $v_e$ in the cycle $C_v$ corresponds to an arc $e$ incident to $v$ in $G$, and vertices in $C_v$ are sorted in the same cyclic order as their corresponding arcs in $G$.
Next we add an arc $u_{uv}v_{uv}$ to $G'$ for every arc $uv\in E(G)$. Observe that $G'$ is a planar graph obtained by replacing each vertex of $G$ by a cycle, and we can keep the same planar drawing as that of $G$. In particular, the resulting graph $G'$ is a plane graph.
It can be seen that every simple path in $G$ corresponds to some simple path in $G'$. Thus, every useful arc in $G$ (w.r.t. $s$ and $t$) is also useful in $G'$ (w.r.t. $s'$ and $t'$). Now let us consider a useless arc $uv$ in $G$ (w.r.t. $s$ and $t$). Assume to the contrary that there is a simple $s',t'$-path $P'$ in $G'$ containing the arc $v_{uv}u_{uv}$, which is the arc corresponding to $uv$.
Since $G$ has no simple $s,t$-path containing $uv$, we know that $P'$ maps to a walk $P$ in $G$ that visits some vertex $w$ at least twice. Thus, $P'$ must visit the cycle $C_w$ at least twice as well. Let us say $P'$ enters $C_w$ at a vertex $w_{e_1}$, leaves $C_w$ from a vertex $w_{e_2}$, enters $C_{w}$ again at a vertex $w_{e_3}$ and then leaves $C_{w}$ from a vertex $w_{e_4}$.
Then we can construct a cycle $Q$ by walking along the $w_{e_1},w_{e_3}$-subpath of $P'$ and then continue to the $w_{e_3},w_{e_1}$-subpath of $C_w$.
Since the original graph $G$ has no clockwise cycle, $Q$ must be counterclockwise. Moreover, since $C_w$ is a clockwise cycle, the vertex $w_{e_4}$ must lie between $w_{e_3}$ and $w_{e_1}$.
At this point, it is not hard to see that the cycle $Q$ must enclose the $s',w_{e_4}$-subpath of $P'$.
So, the only way that $P'$ can leave the cycle $Q$ and reach $t'$ is to cross the $w_{e_2},w_{e_3}$-subpath of $C_w$. But, this is not possible unless $P'$ crosses itself. Hence, we arrive at a contradiction.
Therefore, the graph $G'$ preserves the usefulness of every arc from $G$. \end{proof}
\noindent{\bf Acknowledgement.} We thank Joseph Cheriyan for introducing us the flow network simplification problem. We also thank Karthik~C.S. for pointing out some typos.
Part of this work was done while Bundit Laekhanukit was visiting the Simons Institute for the Theory of Computing. It was partially supported by the DIMACS/Simons Collaboration on Bridging Continuous and Discrete Optimization through NSF grant \#CCF-1740425.
\end{document}
|
arXiv
|
{
"id": "1702.04786.tex",
"language_detection_score": 0.874811053276062,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{abstract} The aim of this short note is to give a simple proof of the non-rationality of the double cover of the three-dimensional projective space branched over a sufficiently general quartic. \end{abstract}
\title{A simple proof of the non-rationality
\
of a general quartic double solid} \section{Introduction} Throughout this work the ground field is supposed to be the complex number field $\mathbb{C}$.
A \textit{quartic double solid} is a projective variety represented as a the double cover of $\mathbb{P}^3$ branched along a smooth quartic. It is known that quartic double solids are unirational but not rational \cite{Beauville1977}, \cite{Tikhomirov1986}, \cite{Voisin1988}, \cite{Clemens1991}. Moreover, a general quartic double solid is not \textit{stably rational} \cite{Voisin2015a}. There are also a lot of results related to rationality problems of \emph{singular} quartic double solids see e.g. \cite{Artin-Mumford-1972}, \cite{Clemens1983}, \cite{Varley1986}, \cite{Debarre1990}, \cite{Przhiyalkovskij-Cheltsov-Shramov-2015}, \cite{CheltsovPrzyjalkowskiShramov2015b}.
The main result of this note is to give a simple proof of the following
\begin{mtheorem}{\bf Theorem.}\label{theorem-main} Let $X$ be the quartic double solid branched over the surface \begin{equation*} \label{equation-1} x_{1}^3x_{2}+ x_{2}^3x_{3}+ x_{3}^3x_{4}+ x_{4}^3x_{1}=0. \end{equation*} Then the intermediate Jacobian $J(X)$ is not a sum of Jacobians of curves. As a consequence, $X$ is not rational. \end{mtheorem}
\begin{mtheorem}{\bf Corollary.}\label{corollary-main} A general quartic double solid is not rational. \end{mtheorem}
Our proof uses methods of A. Beauville \cite{Beauville2012}, \cite{Beauville2013} and Yu. Zarhin \cite{Zarhin2009}. The basic idea is to find a sufficiently symmetric variety in the family. Then the action of the automorphism group provides a good tool to prove non-decomposability the intermediate Jacobian into a sum of Jacobians of curves by using purely \textit{group-theoretic} techniques. Since the Jacobians and their sums form a closed subvariety of the moduli space of principally polarized abelian varieties, this shows that a general quartic double solid is not rational\footnote{Recently V. Przyjalkowski and C. Shramov used similar method to prove non-rationality of some double quadrics \cite{PrzyjalkowskiShramov2016}.}.
\section{Preliminaries} \begin{say}{\bf Notation.} We use standard group-theoretic notation: if $G$ is a group, then ${\operatorname{z}}(G)$ denotes its center, $[G,G]$ its derived subgroup, and ${\operatorname{Syl}}_p(G)$ its (some) Sylow $p$-subgroup. By $\zeta_m$ we denote a primitive $m$-th root of unity. The group generated by elements $\upalpha_1,\upalpha_2,\dots$ is denoted by $\langle \upalpha_1,\upalpha_2,\dots\rangle$. \end{say}
\begin{say} Let $X$ be a three-dimensional smooth projective variety with $H^3(X,\mathscr{O}_X)=0$ and let $J(X)$ be its intermediate Jacobian regarded as a principally polarized abelian variety (see \cite{Clemens-Griffiths}). Then $J(X)$ can be written, uniquely up to permutations, as a direct sum \begin{equation} \label{equation-decomposition} J(X)=A_1 \oplus \dots \oplus A_n, \end{equation} where $A_1, \dots, A_p$ are indecomposable principally polarized abelian varieties (see \cite[Corollary 3.23]{Clemens-Griffiths}). This decomposition induces a decomposition of tangent spaces \begin{equation} \label{equation-decomposition-tangent} \operatorname{T}_{0,J(X)}= \operatorname{T}_{ 0,A_1} \oplus \dots \oplus \operatorname{T}_{0,A_n} \end{equation} Now assume that $X$ is acted on by a finite group $G$. Then $G$ naturally acts on $J(X)$ and $\operatorname{T}_{0,J(X)}$ preserving decompositions \eqref{equation-decomposition} and \eqref{equation-decomposition-tangent}. \end{say}
\begin{mtheorem}{\bf Lemma.}\label{lemma-action-80} Let $C$ be a curve of genus $g\ge 2$ and let $\Gamma\subset \operatorname{Aut}(C)$ be a subgroup of order $2^k\cdot 5$ whose Sylow $5$-subgroup ${\operatorname{Syl}}_5(\Gamma)$ is normal in $\Gamma$. Then the following assertions hold: \begin{enumerate} \item\label{lemma-action-80-k=2} if $k=2$, then $g\ge 3$,
\item\label{lemma-action-80-k=4} if $k=4$, then $g\ge 6$,
\item\label{lemma-action-80-k=5}\label{faithful-action-on-curve} if $k=5$, then $g\ge 11$. \end{enumerate}
\end{mtheorem} \begin{proof} Let $C':=C/{\operatorname{Syl}}_5(\Gamma)$ and $g':=g(C')$. Let $P_1,\dots, P_n\in C'$ be all the branch points. By Hurwitz's formula \[ g+4=5g'+2n. \] The group $\Gamma':=\Gamma/ {\operatorname{Syl}}_5(\Gamma)$ of order $2^k$ faithfully acts on $C'$ and permutes $P_1,\dots, P_n$. \ref{lemma-action-80-k=2} Assume that $k=g=2$. Then $g'=0$, $C'\simeq \mathbb{P}^1$, and $n=3$. At least one of the points $P_1,P_2,P_3$, say $P_1$, must be fixed by $\Gamma'$. But then $\Gamma'$ must be cyclic (of order $4$) and it cannot leave the set $\{P_1,P_2,P_3\}\subset \mathbb{P}^1$ invariant. This proves \ref{lemma-action-80-k=2}.
\ref{lemma-action-80-k=4} Assume that $k=4$ and $g\le 5$. Then $g'\le 1$. If $g'=0$, then $n\in \{3,\, 4\}$ and the group $\Gamma'$ of order $16$ acts on $C'\simeq \mathbb{P}^1$ so that the set $\{P_1,\dots, P_n\}$ is invariant. This is impossible. If $g'=1$, then, as above, $\Gamma'$ acts on an elliptic curve $C'$ leaving a non-empty set of $n\le 2$ points is invariant. This is again impossible and the contradiction proves \ref{lemma-action-80-k=4}.
\ref{lemma-action-80-k=5} Finally, let $k=5$ and $g\le 10$. Then $g'\le 2$ and $n\le 7$. If $g'\le 1$, then we get a contradiction as above. Let $g'=2$, let $C'\to \mathbb{P}^1$ the the canonical map, and let $\Gamma''\subset \operatorname{Aut}(\mathbb{P}^1)$ be the image of $\Gamma'$. Since $\Gamma''$ is a $2$-subgroup in $\operatorname{Aut}(\mathbb{P}^1)$, it is either cyclic or dihedral. On the other hand, $\Gamma''$ permutes the branch points $Q_1,\dots,Q_6\in \mathbb{P}^1$ so that the stabilizer of each $Q_i$ is a subgroup in $\Gamma''$ of index $\le 4$. Clearly, this is impossible. \end{proof}
\section{Symmetric quartic double solid} \begin{say} Let $X$ be the quartic double solid as in Theorem \ref{theorem-main}. Then $X$ is isomorphic to a hypersurface given by \begin{equation} \label{equation} y^2+x_{1}^3x_{2}+ x_{2}^3x_{3}+ x_{3}^3x_{4}+ x_{4}^3x_{1}=0, \end{equation} in the weighted projective space $\mathbb{P}:=\mathbb{P}(1^4,2)$, where $x_1,\dots,x_4,y$ are homogeneous coordinates with $\deg x_i=1$, $\deg y=2$.
Let $\upalpha$ be the automorphism of $X$ induced by the diagonal matrix \[ {\operatorname{diag}} (1,\, \zeta_{40}^{38},\, \zeta_{40}^{4},\, \zeta_{40}^{26};\, \zeta_{40}^{-1}) \] and let $\upbeta$ be the cyclic permutation $(1,2,3,4)$ of coordinates $x_1,x_2,x_3,x_4$. Since \[ \upbeta\upalpha\upbeta^{-1}= {\operatorname{diag}} (\zeta_{40}^{26}, 1, \zeta_{40}^{38},\, \zeta_{40}^{4};\, \zeta_{40}^{-1})= {\operatorname{diag}} (1,\, \zeta_{40}^{14},\, \zeta_{40}^{12},\, \zeta_{40}^{18};\, \zeta_{40}^{27})= \upalpha^{13}, \] these automorphisms generate the group \[ G= \langle \upalpha,\, \upbeta \mid \upalpha^{40}=\upbeta^4=1,\hspace{3pt} \upbeta\upalpha\upbeta^{-1}=\upalpha^{13}\rangle \subset \operatorname{Aut}(X), \quad G\simeq \mathbb{Z}/40 \rtimes\mathbb{Z}/ 4. \] \end{say}
\begin{mtheorem}{\bf Lemma.}\label{subgroup-index10} Let $G$ be as above. Then we have \begin{enumerate} \item \label{subgroup-index10-1} ${\operatorname{z}}(G)=\langle \upalpha^{10}\rangle$ and $[G,G]=\langle \upalpha^4\rangle$, \item \label{subgroup-index10-2} the Sylow $5$-subgroup ${\operatorname{Syl}}_5(G)$ is normal, \item \label{subgroup-index10-3} any subgroup in $G$ of index $10$ contains ${\operatorname{z}}(G)$. \end{enumerate} \end{mtheorem}
\begin{proof} \ref{subgroup-index10-1} can be proved by direct computations and \ref{subgroup-index10-2} is obvious because ${\operatorname{Syl}}_5(G)\subset \langle\upalpha\rangle$. To prove \ref{subgroup-index10-3} consider a subgroup $G'\subset G$ of index $10$. The intersection $G'\cap \langle\upalpha\rangle$ is of index $\le 4$ in $G'$. Hence $G'\cap \langle\upalpha\rangle$ is a $2$-group of order $\ge 4$ and so $\upalpha^{10}\in G'\cap \langle\upalpha\rangle$. \end{proof}
\begin{mtheorem}{\bf Lemma (cf. \cite[0.1(b)]{Voisin1988}).}\label{Lemma-representation0} There exists a natural exact sequence \[ 0 \to H^2(X,\Omega_{X}^1 ) \to H^0(X,-K_X)^\vee {\longrightarrow}\mathbb{C} \to 0. \] \end{mtheorem} \begin{proof} Since $X$ is contained in the smooth locus of $\mathbb{P}$ and $\mathscr{O}_{\mathbb{P}}(X)=\mathscr{O}_{\mathbb{P}}(4)$, we have the following exact sequence \[
0 \longrightarrow \mathscr{O}_X(-4)\longrightarrow \Omega_{\mathbb{P}}^1|_X \longrightarrow \Omega_{X}^1 \longrightarrow 0, \] and so \[
H^2(X, \Omega_{\mathbb{P}}^1|_X) \to H^2(X,\Omega_{X}^1 ) \to H^0(X,\mathscr{O}_X(2))^\vee \to H^3(X, \Omega_{\mathbb{P}}^1|_X)\to 0 \] The Euler exact sequence for $\mathbb{P}=\mathbb{P}(1^4,2)$ has the form \[ 0 \longrightarrow \Omega^1_{\mathbb{P}} \longrightarrow\mathscr{O}_{\mathbb{P}}(-2)\oplus \mathscr{O}_{\mathbb{P}}(-1)^{\oplus 4} \longrightarrow \mathscr{O}_{\mathbb{P}} \longrightarrow 0. \] Restricting it to $X$ we obtain
$H^2(X, \Omega_{\mathbb{P}}^1|_X)=0$ and $H^3(X, \Omega_{\mathbb{P}}^1|_X)=\mathbb{C}$. \end{proof}
\begin{mtheorem}{\bf Lemma.}\label{Lemma-representation} We have the following decomposition of $G$-modules: \[ \operatorname{T}_{0,J(X)}= V_4\oplus V_4'\oplus V_2, \] where $V_4$, $V_4'$ are irreducible faithful $4$-dimensional representations and $V_2$ is an irreducible $2$-dimensional representation with kernel $\langle \upalpha^{8},\,\upbeta^2\rangle$. Moreover, ${\operatorname{z}}(G)$ acts on $V_4$ and $V_4'$ via different characters. \end{mtheorem}
\begin{proof} Clearly, $\operatorname{T}_{0,J(X)}\simeq H^0(J(X),\Omega_{J(X)})^\vee \simeq H^2(X,\Omega_{X}^1)$ and by Lemma \ref{Lemma-representation0} we have an injection
$\operatorname{T}_{0,J(X)}\hookrightarrow H^0(X,-K_X)^\vee$. By the adjunction formula $K_X=(K_{\mathbb{P}}+X)|_X$ and so \[ H^0(X,-K_X)\simeq H^0(\mathbb{P},\mathscr{O}_{\mathbb{P}}(-K_{\mathbb{P}}-X)). \] Consider the affine open subset $U:=\{x_1x_2x_3x_4\neq 0\}$. Then $v=y/x_1^2$ and $z_i=x_i/x_1$, $i=2,3,4$ are affine coordinates in $U\subset \{x_1\neq 0\}\simeq \mathbb A^4$. Let $\upomega$ be the $3$-form \[ \upomega:= \frac {d z_2\wedge d z_3\wedge d z_4 }{\partial \phi/\partial v} = \frac {d z_2\wedge d z_3\wedge d z_4 }{2 v}, \] where $\phi=v^2+z_{2}+ z_{2}^3z_{3}+ z_{3}^3z_{4}+ z_{4}^3$ is the equation of $X$ in $U$. It is easy to check that for any polynomial $\psi(z_2,z_3,z_4)$ of degree $\le 2$ the element $\psi\cdot \upomega ^{-1}$ extends to a section of $H^0(X, -K_X)$. Thus we have \[
H^0(X, -K_X)\simeq \{ \psi(z_2,z_3,z_4)\cdot \upomega ^{-1} \mid \deg \psi\le 2\}. \] It is easy to check that the forms \begin{equation} \label{equation-basis} \upomega ^{-1}, z_2^2 \upomega ^{-1}, z_3^2 \upomega ^{-1}, z_4^2\upomega ^{-1}, z_2 \upomega ^{-1}, z_2z_3 \upomega ^{-1}, z_3z_4 \upomega ^{-1}, z_4 \upomega ^{-1}, z_3 \upomega ^{-1} , z_2z_4\upomega ^{-1} \end{equation} are eigenvectors for $\upalpha$ and $\upbeta$ permutes them. Moreover, the following subspaces \begin{itemize}
\item[] $W_4= \langle \upomega ^{-1},\hspace{5pt}z_2^2 \upomega ^{-1},\hspace{5pt}z_3^2 \upomega ^{-1},\hspace{5pt}z_4^2\upomega ^{-1}\rangle$,
\item[] $W_4'= \langle z_2 \upomega ^{-1},\hspace{5pt}z_2z_3 \upomega ^{-1},\hspace{5pt}z_3z_4 \upomega ^{-1},\hspace{5pt}z_4 \upomega ^{-1}\rangle$,
\item[] $W_2= \langle z_3 \upomega ^{-1} ,\hspace{5pt}z_2z_4\upomega ^{-1}\rangle$.
\end{itemize} are $G$-invariant in $H^0(X, -K_X)$. Moreover, in the basis \eqref{equation-basis} the element $\upalpha$ acts diagonally: \begin{equation} \label{equation-W} \begin{array}{l}
\upalpha|_{W_4} ={\operatorname{diag}} (\zeta_{40}^{11},\, \zeta_{40}^{7},\, \zeta_{40}^{19},\, \zeta_{40}^{23}), \\[4pt]
\upalpha|_{W_4'} ={\operatorname{diag}} (\zeta_{40}^{9},\, \zeta_{40}^{13},\, \zeta_{40}^{},\, \zeta_{40}^{37}), \\[4pt]
\upalpha|_{W_2} ={\operatorname{diag}} (\zeta_{8}^{3},\, \zeta_{8}^{7}), \end{array} \end{equation} and $\upbeta$ acts on each of these subspaces permuting the eigenspaces of $\upalpha$ cyclically.
Thus $\upalpha^{10}$ acts on $W_4$ (resp., $W_4'$) via scalar multiplication by $\zeta_4^{3}$ (resp., $\zeta_4$). Put $V_4:=W_4^\vee$, $V_4':=W_4'^\vee$, $V_2:=W_2^\vee$. \end{proof}
\section{Proof of Theorem \ref{theorem-main}}
\begin{say} Assume to the contrary to Theorem \ref{theorem-main} that $J(X)$ is a direct sum of Jacobians of curves, i.e. in the unique decomposition \eqref{equation-decomposition} we have $A_i\simeq J(C_i)$, where $C_i$ is a curve of genus $\ge 1$ and $J(C_i)$ is its Jacobian regarded as a principally polarized abelian variety. Let $G_i$ be the stabilizer of $A_i$. There is a natural homomorphism $\varsigma_i: G_i \to\operatorname{Aut} (C_i)$. By the Torelli theorem $\varsigma_i$ is injective and we have \begin{equation} \label{equation-Aut-C} \operatorname{Aut} (J(C_i))\simeq \begin{cases} \operatorname{Aut} (C_i)&\text{if $C_i$ is hyperelliptic,} \\ \operatorname{Aut} (C_i)\times \{\pm 1\}&\text{otherwise.} \end{cases} \end{equation} Let us analyze the action of $G$ on the set $\{A_1,\dots, A_n\}$. Up to renumbering we may assume that subvarieties $A_1,\dots, A_m$ form one $G$-orbit (however, the choice of this orbit is not unique in general). Clearly, $m\in \{1,2,4,5,8,10\}$. Denote the stabilizer of $A_i$ by $G_i$. Consider the possibilities for $m$ case by case. \end{say}
\begin{say} \label{lemma-invariant-A} {\bf Case: $m=1$,} that is, $A_1\subset J(X)$ is a $G$-invariant subvariety. Since ${\operatorname{z}}(G)=\langle \upalpha^{10}\rangle$, the only normal subgroup of order 2 in $G$ is $\langle \upalpha^{20}\rangle$. Hence $G$ cannot be decomposed as a direct product of groups of orders $2$ and $80$ (otherwise the order of $\upalpha$ would be $20$). If the action of $G$ on $A_1=J(C_1)$ is faithful, then by \eqref{equation-Aut-C} so is the corresponding action on $C_1$. So, the curve $C_1$ of genus $\le 10$ admits faithful action of the group $G$ of order $2^{5}\cdot 5$. This contradicts Lemma \ref{lemma-action-80}\ref{lemma-action-80-k=5}. Therefore the induced representation on $\operatorname{T}_{0,A_1}$ is not faithful. By Lemma \ref{Lemma-representation}\ $\operatorname{T}_{0,J(C_1)}= V_2$. In this case $g(C_1)=2$ and the action of $G$ on $J(C_1)$ induces a faithful action of the group $\bar G:= G/\langle \upalpha^{8},\,\upbeta^2\rangle$ of order $16$. Since $C_1$ is hyperelliptic, $\bar G$ is contained in $\operatorname{Aut}(C_1)$. If $\bar G$ contains the hyperelliptic involution $\tau$, then $\tau$ generates a normal subgroup of order 2. In this case $\langle \tau\rangle=[\bar G,\bar G]$ and $\bar G/\langle \tau\rangle$ is an abelian non-cyclic group of order $8$. But such a group cannot act faithfully on $C_1/\langle \tau\rangle\simeq \mathbb{P}^1$. Thus $\bar G$ does not contain the hyperelliptic involution. In this case the image of the induced action of $\bar G$ on canonical sections $H^0(C_1,\mathscr{O}_{C_1}(K_{C_1}))$ does not contain scalar matrices. Hence this representation is reducible and so it is trivial on $[\bar G,\bar G]$. On the other hand, the action of $\operatorname{Aut}(C_1)$ on $H^0(C_1,\mathscr{O}_{C_1}(K_{C_1}))$ must be faithful a contradiction. \end{say}
From now on we may assume that the decomposition \eqref{equation-decomposition} contains no $G$-invariant summands.
\begin{say}{\bf Case: $m=5$.} The subspace $\operatorname{T}_{0,A_1}\oplus\cdots\oplus \operatorname{T}_{0,A_5}\subset \operatorname{T}_{0,J(X)}$ is a $G$-invariant of dimension $5$ or $10$. On the other hand, $\operatorname{T}_{0,J(X)}$ contains no invariant subspaces of dimension $5$ by Lemma \ref{Lemma-representation}. Hence, $\operatorname{T}_{0,A_1}\oplus\cdots\oplus \operatorname{T}_{0,A_5}= \operatorname{T}_{0,J(X)}$, $\dim A_i=2$, and $J(X)= \oplus_{i=1}^{5} A_i$. The stabilizer $G_i\subset G$ is a Sylow $2$-subgroup that faithfully acts on $C_i$ (because $C_i$ is hyperelliptic, see \eqref{equation-Aut-C}). Further, $G_i$ permutes the Weierstrass points $P_1,\dots,P_6\in C_i$. Hence a subgroup $G_i'\subset G_i$ of index 2 fixes one of them. In this situation, $G_i'$ must be cyclic. On the other hand, it is easy to see that $G$ does not contain any elements of order $16$, a contradiction. \end{say}
\begin{say}{\bf Case: $m=10$.} Then $A_1,\dots,A_{10}$ are elliptic curves and $G_i\subset G$ is a subgroup of index $10$. By Lemma \ref{subgroup-index10} each $G_i$ contains ${\operatorname{z}}(G)$. Clearly, ${\operatorname{z}}(G)$ acts on $\operatorname{T}_{0,A_i}$ via the same character. Since the subspaces $\operatorname{T}_{0,A_i}$ generate $\operatorname{T}_{0,J(X)}$, the group ${\operatorname{z}}(G)$ acts on $\operatorname{T}_{0,J(X)}$ via scalar multiplication. This contradicts Lemma \ref{Lemma-representation}. \end{say}
\begin{say} \label{orbit=8} {\bf Case: $m=8$.} Then $A_1,\dots,A_8$ are elliptic curves and the stabilizer $G_1\subset G$ is of order $20$. In particular, the Sylow $5$-subgroup ${\operatorname{Syl}}_5(G)$ is contained in $G_1$. Since ${\operatorname{Syl}}_5(G)$ is normal in $G$, we have ${\operatorname{Syl}}_5(G)\subset G_i$ for $i=1,\dots,8$. Since the automorphism group of an elliptic curve contains no order $5$ elements, ${\operatorname{Syl}}_5(G)$ acts trivially on $A_i$. Therefore, ${\operatorname{Syl}}_5(G)$ acts trivially on the $8$-dimensional $G$-invariant subspace $\operatorname{T}_{0,A_1}\oplus \cdots\oplus \operatorname{T}_{0,A_8}$. This contradicts Lemma \ref{Lemma-representation}. \end{say}
\begin{say}\label{orbit=4}{\bf Case: $m=4$.} The intersection $G_1\cap \langle\upalpha\rangle$ is a subgroup of index $\le 4$ in both $G_1$ and $\langle\upalpha\rangle$. Hence, $G_1\ni \upalpha^{4}$ and so $G_1\supset [G,G]$. In particular, $G_1$ is normal and $G_1=\cdots=G_4$. If $\dim A_1=1$, then the element $\upalpha^8$ of order $5$ must act trivially on elliptic curves $A_i\in 0$, $i=1,\dots,4$. Therefore, $\upalpha^8$ acts trivially on the $4$-dimensional space $\operatorname{T}_{0,A_1}\oplus\cdots\oplus \operatorname{T}_{0,A_4}$. This contradicts Lemma \ref{Lemma-representation}.
Thus $\dim A_1=2$. Then $\operatorname{T}_{0,A_1}\oplus\cdots\oplus \operatorname{T}_{0,A_4}=V_4\oplus V_4'$. An eigenvalue of $\upalpha$ on $\operatorname{T}_{0,A_1}\oplus\cdots\oplus \operatorname{T}_{0,A_4}$ must be a primitive $40$-th root of unity (see \eqref{equation-W}). Hence the group $G_1\cap \langle\upalpha\rangle$ acts faithfully on $\operatorname{T}_{0,A_1}$ and $C_1$ (see \eqref{equation-Aut-C}). By Lemma \ref{lemma-action-80}\ref{lemma-action-80-k=2} $G_1\cap \langle\upalpha\rangle$ is of order $10$, i.e. $G_1\cap \langle\upalpha\rangle=\langle \upalpha^4\rangle$ and the kernel $N:=\ker (G_1\to \operatorname{Aut}(C_1))$ is of order $4$. Thus $G_1=\langle \upalpha^4\rangle\times N$. In particular, $G_1$ is abelian. But then the centralizer $\operatorname{C}(\upalpha^8)$ of $\upalpha^8$ contains $N$ and $\langle\upalpha\rangle$. Therefore, $\operatorname{C}(\upalpha^8)=G$ and $\upalpha^8\in {\operatorname{z}}(G)$. This contradicts Lemma \ref{subgroup-index10}\ref{subgroup-index10-1}. \end{say}
Thus we have excluded the cases $m=1,4,5,8,10$. The only remaining possibility is that all the orbits of $G$ on $\{ A_i\}$ are of cardinality $2$.
\begin{say}{\bf Case: $m=2$.} Then $\dim A_1\le 5$ and $G_1$ is a group of order $80$. By replacing the orbit $\{A_1,A_2\}$ with another one we may assume that $\operatorname{T}_{0,A_1}\oplus \operatorname{T}_{0,A_2}\not \subset V_2$ and so $\operatorname{T}_{0,A_1}\oplus \operatorname{T}_{0,A_2}$ coincides with either $V_4$, $V_4'$, or $V_4\oplus V_4'$. In particular, $g(C_1)\ge 2$. Clearly, $G_1\cap \langle\upalpha\rangle$ is of order $40$ or $20$. Hence, $\upalpha^{2}\in G_1$ and so the group $G_1$ cannot be decomposed as a direct product $G_1=\langle \upalpha^{20}\rangle\times H$. By the Torelli theorem $G_1$ faithfully acts on $C_1$. This contradicts Lemma \ref{lemma-action-80}\ref{lemma-action-80-k=4}. \end{say} Proof of Theorem \ref{theorem-main} is now complete.
\begin{proof}[Proof of Corollary \textup{\ref{corollary-main}}] The Jacobians and their sums form a closed subvariety of the moduli space of principally polarized abelian varieties. By Theorem \ref{theorem-main}, in our case, this subvariety does not contain the subvariety formed by Jacobians of quartic double solids. Therefore a general quartic double solid is not rational. \end{proof}
\subsection*{Acknowledgements.} The author would like to thank C. Shramov and the referee for useful comments and corrections. \input jacobian.bbl
\end{document}
|
arXiv
|
{
"id": "1605.02017.tex",
"language_detection_score": 0.583763599395752,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Random Matrices from Linear Codes and the Convergence to Wigner's Semicircle Law} \begin{abstract} Recently we considered a class of random matrices obtained by choosing distinct codewords at random from linear codes over finite fields and proved that under some natural algebraic conditions their empirical spectral distribution converges to Wigner's semicircle law as the length of the codes goes to infinity. One of the conditions is that the dual distance of the codes is at least 5. In this paper, employing more advanced techniques related to Stieltjes transform, we show that the dual distance being at least 5 is sufficient to ensure the convergence, and the convergence rate is of the form $n^{-\beta}$ for some $0 < \beta < 1$, where $n$ is the length of the code.
\end{abstract} \begin{keywords} Group randomness, linear code, dual distance, empirical spectral measure, random matrix theory, Wigner's semicircle law. \end{keywords} \section{Introduction}\label{intro}
Random matrix theory is the study of matrices whose entries are random variables. Of particular interest is the study of eigenvalue statistics of random matrices such as the empirical spectral measure. It has been broadly investigated in a wide variety of areas, including statistics \cite{Wis}, number theory \cite{MET}, economics \cite{econ}, theoretical physics \cite{Wig} and communication theory \cite{TUL}.
Most of the matrix models considered in the literature were matrices whose entries have independent structures. In a series of work (\cite{Tarokh2,Tarokh3,OQBT}), initiated in \cite{Tarokh1}, the authors studied a class of matrices formed by choosing codewords at random from linear codes over finite fields and ultimately proved the convergence of the empirical spectral distribution of their Gram matrices to the Marchenko-Pastur law under the condition that the minimum Hamming distance of the dual codes is at least 5. This is the first result relating the randomness of matrices from linear codes to the algebraic properties of the underlying dual codes, and can be interpreted as a joint randomness test for sequences from linear codes. It implies in particular that sequences from linear codes with desired properties behave like random sequences from the view point of random matrix theory. This is called a ``group randomness'' property in \cite{Tarokh1} and may have many applications (see \cite{WigCode,WigM} from a different perspective).
Recently we considered a distinct normalization of matrices obtained in a similar fashion from linear codes and proved the convergence of the empirical spectral distribution to the Wigner's semicircle law under some natural algebraic conditions of the underlying codes (see \cite{CSC}). This is also a group randomness property of linear codes. In this paper we explore this new phenomenon much further.
\subsection{Statement of Main Results}
To describe our results more precisely, we need some notation. Let $\mathscr{C}=\{\mathcal{C}_i : i \ge 1\}$ be a family of linear codes of length $n_i$ and dimension $k_i$ over the finite field $\mathbb{F}_q$ of $q$ elements ($\mathcal{C}_i$ is called an $[n_i,k_i]_q$ code for short), where $q$ is a prime power. The most interesting case is binary linear codes, corresponding to $q=2$. Denote by $\mathcal{C}_i^\bot$ the dual code of $\mathcal{C}_i$ and $d_i^\bot$ the Hamming distance of $\mathcal{C}_i^\bot$. $d_i^\bot$ is also called the \emph{dual distance} of $\mathcal{C}_i$.
The standard additive character of $\mathbb{F}_q$ extends component-wise to a natural mapping $\psi: \mathbb{F}_q^{n_i} \to (\mathbb{C}^*)^{n_i}$. For each $i$, we choose $p_i$ distinct codewords from $\mathcal{C}_i$ and apply the mapping $\psi$. Endowing with uniform probability on the choice of the $p_i$ codewords, this forms a probability space. Put the $p_i$ distinct sequences as the rows of a $p_i \times n_i$ random matrix $\Phi_{\mathcal{C}_i}$. Denote \begin{equation}\label{Gram} \mathcal{G}_{\mathcal{C}_i}=\frac{1}{n_i}\Phi_{\mathcal{C}_i}\Phi_{\mathcal{C}_i}^*, \end{equation} where $\Phi_{\mathcal{C}_i}^*$ is the conjugate transpose of the matrix $\Phi_{\mathcal{C}_i}$ and define \begin{equation}\label{Mn} M_{\mathcal{C}_i}=\sqrt\frac{n_i}{p_i}(\mathcal{G}_{\mathcal{C}_i}-I_{p_i}). \end{equation} Here $I_{p_i}$ is the $p_i \times p_i$ identity matrix.
For any $n \times n$ matrix $\mathbf{A}$ with eigenvalues $\lambda_1,\ldots,\lambda_n$, the \emph{spectral measure} of $\mathbf{A}$ is defined by $$\mu_\mathbf{A}=\frac{1}{n}\sum_{j=1}^n \delta_{\lambda_j},$$ where $\delta_{\lambda}$ is the Dirac measure at the point $\lambda$. The \emph{empirical spectral distribution} of $\mathbf{A}$ is defined by $$F_\mathbf{A}(x):=\int_{-\infty}^x \mu_\mathbf{A}(\text{d}x).$$
Our first main result is as follows: \begin{theorem} \label{thm} Suppose $p_i, \frac{n_i}{p_i} \to \infty$ simultaneously as $i \to \infty$. If $d^\bot_i \geq 5$ for any $i$, then as $i \to \infty$, we have
\begin{eqnarray} \label{1:eq1} \mu_{n_i}(\mathcal{I}) \to \varrho_\SC(\mathcal{I}) \quad \emph{ in Probability},\end{eqnarray}
and the convergence is uniform for all intervals $\mathcal{I} \subset \mathbb{R}$. Here $\mu_{n_i}$ is the spectral measure of the matrix $M_{\mathcal{C}_i}$ and $\varrho_\mathrm{SC}$ is the probability measure of the semicircle law whose density function is given by \begin{equation}\label{scpdf} \mathrm{d}\varrho_\SC(x)=\frac{1}{2\pi}\sqrt{4-x^2}\mathbbm{1}_{[-2,2]}\,\mathrm{d} x, \end{equation} and $\mathbbm{1}_{[-2,2]}$ is the indicator function of the interval $[-2,2]$. \end{theorem}
We remark that originally in \cite{CSC} the same convergence (\ref{1:eq1}) was proved with an extra condition that there is a fixed constant $c > 0$ independent of $i$ such that \begin{equation} \label{1:srip}
|\langle v,v' \rangle| \leq c\sqrt{n_i}, \quad \mbox{ for any } v \ne v' \in \psi(\mathcal{C}_i). \end{equation} The condition (\ref{1:srip}) is natural as explained in \cite{CSC}, and when $q=2$, it is equivalent to
\[\left|\mathrm{wt}(\mathbf{c})-\frac{n_i}{2}\right| \le \frac{c}{2} \sqrt{n_i}, \quad \forall \mathbf{c} \in \mathcal{C}_i \setminus \{\mathbf{0}\},\] where $\mathrm{wt}(\mathbf{c})$ is the Hamming weight of the codeword $\mathbf{c}$. It is interesting that this extra condition can be dropped. Now the result of Theorem \ref{thm} has the same strength as that of \cite{OQBT} where the condition $d_i^\bot \ge 5$ alone is sufficient to ensure the convergence. It shall be noted that similar to \cite{OQBT}, the condition $d_i^\bot \ge 5$ in Theorem \ref{thm} is optimal because if $d_i^\bot=4 \, \forall i$, then Conclusion (\ref{1:eq1}) is false for first-order binary Reed-Muller codes which have dual distance $4$.
Our second main result shows that the rate of convergence (\ref{1:eq1}) is fast with respect to the length of the codes.
\begin{theorem} \label{thm1-2} Let $\mathcal{C}$ be an $[n,k]_q$ code with dual distance $d^\bot \ge 5$. For fixed constants $\gamma_1,\gamma_2 \in (0,1)$ and $c \ge 1$, suppose $p$ and $n$ satisfy \[c^{-1} n^{\gamma_1} \leq p \leq c \, n^{\gamma_2}.\] Then \begin{equation}\label{SD}
\left|\mu_{n}(\mathcal{I})-\varrho_\SC(\mathcal{I})\right| \prec n^{-\beta} \end{equation} uniformly for all intervals $\mathcal{I} \subset \mathbb{R}$, where $\beta>0$ is given by \begin{equation}\label{beta} \beta:=\min\left\{\frac{\gamma_1}{4},\frac{1-\gamma_2}{8} \right\}. \end{equation} \end{theorem}
We remark that the symbol ``$\prec$'' in (\ref{SD}) is a standard ``stochastic domination'' notation in probability theory (see \cite{Local law} for details), which means that for any $\varepsilon >0$ and any $D>0$, there is a quantity $N(\varepsilon,D,c,\gamma_1,\gamma_2)$, such that whenever $n \geq N(\varepsilon,D,c,\gamma_1,\gamma_2)$, we have \begin{equation}\label{SD2}
\sup_\mathcal{I} \mathbb{P}\left[|\mu_{n}(\mathcal{I})-\varrho_\SC(\mathcal{I})| > n^{-\beta+\varepsilon} \right] \leq n^{-D}. \end{equation} Here $\mathbb{P}$ is the probability within the space of picking $p$ distinct codewords from $\mathcal{C}$ and the supremum is taken over all intervals $\mathcal{I} \subset \mathbb{R}$. Since $\varepsilon, D$ and $N(\varepsilon,D,c,\gamma_1,\gamma_2)$ do not depend on $\mathcal{C}$, the supremum can be taken over all linear codes $\mathcal{C}$ of length $n$ over $\mathbb{F}_q$ with $d^\bot \geq 5$.
We also remark that $d^\bot \geq 5$ is a very mild restriction on linear codes $\mathcal{C}$, and there is an abundance of binary codes that satisfy this condition, for example, the Gold codes (\cite{Gold}), some families of BCH codes (see \cite{DIN1,DIN2}) and many families of cyclic and linear codes studied in the literature (see for example \cite{CHE,TAN}). Such binary linear codes can also be generated by almost perfect nonlinear (APN) functions \cite{Blon,POTT}, a special class of functions with important applications in cryptography.
\subsection{Simulations}
We illustrate Theorems \ref{thm} and \ref{thm1-2} by numerical experiments. We focus on binary Gold codes augmented by the all-1 vector. It is known that binary Gold codes have length $n=2^m-1$, dimension $2m$ and dual distance 5. The augmented binary Gold codes has length $n$, dimension $2m+1$ and dual distance at least 5. Because of the presence of the all-1 vector, the condition (\ref{1:srip}) is not satisfied. For each triple $(m,n,p)$ in the set $\{(5,31,8), (7,127,20),(9,511,35),(11,2047,50)\}$, we randomly pick $p$ codewords from the augmented binary Gold code of length $n=2^m-1$ and form the corresponding matrix, from which we use {\bf Sage} to compute the eigenvalues and plot the empirical spectral distribution along with Wigner's distribution (see Figures \ref{fig1} to \ref{fig4} below). We do the above 10 times for each such triple $(m,n,p)$ and at each time, we find that the plots are almost the same as before: they are all very close to Wigner's semicircle law and as the length $n$ increases, they become less and less distinguishable.
In order to illustrate more clearly the shape of the eigenvalue distribution, we also plot a density graph, which is shown in Figure \ref{fig5}. This is based on picking $p=100$ codewords from a binary Gold code of length $n=32767=2^{15}-1$.
From (\ref{beta}) it is easy to see that $\beta \leq 1/12$ and the upper bound is achieved when $\gamma_1=\gamma_2=1/3$. It might be possible to improve this value $\beta$ and hence obtain a better convergence rate. From the simulation results, however, it is not clear to us what the optimal $\beta$ that one may expect is.
\begin{figure}
\caption{Empirical spectral distribution (ESD) of $[31,11,12]$ augmented binary Gold code versus Wigner semicircle law (SC), with $p=8$}
\label{fig1}
\end{figure} \begin{figure}
\caption{Empirical spectral distribution (ESD) of $[127,15,56]$ augmented binary Gold code versus Wigner semicircle law (SC), with $p=20$}
\label{fig2}
\end{figure} \begin{figure}
\caption{Empirical spectral distribution (ESD) of $[511,19,240]$ augmented binary Gold code versus Wigner semicircle law (SC), with $p=35$}
\label{fig3}
\end{figure} \begin{figure}
\caption{Empirical spectral distribution (ESD) of $[2047,23,992]$ augmented binary Gold code versus Wigner semicircle law (SC), with $p=50$}
\label{fig4}
\end{figure} \begin{figure}
\caption{Empirical spectral density of $[32767,30,16256]$ binary Gold code versus Wigner semicircle density, with $p=100$}
\label{fig5}
\end{figure}
\subsection{Techniques and relation to previous work}
This paper strengthens \cite[Theorem 2]{CSC} on two fronts: in Theorem \ref{thm} we obtain the same convergence by removing the extra condition (\ref{1:srip}), and in Theorem \ref{thm1-2} we obtain a strong and explicit convergence rate with respect to the length of the code, and the results were supported by computer simulations.
The main technique we use in this paper is the Stieltjes transform, a well-developed and standard tool in random matrix theory, and the method is essentially complex analysis. From the view point of random matrix theory, in \cite{BaiY,Bao,Xie} the authors have used Stieltjes transform to study similar matrix models with success, however, our matrices, arising from general linear codes over finite fields with dual distance 5, possess characteristics significantly different from \cite{BaiY,Bao,Xie}. With applications in mind, say, to generate pseudo-random matrices efficiently via linear codes, our matrices are more natural and interesting. None of the methods in previous works seem to apply directly to our setting. Instead we adopt methods from \cite{ConI,SA,Local law} and use a combination of ideas to obtain our final results.
Related to this paper, the authors in \cite{CTX} have used Stieltjes transform to obtain a strong convergence rate which is similar in nature to Theorem \ref{thm1-2} of this paper, hence extending the work \cite{OQBT}, and some of the arguments are similar.
The paper is organized as follows. In Section \ref{pre} we introduce Stieltjes transform and related formulas and lemmas which will play important roles later. The main ideas of proving Theorems \ref{thm} and \ref{thm1-2} share some similarity but technically speaking, they are quite involved, with the latter being even more so. To streamline the idea of the proofs, we assume a major technical statement (Theorem \ref{thm2}) from which we prove Theorems \ref{thm} and \ref{thm1-2} in Sections \ref{proof} and \ref{proof1-2} respectively. Finally we prove the required Theorem \ref{thm2} in Section \ref{proof2}.
\section{Preliminaries}\label{pre}
\subsection{Linear codes over $\mathbb{F}_q$ of dual distance at least 5}
The standard additive character $\psi: \mathbb{F}_q \to \mathbb{C}^*$ is given by \begin{equation}\label{char} \psi(a) = \zeta^{\mathrm{Tr}(a)}, \quad \forall a \in \mathbb{F}_q, \end{equation} where $\mathrm{Tr}$ is the absolute trace mapping from $\mathbb{F}_q$ to its prime subfield $\mathbb{F}_r$ of order $r$ and $\zeta=\exp(2\pi\sqrt{-1}/r)$ is a (complex) primitive $r$-th root of unity. In particular when $q=r=2$, then $\zeta=-1$ and $\psi(a)=(-1)^a$ for $a \in \mathbb{F}_2$. It is known that $\psi$ satisfies the following orthogonality relation: \begin{equation} \label{2:orth} \frac{1}{q}\sum_{x \in \mathbb{F}_q} \psi(ax)= \left\{ \begin{array}{lll} 1 &:& \mbox{ if } a = 0;\\ 0 &:& \mbox{ if } a \in \mathbb{F}_q \setminus \{0\}. \end{array}\right. \end{equation}
Let $\mathcal{C}$ be an $[n,k]_q$ linear code with dual distance $d^\bot \ge 5$. By the sphere-packing bound \cite[Theorem 1.12.1]{FECC}, we have $$\#\C^\bot =q^{n-k} \leq \frac{q^n}{1+n(q-1)+\binom{n}{2}(q-1)^2} =O\left(\frac{q^n}{n^2}\right),$$ here the implied constant in the big O-notation depends only on $q$. From this we can obtain \begin{eqnarray} \label{le:db} \frac{n^2}{q^k}=O(1). \end{eqnarray} Since $\mathcal{C}$ is linear, the orthogonal relation (\ref{2:orth}) further implies that for any $\mathbf{a} \in \mathbb{F}_q^n$, we have \begin{equation}\label{lem} \frac{1}{\#\mathcal{C}}\sum_{\mathbf{c} \in \mathcal{C}}\psi(\mathbf{a}\cdot \mathbf{c})=\left\{\begin{array}{lll} 1 &:& \mbox{ if } \mathbf{a} \in \C^\bot,\\ 0 &:& \mbox{ if } \mathbf{a} \notin \C^\bot. \end{array}\right. \end{equation} Here $\mathbf{a}\cdot\mathbf{c}$ is the usual inner product between the vectors $\mathbf{a}$ and $\mathbf{c}$ in $\mathbb{F}_q^n$.
\subsection{Stieltjes Transform} In this section we recall some basic knowledge of Stieltjes transform. Interested readers may refer to \cite[Chapter B.2]{SA} for more details. Stieltjes transform can be defined for any real function of bounded variation. For the case of interest to us, however, we confine ourselves to functions arising from probability theory.
Let $\mu$ be a probability measure and let $F$ be the corresponding cumulative distribution function. The Stieltjes transform of $F$ or $\mu$ is defined by $$s(z):=\int_{-\infty}^\infty \frac{\mathrm{d} F(x)}{x-z}=\int_{-\infty}^\infty \frac{\mu(\mathrm{d} x)}{x-z}, $$ where $z$ is a complex variable taking values in $\mathbb{C}^+:=\{z \in \mathbb{C}: \Im z > 0\}$, the upper half complex plane. Here $\Im z$ is the imaginary part of $z$.
It is known that $s(z)$ is well-defined for all $z \in \mathbb{C}^+$ and is well-behaved, satisfying the following properties: \begin{itemize} \item[(i).] $s(z) \in \mathbb{C}^+$ for any $z \in \mathbb{C}^+$;
\item[(ii).] $s(z)$ is analytic in $\mathbb{C}^+$ and
\begin{eqnarray} \label{2:lips} \left|\frac{\mathrm{d} s(z)}{\mathrm{d} z}\right| \leq \int_{-\infty}^\infty \frac{\mu(\mathrm{d} x)}{|x-z|^2} \le \frac{1}{\eta^2},\end{eqnarray} where $\eta=\Im z>0$;
\item[(iii).] the probability measure $\mu$ can be recovered from the Stieltjes transform $s(z)$ via the inverse formula (see \cite{SA}): \begin{equation}\label{2:inverse} \mu((x_1,x_2])=F(x_2)-F(x_1)=\lim_{\eta \to 0^+}\frac{1}{\pi}\int_{x_1}^{x_2} \Im(s(E+\mathrm{i}\eta)) \, \mathrm{d} E; \end{equation}
\item[(iv).] the convergence of Stieltjes transforms is equivalent to the convergence of the underlying probability measures (see for example \cite[Theorem B.9]{SA}). \end{itemize}
\subsection{Resolvent Identities and Formulas for Green function entries}
Let $M$ be a Hermitian $p \times p$ matrix whose $(j,k)$-th entry is $M_{jk}$. Denote by $G$ the Green function of $M$, that is, \[G:=G(z)=(M-zI_p)^{-1},\] where $z \in \mathbb{C}^{+}$. The $(j,k)$-th entry of $G$ is $G_{jk}$.
Given any subset $T \subset [1\mathrel{{.}\,{.}}\nobreak p]:=\{1,2,\cdots,p\}$, let $M^{(T)}$ be the $p \times p$ matrix whose $(j,k)$-th entry is given by $(M^{(T)})_{jk}:=\mathbbm{1}_{j,k \notin T}M_{jk}$. In addition, let $G^{(T)}$ be the Green function of $M^{(T)}$, that is, \[ G^{(T)}:=G^{(T)}(z)=(M^{(T)}-zI_p)^{-1}.\]
When $T$ is a singleton, say $\{\ell\}$, it is common to further abbreviate the notation $G^{(\{\ell\})}$ as $G^{(\ell)}$, and similar for other matrices.
Let $\mathbf{m}_{\ell}$ denote the $\ell$-th column of $M$. For $z \in \mathbb{C}^{+}$ and any $\ell \in [1\mathrel{{.}\,{.}}\nobreak p] \setminus T$, we have the \emph{Schur complement formula} (see \cite{SA, Local law}) \begin{equation}\label{diagonal} \frac{1}{G_{\ell\ell}^{(T)}}=M_{\ell\ell}-z-\mathbf{m}_{\ell}^*G^{(T\ell)}\mathbf{m}_{\ell}, \end{equation} where $G^{(T\ell)}:=G^{(T \cup \{\ell\})}$ and $\mathbf{m}_{\ell}^*$ is the conjugate transpose of $\mathbf{m}_{\ell}$.
We also have the following eigenvalue interlacing property (see \cite{SA, Local law})
\begin{equation}\label{Interlacing} |{\bf Tr} G^{(T)}(z)-{\bf Tr} G(z)| \leq C\eta^{-1}, \end{equation} where $z=E+\mathrm{i}\eta \in \mathbb{C}^{+}$, ${\bf Tr}$ is the trace function, and $C$ is a constant depending only on the set $T$.
\subsection{Stieltjes Transform of the Semicircle Law} The Stieltjes transform $s_\mathrm{SC}$ of the semicircle distribution given in (\ref{scpdf}) can be computed as (see \cite{SA}) \begin{equation}\label{SSC} s_\mathrm{SC}(z)=\frac{-z+\sqrt{z^2-4}}{2}. \end{equation} Here and throughout this paper, we always pick the complex square root $\sqrt{\cdot}$ to be the one with positive imaginary part.
It is well-known that $s_\mathrm{SC}(z)$ is the unique function that satisfies the equation \begin{equation}\label{SSC2} u(z)=\frac{1}{-z-u(z)} \end{equation} such that $\Im u(z) > 0$ whenever $\eta:=\Im z > 0$.
\subsection{Convergence of Stieltjes Transform in Probability} In order to bound the convergence rate of a random Stieltjes transform in probability, we need the following well-known McDiarmid's lemma from probability theory (see \cite[Lemma F.3]{Local law}).
\begin{lemma}[McDiarmid]\label{leMic} Let $X_1,\cdots,X_p$ be independent random variables taking values in the spaces $E_1,\cdots,E_p$ respectively. Let $$f: E_1 \times \cdots E_p \to \mathbb{R}$$ be a measurable function and define the random variable $Y=f(X_1,\cdots,X_p)$. Define, for each $k \in [1\mathrel{{.}\,{.}}\nobreak p]$, \begin{equation}\label{ck}
c_k:=\sup |f(x_1,\cdots,x_{k-1},y,x_{k+1},\cdots,x_p)-f(x_1,\cdots,x_{k-1},z,x_{k+1},\cdots,x_p)|, \end{equation} where the supremum is taken over all $x_j \in E_j$ for $j \neq k$ and $y,z \in E_k$. Then for any $\varepsilon > 0$, we have \begin{equation}\label{Mic}
\mathbb{P}\left(|Y-\mathbb{E} Y| \geq \varepsilon \right) \leq 2\exp\left(-\frac{2\varepsilon^2}{c_1^2+\cdots+c_p^2}\right). \end{equation} \end{lemma} We will need the following concentration inequality. We remark that a very similar concentration inequality was proved (see \cite[Lemma F.4]{Local law}). Here for the sake of completeness, we provide a detailed proof.
\begin{lemma}\label{deviation} Let $\mathcal{M}$ be a $p \times n$ random matrix with independent rows, define $S=(n/p)^{1/2}(\mathcal{M}\mathcal{M}^*-I_p)$. Let $s(z)$ be the Stieltjes transform of the empirical spectral distribution of $S$. Then for any $\varepsilon > 0$ and $z =E+i \eta \in \mathbb{C}^+$,
$$\mathbb{P}\left(|s(z)-\mathbb{E} s(z)| \geq \varepsilon \right) \leq 2\exp\left(-\frac{p\eta^2\varepsilon^2}{8}\right).$$ \end{lemma} \begin{proof}[Proof of Lemma \ref{deviation}] Applying Lemma \ref{leMic}, we take $X_j$ to be the $j$-th row of $\mathcal{M}$ and the function $f$ to be the Stieltjes transform $s$. Note that the $(j,k)$-th entry of $S$ is a linear function of the inner product of the $j$-th and $k$-th rows of $\mathcal{M}$. Hence changing one row of $\mathcal{M}$ only gives an additive perturbation of $S$ of rank at most two. Applying the resolvent identity \cite[(2.3)]{Local law}, we see that the Green function is also only affected by an additive perturbation by a matrix of rank at most two and operator norm at most $2\eta^{-1}$. Therefore the quantities $c_k$ in (\ref{ck}) can be bounded by $$c_k \leq \frac{4}{p\eta}.$$ Then the required result follows directly from inserting the above bound to (\ref{Mic}). \end{proof}
\section{Proof of Theorem \ref{thm}}\label{proof}
Throughout the paper, let $\mathcal{C}$ be an $[n,k]_q$ linear code over $\mathbb{F}_q$. We always assume that its dual distance satisfies $d^\bot \ge 5$. Denote $N=q^k$. The standard additive character on $\mathbb{F}_q$ extends component-wise to a natural mapping $\psi: \mathbb{F}_q^{n} \to \mathbb{C}^{n}$. Define $\mathcal{D}=\psi(\mathcal{C})$.
\subsection{Problem set-up}
Theorems \ref{thm} and \ref{thm1-2} are for random matrices in the probability space $\Omega_{p,I}$ of choosing $p$ distinct elements uniformly from $\mathcal{D}$. Denote by $\mathcal{D}^p$ the probability space of choosing $p$ elements from $\mathcal{D}$ independently and uniformly. Because $d^\bot \ge 5$, from (\ref{le:db}) we have $$\frac{\#\mathcal{D}^p}{\#\Omega_{p,I}}=\frac{N^p}{N(N-1)(N-2)\cdots(N-p+1)}=1+O\left(\frac{p^2}{N}\right) \to 1,$$ as $n,p \to \infty$. Thus to prove Theorems \ref{thm} and \ref{thm1-2}, it is equivalent to consider the larger probability space $\mathcal{D}^p$. This will simplify the proofs.
Now let $\Phi_n$ be a $p \times n$ random matrix whose rows are picked from $\mathcal{D}$ uniformly and independently. Denote by $\mathbb{E}$ the expectation with respect to the probability space $\mathcal{D}^p$. We may assume that $p:=p(n)$ is a function of $n$ such that $p,n/p \to \infty$ as $n \to \infty$.
Let \begin{eqnarray} \label{3:gn} \mathcal{G}_n=\frac{1}{n}\Phi_n\Phi_n^*, \quad M_n=\sqrt{\frac{n}{p}}(\mathcal{G}_n-I_p).\end{eqnarray} Let $\mu_n$ be the empirical spectral measure of $M_n$ and let $s_{M_n}(z)$ be its Stieltjes transform, that is, $$s_{M_n}(z)=\frac{1}{p}\sum_{j=1}^p \frac{1}{\lambda_j-z}=\frac{1}{p}{\bf Tr} G. $$ Here $\lambda_1,\cdots,\lambda_p$ are the eigenvalues of the matrix $M_n$, and $G:=G(z)$ is the Green function of $M_n$ given by \[ G(z)=(M_n-zI_p)^{-1}.\] Note that the Stieltjes transform $s_{M_n}(z)$ is itself a random variable in the space $\mathcal{D}^p$. We define \begin{equation}\label{EStietjes} s_n(z):=\mathbb{E} s_{M_n}(z)=\frac{1}{p}\mathbb{E} {\bf Tr} G. \end{equation} Throughout the paper, the complex value $z \in \mathbb{C}^+$ is always written as \[z=E+ \mathrm{i} \eta, \quad \mbox{ where } E, \eta \in \mathbb{R} \mbox{ and } \eta >0. \] For a fixed constant $\tau \in (0,1)$, we define \begin{equation}\label{st}
\Gamma_{\tau}:=\bigg\{z=E+\mathrm{i}\eta: |E| \leq \tau^{-1}, 0< \eta \leq \tau^{-1} \bigg\}. \end{equation} Now we assume a result about the expected Stieltjes transform $s_n(z)$. \begin{theorem}\label{thm2} For any $z \in \Gamma_{\tau}$, we write \begin{eqnarray} \label{3:snz} s_n(z)=\frac{1}{-z-s_n(z)+\Delta(z)}. \end{eqnarray} Then we have \[ \Delta(z)=O_{\tau}\left(\eta^{-3} \left(p^{-1}+\sqrt{p/n}\right)\right).\] \end{theorem}
We emphasize here that this is one of the major technical results in this paper and the proof is a little complicated. This is the only result in the paper that is directly related to the properties of linear codes. It requires $d^\bot \geq 5$ but not the extra condition (\ref{1:srip}) used in \cite{CSC}. To streamline the presentation, here we assume Theorem \ref{thm2}, then Theorem \ref{thm} can be proved easily. The proof of Theorem \ref{thm2} is postponed to Section \ref{proof2}.
\subsection{Proof of Theorem \ref{thm}}\label{sec}
By properties of the Stieltjes transform (see \cite[Theorem B.9]{SA}), to prove Theorem \ref{thm}, it is equivalent to prove the following statement: \emph{For any $\varepsilon>0$, we have}
\begin{eqnarray} \label{3:conP} \mathbb{P}\left(\exists z \in \mathbb{C}^+ \mbox{ such that } \left|s_{M_n}(z) - s_\mathrm{SC}(z) \right| \ge \varepsilon \right) \to 0 \quad \mbox{ as } n \to \infty. \end{eqnarray} We prove Statement (\ref{3:conP}) in several steps.
First, we fix an arbitrary value $z \in \mathbb{C}^+$. The quadratic equation (\ref{3:snz}) has two solutions $$s_n^{\pm}(z)=\frac{-(z-\Delta)\pm \sqrt{(z-\Delta)^2-4}}{2}.$$ As $n \to \infty$, from Theorem \ref{thm2} we have $\Delta(z) \to 0$, so $z-\Delta \in \mathbb{C}^+$ for large enough $n$. Since $s_n(z),s_\mathrm{SC}(z) \in \mathbb{C}^+$, we see that \begin{equation}\label{3:snz2} s_n(z)=s_n^+(z)=s_\mathrm{SC}(z-\Delta). \end{equation} Then by the continuity of $s_\mathrm{SC}$ and by taking $n \to \infty$, we obtain \begin{equation}\label{3:limit}
s_n(z) \to s_\mathrm{SC}(z). \end{equation} Moreover, by Lemma \ref{deviation}, for any fixed $\varepsilon >0$, as $n \to \infty$, we have
\[\mathbb{P}\left(\left|s_{M_n}(z)-s_n(z)\right| \geq \varepsilon\right) \to 0.\] This and (\ref{3:limit}) immediately imply \begin{equation}\label{3:prob}
\mathbb{P}\left(\left|s_{M_n}(z)-s_\mathrm{SC}(z)\right| \geq \varepsilon\right) \to 0. \end{equation}
Noting that (\ref{3:prob}) holds for any fixed $z \in \mathbb{C}^+$ and any $\varepsilon>0$, so to prove (\ref{3:conP}), in the next step we need to show that the convergence is ``uniform'' for all $z \in \mathbb{C}^+$. To do this, we adopt a simple lattice argument.
For any $\tau,\varepsilon \in (0,1)$, define the sets $$\Gamma_{\tau}':=\Gamma_{\tau} \cap \{z=E+\mathrm{i} \eta: \eta \geq \tau\}$$ and $$\mathbf{L}_{\tau,\ep}:=\Gamma_{\tau}' \cap \left\{z=\frac{\tau^2\varepsilon}{4}(a+\mathrm{i} b): (a,b) \in \mathbb{Z}^2\right\}.$$ It is easy to see that $\mathbf{L}_{\tau,\ep} \neq \emptyset$ and $$\#\mathbf{L}_{\tau,\ep} = O_\tau\left(\tau^{-4}\varepsilon^{-2}\right) < \infty.$$ For any fixed $z \in \mathbb{C}^+$, define $\Xi_{n,\varepsilon}(z)$ to be the event
$$\left\{|s_{M_n}(z)-s_\mathrm{SC}(z)| < \varepsilon \right\}.$$ By (\ref{3:prob}), for any $\delta > 0$, there is an $N(z,\tau,\varepsilon,\delta)$ such that $$n > N(z,\tau,\varepsilon,\delta) \implies \mathbb{P}\left(\Xi_{n,\frac{\varepsilon}{2}}(z)^{\bf c}\right) < \frac{\delta}{\#\mathbf{L}_{\tau,\ep}}.$$ Here the set $\Xi_{n,\frac{\varepsilon}{2}}(z)^{\bf c}$ denotes the complement of the event $\Xi_{n,\frac{\varepsilon}{2}}(z)$. Then for any $n$ such that \[ n > N(\tau,\varepsilon,\delta):=\max_{z \in \mathbf{L}_{\tau,\ep}} N(z,\tau,\varepsilon,\delta),\] we have $$\mathbb{P}\left(\left(\bigcap_{z \in \mathbf{L}_{\tau,\ep}} \Xi_{n,\frac{\varepsilon}{2}}(z)\right)^{\bf c}\right)=\mathbb{P}\left(\bigcup_{z \in \mathbf{L}_{\tau,\ep}} \Xi_{n,\frac{\varepsilon}{2}}(z)^{\bf c}\right) < \delta. $$
Finally we consider the event $\bigcap_{z \in \mathbf{L}_{\tau,\ep}} \Xi_{n,\frac{\varepsilon}{2}}(z)$, that is,
\[ \left|s_{M_n}(z')-s_\mathrm{SC}(z') \right|< \frac{\varepsilon}{2} \quad \forall z' \in \mathbf{L}_{\tau,\ep}.\] Recall from (\ref{2:lips}) that the Stieltjes transforms $s_{M_n}(z)$ and $s_\mathrm{SC}(z)$ are both $\tau^{-2}$-Lipschitz on the set $\Gamma_{\tau}'$, and for any $z \in \Gamma_{\tau}'$, we can find one $z' \in \mathbf{L}_{\tau,\ep}$ such that
$$|z-z'| \leq \frac{\tau^2\varepsilon}{4}. $$ So for this $z \in \Gamma_{\tau}'$ we have \begin{align*}
|s_{M_n}(z)-s_\mathrm{SC}(z)| &\leq |s_{M_n}(z)-s_{M_n}(z')|+|s_{M_n}(z')-s_\mathrm{SC}(z')|+|s_\mathrm{SC}(z')-s_\mathrm{SC}(z)|\\
&< \tau^{-2}|z-z'|+\frac{\varepsilon}{2}+\tau^{-2}|z-z'|\\ & \le \varepsilon. \end{align*} This means that $$\bigcap_{z \in \mathbf{L}_{\tau,\ep}} \Xi_{n,\frac{\varepsilon}{2}}(z) \subset \bigcap_{z \in \Gamma_{\tau}'} \Xi_{n,\varepsilon}(z).$$ Therefore $$\mathbb{P}\left(\left(\bigcap_{z \in \Gamma_{\tau}'} \Xi_{n,\varepsilon}(z)\right)^{\bf c}\right) \leq \mathbb{P}\left(\left(\bigcap_{z \in \mathbf{L}_{\tau,\ep}} \Xi_{n,\frac{\varepsilon}{2}}(z)\right)^{\bf c}\right) < \delta$$ for any $n > N(\tau,\varepsilon,\delta)$.
Hence for any $\tau, \varepsilon \in (0,1)$, we have
\begin{eqnarray*} \mathbb{P}\left(\exists z \in \Gamma_{\tau}' \mbox{ such that } \left|s_{M_n}(z) - s_\mathrm{SC}(z) \right| \ge \varepsilon \right) \to 0 \quad \mbox{ as } n \to \infty. \end{eqnarray*} Taking the limit $\tau \to 0^+$, we obtain the desired Statement (\ref{3:conP}). This completes the proof of Theorem \ref{thm}.
\section{Proof of Theorem \ref{thm1-2}}\label{proof1-2} Now for fixed constants $c > 1$ and $\gamma_1,\gamma_2 \in (0,1)$, let us assume \[ c^{-1}n^{\gamma_1} \leq p \leq cn^{\gamma_2}.\] Similar in proving Theorem \ref{thm} in the previous section, here we assume Theorem \ref{thm2}. Then the main idea of proving Theorem \ref{thm1-2} is to provide a refined and quantitative version of Statement (\ref{3:conP}), so in each step of the proofs, we need to keep track of all the varying parameters as $n \to \infty$.
First, the upper bound for $\Delta(z)$ in Theorem \ref{thm2} can be simplified as \[ \Delta(z)=O_{c,\tau}\left(n^{-4\beta}\eta^{-3}\right), \] where the constant $\beta>0$ is explicitly given in (\ref{beta}).
Let us define \begin{eqnarray*} \mathbf{S}_{\tau}:=\Gamma_{\tau} \bigcap \left\{z=E+\mathrm{i} \eta: \eta \ge n^{-\beta+\tau}\right\}. \end{eqnarray*}
From now on, $C_{c,\tau}$ denotes some positive constant depending only on $c$ and $\tau$ whose value may vary at each occurrence. We can estimate the difference $|s_n(z)-s_\mathrm{SC}(z)|$ as follows.
\begin{lemma}\label{EStieltjes} For any $z \in \mathbf{S}_{\tau}$, we have
$$|s_n(z)-s_\mathrm{SC}(z)|=O_{c,\tau}\left(n^{-4\beta}\eta^{-4}\right).$$ \end{lemma} \begin{proof}[Proof of Lemma \ref{EStieltjes}] First, for large enough $n$, noting that $$\Im(z-\Delta)\geq \eta-C_{c,\tau}n^{-4\beta}\eta^{-3} > 0,$$ we see that Equation (\ref{3:snz2}) holds for all $z \in \mathbf{S}_{\tau}$. More precisely, we have $$\Im(z-\Delta) > C_{c,\tau}\eta.$$
By using the fact $\left|\frac{\mathrm{d} s_\mathrm{SC}(z)}{\mathrm{d} z}\right| \leq \eta^{-1}$ which can be easily checked from (\ref{SSC}), we conclude that
$$|s_n(z)-s_\mathrm{SC}(z)|=\left|s_\mathrm{SC}(z-\Delta)-s_\mathrm{SC}(z) \right| \leq C_{c,\tau}\eta^{-1}|\Delta|\leq C_{c,\tau}n^{-4\beta}\eta^{-4}.$$ Then Lemma \ref{EStieltjes} is proved. \end{proof}
Next we estimate the term $|s_{M_n}(z)-s_\mathrm{SC}(z)|$. An $n$-dependent event $\Xi$ is said to hold \emph{with high probability} if for any $D>0$, there is a quantity $N=N(D)>0$ such that $\mathbb{P}(\Xi) \geq 1-n^{-D}$ for any $n > N$.
\begin{theorem}\label{Stieltjes} We have, with high probability,
$$|s_{M_n}(z)-s_\mathrm{SC}(z)| \leq n^\tau(n^{-\beta}+n^{-4\beta}\eta^{-4}) \quad \forall z \in \mathbf{S}_{\tau}.$$ \end{theorem} \begin{proof}[Proof of Theorem \ref{Stieltjes}] By the concentration inequality given in Lemma \ref{deviation}, we have \begin{equation}\label{deviation1}
\mathbb{P}\left(\left|s_{M_n}(z)-s_n(z)\right| \geq n^{\frac{\tau}{2}-\beta}\right)\leq 2\exp\left(-\frac{n^{\gamma_1-4\beta+3\tau}}{8c}\right)\leq 2\exp\left(-\frac{n^{3\tau}}{8c}\right). \end{equation} Noting that the inequality (\ref{deviation1}) holds for any fixed $z \in \mathbf{S}_{\tau}$. In order to prove Theorem \ref{Stieltjes}, we need an upper bound which is uniform for all $z \in \mathbf{S}_{\tau}$. We apply a lattice argument again.
Let $$\mathbf{L}_\tau:=\mathbf{S}_{\tau} \cap \left\{z=n^{-3 \beta} (a+\mathrm{i} b): (a,b) \in \mathbb{Z}^2 \right\}.$$ Note that the set $\mathbf{L}_\tau \ne \emptyset$ and $$\#\mathbf{L}_\tau \leq C_\tau n^{6\beta}.$$ Also, for any $z \in \mathbf{S}_{\tau}$ and $\varepsilon > 0$, define $\mathcal{E}_{n,\varepsilon}(z)$ to be the event
$$ \left\{|s_{M_n}(z)-s_n(z)| \leq n^{\varepsilon-\beta} \right\},$$ and $\mathcal{E}_{n,\varepsilon}(z)^{\bf c}$ the complement. Then (\ref{deviation1}) can be rewritten as $$\mathbb{P}\left(\mathcal{E}_{n,\frac{\tau}{2}}(z)^{\bf c} \right) \leq 2\exp\left(-\frac{n^{3\tau}}{8c}\right).$$ So we have \begin{eqnarray} \mathbb{P}\left(\left(\bigcap_{z \in \mathbf{L}_\tau} \mathcal{E}_{n,\frac{\tau}{2}}(z)\right)^{\bf c}\right)&=&\mathbb{P}\left(\bigcup_{z \in \mathbf{L}_\tau} \mathcal{E}_{n,\frac{\tau}{2}}(z)^{\bf c}\right) \nonumber \\ &\leq & C_\tau n^{6\beta}\exp\left(-\frac{n^{3\tau}}{8c}\right) \leq n^{-D} \label{deviation2} \end{eqnarray} for any $D > 0$ and $n > N(c,\gamma_1,\gamma_2,\tau,D)$.
Finally we consider the event $\bigcap_{z \in \mathbf{L}_\tau} \mathcal{E}_{n,\frac{\tau}{2}}(z)$, that is,
\[|s_{M_n}(z')-s_n(z')| \le n^{\frac{\tau}{2}-\beta} \quad \forall z' \in \mathbf{L}_\tau.\] Noting that for any $z \in \mathbf{S}_{\tau}$, there is $z' \in \mathbf{L}_\tau$ such that
$$|z-z'| \leq n^{-3\beta}$$ and that $s_{M_n}(z)$ and $s_n(z)$ are both $n^{2\beta}$-Lipschitz on $\mathbf{S}_{\tau}$, we obtain, for any $z \in \mathbf{S}_{\tau}$, \begin{align*}
\left|s_{M_n}(z)-s_n(z) \right| &\leq |s_{M_n}(z)-s_{M_n}(z')|+|s_{M_n}(z')-s_n(z')|+|s_n(z')-s_n(z)|\\
&< 2n^{2\beta}|z-z'|+n^{\frac{\tau}{2}-\beta} \leq n^{\tau-\beta}. \end{align*} This means that $$\bigcap_{z \in \mathbf{L}_\tau} \mathcal{E}_{n,\frac{\tau}{2}}(z) \subset \bigcap_{z \in \mathbf{S}_{\tau}} \mathcal{E}_{n,\tau}(z).$$ Hence by (\ref{deviation2}) we have $$\mathbb{P}\left(\bigcap_{z \in \mathbf{S}_{\tau}} \mathcal{E}_{n,\tau}(z)\right) \geq \mathbb{P}\left(\bigcap_{z \in \mathbf{L}_\tau} \mathcal{E}_{n,\frac{\tau}{2}}(z)\right) \geq 1-n^{-D}$$ for all $n > N(c,\gamma_1,\gamma_2,\tau,D)$.
Combining the above inequality with Lemma \ref{EStieltjes} completes the proof of Theorem \ref{Stieltjes}. \end{proof}
\begin{proof}[Proof of Theorem \ref{thm1-2}] As a standard application of the Helffer-Sj\"{o}strand formula via complex analysis, Theorem \ref{thm1-2} can be derived directly from Theorem \ref{Stieltjes}. This is quite well-known, and the computation is routine. Interested readers may refer to \cite[Section 8]{Local law} for a very similar analysis. We omit the details. \end{proof}
\section{Proof of Theorem \ref{thm2}}\label{proof2} In this section we give a detailed proof of Theorem \ref{thm2}, where the condition that $d^\bot \geq 5$ plays an important role.
Recall from the beginning of Section \ref{proof} that $\mathcal{C}$ is a linear code of length $n$ over $\mathbb{F}_q$ with $d^\bot \geq 5$, $\psi$ is the standard additive character on $\mathbb{F}_q$, extended component-wisely to $\mathbb{F}_q^n$, $\mathcal{D}=\psi(\mathcal{C})$, and $\Phi_n$ is a $p \times n$ random matrix whose rows are selected uniformly and independently from $\mathcal{D}$. This makes $\mathcal{D}^p$ a probability space, on which we use $\mathbb{E}$ to denote the expectation. Let $\mathcal{G}_n$ and $M_n$ be defined as in (\ref{3:gn}). Since all the entries of $\Phi_n$ are roots of unity, the diagonal entries of $M_n$ are all zero.
Let $x_{jk}$ be the $(j,k)$-th entry of $\Phi_n$. The following properties of $x_{jk}$, while very simple, depend crucially on the condition that $d^\bot \ge 5$.
\begin{lemma}\label{cor} For any $\ell \in [1\mathrel{{.}\,{.}}\nobreak p]$, we have
(a) $\mathbb{E}(x_{\ell j}\overline{x}_{\ell k})=0$ if $j \neq k$;
(b) $\mathbb{E}(x_{\ell j}x_{\ell t}\overline{x}_{\ell k}\overline{x}_{\ell s})=0$ if the indices $j,t,k,s$ do not come in pairs; If the indices come in pairs, then $|\mathbb{E}(x_{\ell j}x_{\ell t}\overline{x}_{\ell k}\overline{x}_{\ell s})| \leq 1$. \end{lemma} \begin{proof}[Proof of Lemma \ref{cor}] (a) It is easy to see that \[\mathbb{E}(x_{\ell j}\overline{x}_{\ell k})=\frac{1}{N}\sum_{\mathbf{c} \in \mathcal{C}}\psi(c_j-c_k)=\frac{1}{N}\sum_{\mathbf{c} \in \mathcal{C}}\psi(\mathbf{a}_1 \cdot \mathbf{c}),\] where $\mathbf{c}=(c_1,c_2,\cdots,c_n) \in \mathcal{C}$ and $\mathbf{a}_1=(0, \cdots, 0, 1, 0 \cdots, 0, -1, 0 \cdots, 0) \in \mathbb{F}_q^n$. Here in $\mathbf{a}_1$ the 1 and $-1$ appear at the $j$-th and $k$-th entries respectively. Since $d^\bot \geq 5$, we have $\mathbf{a}_1 \notin \C^\bot$, and the desired result follows directly from (\ref{lem}).
(b) It is easy to see that \[\mathbb{E}(x_{\ell j}x_{\ell t}\overline{x}_{\ell k}\overline{x}_{\ell s})=\frac{1}{N}\sum_{\mathbf{c} \in \mathcal{C}}\psi(c_j+c_t-c_k-c_s)=\frac{1}{N}\sum_{\mathbf{c} \in \mathcal{C}}\psi(\mathbf{a}_2 \cdot \mathbf{c}),\]
where the vector $\mathbf{a}_2 \in \mathbb{F}_q^n$ is formed from the all-zero vector by adding $1$s to the $j$-th and $t$-th entries and then adding $-1$s from the $k$-th and $s$-th entries. If the indices $j,t,k,s$ do not come in pairs, then $0 \ne \mathrm{wt}(\mathbf{a}_2) \le 4$. Since $d^\bot \geq 5$, we have $\mathbb{E}(x_{\ell j}x_{\ell t}\overline{x}_{\ell k}\overline{x}_{\ell s})=0$ by (\ref{lem}). The second statement of \emph{(b)} is trivial since $|x_{ij}|=1$ for any $i,j$. \end{proof}
For any $\ell \in [1 \mathrel{{.}\,{.}}\nobreak p]$, let $\Phi_n^{(\ell)}$ be the $p \times n$ matrix obtained from $\Phi_n$ by changing the whole $\ell$-th row to {\bf 0}. Define \[ \G^{(\ell)}:=\frac{1}{n}\Phi_n^{(\ell)}{\Phi_n^{(\ell)}}^*, \quad M_n^{(\ell)}:=\sqrt{\frac{n}{p}}\left(\G^{(\ell)}-I_p\right).\] Denote by $\omega(\ell)$ the $\ell$-th row of $\Phi_n$, and $\mathbf{m}_{\ell}$ the $\ell$-th column of $M_n$. It is easy to see that \begin{eqnarray} \label{4:ml} \mathbf{m}_{\ell}=\frac{1}{\sqrt{pn}}\Phi_n^{(\ell)}\omega(\ell)^*.\end{eqnarray} Let \[ G:=G(z)=\left(M_n-z I_p\right)^{-1}, \quad G^{(\ell)}:=G^{(\ell)}(z)=\left(M_n^{(\ell)}-z I_p\right)^{-1}\] be the Green functions of $M_n$ and $M_n^{(\ell)}$ respectively for the complex variable $z \in \mathbb{C}^+$.
For the Green function $G$, we start with the resolvent identity (\ref{diagonal}) for $T=\emptyset$. Using (\ref{4:ml}), we can express the third term on the right side of (\ref{diagonal}) as \begin{align*} \mathbf{m}_{\ell}^*G^{(\ell)}\mathbf{m}_{\ell}&=\frac{1}{pn}\omega(\ell)\Phi_n^{(\ell)*}G^{(\ell)}\Phi_n^{(\ell)}\omega(\ell)^* \nonumber\\ &=\frac{1}{pn}{\bf Tr} \left({\Phi_n^{(\ell)}}^*G^{(\ell)}\Phi_n^{(\ell)}\omega(\ell)^*\omega(\ell) \right). \end{align*} By the identity $$(\omega(\ell)^*\omega(\ell))_{jk}=\delta_{jk}+(1-\delta_{jk})x_{\ell j}\overline{x}_{\ell k},$$ the right hand side can be further expressed as \begin{align*} \frac{1}{pn}{\bf Tr} \left({\Phi_n^{(\ell)}}^*G^{(\ell)}\Phi_n^{(\ell)}\right)+Z_\ell =\frac{1}{p}{\bf Tr} \left(G^{(\ell)}\G^{(\ell)}\right)+Z_\ell, \end{align*} where \begin{equation}\label{Zl} Z_\ell=\sum_{j \neq k} a_{jk} x_{\ell j}\overline{x}_{\ell k}. \end{equation} Here the indices $j,k$ vary in $[1\mathrel{{.}\,{.}}\nobreak n]$ and $a_{jk}$'s are the $(jk)$-th entry of the $n \times n$ matrix $(a_{jk})$ given by \begin{equation}\label{ajk} (a_{jk})=\frac{1}{pn}{\Phi_n^{(\ell)}}^*G^{(\ell)}\Phi_n^{(\ell)}. \end{equation} Hence the resolvent identity (\ref{diagonal}) yields \begin{align*} \frac{1}{G_{\ell\ell}}&=M_{\ell\ell}-z-\frac{1}{p}{\bf Tr} \left(G^{(\ell)}\G^{(\ell)}\right)-Z_\ell \\ &=-z-\frac{1}{p}{\bf Tr} G^{(\ell)}\left(\sqrt{\frac{p}{n}}M_n^{(\ell)}+I_p\right)-Z_\ell. \end{align*} Expanding the second term on the right, we obtain \begin{align} \frac{1}{G_{\ell\ell}}=-z-s_n(z)+Y_\ell,\label{diagonal2} \end{align} where \begin{equation}\label{Yl} Y_\ell=s_n(z)-\sqrt\frac{p}{n}-\left(\frac{1}{p}+\frac{z}{\sqrt{pn}}\right){\bf Tr} G^{(\ell)}-Z_\ell. \end{equation}
\subsection{Estimates of $Z_\ell$ and $Y_\ell$}
The random variables $Z_{\ell}$ and $Y_{\ell}$ depend on the complex value $z =E+\mathrm{i} \eta \in \mathbb{C}^+$. For any fixed constant $\tau>0$, recall $\Gamma_{\tau}$ defined in (\ref{st}). Throughout this section we always assume $z \in \Gamma_{\tau}$.
\begin{lemma}\label{Zl2} Let $z \in \Gamma_{\tau}$. Then for any $\ell \in [1\mathrel{{.}\,{.}}\nobreak p]$, we have
(a) $\mathbb{E}^{(\ell)}Z_\ell=\mathbb{E}Z_\ell=0$. Here $\mathbb{E}^{(\ell)}$ is the conditional expectation given $\{x_{jk}: j \ne \ell\}$;
(b) $\mathbb{E}|Z_\ell|^2=O_{\tau}(p^{-1}\eta^{-2})$. \end{lemma} \begin{proof}[Proof of Lemma \ref{Zl2}] (a) Since the rows of $\Phi_n$ are independent, the entries $a_{jk}$ as defined in (\ref{ajk}) are independent with $x_{\ell j}$ and $x_{\ell k}$. Hence from the definition of $Z_\ell$ in (\ref{Zl}) and statement (a) of Lemma \ref{cor}, we have $$\mathbb{E}^{(\ell)}Z_\ell=\sum_{j \neq k}a_{jk}\mathbb{E}(x_{\ell j}\overline{x}_{\ell k})=0.$$ The proof of the result on $\mathbb{E}Z_\ell$ is similar by replacing $a_{jk}$ with $\mathbb{E} a_{jk}$.
(b) Expanding $|Z_\ell|^2$ and taking expectation $\mathbb{E}$ inside, noting that the rows of $\Phi_n$ are independent, we have \begin{align*}
\mathbb{E}|Z_\ell|^2&=\mathbb{E}\left|\sum_{j \neq k}a_{jk}x_{\ell j}\overline{x}_{\ell k}\right|^2\\ &=\sum_{\substack{j\neq k\\s\neq t}}\mathbb{E}(a_{jk}\overline{a}_{st})\mathbb{E}(x_{\ell j}x_{\ell t}\overline{x}_{\ell k}\overline{x}_{\ell s}). \end{align*} Since $d^\bot \geq 5$, by using statement (b) of Lemma \ref{cor}, we find
$$\mathbb{E}|Z_\ell|^2 \leq C\sum_{j,k} \mathbb{E}|a_{jk}|^2=C\mathbb{E} {\bf Tr} ((a_{jk})(a_{jk})^*),$$ where $C$ is an absolute constant which may be different in each appearance. Using the definition of $(a_{jk})$ in (\ref{ajk}) we have \begin{align*}
\mathbb{E}|Z_\ell|^2 &\leq\frac{C}{p^2}\mathbb{E} {\bf Tr} \left(\frac{1}{n^2}{\Phi_n^{(\ell)}}^*G^{(\ell)}\Phi_n^{(\ell)}{\Phi_n^{(\ell)}}^*G^{(\ell)}(\bar{z})\Phi_n^{(\ell)}\right)\\ &=\frac{C}{p^2}\mathbb{E} {\bf Tr} [(M_n^{(\ell)}-z)^{-1}\G^{(\ell)}(M_n^{(\ell)}-\bar{z})^{-1}\G^{(\ell)}]\\ &=\frac{C}{p^2}\mathbb{E} \sum_{j=1}^{p} \frac{\left(\sqrt\frac{p}{n}\lambda_j^{(\ell)}+1\right)^2}{(\lambda_j^{(\ell)}-z)(\lambda_j^{(\ell)}-\bar{z})}. \end{align*} Expanding the terms on the right, we can easily obtain \begin{align*}
\mathbb{E}|Z_\ell|^2
&\leq \frac{C}{p^2}\mathbb{E} \sum_{j=1}^{p}\left(\frac{p}{n}\left(1+\frac{|z|^2}{|\lambda_j^{(\ell)}-z|^2}\right)+\frac{1}{|\lambda_j^{(\ell)}-z|^2}\right)\\
&\leq C\left(\frac{1}{n}+\frac{|z|^2}{n\eta^2}+\frac{1}{p\eta^2}\right)\leq \frac{C_{\tau}}{p\eta^2}. \end{align*} Here $\lambda_j^{(\ell)} \in \mathbb{R} (1 \leq j \leq p)$ are the eigenvalues of $M_n^{(\ell)}$, and $C_{\tau}$ is a positive constant depending only on $\tau$ whose value may vary in each occurrence. \end{proof} The above estimations lead to the following estimations of $Y_\ell$. \begin{lemma}\label{Yl2} Let $z \in \Gamma_{\tau}$. Then for any $\ell \in [1\mathrel{{.}\,{.}}\nobreak p]$, we have
(a) $\mathbb{E} Y_\ell=O_{\tau}\left(\eta^{-1}(p^{-1}+(p/n)^{\frac{1}{2}})\right)$;
(b) $\mathbb{E}|Y_\ell|^2=O_{\tau}\left(\eta^{-2}(p^{-1}+p/n)\right)$. \end{lemma} \begin{proof}[Proof of Lemma \ref{Yl2}] (a) Taking expectation on $Y_\ell$ in (\ref{Yl}) and noting that $\mathbb{E} Z_\ell=0$, we get \begin{align*} \mathbb{E} Y_\ell &=\frac{1}{p}\mathbb{E}({\bf Tr} G-{\bf Tr} G^{(\ell)})-\sqrt\frac{p}{n}-\frac{z}{\sqrt{pn}}\mathbb{E}{\bf Tr}G^{(\ell)}. \end{align*}
By the eigenvalue interlacing property in (\ref{Interlacing}) and the trivial bound $|G_{jj}^{(\ell)}| \leq \eta^{-1}$, we get
$$|\mathbb{E} Y_\ell| \leq \frac{C}{p\eta}+\sqrt\frac{p}{n}+\sqrt\frac{p}{n}\frac{|z|}{\eta} \leq \frac{C}{p\eta}+\frac{C_{\tau}}{\eta}\sqrt\frac{p}{n}.$$
(b) We split $\mathbb{E}|Y_\ell|^2$ as \begin{equation}\label{Yl3}
\mathbb{E}|Y_\ell|^2=\mathbb{E}|Y_\ell-\mathbb{E}Y_\ell|^2+|\mathbb{E}Y_\ell|^2=V_1+V_2+|\mathbb{E}Y_\ell|^2, \end{equation} where
$$V_1=\mathbb{E}|Y_\ell-\mathbb{E}^{(\ell)}Y_\ell|^2, \quad V_2=\mathbb{E}|\mathbb{E}^{(\ell)}Y_\ell-\mathbb{E}Y_\ell|^2. $$ We first estimate $V_1$. Using (a) of Lemma \ref{Zl2}, we see that $$Y_\ell-\mathbb{E}^{(\ell)}Y_\ell=-Z_\ell+\mathbb{E}^{(\ell)}Z_\ell=-Z_\ell$$ Hence by (b) of Lemma \ref{Zl2} we obtain \begin{equation}\label{V1}
V_1=\mathbb{E}|Z_\ell|^2=O_{\tau}(p^{-1}\eta^{-2}). \end{equation} Next we estimate $V_2$. Again by Lemma \ref{Zl2} we have \begin{align*} \mathbb{E}^{(\ell)}Y_\ell-\mathbb{E}Y_\ell &=-\left(\frac{1}{p}+\frac{z}{\sqrt{pn}}\right)({\bf Tr} G^{(\ell)}-\mathbb{E}{\bf Tr} G^{(\ell)}). \end{align*} So we have \begin{align}\label{V2}
V_2&=\left|\frac{1}{p}+\frac{z}{\sqrt{pn}}\right|^2\mathbb{E}|{\bf Tr} G^{(\ell)}-\mathbb{E}{\bf Tr} G^{(\ell)}|^2\nonumber\\
&=\left|\frac{1}{p}+\frac{z}{\sqrt{pn}}\right|^2\sum_{m \neq \ell}\mathbb{E}\left|\mathbb{E}^{(T_{m-1})}{\bf Tr} G^{(\ell)}-\mathbb{E}^{(T_m)}{\bf Tr} G^{(\ell)}\right|^2. \end{align} Here we denote $T_0:=\emptyset$ and $T_m:=[1\mathrel{{.}\,{.}}\nobreak m]$ for any $m \in [1\mathrel{{.}\,{.}}\nobreak p]$, and for any subset $T \subset [1\mathrel{{.}\,{.}}\nobreak p]$, we denote $\mathbb{E}^{(T)}$ to be the conditional expectation given $\{x_{jk}: j \notin T\}$. The second equality follows from applying successively the law of total variance to the rows of $\Phi_n$.
For $m \neq \ell$, writing $\gamma_m:=\mathbb{E}^{(T_{m-1})}{\bf Tr} G^{(\ell)}-\mathbb{E}^{(T_m)}{\bf Tr} G^{(\ell)}$, we can easily check that $$\gamma_m=\mathbb{E}^{(T_{m-1})}\sigma_m-\mathbb{E}^{(T_m)}\sigma_m, $$
where $\sigma_m:={\bf Tr} G^{(\ell)}-{\bf Tr} G^{(\ell,m)}$. By (\ref{Interlacing}) we have $|\gamma_m| \leq C\eta^{-1}$. Hence we obtain
$$V_2 \leq C \left(\frac{1}{p^2}+\frac{|z|^2}{pn}\right)\left(\frac{p}{\eta^2}\right) \leq \frac{C_{\tau}}{p\eta^2}.$$
Plugging the estimates of $\mathbb{E}Y_\ell$ in statement (a), $V_1$ in (\ref{V1}) and $V_2$ above into the equation (\ref{Yl3}), we obtain the desired estimate of $\mathbb{E}|Y_\ell|^2$. \end{proof} \subsection{Proof of Theorem \ref{thm2}} We can now complete the proof of Theorem \ref{thm2}. \begin{proof}[Proof of Theorem \ref{thm2}] We write (\ref{diagonal2}) as \begin{eqnarray} \label{4:gll} G_{\ell\ell}=\frac{1}{\alpha_n+Y_\ell},\end{eqnarray} where $$\alpha_n=-z-s_n(z).$$ Taking expectations on both sides of (\ref{4:gll}), we can obtain \begin{equation}\label{diagonal3} \mathbb{E} G_{\ell\ell}=\frac{1}{\alpha_n}+A_\ell=\frac{1}{\alpha_n+\Delta_\ell}, \end{equation} where \begin{equation}\label{Al} A_\ell=\mathbb{E}\left(\frac{1}{\alpha_n+Y_\ell}\right)-\frac{1}{\alpha_n}=-\frac{1}{\alpha_n^2}\mathbb{E} Y_\ell+\frac{1}{\alpha_n^2}\mathbb{E} \left(\frac{Y_\ell^2}{\alpha_n+Y_\ell}\right), \end{equation} and \begin{equation}\label{dl} \Delta_\ell=\left(\frac{1}{\alpha_n}+A_\ell\right)^{-1}-\alpha_n=-\frac{\alpha_n^2 A_\ell}{1+\alpha_n A_\ell}. \end{equation}
For $A_\ell$, since
\begin{eqnarray*} \left|G_{\ell\ell}\right|=\left|\frac{1}{\alpha_n+Y_\ell}\right| \le \eta^{-1},\end{eqnarray*} we obtain \begin{equation}\label{Al2}
\left|\alpha_n^2 A_\ell \right|=\left|-\mathbb{E} Y_\ell+\mathbb{E}\frac{Y_\ell^2}{\alpha_n+Y_\ell}\right|\leq |\mathbb{E} Y_\ell|+\frac{1}{\eta}\mathbb{E}|Y_\ell|^2. \end{equation}
For $\Delta_\ell$, using the fact that $|\alpha_n| \ge \eta$ and Lemma \ref{Yl2} we obtain
\begin{eqnarray*} \label{4:dl} \left|\Delta_\ell \right| \leq \frac{C_{\tau}}{\eta^3}\left(\frac{1}{p}+\sqrt\frac{p}{n}\right) \end{eqnarray*} for any $z \in \Gamma_{\tau}$.
Summing for all $\ell \in [1\mathrel{{.}\,{.}}\nobreak p]$ and then dividing $p$ on both sides of (\ref{diagonal3}), it is easy to see that in writing \[s_n(z)= \frac{1}{\alpha_n+\Delta(z)},\] the quantity $\Delta(z)$ satisfies the same bound as $\Delta_\ell$ above. This completes the proof of Theorem \ref{thm2}. \end{proof}
\section*{Acknowledgments}
The research of M. Xiong was supported by RGC grant number 16303615 from Hong Kong.
\end{document}
|
arXiv
|
{
"id": "1907.00323.tex",
"language_detection_score": 0.6087767481803894,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{abstract} Let $K$ be a finite extension of $\mathbf{Q}_p$ with residue field $\mathbf{F}_q$ and let $P(T) = T^d + a_{d-1}T^{d-1} + \cdots +a_1 T$ where $d$ is a power of $q$ and $a_i \in \mathfrak{m}_K$ for all $i$. Let $u_0$ be a uniformizer of $\mathcal{O}_K$ and let $\{u_n\}_{n \geqslant 0}$ be a sequence of elements of $\overline{\mathbf{Q}}_p$ such that $P(u_{n+1}) = u_n$ for all $n \geqslant 0$. Let $K_\infty$ be the field generated over $K$ by all the $u_n$. If $K_\infty / K$ is a Galois extension, then it is abelian, and our main result is that it is generated by the torsion points of a relative Lubin-Tate group (a generalization of the usual Lubin-Tate groups). The proof of this involves generalizing the construction of Coleman power series, constructing some $p$-adic periods in Fontaine's rings, and using local class field theory. \end{abstract}
\begin{altabstract} Soit $K$ une extension finie de $\mathbf{Q}_p$ de corps r\'esiduel $\mathbf{F}_q$ et $P(T) = T^d + a_{d-1}T^{d-1} + \cdots +a_1 T$ o\`u $d$ est une puissance de $q$ et $a_i \in \mathfrak{m}_K$ pour tout $i$. Soit $u_0$ une uniformisante de $\mathcal{O}_K$ et $\{u_n\}_{n \geqslant 0}$ une suite d'\'el\'ements de $\overline{\mathbf{Q}}_p$ telle que $P(u_{n+1}) = u_n$ pour tout $n \geqslant 0$. Soit $K_\infty$ l'extension de $K$ engendr\'ee par les $u_n$. Si $K_\infty / K$ est Galoisienne, alors elle est ab\'elienne, et notre r\'esultat principal est qu'elle est engendr\'ee par les points de torsion d'un groupe de Lubin-Tate relatif (une g\'en\'eralisation des groupes de Lubin-Tate usuels). Pour prouver cela, nous g\'en\'eralisons la construction des s\'eries de Coleman, construisons des p\'eriodes $p$-adiques dans les anneaux de Fontaine et utilisons la th\'eorie du corps de classes local. \end{altabstract}
\subjclass{11S05; 11S15; 11S20; 11S31; 11S82; 13F25; 14F30}
\keywords{Iterated extension; Coleman power series; Field of norms; $p$-adic dynamical system; $p$-adic Hodge theory; Local class field theory; Lubin-Tate group; Chebyshev polynomial}
\thanks{This material is partially based upon work supported by the NSF under Grant No.\ 0932078 000 while the author was in residence at the MSRI in Berkeley, California, during the Fall 2014 semester.}
\dedicatory{To Glenn Stevens, on the occasion of his 60th birthday}
\maketitle
\tableofcontents
\setlength{\baselineskip}{18pt}
\section*{Introduction}\label{intro}
Let $K$ be a field, let $P(T) \in K[T]$ be a polynomial of degree $d \geqslant 1$, choose $u_0 \in K$ and for $n \geqslant 0$, let $u_{n+1} \in \overline{K}$ be such that $P(u_{n+1}) = u_n$. The field $K_\infty$ generated over $K$ by all the $u_n$ is called an \emph{iterated extension} of $K$. These iterated extensions and the resulting Galois groups have been studied in various contexts, see for instance \cite{O85}, \cite{S92}, \cite{AHM} and \cite{BJ}.
In this article, we focus on a special situation: $p \neq 2$ is a prime number, $K$ is a finite extension of $\mathbf{Q}_p$, with ring of integers $\mathcal{O}_K$, whose maximal ideal is $\mathfrak{m}_K$ and whose residue field is $k$. Let $d$ be a power of $\mathrm{Card}(k)$, and let $P(T) = T^d + a_{d-1} T^{d-1} + \cdots + a_1 T$ be a monic polynomial of degree $d$ with $a_i \in \mathfrak{m}_K$ for $1 \leqslant i \leqslant d-1$. Let $u_0$ be a uniformizer of $\mathcal{O}_K$ and define a sequence $\{ u_n \}_{n \geqslant 0}$ by letting $u_{n+1}$ be a root of $P(T) = u_n$. Let $K_n = K(u_n)$ and $K_\infty=\cup_{n \geqslant 1} K_n$. This iterated extension is called a \emph{Frobenius-iterate extension}, after \cite{CD14} (whose definition is a bit more general than ours). The question that we consider in this article is: which \emph{Galois} extensions $K_\infty/K$ are Frobenius iterate?
This question is inspired by the observation, made in remark 7.16 of \cite{CD14}, that it follows from the main results of ibid.\ and \cite{LTFN} that: if $K_\infty/K$ is Frobenius-iterate and Galois, then it is necessarily abelian. Here, we prove a much more precise result.
First, let us recall that in \cite{DSLT}, de Shalit gives a generalization of the construction of Lubin-Tate formal groups (for which see \cite{LT}). A \emph{relative Lubin-Tate group} is a formal group $\mathrm{S}$ that is attached to an unramified extension $E/F$ and to an element $\alpha$ of $F$ of valuation $[E:F]$. The extension $E_\infty^{\mathrm{S}}/F$ generated over $F$ by the torsion points of this formal group is the subextension of $F^{\mathrm{ab}}$ cut out via local class field theory by the subgroup of $F^\times$ generated by $\alpha$. If $E=F$, we recover the classical Lubin-Tate groups.
\begin{enonce*}{Theorem A} Let $K$ be a finite Galois extension of $\mathbf{Q}_p$, and let $K_\infty/K$ be a Frobenius-iterate extension. If $K_\infty/K$ is Galois, then there exists a subfield $F$ of $K$, and a relative Lubin-Tate group $\mathrm{S}$, relative to the extension $F^{\mathrm{unr}} \cap K$ of $F$, such that if $K_\infty^{\mathrm{S}}$ denotes the extension of $K$ generated by the torsion points of $\mathrm{S}$, then $K_\infty \subset K_\infty^{\mathrm{S}}$ and $K_\infty^{\mathrm{S}} / K_\infty$ is a finite extension. \end{enonce*}
This is theorem \ref{dslt}. Conversely, it is easy to see that the extension coming from a relative Lubin-Tate group is Frobenius-iterate after the first layer (see example \ref{expolit}). The proof of theorem A is quite indirect. We start with the observation that if $K_\infty / K$ is a Frobenius-iterate extension, that is not necessarily Galois, then we can generalize the construction of Coleman's power series (see \cite{C79}). Let $\varprojlim \mathcal{O}_{K_n}$ denote the set of sequences $\{x_n\}_{n \geqslant 0}$ with $x_n \in \mathcal{O}_{K_n}$ and such that $\mathrm{N}_{K_{n+1}/K_n}(x_{n+1}) = x_n$ for all $n \geqslant 0$.
\begin{enonce*}{Theorem B} We have $\{u_n\}_{n \geqslant 0} \in \varprojlim \mathcal{O}_{K_n}$ and if $\{x_n\}_{n \geqslant 0} \in \varprojlim \mathcal{O}_{K_n}$, then there exists a unique power series $\mathrm{Col}_x(T) \in \mathcal{O}_K \dcroc{T}$ such that $x_n = \mathrm{Col}_x(u_n)$ for all $n \geqslant 0$. \end{enonce*}
Suppose now that $K_\infty/K$ is Galois, and let $\Gamma=\mathrm{Gal}(K_\infty/K)$. The results of \cite{CD14} and \cite{LTFN} imply that $K_\infty/K$ is abelian, so that $K_n/K$ is Galois for all $n \geqslant 1$. If $g \in \Gamma$, then $\{g(u_n)\}_{n \geqslant 0} \in \varprojlim \mathcal{O}_{K_n}$, so that by theorem B, we get a power series $\mathrm{Col}_g(T) \in \mathcal{O}_K \dcroc{T}$ such that $g(u_n) = \mathrm{Col}_g(u_n)$ for all $n \geqslant 0$. Let $\widetilde{\mathbf{E}}^+ = \varprojlim_{x \mapsto x^d} \mathcal{O}_{\mathbf{C}_p} / p$, let $K_0 = \mathbf{Q}_p^{\mathrm{unr}} \cap K$ and let $\tilde{\mathbf{A}}^+ = \mathcal{O}_K \otimes_{\mathcal{O}_{K_0}} W(\widetilde{\mathbf{E}}^+)$ be Fontaine's rings of periods (see \cite{FP}). The element $\{u_n\}_{n \geqslant 0}$ gives rise to an element $\overline{u} \in \widetilde{\mathbf{E}}^+$.
\begin{enonce*}{Theorem C} There exists $u \in \tilde{\mathbf{A}}^+$ whose image in $\widetilde{\mathbf{E}}^+$ is $\overline{u}$, and such that $\varphi_d(u) = P(u)$. We have $g(u)= \mathrm{Col}_g(u)$ if $g \in \Gamma$. \end{enonce*}
The power series $\mathrm{Col}_g(T)$ satisfies the functional equation $\mathrm{Col}_g \circ P(T) = P \circ \mathrm{Col}_g(T)$. The study of $p$-adic power series that commute under composition was taken up by Lubin in \cite{L94}. In \S 6 of ibid., Lubin writes that ``experimental evidence seems to suggest that for an invertible series to commute with a noninvertible series, there must be a formal group somehow in the background''. There are a number of results in this direction, see for instance \cite{LMS}, \cite{SS13} and \cite{JS14}. In our setting, the series $\{ \mathrm{Col}_g(T) \}_{g \in \Gamma}$ commute with $P(T)$ and theorem A says that indeed, there is a formal group that accounts for this. Let us now return to the proof of theorem A. We first show that $P'(T) \neq 0$. It is then proved in \S 1 of \cite{L94} that given such a $P(T)$, a power series $\mathrm{Col}_g(T)$ that commutes with $P(T)$ is determined by $\mathrm{Col}_g'(0)$. If we let $\eta(g) = \mathrm{Col}_g'(0)$, we get the following: the map $\eta : \Gamma \to \mathcal{O}_K^\times$ is an injective character.
In order to finish the proof of theorem A, we use some $p$-adic Hodge theory. Let $\mathrm{L}_P(T) \in K\dcroc{T}$ be the \emph{logarithm} attached to $P(T)$ and constructed in \cite{L94}; it converges on the open unit disk, and satisfies $\mathrm{L}_P \circ P(T) = P'(0) \cdot \mathrm{L}_P(T)$ as well as $\mathrm{L}_P \circ \mathrm{Col}_g(T) = \eta(g) \cdot \mathrm{L}_P(T)$ for $g \in \Gamma$. In particular, we can consider $\mathrm{L}_P(u)$ as an element of the ring $\mathbf{B}_{\mathrm{cris}}^+$ (see \cite{FP} for the rings of periods $\mathbf{B}_{\mathrm{cris}}^+$ and $\mathbf{B}_{\mathrm{dR}}$), which satisfies $g(\mathrm{L}_P(u)) = \eta(g) \cdot \mathrm{L}_P(u)$. More generally, if $\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)$, then we can twist $u$ by $\tau$ to get some elements $u_\tau \in \tilde{\mathbf{A}}^+$ and $\mathrm{L}_P^\tau(u_\tau) \in \mathbf{B}_{\mathrm{cris}}^+$, satisfying $g(\mathrm{L}_P^\tau(u_\tau)) = \tau(\eta(g)) \cdot \mathrm{L}_P^\tau(u_\tau)$. The elements $\{\mathrm{L}_P^\tau(u_\tau)\}_\tau$ are crystalline periods for the representation arising from $\eta$. Our main technical result concerning these periods is that the set of $\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)$ such that $\mathrm{L}_P^\tau(u_\tau) \in \mathrm{Fil}^1 \mathbf{B}_{\mathrm{dR}}$ is a subgroup of $\mathrm{Gal}(K/\mathbf{Q}_p)$, and therefore cuts out a subfield $F$ of $K$. This allows us to prove the following.
\begin{enonce*}{Theorem D} There exists a subfield $F$ of $K$, a Lubin-Tate character $\chi_K$ attached to a uniformizer of $K$, and an integer $r \geqslant 1$, such that $\eta = \mathrm{N}_{K/F}(\chi_K)^r$. \end{enonce*}
Theorem A follows from theorem D by local class field theory: the extensions of $K$ corresponding to $\mathrm{N}_{K/F}(\chi_K)$ are precisely those that come from relative Lubin-Tate groups. At the end of \S \ref{lcft}, we give an example for which $r=2$. In this example, the Coleman power series $p$-adically interpolate Chebyshev polynomials.
\section{Relative Lubin-Tate groups} \label{dsltsec}
We recall de Shalit's construction (see \cite{DSLT}) of a family of formal groups that generalize Lubin-Tate groups. Let $F$ be a finite extension of $\mathbf{Q}_p$, with ring of integers $\mathcal{O}_F$ and residue field $k_F$ of cardinality $q$. Take $h \geqslant 1$ and let $E$ be the unramified extension of $F$ of degree $h$. Let $\varphi_q : E \to E$ denote the Frobenius map that lifts $[x \mapsto x^q]$. If $f(T) = \sum_{i \geqslant 0} f_i T^i \in E\dcroc{T}$, let $f^{\varphi_q}(T) = \sum_{i \geqslant 0} \varphi_q(f_i) T^i$.
If $\alpha \in \mathcal{O}_F$ is such that $\mathrm{val}_F(\alpha) = h$, let $\mathcal{F}_\alpha$ be the set of power series $f(T) \in \mathcal{O}_E \dcroc{T}$ such that $f(T) = \pi T + \mathrm{O}(T^2)$ with $\mathrm{N}_{E/F}(\pi) = \alpha$ and such that $f(T) \equiv T^q \bmod{\mathfrak{m}_E \dcroc{T}}$. The set $\mathcal{F}_\alpha$ is nonempty, since $\mathrm{N}_{E/F}(E^\times)$ is the set of elements of $F^\times$ whose valuation is in $h \cdot \mathbf{Z}$. If $\mathrm{N}_{E/F}(\pi) = \alpha$, one can take $f(T)=\pi T+T^q$. The following theorem summarizes some of the results of \cite{DSLT} (see also \S IV of \cite{IK}).
\begin{theo}\label{dsltmain} If $f(T) \in \mathcal{F}_\alpha$, then \begin{enumerate} \item there is a unique formal group law $\mathrm{S}(X,Y) \in \mathcal{O}_E \dcroc{X,Y}$ such that $\mathrm{S}^{\varphi_q} \circ f = f \circ \mathrm{S}$, and the isomorphism class of $\mathrm{S}$ depends only on $\alpha$; \item for all $a \in \mathcal{O}_F$, there exists a unique power series $[a](T) \in \mathcal{O}_E \dcroc{T}$ such that $[a](T) = aT + \mathrm{O}(T^2)$ and $[a](T) \in \mathrm{End}(\mathrm{S})$. \end{enumerate} Let $x_0 = 0$ and for $m \geqslant 0$, let $x_m \in \overline{\mathbf{Q}}_p$ be such that $f^{\varphi_q^m}(x_{m+1}) = x_m$ (with $x_1 \neq 0$). Let $E_m = E(x_m)$ and let $E_\infty^{\mathrm{S}} = \cup_{m \geqslant 1} E_m$. \begin{enumerate} \item The fields $E_m$ depend only on $\alpha$, and not on the choice of $f(T) \in \mathcal{F}_\alpha$; \item The extension $E_m/E$ is Galois, and its Galois group is isomorphic to $(\mathcal{O}_F/\mathfrak{m}_F^m)^\times$; \item $E_\infty^{\mathrm{S}} \subset F^{\mathrm{ab}}$ and $E_\infty^{\mathrm{S}}$ is the subfield of $F^{\mathrm{ab}}$ cut out by $\langle \alpha \rangle \subset F^\times$ via local class field theory. \end{enumerate} \end{theo}
\begin{rema}\label{dshone} If $h=1$, then we recover the usual Lubin-Tate formal groups of \cite{LT}. \end{rema}
\section{Frobenius-iterate extensions} \label{polit}
Let $p \neq 2$ be a prime number, let $K$ be a finite extension of $\mathbf{Q}_p$, with ring of integers $\mathcal{O}_K$, whose maximal ideal is $\mathfrak{m}_K$ and whose residue field is $k$. Let $q=\mathrm{Card}(k)$, and let $\pi$ denote a uniformizer of $\mathcal{O}_K$. Let $d$ be a power of $q$, and let $P(T) = T^d + a_{d-1} T^{d-1} + \cdots + a_1 T$ be a monic polynomial of degree $d$ with $a_i \in \mathfrak{m}_K$ for $1 \leqslant i \leqslant d-1$.
Let $u_0$ be a uniformizer of $\mathcal{O}_K$ and define a sequence $\{ u_n \}_{n \geqslant 0}$ by letting $u_{n+1}$ be a root of $P(T) = u_n$. Let $K_n = K(u_n)$.
\begin{lemm}\label{pinunif} The extension $K_n/K$ is totally ramified of degree $d^n$, $u_n$ is a uniformizer of $\mathcal{O}_{K_n}$ and $\mathrm{N}_{K_{n+1}/K_n} ( u_{n+1} ) = u_n$. \end{lemm}
\begin{proof} The first two assertions follow immediately from the theory of Newton polygons, and the last one from the fact that $P(T) - u_n$ is the minimal polynomial of $u_{n+1}$ over $K_n$, as well as the fact that $d$ is odd since $p \neq 2$. \end{proof}
Let $K_\infty = \cup_{n \geqslant 1} K_n$. This is a totally ramified infinite and pro-$p$ extension of $K$.
\begin{defi}\label{defpolit} We say that an extension $K_\infty/K$ is \emph{$\varphi$-iterate} if it is of the form above. \end{defi}
This definition is inspired by the similar one that is given in definition 1.1 of \cite{CD14}. We require $P(T)$ to be a monic polynomial, instead of a more general power series as in ibid., in order to control the norm of $u_n$ and to ensure the good behavior of $K_{n+1}/K_n$.
\begin{exem}\label{expolit} (i) If $P(T)=T^q$, then $K_\infty/K$ is a $\varphi$-iterate extension, which is the Kummer extension of $K$ corresponding to $\pi$.
(ii) Let $\mathrm{LT}$ be a Lubin-Tate formal $\mathcal{O}_K$-module attached to $\pi$, and $K_n=K(\mathrm{LT}[\pi^n])$. The extension $K_\infty/K_1$ is $\varphi$-iterate with $P(T) = [\pi](T)$.
(iii) More generally, let $\mathrm{S}$ be a relative Lubin-Tate group, relative to an extension $E/F$ and $\alpha \in F$ as in \S \ref{dsltsec}. The extension $E_\infty^{\mathrm{S}}/E_1$ is $\varphi$-iterate with $P(T) = [\alpha](T)$. \end{exem}
\begin{proof} Item (ii) follows from applying (iii) with $K=E=F$, and we now prove (iii). We use the notation of theorem \ref{dsltmain}. Since the isomorphism class of $\mathrm{S}$ and the extension $E_\infty^{\mathrm{S}}/E$ only depend on $\alpha$, we can take $f(T) = \pi T + T^q$ where $\mathrm{N}_{E/F}(\pi) = \alpha$. Let $P(T) = f^{\varphi_q^{h-1}} \circ \cdots \circ f^{\varphi_q} \circ f(T) \in \mathcal{O}_E[T]$, so that $P(T) = [\alpha](T)$. The extension $E_{hm+1}$ is generated by $x_{hm+1}$ over $E_1$, and we have $P(x_{hm+1}) = x_{(h-1)m+1}$. The claim therefore follows from taking $u_m = x_{hm+1}$ for $m \geqslant 0$, and observing that since $\pi + u_0^{q-1} = 0$, $u_0$ is a uniformizer of $\mathcal{O}_{E_1}$. \end{proof}
\section{Coleman power series} \label{colpow}
Let us write $\varprojlim \mathcal{O}_{K_n}$ for the set of sequences $\{x_n\}_{n \geqslant 0}$ such that $x_n \in \mathcal{O}_{K_n}$ and such that $\mathrm{N}_{K_{n+1}/K_n}(x_{n+1})=x_n$ for $n \geqslant 0$. By lemma \ref{pinunif}, the sequence $\{ u_n \}_{n \geqslant 0}$ belongs to $\varprojlim \mathcal{O}_{K_n}$. The goal of this {\S} is to show the following theorem (theorem B).
\begin{theo}\label{colexist} If $\{ x_n \}_{n \geqslant 0} \in \varprojlim \mathcal{O}_{K_n}$, then there exists a uniquely determined power series $\mathrm{Col}_x(T) \in \mathcal{O}_K \dcroc{T}$ such that $x_n = \mathrm{Col}_x(u_n)$ for all $n \geqslant 0$. \end{theo}
Our proof follows the one that is given in \S 13 of \cite{W97}. The unicity is a consequence of the following well-known general principle.
\begin{prop} \label{colunik} If $f(T) \in \mathcal{O}_K \dcroc{T}$ is nonzero, then $f(T)$ has only finitely many zeroes in the open unit disk. \end{prop}
In order to prove the existence part of theorem \ref{colexist}, we start by generalizing Coleman's norm map (see \cite{C79} for the original construction, and \S 2.3 of \cite{F90} for the generalization that we use). The ring $\mathcal{O}_K\dcroc{T}$ is a free $\mathcal{O}_K\dcroc{P(T)}$-module of rank $d$. If $f(T) \in \mathcal{O}_K\dcroc{T}$, let $\mathcal{N}_P(f)(T) \in \mathcal{O}_K\dcroc{T}$ be defined by the requirement that $\mathcal{N}_P(f)(P(T)) = \mathrm{N}_{\mathcal{O}_K\dcroc{T} / \mathcal{O}_K\dcroc{P(T)}} (f(T))$. For example, $\mathcal{N}_P(T)=T$ since $d$ is odd.
\begin{prop}\label{pronmp} The map $\mathcal{N}_P$ has the following properties. \begin{enumerate} \item If $f(T) \in \mathcal{O}_K\dcroc{T}$, then $\mathcal{N}_P(f)(u_n) = \mathrm{N}_{K_{n+1}/K_n} (f(u_{n+1}))$; \item If $k \geqslant 1$ and $f(T) \in 1+ \pi^k \mathcal{O}_K\dcroc{T}$, then $\mathcal{N}_P(f)(T) \in 1+ \pi^{k+1} \mathcal{O}_K\dcroc{T}$; \item If $f(T) \in \mathcal{O}_K \dcroc{T}$, then $\mathcal{N}_P(f)(T) \equiv f(T) \bmod{\pi}$; \item If $f(T) \in \mathcal{O}_K \dcroc{T}^\times$, and $k,m \geqslant 0$, then $\mathcal{N}_P^{m+k}(f) \equiv \mathcal{N}_P^k(f) \bmod{\pi^{k+1}}$. \end{enumerate} \end{prop}
\begin{proof} The determinant of the multiplication-by-$f(T)$ map on the $\mathcal{O}_K\dcroc{P(T)}$-module $\mathcal{O}_K\dcroc{T}$ is $\mathcal{N}_P(f)(P(T))$. By evaluating at $T=u_{n+1}$, we find that the determinant of the multiplication-by-$f(u_{n+1})$ map on the $\mathcal{O}_{K_n}$-module $\mathcal{O}_{K_{n+1}}$ is $\mathcal{N}_P(f)(u_n)$, so that $\mathcal{N}_P(f)(u_n) = \mathrm{N}_{K_{n+1}/K_n} (f(u_{n+1}))$.
We now prove (2). If $f(T) \in \mathcal{O}_K\dcroc{T}$, let $\mathcal{T}_P(f)(T) \in \mathcal{O}_K\dcroc{T}$ be the trace map defined by $\mathcal{T}_P(f)(P(T)) = \mathrm{Tr}_{\mathcal{O}_K\dcroc{T} / \mathcal{O}_K\dcroc{P(T)}} (f(T))$. A straightforward calculation shows that if $h(T) \in \mathcal{O}_K \dcroc{T}$, then $\mathcal{T}_P(h)(T) \in \pi \cdot \mathcal{O}_K \dcroc{T}$. If $f(T) = 1 +\pi^k h(T)$, then $\mathcal{N}_P(f)(T) \equiv 1 + \pi^k \mathcal{T}_P(h)(T) \bmod{\pi^{k+1}}$, so that $\mathcal{N}_P(f)(T) \in 1+ \pi^{k+1} \mathcal{O}_K\dcroc{T}.$
Item (3) follows from a straightfoward calculation in $k\dcroc{T}$ using the fact that $P(T)=T^d$ in $k\dcroc{T}$. Finally, let us prove (4). If $f(T) \in \mathcal{O}_K \dcroc{T}^\times$, then $\mathcal{N}_P(f)/f \equiv 1 \bmod{\pi}$ by (3), so that $\mathcal{N}_P^m(f)/f \equiv 1 \bmod{\pi}$ as well. Item (2) now implies that $\mathcal{N}_P^{m+k}(f) \equiv \mathcal{N}_P^k(f) \bmod{\pi^{k+1}}$. \end{proof}
\begin{proof}[of theorem \ref{colexist}] The power series $\mathrm{Col}_x(T)$ is unique by lemma \ref{colunik}, and we now show its existence. If $x_n$ is not a unit of $\mathcal{O}_{K_n}$, then there exists $e \geqslant 1$ such that $x_n=u_n^e x_n^*$ where $x_n^* \in \mathcal{O}_{K_n}^\times$ for all $n$, and then $\mathrm{Col}_x(T) = T^e \cdot \mathrm{Col}_{x^*}(T)$. We can therefore assume that $x_n$ is a unit of $\mathcal{O}_{K_n}$. For all $j \geqslant 1$, we have $\mathcal{O}_{K_j} = \mathcal{O}_K[u_j]$, so that there exists $g_j(T) \in \mathcal{O}_K[T]$ such that $x_j = g_j(u_j)$. Let $f_j(T) = \mathcal{N}_P^j(g_{2j})$. By proposition \ref{pronmp}, we have $x_n \equiv f_j(u_n) \bmod{\pi^{j+1}}$ for all $n \leqslant j$. The space $\mathcal{O}_K \dcroc{T}$ is compact; let $f(T)$ be a limit point of $\{ f_j \}_{j \geqslant 1}$. We have $x_n = f(u_n)$ for all $n$ by continuity, so that we can take $\mathrm{Col}_x(T) = f(T)$. \end{proof}
\begin{rema}\label{colnpinv} We have $\mathcal{N}_P(\mathrm{Col}_x)(T) = \mathrm{Col}_x(T)$. \end{rema}
\begin{proof} The power series $\mathcal{N}_P(\mathrm{Col}_x)(T) - \mathrm{Col}_x(T)$ is zero at $T=u_n$ for all $n \geqslant 0$ by proposition \ref{pronmp}, so that $\mathcal{N}_P(\mathrm{Col}_x)(T) = \mathrm{Col}_x(T)$ by lemma \ref{colunik}. \end{proof}
\section{Lifting the field of norms} \label{cdn}
In this {\S}, we assume that $K_\infty/K$ is a Galois extension, and let $\Gamma = \mathrm{Gal}(K_\infty / K)$. We recall some results of \cite{CD14} and \cite{LTFN}, and give a more precise formulation of some of them in our specific situation.
\begin{prop}\label{kngal} If $K_\infty/K$ is Galois, then $K_n/K$ is Galois for all $n \geqslant 1$. \end{prop}
\begin{proof} It follows from the main results of \cite{CD14} and of \cite{LTFN} (see remark 7.16 of \cite{CD14}) that if $K_\infty/K$ is a $\varphi$-iterate extension that is Galois, then it is abelian. This implies the proposition (it would be more satisfying to find a direct proof). \end{proof}
If $g \in \Gamma$, proposition \ref{kngal} and theorem \ref{colexist} imply that there is a unique power series $\mathrm{Col}_g(T) \in \mathcal{O}_K \dcroc{T}$ such that $g(u_n) = \mathrm{Col}_g(u_n)$ for all $n \geqslant 0$. In the sequel, we need some ramification-theoretic properties of $K_\infty/K$. They are summarized in the theorem below.
\begin{theo}\label{normpow} There exists a constant $c=c(K_\infty/K) > 0$ such that for any $E \subset F$, finite extensions of $K$ contained in $K_\infty$, and $x \in \mathcal{O}_F$, we have \[ \mathrm{val}_K \left( \frac{\mathrm{N}_{F/E}(x)}{x^{[F:E]}}-1 \right) \geqslant c. \] \end{theo}
\begin{proof} By the main result of \cite{CDAPF}, the extension $K_\infty/K$ is \emph{strictly APF}, so that if we denote by $c(K_\infty/K)$ the constant defined in 1.2.1 of \cite{W83}, then $c(K_\infty/K) > 0$. By 4.2.2.1 of ibid., we have \[ \mathrm{val}_E \left( \frac{\mathrm{N}_{F/E}(x)}{x^{[F:E]}}-1 \right) \geqslant c(F/E), \] By 1.2.3 of ibid., $c(F/E) \geqslant c(K_\infty/E)$ and (see for instance the proof of 4.5 of \cite{CD14} or page 83 of \cite{W83}) $c(K_\infty/E) \geqslant c(K_\infty/K) \cdot [E:K]$. This proves the theorem. \end{proof}
Let $c$ be the constant afforded by theorem \ref{normpow}. We can always assume that $c \leqslant \mathrm{val}_K(p)/(p-1)$. If $E$ is some subfield of $\mathbf{C}_p$, let $\mathfrak{a}_E^c$ denote the set of elements $x$ of $E$ such that $\mathrm{val}_K(x) \geqslant c$. Let $\widetilde{\mathbf{E}}^+= \varprojlim_{x \mapsto x^d} \mathcal{O}_{\mathbf{C}_p} / \mathfrak{a}^c_{\mathbf{C}_p}$. The sequence $\{ u_n \}_{n \geqslant 0}$ gives rise to an element $\overline{u} \in \widetilde{\mathbf{E}}^+$. Recall that by \S 2.1 and \S 4.2 of \cite{W83}, there is an embedding $\iota : \varprojlim \mathcal{O}_{K_n} \to \widetilde{\mathbf{E}}^+$, which is an isomorphism onto $\varprojlim_{x \mapsto x^d} \mathcal{O}_{K_n} / \mathfrak{a}^c_{K_n}$, which is also isomorphic to $k \dcroc{\overline{u}}$. Let $K_0= \mathbf{Q}_p^{\mathrm{unr}} \cap K$ and $\tilde{\mathbf{A}}^+=\mathcal{O}_K \otimes_{\mathcal{O}_{K_0}} W(\widetilde{\mathbf{E}}^+)$. Recall (see \cite{FP}) that we have a map $\theta : \tilde{\mathbf{A}}^+ \to \mathcal{O}_{\mathbf{C}_p}$. If $x \in \tilde{\mathbf{A}}^+$ and $\overline{x} = (x_n)_{n \geqslant 0}$ in $\widetilde{\mathbf{E}}^+$, then $\theta \circ \varphi_d^{-n}(x) = x_n$ in $\mathcal{O}_{\mathbf{C}_p} / \mathfrak{a}^c_{\mathbf{C}_p}$.
\begin{theo}\label{liftubar} There exists a unique $u \in \tilde{\mathbf{A}}^+$ whose image in $\widetilde{\mathbf{E}}^+$ is $\overline{u}$, and such that $\varphi_d(u) = P(u)$. Moreover: \begin{enumerate} \item[(i)] If $n \geqslant 0$, then $\theta \circ \varphi_d^{-n}(u) = u_n$; \item[(ii)] $\mathcal{O}_K \dcroc{u} = \{ x \in \tilde{\mathbf{A}}^+$, $\theta \circ \varphi_d^{-n}(x) \in \mathcal{O}_{K_n}$ for all $n \geqslant 1\}$; \item[(iii)] $g(u) = \mathrm{Col}_g(u)$ if $g \in \Gamma$. \end{enumerate} \end{theo}
\begin{proof} The existence of $u$ and item (i) are proved in lemma 9.3 of \cite{CEV}, where it is shown that $u = \lim_{n \to +\infty} P^{\circ n}(\varphi_d^{-n}([\overline{u}]))$.
Let $R = \{ x \in \tilde{\mathbf{A}}^+$ such that $\theta \circ \varphi_d^{-n}(x) \in \mathcal{O}_{K_n}$ for all $n \geqslant 1\}$. If $x \in R$, then its image in $\widetilde{\mathbf{E}}^+$ lies in $\varprojlim_{x \mapsto x^d} \mathcal{O}_{K_n} / \mathfrak{a}^c_{K_n} = k \dcroc{\overline{u}}$. We have $u \in R$ by proposition \ref{liftubar}, so that the map $R / \pi R \to k \dcroc{\overline{u}}$ is surjective. We then have $R = \mathcal{O}_K \dcroc{u}$, since $R$ is separated and complete for the $\pi$-adic topology, which proves (ii).
The ring $\mathcal{O}_K \dcroc{u}$ is stable under the action of $G_K$ by (ii). If $g \in \Gamma$, there exists $F_g(T) \in \mathcal{O}_K \dcroc{T}$ such that $g(u) = F_g(u)$. We have $g(u_n) = g ( \theta \circ \varphi_d^{-n}(u)) = \theta \circ \varphi_d^{-n} (F_g(u)) = F_g(u_n)$ by (i), so that $g(u_n) = F_g(u_n)$ for all $n$. This implies that $F_g(T) = \mathrm{Col}_g(T)$. \end{proof}
\begin{rema}\label{liftfn} In the terminology of \cite{W83}, $\varprojlim \mathcal{O}_{K_n}$ is the ring of integers of the field of norms $X(K_\infty)$ of the extension $K_\infty/K$, and theorem \ref{liftubar} shows that we can lift $X(K_\infty)$ to characteristic zero, along with the Frobenius map $\varphi_d$ and the action of $\Gamma$. \end{rema}
If $g \in \Gamma$, then $\mathrm{Col}_g \circ P(T) = P \circ \mathrm{Col}_g(T)$ since the two series have the same value at $u_n$ for all $n \geqslant 1$. Let $\eta(g) = \mathrm{Col}_g'(0)$, so that $g \mapsto \eta(g)$ is a character $\eta : \Gamma \to \mathcal{O}_K^\times$
\begin{prop}\label{comnoz} If $F(T) \in T \cdot \mathcal{O}_K \dcroc{T}$ is such that $F'(0) \in 1+p \mathcal{O}_K$, and if $A(T) \in T \cdot \mathcal{O}_K \dcroc{T}$ vanishes at order $k \geqslant 2$ at $0$, and satisifes $A \circ F(T) = F \circ A(T)$, then $F(T) = T$. \end{prop}
\begin{proof} Write $F(T) = f_1 T + \mathrm{O}(T^2)$, and $A(T) = a_k T^k + \mathrm{O}(T^{k+1})$ with $a_k \neq 0$. The equation $F \circ A(T) = A \circ F(T)$ implies that $f_1 a_k = a_k f_1^k$ so that if $k \neq 1$, then $f_1^{k-1}=1$. Since $f_1 \in 1+p\mathcal{O}_K$ and $p \neq 2$, this implies that $f_1=1$. If $F(T) \neq T$, we can write $F(T) = T + T^i h(T)$ for some $i \geqslant 2$ with $h(0) \neq 0$. The equation $F \circ A(T) = A \circ F(T)$ and the equality $A(T + T^i h(T))=\sum_{j \geqslant 0} (T^i h(T))^j A^{(j)}(T)/j!$ imply that $A(T) + A(T)^i h(A(T)) = A(T) + T^i h(T) A'(T) + \mathrm{O} (T^{2i+k-2})$, so that $A(T)^i h(A(T)) = T^i h(T) A'(T) + \mathrm{O} (T^{2i+k-2})$. The term of lowest degree of the LHS is of degree $ki$, while on the RHS it is of degree $i+k-1$. We therefore have $ki=i+k-1$, so that $(k-1)(i-1)=0$ and hence $k=1$. \end{proof}
\begin{coro}\label{dernonul} We have $P'(0) \neq 0$. \end{coro}
\begin{proof} This follows from proposition \ref{comnoz}, since $\mathrm{Col}_g'(0) \in 1+p\mathcal{O}_K$ if $g$ is close enough to $1$, and $\mathrm{Col}_g(T) = T$ if and only if $g=1$ (compare with lemma 4.5 of \cite{LTFN}). \end{proof}
\begin{coro}\label{galisab} The character $\eta : \Gamma \to \mathcal{O}_K^\times$ is injective. \end{coro}
\begin{proof} This follows from proposition 1.1 of \cite{L94}, which says that if $P'(0) \in \mathfrak{m}_K \setminus \{ 0 \}$, then a power series $F(T) \in T \cdot \mathcal{O}_K\dcroc{T}$ that commutes with $P(T)$ is determined by $F'(0)$. This implies that $\mathrm{Col}_g(T)$ is determined by $\eta(g)$, and then $g$ itself is determined by $\mathrm{Col}_g(T)$, since $g(u_n) = \mathrm{Col}_g(u_n)$ for all $n$. \end{proof}
We therefore have a character $\eta : \mathrm{Gal}(\overline{\mathbf{Q}}_p/K) \to \mathcal{O}_K^\times$, such that $K_\infty = \overline{\mathbf{Q}}_p^{\ker \eta}$.
\section{$p$-adic Hodge theory} \label{chareta}
We now assume that $K/\mathbf{Q}_p$ is Galois (for simplicity), and we keep assuming that $K_\infty/K$ is Galois. We use the element $u$ above, and Lubin's logarithm (proposition \ref{lublog} below), to construct crystalline periods for $\eta$.
\begin{prop}\label{lublog} There exists a power series $\mathrm{L}_P(T) \in K\dcroc{T}$ that is holomorphic on the open unit disk, and satisfies \begin{enumerate} \item $\mathrm{L}_P(T) = T + \mathrm{O}(T^2)$; \item $\mathrm{L}_P \circ P(T) = P'(0) \cdot \mathrm{L}_P(T)$; \item $\mathrm{L}_P \circ \mathrm{Col}_g(T) = \eta(g) \cdot \mathrm{L}_P(T)$ if $g \in \Gamma$. \end{enumerate} If we write $P(T)=T \cdot Q(T)$, then \[ \mathrm{L}_P(T) = \lim_{n \to +\infty} \frac{P^{\circ n}(T)} {P'(0)^n} = T \cdot \prod_{n \geqslant 0} \frac{Q(P^{\circ n}(T))}{Q(0)}. \] \end{prop}
\begin{proof} See propositions 1.2, 2.2 and 1.3 of \cite{L94}. \end{proof}
Let $\widetilde{\mathbf{B}}^+_{\mathrm{rig}}$ denote the Fr\'echet completion of $\tilde{\mathbf{A}}^+ [1/\pi]$, so that our $\widetilde{\mathbf{B}}^+_{\mathrm{rig}}$ is $K \otimes_{K_0}$ the ``usual'' $\widetilde{\mathbf{B}}^+_{\mathrm{rig}}$ (for which see \cite{LB2}). If $u \in \tilde{\mathbf{A}}^+$ is the element afforded by proposition \ref{liftubar}, then $\mathrm{L}_P(u)$ converges in $\widetilde{\mathbf{B}}^+_{\mathrm{rig}}$. We have $g(\mathrm{L}_P(u)) = \eta(g) \cdot \mathrm{L}_P(u)$ by proposition \ref{lublog}. If $\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)$, then let $n(\tau)$ be some $n \in \mathbf{Z}$ such that $\tau = \varphi^n$ on $k_K$, and let $u_\tau = (\tau \otimes \varphi^{n(\tau)}) (u) \in \tilde{\mathbf{A}}^+$.
If $F(T) = \sum_{i \geqslant 0} f_i T^i \in K \dcroc{T}$, let $F^\tau(T) = \sum_{i \geqslant 0} \tau(f_i) T^i$. We have $g(\mathrm{L}_P^\tau(u_\tau)) = \tau(\eta(g)) \cdot \mathrm{L}_P^\tau(u_\tau)$ in $\widetilde{\mathbf{B}}^+_{\mathrm{rig}}$. This implies the following result, which is a slight improvement of theorem 4.1 of \cite{LTFN}.
\begin{prop}\label{cisdrp} The character $\eta : \Gamma \to \mathcal{O}_K^\times$ is crystalline, with weights in $\mathbf{Z}_{\geqslant 0}$. \end{prop}
\begin{proof} The fact that $g(\mathrm{L}_P^\tau(u_\tau)) = \tau(\eta(g)) \cdot \mathrm{L}_P^\tau(u_\tau)$ for all $\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)$ implies that $\eta$ gives rise to a $K \otimes_{K_0} \mathbf{B}_{\mathrm{cris}}$-admissible representation. If $V$ is any $p$-adic representation of $G_K$, then \[ \left((K \otimes_{K_0} \mathbf{B}_{\mathrm{cris}}) \otimes_{\mathbf{Q}_p} V \right)^{G_K} = K \otimes_{K_0} (\mathbf{B}_{\mathrm{cris}} \otimes_{\mathbf{Q}_p} V)^{G_K}. \] This implies that a $K \otimes_{K_0} \mathbf{B}_{\mathrm{cris}}$-admissible representation is crystalline. The weights of $\eta$ are $\geqslant 0$ because $\mathrm{L}_P^\tau(u_\tau) \in \mathbf{B}_{\mathrm{dR}}^+$ for all $\tau$. \end{proof}
\begin{lemm}\label{thetutau} We have $\theta \circ \varphi_d^{-n}(u_\tau) = \lim_{k \to +\infty} (P^\tau)^{\circ k}(u_{n+k}^{p^{n(\tau)}})$. \end{lemm}
\begin{proof} The element $u_\tau \in \tilde{\mathbf{A}}^+$ has the property that its image in $\widetilde{\mathbf{E}}^+$ is $\varphi^{n(\tau)}(\overline{u}) = \overline{u}^{p^{n(\tau)}}$, and that $\varphi_d(u_\tau) = P^\tau(u_\tau)$. The lemma then follows from lemma 9.3 of \cite{CEV}. \end{proof}
For simplicity, write $u_\tau^n = \theta \circ \varphi_d^{-n}(u_\tau)$ and $u_\tau^{n,k} = (P^\tau)^{\circ k}(u_{n+k}^{p^{n(\tau)}})$.
\begin{lemm}\label{unifconv} If $M > 0$, there exists $j \geqslant 0$ such that $\mathrm{val}_K(u_\tau^n - u_\tau^{n,j}) \geqslant M$ for $n \geqslant 1$. \end{lemm}
\begin{proof} If $c$ is the constant coming from theorem \ref{normpow}, then $\mathrm{val}_K(u_\tau^n - u_\tau^{n,0}) \geqslant c$ for all $n \geqslant 1$. We prove the lemma by inductively constructing a sequence $\{c_j\}_{j \geqslant 0}$ such that $\mathrm{val}_K(u_\tau^n - u_\tau^{n,j}) \geqslant c_j$ for all $n \geqslant 1$, and such that $c_j \geqslant M$ for $j \gg 0$. Let $c_0=c$ and suppose that for some $j$, we have $\mathrm{val}_K(u_\tau^n - u_\tau^{n,j}) \geqslant c_j$ for all $n \geqslant 1$. We then have \[ \mathrm{val}_K(u_\tau^n - u_\tau^{n,j+1}) = \mathrm{val}_K \left( P^\tau(u_\tau^{n+1}) - P^\tau(u_\tau^{n+1,j})\right). \]
If $R(T) \in \mathcal{O}_K[T]$ and $x,y \in \mathcal{O}_{\overline{\mathbf{Q}}_p}$, then $R(x)-R(y) = (x-y) R'(y) + (x-y)^2 S(x,y)$ with $S(T,U) \in \mathcal{O}_K [T,U]$. This, and the fact that $P'(T) \in \mathfrak{m}_K[T]$, implies that we can take $c_{j+1} = \min(c_j+1, 2c_j)$. The lemma follows. \end{proof}
We now recall a result from \cite{L94}. If $f(T) \in T \cdot \mathcal{O}_K \dcroc{T}$ is such that $f'(0) \in \mathfrak{m}_K \setminus \{ 0 \}$, let $\Lambda(f)$ be the set of the roots of all iterates of $f$. If $u(T) \in T \cdot \mathcal{O}_K \dcroc{T}$ is such that $u'(0) \in \mathcal{O}_K^\times$ and $u'(0)$ is not a root of $1$, let $\Lambda(u)$ be the set of the fixed points of all iterates of $u$.
\begin{lemm}\label{lublamb} If $f$ and $u$ are as above, and if $u \circ f = f \circ u$, then $\Lambda(f)=\Lambda(u)$. \end{lemm}
\begin{proof} This is proposition 3.2 of \cite{L94}. \end{proof}
For each $\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)$, let $r_\tau$ be the weight of $\eta$ at $\tau$.
\begin{prop}\label{algfil} If $\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)$, then the following are equivalent. \begin{enumerate} \item[(i)] $r_\tau \geqslant 1$; \item[(ii)] $\mathrm{L}_P^\tau(u_\tau) \in \mathrm{Fil}^1 \mathbf{B}_{\mathrm{dR}}$; \item[(iii)] $\theta(u_\tau) \in \overline{\mathbf{Q}}_p$; \item[(iv)] $\theta(u_\tau) \in \Lambda(P^\tau)$; \item[(v)] $u_\tau \in \cup_{j \geqslant 0} \ \varphi_d^{-j}(\mathcal{O}_K \dcroc{u})$. \end{enumerate} \end{prop}
\begin{proof} The equivalence between (i) and (ii) is immediate. We now prove that (ii) implies (iii). If $\mathrm{L}_P^\tau(u_\tau) \in \mathrm{Fil}^1 \mathbf{B}_{\mathrm{dR}}$, then $\mathrm{L}_P^\tau(\theta(u_\tau)) = 0$ so that $\theta(u_\tau) \in \overline{\mathbf{Q}}_p$ since it is a root of a convergent power series with coefficients in $K$. We next prove that (iii) implies (iv) (it is clear that (iv) implies (iii)). If $x= \theta(u_\tau)$ then $g(x) = \mathrm{Col}_g^\tau(x)$. If $x \in \overline{\mathbf{Q}}_p$ and if $g$ is close enough to $1$, then $g(x)=x$ so that $x \in \Lambda(\mathrm{Col}_g^\tau)$, and then $x \in \Lambda(P^\tau)$ by lemma \ref{lublamb}. Let us prove that (iv) implies (ii). If there exists $n \geqslant 0$ such that $(P^{\tau})^{\circ n}(\theta(u_\tau)) = 0$, then $(P^{\tau})^{\circ n}(u_\tau) \in \mathrm{Fil}^1 \mathbf{B}_{\mathrm{dR}}$ so that $\mathrm{L}_P^\tau(u_\tau) \in \mathrm{Fil}^1 \mathbf{B}_{\mathrm{dR}}$ as well by proposition \ref{lublog}. Conditions (i), (ii), (iii) and (iv) are therefore equivalent. Condition (v) implies (iii) by using theorem \ref{liftubar} as well as the fact that $\varphi_d(u_\tau) = P^\tau(u_\tau)$.
It remains to prove that (iii) implies (v). Recall that $u_\tau^n = \theta \circ \varphi_d^{-n}(u_\tau)$. It is enough to show that there exists $j \geqslant 0$ such that $u_\tau^n \in \mathcal{O}_{K_{n+j}}$ for all $n$, since by theorem \ref{liftubar}, this implies that $u_\tau \in \varphi_d^{-j}(\mathcal{O}_K \dcroc{u})$. Recall that $u_\tau^{n,k} = (P^\tau)^{\circ k}(u_{n+k}^{p^{n(\tau)}})$. Take $M \geqslant 1 + \mathrm{val}_K((P^\tau)'(u_\tau^n))$ for all $n \gg 0$. By lemma \ref{unifconv}, there exists $j \geqslant 0$ such that $\mathrm{val}_K(u_\tau^n - u_\tau^{n,j}) \geqslant M$ for all $n \geqslant 1$. The element $u_\tau^n$ is a root of $P^\tau(T) = u_\tau^{n-1}$, and therefore $u_\tau^n- u_\tau^{n,j}$ is a root of $P^\tau(u_\tau^{n,j}+T) - u_\tau^{n-1}$. If $u_\tau^{n-1} \in \mathcal{O}_{K_{n+j-1}}$, then the polynomial $R_n(T) = P^\tau(u_\tau^{n,j}+T) - u_\tau^{n-1}$ belongs to $\mathcal{O}_{K_{n+j}}[T]$, and satisfies $\mathrm{val}_K(R_n(0)) \geqslant M + \mathrm{val}_K(R_n'(0))$. By the theory of Newton polygons, $R_n(T)$ has a unique root of slope $\mathrm{val}_K(R_n(0)) - \mathrm{val}_K(R_n'(0)) \geqslant M$, and this root, which is $u_\tau^n - u_\tau^{n,j}$, therefore belongs to $K_{n+j}$. This implies that $u_\tau^n \in \mathcal{O}_{K_{n+j}}$, which finishes the proof by induction on $n$. \end{proof}
If $\tau$ satisfies the equivalent conditions of proposition \ref{algfil}, then we can write $u_\tau = f_\tau ( \varphi_q^{-j_\tau}(u))$ for some $j_\tau \geqslant 0$ and $f_\tau(T) \in \mathcal{O}_K \dcroc{T}$.
\begin{lemm}\label{ftnul} We have $f_\tau(0)=0$, $f_\tau'(0) \neq 0$, $P^\tau \circ f_\tau(T) = f_\tau \circ P(T)$ and $\mathrm{Col}_g^\tau \circ f_\tau(T) = f_\tau \circ \mathrm{Col}_g(T)$. \end{lemm}
\begin{proof} If $u_\tau = f_\tau(\varphi_d^{-j}(u))$, then $P^\tau(u_\tau) = P^\tau \circ f_\tau (\varphi_d^{-j}(u))$ and then $\varphi_d(u_\tau) = f_\tau \circ P (\varphi_d^{-j}(u))$ so that $P^\tau \circ f_\tau(T) = f_\tau \circ P(T)$. Likewise, computing $g(u_\tau)$ in two ways shows that $\mathrm{Col}_g^\tau \circ f_\tau(T) = f_\tau \circ \mathrm{Col}_g(T)$. Evaluating $P^\tau \circ f_\tau(T) = f_\tau \circ P(T)$ at $T=0$ gives $P^\tau(f_\tau(0)) = f_\tau(0)$ so that $f_\tau(0)$ is a root of $P^\tau(T) = T$. The theory of Newton polygons shows that those roots are $0$ and elements of valuation $0$. The latter case is excluded because $\theta \circ \varphi_d^{-n}(u_\tau) = f_\tau ( u_{n+j} ) \in \mathfrak{m}_{K_\infty}$, so that $f_\tau(0) \in \mathfrak{m}_K$. We now prove that $f_\tau'(0) \neq 0$. Write $f(T) = f_k T^k + \mathrm{O}(T^{k+1})$ with $f_k \neq 0$. The fact that $P^\tau \circ f_\tau(T) = f_\tau \circ P(T)$ implies that $\tau(P'(0)) f_k = f_k P'(0)^k$ so that $\tau(P'(0)) = P'(0)^k$. Since $\mathrm{val}_K(P'(0)) > 0$, this implies that $k=1$. \end{proof}
\begin{coro}\label{embssgr} The set of those $\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)$ such that $r_\tau \geqslant 1$ forms a subgroup of $\mathrm{Gal}(K/\mathbf{Q}_p)$, and if $F$ is the subfield of $K$ cut out by this subgroup, then $\eta(g) \in \mathcal{O}_F^\times$. The weight $r_\tau$ is independent of $\tau \in \mathrm{Gal}(K/F)$. \end{coro}
\begin{proof} By proposition \ref{liftubar}, $\tau=\mathrm{Id}$ satisfies condition (iii) of proposition \ref{algfil} above, and therefore condition (i) as well, so that $r_{\mathrm{Id}} \geqslant 1$. If $\sigma, \tau$ satisfy condition (v) of ibid, then we can write $u_\sigma = f_\sigma(\varphi_d^{-j_\sigma}(u))$ and $u_\tau = f_\tau(\varphi_d^{-j_\tau}(u))$ so that $u_{\sigma \tau} = f_\tau^\sigma \circ f_\sigma (\varphi_d^{-(j_\tau+j_\sigma)}(u))$ and therefore $\sigma \tau$ also satisfies condition (v). Since $\mathrm{Gal}(K/\mathbf{Q}_p)$ is a finite group, these two facts imply that the set of $\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)$ such that $r_\tau \geqslant 1$ is a group.
By lemma \ref{ftnul}, we have $P^\tau \circ f_\tau(T) = f_\tau \circ P(T)$. This implies that $P'(0) \in \mathfrak{m}_F$ and also that $(P^\tau)^{\circ n} \circ f_\tau(T) = f_\tau \circ P^{\circ n}(T)$, so that \[ \frac{1}{P'(0)^n}(P^\tau)^{\circ n} \circ f_\tau(T) = \frac{1}{P'(0)^n} f_\tau \circ P^{\circ n}(T), \] which implies by passing to the limit that $\mathrm{L}_P^\tau \circ f_\tau(T) = f_\tau'(0) \cdot \mathrm{L}_P(T)$. Since $\mathrm{Col}_g^\tau \circ f_\tau(T) = f_\tau \circ \mathrm{Col}_g(T)$, we have $g(\mathrm{L}_P^\tau \circ f_\tau(u) ) = \tau(\eta(g)) \cdot (\mathrm{L}_P^\tau \circ f_\tau(u))$. Moreover, $\mathrm{L}_P^\tau \circ f_\tau(u) = f_\tau'(0) \cdot \mathrm{L}_P(u)$, and therefore $\tau(\eta(g)) = \eta(g)$. This is true for every $\tau \in \mathrm{Gal}(K/F)$, so that $\eta(g) \in \mathcal{O}_F^\times$. The fact that $\eta(g) \in \mathcal{O}_F^\times$ implies that $r_\tau$ depends only on $\tau {\mid}_F$ and is therefore independent of $\tau \in \mathrm{Gal}(K/F)$. \end{proof}
\section{Local class field theory} \label{lcft}
We now prove theorem D, and show how local class field theory allows us to derive theorem A from theorem D. We still assume that $K/\mathbf{Q}_p$ is Galois for simplicity. Let $\lambda$ be a uniformizer of $\mathcal{O}_K$ and let $K_\lambda$ denote the extension of $K$ attached to $\lambda$ by local class field theory. This extension is generated over $K$ by the torsion points of a Lubin-Tate formal group defined over $K$ and attached to $\lambda$ (see for instance \cite{LT} and \cite{LCFT}). Let $\chi^K_\lambda : \mathrm{Gal}(K_\lambda / K) \to \mathcal{O}_K^\times$ denote the corresponding Lubin-Tate character.
We still assume that the extension $K_\infty/K$ is Galois, so that it is an abelian totally ramified extension. This implies that there is a uniformizer $\lambda$ of $\mathcal{O}_K$ such that $K_\infty \subset K_\lambda$. Let $\eta : \Gamma \to \mathcal{O}_K^\times$ be the character constructed in \S\ref{chareta}.
\begin{prop}\label{etaexpl} We have $\eta = \prod_{\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)} \tau(\chi^K_\lambda)^{r_\tau}$. \end{prop}
\begin{proof} The character $\eta : \Gamma \to \mathcal{O}_K^\times$ is crystalline, and its weight at $\tau$ is $r_\tau$ by definition. The character $\eta_0 = \eta \cdot (\prod_{\tau \in \mathrm{Gal}(K/\mathbf{Q}_p)} \tau(\chi^K_\lambda)^{r_\tau} )^{-1}$ of $\mathrm{Gal}(K_\lambda/K)$ is therefore crystalline with weights $0$ at all embeddings, so that it is an unramified character of $\mathrm{Gal}(K_\lambda/K)$. Since $K_\lambda/K$ is totally ramified, we have $\eta_0=1$. \end{proof}
Proposition \ref{etaexpl} and corollary \ref{embssgr} imply the following, which is theorem D.
\begin{theo}\label{maineta} There exists $F \subset K$ and $r \in \mathbf{Z}_{\geqslant 1}$ such that $\eta = \mathrm{N}_{K/F}(\chi^K_\lambda)^r$. \end{theo}
We now show how this implies theorem A. If $u \in \mathcal{O}_K^\times$, let $\mu^K_u$ denote the unramified character of $G_K$ that sends the Frobenius map of $k_K$ to $u$. If $F$ is a subfield of $K$, and $\mathrm{N}_{K/F}(\lambda) = \varpi^h u$ with $\varpi$ a uniformizer of $\mathcal{O}_F$ and some $u \in \mathcal{O}_F^\times$, then $\mathrm{N}_{K/F}(\chi^K_\lambda) = \chi^F_\varpi \cdot \mu^K_u$.
\begin{prop}\label{chards} Let $\mathrm{S}$ be a relative Lubin-Tate group, attached to an extension $E/F$, and an element $\alpha = \varpi^h u \in \mathcal{O}_F^\times$. The action of $\mathrm{Gal}(\overline{\mathbf{Q}}_p/E)$ on the torsion points of $\mathrm{S}$ is given by $g(x) = [\chi^F_\varpi \cdot \mu^E_u(g)](x)$. \end{prop}
\begin{proof} See \S 4 of \cite{Y08}. \end{proof}
Let $F$ be the subfield of $K$ afforded by theorem \ref{maineta}, and let $E$ be the maximal unramified extension of $F$ contained in $K$.
\begin{theo}\label{dslt} There exists a relative Lubin-Tate group $\mathrm{S}$, relative to the extension $E/F$, such that if $K_\infty^{\mathrm{S}}$ denotes the extension of $K$ generated by the torsion points of $\mathrm{S}$, then $K_\infty \subset K_\infty^{\mathrm{S}}$ and $K_\infty^{\mathrm{S}} / K_\infty$ is a finite extension. \end{theo}
\begin{proof} Let $\lambda$ be a uniformizer of $K$ such that $K_\infty \subset K_\lambda$ and let $\pi = \mathrm{N}_{K/E}(\lambda)$ and $\alpha = \mathrm{N}_{K/F}(\lambda)$, so that $\pi$ is a uniformizer of $E$ and $\alpha = \mathrm{N}_{E/F}(\pi)$. Let $\mathrm{S}$ be a relative Lubin-Tate group attached to $\alpha$, and let $K_\infty^{\mathrm{S}}$ be the extension of $K$ generated by the torsion points of $\mathrm{S}$. If $g \in \mathrm{Gal}(\overline{\mathbf{Q}}_p/K_\infty^{\mathrm{S}})$, then $\mathrm{N}_{K/F}(\chi_K)(g)=1$ by proposition \ref{chards} and the observation preceding it, so that $\eta(g)=1$ by theorem \ref{maineta}. This implies that $K_\infty \subset K_\infty^{\mathrm{S}}$. By Galois theory and theorem \ref{maineta}, \begin{enumerate} \item $K_\infty^{\mathrm{S}}$ is the field cut out by $\{ g \in G_K \mid \mathrm{N}_{K/F}(\chi_\lambda^K(g)) = 1\}$; \item $K_\infty$ is the field cut out by $\{ g \in G_K \mid \mathrm{N}_{K/F}(\chi_\lambda^K(g))^r = 1\}$. \end{enumerate} This implies that $K_\infty^{\mathrm{S}} / K_\infty$ is a finite Galois extension, whose Galois group injects into $\{ x \in \mathcal{O}_F^\times \mid x^r=1\}$. \end{proof}
This proves theorem A. We conclude this {\S} with an example of a $\varphi$-iterate extension that is Galois, corresponding to a polynomial $P(T) \in \mathbf{Q}_p[T]$ such that $r=2$ and such that the extension $K_\infty^{\mathrm{S}} / K_\infty$ is of degree $2$ in the notation of theorems \ref{maineta} and \ref{dslt}.
\begin{theo}\label{chebtheo} Let $K=\mathbf{Q}_3$, $P(T)=T^3+6T^2+9T$ and $u_0=-3$. The corresponding iterated extension $K_\infty$ is $\mathbf{Q}_3(\mu_{3^\infty})^{\{\pm 1\} \subset \mathbf{Z}_3^\times}$, and $\eta=\chi_{\mathrm{cyc}}^2$. \end{theo}
\begin{proof} For $k \geqslant 1$, let $C_k(T)$ denote the $k$-th Chebyshev polynomial, which is characterized by the fact that $C_k(\cos(\theta)) = \cos(k \theta)$. Let $P_k(T) = 2 ( C_k(T/2+1) - 1)$, so that $P_k(T)$ is a monic polynomial of degree $k$, and $P_k(2(\cos(\theta)-1)) = 2(\cos(k \theta)-1)$. Note that $P(T)=P_3(T)$ and that $u_0 = -3 = 2(\cos(2\pi/3)-1)$. The element $u_n$ is therefore a conjugate of $2(\cos(2\pi/3^{n+1})-1)$. This proves the fact that $K_\infty = \mathbf{Q}_3(\mu_{3^\infty})^{\{\pm 1\} \subset \mathbf{Z}_3^\times}$
If $g \in G_{\mathbf{Q}_3}$, then $g( 2(\cos(2\pi/3^n)-1) = 2(\cos(2\pi\chi_{\mathrm{cyc}}(g) /3^n)-1)$. This implies that $\mathrm{Col}_g(T) = P_k(T)$ if $\chi_{\mathrm{cyc}}(g) = k \in \mathbf{Z}_{\geqslant 1}$. The formula for $\eta$ now follows from this, and the well-known fact that $C_k'(1)=k^2$ if $k \geqslant 1$. \end{proof}
We leave to the reader the generalization of this construction to other $p$ and other Lubin-Tate groups. The results of \S 2 of \cite{LMS} should be useful for this.
\providecommand{\bysame}{\leavevmode ---\ } \providecommand{\og}{``} \providecommand{\fg}{''} \providecommand{\smfandname}{\&} \providecommand{\smfedsname}{\'eds.} \providecommand{\smfedname}{\'ed.} \providecommand{\smfmastersthesisname}{M\'emoire} \providecommand{\smfphdthesisname}{Th\`ese}
\end{document}
|
arXiv
|
{
"id": "1411.7064.tex",
"language_detection_score": 0.5823366641998291,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Reidemeister transformations of the potential function and the solution
} \author{\sc Jinseok Cho and Jun Murakami} \maketitle \begin{abstract} The potential function of the optimistic limit of the colored Jones polynomial and the construction of the solution of the hyperbolicity equations were defined in the authors' previous articles. In this article, we define the Reidemeister transformations of the potential function and the solution by the changes of them under the Reidemeister moves of the link diagram and show the explicit formulas. These two formulas enable us to see the changes of the complex volume formula under the Reidemeister moves. As an application, we can simply specify the discrete faithful representation of the link group by showing a link diagram and one geometric solution. \end{abstract}
\section{Introduction}\label{sec1}
\subsection{Overview}
One of the fundamental theorem of knot theory is the Reidemeister theorem, which states two links are equivalent if and only if their diagrams are related by finite steps of the Reidemeister moves. Therefore, one of the most natural method to obtain a knot invariant is to define a value from a knot diagram and show that the value is invariant under the Reidemeister moves. However, some invariants cannot be defined in this way, especially the ones defined from the hyperbolic structure of the link. This is because, the hyperbolicity equations\footnote{{\it Hyperbolicity equations} are the gluing equations of the edges of a given triangulation together with the equation of the completeness condition. Each solution of the equations determines a boundary-parabolic representation and some of them determines the hyperbolic structure of the link.} and their solutions do not change locally under the Reidemeister moves. Especially, when a solution of certain hyperbolicity equations is given, we cannot see how the equations and the solution change under the Reidemeister moves. This is one of the major obstructions to develop a combinatorial approach to the hyperbolic invariants of links.
On the other hand, {the optimistic limit method was first introduced at \cite{Murakami00b}. Although this method was not defined rigorously, the resulting value was {\it optimistically} expected to be the actual limit of certain quantum invariants. The rigorous definition of the optimistic limit of the Kashaev invariant was proposed at \cite{Yokota10} and the resulting value was proved to be the complex volume of the knot. Although this definition is rigorous and general enough, they requires some unnatural assumptions on the diagram and several technical difficulties. Therefore it was modified to more combinatorial version at \cite{Cho13a}. The optimistic limit used in this article is the one defined at \cite{Cho13b} and the main results are based on \cite{Cho14c}.
In our definition, the triangulation is naturally defined from the link diagram and its hyperbolicity equations, whose solutions determine the boundary-parabolic representations\footnote{ A representation $\rho:\pi_1(L)\rightarrow {\rm PSL}(2,\mathbb{C})$ of the link group $\pi_1(L):=\pi_1(\mathbb{S}^3\backslash L)$ is {\it boundary-parabolic} when any meridian loop of the boundary-tori of the link complement $\mathbb{S}^3\backslash L$ maps to a parabolic element in ${\rm PSL}(2,\mathbb{C})$ under $\rho$.} of the link group, are the partial derivatives $\exp(w_k\frac{\partial W}{\partial w_k})=1$ $(k=1,\ldots,n)$ of certain potential function $W(w_1,\ldots,w_n)$. Note that this potential function is combinatorially defined from the link diagram, so it changes naturally under the Reidemeister moves. Then the optimistic limit is defined by the evaluation of the potential function (with slight modification) at certain solution of the hyperbolicity equations. (Explicit definition is the equation (\ref{optimistic}).)
Let $\mathcal{P}$ be the conjugation quandle consisting of the parabolic elements of ${\rm PSL}(2,\mathbb{C})$ proposed at \cite{Kabaya14}. The shadow-coloring of $\mathcal{P}$ is a way of assigning elements of $\mathcal{P}$ to arcs and regions of the link diagram. The elements on arcs are naturally determined from $\rho$ and the ones on the regions are from certain rules. According to \cite{Kabaya14} and \cite{Cho14c}, we can construct the developing map of a given boundary-parabolic representation $\rho:\pi_1(L)\rightarrow {\rm PSL}(2,\mathbb{C})$ directly from the shadow-coloring. (The explicit construction is in Figure \ref{fig06}. Note that this construction is based on \cite{Neumann99} and \cite{Zickert09}.)}
This construction of the solution has two major advantages. At first, if a boundary-parabolic representation $\rho$ is given, then we can always construct the solution corresponding to $\rho$ for any link diagram. (This was the main theorem of \cite{Cho14c}.) In other words, for the hyperbolicity equations of our triangulation, we can always guarantee the existence of a geometric solution,\footnote{ {\it Geometric solution} is a solution of the hyperbolicity equations which determines the discrete faithful representation. (Unlike the standard definition, we allow some tetrahedron can have negative volume. If we consider the triangulation of $\mathbb{S}^3\backslash (L\cup\{\text{two points}\}$), the negative volume tetrahedra are unavoidable.) Note that geometric solution in our context is not unique.} which is an assumption in many other texts. Furthermore, the constructed solution changes locally under the Reidemeister moves on the link diagram $D$. Note that the variables $w_1,\ldots,w_n$ of the hyperbolicity equations are assinged to regions of the diagram. (See Section \ref{sec12} below.) Assume the solution $(w_1^{(0)},\ldots,w_n^{(0)})$ is constructed from the diagram $D$ together with the representation $\rho$.
{ \begin{definition} A solution $(w_1^{(0)},\ldots,w_n^{(0)})$ is called {\it essential} when $w_k^{(0)}\neq 0$ for all $k=1,\ldots,n$ and $w_k^{(0)}\neq w_m^{(0)}$ for the pairs $w_k^{(0)}$ and $w_m^{(0)}$ assigned to adjacent regions of the diagram $D$. \end{definition}
According to Lemma \ref{lem}, essentialness of a solution is generic property, so we can always construct uncountably many essential solutions from any $D$ and $\rho$. From now on, solutions in this article are always assumed to be essential and the Reidemeister transformations are defined between two essential solutions. (This assumption is guaranteed by Corollary \ref{cor32}.) Note that essentialness of the solution guarantees that the shape parameters defined in Section \ref{sec33} are not in $\{0,1,\infty\}$.
Let $D'$ be the link diagram obtained by applying one Reidemeister move to $D$.} In this article, we will show that if a new variable $w_{n+1}$ is appeared in $D'$, then the values $w_1^{(0)},\ldots,w_n^{(0)}$ of the newly constructed solution from $D'$ and $\rho$ are preserved and the value $w_{n+1}^{(0)}$ is uniquely determined by the other values $w_1^{(0)},\ldots,w_n^{(0)}$. (The explicit relations are in Section \ref{sec4}.) Also, if a region with $w_k$ is removed in $D'$, then we can easily get the solution by removing the value $w_k^{(0)}$ of the variable $w_k$. These changes of the solution will be called {\it the Reidemeister transformations of the solution} in Section \ref{sec12}.
Using the Reidemeister transformations of the potential function together with the solution, we can see how the complex volume formula changes under the Reidemeister moves.
(See Theorem \ref{thm1}.) As an application, we can easily specify the discrete faithful representation by showing one link diagram $D$ and one geometric solution corresponding to the diagram. In particular, if we have another diagram $D'$ of $L$, then we can easily find the geometric solution corresponding to $D'$, {without solving the hyperbolicity equations again,} by applying the Reidemeister transformations of the solution.
Many results of the optimistic limit and other concepts used in this article are scattered in the authors' previous articles. Referring all of them might be quite confusing for readers, so we added many known results here, especially in Sections \ref{sec2}-\ref{sec3}, and sometimes we reprove the known results to clarify the discussion.
\subsection{Reidemeister transformations}\label{sec12}
To describe the exact definition of the Reidemeister transformation, we have to define the potential function first. Consider a link diagram\footnote{ We always assume the diagram $D$ does not contain a trivial knot component which has only over-crossings or under-crossings or no crossing. If it happens, we change the diagram of the trivial component slightly by adding a kink. } $D$ of a link $L$ and assign complex variables $w_1,\ldots,w_n$ to regions of $D$. Then we define the potential function of a crossing $j$ as in Figure \ref{pic01}.
\begin{figure}
\caption{Potential function of the crossing $j$}
\label{pic01}
\end{figure}
{In the definition above, ${\rm Li}_2(z):=-\int_0^z \frac{\log(1-t)}{t}dt$ is the dilogarithm function. Although it is a multi-valued function depending on the choice of the branches of $\log t$ and $\log(1-t)$, the final formula in (\ref{optimistic}) does not depend on choice of the branches. (See Lemma 2.1's of \cite{Cho13a} and \cite{Cho13c}).}
{\it The potential function} of $D$ is defined by \begin{equation}\label{defW} W(w_1,\ldots,w_n):=\sum_{j\text{ : crossings of }D}W^j, \end{equation} and we modify it to \begin{equation}\label{defW_0} W_0(w_1,\ldots,w_n):=W(w_1,\ldots,w_n)-\sum_{k=1}^n \left(w_k\frac{\partial W}{\partial w_k}\right)\log w_k. \end{equation}
Also, we define the set of equations \begin{equation}\label{defH}
\mathcal{I}:=\left\{\left.\exp\left(w_k\frac{\partial W}{\partial w_k}\right)=1\right|k=1,\ldots,n\right\}. \end{equation} Then $\mathcal{I}$ becomes the set of the hyperbolicity equations of the five-term triangulation defined in Section \ref{sec2}. (See Proposition \ref{pro21}.)
Consider a boundary-parabolic representation $\rho:\pi_1(L)\rightarrow{\rm PSL}(2,\mathbb{C})$. Then, using the shadow-coloring of $\mathcal{P}$ induced by $\rho$, we can construct the solution $\bold w^{(0)}=(w_1^{(0)},\ldots,w_n^{(0)})$ of $\mathcal{I}$ satisfying $\rho_{\bold w^{(0)}}=\rho$ up to conjugation, where $\rho_{\bold w^{(0)}}$ is the representation induced by the five-term triangulation together with the solution $\bold w^{(0)}$. (The detail is in Section \ref{sec3}. See Proposition \ref{pro2}. {We will also show that any solution of $\mathcal{I}$ can be constructed by this method in Appendix \ref{app}.}) Furthermore, the solution satisfies \begin{equation}\label{optimistic} W_0(\bold w^{(0)})\equiv i({\rm vol}(\rho)+i\,{\rm cs}(\rho))~~({\rm mod}~\pi^2), \end{equation} where ${\rm vol}(\rho)$ and ${\rm cs}(\rho)$ are the hyperbolic volume and the Chern-Simons invariant of $\rho$, respectively, {which were defined in \cite{Neumann04} and \cite{Zickert09}.} We call ${\rm vol}(\rho)+i\,{\rm cs}(\rho)$ {\it the (hyperbolic) complex volume of $\rho$} and define {\it the optimistic limit of the colored Jones polynomial} by $W_0(\bold w^{(0)})$.
The oriented Reidemeister moves defined in \cite{Polyak10} are in Figure \ref{pic02}. We define the potential functions $V_{R1}$, $V_{R1'}$, $V_{R_2}$, $V_{R2'}$, $V_{R3}$ and $V_{R3'}$ of Figure \ref{pic02} as follows: \begin{align*} V_{R1}(w_a,w_b,w_c)=&{\rm Li}_2(\frac{w_b}{w_a})+{\rm Li}_2(\frac{w_b}{w_c})-{\rm Li}_2(\frac{w_b^2}{w_a w_c})\\
&-{\rm Li}_2(\frac{w_a}{w_b})-{\rm Li}_2(\frac{w_c}{w_b})+\frac{\pi^2}{6}-\log\frac{w_a}{w_b}\log\frac{w_c}{w_b},
\end{align*} \begin{align*} V_{R1'}(w_a,w_b,w_c)=&-{\rm Li}_2(\frac{w_b}{w_a})-{\rm Li}_2(\frac{w_b}{w_c})+{\rm Li}_2(\frac{w_b^2}{w_a w_c})\\
&+{\rm Li}_2(\frac{w_a}{w_b})+{\rm Li}_2(\frac{w_c}{w_b})-\frac{\pi^2}{6}+\log\frac{w_a}{w_b}\log\frac{w_c}{w_b}, \end{align*} \begin{align*} V_{R2}&(w_a,w_b,w_c,w_d,w_e)=\\
&{\rm Li}_2(\frac{w_a}{w_d})+{\rm Li}_2(\frac{w_a}{w_e})-{\rm Li}_2(\frac{w_a w_c}{w_d w_e})-{\rm Li}_2(\frac{w_d}{w_c})-{\rm Li}_2(\frac{w_e}{w_c})-\log\frac{w_d}{w_c}\log\frac{w_e}{w_c}\\
&-{\rm Li}_2(\frac{w_c}{w_b})-{\rm Li}_2(\frac{w_c}{w_e})+{\rm Li}_2(\frac{w_a w_c}{w_b w_e})+{\rm Li}_2(\frac{w_b}{w_a})+{\rm Li}_2(\frac{w_e}{w_a})+\log\frac{w_b}{w_a}\log\frac{w_e}{w_a},
\end{align*} \begin{align*} V_{R2'}&(w_a,w_b,w_c,w_d,w_e)=\\
&{\rm Li}_2(\frac{w_c}{w_d})+{\rm Li}_2(\frac{w_c}{w_e})-{\rm Li}_2(\frac{w_a w_c}{w_d w_e})-{\rm Li}_2(\frac{w_d}{w_a})-{\rm Li}_2(\frac{w_e}{w_a})-\log\frac{w_d}{w_a}\log\frac{w_e}{w_a}\\
&-{\rm Li}_2(\frac{w_a}{w_b})-{\rm Li}_2(\frac{w_a}{w_e})+{\rm Li}_2(\frac{w_a w_c}{w_b w_e})+{\rm Li}_2(\frac{w_b}{w_c})+{\rm Li}_2(\frac{w_e}{w_c})+\log\frac{w_b}{w_c}\log\frac{w_e}{w_c}, \end{align*} \begin{align*} V_{R3}&(w_a,w_b,w_c,w_d,w_e,w_f,w_h)=-\frac{\pi^2}{2}\\
&-{\rm Li}_2(\frac{w_e}{w_d})-{\rm Li}_2(\frac{w_e}{w_f})+{\rm Li}_2(\frac{w_e w_h}{w_d w_f})+{\rm Li}_2(\frac{w_d}{w_h})+{\rm Li}_2(\frac{w_f}{w_h})+\log\frac{w_d}{w_h}\log\frac{w_f}{w_h}\\
&-{\rm Li}_2(\frac{w_f}{w_a})-{\rm Li}_2(\frac{w_f}{w_h})+{\rm Li}_2(\frac{w_b w_f}{w_a w_h})+{\rm Li}_2(\frac{w_a}{w_b})+{\rm Li}_2(\frac{w_h}{w_b})+\log\frac{w_a}{w_b}\log\frac{w_h}{w_b}\\
&-{\rm Li}_2(\frac{w_h}{w_b})-{\rm Li}_2(\frac{w_h}{w_d})+{\rm Li}_2(\frac{w_c w_h}{w_b w_d})+{\rm Li}_2(\frac{w_b}{w_c})+{\rm Li}_2(\frac{w_d}{w_c})+\log\frac{w_b}{w_c}\log\frac{w_d}{w_c},
\end{align*} \begin{align*} V_{(R3)^{-1}}&(w_a,w_b,w_c,w_d,w_e,w_f,w_g)=-\frac{\pi^2}{2}\\
&-{\rm Li}_2(\frac{w_f}{w_a})-{\rm Li}_2(\frac{w_f}{w_e})+{\rm Li}_2(\frac{w_g w_f}{w_a w_e})+{\rm Li}_2(\frac{w_a}{w_g})+{\rm Li}_2(\frac{w_e}{w_g})+\log\frac{w_a}{w_g}\log\frac{w_e}{w_g}\\
&-{\rm Li}_2(\frac{w_e}{w_d})-{\rm Li}_2(\frac{w_e}{w_g})+{\rm Li}_2(\frac{w_c w_e}{w_d w_g})+{\rm Li}_2(\frac{w_d}{w_c})+{\rm Li}_2(\frac{w_g}{w_c})+\log\frac{w_d}{w_c}\log\frac{w_g}{w_c}\\
&-{\rm Li}_2(\frac{w_g}{w_a})-{\rm Li}_2(\frac{w_g}{w_c})+{\rm Li}_2(\frac{w_b w_g}{w_a w_c})+{\rm Li}_2(\frac{w_a}{w_b})+{\rm Li}_2(\frac{w_c}{w_b})+\log\frac{w_a}{w_b}\log\frac{w_c}{w_b}. \end{align*} Note that $V_{R1}$ is the potential function of the diagram obtained after applying $R1$ move in Figure \ref{pic02}(a). All others are defined in the same ways, for example, $V_{R3}$ and $V_{R3'}$ are the potential functions of the right-hand and the left-hand sides of Figure \ref{pic02}(c), respectively.
\begin{figure}
\caption{Oriented Reidemeister moves}
\label{pic02}
\end{figure}
\begin{definition}\label{def1} {\it The Reidemeister transformations}\label{def1} $T_{R1}, T_{(R1)^{-1}},...,T_{R3}, T_{(R3)^{-1}}$ of the potential function $W(\ldots,w_a,w_b,\ldots)$ are defined as follows: \begin{align} T_{R1}(W)(\ldots,w_a,w_b,w_c,\ldots)&=W+V_{R1},\nonumber\\ T_{(R1)^{-1}}(W)(\ldots,w_a,w_b,\ldots)&=W-V_{R1},\nonumber\\ T_{R1'}(W)(\ldots,w_a,w_b,w_c,\ldots)&=W+V_{R1'},\nonumber\\ T_{(R1')^{-1}}(W)(\ldots,w_a,w_b,\ldots)&=W-V_{R1'},\nonumber\\ T_{R2}(W)(\ldots,w_a,w_b,w_c,w_d,w_e,\ldots)&=W+V_{R2},\label{R2a}\\ T_{(R2)^{-1}}(W)(\ldots,w_a,w_b,w_c,\ldots)&=W-V_{R2},\label{R2b}\\ T_{R2'}(W)(\ldots,w_a,w_b,w_c,w_d,w_e,\ldots)&=W+V_{R2'},\label{R2c}\\ T_{(R2')^{-1}}(W)(\ldots,w_a,w_b,w_c,\ldots)&=W-V_{R2'},\label{R2d}\\ T_{R3}(W)(\ldots,w_a,w_b,w_c,w_d,w_e,w_f,w_h,\ldots)&=W+V_{R3}-V_{(R3)^{-1}},\nonumber\\ T_{(R3)^{-1}}(W)(\ldots,w_a,w_b,w_c,w_d,w_e,w_f,w_g,\ldots)&=W-V_{R3}+V_{(R3)^{-1}}.\nonumber \end{align} Note that, when applying $R2$ (or $R2'$) move in (\ref{R2a}) (or (\ref{R2c})), we replace $w_b$ of $W$ with $w_d$ for the potential functions of the crossings adjacent to the region associated with $w_d$. Also, when applying $(R2)^{-1}$ (or $(R2')^{-1}$) move in (\ref{R2b}) (or (\ref{R2d})), we replace $w_d$ of $W$ with $w_b$. \end{definition}
Remark that the Reidemeister transformations of the potential function is nothing but the changes of the potential function defined in (\ref{defW}) under the corresponding Reidemeister moves.
\begin{definition} {\it The Reidemeister transformations}\label{def1} $T_{R1}, T_{(R1)^{-1}},...,T_{R3}, T_{(R3)^{-1}}$ of the {essential} solution $(\ldots,w_a^{(0)},w_b^{(0)},\ldots)$ of $\mathcal{I}$ in (\ref{defH}) is defined as follows: for the first Reidemeister moves \begin{align*} T_{R1}(\ldots,w_a^{(0)},w_b^{(0)},\ldots)=T_{R1'}(\ldots,w_a^{(0)},w_b^{(0)},\ldots) =(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},\ldots), \end{align*} where $w_c^{(0)}=2w_b^{(0)}-w_a^{(0)}$, and \begin{align*} T_{(R1)^{-1}}(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},\ldots)=T_{(R1')^{-1}}(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},\ldots) =(\ldots,w_a^{(0)},w_b^{(0)},\ldots). \end{align*}
For the second Reidemeister moves, we put $T_{R2}(W)$ (or $T_{R2'}(W)$) be the potential function in (\ref{R2a}) (or (\ref{R2c})). Then{ \begin{align*} T_{R2}(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},\ldots)=(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},\ldots)\\ \left(\text{or }T_{R2'}(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},\ldots)=(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},\ldots)\right), \end{align*} where $w_d^{(0)}=w_b^{(0)}$ and $w_e^{(0)}$ is uniquely determined by the equation \begin{equation}\label{TR2} \exp(w_b\frac{\partial T_{R2}(W)}{\partial w_b})=1 \left(\text{or } \exp(w_b\frac{\partial T_{R2'}(W)}{\partial w_b})=1\right), \end{equation} and \begin{align*} T_{(R2)^{-1}}(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},\ldots)=(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},\ldots)\\ \left(\text{or } T_{(R2')^{-1}}(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},\ldots)=(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},\ldots)\right). \end{align*} Note that the equation (\ref{TR2}) can be expressed explicitly by using the parameters around the region of $w_b$. (Explicit expression of (\ref{TR2}) is in Lemma \ref{lem53}.)}
For the third Reidemeister moves, \begin{align*} &T_{R3}(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},w_f^{(0)},w_g^{(0)},\ldots)\\ &~~~=(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},w_f^{(0)},w_h^{(0)},\ldots),\\ &T_{(R3)^{-1}}(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},w_f^{(0)},w_h^{(0)},\ldots)\\ &~~~=(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},w_f^{(0)},w_g^{(0)},\ldots), \end{align*} where $w_h^{(0)}$ or $w_g^{(0)}$ is uniquely determined by the equation \begin{equation}\label{R3rel} w_d^{(0)} w_g^{(0)}-w_c^{(0)} w_e^{(0)}=w_a^{(0)} w_h^{(0)} -w_b^{(0)} w_f^{(0)}. \end{equation}
\end{definition}
\begin{theorem}\label{thm1} For a link diagram $D$ of $L$, let $\bold w=(\ldots,w_a,w_b,\ldots)$ and put $W(\bold w)$ be the potential function of $D$. Consider a boundary-parabolic representation $\rho:\pi_1(L)\rightarrow{\rm PSL}(2,\mathbb{C})$ and let $\bold w^{(0)}$ be the solution of $\mathcal{I}$ constructed by the shadow-coloring of $\mathcal{P}$ satisfying $\rho_{\bold w^{(0)}}=\rho$, up to conjugation. (See Proposition \ref{pro2} for the actual construction of $\bold w^{(0)}$.) Then, for any Reidemeister transformation $T$, $T(\bold w^{(0)})$ is also a solution\footnote{ The resulting solution of a Reidemeister transformation on an essential solution can be nonessential. However, essential solutions are generic, so we can deform both solutions to essential ones by changing the region-colorings slightly. (See Corollary \ref{cor32}.) Therefore, we assume the solutions are always essential.} of $\mathcal{I}$ and the induced representation $\rho_{T(\bold w^{(0)})}$ satisfies $\rho_{T(\bold w^{(0)})}=\rho$, up to conjugation. Furthermore, \begin{equation}\label{volume} T(W)_0(T(\bold w^{(0)}))\equiv W_0(\bold w^{(0)}))\equiv i({\rm vol}(\rho)+i\,{\rm cs}(\rho))~~({\rm mod}~\pi^2), \end{equation} where $T(W)_0$ is the modification of the potential function $T(W)$ by (\ref{defW_0}). \end{theorem}
\begin{proof} This follows from Lemma \ref{LemR1}, Lemma \ref{LemR2}, Lemma \ref{LemR3} and Proposition \ref{pro2}. \end{proof}
Note that when some orientations of the strings in Figure \ref{pic02} are reversed, the Reidemeister transformations of the solutions can be defined by the exactly same formula. (It will be proved in Section \ref{sec5}.) If we change the potential function according to the changes of the orientation, then Theorem \ref{thm1} still works. Therefore, it defines the {\it un-oriented} Reidemeister transformations\footnote{ The Reidemeister triansformations of the potential function still depend on the orientation. As a matter of fact, it is possible to define the potential function of the un-oriented diagram using Section 3.2 of \cite{Cho13b}. However, the formula will be redundantly complicate than the one defined in this article, so we do not introduce it.} of the solutions. We will discuss and prove the un-oriented ones in Section \ref{sec5}. Also, the mirror images of the Reidemeister moves will be discussed in Section \ref{sec5}.
As an example of the Reidemeister transformations, we will show the changes of the geometric solution of a diagram $D$ of the figure-eight knot to its mirror image $\overline{D}$ in Section \ref{sec6}.
\section{Five-term triangulation of $\mathbb{S}^3\backslash (L\cup\{\text{two points}\})$}\label{sec2}
In this section, we describe the five-term triangulation of $\mathbb{S}^3\backslash (L\cup\{\text{two points}\})$. Many parts of explanation come from \cite{Cho13c}.
We place an octahedron ${\rm A}_j{\rm B}_j{\rm C}_j{\rm D}_j{\rm E}_j{\rm F}_j$ on each crossing $j$ of the link diagram as in Figure \ref{twistocta} so that the vertices ${\rm A}_j$ and ${\rm C}_j$ lie on the over-bridge and the vertices ${\rm B}_j$ and ${\rm D}_j$ on the under-bridge of the diagram, respectively. Then we twist the octahedron by gluing edges ${\rm B}_j{\rm F}_j$ to ${\rm D}_j{\rm F}_j$ and ${\rm A}_j{\rm E}_j$ to ${\rm C}_j{\rm E}_j$, respectively. The edges ${\rm A}_j{\rm B}_j$, ${\rm B}_j{\rm C}_j$, ${\rm C}_j{\rm D}_j$ and ${\rm D}_j{\rm A}_j$ are called {\it horizontal edges} and we sometimes express these edges in the diagram as arcs around the crossing in the left-hand side of Figure \ref{twistocta}.
\begin{figure}
\caption{Octahedron on the crossing $j$}
\label{twistocta}
\end{figure}
Then we glue faces of the octahedra following the edges of the link diagram. Specifically, there are three gluing patterns as in Figure \ref{glue pattern}. In each cases (a), (b) and (c), we identify the faces $\triangle{\rm A}_{j}{\rm B}_{j}{\rm E}_{j}\cup\triangle{\rm C}_{j}{\rm B}_{j}{\rm E}_{j}$ to $\triangle{\rm C}_{j+1}{\rm D}_{j+1}{\rm F}_{j+1}\cup\triangle{\rm C}_{j+1}{\rm B}_{j+1}{\rm F}_{j+1}$, $\triangle{\rm B}_{j}{\rm C}_{j}{\rm F}_{j}\cup\triangle{\rm D}_{j}{\rm C}_{j}{\rm F}_{j}$ to $\triangle{\rm D}_{j+1}{\rm C}_{j+1}{\rm F}_{j+1}\cup\triangle{\rm B}_{j+1}{\rm C}_{j+1}{\rm F}_{j+1}$ and $\triangle{\rm A}_{j}{\rm B}_{j}{\rm E}_{j}\cup\triangle{\rm C}_{j}{\rm B}_{j}{\rm E}_{j}$ to $\triangle{\rm C}_{j+1}{\rm B}_{j+1}{\rm E}_{j+1}\cup\triangle{\rm A}_{j+1}{\rm B}_{j+1}{\rm E}_{j+1}$, respectively.
\begin{figure}
\caption{Three gluing patterns}
\label{glue pattern}
\end{figure}
Note that this gluing process identifies vertices $\{{\rm A}_j, {\rm C}_j\}$ to one point, denoted by $-\infty$, and $\{{\rm B}_j, {\rm D}_j\}$ to another point, denoted by $\infty$, and finally $\{{\rm E}_j, {\rm F}_j\}$ to the other points, denoted by ${\rm P}_k$ where $k=1,\ldots,s$ and $s$ is the number of the components of the link $L$. The regular neighborhoods of $-\infty$ and $\infty$ are 3-balls and that of $\cup_{k=1}^s P_k$ is {cone over the tori of the link $L$.} Therefore, if we remove the vertices ${\rm P}_1,\ldots,{\rm P}_s$ from the octahedra, then we obtain a decomposition of $\mathbb{S}^3\backslash L$, denoted by $T$. On the other hand, if we remove all the vertices of the octahedra, the result becomes an ideal decomposition of $\mathbb{S}^3\backslash (L\cup\{\pm\infty\})$. We call the latter {\it the octahedral decomposition} and denote it by $T'$.
To obtain an ideal triangulation from $T'$, we divide each octahedron ${\rm A}_j{\rm B}_j{\rm C}_j{\rm D}_j{\rm E}_j{\rm F}_j$ in Figure \ref{twistocta} into five ideal tetrahedra ${\rm A}_j{\rm B}_j{\rm D}_j{\rm F}_j$, ${\rm B}_j{\rm C}_j{\rm D}_j{\rm F}_j$, ${\rm A}_j{\rm B}_j{\rm C}_j{\rm D}_j$, ${\rm A}_j{\rm B}_j{\rm C}_j{\rm E}_j$ and ${\rm A}_j{\rm C}_j{\rm D}_j{\rm E}_j$. We call the result {\it the five-term triangulation} of $\mathbb{S}^3\backslash (L\cup\{\pm\infty\})$.
Note that if we assign the shape parameter $u\in\mathbb{C}\backslash\{0,1\}$ to an edge of an ideal hyperbolic tetrahedron, then the other edges are also parametrized by $u, u':=\frac{1}{1-u}$ and $u'':=1-\frac{1}{u}$ as in Figure \ref{fig6}.
\begin{figure}
\caption{Parametrization of an ideal tetrahedron with a shape parameter $u$}
\label{fig6}
\end{figure} To determine the shape of the octahedron in Figure \ref{twistocta}, we assign shape parameters to edges of tetrahedra as in Figure \ref{fig7}. Note that $\frac{w_a w_c}{w_b w_d}$ in Figure \ref{fig7}(a) and $\frac{w_b w_d}{w_a w_c}$ in Figure \ref{fig7}(b) are the shape parameters of the tetrahedron ${\rm A}_j{\rm B}_j{\rm C}_j{\rm D}_j$ assigned to the edges ${\rm B}_j{\rm D}_j$ and ${\rm A}_j{\rm C}_j$. Also note that the assignment of shape parameters here does not depend on the orientations of the link diagram.
\begin{figure}
\caption{Assignment of shape parameters}
\label{fig7}
\end{figure}
To obtain the boundary parabolic representation $\rho:\pi_1(\mathbb{S}^3\backslash (L\cup\{\pm\infty\}))\longrightarrow{\rm PSL}(2,\mathbb{C})$, we require two conditions on the ideal triangulation of $\mathbb{S}^3\backslash (L\cup\{\pm\infty\})$; the product of shape parameters on any edge in the triangulation becomes one, and the holonomies induced by meridian and longitude of the boundary torus act as {non-trivial} translations on the torus cusps.
Note that these conditions are all expressed as equations of shape parameters. The former equations are called {\it (Thurston's) gluing equations}, the latter is called {\it completeness condition}, and the whole set of these equations are called {\it the hyperbolicity equations}. Using Yoshida's construction in Section 4.5 of \cite{Tillmann13}, an {essential} solution ${\bold w}^{(0)}$ of the hyperbolicity equations determines a representation $$\rho_{\bold{w}^{(0)}}:\pi_1(\mathbb{S}^3\backslash (L\cup\{\pm\infty\})) =\pi_1(\mathbb{S}^3\backslash L)\longrightarrow{\rm PSL}(2,\mathbb{C}).$$
\begin{proposition}[Proposition 1.1 of \cite{Cho13c}]\label{pro21} The set $\mathcal{I}$ defined in (\ref{defH}) is the set of the hyperbolicity equations of the five-term triangulation, where the shape parameters are assigned as in Figure \ref{fig7}. \end{proposition}
{The proof of this proposition is quite complicate and technical, so we refer \cite{Cho13c} and \cite{Cho13b}. The following lemma was stated and used in \cite{Cho13c} without proof because it is almost trivial. To avoid confusion, we add the proof here.
\begin{lemma}\label{ntri} The holonomy of a meridian induced by an essential solution $\bold{w}^{(0)}=(w_1^{(0)},\ldots,w_n^{(0)})$ of $\mathcal{I}$ is always non-trivial. \end{lemma}
\begin{proof} At first, note that we assumed any component of the link diagram contains the gluing pattern in Figure \ref{glue pattern}(a). (See Footnote 4.) For the local diagram with the meridian loop $m$ and the variables $w_a, w_b$ in Figure \ref{nontricusp}(a), the corresponding cusp diagram of the five-term triangulation becomes Figure \ref{nontricusp}(b).
\begin{figure}
\caption{Holonomy of the meridian $m$}
\label{nontricusp}
\end{figure}
Note that the same shape parameter $\frac{w_b^{(0)}}{w_a^{(0)}}$ is placed in two different positions in Figure \ref{nontricusp}(b). The essentialness of the solution $\bold{w}^{(0)}$ guarantees $\frac{w_b^{(0)}}{w_a^{(0)}}\neq 0,1,\infty$, so the holonomy cannot be trivial.
\end{proof}}
From Proposition \ref{pro21} {and Lemma \ref{ntri}}, a solution $\bold{w}^{(0)}=(w_1^{(0)},\ldots,w_n^{(0)})$ of $\mathcal{I}$ induces a boundary-parabolic representation $\rho_{\bold{w^{(0)}}}$. We will make a special solution $\bold{w}^{(0)}$ from the given representation $\rho$ in the next section, which satisfies $\rho_{\bold{w^{(0)}}}=\rho$ up to conjugation.
\section{Construction of the solution}\label{sec3}
\subsection{Reviews on shadow-coloring}
This section is a summary of definitions and properties we need. For complete descriptions, see Section 2 of \cite{Cho14a}. (All definitions and Lemma \ref{lem} originally came from \cite{Kabaya14}.)
Let $\mathcal{P}$ be the set of parabolic elements of ${\rm PSL}(2,\mathbb{C})={\rm Isom^+}(\mathbb{H}^3)$. We identify $\mathbb{C}^2\backslash\{0\}/\pm$ with $\mathcal{P}$ by \begin{equation}\label{matrixcc} \left(\begin{array}{cc}\alpha &\beta\end{array}\right)
\longleftrightarrow\left(\begin{array}{cc}1+\alpha\beta & \beta^2 \\ -\alpha^2& 1-\alpha\beta\end{array}\right), \end{equation} and define operation $*$ by \begin{eqnarray*}
\left(\begin{array}{cc}\alpha & \beta\end{array}\right)* \left(\begin{array}{cc}\gamma & \delta\end{array}\right)
:=\left(\begin{array}{cc}\alpha & \beta\end{array}\right)
\left(\begin{array}{cc}1+\gamma\delta &\delta^2 \\ -\gamma^2& 1-\gamma\delta\end{array}\right)
\in \mathcal{P}, \end{eqnarray*} where this operation is actually induced by the conjugation as follows: $$ \left(\begin{array}{cc}\alpha & \beta\end{array}\right)* \left(\begin{array}{cc}\gamma & \delta\end{array}\right)\in\mathcal{P} \longleftrightarrow \left(\begin{array}{cc}\gamma&\delta\end{array}\right)
\left(\begin{array}{cc}\alpha & \beta\end{array}\right)
\left(\begin{array}{cc}\gamma &\delta\end{array}\right)^{-1}\in{\rm PSL}(2,\mathbb{C}).$$ The inverse operation $*^{-1}$ is expressed by $$\left(\begin{array}{cc}\alpha & \beta\end{array}\right)*^{-1} \left(\begin{array}{cc}\gamma & \delta\end{array}\right)
= \left(\begin{array}{cc}\alpha & \beta\end{array}\right)
\left(\begin{array}{cc}1-\gamma\delta & -\gamma^2 \\ \delta^2 & 1+\gamma\delta\end{array}\right)\in\mathcal{P},$$ and $(\mathcal{P},*)$ becomes a {\it conjugation quandle}. Here, {\it quandle} implies, for any $a,b,c\in\mathcal{P}$, the map $*b:a\mapsto a*b$ is bijective and $$a*a=a, ~(a*b)*c=(a*c)*(b*c)$$ hold. {\it Conjugation quandle} implies the operation $*$ is defined by the conjugation.
We define {\it the Hopf map} $h:\mathcal{P}\rightarrow\mathbb{CP}^1=\mathbb{C}\cup\{\infty\}$ by $$\left(\begin{array}{cc}\alpha &\beta\end{array}\right)\mapsto \frac{\alpha}{\beta}.$$ Note that the image is the fixed point of the M\"{o}bius transformation $f(z)=\frac{(1+\alpha\beta)z-\alpha^2}{\beta^2 z+(1-\alpha\beta)}$.
For an oriented link diagram $D$ of $L$ and a given boundary-parabolic representation $\rho$, we assign {\it arc-colors} $a_1,\ldots,a_r\in\mathcal{P}$ to arcs of $D$ so that each $a_k$ is the image of the meridian around the arc under the representation $\rho$. Note that, in Figure \ref{fig02}, we have \begin{equation}\label{ope}
a_m=a_l*a_k. \end{equation}
\begin{figure}
\caption{Arc-coloring}
\label{fig02}
\end{figure}
We also assign {\it region-colors} $s_1,\ldots,s_n\in\mathcal{P}$ to regions of $D$ satisfying the rule in Figure \ref{fig03}. Note that, if an arc-coloring is given, then a choice of one region-color determines all the other region-colors.
\begin{figure}
\caption{Region-coloring}
\label{fig03}
\end{figure}
\begin{lemma}[Lemma 2.4 of \cite{Cho14a}]\label{lem} Consider the arc-coloring induced by the boundary-parabolic representation $\rho:\pi_1(L)\rightarrow {\rm PSL}(2,\mathbb{C})$. Then, for any triple $(a_k,s,s*a_k)$ of an arc-color $a_k$ and its surrounding region-colors $s, s*a_k$ as in Figure \ref{fig03}, there exists a region-coloring satisfying \begin{equation*}
h(a_k)\neq h(s)\neq h(s*a_k)\neq h(a_k). \end{equation*} \end{lemma}
\begin{proof} For the given arc-colors $a_1,\ldots,a_r$, we choose region-colors $s_1,\ldots,s_n$ so that \begin{equation}\label{exi}
\{h(a_1),\ldots,h(a_r)\}\cap\{h(s_1),\ldots,h(s_n)\}=\emptyset. \end{equation} This is always possible because, {
each $h(s_k)$ is written as $h(s_k)=M_k(h(s_1))$ by a M\"{o}bius transformation $M_k$, which only depends on the arc-colors $a_1,\dots,a_r$. If we choose $h(s_1)\in\mathbb{CP}^1$ away from the finite set $$\bigcup_{1\leq k\leq n}\left\{M_k^{-1}(h(a_1)),\ldots,M_k^{-1}(h(a_r))\right\},$$ we have $h(s_k)\notin \{h(a_1),\ldots,h(a_r)\}$ for all $k$.}
Now consider Figure \ref{fig03} and assume $h(s*a_k)=h(s)$. Then we obtain \begin{equation}\label{eqnh}
h(s*a_k)=\widehat{a_k}(h(s))=h(s), \end{equation} where $\widehat{a_k}:\mathbb{CP}^1\rightarrow\mathbb{CP}^1$ is the M\"{o}bius transformation \begin{equation}\label{mob} \widehat{a_k}(z)=\frac{(1+\alpha_k\beta_k)z-\alpha_k^2}{\beta_k^2 z+(1-\alpha_k\beta_k)} \end{equation} of $a_k=\left(\begin{array}{cc}\alpha_k &\beta_k\end{array}\right)$. Then (\ref{eqnh}) implies $h(s)$ is the fixed point of $\widehat{a_k}$, which means $h(a_k)=h(s)$ and this contradicts (\ref{exi}).
\end{proof}
We remark that Lemma \ref{lem} holds for any choice of $h(s_1)\in\mathbb{CP}^1$ with only finitely many exceptions. Therefore, if we want to find a region-coloring explicitly, we first choose $h(s_1)\notin\{h(a_1),\ldots,h(a_r)\}$ and then decide $h(s_2), \ldots, h(s_n)$ using \begin{equation}\label{app1} h(s_1*a)=\widehat{a}(h(s_1)),~h(s_1*^{-1} a)=\widehat{a}^{-1}(h(s_1)). \end{equation}
If this choice does not satisfy Lemma \ref{lem}, then we change $h(s_1)$ and do the same process again. This process is very easy and it ends in finite steps. If proper $h(s_1)$ is chosen, then we can easily extend the value $h(s_1)$ to a region-color $s_1$ and find the proper region-coloring $\{s_1,\ldots,s_n\}$. This observation implies the following corollary.
\begin{corollary}\label{cor32}
Consider a sequence of diagrams $D_1,\ldots,D_m$, where each $D_{k+1}$ ($k=1,\ldots,m-1$) is obtained from $D_k$ by applying one of the Reidemeister moves in Figure \ref{pic02} once. Also assume arc-colorings of $D_1,\ldots,D_m$ are given by certain boundary-parabolic representation $\rho:\pi_1(L)\rightarrow {\rm PSL}(2,\mathbb{C})$. (Note that a region-coloring of $D_1$ determines the region-colorings of $D_2,\ldots,D_m$ uniquely.) Then there exists a region-coloring of $D_1$ satisfying Lemma \ref{lem} for all region-colorings of $D_1,\ldots,D_m$. \end{corollary}
\begin{proof} Let $s_1$ be the region-color of the unbounded region of $D_1,\ldots,D_m$. For each $D_k$, the number of values of $h(s_1)\in\mathbb{CP}^1$ that does not satisfy Lemma \ref{lem} is finite. Therefore, we can choose $h(s_1)$ so that Lemma \ref{lem} holds for all $D_1,\ldots,D_m$. By extending $h(s_1)$ to a region-color $s_1$, we can determine the region-colorings of $D_1,\ldots,D_m$ satisfying Lemma \ref{lem} uniquely.
\end{proof}
The arc-coloring induced by $\rho$ together with the region-coloring satisfying Lemma \ref{lem} is called {\it the shadow-coloring induced by} $\rho$. We choose an element $p\in\mathcal{P}$ so that \begin{equation}\label{p} h(p)\notin\{h(a_1),\ldots,h(a_r), h(s_1),\ldots,h(s_n)\}. \end{equation} The geometric shape of the five-term triangulation in Section \ref{sec2} will be determined by the shadow-coloring induced by $\rho$ and $p$ in the next section.
From now on, we fix the representatives of shadow-colors in $\mathbb{C}^2\backslash\{0\}$. As mentioned in \cite{Cho14a}, the representatives of some arc-colors may satisfy (\ref{ope}) up to sign, in other words, $a_m=\pm (a_l*a_k)$. However, the representatives of the region-colors are uniquely determined due to the fact $s*(\pm a)=s*a$ for any region-color $s$ and any arc-color $a$.
For $a=\left(\begin{array}{cc}\alpha_1 &\alpha_2\end{array}\right)$ and $ b=\left(\begin{array}{cc}\beta_1 &\beta_2\end{array}\right)$ in $\mathbb{C}^2\backslash\{0\}$, we define {\it determinant} $\det(a,b)$ by \begin{equation*}
\det(a,b):=\det\left(\begin{array}{cc}\alpha_1 & \alpha_2 \\\beta_1 & \beta_2\end{array}\right)=\alpha_1 \beta_2-\alpha_2 \beta_1. \end{equation*} Then the determinant satisfies $\det(a*c,b*c)=\det(a,b)$ for any $a,b,c\in\mathbb{C}^2\backslash\{0\}$. Furthermore, for $v_0,\ldots,v_3\in\mathbb{C}^2\backslash \{0\}$, the cross-ratio can be expressed using the determinant by $$[h(v_0),h(v_1),h(v_2),h(v_3)]=\frac{\det(v_0,v_3)\det(v_1,v_2)}{\det(v_0,v_2)\det(v_1,v_3)}.$$ (For the proofs, see Section 2 of \cite{Cho14a}.)
\subsection{Geometric shape of the five-term triangulation}\label{sec22}
Note that the five-term triangulation was already defined in Section \ref{sec2}. Consider the crossings in Figure \ref{fig04} with the shadow-colorings induced by $\rho$, and let $w_a,\ldots,w_d$ be the variables assigned to regions of $D$.
\begin{figure}
\caption{Crossings with shadow-colorings and the region variables}
\label{fig04}
\end{figure}
We place tetrahedra at each crossings of $D$ and assign coordinates of them as in Figure \ref{fig06} so as to make them hyperbolic ideal tetrahedra in the upper-space model of the hyperbolic 3-space $\mathbb{H}^3$.
\begin{figure}
\caption{Five-term triangulation at the crossing in Figure \ref{fig04}}
\label{fig06}
\end{figure}
\begin{proposition}[Proposition 2.3 of \cite{Cho14c}]\label{pro} All the tetrahedra in Figure \ref{fig06} are non-degenerate. \end{proposition}
According to Section 2 of \cite{Kabaya12} and the proof of Theorem 5 of \cite{Kabaya14}, the five-term triangulation defined by Figure \ref{fig06} induces the given representation\footnote{The triangulation of \cite{Kabaya12} is different from ours. However, the fundamental domain obtained by the five-term triangulation coincides with that of \cite{Kabaya12}, so it induces the same representation. (Our triangulation is obtained by choosing different subdivision of the same fundamental domain. See Section 2.2 of \cite{Cho14c} for details.)} $\rho:\pi_1(L)\rightarrow{\rm PSL}(2,\mathbb{C})$ and the shape parameters of this triangulation satisfy the gluing equations of all edges. (The face-pairing maps are the isomorphisms induced by the M\"{o}bius transformations of ${a_1},\ldots,a_r\in{\rm PSL}(2,\mathbb{C})$. {Note that this construction is based on the construction method developed at \cite{Neumann99} and \cite{Zickert09}.}) Furthermore, the representation $\rho$ is boundary-parabolic, which implies the shape-parameters satisfy the hyperbolicity equations of the five-term triangulation.
\subsection{Formula of the solution ${\bold w}^{(0)}$}\label{sec33} Consider the crossings in Figure \ref{fig04} and the tetrahedra in Figure \ref{fig06}. For the positive crossing, we assign shape parameters to the edges as follows: \begin{itemize} \item $\displaystyle\frac{w_d}{w_a}$ to $(h(a_k), h(s*a_k))$ of $(h(p*a_k),h(p),h(a_k), h(s*a_k))$, \item $\displaystyle\frac{w_b}{w_c}$ to $(h(a_k), h((s*a_l)*a_k))$ of $-(h(p*a_k),h(p),h(a_k), h((s*a_l)*a_k))$, \item $\displaystyle\frac{w_b}{w_a}$ to $(h(p), h(a_l*a_k))$ of $(h(p), h(a_l*a_k),h(s*a_k), h((s*a_l)*a_k))$ and \item $\displaystyle\frac{w_d}{w_c}$ to $(h(p), h(a_l))$ of $-(h(p), h(a_l),h(s), h(s*a_l))$, respectively. \end{itemize} On the other hand, for the negative crossing, we assign shape parameters to the edges as follows: \begin{itemize} \item $\displaystyle\frac{w_a}{w_b}$ to $(h(a_k), h((s*a_l)*a_k))$ of $-(h(p),h(p*a_k),h(a_k), h((s*a_l)*a_k))$, \item $\displaystyle\frac{w_c}{w_d}$ to $(h(a_k), h(s*a_k))$ of $(h(p),h(p*a_k),h(a_k), h(s*a_k))$, \item $\displaystyle\frac{w_c}{w_b}$ to $(h(p), h(a_l*a_k))$ of $(h(p), h(a_l*a_k), h((s*a_l)*a_k),h(s*a_k))$ and \item $\displaystyle\frac{w_a}{w_d}$ to $(h(p), h(a_l))$ of $-(h(p), h(a_l),h(s*a_l), h(s))$, respectively. \end{itemize} Note that these assignments coincide with the one defined in Figure \ref{fig7}.
\begin{proposition}[Theorem 1.1 of \cite{Cho14c}]\label{pro2} For a region of $D$ with region-color $s_k$ and region-variable $w_k$, we define \begin{equation}\label{main} w_k^{(0)}:=\det(p,s_k). \end{equation} Then, $w_k^{(0)}\not= 0$ and ${\bold w^{(0)}}=(w_1^{(0)},\ldots,w_n^{(0)})$ is an {essential} solution of the hyperbolicity equations in $\mathcal{I}$ of (\ref{defH}). Furthermore, the solution satisfies $\rho_{\bold w^{(0)}}=\rho$ up to conjugation and \begin{equation}\label{volume} W_0(\bold w^{(0)})\equiv i({\rm vol}(\rho)+i\,{\rm cs}(\rho))~~({\rm mod}~\pi^2). \end{equation} \end{proposition}
\begin{proof}
The first property $w_k^{(0)}\not= 0$ is trivial from the definition of $p$ in (\ref{p}). From the discussion below Proposition \ref{pro}, the shape parameters of the five-term triangulation defined by Figure \ref{fig06} satisfy the hyperbolicity equations and the fundamental domain induces the boundary-parabolic representation $\rho$.
On the other hand, direct calculation shows the values $w_k^{(0)}$ defined in (\ref{main}) determines the same shape parameter of the five-term triangulation defined by Figure \ref{fig06}. Specifically, for the first two cases of the positive crossing, the shape parameters assigned to edges $(h(a_k), h(s*a_k))$ and $(h(a_k), h((s*a_l)*a_k))$ are the cross-ratios \begin{eqnarray*} [h(p*a_k),h(p),h(a_k), h(s*a_k)]&=&\frac{\det(p*a_k,s*a_k)\det(p,a_k)}{\det(p*a_k,a_k)\det(p,s*a_k)}\\ &=&\frac{\det(p,s)\det(p,a_k)}{\det(p,a_k)\det(p,s*a_k)}=\frac{w_d^{(0)}}{w_a^{(0)}},\\ {[h}(p*a_k),h(p),h(a_k), h((s*a_l)*a_k)]^{-1} &=&\frac{\det(p*a_k,a_k)\det(p,(s*a_l)*a_k)}{\det(p*a_k,(s*a_l)*a_k)\det(p,a_k)}\\ &=&\frac{\det(p,a_k)\det(p,(s*a_l)*a_k)}{\det(p,s*a_l)\det(p,a_k)}=\frac{w_b^{(0)}}{w_c^{(0)}}, \end{eqnarray*} respectively, and all the other cases can be verified by the same way. From Proposition \ref{pro21}, we conclude that ${\bold w^{(0)}}=(w_1^{(0)},\ldots,w_n^{(0)})$ is an {essential} solution of $\mathcal{I}$.
Finally, the identity (\ref{volume}) was already proved in Theorem 1.2 of \cite{Cho13c}.
\end{proof}
{In Appendix \ref{app}, we will show that any {essential} solution of $\mathcal{I}$ can be constructed by certain shadow-coloring.}
\section{Reidemeister transformations on the solution}\label{sec4}
In this section, we show how the solution ${\bold w^{(0)}}$ of $\mathcal{I}$ defined in Proposition \ref{pro2} changes under the Reidemeister moves. We assume all the region-colorings in this and later sections satisfy Corollary \ref{cor32} {so that the original and the transformed solutions are both essential}.
At first, we introduce very simple, but useful lemma. {Recall that, according to Proposition \ref{pro21}, the set $\mathcal{I}$ defined in (\ref{defH}) is the set of the hyperbolicity equations.} The following lemma shows the hyperbolicity equations do not change under the change of the orientation.
\begin{figure}
\caption{Change of orientation}
\label{fig10}
\end{figure}
\begin{lemma}\label{lem40} For the potential function $W^j(w_a,w_b,w_c,w_d)$ and $W^{j'}(w_a,w_b,w_c,w_d)$ of Figure \ref{fig10}(a) and (b), respectively, we have $$\exp(w_k\frac{\partial W^j}{\partial w_k})=\exp(w_k\frac{\partial W^{j'}}{\partial w_k}),$$ for $k=a,b,c,d$. \end{lemma}
\begin{proof} It is easily verified by direct calculation. For example, $$\exp(w_a\frac{\partial W^j}{\partial w_a})=\frac{(w_b-w_a)(w_d-w_a)}{w_bw_d-w_cw_a} =\exp(w_a\frac{\partial W^{j'}}{\partial w_a}).$$
\end{proof}
\subsection{Reidemeister 1st move}
Consider the Reidemeister 1st moves in Figure \ref{LocalMove1}. Let $\alpha\in\mathcal{P}$ be the arc-color, $s,s*\alpha,(s*\alpha)*\alpha\in\mathcal{P}$ be the region-colors and $w_a,w_b,w_c$ be the variables of the potential function. Then, by (\ref{main}), \begin{equation}\label{eq41} w_a^{(0)}=\det(p,s),~w_b^{(0)}=\det(p,s*\alpha), ~w_c^{(0)}=\det(p,(s*\alpha)*\alpha). \end{equation}
\begin{figure}
\caption{First Reidemeister moves}
\label{LocalMove1}
\end{figure}
\begin{lemma}\label{LemR1} The values $w_a^{(0)},w_b^{(0)},w_c^{(0)}$ defined in (\ref{eq41}) satisfy $$w_c^{(0)}=2w_b^{(0)}-w_a^{(0)}.$$ \end{lemma}
\begin{proof} Using the identification (\ref{matrixcc}), let \begin{equation*}
\alpha=\left(\begin{array}{cc}\alpha_1 & \alpha_2\end{array}\right)
\longleftrightarrow A=\left(\begin{array}{cc}1+\alpha_1\alpha_2 &\alpha_2^2 \\-\alpha_1^2 & 1-\alpha_1\alpha_2\end{array}\right). \end{equation*} Then $s*\alpha=sA\in\mathcal{P}$ and $(s*\alpha)*\alpha=sA^2\in\mathcal{P}$ holds by the definition of the operation $*$. Furthermore, by the Cayley-Hamilton theorem, the matrix $A$ satisfies $$A^2-2A+I=0,$$ where $I$ is the $2\times 2$ identity matrix. Using these, we obtain $$w_c^{(0)}=\det(p,(s*\alpha)*\alpha)=\det(p,sA^2)=2\det(p,sA )-\det(p,s I )=2w_b^{(0)}-w_a^{(0)}.$$ \end{proof}
\subsection{Reidemeister 2nd moves}
Consider the Reidemeister 2nd moves in Figure \ref{LocalMove2}. Let $\alpha$, $\beta$, $\alpha*\beta\in\mathcal{P}$ be the arc-colors, $s$, $s*\alpha$, $s*\beta$, $(s*\alpha)*\beta\in\mathcal{P}$ be the region-colors and $w_a,\ldots,w_e$ be the variables of the potential function. Then, by (\ref{main}), \begin{equation*} w_a^{(0)}=\det(p,s*\beta),~w_b^{(0)}=w_d^{(0)}=\det(p,(s*\alpha)*\beta), ~w_c^{(0)}=\det(p,s*\alpha),~w_e^{(0)}=\det(p,s) \end{equation*} for the case of R2 move in Figure \ref{LocalMove2}(a), and \begin{equation*} w_a^{(0)}=\det(p,s*\alpha),~w_b^{(0)}=w_d^{(0)}=\det(p,s), ~w_c^{(0)}=\det(p,s*\beta),~w_e^{(0)}=\det(p,(s*\alpha)*\beta) \end{equation*} for the case of R2$'$ move in Figure \ref{LocalMove2}(b).
\begin{figure}
\caption{Second Reidemeister moves}
\label{LocalMove2}
\end{figure}
\begin{lemma}\label{LemR2} Let $T(W)(...,w_a,w_b,w_c,w_d,w_e,...)$ be the potential function (\ref{R2a}) (or (\ref{R2c})) of the diagram in Figure \ref{LocalMove2}(a) (or (b)) after applying $R2$ (or $R2'$) move. Then $w_d^{(0)}=w_b^{(0)}$ and $w_e^{(0)}$ is uniquely determined from the values of the parameters around the region of $w_b$ by the equation \begin{equation*} \exp(w_b\frac{\partial T(W)}{\partial w_b})=1. \end{equation*} \end{lemma}
\begin{proof} At first, $$\exp(w_e\frac{\partial T(W)}{\partial w_e})=\frac{w_aw_c-w_bw_e}{w_aw_c-w_dw_e}=1$$ induces $w_d^{(0)}=w_b^{(0)}$.
Consider the case of Figure \ref{LocalMove2}(a). The variable $w_e$ of the potential function $T(W)$ appears only inside the function $V_{R2}(w_a, w_b, w_c, w_d, w_e)$ defined in Section \ref{sec12}, and direct calculation shows $$\exp(w_b\frac{\partial V_{R2}}{\partial w_b})
=\frac{w_a w_c-w_b w_e}{(w_c-w_b)(w_a-w_b)}.$$ Therefore, the equation $\exp(w_b\frac{\partial T(W)}{\partial w_b})=1$ is linear with respect to $w_e$ and it determines $w_e^{(0)}$ uniquely.
The case of Figure \ref{LocalMove2}(b) is trivial from Lemma \ref{lem40} and the above.
\end{proof}
Remark that the equation $\exp(w_d\frac{\partial T(W)}{\partial w_d})=1$ also determines the same value $w_e^{(0)}$.
\subsection{Reidemeister 3rd move}
Consider the Reidemeister 3rd move in Figure \ref{LocalMove3}. Let $\alpha$, $\beta$, $\gamma$, $\beta*\gamma$, $(\alpha*\beta)*\gamma=(\alpha*\gamma)*(\beta*\gamma)\in\mathcal{P}$ be the arc-colors, $s$, $s*\alpha$, $s*\beta$, $s*\gamma$, $(s*\alpha)*\beta$, $(s*\alpha)*\gamma$, $(s*\beta)*\gamma$, $((s*\alpha)*\beta)*\gamma\in\mathcal{P}$ be the region-colors and $w_a,\ldots,w_g,w_h$ be the variables of the potential function. Then, by (\ref{main}), \begin{align} w_a^{(0)}=\det(p,s),~w_b^{(0)}=\det(p,s*\gamma),~w_c^{(0)}=\det(p,(s*\beta)*\gamma),\label{eq43}\\ w_d^{(0)}=\det(p,((s*\alpha)*\beta)*\gamma),~w_e^{(0)}=\det(p,(s*\alpha)*\beta),\nonumber\\ w_f^{(0)}=\det(p,s*\alpha),~w_g^{(0)}=\det(p,s*\beta),~w_h^{(0)}=\det(p,(s*\alpha)*\gamma).\nonumber \end{align}
\begin{figure}
\caption{Third Reidemeister move}
\label{LocalMove3}
\end{figure}
\begin{lemma}\label{LemR3} The values $w_a^{(0)},\ldots,w_h^{(0)}$ defined in (\ref{eq43}) satisfy $$w_d^{(0)}w_g^{(0)}-w_c^{(0)}w_e^{(0)}=w_a^{(0)}w_h^{(0)}-w_b^{(0)}w_f^{(0)}.$$ \end{lemma}
\begin{proof} Using the identification (\ref{matrixcc}), let $$\alpha\leftrightarrow A,~ \beta\leftrightarrow B,~\gamma\leftrightarrow C,$$ {and $$A=\left(\begin{array}{cc}1+a_1 a_2 &a_2^2 \\-a_1^2 & 1-a_1a_2\end{array}\right), B=\left(\begin{array}{cc}1+b_1 b_2 &b_2^2 \\-b_1^2& 1-b_1b_2\end{array}\right), C=\left(\begin{array}{cc}1+c_1 c_2 & c_2^2\\ -c_1^2& 1-c_1c_2\end{array}\right).$$ Also, put $$p=\left(\begin{array}{cc}p_1 &p_2\end{array}\right)\text{ and } s=\left(\begin{array}{cc}s_1 &s_2\end{array}\right).$$ Then direct calculation shows the following identity: \begin{align*} &\det(p,sABC)\det(p,sB)-\det(p,sBC)\det(p,sAB)\\ &=-(c_2p_1-c_1p_2)^2(a_2s_1-a_1s_2)^2\\ &=\det(p,s)\det(p,sAC)-\det(p,sC)\det(p,sA). \end{align*} ( Although this identity looks very elementary, the authors cannot find any other proof except the direct calculation.) Applying (\ref{eq43}) to this identity proves the lemma.}
\end{proof}
\section{Orientation change and the mirror image}\label{sec5}
The proofs of the relations of solutions in Lemma \ref{LemR1}-\ref{LemR3} needed orientation of the link diagram. However, we can show that the same relations still hold for any choice of orientations and for the mirror images. (Exact statements will appear below.) These results are very useful when we consider the actual examples because they reduce the number of the Reidemeister moves.
\begin{lemma}[Uniqueness of the solution]\label{uniquelemma} Let $\bold{w}=(w_1^{(0)},\ldots,w_n^{(0)})$ be the solution of the hyperbolicity equations obtained by the shadow-coloring induced by $\rho$. (See Proposition \ref{pro2} for the construction.) After applying one of the Reidemeister moves R1, R1$'$, R3 and (R3)$^{-1}$ to the link diagram once,
assume the new variable $w_{n+1}$ appeared. Then the value $w_{n+1}^{(0)}$ satisfying $(w_1^{(0)},\ldots,w_{n+1}^{(0)})$ to be a solution of the hyperbolicity equations is uniquely determined by the values $w_1^{(0)},\ldots,w_n^{(0)}$. Likewise, if new variables $w_{n+1}$ and $w_{n+2}$ appeared after applying R2 or R2$\,'$ move once,
then the values $w_{n+1}^{(0)}$ and $w_{n+2}^{(0)}$ satisfying the hyperbolicity equations are uniquely determined by the values $w_1^{(0)},\ldots,w_n^{(0)}$. \end{lemma}
\begin{proof} Note that the main idea was already appeared in the proof of Lemma \ref{LemR2}.
Let $T(W)$ be the potential function of the link diagram obtained after applying the Reidemeister move once. Now consider Figure \ref{pic02}. In Figure \ref{pic02}(a), the value $w_{c}^{(0)}$ is uniquely determined by the equation $\exp(w_a\frac{\partial T(W)}{\partial w_a})=1$. In Figure \ref{pic02}(b), the equation $\exp(w_e\frac{\partial T(W)}{\partial w_e})=1$ uniquely determines the value $w_{d}^{(0)}=w_{b}^{(0)}$ and $w_{e}^{(0)}$ is uniquely determined by the equation $\exp(w_b\frac{\partial T(W)}{\partial w_b})=1$. In Figure \ref{pic02}(c), the values $w_{g}^{(0)}$ and $w_{h}^{(0)}$ are uniquely determined by the equations $\exp(w_d\frac{\partial T(W)}{\partial w_d})=1$ and $\exp(w_a\frac{\partial T(W)}{\partial w_a})=1$, respectively.
\end{proof}
{ Now we introduce {\it the local orientation change} and show how the arc-color changes. From (\ref{matrixcc}), we obtain \begin{equation*} \pm i \left(\begin{array}{cc}\alpha &\beta\end{array}\right)
\longleftrightarrow\left(\begin{array}{cc}1-\alpha\beta & -\beta^2 \\ \alpha^2& 1+\alpha\beta\end{array}\right)
=\left(\begin{array}{cc}1+\alpha\beta & \beta^2 \\ -\alpha^2& 1-\alpha\beta\end{array}\right)^{-1} \end{equation*} and $s*^{-1}(\pm i a_k)=s*a_k.$ Therefore we define the local orientation change as in Figure \ref{orichang}. Note that the arc-color of the reversed orientated arc changes to $\pm i a_k$, but the region-colors are still well-defined. Therefore, the invariance of the region-colors shows the invariance of the solution under the local orientation change.
\begin{figure}
\caption{Local orientation change}
\label{orichang}
\end{figure} }
\begin{figure}
\caption{Un-oriented Reidemeister moves}
\label{pic15}
\end{figure}
\begin{proposition}\label{unR} Consider the un-oriented Reidemeister moves in Figure \ref{pic15} and a link diagram $D$ containing one of Figure \ref{pic15}. Let $(\ldots,w_a^{(0)},w_b^{(0)},\ldots)$ be the solution of the hyperbolicity equations obtained by the shadow-coloring induced by $\rho$. Then, for Figure \ref{pic15}(a), the values of the variables satisfy \begin{equation}\label{eq20} w_c^{(0)}=2w_b^{(0)}-w_a^{(0)}, \end{equation} and, for Figure \ref{pic15}(c), the values satisfy \begin{equation}\label{eq21} w_d^{(0)}w_g^{(0)}-w_c^{(0)}w_e^{(0)}=w_a^{(0)}w_h^{(0)}-w_b^{(0)}w_f^{(0)}. \end{equation} \end{proposition}
\begin{proof} {
By applying the local orientation change whenever it is necessary, we can assign the local orientation of Figure \ref{pic02} to the un-oriented diagrams in Figure \ref{pic15}. Then, from the oriented Reidemeister transformations, we obtain the relations (\ref{eq20}) and (\ref{eq21}). The solution of the hyperbolicity equations is invariant under the local orientation change, so the relations are independent with the choice of the orientations.
}
\end{proof}
As a result of Proposition \ref{unR}, we define the {\it un-oriented Reidemeister 1st and 3rd transformations of solutions} by the same formulas of the oriented version in Definition \ref{def1}. On the other hand, the formula of the oriented Reidemeister 2nd move in Definition \ref{def1} needed an explicit potential function, which depends on the orientation. However, we can formulate the Reidemeister 2nd move without orientations using the following lemma. At first, we define the weight of the corner as in Figure \ref{pic16}.
\begin{figure}
\caption{Weight $x_a^j$ assigned to the corner of the crossing $j$}
\label{pic16}
\end{figure}
\begin{lemma}\label{lem53} For an oriented link diagram $D$, let $W$ be the potential function of $D$ and choose a region R assigned with variable $w_k$. Then, $$\exp(w_k\frac{\partial W}{\partial w_k})=\prod_{j}x_k^j,$$ where $j$ is over all the crossings adjacent to the region R. \end{lemma}
\begin{proof} The proof can be easily obtained by direct calculations. In the case of Figure \ref{pic01}(a), \begin{align*} &\exp(w_a\frac{\partial W^j}{\partial w_a})=\frac{(w_b-w_a)(w_c-w_a)}{w_bw_d-w_cw_a}=x_a^j, &\exp(w_b\frac{\partial W^j}{\partial w_b})=\frac{w_aw_c-w_dw_b}{(w_a-w_b)(w_c-w_b)}=x_b^j,\\ &\exp(w_c\frac{\partial W^j}{\partial w_c})=\frac{(w_b-w_c)(w_d-w_c)}{w_bw_d-w_aw_c}=x_c^j, &\exp(w_d\frac{\partial W^j}{\partial w_d})=\frac{w_aw_c-w_bw_d}{(w_a-w_d)(w_c-w_d)}=x_d^j, \end{align*} and the case of Figure \ref{pic01}(b) can be obtained from Lemma \ref{lem40}. The main equation is obtained by $$\exp(w_k\frac{\partial W}{\partial w_k})=\exp(w_k\frac{\partial (\sum_j W^j)}{\partial w_k}) =\prod_{j}\exp(w_k\frac{\partial W^j}{\partial w_k})=\prod_{j}x_k^j,$$ where $j$ is over all the crossings adjacent to the region R. (Note that if $j$ is not adjacent to the region R, then $w_k\frac{\partial W^j}{\partial w_k}=0$.)
\end{proof}
The weight does not depends on the orientation, so we can describe the equation $\exp(w_b\frac{\partial W}{\partial w_b})=1$ in the right-hand side of Figure \ref{pic15}(b) without orientation. This equation determines the value $w_e^{(0)}$ uniquely and it defines {\it the un-oriented Reidemeister 2nd move of the solution}.
Now consider a link diagram $D$ and its mirror image\footnote{The mirror image is defined as follows: assume the link is in the $(x,y,z)$-space and the diagram $D$ is obtained by the projection along $z$-axis to the $(x,y)$-plane. Then $\overline{D}$ is the mirror image of $D$ by the reflection on the $(y,z)$-plane. For an example, see Figure \ref{pic15}(c) and Figure \ref{pic17}.} $\overline{D}$. When the variables $w_a,w_b,\ldots$ are assigned to $D$, we always assume the same variables are assigned to the same mirrored regions. Let $W(w_a,w_b,\ldots)$ be the potential function of $D$. Then, from the definition of the potential functions in Figure \ref{pic01}, the potential function of $\overline{D}$ becomes $-W(w_a,w_b,\ldots)$. (Note that we are using here the invariance of the potential functions in Figure \ref{pic01} under $w_b\leftrightarrow w_d$.) This suggests that the hyperbolicity equations in $\mathcal{I}$ are invariant under the mirroring.
\begin{lemma}\label{R3mirror} Consider the diagrams in Figure \ref{pic17} and let $(\ldots,w_a^{(0)},\ldots,w_f^{(0)},w_g^{(0)},\ldots)$ and $(\ldots,w_a^{(0)},\ldots,w_f^{(0)},w_h^{(0)},\ldots)$ be the solutions of the hyperbolicity equations obtained by the shadow-coloring induced by $\rho$. The the values satisfy the equation (\ref{eq21}). \end{lemma}
\begin{proof} For Figure \ref{pic15}(c), the relation (\ref{eq21}) holds. The hyperbolicity equations are invariant under the mirroring, {so a solution on the diagram $D$ is also a solution on $\overline{D}$. Therefore, a pair of solutions related by the relation (\ref{eq21}) on Figure \ref{pic15}(c) is also a pair of solution on Figure \ref{pic17} related by the same relation. From the uniqueness of the solution in Lemma \ref{uniquelemma}, this is the only relation induced by a shadow-coloring.}
\end{proof}
\begin{figure}
\caption{Mirror image of Figure \ref{pic15}(c)}
\label{pic17}
\end{figure}
\begin{proposition} Consider a diagram $D'$ is obtained from $D$ by applying one un-oriented Reidemeister move, assume variables $w_1,\ldots,w_n$ are assigned to regions of $D$ and $w_{n+1}$ is assigned to a newly appeared region of $D'$. Then Lemma \ref{uniquelemma} showed that the value $w_{n+1}^{(0)}$ is uniquely determined from $w_1^{(0)},\ldots, w_n^{(0)}$ by a certain equation. If we consider the mirror image $\overline{D}'$ obtained from $\overline{D}$, then the value $w_{n+1}^{(0)}$ of the mirror image is uniquely determined by the same equation. \end{proposition}
\begin{proof} The mirror images of the un-oriented Reidemeister 1st and 2nd moves are just $\pi$-rotations of the original moves, so the same equation holds for each cases. The mirror image of the un-oriented Reidemeister 3rd move was already proved in Lemma \ref{R3mirror}.
\end{proof}
\begin{exa}\label{exa1} Consider Figure \ref{pic17} and let $(\ldots,w_a^{(0)},\ldots,w_f^{(0)},w_g^{(0)},\ldots)$ and $(\ldots,w_a^{(0)},\ldots,w_f^{(0)},w_h^{(0)},\ldots)$ be the solutions of the hyperbolicity equations of each diagrams obtained by the shadow-colorings induced by $\rho$. The the variables satisfy \begin{equation*} w_d^{(0)}w_g^{(0)}-w_c^{(0)}w_e^{(0)}=w_a^{(0)}w_h^{(0)}-w_b^{(0)}w_f^{(0)}, \end{equation*} which is the same equation with (\ref{eq21}). \end{exa}
\section{Examples}\label{sec6}
\begin{exa}\label{exa2} Consider the twisting move in Figure \ref{pic18} and let $(\ldots,w_a^{(0)},w_b^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},\ldots)$ and $(\ldots,w_a^{(0)},w_c^{(0)},w_d^{(0)},w_e^{(0)},w_f^{(0)},w_g^{(0)},\ldots)$ be the solutions of the hyperbolicity equations of each diagrams obtained by the shadow-colorings induced by $\rho$.
\begin{figure}
\caption{Twisting move}
\label{pic18}
\end{figure}
If we apply the twisting move from the left to the right, then $w_f^{(0)}$ and $w_g^{(0)}$ are uniquely determined by $$w_f^{(0)}=2w_a^{(0)}-w_b^{(0)}~\text{ and }~w_d^{(0)} w_g^{(0)}-w_c^{(0)} w_e^{(0)}=w_f^{(0)} w_b^{(0)}-(w_a^{(0)})^2,$$ and if we apply it from the right to the left, then $w_b^{(0)}$ is uniquely determined by $$w_b^{(0)}=2w_a^{(0)}-w_f^{(0)}.$$ \end{exa}
\begin{proof} By adding a kink from the left-hand side of Figure \ref{pic18}, we obtain Figure \ref{pic19} and the equation $$w_f^{(0)}=2w_a^{(0)}-w_b^{(0)}.$$
\begin{figure}
\caption{Adding a kink}
\label{pic19}
\end{figure}
Applying Example \ref{exa1} to Figure \ref{pic19}, we obtain another equation $$w_d^{(0)} w_g^{(0)}-w_c^{(0)} w_e^{(0)}=w_f^{(0)} w_b^{(0)}-(w_a^{(0)})^2.$$ \end{proof}
Now we will show the changes of the solution from one figure-eight knot diagram to its mirror image. At first, consider Figure \ref{example1}.
\begin{figure}
\caption{Figure-eight knot $4_1$ with parameters}
\label{example1}
\end{figure}
Let $\rho:\pi_1(4_1)\rightarrow{\rm PSL}(2,\mathbb{C})$ be the boundary-parabolic representation defined by the arc-colors $$a_1=\left(\begin{array}{cc}0 &t\end{array}\right),~a_2=\left(\begin{array}{cc}1 &0\end{array}\right),~ a_3=\left(\begin{array}{cc}-t &1+t\end{array}\right),~a_4=\left(\begin{array}{cc}-t &t\end{array}\right),$$ where $t$ is a solution of $t^2+t+1=0$, and let one region-color $s_1=\left(\begin{array}{cc}1 &1\end{array}\right)$. Then the other region-colors become \begin{align*} s_2=\left(\begin{array}{cc}0 &1\end{array}\right),~ s_3=\left(\begin{array}{cc}-t-1 &t+2\end{array}\right),~s_4=\left(\begin{array}{cc}-2t-1 &2t+3\end{array}\right),\\ s_5=\left(\begin{array}{cc}-2t-1 &t+4\end{array}\right),~s_6=\left(\begin{array}{cc}1 &t+2\end{array}\right). \end{align*}
The potential function $W(w_1,\ldots,w_6)$ of Figure \ref{example1} is \begin{align*} W=\left\{{\rm Li}_2(\frac{w_1}{w_2})+{\rm Li}_2(\frac{w_1}{w_4})-{\rm Li}_2(\frac{w_1 w_3}{w_2 w_4}) -{\rm Li}_2(\frac{w_2}{w_3})-{\rm Li}_2(\frac{w_4}{w_3})+\frac{\pi^2}{6}-\log\frac{w_2}{w_3}\log\frac{w_4}{w_3}\right\}\\ +\left\{{\rm Li}_2(\frac{w_3}{w_2})+{\rm Li}_2(\frac{w_3}{w_6})-{\rm Li}_2(\frac{w_1 w_3}{w_2 w_6}) -{\rm Li}_2(\frac{w_2}{w_1})-{\rm Li}_2(\frac{w_6}{w_1})+\frac{\pi^2}{6}-\log\frac{w_2}{w_1}\log\frac{w_6}{w_1}\right\}\\ +\left\{-{\rm Li}_2(\frac{w_4}{w_3})-{\rm Li}_2(\frac{w_4}{w_5})+{\rm Li}_2(\frac{w_4 w_6}{w_3 w_5}) +{\rm Li}_2(\frac{w_3}{w_6})+{\rm Li}_2(\frac{w_5}{w_6})-\frac{\pi^2}{6}+\log\frac{w_3}{w_6}\log\frac{w_5}{w_6}\right\}\\ +\left\{-{\rm Li}_2(\frac{w_6}{w_1})-{\rm Li}_2(\frac{w_6}{w_5})+{\rm Li}_2(\frac{w_4 w_6}{w_1 w_5}) +{\rm Li}_2(\frac{w_1}{w_4})+{\rm Li}_2(\frac{w_5}{w_4})-\frac{\pi^2}{6}+\log\frac{w_1}{w_4}\log\frac{w_5}{w_4}\right\}. \end{align*}
By putting $p=\left(\begin{array}{cc}2 &1\end{array}\right)$, we obtain \begin{align} w_1^{(0)}=\det(p,s_1)=1,~w_2^{(0)}=\det(p,s_2)=2,~w_3^{(0)}=\det(p,s_3)=3t+5,\label{sol1}\\ w_4^{(0)}=\det(p,s_4)=6t+7,~w_5^{(0)}=\det(p,s_5)=4t+9,~w_6^{(0)}=\det(p,s_6)=2t+3,\nonumber \end{align} and $(w_1^{(0)},\ldots,w_6^{(0)})$ becomes a solution of
$\mathcal{I}=\{\exp(w_k\frac{\partial W}{\partial w_k})=1~|~k=1,\ldots,6\}$. Furthermore, we obtain \begin{equation*} W_0(w_1^{(0)},\ldots,w_6^{(0)})\equiv i({\rm vol}(\rho)+i\,{\rm cs}(\rho))~~({\rm mod}~\pi^2), \end{equation*} and numerical calculation verifies it by \begin{equation*} W_0(w_1^{(0)},\ldots,w_6^{(0)})=
\left\{\begin{array}{ll}i(2.0299...+0\,i)=i({\rm vol}(4_1)+i\,{\rm cs}(4_1))&\text{ if }t=\frac{-1-\sqrt{3} \,i}{2}, \\
i(-2.0299...+0\,i)=i(-{\rm vol}(4_1)+i\,{\rm cs}(4_1))&\text{ if }t=\frac{-1+\sqrt{3}\,i}{2}. \end{array}\right. \end{equation*}
Note that the above example was already appeared in Section 3.1.\;of \cite{Cho14c}. From Theorem \ref{thm1}, we can easily specify the discrete faithful representation by Figure \ref{example1} together with the solution (\ref{sol1}). (The explicit construction of the representation can be done by applying Yoshida's construction of \cite{Tillmann13} to the five-term triangulation defined in Figure \ref{fig7}.)
Now we will apply (un-oriented) Reidemeister moves to the solution in (\ref{sol1}). Consider the changes of the figure-eight knot diagrams in Figure \ref{mirror4_1}.
\begin{figure}
\caption{Changing the figure-eight knot diagram to its mirror image}
\label{mirror4_1}
\end{figure}
Note that Figure \ref{mirror4_1}(c) is obtained by the Reidemeister 3rd move, Figures \ref{mirror4_1}(d)-(e) are obtained by the twist move defined in Figure \ref{pic18} and Figure \ref{mirror4_1}(g) is obtained by the rotation. Then the values of the variables are determined by \begin{align*} w_7^{(0)}&=\frac{w_1^{(0)}w_5^{(0)}(w_3^{(0)}-w_4^{(0)})^2-(w_1^{(0)}w_3^{(0)}-w_2^{(0)}w_4^{(0)})(w_3^{(0)}w_5^{(0)}-w_4^{(0)}w_6^{(0)})}{w_4^{(0)}(w_3^{(0)}-w_4^{(0)})^2}\\ &=\frac{6t^2+37t+36}{(3t+2)^2}=-5t-3,\\ w_8^{(0)}&=w_4^{(0)}=6t+7,\\ w_9^{(0)}&=\frac{w_4^{(0)}w_7^{(0)}-w_1^{(0)}w_5^{(0)}+w_2^{(0)}w_6^{(0)}}{w_3^{(0)}}=\frac{-30t^2-53t-24}{3t+5}=-7t-3,\\ w_{10}^{(0)}&=2w_2^{(0)}-w_3^{(0)}=-3t-1,\\ w_{11}^{(0)}&=\frac{w_3^{(0)}w_{10}^{(0)}-(w_2^{(0)})^2+w_1^{(0)}w_9^{(0)}}{w_6^{(0)}}=\frac{-9t^2-25t-12}{2t+3}=-6t-5,\\ w_{12}^{(0)}&=2w_1^{(0)}-w_8^{(0)}=-6t-5=w_{11}^{(0)}. \end{align*} (Here, $w_7^{(0)}$ is calculated by the partial derivative of the potential function with respect to $w_4$.)
The potential function of Figure \ref{mirror4_1}(g) becomes $-W(w_1,w_7,w_9,w_{2},w_{10},w_{11})$ and the following numerical calculation \begin{align*} -W_0(w_1^{(0)},w_7^{(0)},w_9^{(0)},w_{2}^{(0)},w_{10}^{(0)},w_{11}^{(0)})&=
\left\{\begin{array}{ll}i(2.0299...+0\,i)&\text{ if }t=\frac{-1-\sqrt{3} \,i}{2} \\
i(-2.0299...+0\,i)&\text{ if }t=\frac{-1+\sqrt{3}\,i}{2} \end{array}\right.\\ &=W_0(w_1^{(0)},w_2^{(0)},w_3^{(0)},w_4^{(0)},w_5^{(0)},w_6^{(0)}) \end{align*} confirms Theorem \ref{thm1}.
\appendix \section{Shadow-coloring induced by a solution}\label{app}
In \cite{Cho14c} and this article, we always start from a given boundary-parabolic representation $\rho:\pi_1(L)\rightarrow{\rm PSL}(2,\mathbb{C})$ and construct a solution $(w_1^{(0)},\ldots,w_n^{(0)})$ of the hyperbolicity equations $\mathcal{I}$ using (\ref{main}). In other words, for any representation $\rho$, we can always construct a solution that induces $\rho$. Therefore natural question arises that whether any {essential} solution of $\mathcal{I}$ can be constructed by the formula (\ref{main}) of certain shadow-coloring. This question is important because, if it is true, then any essential solution of $\mathcal{I}$ is governed by the Reidemeister transformations. Furthermore, we can characterize the solutions of $\mathcal{I}$ by the choices of certain shadow-coloring.
\begin{theorem} For any {essential} solution
$\bold{w}^{(0)}=(w_1^{(0)},\ldots,w_n^{(0)})$ of $\mathcal{I}$, there exist an arc-coloring $\{a_1,\ldots,a_r\}$, a region coloring $\{s_1,\ldots,s_n\}$ and an element $p\in\mathbb{C}^2\backslash\{0\}$ satisfying \begin{equation*} w_k^{(0)}=\det(p,s_k), \end{equation*} for $k=1,\ldots,n$. \end{theorem}
\begin{proof} From the discussion above of Proposition \ref{pro21}, the solution $\bold{w}^{(0)}$ induces the boundary-parabolic representation $\rho$. After fixing an oriented diagram $D$ of the link $L$, the representation $\rho$ induces unique arc-coloring $\{a_1,\ldots,a_r\}$ up to conjugation. By using proper conjugation, we may assume $$\infty\notin\{h(a_1),\ldots,h(a_r)\}.$$ Then we define $p=\left(\begin{array}{cc}1 &0\end{array}\right)$.
The main idea of this proof is to show the following region-coloring $$s_k=w_k^{(0)}\left(\begin{array}{cc}h(s_k) &1\end{array}\right)$$ for $k=1,\ldots,n$, is what we want. To prove rigorously, assume the regions with $s_1$ and $s_2$ are adjacent, as in Figure \ref{fig22}.
\begin{figure}
\caption{Defining the region-color $s_1$}
\label{fig22}
\end{figure}
Define $s_1$ by $$s_1=w_1^{(0)}\left(\begin{array}{cc}\left(\alpha_l \beta_l-1+\frac{w_2^{(0)}}{w_1^{(0)}}\right)/\beta_l^2 & 1 \end{array}\right).$$ Then $h(s_1)=\left(\alpha_l \beta_l-1+\frac{w_2^{(0)}}{w_1^{(0)}}\right)/\beta_l^2$. We can decide $h(s_2),\ldots,h(s_n)$ from $h(s_1)$ using the arc-coloring $\{a_1,\ldots,a_r\}$ together with (\ref{app1}), and we denote the region-colorings by \begin{equation}\label{app2} s_k=x_k\left(\begin{array}{cc}h(s_k) &1\end{array}\right) \end{equation} for $k=1,\ldots,n$. (Therefore, $x_1=w_1^{(0)}$ and $x_2,\ldots,x_n$ are uniquely determined.) Then, from the second row of the following matrices $$s_2=x_2\left(\begin{array}{cc}h(s_2) &1 \end{array}\right)=s_1*\alpha_l =x_1\left(\begin{array}{cc}h(s_1) &1 \end{array}\right) \left(\begin{array}{cc}1+\alpha_l\beta_l &\beta_l^2 \\-\alpha_l^2& 1-\alpha_l\beta_l\end{array}\right),$$ we obtain $$x_2=x_1(\beta_l^2 h(s_1)+1-\alpha_l\beta_l)=x_1\frac{w_2^{(0)}}{w_1^{(0)}}=w_2^{(0)}.$$
Now let the developing map induced by the solution $\bold{w}^{(0)}$ be $D_1$, and the one induced by the shadow-coloring $\{a_1,\ldots,a_r,s_1,\ldots,s_n,p\}$ together with (\ref{app2}) be $D_2$. (The definition of the developing map we are using here is {\it Definition 4.10} from Section 4 of \cite{Zickert09}.) Note that $D_1$ can be constructed by gluing tetrahedra with the shape parameters determined by $\bold{w}^{(0)}$, and $D_2$ is constructed explicitly at Figure \ref{fig06}. (The developing map $D_2$ satisfies the condition (4.2) in {\it Proof} of THEOREM 4.11 of \cite{Zickert09}.) To make $D_1=D_2$, consider the five-term triangulation defined in Section \ref{sec2} and see Figure \ref{fig23}. \begin{figure}
\caption{Figure \ref{fig22} together with the crossing $j$}
\label{fig23}
\end{figure}
We also consider the octahedron ${\rm A}_j{\rm B}_j{\rm C}_j{\rm D}_j{\rm E}_j{\rm F}_j$ at the crossing $j$ as in Figure \ref{fig7}. (Note that Figure \ref{fig23} is the left-hand side of Figure \ref{fig7}.) The property $x_1=w_1^{(0)}$ and $x_2=w_2^{(0)}$ imply that, in case of Figures \ref{fig7}(a) and \ref{fig23}(a), the shape parameter of $\Delta{\rm D}_j{\rm B}_j{\rm F}_j{\rm A}_j$ assigned to ${\rm F}_j{\rm A}_j$ is $\frac{w_1^{(0)}}{w_2^{(0)}}$ and the shape parameter of $\Delta{\rm D}_j{\rm E}_j{\rm A}_j{\rm C}_j$ assigned to ${\rm D}_j{\rm E}_j$ is $\frac{w_2^{(0)}}{w_1^{(0)}}$. These shape parameters coincide with the shape parameters determined by the solution $\bold{w}^{(0)}$. Hence, for the lifts $\widetilde{{\rm A}}_j$, $\widetilde{{\rm B}}_j$, ..., $\widetilde{{\rm F}}_j$ of the vertices, we can put \begin{equation}\label{app3} D_1(\widetilde{{\rm D}}_j)=D_2(\widetilde{{\rm D}}_j),~D_1(\widetilde{{\rm B}}_j)=D_2(\widetilde{{\rm B}}_j),~ D_1(\widetilde{{\rm F}}_j)=D_2(\widetilde{{\rm F}}_j), D_1(\widetilde{{\rm A}}_j)=D_2(\widetilde{{\rm A}}_j) \end{equation} in the case of Figures \ref{fig7}(a) and \ref{fig23}(a), and put \begin{equation}\label{app4} D_1(\widetilde{{\rm D}}_j)=D_2(\widetilde{{\rm D}}_j),~ D_1(\widetilde{{\rm E}}_j)=D_2(\widetilde{{\rm E}}_j),~ D_1(\widetilde{{\rm A}}_j)=D_2(\widetilde{{\rm A}}_j),~ D_1(\widetilde{{\rm C}}_j)=D_2(\widetilde{{\rm C}}_j) \end{equation} in the case of Figures \ref{fig7}(b) and \ref{fig23}(b).
Note that the developing maps $D_1$ and $D_2$ are defined from the same representation $\rho$. Therefore, from the uniqueness theorem of the developing map in THEOREM 4.11 of \cite{Zickert09}, $D_1$ and $D_2$ agree on the ideal points corresponding to the {\it nontrivial} ends. (See {\it Definition 4.3} of \cite{Zickert09} for the definitions of the nontrivial and trivial ends.) Furthermore, the five-term triangulation we are using has two trivial ends and we denoted the corresponding points by $\pm\infty$ in Section \ref{sec2}. At the octahedron in Figure \ref{fig7}, the ideal points $\widetilde{{\rm A}}_j$ and $\widetilde{{\rm C}}_j$ corresponds to $\infty$ and $\widetilde{{\rm B}}_j$ and $\widetilde{{\rm D}}_j$ corresponds to $-\infty$. From (\ref{app3}) and (\ref{app4}), the two developing maps $D_1$ and $D_2$ coincide not only at the points corresponding to the nontrivial ends, but also at the points corresponding to the trivial ends. Therefore, we obtain $D_1=D_2$.
The coincidence of the two developing maps implies the shape parameters of the tetrahedra in the five-term triangulation coincide everywhere. Therefore, from (\ref{main}), we obtain $$\frac{x_k}{x_m}=\frac{w_k^{(0)}}{w_m^{(0)}}$$ for any adjacent regions with $w_k$ and $w_m$. (Note that $\left(\frac{x_k}{x_m}\right)^{\pm}$ is the shape parameter determined by the formula (\ref{main}) and the developing map $D_2$.) We already know $x_1=w_1^{(0)}$ and $x_2=w_2^{(0)}$, so we obtain $x_k=w_k^{(0)}$ for $k=1,\ldots,n$, and the shadow-coloring $\{a_1,\ldots,a_r,s_1,\ldots,s_n,p\}$ is what we want.
\end{proof}
\end{ack}
{
\begin{flushleft}
Busan National University of Education\\ Busan 47503, Republic of Korea\\ E-mail: [email protected]\\
Department of Mathematics, Faculty of Science and Engineering, Waseda University\\ 3-4-1 Ohkumo, Shinjuku-ku, Tokyo 169-855, Japan\\ E-mail: [email protected]\\ \end{flushleft}}
\end{document}
|
arXiv
|
{
"id": "1509.02378.tex",
"language_detection_score": 0.5756224989891052,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\maketitle
\begin{abstract} We will show a rigidity of a K\"ahler potential of the Poincar\'e metric with a constant length differential. \end{abstract}
\section{Introduction}
From the fundamental result of Donnelly-Fefferman~\cite{DF}, the vanishing of the space of $L^2$ harmonic $(p,q)$ forms has been an important research theme in the theory of complex domains. Since M.~Gromov (\cite{Gromov}, see also \cite{Donnelly1994}) suggested the concept of the K\"ahler hyperbolicity and gave a connection to the vanishing theorem, there have been many studies on the K\"ahler hyperbolicity of the Bergman metric, which is a fundamental K\"ahler structure of bounded pseudoconvex domains. The K\"ahler structure $\omega$ is \emph{K\"ahler hyperbolic} if there is a global $1$-form $\eta$ with $d\eta=\omega$ and $\sup\norm{\eta}_\omega<\infty$.
In \cite{Donnelly1997}, H. Donnelly showed the K\"ahler hyperbolicity of Bergman metric on some class of weakly pseudoconvex domains. For bounded homogeneous domain $D$ in $\mathbb{C}^n$ and its Bergman metric $\omega_D$ especially, he used a classical result of Gindikin~\cite{Gindikin} to show that $\sup\norm{d\log K_D}_{\omega_D}<\infty$. Here $K_D$ is the Bergman kernel function of $D$ so $\log K_D$ is a canonical potential of $\omega_D$.
In their paper \cite{Kai-Ohsawa}, S.~Kai and T.~Ohsawa gave another approach. They proved that every bounded homogeneous domain has a K\"ahler potential of the Bergman metric whose differential has a constant length. \begin{theorem}[Kai-Ohsawa \cite{Kai-Ohsawa}]\label{thm:KO1} For a bounded homogeneous domain $D$ in $\mathbb{C}^n$, there exists a positive real valued function $\varphi$ on $D$ such that $\log\varphi$ is a K\"ahler potential of the Bergman metric $\omega_D$ and $\norm{d\log\varphi}_{\omega_D}$ is constant. \end{theorem}
It can be obtained by the facts that each homogeneous domain is biholomorphic to a Siegel domain (see \cite{VGP}) and a homogeneous Siegel domain is affine homogeneous (see \cite{KMO}).
More precisely, let us consider a bounded homogeneous domain $D$ in $\mathbb{C}^n$ and a biholomorphism $F:D\to S$ for a Siegel domain $S$. For the Bergman kernel function $K_S$ of $S$ which is a canonical potential of the Bergman metric $\omega_S$, it is easy to show that $d\log K_S$ has a constant length with respect to $\omega_S$ from the affine homogeneity of $S$ (the group of affine holomorphic automorphisms acts transitively on $S$). Since $\log K_S$ is a K\"ahler potential of $\omega_S$, the transformation formula of the Bergman kernel implies that the pullback $F^*\log K_S=\log K_S\circ F$ is also a K\"ahler potential of $\omega_D$. Using the fact that $F:(D,\omega_D)\to(S,\omega_S)$ is an isometry, we have $\norm{d(F^*\log K_S)}_{\omega_D}=\norm{d\log K_S}_{\omega_S}\circ F$. As a function $\varphi$ in Theorem~\ref{thm:KO1}, we can choose the pullback $K_S\circ F$ of the Bergman kernel function of the Siegel domain.
At this junction, it is natural to ask: \begin{quote} \textit{If there is a K\"ahler potential $\log\varphi$ with a constant $\norm{d\log\varphi}_{\omega_D}$, is it always obtained by the pullback of the Bergman kernel function of the Siegel domain?} \end{quote} The aim of this paper is to discuss of this question in the $1$-dimensional case.
The only bounded homogeneous domain in $\mathbb{C}$ is the unit disc $\Delta=\{z\in\mathbb{C}:\abs{z}<1\}$ up to the biholomorphic equivalence and the $1$-dimensional correspondence of the Bergman metric, namely a holomorphically invariant hermitian structure, is only the Poincar\'e metric. Hence the main theorem as follows gives a positive answer to the question. \begin{theorem}\label{thm:main thm rough} Let $\omega_\Delta$ be the Poincar\'e metric of the unit disc $\Delta$. Suppose that there exists a positive real valued function $\varphi:\Delta\to\mathbb{R}$ such that $\log\varphi$ is a K\"ahler potential of the Poincar\'e metric and $\norm{d\log\varphi}_{\omega_\Delta}$ is constant on $\Delta$. Then $\varphi$ is the pullback of the canonical potential on the half-plane $\mathbf{H}=\{z\in\mathbb{C}: \mathrm{Re}\, z<0\}$. \end{theorem}
Note that $1$-dimensional Siegel domain is just the half-plane. We will introduce the Poincar\'e metric and related notions in Section~\ref{sec:2}. As an application of the main theorem, we can characterize the half-plane by the canonical potential.
\begin{corollary}\label{cor:main cor rough} Let $D$ be a simply connected, proper domain in $\mathbb{C}$ with a Poincar\'e metric $\omega_D=i\lambda dz \wedge d\bar z$. If $\norm{d\log \lambda}_{\omega_D}$ is constant on $D$, then $D$ is affine equivalent to the half-plane $\mathbf{H}=\{z\in\mathbb{C}: \mathrm{Re}\, z<0\}$. \end{corollary}
In Section~\ref{sec:2}, we will introduce notions and concrete version of the main theorem. Then we will study the existence of a nowhere vanishing complete holomorphic vector field which is tangent to a potential whose differential is of constant length (Section~\ref{sec:3}). Using relations between complete holomorphic vector fields and model potentials in Section~\ref{sec:4}, we will prove theorems.
\section{Background materials}\label{sec:2}
Let $X$ be a Riemann surface. The Poincar\'e metric of $X$ is a complete hermitian metric with a constant Gaussian curvature, $-4$. The Poincar\'e metric exists on $X$ if and only if $X$ is a quotient of the unit disc. If $X$ is covered by $\Delta$, the Poincar\'e metric can be induced by the covering map $\pi:\Delta\to X$ and it is uniquely determined. Throughout of this paper, the K\"ahler form of the Poincar\'e metric of $X$, denoted by $\omega_X$, stands for the metric also. When $\omega_X=i\lambda dz\wedge d\bar z$ in the local holomorphic coordinate function $z$, the curvature can be written by \begin{equation*} \kappa=-\frac{2}{\lambda}\spd{}{z}{\bar z} \log \lambda \;. \end{equation*} So the curvature condition $\kappa\equiv -4$ implies that \begin{equation*} \spd{}{z}{\bar z} \log \lambda = 2\lambda \;, \end{equation*} equivalently \begin{equation*} dd^c\log\lambda = 2\omega_X \;, \end{equation*} where $d^c=\frac{i}{2}(\overline\partial-\partial)$. That means the function $\frac{1}{2}\log\lambda$ is a local K\"ahler potential of $\omega_X$. Any other local potential of $\omega_X$ is always of the form $\frac{1}{2}\log\lambda+\log\abs{f}^2$ where $f$ is a local holomorphic function on the domain of $z$. We call $\frac{1}{2}\log\lambda$ the \emph{canonical potential} with respect to the coordinate function $z$. For a domain $D$ in $\mathbb{C}$, the canonical potential of $D$ means the canonical potential with respect to the standard coordinate function of $\mathbb{C}$.
Let us consider the Poincar\'e metric $\omega_\Delta$ of the unit disc $\Delta$: \begin{equation*} \omega_\Delta=i\frac{1}{\paren{1-\abs{z}^2}^2}dz\wedge d\bar z = i\lambda_\Delta dz\wedge d\bar z \;. \end{equation*} The canonical potential $\lambda_\Delta$ satisfies \begin{equation*} \norm{d\log\lambda_\Delta}_{\omega_\Delta}^2=\norm{\pd{\log\lambda_\Delta}{z}dz+\pd{\log\lambda_\Delta}{\bar z}d\bar z}_{\omega_\Delta}^2 =\pd{\log\lambda_\Delta}{z}\pd{\log\lambda_\Delta}{\bar z}\frac{1}{\lambda_\Delta}=4\abs{z}^2 \;, \end{equation*} so does not have a constant length. By the same way of Kai-Ohsawa~\cite{Kai-Ohsawa}, we can get a model for $\varphi$ in Theorem~\ref{thm:KO1} for the unit disc, \begin{equation}\label{eqn:model} \varphi_\theta(z)=\frac{\abs{1+e^{i\theta}z}^4}{\paren{1-\abs{z}^2}^2} \quad\text{for $\theta\in\mathbb{R}$} \end{equation} as a pullback of the canonical potential $\lambda_\mathbf{H}=1/\abs{\mathrm{Re}\, w}^2$ on the left-half plane $\mathbf{H}=\{w:\mathrm{Re}\, w<0\}$ by the Cayley transforms (see \eqref{eqn:CT} for instance). The term $\theta$ depends on the choice of the Cayley transform. Since $\log\varphi_\theta=\log\lambda_\Delta+\log\abs{1+e^{i\theta}z}^4$, the function $\frac{1}{2}\log\varphi_\theta$ is a K\"ahler potential. Moreover \begin{equation*} \norm{d\log\varphi_\theta}_{\omega_\Delta}^2 \equiv 4 \;. \end{equation*} At this moment, we introduce a significant result of Kai-Ohsawa. \begin{theorem}[Kai-Ohsawa \cite{Kai-Ohsawa}]\label{thm:KO2} For a bounded homogeneous domain $D$ in $\mathbb{C}^n$, suppose that there is a K\"ahler potential $\log\psi$ of the Bergman metric $\omega_D$ with a constant $\norm{d\log\psi}_{\omega_D}$, then $\norm{d\log\psi}_{\omega_D}=\norm{d\log\varphi}_{\omega_D}$ where $\varphi$ is as in Theorem~\ref{thm:KO1}. \end{theorem} Suppose that a positively real valued $\varphi$ on $\Delta$ satisfies that $dd^c\log\varphi=2\omega_\Delta$ and $\norm{d\log\varphi}_{\omega_\Delta}^2\equiv c$ for some constant $c$. Theorem~\ref{thm:KO2} implies that $c$ must be $4$. Therefore, we can rewrite Theorem~\ref{thm:main thm rough} by \begin{theorem}\label{thm:main thm} If there exists a function $\varphi:\Delta\to\mathbb{R}$ satisfying \begin{equation}\label{eqn:basic condition} dd^c\log\varphi=2\omega_\Delta \quad\text{and}\quad \norm{d\log\varphi}_{\omega_\Delta}^2\equiv4 \;. \end{equation} Then $\varphi=r\varphi_\theta$ as in \eqref{eqn:model} for some $r>0$ and $\theta\in\mathbb{R}$. \end{theorem}
Corollary~\ref{cor:main cor rough} can be also written by
\begin{corollary}\label{cor:main cor} Let $D$ be a simply connected, proper domain in $\mathbb{C}$ with a Poincar\'e metric $\omega_D=i\lambda dz \wedge d\bar z$. If $\norm{d\log \lambda}_{\omega_D}^2\equiv 4$, then $D$ is affine equivalent to the half-plane $\mathbf{H}=\{z\in\mathbb{C}: \mathrm{Re}\, z<0\}$. \end{corollary}
\section{Existence of nowhere vanishing complete holomorphic vector field}\label{sec:3}
In this section, we will study an existence of a complete holomorphic tangent vector field on a Riemann surface $X$ which admits a K\"ahler potential of the Poincar\'e metric with a constant length differential.
By a holomorphic tangent vector field of a Riemann surface $X$, we means a holomorphic section $\mathcal{W}$ to the holomorphic tangent bundle $T^{1,0}X$. If the corresponding real tangent vector field $\mathrm{Re}\, \mathcal{W}=\mathcal{W}+\overline{\mathcal{W}}$ is complete, we also say $\mathcal{W}$ is complete. Thus the complete holomorphic tangent vector field generates a $1$-parameter family of holomorphic transformations.
In this section, we will show that \begin{theorem}\label{thm:existence} Let $X$ be a Riemann surface with the Poincar\'e metric $\omega_X$. If there is a function $\varphi:X\to\mathbb{R}$ with \begin{equation}\label{eqn:condition on surface} dd^c\log\varphi = 2\omega_X \quad\text{and}\quad \norm{d\log\varphi}_{\omega_X}^2\equiv 4 \end{equation} then there is a nowhere vanishing complete holomorphic vector field $\mathcal{W}$ such that $(\mathrm{Re}\, \mathcal{W})\varphi\equiv 0$. \end{theorem}
\begin{proof} Take a local holomorphic coordinate function $z$ and let $\omega_X=i\lambda dz\wedge d\bar z$. The equation~\eqref{eqn:condition on surface} can be written by \begin{equation*} \paren{\log\varphi}_{z\bar z} = 2\lambda \quad\text{and}\quad \paren{\log\varphi}_z \paren{\log\varphi}_{\bar z} = 4\lambda \end{equation*} Here, $\paren{\log\varphi}_z=\pd{}{z}\log\varphi$, $\paren{\log\varphi}_{\bar z}=\pd{}{\bar z}\log\varphi$ and $\paren{\log\varphi}_{z\bar z}=\spd{}{z}{\bar z}\log\varphi$. This implies that \begin{align*} \paren{\varphi^{-1/2}}_{z}
&=
\pd{}{z}\varphi^{-1/2}
=
-\frac{1}{2}\varphi^{-1/2} \paren{\log\varphi}_{z} \; ; \\ \paren{\varphi^{-1/2}}_{z\bar z}
&=
\spd{}{z}{\bar z}\varphi^{-1/2}
=
-\frac{1}{2}\varphi^{-1/2} \paren{\log\varphi}_{z\bar z}
+\frac{1}{4}\varphi^{-1/2}\paren{\log\varphi}_z \paren{\log\varphi}_{\bar z}
\\
&=
-\frac{1}{2}\varphi^{-1/2} \paren{
\paren{\log\varphi}_{z\bar z}
-\frac{1}{2}\paren{\log\varphi}_z \paren{\log\varphi}_{\bar z}
}
\\
&= 0 \; . \end{align*} Thus we have that the function $\varphi^{-1/2}$ is harmonic so $\paren{\varphi^{-1/2}}_{z}$ is holomorphic.
Let us consider a local holomorphic vector field, \begin{equation*} \mathcal{W}=\frac{i}{\paren{\varphi^{-1/2}}_z}\pd{}{z} =\frac{-2i\varphi^{3/2}}{\varphi_z}\pd{}{z} =\frac{-2i\varphi^{1/2}}{\paren{\log\varphi}_z}\pd{}{z} \;. \end{equation*} In any other local holomorphic coordinate function $w$, we have \begin{equation*} \mathcal{W}=\frac{i}{\paren{\varphi^{-1/2}}_z}\pd{}{z} =\frac{i}{\paren{\varphi^{-1/2}}_w\pd{w}{z}}\pd{w}{z}\pd{}{w} =\frac{i}{\paren{\varphi^{-1/2}}_w}\pd{}{w} \;. \end{equation*} so $W$ is globally defined on $X$. Now we will show that $\mathcal{W}$ satisfies conditions in the theorem.
Since \begin{equation*} \norm{\varphi^{-1/2}\mathcal{W}}_{\omega_X}^2 =\norm{\frac{-2i}{\paren{\log\varphi}_{z}}\pd{}{z} }_{\omega_X}^2 =\frac{4\lambda}{\paren{\log\varphi}_{z}\paren{\log\varphi}_{\bar z}} =1 \;, \end{equation*} the vector field $\varphi^{-1/2}\mathcal{W}$ has a unit length with respect to the complete metric $\omega_X$, so the corresponding real vector field $\mathrm{Re}\, \varphi^{-1/2}\mathcal{W}=\varphi^{-1/2}(\mathcal{W}+\overline{\mathcal{W}})$ is complete. Moreover \begin{equation*} (\mathrm{Re}\, \mathcal{W})\varphi = \frac{-2i\varphi^{3/2}}{\varphi_{z}}\varphi_z +\frac{2i\varphi^{3/2}}{\varphi_{\bar z}}\varphi_{\bar z} = 0 \;. \end{equation*} Hence it remains to show the completeness of $\mathcal{W}$. Take any integral curve $\gamma:\mathbb{R}\to X$ of $\varphi^{-1/2}\mathrm{Re}\,\mathcal{W}$. It satisfies \begin{equation*} \paren{\varphi^{-1/2}(\mathrm{Re}\, \mathcal{W})}\circ \gamma = \dot\gamma \end{equation*} equivalently \begin{equation*} (\mathrm{Re}\, \mathcal{W})\circ \gamma = \paren{\varphi^{1/2}\circ\gamma} \dot\gamma \end{equation*} The condition $(\mathrm{Re}\, \mathcal{W})\varphi\equiv 0$, equivalently $\varphi^{-1/2}(\mathrm{Re}\, \mathcal{W})\varphi\equiv 0$, implies that the curve $\gamma$ is on a level set of $\varphi$ so $\varphi^{1/2}\circ\gamma\equiv C$ for some constant $C$. The curve $\sigma:\mathbb{R}\to X$ defined by $\sigma(t)=\gamma(Ct)$ satisfies \begin{equation*} (\mathrm{Re}\, \mathcal{W})\circ \sigma (t) = (\mathrm{Re}\, \mathcal{W})(\gamma(Ct)) = C\dot\gamma(Ct) =\dot\sigma(t) \end{equation*} This means that $\sigma:\mathbb{R}\to X$ is the integral curve of $\mathrm{Re}\, \mathcal{W}$; therefore $\mathrm{Re}\, \mathcal{W}$ is complete. This completes the proof. \end{proof}
\section{Complete holomorphic vector fields on the unit disc}\label{sec:4} In this section, we introduce parabolic and hyperbolic vector fields on the unit disc and discuss their relation to the model potential, \begin{equation}\label{eqn:model0} \varphi_0=\frac{\abs{1+z}^4}{\paren{1-\abs{z}^2}^2} \end{equation} where it is $\varphi_\theta$ in \eqref{eqn:model} with $\theta=0$. \subsection{Nowhere vanishing complete holomorphic vector fields from the left-half plane} On the left-half plane $\mathbf{H}=\{w\in\mathbb{C}:\mathrm{Re}\, w<0\}$, there are two kinds of affine transformations: \begin{equation*} \mathcal{D}_s(w)=e^{2s} w \quad\text{and}\quad \mathcal{T}_s(w)=w+2is \end{equation*} for $s\in\mathbb{R}$. Their infinitesimal generators are \begin{equation*} \mathcal{D}= 2w\pd{}{w} \quad\text{and}\quad \mathcal{T}=2i\pd{}{w} \end{equation*} which are nowhere vanishing complete holomorphic vector fields of $\mathbf{H}$. Note that \begin{equation}\label{eqn:relation} (\mathcal{T}_s)_*\mathcal{D} = 2(w-2is)\pd{}{w} = \mathcal{D}-2s\mathcal{T} \quad\text{and}\quad (\mathcal{T}_s)_*\mathcal{T} =2i\pd{}{w}= \mathcal{T} \end{equation} for any $s$.
For the Cayley transform $F:\mathbf{H}\to\Delta$ defined by \begin{equation}\label{eqn:CT} \begin{aligned} F:\mathbf{H}&\longrightarrow\Delta \\ w&\longmapsto z=\frac{1+w}{1-w} \;, \end{aligned} \end{equation} we can take two nowhere vanishing complete holomorphic vector fields of $\Delta$: \begin{equation*} \mathcal{H}=F_*(\mathcal{D})=(z^2-1)\pd{}{z} \end{equation*} and \begin{equation*} \mathcal{P}=F_*(\mathcal{T})=i(z+1)^2\pd{}{z}\;. \end{equation*} When we define $\mathcal{H}_s=F\circ\mathcal{D}_s\circ F^{-1}$ and $\mathcal{P}_s=F\circ\mathcal{T}_s\circ F^{-1}$, vector fields $\mathcal{H}$ and $\mathcal{P}$ are infinitesimal generators of $\mathcal{H}_s$ and $\mathcal{P}_s$, respectively. Moreover Equation~\eqref{eqn:relation} can be written by \begin{equation}\label{eqn:relation'} (\mathcal{P}_s)_*\mathcal{H} = \mathcal{H}-2s\mathcal{P} \quad\text{and}\quad (\mathcal{P}_s)_*\mathcal{P}= \mathcal{P} \;. \end{equation} There is another complete holomorphic vector field $\mathcal{R}=iz\pdl{}{z}$ generating the rotational symmetry \begin{equation}\label{eqn:rotation} \mathcal{R}_s(z)=e^{is}z \;. \end{equation} The holomorphic automorphism group of $\Delta$ is a real $3$-dimension connected Lie group (cf. see \cite{Cartan,Narasimhan}), we can conclude that any complete holomorphic vector field can be a real linear combination of $\mathcal{H}$, $\mathcal{P}$ and $\mathcal{R}$. Since $\mathcal{H}(-1)=\mathcal{P}(-1)=0$ and $\mathcal{R}(-1)=-i\pdl{}{z}$, we have \begin{lemma}\label{lem:lc} If $\mathcal{W}$ is a complete holomorphic vector field of $\Delta$ satisfying $\mathcal{W}(-1)=0$, then there exist $a,b\in\mathbb{R}$ with $\mathcal{W}=a\mathcal{H}+b\mathcal{P}$. \end{lemma}
\subsection{Hyperbolic vector fields} In this subsection, we will show that the hyperbolic vector field $\mathcal{H}$ can not be tangent to a K\"ahler potential with a constant length differential.
By the simple computation, \begin{equation*} \mathcal{H}(\log\varphi_0)
= (z^2-1)\frac{2(1+\bar z)}{(1+z)(1-\abs{z}^2)}
=2\frac{\abs{z}^2+z-\bar z-1}{(1-\abs{z}^2)}
\;, \end{equation*} we get \begin{equation*} (\mathrm{Re}\,\mathcal{H})\log\varphi_0 \equiv-4 \;. \end{equation*}
That means $\mathrm{Re}\,\mathcal{H}$ is nowhere tangent to $\varphi_0$. Moreover \begin{lemma}\label{lem:hyperbolic} Let $\varphi:\Delta\to\mathbb{R}$ with $dd^c \log\varphi=2\omega_\Delta$ and $\norm{d\log\varphi}_{\omega_\Delta}^2\equiv 4$. If $(\mathrm{Re}\,\mathcal{H}) \log\varphi\equiv c$ for some $c$, then $c=\pm4$. \end{lemma} \begin{proof} Since $dd^c\log\varphi_0=2\omega_\Delta$ also, the function $\log\varphi-\log\varphi_0$ is harmonic; hence we may let $\log\varphi=\log\varphi_0+f+\bar f$ for some holomorphic function $f:\Delta\to\mathbb{C}$. Then the condition $(\mathrm{Re}\,\mathcal{H}) \log\varphi\equiv c$ can be written by \begin{equation}\label{eqn:basic identity} (\mathrm{Re}\,\mathcal{H})\log\varphi = -4+(z^2-1)f'+(\bar{z}^2-1)\bar f' \equiv c \;. \end{equation} This implies that $(z^2-1)f'$ is constant. Thus we can let \begin{equation}\label{eqn:hmm} f'=\frac{C}{z^2-1} \end{equation} for some $C\in\mathbb{C}$. Since \begin{equation*} \pd{}{z}\log\varphi=f'+\pd{}{z}\log\varphi_0 = f'+ \frac{2(1+\bar z)}{(1+z)(1-\abs{z}^2)}\;, \end{equation*} we have \begin{multline*} \norm{d\log\varphi}_{\omega_\Delta}^2=\paren{\pd{}{z}\log\varphi}\paren{\pd{}{\bar z}\log\varphi}\frac{1}{\lambda_\Delta} \\
=\abs{f'}^2(1-|z|^2)^2 +\frac{2(1+\bar z)(1-\abs{z}^2)}{(1+z)}\bar f' +\frac{2(1+z)(1-\abs{z}^2)}{(1+\bar z)}f' +\norm{d\log\varphi_0}_{\omega_\Delta}^2 \;. \end{multline*} From the condition $\norm{d\log\varphi}_{\omega_\Delta}^2\equiv 4\equiv\norm{d\log\varphi_0}_{\omega_\Delta}^2$, it follows \begin{equation*}
\abs{f'}^2(1-|z|^2)^2 = -\frac{2(1+\bar z)(1-\abs{z}^2)}{(1+z)}\bar f' -\frac{2(1+z)(1-\abs{z}^2)}{(1+\bar z)}f' \;, \end{equation*} equivalently \begin{equation}\label{eqn:identity}
\frac{1}{2}\abs{f'}^2(1-|z|^2) = -\frac{(1+\bar z)}{(1+z)}\bar f' -\frac{(1+z)}{(1+\bar z)}f' \;. \end{equation} Applying \eqref{eqn:hmm} to the right side above, \begin{multline*} -\frac{(1+\bar z)}{(1+z)}\bar f' -\frac{(1+z)}{(1+\bar z)}f' =\frac{(1+\bar z)}{(1+z)}\frac{\bar C}{1-\bar z^2} +\frac{(1+z)}{(1+\bar z)}\frac{C}{1-z^2}
\\ =\frac{(1+\bar z-z-\abs{z}^2)\bar C + (1-\bar z+z-\abs{z}^2)C}{\abs{1-z^2}^2} \;. \end{multline*}
Let $C=a+bi$ for $a,b\in\mathbb{R}$, then \begin{equation*} (1+\bar z-z-\abs{z}^2)\bar C + (1-\bar z+z-\abs{z}^2)C
= 2a(1-\abs{z}^2) +2bi(z-\bar z) \;. \end{equation*} Now Equation~\eqref{eqn:identity} can be written by \begin{equation*}
\frac{1}{2}\frac{\abs{C}^2}{\abs{z^2-1}^2}(1-|z|^2) =\frac{2a(1-\abs{z}^2) +2bi(z-\bar z)}{\abs{1-z^2}^2} \;, \end{equation*} so we have \begin{equation*} (\abs{C}^2-4a)(1-\abs{z}^2) =4bi(z-\bar z) \end{equation*} on $\Delta$. Take $\partial\bar\partial$ to above, we have \begin{equation*} \abs{C}^2-4a=0 \;. \end{equation*} Simultaneously $b=0$ so $C=a$. Now we have $a^2=4a$. Such $a$ is $0$ or $4$. If $f'=4/(z^2-1)$, then $c=4$ from \eqref{eqn:basic identity}. If $f'=0$, then $c=-4$. \end{proof}
\subsection{Parabolic vector fields} Since \begin{equation*} \mathcal{P}(\log\varphi_0)
= i(z+1)^2\frac{2(1+\bar z)}{(1+z)(1-\abs{z}^2)}
=2i\frac{\abs{1+z}^2}{1-\abs{z}^2}
\;, \end{equation*} we have \begin{equation*} (\mathrm{Re}\,\mathcal{P})\log\varphi_0 \equiv 0 \;. \end{equation*} That means that the parabolic vector field $\mathcal{P}$ is tangent to $\varphi_0$. The vector field $\mathcal{P}$ is indeed the nowhere vanishing complete holomorphic vector field as constructed in Theorem~\ref{thm:existence} corresponding to $\varphi_0$. The main result of this section is the following.
\begin{lemma}\label{lem:parabolic} Let $\varphi:\Delta\to\mathbb{R}$ with $dd^c \log\varphi=2\omega_\Delta$ and $\norm{d\log\varphi}_{\omega_\Delta}^2\equiv 4$. If $(\mathrm{Re}\,\mathcal{P}) \log\varphi\equiv c$ for some $c$, then $c=0$ and $\varphi=r\varphi_0$ for some $r>0$. \end{lemma} \begin{proof} By the same way in the proof of Lemma~\ref{lem:hyperbolic}, we let $\log\varphi=\log\varphi_0+f+\bar f$ for some holomorphic $f:\Delta\to\mathbb{C}$. Since \begin{equation}\label{eqn:basic identity1} (\mathrm{Re}\,\mathcal{P})\log\varphi = i(z+1)^2f'-i(\bar{z}+1)^2\bar f' \equiv c \end{equation} it follows that $(z+1)^2f'$ is constant. Thus we have \begin{equation}\label{eqn:hmm1} f'=\frac{C}{(z+1)^2} \end{equation} for some $C\in\mathbb{C}$. Since \eqref{eqn:identity} also holds, we can apply \eqref{eqn:hmm1} to the right side of \eqref{eqn:identity} to get \begin{multline*} -\frac{(1+\bar z)}{(1+z)}\bar f' -\frac{(1+z)}{(1+\bar z)}f' = -\frac{(1+\bar z)}{(1+z)}\frac{\bar C}{(\bar z+1)^2} -\frac{(1+z)}{(1+\bar z)}\frac{C}{(z+1)^2} \\ =\frac{-\bar C}{\abs{1+z}^2} +\frac{-C}{\abs{1+z}^2} =\frac{-\bar C -C}{\abs{1+z}^2} \end{multline*} Now Equation \eqref{eqn:identity} is can be written by \begin{equation*}
\frac{\abs{C}^2}{\abs{z+1}^4}(1-|z|^2)=2\frac{-\bar C -C}{\abs{1+z}^2} \end{equation*} equivalently \begin{equation*}
\abs{C}^2(1-|z|^2) =-\paren{2\bar C +2C}\abs{1+z}^2\;. \end{equation*} Evaluating $z=0$, we have $\abs{C}^2=-2\bar C-2C$. And taking
$\partial\bar\partial$ to above, we have $-\abs{C}^2=-2\bar C-2C$. It follows that $C=0$ so $f$ is constant. Moreover Equation \eqref{eqn:basic identity1} implies that $c=0$. \end{proof}
\section{Proof of the main theorem}\label{sec:5}
Now we prove Theorem~\ref{thm:main thm} and Corollary~\ref{cor:main cor}
\noindent\textit{Proof of Theorem~\ref{thm:main thm}.} Let $\varphi:\Delta\to\mathbb{R}$ be a function with \begin{equation*} dd^c\log\varphi=2\omega_\Delta \quad\text{and}\quad \norm{d\log\varphi}_{\omega_\Delta}^2\equiv 4 \;. \end{equation*} By Theorem~\ref{thm:existence}, we can take a nowhere vanishing complete holomorphic vector field $\mathcal{W}$ with $(\mathrm{Re}\, \mathcal{W})\varphi\equiv0$. Since every automorphism of $\Delta$ has at least one fixed point on $\overline\Delta$ and $\mathcal{W}$ is nowhere vanishing on $\Delta$, any nontrivial automorphism generated by $\mathrm{Re}\, \mathcal{W}$ has no fixed point in $\Delta$ and should have a common fixed point $p$ at the boundary $\partial\Delta$. This means $p$ is a vanishing point of $\mathcal{W}$. Consider a rotational symmetry $\mathcal{R}_\theta$ in \eqref{eqn:rotation} satisfying $\mathcal{R}_\theta(-1)=p$. We will show that $\varphi\circ\mathcal{R}_\theta=r\varphi_0$ where $\varphi_0$ is as in \eqref{eqn:model0} and $r>0$. This implies that $\varphi=r\varphi_{-\theta}$.
Now we can simply denote by $\varphi=\varphi\circ\mathcal{R}_\theta$ and $\mathcal{W}=(\mathcal{R}_\theta^{-1})_*\mathcal{W}$. Since $-1$ is a vanishing point of $\mathcal{W}$, Lemma~\ref{lem:lc} implies \begin{equation*} \mathcal{W}=a\mathcal{H}+b\mathcal{P} \end{equation*} for some real numbers $a$, $b$.
Suppose that $a\neq0$. Equation~\eqref{eqn:relation'} implies that \begin{equation*} (\mathcal{P}_s)_* \mathcal{W} = (\mathcal{P}_s)_*(a\mathcal{H}+b\mathcal{P}) = a\mathcal{H}-2as\mathcal{P}+b\mathcal{P} = a\mathcal{H}+(b-2as)\mathcal{P}\;. \end{equation*} Take $s=b/2a$, then $\widetilde\mathcal{W}=(\mathcal{P}_s)_* \mathcal{W} =a\mathcal{H}$. Let $\tilde\varphi=\varphi\circ\mathcal{P}_{-s}$ for this $s$. Then $\tilde\varphi$ satisfies conditions in Theorem~\ref{thm:main thm} and $(\mathrm{Re}\, \widetilde\mathcal{W})\tilde\varphi\equiv 0$. But Lemma~\ref{lem:hyperbolic} said that $(\mathrm{Re}\, \widetilde\mathcal{W})\tilde\varphi=a(\mathrm{Re}\, \mathcal{H})\tilde\varphi\equiv \pm4a\tilde\varphi$. It contradicts to $(\mathrm{Re}\, \mathcal{W})\varphi\equiv 0$ equivalently $(\mathrm{Re}\, \widetilde\mathcal{W})\tilde\varphi\equiv 0$. Thus $a=0$.
Now $\mathcal{W}=b\mathcal{P}$. Since $\mathcal{W}$ is nowhere vanishing already, $b\neq 0$. The condition $(\mathrm{Re}\,\mathcal{W})\varphi\equiv0$ implies $(\mathrm{Re}\,\mathcal{P})\varphi\equiv 0$. Lemma~\ref{lem:parabolic} says that $\varphi=r\varphi_0$ for some positive $r$. This completes the proof. \qed
\noindent\textit{Proof of Corollary~\ref{cor:main cor}.} Let $D$ be a simply connected proper domain in $\mathbb{C}$ and let $\omega_D=i\lambda_D dz \wedge d\bar z$ be its Poincar\'e metric with $\norm{d\log \lambda_D}_{\omega_D}^2\equiv 4$. By Theorem~\ref{thm:existence}, there is a nowhere vanishing complete holomorphic vector field $\mathcal{W}$ with $(\mathrm{Re}\,\mathcal{W})\lambda_D\equiv0$. Take a biholomorphism $G:\Delta\to D$ and let \begin{equation*} \varphi=\lambda_D\circ G \quad\text{and}\quad \mathcal{Z}=(G^{-1})_*\mathcal{W} \,. \end{equation*} Note that $(\mathrm{Re}\,\mathcal{Z})\varphi\equiv0$ by assumption. Using the rotational symmetry $\mathcal{R}_\theta$ of $\Delta$ which is also affine, we may assume that $\mathcal{Z}(-1)=0$ and we will prove that $G$ is a Cayley transform.
Since $G:(\Delta,\omega_\Delta)\to(D,\omega_D)$ is an isometry, we have $G^*\omega_D=\omega_\Delta$, equivalently \begin{equation*} \varphi=\frac{\lambda_\Delta}{\abs{G'}^2} \;. \end{equation*} Moreover $d\log\varphi=d(G^*\log\lambda_D)$ implies that $\norm{d\log\varphi}_{\omega_D}^2=\norm{d(G^*\log\lambda_D)}_{\omega_D}^2\equiv4$. By Theorem~\ref{thm:main thm}, we have \begin{equation*} \frac{\lambda_\Delta}{\abs{G'}^2}=\varphi=r\varphi_0=r\lambda_\Delta \abs{1+z}^4 \end{equation*} for some positive $r$. This means that $G'=e^{i\theta'}/\sqrt{r}(1+z)^2$ for some $\theta'\in\mathbb{R}$ so that \begin{equation*} G=\frac{e^{i\theta'}}{2\sqrt{r}}\frac{z-1}{z+1}+C \end{equation*} Since the function $z\mapsto (z-1)/(z+1)$ is the inverse mapping of the Cayley transform $F:\mathbf{H}\to\Delta$ in \eqref{eqn:CT}, we have \begin{align*} G\circ F:\mathbf{H}&\to D \\ z&\mapsto \frac{e^{i\theta'}}{2\sqrt{r}}z+C \;. \end{align*} This implies that $D=G(F(\mathbf{H}))$ is affine equivalent to $\mathbf{H}$. \qed
\end{document}
|
arXiv
|
{
"id": "2008.01991.tex",
"language_detection_score": 0.5289132595062256,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\thanks[SIR]{ GR is partially, and NP is totally, supported by the project MIUR-SIR CMACBioSeq (``Combinatorial methods for analysis and compression of biological sequences'') grant n.~RBSI146R5L.}
\title{Space-Efficient Construction of Compressed Suffix Trees}
\author{ Nicola Prezza\thanksref{SIR}}
\address{ Department of Computer Science, University of Pisa, Italy {\tt [email protected]}}
\author{Giovanna Rosone\thanksref{SIR}}
\address{ Department of Computer Science, University of Pisa, Italy {\tt [email protected]}}
\begin{abstract}
We show how to build several data structures of central importance to string processing, taking as input the Burrows-Wheeler transform (BWT) and using small extra working space. Let $n$ be the text length and $\sigma$ be the alphabet size. We first provide two algorithms that enumerate all LCP values and suffix tree intervals in $O(n\log\sigma)$ time using just $o(n\log\sigma)$ bits of working space on top of the input BWT. Using these algorithms as building blocks, for any parameter $0 < \epsilon \leq 1$ we show how to build the PLCP bitvector and the balanced parentheses representation of the suffix tree topology in $O\left(n(\log\sigma + \epsilon^{-1}\cdot \log\log n)\right)$ time using at most $n\log\sigma \cdot(\epsilon + o(1))$ bits of working space on top of the input BWT and the output.
In particular, this implies that we can build a compressed suffix tree from the BWT using just succinct working space (i.e. $o(n\log\sigma)$ bits) and any time in $\Theta(n\log\sigma) + \omega(n\log\log n)$. This improves the previous most space-efficient algorithms, which worked in $O(n)$ bits and $O(n\log n)$ time. We also consider the problem of merging BWTs of string collections, and provide a solution running in $O(n\log\sigma)$ time and using just $o(n\log\sigma)$ bits of working space. An efficient implementation of our LCP construction and BWT merge algorithms use (in RAM) as few as $n$ bits on top of a packed representation of the input/output and process data as fast as $2.92$ megabases per second. \begin{keyword} Burrows-Wheeler transform, compressed suffix tree, LCP, PLCP. \end{keyword}
\end{abstract}
\end{frontmatter}
\section{Introduction and Related Work} The increasingly-growing production of large string collections---especially in domains such as biology, where new generation sequencing technologies can nowadays generate Gigabytes of data in few hours---is lately generating much interest towards fast and space-efficient algorithms able to index this data. The Burrows-Wheeler Transform~\cite{burrows1994block} and its extension to sets of strings~\cite{MantaciRRS07,BauerCoxRosoneTCS2013} is becoming the gold-standard in the field: even when not compressed, its size is one order of magnitude smaller than classic suffix arrays (while preserving many of their indexing capabilities). This generated considerable interest towards fast and space-efficient BWT construction algorithms~\cite{BauerCoxRosoneTCS2013,Karkkainen:2007:FBS:1314704.1314852,10.1007/978-3-319-15579-1_46,10.1007/978-3-319-02432-5_5,8712716,Kempa:2019:SSS:3313276.3316368,navarro2014optimal,10.1007/978-3-319-15579-1_46}. As a result, the problem of building the BWT is well understood to date. The fastest algorithm solving this problem operates in sublinear $O(n/\sqrt{\log n})$ time and $O(n)$ bits of space on a binary text of length $n$ by exploiting word parallelism~\cite{Kempa:2019:SSS:3313276.3316368}. The authors also provide a conditional lower bound suggesting that this running time might be optimal. The most space-efficient algorithm terminates in $O(n\log n/\log\log n)$ time and uses just $o(n\log\sigma)$ bits of space (succinct) on top of the input and output~\cite{navarro2014optimal}, where $\sigma$ is the alphabet's size. In the average case, this running time can be improved to $O(n)$ on constant-sized alphabets while still operating within succinct space~\cite{10.1007/978-3-319-15579-1_46}.
In some cases, a BWT alone is not sufficient to complete efficiently particular string-processing tasks. For this reason, the functionalities of the BWT are often extended by flanking to it additional structures such as the Longest Common Prefix (LCP) array~\cite{CGRS_JDA_2016} (see e.g.~\cite{prezza2018detecting,prezza2019,GuerriniRosoneAlcob2019,TUSTUMI2016} for bioinformatic applications requiring this additional component).
A disadvantage of the LCP array is that it requires $O(n\log n)$ bits to be stored in plain form. To alleviate this problem, usually the PLCP array~\cite{Sadakane:2002:SRL:545381.545410}---an easier-to-compress permutation of the LCP array---is preferred. The PLCP relies on the idea of storing LCP values in text order instead of suffix array order. As shown by Kasai et al.~\cite{10.1007/3-540-48194-X_17}, this permutation is almost increasing ($PLCP[i+1] \geq PLCP[i]-1$) and can thus be represented in just $2n$ bits in a bitvector known as the \emph{PLCP bitvector}. More advanced applications might even require full suffix tree functionality. In such cases, compressed suffix trees~\cite{Grossi:2005:CSA:1093654.1096192,Sadakane:2007:CST:1326296.1326297} (CSTs) are the preferred choice when the space is at a premium. A typical compressed suffix tree is formed by a compressed suffix array (CSA), the PLCP bitvector, and a succinct representation of the suffix tree topology~\cite{Sadakane:2007:CST:1326296.1326297} (there exist other designs, see Ohlebusch et al.~\cite{10.1007/978-3-642-16321-0_34} for an exhaustive survey). To date, several practical algorithms have been developed to solve the task of building \emph{de novo} such additional components~\cite{CGRS_JDA_2016,holt2014constructing,holt2014merging,Bonizzoni2018,EgidiAMB2019,Belazzougui:2014:LTC:2591796.2591885,10.1007/978-3-642-02441-2_17,Valimaki:2010:ECS:1498698.1594228}, but little work has been devoted to the task of computing them from the BWT in little working space (internal and external). Considering the advanced point reached by state-of-the-art BWT construction algorithms, it is worth to explore whether such structures can be built more efficiently starting from the BWT, rather than from the raw input text.
\textbf{CSA} As far as the CSA is concerned, this component can be easily built from the BWT using small space as it is formed (in its simplest design) by just a BWT with rank/select functionality enhanced with a suffix array sampling, see also~\cite{Belazzougui:2014:LTC:2591796.2591885}.
\textbf{LCP} We are aware of only one work building the LCP array in small space from the BWT: Beller et al.~\cite{beller2013computing} show how to build the LCP array in $O(n\log\sigma)$ time and $O(n)$ bits of working space on top of the input BWT and the output. Other works~\cite{munro2017space,Belazzougui:2014:LTC:2591796.2591885} show how to build the LCP array directly from the text in $O(n)$ time and $O(n\log\sigma)$ bits of space (compact).
\textbf{PLCP} K{\"a}rkk{\"a}inen et al.~\cite{10.1007/978-3-642-02441-2_17} show that the PLCP bitvector can be built in $O(n\log n)$ time using $n$ bits of working space on top of the text, the suffix array, and the output PLCP. Kasai at al.'s lemma also stands at the basis of a more space-efficient algorithm from V\"{a}lim\"{a}ki et al.~\cite{Valimaki:2010:ECS:1498698.1594228}, which computes the PLCP from a CSA in $O(n\log n)$ time using constant working space on top of the CSA and the output. Belazzougui~\cite{Belazzougui:2014:LTC:2591796.2591885} recently presented an algorithm for building the PLCP bitvector from the text in optimal $O(n)$ time and compact space ($O(n\log\sigma)$ bits).
\textbf{Suffix tree topology} The remaining component required to build a compressed suffix tree (in the version described by Sadakane~\cite{Sadakane:2007:CST:1326296.1326297}) is the suffix tree topology, represented either in BPS~\cite{Munro:1997:SRB:795663.796328} (balanced parentheses) or DFUDS~\cite{Benoit2005} (depth first unary degree sequence), using $4n$ bits. As far as the BPS representation is concerned, Hon et al.~\cite{Hon:2009:BTB:1654348.1654351} show how to build it from a CSA in $O(n(\log\sigma + \log^\epsilon n))$ time and compact space for any constant $\epsilon>0$. Belazzougui~\cite{Belazzougui:2014:LTC:2591796.2591885} improves this running time to the optimal $O(n)$, still working within compact space. V\"{a}lim\"{a}ki et al.~\cite{Valimaki:2010:ECS:1498698.1594228} describe a linear-time algorithm that improves the space to $O(n)$ bits on top of the LCP array (which however needs to be represented in plain form), while Ohlebusch et al.~\cite{10.1007/978-3-642-16321-0_34} show how to build the DFUDS representation of the suffix tree topology in $O(t_{lcp}\cdot n)$ time using $n+o(n)$ bits of working space on top of a structure supporting access to LCP array values in $O(t_{lcp})$ time.
Summing up, the situation for building compressed suffix trees from the BWT is the following: algorithms working in optimal linear time require $O(n\log\sigma)$ bits of working space. Algorithms reducing this space to $O(n)$ (on top of a CSA) are only able to build the suffix tree topology within $O(n\cdot t_{lcp})$ time, which is $\Omega(n\log^{\epsilon}n)$ with the current best techniques, and the PLCP bitvector in $O(n\log n)$ time. No algorithm can build all the three CST components within $o(n\log\sigma)$ bits of working space on top of the input BWT and the output. Combining the most space-efficient existing algorithms, the following two trade-offs can therefore be achieved for building all compressed suffix tree components from the BWT: \begin{itemize}
\item $O(n\log\sigma)$ bits of working space and $O(n)$ time, or
\item $O(n)$ bits of working space and $O(n\log n)$ time. \end{itemize}
\paragraph{Our contributions}
In this paper, we give new space-time trade-offs that allow building the CST's components in smaller working space (and in some cases even faster) with respect to the existing solutions. We start by combining Beller et al.'s algorithm~\cite{beller2013computing} with the suffix-tree enumeration procedure of Belazzougui~\cite{Belazzougui:2014:LTC:2591796.2591885} to obtain an algorithm that enumerates (i) all pairs $(i,LCP[i])$, and (ii) all suffix tree intervals in $O(n\log\sigma)$ time using just $o(n\log\sigma)$ bits of working space on top of the input BWT. We use this procedure to obtain algorithms that build (working space is on top of the input BWT and the output):
\begin{enumerate}
\item\label{goalLCP} The LCP array of a string collection in $O(n\log\sigma)$ time and $o(n\log\sigma)$ bits of working space (see Section \ref{sec:LCP}).
\item\label{goalPLCP} the PLCP bitvector and the BPS representation of the suffix tree topology in $O\left(n(\log\sigma + \epsilon^{-1}\cdot \log\log n)\right)$ time and $n\log\sigma \cdot (\epsilon + o(1))$ bits of working space, for any user-defined parameter $0 < \epsilon \leq 1$ (see Section \ref{sec:PLCP} and \ref{sec:ST topology}).
\item\label{goalMerge} The BWT of the union of two string collections of total size $n$ in $O(n\log\sigma)$ time and $o(n\log\sigma)$ bits of working space, given the BWTs of the two collections as input (see Section \ref{sec:algo2}). \end{enumerate}
Contribution (\ref{goalLCP}) is the first showing that the LCP array can be induced from the BWT using succinct working space \emph{for any alphabet size}.
Contribution (\ref{goalPLCP}) can be used to build a compressed suffix tree from the BWT using just $o(n\log\sigma)$ bits of working space and any time in $O(n\log\sigma) + \omega(n\log\log n)$---for example, $O(n(\log\sigma + (\log\log n)^{1+\delta}))$, for any $\delta>0$. On small alphabets, this improves both working space and running time of existing $O(n)$-bits solutions.
Also contribution (\ref{goalMerge}) improves the state-of-the-art, due to Belazzougui et al.~\cite{Belazzougui:2014:LTC:2591796.2591885,belazzougui2016linear}. In those papers, the authors show how to merge the BWTs of two texts $T_1, T_2$ and obtain the BWT of the collection $\{T_1, T_2\}$ in $O(nk)$ time and $n\log\sigma(1+1/k) + 11n + o(n)$ bits of working space for any $k \geq 1$~\cite[Thm. 7]{belazzougui2016linear}. When $k=\log\sigma$, this running time is the same as our result (\ref{goalMerge}), but the working space is much higher on small alphabets.
We implemented and tested our algorithms (\ref{goalLCP}, \ref{goalMerge}) on DNA alphabet. Our tools use (in RAM) as few as $n$ bits on top of a packed representation of the input/output, and process data as fast as $2.92$ megabases per second.
Contributions (\ref{goalLCP}, \ref{goalMerge}) are part of a preliminary version~\cite{prezza_et_al:LIPIcs:2019:10478} of this paper. This paper also extends such results with the suffix tree interval enumeration procedure and with the algorithms of contribution (\ref{goalPLCP}) for building the PLCP bitvector and the BPS representation of the suffix tree topology.
\section{Basic Concepts}\label{sec:notation}
Let $\Sigma =\{c_1, c_2, \ldots, c_\sigma\}$ be a finite ordered alphabet of size $\sigma$ with $\# = c_1< c_2< \ldots < c_\sigma$, where $<$ denotes the standard lexicographic order. Given a text $T=t_1 t_2 \cdots t_n \in \Sigma^*$ we denote by $|T|$ its length $n$. We assume that the input text is terminated by the special symbol (terminator) $\#$, which does not appear elsewhere in $T$. We use $\epsilon$ to denote the empty string. A \emph{factor} (or \emph{substring}) of $T$ is written as $T[i,j] = t_i \cdots t_j$ with $1\leq i \leq j \leq n$. When declaring an array $A$, we use the same notation $A[1,n]$ to indicate that the array has $n$ entries indexed from $1$ to $n$. A \emph{right-maximal} substring $W$ of $T$ is a string for which there exist at least two distinct characters $a,b$ such that $Wa$ and $Wb$ occur in $T$.
The \emph{suffix array} SA of a string $T$ (see \cite{PuglisiTurpin2008} for a survey) is an array containing the permutation of the integers $1,2, \ldots, n$ that arranges the starting positions of the suffixes of $T$ into lexicographical order, i.e., for all $1 \leq i < j \leq n$, $SA[i] < SA[j]$.
The \emph{inverse suffix array} $ISA[1, n]$ is the inverse permutation of $SA$, i.e., $ISA[i] = j$ if and only if $SA[j] = i$.
The Burrows-Wheeler Transform of a string $T$ is a reversible transformation that permutates its symbols, i.e. $BWT[i]=T[SA[i]-1]$ if $SA[i] > 1$ or $\#$ otherwise.
In some of our results we deal with \emph{string collections}. There exist some natural extensions of the suffix array and the Burrows-Wheeler Transform to a collection of strings.
Let $\mathcal{S}\xspace = \{T_1, \dots, T_m\}$ be a string collection of total length $n$, where each $T_i$ is terminated by a character $\#$ (the terminator) lexicographically smaller than all other alphabet's characters. In particular, a collection is an ordered multiset, and we denote $\mathcal{S}\xspace[i] = T_i$.
We define lexicographic order among the strings' suffixes in the usual way, except that, \emph{only while sorting}, each terminator $\#$ of the $i$-th string $\mathcal{S}\xspace[i]$ is considered (implicitly) a different symbol $\#_i$, with $\#_i < \#_j$ if and only if $i<j$.
Equivalently, in case of equal suffixes ties are broken by input's order: if $T_i[k,|T_i|-1]=T_j[k',|T_j|-1]$, then we define $T_i[k,|T_i|] < T_j[k',|T_j|]$ if and only if $i < j$.
The \emph{generalized suffix array} $GSA[1,n]$ (see~\cite{Shi:1996,CGRS_JDA_2016,Louza2017}) of $\mathcal{S}\xspace$ is an array of pairs $GSA[i] = \langle j,k \rangle$ such that
$\mathcal{S}\xspace[j][k,|\mathcal{S}\xspace[j]|]$ is the $i$-th lexicographically smallest suffix of strings in $\mathcal{S}\xspace$, where we break ties by input position (i.e. $j$ in the notation above). Note that, if the collection is formed by a single string $T$, then the first component in $GSA$'s pairs is always equal to 1, and the second components form the suffix array of $T$.
We denote by $\mathtt{range(W)} = \langle \mathtt{left(W)}, \mathtt{right(W)} \rangle$, also referred to as \emph{suffix array (SA) interval of $W$, or simply $W$-interval}, the maximal pair $\langle L,R \rangle$ such that all suffixes in $GSA[L,R]$ are prefixed by $W$. We use the same notation with the suffix array of a single string $T$. Note that the number of suffixes lexicographically smaller than $W$ in the collection is $L-1$. We extend this definition also to cases where $W$ is not present in the collection: in this case, the (empty) range is $\langle L, L-1\rangle$ and we still require that $L-1$ is the number of suffixes lexicographically smaller than $W$ in the collection (or in the string).
The \emph{extended Burrows-Wheeler Transform}
$BWT[1,n]$ \cite{MantaciRRS07,BauerCoxRosoneTCS2013} of $\mathcal{S}\xspace$ is the character array defined as $BWT[i] = \mathcal{S}\xspace[j][k-1\ \mathtt{mod}\ |\mathcal{S}\xspace[j]|]$, where $\langle j,k \rangle = GSA[i]$.
To simplify notation, we indicate with ``$BWT$'' both the Burrows-Wheeler Transform of a string and of a string collection. The used transform will be clear from the context.
The \emph{longest common prefix} (LCP) array of a string $s$ \cite{ManberMyers1993} (resp. a collection $\mathcal{S}\xspace$ of strings, see \cite{CGRS_JDA_2016,Louza2017,EgidiAMB2019}) is an array storing the length of the longest common prefixes between two consecutive suffixes of $s$ (resp. $\mathcal{S}\xspace$) in lexicographic order (with $LCP[1]=0$). When applied to a string collection, we take the longest common prefix of two equal suffixes of length $\ell$ to be equal to $\ell-1$ (i.e. as if their terminators were different).
Given two collections $\mathcal{S}\xspace_1, \mathcal{S}\xspace_2$ of total length $n$, the Document Array of their union is the binary array $DA[1,n]$ such that $DA[i] = 0$ if and only if the $i$-th smallest suffix comes from $\mathcal{S}\xspace_1$. When merging suffixes of the two collections, ties are broken by collection number (i.e. suffixes of $\mathcal{S}\xspace_1$ are smaller than suffixes of $\mathcal{S}\xspace_2$ in case of ties).
The $C$-array of a string (or collection) $S$ is an array $C[1,\sigma]$ such that $C[i]$ contains the number of characters lexicographically smaller than $i$ in $S$, plus one ($S$ will be clear from the context). Equivalently, $C[c]$ is the starting position of suffixes starting with $c$ in the suffix array of the string. When $S$ (or any of its permutations) is represented with a balanced wavelet tree, then we do not need to store explicitly $C$, and $C[c]$ can be computed in $O(\log\sigma)$ time with no space overhead on top of the wavelet tree (see~\cite{navarro2012wavelet}).
Function $\mathtt{S.rank_c(i)}$ returns the number of characters equal to $c$ in $S[1,i-1]$. When $S$ is represented by a wavelet tree, \emph{rank} can be computed in $O(\log \sigma)$ time.
Function $\mathtt{getIntervals(L,R,BWT)}$, where $BWT$ is the extended Burrows-Wheeler transform of a string collection $\mathcal{S}\xspace$ and $\langle L,R\rangle$ is the suffix array interval of some string $W$ appearing as a substring of some element of $\mathcal{S}\xspace$, returns all suffix array intervals of strings $cW$, with $c\neq \#$, that occur in $\mathcal{S}\xspace$. When $BWT$ is represented with a balanced wavelet tree, we can implement this function so that it terminates in $O(\log\sigma)$ time per returned interval~\cite{beller2013computing}. The function can be made to return the output intervals on-the-fly, one by one (in an arbitrary order), without the need to store them all in an auxiliary vector, with just $O(\log n)$ bits of additional overhead in space~\cite{beller2013computing} (this requires to DFS-visit the sub-tree of the wavelet tree induced by $BWT[L,R]$; the visit requires only $\log\sigma$ bits to store the current path in the tree).
An extension of the above function that navigates in parallel two BWTs is immediate. Function $\mathtt{getIntervals(L_1,R_1,L_2, R_2, BWT_1, BWT_2)}$ takes as input two ranges of a string $W$ on the BWTs of two collections, and returns the pairs of ranges on the two BWTs corresponding to all left-extensions $cW$ of $W$ ($c\neq \#$) such that $cW$ appears in at least one of the two collections. To implement this function, it is sufficient to navigate in parallel the two wavelet trees as long as at least one of the two intervals is not empty.
Let $S$ be a string. The function $S.\mathtt{rangeDistinct(i,j)}$ returns the set of distinct alphabet characters \emph{different than the terminator} $\#$ in $S[i,j]$. Also this function can be implemented in $O(\log\sigma)$ time per returned element when $S$ is represented with a wavelet tree (again, this requires a DFS-visit of the sub-tree of the wavelet tree induced by $S[i,j]$).
$BWT.\mathtt{bwsearch(\langle L,R \rangle, c)}$ is the function that, given the suffix array interval $\langle L,R \rangle$ of a string $W$ occurring in the collection, returns the suffix array interval of $cW$ by using the BWT of the collection~\cite{ferragina2000opportunistic}. This function requires access to array $C$ and \emph{rank} support on $BWT$, and runs in $O(\log\sigma)$ time when $BWT$ is represented with a balanced wavelet tree.
To conclude, our algorithms will take as input a wavelet tree representing the BWT. As shown in the next lemma by Claude et al., this is not a restriction:
\begin{lemma}[\cite{claude2015wavelet}]\label{thm:BWT->WT}
Given a word-packed string of length $n$ on alphabet $[1,\sigma]$, we can replace it with its wavelet matrix~\cite{claude2015wavelet} in $O(n\log\sigma)$ time using $n$ bits of additional working space. \end{lemma}
Wavelet matrices~\cite{claude2015wavelet} are a particular space-efficient representation of wavelet trees taking $n\log\sigma \cdot (1+o(1))$ bits of space and supporting all their operations within the same running times. Since the output of all our algorithms will take at least $n$ bits, it will always be possible to re-use a portion of the output's space (before computing it) to fit the extra $n$ bits required by Lemma \ref{thm:BWT->WT}.
\section{Belazzougui's Enumeration Algorithm}\label{sec:belazzougui}
In~\cite{Belazzougui:2014:LTC:2591796.2591885}, Belazzougui showed that a BWT with \emph{rank} and \emph{range distinct} functionality (see Section \ref{sec:notation}) is sufficient to enumerate in small space a rich representation of the internal nodes of the suffix tree of a text $T$. For the purposes of this article, we assume that the BWT is represented using a wavelet tree (whereas Belazzougui's original result is more general), and thus that all queries take $O(\log \sigma)$ time.
\begin{theorem}[Belazzougui \cite{Belazzougui:2014:LTC:2591796.2591885}]\label{th:Belazzougui}
Given the Burrows-Wheeler Transform of a text $T\in[1,\sigma]^n$ represented with a wavelet tree, we can enumerate the following information for each distinct right-maximal substring $W$ of $T$: (i) $|W|$, and (ii) $range(Wc_i)$ for all $c_1 < \dots < c_k$ such that $Wc_i$ occurs in $T$. The process runs in $O(n\log\sigma)$ time and uses $O(\sigma^2\log^2n)$ bits of working space on top of the BWT. \end{theorem}
To keep the article self-contained, in this section we describe the algorithm at the core of the above result. Remember that explicit suffix tree nodes correspond to right-maximal substrings.
The first idea is to represent any substring $W$ (not necessarily right-maximal) as follows. Let $\mathtt{chars_W[1,k_W]}$ be the alphabetically-sorted character array such that $W\cdot \mathtt{chars_W[i]}$ is a substring of $T$ for all $i=1,\dots, k_W$, where $k_W$ is the number of right-extensions of $W$. We require $\mathtt{chars_W}$ to be also complete: if $Wc$ is a substring of $T$, then $c\in \mathtt{chars_W}$. Let moreover $\mathtt{first_W[1,k_W+1]}$ be the array such that $\mathtt{first_W[i]}$ is the starting position of (the range of) $W\cdot \mathtt{chars_W[i]}$ in the suffix array of $T$ for $i=1,\dots, k_W$, and $\mathtt{first_W[k_W+1]}$ is the end position of $W$ in the suffix array of $T$. The representation for $W$ is (differently from~\cite{Belazzougui:2014:LTC:2591796.2591885}, we omit $\mathtt{chars_W}$ from the representation and we add $|W|$; these modifications will turn useful later):
$$\mathtt{repr(W) = \langle \mathtt{first_W},\ |W| \rangle}$$
Note that, if $W$ is not right-maximal nor a text suffix, then $W$ is followed by $k_W=1$ distinct characters in $T$ and the above representation is still well-defined. When $W$ is right-maximal, we will also say that $\mathtt{repr(W)}$ is the representation of a suffix tree explicit node (i.e. the node reached by following the path labeled $W$ from the root).
\paragraph{Weiner Link Tree Visit}\label{app:belazzougui}
The enumeration algorithm works by visiting the Weiner Link tree of $T$ starting from the root's representation, that is, $\mathtt{repr(\epsilon) = \langle \mathtt{first_\epsilon},\ 0 \rangle}$, where $\mathtt{first_\epsilon} = \langle C[c_1], \dots, C[c_\sigma], n \rangle$ (see Section \ref{sec:notation} for a definition of the $C$-array) and $c_1, \dots, c_\sigma$ are the sorted alphabet's characters. Since the suffix tree and the Weiner link tree share the same set of nodes, this is sufficient to enumerate all suffix tree nodes. The visit uses a stack storing representations of suffix tree nodes, initialized with $\mathtt{repr(\epsilon)}$. At each iteration, we pop the head $\mathtt{repr(W)}$ from the stack and we push $\mathtt{repr(cW)}$ such that $cW$ is right-maximal in $T$.
To keep the stack's size under control, once computed $\mathtt{repr(cW)}$ for the right-maximal left-extensions $cW$ of $W$ we push them on the stack in decreasing order of range length $\mathtt{range(cW)}$ (i.e. the node with the smallest range is pushed last). This guarantees that the stack will always contain at most $O(\sigma\log n)$ elements~\cite{Belazzougui:2014:LTC:2591796.2591885}. Since each element takes $O(\sigma\log n)$ bits to be represented, the stack's size never exceeds $O(\sigma^2\log^2 n)$ bits.
\paragraph{Computing Weiner Links}
We now show how to efficiently compute the node representation $\mathtt{repr(cW)}$ from $\mathtt{repr(W)}$ for the characters $c$ such that $cW$ is right-maximal in $T$. In~\cite{Belazzougui:2014:LTC:2591796.2591885,belazzougui2016linear} this operation is supported efficiently by first enumerating all \emph{distinct} characters in each range $BWT[\mathtt{first_W[i], first_W[i+1]}]$ for $i=1, \dots, k_W$, using function $\mathtt{BWT.rangeDistinct(first_W[i], first_W[i+1])}$ (see Section \ref{sec:notation}).
Equivalently, for each $a\in \mathtt{chars_W}$ we want to list all distinct left-extensions $cWa$ of $Wa$. Note that, in this way, we may also visit implicit suffix tree nodes (i.e. some of these left-extensions could be not right-maximal). Stated otherwise, we are traversing all explicit \emph{and} implicit Weiner links. Since the number of such links is linear~\cite{Belazzougui:2014:LTC:2591796.2591885,belazzougui2014alphabet} (even including implicit Weiner links\footnote{To see this, first note that the number of right-extensions $Wa$ of $W$ that have only one left-extension $cWa$ is at most equal to the number of right-extensions of $W$; globally, this is at most the number of suffix tree's nodes (linear). Any other right-extension $Wa$ that has at least two distinct left-extensions $cWa$ and $bWa$ is, by definition, left maximal and corresponds therefore to a node in the suffix tree of the reverse of $T$. It follows that all left-extensions of $Wa$ can be charged to an edge of the suffix tree of the reverse of $T$ (again, the number of such edges is linear).}), globally the number of distinct characters returned by $\mathtt{rangeDistinct}$ operations is $O(n)$. An implementation of $\mathtt{rangeDistinct}$ on wavelet trees is discussed in \cite{beller2013computing} with the procedure \texttt{getIntervals} (this procedure actually returns more information: the suffix array range of each $cWa$). This implementation runs in $O(\log\sigma)$ time per returned character. Globally, we therefore spend $O(n\log\sigma)$ time using a wavelet tree. We now need to compute $\mathtt{repr(cW)}$ for all left-extensions of $W$ and keep only the right-maximal ones. Let $x=\mathtt{repr(W)}$ and $\mathtt{BWT.Weiner(x)}$ be the function that returns the representations of such strings (used in Line \ref{range distinct2} of Algorithm \ref{alg:fill nodes}). This function can be implemented by observing that $$ \begin{array}{lcl}
\mathtt{range(cWa)} & = \mathtt{\langle}& \mathtt{C[c] + BWT.rank_c(left(Wa))}, \\
&& \mathtt{ C[c] + BWT.rank_c(right(Wa)+1)-1 \ \rangle} \\ \end{array} $$
where $a=\mathtt{chars_W[i]}$ for $1\leq i < |\mathtt{first_W}|$, and noting that $\mathtt{left(Wa)}$ and $\mathtt{right(Wa)}$ are available in $\mathtt{repr(W)}$. Note also that we do not actually need to know the value of characters $\mathtt{chars_W[i]}$ to compute the ranges of each $cW\cdot \mathtt{chars_W[i]}$; this is the reason why we can omit $\mathtt{chars_W}$ from $\mathtt{repr(W)}$. Using a wavelet tree, the above operation takes $O(\log\sigma)$ time. By the above observations, the number of strings $cWa$ such that $W$ is right-maximal is bounded by $O(n)$. Overall, computing $\mathtt{repr(cW)} = \langle \mathtt{first_{cW}}, |W|+1 \rangle$ for all left-extensions $cW$ of all right-maximal strings $W$ takes therefore $O(n\log\sigma)$ time. Within the same running time, we can check which of those extensions is right maximal (i.e. those such that $|\mathtt{first_{cW}}|\geq 2$), sort them in-place by interval length (we always sort at most $\sigma$ node representations, therefore also sorting takes globally $O(n\log\sigma)$ time), and push them on the stack.
\section{Beller et al.'s Algorithm}\label{sec:beller}
The second ingredient used in our solutions is the following result, due to Beller et al (we slightly re-formulate their result to fit our purposes, read below for a description of the differences):
\begin{theorem}[Beller et al.\cite{beller2013computing}]\label{th:Beller}
Given the Burrows-Wheeler Transform of a text $T$ represented with a wavelet tree, we can enumerate all pairs $(i,LCP[i])$ in $O(n\log\sigma)$ time using $5n$ bits of working space on top of the BWT. \end{theorem}
Theorem \ref{th:Beller} represents the state of the art for computing the LCP array from the BWT.
Also Beller et al.'s algorithm works by enumerating a (linear) subset of the BWT intervals. LCP values are induced from a particular visit of those intervals. Belazzougui's and Beller et al.'s algorithms have, however, two key differences which make the former more space-efficient on small alphabets, while the latter more space-efficient on large alphabets: (i) Beller et al. use a queue (FIFO) instead of a stack (LIFO), and (ii) they represent $W$-intervals with just the pair of coordinates $\mathtt{range(W)}$ and the value $|W|$. In short, while Beller et al.'s queue might grow up to size $\Theta(n)$, the use of intervals (instead of the more complex representation used by Belazzougui) makes it possible to represent it using $O(1)$ bitvectors of length $n$. On the other hand, the size of Belazzougui's stack can be upper-bounded by $O(\sigma\log n)$, but its elements take more space to be represented.
We now describe in detail Beller et al.'s result. We keep a bitvector $U[1,n]$ such that $U[i]=0$ if and only if the pair $(i,LCP[i])$ has not been output yet. In their original algorithm, Beller et al. use the LCP array itself to mark undefined LCP entries. In our case, we don't want to store the whole LCP array (for reasons that will be clear in the next sections) and thus we only record which LCP values have been output. Bitvector $U$ accounts for the additional $n$ bits used by Theorem \ref{th:Beller} with respect to the original result described in~\cite{beller2013computing}. At the beginning, $U[i]=0$ for all $i=1, \dots, n$. Beller et al.'s algorithm starts by inserting in the queue the triple $\langle1,n,0 \rangle$, where the first two components are the BWT interval of $\epsilon$ (the empty string) and the third component is its length. From this point, the algorithm keeps performing the following operations until the queue is empty. We remove the first (i.e. the oldest) element $\langle L,R,\ell\rangle$ from the queue, which (by induction) is the interval and length of some string $W$: $\mathtt{range(W)}= \langle L,R \rangle$ and $|W|=\ell$. Using operation $\mathtt{getIntervals(L,R,BWT)}$~\cite{beller2013computing} (see Section \ref{sec:notation}) we left-extend the BWT interval $\langle L,R\rangle$ with the characters $c_1, \dots, c_k$ in $\mathtt{BWT.rangeDistinct(L,R)}$, obtaining the triples $\langle L_1, R_1, \ell+1 \rangle, \dots, \langle L_k, R_k, \ell+1 \rangle$ corresponding to the strings $c_1W, \dots, c_kW$. For each such triple $\langle L_i, R_i, \ell+1\rangle$, if $R_i\neq n$ and $U[R_i+1] = 0$ then we set $U[R_i+1] \leftarrow 1$, we output the LCP pair $(R_i+1, \ell)$ and push $\langle L_i, R_i, \ell+1\rangle$ on the queue. Importantly, note that we can push the intervals returned by $\mathtt{getIntervals(L,R,BWT)}$ in the queue in any order; as discussed in Section \ref{sec:notation}, this step can be implemented with just $O(\log n)$ bits of space overhead with a DFS-visit of the wavelet tree's sub-tree induced by $BWT[L,R]$ (i.e. the intervals are not stored temporarily anywhere: they are pushed as soon as they are generated).
\paragraph{Queue implementation} To limit space usage, Beller et al. use the following queue representations. First note that, at each time point, the queue's triples are partitioned into a (possibly empty) sequence with associated length (i.e. the third element in the triples) $\ell+1$, followed by a sequence with associated length $\ell$, for some $\ell$. To simplify the description, let us assume that these two sequences are kept as two distinct queues, indicated in the following as $Q_\ell$ and $Q_{\ell+1}$. At any stage of the algorithm, we pop from $Q_\ell$ and push into $Q_{\ell+1}$. It follows that there is no need to store strings' lengths in the triples themselves (i.e. the queue's elements become just ranges), since the length of each element in $Q_\ell$ is $\ell$. When $Q_\ell$ is empty, we create a new empty queue $Q_{\ell+2}$, pop from $Q_{\ell+1}$, and push into $Q_{\ell+2}$ (and so on). Beller et al. represent $Q_\ell$ as follows. While pushing elements in $Q_\ell$, as long as its size does not exceed $n/\log n$ we represent it as a vector of pairs (of total size at most $O(n)$ bits). This representation supports push/pop operations in (amortized) constant time and takes at most $O(\log n \cdot n/\log n) = O(n)$ bits of space. As soon as $Q_\ell$'s size exceeds $n/\log n$, we switch to a representation that uses two packed bitvectors of length $n$ storing, respectively, the left- and right-most boundaries of the ranges in the queue.
Note that this representation can be safely used since the pairs in $Q_\ell$ are suffix array ranges of strings of some fixed length $\ell$, therefore there cannot be overlapping intervals. Pushing an interval into such a queue takes constant time (it just requires setting two bits). Popping all the $t = |Q_\ell|$ intervals, on the other hand, can easily be implemented in $O(t+ n/\log n)$ time by scanning the bitvectors and exploiting word-parallelism (see \cite{beller2013computing} for all details). Since Beller et al.'s procedure visits $O(n)$ SA intervals, $Q_\ell$ will exceed size $n/\log n$ for at most $O(\log n)$ values of $\ell$. It follows that also with this queue representation pop operations take amortized constant time.
\paragraph{Time complexity} It is easy to see that the algorithm inserts in total a linear number of intervals in the queue since an interval $\langle L_i, R_i, \ell+1 \rangle$ is inserted only if $U[R_i+1] = 0$, and successively $U[R_i+1]$ is set to $1$. Clearly, this can happen at most $n$ times. In~\cite{beller2013computing} the authors moreover show that, even when counting the left-extensions of those intervals (computed after popping each interval from the queue), the total number of generated intervals stays linear. Overall, the algorithm runs therefore in $O(n\log\sigma)$ time (as discussed in Section \ref{sec:notation}, $\mathtt{getIntervals}$ runs in $O(\log\sigma)$ time per returned element).
\section{Enumerating LCP values}\label{sec:LCP}
In this section we prove our first main result: how to enumerate LCP pairs $(i,LCP[i])$ using succinct working space on top of a wavelet tree representing the BWT. Later we will use this procedure to build the LCP and PLCP arrays in small space on top of a plain representation of the BWT. We give our lemma in the general form of string collections, which will require adapting the algorithms seen in the previous sections to this more general setting. Our first observation is that Theorem \ref{th:Belazzougui}, extended to string collections as described below, can be directly used to enumerate LCP pairs $(i,LCP[i])$ using just $O(\sigma^2\log^2n)$ bits of working space on top of the input and output. We combine this procedure with an extended version of Beller et al.'s algorithm working on string collections in order to get small working space for all alphabets. Algorithms \ref{alg:fill nodes} and \ref{alg:fill leaves} report our complete procedure; read below for an exhaustive description. We obtain our first main result:
\begin{lemma}\label{thm:LCP collection}
Given a wavelet tree for the Burrows-Wheeler Transform of a collection $\mathcal{S}\xspace = \{T_1, \dots, T_m\}$ of total length $n$ on alphabet $[1,\sigma]$, we can enumerate all pairs $(i, LCP[i])$ in $O(n\log\sigma)$ time using $o(n\log\sigma)$ bits of working space on top of the BWT. \end{lemma} \begin{proof}
If $\sigma < \sqrt{n}/\log^2n$ then $\sigma^2\log^2n = o(n)$ and our extension of Theorem \ref{th:Belazzougui} gives us $o(n\log\sigma)$ additional working space. If $\sigma \geq \sqrt{n}/\log^2n$ then $\log\sigma = \Theta(\log n)$ and we can use our extension to string collections of Theorem \ref{th:Beller}, which yields extra working space $O(n) = o(n\log n) = o(n\log\sigma)$. Note that, while we used the threshold $\sigma < \sqrt{n}/\log^2n$, any threshold of the form $\sigma < \sqrt{n}/\log^{1+\epsilon}n$, with $\epsilon>0$ would work. The only constraint is that $\epsilon>0$, since otherwise for $\epsilon=0$ the working space would become $O(n\log\sigma)$ for constant $\sigma$ (not good since we aim at $o(n\log\sigma)$). \end{proof}
We now describe all the details of our extensions of Theorems \ref{th:Belazzougui} and \ref{th:Beller} used in the proof of Lemma \ref{thm:LCP collection}. Procedure \texttt{BGOS(BWT)} in Line \ref{beller et al.} of Algorithm \ref{alg:fill nodes} is a call to Beller et al.'s algorithm, modified as follows. First, we enumerate the LCP pairs $(C[c], 0)$ for all $c\in\Sigma$. Then, we push in the queue $\langle \mathtt{range(c), 1} \rangle$ for all $c\in\Sigma$ and start the main algorithm. Note moreover that (see Section \ref{sec:notation}) from now on we never left-extend ranges with $\#$.
Recall that each string of a text collection $\mathcal{S}\xspace$ is ended by a terminator $\#$ (common to all strings). Consider now the LCP and GSA
arrays of $\mathcal{S}\xspace$. We divide LCP values in two types.
Let $GSA[i] = \langle j,k \rangle$, with $i>1$, indicate that the $i$-th suffix in the lexicographic ordering of all suffixes of strings in $\mathcal{S}\xspace$ is $\mathcal{S}\xspace[j][k,|\mathcal{S}\xspace[j]|]$. A LCP value $\mathtt{LCP[i]}$ is of \emph{node type} when the $i$-th and $(i-1)$-th suffixes are distinct: $\mathcal{S}\xspace[j][k,|\mathcal{S}\xspace[j]|] \neq \mathcal{S}\xspace[j'][k',|\mathcal{S}\xspace[j']|]$, where $GSA[i] = \langle j,k \rangle$ and $GSA[i-1] = \langle j',k' \rangle$. Those two suffixes differ before the terminator is reached in both suffixes (it might be reached in one of the two suffixes, however); we use the name \emph{node-type} because $i-1$ and $i$ are the last and first suffix array positions of the ranges of two adjacent children of some suffix tree node, respectively (i.e. the node corresponding to string $\mathcal{S}\xspace[j][k,k+LCP[i]-1]$). Note that it might be that one of the two suffixes, $\mathcal{S}\xspace[j][k,|\mathcal{S}\xspace[j]|]$ or $\mathcal{S}\xspace[j'][k',|\mathcal{S}\xspace[j']|]$, is the string ``$\#$''. Similarly, a \emph{leaf-type} LCP value $\mathtt{LCP[i]}$ is such that the $i$-th and $(i-1)$-th suffixes are equal: $\mathcal{S}\xspace[j][k,|\mathcal{S}\xspace[j]|] = \mathcal{S}\xspace[j'][k',|\mathcal{S}\xspace[j']|]$. We use the name \emph{leaf-type} because, in this case, it must be the case that $i \in [L+1,R]$, where $\langle L,R \rangle$ is the suffix array range of some suffix tree leaf (it might be that $R>L$ since there might be repeated suffixes in the collection). Note that, in this case, $\mathcal{S}\xspace[j][k,|\mathcal{S}\xspace[j]|] = \mathcal{S}\xspace[j'][k',|\mathcal{S}\xspace[j']|]$ could coincide with $\#$. Entry $LCP[0]$ escapes the above classification, so we output it separately.
Our idea is to compute first node-type and then leaf-type LCP values. We argue that Beller et al.'s algorithm already computes the former kind of LCP values. When this algorithm uses too much space (i.e. on small alphabets), we show that Belazzougui's enumeration strategy can be adapted to reach the same goal: by the very definition of node-type LCP values, they lie between children of some suffix tree node $x$, and their value corresponds to the string depth of $x$. This strategy is described in Algorithm \ref{alg:fill nodes}. Function $\mathtt{BWT.Weiner(x)}$ in Line \ref{range distinct2} takes as input the representation of a suffix tree node $x$ and returns all explicit nodes reached by following Weiner links from $x$ (an implementation of this function is described in Section~\ref{sec:belazzougui}). Leaf-type LCP values, on the other hand, can easily be computed by enumerating intervals corresponding to suffix tree leaves. To reach this goal, it is sufficient to enumerate ranges of suffix tree leaves starting from $\mathtt{range(\#)}$ and recursively left-extending with backward search with characters different from $\#$ whenever possible. For each range $\langle L,R \rangle$ obtained in this way, we set each entry $LCP[L+1,R]$ to the string depth (terminator excluded) of the corresponding leaf. This strategy is described in Algorithm \ref{alg:fill leaves}. In order to limit space usage, we use again a stack or a queue to store leaves and their string depth (note that each leaf takes $O(\log n)$ bits to be represented): we use a queue when $\sigma > n/\log^3n$, and a stack otherwise. The queue is the same used by Beller et al.\cite{beller2013computing} and described in Section \ref{sec:beller}. This guarantees that the bit-size of the queue/stack never exceeds $o(n\log\sigma)$ bits: since leaves take just $O(\log n)$ bits to be represented and the stack's size never contains more than $O(\sigma\cdot \log n)$ leaves, the stack's bit-size never exceeds $O(n/\log n) = o(n)$ when $\sigma \leq n/\log^3n$. Similarly, Beller et al's queue always takes at most $O(n)$ bits of space, which is $o(n\log\sigma)$ for $\sigma > n/\log^3n$.
Note that in Lines \ref{getIntervals1}-\ref{push2} we can afford storing temporarily the $k$ resulting intervals since, in this case, the alphabet's size is small enough.
To sum up, our full procedure works as follows: (1), we output node-type LCP values using procedure \texttt{Node-Type}$(\mathtt{BWT})$ described in Algorithm \ref{alg:fill nodes}, and (2) we output leaf-type LCP values using procedure \texttt{Leaf-Type}$(\mathtt{BWT})$ described in Algorithm \ref{alg:fill leaves}.
\begin{algorithm}
\begin{algorithmic}[1]
\If{$\sigma > \sqrt n/\log^2n$}
\State $\mathtt{BGOS(BWT)}$ \Comment Run Beller et al.'s algorithm\label{beller et al.}
\Else
\State $\mathtt P \leftarrow \texttt{new\_stack()}$\Comment Initialize new stack\label{new stack2}
\State $\mathtt P\mathtt{.push( repr(\epsilon))}$\Comment Push representation of $\epsilon$ \label{push3}
\While{$\mathtt{\mathbf{not}\ P.empty()}$}\label{while2}
\State $\langle \mathtt{first_W},\ \ell \rangle \leftarrow \mathtt{P.pop()}$\Comment Pop highest-priority element\label{pop2}
\State $t \leftarrow |\mathtt{first_W}|-1$\Comment Number of children of ST node\label{nchild}
\For{$i = 2, \dots, t$}
\State \textbf{output} $(\mathtt{first_W}[i], \ell)$\Comment Output LCP value\label{LCP in Node}
\EndFor
\State $x_1,\dots, x_k \leftarrow \mathtt{BWT.Weiner(\langle first_W,\ \ell \rangle)}$\Comment Follow Weiner Links\label{range distinct2}
\State $x'_1, \dots, x'_k \leftarrow \mathtt{sort}(x_1, \dots, x_k)$\Comment Sort by interval length\label{sort2}
\For{$i=k\dots 1$}
\State $\mathtt{P.push}(x'_i)$\Comment Push representations\label{push4}
\EndFor
\EndWhile
\EndIf
\caption{\texttt{Node-Type(BWT)}}\label{alg:fill nodes} \end{algorithmic} \end{algorithm}
\begin{algorithm}
\begin{algorithmic}[1]
\For{$i=left(\#),\dots, right(\#)$}
\State \textbf{output} $(i,0)$
\EndFor
\If{$\sigma > n/\log^3n$}
\State $\mathtt P \leftarrow \mathtt{new\_queue()}$\Comment{Initialize new queue}\label{new queue1}
\Else
\State $\mathtt P \leftarrow \mathtt{new\_stack()}$\Comment{Initialize new stack}\label{new stack1}
\EndIf
\State $\mathtt P\mathtt{.push( BWT.range(\#),0)}$\Comment{Push range of terminator and LCP value 0}\label{push1}
\While{$\mathtt{\mathbf{not}\ P.empty()}$}\label{while1}
\State $\langle \langle L,R \rangle, \ell \rangle \leftarrow \mathtt{P.pop()}$\Comment{Pop highest-priority element}\label{pop1}
\For{$i=L+1\dots R$}
\State \textbf{output} $(i, \ell)$\Comment{Output LCP inside range of ST leaf}\label{LCP in Leaves}
\EndFor
\If{$\sigma > n/\log^3n$}
\State $\mathtt{P.push(getIntervals(L, R, BWT), \ell+1)}$\Comment{Pairs $\langle$interval,$\ell+1\rangle$}\label{push7}
\Else
\State $\langle L_i, R_i\rangle_{i=1, \dots, k} \leftarrow \mathtt{getIntervals(L, R,BWT)}$\label{getIntervals1}
\State $\langle L'_i, R'_i\rangle_{i=1, \dots, k} \leftarrow \mathtt{sort}(\langle L_i, R_i\rangle_{i=1, \dots, k})$\Comment{Sort by interval length}\label{sort1}
\For{$i=k\dots 1$}
\State $\mathtt{P.push}(\langle L'_i, R'_i\rangle, \ell+1)$\Comment{Push in order of decreasing length}\label{push2}
\EndFor
\EndIf
\EndWhile
\caption{\texttt{Leaf-Type(BWT)}}\label{alg:fill leaves} \end{algorithmic} \end{algorithm}
The correctness, completeness, and complexity of our procedure are proved in the following Lemma:
\begin{lemma}\label{lemma:proof of thm1}
Algorithms \ref{alg:fill nodes} and \ref{alg:fill leaves} correctly output all LCP pairs $(i,LCP[i])$ of the collection in $O(n\log\sigma)$ time using $o(n\log\sigma)$ bits of working space on top of the input BWT. \end{lemma} \begin{proof}
\emph{Correctness - Algorithm \ref{alg:fill nodes}}. We start by proving that Beller et al.'s procedure in Line \ref{beller et al.} of Algorithm \ref{alg:fill nodes} (procedure \texttt{BGOS(BWT)}) outputs all the node-type LCP entries correctly. The proof proceeds by induction on the LCP value $\ell$ and follows the original proof of~\cite{beller2013computing}.
At the beginning, we insert in the queue all $c$-intervals, for $c\in\Sigma$. For each such interval $\langle L,R \rangle$ we output $LCP[R+1]=\ell = 0$. It is easy to see that after this step all and only the node-type LCP values equal to 0 have been correctly computed.
Assume, by induction, that all node-type LCP values less than or equal to $\ell$ have been correctly output, and that we are about to extract from the queue the first triple $\langle L,R,\ell+1 \rangle$ having length $\ell+1$. For each extracted triple with length $\ell+1$ associated to a string $W$, consider the triple $\langle L',R',\ell+2 \rangle$ associated to one of its left-extensions $cW$. If $LCP[R'+1]$ has been computed, i.e. if $U[R'+1]=1$, then we have nothing to do. However, if $U[R'+1]=0$, then it must be the case that (i) the corresponding LCP value satisfies $LCP[R'+1] \geq \ell+1$, since by induction we have already computed all node-type LCP values smaller than or equal to $\ell$, and (ii) $LCP[R'+1]$ is of node-type, since otherwise the BWT interval of $cW$ would also include position $R'+1$. On the other hand, it cannot be the case that $LCP[R'+1] > \ell+1$ since otherwise the $cW$-interval would include position $R'+1$. We therefore conclude that $LCP[R'+1] = \ell+1$ must hold.
\emph{Completeness - Algorithm \ref{alg:fill nodes}}. The above argument settles correctness; to prove completeness, assume that, at some point, $U[i] = 0$ and the value of $LCP[i]$ to be computed and output is $\ell+1$. We want to show that we will pull a triple $\langle L,R,\ell+1 \rangle$ from the queue corresponding to a string $W$ (note that $\ell+1=|W|$ and, moreover, $W$ could end with $\#$) such that one of the left-extensions $aW$ of $W$ satisfies $\mathtt{range(aW)} = \langle L',i-1 \rangle$, for some $L'$. This will show that, at some point, we will output the LCP pair $(i, \ell+1)$.
We proceed by induction on $|W|$.
Note that we separately output all LCP values equal to 0.
The base case $|W|=1$ is easy: by the way we initialized the queue, $\langle \mathtt{range(c)}, 1\rangle$, for all $c\in\Sigma$, are the first triples we pop. Since we left-extend these ranges with all alphabet's characters except $\#$, it is easy to see that all LCP values equal to 1 have been output. From now on we can therefore assume that we are working on LCP values equal to $\ell+1>1$, i.e. $W=b\cdot V$, for $b\in\Sigma-\{\#\}$ and $V\in \Sigma^+$.
Let $abV$ be
the length-$(\ell+2)$ left-extension of $W=bV$
such that $\mathtt{right(abV)+1} = i$. Since, by our initial hypothesis, $\mathtt{LCP[i]} = \ell+1$, the collection contains also a suffix $aU$ lexicographically larger than $abV$ and such that $\mathtt{LCP(aU,abV)} = \ell+1$. But then, it must be the case that $\mathtt{LCP(right(bV)+1)} = \ell$ (it cannot be smaller by the existence of $U$ and it cannot be larger since $|bV|=\ell+1$). By inductive hypothesis, this value was set after popping a triple $\langle L'', R'', \ell\rangle$ corresponding to string $V$, left-extending $V$ with $b$, and pushing $\langle \mathtt{range(bV)}, \ell+1 \rangle$ in the queue.
This ends the completeness proof since we showed that $\langle \mathtt{range(bV)}, \ell+1 \rangle$ is in the queue, so at some point we will pop it, extend it with $a$, and output $(right(abV)+1,\ell+1) = (i,\ell+1)$.
If the queue uses too much space, then Algorithm \ref{alg:fill nodes} switches to a stack and Lines \ref{new stack2}-\ref{push4} are executed instead of Line \ref{beller et al.}. Note that this pseudocode fragment corresponds to Belazzougui's enumeration algorithm, except that now we also set LCP values in Line \ref{LCP in Node}. By the enumeration procedure's correctness, we have that, in Line \ref{LCP in Node}, $\langle \mathtt{first_W[1]}, \mathtt{first_W[t+1]} \rangle$ is the SA-range of a right-maximal string $W$ with $\ell = |W|$, and $\mathtt{first_W[i]}$ is the first position of the SA-range of $Wc_i$, with $i=1,\dots,t$, where $c_1, \dots, c_2$ are all the (sorted) right-extensions of $W$. Then, clearly each LCP value in Line \ref{LCP in Node} is of node-type and has value $\ell$, since it is the LCP between two strings prefixed by $W\cdot \mathtt{chars_W[i-1]}$ and $W\cdot \mathtt{chars_W[i]}$. Similarly, completeness of the procedure follows from the completeness of the enumeration algorithm. Let $LCP[i]$ be of node-type. Consider the prefix $Wb$ of length $LCP[i]+1$ of the $i$-th suffix in the lexicographic ordering of all strings' suffixes. Since $LCP[i] = |W|$, the $(i-1)$-th suffix is of the form $Wa$, with $b\neq a$, and $W$ is right-maximal. But then, at some point our enumeration algorithm will visit the representation of $W$, with $|W|=\ell$. Since $i$ is the first position of the range of $Wb$, we have that $i= \mathtt{first_W[j]}$ for some $j \geq 2$, and Line \ref{LCP in Node} correctly outputs the LCP pair $(first_W[j], |W|) = (i,|W|)$.
\emph{Correctness and completeness - Algorithm \ref{alg:fill leaves}}. Proving correctness and completeness of this procedure is much easier. It is sufficient to note that the \texttt{while} loop iterates over all ranges $\langle L,R \rangle$ of strings ending with $\#$ and not containing $\#$ anywhere else (note that we start from the range of $\#$ and we proceed by recursively left-extending this range with symbols different than $\#$). Then, for each such range we conclude that $LCP[L+1,R]$ is equal to $\ell$, i.e. the string depth of the corresponding string (excluding the final character $\#$). By their definition, all leaf-type LCP values are correctly computed in this way.
\emph{Complexity - Algorithm \ref{alg:fill nodes}}. If $\sigma > \sqrt n/\log^2 n$, then we run Beller et al's algorithm, which terminates in $O(n\log\sigma)$ time and uses $O(n) = o(n\log\sigma)$ bits of additional working space. Otherwise, we perform a linear number of operations on the stack since, as observed in Section \ref{sec:belazzougui}, the number of Weiner links is linear. By the same analysis of Section \ref{sec:belazzougui}, the operation in Line \ref{range distinct2} takes $O(k\log\sigma)$ amortized time on wavelet trees, and sorting in Line \ref{sort2} (using any comparison-sorting algorithm sorting $m$ integers in $O(m\log m)$ time) takes $O(k\log\sigma)$ time. Note that in this sorting step we can afford storing in temporary space nodes $x_1, \dots, x_k$ since this takes additional space $O(k\sigma\log n) = O(\sigma^2\log n) = O(n/\log^3n) = o(n)$ bits.
All these operations sum up to $O(n\log\sigma)$ time. Since the stack always takes at most $O(\sigma^2\log^2n)$ bits and $\sigma \leq \sqrt n/\log^2 n$, the stack's size never exceeds $O(n/\log^2n) = o(n)$ bits.
\emph{Complexity - Algorithm \ref{alg:fill leaves}}. Note that, in the \texttt{while} loop, we start from the interval of $\#$ and recursively left-extend with characters different than $\#$ until this is possible. It follows that we visit the intervals of all strings of the form $W\#$ such that $\#$ does not appear inside $W$. Since these intervals form a cover of $[1,n]$, their number (and therefore the number of iterations in the \texttt{while} loop) is also bounded by $n$. This is also the maximum number of operations performed on the queue/stack. Using Beller et al.'s implementation for the queue and a simple vector for the stack, each operation takes constant amortized time. Operating on the stack/queue takes therefore overall $O(n)$ time.
For each interval $\langle L,R \rangle$ popped from the queue/stack, in Line \ref{LCP in Leaves} we output $R-L-2$ LCP values. As observed above, these intervals form a cover of $[1,n]$ and therefore Line \ref{LCP in Leaves} is executed no more than $n$ times. Line \ref{getIntervals1} takes time $O(k\log\sigma)$. Finally, in Line \ref{sort1} we sort at most $\sigma$ intervals. Using any fast comparison-based sorting algorithm, this costs overall at most $O(n\log\sigma)$ time.
As far as the space usage of Algorithm \ref{alg:fill leaves} is concerned, note that we always push just pairs interval/length ($O(\log n)$ bits) in the queue/stack. If $\sigma > n/\log^3n$, we use Beller et al.'s queue, taking at most $O(n) = o(n\log\sigma)$ bits of space. Otherwise, the stack's size never exceeds $O(\sigma\cdot \log n)$ elements, with each element taking $O(\log n)$ bits. This amounts to $O(\sigma\cdot \log^2 n) = O(n/\log n) = o(n)$ bits of space usage. Moreover, in Lines \ref{getIntervals1}-\ref{sort1} it holds $\sigma\leq n/\log^3n$ so we can afford storing temporarily all intervals returned by $\mathtt{getIntervals}$ in $O(k\log n) = O(\sigma\log n) = O(n/\log^2n) = o(n)$ bits. \end{proof}
Combining Lemma \ref{thm:LCP collection} and Lemma \ref{thm:BWT->WT}, we obtain:
\begin{theorem}\label{thm:LCP collection succinct}
Given the word-packed Burrows-Wheeler Transform of a collection $\mathcal{S}\xspace = \{T_1, \dots, T_m\}$ of total length $n$ on alphabet $[1,\sigma]$, we can build the LCP array of the collection in $O(n\log\sigma)$ time using $o(n\log\sigma)$ bits of working space on top of the BWT. \end{theorem}
\section{Enumerating Suffix Tree Intervals}\label{sec:intervals}
In this section we show that the procedures described in Section \ref{sec:LCP} can be used to enumerate all suffix tree intervals---that is, the suffix array intervals of all right-maximal text substrings---taking as input the BWT of a text. Note that in this section we consider just simple texts rather than string collections as later we will use this procedure to build the compressed suffix tree of a text.
When $\sigma \leq \sqrt n/\log^2n$, we can directly use Belazzougui's procedure (Theorem \ref{th:Belazzougui}), which already solves the problem. For larger alphabets, we modify Beller et al's procedure (Theorem \ref{th:Beller}) to also generate suffix tree's intervals as follows.
When $\sigma > \sqrt n/\log^2n$, we modify Beller et al.'s procedure to enumerate suffix tree intervals using $O(n) = o(n\log\sigma)$ bits of working space, as follows.
We recall that (see Section \ref{sec:beller}), Beller et al's procedure can be conveniently described using two separate queues: $Q_{\ell}$ and $Q_{\ell+1}$. At each step, we pop from $Q_\ell$ an element $\langle\langle L,R \rangle, |W| \rangle$ with $\langle L,R \rangle = \mathtt{range(W)}$ and $|W|=\ell$ for some string $W$, left-extend the range with all $a\in \mathtt{BWT.rangeDistinct(L,R)}$, obtaining the ranges $\mathtt{range(aW)} = \langle L_a, R_a\rangle$ and, only if $U[R_a+1]=0$, set $U[R_a+1]\leftarrow 1$, output the LCP pair $(R_a+1, |W|)$, and push $\langle \langle L_a, R_a\rangle, |W|+1 \rangle$ into $Q_{\ell+1}$. Note that, since $LCP[R_a+1] = |W|$ we have that the $R_a$-th and $(R_a+1)$-th smallest suffixes start, respectively, with $aXc$ and $aXd$ for some $c<d\in\Sigma$, where $W=Xc$. This implies that $aX$ is right-maximal. It is also clear that, from the completeness of Beller et al.'s procedure, all right-maximal text substrings are visited by the procedure, since otherwise the LCP values equal to $\ell = |aX|$ inside $\mathtt{range(aX)}$ would not be generated. It follows that, in order to generate all suffix tree intervals \emph{once}, we need two extra ingredients: (i) whenever we pop from $Q_\ell$ an element $\langle\langle L,R \rangle, |W| \rangle$ corresponding to a string $W = Xc$, we also need the range of $X$, and (ii) we need to quickly check if a given range $\mathtt{range(aX)}$ of a right-maximal substring $aX$ has already been output. Point (ii) is necessary since, using only the above procedure (augmented with point (i)), $\mathtt{range(aX)}$ will be output for each of its right-extensions (except the lexicographically largest, which does not cause the generation of an LCP pair).
Remember that, in order to keep space usage under control (i.e. $O(n)$ bits), we represent $Q_\ell$ as a standard queue of pairs $\langle \mathtt{range(W)}, |W| \rangle$ if and only if $|Q_\ell| < n/\log n$. For now, let us assume that the queue size does not exceed this quantity (the other case will be considered later).
In this case, to implement point (i) we simply augment queue pairs as $\langle \mathtt{range(W)}, \mathtt{range(X)}, |W| \rangle$, where $W=Xc$ for some $c\in\Sigma$. When left-extending $W$ with a character $a$, we also left-extend $X$ with $a$, obtaining $\mathtt{range(aX)}$. Let $\mathtt{range(aW)} = \langle L_a, R_a\rangle$. At this point, if $U[R_a+1]=0$ we do the following: \begin{enumerate}
\item we set $U[R_a+1]\leftarrow 1$,
\item we push $\langle \mathtt{range(aW)}, \mathtt{range(aX)}, |W|+1 \rangle$ in $Q_{\ell+1}$, and
\item if $\mathtt{range(aX)}$ has not already been generated, we output $\mathtt{range(aX)}$. \end{enumerate}
Note that steps (1) and (2) correspond to Beller et al.'s procedure. The test in step (3) (that is, point (ii) above) can be implemented as follows. Note that a suffix array range $\mathtt{range(aX)} = \langle L,R\rangle$ can be identified unambiguously by the two integers $L$ and $|aX| = \ell$. Note also that we generate suffix tree intervals in increasing order of string depth (i.e. when popping elements from $Q_\ell$, we output suffix array intervals of string depth $\ell$). It follows that we can keep a bitvector $GEN_\ell$ of length $n$ recording in $GEN_\ell[i]$ whether or not the suffix array interval of the string of length $\ell$ whose first coordinate is $i$ has already been output. Each time we change the value of a bit $GEN_\ell[i]$ from 0 to 1, we also push $i$ into a stack $SET_\ell$. Let us assume for now that also $SET_\ell$'s size does not exceed $n/\log n$ (later we will consider a different representation for the other case). Then, also the bit-size of $SET_\ell$ will never exceed $O(n)$ bits. After $Q_\ell$ has been emptied, for each $i\in SET_\ell$ we set $GEN_\ell[i] \leftarrow 0$. This makes all $GEN_\ell$'s entries equal to 0, and we can thus re-use its space for $GEN_{\ell+1}$ at the next stage (i.e. when popping elements from $Q_{\ell+1}$).
Now, let us consider the case $|Q_\ell| \geq n/\log n$. The key observation is that $Q_\ell$ exceeds this value for at most $O(\log n)$ values of $\ell$, therefore we can afford spending extra $O(n/\log n)$ time to process each of these queues. As seen in Section \ref{sec:beller}, whenever $Q_\ell$'s size exceeds $n/\log n$ (while pushing elements in it) we switch to a different queue representation using packed bitvectors.
Point (i) can be solved by storing two additional bitvectors as follows. Suppose we are about to push the triple $\langle \mathtt{range(W)}, \mathtt{range(X)}, |W| \rangle$ in $Q_\ell$, where $W=Xc$ for some $c\in\Sigma$. The solution seen in Section \ref{sec:beller} consisted in marking, in two packed bitvectors $\mathtt{open[1,n]}$ and $\mathtt{close[1,n]}$, the start and end points of $\mathtt{range(W)}$. Now, we just use two additional packed bitvectors $\mathtt{\overline{open}[1,n]}$ and $\mathtt{\overline{close}[1,n]}$ to also mark the start and end points of $\mathtt{range(X)}$. As seen in Section \ref{sec:beller}, intervals are extracted from $Q_\ell$ by scanning $\mathtt{open[1,n]}$ and $\mathtt{close[1,n]}$ in $O(n/\log n + |Q_\ell|)$ time (exploiting word-parallelism). Note that $W$ is a right-extension of $X$, therefore $\mathtt{range(W)}$ is contained in $\mathtt{range(X)}$. It follows that we can scan in parallel the bitvectors $\mathtt{open[1,n]}$, $\mathtt{close[1,n]}$, $\mathtt{\overline{open}[1,n]}$, and $\mathtt{\overline{close}[1,n]}$ and retrieve, for each $\mathtt{range(W)}$ extracted from the former two bitvectors, the (unique in the queue) interval $\mathtt{range(X)}$ enclosing $\mathtt{range(W)}$ (using the latter two bitvectors).
More formally, whenever finding a bit set at $\mathtt{\overline{open}[i]}$, we search $\mathtt{\overline{close}[i,n]}$ to find the next bit set. Let us call $j$ the position containing such bit set. Then, we similarly scan $\mathtt{open[i,j]}$ and $\mathtt{close[i,j]}$ to generate all intervals $\langle l,r \rangle$ enclosed by $\langle i,j \rangle$, and for each of them generate the triple $\langle \langle l,r \rangle, \langle i,j \rangle, \ell \rangle$. Again, exploiting word-parallelism the process takes $O(n/\log n + |Q_\ell|)$ time to extract all triples $\langle \mathtt{range(W)}, \mathtt{range(X)}, |W| \rangle$ from $Q_\ell$.
A similar solution can be used to solve point (ii) for large $SET_\ell$. Whenever $SET_\ell$ exceeds size $n/\log n$, we simply empty it and just use bitvector $GEN_\ell$. This time, however, this bitvector is packed in $O(n/\log n)$ words. It can therefore be erased (i.e. setting all its entries to 0) in $O(n/\log n)$ time, and we do not need to use the stack $SET_\ell$ at all. Since (a) we insert an element in some $SET_\ell$ only when outputting a suffix tree range and (b) in total we output $O(n)$ such ranges, $SET_\ell$ can exceed size $n/\log n$ for at most $O(\log n)$ values of $\ell$. We conclude that also the cost of creating and processing all $GEN_\ell$ and $SET_\ell$ amortizes to $O(n)$.
To sum up, the overall procedure runs in $O(n\log\sigma)$ time and uses $O(n)$ bits of space. By combining it with Belazzougui's procedure as seen above (i.e. choosing the right procedure according to the alphabet's size), we obtain:
\begin{lemma}\label{lem:ST intervals}
Given a wavelet tree representing the Burrows-Wheeler transform of a text $T$ of length $n$ on alphabet $[1,\sigma]$, in $O(n\log\sigma)$ time and $o(n\log\sigma)$ bits of working space we can enumerate the suffix array intervals corresponding to all right maximal text's substrings. \end{lemma}
\section{Building the PLCP Bitvector}\label{sec:PLCP}
The PLCP array is defined as $PLCP[i] = LCP[ISA[i]]$, and can thus be used to retrieve LCP values as $LCP[i] = PLCP[SA[i]]$ (note that this requires accessing the suffix array). Kasai et al. showed in~\cite{10.1007/3-540-48194-X_17} that PLCP is almost increasing: $PLCP[i+1] \geq PLCP[i]-1$. This allows representing it in small space as follows. Let $\mathtt{plcp[1,2n]}$ denote the bitvector having a bit set at each position $PLCP[i]+2i$, for $i=1, \dots, n$ (and 0 in all other positions). Since $PLCP[i+1] \geq PLCP[i]-1$, the quantity $PLCP[i]+2i$ is different for each $i$. By definition, $PLCP[i]$ can be written as $j-2i$, where $j$ is the position of the $i$-th bit set in $\mathtt{plcp}$; this shows that each PLCP entry can be retrieved in constant time using the bitvector $\mathtt{plcp}$, augmented to support constant-time \emph{select} queries.
We now show how to build the $\mathtt{plcp}$ bitvector in small space using the LCP enumeration procedure of Section \ref{sec:LCP}. Our procedure relies on the concept of \emph{irreducible LCP values}:
\begin{definition}\label{def:irreducible} $LCP[i]$ is said to be \emph{irreducible} if and only if either $i=0$ or $BWT[i] \neq BWT[i-1]$ hold. \end{definition}
We call \emph{reducible} a non-irreducible LCP value. We extend the above definition to PLCP values, saying that $PLCP[i]$ is irreducible if and only if $LCP[ISA[i]]$ is irreducible. The following Lemma, shown in~\cite{10.1007/978-3-540-27810-8_32}, is easy to prove (see also ~\cite[Lem. 4]{10.1007/978-3-642-02441-2_17}):
\begin{lemma}[\cite{10.1007/978-3-540-27810-8_32}, Lem. 1]\label{lem:reducible}
If $PLCP[i]$ is reducible, then $PLCP[i] = PLCP[i - 1] - 1$. \end{lemma}
We also make use of the following Theorem from K{\"a}rkk{\"a}inen et al.~\cite{10.1007/978-3-642-02441-2_17}:
\begin{theorem}[\cite{10.1007/978-3-642-02441-2_17}, Thm. 1]\label{lem:sum irreducible}
The sum of all irreducible lcp values is at most $2n\log n$. \end{theorem}
Our strategy is as follows. We divide $BWT[1,n]$ in $\lceil n/B \rceil$ blocks $BWT[(i-1)\cdot B+1,i\cdot B]$, $i=1, \dots, \lceil n/B \rceil$ of size $B$ (assume for simplicity that $B$ divides $n$). For each block $i=1, \dots, \lceil n/B \rceil$, we use Lemma \ref{thm:LCP collection} to enumerate all pairs $(j,LCP[j])$. Whenever we generate a pair $(j,LCP[j])$ such that (i) $j$ falls in the current block's range $[(i-1)\cdot B+1,i\cdot B]$, (ii) $LCP[j]> \log^3 n$, and (iii) $LCP[j]$ is irreducible (this can be checked easily using Definition \ref{def:irreducible}), we store $(j,LCP[j])$ in a temporary array $\mathtt{LARGE\_LCP}$ (note: each such pair requires $O(\log n)$ bits to be stored). By Theorem \ref{lem:sum irreducible}, there cannot be more than $2n/\log^2 n$ irreducible LCP values being larger than $\log^3 n$, that is, $\mathtt{LARGE\_LCP}$ will never contain more than $2n/\log^2 n$ values and its bit-size will never exceed $O(n/\log n) = o(n)$ bits. We also mark all such relative positions $j-(i-1)\cdot B$ in a bitvector of length $B$ with rank support and radix-sort $\mathtt{LARGE\_LCP}$ in $O(B)$ time to guarantee constant-time access to $LCP[j]$ whenever conditions (i-iii) hold true for index $j$. On the other hand, if (i) $j$ falls in the current block's range $[(i-1)\cdot B+1,i\cdot B]$, (ii) $LCP[j] \leq \log^3 n$, and (iii) $LCP[j]$ is irreducible then we can store $LCP[j]$ in another temporary vector $\mathtt{SMALL\_LCP[1,B]}$ as follows: $\mathtt{SMALL\_LCP}[j-(i-1)\cdot B] \leftarrow LCP[j]$ (at the beginning, the vector is initialized with undefined values). By condition (ii), $\mathtt{SMALL\_LCP}$ can be stored in $O(B\log\log n)$ bits. Using $\mathtt{LARGE\_LCP}$ and $\mathtt{SMALL\_LCP}$, we can access in constant time all irreducible values $LCP[j]$ whenever $j$ falls in the current block $[(i-1)\cdot B+1,i\cdot B]$. At this point, we enumerate all pairs $(i,ISA[i])$ in text order (i.e. for $i=1, \dots, n$) using the FL function on the BWT. Whenever one of those pairs $(i,ISA[i]) = (i,j)$ is such that (i) $j$ falls in the current block's range $[(i-1)\cdot B+1,i\cdot B]$ and (ii) $LCP[j]$ is irreducible, we retrieve $LCP[j]$ in constant time as seen above and we set $\mathtt{plcp[2i+LCP[j]]} \leftarrow 1$; the correctness of this assignment follows from the fact that $j=ISA[i]$, thus $LCP[j] = PLCP[i]$. Using Lemma \ref{lem:reducible}, we can moreover compute the reducible PLCP values that follow $PLCP[i]$ in text order (up to the next irreducible value), and set the corresponding bits in $\mathtt{plcp}$. After repeating the above procedure for all blocks $BWT[(i-1)\cdot B+1,i\cdot B]$, $i=1, \dots, \lceil n/B \rceil$, we terminate the computation of bitvector $\mathtt{plcp}$. For each block, we spend $O(n\log\sigma)$ time (one application of Lemma \ref{thm:LCP collection} and one BWT navigation to generate all pairs $(i,ISA[i])$). We also spend $O(n/\log^2 n)$ time to allocate the instances of $\mathtt{LARGE\_LCP}$ across all blocks. Overall, we spend $O((n^2/B)\log\sigma + n\log\sigma)$ time across all blocks. The space used is $o(n) + O(B\cdot \log\log n)$ bits on top of the BWT. By setting $B = (\epsilon\cdot n\log\sigma)/\log\log n$ we obtain our result:
\begin{lemma}\label{lem:PLCP}
Given a wavelet tree for the Burrows-Wheeler transform of a text $T$ of length $n$ on alphabet $[1,\sigma]$, for any parameter $0<\epsilon \leq 1$ we can build the PLCP bitvector in $O(n(\log\sigma + \epsilon^{-1}\log\log n))$ time and $\epsilon \cdot n\log\sigma + o(n)$ bits of working space on top of the input BWT and the optput. \end{lemma}
\section{Building the Suffix Tree Topology}\label{sec:ST topology}
In order to build the suffix tree topology we use a strategy analogous to the one proposed by Belazzougui~\cite{Belazzougui:2014:LTC:2591796.2591885}. The main observation is that, given a procedure that enumerates suffix tree intervals, for each interval $[l,r]$ we can increment a counter $\mathtt{Open[l]}$ and a counter $\mathtt{Close[r]}$, where $\mathtt{Open}$ and $\mathtt{Close}$ are integer vectors of length $n$. Then, the BPS representation of the suffix tree topology can be built by scanning left-to right the two arrays and, for each $i=1, \dots, n$, append $\mathtt{Open[i]}$ open parentheses followed by $\mathtt{Close[i]}$ close parentheses to the BPS representation. The main drawback of this solution is that it takes too much space: $2n\log n$ bits to store the two arrays. Belazzougui solves this problem by noticing that the sum of all the values in the two arrays is the length of the final BPS representation, that is, at most $4n$. This makes it possible to represent the arrays in just $O(n)$ bits of space by representing (the few) large counters in plain form and (the many) small counters using delta encoding (while still supporting updates in constant time).
Our goal in this section is to reduce the working space from $O(n)$ to a (small) fraction of $n\log \sigma$. A first idea could be to iterate Belazzougui's strategy on chunks of the interval $[1,n]$. Unfortunately, this does not immediately give the correct solution as a chunk could still account for up to $\Theta(n)$ parentheses, no matter what the length of the chunk is; as a result, Belazzougui's representation could still take $O(n)$ bits of space for a chunk (when using large enough chunks to keep the running time under control as seen in the previous section). We use a solution analogous to the one discussed in the previous section. This solution corresponds to the first part of Belazzougui's strategy (in particular, we will store small counters in plain form instead of using delta encoding). We divide $BWT[1,n]$ in $\lceil n/B \rceil$ blocks $BWT[(i-1)\cdot B+1,i\cdot B]$, $i=1, \dots, \lceil n/B \rceil$ of size $B$ (assume for simplicity that $B$ divides $n$). For each block $i=1, \dots, \lceil n/B \rceil$, we use Lemma \ref{lem:ST intervals} to enumerate all suffix tree intervals $[l,r]$. We keep two arrays $\mathtt{Open[1,B]}$ and $\mathtt{Close[1,B]}$ storing integers of $2\log\log n$ bits each. Whenever the beginning $l$ of a suffix tree interval $[l,r]$ falls inside the current block $[(i-1)\cdot B+1,i\cdot B]$, we increment $\mathtt{Open[l- (i-1)\cdot B]}$ (the description is analogous for index $r$ and array $\mathtt{Close}$). If $\mathtt{Open[l- (i-1)\cdot B]}$ reaches the maximum value $2^{2\log\log n-1}$, we no longer increment it. Adopting Belazzougui's terminology, we call such a bucket ``saturated''. After having generated all suffix tree intervals, let $k$ be the number of saturated counters. We allocate a vector $\mathtt{LARGE\_COUNTERS}$ storing $k$ integers of $\log n + 2$ bits each (enough to store the value $4n$, i.e. an upper-bound to the value that a counter can reach). We also allocate a bitvector of length $B$ marking saturated counters, and process it to support constant-time rank queries. This allows us to obtain in constant time the location in $\mathtt{LARGE\_COUNTERS}$ corresponding to any saturated counter in the block. We generate all suffix tree intervals for a second time using again Lemma \ref{lem:ST intervals}, this time incrementing (in $\mathtt{LARGE\_COUNTERS}$) only locations corresponding to saturated counters. Since the BPS sequence has length at most $4n$ and a counter saturates when it reaches value $\Theta(\log^2 n)$, we have that $k = O(n/\log^2n)$ and thus $\mathtt{LARGE\_COUNTERS}$ takes at most $O(n/\log n) = o(n)$ bits to be stored. The rest of the analysis is identical to the algorithm described in the previous section. For each block, we spend $O(n\log\sigma)$ time (two applications of Lemma \ref{lem:ST intervals}). We also spend $O(n/\log^2 n)$ time to allocate the instances of $\mathtt{LARGE\_COUNTERS}$ across all blocks. Overall, we spend $O((n^2/B)\log\sigma + n\log\sigma)$ time across all blocks. The space used is $o(n) + O(B\cdot \log\log n)$ bits on top of the BWT. By setting $B = (\epsilon\cdot n\log\sigma)/\log\log n$ we obtain:
\begin{lemma}\label{lem:BPS}
Given a wavelet tree for the Burrows-Wheeler transform of a text $T$ of length $n$ on alphabet $[1,\sigma]$, for any parameter $0<\epsilon \leq 1$ we can build the BPS representation of the suffix tree topology in $O(n(\log\sigma + \epsilon^{-1}\log\log n))$ time and $\epsilon \cdot n\log\sigma + o(n)$ bits of working space on top of the input BWT and the optput. \end{lemma}
To conclude, we note that our procedures can be immediately used to build space-efficiently the compressed suffix tree described by Sadakane~\cite{Sadakane:2007:CST:1326296.1326297} starting from the BWT. The only missing ingredients are (i) to augment the BWT with a suffix array sample in order to turn it into a CSA, and (ii) to pre-process the PLCP and BPS sequences to support fast queries (\emph{select} on the PLCP and navigational queries on the BPS). Step (i) can be easily performed in $O(n\log\sigma)$ time and $n+o(n)$ bits of working space with a folklore solution that iteratively applies function LF to navigate all BWT's positions and collect one suffix array sample every $O(\log^{1+\delta} n/\log\sigma)$ text positions, for any fixed $\delta>0$ (using a succinct bitvector to mark sampled positions). The resulting CSA takes $n\log\sigma + o(n\log\sigma)$ bits of space and allows computing any $SA[i]$ in $O(\log^{1+\delta} n)$ time. Step (ii) can be performed in $O(n)$ time and $o(n)$ bits of working space using textbook solutions (see \cite{Navarro:2016:CDS:3092586}). Combining this with Lemmas \ref{thm:BWT->WT}, \ref{lem:PLCP}, and \ref{lem:BPS}, we obtain:
\begin{theorem}\label{lem:BPS}
Given the word-packed BWT of a text $T$ of length $n$ on alphabet $[1,\sigma]$, for any parameter $0<\epsilon \leq 1$ we can replace it
in $O(n(\log\sigma + \epsilon^{-1}\log\log n))$ time and $\epsilon \cdot n\log\sigma + o(n)$ bits of working space
with a compressed suffix tree taking $n\log\sigma + 6n + o(n\log\sigma)$ bits of space and supporting all operations in $O(\tt{polylog}\ n)$ time. \end{theorem}
\section{Merging BWTs in Small Space}\label{sec:algo2}
In this section we use our space-efficient BWT-navigation strategies to tackle an additional problem: to merge the BWTs of two string collections. In~\cite{Belazzougui:2014:LTC:2591796.2591885,belazzougui2016linear}, Belazzougui et al. show that Theorem \ref{th:Belazzougui} can be adapted to merge the BWTs of two texts $T_1, T_2$ and obtain the BWT of the collection $\{T_1, T_2\}$ in $O(nk)$ time and $n\log\sigma(1+1/k) + 11n + o(n)$ bits of working space for any $k \geq 1$~\cite[Thm. 7]{belazzougui2016linear}. We show that our strategy enables a more space-efficient algorithm for the task of merging BWTs of collections.
The following theorem, whose proof is reported later in this section, merges two BWTs by computing the binary DA of their union. After that, the merged BWT can be streamed to external memory (the DA tells how to interleave characters from the input BWTs) and does not take additional space in internal memory. Similarly to what we did in the proof of Theorem \ref{thm:LCP collection succinct}, this time we re-use the space of the Document Array to accommodate the extra $n$ bits needed to replace the BWTs of the two collections with their wavelet matrices. This is the main result of this section:
\begin{theorem}\label{th:merge}
Given the Burrows-Wheeler Transforms of two collections $\mathcal{S}\xspace_1$ and $\mathcal{S}\xspace_2$ of total length $n$ on alphabet $[1,\sigma]$, we can compute the Document Array
of $\mathcal{S}\xspace_1 \cup \mathcal{S}\xspace_2$ in $O(n\log\sigma)$ time using $o(n\log\sigma)$ bits of working space on top of the input BWTs and the output DA. \end{theorem}
We also briefly discuss how to extend Theorem \ref{th:merge} to build the LCP array of the merged collection. In Section \ref{sec:experiments} we present an implementation of our algorithms and an experimental comparison with \texttt{eGap}\xspace~\cite{egidi2017lightweight}, the state-of-the-art tool designed for the same task of merging BWTs while inducing the LCP of their union.
The procedure of Algorithm \ref{alg:fill leaves} can be extended to merge BWTs of two collections $\mathcal{S}\xspace_1$, $\mathcal{S}\xspace_2$ using $o(n\log\sigma)$ bits of working space on top of the input BWTs and output Document Array (here, $n$ is the cumulative length of the two BWTs). The idea is to simulate a navigation of the \emph{leaves} of the generalized suffix tree of $\mathcal{S}\xspace_1 \cup \mathcal{S}\xspace_2$ (note: for us, a collection is an ordered multi-set of strings). Our procedure differs from that described in~\cite[Thm. 7]{belazzougui2016linear} in two ways. First, they navigate a subset of the suffix tree \emph{nodes} (so-called \emph{impure} nodes, i.e. the roots of subtrees containing suffixes from distinct strings), whereas we navigate leaves. Second, their visit is implemented by following Weiner links. This forces them to represent the nodes with the ``heavy'' representation $\mathtt{repr}$ of Section \ref{sec:belazzougui}, which is not efficient on large alphabets. On the contrary, leaves can be represented simply as ranges and allow for a more space-efficient queue/stack representation.
We represent each leaf by a pair of intervals, respectively on $BWT(\mathcal{S}\xspace_1)$ and $BWT(\mathcal{S}\xspace_2)$, of strings of the form $W\#$. Note that: (i) the suffix array of $\mathcal{S}\xspace_1 \cup \mathcal{S}\xspace_2$ is covered by the non-overlapping intervals of strings of the form $W\#$, and (ii) for each such string $W\#$, the interval $\mathtt{range(W\#)} = \langle L,R \rangle$ in $GSA(\mathcal{S}\xspace_1 \cup \mathcal{S}\xspace_2)$ can be partitioned as $\langle L, M \rangle \cdot \langle M+1, R\rangle$, where $\langle L,M\rangle$ contains only suffixes from $\mathcal{S}\xspace_1$ and $\langle M+1,R \rangle$ contains only suffixes from $\mathcal{S}\xspace_2$ (one of these two intervals could be empty). It follows that we can navigate in parallel the leaves of the suffix trees of $\mathcal{S}\xspace_1$ and $\mathcal{S}\xspace_2$ (using again a stack or a queue containing pairs of intervals on the two BWTs), and fill the Document Array $DA[1,n]$, an array that will tell us whether the $i$-th entry of $BWT(\mathcal{S}\xspace_1 \cup \mathcal{S}\xspace_2)$ comes from $BWT(\mathcal{S}\xspace_1)$ ($DA[i] = 0$) or $BWT(\mathcal{S}\xspace_2)$ ($DA[i] = 1$). To do this, let $\langle L_1, R_1\rangle$ and $\langle L_2, R_2\rangle$ be the ranges on the suffix arrays of $\mathcal{S}\xspace_1$ and $\mathcal{S}\xspace_2$, respectively, of a suffix $W\#$ of some string in the collections. Note that one of the two intervals could be empty: $R_j<L_j$. In this case, we still require that $L_j-1$ is the number of suffixes in $\mathcal{S}\xspace_j$ that are smaller than $W\#$. Then, in the collection $\mathcal{S}\xspace_1 \cup \mathcal{S}\xspace_2$ there are $L_1 + L_2 - 2$ suffixes smaller than $W\#$, and $R_1 + R_2$ suffixes smaller than or equal to $W\#$. It follows that the range of $W\#$ in the suffix array of $\mathcal{S}\xspace_1 \cup \mathcal{S}\xspace_2$ is $\langle L_1+L_2-1, R_1+R_2\rangle$, where the first $R_1-L_1+1$ entries correspond to suffixes of strings from $\mathcal{S}\xspace_1$. Then, we set $DA[L_1+L_2-1, L_2 + R_1-1] \leftarrow 0$ and $DA[L_2 + R_1,R_1+R_2] \leftarrow 1$. The procedure starts from the pair of intervals corresponding to the ranges of the string ``$\#$'' in the two BWTs, and proceeds recursively by left-extending the current pair of ranges $\langle L_1, R_1\rangle$, $\langle L_2, R_2\rangle$ with the symbols in $\mathtt{BWT_1.rangeDistinct(L_1,R_1)} \cup \mathtt{BWT_2.rangeDistinct(L_2,R_2)}$. The detailed procedure is reported in Algorithm \ref{alg:merge}. The leaf visit is implemented, again, using a stack or a queue; this time however, these containers are filled with pairs of intervals $\langle L_1, R_1\rangle$, $\langle L_2, R_2\rangle$. We implement the stack simply as a vector of quadruples $\langle L_1, R_1, L_2, R_2\rangle$. As far as the queue is concerned, some care needs to be taken when representing the pairs of ranges using bitvectors as seen in Section \ref{sec:beller} with Beller et al.'s representation. Recall that, at any time, the queue can be partitioned in two sub-sequences associated with LCP values $\ell$ and $\ell+1$ (we pop from the former, and push in the latter). This time, we represent each of these two subsequences as a vector of quadruples (pairs of ranges on the two BWTs) as long as the number of quadruples in the sequence does not exceed $n/\log n$. When there are more quadruples than this threshold, we switch to a bitvector representation defined as follows.
Let $|BWT(\mathcal{S}\xspace_1)|=n_1$, $|BWT(\mathcal{S}\xspace_2)|=n_2$, and $|BWT(\mathcal{S}\xspace_1\cup \mathcal{S}\xspace_2)| = n = n_1+n_2$. We keep two bitvectors $\mathtt{Open[1,n]}$ and $\mathtt{Close[1,n]}$ storing opening and closing parentheses of intervals in $BWT(\mathcal{S}\xspace_1\cup \mathcal{S}\xspace_2)$. We moreover keep two bitvectors $\mathtt{NonEmpty_1[1,n]}$ and $\mathtt{NonEmpty_2[1,n]}$ keeping track, for each $i$ such that $\mathtt{Open[i]=1}$, of whether the interval starting in $BWT(\mathcal{S}\xspace_1\cup \mathcal{S}\xspace_2)[i]$ contains suffixes of reads coming from $\mathcal{S}\xspace_1$ and $\mathcal{S}\xspace_2$, respectively. Finally, we keep four bitvectors $\mathtt{Open_j[1,n_j]}$ and $\mathtt{Close_j[1,n_j]}$, for $j=1,2$, storing non-empty intervals on $BWT(\mathcal{S}\xspace_1)$ and $BWT(\mathcal{S}\xspace_2)$, respectively. To insert a pair of intervals $\langle L_1, R_1\rangle,\ \langle L_2, R_2\rangle$ in the queue, let $\langle L,R \rangle = \langle L_1+L_2-1, R_1+R_2\rangle$. We set $\mathtt{Open[L]} \leftarrow 1$ and $\mathtt{Close[R]} \leftarrow 1$. Then, for $j=1,2$, we set $\mathtt{NonEmpty_j[L]} \leftarrow 1$, $\mathtt{Open_j[L_j]} \leftarrow 1$ and $\mathtt{Close_j[R_j]} \leftarrow 1$ if and only if $R_j\geq L_j$. This queue representation takes $O(n)$ bits. By construction, for each bit set in $\mathtt{Open}$ at position $i$, there is a corresponding bit set in $\mathtt{Open_j}$ if and only if $\mathtt{NonEmpty_j[i]} = 1$ (moreover, corresponding bits set appear in the same order in $\mathtt{Open}$ and $\mathtt{Open_j}$). It follows that a left-to-right scan of these bitvectors is sufficient to identify corresponding intervals on $BWT(\mathcal{S}\xspace_1\cup \mathcal{S}\xspace_2)$, $BWT(\mathcal{S}\xspace_1)$, and $BWT(\mathcal{S}\xspace_2)$. By packing the bits of the bitvectors in words of $\Theta(\log n)$ bits, the $t$ pairs of intervals contained in the queue can be extracted in $O(t+ n/\log n)$ time (as described in~\cite{beller2013computing}) by scanning in parallel the bitvectors forming the queue. Particular care needs to be taken only when we find the beginning of an interval $\mathtt{Open[L]=1}$ with $\mathtt{NonEmpty_1[L]} = 0$ (the case $\mathtt{NonEmpty_2[L]} = 0$ is symmetric). Let $L_2$ be the beginning of the corresponding non-empty interval on $BWT(\mathcal{S}\xspace_2)$. Even though we are not storing $L_1$ (because we only store nonempty intervals), we can retrieve this value as $L_1=L-L_2+1$. Then, the empty interval on $BWT(\mathcal{S}\xspace_1)$ is $\langle L_1, L_1-1\rangle$.
The same arguments used in the previous section show that the algorithm runs in $O(n\log\sigma)$ time and uses $o(n\log\sigma)$ bits of space on top of the input BWTs and output Document Array. This proves Theorem \ref{th:merge}.
To conclude, we note that the algorithm can be easily extended to compute the LCP array of the merged collection while merging the BWTs. This requires adapting Algorithm \ref{alg:fill nodes} to work on pairs of suffix tree nodes (as we did in Algorithm \ref{alg:merge} with pairs of leaves). Results on an implementation of the extended algorithm are discussed in the next section. From the practical point of view, note that it is more advantageous to induce the LCP of the merged collection while merging the BWTs (rather than first merging and then inducing the LCP using the algorithm of the previous section), since leaf-type LCP values can be induced directly while computing the document array.
\begin{algorithm}
\begin{algorithmic}[1]
\If{$\sigma > n/\log^3n$}
\State $\mathtt P \leftarrow \mathtt{new\_queue()}$\Comment{Initialize new queue of interval pairs}\label{new queue3}
\Else
\State $\mathtt P \leftarrow \mathtt{new\_stack()}$\Comment{Initialize new stack of interval pairs}\label{new stack3}
\EndIf
\State $\mathtt P\mathtt{.push(BWT_1.range(\#),BWT_2.range(\#))}$\Comment{Push SA-ranges of terminator}\label{push5}
\While{$\mathtt{\mathbf{not}\ P.empty()}$}\label{while3}
\State $\langle L_1,R_1, L_2, R_2 \rangle \leftarrow \mathtt{P.pop()}$\Comment{Pop highest-priority element}\label{pop3}
\For{$i=L_1+L_2-1\dots L_2+R_1-1$}
\State $\mathtt{DA}[i] \leftarrow 0$\Comment{Suffixes from $\mathcal{S}\xspace_1$}\label{DA1}
\EndFor
\For{$i=L_2+R_1\dots R_1+R_2$}
\State $\mathtt{DA}[i] \leftarrow 1$\Comment{Suffixes from $\mathcal{S}\xspace_2$}\label{DA2}
\EndFor
\If{$\sigma > n/\log^3n$}
\State$\mathtt{P.push(getIntervals(L_1, R_1, L_2, R_2, BWT_1, BWT_2))}$\Comment{New intervals}\label{push8}
\Else
\State $c_1^1, \dots, c_{k_1}^1 \leftarrow \mathtt{BWT_1.rangeDistinct(L_1,R_1)}$\label{range distinct3}
\State $c_1^2, \dots, c_{k_2}^2 \leftarrow \mathtt{BWT_2.rangeDistinct(L_2,R_2)}$\label{range distinct4}
\State $\{c_1\dots c_k\} \leftarrow \{c_1^1, \dots, c_{k_1}^1\} \cup \{c_1^2, \dots, c_{k_2}^2\}$\label{range distinct5}
\For{$i=1\dots k$}
\State $\langle L_1^i, R_1^i\rangle \leftarrow \mathtt{BWT_1.bwsearch(\langle L_1, R_1\rangle, c_i)}$\Comment{Backward search step}\label{BWS2}
\EndFor
\For{$i=1\dots k$}
\State $\langle L_2^i, R_2^i\rangle \leftarrow \mathtt{BWT_2.bwsearch(\langle L_2, R_2\rangle, c_i)}$\Comment{Backward search step}\label{BWS3}
\EndFor
\State $\langle \hat L_1^i, \hat R_1^i, \hat L_2^i, \hat R_2^i, \rangle_{i=1, \dots, k} \leftarrow \mathtt{sort}(\langle L_1^i, R_1^i, L_2^i, R_2^i, \rangle_{i=1, \dots, k})$\label{sort3}
\For{$i=k\dots 1$}
\State $\mathtt{P.push}(\hat L_1^i, \hat R_1^i, \hat L_2^i, \hat R_2^i)$\Comment{Push in order of decreasing length}\label{push6}
\EndFor
\EndIf
\EndWhile
\caption{$\mathtt{Merge(BWT_1,BWT_2, DA)}$}\label{alg:merge} \end{algorithmic} \end{algorithm}
Note that Algorithm \ref{alg:merge} is similar to Algorithm \ref{alg:fill leaves}, except that now we manipulate pairs of intervals. In Line \ref{sort3}, we sort quadruples according to the length $R_1^i + R_2^i - (L_1^i + L_2^i) +2$ of the combined interval on $BWT(\mathcal{S}\xspace_1\cup \mathcal{S}\xspace_2)$. Finally, note that Backward search can be performed correctly also when the input interval is empty: $\mathtt{BWT_j.bwsearch(\langle L_j, L_j-1 \rangle, c)}$, where $L_j-1$ is the number of suffixes in $\mathcal{S}\xspace_j$ smaller than some string $W$, correctly returns the pair $\langle L', R'\rangle$ such that $L'$ is the number of suffixes in $\mathcal{S}\xspace_j$ smaller than $cW$: this is true when implementing backward search with a $rank_c$ operation on position $L_j$; then, if the original interval is empty we just set $R'=L'-1$ to keep the invariant that $R'-L'+1$ is the interval's length.
\section{Implementation and Experimental Evaluation}\label{sec:experiments}
We implemented our LCP construction and BWT merge algorithms on DNA alphabet in \url{https://github.com/nicolaprezza/bwt2lcp}\xspace using the language C++. Due to the small alphabet size, it was actually sufficient to implement our extension of Belazzougui's enumeration algorithm (and not the strategy of Beller et al., which becomes competitive only on large alphabets). The repository features a new packed string on DNA alphabet $\Sigma_{DNA}=\{A,C,G,T,\#\}$ using 4 bits per character and able to compute the quintuple $\langle BWT.rank_c(i) \rangle_{i\in \Sigma_{DNA}}$ with just one cache miss. This is crucial for our algorithms, since at each step we need to left-extend ranges by all characters. This structure divides the text in blocks of 128 characters. Each block is stored using 512 cache-aligned bits (the typical size of a cache line), divided as follows. The first 128 bits store four 32-bits counters with the partial ranks of A, C, G, and T before the block (if the string is longer than $2^{32}$ characters, we further break it into superblocks of $2^{32}$ characters; on reasonably-large inputs, the extra rank table fits in cache and does not cause additional cache misses). The following three blocks of 128 bits store the first, second, and third bits, respectively, of the characters' binary encodings (each character is packed in 3 bits). Using this layout, the rank of each character in the block can be computed with at most three masks, a bitwise AND (actually less, since we always compute the rank of all five characters and we re-use partial results whenever possible), and a \texttt{popcount} operation. We also implemented a packed string on the augmented alphabet $\Sigma_{DNA}^+=\{A,C,G,N,T,\#\}$ using $4.38$ bits per character and offering the same cache-efficiency guarantees. In this case, a 512-bits block stores 117 characters, packed as follows. As seen above, the first 128 bits store four 32-bits counters with the partial ranks of A, C, G, and T before the block. Each of the following three blocks of 128 bits is divided in a first part of 117 bits and a second part of 11 bits. The first parts store the first, second, and third bits, respectively, of the characters' binary encodings. The three parts of 11 bits, concatenated together, store the rank of N's before the block. This layout minimizes the number of bitwise operations (in particular, shifts and masks) needed to compute a parallel rank.
Several heuristics have been implemented to reduce the number of cache misses in practice. In particular, we note that in Algorithm \ref{alg:fill leaves} we can avoid backtracking when the range size becomes equal to one; the same optimization can be implemented in Algorithm \ref{alg:merge} when also computing the LCP array, since leaves of size one can be identified during navigation of internal suffix tree nodes. Overall, we observed (using a memory profiler) that in practice the combination of Algorithms \ref{alg:fill nodes}-\ref{alg:fill leaves} generates at most $1.5n$ cache misses, $n$ being the total collection's size. The extension of Algorithm \ref{alg:merge} that computes also LCP values generates twice this number of cache misses (this is expected, since the algorithm navigates two BWTs).
We now report some preliminary experiments on our algorithms: \texttt{bwt2lcp}\xspace (Algorithms \ref{alg:fill nodes}-\ref{alg:fill leaves}) and \texttt{merge}\xspace (Algorithm \ref{alg:merge}, extended to compute also the LCP array). All tests were done on a DELL PowerEdge R630 machine, used in non exclusive mode. Our platform is a $24$-core machine with Intel(R) Xeon(R) CPU E5-2620 v3 at $2.40$ GHz, with $128$ GiB of shared memory and 1TB of SSD. The system is Ubuntu 14.04.2 LTS. The code was compiled using gcc 8.1.0 with flags \texttt{-Ofast} \texttt{-fstrict-aliasing}.
\begin{table}[t]
\centering
{\scriptsize
\begin{tabular}{|@{\ }l@{\ }|@{\ }c@{\ }|@{\ }c@{\ }|@{\ }c@{\ }|@{\ }c@{\ }|c@{\ }|c@{\ }|}
\hline
Name & Size & $\sigma$ & N. of & Max read & Bytes for\\
& GiB & & reads & length & lcp values\\
\hline
NA12891.8 & 8.16 & 5 & 85,899,345 & 100 & 1 \\
\hline
shortreads & 8.0 & 6 & 85,899,345 & 100 & 1 \\
\hline
pacbio & 8.0 & 6 & 942,248 & 71,561 & 4 \\
\hline
pacbio.1000 & 8.0 & 6 & 8,589,934 & 1000 & 2\\
\hline
NA12891.24 & 23.75 & 6 & 250,000,000 & 100 & 1 \\
\hline
NA12878.24 & 23.75 & 6 & 250,000,000 & 100 & 1 \\
\hline
\end{tabular}
}
\caption{Datasets used in our experiments. Size accounts only for the alphabet's characters. The alphabet's size $\sigma$ includes the terminator.}
\label{tableDataset} \end{table}
\begin{table}[t]
\centering
{\scriptsize
\begin{tabular}{|c|c|c|c|c|c|c|c|c|}
\hline
& \multicolumn{2}{c|}{Preprocessing} & \multicolumn{2}{c|}{\texttt{eGap}\xspace} & \multicolumn{2}{c|}{\texttt{merge}\xspace} \\
\hline
Name & Wall Clock & RAM & Wall Clock & RAM & Wall Clock & RAM \\
& (h:mm:ss) & (GiB) & (h:mm:ss) & (GiB) & (h:mm:ss) & (GiB) \\
\hline
NA12891.8 & 1:15:57 & 2.84 & \multirow{2}{*}{10:15:07} & \multirow{2}{*}{18.09 (-m 32000)} & \multirow{2}{*}{3:16:40} & \multirow{2}{*}{26.52} \\
\cline{1-3}
NA12891.8.RC & 1:17:55 & 2.84 & & & & \\
\hline
shortreads & 1:14:51 & 2.84 & \multirow{2}{*}{11:03:10} & \multirow{2}{*}{16.24 (-m 29000)} & \multirow{2}{*}{3:36:21} & \multirow{2}{*}{26.75} \\
\cline{1-3}
shortreads.RC & 1:19:30 & 2.84 & & & & \\
\hline
pacbio.1000 & 2:08:56 & 31.28 & \multirow{2}{*}{5:03:01} & \multirow{2}{*}{21.23 (-m 45000)} & \multirow{2}{*}{4:03:07} & \multirow{2}{*}{42.75} \\
\cline{1-3}
pacbio.1000.RC & 2:15:08 & 31.28 & & & & \\
\hline
pacbio & 2:27:08 & 31.25 & \multirow{2}{*}{2:56:31} & \multirow{2}{*}{33.40 (-m 80000)} & \multirow{2}{*}{4:38:27} & \multirow{2}{*}{74.76} \\
\cline{1-3}
pacbio.RC & 2:19:27 & 31.25 & & & & \\
\hline
NA12878.24 & 4:24:27 & 7.69 & \multirow{2}{*}{31:12:28} & \multirow{2}{*}{47.50 (-m 84000)} & \multirow{2}{*}{6:41:35} & \multirow{2}{*}{73.48} \\
\cline{1-3}
NA12891.24 & 4:02:42 & 7.69 & & & & \\
\hline
\end{tabular}
\caption{In this experiment, we merge pairs of BWTs and induce the LCP of their union using \texttt{eGap}\xspace and \texttt{merge}\xspace. We also show the resources used by the pre-processing step (building the BWTs) for comparison. Wall clock is the elapsed time from start to completion of the instance, while RAM (in GiB) is the peak Resident Set Size (RSS). All values were taken using the \texttt{/usr/bin/time} command. During the preprocessing step on the collections pacBio.1000 and pacBio, the available memory in MB (parameter m) of \texttt{eGap}\xspace was set to 32000 MB. In the merge step this parameter was set to about to the memory used by \texttt{merge}\xspace.
\texttt{eGap}\xspace and \texttt{merge}\xspace take as input the same BWT file.}
\label{tab:merge}
} \end{table}
Table \ref{tableDataset} summarizes the datasets used in our experiments. ``NA12891.8''\footnote{{\scriptsize \url{ftp://ftp.1000genomes.ebi.ac.uk/vol1/ftp/phase3/data/NA12891/sequence_read/SRR622458_1.filt.fastq.gz}}} contains Human DNA reads on the alphabet $\Sigma_{DNA}$
where we have removed reads containing the nucleotide $N$. ``shortreads'' contains Human DNA short reads on the extended alphabet $\Sigma_{DNA}^+$.
``pacbio'' contains PacBio RS II reads from the species \emph{Triticum aestivum} (wheat).
``pacbio.1000'' are the strings from ``pacbio'' trimmed to length 1,000. All the above datasets except the first have been download from \url{https://github.com/felipelouza/egap/tree/master/dataset}. To conclude, we added two collections, ``NA12891.24'' and ``NA12878.24'' obtained by taking the first $250,000,000$ reads from individuals NA12878\footnote{{\scriptsize \url{ftp://ftp.1000genomes.ebi.ac.uk/vol1/ftp/phase3/data/NA12878/sequence_read/SRR622457_1.filt.fastq.gz}}} and NA12891. All datasets except ``NA12891.8'' are on the alphabet $\Sigma_{DNA}^+$. In Tables \ref{tab:merge} and \ref{tab:induce}, the suffix ``.RC'' added to a dataset's name indicates the reverse-complemented dataset.
We compare our algorithms with \texttt{eGap}\xspace\footnote{{\scriptsize\url{https://github.com/felipelouza/egap}}} and BCR\footnote{{\scriptsize\url{https://github.com/giovannarosone/BCR_LCP_GSA}}}, two tools designed to build the BWT and LCP of a set of DNA reads. Since no tools for inducing the LCP from the BWT of a set of strings are available in the literature, in Table \ref{tab:induce} we simply compare the resources used by \texttt{bwt2lcp}\xspace with the time and space requirements of \texttt{eGap}\xspace and BCR when building the BWT. In \cite{EgidiAMB2019}, experimental results show that BCR works better on short reads and collections with a large average LCP, while \texttt{eGap}\xspace works better when the datasets contain long reads and relatively small average LCP. For this reason, in the preprocessing step we have used BCR for the collections containing short reads and \texttt{eGap}\xspace for the other collections. \texttt{eGap}\xspace, in addition, is capable of merging two or more BWTs while inducing the LCP of their union. In this case, we can therefore directly compare the performance of \texttt{eGap}\xspace with our tool \texttt{merge}\xspace; results are reported in Table \ref{tab:merge}. Since the available RAM is greater than the size of the input, we have used the semi-external strategy of \texttt{eGap}\xspace.
Notice that an entirely like-for-like comparison between our tools and \texttt{eGap}\xspace is not completely feasible, being \texttt{eGap}\xspace a semi-external memory tool (our tools, instead, use internal memory only). While in our tables we report RAM usage only, it is worth to notice that \texttt{eGap}\xspace uses a considerable amount of disk working space. For example, the tool uses $56$GiB of disk working space when run on a $8$GiB input (in general, the disk usage is of $7n$ bytes).
\begin{table}[t]
{\scriptsize
\centering
\begin{tabular}{|c|c|c|c|c|c|c|}
\hline
& \multicolumn{2}{c|}{Preprocessing} & \multicolumn{2}{c|}{\texttt{bwt2lcp}\xspace} \\
\hline
Name & Wall Clock & RAM & Wall Clock & RAM \\
& (h:mm:ss) & GiB & (h:mm:ss) & (GiB) \\
\hline
NA12891.8 $\cup$ NA12891.8.RC (BCR) & 2:43:02 & 5.67 & 1:40:01 & 24.48 \\
\hline
shortread $\cup$ shortread.RC (BCR) & 2:47:07 & 5.67 & 2:14:41 & 24.75 \\
\hline
pacbio.1000 $\cup$ pacbio.1000.RC (\texttt{eGap}\xspace -m 32000) & 7:07:46 & 31.28 & 1:54:56 & 40.75 \\
\hline
pacbio $\cup$ pacbio.RC (\texttt{eGap}\xspace -m 80000) & 6:02:37 & 78.125 & 2:14:37 & 72.76 \\
\hline
NA12878.24 $\cup$ NA12891.24 (BCR) & 8:26:34 & 16.63 & 6:41:35 & 73.48 \\
\hline
\end{tabular}
\caption{In this experiment, we induced the LCP array from the BWT of a collection (each collection is the union of two collections from Table \ref{tab:merge}). We also show pre-processing requirements (i.e. building the BWT) of the better performing tool between BCR and \texttt{eGap}\xspace.}
\label{tab:induce}
} \end{table}
Our tools exhibit a dataset-independent linear time complexity, whereas \texttt{eGap}\xspace's running time depends on the average LCP. Table \ref{tab:induce} shows that our tool \texttt{bwt2lcp}\xspace induces the LCP from the BWT faster than building the BWT itself. When 'N's are not present in the dataset, \texttt{bwt2lcp}\xspace processes data at a rate of $2.92$ megabases per second and uses $0.5$ Bytes per base in RAM in addition to the LCP. When 'N's are present, the throughput decreases to $2.12$ megabases per second and the tool uses $0.55$ Bytes per base in addition to the LCP. As shown in Table \ref{tab:merge}, our tool \texttt{merge}\xspace is from $1.25$ to $4.5$ times faster than \texttt{eGap}\xspace on inputs with large average LCP, but $1.6$ times slower when the average LCP is small (dataset ``pacbio''). When 'N's are not present in the dataset, \texttt{merge}\xspace processes data at a rate of $1.48$ megabases per second and uses $0.625$ Bytes per base in addition to the LCP. When 'N's are present, the throughput ranges from $1.03$ to $1.32$ megabases per second and the tool uses $0.673$ Bytes per base in addition to the LCP. When only computing the merged BWT (results not shown here for space reasons), \texttt{merge}\xspace uses in total $0.625$/$0.673$ Bytes per base in RAM (without/with 'N's) and is about $1.2$ times faster than the version computing also the LCP.
\end{document}
|
arXiv
|
{
"id": "1908.04686.tex",
"language_detection_score": 0.7435610294342041,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[GRADED ANNIHILATORS AND TIGHT CLOSURE]{GRADED ANNIHILATORS OF MODULES OVER THE FROBENIUS SKEW POLYNOMIAL RING, AND TIGHT CLOSURE} \author{RODNEY Y. SHARP} \address{Department of Pure Mathematics, University of Sheffield, Hicks Building, Sheffield S3 7RH, United Kingdom\\ {\it Fax number}: 0044-114-222-3769} \email{[email protected]}
\thanks{The author was partially supported by the Engineering and Physical Sciences Research Council of the United Kingdom (Overseas Travel Grant Number EP/C538803/1).}
\subjclass[2000]{Primary 13A35, 16S36, 13D45, 13E05, 13E10; Secondary 13H10}
\date{\today}
\keywords{Commutative Noetherian ring, prime characteristic, Frobenius homomorphism, tight closure, (weak) test element, (weak) parameter test element, skew polynomial ring; local cohomology; Cohen--Macaulay local ring.}
\begin{abstract} This paper is concerned with the tight closure of an ideal ${\mathfrak{a}}$ in a commutative Noetherian local ring $R$ of prime characteristic $p$. Several authors, including R. Fedder, K.-i. Watanabe, K. E. Smith, N. Hara and F. Enescu, have used the natural Frobenius action on the top local cohomology module of such an $R$ to good effect in the study of tight closure, and this paper uses that device. The main part of the paper develops a theory of what are here called `special annihilator submodules' of a left module over the Frobenius skew polynomial ring associated to $R$; this theory is then applied in the later sections of the paper to the top local cohomology module of $R$ and used to show that, if $R$ is Cohen--Macaulay, then it must have a weak parameter test element, even if it is not excellent. \end{abstract}
\maketitle
\setcounter{section}{-1} \section{\sc Introduction} \label{in}
Throughout the paper, $R$ will denote a commutative Noetherian ring of prime characteristic $p$. We shall always denote by $f:R\longrightarrow R$ the Frobenius homomorphism, for which $f(r) = r^p$ for all $r \in R$. Let ${\mathfrak{a}}$ be an ideal of $R$. The {\em $n$-th Frobenius power\/} ${\mathfrak{a}}^{[p^n]}$ of ${\mathfrak{a}}$ is the ideal of $R$ generated by all $p^n$-th powers of elements of ${\mathfrak{a}}$.
We use $R^{\circ}$ to denote the complement in $R$ of the union of the minimal prime ideals of $R$. An element $r \in R$ belongs to the {\em tight closure ${\mathfrak{a}}^*$ of ${\mathfrak{a}}$\/} if and only if there exists $c \in R^{\circ}$ such that $cr^{p^n} \in {\mathfrak{a}}^{[p^n]}$ for all $n \gg 0$. We say that ${\mathfrak{a}}$ is {\em tightly closed\/} precisely when ${\mathfrak{a}}^* = {\mathfrak{a}}$. The theory of tight closure was invented by M. Hochster and C. Huneke \cite{HocHun90}, and many applications have been found for the theory: see \cite{Hunek96} and \cite{Hunek98}, for example.
In the case when $R$ is local, several authors have used, as an aid to the study of tight closure, the natural Frobenius action on the top local cohomology module of $R$: see, for example, R. Fedder \cite{F87}, Fedder and K.-i. Watanabe \cite{FW87}, K. E. Smith \cite{S94}, N. Hara and Watanabe \cite{HW96} and F. Enescu \cite{Enesc03}. This device is employed in this paper. The natural Frobenius action provides the top local cohomology module of $R$ with a natural structure as a left module over the skew polynomial ring $R[x,f]$ associated to $R$ and $f$. Sections \ref{nt} and \ref{ga} develop a theory of what are here called `special annihilator submodules' of a left $R[x,f]$-module $H$. To explain this concept, we need the definition of the {\em graded annihilator\/} $\grann_{R[x,f]}H$ of $H$. Now $R[x,f]$ has a natural structure as a graded ring, and $\grann_{R[x,f]}H$ is defined to be the largest graded two-sided ideal of $R[x,f]$ that annihilates $H$. On the other hand, for a graded two-sided ideal ${\mathfrak{B}}$ of $R[x,f]$, the {\em annihilator of ${\mathfrak{B}}$ in $H$\/} is defined as $$ \ann_H{\mathfrak{B}} := \{ h \in H : \theta h = 0 \mbox{~for all~}\theta \in {\mathfrak{B}}\}. $$ I say that an $R[x,f]$-submodule of $H$ is a {\em special annihilator submodule\/} of $H$ if it has the form $\ann_H{\mathfrak{B}}$ for some graded two-sided ideal ${\mathfrak{B}}$ of $R[x,f]$.
There is a natural bijective inclusion-reversing correspondence between the set of all special annihilator submodules of $H$ and the set of all graded annihilators of submodules of $H$. A large part of this paper is concerned with exploration and exploitation of this correspondence. It is particularly satisfactory in the case where the left $R[x,f]$-module $H$ is $x$-torsion-free, for then it turns out that the set of all graded annihilators of submodules of $H$ is in bijective correspondence with a certain set of radical ideals of $R$, and one of the main results of \S \ref{ga} is that this set is finite in the case where $H$ is Artinian as an $R$-module. The theory that emerges has some uncanny similarities to tight closure theory. Use is made of the Hartshorne--Speiser--Lyubeznik Theorem (see R. Hartshorne and R. Speiser \cite[Proposition 1.11]{HarSpe77}, G. Lyubeznik \cite[Proposition 4.4]{Lyube97}, and M. Katzman and R. Y. Sharp \cite[1.4 and 1.5]{KS}) to pass between a general left $R[x,f]$-module that is Artinian over $R$ and one that is $x$-torsion-free.
In \S \ref{tc}, this theory of special annihilator submodules is applied to prove an existence theorem for weak parameter test elements in a Cohen--Macaulay local ring of characteristic $p$. To explain this, I now review some definitions concerning weak test elements.
A {\em $p^{w_0}$-weak test element\/} for $R$ (where $w_0$ is a non-negative integer) is an element $c' \in R^{\circ}$ such that, for every ideal ${\mathfrak{b}}$ of $R$ and for $r \in R$, it is the case that $r \in {\mathfrak{b}}^*$ if and only if $c'r^{p^n} \in {\mathfrak{b}}^{[p^n]}$ for all $n \geq w_0$. A $p^0$-weak test element is called a {\em test element\/}.
A proper ideal ${\mathfrak{a}}$ in $R$ is said to be a {\em parameter ideal\/} precisely when it can be generated by $\height {\mathfrak{a}}$ elements. Parameter ideals play an important r\^ole in tight closure theory, and Hochster and Huneke introduced the concept of parameter test element for $R$. A {\em $p^{w_0}$-weak parameter test element\/} for $R$ is an element $c' \in R^{\circ}$ such that, for every parameter ideal ${\mathfrak{b}}$ of $R$ and for $r \in R$, it is the case that $r \in {\mathfrak{b}}^*$ if and only if $c'r^{p^n} \in {\mathfrak{b}}^{[p^n]}$ for all $n \geq w_0$. A $p^0$-weak parameter test element is called a {\em parameter test element\/}.
It is a result of Hochster and Huneke \cite[Theorem (6.1)(b)]{HocHun94} that an algebra of finite type over an excellent local ring of characteristic $p$ has a $p^{w_0}$-weak test element for some non-negative integer $w_0$; furthermore, such an algebra which is reduced actually has a test element. Of course, a (weak) test element is a (weak) parameter test element.
One of the main results of this paper is Theorem \ref{tc.2}, which shows that every Cohen--Macaulay local ring of characteristic $p$, even if it is not excellent, has a $p^{w_0}$-weak parameter test element for some non-negative integer $w_0$.
Lastly, the final \S \ref{en} establishes some connections between the theory developed in this paper and the $F$-stable primes of F. Enescu \cite{Enesc03}.
\section{\sc Graded annihilators and related concepts} \label{nt}
\begin{ntn} \label{nt.1} Throughout, $R$ will denote a commutative Noetherian ring of prime characteristic $p$. We shall work with the
skew polynomial ring $R[x,f]$ associated to $R$ and $f$ in the indeterminate $x$ over $R$. Recall that $R[x,f]$ is, as a left $R$-module, freely generated by $(x^i)_{i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ (I use $\mathbb N$ and $\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ to denote the set of positive integers and the set of non-negative integers, respectively),
and so consists
of all polynomials $\sum_{i = 0}^n r_i x^i$, where $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$
and $r_0,\ldots,r_n \in R$; however, its multiplication is subject to the
rule
$$
xr = f(r)x = r^px \quad \mbox{~for all~} r \in R\/.
$$ Note that $R[x,f]$ can be considered as a positively-graded ring $R[x,f] = \bigoplus_{n=0}^{\infty} R[x,f]_n$, with $R[x,f]_n = Rx^n$ for all $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. The ring $R[x,f]$ will be referred to as the {\em Frobenius skew polynomial ring over $R$.}
Throughout, we shall let $G$ and $H$ denote left $R[x,f]$-modules. The {\em annihilator of $H$\/} will be denoted by $\ann_{R[x,f]}H$ or $\ann_{R[x,f]}(H)$. Thus $$ \ann_{R[x,f]}(H) = \{ \theta \in R[x,f] : \theta h = 0 \mbox{~for all~} h \in H\}, $$ and this is a (two-sided) ideal of $R[x,f]$. For a two-sided ideal ${\mathfrak{B}}$ of $R[x,f]$, we shall use $\ann_H{\mathfrak{B}}$ or $\ann_H({\mathfrak{B}})$ to denote the {\em annihilator of ${\mathfrak{B}}$ in $H$}. Thus $$ \ann_H{\mathfrak{B}} = \ann_H({\mathfrak{B}}) = \{ h \in H : \theta h = 0 \mbox{~for all~}\theta \in {\mathfrak{B}}\}, $$ and this is an $R[x,f]$-submodule of $H$. \end{ntn}
\begin{defrmks} \label{nt.2} We say that the left $R[x,f]$-module $H$ is {\em $x$-torsion-free\/} if $xh = 0$, for $h \in H$, only when $h = 0$. The set $\Gamma_x(H) := \left\{ h \in H : x^jh = 0 \mbox{~for some~} j \in \mathbb N \right\}$ is an $R[x,f]$-submodule of $H$, called the\/ {\em $x$-torsion submodule} of $H$. The $R[x,f]$-module $H/\Gamma_x(H)$ is $x$-torsion-free. \end{defrmks}
\begin{rmk} \label{nt.2b} Let ${\mathfrak{B}}$ be a subset of $R[x,f]$. It is easy to see that ${\mathfrak{B}}$ is a graded two-sided ideal of $R[x,f]$ if and only if there is an ascending chain $({\mathfrak{b}}_n)_{n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ of ideals of $R$ (which must, of course, be eventually stationary) such that ${\mathfrak{B}} = \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{b}}_n x^n$. We shall sometimes denote the ultimate constant value of the ascending sequence $({\mathfrak{b}}_n)_{n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ by $\lim_{n \rightarrow \infty}{\mathfrak{b}}_n$.
Note that, in particular, if ${\mathfrak{b}}$ is an ideal of $R$, then ${\mathfrak{b}} R[x,f] = \bigoplus_{n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{b}} x^n$ is a graded two-sided ideal of $R[x,f]$. It was noted in \ref{nt.1} that the annihilator of a left $R[x,f]$-module is a two-sided ideal. \end{rmk}
\begin{lem} [Y. Yoshino {\cite[Corollary (2.7)]{Yoshi94}}] \label{nt.2c} The ring $R[x,f]$ satisfies the ascending chain condition on graded two-sided ideals. \end{lem}
\begin{proof} This can be proved by the argument in Yoshino's proof of \cite[Corollary (2.7)]{Yoshi94}. \end{proof}
\begin{defs} \label{nt.2d} We define the {\em graded annihilator\/} $\grann_{R[x,f]}H$ of the left $R[x,f]$-module $H$ by $$ \grann_{R[x,f]}H = \left\{ \sum_{i=0}^n r_ix^i \in R[x,f] : n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi, \mbox{~and~} r_i \in R,\, r_ix^i \in \ann_{R[x,f]}H \mbox{~for all~} i = 0, \ldots, n\right\}. $$ Thus $\grann_{R[x,f]}H$ is the largest graded two-sided ideal of $R[x,f]$ contained in $\ann_{R[x,f]}H$; also, if we write $\grann_{R[x,f]}H = \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{b}}_n x^n$ for a suitable ascending chain $({\mathfrak{b}}_n)_{n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ of ideals of $R$, then ${\mathfrak{b}}_0 = (0:_RH)$, the annihilator of $H$ as an $R$-module.
We say that an $R[x,f]$-submodule of $H$ is a {\em special annihilator submodule of $H$\/} if it has the form $\ann_H({\mathfrak{B}})$ for some {\em graded\/} two-sided ideal ${\mathfrak{B}}$ of $R[x,f]$. We shall use $\mathcal{A}(H)$ to denote the set of special annihilator submodules of $H$. \end{defs}
\begin{defrmks} \label{nt.05d} There are some circumstances in which $\grann_{R[x,f]}H = \ann_{R[x,f]}H$: for example, this would be the case if $H$ was a $\mathbb Z$-graded left $R[x,f]$-module. Work of Y. Yoshino in \cite[\S 2]{Yoshi94} provides us with further examples.
Following Yoshino \cite[Definition (2.1)]{Yoshi94}, we say that $R$ {\em has sufficiently many units\/} precisely when, for each $n \in \mathbb N$, there exists $r_n \in R$ such that all $n$ elements $(r_n)^{p^i} - r_n~(1 \leq i \leq n)$ are units of $R$. Yoshino proved in \cite[Lemma (2.2)]{Yoshi94} that if either $R$ contains an infinite field, or $R$ is local and has infinite residue field, then $R$ has sufficiently many units. He went on to show in \cite[Theorem (2.6)]{Yoshi94} that, if $R$ has sufficiently many units, then each two-sided ideal of $R[x,f]$ is graded.
Thus if $R$ has sufficiently many units, then $\grann_{R[x,f]}H = \ann_{R[x,f]}H$, even if $H$ is not graded. \end{defrmks}
\begin{lem} \label{nt.3} Let ${\mathfrak{B}}$ and ${\mathfrak{B}} '$ be graded two-sided ideals of $R[x,f]$ and let $N$ and $N'$ be $R[x,f]$-submodules of the left $R[x,f]$-module $H$.
\begin{enumerate} \item If ${\mathfrak{B}} \subseteq {\mathfrak{B}} '$, then $\ann_H({\mathfrak{B}}) \supseteq \ann_H({\mathfrak{B}} ')$. \item If $N \subseteq N'$, then $\grann_{R[x,f]}N \supseteq \grann_{R[x,f]}N '$. \item We have ${\mathfrak{B}} \subseteq \grann_{R[x,f]}\left(\ann_H({\mathfrak{B}})\right)$. \item We have $N \subseteq \ann_H\!\left(\grann_{R[x,f]}N\right).$ \item There is an order-reversing bijection, $\Gamma\/,$ from the set $\mathcal{A}(H)$ of special annihilator submo\-d\-ules of $H$ to the set of graded annihilators of submodules of $H$ given by $$ \Gamma : N \longmapsto \grann_{R[x,f]}N. $$ The inverse bijection, $\Gamma^{-1},$ also order-reversing, is given by
$$ \Gamma^{-1} : {\mathfrak{B}} \longmapsto \ann_H({\mathfrak{B}}). $$ \end{enumerate} \end{lem}
\begin{proof} Parts (i), (ii), (iii) and (iv) are obvious.
(v) Application of part (i) to the inclusion in part (iii) yields that $$ \ann_H({\mathfrak{B}}) \supseteq \ann_H\!\left( \grann_{R[x,f]}\left(\ann_H({\mathfrak{B}})\right) \right)\mbox{;} $$ however, part (iv) applied to the $R[x,f]$-submodule $\ann_H({\mathfrak{B}})$ of $H$ yields that $$ \ann_H({\mathfrak{B}}) \subseteq \ann_H\!\left( \grann_{R[x,f]}\left(\ann_H({\mathfrak{B}})\right) \right)\mbox{;} $$ hence $ \ann_H({\mathfrak{B}}) = \ann_H\!\left( \grann_{R[x,f]}\left(\ann_H({\mathfrak{B}})\right) \right). $ Similar considerations show that $$ \grann_{R[x,f]}N = \grann_{R[x,f]}\left(\ann_H\!\left(\grann_{R[x,f]}N\right)\right). $$ \end{proof}
\begin{rmk} \label{nt.4} It follows from Lemma \ref{nt.3} that, if $N$ is a special annihilator submodule of $H$, then it is the annihilator (in $H$) of its own graded annihilator. Likewise, a graded two-sided ideal ${\mathfrak{B}}$ of $R[x,f]$ which is the graded annihilator of some $R[x,f]$-submodule of $H$ must be the graded annihilator of $\ann_H({\mathfrak{B}})$. \end{rmk}
Much use will be made of the following lemma.
\begin{lem} \label{nt.5} Assume that the left $R[x,f]$-module $G$ is $x$-torsion-free.
Then there is a radical ideal ${\mathfrak{b}}$ of $R$ such that $\grann_{R[x,f]}G = {\mathfrak{b}} R[x,f] = \bigoplus _{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{b}} x^n$. \end{lem}
\begin{proof} There is a family $({\mathfrak{b}}_n)_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ of ideals of $R$ such that ${\mathfrak{b}}_n \subseteq {\mathfrak{b}}_{n+1}$ for all $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $\grann_{R[x,f]}G = \bigoplus _{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{b}}_n x^n$. There exists $n_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ such that ${\mathfrak{b}}_n = {\mathfrak{b}}_{n_0}$ for all $n \geq n_0$. Set ${\mathfrak{b}} := {\mathfrak{b}}_{n_0}$. It is enough for us to show that, if $r \in R$ and $e \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ are such that $r^{p^e} \in {\mathfrak{b}}$, then $r \in {\mathfrak{b}}_0$.
To this end, let $h \in \mathbb N$ be such that $h \geq \max \{e,n_0\}$. Then, for all $g \in G$, we have $x^hrg = r^{p^h}x^hg = 0$, since $r^{p^h} \in {\mathfrak{b}} = {\mathfrak{b}}_h$. Since $G$ is $x$-torsion-free, it follows that $rG = 0$, so that $r \in {\mathfrak{b}}_0$. \end{proof}
\begin{defi} \label{nt.6} Assume that the left $R[x,f]$-module $G$ is $x$-torsion-free. An ideal ${\mathfrak{b}}$ of $R$ is called a {\em $G$-special $R$-ideal} if there is an $R[x,f]$-submodule $N$ of $G$ such that $\grann_{R[x,f]}N = {\mathfrak{b}} R[x,f] = \bigoplus _{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{b}} x^n$. It is worth noting that, then, the ideal ${\mathfrak{b}}$ is just $(0:_RN)$.
We shall denote the set of $G$-special $R$-ideals by $\mathcal{I}(G)$. Note that, by Lemma \ref{nt.5}, all the ideals in $\mathcal{I}(G)$ are radical. \end{defi}
We can now combine together the results of Lemmas \ref{nt.3}(v) and \ref{nt.5} to obtain the following result, which is fundamental for the work in this paper.
\begin{prop} \label{nt.7} Assume that the left $R[x,f]$-module $G$ is $x$-torsion-free.
There is an order-reversing bijection, $\Delta : \mathcal{A}(G) \longrightarrow \mathcal{I}(G),$ from the set $\mathcal{A}(G)$ of special annihilator submodules of $G$ to the set $\mathcal{I}(G)$ of $G$-special $R$-ideals given by $$ \Delta : N \longmapsto \left(\grann_{R[x,f]}N\right)\cap R = (0:_RN). $$
The inverse bijection, $\Delta^{-1} : \mathcal{I}(G) \longrightarrow \mathcal{A}(G),$ also order-reversing, is given by $$ \Delta^{-1} : {\mathfrak{b}} \longmapsto \ann_G\left({\mathfrak{b}} R[x,f])\right). $$ When $N \in \mathcal{A}(G)$ and ${\mathfrak{b}} \in \mathcal{I}(G)$ are such that $\Delta(N) = {\mathfrak{b}}$, we shall say simply that `{\em $N$ and ${\mathfrak{b}}$ correspond}'. \end{prop}
\begin{cor} \label{nt.9} Assume that the left $R[x,f]$-module $G$ is $x$-torsion-free.
Then both the sets $\mathcal{A}(G)$ and $\mathcal{I}(G)$ are closed under taking arbitrary intersections. \end{cor}
\begin{proof} Let $\left(N_{\lambda}\right)_{\lambda \in \Lambda}$ be an arbitrary family of special annihilator submodules of $G$. For each $\lambda \in \Lambda$, let ${\mathfrak{b}}_{\lambda}$ be the $G$-special $R$-ideal corresponding to $N_{\lambda}$. In view of Proposition \ref{nt.7}, it is sufficient for us to show that $\bigcap_{\lambda \in \Lambda}N_{\lambda} \in \mathcal{A}(G)$ and ${\mathfrak{b}} := \bigcap_{\lambda \in \Lambda}{\mathfrak{b}}_{\lambda} \in \mathcal{I}(G)$.
To prove these, simply note that $$ {\textstyle \bigcap_{\lambda \in \Lambda}N_{\lambda} = \bigcap_{\lambda \in \Lambda}\ann_G\left({\mathfrak{b}}_{\lambda} R[x,f])\right) = \ann_G\left(\left(\sum_{\lambda \in \Lambda}{\mathfrak{b}}_{\lambda}\right)\! R[x,f]\right)} $$ and that $\sum_{\lambda \in \Lambda}N_{\lambda}$ is an $R[x,f]$-submodule of $G$ such that $$ {\textstyle \grann_{R[x,f]}\left(\sum_{\lambda \in \Lambda}N_{\lambda}\right) = \bigcap_{\lambda \in \Lambda}\grann_{R[x,f]}N_{\lambda} = \bigcap_{\lambda \in \Lambda}\left({\mathfrak{b}}_{\lambda}R[x,f] \right) = {\mathfrak{b}} R[x,f].} $$ \end{proof}
\begin{rmk} \label{nt.8} Suppose that the left $R[x,f]$-module $G$ is $x$-torsion-free.
It is worth pointing out now that, since $R$ is Noetherian, so that the set $\mathcal{I}(G)$ of $G$-special $R$-ideals satisfies the ascending chain condition, it is a consequence of Proposition \ref{nt.7} that the set $\mathcal{A}(G)$ of special annihilator submodules of $G$, partially ordered by inclusion, satisfies the descending chain condition. This is the case even if $G$ is not finitely generated. Note that (by \cite[Theorem (1.3)]{Yoshi94}), the (noncommutative) ring $R[x,f]$ is neither left nor right Noetherian if $\dim R > 0$. \end{rmk}
\section{\sc Examples relevant to the theory of tight closure} \label{ex}
The purpose of this section is to present some motivating examples, from the theory of tight closure, of some of the concepts introduced in \S \ref{nt}. Throughout this section, we shall again employ the notation of \ref{nt.1}, and ${\mathfrak{a}}$ will always denote an ideal of $R$. Recall that the {\em Frobenius closure ${\mathfrak{a}}^F$ of ${\mathfrak{a}}$} is the ideal of $R$ defined by $$ {\mathfrak{a}} ^F := \big\{ r \in R : \mbox{there exists~} n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi \mbox{~such that~} r^{p^n} \in {\mathfrak{a}}^{[p^n]}\big\}. $$
\begin{rmk} \label{ex.0} Let $({\mathfrak{b}}_n)_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ be a family of ideals of $R$ such that ${\mathfrak{b}}_n \subseteq f^{-1}({\mathfrak{b}}_{n+1})$ for all $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. Then $\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{b}}_nx^n$ is a graded left ideal of $R[x,f]$, and so we may form the graded left $R[x,f]$-module $ R[x,f]/\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{b}}_nx^n$. This may be viewed as $\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}R/{\mathfrak{b}}_n$, where, for $r \in R$ and $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, the result of multiplying the element $r + {\mathfrak{b}}_n$ of the $n$-th component by $x$ is the element $r^{p} + {\mathfrak{b}}_{n+1}$ of the $(n+1)$-th component.
Note that the left $R[x,f]$-module $R[x,f]/\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{b}}_nx^n$ is $x$-torsion-free if and only if ${\mathfrak{b}}_n = f^{-1}({\mathfrak{b}}_{n+1})$ for all $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, that is, if and only if $({\mathfrak{b}}_n)_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ is an $f$-sequence in the sense of \cite[Definition 4.1(ii)]{SN}. \end{rmk}
\begin{ntn} \label{ex.0a} Since $R[x,f]{\mathfrak{a}} = \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{a}}^{[p^n]}x^n$, we can view the graded left $R[x,f]$-module $$R[x,f]/R[x,f]{\mathfrak{a}}$$ as $\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}R/{\mathfrak{a}}^{[p^n]}$ in the manner described in \ref{ex.0}. We shall denote the graded left $R[x,f]$-module $\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}R/{\mathfrak{a}}^{[p^n]}$ by $H({\mathfrak{a}})$.
Recall from \cite[4.1(iii)]{SN} that $\left(({\mathfrak{a}}^{[p^n]})^F\right)_{n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ is the {\em canonical $f$-sequence associated to ${\mathfrak{a}}$}. We shall denote $\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}R/({\mathfrak{a}}^{[p^n]})^F$, considered as a graded left $R[x,f]$-module in the manner described in \ref{ex.0}, by $G({\mathfrak{a}})$. Note that $G({\mathfrak{a}})$ is $x$-torsion-free. \end{ntn}
\begin{lem} \label{ex.0b} With the notation of\/ {\rm \ref{ex.0a}}, we have $\Gamma_x(H({\mathfrak{a}})) = \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}({\mathfrak{a}}^{[p^n]})^F/{\mathfrak{a}}^{[p^n]}$, so that there is an isomorphism of graded left $R[x,f]$-modules $$ H({\mathfrak{a}})/ \Gamma_x(H({\mathfrak{a}})) \cong G({\mathfrak{a}}). $$ \end{lem}
\begin{proof} Let $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $r \in R$. Then the element $r + {\mathfrak{a}}^{[p^n]}$ of the $n$-th component of $H({\mathfrak{a}})$ belongs to $\Gamma_x(H({\mathfrak{a}}))$ if and only if there exists $m \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ such that $x^m(r + {\mathfrak{a}}^{[p^n]})= r^{p^m} + ({\mathfrak{a}}^{[p^{n}]})^{[p^{m}]} = 0$, that is, if and only if $r \in ({\mathfrak{a}}^{[p^n]})^F$. \end{proof}
\begin{prop} \label{ex.1} We use the notation of\/ {\rm \ref{ex.0a}}.
Suppose that there exists a $p^{w_0}$-weak test element $c$ for $R$, for some $w_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. Then \begin{enumerate} \item $ \ann_{H({\mathfrak{a}})}\left( \bigoplus_{n\geq w_0} Rcx^n\right) = \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}({\mathfrak{a}}^{[p^n]})^*/{\mathfrak{a}}^{[p^n]}\mbox{;} $
\item $ \ann_{G({\mathfrak{a}})}\left(\bigoplus_{n\in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} Rcx^n\right) = \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}({\mathfrak{a}}^{[p^n]})^*/({\mathfrak{a}}^{[p^n]})^F. $ \end{enumerate} \end{prop}
\begin{proof} (i) Let $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $r \in R$. Then the element $r + {\mathfrak{a}}^{[p^j]}$ of the $j$-th component of $H({\mathfrak{a}})$ belongs to $\ann_{H({\mathfrak{a}})}\left(\bigoplus_{n\geq w_0} Rcx^n\right)$ if and only if $cr^{p^n} \in ({\mathfrak{a}}^{[p^j]})^{[p^n]}$ for all $n \geq w_0$, that is, if and only if $r \in ({\mathfrak{a}}^{[p^j]})^*$.
(ii) By part (i), $$ {\textstyle \bigoplus_{n\in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}}({\mathfrak{a}}^{[p^n]})^*/({\mathfrak{a}}^{[p^n]})^F \subseteq \ann_{G({\mathfrak{a}})}\left({\textstyle \bigoplus_{n\geq w_0}} Rcx^n\right). $$ Note that $\ann_{G({\mathfrak{a}})}\left(\bigoplus_{n\geq w_0} Rcx^n\right)$ is a graded $R[x,f]$-submodule of $G({\mathfrak{a}})$. Let $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $r \in R$ be such that $r + ({\mathfrak{a}}^{[p^j]})^F$ belongs to the $j$-th component of $\ann_{G({\mathfrak{a}})}\left(\bigoplus_{n\geq w_0} Rcx^n\right)$. Then, for all $n \geq w_0$, we have $cr^{p^n} \in ({\mathfrak{a}}^{[p^{j+n}]})^F = (({\mathfrak{a}}^{[p^{j}]})^{[p^{n}]})^F$. Therefore, by \cite[Lemma 0.1]{KS}, we have $r \in ({\mathfrak{a}}^{[p^{j}]})^*$.
It follows from this that $$ {\textstyle \bigoplus_{n\in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}}({\mathfrak{a}}^{[p^n]})^*/({\mathfrak{a}}^{[p^n]})^F = \ann_{G({\mathfrak{a}})}\left({\textstyle \bigoplus_{n\geq w_0}} Rcx^n\right), $$ and so $\bigoplus_{n\in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}({\mathfrak{a}}^{[p^n]})^*/({\mathfrak{a}}^{[p^n]})^F$ is a special annihilator submodule of the $x$-torsion-free graded left $R[x,f]$-module $G({\mathfrak{a}})$. Let ${\mathfrak{b}}$ be the $G({\mathfrak{a}})$-special $R$-ideal corresponding to this member of $\mathcal{A}(G({\mathfrak{a}}))$. The above-displayed equation shows that $Rc \subseteq {\mathfrak{b}}$. Hence, by Proposition \ref{nt.7}, \begin{align*} {\textstyle \bigoplus_{n\in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}}({\mathfrak{a}}^{[p^n]})^*/({\mathfrak{a}}^{[p^n]})^F & = \ann_{G({\mathfrak{a}})}\left({\textstyle \bigoplus_{n\geq w_0}} Rcx^n\right) \supseteq \ann_{G({\mathfrak{a}})}\left({\textstyle \bigoplus_{n\in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}} Rcx^n\right) \\ & \supseteq \ann_{G({\mathfrak{a}})}\left({\textstyle \bigoplus_{n\in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}} {\mathfrak{b}} x^n\right) = {\textstyle \bigoplus_{n\in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}}({\mathfrak{a}}^{[p^n]})^*/({\mathfrak{a}}^{[p^n]})^F. \end{align*} \end{proof}
\begin{defi} \label{ex.1t} The {\em weak test ideal $\tau'(R)$ of $R$\/} is defined to be the ideal generated by $0$ and all weak test elements for $R$. (By a `weak test element' for $R$ we mean a $p^{w_0}$-weak test element for $R$ for some $w_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$.)
It is easy to see that each element of $\tau'(R) \cap R^{\circ}$ is a weak test element for $R$. \end{defi}
\begin{thm} \label{ex.2} We use the notation of\/ {\rm \ref{ex.0a}}. Suppose that there exists a $p^{w_0}$-weak test element $c$ for $R$, for some $w_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$.
Let $H$ be the positively-graded left $R[x,f]$-module given by $$ H := \bigoplus_{{\mathfrak{a}} \textit{~is an ideal of~} R}H({\mathfrak{a}}) = \bigoplus_{{\mathfrak{a}} \textit{~is an ideal of~} R}\Big(\textstyle{\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}} R/{\mathfrak{a}}^{[p^n]}\Big). $$ Set $T := {\displaystyle \bigoplus_{{\mathfrak{a}} \textit{~is an ideal of~} R}} \Big(\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}({\mathfrak{a}}^{[p^n]})^*/{\mathfrak{a}}^{[p^n]}\Big)$.
\begin{enumerate} \item Then $ T = \ann_{H}\left( \bigoplus_{n\geq w_0} Rcx^n\right), $ and so is a special annihilator submodule of $H$. \item Write $\grann_{R[x,f]}T = \bigoplus_{n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{c}}_nx^n$ for a suitable ascending chain $({\mathfrak{c}}_n)_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ of ideals of $R$. Then $\lim_{n \rightarrow \infty}{\mathfrak{c}}_n = \tau'(R)$, the weak test ideal for $R$.
\item Furthermore, $T$ contains every special annihilator submodule $T'$ of $H$ for which the graded annihilator $\grann_{R[x,f]}T' = \bigoplus_{n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{b}}_nx^n$ has\/ $\height (\lim_{n \rightarrow \infty}{\mathfrak{b}}_n) \geq 1$. (The height of the improper ideal $R$ is considered to be $\infty$.) \end{enumerate} \end{thm}
\begin{proof} (i) This is immediate from Proposition \ref{ex.1}(i).
(ii) Write ${\mathfrak{c}} := \lim_{n \rightarrow \infty}{\mathfrak{c}}_n$. Since there exists a weak test element for $R$, the ideal $\tau'(R)$ can be generated by finitely many weak test elements for $R$, say by $c_i~(i = 1, \ldots, h)$, where $c_i$ is a $p^{w_i}$-weak test element for $R$ (for $i= 1, \ldots, h$). Set $\widetilde{w} = \max\{w_1, \ldots, w_h\}$. It is immediate from part (i) that $\bigoplus_{n \geq \widetilde{w}}\tau'(R)x^n \subseteq \grann_{R[x,f]}T$, and so $\tau'(R) \subseteq {\mathfrak{c}}$. Therefore $\height {\mathfrak{c}} \geq 1$, so that ${\mathfrak{c}} \cap R^{\circ} \neq \emptyset$ by prime avoidance, and ${\mathfrak{c}}$ can be generated by its elements in $R^{\circ}$.
There exists $m_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ such that ${\mathfrak{c}}_n = {\mathfrak{c}}$ for all $n \geq m_0$. Let $c' \in {\mathfrak{c}} \cap R^{\circ}$. Thus $T$ is annihilated by $c'x^n$ for all $n \geq m_0$; therefore, for each ideal ${\mathfrak{a}}$ of $R$, and for all $r \in {\mathfrak{a}}^*$, we have $c'r^{p^n} \in {\mathfrak{a}}^{[p^n]}$ for all $n \geq m_0$, so that $c'$ is a $p^{m_0}$-weak test element for $R$. Therefore $c' \in \tau'(R)$. Since ${\mathfrak{c}}$ can be generated by elements in ${\mathfrak{c}} \cap R^{\circ}$, it follows that ${\mathfrak{c}} \subseteq \tau'(R)$.
(iii) Since $T' = \ann_{H}\Big(\bigoplus_{n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{b}}_nx^n\Big)$, it follows that $$ T' = \bigoplus_{{\mathfrak{a}} \text{~is an ideal of~} R} \Big({\textstyle \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}}{\mathfrak{a}}_n/{\mathfrak{a}}^{[p^n]}\Big), $$ where, for each ideal ${\mathfrak{a}}$ of $R$ and each $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, the ideal ${\mathfrak{a}}_n$ of $R$ contains ${\mathfrak{a}}^{[p^n]}$. Suppose that $\lim_{n \rightarrow \infty}{\mathfrak{b}}_n = {\mathfrak{b}}$ and that $v_0 \in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ is such that ${\mathfrak{b}}_n = {\mathfrak{b}}$ for all $n \geq v_0$. Since $\height {\mathfrak{b}} \geq 1$, there exists $\overline{c} \in {\mathfrak{b}}\cap R^{\circ}$, by prime avoidance. Let ${\mathfrak{a}}$ be an ideal of $R$ and let $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. Then, for each $r \in {\mathfrak{a}}_n$, the element $r + {\mathfrak{a}}^{[p^n]}$ of the $n$-th component of $H({\mathfrak{a}})$ is annihilated by $\overline{c}x^j$ for all $j \geq v_0$. This means that $\overline{c}r^{p^j} \in ({\mathfrak{a}}^{[p^n]})^{[p^j]}$ for all $j \geq v_0$, so that $r \in ({\mathfrak{a}}^{[p^n]})^*$. Therefore $T' \subseteq T$. \end{proof}
\begin{thm} \label{ex.3} We use the notation of\/ {\rm \ref{ex.0a}}. Suppose that there exists a $p^{w_0}$-weak test element $c$ for $R$, for some $w_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$.
Let $G$ be the positively-graded $x$-torsion-free left $R[x,f]$-module given by $$ G := \bigoplus_{{\mathfrak{a}} \textit{~is an ideal of~} R}G({\mathfrak{a}}) = \bigoplus_{{\mathfrak{a}} \textit{~is an ideal of~} R}\Big({\textstyle \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}} R/({\mathfrak{a}}^{[p^n]})^F\Big). $$ Set $U := {\displaystyle \bigoplus_{{\mathfrak{a}} \textit{~is an ideal of~} R}} \Big(\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}({\mathfrak{a}}^{[p^n]})^*/({\mathfrak{a}}^{[p^n]})^F\Big)$.
\begin{enumerate} \item Then $ U = \ann_{G}\left( \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} Rcx^n\right), $ and so is a special annihilator submodule of $G$. \item Let ${\mathfrak{b}}$ be the $G$-special $R$-ideal corresponding to $U$. Then ${\mathfrak{b}}$ is the smallest member of $\mathcal{I}(G)$ of positive height. \end{enumerate} \end{thm}
\begin{proof} (i) This is immediate from Proposition \ref{ex.1}(ii).
(ii) Note that $Rc \subseteq {\mathfrak{b}}$, by part (i); therefore $\height {\mathfrak{b}} \geq 1$. To complete the proof, we show that, if ${\mathfrak{b}}' \in \mathcal{I}(G)$ has $\height {\mathfrak{b}}' \geq 1$, then ${\mathfrak{b}} \subseteq {\mathfrak{b}}'$. By prime avoidance, there exists $\widetilde{c} \in {\mathfrak{b}}' \cap R^{\circ}$. Let $U' \in \mathcal{A}(G)$ correspond to ${\mathfrak{b}}'$ (in the correspondence of Proposition \ref{nt.7}). Since $U' = \ann_G{\mathfrak{b}}' R[x,f]$, it follows that $$ U' = \bigoplus_{{\mathfrak{a}} \text{~is an ideal of~} R} \Big({\textstyle \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}}{\mathfrak{a}}_n/({\mathfrak{a}}^{[p^n]})^F\Big), $$ where, for each ideal ${\mathfrak{a}}$ of $R$ and each $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, the ideal ${\mathfrak{a}}_n$ of $R$ contains $({\mathfrak{a}}^{[p^n]})^F$. Let ${\mathfrak{a}}$ be an ideal of $R$ and let $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. Then, for each $r \in {\mathfrak{a}}_n$, the element $r + ({\mathfrak{a}}^{[p^n]})^F$ of the $n$-th component of $G({\mathfrak{a}})$ is annihilated by $\widetilde{c}x^j$ for all $j \geq 0$. This means that $\widetilde{c}r^{p^j} \in (({\mathfrak{a}}^{[p^n]})^{[p^j]})^F$ for all $j \geq 0$, so that $r \in ({\mathfrak{a}}^{[p^n]})^*$ by \cite[Lemma 0.1(i)]{KS}. Therefore $U' \subseteq U$, so that ${\mathfrak{b}}' \supseteq {\mathfrak{b}}$. \end{proof}
\section{\sc Properties of special annihilator submodules in the $x$-torsion-free case} \label{ga}
Throughout this section, we shall employ the notation of \ref{nt.1}. The aim is to develop the theory of special annihilator submodules of an $x$-torsion-free left $R[x,f]$-module.
\begin{lem} \label{ga.1} Suppose that $G$ is $x$-torsion-free. Let $N$ be a special annihilator submodule of $G$. Then the left $R[x,f]$-module $G/N$ is also $x$-torsion-free. \end{lem}
\begin{proof} By Lemma \ref{nt.5} and Proposition \ref{nt.7}, there is a radical ideal ${\mathfrak{b}}$ of $R$ such that $N = \ann_G\left({\mathfrak{b}} R[x,f]\right).$ Let $g \in G$ be such that $xg \in N$. Therefore, for all $r \in {\mathfrak{b}}$ and all $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, we have $rx^j(xg) = 0$, that is $rx^{j+1}g = 0$. Also, for $r \in {\mathfrak{b}}$, since $r(xg) = 0$, we have $x(rg) = r^pxg = 0$, and so $rg = 0$ because $G$ is $x$-torsion-free. Thus $g \in \ann_G\left(\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{b}} x^n\right) = N$. It follows that $G/N$ is $x$-torsion-free. \end{proof}
\begin{lem} \label{ga.1a} Suppose that $G$ is $x$-torsion-free. Let ${\mathfrak{a}}$ be an ideal of $R$, and set $L := \ann_G\left({\mathfrak{a}} R[x,f]\right) \in \mathcal{A}(G)$. Then $L =
\ann_G\left(\sqrt{{\mathfrak{a}}} R[x,f]\right)$. \end{lem}
\begin{proof} Let ${\mathfrak{d}} \in \mathcal{I}(G)$ correspond to $L$. Note that ${\mathfrak{d}}$ is radical, by Lemma \ref{nt.5}; also, ${\mathfrak{a}} \subseteq {\mathfrak{d}}$. Hence $ {\mathfrak{a}} \subseteq \sqrt{{\mathfrak{a}}} \subseteq \sqrt{{\mathfrak{d}}} = {\mathfrak{d}}$. Since $\ann_G\left({\mathfrak{a}} R[x,f]\right) = \ann_G\left({\mathfrak{d}} R[x,f]\right)$, we must have $\ann_G\left({\mathfrak{a}} R[x,f]\right) = \ann_G\left(\sqrt{{\mathfrak{a}}} R[x,f]\right).$ \end{proof}
\begin{prop} \label{ga.3} Suppose that $G$ is $x$-torsion-free. Let ${\mathfrak{a}}$ be an ideal of $R$, and set $$L := \ann_G\left({\mathfrak{a}} R[x,f]\right) \in \mathcal{A}(G).$$ Note that $G/L$ is $x$-torsion-free, by Lemma\/ {\rm \ref{ga.1}}. Let $N$ be an $R$-submodule of $G$ such that $L \subseteq N \subseteq G$.
\begin{enumerate} \item If $N = \ann_G\left({\mathfrak{b}} R[x,f]\right) \in \mathcal{A}(G)$, where ${\mathfrak{b}}$ is an ideal of $R$ contained in ${\mathfrak{a}}$, then $$N/L = \ann_{G/L}\left(({\mathfrak{b}}:{\mathfrak{a}})R[x,f]\right) \in \mathcal{A}(G/L).$$ Furthermore, if the ideal in $\mathcal{I}(G)$ corresponding to $N$ is ${\mathfrak{b}}$, then $({\mathfrak{b}}:{\mathfrak{a}})$ is the ideal in $\mathcal{I}(G/L)$ corresponding to $N/L$.
\item If $N/L = \ann_{G/L}\left({\mathfrak{c}} R[x,f]\right) \in \mathcal{A}(G/L)$, where ${\mathfrak{c}}$ is an ideal of $R$, then $$N = \ann_{G}\left({\mathfrak{a}}{\mathfrak{c}} R[x,f]\right) = \ann_{G}\left(({\mathfrak{a}}\cap{\mathfrak{c}}) R[x,f]\right) \in \mathcal{A}(G).$$ Furthermore, if ${\mathfrak{a}}$ is the ideal in $\mathcal{I}(G)$ corresponding to $L$ and ${\mathfrak{c}}$ is the ideal in $\mathcal{I}(G/L)$ corresponding to $N/L$, then ${\mathfrak{a}}\cap{\mathfrak{c}}$ is the ideal in $\mathcal{I}(G)$ corresponding to $N$.
\item There is an order-preserving bijection from $\{ N \in \mathcal{A}(G) : N \supseteq L\}$ to $\mathcal{A}(G/L)$ given by $N \mapsto N/L$. \end{enumerate} \end{prop}
\begin{proof} By Lemma \ref{ga.1}, the left $R[x,f]$-module $G/L$ is $x$-torsion-free.
(i) Let $g \in N$. Let $i,j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $r \in ({\mathfrak{b}}:{\mathfrak{a}})$, $u \in {\mathfrak{a}}$. Then $ux^i(rx^jg) = ur^{p^i}x^{i+j}g = 0$ because $ ur^{p^i} \in {\mathfrak{b}}$ and ${\mathfrak{b}} x^{i+j}$ annihilates $N$. This is true for all $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $u \in {\mathfrak{a}}$. Therefore $ rx^jg \in \ann_G\left({\mathfrak{a}} R[x,f]\right) = L. $ Since this is true for all $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $r \in ({\mathfrak{b}}:{\mathfrak{a}})$, we see that $N/L \subseteq \ann_{G/L}\left(({\mathfrak{b}}:{\mathfrak{a}}) R[x,f]\right)$.
Now suppose that $g \in G$ is such that $g + L \in \ann_{G/L}\left(({\mathfrak{b}}:{\mathfrak{a}}) R[x,f]\right)$. Let $r \in {\mathfrak{b}}$ and $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. Then $r \in ({\mathfrak{b}}:{\mathfrak{a}})$ and so $rx^{i+1}$ annihilates $g + L \in G/L$. Hence $rx^{i+1}g \in L$. Since ${\mathfrak{b}} \subseteq {\mathfrak{a}}$, we see that $r^{p-1}rx^{i+1}g = 0$, so that $xrx^ig = 0$. As $G$ is $x$-torsion-free, it follows that $rx^ig = 0$. As this is true for all $r \in {\mathfrak{b}}$ and $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, we see that $ g \in \ann_G\left({\mathfrak{b}} R[x,f]\right) = N. $ Hence $N/L = \ann_{G/L}\left(({\mathfrak{b}}:{\mathfrak{a}}) R[x,f]\right)$.
To prove the final claim, we have to show that $\grann_{R[x,f]}(N/L) = \bigoplus _{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} ({\mathfrak{b}}:{\mathfrak{a}}) x^n$, given that $\grann_{R[x,f]}N = \bigoplus _{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{b}} x^n$. In view of the preceding paragraph, it remains only to show that $$\grann_{R[x,f]}(N/L) \subseteq ({\mathfrak{b}}:{\mathfrak{a}})R[x,f].$$
Let $r \in R$ be such that $rx^i \in \grann_{R[x,f]}(N/L)$ for all $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. Let $g \in N$. Then $rx^ig \in L$ for all $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, and so ${\mathfrak{a}} rx^ig = 0$ for all $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. As this is true for all $g \in N$ and for all $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, it follows that ${\mathfrak{a}} r \subseteq \left(\grann_{R[x,f]}N\right) \cap R = {\mathfrak{b}}$. Hence $r \in ({\mathfrak{b}}:{\mathfrak{a}})$.
(ii) Let $g \in N$. Then $ux^ig \in L$ for all $u \in {\mathfrak{c}}$ and $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, and so $rux^ig = 0$ for all $r \in {\mathfrak{a}}$, $u \in {\mathfrak{c}}$ and $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. Hence $N \subseteq \ann_G\left(\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{a}}{\mathfrak{c}} x^n\right) = \ann_G\left({\mathfrak{a}}{\mathfrak{c}} R[x,f]\right)$.
Now let $g \in \ann_G\left({\mathfrak{a}}{\mathfrak{c}} R[x,f]\right)$. Then, for all $r \in {\mathfrak{a}}$, $u \in {\mathfrak{c}}$ and $i,j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, we have $rx^i(ux^jg) = ru^{p^i}x^{i+j}g = 0$, and so $ux^jg \in L$ for all $u \in {\mathfrak{c}}$ and $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$. Hence $g + L \in \ann_{G/L}\left(\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{c}} x^n\right) = N/L$, and $g \in N$. It follows that $ N = \ann_G\left({\mathfrak{a}}{\mathfrak{c}} R[x,f]\right). $ Also, by Lemma \ref{ga.1a}, we have $$\ann_G\left({\mathfrak{a}}{\mathfrak{c}} R[x,f]\right) = \ann_G\left(({\mathfrak{a}}\cap{\mathfrak{c}}) R[x,f]\right),$$ because ${\mathfrak{a}} {\mathfrak{c}}$ and ${\mathfrak{a}} \cap {\mathfrak{c}}$ have the same radical.
To prove the final claim, we have to show that $\grann_{R[x,f]}N = ({\mathfrak{a}}\cap{\mathfrak{c}})R[x,f]$, given that $$\grann_{R[x,f]}(N/L) = {\mathfrak{c}} R[x,f] \quad \mbox{~and~} \quad \grann_{R[x,f]}(L) = {\mathfrak{a}} R[x,f].$$ In view of the preceding paragraph, it remains only to show that $\grann_{R[x,f]}N \subseteq ({\mathfrak{a}}\cap{\mathfrak{c}})R[x,f].$ However, this is clear, because $ \grann_{R[x,f]}N \subseteq \grann_{R[x,f]}L \cap \grann_{R[x,f]}(N/L). $
(iii) This is now immediate from parts (i) and (ii). \end{proof}
\begin{rmk} \label{ga.3r} It follows from Proposition \ref{ga.3}(ii) (and with the hypotheses and notation thereof) that, if ${\mathfrak{a}}$ is an ideal of $R$ and $L := \ann_G\left({\mathfrak{a}} R[x,f]\right)$, then $ \ann_{G/L}\left({\mathfrak{a}} R[x,f]\right) = 0$. \end{rmk}
Because the special $R$-ideals introduced in Definition \ref{nt.6} are radical, the following lemma will be very useful.
\begin{lem} \label{ga.2} Let ${\mathfrak{a}}$ and ${\mathfrak{b}}$ be proper radical ideals of $R$, and let their (unique) minimal primary decompositions be $$ {\mathfrak{a}} = {\mathfrak{r}}_1 \cap \ldots \cap {\mathfrak{r}}_k \cap {\mathfrak{p}}_1 \cap \ldots \cap {\mathfrak{p}}_t \cap {\mathfrak{p}}_1' \cap \ldots \cap {\mathfrak{p}}_u' $$ and $$ {\mathfrak{b}} = {\mathfrak{r}}_1 \cap \ldots \cap {\mathfrak{r}}_k \cap {\mathfrak{q}}_1 \cap \ldots \cap {\mathfrak{q}}_v \cap {\mathfrak{q}}_1' \cap \ldots \cap {\mathfrak{q}}_w', $$ where the notation is such that $$ \left\{{\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_t, {\mathfrak{p}}_1', \ldots, {\mathfrak{p}}_u'\right\} \cap \left\{{\mathfrak{q}}_1, \ldots, {\mathfrak{q}}_v, {\mathfrak{q}}_1', \ldots, {\mathfrak{q}}_w'\right\} = \emptyset, $$ and such that none of ${\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_t$ contains an associated prime of ${\mathfrak{b}}$, each of ${\mathfrak{p}}_1', \ldots, {\mathfrak{p}}_u'$ contains an associated prime of ${\mathfrak{b}}$, none of ${\mathfrak{q}}_1, \ldots, {\mathfrak{q}}_v$ contains an associated prime of ${\mathfrak{a}}$, and each of ${\mathfrak{q}}_1', \ldots, {\mathfrak{q}}_w'$ contains an associated prime of ${\mathfrak{a}}$. (Note that some, but not all, of the integers $k$, $t$ and $u$ might be zero; a similar comment applies to the primary decomposition of ${\mathfrak{b}}$.) Then \begin{enumerate} \item ${\mathfrak{a}} \cap {\mathfrak{b}} = {\mathfrak{r}}_1 \cap \ldots \cap {\mathfrak{r}}_k \cap {\mathfrak{p}}_1 \cap \ldots \cap {\mathfrak{p}}_t \cap {\mathfrak{q}}_1 \cap \ldots \cap {\mathfrak{q}}_v$ is the minimal primary decomposition; \item if ${\mathfrak{a}} \not\subseteq {\mathfrak{b}}$, the equation $ ({\mathfrak{b}} : {\mathfrak{a}}) = {\mathfrak{q}}_1 \cap \ldots \cap {\mathfrak{q}}_v$ gives the minimal primary decomposition. \end{enumerate} \end{lem}
\begin{proof} (i) Each of ${\mathfrak{p}}_1', \ldots, {\mathfrak{p}}_u'$ must contain one of ${\mathfrak{q}}_1, \ldots, {\mathfrak{q}}_v$; likewise, each of ${\mathfrak{q}}_1', \ldots, {\mathfrak{q}}_w'$ must contain one of ${\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_t$ . The claim then follows easily.
(ii) Since $ ({\mathfrak{b}} : {\mathfrak{a}}) = ({\mathfrak{b}} \cap {\mathfrak{a}} : {\mathfrak{a}})$, it is clear from part (i) that $$ {\mathfrak{q}}_1 \cap \ldots \cap {\mathfrak{q}}_v \subseteq ({\mathfrak{b}} : {\mathfrak{a}}). $$ Now let $r \in ({\mathfrak{b}} : {\mathfrak{a}}) = ({\mathfrak{b}} \cap {\mathfrak{a}} : {\mathfrak{a}})$. Then, for each $i = 1, \ldots, v$, we have $r{\mathfrak{a}} \subseteq {\mathfrak{q}}_i$, whereas ${\mathfrak{a}} \not\subseteq {\mathfrak{q}}_i$; hence $r \in {\mathfrak{q}}_i$ because ${\mathfrak{q}}_i$ is prime. \end{proof}
\begin{thm} \label{ga.4} Suppose that $G$ is $x$-torsion-free. Let $N := \ann_G\left({\mathfrak{b}} R[x,f]\right) \in \mathcal{A}(G)$, where the ideal ${\mathfrak{b}} \in \mathcal{I}(G)$ corresponds to $N$. Assume that $N \neq 0$, and let ${\mathfrak{b}} = {\mathfrak{p}}_1 \cap \ldots \cap {\mathfrak{p}}_t$ be the minimal primary decomposition of the (radical) ideal ${\mathfrak{b}}$.
Suppose that $t > 1$, and consider any partition $\{1,\ldots,t\} = U \cup V$, where $U$ and $V$ are two non-empty disjoint sets. Set $ {\mathfrak{a}} = \bigcap_{i\in U} {\mathfrak{p}}_i$ and ${\mathfrak{c}} = \bigcap_{i\in V} {\mathfrak{p}}_i$. Let $L := \ann_G\left({\mathfrak{a}} R[x,f]\right) \in \mathcal{A}(G)$. Then
\begin{enumerate} \item $0 \subset L \subset N$ (the symbol `$\subset$' is reserved to denote strict inclusion); \item $N/L = \ann_{G/L}\left({\mathfrak{c}} R[x,f]\right) \in \mathcal{A}(G/L)$ with corresponding ideal ${\mathfrak{c}} \in \mathcal{I}(G/L)$; and \item $\grann_{R[x,f]}L = {\mathfrak{a}} R[x,f]$, so that ${\mathfrak{a}} \in \mathcal{I}(G)$ corresponds to $L$. \end{enumerate} \end{thm}
\begin{proof} (i) It is clear that $ L \subseteq N$. Suppose that $L = 0$ and seek a contradiction. Let $g \in N$. Let $i,j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $r \in {\mathfrak{c}}$, $u \in {\mathfrak{a}}$. Then $ux^i(rx^jg) = ur^{p^i}x^{i+j}g = 0$ because $ ur^{p^i} \in {\mathfrak{b}}$ and ${\mathfrak{b}} x^{i+j}$ annihilates $N$. This is true for all $i \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $u \in {\mathfrak{a}}$. Therefore $ rx^jg \in \ann_G\left(\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{a}} x^n\right) = L = 0. $ It follows that $\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{c}} x^n \subseteq \grann_{R[x,f]}N = \bigoplus _{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{b}} x^n$, so that ${\mathfrak{c}} \subseteq {\mathfrak{b}}$. But ${\mathfrak{b}} \subseteq {\mathfrak{c}}$, and so ${\mathfrak{c}} = {\mathfrak{b}}$. However, this contradicts the fact that ${\mathfrak{b}} = {\mathfrak{p}}_1 \cap \ldots \cap {\mathfrak{p}}_t$ is the unique minimal primary decomposition of ${\mathfrak{b}}$. Therefore $L \neq 0$.
Now suppose that $L = N$ and again seek a contradiction. Then $\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{a}} x^n \subseteq \grann_{R[x,f]}N = \bigoplus _{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi} {\mathfrak{b}} x^n$, so that ${\mathfrak{a}} \subseteq {\mathfrak{b}}$. But ${\mathfrak{b}} \subseteq {\mathfrak{a}}$, and so ${\mathfrak{a}} = {\mathfrak{b}}$, and this again leads to a contradiction. Therefore $L \neq N$.
(ii) Since ${\mathfrak{b}} \subseteq {\mathfrak{a}}$, it is immediate from Proposition \ref{ga.3}(i) that $N/L = \ann_{G/L}\left(({\mathfrak{b}}:{\mathfrak{a}})R[x,f]\right) \in \mathcal{A}(G/L)$ and that the ideal $({\mathfrak{b}}:{\mathfrak{a}}) \in \mathcal{I}(G/L)$ corresponds to $N/L$. However, it follows from Lemma \ref{ga.2}(ii) that $({\mathfrak{b}}:{\mathfrak{a}}) = \bigcap_{i\in V} {\mathfrak{p}}_i = {\mathfrak{c}}$.
(iii) Let ${\mathfrak{d}} \in \mathcal{I}(G)$ correspond to $L$. Note that ${\mathfrak{a}} = \bigcap_{i\in U} {\mathfrak{p}}_i \subseteq {\mathfrak{d}}$. By Proposition \ref{ga.3}(i), the ideal in $\mathcal{I}(G/L)$ corresponding to $N/L$ is $({\mathfrak{b}}:{\mathfrak{d}})$. Therefore, by part (ii), we have $({\mathfrak{b}}:{\mathfrak{d}}) = {\mathfrak{c}}$. But, by Proposition \ref{ga.3}(ii), the ideal in $\mathcal{I}(G)$ corresponding to $N$ is ${\mathfrak{d}}\cap{\mathfrak{c}}$. Therefore ${\mathfrak{b}} = {\mathfrak{d}} \cap {\mathfrak{c}}$, and so ${\mathfrak{d}} \neq R$.
Now ${\mathfrak{d}}$ is a radical ideal of $R$. By Lemma \ref{ga.2}(i), each ${\mathfrak{p}}_j$, for $j \in U$, is an associated prime of ${\mathfrak{d}}$. Hence ${\mathfrak{d}} \subseteq \bigcap_{j\in U} {\mathfrak{p}}_j = {\mathfrak{a}}$. But we already know that ${\mathfrak{a}} \subseteq {\mathfrak{d}}$, and so ${\mathfrak{d}} = {\mathfrak{a}}$. \end{proof}
\begin{cor} \label{ga.10} Suppose that $G$ is $x$-torsion-free. Then the set of $G$-special $R$-ideals is precisely the set of all finite intersections of prime $G$-special $R$-ideals (provided one includes the empty intersection, $R$, which corresponds to the zero special annihilator submodule of $G$). In symbols, $$ \mathcal{I}(G) = \left\{ {\mathfrak{p}}_1 \cap \ldots \cap {\mathfrak{p}}_t : t \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi \mbox{~and~} {\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_t \in \mathcal{I}(G)\cap\Spec(R)\right\}. $$ \end{cor}
\begin{proof} By Corollary \ref{nt.9}, the set $\mathcal{I}(G)$ is closed under taking intersections. A proper ideal ${\mathfrak{a}} \in \mathcal{I}(G)$ is radical and it follows from Theorem \ref{ga.4} that each (necessarily prime) primary component of ${\mathfrak{a}}$ also belongs to $\mathcal{I}(G)$. This is enough to complete the proof. \end{proof}
\begin{lem} \label{ga.11} Suppose that $G$ is $x$-torsion-free. Let ${\mathfrak{p}}$ be a maximal member of $\mathcal{I}(G) \setminus \{R\}$ with respect to inclusion, and let $L \in \mathcal{A}(G)$ be the corresponding special annihilator submodule of $G$. Thus $L$ is a minimal member of the set of non-zero special annihilator submodules of $G$.
Then ${\mathfrak{p}}$ is prime, and any non-zero $g\in L$ satisfies $\grann_{R[x,f]}R[x,f]g = {\mathfrak{p}} R[x,f]$. \end{lem}
\begin{proof} It follows from Corollary \ref{ga.10} that ${\mathfrak{p}}$ is prime.
Since $R[x,f]g$ is a non-zero $R[x,f]$-submodule of $L$, there is a proper radical ideal ${\mathfrak{a}} \in \mathcal{I}(G)$ such that $$ {\mathfrak{a}} R[x,f] = \grann_{R[x,f]}R[x,f]g \supseteq \grann_{R[x,f]}L = {\mathfrak{p}} R[f,x]. $$ Since ${\mathfrak{p}}$ is a maximal member of $\mathcal{I}(G) \setminus \{R\}$, we must have ${\mathfrak{a}} = {\mathfrak{p}}$. \end{proof}
Our next major aim is to show that, in the situation of Corollary \ref{ga.10}, the set $\mathcal{I}(G)$ is finite if $G$ has the property that, for each special annihilator submodule $L$ of $G$ (including $0 = \ann_GR[x,f]$), the $x$-torsion-free residue class module $G/L$ (see Lemma \ref{ga.1}) does not contain, as an $R[x,f]$-submodule, an infinite direct sum of non-zero special annihilator submodules of $G/L$. This may seem rather a complicated hypothesis, and so we point out now that it is satisfied if $G$ is a Noetherian or Artinian left $R[x,f]$-module, and therefore if $G$ is a Noetherian or Artinian $R$-module. These ideas will be applied, later in the paper, to an example in which $G$ is Artinian as an $R$-module.
The following lemma will be helpful in an inductive argument in the proof of Theorem \ref{ga.12}.
\begin{lem} \label{ga.12p} Suppose that $G$ is $x$-torsion-free, and that the set $\mathcal{I}(G) \setminus \{R\}$ is non-empty and has finitely many maximal members: suppose that there are $n$ of these and denote them by ${\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n$. (The ideals ${\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n$ are prime, by \/ {\rm \ref{ga.11}}.) Let $L := \ann_G\left({\mathfrak{p}}_1 \cap \cdots \cap {\mathfrak{p}}_n\right)R[x,f]$. Then the left $R[x,f]$-module $G/L$ is $x$-torsion-free, and $$ \mathcal{I}(G/L) \cap \Spec (R) = \mathcal{I}(G) \cap \Spec (R) \setminus \{{\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n\}. $$ \end{lem}
\begin{proof} Note that $\bigcap_{i=1}^n {\mathfrak{p}}_i \in \mathcal{I}(G)$, by Corollary \ref{nt.9}. Therefore $\grann_{R[x,f]}L = \left(\bigcap_{i=1}^n {\mathfrak{p}}_i\right)R[x,f]$ and $L$ corresponds to $\bigcap_{i=1}^n {\mathfrak{p}}_i$.
That $G/L$ is $x$-torsion-free follows from Lemma \ref{ga.1}. By Proposition \ref{ga.3}(iii), $$ \mathcal{A}(G/L) = \left\{ N/L : N \in \mathcal{A}(G) \mbox{~and~} L \subseteq N \right\}. $$ Let $N \in \mathcal{A}(G)$ with $L \subset N$, and let ${\mathfrak{b}} \in \mathcal{I}(G)$ correspond to $N$. Note that ${\mathfrak{b}} \subset \bigcap_{i=1}^n {\mathfrak{p}}_i$, and that no associated prime of ${\mathfrak{b}}$ can contain properly any of ${\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n$. Therefore the minimal primary decomposition of the radical ideal ${\mathfrak{b}}$ will have the form $$ {\mathfrak{b}} = \left( {\textstyle \bigcap_{i\in I}} {\mathfrak{p}}_i \right) \cap {\mathfrak{q}}_1 \cap \ldots \cap {\mathfrak{q}}_v, $$ where $I$ is some (possibly empty) subset of $\{1, \ldots, n\}$ and none of ${\mathfrak{q}}_1, \ldots, {\mathfrak{q}}_v$ contains any of ${\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n$. Note that ${\mathfrak{q}}_1, \ldots, {\mathfrak{q}}_v$ must all belong to $\mathcal{I}(G) \cap \Spec (R) \setminus \{{\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n\}$. Proposition \ref{ga.3}(i), this time used in conjunction with Lemma \ref{ga.2}(ii), now shows that $N/L \in \mathcal{A}(G/L)$ and the ideal of $\mathcal{I}(G/L)$ corresponding to $N/L$ is $$ \left({\mathfrak{b}} : {\mathfrak{p}}_1 \cap \cdots \cap {\mathfrak{p}}_n\right) = {\mathfrak{q}}_1 \cap \ldots \cap {\mathfrak{q}}_v. $$ Note also that, if ${\mathfrak{q}} \in \mathcal{I}(G) \cap \Spec (R) \setminus \{{\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n\}$ and $$ J := \left\{ j \in \{1, \ldots, n\} : {\mathfrak{p}}_j \not\supset {\mathfrak{q}} \right\}, $$ then ${\mathfrak{c}} := \left(\bigcap_{j\in J} {\mathfrak{p}}_j \right) \cap {\mathfrak{q}} \in \mathcal{I}(G)$ and ${\mathfrak{c}} \subset \bigcap_{i=1}^n{\mathfrak{p}}_i$. It now follows from Corollary \ref{ga.10} that $$ \mathcal{I}(G/L) \cap \Spec (R) = \mathcal{I}(G) \cap \Spec (R) \setminus \{{\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n\}, $$ as required. \end{proof}
\begin{thm} \label{ga.12} Suppose that $G$ is $x$-torsion-free. Assume that $G$ has the property that, for each special annihilator submodule $L$ of $G$ (including $0 = \ann_GR[x,f]$), the $x$-torsion-free residue class module $G/L$ does not contain, as an $R[x,f]$-submodule, an infinite direct sum of non-zero special annihilator submodules of $G/L$.
Then the set $\mathcal{I}(G)$ of $G$-special $R$-ideals is finite. \end{thm}
\begin{proof} By Corollary \ref{ga.10}, it is enough for us to show that the set $\mathcal{I}(G)\cap \Spec (R)$ is finite; we may suppose that the latter set is not empty, so that it has maximal members with respect to inclusion. In the first part of the proof, we show that $\mathcal{I}(G)\cap \Spec (R)$ has only finitely many such maximal members.
Let $\left({\mathfrak{p}}_{\lambda}\right)_{\lambda \in \Lambda}$ be a labelling of the set of maximal members of $\mathcal{I}(G)\cap \Spec (R)$, arranged so that ${\mathfrak{p}}_{\lambda} \neq {\mathfrak{p}}_{\mu}$ whenever $\lambda$ and $\mu$ are different elements of $\Lambda$. For each $\lambda \in \Lambda$, let $S_{\lambda}$ be the member of $\mathcal{A}(G)$ corresponding to ${\mathfrak{p}}_{\lambda}$.
Consider $\lambda, \mu \in \Lambda$ with $\lambda \neq \mu$. By Lemma \ref{ga.11}, a non-zero $g \in S_{\lambda} \cap S_{\mu}$ would have to satisfy $\grann_{R[x,f]}R[x,f]g = {\mathfrak{p}}_{\lambda} R[x,f] = {\mathfrak{p}}_{\mu} R[x,f]$. Since ${\mathfrak{p}}_{\lambda} \neq {\mathfrak{p}}_{\mu}$, this is impossible. Therefore $S_{\lambda} \cap S_{\mu} = 0$ and the sum $S_{\lambda} + S_{\mu}$ is direct.
Suppose, inductively, that $n \in \mathbb N$ and we have shown that, whenever $\lambda_1, \ldots, \lambda_n$ are $n$ distinct members of $\Lambda$, then the sum $\sum_{i=1}^nS_{\lambda_i}$ is direct. We can now use Lemma \ref{ga.11} to see that, if $g_i \in S_{\lambda_i}$ for $i = 1, \ldots, n$, then $$ \grann_{R[x,f]}R[x,f](g_1 + \cdots + g_n) = \bigcap_{\stackrel{\scriptstyle i=1}{g_i \neq 0}}^n{\mathfrak{p}}_{\lambda_i}R[x,f], $$ and then to deduce that, for $\lambda_{n+1} \in \Lambda \setminus\{\lambda_1, \ldots, \lambda_n\}$, we must have $ \left( \bigoplus_{i=1}^n S_{\lambda_i} \right) \bigcap S_{\lambda_{n+1}} = 0, $ so that the sum $S_{\lambda_1} + \cdots + S_{\lambda_n} + S_{\lambda_{n+1}}$ is direct.
It follows that the sum $\sum_{\lambda \in \Lambda}S_{\lambda}$ is direct; since each $S_{\lambda}$ is non-zero, the hypothesis about $G/0$ (that is, about $G$) ensures that $\Lambda$ is finite.
We have thus shown that $\mathcal{I}(G)\cap \Spec (R)$ has only finitely many maximal members. Note that $\max\{ \height {\mathfrak{p}} : {\mathfrak{p}} \mbox{~is a maximal member of~} \mathcal{I}(G)\cap \Spec (R) \}$ is an upper bound for the lengths of chains $$ {\mathfrak{p}}_0 \subset {\mathfrak{p}}_1 \subset \cdots \subset {\mathfrak{p}}_w $$ of prime ideals in $\mathcal{I}(G)\cap \Spec (R)$. We argue by induction on the maximum $t$ of these lengths. When $t = 0$, all members of $\mathcal{I}(G)\cap \Spec (R)$ are maximal members of that set, and so, by the first part of this proof, $\mathcal{I}(G)\cap \Spec (R)$ is finite. Now suppose that $t > 0$, and that it has been proved that $\mathcal{I}(G)\cap \Spec (R)$ is finite for smaller values of $t$.
We know that there are only finitely many maximal members of $\mathcal{I}(G)\cap \Spec (R)$; suppose that there are $n$ of these and denote them by ${\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n$. Let $L := \ann_G\left({\mathfrak{p}}_1 \cap \cdots \cap {\mathfrak{p}}_n\right)R[x,f]$. We can now use Lemma \ref{ga.12p} to deduce that the left $R[x,f]$-module $G/L$ is $x$-torsion-free and $$ \mathcal{I}(G/L) \cap \Spec (R) = \mathcal{I}(G) \cap \Spec (R) \setminus \{{\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n\}. $$ It follows from this and Proposition \ref{ga.3}(ii) that the inductive hypothesis can be applied to $G/L$, and so we can deduce that the set $$\mathcal{I}(G) \cap \Spec (R) \setminus \{{\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_n\}$$ is finite. Hence $\mathcal{I}(G) \cap \Spec (R)$ is a finite set and the inductive step is complete. \end{proof}
\begin{cor} \label{ga.12c} Suppose that the left $R[x,f]$-module $G$ is $x$-torsion-free and either Artinian or Noetherian as an $R$-module. Then the set $\mathcal{I}(G)$ of $G$-special $R$-ideals is finite. \end{cor}
\begin{thm} \label{ga.13} Suppose that $G$ is $x$-torsion-free and that the set $\mathcal{I}(G)$ of $G$-special $R$-ideals is finite. Then there exists a (uniquely determined) ideal ${\mathfrak{b}} \in \mathcal{I}(G)$ with the properties that $\height {\mathfrak{b}} \geq 1$ (the improper ideal $R$ is considered to have infinite height) and ${\mathfrak{b}} \subset {\mathfrak{c}}$ for every other ideal ${\mathfrak{c}} \in \mathcal{I}(G)$ with $\height {\mathfrak{c}} \geq 1$. Furthermore, for $g \in G$, the following statements are equivalent:
\begin{enumerate} \item $g$ is annihilated by ${\mathfrak{b}} R[x,f] = \bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{b}} x^n$; \item there exists $c \in R^{\circ}\cap {\mathfrak{b}}$ such that $cx^ng = 0$ for all $n \gg 0$; \item there exists $c \in R^{\circ}$ such that $cx^ng = 0$ for all $n \gg 0$. \end{enumerate} \end{thm}
\begin{proof} By Corollary \ref{ga.10}, we have $$ \mathcal{I}(G) = \left\{ {\mathfrak{p}}_1 \cap \ldots \cap {\mathfrak{p}}_t : t \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi \mbox{~and~} {\mathfrak{p}}_1, \ldots, {\mathfrak{p}}_t \in \mathcal{I}(G)\cap\Spec(R)\right\}. $$ Since $\mathcal{I}(G)$ is finite, it is immediate that $$ {\mathfrak{b}} := \bigcap_{\stackrel{\scriptstyle {\mathfrak{p}} \in \mathcal{I}(G)\cap\Spec(R)}{\height {\mathfrak{p}} \geq 1}}{\mathfrak{p}} $$ is the smallest ideal in $\mathcal{I}(G)$ of height greater than $0$. Since $\height {\mathfrak{b}} \geq 1$, so that there exists $c \in {\mathfrak{b}} \cap R^{\circ}$ by prime avoidance, it is clear that (i) $\Rightarrow$ (ii) and (ii) $\Rightarrow$ (iii).
(iii) $\Rightarrow$ (i) Let $n_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ and $c \in R^{\circ}$ be such that $cx^ng = 0$ for all $n \geq n_0$. Then, for all $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, we have $x^{n_0}cx^jg = c^{p^{n_0}}x^{n_0 + j}g = 0$, so that $cx^jg = 0$ because $G$ is $x$-torsion-free.
Therefore $g \in \ann_G(RcR[x,f])$. Now $\ann_G(RcR[x,f])\in \mathcal{A}(G)$: let ${\mathfrak{a}} \in \mathcal{I}(G)$ be the corresponding $G$-special $R$-ideal. Since $c \in {\mathfrak{a}}$, we must have $\height {\mathfrak{a}} \geq 1$. Therefore ${\mathfrak{b}} \subseteq {\mathfrak{a}}$, by definition of ${\mathfrak{b}}$, and so $$ g \in \ann_G(RcR[x,f]) = \ann_G({\mathfrak{a}} R[x,f]) \subseteq \ann_G({\mathfrak{b}} R[x,f]). $$ \end{proof}
Corollary \ref{ga.12c} and Theorem \ref{ga.13} give hints about how this work will be exploited, in Section \ref{tc} below, to obtain results in the theory of tight closure. The aim is to apply Corollary \ref{ga.12c} and Theorem \ref{ga.13} to $H^d_{{\mathfrak{m}}}(R)/\Gamma_x(H^d_{{\mathfrak{m}}}(R))$, where $(R,{\mathfrak{m}})$ is a local ring of dimension $d > 0$; the local cohomology module $H^d_{{\mathfrak{m}}}(R)$, which is well known to be Artinian as an $R$-module, carries a natural structure as a left $R[x,f]$-module. The passage between $H^d_{{\mathfrak{m}}}(R)$ and its $x$-torsion-free residue class $R[x,f]$-module $H^d_{{\mathfrak{m}}}(R)/\Gamma_x(H^d_{{\mathfrak{m}}}(R))$ is facilitated by the following extension, due to G. Lyubeznik, of a result of R. Hartshorne and R. Speiser. It shows that, when $R$ is local, an $x$-torsion left $R[x,f]$-module which is Artinian (that is, `cofinite' in the terminology of Hartshorne and Speiser) as an $R$-module exhibits a certain uniformity of behaviour.
\begin{thm} [G. Lyubeznik {\cite[Proposition 4.4]{Lyube97}}] \label{hs.4} {\rm (Compare Hartshorne--Speiser \cite[Proposition 1.11]{HarSpe77}.)} Suppose that $(R,{\mathfrak{m}})$ is local, and let $H$ be a left $R[x,f]$-module which is Artinian as an $R$-module. Then there exists $e \in \mathbb N_0$ such that $x^e\Gamma_x(H) = 0$. \end{thm}
Hartshorne and Speiser first proved this result in the case where $R$ is local and contains its residue field which is perfect. Lyubeznik applied his theory of $F$-modules to obtain the result without restriction on the local ring $R$ of characteristic $p$.
\begin{defi} \label{hslno} Suppose that $(R,{\mathfrak{m}})$ is local, and let $H$ be a left $R[x,f]$-module which is Artinian as an $R$-module. By the Hartshorne--Speiser--Lyubeznik Theorem \ref{hs.4}, there exists $e \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ such that $x^e\Gamma_x(H) = 0$: we call the smallest such $e$ the {\em Hartshorne--Speiser--Lyubeznik number\/}, or {\em HSL-number\/} for short, of $H$. \end{defi}
It will be helpful to have available an extension of this idea.
\begin{defi} \label{ga.14} We say that the left $R[x,f]$-module $H$ {\em admits an HSL-number\/} if there exists $e \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ such that $x^e\Gamma_x(H) = 0$; then we call the smallest such $e$ the {\em HSL-number\/} of $H$. \end{defi}
We have seen above in \ref{hs.4} and \ref{hslno} that if $H$ is Artinian as an $R$-module, then it admits an HSL-number. Note also that if $H$ is Noetherian as an $R$-module, then it admits an HSL-number, because $\Gamma_x(H)$ is an $R[x,f]$-submodule of $H$, and so is an $R$-submodule and therefore finitely generated.
\begin{cor} \label{ga.15} Suppose that the left $R[x,f]$-module $H$ admits an HSL-number $m_0$, and that the $x$-torsion-free left $R[x,f]$-module $G := H/\Gamma_x(H)$ has only finitely many $G$-special $R$-ideals. Let ${\mathfrak{b}}$ be the smallest ideal in $\mathcal{I}(G)$ of positive height (see\/ {\rm \ref{ga.13}}). For $h \in H$, the following statements are equivalent:
\begin{enumerate} \item $h$ is annihilated by $\bigoplus_{n\geq m_0}{\mathfrak{b}}^{[p^{m_0}]} x^n$; \item there exists $c \in R^{\circ}\cap {\mathfrak{b}}$ such that $cx^nh = 0$ for all $n \geq m_0$; \item there exists $c \in R^{\circ}\cap {\mathfrak{b}}$ such that $cx^nh = 0$ for all $n \gg 0$; \item there exists $c \in R^{\circ}$ such that $cx^nh = 0$ for all $n \gg 0$. \end{enumerate} \end{cor}
\begin{proof} Since ${\mathfrak{b}} \cap R^{\circ} \neq 0$ by prime avoidance, it is clear that (i) $\Rightarrow$ (ii), (ii) $\Rightarrow$ (iii) and (iii) $\Rightarrow$ (iv).
(iv) $\Rightarrow$ (i) Since $cx^n(h + \Gamma_x(H)) = 0$ in $G$ for all $n \gg 0$, it follows from Theorem \ref{ga.13} that $h + \Gamma_x(H)$ is annihilated by ${\mathfrak{b}} R[x,f]$. Therefore, for all $r \in {\mathfrak{b}}$ and $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, we have $rx^j(h + \Gamma_x(H)) = 0$, so that $rx^jh \in \Gamma_x(H)$ and $r^{p^{m_0}}x^{m_0+j}h = x^{m_0}rx^jh = 0$. Therefore $h \in \ann_H \left(\bigoplus_{n\geq m_0}{\mathfrak{b}}^{[p^{m_0}]} x^n\right)$. \end{proof}
\section{\sc Applications to tight closure} \label{tc}
The aim of this section is to apply results from Section \ref{ga} to the theory of tight closure in the local ring $(R,{\mathfrak{m}})$ of dimension $d > 0$. As was mentioned in Section \ref{ga}, we shall be concerned with the top local cohomology module $H^d_{{\mathfrak{m}}}(R)$, which has a natural structure as a left $R[x,f]$-module, and its $x$-torsion-free residue class module $H^d_{{\mathfrak{m}}}(R)/\Gamma_x(H^d_{{\mathfrak{m}}}(R))$. The (well-known) left $R[x,f]$-module structure carried by $H^d_{{\mathfrak{m}}}(R)$ is described in detail in \cite[2.1 and 2.3]{KS}.
\begin{rmd} \label{tc.1} Suppose that $(R,{\mathfrak{m}})$ is a local ring of dimension $d > 0$. The above-mentioned natural left $R[x,f]$-module structure carried by $H^d_{{\mathfrak{m}}}(R)$ is independent of any choice of a system of parameters for $R$. However, if one does choose a system of parameters $a_1, \ldots, a_d$ for $R$, then one can obtain a quite concrete representation of the local cohomology module $H^d_{{\mathfrak{m}}}(R)$ and, through this, an explicit formula for the effect of multiplication by the indeterminate $x \in R[x,f]$ on an element of $H^d_{{\mathfrak{m}}}(R)$.
Denote by $a_1, \ldots, a_d$ a system of parameters for $R$.
\begin{enumerate} \item Represent $H^d_{{\mathfrak{m}}}(R)$ as the $d$-th cohomology module of the \u{C}ech complex of $R$ with respect to $a_1, \ldots, a_d$, that is, as the residue class module of $R_{a_1 \ldots a_d}$ modulo the image, under the \u{C}ech complex `differentiation' map, of $\bigoplus_{i=1}^dR_{a_1 \ldots a_{i-1}a_{i+1}\ldots a_d}$. See \cite[\S 5.1]{LC}. We use `$\left[\phantom{=} \right]$' to denote natural images of elements of $R_{a_1\ldots a_d}$ in this residue class module. Note that, for $i \in \{1, \ldots, d\}$, we have $$ \left[\frac{a_i^k}{(a_1 \ldots a_d)^k}\right] = 0 \quad \mbox{~for all~} k \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi. $$
Denote the product $a_1 \ldots a_d$ by $a$. A typical element of $H^d_{{\mathfrak{m}}}(R)$ can be represented as $ \left[r/a^j\right]$ for some $r \in R$ and $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$; moreover, for $r, r_1 \in R$ and $j, j_1 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, we have $ \left[r/a^j\right] = \left[r_1/a^{j_1}\right] $ if and only if there exists $k \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ such that $k \geq \max\{j,j_1\}$ and $ a^{k-j}r - a^{k-j_1}r_1 \in (a_1^k, \ldots, a_d^k)R. $ In particular, if $a_1, \ldots, a_d$ form an $R$-sequence (that is, if $R$ is Cohen--Macaulay), then $ \left[r/a^j\right] = 0$ if and only if $r \in (a_1^j, \ldots, a_d^j)R$, by \cite[Theorem 3.2]{O'Car83}, for example.
\item The left $R[x,f]$-module structure on $H^d_{{\mathfrak{m}}}(R)$ is such that $$ x\left[\frac{r}{(a_1\ldots a_d)^j}\right] = \left[\frac{r^p}{(a_1\ldots a_d)^{jp}}\right] \quad \mbox{~for all~} r \in R \mbox{~and~} j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi. $$ The reader might like to consult \cite[2.3]{KS} for more details, and should in any case note that this left $R[x,f]$-module structure does not depend on the choice of system of parameters $a_1, \ldots, a_d$. \end{enumerate} \end{rmd}
\begin{rmk} \label{tc.1r} Let the situation and notation be as in \ref{tc.1}. Here we relate the left $R[x,f]$-module structure on $H := H^d_{{\mathfrak{m}}}(R)$ described in \ref{tc.1} to the tight closure in $H$ of its zero submodule. See \cite[Definition (8.2)]{HocHun90} for the definition of the tight closure in an $R$-module of one of its submodules. Let $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$.
\begin{enumerate} \item The $n$-th component $Rx^n$ of $R[x,f]$ is isomorphic, as an $(R,R)$-bimodule, to $R$ considered as a left $R$-module in the natural way and as a right $R$-module via $f^n$, the $n$-th power of the Frobenius ring homomorphism. Let $L$ be a submodule of the $R$-module $M$. It follows that an element $m \in M$ belongs to $L^*_M$, the {\em tight closure of $L$ in $M$\/}, if and only if there exists $c \in R^\circ$ such that $cx^n \otimes m$ belongs, for all $n \gg 0$, to the image of $R[x,f]\otimes_R L$ in $R[x,f]\otimes_R M$ under the map induced by inclusion.
\item Let $S$ be a multiplicatively closed subset of $R$. It is straightforward to check that there is an isomorphism of $R$-modules $$\gamma_n: Rx^n \otimes_R S^{-1}R \stackrel{\cong}{\longrightarrow} S^{-1}R$$ for which $\gamma_n ( bx^n \otimes (r/s) ) = br^{p^n}/s^{p^n}$ for all $b,r \in R$ and $s \in S$; the inverse of $\gamma_n$ satisfies $(\gamma_n)^{-1} (r/s) = rs^{p^n-1}x^n \otimes (1/s)$ for all $r \in R$ and $s \in S$.
\item Now represent $H := H^d_{{\mathfrak{m}}}(R)$ as the $d$-th cohomology module of the \u{C}ech complex of $R$ with respect to the system of parameters $a_1, \ldots, a_d$, as in \ref{tc.1}(i). We can use isomorphisms like that described in part (ii), together with the right exactness of tensor product, to see that (when we think of $H$ simply as an $R$-module) there is an isomorphism of $R$-modules $\delta_n: Rx^n \otimes_R H \stackrel{\cong}{\longrightarrow} H$ for which $$ \delta_n \left( bx^n \otimes \left[\frac{r}{(a_1\ldots a_d)^j}\right] \right) = \left[\frac{br^{p^n}}{(a_1\ldots a_d)^{jp^n}}\right] \quad \mbox{~for all~} b,r \in R \mbox{~and~} j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi. $$ Thus, in terms of the natural left $R[x,f]$-module structure on $H$, we have $\delta_n \left( bx^n \otimes h\right) = bx^nh$ for all $b \in R$ and $h \in H$.
\item It thus follows that, for $h \in H$, we have $h \in 0^*_{H}$ if and only if there exists $c \in R^{\circ}$ such that $cx^nh = 0$ for all $n \gg 0$.
\item Observe that $\Gamma_x(H) \subseteq 0^*_H$.
\item Suppose that $(R,{\mathfrak{m}})$ is Cohen--Macaulay, and use the notation of part (iii) again; write $a := a_1\ldots a_d$. Let $r \in R$, $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, and let $h := \left[r/a^j\right]$ in $H$. It follows from \ref{tc.1}(i) that $r \in ((a_1^j, \ldots, a_d^j)R)^*$ if and only if there exists $c \in R^{\circ}$ such that $cx^nh = 0$ for all $n \gg 0$. Thus, by part (iv) above, $r \in ((a_1^j, \ldots, a_d^j)R)^*$ if and only if $h \in 0^*_{H}$.
\item Let the situation and notation be as in part (vi) above. Then the $R$-homomorphism $\nu_j : R/(a_1^j, \ldots, a_d^j)R \longrightarrow H$ for which $\nu_j(r' + (a_1^j, \ldots, a_d^j)R) = \left[r'/a^j\right]$ for all $r'\in R$ is a monomorphism (by \ref{tc.1}(i)). Furthermore, the induced homogeneous $R[x,f]$-homomorphism $$ R[x,f]\otimes_R\nu_j : R[x,f]\otimes_R\left(R/(a_1^j, \ldots, a_d^j)R\right) \longrightarrow R[x,f]\otimes_RH $$ of graded left $R[x,f]$-modules is also a monomorphism: this is because a homogeneous element of $\Ker \left(R[x,f]\otimes_R\nu_j\right)$ must have the form $r'x^k\otimes(1 + (a_1^j, \ldots, a_d^j)R)$ for some $r' \in R$ and $k \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$; since $r'x^k\otimes\left[1/a^j\right] = 0$, it follows from \ref{tc.1r}(iii) and \ref{tc.1}(i) that $r' \in (a_1^{jp^k}, \ldots, a_d^{jp^k})R$, so that $r'x^k\otimes(1 + (a_1^j, \ldots, a_d^j)R) = 0$. \end{enumerate} \end{rmk}
\begin{lem} \label{tc.1s} Suppose that $(R,{\mathfrak{m}})$ is a local ring of dimension $d > 0$; set $H := H^d_{{\mathfrak{m}}}(R)$ and $G := H/\Gamma_x(H)$. Let $h \in H$. Then the following statements are equivalent:
\begin{enumerate} \item $h \in 0^*_{H}$; \item $h + \Gamma_x(H) \in 0^*_{G}$; \item there exists $c \in R^{\circ}$ such that $cx^n(h+\Gamma_x(H)) = 0$ in $G$ for all $n \gg 0$; \item there exists $c \in R^{\circ}$ such that $cx^n(h+\Gamma_x(H)) = 0$ in $G$ for all $n \geq 0$. \end{enumerate} \end{lem}
\begin{proof} Let $m_0$ denote the HSL-number of $H$ (see\/ {\rm \ref{hslno}}).
(i) $\Rightarrow$ (ii) This is immediate from the fact that $0^*_{H} \subseteq (\Gamma_x(H))^*_H$ once it is recalled from \cite[Remark (8.4)]{HocHun90} that $h + \Gamma_x(H) \in 0^*_G$ if and only if $h \in (\Gamma_x(H))^*_H$.
(ii) $\Rightarrow$ (iii) Suppose that $h + \Gamma_x(H) \in 0^*_{G}$, so that $h \in (\Gamma_x(H))^*_H$. Under the isomorphism $\delta_n: Rx^n \otimes_R H \stackrel{\cong}{\longrightarrow}H$ of \ref{tc.1r}(iii) (where $n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$), the image of $Rx^n \otimes_R \Gamma_x(H)$ is mapped into $\Gamma_x(H)$. Therefore there exists $c \in R^{\circ}$ such that $cx^nh \in\Gamma_x(H)$ for all $n \gg 0$, that is, such that $cx^n(h+\Gamma_x(H)) = 0$ in $G$ for all $n \gg 0$.
(iii) $\Rightarrow$ (iv) Suppose that there exist $c \in R^{\circ}$ and $n_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ such that $cx^n(h+\Gamma_x(H)) = 0$ in $G$ for all $n \geq n_0$. Then, for all $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, we have $$ x^{n_0}cx^j(h+\Gamma_x(H)) = c^{p^{n_0}}x^{n_0 + j}(h+\Gamma_x(H)) = 0 \quad \mbox{~in~} G. $$ Since $G$ is $x$-torsion-free, we see that $cx^j(h+\Gamma_x(H)) = 0$ for all $j \geq 0$.
(iv) $\Rightarrow$ (i) Suppose that there exists $c \in R^{\circ}$ such that $cx^n(h+\Gamma_x(H)) = 0$ in $G$ for all $n \geq 0$. Then $cx^nh \in \Gamma_x(H)$ for all $n \geq 0$, so that $x^{m_0}cx^nh = 0$ for all $n \geq 0$. This implies that $c^{p^{m_0}}x^{m_0 + n}h = 0$ for all $n \geq 0$, so that $h \in 0^*_{H}$ by \ref{tc.1r}(iv). \end{proof}
\begin{defi} \label{tc.1t} The {\em weak parameter test ideal $\sigma'(R)$ of $R$\/} is defined to be the ideal generated by $0$ and all weak parameter test elements for $R$. (By a `weak parameter test element' for $R$ we mean a $p^{w_0}$-weak parameter test element for $R$ for some $w_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$.)
It is easy to see that each element of $\sigma'(R) \cap R^{\circ}$ is a weak parameter test element for $R$. \end{defi}
The next theorem is one of the main results of this paper.
\begin{thm} \label{tc.2} Let $(R,{\mathfrak{m}})$ (as in\/ {\rm \ref{nt.1}}) be a Cohen--Macaulay local ring of dimension $d > 0$. Set $H := H^d_{{\mathfrak{m}}}(R)$, a left $R[x,f]$-module which is Artinian as an $R$-module; let $m_0$ be its HSL-number (see\/ {\rm \ref{hslno}}), and let $q_0 := p^{m_0}$.
Set $G := H/\Gamma_x(H)$, an $x$-torsion-free left $R[x,f]$-module. By\/ {\rm \ref{ga.12c}} and\/ {\rm \ref{ga.13}}, there exists a (uniquely determined) smallest ideal ${\mathfrak{b}}$ of height at least $1$ in the set $\mathcal{I}(G)$ of $G$-special $R$-ideals.
Let $c$ be any element of ${\mathfrak{b}} \cap R^{\circ}$. Then $c^{q_0}$ is a $q_0$-weak parameter test element for $R$. In particular, $R$ has a $q_0$-weak parameter test element. In fact, the weak parameter test ideal $\sigma'(R)$ of $R$ satisfies ${\mathfrak{b}}^{[q_0]} \subseteq \sigma'(R) \subseteq {\mathfrak{b}}$. \end{thm}
\begin{note} It should be noted that, in Theorem \ref{tc.2}, it is not assumed that $R$ is excellent. There are examples of Gorenstein local rings of characteristic $p$ which are not excellent: see \cite[p.\ 260]{HMold}. \end{note}
\begin{proof} We have to show that, for an arbitrary parameter ideal ${\mathfrak{a}}$ of $R$ and $r \in {\mathfrak{a}}^*$, we have $c^{q_0}r^{p^n} \in {\mathfrak{a}}^{[p^n]}$ for all $n \geq m_0$. In the first part of the proof, we establish this in the case where ${\mathfrak{a}}$ is an ideal ${\mathfrak{q}}$ generated by a full system of parameters $a_1, \ldots, a_d$ for $R$.
Let $r \in {\mathfrak{q}}^*$, so that there exists $\widetilde{c} \in R^{\circ}$ such that $\widetilde{c}r^{p^n} \in {\mathfrak{q}}^{[p^n]}$ for all $n \gg 0$. Use $a_1, \ldots, a_d$ in the notation of \ref{tc.1}(i) for $H^d_{{\mathfrak{m}}}(R) = H$, and write $a := a_1\ldots a_d$. We have $ \widetilde{c}x^n\left[r/a\right] = \left[\widetilde{c}r^{p^n}\!/a^{p^n}\right] = 0 $ in $H$ for all $n \gg 0$. Set $h := \left[r/a\right] \in H$. Thus $\widetilde{c}x^nh = 0$ for all $n \gg 0$.
It therefore follows from Corollary \ref{ga.15} that $h$ is annihilated by $ \bigoplus_{n\geq m_0}{\mathfrak{b}}^{[p^{m_0}]} x^n$, so that, in particular, $c^{p^{m_0}}x^nh = 0$ for all $n \geq m_0$. Hence, in $H$, $$ \left[\frac{c^{q_0}r^{p^{n}}}{(a_1 \ldots a_d)^{p^{ n}}}\right] = c^{p^{m_0}}x^{n}\left[\frac{r}{a_1 \ldots a_d}\right] = c^{p^{m_0}}x^n h = 0 \quad \mbox{~for all~} n \geq m_0. $$ Since $R$ is Cohen--Macaulay, we can now deduce from \ref{tc.1}(i) that $c^{q_0}r^{p^n} \in {\mathfrak{q}}^{[p^n]}$ for all $n \geq m_0$, as required (for ${\mathfrak{q}}$).
Now let ${\mathfrak{a}}$ be an arbitrary parameter ideal of $R$. A proper ideal in a Cohen--Macaulay local ring is a parameter ideal if and only if it can be generated by part of a system of parameters. In view of the first part of this proof, we can, and do, assume that $\height {\mathfrak{a}} < d$. There exist a system of parameters $a_1, \ldots, a_d$ for $R$ and an integer $i \in \{0, \ldots, d-1\}$ such that ${\mathfrak{a}} = (a_1,\ldots, a_i)R$. Let $r \in {\mathfrak{a}}^*$. Then, for each $v \in \mathbb N$, we have $r \in ((a_1,\ldots, a_i,a_{i+1}^v, \ldots, a_d^v)R)^*$, and, since $a_1,\ldots, a_i,a_{i+1}^v, \ldots, a_d^v$ is a system of parameters for $R$, it follows from the first part of this proof that $$ c^{q_0}r^{p^n} \in ((a_1,\ldots, a_i,a_{i+1}^v, \ldots, a_d^v)R)^{[p^n]} = \left(a_1^{p^n},\ldots, a_i^{p^n},a_{i+1}^{vp^n}, \ldots, a_d^{vp^n}\right)\!R \quad \mbox{~for all~} n \geq m_0. $$ Therefore, for all $n \geq m_0$, $$ c^{q_0}r^{p^n} \in \bigcap_{v \in \mathbb N} \left(a_1^{p^n},\ldots, a_i^{p^n},a_{i+1}^{vp^n}, \ldots, a_d^{vp^n}\right)\!R \subseteq \bigcap_{v \in \mathbb N} \left({\mathfrak{a}}^{[p^{n}]} + {\mathfrak{m}}^{vp^{n}} \right) = {\mathfrak{a}}^{[p^{n}]} $$ by Krull's Intersection Theorem. This shows that $c^{q_0}$ is a $q_0$-weak parameter test element for $R$, so that, since ${\mathfrak{b}}$ can be generated by elements in ${\mathfrak{b}} \cap R^{\circ}$, it follows that ${\mathfrak{b}}^{[q_0]} \subseteq \sigma'(R)$.
Now let $c \in \sigma'(R)\cap R^{\circ}$; we suppose that $c \not\in {\mathfrak{b}}$ and seek a contradiction. Thus ${\mathfrak{b}} \subset {\mathfrak{b}} + Rc$. Let $L : = \ann_G({\mathfrak{b}} R[x,f])$ and $L' := \ann_G(({\mathfrak{b}} + Rc)R[x,f])$, two special annihilator submodules of the $x$-torsion-free left $R[x,f]$-module $G$. Since $L$ corresponds to the $G$-special $R$-ideal ${\mathfrak{b}}$, we must have $L' \subset L$, since otherwise we would have $$ ({\mathfrak{b}} + Rc)R[x,f] \subseteq \grann_{R[x,f]}L' = \grann_{R[x,f]}L = {\mathfrak{b}} R[x,f]. $$ Therefore there exists $h \in H$ such that $h + \Gamma_x(H)$ is annihilated by ${\mathfrak{b}} R[x,f]$ but not by $({\mathfrak{b}} + Rc) R[x,f]$. Since $\height {\mathfrak{b}} \geq 1$, it follows from Lemma \ref{tc.1s} that $h \in 0^*_H$.
Choose a system of parameters $a_1, \ldots, a_d$ for $R$; use $a_1, \ldots, a_d$ in the notation of \ref{tc.1}(i) for $H^d_{{\mathfrak{m}}}(R) = H$, and write $a := a_1\ldots a_d$. There exist $r \in R$ and $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ such that $h = [r/a^j]$. By \ref{tc.1r}(vi), we have $r \in ((a_1^j, \ldots, a_d^j)R)^*$. Since $c$ is a weak parameter test element for $R$, we see that $cr^{p^n} \in (a_1^{jp^n}, \ldots, a_d^{jp^n})R$ for all $n \gg 0$, so that $cx^nh = 0$ for all $n \gg 0$. Thus there is some $n_0 \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$ such that $cx^n(h + \Gamma_x(H)) = 0$ in $G$ for all $n \geq n_0$. Therefore, for all $j \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi$, we have $$ x^{n_0}cx^j(h + \Gamma_x(H)) = c^{p^{n_0}}x^{n_0 + j}(h + \Gamma_x(H)) = 0, $$ so that $cx^j(h + \Gamma_x(H)) = 0$ because $G$ is $x$-torsion-free. Thus $h + \Gamma_x(H)$ is annihilated by $RcR[x,f]$ as well as by ${\mathfrak{b}} R[x,f]$. This is a contradiction.
Therefore ${\mathfrak{b}}^{[q_0]} \subseteq \sigma'(R) \subseteq {\mathfrak{b}}$, since $\sigma'(R)$ can be generated by its elements that lie in $R^{\circ}$. \end{proof}
Use $R'$ to denote $R$ (as in\/ {\rm \ref{nt.1}}) regarded as an $R$-module by means of $f$. With this notation, $f : R \longrightarrow R'$ becomes a homomorphism of $R$-modules. Recall that, when $(R,{\mathfrak{m}})$ is a local ring of dimension $d > 0$, we say that $R$ is {\em $F$-injective\/} precisely when the induced homomorphisms $ H^i_{{\mathfrak{m}}}(f) : H^i_{{\mathfrak{m}}}(R) \longrightarrow H^i_{{\mathfrak{m}}}(R')$ are injective for all $i = 0, \ldots, d$. See R. Fedder and K-i. Watanabe \cite[Definition 1.7]{FW87} and the ensuing discussion.
\begin{cor} \label{tc.3} Let $(R,{\mathfrak{m}})$ be an $F$-injective Cohen--Macaulay local ring of dimension $d > 0$. The left $R[x,f]$-module $H := H^d_{{\mathfrak{m}}}(R)$ is $x$-torsion-free. By Theorem\/ {\rm \ref{ga.13}}, there exists a (uniquely determined) smallest ideal ${\mathfrak{b}}$ of height at least $1$ in the set $\mathcal{I}(H)$ of $H$-special $R$-ideals.
Let $c$ be any element of ${\mathfrak{b}} \cap R^{\circ}$. Then $c$ is a parameter test element for $R$. In fact, ${\mathfrak{b}}$ is the parameter test ideal of $R$ (see\/ {\rm \cite[Definition 4.3]{Smith95}}). \end{cor}
\begin{note} It should be noted that, in Corollary \ref{tc.3}, it is not assumed that $R$ is excellent. \end{note}
\begin{proof} With the notation of Theorem \ref{tc.2}, the HSL-number $m_0$ of $H$ is $0$ when $R$ is $F$-injective, and so $q_0 = 1$ and $G \cong H$ in this case. By Theorem \ref{tc.2}, each element $c \in {\mathfrak{b}} \cap R^{\circ}$ is a $q_0$-weak parameter test element for $R$, that is, a parameter test element for $R$. Since $R$ has a parameter test element, its parameter test ideal $\sigma(R)$ is equal to the ideal of $R$ generated by all parameter test elements. By Theorem \ref{tc.2}, we therefore have $ {\mathfrak{b}} = {\mathfrak{b}}^{[q_0]} \subseteq \sigma(R) \subseteq \sigma'(R) \subseteq {\mathfrak{b}}. $ \end{proof}
\begin{cor} \label{tc.4} Let $(R,{\mathfrak{m}})$ be an $F$-injective Gorenstein local ring of dimension $d > 0$. The left $R[x,f]$-module $H := H^d_{{\mathfrak{m}}}(R)$ is $x$-torsion-free. By Theorem\/ {\rm \ref{ga.13}}, there exists a (uniquely determined) smallest ideal ${\mathfrak{b}}$ of height at least $1$ in the set $\mathcal{I}(H)$ of $H$-special $R$-ideals.
Let $c$ be any element of ${\mathfrak{b}} \cap R^{\circ}$. Then $c$ is a test element for $R$. In fact, ${\mathfrak{b}}$ is the test ideal of $R$. \end{cor}
\begin{note} It should be noted that, in Corollary \ref{tc.4}, it is not assumed that $R$ is excellent. \end{note}
\begin{proof} This follows immediately from Corollary \ref{tc.3} once it is recalled that an $F$-injective Cohen--Macaulay local ring is reduced and that a parameter test element for a reduced Gorenstein local ring $R$ of characteristic $p$ is automatically a test element for $R$: see the proof of \cite[Proposition 4.1]{Hunek98}. \end{proof}
\section{\sc Special $R$-ideals and Enescu's $F$-stable primes} \label{en}
The purpose of this section is to establish connections between the work in \S \ref{ga} and \S \ref{tc} above and F. Enescu's $F$-stable primes of an $F$-injective Cohen--Macaulay local ring $(R,{\mathfrak{m}})$, defined in \cite[\S 2]{Enesc03}.
\begin{ntn} \label{en.1} Throughout this section, $(R,{\mathfrak{m}})$ will be assumed to be a Cohen--Macaulay local ring of dimension $d > 0$, and we shall let $a_1, \ldots, a_d$ denote a fixed system of parameters for $R$, and set ${\mathfrak{q}} := (a_1, \ldots, a_d)R$. We shall use $a_1, \ldots, a_d$ in the notation of \ref{tc.1}(i) for $H := H^d_{{\mathfrak{m}}}(R)$, and write $a := a_1\ldots a_d$.
For each $b \in R$, we define (following Enescu \cite[Definition 1.1]{Enesc03}) the ideal ${\mathfrak{q}}(b)$ by $$ {\mathfrak{q}}(b) := \left\{ c \in R : cb^{p^n} \in {\mathfrak{q}}^{[p^n]} \mbox{~for all~} n \gg 0 \right\}. $$ (Actually, Enescu only made this definition when $b \not\in {\mathfrak{q}}$; however, the right-hand side of the above display is equal to $R$ when $b \in {\mathfrak{q}}$, and there is no harm in our defining ${\mathfrak{q}}(b)$ to be $R$ in this case.) In view of \ref{tc.1}(i), the ideal ${\mathfrak{q}}(b)$ is equal to the ultimate constant value of the ascending chain $({\mathfrak{b}}_n)_{n \in \relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}$ of ideals of $R$ for which $\bigoplus_{n\in\relax\ifmmode{\mathbb N_{0}}\else$\mathbb N_{0}$\fi}{\mathfrak{b}}_n x^n = \grann_{R[x,f]}R[x,f][b/a]$, the graded annihilator of the $R[x,f]$-submodule of $H$ generated by $[b/a]$.
Now consider the special case in which $R$ is (also) $F$-injective. Then the left $R[x,f]$-module $H$ is $x$-torsion-free, and so it follows from Lemma \ref{nt.5} that, for each $b \in R$, the ideal ${\mathfrak{q}}(b)$ is radical and $\grann_{R[x,f]}R[x,f][b/a] = {\mathfrak{q}}(b)R[x,f]$; thus ${\mathfrak{q}}(b)$ is an $H$-special $R$-ideal. We again follow Enescu and set $$ Z_{{\mathfrak{q}},R} := \{ {\mathfrak{q}}(b) : b \in R \setminus {\mathfrak{q}} \}. $$ \end{ntn}
Enescu proved, in \cite[Theorem 2.1]{Enesc03}, that (when $(R,{\mathfrak{m}})$ is Cohen--Macaulay and $F$-injective) the set of maximal members of $Z_{{\mathfrak{q}},R}$ is independent of the choice of ${\mathfrak{q}}$, is finite, and consists of prime ideals. The next theorem shows that the set of maximal members of $Z_{{\mathfrak{q}},R}$ is actually equal to the set of maximal members of $\mathcal{I}(H) \setminus \{R\}$: we saw in Lemma \ref{ga.11} that this set consists of prime ideals, and in Corollary \ref{ga.12c} that it is finite.
\begin{thm} \label{en.2} Let the situation and notation be as in\/ {\rm \ref{en.1}}, and suppose that the Cohen--Macaulay local ring $(R,{\mathfrak{m}})$ is $F$-injective. Then the set of maximal members of $Z_{{\mathfrak{q}},R}$ is equal to the set of maximal members of $\mathcal{I}(H) \setminus \{R\}$. \end{thm}
\begin{proof} The comments in \ref{en.1} show that $Z_{{\mathfrak{q}},R} \subseteq \mathcal{I}(H)$; clearly, no member of $Z_{{\mathfrak{q}},R}$ can be equal to $R$. It is therefore sufficient for us to show that a maximal member ${\mathfrak{p}}$ of $\mathcal{I}(H) \setminus \{R\}$ must belong to $Z_{{\mathfrak{q}},R}$.
Let $L \in \mathcal{A}(H)$ be the special annihilator submodule of $H$ corresponding to ${\mathfrak{p}}$. Now $H$ is an Artinian $R$-module: let $h$ be a non-zero element of the socle of $L$. By Lemma \ref{ga.11}, we have $\grann_{R[x,f]}R[x,f]h = {\mathfrak{p}} R[x,f]$. However, for each $j \in \mathbb N$, we have $R[1/a^j] \cong R/(a_1^j, \ldots, a_d^j)R$, by \ref{tc.1r}(vii), so that $$ \Hom_R(R/{\mathfrak{m}},R[1/a^j]) \cong \Hom_R(R/{\mathfrak{m}},R/(a_1^j, \ldots, a_d^j)R) \cong \Ext^d_R(R/{\mathfrak{m}},R) $$ by \cite[Lemma 1.2.4]{BH}. It follows that it is possible to write $h$ in the form $h = [r/a]$ for some $r \in R$, and therefore ${\mathfrak{p}} = {\mathfrak{q}}(r) \in Z_{{\mathfrak{q}},R}$. \end{proof}
\end{document}
|
arXiv
|
{
"id": "0605329.tex",
"language_detection_score": 0.6678232550621033,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[A class of nonlocal elliptic problems in the half space with a hole] {Existence of positive solution for a class of nonlocal elliptic problems in the half space with a hole} \author{Xing Yi} \thanks{*College of Mathematics and Computer Science, Key Laboratory of High Performance Computing and Stochastic Information Processing (Ministry of Education of China), Hunan Normal University, Changsha, Hunan 410081, P. R. China ([email protected])} \maketitle
\vskip 0.3in {\bf Abstract}\quad This work concerns with the existence of solutions for the following class of nonlocal elliptic problems \begin{eqnarray}\label{eq:0.1} &&\left\{\begin{array}{l}
(-\Delta)^{s} u+u=|u|^{p-2} u \text { in } \Omega_{r} \\ u \geq 0 \quad \text { in }\Omega_{r} \text { and } u \neq 0 \\ u=0 \quad \mathbb{R}^{N} \backslash \Omega_{r} \end{array}\right., \end{eqnarray} involving the fractional Laplacian operator $(-\Delta)^{s},$ where $s \in(0,1), N>2 s$, $\Omega_{r}$ is the half space with a hole in $\mathbb{R}^N$ and $p \in\left(2,2_{s}^{*}\right) .$ The main technical approach is based on variational and topological methods. \vskip 0.1in \noindent{\it Keywords:}\quad Nonlocal elliptic problems, Positive high energy solution, Half space with a hole
\noindent {\bf AMS} classification: 58J05, 35J60.
\renewcommand{\thesection.\arabic{equation}}{\thesection.\arabic{equation}} \setcounter{equation}{0}\Section*{1. Introduction} \setcounter{section}{1}\setcounter{equation}{0} Let $\mathbb{R}_{+}^{N}=\left\{\left(x^{\prime}, x_{N}\right) \in \mathbb{R}^{N-1} \times \mathbb{R} \mid 0<x_{N}<\infty\right\}$ be the upper half space. $\Omega_{r}$ is an unbounded smooth domain such that \[\overline{\Omega_{r}}\subset \mathbb{R}_{+}^{N},\]
and \[ \mathbb{R}_{+}^{N}\setminus\overline{\Omega_{r}}\subset B_{\rho}(a_{r})\subset \mathbb{R}_{+}^{N}\ \] with $a_{r}=(a,r)\in \mathbb{R}_{+}^{N}$.
Indeed, $\Omega_r$ is the upper half space with a hole.
We consider the following fractional elliptic problem: \begin{eqnarray}\label{eq:1.1} &&\left\{\begin{array}{l}
(-\Delta)^{s} u+u=|u|^{p-2} u \text { in } \Omega \\ u \geq 0 \quad \text { in }\Omega_{r} \text { and } u \neq 0 \\ u=0 \quad \mathbb{R}^{N} \backslash \Omega \end{array}\right., \end{eqnarray} where $ \Omega=\Omega_r,\ s \in(0,1), N>2 s$, $p \in\left(2,2_{s}^{*}\right),$ where $2_{s}^{*}=\frac{2 N}{N-2 s}$ is the fractional critical Sobolev exponent and $(-\Delta)^{s}$ is the classical fractional Laplace operator.
When $s \nearrow 1^{-},$ problem(\ref{eq:1.1}) is related to the following elliptic problem \begin{eqnarray}\label{eq:1.2}
-\triangle u+u=|u|^{p-1}u,\quad x\in \Omega, \quad u\in H_{0}^{1}(\Omega) \end{eqnarray} . When $\Omega$ is a bounded domain, by applying the compactness of the embedding $H_{0}^{1}(\Omega)\hookrightarrow L^{p}(\Omega), 1< p<\frac{2N}{N-2}$, there is a positive solution of (\ref{eq:1.2}). If $\Omega$ is an unbounded domain, we can not obtain a solution for problem (\ref{eq:1.2}) by using Mountain-Pass Theorem directly because the embedding $H_{0}^{1}(\Omega)\hookrightarrow L^{p}(\Omega), 1< p<\frac{2N}{N-2}$ is not compactness. However, if $\Omega=\mathbb{R}^{N}$, Berestycki-Lions \cite{21},
proved that there is a radial positive solution of equation (\ref{eq:1.2})
by applying the compactness of the embedding $H_{r}^{1}(\mathbb{R}^{N})\hookrightarrow L^{p}(R^{N}),2<p<\frac{2N}{N-2}$, where $H_r^{1}(\mathbb{R}^{N})$ consists of the radially symmetric functions in $H^{1}(\mathbb{R}^{N})$.
By the P.L.Lions's Concentration-Compactness Principle \cite{7}, there exists an unique positive solution for problem (\ref{eq:1.2}) in $\mathbb{R}^{N}$.
By moving Plane method, Gidas-Ni-Nirenberg \cite{3} also proved that every positive solution of equation
\begin{eqnarray}\label{eq:1.3}
-\triangle u+u=|u|^{p-1}u,\quad x\in \mathbb{R}^{N}, \quad u\in H^{1}(\mathbb{R}^{N}) \end{eqnarray}
is radially symmetric with respect to some point in $\mathbb{R}^{N}$ satisfying \begin{eqnarray} u(r)re^{r}=\gamma+o(1)\ as \ r \rightarrow \infty. \end{eqnarray} Kwong \cite{188} proved that the positive solution of (\ref{eq:1.3}) is unique up to translations.
In fact, Esteban and Lions \cite{1} proved that there is not any nontrivial solution of equation (\ref{eq:1.2}) when $\Omega$ is an Esteban-Lions domain (for example $\mathbb{R}_+^3$). Thus, we want to change the topological property of the domain $\Omega$ to look for a solution of problem (\ref{eq:1.2}). Wang \cite{4} proved that if $\rho$ is sufficiently small and $z_{0N}\rightarrow\infty$, then Eq(\ref{eq:1.2}) admits a positive higher energy solution in $\mathbb{R}_{+}^{N} \backslash \overline{B_{\rho}\left(z_{0}^{\prime}, z_{0 N}\right)}$. Such problem has been extensively studied in recent years, see for instance, \cite{9,55} and references therein. From the above researches,
we believed that the existence of the solution to the equation (\ref{eq:1.2}) will be affected by the topological property of the domain $\Omega$.
Recently, the case $s \in(0,1)$ has received a special attention, because involves the fractional Laplacian operator $(-\Delta)^{s}$, which arises in a quite natural way in many different contexts, such as, among the others, the thin obstacle problem, optimization, finance, phase transitions, stratified materials, anomalous diffusion, crystal dislocation, soft thin films, semipermeable membranes, flame propagation, conservation laws, ultra-relativistic limits of quantum mechanics, quasigeostrophic flows, multiple scattering, minimal surfaces, materials science and water waves, for more detail see \cite{x1,e16,x2,D21,x4}.
When $ \Omega \subset \mathbb{R}^{N}$ is an exterior domain, i.e. an unbounded domain with smooth boundary $\partial \Omega \neq \emptyset$ such that $\mathbb{R}^{N} \backslash \Omega$ is bounded, $s \in(0,1), N>2 s$, $p \in\left(2,2_{s}^{*}\right),$ the above problem has been studied by O. Alves, Giovanni Molica Bisci , César E. Torres Ledesma in \cite{21} proving that (1.1) does not have a ground state solution. This fact represents a serious difficulty when dealing with this kind of nonlinear fractional elliptic phenomena. More precisely, the authors analyzed the behavior of Palais-Smale sequences and showed a precise estimate of the energy levels where the Palais-Smale condition fails, which made it possible to show that the problem (1.1) has at least one positive solution, for $\mathbb{R}^{N} \backslash \Omega$ small enough. A key point in the approach explored in \cite{16,9} is the existence and uniqueness, up to a translation, of a positive solution $Q$ of the limit problem associated with (\ref{eq:1.1}) given by \begin{eqnarray}\label{eq:1.6}
(-\Delta)^{s} u+u=|u|^{p-2} u \text { in } \mathbb{R}^{N}, \end{eqnarray}
for every $p \in\left(2,2_{s}^{*}\right)$. Moreover, $Q$ is radially symmetric about the origin and monotonically decreasing in $|x| .$ On the contrary of the classical elliptic case, the exponential decay at infinity is not used in order to prove the existence of a nonnegative solution for (\ref{eq:1.1}).
When $\Omega \subset \mathbb{R}^{N}$ is $\mathbb{R}_+^N$, by moving Plane method, Wenxiong Chen, Yan Li and Pei Ma \cite{11d6}(p123 Theorem 6.8.3 ) proved that there is no nontrivial solution of problem (\ref{eq:1.1}).
It is interesting in considering the existence of the high energy equation for the problem (\ref{eq:1.1}) in the half space with a hole in $\mathbb{R}^{N}$. \begin{theorem} There is $\rho_{0}>0,r_{0}>0$ such that if $0<\rho\leq\rho_{0} $ and $r \geq r_{0}$, then there is a positive solution of equation (\ref{eq:0.1}). \end{theorem}
The paper is organized as follows. In section 2, we give some preliminary results. The Compactness lemma will be given in Section 3. At last, we give the proof of Theorem 1.1.
\setcounter{equation}{0}\Section{Some preliminary results } For $s \in(0,1)$ and $N>2 s,$ the fractional Sobolev space of order $s$ on $\mathbb{R}^{N}$ is defined by $$
H^{s}\left(\mathbb{R}^{N}\right):=\left\{u \in L^{2}\left(\mathbb{R}^{N}\right): \int_{\mathbb{R}^{N}} \int_{\mathbb{R}^{N}} \frac{|u(x)-u(z)|^{2}}{|x-z|^{N+2 s}} d z d x<\infty\right\} $$ endowed with the norm $$
\|u\|_{s}:=\left(\int_{\mathbb{R}^{N}}|u(x)|^{2} d x+\int_{\mathbb{R}^{N}} \int_{\mathbb{R}^{N}} \frac{|u(x)-u(z)|^{2}}{|x-z|^{N+2 s}} d z d x\right)^{1 / 2} .$$ We recall the fractional version of the Sobolev embeddings (see \cite{q16}).
\begin{theorem} Let $s \in(0,1),$ then there exists a positive constant $C=C(N, s)>0$ such that $$
\|u\|_{L^{2_{s}^{*}}\left(\mathbb{R}^{N}\right)}^{2} \leq C \int_{\mathbb{R}^{N}} \int_{\mathbb{R}^{N}} \frac{|u(x)-u(y)|^{2}}{|x-y|^{N+2 s}} d y d x $$ and then $H^{s}\left(\mathbb{R}^{N}\right) \hookrightarrow L^{q}\left(\mathbb{R}^{N}\right)$ is continuous for all $q \in\left[2,2_{s}^{*}\right] .$ Moreover, if $\Theta \subset \mathbb{R}^{N}$ is a bounded domain, we have that the embedding $H^{s}\left(\mathbb{R}^{N}\right) \hookrightarrow L^{q}(\Theta)$ is compact for any $q \in\left[2,2_{s}^{*}\right) .$ \end{theorem} Hereafter, we denote by $X_{0}^{s} \subset H^{s}\left(\mathbb{R}^{N}\right)$ the subspace defined by $$ X_{0}^{s}:=\left\{u \in H^{s}\left(\mathbb{R}^{N}\right): u=0 \text { a.e. in } \mathbb{R}^{N} \backslash \Omega\right\}. $$
We endow $X_{0}^{s}$ with the norm $\|\cdot\|_{s}$. Moreover we introduce the following norm $$
\|u\|:=\left(\int_{\Omega_{r}}|u(x)|^{2} d x+\iint_{\mathcal{Q}} \frac{|u(x)-u(z)|^{2}}{|x-z|^{N+2 s}} d z d x\right)^{\frac{1}{2}} $$
where $\mathcal{Q}:=\mathbb{R}^{2 N} \backslash\left(\Omega_{r}^{c} \times \Omega_{r}^{c}\right)$. We point out that $\|u\|_{s}=\|u\|$ for any $u \in X_{0}^{s}$. Since $\partial \Omega$ is bounded and smooth, by [\cite{D21}, Theorem 2.6], we have the following result. \begin{theorem}
The space $C_{0}^{\infty}(\Omega)$ is dense in $\left(X_{0}^{s},\|\cdot\|\right) .$ \end{theorem} In what follows, we denote by $H^{s}(\Omega)$ the usual fractional Sobolev space endowed with the norm $$
\|u\|_{H^{s}}:=\left(\int_{\Omega}|u(x)|^{2} d x+\int_{\Omega} \int_{\Omega} \frac{|u(x)-u(z)|^{2}}{|x-z|^{N+2 s}} d z d x\right)^{\frac{1}{2}}. $$ Related to these fractional spaces, we have the following properties
{\bf Proposition 2.3.}The following assertions hold true:
(i) If $v \in X_{0}^{s}$, we have that $v \in H^{s}(\Omega)$ and $$
\|v\|_{H^{s}} \leq\|v\|_{s}=\|v\| . $$
(ii) Let $\Theta$ an open set with continuous boundary. Then, there exists a positive constant $\mathfrak{C}=$ $\mathfrak{C}(N, s),$ such that $$
\|v\|_{L^{2_{s}^{*}}(\Theta)}^{2}=\|v\|_{L^{2^{*}}\left(\mathbb{R}^{N}\right)}^{2} \leq \mathfrak{C} \iint_{\mathbb{R}^{2 N}} \frac{|v(x)-v(z)|^{2}}{|x-z|^{N+2 s}} d z d x $$ for every $v \in X_{0}^{s} ;$ see [\cite{e16}, Theorem 6.5$]$.
From now on, $M_{\infty}$ denotes the following constant \begin{eqnarray}\label{eq:1.s6}
M_{\infty}:=\inf \left\{\|u\|_{s}^{2}: u \in H^{s}\left(\mathbb{R}^{N}\right) \text { and } \int_{\mathbb{R}^{N}}|u(x)|^{p} d x=1\right\}, \end{eqnarray} which is positive by Theorem 2.1. Furthermore, for any $v \in H^{s}\left(\mathbb{R}^{N}\right)$ and $z \in \mathbb{R}^{N},$ we set the function $$ v^{z}(x):=v(x+z). $$ Then, by doing the change of variable $\tilde{x}=x+z$ and $\tilde{y}=y+z,$ it is easily seen that
$$\left\|v^{z}\right\|_{s}^{2}=\|v\|_{s}^{2} \text { as well as }\left\|v^{z}\right\|_{L^{p}\left(\mathbb{R}^{N}\right)}=\|v\|_{L^{p}\left(\mathbb{R}^{N}\right)} .$$ Arguing as in \cite{b16} the following result holds true.
\begin{theorem} Let $\left\{u_{n}\right\} \subset H^{s}\left(\mathbb{R}^{N}\right)$ be a minimizing sequence such that $$
\left\|u_{n}\right\|_{L^{p}\left(\mathbb{R}^{N}\right)}=1 \text { and }\left\|u_{n}\right\|_{s}^{2} \rightarrow M_{\infty} \text { as } n \rightarrow+\infty. $$ Then, there is a sequence $\left\{y_{n}\right\} \subset \mathbb{R}^{N}$ such that $\left\{u_{n}^{y_{n}}\right\}$ has a convergent subsequence, and so, $M_{\infty}$ is attained. \end{theorem} As a byproduct of the above result the next corollary is obtained.
{\bf Corollary 1} There is $v \in H^{s}\left(\mathbb{R}^{N}\right)$ such that $\|v\|_{s}=M_{\infty}$ and $\|v\|_{L^{p}\left(\mathbb{R}^{N}\right)}=1 .$
Let $\varphi$ be a minimizer of $(\ref{eq:1.s6}),$ that is $$
\varphi \in H^{s}\left(\mathbb{R}^{N}\right), \quad \int|\varphi|^{p} d x=1 \text { and } M_{\infty}=\|\varphi\|_{s}^{2}. $$ Take \begin{eqnarray} &&\xi \in C^{\infty}(\mathbb{R}^{+},\mathbb{R}),\eta\in C^{\infty}(\mathbb{R},\mathbb{R}), \end{eqnarray} such that \[\xi(t)=\left\{
\begin{array}{ll} 0,0\leq t\leq \rho,\\[2mm] 1,t\geq 2\rho, \end{array} \right.\] \[ \eta(t)=\left\{
\begin{array}{ll} 0,t\leq 0,\\[2mm] 1,t\geq 1, \end{array} \right.\] and \[0\leq \zeta\leq 1,0\leq \eta\leq 1. \]
Now, we define
\[f_{y}(x)=\xi(|x-a_{r}|)\eta(x_{N})\varphi(x-y),\] and
\[\Psi_{y}(x)=\frac{f_{y}(x)}{\left\|f_{y}\right\|_{L^{p}\left(\mathbb{R}^{N}\right)}}=c_{y} f_{y}(x) \text { where } c_{y}=\frac{1}{\left\|f_{y}\right\|_{L^{p}\left(\mathbb{R}^{N}\right)}} .\] Throughout this section we endow $X_{0}^{s}$ with the norm $$
\|u\|:=\left(\iint_{\mathcal{Q}} \frac{|u(x)-u(y)|^{2}}{|x-y|^{n+2 s}} d y d x+\int_{\Omega_{r}}|u|^{2} d x\right)^{1 / 2} $$ and denote by $M>0$ the number \begin{eqnarray}\label{eq:1w.sq6}
M:=\inf \left\{\|u\|^{2}: u \in X_{0}^{s}, \int_{\Omega_{r}}|u(x)|^{p} d x=1\right\} . \end{eqnarray} \begin{lemma}\label{lm:2.4} Let $y=(y^{'},y_{N})$ , we have\\
(1)$\|f_{y}-\varphi(x-y)\|_{L^{p}(\mathbb{R}^{N})}=o(1)$, $|y-a_{r}|\rightarrow \infty$, and $y_{N}\rightarrow +\infty$, or $y_{N} \rightarrow \infty$ and $\rho \rightarrow 0$;
(2)$\|f_{y}-\varphi(x-y)\|=o(1),|y-a_{r}|\rightarrow \infty$, and $y_{N}\rightarrow +\infty$, or $y_{N} \rightarrow +\infty$ and $\rho \rightarrow 0$.\label{11} \end{lemma} \begin{proof} Similarly as \cite{9,4}, we have
(i) After the change of variables $z=x-y,$ one has \[ \begin{array}{ll}
\ \|f_{y}-\varphi(x-y)\|^{p}_{L^{p}(\mathbb{R}^{N})}
&=\int_{\mathbb{R}^{N}}|\xi(|x-a_{r}|)\eta(x_{N})-1|^{p}|\varphi(x-y)|^{p}\ dx\\[2mm]
&=\int_{\mathbb{R}^{N}}|\xi(|x+y-a_{r}|)\eta(x_{N}+y)-1|^{p}|\varphi(z)|^{p}\ dz \end{array} \]
Let $g_{y}(z)=|\xi(|x+y-a_{r}|)\eta(x_{N}+y)-1|^{p}|\varphi(z)|^{p}.$ Since $|y-a_{r}|\rightarrow \infty$, and $y_{N}\rightarrow +\infty$, it follows that $$ g_{y}(z) \rightarrow 0 \text { a.e. in } \mathbb{R}^{N} $$ Now, taking into account that $$
g_{y}(z)=|\xi(|x+y-a_{r}|)\eta(x_{N}+y)-1|^{p}|\varphi(z)|^{p} \leq 2^{p}|\varphi(z)|^{p} \in L^{1}\left(\mathbb{R}^{N}\right), $$ the Lebesgue's dominated convergence theorem yields $$
\int_{\mathbb{R}^{N}} g_{y}(z) d z \rightarrow 0 \text { as } |y-a_{r}|\rightarrow \infty, \ y_{N}\rightarrow +\infty. $$ Therefore $$
\|f_{y}-\varphi(x-y)\|_{L^{p}(\mathbb{R}^{N})}=o(1), |y-a_{r}|\rightarrow \infty, \ y_{N}\rightarrow +\infty $$ \[ \begin{array}{ll}
\ \|f_{y}-\varphi(x-y)\|^{p}_{L^{p}(\mathbb{R}^{N})}
&=\int_{B_{2\rho(a_{r})}\cup\{x_{N}\leq 1\}}|\xi(|x-a_{r}|)\eta(x_{N})-1|^{p}|\varphi(x-y)|^{p}\ dx\\[2mm]
&=\int_{B_{2\rho(a_{r})}}|\xi(|x-a_{r}|)\eta(x_{N})-1|^{p}|\varphi(x-y)|^{p}\ dx\\[2mm]
&+\int_{\{x_{N}\leq 1\}}|\xi(|x-a_{r}|)\eta(x_{N})-1|^{p}|\varphi(x-y)|^{p}\ dx \end{array} \] and \[ \begin{array}{ll}
\ \int_{B_{2\rho(a_{r})}}|\xi(|x-a_{r}|)\eta(x_{N})-1|^{p}|\varphi(x-y)|^{p}\ dx &\leq C mesB_{2\rho(a_{r})}\max_{x\in\mathbb{R}^{N}}\varphi(x)\rightarrow 0\ as\ \rho \rightarrow 0, \end{array} \] \[ \begin{array}{ll}
\ \int_{\{x_{N}\leq 1\}}|\xi(|x-a_{r}|)\eta(x_{N})-1|^{p}|\varphi(x-y)|^{p}\ dx
&=\int_{\{x_{N}\leq 1\}}|\eta(x_{N})-1|^{p}|\varphi(x-y)|^{p}\ dx\\[2mm]
&=\int_{\{z_{3}\leq y_{N}+1\}}|\eta (x_{N}+y_{N})-1|^{p}|\varphi(z)|^{p}\ dz\\[2mm] & \rightarrow 0 \ as\ y_{N}\rightarrow +\infty,\ \rho \rightarrow 0. \end{array} \] Therefore $$
\|f_{y}-\varphi(x-y)\|_{L^{p}(\mathbb{R}^{N})}=o(1)\ y_{N} \rightarrow \infty\ and \ \rho \rightarrow 0. $$
(ii) Now, we claim that $$
\int_{\mathbb{R}^{N}} \int_{\mathbb{R}^{N}} \frac{\left|(\xi(|x-a_{r}|)\eta(x_{N})-1) \varphi\left(x-y\right)-(\xi(|z-a_{r}|)\eta(z_{N})-1) \varphi\left(z-y\right)\right|^{2}}{|x-z|^{N+2 s}} d z d x=o_{n}(1) $$ Indeed, let $$
\Upsilon_{u}(x, y):=\frac{u(x)-u(z)}{|x-z|^{\frac{N}{2}+s}} $$ Then, after the change of variables $\tilde{x}=x-y$ and $\tilde{y}=z-y,$ one has \[\begin{array}{ll}
\int_{\mathbb{R}^{N}} \int_{\mathbb{R}^{N}} \frac{\left|(\xi(|x-a_{r}|)\eta(x_{N})-1) \varphi\left(x-y\right)-(\xi(|z-a_{r}|)\eta(z_{N})-1) \varphi\left(z-y\right)\right|^{2}}{|x-z|^{N+2 s}} d z d x
& =\int_{\mathbb{R}^{N}} \int_{\mathbb{R}^{N}}\left|\Upsilon_{n}(x, z)\right|^{2} d z d x.
\end{array} \] where $$
\Phi_{n}(x, z):=\frac{\left(\xi(|x+y-a_{r}|)\eta(x_{N}+y_{N}))-1\right) \varphi(x)-\left(\xi(|z+y-a_{r}|)\eta(z_{N}+y_{N}))-1\right) \varphi(z)}{|x-z|^{\frac{N}{2}+s}} $$
Recalling that $|y-a_{r}|\rightarrow \infty$, and $y_{N}\rightarrow +\infty$, we also have $$ \Upsilon_{n}(x, y) \rightarrow 0 \text { a.e. in } \mathbb{R}^{N} \times \mathbb{R}^{N} $$ On the other hand, a direct application of the mean value theorem yields \begin{eqnarray}\label{ezq:0b.1} \begin{aligned}
\left|\Upsilon_{n}(x, z)\right| & \leq\left|\left(\xi(|x+y-a_{r}|)\eta(x_{N}+y_{N}))-1\right) \| \Upsilon_{\varphi}(x, y)\right|+|\varphi(z)|\left|\Upsilon_{1-\xi\eta}\left(x+y, z+y\right)\right| \\
& \leq\left|\Upsilon_{\varphi}(x, z)\right|+\frac{C|\varphi(z)|}{|x-z|^{\frac{N}{2}+s-1}} \chi_{B(z, 1)}(x)+\frac{2|\varphi(z)|}{|x-z|^{\frac{N}{2}+s}} \chi_{B^{c}(z, 1)}(x), \end{aligned} \end{eqnarray} for almost every $(x, z) \in \mathbb{R}^{N} \times \mathbb{R}^{N} .$ Now, it is easily seen that the right hand side in (\ref{ezq:0b.1}) is $L^{2}$ -integrable. Thus, By the Lebesgue's dominated convergence theorem and i, it follows that
$$\|f_{y}-\varphi(x-y)\|=o(1),|y-a_{r}|\rightarrow \infty,\ and\ y_{N}\rightarrow +\infty$$
By i,we have $\|f_{y}-\varphi(x-y)\|_{L^{2}(\mathbb{R}^{N})}=o(1)$, $|y-a_{r}|\rightarrow \infty$, and $y_{N}\rightarrow +\infty$, or $y_{N} \rightarrow \infty$ and $\rho \rightarrow 0$; \[\begin{array}{ll}
\ \|f_{y}-\varphi(x-y)\|^{2}
&= \int_{\mathbb{R}^{N}} \int_{\mathbb{R}^{N}} \frac{\left|\left(f_{y}-\varphi(\cdot-y)\right)(x)-\left(f_{y}-\varphi(\cdot-y)\right)(z)\right|^{2}}{|x-z|^{N+2 s}} d z d x \\[2mm]
&+\int_{\mathbb{R}^{N}}\left|f_{y}(x)-\varphi(x-y)\right|^{2} d x . \end{array} \] Setting $$
I_{1}:=\iint_{\mathbb{R}^{2 N}} \frac{(\xi(|x-a_{r}|)\eta(x_{N}))-\xi(|z-a_{r}|)\eta(z_{3}))|^{2}|\varphi(x-y)|^{2}}{|x-z|^{N+2 s}} d z d x $$ and $$
I_{2}:=\iint_{\mathbb{R}^{2 N}} \frac{\left.\left|\xi(|z-a_{r}|)\eta(z_{3}))-1\right|^{2} \mid \varphi(x-y)-\varphi(z-y)\right)\left.\right|^{2}}{|x-z|^{N+2 s}} d z d x $$ the following inequality holds $$
\int_{\mathbb{R}^{N}} \int_{\mathbb{R}^{N}} \frac{\left|\left(f_{y}-\varphi(\cdot-y)\right)(x)-\left(f_{y}-\varphi(\cdot-y)\right)(z)\right|^{2}}{|x-z|^{N+2 s}} d z d x \leq I_{1}+I_{2}. $$ Moreover, by definition of $\xi,$ we also have $$
\xi(|z+y-a_{r}|)\eta(z_{3}+y_{N})-\left.1\right|^{2} \frac{|\varphi(x)-\varphi(z)|^{2}}{|x-z|^{N+2 s}} \leq 4 \frac{|\varphi(x)-\varphi(z)|^{2}}{|x-z|^{N+2 s}} \in L^{1}\left(\mathbb{R}^{N} \times \mathbb{R}^{N}\right) $$ and
$$\xi(|z+y-a_{r}|)\eta(z_{3}+y_{N})-1|^{2} \frac{|\varphi(x)-\varphi(z)|^{2}}{|x-y|^{N+2 s}} \rightarrow 0 \text { a.e. in } \mathbb{R}^{N} \times \mathbb{R}^{N}$$ as $y_{N} \rightarrow \infty$ and $\rho \rightarrow 0$. Hence, the Lebesgue's theorem ensures that $$ I_{2}\rightarrow 0 \text { as } \rho \rightarrow 0. $$ Now, by [\cite{m21}, Lemma 2.3], for every $y \in \mathbb{R}^{N}$, one has $$
I_{1}=\iint_{\mathbb{R}^{2 N}} \frac{(\xi(|x-a_{r}|)\eta(x_{N}))-\xi(|z-a_{r}|)\eta(z_{3}))|^{2}|\varphi(x-y)|^{2}}{|x-z|^{N+2 s}} d z d x \rightarrow 0 \text { as } \rho \rightarrow 0 $$. Therefore $$
\|f_{y}-\varphi(x-y)\|=o(1), y_{N} \rightarrow +\infty , \rho \rightarrow 0. $$
\end{proof}
\begin{lemma}
The equality $M_{\infty}=M$ holds true. Hence, there is no $u \in X_{0}^{s}$ such that $\|u\|^{2}=M$ and $\|u\|_{L^{p}\left(\mathbb{R}^{N}\right)}=1$, and so, the minimization problem (\ref{eq:1w.sq6}) does not have solution. \end{lemma}
\begin{proof} The proof is similar to \cite{4}, and we only give a sketch here. By Proposition 2.3 - part (i) it follows that $$ M_{\infty} \leq M $$
Take a sequence $y^{n}$ in $\Omega_{r}$ such that\\
\[|y^{n}-a_{r}|\rightarrow \infty,and\ y_{N}^{n} \rightarrow +\infty\ as\ n\rightarrow \infty.\] Then by lemma \ref{lm:2.4}, we have
\[\|f_{y^{n}}-\varphi(x-y^{n})\|_{L^{p}(R^{N})}=o(1),|y^{n}-a_{r}|\rightarrow \infty,and\ y^{n}_{3}\rightarrow +\infty,\]
\[\|f_{y^{n}}-\varphi(x-y^{n})\|_{s}=o(1),|y^{n}-a_{r}|\rightarrow \infty, and \ y^{n}_{3}\rightarrow +\infty,\]
\[c_{y^{n}}=\frac{1}{\left\|f_{y^{n}}\right\|_{L^{p}\left(\mathbb{R}^{N}\right)}} \rightarrow 1,|y^{n}-a_{r}|\rightarrow \infty,and\ y^{n}_{3}\rightarrow +\infty\] Now, since $\varphi$ is a minimizer of $(\ref{eq:1.s6}),$ one has $$
\left\|f_{y^{n}}\right\|_{s}^{2}=\left\|\varphi\left(\cdot-y_{n}\right)\right\|_{s}^{2}+o_{n}(1)=\|\varphi\|_{s}^{2}+o_{n}(1)=M_{\infty}+o_{n}(1) $$ Similar arguments ensure that $$
\left\|\Psi_{n}\right\|_{s}^{2}=\left\|\Psi_{n}\right\|^{2}=M_{\infty}+o_{n}(1) $$
and$$\left\|\Psi_{n}\right\|_{L^{p}\left(\mathbb{R}^{N}\right)}=1$$ So\[ M \leq M_{\infty}\]. We then conclude that $M= M_{\infty}$. Now, suppose by contradiction that there is $v_{0} \in X_{0}^{s}$ satisfying $$
\left\|v_{0}\right\|=M \text { and }\left\|v_{0}\right\|_{L^{p}(\Omega)}=1 $$
Without loss of generality, we can assume that $v_{0} \geq 0$ in $\Omega$. Note that by $M= M_{\infty}$, since $v_{0} \in$ $H^{s}\left(\mathbb{R}^{N}\right)$ and $\left\|v_{0}\right\|=\left\|v_{0}\right\|_{s},$ it follows that $v_{0}$ is a minimizer for $(\ref{eq:1.s6}),$ and so, a solution of roblem \begin{eqnarray}\label{eq:0.v1} \left\{\begin{aligned} (-\Delta)^{s} u+u &=M_{\infty} u^{p-1} \text { in } \mathbb{R}^{N} \\ u & \in H^{s}\left(\mathbb{R}^{N}\right) . \end{aligned}\right. \end{eqnarray} Therefore, by the maximum principle we get that $v_{0}>0$ in $\mathbb{R}^{N}$, which is impossible, because $v_{0}=0$ in $\mathbb{R}^{N} \backslash \Omega_{r}$. This completes the proof. \end{proof}
\setcounter{equation}{0}\Section{A Compactness lemma } In this section we prove a compactness result involving the energy functional $I: X_{0}^{s} \rightarrow \mathbb{R}$ associated to the main problem (\ref{eq:0.1}) and given by $$
I(u):=\frac{1}{2}\left(\iint_{\mathcal{Q}} \frac{|u(x)-u(y)|^{2}}{|x-y|^{N+2 s}} d y d x+\int_{\Omega_{r}}|u|^{2} d x\right)-\frac{1}{p} \int_{\Omega_{r}}|u|^{p} d x $$ In order to do this, we consider the problem \begin{eqnarray}\label{eq:0.x1} \left\{\begin{aligned}
(-\Delta)^{s} u+u &=|u|^{p-2} u \text { in } \mathbb{R}^{N} \\ u & \in H^{s}\left(\mathbb{R}^{N}\right) \end{aligned}\right. \end{eqnarray} whose energy functional $I_{\infty}: H^{s}\left(\mathbb{R}^{N}\right) \rightarrow \mathbb{R}$ has the form $$
I_{\infty}(u):=\frac{1}{2}\left(\int_{\mathbb{R}^{N}} \int_{\mathbb{R}^{N}} \frac{|u(x)-u(y)|^{2}}{|x-y|^{N+2 s}} d y d x+\int_{\mathbb{R}^{N}}|u|^{2} d x\right)-\frac{1}{p} \int_{\mathbb{R}^{N}}|u|^{p} d x. $$ With the above notations we are able to prove the following compactness result. \begin{lemma}\label{eq:0cb.1} Let $\left\{u_{n}\right\} \subset X_{0}^{s}$ be a sequence such that \begin{eqnarray}\label{eq:0bx.1} I\left(u_{n}\right) \rightarrow c \text { and } I^{\prime}\left(u_{n}\right) \rightarrow 0 \text { as } n \rightarrow \infty. \end{eqnarray}
Then there are a nonnegative integer $k, k$ sequences $\left\{y_{n}^{i}\right\}$ of points of the form $\left(x_{n}^{\prime}, m_{n}+1 / 2\right)$ for integers $m_{n}, i=1,2, \cdots, k, $ $u_{0} \in X_{0}^{s}$ solving equation (\ref{eq:0.1}) and nontrivial functions $u^{1}, \cdots, u^{k}$ in $H^{s}\left(\mathbb{R}^{N}\right)$ solving equation (\ref{eq:0.x1}). Moreover there is a subsequence $\left\{u_{n}\right\}$ satisfying
$$ \begin{array}{l} \text { (1) }u_{n}(x)=u^{0}(x)+u^{1}\left(x-x_{n}^{1}\right)+\cdots+u^{k}\left(x-x_{n}^{k}\right)+o(1) strongly, where x_{n}^{i}=y_{n}^{1}+\cdots+y_{n}^{i} \rightarrow \infty,\\ i=1,2, \cdots, k \\
\text { (2) }\left\|u_{n}\right\|^{2}=\left\|u^{0}\right\|_{\Omega_{r}}^{2}+\left\|u^{1}\right\|^{2}+\cdots+\left\|u^{k}\right\|^{2}+o(1) \\ \text { (3) } I\left(u_{n}\right)=I\left(u^{0}\right)+I_{\infty}\left(u^{1}\right)+\cdots+I_{\infty}\left(u^{k}\right)+o(1) \end{array} $$ If $u_{n} \geqslant 0$ for $n=1,2, \cdots,$ then $u^{1}, \cdots, u^{k}$ can be chosen as positive solutions, and $u^{0} \geqslant 0$ \end{lemma}
\begin{proof}
See \cite{21,4}.
\end{proof}
{\bf COROLLARY 2} Let $\left\{u_{n}\right\} \subset M_{\Omega_{r}}$ satisfy $\|u_{n}\|_{\Omega_{r}}^{2}=c+o(1)$ and $M<c<2^{(p-2) / p} M.$ Then $\left\{u_{n}\right\}$ contains a strongly convergent subsequence. \begin{proof}
See \cite{21,4}.
\end{proof}
\renewcommand{\thesection.\arabic{equation}}{\thesection.\arabic{equation}} \setcounter{equation}{0}\Section{ Proof of Theorem 1} Set $$ \chi(t)=\left\{\begin{array}{ll} 1 & \text { if } 0 \leqslant t \leqslant 1 \\ \frac{1}{t} & \text { if } 1 \leqslant t<\infty \end{array}\right. $$ and define $\beta: H^{s}\left(\mathbb{R}^{N}\right) \rightarrow \mathbb{R}^{N}$\cite{4} by $$
\beta(u)=\int_{\mathbb{R}^{N}} u^{2}(x) \chi(|x|) x d x. $$ For $r \geqslant r_{1}$, let $$
\begin{array}{l}V_{r}=\left\{\left.u \in H_{0}^{1}(\Omega_{r}\right)|\int_{\Omega_{r}}|u|^{p}=1, \beta(u)=a_{r}\right\} \\ c_{r}=\inf _{u \in V_{r}}\|u\|_{\Omega_{r}}^{2}.\end{array} $$ \begin{lemma} $c_{r}>M$. \end{lemma} ProoF: It is easy to see that $c_{r} \geqslant M.$ Suppose $c_{r}=\alpha .$ Take a sequence $\left\{v_{m}\right\} \subset X_{0}^{s}$ s.t. $$ \begin{array}{l}
\left\|v_{m}\right\|_{L_{p} \left(\Omega_{r}\right)}=1, \beta\left(v_{m}\right)=a_{r} \quad \text { for } \quad m=1,2, \cdots, \\
\left\|v_{m}\right\|^{2}=M+o(1). \end{array} $$ Let $u_{m}=M^{1 /(p-2)} v_{m}$ for $m=1,2, \cdots$. Then $$I^{\prime}\left(u_{n}\right)=o_{n}(1) \text { in }\left(X_{0}^{s}\right)^{*}$$ and $$I\left(u_{n}\right)=\left(\frac{1}{2}-\frac{1}{p}\right) M^{\frac{p}{p-2}}+o_{n}(1).$$ By the maximum principle, $\left\{u_{m}\right\}$ does not contain any convergent subsequence. $\mathrm{By}$ lemma (\ref{eq:0cb.1}) there is a sequence $\left\{x_{m}\right\}$ of the form $\left(x_{m}^{\prime}, m+\frac{1}{2}\right)$ for integers $m$ such that $$ \begin{array}{c}
\left|x_{m}\right| \longrightarrow \infty \\ u_{m}(x)=\varphi\left(x-x_{m}\right)+o(1) \text { strongly. } \end{array} $$ Since $\varphi$ is radially symmetric, we may take $m$ to be positive. Next, we consider the following sets $$ \mathbb{R}_{+}^{N}:=\left\{x \in \mathbb{R}^{N}:\left\langle x, x_{m}\right\rangle>0\right\} \text { and }\mathbb{R}_{-}^{N}:=\mathbb{R}^{N} \backslash\mathbb{R}_{+}^{N}. $$
We may assume that $$ \begin{aligned}
\left|x_{m}\right| \geqslant 4 \text { from } m=1,2, \cdots . \text { Now } & \\
\qquad\left\langle\beta\left(\varphi\left(x-x_{m}\right)\right), x_{m}\right\rangle=& \int_{\mathbb{R}^{N}} \varphi^{2}\left(x-x_{m}\right) \chi(|x|)\left\langle x, x_{m}\right\rangle d x \\
=& \int_{\mathbb{R}_{+}^{N}} \varphi^{2}\left(x-x_{m}\right) \chi(|x|)\left\langle x, x_{m}\right\rangle d x \\
&+\int_{\left(\mathbb{R}_{-}^{N}\right)} \varphi^{2}\left(x-x_{m}\right) \chi(|x|)\left\langle x, x_{m}\right\rangle d x \\
\geqslant & \int_{B_{1}\left(x_{m}\right)} \varphi^{2}\left(x-x_{m}\right) \chi(|x|)\left\langle x, x_{m}\right\rangle d x \\
&+\int_{\mathbb{R}_{-}^{N}} \varphi^{2}\left(x-x_{m}\right) \chi(|x|)\left\langle x, x_{m}\right\rangle d x. \end{aligned} $$ Note that there are $c_{1}>0, c_{2}>0$ such that for $x \in B_{1}\left(x_{m}\right),$ we have $$ \begin{array}{l} \varphi^{2}\left(x-x_{m}\right) \geqslant c_{1} \\
\quad\left\langle x, x_{m}\right\rangle \geqslant c_{2}|x|\left|x_{m}\right| \text { for } m=1,2, \cdots . \end{array} $$ Thus $$ \begin{aligned}
\int_{B_{1}\left(x_{m}\right)} \varphi^{2}\left(x-x_{m}\right) \chi(|x|)\left\langle x, x_{m}\right\rangle d x & \geqslant c_{1} c_{2} \int_{B_{1}\left(x_{m}\right)} \chi(|x|)|x|\left|x_{m}\right| d x \\
& \geqslant c_{3}\left|x_{m}\right|^{N+1}, \quad c_{3}>0 \quad \text { a constant. } \end{aligned} $$ Recalling that for each $x\in\mathbb{R}_{-}^{N}$, $$
\left|x-y_{n}\right| \geq|x| $$ it follows that $$
\left|u\left(x-y_{n}\right)\right|^{2} \chi(|x|)|x| \leq R|u(|x|)|^{2} \in L^{1}\left(\mathbb{R}^{N}\right)(R>0) $$ (see \cite{21} lemma 4.3).
This fact, combined with the limit $$
u\left(x-y_{n}\right) \rightarrow 0 \text { as }\left|y_{n}\right| \rightarrow+\infty $$ implies that $$
\int_{\mathbb{R}_{-}^{N}}\left|u\left(x-y_{n}\right)\right|^{2} \chi(|x|)|x| d x=o_{n}(1). $$ We conclude that $$ \begin{aligned}
M^{1 /(p-2)}\left|a_{r}\right| & \geqslant\left\langle\beta\left(u_{m}\right), \frac{x_{m}}{\left|x_{m}\right|}\right\rangle \\
&=\left\langle\beta\left(\varphi\left(x-x_{m}\right)\right), \frac{x_{m}}{\left|x_{m}\right|}\right\rangle+o(1) \\
& \geqslant c_{3}\left|x_{m}\right|^{N}+o(1) \end{aligned} $$ a contradiction. Thus $c_{r}>M$.
{\bf REMARK 1} By Lemma \ref{lm:2.4}(1), there is $r_{1}>0$ such that $$
\frac{1}{2} \leqslant\left\|f_{y}\right\|_{L^{p}\left(\Omega_{r}\right)} \leqslant \frac{3}{2} $$
where $r \geqslant r_{1}$ and $\left|y-a_{r}\right| \geqslant r / 2$ and $y_{N} \geqslant r / 2$.
{\bf REMARK 2}. By Lemma \ref{lm:2.4}(2), there is $r_{2} \geqslant r_{1}$ such that
$M<\left\|\Psi_{y}\right\|^{2}<\frac{c_{r}+M}{2}$
where $r \geqslant r_{2}$ and $\left|y-a_{r}\right| \geqslant r / 2$ and $y_{N} \geqslant r / 2$. \begin{lemma}\label{eq:1.f3} There is $r_{3} \geqslant r_{2}$ such that if $r \geqslant r_{3},$ then $$ \left\langle\beta\left(\varphi_{y}\right), y\right\rangle>0 \quad \text { for } \quad y \in \partial\left(B_{r / 2}\left(a_{r}\right)\right). $$ \end{lemma}
\begin{proof} By lemma \ref{lm:2.4}, we have $ 2 / 3 \leqslant c_{y} \leqslant 2 .$ For $r \geqslant r_{2},$ let $$ \begin{aligned}
A_{\left((3 / 8) r_{1}(5 / 8) r\right)} &=\left\{x \in \mathbb{R}^{N}\left|\frac{3}{8} r \leqslant\right| x-a_{r} \mid \leqslant \frac{5}{8} r\right\}, \\ \mathbb{R}_{+}^{N}(y) &=\left\{x \in \mathbb{R}^{N} \mid\langle x, y\rangle>0\right\} \\ \mathbb{R}_{-}^{N}(y) &=\left\{x \in \mathbb{R}^{N} \mid\langle x, y\rangle<0\right\} \end{aligned} $$ $$\begin{aligned}
\left\langle\beta\left(\varphi_{y}\right), y\right\rangle=c_{y} &\left[\int_{\mathbb{R}_{+}^{N}(y)} \xi^{2}\left(\left|x-a_{r}\right|\right) \eta^{2}\left(x_{N}\right) \varphi^{2}(x-y) \chi(|x|)\langle x, y\rangle d x\right.\\
&\left.+\int_{\mathbb{R}_{-}^{N}(y)} \xi^{2}\left(\left|x-a_{r}\right|\right) \eta^{2}\left(x_{N}\right) \varphi^{2}(x-y) \chi(|x|)\langle x, y\rangle d x\right] \\
\geqslant \frac{2}{3} &\left[\int_{A((3 / 8) r,(5 / 8) r)} \varphi^{2}(x-y) \chi(|x|)\langle x, y\rangle\right.\\
&+\int_{\mathbb{R}_{-}^{N}(y)} \varphi^{2}(x-y) \chi(|x|)\langle x, y\rangle d x \end{aligned}$$ $$\begin{aligned}
\int_{A((3 / 8) r,(5 / 8) r)} \varphi^{2}(x-y) \chi(|x|)\langle x, y\rangle d x & \geqslant c_{6} \int_{A((3 / 8) r,(5 / 8) r)} \chi(|x|)|x||y| d x \text { for } c_{6}>0 \\
& \geqslant c_{6}|y|\left[\left(\frac{5}{8} r\right)^{N}-\left(\frac{3}{8} r\right)^{N}\right] \\ & \geqslant c_{7} r^{N+1} \text { for } c_{7}>0 . \\
\int_{R_{-}^{N}(y)} \varphi^{2}(x-y) \chi(|x|)\langle x, y\rangle d x =o_{n}(1). \end{aligned}$$
Therefore, there is $r_{3} \geqslant r_{2},$ such that if $r \geqslant r_{3},\left|y-a_{r}\right|=r / 2$ $$ \left\langle\beta\left(\Psi_{y}\right), y\right\rangle \geqslant c_{7} r^{N+1}-o_{n}(1)>0. $$ This completes the proof.
\end{proof}
By Lemma \ref{lm:2.4} and Lemma \ref{eq:1.f3}, fix $\rho_{0}>0, r_{0} \geqslant r_{3}$ such that if $0<\rho \leqslant \rho_{0}, r \geqslant r_{0}$ then $\left\|\varphi_{y}\right\|_{\Omega_{r}}^{2}<2^{(p-2) / p} \alpha$ for $y \in \overline{B_{r / 2}}\left(a_{r}\right) .$ From now on, fix $\rho_{0}, r_{0},$ for $r \geqslant r_{0} .$ Let $$ \begin{array}{l}
B=\left\{\Psi_{y}|| y-a_{r} \mid \leqslant \frac{r}{2}\right\} \\
\Gamma=\left\{h \in C\left(V_{r}, V_{r}\right) \mid h(u)=u \quad \text { if } \quad\|u\|^{2}<\frac{c_{r}+\alpha}{2}\right\}. \end{array} $$ \begin{lemma} $$h(B) \cap V_{r} \neq \emptyset \text { for each } h \in \Gamma.$$ \end{lemma}
\begin{proof} Let $h \in \Gamma$ and $H(x)=\beta \circ h \circ \varphi_{x}: \mathbb{R}^{N} \rightarrow \mathbb{R}^{N}$. Consider the homotopy, for $0 \leqslant t \leqslant 1$ $$ F(t, x)=(1-t) H(x)+t I(x) \quad \text { for } \quad x \in \mathbb{R}^{N}. $$ If $x \in \partial\left(B_{r / 2}\left(a_{r}\right)\right),$ then, by Remark 8 and Lemma 9, $$ \begin{array}{c} \left\langle\beta\left(\Psi_{x}\right), x\right\rangle>0 \\
\alpha<\left\|\Psi_{x}\right\|^{2}<\frac{c_{r}+\alpha}{2}. \end{array} $$ Then $$ \begin{aligned} \langle F(t, x), x\rangle &=\langle(1-t) H(x), x\rangle+\langle t x, x\rangle \\ &=(1-t)\left\langle\beta\left(\Psi_{x}\right), x\right\rangle+t\langle x, x\rangle \\ &>0. \end{aligned} $$ Thus $F(t, x) \neq 0$ for $x \in \partial\left(B_{r / 2}\left(a_{r}\right)\right) .$ By the homotopic invariance of the degree $$ d\left(H(x), B_{r / 2}\left(a_{r}\right), a_{r}\right)=d\left(I, B_{r / 2}\left(a_{r}\right), a_{r}\right)=1. $$ There is $x \in B_{r / 2}\left(a_{r}\right)$ such that $$ a_{r}=H(x)=\beta\left(h \circ \Psi_{x}\right). $$ Thus $h(B) \cap V_{r} \neq \emptyset$ for each $h \in \Gamma$. Now we are in the position to prove Theorem A: Consider the class of mappings $$
F=\left\{h \in C\left(\overline{B_{r / 2}\left(a_{r}\right)}\right), H^{1}\left(R_{N}\right):\left.h\right|_{\partial B_{r / 2}\left(a_{r}\right)}=\Psi_{y}\right\} $$ and set $$
c=\inf _{h \in F} \frac{\sup }{y \in B_{r / 2}\left(a_{r}\right)}\|h(y)\|_{\Omega_{r}}^{2}. $$ It follows from the above Lemmas, with the appropriate choice of $r$ that $$
\alpha<c_{r}=\inf _{u \in V_{\gamma}}\|u\|_{\Omega_{r}}^{2} \leqslant c<2^{(p-2) / p} \alpha $$ and $$
\max _{\partial B_{r / 2}\left(a_{r}\right)}\|h(y)\|_{\Omega_{r}}^{2}<\max _{B_{r / 2}\left(a_{r}\right)}\|h(y)\|_{\Omega_{r}}^{2}. $$ Theorem 1,1 then follows by applying the version of the mountain pass theorem from Brezis-Nirenberg \cite{d6}.
\end{proof} \vskip 0.3in
\end{document}
|
arXiv
|
{
"id": "2104.12940.tex",
"language_detection_score": 0.5151960253715515,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{On finite factors of centralizers of parabolic subgroups in Coxeter groupsootnotetext{MSC2000: 20F55 (primary), 20E34 (secondary)} \begin{abstract} It has been known that the centralizer $Z_W(W_I)$ of a parabolic subgroup $W_I$ of a Coxeter group $W$ is a split extension of a naturally defined reflection subgroup by a subgroup defined by a $2$-cell complex $\mathcal{Y}$. In this paper, we study the structure of $Z_W(W_I)$ further and show that, if $I$ has no irreducible components of type $A_n$ with $2 \leq n < \infty$, then every element of finite irreducible components of the inner factor is fixed by a natural action of the fundamental group of $\mathcal{Y}$. This property has an application to the isomorphism problem in Coxeter groups. \end{abstract}
\section{Introduction} \label{sec:intro}
A pair $(W,S)$ of a group $W$ and its (possibly infinite) generating set $S$ is called a \emph{Coxeter system} if $W$ admits the following presentation \begin{displaymath} W=\langle S \mid (st)^{m(s,t)}=1 \mbox{ for all } s,t \in S \mbox{ with } m(s,t)<\infty \rangle \enspace, \end{displaymath} where $m \colon (s,t) \mapsto m(s,t) \in \{1,2,\dots\} \cup \{\infty\}$ is a symmetric mapping in $s,t \in S$ with the property that we have $m(s,t)=1$ if and only if $s=t$. A group $W$ is called a \emph{Coxeter group} if $(W,S)$ is a Coxeter system for some $S \subseteq W$. Since Coxeter systems and some associated objects, such as root systems, appear frequently in various topics of mathematics, algebraic or combinatorial properties of Coxeter systems and those associated objects have been investigated very well, forming a long history and establishing many beautiful theories (see e.g., \cite{Hum} and references therein). For example, it has been well known that, given an arbitrary Coxeter system $(W,S)$, the mapping $m$ by which the above group presentation defines the same group $W$ is uniquely determined.
In recent decades, not only the properties of a Coxeter group $W$ associated to a specific generating set $S$, but also the group-theoretic properties of an arbitrary Coxeter group $W$ itself have been studied well. One of the recent main topics in the study of group-theoretic properties of Coxeter groups is the \emph{isomorphism problem}, that is, the problem of determining which of the Coxeter groups are isomorphic to each other as abstract groups. In other words, the problem is to investigate the possible \lq\lq types'' of generating sets $S$ for a given Coxeter group $W$. For example, it has been known that for a Coxeter group $W$ in certain classes, the set of reflections $S^W := \{wsw^{-1} \mid w \in W \mbox{ and } s \in S\}$ associated to any possible generating set $S$ of $W$ (as a Coxeter group) is equal to each other and independent of the choice of $S$ (see e.g., \cite{Bah}). A Coxeter group $W$ having this property is called \emph{reflection independent}. A simplest nontrivial example of a Coxeter group which is not reflection independent is Weyl group of type $G_2$ (or the finite Coxeter group of type $I_2(6)$) with two simple reflections $s,t$, which admits another generating set $\{s,ststs,(st)^3\}$ of type $A_1 \times A_2$ involving an element $(st)^3$ that is not a reflection with respect to the original generating set. One of the main branches of the isomorphism problem in Coxeter groups is to determine the possibilities of a group isomorphism between two Coxeter groups which preserves the sets of reflections (with respect to some specified generating sets). Such an isomorphism is called \emph{reflection-preserving}.
In a recent study by the author of this paper, it is revealed that some properties of the centralizers $Z_W(r)$ of reflections $r$ in a Coxeter group $W$ (with respect to a generating set $S$) can be applied to the study of reflection independent Coxeter groups and reflection-preserving isomorphisms. An outline of the idea is as follows. First, by a general result on the structures of the centralizers of parabolic subgroups \cite{Nui11} or the normalizers of parabolic subgroups \cite{Bri-How} in Coxeter groups applied to the case of a single reflection, we have a decomposition $Z_W(r) = \langle r \rangle \times (W^{\perp r} \rtimes Y_r)$, where $W^{\perp r}$ denotes the subgroup generated by all the reflections except $r$ itself that commute with $r$, and $Y_r$ is a subgroup isomorphic to the fundamental group of a certain graph associated to $(W,S)$. The above-mentioned general results also give a canonical presentation of $W^{\perp r}$ as a Coxeter group. Then the unique maximal reflection subgroup (i.e., subgroup generated by reflections) of $Z_W(r)$ is $\langle r \rangle \times W^{\perp r}$. Now suppose that $W^{\perp r}$ has no finite irreducible components. In this case, the maximal reflection subgroup of $Z_W(r)$ has only one finite irreducible component, that is $\langle r \rangle$. Now it can be shown that, if the image $f(r)$ of $r$ by a group isomorphism $f$ from $W$ to another Coxeter group $W'$ is not a reflection with respect to a generating set of $W'$, then the finite irreducible components of the unique maximal reflection subgroup of the centralizer of $f(r)$ in $W'$ have more elements than $\langle r \rangle$, which is a contradiction. Hence, in such a case of $r$, the image of $r$ by any group isomorphism from $W$ to another Coxeter group is always a reflection. See the author's preprint \cite{Nui_ref} for more detailed arguments.
As we have seen in the previous paragraph, it is worthy to look for a class of Coxeter groups $W$ for which the above subgroup $W^{\perp r}$ of the centralizer $Z_W(r)$ of each reflection $r$ has no finite irreducible components. The aim of this paper is to establish a tool for finding Coxeter groups having the desired property. The main theorem (in a special case) of this paper can be stated as follows: \begin{quote} \textbf{Main Theorem (in a special case).} Let $r \in W$ be a reflection, and let $s_{\gamma}$ be a generator of $W^{\perp r}$ (as a Coxeter group) which belongs to a finite irreducible component of $W^{\perp r}$. Then $s_{\gamma}$ commutes with every element of $Y_r$. (See the previous paragraph for the notations.) \end{quote} By virtue of this result, to show that $W^{\perp r}$ has no finite irreducible components, it suffices to find (by using the general structural results in \cite{Nui11} or \cite{Bri-How}) for each generator $s_{\gamma}$ of $W^{\perp r}$ an element of $Y_r$ that does not commute with $s_{\gamma}$. A detailed argument along this strategy is given in the preprint \cite{Nui_ref}.
In fact, the main theorem (Theorem \ref{thm:YfixesWperpIfin}) of this paper is not only proven for the above-mentioned case of single reflection $r$, but also generalized to the case of centralizers $Z_W(W_I)$ of parabolic subgroups $W_I$ generated by some subsets $I \subseteq S$, with the property that $I$ has no irreducible components of type $A_n$ with $2 \leq n < \infty$. (We notice that there exists a counterexample when the assumption on $I$ is removed; see Section \ref{sec:counterexample} for details.) In the generalized statement, the group $W^{\perp r}$ is replaced naturally with the subgroup of $W$ generated by all the reflections except those in $I$ that commute with every element of $I$, while the group $Y_r$ is replaced with a subgroup of $W$ isomorphic to the fundamental group of a certain $2$-cell complex defined in \cite{Nui11}. We emphasize that, although the general structures of these subgroups of $Z_W(W_I)$ have been described in \cite{Nui11} (or \cite{Bri-How}), the main theorem of this paper is still far from being trivial; moreover, to the author's best knowledge, no other results on the structures of the centralizers $Z_W(W_I)$ which is in a significantly general form and involves much detailed information than those given in the general structural results \cite{Bri-How,Nui11} have been known in the literature.
The paper is organized as follows. In Section \ref{sec:Coxetergroups}, we summarize some fundamental properties and definitions for Coxeter groups. In Section \ref{sec:properties_centralizer}, we summarize some properties of the centralizers of parabolic subgroups relevant to our argument in the following sections, which have been shown in some preceding works (mainly in \cite{Nui11}). In Section \ref{sec:main_result}, we give the statement of the main theorem of this paper (Theorem \ref{thm:YfixesWperpIfin}), and give a remark on its application to the isomorphism problem in Coxeter groups (also mentioned in a paragraph above). The proof of the main theorem is divided into two main steps: First, Section \ref{sec:proof_general} presents some auxiliary results which do not require the assumption, put in the main theorem, on the subset $I$ of $S$ that $I$ has no irreducible components of type $A_n$ with $2 \leq n < \infty$. Then, based on the results in Section \ref{sec:proof_general}, Section \ref{sec:proof_special} deals with the special case as in the main theorem that $I$ has no such irreducible components, and completes the proof of the main theorem. The proof of the main theorem makes use of the list of positive roots given in Section \ref{sec:Coxetergroups} several times. Finally, in Section \ref{sec:counterexample}, we describe in detail a counterexample of our main theorem when the assumption that $I$ has no irreducible components of type $A_n$ with $2 \leq n < \infty$ is removed.
\paragraph*{Acknowledgments.} The author would like to express his deep gratitude to everyone who helped him, especially to Professor Itaru Terada who was the supervisor of the author during the graduate course in which a part of this work was done, and to Professor Kazuhiko Koike, for their invaluable advice and encouragement. The author would also like to the anonymous referee for the precious comments, especially for suggestion to reduce the size of the counterexample shown in Section \ref{sec:counterexample} which was originally of larger size. A part of this work was supported by JSPS Research Fellowship (No.~16-10825).
\section{Coxeter groups} \label{sec:Coxetergroups}
The basics of Coxeter groups summarized here are found in \cite{Hum} unless otherwise noticed. For some omitted definitions, see also \cite{Hum} or the author's preceding paper \cite{Nui11}.
\subsection{Basic notions} \label{sec:defofCox}
A pair $(W,S)$ of a group $W$ and its (possibly infinite) generating set $S$ is called a \emph{Coxeter system}, and $W$ is called a \emph{Coxeter group}, if $W$ admits the following presentation \begin{displaymath} W=\langle S \mid (st)^{m(s,t)}=1 \mbox{ for all } s,t \in S \mbox{ with } m(s,t)<\infty \rangle \enspace, \end{displaymath} where $m \colon (s,t) \mapsto m(s,t) \in \{1,2,\dots\} \cup \{\infty\}$ is a symmetric mapping in $s,t \in S$ with the property that we have $m(s,t)=1$ if and only if $s=t$. Let $\Gamma$ denote the \emph{Coxeter graph} of $(W,S)$, which is a simple undirected graph with vertex set $S$ in which two vertices $s,t \in S$ are joined by an edge with label $m(s,t)$ if and only if $m(s,t) \geq 3$ (by usual convention, the label is omitted when $m(s,t)=3$; see Figure \ref{fig:finite_irreducible_Coxeter_groups} below for example). If $\Gamma$ is connected, then $(W,S)$ is called \emph{irreducible}. Let $\ell$ denote the length function of $(W,S)$. For $w,u \in W$, we say that $u$ is a \emph{right divisor} of $w$ if $\ell(w) = \ell(wu^{-1}) + \ell(u)$. For each subset $I \subseteq S$, the subgroup $W_I := \langle I \rangle$ of $W$ generated by $I$ is called a \emph{parabolic subgroup} of $W$. Let $\Gamma_I$ denote the Coxeter graph of the Coxeter system $(W_I,I)$.
For two subsets $I,J \subseteq S$, we say that $I$ is \emph{adjacent to} $J$ if an element of $I$ is joined by an edge with an element of $J$ in the Coxeter graph $\Gamma$. We say that $I$ is \emph{apart from} $J$ if $I \cap J = \emptyset$ and $I$ is not adjacent to $J$. For the terminologies, we often abbreviate a set $\{s\}$ with a single element of $S$ to $s$ for simplicity.
\subsection{Root systems and reflection subgroups} \label{sec:rootsystem}
Let $V$ denote the \emph{geometric representation space} of $(W,S)$, which is an $\mathbb{R}$-linear space equipped with a basis $\Pi = \{\alpha_s \mid s \in S\}$ and a $W$-invariant symmetric bilinear form $\langle \,,\, \rangle$ determined by \begin{displaymath} \langle \alpha_s, \alpha_t \rangle = \begin{cases} -\cos(\pi / m(s,t)) & \mbox{if } m(s,t) < \infty \enspace; \\ -1 & \mbox{if } m(s,t) = \infty \enspace, \end{cases} \end{displaymath} where $W$ acts faithfully on $V$ by $s \cdot v=v-2\langle \alpha_s, v\rangle \alpha_s$ for $s \in S$ and $v \in V$. Then the \emph{root system} $\Phi=W \cdot \Pi$ consists of unit vectors with respect to the bilinear form $\langle \,,\, \rangle$, and $\Phi$ is the disjoint union of $\Phi^+ := \Phi \cap \mathbb{R}_{\geq 0}\Pi$ and $\Phi^- := -\Phi^+$ where $\mathbb{R}_{\geq 0}\Pi$ signifies the set of nonnegative linear combinations of elements of $\Pi$. Elements of $\Phi$, $\Phi^+$, and $\Phi^-$ are called \emph{roots}, \emph{positive roots}, and \emph{negative roots}, respectively. For a subset $\Psi \subseteq \Phi$ and an element $w \in W$, define \begin{displaymath} \Psi^+ := \Psi \cap \Phi^+ \,,\, \Psi^- := \Psi \cap \Phi^- \,,\, \Psi[w] :=\{\gamma \in \Psi^+ \mid w \cdot \gamma \in \Phi^-\} \enspace. \end{displaymath}
It is well known that the length $\ell(w)$ of $w$ is equal to $|\Phi[w]|$.
For an element $v=\sum_{s \in S}c_s\alpha_s$ of $V$, define the \emph{support} $\mathrm{Supp}\,v$ of $v$ to be the set of all $s \in S$ with $c_s \neq 0$. For a subset $\Psi$ of $\Phi$, define the support $\mathrm{Supp}\,\Psi$ of $\Psi$ to be the union of $\mathrm{Supp}\,\gamma$ over all $\gamma \in \Psi$. For each $I \subseteq S$, define \begin{displaymath} \Pi_I := \{\alpha_s \mid s \in I\} \subseteq \Pi \,,\, V_I := \mathrm{span}\,\Pi_I \subseteq V \,,\, \Phi_I := \Phi \cap V_I \enspace. \end{displaymath} It is well known that $\Phi_I$ coincides with the root system $W_I \cdot \Pi_I$ of $(W_I,I)$. We notice the following well-known fact: \begin{lem} \label{lem:support_is_irreducible} The support of any root $\gamma \in \Phi$ is irreducible. \end{lem} \begin{proof} Note that $\gamma \in \Phi_I = W_I \cdot \Pi_I$, where $I= \mathrm{Supp}\,\gamma$. On the other hand, it follows by induction on the length of $w$ that, for any $w \in W_I$ and $s \in I$, the support of $w \cdot \alpha_s$ is contained in the irreducible component of $I$ containing $s$. Hence the claim follows. \end{proof}
For a root $\gamma=w \cdot \alpha_s \in \Phi$, let $s_\gamma := wsw^{-1}$ be the \emph{reflection} along $\gamma$, which acts on $V$ by $s_\gamma \cdot v=v-2 \langle \gamma, v \rangle \gamma$ for $v \in V$. For any subset $\Psi \subseteq \Phi$, let $W(\Psi)$ denote the \emph{reflection subgroup} of $W$ generated by $\{s_{\gamma} \mid \gamma \in \Psi\}$. It was shown by Deodhar \cite{Deo_refsub} and by Dyer \cite{Dye} that $W(\Psi)$ is a Coxeter group. To determine their generating set $S(\Psi)$ for $W(\Psi)$, let $\Pi(\Psi)$ denote the set of all \lq\lq simple roots'' $\gamma \in (W(\Psi) \cdot \Psi)^+$ in the \lq\lq root system'' $W(\Psi) \cdot \Psi$ of $W(\Psi)$, that is, all the $\gamma$ for which any expression $\gamma=\sum_{i=1}^{r}c_i\beta_i$ with $c_i>0$ and $\beta_i \in (W(\Psi) \cdot \Psi)^+$ satisfies that $\beta_i=\gamma$ for every index $i$. Then the set $S(\Psi)$ is given by \begin{displaymath} S(\Psi) := \{s_\gamma \mid \gamma \in \Pi(\Psi)\} \enspace. \end{displaymath} We call $\Pi(\Psi)$ the \emph{simple system} of $(W(\Psi),S(\Psi))$. Note that the \lq\lq root system'' $W(\Psi) \cdot \Psi$ and the simple system $\Pi(\Psi)$ for $(W(\Psi),S(\Psi))$ have several properties that are similar to the usual root systems $\Phi$ and simple systems $\Pi$ for $(W,S)$; see e.g., Theorem 2.3 of \cite{Nui11} for the detail. In particular, we have the following result: \begin{thm} [{e.g., \cite[Theorem 2.3]{Nui11}}] \label{thm:reflectionsubgroup_Deodhar} Let $\Psi \subseteq \Phi$, and let $\ell_\Psi$ be the length function of $(W(\Psi),S(\Psi))$. Then for $w \in W(\Psi)$ and $\gamma \in (W(\Psi) \cdot \Psi)^+$, we have $\ell_\Psi(ws_\gamma)<\ell_\Psi(w)$ if and only if $w \cdot \gamma \in \Phi^-$. \end{thm}
We say that a subset $\Psi \subseteq \Phi^+$ is a \emph{root basis} if for each pair $\beta,\gamma \in \Psi$, we have \begin{displaymath} \begin{cases} \langle \beta,\gamma \rangle=-\cos(\pi/m) & \mbox{if } s_\beta s_\gamma \mbox{ has order } m<\infty \enspace;\\ \langle \beta,\gamma \rangle \leq -1 & \mbox{if } s_\beta s_\gamma \mbox{ has infinite order}. \end{cases} \end{displaymath} For example, it follows from Theorem \ref{thm:conditionforrootbasis} below that the simple system $\Pi(\Psi)$ of $(W(\Psi),S(\Psi))$ is a root basis for any $\Psi \subseteq \Phi$. For two root bases $\Psi_1,\Psi_2 \subseteq \Phi^+$, we say that a mapping from $\Psi_1 = \Pi(\Psi_1)$ to $\Psi_2 = \Pi(\Psi_2)$ is an isomorphism if it induces an isomorphism from $S(\Psi_1)$ to $S(\Psi_2)$. We show some properties of root bases: \begin{thm} [{\cite[Theorem 4.4]{Dye}}] \label{thm:conditionforrootbasis} Let $\Psi \subseteq \Phi^+$. Then we have $\Pi(\Psi)=\Psi$ if and only if $\Psi$ is a root basis. \end{thm} \begin{prop} [{\cite[Corollary 2.6]{Nui11}}] \label{prop:fintyperootbasis}
Let $\Psi \subseteq \Phi^+$ be a root basis with $|W(\Psi)|<\infty$. Then $\Psi$ is a basis of a positive definite subspace of $V$ with respect to the bilinear form $\langle \,,\, \rangle$. \end{prop} \begin{prop} [{\cite[Proposition 2.7]{Nui11}}] \label{prop:finitesubsystem}
Let $\Psi \subseteq \Phi^+$ be a root basis with $|W(\Psi)|<\infty$, and $U= \mathrm{span}\,\Psi$. Then there exist an element $w \in W$ and a subset $I \subseteq S$ satisfying that $|W_I|<\infty$ and $w \cdot (U \cap \Phi^+)=\Phi_I^+$. Moreover, the action of this $w$ maps $U \cap \Pi$ into $\Pi_I$. \end{prop}
\subsection{Finite parabolic subgroups} \label{sec:longestelement}
We say that a subset $I \subseteq S$ is of \emph{finite type} if $|W_I|<\infty$. The finite irreducible Coxeter groups have been classified as summarized in \cite[Chapter 2]{Hum}. Here we determine a labelling $r_1,r_2,\dots,r_n$ (where $n = |I|$) of elements of an irreducible subset $I \subseteq S$ of each finite type in the following manner, where the values $m(r_i,r_j)$ not listed here are equal to $2$ (see Figure \ref{fig:finite_irreducible_Coxeter_groups}): \begin{description} \item[Type $A_n$ ($1 \leq n < \infty$):] $m(r_i,r_{i+1})=3$ ($1 \leq i \leq n-1$); \item[Type $B_n$ ($2 \leq n < \infty$):] $m(r_i,r_{i+1})=3$ ($1 \leq i \leq n-2$) and $m(r_{n-1},r_n)=4$; \item[Type $D_n$ ($4 \leq n < \infty$):] $m(r_i,r_{i+1})=m(r_{n-2},r_n)=3$ ($1 \leq i \leq n-2$); \item[Type $E_n$ ($n=6,7,8$):] $m(r_1,r_3)=m(r_2,r_4)=m(r_i,r_{i+1})=3$ ($3 \leq i \leq n-1$); \item[Type $F_4$:] $m(r_1,r_2)=m(r_3,r_4)=3$ and $m(r_2,r_3)=4$; \item[Type $H_n$ ($n=3,4$):] $m(r_1,r_2)=5$ and $m(r_i,r_{i+1})=3$ ($2 \leq i \leq n-1$); \item[Type $I_2(m)$ ($5 \leq m < \infty$):] $m(r_1,r_2)=m$. \end{description} We call the above labelling $r_1,\dots,r_n$ the \emph{standard labelling} of $I$. \begin{figure}
\caption{Coxeter graphs of the finite irreducible Coxeter groups (here we write $i$ instead of $r_i$ for each vertex)}
\label{fig:finite_irreducible_Coxeter_groups}
\end{figure}
Let $w_0(I)$ denote the (unique) longest element of a finite parabolic subgroup $W_I$. It is well known that $w_0(I)^2 = 1$ and $w_0(I) \cdot \Pi_I = -\Pi_I$. Now let $I$ be irreducible of finite type. If $I$ is of type $A_n$ ($n \geq 2$), $D_k$ ($k$ odd), $E_6$ or $I_2(m)$ ($m$ odd), then the automorphism of the Coxeter graph $\Gamma_I$ of $W_I$ induced by (the conjugation action of) $w_0(I)$ is the unique nontrivial automorphism of $\Gamma_I$. Otherwise, $w_0(I)$ lies in the center $Z(W_I)$ of $W_I$ and the induced automorphism of $\Gamma_I$ is trivial, in which case we say that $I$ is of \emph{$(-1)$-type}. Moreover, if $W_I$ is finite but not irreducible, then $w_0(I)=w_0(I_1) \dotsm w_0(I_k)$ where the $I_i$ are the irreducible components of $I$.
\section{Known properties of the centralizers} \label{sec:properties_centralizer}
This section summarizes some known properties (mainly proven in \cite{Nui11}) of the centralizers $Z_W(W_I)$ of parabolic subgroups $W_I$ in Coxeter groups $W$, especially those relevant to the argument in this paper.
First, we fix an abstract index set $\Lambda$ with $|\Lambda| = |I|$, and define $S^{(\Lambda)}$ to be the set of all injective mappings $x \colon \Lambda \to S$. For $x \in S^{(\Lambda)}$ and $\lambda \in \Lambda$, we put $x_\lambda = x(\lambda)$; thus $x$ may be regarded as a duplicate-free \lq\lq $\Lambda$-tuple'' $(x_\lambda)=(x_\lambda)_{\lambda \in \Lambda}$ of elements of $S$. For each $x \in S^{(\Lambda)}$, let $[x]$ denote the image of the mapping $x$; $[x] = \{x_{\lambda} \mid \lambda \in \Lambda\}$. In the following argument, we fix an element $x_I \in S^{(\Lambda)}$ with $[x_I] = I$. We define \begin{displaymath} C_{x,y} := \{w \in W \mid \alpha_{x_\lambda}=w \cdot \alpha_{y_\lambda} \mbox{ for every } \lambda \in \Lambda\} \mbox{ for } x,y \in S^{(\Lambda)} \enspace. \end{displaymath} Note that $C_{x,y} \cdot C_{y,z} \subseteq C_{x,z}$ and $C_{x,y}{}^{-1} = C_{y,x}$ for $x,y,z \in S^{(\Lambda)}$. Now we define \begin{displaymath} w \ast y_{\lambda} := x_{\lambda} \mbox{ for } x,y \in S^{(\Lambda)}, w \in C_{x,y} \mbox{ and } \lambda \in \Lambda \enspace, \end{displaymath} therefore we have $w \cdot \alpha_s = \alpha_{w \ast s}$ for any $w \in C_{x,y}$ and $s \in [y]$. (This $\ast$ can be interpreted as the conjugation action of elements of $C_{x,y}$ to the elements of $[y]$.) Moreover, we define \begin{displaymath} w \ast y := x \mbox{ for } x,y \in S^{(\Lambda)} \mbox{ and } w \in C_{x,y} \end{displaymath} (this $\ast$ can be interpreted as the diagonal action on the $\Lambda$-tuples). We define $C_I = C_{x_I,x_I}$, therefore we have \begin{displaymath} C_I = \{w \in W \mid w \cdot \alpha_s=\alpha_s \mbox{ for every } s \in I\} \enspace, \end{displaymath} which is a normal subgroup of $Z_W(W_I)$.
To describe generators of $C_I$, we introduce some notations. For subsets $J,K \subseteq S$, let $J_{\sim K}$ denote the set of elements of $J \cup K$ that belongs to the same connected component of $\Gamma_{J \cup K}$ as an element of $K$. Now for $x \in S^{(\Lambda)}$ and $s \in S \smallsetminus [x]$ for which $[x]_{\sim s}$ is of finite type, there exists a unique $y \in S^{(\Lambda)}$ for which the element \begin{displaymath} w_x^s := w_0([x]_{\sim s})w_0([x]_{\sim s} \smallsetminus \{s\}) \end{displaymath} belongs to $C_{y,x}$. In this case, we define \begin{displaymath} \varphi(x,s) := y \enspace, \end{displaymath} therefore $\varphi(x,s) = w_x^s \ast x$ in the above notations. We have the following result: \begin{prop} [{see \cite[Theorem 3.5(iii)]{Nui11}}] \label{prop:generator_C} Let $x,y \in S^{(\Lambda)}$ and $w \in C_{x,y}$. Then there are a finite sequence $z_0 = y,z_1,\dots,z_{n-1},z_n = x$ of elements of $S^{(\Lambda)}$ and a finite sequence $s_0,s_1,\dots,s_{n-1}$ of elements of $S$ satisfying that $s_i \not\in [z_i]$, $[z_i]_{\sim s_i}$ is of finite type and $\varphi(z_i,s_i) = z_{i+1}$ for each index $0 \leq i \leq n-1$, and we have $w = w_{z_{n-1}}^{s_{n-1}} \cdots w_{z_1}^{s_1} w_{z_0}^{s_0}$. \end{prop}
For subsets $J,K \subseteq S$, define \begin{displaymath} \Phi_J^{\perp K} := \{\gamma \in \Phi_J \mid \langle \gamma,\alpha_s \rangle = 0 \mbox{ for every } s \in K\} \,,\, W_J^{\perp K} := W(\Phi_J^{\perp K}) \end{displaymath} (see Section \ref{sec:rootsystem} for notations). Then $(W_J^{\perp K},R^{J,K})$ is a Coxeter system with root system $\Phi_J^{\perp K}$ and simple system $\Pi^{J,K}$, where \begin{displaymath} R^{J,K} := S(\Phi_J^{\perp K}) \,,\, \Pi^{J,K} := \Pi(\Phi_J^{\perp K}) \end{displaymath} (see \cite[Section 3.1]{Nui11}). In the notations, the symbol $J$ will be omitted when $J=S$; hence we have \begin{displaymath} W^{\perp I}=W_S^{\perp I}=\langle \{s_\gamma \mid \gamma \in \Phi^{\perp I}\} \rangle \enspace. \end{displaymath} On the other hand, we define \begin{displaymath} Y_{x,y} := \{w \in C_{x,y} \mid w \cdot (\Phi^{\perp [y]})^+ \subseteq \Phi^+\} \mbox{ for } x,y \in S^{(\Lambda)} \enspace. \end{displaymath} Note that $Y_{x,y} = \{w \in C_{x,y} \mid (\Phi^{\perp [x]})^+=w \cdot (\Phi^{\perp [y]})^+\}$ (see \cite[Section 3.1]{Nui11}). Note also that $Y_{x,y} \cdot Y_{y,z} \subseteq Y_{x,z}$ and $Y_{x,y}{}^{-1} = Y_{y,x}$ for $x,y,z \in S^{(\Lambda)}$. Now we define $Y_I = Y_{x_I,x_I}$, therefore we have \begin{displaymath} Y_I = \{w \in C_I \mid (\Phi^{\perp I})^+ = w \cdot (\Phi^{\perp I})^+\} \enspace. \end{displaymath} We have the following results: \begin{prop} [{see \cite[Lemma 4.1]{Nui11}}] \label{prop:charofBphi} For $x \in S^{(\Lambda)}$ and $s \in S \smallsetminus [x]$, the three conditions are equivalent: \begin{enumerate} \item $[x]_{\sim s}$ is of finite type, and $\varphi(x,s) = x$; \item $[x]_{\sim s}$ is of finite type, and $\Phi^{\perp [x]}[w_x^s] \neq \emptyset$; \item $\Phi_{[x] \cup \{s\}}^{\perp [x]} \neq \emptyset$. \end{enumerate} If these three conditions are satisfied, then we have $\Phi^{\perp [x]}[w_x^s]=(\Phi_{[x] \cup \{s\}}^{\perp [x]})^+=\{\gamma(x,s)\}$ for a unique positive root $\gamma(x,s)$ satisfying $s_{\gamma(x,s)}=w_x^s$. \end{prop} \begin{prop} \label{prop:factorization_C} Let $x,y \in S^{(\Lambda)}$. \begin{enumerate} \item \label{item:prop_factorization_C_decomp} (See \cite[Theorem 4.6(i)(iv)]{Nui11}.) The group $C_{x,x}$ admits a semidirect product decomposition $C_{x,x} = W^{\perp [x]} \rtimes Y_{x,x}$. Moreover, if $w \in Y_{x,y}$, then the conjugation action by $w$ defines an isomorphism $u \mapsto wuw^{-1}$ of Coxeter systems from $(W^{\perp [y]},R^{[y]})$ to $(W^{\perp [x]},R^{[x]})$. \item \label{item:prop_factorization_C_generator_Y} (See \cite[Theorem 4.6(ii)]{Nui11}.) Let $w \in Y_{x,y}$. Then there are a finite sequence $z_0 = y,z_1,\dots,z_{n-1},z_n = x$ of elements of $S^{(\Lambda)}$ and a finite sequence $s_0,s_1,\dots,s_{n-1}$ of elements of $S$ satisfying that $z_{i+1} \neq z_i$, $s_i \not\in [z_i]$, $[z_i]_{\sim s_i}$ is of finite type and $w_{z_i}^{s_i} \in Y_{z_{i+1},z_i}$ for each index $0 \leq i \leq n-1$, and we have $w = w_{z_{n-1}}^{s_{n-1}} \cdots w_{z_1}^{s_1} w_{z_0}^{s_0}$. \item \label{item:prop_factorization_C_generator_perp} (See \cite[Theorem 4.13]{Nui11}.) The generating set $R^{[x]}$ of $W^{\perp [x]}$ consists of elements of the form $w s_{\gamma(y,s)} w^{-1}$ satisfying that $y \in S^{(\Lambda)}$, $w \in Y_{x,y}$ and $\gamma(y,s)$ is a positive root as in the statement of Proposition \ref{prop:charofBphi} (hence $[y]_{\sim s}$ is of finite type and $\varphi(y,s) = y$). \end{enumerate} \end{prop} \begin{prop} [{see \cite[Proposition 4.8]{Nui11}}] \label{prop:Yistorsionfree} For any $x \in S^{(\Lambda)}$, the group $Y_{x,x}$ is torsion-free. \end{prop}
For the structure of the entire centralizer $Z_W(W_I)$, a general result (Theorem 5.2 of \cite{Nui11}) implies the following proposition in a special case (a proof of the proposition from Theorem 5.2 of \cite{Nui11} is straightforward by noticing the fact that, under the hypothesis of the following proposition, the group $\mathcal{A}$ defined in the last paragraph before Theorem 5.2 of \cite{Nui11} is trivial and hence the group $B_I$ used in Theorem 5.2 of \cite{Nui11} coincides with $Y_I$): \begin{prop} [{see \cite[Theorem 5.2]{Nui11}}] \label{prop:Z_for_special_case} If every irreducible component of $I$ of finite type is of $(-1)$-type (see Section \ref{sec:longestelement} for the terminology), then we have $Z_W(W_I) = Z(W_I) \times (W^{\perp I} \rtimes Y_I)$. \end{prop}
We also present an auxiliary result, which will be used later: \begin{lem} [{see \cite[Lemma 3.2]{Nui11}}] \label{lem:rightdivisor} Let $w \in W$ and $J,K \subseteq S$, and suppose that $w \cdot \Pi_J \subseteq \Pi$ and $w \cdot \Pi_K \subseteq \Phi^-$. Then $J \cap K=\emptyset$, the set $J_{\sim K}$ is of finite type, and $w_0(J_{\sim K})w_0(J_{\sim K} \smallsetminus K)$ is a right divisor of $w$ (see Section \ref{sec:defofCox} for the terminology). \end{lem}
\section{Main results} \label{sec:main_result}
In this section, we state the main results of this paper, and give some relevant remarks. The proof will be given in the following sections.
The main results deal with the relations between the \lq\lq finite part'' of the reflection subgroup $W^{\perp I}$ and the subgroup $Y_I$ of the centralizer $Z_W(W_I)$. In general, for any Coxeter group $W$, the product of the finite irreducible components of $W$ is called the \emph{finite part} of $W$; here we write it as $W_{\mathrm{fin}}$. Then, since $W^{\perp I}$ is a Coxeter group (with generating set $R^I$ and simple system $\Pi^I$) as mentioned in Section \ref{sec:properties_centralizer}, $W^{\perp I}$ has its own finite part $W^{\perp I}{}_{\mathrm{fin}}$.
To state the main theorem, we introduce a terminology: We say that a subset $I$ of $S$ is \emph{$A_{>1}$-free} if $I$ has no irreducible components of type $A_n$ with $2 \leq n < \infty$. Then the main theorem of this paper is stated as follows: \begin{thm} \label{thm:YfixesWperpIfin} Let $I$ be an $A_{>1}$-free subset of $S$ (see above for the terminology). Then for each $\gamma \in \Pi^I$ with $s_\gamma \in W^{\perp I}{}_{\mathrm{fin}}$, we have $w \cdot \gamma = \gamma$ for every $w \in Y_I$. Hence each element of the subgroup $Y_I$ of $Z_W(W_I)$ commutes with every element of $W^{\perp I}{}_{\mathrm{fin}}$. \end{thm} Among the several cases for the subset $I$ of $S$ covered by Theorem \ref{thm:YfixesWperpIfin}, we emphasize the following important special case: \begin{cor} \label{cor:YfixesWperpIfin} Let $I \subseteq S$. If every irreducible component of $I$ of finite type is of $(-1)$-type (see Section \ref{sec:longestelement} for the terminology), then we have \begin{displaymath} Z_W(W_I) = Z(W_I) \times W^{\perp I}{}_{\mathrm{fin}} \times (W^{\perp I}{}_{\mathrm{inf}} \rtimes Y_I) \enspace, \end{displaymath} where $W^{\perp I}{}_{\mathrm{inf}}$ denotes the product of the infinite irreducible components of $W^{\perp I}$ (hence $W^{\perp I} = W^{\perp I}{}_{\mathrm{fin}} \times W^{\perp I}{}_{\mathrm{inf}}$). \end{cor} \begin{proof} Note that the assumption on $I$ in Theorem \ref{thm:YfixesWperpIfin} is now satisfied. In this situation, Proposition \ref{prop:Z_for_special_case} implies that $Z_W(W_I) = Z(W_I) \times (W^{\perp I} \rtimes Y_I)$. Now by Theorem \ref{thm:YfixesWperpIfin}, both $Y_I$ and $W^{\perp I}{}_{\mathrm{inf}}$ centralize $W^{\perp I}{}_{\mathrm{fin}}$, therefore the latter factor of $Z_W(W_I)$ decomposes further as $W^{\perp I}{}_{\mathrm{fin}} \times (W^{\perp I}{}_{\mathrm{inf}} \rtimes Y_I)$. \end{proof} We notice that the conclusion of Theorem \ref{thm:YfixesWperpIfin} will not generally hold when we remove the $A_{>1}$-freeness assumption on $I$. A counterexample will be given in Section \ref{sec:counterexample}.
Here we give a remark on an application of the main results to a study of the isomorphism problem in Coxeter groups. An important branch in the research on the isomorphism problem in Coxeter groups is to investigate, for two Coxeter systems $(W,S)$, $(W',S')$ and a group isomorphism $f \colon W \to W'$, the possibilities of \lq\lq shapes'' of the images $f(r) \in W'$ by $f$ of reflections $r \in W$ (with respect to the generating set $S$); for example, whether $f(r)$ is always a reflection in $W'$ (with respect to $S'$) or not. Now if $r \in S$, then Corollary \ref{cor:YfixesWperpIfin} and Proposition \ref{prop:Yistorsionfree} imply that the unique maximal reflection subgroup of the centralizer of $r$ in $W$ is $\langle r \rangle \times W^{\perp \{r\}}$, which has finite part $\langle r \rangle \times W^{\perp \{r\}}{}_{\mathrm{fin}}$. Moreover, the property of $W^{\perp \{r\}}{}_{\mathrm{fin}}$ shown in Theorem \ref{thm:YfixesWperpIfin} can imply that the factor $W^{\perp \{r\}}{}_{\mathrm{fin}}$ becomes \lq\lq frequently'' almost trivial. In such a case, the finite part of the unique maximal reflection subgroup of the centralizer of $f(r)$ in $W'$ should be very small, which can be shown to be impossible if $f(r)$ is too far from being a reflection. Thus the possibilities of the shape of $f(r)$ in $W'$ can be restricted by using Theorem \ref{thm:YfixesWperpIfin}. See \cite{Nui_ref} for a detailed study along this direction. The author hope that such an argument can be generalized to the case that $r$ is not a reflection but an involution of \lq\lq type'' which is $A_{>1}$-free (in a certain appropriate sense).
\section{Proof of Theorem \ref{thm:YfixesWperpIfin}: General properties} \label{sec:proof_general}
In this and the next sections, we give a proof of Theorem \ref{thm:YfixesWperpIfin}. First, this section gives some preliminary results that hold for an arbitrary $I \subseteq S$ (not necessarily $A_{>1}$-free; see Section \ref{sec:main_result} for the terminology). Then the next section will focus on the case that $I$ is $A_{>1}$-free as in Theorem \ref{thm:YfixesWperpIfin} and complete the proof of Theorem \ref{thm:YfixesWperpIfin}.
\subsection{Decompositions of elements of $Y_{z,y}$} \label{sec:finitepart_decomposition_Y}
It is mentioned in Proposition \ref{prop:factorization_C}(\ref{item:prop_factorization_C_generator_Y}) that each element $u \in Y_{z,y}$ with $y,z \in S^{(\Lambda)}$ admits a kind of decomposition into elements of some $Y$. Here we introduce a generalization of such decompositions, which will play an important role below. We give a definition: \begin{defn} \label{defn:standard_decomposition} Let $u \in Y_{z,y}$ with $y,z \in S^{(\Lambda)}$. We say that an expression $\mathcal{D} := \omega_{n-1} \cdots \omega_1\omega_0$ of $u$ is a \emph{semi-standard decomposition} of $u$ with respect to a subset $J$ of $S$ if there exist $y^{(i)} = y^{(i)}(\mathcal{D}) \in S^{(\Lambda)}$ for $0 \leq i \leq n$, $t^{(i)} = t^{(i)}(\mathcal{D}) \in S$ for $0 \leq i \leq n-1$ and $J^{(i)} = J^{(i)}(\mathcal{D}) \subseteq S$ for $0 \leq i \leq n$, with $y^{(0)} = y$, $y^{(n)} = z$ and $J^{(0)} = J$, satisfying the following conditions for each index $0 \leq i \leq n-1$: \begin{itemize} \item We have $t^{(i)} \not\in [y^{(i)}] \cup J^{(i)}$ and $t^{(i)}$ is adjacent to $[y^{(i)}]$. \item The subset $K^{(i)} = K^{(i)}(\mathcal{D}) := ([y^{(i)}] \cup J^{(i)})_{\sim t^{(i)}}$ of $S$ is of finite type (see Section \ref{sec:properties_centralizer} for the notation). \item We have $\omega_i = \omega_{y^{(i)},J^{(i)}}^{t^{(i)}} := w_0(K^{(i)})w_0(K^{(i)} \smallsetminus \{t^{(i)}\})$. \item We have $\omega_i \in Y_{y^{(i+1)},y^{(i)}}$ and $\omega_i \cdot \Pi_{J^{(i)}} = \Pi_{J^{(i+1)}}$. \end{itemize} We call the above subset $K^{(i)}$ of $S$ the \emph{support} of $\omega_i$. We call a component $\omega_i$ of $\mathcal{D}$ a \emph{wide transformation} if its support $K^{(i)}$ intersects with $J^{(i)} \smallsetminus [y^{(i)}]$; otherwise, we call $\omega_i$ a \emph{narrow transformation}, in which case we have $\omega_i = \omega_{y^{(i)},J^{(i)}}^{t^{(i)}} = w_{y^{(i)}}^{t^{(i)}}$. Moreover, we say that $\mathcal{D} = \omega_{n-1} \cdots \omega_1\omega_0$ is a \emph{standard decomposition} of $u$ if $\mathcal{D}$ is a semi-standard decomposition of $u$ and $\ell(u) = \sum_{j=0}^{n-1} \ell(\omega_j)$. The integer $n$ is called the \emph{length} of $\mathcal{D}$ and is denoted by $\ell(\mathcal{D})$. \end{defn} \begin{exmp} \label{exmp:semi-standard_decomposition} We give an example of a semi-standard decomposition. Let $(W,S)$ be a Coxeter system of type $D_7$, with standard labelling $r_1,\dots,r_7$ of elements of $S$ given in Section \ref{sec:longestelement}. We put $n := 4$, and define the objects $y^{(i)}$, $t^{(i)}$ and $J^{(i)}$ as in Table \ref{tab:example_semi-standard_decomposition}, where we abbreviate each $r_i$ to $i$ for simplicity. In this case, the subsets $K^{(i)}$ of $S$ introduced in Definition \ref{defn:standard_decomposition} are determined as in the last row of Table \ref{tab:example_semi-standard_decomposition}. We have \begin{displaymath} \begin{split} \omega_0 &= w_0(\{r_1,r_2,r_3,r_4,r_5\})w_0(\{r_1,r_2,r_3,r_5\}) = r_2r_3r_4r_5r_1r_2r_3r_4 \enspace, \\ \omega_1 &= w_0(\{r_3,r_4,r_5,r_6\})w_0(\{r_3,r_4,r_5\}) = r_3r_4r_5r_6 \enspace, \\ \omega_2 &= w_0(\{r_4,r_5,r_6,r_7\})w_0(\{r_4,r_5,r_6\}) = r_7r_5r_4r_6r_5r_7 \enspace, \\ \omega_3 &= w_0(\{r_3,r_4,r_5,r_6\})w_0(\{r_4,r_5,r_6\}) = r_6r_5r_4r_3 \enspace. \end{split} \end{displaymath} Let $u$ denote the element $\omega_3\omega_2\omega_1\omega_0$ of $W$. Then it can be shown that $u \in Y_{z,y}$ where $y := y^{(0)} = (r_1,r_2,r_3)$ and $z := y^{(n)} = (r_5,r_4,r_3)$, and the expression $\mathcal{D} = \omega_3\omega_2\omega_1\omega_0$ is a semi-standard decomposition of $u$ of length $4$ with respect to $J := J^{(0)} = \{r_5\}$. Moreover, $\mathcal{D}$ is in fact a standard decomposition of $u$ (which is the same as the one obtained by using Proposition \ref{prop:standard_decomposition_existence} below). Among the four component $\omega_i$, the first one $\omega_0$ is a wide transformation and the other three $\omega_1,\omega_2,\omega_3$ are narrow transformations. \end{exmp} \begin{table}[hbt] \centering \caption{The data for the example of semi-standard decompositions} \label{tab:example_semi-standard_decomposition}
\begin{tabular}{|c||c|c|c|c|c|} \hline $i$ & $4$ & $3$ & $2$ & $1$ & $0$ \\ \hline\hline $y^{(i)}$ & $(5,4,3)$ & $(6,5,4)$ & $(4,5,6)$ & $(3,4,5)$ & $(1,2,3)$ \\ \hline $t^{(i)}$ & --- & $3$ & $7$ & $6$ & $4$ \\ \hline $J^{(i)}$ & $\{1\}$ & $\{1\}$ & $\{1\}$ & $\{1\}$ & $\{5\}$ \\ \hline $K^{(i)}$ & --- & $\{3,4,5,6\}$ & $\{4,5,6,7\}$ & $\{3,4,5,6\}$ & $\{1,2,3,4,5\}$ \\ \hline \end{tabular} \end{table}
The next proposition shows existence of standard decompositions: \begin{prop} \label{prop:standard_decomposition_existence} Let $u \in Y_{z,y}$ with $y,z \in S^{(\Lambda)}$, and let $J \subseteq S$ satisfying that $u \cdot \Pi_J \subseteq \Pi$. Then there exists a standard decomposition of $u$ with respect to $J$. \end{prop} \begin{proof} We proceed the proof by induction on $\ell(u)$. For the case $\ell(u) = 0$, i.e., $u = 1$, the empty expression satisfies the conditions for a standard decomposition of $u$. From now, we consider the case $\ell(u) > 0$. Then there is an element $t = t^{(0)} \in S$ satisfying that $u \cdot \alpha_t \in \Phi^-$. Since $u \in Y_{z,y}$ and $u \cdot \Pi_J \subseteq \Pi \subseteq \Phi^+$, we have $t \not\in [y] \cup J$ and $\alpha_t \not\in \Phi^{\perp [y]}$, therefore $t$ is adjacent to $[y]$. Now by Lemma \ref{lem:rightdivisor}, $K = K^{(0)} := ([y] \cup J)_{\sim t}$ is of finite type and $\omega_0 := \omega_{y,J}^{t}$ is a right divisor of $u$ (see Section \ref{sec:defofCox} for the terminology). By the definition of $\omega_{y,J}^{t}$ in Definition \ref{defn:standard_decomposition}, there exist unique $y^{(1)} \in S^{(\Lambda)}$ and $J^{(1)} \subseteq S$ satisfying that $y^{(1)} = \omega_0 \ast y$ (see Section \ref{sec:properties_centralizer} for the notation) and $\omega_0 \cdot \Pi_J = \Pi_{J^{(1)}}$. Moreover, since $\omega_0$ is a right divisor of $u$, it follows that $\Phi[\omega_0] \subseteq \Phi[u]$ (see e.g., Lemma 2.2 of \cite{Nui11}), therefore $\Phi^{\perp [y]}[\omega_0] \subseteq \Phi^{\perp [y]}[u] = \emptyset$ and $\omega_0 \in Y_{y^{(1)},y}$. Put $u' = u\omega_0{}^{-1}$. Then we have $u' \in Y_{z,y^{(1)}}$, $u' \cdot \Pi_{J^{(1)}} \subseteq \Pi$ and $\ell(u') = \ell(u) - \ell(\omega_0) < \ell(u)$ (note that $\omega_0 \neq 1$). Hence the concatenation of $\omega_0$ to a standard decomposition of $u' \in Y_{z,y^{(1)}}$ with respect to $J^{(1)}$ obtained by the induction hypothesis gives a desired standard decomposition of $u$. \end{proof}
We present some properties of (semi-)standard decompositions. First, we have the following: \begin{lem} \label{lem:another_decomposition_Y_no_loop} For any semi-standard decomposition $\omega_{n-1} \cdots \omega_1\omega_0$ of an element of $W$, for each $0 \leq i \leq n-1$, there exists an element of $\Pi_{K^{(i)} \smallsetminus \{t^{(i)}\}}$ which is not fixed by $\omega_i$. \end{lem} \begin{proof} Assume contrary that $\omega_i$ fixes $\Pi_{K^{(i)} \smallsetminus \{t^{(i)}\}}$ pointwise. Then by applying Proposition \ref{prop:charofBphi} to the pair of $[y^{(i)}] \cup J^{(i)}$ and $t^{(i)}$ instead of the pair of $[x]$ and $s$, it follows that there exists a root $\gamma \in (\Phi_{K^{(i)}}^{\perp K^{(i)} \smallsetminus \{t^{(i)}\}})^+$ with $\omega_i \cdot \gamma \in \Phi^-$ (note that, in this case, the element $w_x^s$ in Proposition \ref{prop:charofBphi} coincides with $\omega_i$). By the definition of the support $K^{(i)}$ of $\omega_i$, $K^{(i)}$ is apart from $[y^{(i)}] \smallsetminus K^{(i)}$, therefore this root $\gamma$ also belongs to $(\Phi^{\perp [y^{(i)}]})^+$. Hence we have $\Phi^{\perp [y^{(i)}]}[\omega_i] \neq \emptyset$, contradicting the property $\omega_i \in Y_{y^{(i+1)},y^{(i)}}$ in Definition \ref{defn:standard_decomposition}. Hence Lemma \ref{lem:another_decomposition_Y_no_loop} holds. \end{proof}
For a semi-standard decomposition $\mathcal{D} = \omega_n \cdots \omega_1\omega_0$ of $u \in Y_{z,y}$, let $0 \leq i_1 < i_2 < \cdots < i_k \leq n$ be the indices $i$ with the property that $[y^{(i+1)}(\mathcal{D})] = [y^{(i)}(\mathcal{D})]$ and $J^{(i+1)}(\mathcal{D}) = J^{(i)}(\mathcal{D})$. Then we define the \emph{simplification} $\widehat{\mathcal{D}}$ of $\mathcal{D}$ to be the expression $\omega_n \cdots \hat{\omega_{i_k}} \cdots \hat{\omega_{i_1}} \cdots \hat{\omega_{i_0}} \cdots \omega_0$ obtained from $\mathcal{D} = \omega_n \cdots \omega_1\omega_0$ by removing all terms $\omega_{i_j}$ with $1 \leq j \leq k$. Let $\widehat{u}$ denote the element of $W$ expressed by the product $\widehat{\mathcal{D}}$. The following lemma is straightforward to prove: \begin{lem} \label{lem:another_decomposition_Y_reduce_redundancy} In the above setting, let $\sigma$ denote the mapping from $\{0,1,\dots,n-k\}$ to $\{0,1,\dots,n\}$ satisfying that $\widehat{\mathcal{D}} = \omega_{\sigma(n-k)} \cdots \omega_{\sigma(1)}\omega_{\sigma(0)}$. Then we have $\widehat{u} \in Y_{\widehat{z},y}$ for some $\widehat{z} \in S^{(\Lambda)}$ with $[\widehat{z}] = [z]$; $\widehat{\mathcal{D}}$ is a semi-standard decomposition of $\widehat{u}$ with respect to $J^{(0)}(\widehat{\mathcal{D}}) = J^{(0)}(\mathcal{D})$; we have $J^{(n-k+1)}(\widehat{\mathcal{D}}) = J^{(n+1)}(\mathcal{D})$; and for each $0 \leq j \leq n-k$, we have $[y^{(j)}(\widehat{\mathcal{D}})] = [y^{(\sigma(j))}(\mathcal{D})]$, $[y^{(j+1)}(\widehat{\mathcal{D}})] = [y^{(\sigma(j)+1)}(\mathcal{D})]$, $J^{(j)}(\widehat{\mathcal{D}}) = J^{(\sigma(j))}(\mathcal{D})$ and $J^{(j+1)}(\widehat{\mathcal{D}}) = J^{(\sigma(j)+1)}(\mathcal{D})$. \end{lem} \begin{exmp} \label{exmp:simplification} For the case of Example \ref{exmp:semi-standard_decomposition}, the simplification $\widehat{\mathcal{D}}$ of the standard decomposition $\mathcal{D} = \omega_3\omega_2\omega_1\omega_0$ of $u$ is obtained by removing the third component $\omega_2$, therefore $\widehat{\mathcal{D}} = \omega_3\omega_1\omega_0$. We have \begin{displaymath} \begin{split} &y^{(0)}(\widehat{\mathcal{D}}) = y^{(0)}(\mathcal{D}) = (r_1,r_2,r_3) \,,\, y^{(1)}(\widehat{\mathcal{D}}) = y^{(1)}(\mathcal{D}) = (r_3,r_4,r_5) \enspace, \\ &y^{(2)}(\widehat{\mathcal{D}}) = y^{(2)}(\mathcal{D}) = (r_4,r_5,r_6) \,,\, y^{(3)}(\widehat{\mathcal{D}}) = (r_3,r_4,r_5) = \widehat{z} \enspace. \end{split} \end{displaymath} Now since $\omega_3$ is the inverse of $\omega_1$, the semi-standard decomposition $\widehat{\mathcal{D}}$ of $\widehat{u}$ is not a standard decomposition of $\widehat{u}$. \end{exmp}
Moreover, we have the following result: \begin{lem} \label{lem:another_decomposition_Y_shift_x} Let $\mathcal{D} = \omega_n \cdots \omega_1\omega_0$ be a semi-standard decomposition of an element $u \in W$. Let $r \in [y^{(0)}]$, and suppose that the support of each $\omega_i$ is apart from $r$. Moreover, let $s \in J^{(0)}$, $s' \in J^{(n+1)}$ and suppose that $u \ast s = s'$. Then we have $r \in [y^{(n+1)}]$, $u \ast r = r$ and $u \in Y_{z',z}$, where $z$ and $z'$ are elements of $S^{(\Lambda)}$ obtained from $y^{(0)}$ and $y^{(n+1)}$ by replacing $r$ with $s$ and with $s'$, respectively. \end{lem} \begin{proof} We use induction on $n \geq 0$. Put $\mathcal{D}' = \omega_{n-1} \cdots \omega_1\omega_0$, and let $u' \in Y_{y^{(n)},y^{(0)}}$ be the element expressed by the product $\mathcal{D}'$. Let $s'' := u' \ast s \in J^{(n)}$. By the induction hypothesis, we have $r \in [y^{(n)}]$, $u' \ast r = r$ and $u' \in Y_{z'',z}$, where $z''$ is the element of $S^{(\Lambda)}$ obtained from $y^{(n)}$ by replacing $r$ with $s''$. Now, since the support $K^{(n)}$ of $\omega_n$ is apart from $r \in [y^{(n)}]$, it follows that $r \in [y^{(n+1)}]$ and $\omega_n \ast r = r$, therefore $u \ast r = \omega_n u' \ast r = r$. On the other hand, we have $z' = \omega_n \ast z''$ by the construction of $z'$ and $z''$. Moreover, by the definition of $\omega_n$, the set $K^{(n)}$ is apart from $([y^{(n)}] \cup J^{(n)}) \smallsetminus K^{(n)}$, therefore $K^{(n)}$ is also apart from the subset $([z''] \cup J^{(n)}) \smallsetminus K^{(n)}$ of $([y^{(n)}] \cup J^{(n)}) \smallsetminus K^{(n)}$. Since $[y^{(n)}] \cap K^{(n)} \subseteq [z''] \cap K^{(n)}$, it follows that $\Phi^{\perp [z'']}[\omega_n] = \Phi_{K^{(n)}}^{\perp [z''] \cap K^{(n)}}[\omega_n] \subseteq \Phi_{K^{(n)}}^{\perp [y^{(n)}] \cap K^{(n)}}[\omega_n] = \Phi^{\perp [y^{(n)}]}[\omega_n] = \emptyset$ (note that $\omega_n \in Y_{y^{(n+1)},y^{(n)}}$), therefore we have $\omega_n \in Y_{z',z''}$. Hence we have $u = \omega_n u' \in Y_{z',z}$, concluding the proof. \end{proof}
\subsection{Reduction to a special case} \label{sec:proof_reduction}
Here we give a reduction of our proof of Theorem \ref{thm:YfixesWperpIfin} to a special case where the possibility of the subset $I \subseteq S$ is restricted in a certain manner.
First, for $J \subseteq S$, let $\iota(J)$ denote temporarily the union of the irreducible components of $J$ that are not of finite type, and let $\overline{\iota}(J)$ denote temporarily the set of elements of $S$ that are not apart from $\iota(J)$ (hence $J \cap \overline{\iota}(J) = \iota(J)$). For example, when $(W,S)$ is given by the Coxeter graph in Figure \ref{fig:example_notation_iota} (where we abbreviate each $r_i \in S$ to $i$) and $J = \{r_1,r_3,r_4,r_5,r_6\}$ (indicated in Figure \ref{fig:example_notation_iota} by the black vertices), we have $\iota(J) = \{r_1,r_5,r_6\}$ and $\overline{\iota}(J) = \{r_1,r_2,r_5,r_6,r_7\}$, therefore $J \cap \overline{\iota}(J) = \{r_1,r_5,r_6\} = \iota(J)$ as mentioned above. Now we have the following: \begin{figure}
\caption{An example for the notations $\iota(J)$ and $\overline{\iota}(J)$; here $J = \{1,3,4,5,6\}$}
\label{fig:example_notation_iota}
\end{figure} \begin{lem} \label{lem:proof_reduction_root_perp} Let $I$ be an arbitrary subset of $S$. Then we have $w \in W_{S \smallsetminus \overline{\iota}(I)}$ for any $w \in Y_{y,x_I}$ with $y \in S^{(\Lambda)}$, and we have $\Phi^{\perp I} = \Phi_{S \smallsetminus \overline{\iota}(I)}^{\perp I \smallsetminus \overline{\iota}(I)}$. \end{lem} \begin{proof} First, let $w \in Y_{y,x_I}$ with $y \in S^{(\Lambda)}$. Then by Proposition \ref{prop:factorization_C}(\ref{item:prop_factorization_C_generator_Y}), there are a finite sequence $z_0 = x_I,z_1,\dots,z_{n-1},z_n = y$ of elements of $S^{(\Lambda)}$ and a finite sequence $s_0,s_1,\dots,s_{n-1}$ of elements of $S$ satisfying that $z_{i+1} \neq z_i$, $s_i \not\in [z_i]$, $[z_i]_{\sim s_i}$ is of finite type and $w_{z_i}^{s_i} \in Y_{z_{i+1},z_i}$ for each index $0 \leq i \leq n-1$, and we have $w = w_{z_{n-1}}^{s_{n-1}} \cdots w_{z_1}^{s_1} w_{z_0}^{s_0}$. We show, by induction on $0 \leq i \leq n-1$, that $\iota([z_{i+1}]) = \iota(I)$, $\overline{\iota}([z_{i+1}]) = \overline{\iota}(I)$, and $w_{z_i}^{s_i} \in W_{S \smallsetminus \overline{\iota}(I)}$. It follows from the induction hypothesis when $i > 0$, and is trivial when $i = 0$, that $\iota([z_i]) = \iota(I)$ and $\overline{\iota}([z_i]) = \overline{\iota}(I)$. Since $s_i \not\in [z_i]$ and $[z_i]_{\sim s_i}$ is of finite type, it follows from the definition of $\overline{\iota}$ that $[z_i]_{\sim s_i} \subseteq S \smallsetminus \overline{\iota}([z_i])$, therefore we have $w_{z_i}^{s_i} \in W_{S \smallsetminus \overline{\iota}([z_i])} = W_{S \smallsetminus \overline{\iota}(I)}$, $\iota([z_{i+1}]) = \iota([z_i]) = \iota(I)$, and $\overline{\iota}([z_{i+1}]) = \overline{\iota}([z_i]) = \overline{\iota}(I)$, as desired. This implies that $w = w_{z_{n-1}}^{s_{n-1}} \cdots w_{z_1}^{s_1} w_{z_0}^{s_0} \in W_{S \smallsetminus \overline{\iota}(I)}$, therefore the first part of the claim holds.
For the second part of the claim, the inclusion $\supseteq$ is obvious by the definitions of $\iota(I)$ and $\overline{\iota}(I)$. For the other inclusion, it suffices to show that $\Phi^{\perp I} \subseteq \Phi_{S \smallsetminus \overline{\iota}(I)}$, or equivalently $\Pi^I \subseteq \Phi_{S \smallsetminus \overline{\iota}(I)}$. Let $\gamma \in \Pi^I$. By Proposition \ref{prop:factorization_C}(\ref{item:prop_factorization_C_generator_perp}), we have $\gamma = w \cdot \gamma(y,s)$ for some $y \in S^{(\Lambda)}$, $w \in Y_{x_I,y}$ and a root $\gamma(y,s)$ introduced in the statement of Proposition \ref{prop:charofBphi}. Now by applying the result of the previous paragraph to $w^{-1} \in Y_{y,x_I}$, it follows that $\iota([y]) = \iota(I)$, $\overline{\iota}([y]) = \overline{\iota}(I)$, and $w \in W_{S \smallsetminus \overline{\iota}(I)}$. Moreover, since $[y]_{\sim s}$ is of finite type (see Proposition \ref{prop:charofBphi}), a similar argument implies that $[y]_{\sim s} \subseteq S \smallsetminus \overline{\iota}([y]) = S \smallsetminus \overline{\iota}(I)$ and $w_y^s \in W_{S \smallsetminus \overline{\iota}(I)}$, therefore $\gamma(y,s) \in \Phi_{S \smallsetminus \overline{\iota}(I)}$. Hence we have $\gamma = w \cdot \gamma(y,s) \in \Phi_{S \smallsetminus \overline{\iota}(I)}$, concluding the proof of Lemma \ref{lem:proof_reduction_root_perp}. \end{proof}
For an arbitrary subset $I$ of $S$, suppose that $\gamma \in \Pi^I$, $s_\gamma \in W^{\perp I}{}_{\mathrm{fin}}$, and $w \in Y_I$. Then by the second part of Lemma \ref{lem:proof_reduction_root_perp}, we have $\gamma \in \Pi^I = \Pi^{S \smallsetminus \overline{\iota}(I),I \smallsetminus \overline{\iota}(I)}$ and $s_\gamma$ also belongs to the finite part of $W_{S \smallsetminus \overline{\iota}(I)}^{\perp I \smallsetminus \overline{\iota}(I)}$. Moreover, we have $w \in W_{S \smallsetminus \overline{\iota}(I)}$ by the first part of Lemma \ref{lem:proof_reduction_root_perp}, therefore $w$ also belongs to the group $Y_{I \smallsetminus \overline{\iota}(I)}$ constructed from the pair $S \smallsetminus \overline{\iota}(I)$, $I \smallsetminus \overline{\iota}(I)$ instead of the pair $S$, $I$. Hence we have the following result: If the conclusion of Theorem \ref{thm:YfixesWperpIfin} holds for the pair $S \smallsetminus \overline{\iota}(I)$, $I \smallsetminus \overline{\iota}(I)$ instead of the pair $S$, $I$, then the conclusion of Theorem \ref{thm:YfixesWperpIfin} also holds for the pair $S$, $I$. Note that $I \smallsetminus \overline{\iota}(I) = I \smallsetminus \iota(I)$ is the union of the irreducible components of $I$ of finite type. As a consequence, we may assume without loss of generality that every irreducible component of $I$ is of finite type (note that the $A_{>1}$-freeness in the hypothesis of Theorem \ref{thm:YfixesWperpIfin} is preserved by considering $I \smallsetminus \iota(I)$ instead of $I$).
From now on, we assume that every irreducible component of $I$ is of finite type, as mentioned in the last paragraph. For any $J \subseteq S$, we say that a subset $\Psi$ of the simple system $\Pi^J$ of $W^{\perp J}$ is an \emph{irreducible component} of $\Pi^J$ if $S(\Psi) = \{s_\beta \mid \beta \in \Psi\}$ is an irreducible component of the generating set $R^J$ of $W^{\perp J}$. Now, as in the statement of Theorem \ref{thm:YfixesWperpIfin}, let $w \in Y_I$ and $\gamma \in \Pi^I$, and suppose that $s_{\gamma} \in W^{\perp I}{}_{\mathrm{fin}}$. Let $\Psi$ denote the union of the irreducible components of $\Pi^I$ containing some $w^k \cdot \gamma$ with $k \in \mathbb{Z}$. Then we have the following: \begin{lem} \label{lem:proof_reduction_Psi_finite}
In this setting, $\Psi$ is of finite type; in particular, $|\Psi| < \infty$. Moreover, the two subsets $I \smallsetminus \mathrm{Supp}\,\Psi$ and $\mathrm{Supp}\,\Psi$ of $S$ are not adjacent. \end{lem} \begin{proof}
First, there exists a finite subset $K$ of $S$ for which $w \in W_K$ and $\gamma \in \Phi_K$. Then, the number of mutually orthogonal roots of the form $w^k \cdot \gamma$ is at most $|K| < \infty$, since those roots are linearly independent and contained in the $|K|$-dimensional space $V_K$. This implies that the number of irreducible components of $\Pi^I$ containing some $w^k \cdot \gamma$, which are of finite type by the property $s_\gamma \in W^{\perp I}{}_{\mathrm{fin}}$ and Proposition \ref{prop:factorization_C}(\ref{item:prop_factorization_C_decomp}), is finite. Therefore, the union $\Psi$ of those irreducible components is also of finite type. Hence the first part of the claim holds.
For the second part of the claim, assume contrary that some $s \in I \smallsetminus \mathrm{Supp}\,\Psi$ and $t \in \mathrm{Supp}\,\Psi$ are adjacent. By the definition of $\mathrm{Supp}\,\Psi$, we have $t \in \mathrm{Supp}\,\beta \subseteq \mathrm{Supp}\,\Psi$ for some $\beta \in \Psi$. Now we have $s \not\in \mathrm{Supp}\,\beta$. Let $c > 0$ be the coefficient of $\alpha_t$ in $\beta$. Then the property $s \not\in \mathrm{Supp}\,\beta$ implies that $\langle \alpha_s,\beta \rangle \leq c \langle \alpha_s,\alpha_t \rangle < 0$, contradicting the property $\beta \in \Phi^{\perp I}$. Hence the claim holds, concluding the proof of Lemma \ref{lem:proof_reduction_Psi_finite}. \end{proof}
We temporarily write $L = I \cap \mathrm{Supp}\,\Psi$, and put $\Psi' = \Psi \cup \Pi_L$. Then we have $\mathrm{Supp}\,\Psi' = \mathrm{Supp}\,\Psi$, therefore by Lemma \ref{lem:proof_reduction_Psi_finite}, $I \smallsetminus \mathrm{Supp}\,\Psi'$ and $\mathrm{Supp}\,\Psi'$ are not adjacent. On the other hand, we have $|\Psi| < \infty$ by Lemma \ref{lem:proof_reduction_Psi_finite}, therefore $\mathrm{Supp}\,\Psi' = \mathrm{Supp}\,\Psi$ is a finite set. By these properties and the above-mentioned assumption that every irreducible component of $I$ is of finite type, it follows that $\Pi_L$ is of finite type as well as $\Psi$. Note that $\Psi \subseteq \Pi^I \subseteq \Phi^{\perp L}$. Hence the two root bases $\Psi$ and $\Pi_L$ are orthogonal, therefore their union $\Psi'$ is also a root basis by Theorem \ref{thm:conditionforrootbasis}, and we have $|W(\Psi')|<\infty$. By Proposition \ref{prop:fintyperootbasis}, $\Psi'$ is a basis of a subspace $U := \mathrm{span}\,\Psi'$ of $V_{\mathrm{Supp}\,\Psi'}$. By applying Proposition \ref{prop:finitesubsystem} to $W_{\mathrm{Supp}\,\Psi'}$ instead of $W$, it follows that there exist $u \in W_{\mathrm{Supp}\,\Psi'}$ and $J \subseteq \mathrm{Supp}\,\Psi'$ satisfying that $W_J$ is finite, $u \cdot (U \cap \Phi^+) = \Phi_J^+$ and $u \cdot (U \cap \Pi) \subseteq \Pi_J$. Now we have the following: \begin{lem} \label{lem:proof_reduction_conjugate_by_Y} In this setting, if we choose such an element $u$ of minimal length, then there exists an element $y \in S^{(\Lambda)}$ satisfying that $u \in Y_{y,x_I}$, the sets $[y] \smallsetminus J$ and $J$ are not adjacent, and $(u \cdot \Psi) \cup \Pi_{[y] \cap J}$ is a basis of $V_J$. \end{lem} \begin{proof} Since $\Psi'$ is a basis of $U$, the property $u \cdot (U \cap \Phi^+) = \Phi_J^+$ implies that $u \cdot \Psi'$ is a basis of $V_J$. Now we have $u \cdot \Pi_L \subseteq \Pi_J$ since $\Pi_L \subseteq U \cap \Pi$, while $u$ fixes $\Pi_{I \smallsetminus L}$ pointwise since the sets $I \smallsetminus \mathrm{Supp}\,\Psi' = I \smallsetminus L$ and $\mathrm{Supp}\,\Psi'$ are not adjacent. By these properties, there exists an element $y \in S^{(\Lambda)}$ satisfying that $y = u \ast x_I$, $[y] \cap \mathrm{Supp}\,\Psi' \subseteq J$ and $[y] \smallsetminus \mathrm{Supp}\,\Psi' = I \smallsetminus \mathrm{Supp}\,\Psi'$. Since $J \subseteq \mathrm{Supp}\,\Psi'$, it follows that $[y] \smallsetminus J$ and $J$ are not adjacent. On the other hand, since $u \cdot \Pi_{I \smallsetminus L} = \Pi_{I \smallsetminus L}$, $u \cdot (U \cap \Phi^+) = \Phi_J^+$ and $\Pi_{I \smallsetminus L} \cap U = \emptyset$, it follows that $\Pi_{I \smallsetminus L} \cap \Phi_J^+ = \emptyset$, therefore we have $u \cdot \Pi_L = \Pi_{[y] \cap J}$. Hence $u \cdot \Psi' = (u \cdot \Psi) \cup \Pi_{[y] \cap J}$ is a basis of $V_J$.
Finally, we show that such an element $u$ of minimal length satisfies that $u \cdot \Pi^I \subseteq \Phi^+$, hence $u \cdot (\Phi^{\perp I})^+ \subseteq \Phi^+$ and $u \in Y_{y,x_I}$. We have $u \cdot \Psi \subseteq u \cdot (U \cap \Phi^+) = \Phi_J^+$. Secondly, for any $\beta \in \Pi^I \smallsetminus \Psi$, assume contrary that $u \cdot \beta \in \Phi^-$. Then we have $\beta \in \Phi_{\mathrm{Supp}\,\Psi'}$ since $u \in W_{\mathrm{Supp}\,\Psi'}$, therefore $s_{\beta} \in W_{\mathrm{Supp}\,\Psi'}$. On the other hand, since $\Psi$ is the union of some irreducible components of $\Pi^I$, it follows that $\beta$ is orthogonal to $\Psi$, hence orthogonal to $\Psi'$. By these properties, the element $u s_{\beta}$ also satisfies the above characteristics of the element $u$. However, now the property $u \cdot \beta \in \Phi^-$ implies that $\ell(u s_{\beta}) < \ell(u)$ (see Theorem \ref{thm:reflectionsubgroup_Deodhar}), contradicting the choice of $u$. Hence we have $u \cdot \beta \in \Phi^+$ for every $\beta \in \Pi^I \smallsetminus \Psi$, therefore $u \cdot \Pi^I \subseteq \Phi^+$, concluding the proof of Lemma \ref{lem:proof_reduction_conjugate_by_Y}. \end{proof} For an element $u \in Y_{y,x_I}$ as in Lemma \ref{lem:proof_reduction_conjugate_by_Y}, Proposition \ref{prop:factorization_C}(\ref{item:prop_factorization_C_decomp}) implies that $u \cdot \gamma \in \Pi^{[y]}$ and $s_{u \cdot \gamma} = u s_\gamma u^{-1} \in W^{\perp [y]}{}_{\mathrm{fin}}$. Now $w$ fixes the root $\gamma$ if and only if the element $uwu^{-1} \in Y_{y,y}$ fixes the root $u \cdot \gamma$. Moreover, the conjugation by $u$ defines an isomorphism of Coxeter systems $(W_I,I) \to (W_{[y]},[y])$. Hence, by considering $[y] \subseteq S$, $uwu^{-1} \in Y_{[y]}$, $u \cdot \gamma \in \Pi^{[y]}$ and $u \cdot \Psi \subseteq \Pi^{[y]}$ instead of $I$, $w$, $\gamma$ and $\Psi$ if necessary, we may assume without loss of generality the following conditions: \begin{description} \item[(A1)] Every irreducible component of $I$ is of finite type. \item[(A2)] There exists a subset $J \subseteq S$ of finite type satisfying that $I \smallsetminus J$ and $J$ are not adjacent and $\Psi \cup \Pi_{I \cap J}$ is a basis of $V_J$. \end{description} Moreover, if an irreducible component $J'$ of $J$ is contained in $I$, then a smaller subset $J \smallsetminus J'$ instead of $J$ also satisfies the assumption (A2); indeed, now $\Pi_{J'} \subseteq \Pi_{I \cap J}$ spans $V_{J'}$, and since $\Psi \cup \Pi_{I \cap J}$ is a basis of $V_J$ and the support of any root is irreducible (see Lemma \ref{lem:support_is_irreducible}), it follows that the support of any element of $\Psi \cup \Pi_{I \cap (J \smallsetminus J')}$ does not intersect with $J'$. Hence, by choosing a subset $J \subseteq S$ in (A2) as small as possible, we may also assume without loss of generality the following condition: \begin{description} \item[(A3)] Any irreducible component of $J$ is not contained in $I$. \end{description} We also notice the following properties: \begin{lem} \label{lem:Psi_is_full} In this setting, we have $\Psi = \Pi^{J,I \cap J}$, hence $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$ is a basis of $V_J$. \end{lem} \begin{proof} The inclusion $\Psi \subseteq \Pi^{J,I \cap J}$ follows from the definition of $\Psi$ and the condition (A2). Now assume contrary that $\beta \in \Pi^{J,I \cap J} \smallsetminus \Psi$. Then we have $\beta \in \Pi^I$ by (A2). Since $\Psi$ is the union of some irreducible components of $\Pi^I$, it follows that $\beta$ is orthogonal to $\Psi$ as well as to $\Pi_{I \cap J}$. This implies that $\beta$ belongs to the radical of $V_J$, which should be trivial by Proposition \ref{prop:fintyperootbasis}. This is a contradiction. Hence the claim holds. \end{proof} \begin{lem} \label{lem:property_w} In this setting, the element $w \in Y_I$ satisfies that $w \cdot \Phi_J = \Phi_J$, and the subgroup $\langle w \rangle$ generated by $w$ acts transitively on the set of the irreducible components of $\Pi^{J,I \cap J}$. \end{lem} \begin{proof} The second part of the claim follows immediately from the definition of $\Psi$ and Lemma \ref{lem:Psi_is_full}. It also implies that $w \cdot \Pi^{J,I \cap J} = \Pi^{J,I \cap J}$, while $w \cdot \Pi_{I \cap J} = \Pi_{I \cap J}$ since $w \in Y_I$. Moreover, $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$ is a basis of $V_J$ by Lemma \ref{lem:Psi_is_full}. This implies that $w \cdot V_J = V_J$, therefore we have $w \cdot \Phi_J = \Phi_J$. Hence the claim holds. \end{proof}
\subsection{A key lemma} \label{sec:proof_key_lemma}
Let $I^{\perp}$ denote the set of all elements of $S$ that are apart from $I$. Then there are two possibilities: $\Pi^{J,I \cap J} \not\subseteq \Phi_{I^{\perp}}$, or $\Pi^{J,I \cap J} \subseteq \Phi_{I^{\perp}}$. Here we present a key lemma regarding the former possibility (recall the three conditions (A1)--(A3) specified above): \begin{lem} \label{lem:finitepart_first_case_irreducible} If $\Pi^{J,I \cap J} \not\subseteq \Phi_{I^{\perp}}$, then we have $I \cap J \neq \emptyset$ and $J$ is irreducible. \end{lem} \begin{proof} First, take an element $\beta \in \Pi^{J,I \cap J} \smallsetminus \Phi_{I^{\perp}}$. Then we have $\beta \not\in \Phi_I$ since $\Pi^{J,I \cap J} \subseteq \Phi^{\perp I}$. Moreover, since the support $\mathrm{Supp}\,\beta$ of $\beta$ is irreducible (see Lemma \ref{lem:support_is_irreducible}), there exists an element $s \in \mathrm{Supp}\,\beta \smallsetminus I$ which is adjacent to an element of $I$, say $s' \in I$. Now the property $\beta \in \Phi^{\perp I}$ implies that $s' \in \mathrm{Supp}\,\beta$, since otherwise we have $\langle \beta,\alpha_{s'} \rangle \leq c \langle \alpha_s,\alpha_{s'} \rangle < 0$ where $c > 0$ is the coefficient of $\alpha_s$ in $\beta$. Hence we have $s' \in \mathrm{Supp}\,\Pi^{J,I \cap J} \subseteq J$.
Let $K$ denote the irreducible component of $J$ containing $s'$. Put $\Psi' = \Pi^{J,I \cap J} \cap \Phi_K$. Then, since $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$ is a basis of $V_J$ by Lemma \ref{lem:Psi_is_full} and the support of any root is irreducible (see Lemma \ref{lem:support_is_irreducible}), it follows that $\beta \in \Psi'$, $\Psi'$ is orthogonal to $\Pi^{J,I \cap J} \smallsetminus \Psi'$ and $\Psi' \cup \Pi_{I \cap K}$ is a basis of $V_K$. Now $\Psi'$ is the union of some irreducible components of $\Pi^{J,I \cap J}$. We show that $J$ is irreducible if we have $w \cdot \Phi_K = \Phi_K$. In this case, we have $w \cdot \Psi' = \Psi'$, therefore $\Pi^{J,I \cap J} = \Psi' \subseteq \Phi_K$ by the second part of Lemma \ref{lem:property_w}. Now by the condition (A3), $J$ has no irreducible components other than $K$ (indeed, if such an irreducible component $J'$ of $J$ exists, then the property $\Pi^{J,I \cap J} \subseteq \Phi_K$ implies that the space $V_{J'}$ should be spanned by a subset of $\Pi_{I \cap J}$, therefore $J' \subseteq I$). Hence $J = K$ is irreducible.
Thus it suffices to show that $w \cdot \Phi_K = \Phi_K$. For the purpose, it also suffices to show that $w \cdot \Phi_K \subseteq \Phi_K$ (since $K$ is of finite type as well as $J$), or equivalently $w \cdot \Pi_K \subseteq \Phi_K$. Moreover, by the three properties that $K$ is irreducible, $K \cap I \neq \emptyset$ and $w \cdot \Pi_{K \cap I} = \Pi_{K \cap I}$, it suffices to show that $w \cdot \alpha_{t'} \in \Phi_K$ provided $t' \in K$ is adjacent to some $t \in K$ with $w \cdot \alpha_t \in \Phi_K$. Now note that $w \cdot \Phi_J = \Phi_J$ by Lemma \ref{lem:property_w}. Assume contrary that $w \cdot \alpha_{t'} \not\in \Phi_K$. Then we have $w \cdot \alpha_{t'} \in \Phi_J \smallsetminus \Phi_K = \Phi_{J \smallsetminus K}$ since $K$ is an irreducible component of $J$, therefore $w \cdot \alpha_{t'}$ is orthogonal to $w \cdot \alpha_t \in \Phi_K$. This contradicts the property that $t'$ is adjacent to $t$, since $w$ leaves the bilinear form $\langle\,,\,\rangle$ invariant. Hence we have $w \cdot \alpha_{t'} \in \Phi_K$, as desired. \end{proof}
\section{Proof of Theorem \ref{thm:YfixesWperpIfin}: On the special case} \label{sec:proof_special}
In this section, we introduce the assumption in Theorem \ref{thm:YfixesWperpIfin} that $I$ is $A_{>1}$-free, and continue the argument in Section \ref{sec:proof_general}. Recall the properties (A1), (A2) and (A3) of $I$, $J$ and $\Psi = \Pi^{J,I \cap J}$ (see Lemma \ref{lem:Psi_is_full}) given in Section \ref{sec:proof_reduction}. Our aim here is to prove that $w$ fixes $\Pi^{J,I \cap J}$ pointwise, which implies our goal $w \cdot \gamma = \gamma$ since $\gamma \in \Psi = \Pi^{J,I \cap J}$ by the definition of $\Psi$. We divide the following argument into two cases: $\Pi^{J,I \cap J} \not\subseteq \Phi_{I^{\perp}}$, or $\Pi^{J,I \cap J} \subseteq \Phi_{I^{\perp}}$ (see Section \ref{sec:proof_key_lemma} for the definition of $I^{\perp}$).
\subsection{The first case $\Pi^{J,I \cap J} \not\subseteq \Phi_{I^{\perp}}$} \label{sec:proof_special_first_case}
Here we consider the case that $\Pi^{J,I \cap J} \not\subseteq \Phi_{I^{\perp}}$. In this case, the subset $J \subseteq S$ of finite type is irreducible by Lemma \ref{lem:finitepart_first_case_irreducible}, therefore we can apply the classification of finite irreducible Coxeter groups. Let $J = \{r_1,r_2,\dots,r_N\}$, where $N = |J|$, be the standard labelling of $J$ (see Section \ref{sec:longestelement}). We write $\alpha_i = \alpha_{r_i}$ for simplicity.
We introduce some temporal terminology. We say that an element $y \in S^{(\Lambda)}$ satisfies \emph{Property P} if $[y] \smallsetminus J = I \smallsetminus J$ (hence $[y] \smallsetminus J$ is apart from $J$ by the condition (A2)) and $\Pi^{J,[y] \cap J} \cup \Pi_{[y] \cap J}$ is a basis of $V_J$. For example, $x_I$ itself satisfies Property P. For any $y \in S^{(\Lambda)}$ satisfying Property P and any element $s \in J \smallsetminus [y]$ with $\varphi(y,s) \neq y$, we say that the isomorphism $t \mapsto w_y^s \ast t$ from $[y] \cap J$ to $[\varphi(y,s)] \cap J$ is a \emph{local transformation} (note that now $[y]_{\sim s} \subseteq J$ and $w_y^s \in W_J$ by the above-mentioned property that $[y] \smallsetminus J$ is apart from $J$). By abusing the terminology, in such a case we also call the correspondence $y \mapsto \varphi(y,s)$ a local transformation. Note that, in this case, $\varphi(y,s)$ also satisfies Property P, we have $w_y^s \in Y_{\varphi(y,s),y}$ and $w_y^s \ast t = t$ for any $t \in [y] \smallsetminus J$, and the action of $w_y^s$ induces an isomorphism from $\Pi^{J,[y] \cap J}$ to $\Pi^{J,[\varphi(y,s)] \cap J}$.
Since $w \cdot \Pi^{J,I \cap J} = \Pi^{J,I \cap J}$, the claim is trivial if $|\Pi^{J,I \cap J}| = 1$. From now, we consider the case that $|\Pi^{J,I \cap J}| \geq 2$, therefore we have $N = |J| \geq |I \cap J| + 2 \geq 3$ (note that $I \cap J \neq \emptyset$ by Lemma \ref{lem:finitepart_first_case_irreducible}). In particular, $J$ is not of type $I_2(m)$. On the other hand, we have the following results: \begin{lem} \label{lem:J_not_A_N} In this setting, $J$ is not of type $A_N$. \end{lem} \begin{proof} We show that $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$ cannot span $V_J$ if $J$ is of type $A_N$, which deduces a contradiction and hence concludes the proof. By the $A_{>1}$-freeness of $I$, each irreducible component of $I \cap J$ (which is also an irreducible component of $I$) is of type $A_1$. Now by applying successive local transformations, we may assume without loss of generality that $r_1 \in I$ (indeed, if the minimal index $i$ with $r_i \in I$ satisfies $i \geq 2$, then we have $\varphi(x_I,r_{i-1}) \ast r_i = r_{i-1}$). In this case, we have $r_2 \not\in I$, while we have $\Phi_J^{\perp I} \subseteq \Phi_{J \smallsetminus \{r_1,r_2\}}$ by the fact that any positive root in the root system $\Phi_J$ of type $A_N$ is of the form $\alpha_i + \alpha_{i+1} + \cdots + \alpha_{i'}$ with $1 \leq i \leq i' \leq N$. This implies that the subset $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$ of $\Phi_J^{\perp I} \cup \Pi_{I \cap J}$ cannot span $V_J$, as desired. \end{proof} To prove the next lemma (and some other results below), we give a list of all positive roots of the Coxeter group of type $E_8$. The list is divided into six parts (Tables \ref{tab:positive_roots_E_8_1}--\ref{tab:positive_roots_E_8_6}). In the lists, we use the standard labelling $r_1,\dots,r_8$ of generators. The coefficients of each root are placed at the same relative positions as the corresponding vertices of the Coxeter graph of type $E_8$ in Figure \ref{fig:finite_irreducible_Coxeter_groups}; for example, the last root $\gamma_{120}$ in Table \ref{tab:positive_roots_E_8_6} is $2\alpha_1 + 3\alpha_2 + 4\alpha_3 + 6\alpha_4 + 5\alpha_5 + 4\alpha_6 + 3\alpha_7 + 2\alpha_8$ (which is the highest root of type $E_8$). For the columns for actions of generators (4th to 11th columns), a blank cell means that the generator $r_j$ fixes the root $\gamma_i$ (or equivalently, $\langle \alpha_j,\gamma_i \rangle = 0$); while a cell filled by \lq\lq ---'' means that $\gamma_i = \alpha_j$. Moreover, the positive roots of the parabolic subgroup of type $E_6$ (respectively, $E_7$) generated by $\{r_1,\dots,r_6\}$ (respectively, $\{r_1,\dots,r_7\}$) correspond to the rows indicated by \lq\lq $E_6$'' (respectively, \lq\lq $E_7$''). By the data for actions of generators, it can be verified that the list indeed exhausts all the positive roots. \begin{table}[p] \centering \caption{List of positive roots for Coxeter group of type $E_8$ (part $1$)} \label{tab:positive_roots_E_8_1}
\begin{tabular}{|c||c|c|c|c|c|c|c|c|c|c|cc} \cline{1-11}
height & $i$ & root $\gamma_i$ & \multicolumn{8}{|c|}{index $k$ with $r_j \cdot \gamma_i = \gamma_k$} \\ \cline{4-11} & & & $r_1$ & $r_2$ & $r_3$ & $r_4$ & $r_5$ & $r_6$ & $r_7$ & $r_8$ \\ \cline{1-11}\cline{1-11} $1$ & $1$ & \Eroot{1}{0}{0}{0}{0}{0}{0}{0} & --- & & $9$ & & & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $2$ & \Eroot{0}{1}{0}{0}{0}{0}{0}{0} & & --- & & $10$ & & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $3$ & \Eroot{0}{0}{1}{0}{0}{0}{0}{0} & $9$ & & --- & $11$ & & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $4$ & \Eroot{0}{0}{0}{1}{0}{0}{0}{0} & & $10$ & $11$ & --- & $12$ & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $5$ & \Eroot{0}{0}{0}{0}{1}{0}{0}{0} & & & & $12$ & --- & $13$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $6$ & \Eroot{0}{0}{0}{0}{0}{1}{0}{0} & & & & & $13$ & --- & $14$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $7$ & \Eroot{0}{0}{0}{0}{0}{0}{1}{0} & & & & & & $14$ & --- & $15$ & & $E_7$ \\ \cline{2-11} & $8$ & \Eroot{0}{0}{0}{0}{0}{0}{0}{1} & & & & & & & $15$ & --- \\ \cline{1-11} $2$ & $9$ & \Eroot{1}{0}{1}{0}{0}{0}{0}{0} & $3$ & & $1$ & $16$ & & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $10$ & \Eroot{0}{1}{0}{1}{0}{0}{0}{0} & & $4$ & $17$ & $2$ & $18$ & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $11$ & \Eroot{0}{0}{1}{1}{0}{0}{0}{0} & $16$ & $17$ & $4$ & $3$ & $19$ & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $12$ & \Eroot{0}{0}{0}{1}{1}{0}{0}{0} & & $18$ & $19$ & $5$ & $4$ & $20$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $13$ & \Eroot{0}{0}{0}{0}{1}{1}{0}{0} & & & & $20$ & $6$ & $5$ & $21$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $14$ & \Eroot{0}{0}{0}{0}{0}{1}{1}{0} & & & & & $21$ & $7$ & $6$ & $22$ & & $E_7$ \\ \cline{2-11} & $15$ & \Eroot{0}{0}{0}{0}{0}{0}{1}{1} & & & & & & $22$ & $8$ & $7$ \\ \cline{1-11} $3$ & $16$ & \Eroot{1}{0}{1}{1}{0}{0}{0}{0} & $11$ & $23$ & & $9$ & $24$ & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $17$ & \Eroot{0}{1}{1}{1}{0}{0}{0}{0} & $23$ & $11$ & $10$ & & $25$ & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $18$ & \Eroot{0}{1}{0}{1}{1}{0}{0}{0} & & $12$ & $25$ & & $10$ & $26$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $19$ & \Eroot{0}{0}{1}{1}{1}{0}{0}{0} & $24$ & $25$ & $12$ & & $11$ & $27$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $20$ & \Eroot{0}{0}{0}{1}{1}{1}{0}{0} & & $26$ & $27$ & $13$ & & $12$ & $28$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $21$ & \Eroot{0}{0}{0}{0}{1}{1}{1}{0} & & & & $28$ & $14$ & & $13$ & $29$ & & $E_7$ \\ \cline{2-11} & $22$ & \Eroot{0}{0}{0}{0}{0}{1}{1}{1} & & & & & $29$ & $15$ & & $14$ \\ \cline{1-11} \end{tabular} \end{table} \begin{table}[p] \centering \caption{List of positive roots for Coxeter group of type $E_8$ (part $2$)} \label{tab:positive_roots_E_8_2}
\begin{tabular}{|c||c|c|c|c|c|c|c|c|c|c|cc} \cline{1-11}
height & $i$ & root $\gamma_i$ & \multicolumn{8}{|c|}{index $k$ with $r_j \cdot \gamma_i = \gamma_k$} \\ \cline{4-11} & & & $r_1$ & $r_2$ & $r_3$ & $r_4$ & $r_5$ & $r_6$ & $r_7$ & $r_8$ \\ \cline{1-11} $4$ & $23$ & \Eroot{1}{1}{1}{1}{0}{0}{0}{0} & $17$ & $16$ & & & $30$ & & & & $E_6$ & $E_7$ \\ \cline{2-11} & $24$ & \Eroot{1}{0}{1}{1}{1}{0}{0}{0} & $19$ & $30$ & & & $16$ & $31$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $25$ & \Eroot{0}{1}{1}{1}{1}{0}{0}{0} & $30$ & $19$ & $18$ & $32$ & $17$ & $33$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $26$ & \Eroot{0}{1}{0}{1}{1}{1}{0}{0} & & $20$ & $33$ & & & $18$ & $34$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $27$ & \Eroot{0}{0}{1}{1}{1}{1}{0}{0} & $31$ & $33$ & $20$ & & & $19$ & $35$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $28$ & \Eroot{0}{0}{0}{1}{1}{1}{1}{0} & & $34$ & $35$ & $21$ & & & $20$ & $36$ & & $E_7$ \\ \cline{2-11} & $29$ & \Eroot{0}{0}{0}{0}{1}{1}{1}{1} & & & & $36$ & $22$ & & & $21$ \\ \cline{1-11} $5$ & $30$ & \Eroot{1}{1}{1}{1}{1}{0}{0}{0} & $25$ & $24$ & & $37$ & $23$ & $38$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $31$ & \Eroot{1}{0}{1}{1}{1}{1}{0}{0} & $27$ & $38$ & & & & $24$ & $39$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $32$ & \Eroot{0}{1}{1}{2}{1}{0}{0}{0} & $37$ & & & $25$ & & $40$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $33$ & \Eroot{0}{1}{1}{1}{1}{1}{0}{0} & $38$ & $27$ & $26$ & $40$ & & $25$ & $41$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $34$ & \Eroot{0}{1}{0}{1}{1}{1}{1}{0} & & $28$ & $41$ & & & & $26$ & $42$ & & $E_7$ \\ \cline{2-11} & $35$ & \Eroot{0}{0}{1}{1}{1}{1}{1}{0} & $39$ & $41$ & $28$ & & & & $27$ & $43$ & & $E_7$ \\ \cline{2-11} & $36$ & \Eroot{0}{0}{0}{1}{1}{1}{1}{1} & & $42$ & $43$ & $29$ & & & & $28$ \\ \cline{1-11} $6$ & $37$ & \Eroot{1}{1}{1}{2}{1}{0}{0}{0} & $32$ & & $44$ & $30$ & & $45$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $38$ & \Eroot{1}{1}{1}{1}{1}{1}{0}{0} & $33$ & $31$ & & $45$ & & $30$ & $46$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $39$ & \Eroot{1}{0}{1}{1}{1}{1}{1}{0} & $35$ & $46$ & & & & & $31$ & $47$ & & $E_7$ \\ \cline{2-11} & $40$ & \Eroot{0}{1}{1}{2}{1}{1}{0}{0} & $45$ & & & $33$ & $48$ & $32$ & $49$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $41$ & \Eroot{0}{1}{1}{1}{1}{1}{1}{0} & $46$ & $35$ & $34$ & $49$ & & & $33$ & $50$ & & $E_7$ \\ \cline{2-11} & $42$ & \Eroot{0}{1}{0}{1}{1}{1}{1}{1} & & $36$ & $50$ & & & & & $34$ \\ \cline{2-11} & $43$ & \Eroot{0}{0}{1}{1}{1}{1}{1}{1} & $47$ & $50$ & $36$ & & & & & $35$ \\ \cline{1-11} \end{tabular} \end{table} \begin{table}[p] \centering \caption{List of positive roots for Coxeter group of type $E_8$ (part $3$)} \label{tab:positive_roots_E_8_3}
\begin{tabular}{|c||c|c|c|c|c|c|c|c|c|c|cc} \cline{1-11}
height & $i$ & root $\gamma_i$ & \multicolumn{8}{|c|}{index $k$ with $r_j \cdot \gamma_i = \gamma_k$} \\ \cline{4-11} & & & $r_1$ & $r_2$ & $r_3$ & $r_4$ & $r_5$ & $r_6$ & $r_7$ & $r_8$ \\ \cline{1-11} $7$ & $44$ & \Eroot{1}{1}{2}{2}{1}{0}{0}{0} & & & $37$ & & & $51$ & & & $E_6$ & $E_7$ \\ \cline{2-11} & $45$ & \Eroot{1}{1}{1}{2}{1}{1}{0}{0} & $40$ & & $51$ & $38$ & $52$ & $37$ & $53$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $46$ & \Eroot{1}{1}{1}{1}{1}{1}{1}{0} & $41$ & $39$ & & $53$ & & & $38$ & $54$ & & $E_7$ \\ \cline{2-11} & $47$ & \Eroot{1}{0}{1}{1}{1}{1}{1}{1} & $43$ & $54$ & & & & & & $39$ \\ \cline{2-11} & $48$ & \Eroot{0}{1}{1}{2}{2}{1}{0}{0} & $52$ & & & & $40$ & & $55$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $49$ & \Eroot{0}{1}{1}{2}{1}{1}{1}{0} & $53$ & & & $41$ & $55$ & & $40$ & $56$ & & $E_7$ \\ \cline{2-11} & $50$ & \Eroot{0}{1}{1}{1}{1}{1}{1}{1} & $54$ & $43$ & $42$ & $56$ & & & & $41$ \\ \cline{1-11} $8$ & $51$ & \Eroot{1}{1}{2}{2}{1}{1}{0}{0} & & & $45$ & & $57$ & $44$ & $58$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $52$ & \Eroot{1}{1}{1}{2}{2}{1}{0}{0} & $48$ & & $57$ & & $45$ & & $59$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $53$ & \Eroot{1}{1}{1}{2}{1}{1}{1}{0} & $49$ & & $58$ & $46$ & $59$ & & $45$ & $60$ & & $E_7$ \\ \cline{2-11} & $54$ & \Eroot{1}{1}{1}{1}{1}{1}{1}{1} & $50$ & $47$ & & $60$ & & & & $46$ \\ \cline{2-11} & $55$ & \Eroot{0}{1}{1}{2}{2}{1}{1}{0} & $59$ & & & & $49$ & $61$ & $48$ & $62$ & & $E_7$ \\ \cline{2-11} & $56$ & \Eroot{0}{1}{1}{2}{1}{1}{1}{1} & $60$ & & & $50$ & $62$ & & & $49$ \\ \cline{1-11} $9$ & $57$ & \Eroot{1}{1}{2}{2}{2}{1}{0}{0} & & & $52$ & $63$ & $51$ & & $64$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $58$ & \Eroot{1}{1}{2}{2}{1}{1}{1}{0} & & & $53$ & & $64$ & & $51$ & $65$ & & $E_7$ \\ \cline{2-11} & $59$ & \Eroot{1}{1}{1}{2}{2}{1}{1}{0} & $55$ & & $64$ & & $53$ & $66$ & $52$ & $67$ & & $E_7$ \\ \cline{2-11} & $60$ & \Eroot{1}{1}{1}{2}{1}{1}{1}{1} & $56$ & & $65$ & $54$ & $67$ & & & $53$ \\ \cline{2-11} & $61$ & \Eroot{0}{1}{1}{2}{2}{2}{1}{0} & $66$ & & & & & $55$ & & $68$ & & $E_7$ \\ \cline{2-11} & $62$ & \Eroot{0}{1}{1}{2}{2}{1}{1}{1} & $67$ & & & & $56$ & $68$ & & $55$ \\ \cline{1-11} \end{tabular} \end{table} \begin{table}[p] \centering \caption{List of positive roots for Coxeter group of type $E_8$ (part $4$)} \label{tab:positive_roots_E_8_4}
\begin{tabular}{|c||c|c|c|c|c|c|c|c|c|c|cc} \cline{1-11}
height & $i$ & root $\gamma_i$ & \multicolumn{8}{|c|}{index $k$ with $r_j \cdot \gamma_i = \gamma_k$} \\ \cline{4-11} & & & $r_1$ & $r_2$ & $r_3$ & $r_4$ & $r_5$ & $r_6$ & $r_7$ & $r_8$ \\ \cline{1-11} $10$ & $63$ & \Eroot{1}{1}{2}{3}{2}{1}{0}{0} & & $69$ & & $57$ & & & $70$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $64$ & \Eroot{1}{1}{2}{2}{2}{1}{1}{0} & & & $59$ & $70$ & $58$ & $71$ & $57$ & $72$ & & $E_7$ \\ \cline{2-11} & $65$ & \Eroot{1}{1}{2}{2}{1}{1}{1}{1} & & & $60$ & & $72$ & & & $58$ \\ \cline{2-11} & $66$ & \Eroot{1}{1}{1}{2}{2}{2}{1}{0} & $61$ & & $71$ & & & $59$ & & $73$ & & $E_7$ \\ \cline{2-11} & $67$ & \Eroot{1}{1}{1}{2}{2}{1}{1}{1} & $62$ & & $72$ & & $60$ & $73$ & & $59$ \\ \cline{2-11} & $68$ & \Eroot{0}{1}{1}{2}{2}{2}{1}{1} & $73$ & & & & & $62$ & $74$ & $61$ \\ \cline{1-11} $11$ & $69$ & \Eroot{1}{2}{2}{3}{2}{1}{0}{0} & & $63$ & & & & & $75$ & & $E_6$ & $E_7$ \\ \cline{2-11} & $70$ & \Eroot{1}{1}{2}{3}{2}{1}{1}{0} & & $75$ & & $64$ & & $76$ & $63$ & $77$ & & $E_7$ \\ \cline{2-11} & $71$ & \Eroot{1}{1}{2}{2}{2}{2}{1}{0} & & & $66$ & $76$ & & $64$ & & $78$ & & $E_7$ \\ \cline{2-11} & $72$ & \Eroot{1}{1}{2}{2}{2}{1}{1}{1} & & & $67$ & $77$ & $65$ & $78$ & & $64$ \\ \cline{2-11} & $73$ & \Eroot{1}{1}{1}{2}{2}{2}{1}{1} & $68$ & & $78$ & & & $67$ & $79$ & $66$ \\ \cline{2-11} & $74$ & \Eroot{0}{1}{1}{2}{2}{2}{2}{1} & $79$ & & & & & & $68$ & \\ \cline{1-11} $12$ & $75$ & \Eroot{1}{2}{2}{3}{2}{1}{1}{0} & & $70$ & & & & $80$ & $69$ & $81$ & & $E_7$ \\ \cline{2-11} & $76$ & \Eroot{1}{1}{2}{3}{2}{2}{1}{0} & & $80$ & & $71$ & $82$ & $70$ & & $83$ & & $E_7$ \\ \cline{2-11} & $77$ & \Eroot{1}{1}{2}{3}{2}{1}{1}{1} & & $81$ & & $72$ & & $83$ & & $70$ \\ \cline{2-11} & $78$ & \Eroot{1}{1}{2}{2}{2}{2}{1}{1} & & & $73$ & $83$ & & $72$ & $84$ & $71$ \\ \cline{2-11} & $79$ & \Eroot{1}{1}{1}{2}{2}{2}{2}{1} & $74$ & & $84$ & & & & $73$ & \\ \cline{1-11} \end{tabular} \end{table} \begin{table}[p] \centering \caption{List of positive roots for Coxeter group of type $E_8$ (part $5$)} \label{tab:positive_roots_E_8_5}
\begin{tabular}{|c||c|c|c|c|c|c|c|c|c|c|cc} \cline{1-11}
height & $i$ & root $\gamma_i$ & \multicolumn{8}{|c|}{index $k$ with $r_j \cdot \gamma_i = \gamma_k$} \\ \cline{4-11} & & & $r_1$ & $r_2$ & $r_3$ & $r_4$ & $r_5$ & $r_6$ & $r_7$ & $r_8$ \\ \cline{1-11} $13$ & $80$ & \Eroot{1}{2}{2}{3}{2}{2}{1}{0} & & $76$ & & & $85$ & $75$ & & $86$ & & $E_7$ \\ \cline{2-11} & $81$ & \Eroot{1}{2}{2}{3}{2}{1}{1}{1} & & $77$ & & & & $86$ & & $75$ \\ \cline{2-11} & $82$ & \Eroot{1}{1}{2}{3}{3}{2}{1}{0} & & $85$ & & & $76$ & & & $87$ & & $E_7$ \\ \cline{2-11} & $83$ & \Eroot{1}{1}{2}{3}{2}{2}{1}{1} & & $86$ & & $78$ & $87$ & $77$ & $88$ & $76$ \\ \cline{2-11} & $84$ & \Eroot{1}{1}{2}{2}{2}{2}{2}{1} & & & $79$ & $88$ & & & $78$ & \\ \cline{1-11} $14$ & $85$ & \Eroot{1}{2}{2}{3}{3}{2}{1}{0} & & $82$ & & $89$ & $80$ & & & $90$ & & $E_7$ \\ \cline{2-11} & $86$ & \Eroot{1}{2}{2}{3}{2}{2}{1}{1} & & $83$ & & & $90$ & $81$ & $91$ & $80$ \\ \cline{2-11} & $87$ & \Eroot{1}{1}{2}{3}{3}{2}{1}{1} & & $90$ & & & $83$ & & $92$ & $82$ \\ \cline{2-11} & $88$ & \Eroot{1}{1}{2}{3}{2}{2}{2}{1} & & $91$ & & $84$ & $92$ & & $83$ & \\ \cline{1-11} $15$ & $89$ & \Eroot{1}{2}{2}{4}{3}{2}{1}{0} & & & $93$ & $85$ & & & & $94$ & & $E_7$ \\ \cline{2-11} & $90$ & \Eroot{1}{2}{2}{3}{3}{2}{1}{1} & & $87$ & & $94$ & $86$ & & $95$ & $85$ \\ \cline{2-11} & $91$ & \Eroot{1}{2}{2}{3}{2}{2}{2}{1} & & $88$ & & & $95$ & & $86$ & \\ \cline{2-11} & $92$ & \Eroot{1}{1}{2}{3}{3}{2}{2}{1} & & $95$ & & & $88$ & $96$ & $87$ & \\ \cline{1-11} $16$ & $93$ & \Eroot{1}{2}{3}{4}{3}{2}{1}{0} & $97$ & & $89$ & & & & & $98$ & & $E_7$ \\ \cline{2-11} & $94$ & \Eroot{1}{2}{2}{4}{3}{2}{1}{1} & & & $98$ & $90$ & & & $99$ & $89$ \\ \cline{2-11} & $95$ & \Eroot{1}{2}{2}{3}{3}{2}{2}{1} & & $92$ & & $99$ & $91$ & $100$ & $90$ & \\ \cline{2-11} & $96$ & \Eroot{1}{1}{2}{3}{3}{3}{2}{1} & & $100$ & & & & $92$ & & \\ \cline{1-11} $17$ & $97$ & \Eroot{2}{2}{3}{4}{3}{2}{1}{0} & $93$ & & & & & & & $101$ & & $E_7$ \\ \cline{2-11} & $98$ & \Eroot{1}{2}{3}{4}{3}{2}{1}{1} & $101$ & & $94$ & & & & $102$ & $93$ \\ \cline{2-11} & $99$ & \Eroot{1}{2}{2}{4}{3}{2}{2}{1} & & & $102$ & $95$ & & $103$ & $94$ & \\ \cline{2-11} & $100$ & \Eroot{1}{2}{2}{3}{3}{3}{2}{1} & & $96$ & & $103$ & & $95$ & & \\ \cline{1-11} \end{tabular} \end{table} \begin{table}[p] \centering \caption{List of positive roots for Coxeter group of type $E_8$ (part $6$)} \label{tab:positive_roots_E_8_6}
\begin{tabular}{|c||c|c|c|c|c|c|c|c|c|c|cc} \cline{1-11}
height & $i$ & root $\gamma_i$ & \multicolumn{8}{|c|}{index $k$ with $r_j \cdot \gamma_i = \gamma_k$} \\ \cline{4-11} & & & $r_1$ & $r_2$ & $r_3$ & $r_4$ & $r_5$ & $r_6$ & $r_7$ & $r_8$ \\ \cline{1-11} $18$ & $101$ & \Eroot{2}{2}{3}{4}{3}{2}{1}{1} & $98$ & & & & & & $104$ & $97$ \\ \cline{2-11} & $102$ & \Eroot{1}{2}{3}{4}{3}{2}{2}{1} & $104$ & & $99$ & & & $105$ & $98$ & \\ \cline{2-11} & $103$ & \Eroot{1}{2}{2}{4}{3}{3}{2}{1} & & & $105$ & $100$ & $106$ & $99$ & & \\ \cline{1-11} $19$ & $104$ & \Eroot{2}{2}{3}{4}{3}{2}{2}{1} & $102$ & & & & & $107$ & $101$ & \\ \cline{2-11} & $105$ & \Eroot{1}{2}{3}{4}{3}{3}{2}{1} & $107$ & & $103$ & & $108$ & $102$ & & \\ \cline{2-11} & $106$ & \Eroot{1}{2}{2}{4}{4}{3}{2}{1} & & & $108$ & & $103$ & & & \\ \cline{1-11} $20$ & $107$ & \Eroot{2}{2}{3}{4}{3}{3}{2}{1} & $105$ & & & & $109$ & $104$ & & \\ \cline{2-11} & $108$ & \Eroot{1}{2}{3}{4}{4}{3}{2}{1} & $109$ & & $106$ & $110$ & $105$ & & & \\ \cline{1-11} $21$ & $109$ & \Eroot{2}{2}{3}{4}{4}{3}{2}{1} & $108$ & & & $111$ & $107$ & & & \\ \cline{2-11} & $110$ & \Eroot{1}{2}{3}{5}{4}{3}{2}{1} & $111$ & $112$ & & $108$ & & & & \\ \cline{1-11} $22$ & $111$ & \Eroot{2}{2}{3}{5}{4}{3}{2}{1} & $110$ & $113$ & $114$ & $109$ & & & & \\ \cline{2-11} & $112$ & \Eroot{1}{3}{3}{5}{4}{3}{2}{1} & $113$ & $110$ & & & & & & \\ \cline{1-11} $23$ & $113$ & \Eroot{2}{3}{3}{5}{4}{3}{2}{1} & $112$ & $111$ & $115$ & & & & & \\ \cline{2-11} & $114$ & \Eroot{2}{2}{4}{5}{4}{3}{2}{1} & & $115$ & $111$ & & & & & \\ \cline{1-11} $24$ & $115$ & \Eroot{2}{3}{4}{5}{4}{3}{2}{1} & & $114$ & $113$ & $116$ & & & & \\ \cline{1-11} $25$ & $116$ & \Eroot{2}{3}{4}{6}{4}{3}{2}{1} & & & & $115$ & $117$ & & & \\ \cline{1-11} $26$ & $117$ & \Eroot{2}{3}{4}{6}{5}{3}{2}{1} & & & & & $116$ & $118$ & & \\ \cline{1-11} $27$ & $118$ & \Eroot{2}{3}{4}{6}{5}{4}{2}{1} & & & & & & $117$ & $119$ & \\ \cline{1-11} $28$ & $119$ & \Eroot{2}{3}{4}{6}{5}{4}{3}{1} & & & & & & & $118$ & $120$ \\ \cline{1-11} $29$ & $120$ & \Eroot{2}{3}{4}{6}{5}{4}{3}{2} & & & & & & & & $119$ \\ \cline{1-11} \end{tabular} \end{table}
Then we have the following: \begin{lem} \label{lem:possibility_J_is_E_6}
In this setting, if $J$ is of type $E_6$, then $|I \cap J| = 1$. \end{lem} \begin{proof}
By the property $N \geq |I \cap J| + 2$ and the $A_{>1}$-freeness of $I$, it follows that $I \cap J$ is either $\{r_2,r_3,r_4,r_5\}$ (of type $D_4$) or the union of irreducible components of type $A_1$. In the former case, we have $\Phi_J^{\perp I} = \emptyset$ (see Tables \ref{tab:positive_roots_E_8_1}--\ref{tab:positive_roots_E_8_6}), a contradiction. Therefore, $I \cap J$ consists of irreducible components of type $A_1$.
Now assume contrary that $I \cap J$ is not irreducible. Then, by applying successive local transformations and by using symmetry, we may assume without loss of generality that $r_1 \in I$ (cf., the proof of Lemma \ref{lem:J_not_A_N}). Now we have $\Pi^{J,\{r_1\}} = \{\alpha_2,\alpha_4,\alpha_5,\alpha_6,\alpha'\}$ which is the standard labelling of type $A_5$, where $\alpha'$ is the root $\gamma_{44}$ in Table \ref{tab:positive_roots_E_8_3}. Note that $\Pi_{(I \cap J) \smallsetminus \{r_1\}} \subseteq \Pi^{J,\{r_1\}}$. Now the same argument as Lemma \ref{lem:J_not_A_N} implies that the subspace $V'$ spanned by $\Pi^{J,I \cap J} \cup \Pi_{(I \cap J) \smallsetminus \{r_1\}}$ is a proper subspace of the space spanned by $\Pi^{J,\{r_1\}}$, therefore $\dim V' < 5$. This implies that the subspace spanned by $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$, which is the sum of $V'$ and $\mathbb{R}\alpha_1$, has dimension less than $6 = \dim V_J$, contradicting the fact that $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$ spans $V_J$ (see Lemma \ref{lem:Psi_is_full}). Hence $I \cap J$ is irreducible, therefore the claim holds. \end{proof}
We also give a list of all positive roots of the Coxeter group of type $D_n$ (Table \ref{tab:positive_roots_D_n}) in order to prove the next lemma (and some other results below). Some notations are similar to the above case of type $E_8$. For the data for actions of generators on the roots, if the action $r_k \cdot \gamma$ does not appear in the list, then it means either $r_k$ fixes $\gamma$ (or equivalently, $\gamma$ is orthogonal to $\alpha_k$), or $\gamma = \alpha_k$. Again, these data imply that the list indeed exhausts all the positive roots. \begin{table}[hbt] \centering \caption{List of positive roots for Coxeter group of type $D_n$} \label{tab:positive_roots_D_n}
\begin{tabular}{|c|c|} \hline roots & actions of generators \\ \hline $\gamma^{(1)}_{i,j} := \sum_{h=i}^{j} \alpha_h$ & $r_{i-1} \cdot \gamma^{(1)}_{i,j} = \gamma^{(1)}_{i-1,j}$ ($i \geq 2$) \\ ($1 \leq i \leq j \leq n-2$) & $r_i \cdot \gamma^{(1)}_{i,j} = \gamma^{(1)}_{i+1,j}$ ($i \leq j-1$) \\ ($\gamma^{(1)}_{i,i} = \alpha_i$) & $r_j \cdot \gamma^{(1)}_{i,j} = \gamma^{(1)}_{i,j-1}$ ($i \leq j-1$) \\ & $r_{j+1} \cdot \gamma^{(1)}_{i,j} = \gamma^{(1)}_{i,j+1}$ ($j \leq n-3$) \\ & $r_{n-1} \cdot \gamma^{(1)}_{i,n-2} = \gamma^{(2)}_i$ \\ & $r_n \cdot \gamma^{(1)}_{i,n-2} = \gamma^{(3)}_i$ \\ \hline $\gamma^{(2)}_i := \sum_{h=i}^{n-1} \alpha_h$ & $r_{i-1} \cdot \gamma^{(2)}_i = \gamma^{(2)}_{i-1}$ ($i \geq 2$) \\ ($1 \leq i \leq n-1$) & $r_i \cdot \gamma^{(2)}_i = \gamma^{(2)}_{i+1}$ ($i \leq n-2$) \\ ($\gamma^{(2)}_{n-1} = \alpha_{n-1}$) & $r_{n-1} \cdot \gamma^{(2)}_i = \gamma^{(1)}_{i,n-2}$ ($i \leq n-2$) \\ & $r_n \cdot \gamma^{(2)}_i = \gamma^{(4)}_{i,n-1}$ ($i \leq n-2$) \\ \hline $\gamma^{(3)}_i := \sum_{h=i}^{n-2} \alpha_h + \alpha_n$ & $r_{i-1} \cdot \gamma^{(3)}_i = \gamma^{(3)}_{i-1}$ ($i \geq 2$) \\ ($1 \leq i \leq n-1$) & $r_i \cdot \gamma^{(3)}_i = \gamma^{(3)}_{i+1}$ ($i \leq n-2$) \\ ($\gamma^{(3)}_{n-1} = \alpha_n$) & $r_n \cdot \gamma^{(3)}_i = \gamma^{(1)}_{i,n-2}$ ($i \leq n-2$) \\ & $r_{n-1} \cdot \gamma^{(3)}_i = \gamma^{(4)}_{i,n-1}$ ($i \leq n-2$) \\ \hline $\gamma^{(4)}_{i,j} := \sum_{h=i}^{j-1} \alpha_h + \sum_{h=j}^{n-2} 2\alpha_h + \alpha_{n-1} + \alpha_n$ & $r_{i-1} \cdot \gamma^{(4)}_{i,j} = \gamma^{(4)}_{i-1,j}$ ($i \geq 2$) \\ ($1 \leq i < j \leq n-1$) & $r_i \cdot \gamma^{(4)}_{i,j} = \gamma^{(4)}_{i+1,j}$ ($i \leq j-2$) \\ & $r_{j-1} \cdot \gamma^{(4)}_{i,j} = \gamma^{(4)}_{i,j-1}$ ($i \leq j-2$) \\ & $r_j \cdot \gamma^{(4)}_{i,j} = \gamma^{(4)}_{i,j+1}$ ($j \leq n-2$) \\ & $r_{n-1} \cdot \gamma^{(4)}_{i,n-1} = \gamma^{(3)}_i$ \\ & $r_n \cdot \gamma^{(4)}_{i,n-1} = \gamma^{(2)}_i$ \\ \hline \end{tabular} \end{table}
Then we have the following: \begin{lem} \label{lem:possibility_J_is_D_N} In this setting, suppose that $J$ is of type $D_N$. \begin{enumerate} \item \label{item:lem_possibility_J_is_D_N_case_1}
If $I \cap J$ has an irreducible component of type $D_k$ with $k \geq 4$ and $N - k$ is odd, then we have $|I \cap J| \leq k + (N-k-3)/2$. \item \label{item:lem_possibility_J_is_D_N_case_2}
If $N$ is odd, $I \cap J$ does not have an irreducible component of type $D_k$ with $k \geq 4$ and $\{r_{N-1},r_N\} \not\subseteq I$, then we have $|I \cap J| \leq (N-3)/2$. \item \label{item:lem_possibility_J_is_D_N_case_3}
If $N$ is odd, $I \cap J$ does not have an irreducible component of type $D_k$ with $k \geq 4$ and $\{r_{N-1},r_N\} \subseteq I$, then we have $|I \cap J| \leq (N-1)/2$. \end{enumerate} \end{lem} \begin{proof}
Assume contrary that the hypothesis of one of the three cases in the statement is satisfied but the inequality in the conclusion does not hold. We show that $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$ cannot span $V_J$, which is a contradiction and therefore concludes the proof. First, recall the property $N \geq |I \cap J| + 2$ and the $A_{>1}$-freeness of $I$. Then, in the case \ref{item:lem_possibility_J_is_D_N_case_1}, by applying successive local transformations, we may assume without loss of generality that $I \cap J$ consists of elements $r_{2j}$ with $1 \leq j \leq (N-k-1)/2$ and $r_j$ with $N-k+1 \leq j \leq N$. Similarly, in the case \ref{item:lem_possibility_J_is_D_N_case_2} (respectively, the case \ref{item:lem_possibility_J_is_D_N_case_3}), by applying successive local transformations and using symmetry, we may assume without loss of generality that $I \cap J$ consists of elements $r_{2j}$ with $1 \leq j \leq (N-1)/2$ (respectively, $r_{2j}$ with $1 \leq j \leq (N-1)/2$ and $r_N$). In any case, we have $\Phi_J^{\perp I} \subseteq \Phi_{J \smallsetminus \{r_1\}}$ (see Table \ref{tab:positive_roots_D_n}), therefore the subspace spanned by $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$ is contained in $V_{J \smallsetminus \{r_1\}}$. Hence $\Pi^{J,I \cap J} \cup \Pi_{I \cap J}$ cannot span $V_J$, concluding the proof. \end{proof}
We divide the following argument into two cases.
\subsubsection{Case $w \cdot \Pi_J \not\subseteq \Phi^+$} \label{sec:proof_special_first_case_subcase_1}
In order to prove that $w \cdot \Pi_J \subseteq \Phi^+$, here we assume contrary that $w \cdot \Pi_J \not\subseteq \Phi^+$ and deduce a contradiction.
In this setting, we construct a decomposition of $w$ in the following manner. Take an element $s \in J$ with $w \cdot \alpha_s \in \Phi^-$. By Lemma \ref{lem:rightdivisor}, the element $w_{x_I}^s$ is a right divisor of $w$. This implies that $\Phi^{\perp I}[w_{x_I}^s] \subseteq \Phi^{\perp I}[w] = \emptyset$ (see Lemma 2.2 of \cite{Nui11} for the first inclusion), therefore we have $w_{x_I}^s \in Y_{y,x_I}$ where we put $y := \varphi(x_I,s) \in S^{(\Lambda)}$. By Proposition \ref{prop:charofBphi}, we have $y \neq x_I$. This element $w_{x_I}^s$ induces a local transformation $x_I \mapsto y$. Now if $w(w_{x_I}^s)^{-1} \cdot \Pi_J \not\subseteq \Phi^+$, then we can similarly factor out from $w(w_{x_I}^s)^{-1}$ a right divisor of the form $w_y^t \in Y_{\varphi(y,t),y}$ with $t \in J$. Iterating this process, we finally obtain a decomposition of $w$ of the form $w = u w_{y_{n-1}}^{s_{n-1}} \cdots w_{y_1}^{s_1} w_{y_0}^{s_0}$ satisfying that $n \geq 1$, $u \in Y_{x_I,z}$ with $z \in S^{(\Lambda)}$, $w_{y_i}^{s_i} \in Y_{y_{i+1},y_i} \cap W_J$ for every $0 \leq i \leq n-1$ where we put $y_0 = x_I$ and $y_n = z$, and $u \cdot \Pi_J \subseteq \Phi^+$.
Put $u' := w_{y_{n-1}}^{s_{n-1}} \cdots w_{y_1}^{s_1} w_{y_0}^{s_0} \neq 1$. By the construction, the action of $u' \in Y_{z,x_I} \cap W_J$ induces (as the composition of successive local transformations) an isomorphism $\sigma \colon I \cap J \to [z] \cap J$, $t \mapsto u' \ast t$, while $u'$ fixes every element of $\Pi_{I \smallsetminus J}$. Now $\sigma$ is not an identity mapping; otherwise, we have $z = x_I$ and $1 \neq u' \in Y_{x_I,x_I}$, while $u'$ has finite order since $|W_J| < \infty$, contradicting Proposition \ref{prop:Yistorsionfree}. On the other hand, we have $u \cdot \Phi_J = wu'{}^{-1} \cdot \Phi_J = w \cdot \Phi_J = \Phi_J$, therefore $u \cdot \Phi_J^+ = \Phi_J^+$ since $u \cdot \Pi_J \subseteq \Phi^+$. This implies that $u \cdot \Pi_J = \Pi_J$, therefore the action of $u$ defines an automorphism $\tau$ of $J$. Since $w = u u' \in Y_I$, the composite mapping $\tau \circ \sigma$ is the identity mapping on $I \cap J$, while $\sigma$ is not identity as above. As a consequence, we have $\tau^{-1}|_{I \cap J} = \sigma$ and hence $\tau^{-1}$ is a nontrivial automorphism of $J$, therefore the possibilities of the type of $J$ are $D_N$, $E_6$ and $F_4$ (recall that $J$ is neither of type $A_N$ nor of type $I_2(m)$). \begin{lem} \label{lem:proof_special_first_case_subcase_1_not_F_4} In this setting, $J$ is not of type $F_4$. \end{lem} \begin{proof}
Assume contrary that $J = \{r_1,r_2,r_3,r_4\}$ is of type $F_4$. In this case, each of $r_1$ and $r_2$ is not conjugate in $W_J$ to one of $r_3$ and $r_4$ by the well-known fact that the conjugacy classes for the simple reflections $r_i$ are determined by the connected components of the graph obtained from the Coxeter graph by removing all edges having non-odd labels. Therefore, the mapping $\tau^{-1}|_{I \cap J} = \sigma$ induced by the action of $u' \in W_J$ cannot map an element $r_i$ ($1 \leq i \leq 4$) to $r_{5-i}$. This contradicts the fact that $\tau^{-1}$ is a nontrivial automorphism of $J$. Hence the claim holds. \end{proof} From now, we consider the remaining case that $J$ is either of type $D_N$ with $4 \leq N < \infty$ or of type $E_6$. Take a standard decomposition $\mathcal{D} = \omega_{\ell(\mathcal{D})-1} \cdots \omega_1\omega_0$ of $u \in Y_{x_I,z}$ with respect to $J$ (see Proposition \ref{prop:standard_decomposition_existence}). Note that $J$ is irreducible and $J \not\subseteq [z]$. This implies that, if $0 \leq i \leq \ell(\mathcal{D})-1$ and $\omega_j$ is a narrow transformation for every $0 \leq j \leq i$, then it follows by induction on $0 \leq j \leq i$ that the support of $\omega_j$ is apart from $J$, the product $\omega_j \cdots \omega_1\omega_0$ fixes $\Pi_J$ pointwise, $[y^{(j+1)}] \cap J = [z] \cap J$, and $[y^{(j+1)}] \smallsetminus J$ is not adjacent to $J$ (note that $[z] \smallsetminus J = I \smallsetminus J$ is not adjacent to $J$). By these properties, since $u$ does not fix $\Pi_J$ pointwise, $\mathcal{D}$ contains at least one wide transformation. Let $\omega := \omega_i$ be the first (from right) wide transformation in $\mathcal{D}$, and write $y = y^{(i)}(\mathcal{D})$, $t = t^{(i)}(\mathcal{D})$ and $K = K^{(i)}(\mathcal{D})$ for simplicity. Note that $J^{(i)}(\mathcal{D}) = J$ by the above argument. Note also that $\Pi^{K,[y] \cap K} \subseteq \Pi^{[y]}$, since $[y] \smallsetminus K$ is not adjacent to $K$ by the definition of $K$. Now the action of $\omega_{i-1} \cdots \omega_1\omega_0 u' \in Y_{y,x_I}$ induces an isomorphism $\Pi^I \to \Pi^{[y]}$ which maps $\Pi^{J,I \cap J}$ onto $\Pi^{J,[y] \cap J} = \Pi^{J,[z] \cap J}$. Hence we have the following (recall that $\Pi^{J,I \cap J}$ is the union of some irreducible components of $\Pi^I$): \begin{lem} \label{lem:proof_special_first_case_subcase_1_local_irreducible_components} In this setting, $\Pi^{J,[y] \cap J}$ is isomorphic to $\Pi^{J,I \cap J}$ and is the union of some irreducible components of $\Pi^{[y]}$. In particular, each element of $\Pi^{J,[y] \cap J}$ is orthogonal to any element of $\Pi^{K,[y] \cap K} \smallsetminus \Phi_J$. \end{lem} Now note that $K = ([y] \cup J)_{\sim t}$ is irreducible and of finite type, and $t$ is adjacent to $[y]$. Moreover, by Lemma \ref{lem:another_decomposition_Y_no_loop}, the element $\omega = \omega_{y,J}^{t}$ does not fix $\Pi_{K \smallsetminus \{t\}}$ pointwise. By these properties and symmetry, we may assume without loss of generality that the possibilities of $K$ are as follows: \begin{enumerate} \item \label{item:proof_special_first_case_subcase_1_J_E_6} $J$ is of type $E_6$, and; \begin{enumerate} \item \label{item:proof_special_first_case_subcase_1_J_E_6_K_E_8} $K = J \cup \{t,t'\}$ is of type $E_8$ where $t$ is adjacent to $r_6$ and $t'$, and $t' \in [y]$, \item \label{item:proof_special_first_case_subcase_1_J_E_6_K_E_7} $K = J \cup \{t\}$ is of type $E_7$ where $t$ is adjacent to $r_6$, and $r_6 \in [y]$, \end{enumerate} \item \label{item:proof_special_first_case_subcase_1_J_D_7} $J$ is of type $D_7$, $K = J \cup \{t\}$ is of type $E_8$ where $t$ is adjacent to $r_7$, and $r_7 \in [y]$, \item \label{item:proof_special_first_case_subcase_1_J_D_5} $J$ is of type $D_5$, and; \begin{enumerate} \item \label{item:proof_special_first_case_subcase_1_J_D_5_K_E_7} $K = J \cup \{t,t'\}$ is of type $E_7$ where $t$ is adjacent to $r_5$ and $t'$, and $t' \in [y]$, \item \label{item:proof_special_first_case_subcase_1_J_D_5_K_E_6} $K = J \cup \{t\}$ is of type $E_6$ where $t$ is adjacent to $r_5$, and $r_5 \in [y]$, \end{enumerate} \item \label{item:proof_special_first_case_subcase_1_J_D_N} $J$ is of type $D_N$, $K = J \cup \{t\}$ is of type $D_{N+1}$ where $t$ is adjacent to $r_1$, and $r_1 \in [y]$. \end{enumerate}
We consider Case \ref{item:proof_special_first_case_subcase_1_J_E_6_K_E_8}. We have $|[y] \cap J| = |I \cap J| = 1$ by Lemma \ref{lem:possibility_J_is_E_6}. Now by Tables \ref{tab:positive_roots_E_8_1}--\ref{tab:positive_roots_E_8_6} (where $r_7 = t$ and $r_8 = t'$), we have $\langle \beta,\beta' \rangle \neq 0$ for some $\beta \in \Pi^{J,[y] \cap J}$ and $\beta' \in \Pi^{K,[y] \cap K} \smallsetminus \Phi_J$ (namely, $(\beta,\beta') = (\alpha_4,\gamma_{84})$ when $[y] \cap J = \{r_1\}$; $(\beta,\beta') = (\gamma_{16},\gamma_{74})$ when $[y] \cap J = \{r_3\}$; and $(\beta,\beta') = (\alpha_1,\gamma_{74})$ when $[y] \cap J = \{r_j\}$ with $j \in \{2,4,5,6\}$, where the roots $\gamma_k$ are as in Tables \ref{tab:positive_roots_E_8_1}--\ref{tab:positive_roots_E_8_6}). This contradicts Lemma \ref{lem:proof_special_first_case_subcase_1_local_irreducible_components}.
We consider Case \ref{item:proof_special_first_case_subcase_1_J_E_6_K_E_7}. We have $|[y] \cap J| = |I \cap J| = 1$ by Lemma \ref{lem:possibility_J_is_E_6}, hence $[y] \cap J = \{r_6\}$. Now we have $\alpha_5 + \alpha_6 + \alpha_t \in \Pi^{K,[y] \cap K} \smallsetminus \Phi_J$, $\alpha_4 \in \Pi^{J,[y] \cap J}$, and these two roots are not orthogonal, contradicting Lemma \ref{lem:proof_special_first_case_subcase_1_local_irreducible_components}.
We consider Case \ref{item:proof_special_first_case_subcase_1_J_D_7}. Note that $N = 7 \geq |I \cap J| + 2 = |[y] \cap J| + 2$, therefore $|[y] \cap J| \leq 5$. By Lemma \ref{lem:possibility_J_is_D_N} and $A_{>1}$-freeness of $I$, it follows that the possibilities of $[y] \cap J$ are as listed in Table \ref{tab:proof_special_first_case_subcase_1_J_D_7}, where we put $(r'_1,r'_2,r'_3,r'_4,r'_5,r'_6,r'_7,r'_8) = (t,r_6,r_7,r_5,r_4,r_3,r_2,r_1)$ (hence $K = \{r'_1,\dots,r'_8\}$ is the standard labelling of type $E_8$). Now by Tables \ref{tab:positive_roots_E_8_1}--\ref{tab:positive_roots_E_8_6}, we have $\langle \beta,\beta' \rangle \neq 0$ for some $\beta \in \Pi^{J,[y] \cap J}$ and $\beta' \in \Pi^{K,[y] \cap K} \smallsetminus \Phi_J$ as listed in Table \ref{tab:proof_special_first_case_subcase_1_J_D_7}, where we write $\alpha'_j = \alpha_{r'_j}$ and the roots $\gamma_k$ are as in Tables \ref{tab:positive_roots_E_8_1}--\ref{tab:positive_roots_E_8_6}. This contradicts Lemma \ref{lem:proof_special_first_case_subcase_1_local_irreducible_components}.
\begin{table}[hbt] \centering \caption{List of roots for Case \ref{item:proof_special_first_case_subcase_1_J_D_7}} \label{tab:proof_special_first_case_subcase_1_J_D_7}
\begin{tabular}{|c|c|c|} \hline $[y] \cap J$ & $\beta$ & $\beta'$ \\ \hline $r'_3 \in [y] \cap J \subseteq \{r'_3,r'_6,r'_7,r'_8\}$ & $\alpha'_2$ & $\gamma_{16}$ \\ \hline $\{r'_3,r'_5\}$ & $\alpha'_2$ & $\gamma_{31}$ \\ \hline $\{r'_2,r'_3\} \subseteq [y] \cap J \subseteq \{r'_2,r'_3,r'_4,r'_5,r'_6\}$ & $\alpha'_8$ & $\gamma_{97}$ \\ \cline{1-2} $\{r'_2,r'_3,r'_7\}$ & $\gamma_{22}$ & \\ \hline $\{r'_2,r'_3,r'_8\}$ & $\alpha'_6$ & $\gamma_{104}$ \\ \hline \end{tabular} \end{table}
We consider Case \ref{item:proof_special_first_case_subcase_1_J_D_5_K_E_7}. Note that $N = 5 \geq |I \cap J| + 2 = |[y] \cap J| + 2$, therefore $|[y] \cap J| \leq 3$. By $A_{>1}$-freeness of $I$, every irreducible component of $[y] \cap J$ is of type $A_1$. Now by Lemma \ref{lem:possibility_J_is_D_N}, the possibilities of $[y] \cap J$ are as listed in Table \ref{tab:proof_special_first_case_subcase_1_J_D_5_K_E_7}, where we put $(r'_1,r'_2,r'_3,r'_4,r'_5,r'_6,r'_7) = (r_1,r_4,r_2,r_3,r_5,t,t')$ (hence $K = \{r'_1,\dots,r'_7\}$ is the standard labelling of type $E_7$). Now by Tables \ref{tab:positive_roots_E_8_1}--\ref{tab:positive_roots_E_8_6}, we have $\langle \beta,\beta' \rangle \neq 0$ for some $\beta \in \Pi^{J,[y] \cap J}$ and $\beta' \in \Pi^{K,[y] \cap K} \smallsetminus \Phi_J$ as listed in Table \ref{tab:proof_special_first_case_subcase_1_J_D_5_K_E_7}, where we write $\alpha'_j = \alpha_{r'_j}$ and the roots $\gamma_k$ are as in Tables \ref{tab:positive_roots_E_8_1}--\ref{tab:positive_roots_E_8_6}. This contradicts Lemma \ref{lem:proof_special_first_case_subcase_1_local_irreducible_components}.
\begin{table}[hbt] \centering \caption{List of roots for Case \ref{item:proof_special_first_case_subcase_1_J_D_5_K_E_7}} \label{tab:proof_special_first_case_subcase_1_J_D_5_K_E_7}
\begin{tabular}{|c|c|c|} \hline $[y] \cap J$ & $\beta$ & $\beta'$ \\ \hline $[y] \cap J \subseteq \{r'_2,r'_4,r'_5\}$ & $\alpha'_1$ & $\gamma_{61}$ \\ \cline{1-2} $\{r'_3\}$ & $\gamma_{16}$ & \\ \hline $\{r'_1\}$ & $\alpha'_4$ & $\gamma_{71}$ \\ \hline \end{tabular} \end{table}
We consider Case \ref{item:proof_special_first_case_subcase_1_J_D_5_K_E_6}. By the same reason as Case \ref{item:proof_special_first_case_subcase_1_J_D_5_K_E_7}, every irreducible component of $[y] \cap J$ is of type $A_1$. Now by Lemma \ref{lem:possibility_J_is_D_N}, we have only two possibilities of $[y] \cap J$; $\{r_5\}$ and $\{r_4,r_5\}$. In the first case $[y] \cap J = \{r_5\}$, we have $\alpha_2 \in \Pi^{J,[y] \cap J}$, $\alpha_3 + \alpha_5 + \alpha_t \in \Pi^{K,[y] \cap K} \smallsetminus \Phi_J$, and these two roots are not orthogonal, contradicting Lemma \ref{lem:proof_special_first_case_subcase_1_local_irreducible_components}. Hence we consider the second case $[y] \cap J = \{r_4,r_5\}$. In this case, the action of the first wide transformation $\omega$ in $\mathcal{D}$ maps the elements $r_1$, $r_2$, $r_3$, $r_4$ and $r_5$ to $t$, $r_5$, $r_3$, $r_2$ and $r_4$, respectively (note that $\{t,r_5,r_3,r_2,r_4\}$ is the standard labelling of type $D_5$). Now, by a similar argument as above, the possibility of the second wide transformation $\omega_{i'}$ in $\mathcal{D}$ (if exists) is as in Case \ref{item:proof_special_first_case_subcase_1_J_D_5_K_E_6}, where $t'' := t^{(i')}(\mathcal{D})$ is adjacent to either $r_2$ or $r_4$ (note that Case \ref{item:proof_special_first_case_subcase_1_J_D_5_K_E_7} cannot occur as discussed above, while Case \ref{item:proof_special_first_case_subcase_1_J_D_N} cannot occur by the shape of $J$ and the property $r_1 \not\in [y] \cap J$). This implies that the action of $\omega_{i'}$ either maps the elements $t$, $r_5$, $r_3$, $r_4$ and $r_2$ to $t''$, $r_2$, $r_3$, $r_5$ and $r_4$, respectively (forming a subset of type $D_5$ with the ordering being the standard labelling), or maps the elements $t$, $r_5$, $r_3$, $r_2$ and $r_4$ to $t''$, $r_4$, $r_3$, $r_5$ and $r_2$, respectively (forming a subset of type $D_5$ with the ordering being the standard labelling). By iterating the same argument, it follows that the sequence of elements $(r_2,r_3,r_4,r_5)$ is mapped by successive wide transformations in $\mathcal{D}$ to one of the following three sequences; $(r_2,r_3,r_4,r_5)$, $(r_5,r_3,r_2,r_4)$ and $(r_4,r_3,r_5,r_2)$. Hence $u$ itself should map $(r_2,r_3,r_4,r_5)$ to one of the above three sequences; while the action of $u$ induces the nontrivial automorphism $\tau$ of $J$, which maps $(r_1,r_2,r_3,r_4,r_5)$ to $(r_1,r_2,r_3,r_5,r_4)$. This is a contradiction.
Finally, we consider the case \ref{item:proof_special_first_case_subcase_1_J_D_N}. First we have the following lemma: \begin{lem} \label{lem:proof_special_first_case_subcase_1_J_D_N} In this setting, suppose further that there exists an integer $k \geq 1$ satisfying that $2k \leq N - 3$, $r_{2j - 1} \in [y]$ and $r_{2j} \not\in [y]$ for every $1 \leq j \leq k$, and $r_{2k + 1} \not\in [y]$. Then there exist a root $\beta \in \Pi^{J,[y] \cap J}$ and a root $\beta' \in \Pi^{K,[y] \cap K} \smallsetminus \Phi_J$ with $\langle \beta,\beta' \rangle \neq 0$. \end{lem} \begin{proof}
Put $J' := \{r_j \mid 2k+1 \leq j \leq N\}$. First, we have $\beta' := \alpha_t + \sum_{j=1}^{2k} \alpha_j \in \Pi^{K,[y] \cap K} \smallsetminus \Phi_J$ in this case. On the other hand, $\Pi^{J,[y] \cap J} \smallsetminus \Phi_{J'}$ consists of $k$ roots $\gamma^{(4)}_{2j-1,2j}$ with $1 \leq j \leq k$ (see Table \ref{tab:positive_roots_D_n} for the notation), while $\Pi_{[y] \cap J} \smallsetminus \Phi_{J'}$ consists of $k$ roots $\alpha_{2j-1}$ with $1 \leq j \leq k$. Hence $|(\Pi^{J,[y] \cap J} \cup \Pi_{[y] \cap J}) \smallsetminus \Phi_{J'}| = 2k$. Since $\Pi^{J,[y] \cap J} \cup \Pi_{[y] \cap J}$ is a basis of the space $V_J$ of dimension $N$, it follows that the subset $(\Pi^{J,[y] \cap J} \cup \Pi_{[y] \cap J}) \cap \Phi_{J'}$ spans a subspace of dimension $N - 2k = |J'|$. This implies that $(\Pi^{J,[y] \cap J} \cup \Pi_{[y] \cap J}) \cap \Phi_{J'} \not\subseteq \Phi_{J' \smallsetminus \{r_{2k+1}\}}$, therefore (since $\alpha_{2k+1} \not\in \Pi_{[y] \cap J}$) we have $\Pi^{J,[y] \cap J} \cap \Phi_{J'} \not\subseteq \Phi_{J' \smallsetminus \{r_{2k+1}\}}$, namely there exists a root $\beta \in \Pi^{J,[y] \cap J} \cap \Phi_{J'}$ which has non-zero coefficient of $\alpha_{2k+1}$. These $\beta$ and $\beta'$ satisfy $\langle \beta,\beta' \rangle \neq 0$ by the construction, concluding the proof. \end{proof}
By Lemma \ref{lem:proof_special_first_case_subcase_1_J_D_N} and Lemma \ref{lem:proof_special_first_case_subcase_1_local_irreducible_components}, the hypothesis of Lemma \ref{lem:proof_special_first_case_subcase_1_J_D_N} should not hold. By this fact, $A_{>1}$-freeness of $I$ and the property $N \geq |I \cap J| + 2 = |[y] \cap J| + 2$, it follows that the possibilities of $[y] \cap J$ are as follows (up to the symmetry $r_{N-1} \leftrightarrow r_N$); (I) $[y] \cap J = J \smallsetminus \{r_{2j} \mid 1 \leq j \leq k\}$ for an integer $k$ with $2 \leq k \leq (N-2)/2$ and $2k \neq N-3$; (II) $N$ is odd and $[y] \cap J = \{r_{2j-1} \mid 1 \leq j \leq (N-1)/2\}$; (III) $N$ is even and $[y] \cap J = \{r_{2j-1} \mid 1 \leq j \leq (N-2)/2\}$; (IV) $N$ is even and $[y] \cap J = \{r_{2j-1} \mid 1 \leq j \leq N/2\}$. For Case (I), by the shape of $J$ and $[y] \cap J$, it follows that $I \cap J = [y] \cap J$, and each local transformation can permute the irreducible components of $I \cap J$ containing neither $r_{N-1}$ nor $r_N$ but it fixes pointwise the irreducible component(s) of $I \cap J$ containing $r_{N-1}$ or $r_N$. This contradicts the fact that $\sigma = \tau^{-1}|_{I \cap J}$ for a nontrivial automorphism $\tau^{-1}$ of $J$ (note that $\tau^{-1}$ exchanges $r_{N-1}$ and $r_N$). Case (II) contradicts Lemma \ref{lem:possibility_J_is_D_N}(\ref{item:lem_possibility_J_is_D_N_case_2}). For Case (III), the roots $\alpha_{N-1} \in \Pi^{J,[y] \cap J}$ and $\alpha_t + \sum_{j=1}^{N-2} \alpha_j \in \Pi^{K,[y] \cap K} \smallsetminus \Phi_J$ are not orthogonal, contradicting Lemma \ref{lem:proof_special_first_case_subcase_1_local_irreducible_components}.
Finally, for the remaining case, i.e., Case (IV), by the shape of $J$ and $[y] \cap J$, it follows that $I \cap J = [y] \cap J$ and each local transformation leaves the set $I \cap J$ invariant. By this result and the property that $\sigma = \tau^{-1}|_{I \cap J}$ for a nontrivial automorphism $\tau^{-1}$ of $J$, only the possibility of $[y] \cap J$ is that $N = 4$ and $[y] \cap J = I \cap J = \{r_1,r_3\}$, and $\sigma$ exchanges $r_1$ and $r_3$. Now we arrange the standard decomposition $\mathcal{D}$ of $u$ as $u = \omega''_{\ell} \omega'_{\ell-1} \omega''_{\ell-1} \cdots \omega'_2 \omega''_2 \omega'_1 \omega''_1$, where each $\omega'_j$ is a wide transformation and each $\omega''_j$ is a (possibly empty) product of narrow transformations. Let each wide transformation $\omega'_j$ belong to $Y_{z'_j,z_j}$ with $z_j,z'_j \in S^{(\Lambda)}$. In particular, we have $\omega'_1 = \omega$ and $z_1 = y$. Now we give the following lemma: \begin{lem} \label{lem:proof_special_first_case_subcase_1_J_D_N_case_IV} In this setting, the following properties hold for every $1 \leq j \leq \ell - 1$: The action of the element $u_j := \omega''_j \omega'_{j-1} \omega''_{j-1} \cdots \omega'_1 \omega''_1$ maps $(r_1,r_2,r_3,r_4)$ to $(r_1,r_2,r_3,r_4)$ when $j$ is odd and to $(r_1,r_2,r_4,r_3)$ when $j$ is even; the subsets $J$ and $[z_j] \smallsetminus J$ are not adjacent; the support of $\omega'_j$ is as in Case \ref{item:proof_special_first_case_subcase_1_J_D_N} above, with $t$ replaced by some element $t_j \in S$; and $\omega'_j$ maps $(r_1,r_2,r_3,r_4)$ to $(r_1,r_2,r_4,r_3)$. \end{lem} \begin{proof} We use induction on $j$. By the definition of narrow transformations, the first and the second parts of the claim hold obviously when $j = 1$ and follow from the induction hypothesis when $j > 1$. In particular, we have $u_j \cdot \Pi_J = \Pi_J$. Put $(h,h') := (3,4)$ when $j$ is odd and $(h,h') := (4,3)$ when $j$ is even. Then we have $[z_j] \cap J = \{r_1,r_h\}$. Now, by using the above argument, it follows that the support of $\omega'_j$ is of the form $\{r_1,r_2,r_3,r_4,t_j\}$ which is the standard labelling of type $D_5$, where $t_j$ is adjacent to one of the two elements of $[z_j] \cap J$. We show that $t_j$ is adjacent to $r_1$, which already holds when $j = 1$ (note that $t_j = t$ when $j = 1$). Suppose $j > 1$ and assume contrary that $t_j$ is adjacent to $r_h$. In this case, $t_j$ is apart from $[z_j] \smallsetminus \{r_h\}$. On the other hand, we have $[z'_{j-1}] \cap J = \{r_1,r_h\}$, the subsets $[z'_{j-1}] \smallsetminus J$ and $J$ are not adjacent, and the support of each narrow transformation in $\omega''_j$ is apart from to $J$. Moreover, by the induction hypothesis, we have $[z_{j-1}] \cap J = \{r_1,r_{h'}\}$ and the action of $\omega'_{j-1}$ maps $(r_1,r_2,r_h,r_{h'})$ to $(r_1,r_2,r_{h'},r_h)$ while it fixes every element of $[z_{j-1}] \smallsetminus J$. This implies that $\omega''_j \in Y_{z'',z_{j-1}}$ for the element $z'' \in S^{(\Lambda)}$ obtained from $z_j$ by replacing $r_h$ with $r_{h'}$. Now we have $\alpha_{t_j} \in \Pi^{[z'']}$ since $t_j$ is not adjacent to $[z''] = ([z_j] \smallsetminus \{r_h\}) \cup \{r_{h'}\}$, therefore $\beta' := (\omega''_j)^{-1} \cdot \alpha_{t_j} \in \Pi^{[z_{j-1}]}$. This root belongs to $\Phi_{S \smallsetminus J}$ and has non-zero coefficient of $\alpha_{t_j}$, since the support of each narrow transformation in $\omega''_j$ is not adjacent to $J$ and hence does not contain $t_j$. Therefore, the roots $\beta' \in \Pi^{[z_{j-1}]} \smallsetminus \Pi^{J,[z_{j-1}] \cap J}$ and $\alpha_1 + 2\alpha_2 + \alpha_3 + \alpha_4 \in \Pi^{J,[z_{j-1}] \cap J}$ are not orthogonal. This contradicts the fact that $\Pi^{J,[y] \cap J}$ is the union of some irreducible components of $\Pi^{[y]}$ (see Lemma \ref{lem:proof_special_first_case_subcase_1_local_irreducible_components}) and the isomorphism $\Pi^{[y]} \to \Pi^{[z_{j-1}]}$ induced by the action of $\omega''_{j-1}\omega'_{j-2}\omega''_{j-2} \cdots \omega''_2\omega'_1$ maps $\Pi^{J,[y] \cap J}$ to $\Pi^{J,[z_{j-1}] \cap J}$ (since the action of this element leaves the set $\Pi_J$ invariant). This contradiction proves that $t_j$ is adjacent to $r_1$, therefore the third part of the claim holds. Finally, the fourth part of the claim follows immediately from the third part. Hence the proof of Lemma \ref{lem:proof_special_first_case_subcase_1_J_D_N_case_IV} is concluded. \end{proof} By Lemma \ref{lem:proof_special_first_case_subcase_1_J_D_N_case_IV}, the action of the element $\omega'_{\ell-1} \omega''_{\ell-1} \cdots \omega'_2 \omega''_2 \omega'_1 \omega''_1$, hence of $u = \omega''_{\ell}\omega'_{\ell-1}u_{\ell-1}$, maps the elements $(r_1,r_2,r_3,r_4)$ to either $(r_1,r_2,r_3,r_4)$ or $(r_1,r_2,r_4,r_3)$. This contradicts the above-mentioned fact that $\sigma$ exchanges $r_1$ and $r_3$.
Summarizing, we have derived a contradiction in each of the six possible cases, Cases \ref{item:proof_special_first_case_subcase_1_J_E_6_K_E_8}--\ref{item:proof_special_first_case_subcase_1_J_D_N}. Hence we have proven that the assumption $w \cdot \Pi_J \not\subseteq \Phi^+$ implies a contradiction, as desired.
\subsubsection{Case $w \cdot \Pi_J \subseteq \Phi^+$} \label{sec:proof_special_first_case_subcase_2}
By the result of Section \ref{sec:proof_special_first_case_subcase_1}, we have $w \cdot \Pi_J \subseteq \Phi^+$. Since $w \cdot \Phi_J = \Phi_J$ by Lemma \ref{lem:property_w}, it follows that $w \cdot \Phi_J^+ \subseteq \Phi_J^+$, therefore $w \cdot \Phi_J^+ = \Phi_J^+$ (note that $|\Phi_J| < \infty$). Hence the action of $w$ defines an automorphism $\tau$ of $J$ (in particular, $w \cdot \Pi_J = \Pi_J$). To show that $\tau$ is the identity mapping (which implies the claim that $w$ fixes $\Pi^{J,I \cap J}$ pointwise), assume contrary that $\tau$ is a nontrivial automorphism of $J$. Then the possibilities of the type of $J$ are as follows: $D_N$, $E_6$ and $F_4$ (recall that $J$ is neither of type $A_N$ nor of type $I_2(m)$). Moreover, since the action of $w \in Y_I$ fixes every element of $I \cap J$, the subset $I \cap J$ of $J$ is contained in the fixed point set of $\tau$. This implies that $J$ is not of type $F_4$, since the nontrivial automorphism of a Coxeter graph of type $F_4$ has no fixed points.
Suppose that $J$ is of type $E_6$. Then, by the above argument on the fixed points of $\tau$ and Lemma \ref{lem:possibility_J_is_E_6}, we have $I \cap J = \{r_2\}$ or $I \cap J = \{r_4\}$. Now take a standard decomposition of $w$ with respect to $J$ (see Proposition \ref{prop:standard_decomposition_existence}). Then no wide transformation can appear due to the shape of $J$ and the position of $I \cap J$ in $J$ (indeed, we cannot obtain a subset of finite type by adding to $J$ an element of $S$ adjacent to $I \cap J$). This implies that the decomposition of $w$ consists of narrow transformations only, therefore $w$ fixes $\Pi_J$ pointwise, contradicting the fact that $\tau$ is a nontrivial automorphism.
Secondly, suppose that $J$ is of type $D_N$ with $N \geq 5$. Then, by the above argument on the fixed points of $\tau$, we have $I \cap J \subseteq J \smallsetminus \{r_{N-1},r_N\}$, therefore every irreducible component of $I \cap J$ is of type $A_1$ (by $A_{>1}$-freeness of $I$). Now take a standard decomposition $\mathcal{D}$ of $w$ with respect to $J$ (see Proposition \ref{prop:standard_decomposition_existence}). Note that $\mathcal{D}$ involves at least one wide transformation, since $\tau$ is not the identity mapping. By the shape of $J$ and the position of $I \cap J$ in $J$, only the possibility of the first (from right) wide transformation $\omega = \omega_i$ in $\mathcal{D}$ is as follows: $K = J \cup \{t\}$ is of type $D_{N+1}$, $t$ is adjacent to $r_1$, and $r_1 \in [y]$, where we put $y = y^{(i)}(\mathcal{D})$, $t = t^{(i)}(\mathcal{D})$, and $K = K^{(i)}(\mathcal{D})$. Now the claim of Lemma \ref{lem:proof_special_first_case_subcase_1_J_D_N} in Section \ref{sec:proof_special_first_case_subcase_1} also holds in this case, while $\Pi^{J,[y] \cap J}$ is the union of some irreducible components of $\Pi^{[y]}$ by the same reason as in Section \ref{sec:proof_special_first_case_subcase_1}. Hence the hypothesis of Lemma \ref{lem:proof_special_first_case_subcase_1_J_D_N} should not hold. This argument and the properties that $N \geq |I \cap J| + 2 = |[y] \cap J| + 2$ and $I \cap J \subseteq J \smallsetminus \{r_{N-1},r_N\}$ imply that the possibilities of $[y] \cap J$ are the followings: $N$ is odd and $[y] \cap J$ consists of elements $r_{2j-1}$ with $1 \leq j \leq (N-1)/2$; or, $N$ is even and $[y] \cap J$ consists of elements $r_{2j-1}$ with $1 \leq j \leq (N-2)/2$. The former possibility contradicts Lemma \ref{lem:possibility_J_is_D_N}(\ref{item:lem_possibility_J_is_D_N_case_2}). On the other hand, for the latter possibility, the roots $\alpha_{N-1} \in \Pi^{J,[y] \cap J}$ and $\alpha_t + \sum_{j=1}^{N-2} \alpha_j \in \Pi^{[y]} \smallsetminus \Pi^{J,[y] \cap J}$ are not orthogonal, contradicting the above-mentioned fact that $\Pi^{J,[y] \cap J}$ is the union of some irreducible components of $\Pi^{[y]}$. Hence we have a contradiction for any of the two possibilities.
Finally, we consider the remaining case that $J$ is of type $D_4$. By the property $N = 4 \geq |I \cap J| + 2$ and $A_{>1}$-freeness of $I$, it follows that $I \cap J$ consists of at most two irreducible components of type $A_1$. On the other hand, by the shape of $J$, the fixed point set of the nontrivial automorphism $\tau$ of $J$ is of type $A_1$ or $A_2$. Since $I \cap J$ is contained in the fixed point set of $\tau$ as mentioned above, it follows that $|I \cap J| = 1$. If $I \cap J = \{r_1\}$, then we have $\Pi^{J,I \cap J} = \{\alpha_3,\alpha_4,\beta\}$ where $\beta = \alpha_1 + 2\alpha_2 + \alpha_3 + \alpha_4$ (see Table \ref{tab:positive_roots_D_n}), and every element of $\Pi^{J,I \cap J}$ forms an irreducible component of $\Pi^{J,I \cap J}$. However, now the property $w \cdot \Pi_J = \Pi_J$ implies that $w$ fixes $\alpha_2$ and permutes the three simple roots $\alpha_1$, $\alpha_3$ and $\alpha_4$, therefore $w \cdot \beta = \beta$, contradicting the fact that $\langle w \rangle$ acts transitively on the set of the irreducible components of $\Pi^{J,I \cap J}$ (see Lemma \ref{lem:property_w}). By symmetry, the same result holds when $I \cap J = \{r_3\}$ or $\{r_4\}$. Hence we have $I \cap J = \{r_2\}$. Take a standard decomposition of $w$ with respect to $J$ (see Proposition \ref{prop:standard_decomposition_existence}). Then no wide transformation can appear due to the shape of $J$ and the position of $I \cap J$ in $J$ (indeed, we cannot obtain a subset of finite type by adding to $J$ an element of $S$ adjacent to $I \cap J$). This implies that the decomposition of $w$ consists of narrow transformations only, therefore $w$ fixes $\Pi_J$ pointwise, contradicting the fact that $\tau$ is a nontrivial automorphism.
Summarizing, we have derived in any case a contradiction from the assumption that $\tau$ is a nontrivial automorphism. Hence it follows that $\tau$ is the identity mapping, therefore our claim has been proven in the case $\Pi^{J,I \cap J} \not\subseteq \Phi_{I^{\perp}}$.
\subsection{The second case $\Pi^{J,I \cap J} \subseteq \Phi_{I^{\perp}}$} \label{sec:proof_special_second_case}
In this subsection, we consider the remaining case that $\Pi^{J,I \cap J} \subseteq \Phi_{I^{\perp}}$. In this case, we have $\Pi_{I^{\perp}} \subseteq \Pi^I$, therefore $\Pi^{J,I \cap J} = \Pi_{J \smallsetminus I}$. Let $L$ be an irreducible component of $J \smallsetminus I$. Then $L$ is of finite type. The aim of the following argument is to show that $w$ fixes $\Pi_L$ pointwise; indeed, if this is satisfied, then we have $\Pi^{J,I \cap J} = \Pi_{J \smallsetminus I} = \Pi_L$ since $\langle w \rangle$ acts transitively on the set of irreducible components of $\Pi^{J,I \cap J}$ (see Lemma \ref{lem:property_w}), therefore $w$ fixes $\Pi^{J,I \cap J}$ pointwise, as desired. Note that $w \cdot \Pi_L \subseteq \Pi_{J \smallsetminus I}$, since now $w$ leaves the set $\Pi^{J,I \cap J} = \Pi_{J \smallsetminus I}$ invariant.
\subsubsection{Possibilities of semi-standard decompositions} \label{sec:proof_special_second_case_transformations}
Here we investigate the possibilities of narrow and wide transformations in a semi-standard decomposition of the element $w$, in a somewhat wider context. Let $\mathcal{D} = \omega_{\ell(\mathcal{D})-1} \cdots \omega_1\omega_0$ be a semi-standard decomposition of an element $u$ of $W$, with the property that $[y^{(0)}]$ is isomorphic to $I$, $J^{(0)}$ is irreducible and of finite type, and $J^{(0)}$ is apart from $[y^{(0)}]$. Note that any semi-standard decomposition of the element $w \in Y_I$ with respect to the set $L$ defined above satisfies the condition. Note also that $\mathcal{D}^{-1} := (\omega_0)^{-1}(\omega_1)^{-1} \cdots (\omega_{\ell(\mathcal{D})-1})^{-1}$ is also a semi-standard decomposition of $u^{-1}$, and $(\omega_i)^{-1}$ is a narrow (respectively, wide) transformation if and only if $\omega_i$ is a narrow (respectively, wide) transformation.
The proof of the next lemma uses a concrete description of root systems of all finite irreducible Coxeter groups except types $A$ and $I_2(m)$. Table \ref{tab:positive_roots_B_n} shows the list for type $B_n$, where the notational conventions are similar to the case of type $D_n$ (Table \ref{tab:positive_roots_D_n}). For the list for type $F_4$ (Table \ref{tab:positive_roots_F_4}), the list includes only one of the two conjugacy classes of positive roots (denoted by $\gamma_i^{(1)}$), and the other positive roots (denoted by $\gamma_i^{(2)}$) are obtained by using the symmetry $r_1 \leftrightarrow r_4$, $r_2 \leftrightarrow r_3$. In the list, $[c_1,c_2,c_3,c_4]$ signifies a positive root $c_1 \alpha_1 + c_2 \alpha_2 + c_3\alpha_3 + c_4\alpha_4$, and the description in the columns for actions of generators is similar to the case of type $E_8$ (Tables \ref{tab:positive_roots_E_8_1}--\ref{tab:positive_roots_E_8_6}). The list for type $H_4$ is divided into two parts (Tables \ref{tab:positive_roots_H_4_1} and \ref{tab:positive_roots_H_4_2}). In the list, $[c_1,c_2,c_3,c_4]$ signifies a positive root $c_1 \alpha_1 + c_2 \alpha_2 + c_3\alpha_3 + c_4\alpha_4$, where we put $c = 2\cos(\pi/5)$ for simplicity and therefore $c^2 = c+1$. The other description is in a similar manner as the case of type $E_8$, and the marks \lq\lq $H_3$'' indicate the positive roots of the parabolic subgroup of type $H_3$ generated by $\{r_1,r_2,r_3\}$. \begin{table}[hbt] \centering \caption{List of positive roots for Coxeter group of type $B_n$} \label{tab:positive_roots_B_n}
\begin{tabular}{|c|c|} \hline roots & actions of generators \\ \hline $\gamma^{(1)}_{i,j} := \sum_{h=i}^{j} \alpha_h$ & $r_{i-1} \cdot \gamma^{(1)}_{i,j} = \gamma^{(1)}_{i-1,j}$ ($i \geq 2$) \\ ($1 \leq i \leq j \leq n-1$) & $r_i \cdot \gamma^{(1)}_{i,j} = \gamma^{(1)}_{i+1,j}$ ($i \leq j-1$) \\ ($\gamma^{(1)}_{i,i} = \alpha_i$) & $r_j \cdot \gamma^{(1)}_{i,j} = \gamma^{(1)}_{i,j-1}$ ($i \leq j-1$) \\ & $r_{j+1} \cdot \gamma^{(1)}_{i,j} = \gamma^{(1)}_{i,j+1}$ ($j \leq n-2$) \\ & $r_n \cdot \gamma^{(1)}_{i,n-1} = \gamma^{(2)}_{i,n}$ \\ \hline $\gamma^{(2)}_{i,j} := \sum_{h=i}^{j-1} \alpha_h + \sum_{h=j}^{n-1} 2\alpha_h + \sqrt{2}\alpha_n$ & $r_{i-1} \cdot \gamma^{(2)}_{i,j} = \gamma^{(2)}_{i-1,j}$ ($i \geq 2$) \\ ($1 \leq i < j \leq n$) & $r_i \cdot \gamma^{(2)}_{i,j} = \gamma^{(2)}_{i+1,j}$ ($i \leq j-2$) \\ & $r_{j-1} \cdot \gamma^{(2)}_{i,j} = \gamma^{(2)}_{i,j-1}$ ($i \leq j-2$) \\ & $r_j \cdot \gamma^{(2)}_{i,j} = \gamma^{(2)}_{i,j+1}$ ($j \leq n-1$) \\ & $r_n \cdot \gamma^{(2)}_{i,n} = \gamma^{(1)}_{i,n-1}$ \\ \hline $\gamma^{(3)}_i := \sum_{h=i}^{n-1} \sqrt{2}\alpha_h + \alpha_n$ & $r_{i-1} \cdot \gamma^{(3)}_i = \gamma^{(3)}_{i-1}$ ($i \geq 2$) \\ ($1 \leq i \leq n$) & $r_i \cdot \gamma^{(3)}_i = \gamma^{(3)}_{i+1}$ ($i \leq n-1$) \\ ($\gamma^{(3)}_n = \alpha_n$) & \\ \hline \end{tabular} \end{table} \begin{table}[hbt] \centering \caption{List of positive roots for Coxeter group of type $F_4$} \label{tab:positive_roots_F_4} The data of the remaining positive roots $\gamma^{(2)}_i$ are obtained by replacing $[c_1,c_2,c_3,c_4]$ with $[c_4,c_3,c_2,c_1]$ and replacing each $r_j$ with $r_{5-j}$.\\
\begin{tabular}{|c||c|c|c|c|c|c|} \cline{1-7}
height & $i$ & root $\gamma^{(1)}_i$ & \multicolumn{4}{|c|}{$k$; $r_j \cdot \gamma^{(1)}_i = \gamma^{(1)}_k$} \\ \cline{4-7} & & & $r_1$ & $r_2$ & $r_3$ & $r_4$ \\ \cline{1-7} $1$ & $1$ & $[1,0,0,0]$ & --- & $3$ & & \\ \cline{2-7} & $2$ & $[0,1,0,0]$ & $3$ & --- & $4$ & \\ \cline{1-7} $2$ & $3$ & $[1,1,0,0]$ & $2$ & $1$ & $5$ & \\ \cline{2-7} & $4$ & $[0,1,\sqrt{2},0]$ & $5$ & & $2$ & $6$ \\ \cline{1-7} $3$ & $5$ & $[1,1,\sqrt{2},0]$ & $4$ & $7$ & $3$ & $8$ \\ \cline{2-7} & $6$ & $[0,1,\sqrt{2},\sqrt{2}]$ & $8$ & & & $4$ \\ \cline{1-7} $4$ & $7$ & $[1,2,\sqrt{2},0]$ & & $5$ & & $9$ \\ \cline{2-7} & $8$ & $[1,1,\sqrt{2},\sqrt{2}]$ & $6$ & $9$ & & $5$ \\ \cline{1-7} $5$ & $9$ & $[1,2,\sqrt{2},\sqrt{2}]$ & & $8$ & $10$ & $7$ \\ \cline{1-7} $6$ & $10$ & $[1,2,2\sqrt{2},\sqrt{2}]$ & & $11$ & $9$ & \\ \cline{1-7} $7$ & $11$ & $[1,3,2\sqrt{2},\sqrt{2}]$ & $12$ & $10$ & & \\ \cline{1-7} $8$ & $12$ & $[2,3,2\sqrt{2},\sqrt{2}]$ & $11$ & & & \\ \cline{1-7} \end{tabular} \end{table} \begin{table}[hbt] \centering \caption{List of positive roots for Coxeter group of type $H_4$ (part $1$), where $c = 2\cos(\pi/5)$, $c^2 = c + 1$} \label{tab:positive_roots_H_4_1}
\begin{tabular}{|c||c|c|c|c|c|c|c} \cline{1-7}
height & $i$ & root $\gamma_i$ & \multicolumn{4}{|c|}{$k$; $r_j \cdot \gamma_i = \gamma_k$} \\ \cline{4-7} & & & $r_1$ & $r_2$ & $r_3$ & $r_4$ \\ \cline{1-7} $1$ & $1$ & $[1,0,0,0]$ & --- & $5$ & & & $H_3$ \\ \cline{2-7} & $2$ & $[0,1,0,0]$ & $6$ & --- & $7$ & & $H_3$ \\ \cline{2-7} & $3$ & $[0,0,1,0]$ & & $7$ & --- & $8$ & $H_3$ \\ \cline{2-7} & $4$ & $[0,0,0,1]$ & & & $8$ & --- \\ \cline{1-7} $2$ & $5$ & $[1,c,0,0]$ & $9$ & $1$ & $10$ & & $H_3$ \\ \cline{2-7} & $6$ & $[c,1,0,0]$ & $2$ & $9$ & $11$ & & $H_3$ \\ \cline{2-7} & $7$ & $[0,1,1,0]$ & $11$ & $3$ & $2$ & $12$ & $H_3$ \\ \cline{2-7} & $8$ & $[0,0,1,1]$ & & $12$ & $4$ & $3$ \\ \cline{1-7} $3$ & $9$ & $[c,c,0,0]$ & $5$ & $6$ & $13$ & & $H_3$ \\ \cline{2-7} & $10$ & $[1,c,c,0]$ & $13$ & & $5$ & $14$ & $H_3$ \\ \cline{2-7} & $11$ & $[c,1,1,0]$ & $7$ & $15$ & $6$ & $16$ & $H_3$ \\ \cline{2-7} & $12$ & $[0,1,1,1]$ & $16$ & $12$ & & $7$ \\ \cline{1-7} $4$ & $13$ & $[c,c,c,0]$ & $10$ & $17$ & $9$ & $18$ & $H_3$ \\ \cline{2-7} & $14$ & $[1,c,c,c]$ & $18$ & & & $10$ \\ \cline{2-7} & $15$ & $[c,c+1,1,0]$ & $19$ & $11$ & $17$ & $20$ & $H_3$ \\ \cline{2-7} & $16$ & $[c,1,1,1]$ & $12$ & $20$ & & $11$ \\ \cline{1-7} $5$ & $17$ & $[c,c+1,c,0]$ & $21$ & $13$ & $15$ & $22$ & $H_3$ \\ \cline{2-7} & $18$ & $[c,c,c,c]$ & $14$ & $22$ & & $13$ \\ \cline{2-7} & $19$ & $[c+1,c+1,1,0]$ & $15$ & & $21$ & $23$ & $H_3$ \\ \cline{2-7} & $20$ & $[c,c+1,1,1]$ & $23$ & $16$ & $24$ & $15$ \\ \cline{1-7} $6$ & $21$ & $[c+1,c+1,c,0]$ & $17$ & $25$ & $19$ & $26$ & $H_3$ \\ \cline{2-7} & $22$ & $[c,c+1,c,c]$ & $26$ & $18$ & $27$ & $17$ \\ \cline{2-7} & $23$ & $[c+1,c+1,1,1]$ & $20$ & & $28$ & $19$ \\ \cline{2-7} & $24$ & $[c,c+1,c+1,1]$ & $28$ & & $20$ & $27$ \\ \cline{1-7} \end{tabular} \end{table} \begin{table}[p] \centering \caption{List of positive roots for Coxeter group of type $H_4$ (part $2$), where $c = 2\cos(\pi/5)$, $c^2 = c + 1$} \label{tab:positive_roots_H_4_2}
\begin{tabular}{|c||c|c|c|c|c|c|c} \cline{1-7}
height & $i$ & root $\gamma_i$ & \multicolumn{4}{|c|}{$k$; $r_j \cdot \gamma_i = \gamma_k$} \\ \cline{4-7} & & & $r_1$ & $r_2$ & $r_3$ & $r_4$ \\ \cline{1-7} $7$ & $25$ & $[c+1,2c,c,0]$ & & $21$ & & $29$ & $H_3$ \\ \cline{2-7} & $26$ & $[c+1,c+1,c,c]$ & $22$ & $29$ & $30$ & $21$ \\ \cline{2-7} & $27$ & $[c,c+1,c+1,c]$ & $30$ & & $22$ & $24$ \\ \cline{2-7} & $28$ & $[c+1,c+1,c+1,1]$ & $24$ & $31$ & $23$ & $30$ \\ \cline{1-7} $8$ & $29$ & $[c+1,2c,c,c]$ & & $26$ & $32$ & $25$ \\ \cline{2-7} & $30$ & $[c+1,c+1,c+1,c]$ & $27$ & $33$ & $26$ & $28$ \\ \cline{2-7} & $31$ & $[c+1,2c+1,c+1,1]$ & $34$ & $28$ & & $33$ \\ \cline{1-7} $9$ & $32$ & $[c+1,2c,2c,c]$ & & $35$ & $29$ & \\ \cline{2-7} & $33$ & $[c+1,2c+1,c+1,c]$ & $36$ & $30$ & $35$ & $31$ \\ \cline{2-7} & $34$ & $[2c+1,2c+1,c+1,1]$ & $31$ & $37$ & & $36$ \\ \cline{1-7} $10$ & $35$ & $[c+1,2c+1,2c,c]$ & $38$ & $32$ & $33$ & \\ \cline{2-7} & $36$ & $[2c+1,2c+1,c+1,c]$ & $33$ & $39$ & $38$ & $34$ \\ \cline{2-7} & $37$ & $[2c+1,2c+2,c+1,1]$ & & $34$ & $40$ & $39$ \\ \cline{1-7} $11$ & $38$ & $[2c+1,2c+1,2c,c]$ & $35$ & $41$ & $36$ & \\ \cline{2-7} & $39$ & $[2c+1,2c+2,c+1,c]$ & & $36$ & $42$ & $37$ \\ \cline{2-7} & $40$ & $[2c+1,2c+2,c+2,1]$ & & & $37$ & $43$ \\ \cline{1-7} $12$ & $41$ & $[2c+1,3c+1,2c,c]$ & $44$ & $38$ & $45$ & \\ \cline{2-7} & $42$ & $[2c+1,2c+2,2c+1,c]$ & & $45$ & $39$ & $46$ \\ \cline{2-7} & $43$ & $[2c+1,2c+2,c+2,c+1]$ & & & $46$ & $40$ \\ \cline{1-7} $13$ & $44$ & $[2c+2,3c+1,2c,c]$ & $41$ & & $47$ & \\ \cline{2-7} & $45$ & $[2c+1,3c+1,2c+1,c]$ & $47$ & $42$ & $41$ & $48$ \\ \cline{2-7} & $46$ & $[2c+1,2c+2,2c+1,c+1]$ & & $48$ & $43$ & $42$ \\ \cline{1-7} $14$ & $47$ & $[2c+2,3c+1,2c+1,c]$ & $45$ & $49$ & $44$ & $50$ \\ \cline{2-7} & $48$ & $[2c+1,3c+1,2c+1,c+1]$ & $50$ & $46$ & & $45$ \\ \cline{1-7} $15$ & $49$ & $[2c+2,3c+2,2c+1,c]$ & $51$ & $47$ & & $52$ \\ \cline{2-7} & $50$ & $[2c+2,3c+1,2c+1,c+1]$ & $48$ & $52$ & & $47$ \\ \cline{1-7} $16$ & $51$ & $[3c+1,3c+2,2c+1,c]$ & $49$ & & & $53$ \\ \cline{2-7} & $52$ & $[2c+2,3c+2,2c+1,c+1]$ & $53$ & $50$ & $54$ & $49$ \\ \cline{1-7} $17$ & $53$ & $[3c+1,3c+2,2c+1,c+1]$ & $52$ & & $55$ & $51$ \\ \cline{2-7} & $54$ & $[2c+2,3c+2,2c+2,c+1]$ & $55$ & & $52$ & \\ \cline{1-7} $18$ & $55$ & $[3c+1,3c+2,2c+2,c+1]$ & $54$ & $56$ & $53$ & \\ \cline{1-7} $19$ & $56$ & $[3c+1,3c+3,2c+2,c+1]$ & $57$ & $55$ & & \\ \cline{1-7} $20$ & $57$ & $[3c+2,3c+3,2c+2,c+1]$ & $56$ & $58$ & & \\ \cline{1-7} $21$ & $58$ & $[3c+2,4c+2,2c+2,c+1]$ & & $57$ & $59$ & \\ \cline{1-7} $22$ & $59$ & $[3c+2,4c+2,3c+1,c+1]$ & & & $58$ & $60$ \\ \cline{1-7} $23$ & $60$ & $[3c+2,4c+2,3c+1,2c]$ & & & & $59$ \\ \cline{1-7} \end{tabular} \end{table}
Then, for the wide transformations in $\mathcal{D}$, we have the following: \begin{lem} \label{lem:proof_special_second_case_transformations_wide} In this setting, if $\omega_i$ is a wide transformation, then there exist only the following two possibilities, where $K^{(i)} = \{r_1,r_2,\dots,r_N\}$ is the standard labelling of $K^{(i)}$ given in Section \ref{sec:longestelement}: \begin{enumerate} \item $K^{(i)}$ is of type $A_N$ with $N \geq 3$, $t^{(i)} = r_2$, $[y^{(i)}] \cap K^{(i)} = \{r_1\}$ and $J^{(i)} = \{r_3,\dots,r_N\}$; now the action of $\omega_i$ maps $r_1$ to $r_N$ and $(r_3,r_4,\dots,r_N)$ to $(r_1,r_2,\dots,r_{N-2})$; \item $K^{(i)}$ is of type $E_7$, $t^{(i)} = r_6$, $[y^{(i)}] \cap K^{(i)} = \{r_1,r_2,r_3,r_4,r_5\}$ and $J^{(i)} = \{r_7\}$; now the action of $\omega_i$ maps $(r_1,r_2,r_3,r_4,r_5)$ to $(r_1,r_5,r_3,r_4,r_2)$ and $r_7$ to $r_7$. \end{enumerate} Hence, if $\mathcal{D}$ involves a wide transformation, then $J^{(0)}$ is of type $A_{N'}$ with $1 \leq N' < \infty$. \end{lem} \begin{proof} The latter part of the claim follows from the former part and the fact that the sets $J^{(i)}$ for $0 \leq i \leq \ell(\mathcal{D})$ are all isomorphic to each other. For the former part, note that $J^{(i)}$ is an irreducible subset of $K^{(i)}$ which is not adjacent to $[y^{(i)}]$ (by the above condition that $J^{(0)}$ is apart from $[y^{(0)}]$), $t^{(i)}$ is adjacent to $[y^{(i)}] \cap K^{(i)}$, and $\omega_i$ cannot fix the set $\Pi_{K^{(i)} \smallsetminus \{t\}}$ pointwise (see Lemma \ref{lem:another_decomposition_Y_no_loop}). Moreover, since $I$ is $A_{>1}$-free, $[y^{(i)}]$ is also $A_{>1}$-free. By these properties, a case-by-case argument shows that the possibilities of $K^{(i)}$, $[y^{(i)}]$ and $t^{(i)}$ are as enumerated in Table \ref{tab:lem_proof_special_second_case_transformations_wide} up to symmetry (note that $J^{(i)} = K^{(i)} \smallsetminus ([y^{(i)}] \cup \{t^{(i)}\})$). Now, for each case in Table \ref{tab:lem_proof_special_second_case_transformations_wide} except the two cases specified in the statement, it follows by using the tables for the root systems of finite irreducible Coxeter groups that there exists a root $\beta \in (\Phi_{K^{(i)}}^{\perp [y^{(i)}] \cap K^{(i)}})^+$ that has non-zero coefficient of $\alpha_{t^{(i)}}$, as listed in Table \ref{tab:lem_proof_special_second_case_transformations_wide} (where the notations for the roots $\beta$ are as in the tables). This implies that $\omega_i \cdot \beta \in \Phi^-$. Moreover, the definition of $K^{(i)}$ implies that the set $[y^{(i)}] \smallsetminus K^{(i)}$ is apart from $K^{(i)}$, therefore $\beta \in \Phi^{\perp [y^{(i)}]}$ and $\Phi^{\perp [y^{(i)}]}[\omega_i] \neq \emptyset$. However, this contradicts the property $\omega_i \in Y_{y^{(i+1)},y^{(i)}}$. Hence one of the two conditions specified in the statement should be satisfied, concluding the proof of Lemma \ref{lem:proof_special_second_case_transformations_wide}. \end{proof} \begin{table}[hbt] \centering \caption{List for the proof of Lemma \ref{lem:proof_special_second_case_transformations_wide}} \label{tab:lem_proof_special_second_case_transformations_wide}
\begin{tabular}{|c|c|c|c|} \hline type of $K^{(i)}$ & $[y^{(i)}] \cap K^{(i)}$ & $t^{(i)}$ & $\beta$ \\ \hline $A_N$ ($N \geq 3$) & $\{r_1\}$ & $r_2$ & --- \\ \hline $B_N$ ($N \geq 4$) & $\{r_{k+1},\dots,r_N\}$ ($3 \leq k \leq N-1$) & $r_k$ & $\gamma^{(3)}_{k}$ \\ \hline $D_N$ & $\{r_{N-1},r_N\}$ & $r_{N-2}$ & $\gamma^{(4)}_{1,2}$ \\ \cline{2-3} & $\{r_{k+1},\dots,r_{N-1},r_N\}$ ($2 \leq k \leq N-4$) & $r_k$ & \\ \hline $E_N$ ($6 \leq N \leq 8$) & $\{r_1\}$ & $r_3$ & $\gamma_{44}$ \\ \hline $E_7$ & $\{r_1,r_2,r_3,r_4,r_5\}$ & $r_6$ & --- \\ \cline{2-4} & $\{r_7\}$ & $r_6$ & $\gamma_{61}$ \\ \hline $E_8$ & $\{r_1,r_2,r_3,r_4,r_5\}$ & $r_6$ & $\gamma_{119}$ \\ \cline{2-3} & $\{r_1,r_2,r_3,r_4,r_5,r_6\}$ & $r_7$ & \\ \cline{2-4} & $\{r_8\}$ & $r_7$ & $\gamma_{74}$ \\ \hline $F_4$ & $\{r_1\}$ & $r_2$ & $\gamma^{(1)}_7$ \\ \hline $H_4$ & $\{r_1\}$ & $r_2$ & $\gamma_{40}$ \\ \cline{2-3} & $\{r_1,r_2\}$ & $r_3$ & \\ \cline{2-4} & $\{r_4\}$ & $r_3$ & $\gamma_{32}$ \\ \hline \end{tabular} \end{table}
On the other hand, for the narrow transformations in $\mathcal{D}$, we have the following: \begin{lem} \label{lem:proof_special_second_case_transformations_narrow} In this setting, suppose that $\omega_i$ is a narrow transformation, $[y^{(i+1)}] \neq [y^{(i)}]$, and $K^{(i)} \cap [y^{(i)}] = K^{(i)} \smallsetminus \{t^{(i)}\}$ has an irreducible component of type $A_1$. Then $K^{(i)}$ is of type $A_2$ or of type $I_2(m)$ with $m$ an odd number. \end{lem} \begin{proof} First, by the condition $[y^{(i+1)}] \neq [y^{(i)}]$ and the definition of $\omega_i$, the action of the longest element of $W_{K^{(i)}}$ induces a nontrivial automorphism of $K^{(i)}$ which does not fix the element $t^{(i)}$. This property restricts the possibilities of $K^{(i)}$ to one of the followings (where we use the standard labelling of $K^{(i)}$): $K^{(i)} = \{r_1,\dots,r_N\}$ is of type $A_N$ and $t^{(i)} \neq r_{(N+1)/2}$; $K^{(i)} = \{r_1,\dots,r_N\}$ is of type $D_N$ with $N$ odd and $t^{(i)} \in \{r_{N-1},r_N\}$; $K^{(i)} = \{r_1,\dots,r_6\}$ is of type $E_6$ and $t^{(i)} \not\in \{r_2,r_4\}$; or $K^{(i)}$ is of type $I_2(m)$ with $m$ odd. Secondly, by considering the $A_{>1}$-freeness of $I$ (hence of $[y^{(i)}]$), the possibilities are further restricted to the followings: $K^{(i)}$ is of type $A_2$; $K^{(i)}$ is of type $E_6$ and $t^{(i)} \in \{r_1,r_6\}$; and $K^{(i)}$ is of type $I_2(m)$ with $m$ odd. Moreover, by the hypothesis that $K^{(i)} \cap [y^{(i)}]$ has an irreducible component of type $A_1$, the above possibility of type $E_6$ is denied. Hence the claim holds. \end{proof}
\subsubsection{Proof of the claim} \label{sec:finitepart_secondcase_N_2}
From now, we prove our claim that $w$ fixes the set $\Pi_L$ pointwise. First, we have $w \cdot \Pi_L \subseteq \Pi_{J \smallsetminus I}$ as mentioned above, therefore Proposition \ref{prop:standard_decomposition_existence} implies that there exists a standard decomposition of $w$ with respect to $L$. Moreover, $L$ is apart from $I = [x_I]$, since $\Pi_L$ is an irreducible component of $\Pi^I$. Now if $L$ is not of type $A_N$ with $1 \leq N < \infty$, then Lemma \ref{lem:proof_special_second_case_transformations_wide} implies that the standard decomposition of $w$ involves no wide transformations, therefore $w$ fixes $\Pi_L$ pointwise, as desired (note that any narrow transformation $\omega_i$ fixes $\Pi_{J^{(i)}}$ pointwise by the definition). Hence, from now, we consider the case that $L$ is of type $A_N$ with $1 \leq N < \infty$.
First, we present some definitions: \begin{defn} \label{defn:admissible_for_N_large} Suppose that $2 \leq N < \infty$. Let $\mathcal{D} = \omega_{\ell(\mathcal{D})-1} \cdots \omega_1\omega_0$ be a semi-standard decomposition of an element of $W$. We say that a sequence $s_1,s_2,\dots,s_{\mu}$ of distinct elements of $S$ is \emph{admissible of type $A_N$} with respect to $\mathcal{D}$, if $J^{(0)}$ is of type $A_N$, $\mu \equiv N \pmod{2}$, and the following conditions are satisfied, where we put $M := \{s_1,s_2\dots,s_{\mu}\}$ (see Figure \ref{fig:finitepart_secondcase_N_2_definition}): \begin{enumerate} \item \label{item:admissible_N_large_J_irreducible_component} $\Pi_{J^{(0)}}$ is an irreducible component of $\Pi^{[y^{(0)}]}$. \item \label{item:admissible_N_large_M_line} $m(s_j,s_{j+1}) = 3$ for every $1 \leq j \leq \mu-1$. \item \label{item:admissible_N_large_J_in_M} For each $0 \leq h \leq \ell(\mathcal{D})$, there exists an odd number $\lambda(h)$ with $1 \leq \lambda(h) \leq \mu - N + 1$ satisfying the following conditions, where we put $\rho(h) := \lambda(h) + N - 1$: \begin{displaymath} J^{(h)} = \{s_j \mid \lambda(h) \leq j \leq \rho(h)\} \enspace, \end{displaymath} \begin{displaymath} \begin{split} [y^{(h)}] \cap M &= \{s_j \mid 1 \leq j \leq \lambda(h) - 2 \mbox{ and } j \equiv 1 \pmod{2}\} \\ &\cup \{s_j \mid \rho(h) + 2 \leq j \leq \mu \mbox{ and } j \equiv \mu \pmod{2}\} \enspace. \end{split} \end{displaymath} \item \label{item:admissible_N_large_y_isolated} For each $0 \leq h \leq \ell(\mathcal{D})$, every element of $[y^{(h)}] \cap M$ forms an irreducible component of $[y^{(h)}]$ of type $A_1$. \item \label{item:admissible_N_large_narrow_transformation} For each $0 \leq h \leq \ell(\mathcal{D})-1$, if $\omega_h$ is a narrow transformation, then one of the following two conditions is satisfied: \begin{itemize} \item $K^{(h)}$ intersects with $[y^{(h)}] \cap M$, and $[y^{(h+1)}] = [y^{(h)}]$; \item $K^{(h)}$ is apart from $[y^{(h)}] \cap M$ (hence $[y^{(h+1)}] \cap M = [y^{(h)}] \cap M$). \end{itemize} \item \label{item:admissible_N_large_wide_transformation} For each $0 \leq h \leq \ell(\mathcal{D})-1$, if $\omega_h$ is a wide transformation, then one of the following two conditions is satisfied: \begin{itemize} \item $\lambda(h+1) = \lambda(h) - 2$, $K^{(h)} = J^{(h)} \cup \{s_{\lambda(h)-2},s_{\lambda(h)-1}\}$ is of type $A_{N+2}$, $t^{(h)} = s_{\lambda(h)-1}$, and the action of $\omega_h$ maps $s_{\lambda(h)+j} \in J^{(h)}$ ($0 \leq j \leq N-1$) to $s_{\lambda(h+1)+j}$ and maps $s_{\lambda(h)-2} \in [y^{(h)}]$ to $s_{\rho(h)+2}$; \item $\lambda(h+1) = \lambda(h) + 2$, $K^{(h)} = J^{(h)} \cup \{s_{\rho(h)+1},s_{\rho(h)+2}\}$ is of type $A_{N+2}$, $t^{(h)} = s_{\rho(h)+1}$, and the action of $\omega_h$ maps $s_{\lambda(h)+j} \in J^{(h)}$ ($0 \leq j \leq N-1$) to $s_{\lambda(h+1)+j}$ and maps $s_{\rho(h)+2} \in [y^{(h)}]$ to $s_{\lambda(h)-2}$. \end{itemize} \end{enumerate} Moreover, we say that such a sequence $s_1,s_2,\dots,s_{\mu}$ is \emph{tight} if $M = \bigcup_{h=0}^{\ell(\mathcal{D})} J^{(h)}$. \end{defn} \begin{figure}
\caption{Admissible sequence when $N \geq 2$; here $N = 7$, black circles in the top and the bottom rows indicate elements of $[y^{(h)}] \cap M$ and $[y^{(h+1)}] \cap M$, respectively, and $\omega_h$ is a wide transformation with $t^{(h)} = s_{\lambda(h)-1}$}
\label{fig:finitepart_secondcase_N_2_definition}
\end{figure} \begin{defn} \label{defn:admissible_for_N_1} Suppose that $N = 1$. Let $\mathcal{D} = \omega_{\ell(\mathcal{D})-1} \cdots \omega_1\omega_0$ be a semi-standard decomposition of an element of $W$. We say that a sequence $s_1,s_2,\dots,s_{\mu}$ of distinct elements of $S$ is \emph{admissible of type $A_1$} with respect to $\mathcal{D}$, if $J^{(0)}$ is of type $A_1$ and the following conditions are satisfied, where we put $M = \{s_1,s_2\dots,s_{\mu}\}$ (see Figure \ref{fig:finitepart_secondcase_N_1_definition}): \begin{enumerate} \item \label{item:admissible_N_1_J_irreducible_component} $\Pi_{J^{(0)}}$ is an irreducible component of $\Pi^{[y^{(0)}]}$. \item \label{item:admissible_N_1_J_in_M} For each $0 \leq h \leq \ell(\mathcal{D})$, we have $J^{(h)} \subseteq M$ and $M \smallsetminus J^{(h)} \subseteq [y^{(h)}]$. \item \label{item:admissible_N_1_y_isolated} For each $0 \leq h \leq \ell(\mathcal{D})$, every element of $[y^{(h)}] \cap M$ forms an irreducible component of $[y^{(h)}]$ of type $A_1$. \item \label{item:admissible_N_1_narrow_transformation} For each $0 \leq h \leq \ell(\mathcal{D})-1$, if $\omega_h$ is a narrow transformation, then one of the following two conditions is satisfied: \begin{itemize} \item $K^{(h)}$ intersects with $[y^{(h)}] \cap M$, and $[y^{(h+1)}] = [y^{(h)}]$; \item $K^{(h)}$ is apart from $[y^{(h)}] \cap M$, hence $[y^{(h+1)}] \cap M = [y^{(h)}] \cap M$. \end{itemize} \item \label{item:admissible_N_1_wide_transformation} For each $0 \leq h \leq \ell(\mathcal{D})-1$, if $\omega_h$ is a wide transformation, then one of the following two conditions is satisfied: \begin{itemize} \item $J^{(h+1)} \neq J^{(h)}$, $K^{(h)}$ is of type $A_3$, $K^{(h)} \smallsetminus \{t^{(h)}\} = J^{(h)} \cup J^{(h+1)}$, $J^{(h+1)} \subseteq [y^{(h)}] \cap M$, and the action of $\omega_h$ exchanges the unique element of $J^{(h)}$ and the unique element of $J^{(h+1)}$; \item $J^{(h+1)} = J^{(h)}$ and $[y^{(h+1)}] = [y^{(h)}]$. \end{itemize} \end{enumerate} Moreover, we say that such a sequence $s_1,s_2,\dots,s_{\mu}$ is \emph{tight} if $M = \bigcup_{h=0}^{\ell(\mathcal{D})} J^{(h)}$. \end{defn} \begin{figure}
\caption{Admissible sequence when $N = 1$; here $\omega_h$ is a wide transformation of the first type in Definition \ref{defn:admissible_for_N_1}(\ref{item:admissible_N_1_wide_transformation}), the circles in each row signify elements of $M$, and the diamond signifies the element $t^{(h)}$}
\label{fig:finitepart_secondcase_N_1_definition}
\end{figure}
Note that, if a sequence $s_1,s_2,\dots,s_{\mu}$ is admissible of type $A_N$ with respect to a semi-standard decomposition $\mathcal{D} = \omega_{\ell(\mathcal{D})-1} \cdots \omega_1\omega_0$, then the subsequence of $s_1,s_2,\dots,s_{\mu}$ consisting of the elements of $\bigcup_{j=0}^{\ell(\mathcal{D})} J^{(j)}$ is admissible of type $A_N$ with respect to $\mathcal{D}$ and is tight (for the case $N \geq 2$, the property of wide transformations in Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_wide_transformation}) implies that $\bigcup_{j=0}^{\ell(\mathcal{D})} J^{(j)} = \{s_i \mid \lambda(k) \leq i \leq \rho(k')\}$ for some $k,k' \in \{0,1,\dots,\ell(\mathcal{D})\}$). Moreover, the sequence $s_1$, $s_2,\dots,s_{\mu}$ is also admissible of type $A_N$ with respect to $\mathcal{D}^{-1}$.
The above definitions are relevant to our purpose in the following manner: \begin{lem} \label{lem:claim_holds_when_admissible} Let $\mathcal{D} = \omega_{\ell(\mathcal{D})-1} \cdots \omega_1\omega_0$ be a semi-standard decomposition of $w$ with respect to $L$. If there exists a sequence which is admissible of type $A_N$ with respect to $\mathcal{D}$, then $w$ fixes $\Pi_L$ pointwise. \end{lem} \begin{proof} First, note that $y^{(\ell(\mathcal{D}))} = x_I = y^{(0)}$ since $w \in Y_I$, therefore $[y^{(\ell(\mathcal{D}))}] \cap M = [y^{(0)}] \cap M$ where $M$ is as defined in Definition \ref{defn:admissible_for_N_large} (when $N \geq 2$) or Definition \ref{defn:admissible_for_N_1} (when $N = 1$). Now it follows from the properties in Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_J_in_M}) when $N \geq 2$, or Definition \ref{defn:admissible_for_N_1}(\ref{item:admissible_N_1_J_in_M}) when $N = 1$, that $J^{(\ell(\mathcal{D}))} = J^{(0)} = L$. Hence $w$ fixes $\Pi_L$ pointwise when $N = 1$. Moreover, when $N \geq 2$, the property in Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_wide_transformation}) implies that $\omega_h \ast s_{\lambda(h)+j} = s_{\lambda(h+1)+j}$ for every $0 \leq h \leq \ell(\mathcal{D})-1$ and $0 \leq j \leq N-1$. Now by this property and the above-mentioned property $J^{(\ell(\mathcal{D}))} = J^{(0)}$, it follows that $w$ fixes the set $\Pi_{J^{(0)}} = \Pi_{L}$ pointwise. Hence the proof is concluded. \end{proof}
As mentioned above, a standard decomposition of $w$ with respect to $L$ exists. Therefore, by virtue of Lemma \ref{lem:claim_holds_when_admissible}, it suffices to show that there exists a sequence which is admissible with respect to this standard decomposition. More generally, we prove the following proposition (note that the above-mentioned standard decomposition of $w$ satisfies the assumption in this proposition): \begin{prop} \label{prop:admissible_sequence_exists} Let $\mathcal{D} = \omega_{\ell(\mathcal{D})-1} \cdots \omega_1\omega_0$ be a semi-standard decomposition of an element. Suppose that $J^{(0)}$ is of type $A_N$ with $1 \leq N < \infty$, and $\Pi_{J^{(0)}}$ is an irreducible component of $\Pi^{[y^{(0)}]}$. Then there exists a sequence which is admissible of type $A_N$ with respect to $\mathcal{D}$. \end{prop} To prove Proposition \ref{prop:admissible_sequence_exists}, we give the following key lemma, which will be proven below: \begin{lem} \label{lem:admissible_sequence_extends} Let $n \geq 0$. Let $\mathcal{D} = \omega_n\omega_{n-1} \cdots \omega_1\omega_0$ be a semi-standard decomposition of an element, and put $\mathcal{D}' := \omega_{n-1} \cdots \omega_1\omega_0$, which is also a semi-standard decomposition of an element satisfying that $y^{(0)}(\mathcal{D}') = y^{(0)}(\mathcal{D})$ and $J^{(0)}(\mathcal{D}') = J^{(0)}(\mathcal{D})$. Suppose that $s_1,\dots,s_{\mu}$ is a sequence which is admissible of type $A_N$ with respect to $\mathcal{D}'$. For simplicity, put $y^{(j)} = y^{(j)}(\mathcal{D})$, $J^{(j)} = J^{(j)}(\mathcal{D})$, $t^{(j)} = t^{(j)}(\mathcal{D})$, and $K^{(j)} = K^{(j)}(\mathcal{D})$ for each index $j$. \begin{enumerate} \item \label{item:lem_admissible_sequence_extends_narrow} If $\omega_n$ is a narrow transformation, then we have either $[y^{(n+1)}] = [y^{(n)}]$, or $K^{(n)}$ is apart from $[y^{(n)}] \cap \bigcup_{j=0}^{n} J^{(j)}$. \item \label{item:lem_admissible_sequence_extends_wide_N_1_fix} If $N = 1$, $\omega_n$ is a wide transformation and $J^{(n+1)} = J^{(n)}$, then we have $[y^{(n+1)}] = [y^{(n)}]$. \item \label{item:lem_admissible_sequence_extends_wide_N_1} If $N = 1$, $\omega_n$ is a wide transformation and $J^{(n+1)} \neq J^{(n)}$, then $K^{(n)}$ is of type $A_3$, $K^{(n)} \smallsetminus (J^{(n)} \cup \{t^{(n)}\}) \subseteq [y^{(n)}]$, and the action of $\omega_n$ exchanges the unique element of $J^{(n)}$ and the unique element of $K^{(n)} \smallsetminus (J^{(n)} \cup \{t^{(n)}\})$ (the latter belonging to $[y^{(n)}] \cap J^{(n+1)}$). \item \label{item:lem_admissible_sequence_extends_wide_N_large} If $N \geq 2$ and $\omega_n$ is a wide transformation, then $K^{(n)}$ is of type $A_{N+2}$, the unique element $s'$ of $K^{(n)} \smallsetminus (J^{(n)} \cup \{t^{(n)}\})$ belongs to $[y^{(n)}]$, and one of the following two conditions is satisfied: \begin{enumerate} \item \label{item:lem_admissible_sequence_extends_wide_N_large_left} $t^{(n)}$ is adjacent to $s'$ and $s_{\lambda(n)}$, and the action of $\omega_n$ maps the elements $s_{\lambda(n)}$, $s_{\lambda(n)+1}$, $s_{\lambda(n)+2},\dots,s_{\rho(n)}$ and $s'$ to $s'$, $t^{(n)}$, $s_{\lambda(n)},\dots,s_{\rho(n)-2}$ and $s_{\rho(n)}$, respectively. Moreover; \begin{enumerate} \item \label{item:lem_admissible_sequence_extends_wide_N_large_left_branch} if $\lambda(n) \geq 3$ and $s_{\lambda(n)-2} \in \bigcup_{j=0}^{n} J^{(j)}$, then we have $s' = s_{\lambda(n)-2}$ and $t^{(n)} = s_{\lambda(n)-1}$; \item \label{item:lem_admissible_sequence_extends_wide_N_large_left_terminal} otherwise, we have $s' \not\in \bigcup_{j=0}^{n} J^{(j)}$. \end{enumerate} \item \label{item:lem_admissible_sequence_extends_wide_N_large_right} $t^{(n)}$ is adjacent to $s'$ and $s_{\rho(n)}$, and the action of $\omega_n$ maps the elements $s_{\rho(n)}$, $s_{\rho(n)-1}$, $s_{\rho(n)-2},\dots,s_{\lambda(n)}$ and $s'$ to $s'$, $t^{(n)}$, $s_{\rho(n)},\dots,s_{\lambda(n)+2}$ and $s_{\lambda(n)}$, respectively. Moreover; \begin{enumerate} \item \label{item:lem_admissible_sequence_extends_wide_N_large_right_branch} if $\rho(n) \leq \mu - 2$ and $s_{\rho(n)+2} \in \bigcup_{j=0}^{n} J^{(j)}$, then we have $s' = s_{\rho(n)+2}$ and $t^{(n)} = s_{\rho(n)+1}$; \item \label{item:lem_admissible_sequence_extends_wide_N_large_right_terminal} otherwise, we have $s' \not\in \bigcup_{j=0}^{n} J^{(j)}$. \end{enumerate} \end{enumerate} \end{enumerate} \end{lem} Then Proposition \ref{prop:admissible_sequence_exists} is deduced by applying Lemma \ref{lem:admissible_sequence_extends} and the next lemma to the semi-standard decompositions $\mathcal{D}_{\nu} := \omega_{\nu-1} \cdots \omega_1\omega_0$ ($0 \leq \nu \leq \ell(\mathcal{D})$) successively (note that, when $\nu = 0$, i.e., $\mathcal{D}_{\nu}$ is an empty expression, the sequence $s_1,\dots,s_N$, where $J^{(0)} = \{s_1,\dots,s_N\}$ is the standard labelling of type $A_N$, is admissible of type $A_N$ with respect to $\mathcal{D}_{\nu}$): \begin{lem} \label{lem:admissible_sequence_existence_from_lemma} In the situation of Lemma \ref{lem:admissible_sequence_extends}, we define a sequence $\sigma$ of elements of $S$ in the following manner: For Cases \ref{item:lem_admissible_sequence_extends_narrow}, \ref{item:lem_admissible_sequence_extends_wide_N_1_fix}, \ref{item:lem_admissible_sequence_extends_wide_N_large_left_branch} and \ref{item:lem_admissible_sequence_extends_wide_N_large_right_branch}, let $\sigma$ be the sequence $s_1,\dots,s_{\mu}$; for Case \ref{item:lem_admissible_sequence_extends_wide_N_1}, let $s'$ be the unique element of $K^{(n)} \smallsetminus (J^{(n)} \cup \{t^{(n)}\}) = J^{(n+1)}$, and let $\sigma$ be the sequence $s_1,\dots,s_{\mu},s'$ when $s' \not\in \{s_1,\dots,s_{\mu}\}$ and the sequence $s_1,\dots,s_{\mu}$ when $s' \in \{s_1,\dots,s_{\mu}\}$; for Case \ref{item:lem_admissible_sequence_extends_wide_N_large_left_terminal}, let $\sigma$ be the sequence $s'$, $t^{(n)}$, $s_{\lambda(n)}$, $s_{\lambda(n)+1},\dots,s_{\rho'}$, where $\rho'$ denotes the largest index $1 \leq \rho' \leq \mu$ with $s_{\rho'} \in \bigcup_{j=0}^{n} J^{(j)}$; for the case \ref{item:lem_admissible_sequence_extends_wide_N_large_right_terminal}, let $\sigma$ be the sequence $s'$, $t^{(n)}$, $s_{\rho(n)}$, $s_{\rho(n)-1},\dots,s_{\lambda'}$, where $\lambda'$ denotes the smallest index $1 \leq \lambda' \leq \mu$ with $s_{\lambda'} \in \bigcup_{j=0}^{n} J^{(j)}$. Then $\sigma$ is admissible of type $A_N$ with respect to $\mathcal{D} = \omega_n \cdots \omega_1\omega_0$. \end{lem}
Now our remaining task is to prove Lemma \ref{lem:admissible_sequence_extends} and Lemma \ref{lem:admissible_sequence_existence_from_lemma}. For the purpose, we present an auxiliary result: \begin{lem} \label{lem:admissible_sequence_non_adjacent_pairs} Let $s_1,\dots,s_{\mu}$ be a sequence which is admissible of type $A_N$, where $N \geq 2$, with respect to a semi-standard decomposition $\mathcal{D}$ of an element of $W$. Suppose that the sequence $s_1,\dots,s_{\mu}$ is tight. If $1 \leq j_1 < j_2 \leq \mu$, $j_2 - j_1 \geq 2$, and either $j_1 \equiv 1 \pmod{2}$ or $j_2 \equiv \mu \pmod{2}$, then $s_{j_1}$ is not adjacent to $s_{j_2}$. \end{lem} \begin{proof} By symmetry, we may assume without loss of generality that $j_1 \equiv 1 \pmod{2}$. Put $\mathcal{D} = \omega_{n-1} \cdots \omega_1\omega_0$. Since the sequence $s_1,\dots,s_\mu$ is tight, there exists an index $0 \leq h \leq n$ with $s_{j_2} \in J^{(h)}$. Now the properties \ref{item:admissible_N_large_M_line} and \ref{item:admissible_N_large_J_in_M} in Definition \ref{defn:admissible_for_N_large} imply that $J^{(h)} = \{s_{\lambda(h)},s_{\lambda(h)+1},\dots,s_{\rho(h)}\}$ is the standard labelling of type $A_N$, therefore the claim holds if $s_{j_1} \in J^{(h)}$ (note that $j_2 - j_1 \geq 2$). On the other hand, if $s_{j_1} \not\in J^{(h)}$, then the property \ref{item:admissible_N_large_J_in_M} in Definition \ref{defn:admissible_for_N_large} and the fact $j_1 < j_2$ imply that $j_1 < \lambda(h)$, therefore $s_{j_1} \in [y^{(h)}]$ since $j_1 \equiv 1 \pmod{2}$. Hence the claim follows from the fact that $J^{(h)}$ is apart from $[y^{(h)}]$ (see the property \ref{item:admissible_N_large_J_irreducible_component} in Definition \ref{defn:admissible_for_N_large}). \end{proof}
From now, we prove the pair of Lemma \ref{lem:admissible_sequence_extends} and Lemma \ref{lem:admissible_sequence_existence_from_lemma} by induction on $n \geq 0$. First, we give a proof of Lemma \ref{lem:admissible_sequence_existence_from_lemma} for $n = n_0$ by assuming Lemma \ref{lem:admissible_sequence_extends} for $0 \leq n \leq n_0$. Secondly, we will give a proof of Lemma \ref{lem:admissible_sequence_extends} for $n = n_0$ by assuming Lemma \ref{lem:admissible_sequence_extends} for $0 \leq n < n_0$ and Lemma \ref{lem:admissible_sequence_existence_from_lemma} for $0 \leq n < n_0$. \begin{proof} [Proof of Lemma \ref{lem:admissible_sequence_existence_from_lemma} (for $n = n_0$) from Lemma \ref{lem:admissible_sequence_extends} (for $n \leq n_0$).] When $n_0 = 0$, the claim is obvious from the property of $\omega_{n_0}$ specified in Lemma \ref{lem:admissible_sequence_extends}. From now, we suppose that $n_0 > 0$. We may assume without loss of generality that the sequence $s_1,\dots,s_{\mu}$ (denoted here by $\sigma'$) which is admissible with respect to $\mathcal{D}'$ is tight, therefore we have $M' := \{s_1,\dots,s_{\mu}\} = \bigcup_{j=0}^{n_0} J^{(j)}$. We divide the proof according to the possibility of $\omega_{n_0}$ listed in Lemma \ref{lem:admissible_sequence_extends}. By symmetry, we may omit the argument for Case \ref{item:lem_admissible_sequence_extends_wide_N_large_right} without loss of generality.
In Case \ref{item:lem_admissible_sequence_extends_narrow}, since $M' = \bigcup_{j=0}^{n_0} J^{(j)}$ as above, $\omega_{n_0}$ satisfies the condition for $\sigma'$ in Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_narrow_transformation}) (when $N \geq 2$) or Definition \ref{defn:admissible_for_N_1}(\ref{item:admissible_N_1_narrow_transformation}) (when $N = 1$), hence $\sigma = \sigma'$ is admissible of type $A_N$ with respect to $\mathcal{D}$. Similarly, in Case \ref{item:lem_admissible_sequence_extends_wide_N_1_fix}, Case \ref{item:lem_admissible_sequence_extends_wide_N_large_left_branch}, and Case \ref{item:lem_admissible_sequence_extends_wide_N_1} with $s' \in M'$, respectively, the wide transformation $\omega_{n_0}$ satisfies the condition for $\sigma'$ in Definition \ref{defn:admissible_for_N_1}(\ref{item:admissible_N_1_wide_transformation}), Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_wide_transformation}), and Definition \ref{defn:admissible_for_N_1}(\ref{item:admissible_N_1_wide_transformation}), respectively. Hence $\sigma = \sigma'$ is admissible of type $A_N$ with respect to $\mathcal{D}$ in these three cases.
From now, we consider the remaining two cases: Case \ref{item:lem_admissible_sequence_extends_wide_N_1} with $s' \not\in M'$, and Case \ref{item:lem_admissible_sequence_extends_wide_N_large_left_terminal}. Note that, in Case \ref{item:lem_admissible_sequence_extends_wide_N_large_left_terminal}, the tightness of $\sigma'$ implies that $\lambda(n_0) = 1$ and $\rho' = \mu$, therefore $\sigma$ is the sequence $s'$, $t^{(n_0)}$, $s_1,\dots,s_{\mu}$. Moreover, in this case the unique element $s'$ of $K^{(n_0)} \cap [y^{(n_0)}]$ does not belong to $\bigcup_{j=0}^{n_0} J^{(j)} = M'$, therefore $t^{(n_0)}$ cannot be adjacent to $[y^{(n_0)}] \cap M'$; hence $t^{(n_0)} \not\in M'$ by the property of $\sigma'$ in Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_J_in_M}). Note also that, in both of the two cases, we have $s' \in J^{(n_0+1)}$ and $\{s'\}$ is an irreducible component of $[y^{(n_0)}]$.
We prove by induction on $0 \leq \nu \leq n_0$ that the sequence $\sigma$ is admissible of type $A_N$ with respect to $\mathcal{D}_{\nu}$ and $s' \in [y^{(\nu+1)}(\mathcal{D}_{\nu})]$, where \begin{displaymath} \mathcal{D}_{\nu} = \omega'_{\nu}\omega'_{\nu-1} \cdots \omega'_1\omega'_0 := (\omega_{n_0-\nu})^{-1}(\omega_{n_0-\nu+1})^{-1} \cdots (\omega_{n_0-1})^{-1}(\omega_{n_0})^{-1} \end{displaymath} is a semi-standard decomposition of an element with respect to $J^{(n_0+1)}$. Note that $y^{(j)}(\mathcal{D}_{\nu}) = y^{(n_0-j+1)}$, $J^{(j)}(\mathcal{D}_{\nu}) = J^{(n_0-j+1)}$, $t^{(j)}(\mathcal{D}_{\nu}) = t^{(n_0-j+1)}$ and $K^{(j)}(\mathcal{D}_{\nu}) = K^{(n_0-j+1)}$ for each index $j$. When $\nu = 0$, this claim follows immediately from the property of $\omega_{n_0}$ specified in Lemma \ref{lem:admissible_sequence_extends}, properties of $\sigma'$ and the definition of $\sigma$. Suppose that $\nu > 0$. Note that $s' \in [y^{(\nu)}(\mathcal{D}_{\nu-1})]$ (which is equal to $[y^{(\nu)}(\mathcal{D}_{\nu})] = [y^{(n_0-\nu+1)}]$) by the induction hypothesis. First, we consider the case that $\omega'_{\nu}$ (or equivalently, $\omega_{n_0-\nu}$) is a wide transformation. In this case, the possibility of $\omega_{n_0-\nu}$ is as specified in the condition of $\sigma'$ in Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_wide_transformation}) (when $N \geq 2$) or Definition \ref{defn:admissible_for_N_1}(\ref{item:admissible_N_1_wide_transformation}) (when $N = 1$), where $h = n_0-\nu$; in particular, we have $K^{(n_0-\nu)} \smallsetminus \{t^{(n_0-\nu)}\} \subseteq M'$, therefore $[y^{(n_0-\nu+1)}] \smallsetminus M' = [y^{(n_0-\nu)}] \smallsetminus M'$. Hence the element $s'$ of $[y^{(n_0-\nu+1)}] \smallsetminus M'$ belongs to $[y^{(n_0-\nu)}] = [y^{(\nu+1)}(\mathcal{D}_{\nu})]$, and the property of $\omega_{n_0-\nu} = (\omega'_{n_0})^{-1}$ implies that $\sigma$ is admissible of type $A_N$ with respect to $\mathcal{D}_{\nu}$ as well as $\mathcal{D}_{\nu-1}$. Secondly, we consider the case that $\omega'_{\nu}$ (or equivalently, $\omega_{n_0-\nu}$) is a narrow transformation. By applying Lemma \ref{lem:admissible_sequence_extends} (for $n = \nu$) to the pair $\mathcal{D}_{\nu}$, $\mathcal{D}_{\nu-1}$ and the sequence $\sigma$, it follows that either $[y^{(\nu+1)}(\mathcal{D}_{\nu})] = [y^{(\nu)}(\mathcal{D}_{\nu})]$, or the support of $\omega'_{\nu}$ is apart from $[y^{(\nu)}(\mathcal{D}_{\nu})] \cap \bigcup_{j=0}^{\nu} J^{(j)}(\mathcal{D}_{\nu})$. Now in the former case, we have $s' \in [y^{(\nu)}(\mathcal{D}_{\nu})] = [y^{(\nu+1)}(\mathcal{D}_{\nu})]$. On the other hand, in the latter case, we have $s' \in [y^{(\nu)}(\mathcal{D}_{\nu})] \cap \bigcup_{j=0}^{\nu} J^{(j)}(\mathcal{D}_{\nu})$ since $s' \in [y^{(\nu)}(\mathcal{D}_{\nu})]$ as above and $s' \in J^{(0)}(\mathcal{D}_{\nu}) = J^{(n_0+1)}$ by the choice of $s'$, therefore $s'$ is apart from the support of $\omega'_{\nu}$. Hence, it follows in any case that $s' \in [y^{(\nu+1)}(\mathcal{D}_{\nu})]$; and the property of $\omega_{n_0-\nu} = (\omega'_{\nu})^{-1}$ specified by the condition of $\sigma'$ in Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_narrow_transformation}) (when $N \geq 2$) or Definition \ref{defn:admissible_for_N_1}(\ref{item:admissible_N_1_narrow_transformation}) (when $N = 1$), where $h = n_0-\nu$, implies that $\sigma$ is admissible of type $A_N$ with respect to $\mathcal{D}_{\nu}$ as well as $\mathcal{D}_{\nu-1}$. Hence the claim of this paragraph follows.
By using the result of the previous paragraph with $\nu = n_0$, the sequence $\sigma$ is admissible of type $A_N$ with respect to $\mathcal{D}_{n_0} = \mathcal{D}^{-1}$, hence with respect to $\mathcal{D}$ as well. This completes the proof. \end{proof}
By virtue of the above result, our remaining task is finally to prove Lemma \ref{lem:admissible_sequence_extends} for $n = n_0$ by assuming Lemma \ref{lem:admissible_sequence_extends} for $0 \leq n < n_0$ and Lemma \ref{lem:admissible_sequence_existence_from_lemma} for $0 \leq n < n_0$ (in particular, with no assumptions when $n_0 = 0$). Put $M' := \{s_1,\dots,s_{\mu}\}$. In the proof, we may assume without loss of generality that the sequence $s_1,\dots,s_{\mu}$ (denoted here by $\sigma'$) which is admissible with respect to $\mathcal{D}'$ is tight (hence we have $J^{(0)} = M'$ when $n_0 = 0$). Now by Lemma \ref{lem:proof_special_second_case_transformations_wide}, the claim of Lemma \ref{lem:admissible_sequence_extends} holds for the case that $N = 1$ and $\omega_{n_0}$ is a wide transformation. From now, we consider the other case that either $N \geq 2$ or $\omega_{n_0}$ is a narrow transformation. Assume contrary that the claim of Lemma \ref{lem:admissible_sequence_extends} does not hold. Then, by Lemma \ref{lem:proof_special_second_case_transformations_wide}, Lemma \ref{lem:proof_special_second_case_transformations_narrow} and the properties of the tight sequence $\sigma'$ in Definition \ref{defn:admissible_for_N_large} (when $N \geq 2$) or Definition \ref{defn:admissible_for_N_1} (when $N = 1$), it follows that the possibilities for the $\omega_{n_0}$ is as follows (up to symmetry): \begin{description} \item[Case (I):] $\omega_{n_0}$ is a narrow transformation, $K^{(n_0)}$ is of type $A_2$ or type $I_2(m)$ with $m$ odd, and we have $s_{\eta} \in K^{(n_0)} \cap [y^{(n_0)}]$ for some index $1 \leq \eta \leq \mu$; hence $t^{(n_0)} \not\in M'$, $K^{(n_0)} = \{s_{\eta},t^{(n_0)}\}$ and the action of $\omega_{n_0}$ exchanges the two elements of $K^{(n_0)}$. \item[Case (II):] $N \geq 2$, $\omega_{n_0}$ is a wide transformation, $K^{(n_0)}$ is of type $A_{N+2}$, and $t^{(n_0)}$ is adjacent to $s_{\lambda(n_0)}$ and the unique element $s'$ of $[y^{(n_0)}] \cap K^{(n_0)}$; hence the action of $\omega_{n_0}$ maps the elements $s_{\lambda(n_0)}$, $s_{\lambda(n_0)+1}$, $s_{\lambda(n_0)+2},\cdots,s_{\rho(n_0)}$ and $s'$ to $s'$, $t^{(n_0)}$, $s_{\lambda(n_0)},\dots,s_{\rho(n_0)-2}$ and $s_{\rho(n_0)}$, respectively. Moreover, $t^{(n_0)} \not\in M'$, and \begin{description} \item[Case (II-1):] $s' = s_{j_0}$ for an index $\rho(n_0)+2 \leq j_0 \leq \mu$ with $j_0 \equiv \mu \pmod{2}$; \item[Case (II-2):] $\lambda(n_0) \geq 3$ and $s' \not\in \{s_{\lambda(n_0)-2},s_{\lambda(n_0)-1},\dots,s_{\mu}\}$; \item[Case (II-3):] $\lambda(n_0) \geq 3$ and $s' = s_{\lambda(n_0)-2}$. \end{description} \end{description} In particular, by the tightness of $\sigma'$, the conditions in the above four cases cannot be satisfied when $n_0 = 0$. Hence the claim holds when $n_0 = 0$. From now, we suppose that $n_0 > 0$.
For each of the four cases, we determine an element $\overline{s} \in [y^{(n_0)}] \cap M'$ and an element $\overline{t} \in S \smallsetminus [y^{(n_0)}]$ in the following manner: $\overline{s} = s_{\eta}$ and $\overline{t} = t^{(n_0)}$ in Case (I); $\overline{s} = s_{j_0}$ and $\overline{t} = t^{(n_0)}$ in Case (II-1); $\overline{s} = s_{\lambda(n_0)-2}$ and $\overline{t} = s_{\lambda(n_0)-1}$ in Case (II-2); and $\overline{s} = s_{\lambda(n_0)-2}$ and $\overline{t} = t^{(n_0)}$ in Case (II-3). Note that $\overline{s}$ and $\overline{t}$ are adjacent by the definition. Since $\sigma'$ is tight, there exists an index $0 \leq h_0 \leq n_0-1$ with $\overline{s} \in J^{(h_0)}$; let $h_0$ be the largest index with this property. By the definition of $h_0$, $\omega_{h_0}$ is a wide transformation and $J^{(h_0+1)} \neq J^{(h_0)}$. Let $\overline{r}$ denote the element of $J^{(h_0+1)}$ with $\omega_{h_0} \ast \overline{s} = \overline{r}$. Then we have $\overline{r} \in [y^{(h_0)}]$ by the property of $\omega_{h_0}$ and the choice of $\overline{s}$.
Let $\overline{\mathcal{D}} := \omega'_{n'-1} \cdots \omega'_1\omega'_0$ denote the simplification of \begin{displaymath} (\omega_{n_0-1} \cdots \omega_{h_0+2}\omega_{h_0+1})^{-1} = (\omega_{h_0+1})^{-1}(\omega_{h_0+2})^{-1} \cdots (\omega_{n_0-1})^{-1} \end{displaymath} (see Section \ref{sec:finitepart_decomposition_Y} for the terminology), and let $\overline{u}$ be the element of $W$ expressed by the product $\overline{\mathcal{D}}$. Here we present the following lemma: \begin{lem} \label{lem:proof_lem_admissible_sequence_extends_simplification} In this setting, the support of each transformation in $\overline{\mathcal{D}}$ does not contain $\overline{t}$ and is apart from $\overline{s}$. \end{lem} \begin{proof} We prove by induction on $0 \leq \nu' \leq n'-1$ that the support $K'$ of $\omega'_{\nu'}$ does not contain $\overline{t}$ and is apart from $\overline{s}$. Let $(\omega_{\nu})^{-1}$ be the term in $(\omega_{h_0+1})^{-1}(\omega_{h_0+2})^{-1} \cdots (\omega_{n_0-1})^{-1}$ corresponding to the term $\omega'_{\nu'}$ in the simplification $\overline{\mathcal{D}}$. First, by the definition of simplification and the property of narrow transformations specified in Definition \ref{defn:admissible_for_N_large} (when $N \geq 2$) or Definition \ref{defn:admissible_for_N_1} (when $N = 1$), $K'$ is apart from $[y^{(\nu')}(\overline{\mathcal{D}})] \cap M' = [y^{(\nu+1)}] \cap M'$ (see Lemma \ref{lem:another_decomposition_Y_reduce_redundancy} for the equality) if $\omega'_{\nu'}$ (or equivalently, $\omega_{\nu}$) is a narrow transformation. Now we have $\overline{s} \in [y^{(n_0)}] = [y^{(0)}(\overline{\mathcal{D}})]$ and $\overline{s} \in M'$ by the definition, therefore the induction hypothesis implies that $\overline{s} \in [y^{(\nu')}(\overline{\mathcal{D}})] \cap M'$. Hence $K'$ is apart from $\overline{s}$ if $\omega'_{\nu'}$ is a narrow transformation. This also implies that $\overline{t} \not\in K'$ if $\omega'_{\nu'}$ is a narrow transformation, since $\overline{t}$ is adjacent to $\overline{s}$.
From now, we consider the other case that $\omega'_{\nu'}$ (or equivalently, $\omega_{\nu}$) is a wide transformation. Recall that $\overline{s} \in [y^{(\nu')}(\overline{\mathcal{D}})]$ as mentioned above. Then, by the property of wide transformation $\omega_{\nu}$ specified in Definition \ref{defn:admissible_for_N_large} (when $N \geq 2$) or Definition \ref{defn:admissible_for_N_1} (when $N = 1$) and the definition of simplification, it follows that $\overline{s} \in J^{(\nu'+1)}(\overline{\mathcal{D}})$ provided $K'$ is not apart from $\overline{s}$. On the other hand, by the definition of $h_0$, we have $\overline{s} \not\in J^{(j)}$ for any $h_0+1 \leq j \leq n_0$. This implies that $K'$ should be apart from $\overline{s}$; therefore we have $\overline{t} \not\in K'$, since $\overline{t}$ is adjacent to $\overline{s}$. Hence the proof of Lemma \ref{lem:proof_lem_admissible_sequence_extends_simplification} is concluded. \end{proof} Now, in all the cases except Case (II-2), the following property holds: \begin{lem} \label{lem:proof_lem_admissible_sequence_extends_root_other_cases} In Cases (I), (II-1) and (II-3), there exists a root $\beta \in \Pi^{[y^{(h_0)}]}$ in which the coefficient of $\alpha_{\overline{s}}$ is zero and the coefficient of $\alpha_{\overline{t}} = \alpha_{t^{(n_0)}}$ is non-zero. \end{lem} \begin{proof} First, Lemma \ref{lem:another_decomposition_Y_reduce_redundancy} implies that $\overline{u} \cdot \Pi_{J^{(n_0)}} = \Pi_{J^{(h_0+1)}}$ and $[y'] = [y^{(h_0+1)}]$ where $y' := y^{(n')}(\overline{\mathcal{D}})$. Put $r' := \overline{u}^{-1} \ast \overline{r} \in J^{(n_0)}$. Then by Lemma \ref{lem:proof_lem_admissible_sequence_extends_simplification} and Lemma \ref{lem:another_decomposition_Y_shift_x}, we have $\overline{u} \in Y_{z',z}$, where $z$ and $z'$ are elements of $S^{(\Lambda)}$ obtained from $y^{(0)}(\overline{\mathcal{D}}) = y^{(n_0)}$ and $y'$ by replacing the element $\overline{s}$ with $r'$ and $\overline{r}$, respectively. Now by the property of the wide transformation $\omega_{h_0}$, it follows that $y^{(h_0)}$ is obtained from $y^{(h_0+1)}$ by replacing $\overline{s}$ with $\overline{r}$; hence we have $[z'] = [y^{(h_0)}]$.
We show that there exists a root $\beta' \in \Pi^{[z]}$ in which the coefficient of $\alpha_{\overline{s}}$ is zero and the coefficient of $\alpha_{t^{(n_0)}}$ is non-zero. In Case (I), $t^{(n_0)}$ is apart from both $[y^{(n_0)}] \smallsetminus \{\overline{s}\}$ and $J^{(n_0)}$, while we have $[z] \subseteq ([y^{(n_0)}] \smallsetminus \{\overline{s}\}) \cup J^{(n_0)}$ by the definition; hence $\beta' := \alpha_{t^{(n_0)}}$ satisfies the required condition. In Case (II-1), we have $\overline{r} = s_{\lambda(h_0+1)}$ by the property of $\omega_{h_0}$, therefore $r' = s_{\lambda(n_0)}$ by the property of wide transformations in $\overline{\mathcal{D}}$ (see Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_wide_transformation})). Put $\beta' := \alpha_{t^{(n_0)}} + \alpha_{s_{\lambda(n_0)}} + \alpha_{s_{\lambda(n_0)+1}} \in \Pi^{K^{(n_0)},\{r'\}}$ (note that $N \geq 2$ and $K^{(n_0)}$ is of type $A_{N+2}$). Now $K^{(n_0)}$ is apart from $[y^{(n_0)}] \smallsetminus \{\overline{s}\} = [z] \smallsetminus \{r'\}$, therefore we have $\beta' \in \Pi^{[z]}$ and $\beta'$ satisfies the required condition. Moreover, in Case (II-3), we have $\overline{r} = s_{\rho(h_0+1)}$ by the property of $\omega_{h_0}$, therefore $r' = s_{\rho(n_0)}$ by the property of wide transformations in $\overline{\mathcal{D}}$ (see Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_wide_transformation})). Now, since $N \geq 2$ and $K^{(n_0)}$ is of type $A_{N+2}$, $t^{(n_0)}$ is not adjacent to $r'$, while $K^{(n_0)}$ is apart from $[y^{(n_0)}] \smallsetminus \{\overline{s}\} = [z] \smallsetminus \{r'\}$. Hence $\beta' := \alpha_{t^{(n_0)}}$ satisfies the required condition.
By Lemma \ref{lem:proof_lem_admissible_sequence_extends_simplification}, the action of $\overline{u}$ does not change the coefficients of $\alpha_{\overline{s}}$ and $\alpha_{\overline{t}}$. Hence by the result of the previous paragraph, the root $\beta := \overline{u} \cdot \beta' \in \Pi^{[z']} = \Pi^{[y^{(h_0)}]}$ satisfies the required condition, concluding the proof of Lemma \ref{lem:proof_lem_admissible_sequence_extends_root_other_cases}. \end{proof} Since $\overline{t} \not\in J^{(h_0)}$ and $\overline{t}$ is adjacent to $\overline{s}$, the root $\beta \in \Pi^{[y^{(h_0)}]}$ given by Lemma \ref{lem:proof_lem_admissible_sequence_extends_root_other_cases} does not belong to $\Pi_{J^{(h_0)}}$ and is not orthogonal to $\alpha_{\overline{s}}$. However, since $\overline{s} \in J^{(h_0)}$, this contradicts the fact that $\Pi_{J^{(h_0)}}$ is an irreducible component of $\Pi^{[y^{(h_0)}]}$ (see Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_J_irreducible_component}) when $N \geq 2$, or Definition \ref{defn:admissible_for_N_1}(\ref{item:admissible_N_1_J_irreducible_component}) when $N = 1$). Hence we have derived a contradiction in the three cases in Lemma \ref{lem:proof_lem_admissible_sequence_extends_root_other_cases}.
From now, we consider the remaining case, i.e., Case (II-2). In this case, the following property holds: \begin{lem} \label{lem:proof_lem_admissible_sequence_extends_simplification_2} In this setting, the support of each transformation in $\overline{\mathcal{D}}$ does not contain $t^{(n_0)}$ and is apart from $s'$. \end{lem} \begin{proof} For each $0 \leq i \leq n_0 - h_0 - 1$, let $\mathcal{D}_i$ denote the semi-standard decomposition of an element defined by \begin{displaymath} \mathcal{D}_i = \omega''_i \cdots \omega''_1\omega''_0 := (\omega_{n_0-i})^{-1} \cdots (\omega_{n_0-1})^{-1}(\omega_{n_0})^{-1} \enspace. \end{displaymath} For each $0 \leq i \leq n_0 - h_0 - 1$, let $\sigma_i$ denote the sequence $s'$, $t^{(n_0)}$, $s_{\lambda(n_0)}$, $s_{\lambda(n_0)+1},\dots,s_{\overline{\rho}(i)}$, where $\overline{\rho}(i)$ denotes the largest index $s_{\lambda(n_0)} \leq \overline{\rho}(i) \leq \mu$ with $s_{\overline{\rho}(i)} \in \bigcup_{j=0}^{i+1} J^{(j)}(\mathcal{D}_i)$ ($= \bigcup_{j=n_0-i}^{n_0+1} J^{(j)}$). We prove the following properties by induction on $1 \leq i \leq n_0 - h_0 - 1$: The sequence $\sigma_i$ is admissible with respect to $\mathcal{D}_i$; we have $s' \in [y^{(i+1)}(\mathcal{D}_i)]$; and we have either $[y^{(i+1)}(\mathcal{D}_i)] = [y^{(i)}(\mathcal{D}_i)]$ and $J^{(i+1)}(\mathcal{D}_i) = J^{(i)}(\mathcal{D}_i)$, or the support $K'' = K^{(i)}(\mathcal{D}_i)$ of $\omega''_i$ is apart from $s'$. Note that, by the properties of $\omega_{n_0}$ and $\sigma'$, we have $s' \in [y^{(n_0)}] = [y^{(1)}(\mathcal{D}_0)]$, and the sequence $\sigma_0$ (which is $s'$, $t^{(n_0)}$, $s_{\lambda(n_0)},\dots,s_{\rho(n_0)}$) is admissible with respect to $\mathcal{D}_0$.
By the induction hypothesis and Lemma \ref{lem:admissible_sequence_extends} for $n = i$ applied to the sequence $\sigma_{i-1}$ and the pair $\mathcal{D}_i$ and $\mathcal{D}_{i-1}$ (note that $i \leq n_0 - h_0 - 1 \leq n_0 - 1$), it follows that the possibilities of $\omega''_i = (\omega_{n_0-i})^{-1}$ are as listed in Lemma \ref{lem:admissible_sequence_extends}. Now if $\omega''_i$ is a narrow transformation, then as in Case \ref{item:lem_admissible_sequence_extends_narrow} of Lemma \ref{lem:admissible_sequence_extends}, we have either $[y^{(i+1)}(\mathcal{D}_i)] = [y^{(i)}(\mathcal{D}_i)]$, or $K''$ is apart from $s'$ (note that $s' \in [y^{(i)}(\mathcal{D}_i)]$ by the induction hypothesis, while $s' \in J^{(0)}(\mathcal{D}_i) = J^{(n_0+1)}$). On the other hand, suppose that $\omega''_i = (\omega_{n_0-i})^{-1}$ is a wide transformation. Then, by the property of $\sigma'$, the support $K''$ of the wide transformation $\omega_{n_0-i}$ is contained in $M'$, therefore $s' \not\in K''$. This implies that $K''$ is apart from $s'$, since we have $s' \in [y^{(i)}(\mathcal{D}_i)]$ by the induction hypothesis. Moreover, in any case of $\omega''_i$, we have $s' \in [y^{(i+1)}(\mathcal{D}_i)]$ by the above-mentioned fact $s' \in [y^{(i)}(\mathcal{D}_i)]$ and the above argument. On the other hand, the sequence $\sigma$ in Lemma \ref{lem:admissible_sequence_existence_from_lemma} corresponding to the current case is equal to $\sigma_i$, therefore $\sigma_i$ is admissible with respect to $\mathcal{D}_i$ by Lemma \ref{lem:admissible_sequence_existence_from_lemma} for $n = i$ (note again that $i \leq n_0 - 1$). Hence the claim of the previous paragraph holds.
By the above result, the simplification $\overline{D} = \omega'_{n'-1} \cdots \omega'_0$ of $\omega''_{n_0-h_0-1} \cdots \omega''_2\omega''_1$ satisfies the following conditions: For each $0 \leq \nu' \leq n'-1$, we have $s' \in [y^{(\nu')}(\overline{D})]$, and the support of $\omega'_{\nu'}$ is apart from $s'$. Since $t^{(n_0)}$ is adjacent to $s'$, this implies that the support of each $\omega'_{\nu'}$ does not contain $t^{(n_0)}$. Hence the proof of Lemma \ref{lem:proof_lem_admissible_sequence_extends_simplification_2} is concluded. \end{proof}
By Lemma \ref{lem:proof_lem_admissible_sequence_extends_simplification_2}, we have $s' \in [y^{(n')}(\overline{\mathcal{D}})] = [y^{(h_0+1)}]$, therefore the set $K^{(h_0)}$ of type $A_{N+2}$ consisting of $s_{\lambda(n_0)-2}$, $s_{\lambda(n_0)-1},\dots,s_{\rho(n_0)}$ is apart from $s'$. On the other hand, since $s_{\lambda(n_0)-2} \in [y^{(n_0)}]$, the set $K^{(n_0)}$ of type $A_{N+2}$ is apart from $s'$. From now, by using these properties, we construct a root $\beta' \in \Pi^{[y^{(n_0)}]} \smallsetminus \Pi_{J^{(n_0)}}$ which is not orthogonal to $\alpha_{s_{\lambda(n_0)+1}} \in \Pi_{J^{(n_0)}}$ (note that $N \geq 2$), in the following five steps.
\textbf{Step 1.} Note that the set $K^{(n_0)}$ is apart from $[y^{(n_0)}] \smallsetminus K^{(n_0)}$. Put $z^{(0)} := y^{(n_0)}$. Then we have $u_1 := w_{z^{(0)}}^{t^{(n_0)}} = s' t^{(n_0)} \in Y_{z^{(1)},z^{(0)}}$, where $z^{(1)} \in S^{(\Lambda)}$ is obtained from $z^{(0)}$ by replacing $s'$ with $t^{(n_0)}$. Similarly, we have $u_2 := w_{z^{(1)}}^{s_{\lambda(n_0)}} = t^{(n_0)} s_{\lambda(n_0)} \in Y_{z^{(2)},z^{(1)}}$, where $z^{(2)} \in S^{(\Lambda)}$ is obtained from $z^{(1)}$ by replacing $t^{(n_0)}$ with $s_{\lambda(n_0)}$. Now, since $\beta_0 := \alpha_{s_{\lambda(n_0)}}$ and $\beta'_0 := \alpha_{s_{\lambda(n_0)+1}}$ are non-orthogonal elements of $\Pi_{J^{(n_0)}} \subseteq \Pi^{[z^{(0)}]}$, the roots $\beta_2 := u_2u_1 \cdot \beta_0 = \alpha_{s'}$ and $\beta'_2 := u_2u_1 \cdot \beta'_0 = \alpha_{t^{(n_0)}} + \alpha_{s_{\lambda(n_0)}} + \alpha_{s_{\lambda(n_0)+1}}$ are non-orthogonal elements of $\Pi^{[z^{(2)}]}$.
\textbf{Step 2.} By the construction, $z^{(2)}$ is obtained from $y^{(n_0)}$ by replacing $s'$ with $s_{\lambda(n_0)}$. On the other hand, we have $J^{(n_0)} = J^{(h_0+1)}$ and $\overline{u} \ast s_{\lambda(n_0)} = s_{\lambda(n_0)}$ by the property of wide transformations in $\overline{\mathcal{D}}$. Now by Lemma \ref{lem:another_decomposition_Y_shift_x}, we have $u_3 := \overline{u} \in Y_{z^{(3)},z^{(2)}}$, where $z^{(3)} \in S^{(\Lambda)}$ is obtained from $y^{(n')}(\overline{\mathcal{D}})$ by replacing $s'$ with $s_{\lambda(n_0)}$. Note that $[z^{(3)}] = ([y^{(n')}(\overline{\mathcal{D}})] \smallsetminus \{s'\}) \cup \{s_{\lambda(n_0)}\} = ([y^{(h_0+1)}] \smallsetminus \{s'\}) \cup \{s_{\lambda(n_0)}\}$. Put $\beta_3 := u_3 \cdot \beta_2$ and $\beta'_3 := u_3 \cdot \beta'_2$. Then we have $\beta_3,\beta'_3 \in \Pi^{[z^{(3)}]}$ and $\langle \beta_3,\beta'_3 \rangle \neq 0$. Moreover, by Lemma \ref{lem:proof_lem_admissible_sequence_extends_simplification} and Lemma \ref{lem:proof_lem_admissible_sequence_extends_simplification_2}, $u_3$ fixes $\alpha_{s'}$, hence $\beta_3 = \alpha_{s'}$; and the action of $u_3$ does not change the coefficients of $\alpha_{s'}$, $\alpha_{t^{(n_0)}}$, $\alpha_{\overline{s}}$ and $\alpha_{\overline{t}}$, hence the coefficients of these four simple roots in $\beta'_3$ are $0$, $1$, $0$ and $0$, respectively. This also implies that the coefficient of $\alpha_{s_{\lambda(n_0)}}$ in $\beta'_3$ is non-zero, since $t^{(n_0)}$ is adjacent to $s_{\lambda(n_0)} \in [z^{(3)}]$.
\textbf{Step 3.} Note that the set $K^{(h_0)}$ is apart from $[y^{(h_0+1)}] \smallsetminus K^{(h_0)}$, hence from $[z^{(3)}] \smallsetminus K^{(h_0)}$. Then we have $u_4 := w_{z^{(3)}}^{\overline{t}} = \overline{t} s_{\lambda(n_0)} \overline{s} \overline{t} \in Y_{z^{(4)},z^{(3)}}$, where $z^{(4)} \in S^{(\Lambda)}$ is obtained from $z^{(3)}$ by exchanging $s_{\lambda(n_0)}$ and $\overline{s}$. Now we have $\beta_4 := u_4 \cdot \beta_3 = \alpha_{s'} \in \Pi^{[z^{(4)}]}$, $\beta'_4 := u_4 \cdot \beta'_3 \in \Pi^{[z^{(4)}]}$ and $\langle \beta_4,\beta'_4 \rangle \neq 0$. Moreover, by the property of coefficients in $\beta'_3$ mentioned in Step 2 and the fact that $\overline{t}$ is adjacent to $s_{\lambda(n_0)}$ and $\overline{s}$, it follows that the coefficient of $\alpha_{\overline{s}}$ in $\beta'_4$ is non-zero.
\textbf{Step 4.} Since $[z^{(4)}] = [z^{(3)}]$, there exists an element $z^{(5)} \in S^{(\Lambda)}$ satisfying that $[z^{(5)}] = [z^{(2)}]$ and $u_5 := \overline{u}{}^{-1} \in Y_{z^{(5)},z^{(4)}}$. We have $\beta_5 := u_5 \cdot \beta_4 \in \Pi^{[z^{(5)}]}$, $\beta'_5 := u_5 \cdot \beta'_4 \in \Pi^{[z^{(5)}]}$ and $\langle \beta_5,\beta'_5 \rangle \neq 0$. Now by Lemma \ref{lem:proof_lem_admissible_sequence_extends_simplification} and Lemma \ref{lem:proof_lem_admissible_sequence_extends_simplification_2}, $u_5$ fixes $\alpha_{s'}$, hence $\beta_5 = \alpha_{s'}$; and the action of $u_5$ does not change the coefficient of $\alpha_{\overline{s}}$, hence the coefficient of $\alpha_{\overline{s}}$ in $\beta'_5$ is non-zero.
\textbf{Step 5.} Put $u_6 := u_2{}^{-1}$ and $u_7 := u_1{}^{-1}$. Since $[z^{(5)}] = [z^{(2)}]$ as above, there exists an element $z^{(7)} \in S^{(\Lambda)}$ satisfying that $[z^{(7)}] = [z^{(0)}] = [y^{(n_0)}]$ and $u_7u_6 \in Y_{z^{(7)},z^{(5)}}$. Now we have $\beta_7 := u_7u_6 \cdot \beta_5 = \alpha_{s'}$, since $\beta_5 = \beta_2$. On the other hand, put $\beta'_7 := u_7u_6 \cdot \beta'_5$. Then we have $\beta'_7 \in \Pi^{[z^{(7)}]} = \Pi^{[y^{(n_0)}]}$ and $\langle \beta_7,\beta'_7 \rangle \neq 0$. Moreover, since $u_7u_6 \in W_{S \smallsetminus \{\overline{s}\}}$, the coefficient of $\alpha_{\overline{s}}$ in $\beta'_7$ is the same as the coefficient of $\alpha_{\overline{s}}$ in $\beta'_5$, which is non-zero as mentioned in Step 4.
Hence we have constructed a root $\beta' = \beta'_7$ satisfying the above condition. However, this contradicts the fact that $\Pi_{J^{(n_0)}}$ is an irreducible component of $\Pi^{[y^{(n_0)}]}$ (see Definition \ref{defn:admissible_for_N_large}(\ref{item:admissible_N_large_J_irreducible_component})).
Summarizing, we have derived a contradiction in any of the four cases, Case (I)--Case (II-3), therefore Lemma \ref{lem:admissible_sequence_extends} for $n = n_0$ holds. Hence our claim has been proven in the case $\Pi^{J,I \cap J} \subseteq \Phi_{I^{\perp}}$.
This completes the proof of Theorem \ref{thm:YfixesWperpIfin}.
\section{A counterexample for the general case} \label{sec:counterexample}
In this section, we present an example which shows that our main theorem, Theorem \ref{thm:YfixesWperpIfin}, will not generally hold when the assumption on the $A_{>1}$-freeness of $I \subseteq S$ is removed.
We consider a Coxeter system $(W,S)$ of rank $7$ with Coxeter graph $\Gamma$ in Figure \ref{fig:counterexample}, where the vertex labelled by an integer $i$ corresponds to a generator $s_i \in S$. Put $I = \{s_4,s_5\}$ which is of type $A_2$ (hence is not $A_{>1}$-free). \begin{figure}
\caption{Coxeter graph $\Gamma$ and subset $I \subseteq S$ for the counterexample; here the two duplicated circles correspond to $I = \{s_4,s_5\}$}
\label{fig:counterexample}
\end{figure}
To determine the simple system $\Pi^I$ of $W^{\perp I}$, Proposition \ref{prop:factorization_C}(\ref{item:prop_factorization_C_generator_perp}) implies that each element of $\Pi^I$ is written as $u \cdot \gamma(y,s)$, where $y \in S^{(\Lambda)}$, $u \in Y_{x_I,y}$, $s \in S \smallsetminus [y]$, $[y]_{\sim s}$ is of finite type, $\varphi(y,s) = y$, and $\gamma(y,s)$ is the unique element of $(\Phi_{[y] \cup \{s\}}^{\perp [y]})^+$ as in Proposition \ref{prop:charofBphi}. In this case, the element $u^{-1} \in Y_{y,x_I}$ admits a decomposition as in Proposition \ref{prop:factorization_C}(\ref{item:prop_factorization_C_generator_Y}). In particular, such an element $y$ can be obtained from $x_I$ by applying a finite number of operations of the form $z \mapsto \varphi(z,t)$ with an appropriate element $t \in S$. Table \ref{tab:list_counterexample} gives a list of all the element $y \in S^{(\Lambda)}$ obtained in this way. In the second and the fourth columns of the table, we abbreviate each $s_i$ ($1 \leq i \leq 7$) to $i$ for simplicity. This table shows, for each $y$, all the elements $t \in S \smallsetminus [y]$ satisfying that $[y]_{\sim t}$ is of finite type and $\varphi(y,t) \neq y$, as well as the corresponding element $\varphi(y,t) \in S^{(\Lambda)}$ (more precisely, the subset $[\varphi(y,t)]$ of $S$). Now the list of the $y$ in the table is closed by the operations $y \mapsto \varphi(y,t)$, while it involves the starting point $x_I$ (No.~I in Table \ref{tab:list_counterexample}), therefore the list indeed includes a complete list of the possible $y$. \begin{table}[hbt] \centering \caption{List for the counterexample} \label{tab:list_counterexample}
\begin{tabular}{|c||c|c||c|c|} \hline No. & $[y]$ & $\gamma \in \Phi^{\perp [y]}$ & $t$ & $\varphi(y,t)$ \\ \hline I & $\{4,5\}$ & $[10|0\underline{00}|00]$, $[01|0\underline{00}|00]$ & $3$ & II \\ \cline{4-5} & & & $6$ & III \\ \cline{4-5}
& & & $7$ & IV \\ \hline II & $\{3,4\}$ & $[10|\underline{11}1|00]$, $[01|\underline{11}1|00]$ & $1$ & V \\ \cline{4-5} & & & $2$ & VI \\ \cline{4-5}
& & & $5$ & I \\ \hline III & $\{5,6\}$ & $[10|00\underline{0}|\underline{0}0]$, $[01|00\underline{0}|\underline{0}0]$ & $4$ & I \\ \cline{4-5}
& & & $7$ & IV \\ \hline IV & $\{5,7\}$ & $[10|00\underline{0}|0\underline{0}]$, $[01|00\underline{0}|0\underline{0}]$ & $4$ & I \\ \cline{4-5}
& & & $6$ & III \\ \hline V & $\{1,3\}$ & $[\underline{0}0|\underline{0}01|00]$, $[\underline{1}1|\underline{2}21|00]$ & $2$ & VI \\ \cline{4-5}
& & & $4$ & II \\ \hline VI & $\{2,3\}$ & $[\underline{0}0|\underline{0}01|00]$, $[\underline{1}1|\underline{2}21|00]$ & $1$ & V \\ \cline{4-5} & & & $4$ & II \\ \hline \end{tabular} \end{table}
On the other hand, Table \ref{tab:list_counterexample} also includes some elements of $(\Phi^{\perp [y]})^+$ for each possible $y \in S^{(\Lambda)}$. In the third column of the table, we abbreviate a root $\sum_{i=1}^{7} c_i \alpha_{s_i}$ to $[c_1c_2|c_3c_4c_5|c_6c_7]$. Moreover, a line is drawn under the coefficient $c_i$ of $\alpha_{s_i}$ if $s_i$ belongs to $[y]$. Now for each $y$, each root $\gamma \in (\Phi^{\perp [y]})^+$ and each $t$ appearing in the table, the root $w_y^t \cdot \gamma \in (\Phi^{\perp [\varphi(y,t)]})^+$ also appears in the row corresponding to the element $\varphi(y,t) \in S^{(\lambda)}$. Moreover, for each $y$ in the table, if an element $s \in S \smallsetminus [y]$ satisfies that $[y]_{\sim s}$ is of finite type and $\varphi(y,s) = y$, then the corresponding root $\gamma(y,s)$ always appears in the row corresponding to the $y$. By these properties, the above-mentioned characterization of the elements of $\Pi^I$ and the decompositions of elements of $Y_{x_I,y}$ given by Proposition \ref{prop:factorization_C}(\ref{item:prop_factorization_C_generator_Y}), it follows that all the elements of $\Pi^I$ indeed appear in the list. Hence we have $\Pi^I = \{\alpha_{s_1},\alpha_{s_2}\}$ (see the row I in Table \ref{tab:list_counterexample}), therefore both elements of $\Pi^I$ satisfy that the corresponding reflection belongs to $W^{\perp I}{}_{\mathrm{fin}}$.
Moreover, we consider the following sequence of operations: \begin{displaymath} \begin{split} x_I {}:={} & (s_4,s_5) \overset{3}{\to} (s_3,s_4) \overset{1}{\to} (s_1,s_3) \overset{2}{\to} (s_3,s_2) \overset{4}{\to} (s_4,s_3) \\ &\overset{5}{\to} (s_5,s_4) \overset{6}{\to} (s_6,s_5) \overset{7}{\to} (s_5,s_7) \overset{4}{\to} (s_4,s_5) = x_I \enspace, \end{split} \end{displaymath} where we write $z \overset{i}{\to} z'$ to signify the operation $z \mapsto z' = \varphi(z,s_i)$. Then a direct calculation shows that the element $w$ of $Y_I$ defined by the product of the elements $w_z^t$ corresponding to the above operations satisfies that $w \cdot \alpha_{s_1} = \alpha_{s_2}$. Hence the conclusion of Theorem \ref{thm:YfixesWperpIfin} does not hold in this case where the assumption on the $A_{>1}$-freeness of $I$ is not satisfied.
\noindent \textbf{Koji Nuida}\\ Present address: Research Institute for Secure Systems, National Institute of Advanced Industrial Science and Technology (AIST), AIST Tsukuba Central 2, 1-1-1 Umezono, Tsukuba, Ibaraki 305-8568, Japan\\ E-mail: k.nuida[at]aist.go.jp
\end{document}
|
arXiv
|
{
"id": "1201.1610.tex",
"language_detection_score": 0.6318438053131104,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{A Budget-Adaptive Allocation Rule for Optimal Computing Budget Allocation}
\author{Zirui Cao, Haowei Wang, Ek Peng Chew, Haobin Li, and Kok Choon Tan \thanks{
This research paper has been made possible by the funding support from the Singapore Maritime Institute \& the Center of Excellence in Modelling and Simulation for Next Generation Ports (Singapore Maritime Institute grant: SMI-2022-SP-02). } \thanks{Z. Cao, E.P. Chew, and H. Li are with the Department of Industrial Systems Engineering and Management, National University of Singapore, Singapore, CO 117576 Singapore (e-mail: [email protected]; [email protected]; li\[email protected]). } \thanks{H. Wang is with the Rice-Rick Digitalization PTE. Ltd., Singapore, CO 308900 Singapore (e-mail: haowei\[email protected]). }
\thanks{K.C. Tan is with the Department of Analytics and Operations, National University of Singapore, Singapore, CO 119245 Singapore (e-mail: [email protected]). }}
\maketitle
\begin{abstract} Simulation-based ranking and selection (R\&S) is a popular technique for optimizing discrete-event systems (DESs). It evaluates the mean performance of system designs by simulation outputs and aims to identify the best system design from a finite set of alternatives by intelligently allocating a limited simulation budget. In R\&S, the optimal computing budget allocation (OCBA) is an efficient budget allocation rule that asymptotically maximizes the probability of correct selection (PCS). However, the OCBA allocation rule ignores the impact of budget size, which plays an important role in finite budget allocation. To address this, we develop a budget allocation rule that is adaptive to the simulation budget. Theoretical results show that the proposed allocation rule can dynamically determine the ratio of budget allocated to designs based on different simulation budgets and achieve asymptotic optimality. Furthermore, the finite-budget properties possessed by our proposed allocation rule highlight the significant differences between finite budget and sufficiently large budget allocation strategies. Based on the proposed budget-adaptive allocation rule, two heuristic algorithms are developed. In the numerical experiments, we use both synthetic examples and a case study to show the superior efficiency of our proposed allocation rule. \end{abstract}
\begin{IEEEkeywords} Budget-adaptive allocation rule, discrete-event systems, optimal computing budget allocation (OCBA), ranking and selection, simulation.
\end{IEEEkeywords}
\section{Introduction} \label{sec:introduction} \IEEEPARstart{D}{iscrete}-event systems (DESs) are a widely-used technical abstraction for complex systems (see \cite{zeigler2000theory}), such as traffic control systems, manufacturing systems, and communication systems. When the complexity of DESs is high and analytical models are unavailable, a powerful tool for evaluating the performance of DESs is discrete-event system simulation (see \cite{banks2005discrete}). In this paper, we consider a problem of identifying the best system design from a finite set of competing alternatives, where “best” is defined with respect to the smallest or largest mean performance. The performance of each design is unknown and can be learnt by samples, i.e., by the sample mean of simulation outputs. Such problem is often called statistical ranking and selection (R\&S) problem (see \cite{kim2006selecting,chen2011stochastic,fu2015handbook}) or Ordinal Optimization (OO) problem (see \cite{ho1992ordinal} and \cite{ho2008ordinal}).
In R\&S, sampling efficiency is of significant concern as the simulation budget is often limited by its high expense. For example, the running time for a single simulation replication of the 24-hour dynamics of the scheduling process in a transportation network with 20 intersections can be about 2 hours, and it can take 30 minutes to obtain an accurate estimate of a maintenance strategy's average cost by running 1,000 independent simulation replications for a re-manufacturing system (see \cite{ho2008ordinal}). With finite simulation replications, it is impossible to guarantee a correct selection of the best design occurs with probability 1. This nature of the problem motivates the need of implementing R\&S techniques to intelligently allocate simulation replications to designs for efficiently identifying the best design. In our problem, we consider a fixed budget setting, and the probability of correct selection (PCS), a primary criterion in R\&S literature, is used to measure the quality of budget allocation rules. The goal is to derive a budget allocation rule that can maximize the PCS subjecting to a constraint simulation budget.
Although the simulation budget is limited and of vital importance, many R\&S algorithms allocate a simulation budget either by asymptotically optimal or by one-step-ahead optimal allocation rules, both of which can not adapt to the simulation budget. Intuitively, we argue that a desirable budget allocation rule should be adaptive to the simulation budget. We use an example to illustrate the significant impact of simulation budget on the optimal budget allocation rule. Suppose that there are three designs (design 1, 2, and 3) with normal sampling distributions $N(1,6^2)$, $N(2,6^2)$, and $N(3,6^2)$, respectively. The optimal budget allocation ratios, which can maximize the PCS, are derived by using an exhaustive search, and the PCS is estimated through Monte Carlo simulation. As shown in table \ref{table: optimal small budget allocation}, the optimal budget allocation ratios can be drastically different when the simulation budget changes. This observation is also consistent with theoretical analyses on the optimal budget allocation rules under different simulation budgets. The optimal computing budget allocation (OCBA) asymptotically maximizes the PCS, and it tends to allocate large budget allocation ratios to competitive designs, where competitive designs include the best design and non-best designs that are hard to distinguish from the best (see \cite{chen2000simulation}). However, when the simulation budget is relatively small, assigning large budget allocation ratios to competitive designs, according to the OCBA allocation rule, may decease the PCS (see \cite{peng2015non}). Such scenario is referred to as the low-confidence scenario (see \cite{peng2017gradient}) and also takes place in the expected value of information (EVI) in \cite{chick2010sequential} and knowledge gradient (KG) polices in \cite{frazier2008knowledge}. To avoid the decrease of PCS, the budget allocation ratios of competitive designs should be discounted and the budget allocation ratios of non-competitive designs should be increased (see \cite{peng2017gradient} and \cite{shi2022dynamic}). This counter-intuitive result emphasizes the significant impact of simulation budget on the budget allocation rule. It motivates the need of deriving a desirable budget allocation rule that considers and adapts to the simulation budget.
\begin{table}[thp] \caption{Optimal static budget allocation ratios for three alternative designs under different simulation budgets} \label{table: optimal small budget allocation}
\centering \begin{tabular}{@{}cccc@{}} \hline Simulation budget & Design 1 & Design 2 & Design 3 \\ \hline 20 & 0.100 & 0.350 & 0.550 \\ 50 & 0.320 & 0.360 & 0.320 \\ 200 & 0.400 & 0.365 & 0.235 \\ 500 & 0.436 & 0.396 & 0.168 \\ \hline \end{tabular} \end{table}
In this work, we consider a simulation-based R\&S problem under a fixed budget setting. We formulate the budget allocation problem as an OCBA problem. Instead of assuming a sufficiently large simulation budget like OCBA does, we consider a finite simulation budget and derive an allocation rule that is adaptive to the simulation budget. Theoretical results show that, compared with the OCBA allocation rule in \cite{chen2000simulation}, our proposed allocation rule can discount the budget allocation ratios of non-best designs that are hard to distinguish from the best design while increase the budget allocation ratios of those that are easy to distinguish. These adjustments are based on the simulation budget. When the simulation budget goes to infinity, the proposed allocation rule reduces to the OCBA allocation rule, which is asymptotically optimal. The finite-budget properties possessed by our proposed allocation rule highlight the significant differences between finite budget and sufficiently large budget allocation strategies (see \cite{peng2015non} and \cite{shi2022dynamic}).
To summarize, the main contributions of this paper are as follows: \begin{itemize}
\item[1.] We incorporate the simulation budget into the budget allocation rule and explicitly present the distinct behavior of the proposed budget-adaptive allocation rule under both finite and sufficiently large budgets. The finite-budget properties possessed by our proposed allocation rule is important as the simulation budget is often limited in practice.
\item[2.] Based on two approaches implementing the proposed budget-adaptive allocation rule, we design two heuristic algorithms, called final-budget anchorage allocation (FAA) and dynamic anchorage allocation (DAA). Specifically, FAA fixes the simulation budget to maximize the final PCS while DAA dynamically anchors the next simulation replication to maximize the PCS after the additional allocation. \end{itemize}
\section{Related Literature} \label{related literature} In simulation optimization, the objective is to select the best design among a finite set of alternatives with respect to a performance metric, e.g., PCS. Due to the noisy simulation outputs, it is impossible to surely identify the best design within finite observations. Therefore, a strategy that intends to intelligently allocate simulation replications among designs is supposed to be developed. This problem falls into the actively studied ranking and selection (R\&S) problem.
There are two branches of problem settings in R\&S literature. One is fixed-confidence setting and the other is fixed-budget setting. Fixed-confidence R\&S primarily focuses on the indifference zone (IZ) formulation and tries to guarantee a pre-specified level of PCS while using as little simulation budget as possible (see \cite{kim2001fully}). In later work, the IZ formulation is implemented to develop Frequentist procedures that can adapt to fully sequential setting (see \cite{hong2005tradeoff,batur2006fully,hong2007selecting}). Indifference-zone-free procedure that does not require an IZ parameter is proposed in \cite{fan2016indifference}. More recently, IZ procedures for large-scale R\&S problems in parallel computing environment are developed (see \cite{luo2015fully,zhong2022knockout,hong2022solving}). The fixed-budget R\&S procedures are designed to optimize a certain kind of performance metric by efficiently allocating a fixed simulation budget. In the fixed-budget setting, there are procedures that allocate a simulation budget according to an asymptotically optimal allocation rule, such as OCBA (see \cite{chen2000simulation}), the large deviation allocation (see \cite{glynn2004large}), and the optimal expected opportunity cost allocation (OEA) (see \cite{gao2016new}); and procedures that myopically maximize the expected one-step-ahead improvement, such as the expected value of information (EVI) (see \cite{chick2010sequential}), the knowledge gradient (see \cite{frazier2008knowledge}), and the approximately optimal myopic allocation policy (AOMAP) (see \cite{peng2016myopic}). In particular, the approximately optimal allocation policy (AOAP) achieves both one-step-ahead optimality and asymptotic optimality (see \cite{peng2018ranking}). In most cases, fixed-budget R\&S procedures require less simulation budget than fixed-confidence R\&S procedures to achieve the same level of PCS due to their better adaptiveness to the observed simulation outputs, however, they can not provide a statistical guarantee as fixed-confidence R\&S procedures do.
There is a unique stream of literature in R\&S focusing on the asymptotic behavior of allocation rules. The premise of these allocations is that if such allocations perform optimally when the simulation budget is sufficiently large, then they should also have satisfactory performances when the simulation budget is finite. OCBA is such a typical method that allocates a simulation budget according to an asymptotically optimal allocation rule when sampling distributions are normal (see \cite{chen2000simulation}). In later work, Gylnn and Juneja \cite{glynn2004large} applies the Large Deviation theory extending the analyses to a more general setting where sampling distributions are non-Gaussian. Gao and Chen \cite{gao2016new} present a budget allocation rule that uses the expected opportunity cost (EOC) as the quality measure of their procedure and is shown to be asymptotically optimal. Peng \textit{et al.} \cite{peng2016dynamic} formulate the problem in a stochastic dynamic program framework and derive an approximately optimal design selection policy as well as an asymptotically optimal budget allocation policy. Furthermore, this stream of methods, which explore the asymptotic behavior of allocations, are extended to solving many variants of R\&S problem, such as the subset selection problem (e.g., \cite{gao2016new}, and \cite{chen2008efficient,zhang2015simulation,zhang2022efficient}), ranking and selection with input uncertainty problem (e.g., \cite{gao2017robust} and \cite{xiao2018simulation}), ranking and selection with multiple objectives problem (e.g., \cite{lee2010finding}), stochastically constrained ranking and selection problem (e.g., \cite{hunter2013optimal} and \cite{pasupathy2014stochastically}), and contextual ranking and selection problem (e.g., \cite{li2022efficient}). The most common simplification made by such methods is to consider a sufficiently large budget, which leads to solving a simplified budget allocation problem. However, this simplification results in derived allocation rules ignoring the impact of budget size on the budget allocation strategy.
While a huge number of works contribute to developing asymptotically optimal budget allocation rules, few works investigate the impact of simulation budget on the budget allocation strategy. Typical myopic allocation rules (e.g., \cite{frazier2008knowledge}, \cite{chick2010sequential}, \cite{peng2016myopic}, and \cite{peng2018ranking}) optimize one-step-ahead improvement. In particular, Peng \textit{et al.} \cite{peng2017gradient} consider a low-confidence scenario and propose a gradient-based myopic allocation rule, which takes the induced correlations into account and performs well in such scenario. In later work, a myopic allocation rule that possesses both one-step-ahead optimality and asymptotic optimality is developed in \cite{peng2018ranking}. However, existing myopic allocation rules can not adapt to the simulation budget and some of them are not asymptotically optimal, even though they have excellent performances especially when the simulation budget is relatively small. More recently, Qin, Hong, and Fan \cite{qin2022non} formulate the budget allocation problem as a dynamic program (DP) problem and develop a non-myopic knowledge gradient (KG) policy, which can look multiple steps ahead. Shi \textit{et al.} \cite{shi2022dynamic} propose a dynamic budget-adaptive allocation rule for feasibility determination (FD) problem, a variant of R\&S problem, and show their allocation rule possesses both finite-budget properties and asymptotic optimality. None of existing works consider and develop a budget allocation rule that can not only adapt to the simulation budget but also achieve asymptotic optimality, for solving R\&S problems under a fixed budget setting.
The rest of the paper is organized as follows. The problem statement and formulation are presented in Section \ref{problem formulation}. In Section \ref{optimal budget allocation strategy}, we generalize some asymptotic results in the OCBA paradigm to the case of a general budget size; derive a budget allocation rule that is adaptive to the simulation budget; and explicitly present its desirable finite-budget properties and asymptotic optimality. Then, two heuristic algorithms implementing the proposed budget-adaptive allocation rule are developed. Numerical experiments on synthetic examples and a case study are conducted in Section \ref{numerical experiments}. In the end, Section \ref{conclusion} concludes the paper.
\section{Problem Formulation} \label{problem formulation} We introduce the following notations in our paper. \begin{table}[H] \normalsize
\begin{tabular}{ll} $k$ & The total number of designs; \\ $\mathcal{K}$ & The set of designs, i.e., $\mathcal{K} = \{1,2,\dots,k\}$; \\ $T$ & Simulation budget; \\ $X_{i,j}$ & The $j$-th simulation output sample of design $i$; \\ $\mu_i$ & Mean of the performance of design $i$, i.e.,\\
& $ \mu_i = \mathbb{E}[X_{i,j}]$; \\ $\sigma^2_i$ & Variance of the performance of design $i$, i.e., \\
& $\sigma^2_i = \text{Var}[X_{i,j}]$; \\ $b$ & Real best design, i.e., $b = \arg \min_{i \in \mathcal{K}} \mu_i$; \\ $\mathcal{K}^{\prime}$ & The set of non-best designs, i.e., $\mathcal{K}^{\prime} = \mathcal{K} \backslash b$; \\ $w_i$ & The proportion of simulation budget allocated to \\
& design $i$; \\ $N_i$ & The number of simulation replications allocated to \\
& design $i$, i.e., $N_i = w_i T$; \\ $\hat{\mu}_i$ & Sample mean of the performance of design $i$, i.e., \\
& $\hat{\mu}_i = (1/N_i) \sum_{j=1}^{N_i} X_{i,j}$; \\ $\hat{b}$ & Observed best design, i.e., $\hat{b} = \arg \min_{i \in \mathcal{K}} \hat{\mu}_i$. \end{tabular} \end{table} Suppose that there are $k$ system designs in contention. For each design $i \in \mathcal{K}$, its mean performance $\mu_i$ is unknown and can only be estimated by sampling replications via a stochastic simulation model. The goal of R\&S is to identify the real best design $b$, where “best” is defined with respective to the smallest mean. Assume that the best design $b$ is unique, i.e., $\mu_b < \mu_i$, for $ i \in \mathcal{K}^{\prime}$. This assumption basically requires the best design can be distinguished from others. Since common random numbers and correlated sampling are not considered in the paper, we assume the simulation output samples are independent across different designs and replications, i.e., $X_{i,j}$ is independent for all $i$ and $j$. The most common assumption on the sampling distribution is that the simulation observations of each design $i$ are i.i.d. normally distributed with mean $\mu_i$ and variance $\sigma_i^2$, i.e., $X_{i,j} \sim N(\mu_i,\sigma_i^2)$, for $i \in \mathcal{K}$ and $j \in \mathbb{Z}^+$. For non-Gaussian distributions, the normal assumption can be justified by a central limit theorem, e.g., by the use of batching (see \cite{kim2006selecting}). For simplicity, we ignore the integer constraints on $N_i$, and in practice $ \lfloor N_i \rfloor $ simulation replications are allocated to design $i$, where $i \in \mathcal{K}$ and $ \lfloor \cdot \rfloor $ is the flooring function.
After the simulation budget $T$ is depleted, the observed best design $\hat{b}$ (with the smallest sample mean) is selected. The event of correct selection occurs when the selected design, design $\hat{b}$, is the real best design, design $b$. Thus, we define the probability of correct selection (PCS) as \begin{align*}
\text{PCS} &= \Pr ( \hat{b} = b )\\
&= \Pr \left( \bigcap_{i \in \mathcal{K}^{\prime}} \left\{ \hat{\mu}_b < \hat{\mu}_{i} \right\} \right). \end{align*} The problem of interest is to determine $w = (w_1,w_2,\dots,w_k)$, such that by the time the simulation budget is exhausted and we select the observed best design, the PCS is maximized. Following the OCBA paradigm, we model the budget allocation problem as follows: \begin{equation*}
\begin{split}
Problem \ \mathcal{P}: \ &\max_{w} \ \text{PCS} \\
&\text{s.t.} \ \sum\limits_{i \in \mathcal{K}} w_i = 1,\\
& \quad \quad w_i \geq 0, \quad i \in \mathcal{K}. \end{split} \end{equation*}
Under general settings, the major difficulty in solving Problem $\mathcal{P}$ is that there is no closed-form expression for the PCS. Although Monte Carlo simulation can be used to approximate the PCS, its computational cost is usually unaffordable, especially when the simulated systems are of high complexity. To manage the difficulty, we evaluate the $\text{PCS}$ in an efficient way \begin{align} \label{Bonferroni inequality}
\text{PCS} &= \Pr \left( \bigcap_{i \in \mathcal{K}^{\prime}} \left\{ \hat{\mu}_b < \hat{\mu}_{i} \right\} \right) \notag \\
& = 1 - \Pr \left( \bigcup_{i \in \mathcal{K}^{\prime}} \left\{ \hat{\mu}_i \leq \hat{\mu}_{b} \right\} \right) \notag \\
& \geq 1 - \sum_{i \in \mathcal{K}^{\prime}} \Pr \left( \hat{\mu}_i \leq \hat{\mu}_{b} \right) \notag \\
&= 1 - \sum_{i \in \mathcal{K}^{\prime}} \Pr \left( Z \leq - \frac{\mu_i - \mu_b}{ \sqrt{ \frac{\sigma_i^2}{w_i T} + \frac{\sigma_b^2}{w_b T} } } \right) \notag \\
&= 1 - \sum_{i \in \mathcal{K}^{\prime}} \Phi \left( - \frac{ \delta_{i,b} }{ \sigma_{i,b} } \right) \notag \\
& = \text{APCS}, \notag \end{align} where the inequality holds by Bonferroni inequality, $Z$ is a random variable follows the standard normal distribution, $\Phi$ denotes the cumulative distribution function (c.d.f.) of the standard normal random variable, $\delta_{i,b} = \mu_i - \mu_b $, and $\sigma_{i,b} = \sqrt{\sigma_i^2/(w_i T) + \sigma_b^2/(w_b T)} $. Of note, the \text{APCS}, which serves as a cheap lower bound for the $\text{PCS}$, is frequently used in the OCBA paradigm and converges to 1 as the simulation budget $T$ goes to infinity (see \cite{chen2000simulation} and \cite{zhang2015simulation}). Instead of solving Problem $\mathcal{P}$, with the new objective $\text{APCS}$, we consider the following optimization problem: \begin{equation*}
\begin{split}
Problem \ \mathcal{P}1: \ &\max_{w} \ \text{APCS} \\
&\text{s.t.} \ \sum\limits_{i \in \mathcal{K}} w_i = 1,\\
& \quad \quad w_i \geq 0, \quad i \in \mathcal{K}. \end{split} \end{equation*}
\section{Budget Allocation Strategy} \label{optimal budget allocation strategy} In this section, we first generalize the convexity of Problem $\mathcal{P}$1 to the case of a general budget size and recover the OCBA allocation rule. Then, we derive a new budget allocation rule that is adaptive to the simulation budget and analyze its finite-budget properties and asymptotic optimality. Based on the proposed budget allocation rule, two heuristic algorithms are developed.
Of note, we highlight the most important implication of this section: simulation budget significantly impacts the budget allocation strategy. To enhance readability, all the proofs are relegated to the Appendix.
\subsection{Optimal Computing Budget Allocation}
In the OCBA paradigm, the derivation of optimality conditions essentially requires Problem $\mathcal{P}$1 to be a convex optimization problem. Zhang \textit{et al.} \cite{zhang2015simulation} consider a top-$m$ designs selection problem and show their APCS bound is concave when the simulation budget $T$ is sufficiently large. By letting $m = 1$, this result applies to Problem $\mathcal{P}$1. We generalize this result and rigorously show in Lemma 1 that the concavity of the APCS bound indeed holds for any simulation budget, thereby establishing Problem $\mathcal{P}$1 as a convex optimization problem.
\textit{Lemma 1}:
$\text{APCS}$ is concave and therefore Problem $\mathcal{P}1$ is a convex optimization problem.
\textit{Proof:} See Appendix A.1.
With the convexity, Problem $\mathcal{P}1$ can be much more easily solved by optimization solvers than the original Problem $\mathcal{P}$. Since the sampling efficiency is of significant concern in R\&S, we try to derive a solution with an analytical form for Problem $\mathcal{P}1$. The solution satisfying the Karush-Kuhn-Tucker (KKT) conditions is the optimal solution to Problem $\mathcal{P}1$ (see \cite{boyd2004convex}). Theorem 1 gives the optimality conditions of Problem $\mathcal{P}1$, and some insights can be made from it in the following analyses.
\textit{Theorem 1}:
If the solution $w = (w_1,w_2,\dots,w_k)$ maximizes the $\text{APCS}$ in Problem $\mathcal{P}1$, it satisfies the following optimality conditions $\mathcal{C}_1,\mathcal{C}_2,\mathcal{C}_3$, and $\mathcal{C}_4$ \begin{itemize}
\item[$\mathcal{C}_1$:] $w_b = \sigma_b \sqrt{\sum_{i \in \mathcal{K}^{\prime}} w_i^{2}/\sigma_i^2}$,
\item[$\mathcal{C}_2$:] $ - \frac{\delta_{i,b}^2}{ 2(\sigma_i^2/w_i + \sigma_b^2/w_b) } T + \log \frac{\delta_{i,b}\sigma_i^2}{(\sigma_i^2/w_i + \sigma_b^2/w_b)^{\frac{3}{2}}} - 2 \log w_i = \lambda, \quad i \in \mathcal{K}^{\prime}$,
\item[$\mathcal{C}_3$:] $\sum_{i \in \mathcal{K}} w_{i}=1$,
\item[$\mathcal{C}_4$:] $w_i \geq 0, \quad i \in \mathcal{K}$, \end{itemize} where $\lambda$ is a constant.
\textit{Proof:} See Appendix B.1.
When the simulation budget is sufficiently large, i.e., $T \rightarrow \infty$, the effect of the two terms $\log (\delta_{i,b}\sigma_i^2/(\sigma_i^2/w_i + \sigma_b^2/w_b)^{\frac{3}{2}})$ and $ - 2 \log w_i$ in condition $\mathcal{C}_2$ can be negligibly small compared with $ - [\delta_{i,b}^2 / (2(\sigma_i^2/w_i + \sigma_b^2/w_b))] \times T$. When $T \rightarrow \infty$, this observation motivates us to simplify condition $\mathcal{C}_2$ as \begin{equation} \label{simplified C2}
- \frac{\delta_{i,b}^2}{ 2(\sigma_i^2/w_i + \sigma_b^2/w_b) } T = \lambda, \quad i \in \mathcal{K}^{\prime}. \end{equation} By combining conditions $\mathcal{C}_1,\mathcal{C}_3,\mathcal{C}_4$, and \eqref{simplified C2}, we can verify that $w$ satisfies \begin{equation} \label{large deviation ratio} \begin{split}
w_b &= \sigma_b \sqrt{\sum\limits_{i \in \mathcal{K}^{\prime}} \frac{w_i^{2}}{\sigma_i^2}}, \\
\frac{\delta_{i,b}^2}{(\sigma_i^2/w_i + \sigma_b^2/w_b) } &= \frac{\delta_{j,b}^2}{(\sigma_j^2/w_j + \sigma_b^2/w_b) }, i,j \in \mathcal{K}^{\prime} \ \text{and} \ i \neq j , \end{split} \end{equation} the form of which corresponds to the asymptotic optimality conditions in \cite{glynn2004large} derived by maximizing the asymptotic convergence rate of probability of false selection (PFS). Compared with Theorem 1, Equation \eqref{large deviation ratio} is a great simplification but still requires a numerical solver to determine $w$.
To distinguish the best design from others, one would expect to spend most budget on the best design, i.e., $w_b \gg w_i$, for $i \in \mathcal{K}^{\prime}$. For a top-$m$ designs selection problem, Zhang \textit{et al.} \cite{zhang2015simulation} investigate the ratio of asymptotically optimal allocation ratio of non-critical designs to that of critical designs, and they show the upper bound of the ratio's growth rate is in the order of $1/\sqrt{k}$. By letting $m = 1$, one can show this result applies to $w$, that is, when $T \rightarrow \infty$, $w_i/w_b = \mathcal{O}(1/\sqrt{k})$, for $i \in \mathcal{K}^{\prime}$. The notation $f(x) = \mathcal{O}(g(x))$ means that $g(x)$ can be viewed as the upper bound of the growth rate of $f(x)$, i.e., there exists positive constants $M$ and $x_0$ such that $|f(x)| \leq M \cdot g(x)$, $\forall x \geq x_0$. When the simulation budget is finite, i.e., $T < \infty$, applying the result $w_i/w_b = \mathcal{O}(1/\sqrt{k})$, for $i \in \mathcal{K}^{\prime}$, is less straightforward. Proposition 1 formally generalizes this result to the case of a general budget size.
\textit{Proposition 1}:
Suppose that the variances of all designs are lower bounded by a positive constant $\underline{\sigma}^2 > 0$ and upper bounded by another positive constant $\bar{\sigma}^2 > 0$, that is, $0 < \underline{\sigma}^2 \leq \sigma^2_i \leq \bar{\sigma}^2$, for $i \in \mathcal{K}$. If the solution $w = (w_1,w_2,\dots,w_k)$ satisfies Theorem 1, there exists a positive constant $c > 0$, such that \begin{equation*}
\frac{w_i}{w_b} \leq \frac{\bar{\sigma}}{\underline{\sigma}} \frac{c}{\sqrt{k-1}}, \quad i \in \mathcal{K}^{\prime}, \end{equation*} and therefore, $w_i/w_b = \mathcal{O}(1/\sqrt{k})$, for $i \in \mathcal{K}^{\prime}$.
\textit{Proof:} See Appendix C.1.
Based on Proposition 1, we further simplify \eqref{large deviation ratio} by considering a sufficiently large number of designs $k$, such that $w_b \gg w_i$, for $i \in \mathcal{K}^{\prime}$. Then, we obtain
\begin{equation} \label{OCBA allocation rule}
w^*_i = \frac{I_i}{\sum_{i \in \mathcal{K}} I_i}, \quad i \in \mathcal{K}, \end{equation} where \begin{equation*}
I_i =\left\{\begin{array}{ll} \frac{\sigma_i^2}{\delta_{i,b}^2} & \quad \text{if\ } i \in \mathcal{K}^{\prime} \\ \sigma_b \sqrt{\sum_{i \in \mathcal{K}^{\prime}}\frac{I_i^{2} }{\sigma_i^2}} & \quad \text{if\ }i = b \end{array}\right. \end{equation*} This corresponds to the OCBA allocation rule in \cite{chen2000simulation}. Note that Equation \eqref{OCBA allocation rule} only has slight difference from \eqref{large deviation ratio}. For presentation simplicity, we refer both \eqref{large deviation ratio} and \eqref{OCBA allocation rule} as asymptotic optimality conditions.
\textit{Remark 1}:
The asymptotically optimal solution $w^*$ tends to assign high budget allocation ratios to non-best designs with large $I_i$, where $I_i$ is the non-best design $i$'s variance to the difference in means (between it and the best design). This result implies that, when the simulation budget approaches infinity, more simulation budget should be allocated to non-best designs that are hard to distinguish from the best, while less simulation budget should be allocated to those that are easy to distinguish.
The asymptotically optimal solution $w^*$ is independent of simulation budget and ignores the impact of budget size on it. This observation motivates the need to derive a desirable budget allocation rule which is adaptive to the simulation budget.
\subsection{Budget-Adaptive Allocation Rule}
In this subsection, we develop a budget-adaptive allocation rule that incorporates the simulation budget and approximately maximizes the APCS in Problem $\mathcal{P}$1. Instead of letting the simulation budget go to infinity, we consider a finite simulation budget, i.e., $T < \infty$. While the APCS bound in Problem $\mathcal{P}$1 is typically loose when the simulation budget is finite, solving Problem $\mathcal{P}$1 can provide a solution in analytical form. This not only aids in understanding the impact of budget size on the budget allocation strategy but also offers great improvement of computational efficiency compared with exactly calculating the PCS in Problem $\mathcal{P}$.
To derive a solution in analytical form, similarly, we consider a sufficiently large number of designs $k$, such that $w_b \gg w_i$, for $i \in \mathcal{K}^{\prime}$. Then, condition $\mathcal{C}_2$ in Theorem 1 can be simplified as \begin{equation} \label{simplified condition C2-1}
\log I_i + \log w_i + \frac{T}{I_i} w_i = \lambda, \quad i \in \mathcal{K}^{\prime}. \end{equation}
\textit{Remark 2}:
If the simulation budget is sufficiently small, i.e., $T \rightarrow 0$, the effect of $Tw_i/I_i$ in \eqref{simplified condition C2-1} can be negligible. We then combine conditions $\mathcal{C}_1$, $\mathcal{C}_3$, $\mathcal{C}_4$, and \eqref{simplified condition C2-1}, and obtain a solution $\frac{H_i}{\sum_{i \in \mathcal{K}}H_i}$, where $H_i = I_i^{-1}$, for $i \in \mathcal{K}^{\prime}$, and $H_b = \sigma_b \sqrt{\sum_{i \in \mathcal{K}^{\prime}}H_i^2/\sigma_i^2}$. This solution tends to assign low budget allocation ratios to non-best designs with large $I_i$. More specifically, when the simulation budget is small enough, less simulation budget should be allocated to non-best designs that are hard to distinguish from the best, while more simulation budget should be allocated to those that are easy to distinguish. This result is contrary to the asymptotically optimal solution $w^*$. Although the simulation budget can not approach 0 in practice, this result highlights the significant differences between finite budget and sufficiently large budget allocation strategies.
Furthermore, we approximate the term $\log w_i$ in \eqref{simplified condition C2-1} by its first-order Taylor series expansion at point $w^*_i$, for $i \in \mathcal{K}^{\prime}$ \begin{equation} \label{first order taylor expansion}
\log w_i \approx \log w^*_i + \left(w_i - w^*_i \right)/w^*_i. \end{equation} In \eqref{first order taylor expansion}, the asymptotically optimal solution $w^* = (w^*_1,w^*_2,\allowbreak\dots,w^*_k)$ is regarded as a “good” approximation to the real optimal solution $w = (w_1,w_2,\dots,w_k)$. As the simulation budget $T$ goes to infinity, this approximation tends to be accurate because $w$ would be identical to $w^*$. Therefore, we substitute the term $\log w_i$ with its approximation provided by \eqref{first order taylor expansion} and get the approximated optimality conditions for Problem $\mathcal{P}1$: \begin{itemize}
\item[$\mathcal{C}_1$:] $w_b = \sigma_b \sqrt{\sum_{i \in \mathcal{K}^{\prime}} w_i^{2}/\sigma_i^2}$,
\item[$\widehat{\mathcal{C}}_2$:] $ 2 \log I_i + \left(\frac{T}{I_i} + \frac{1}{w^*_i} \right) w_i = \lambda, \quad i \in \mathcal{K}^{\prime}$,
\item[$\mathcal{C}_3$:] $\sum_{i \in \mathcal{K}} w_{i}=1$,
\item[$\mathcal{C}_4$:] $w_i \geq 0, \quad i \in \mathcal{K}$, \end{itemize} where $\lambda$ is a constant. The condition $\mathcal{C}_2$ in Theorem 1 is approximated by condition $\widehat{\mathcal{C}_2}$, which uses the first-order Taylor series expansions of terms $\log w_i$ at points $w^*_i$, for $i \in \mathcal{K}^{\prime}$. To further simplify the problem, we temporarily omit the non-negativity constraints (condition $\mathcal{C}_4$) and consider conditions $\mathcal{C}_1$, $\widehat{\mathcal{C}_2}$, and $\mathcal{C}_3$ in Lemma 2. In Lemma 3, the non-negativity constraints are discussed to guarantee the feasibility of the solution obtained in Lemma 2.
\textit{Lemma 2}:
If the solution $W(T) = (W_1(T),W_2(T),\dots,\allowbreak W_k(T))$ solves conditions $\mathcal{C}_1$, $\widehat{\mathcal{C}}_2$, and $\mathcal{C}_3$, it satisfies \begin{equation}
W_i(T) =\left\{\begin{array}{ll} w^*_i \alpha_i(T) & \quad \text{if\ } i \in \mathcal{K}^{\prime} \\ \sigma_b \sqrt{\sum_{i \in \mathcal{K}^{\prime}}\frac{(W_i(T))^{2} }{\sigma_i^2}} & \quad \text{if\ } i = b \end{array}\right. \end{equation} where \begin{align}
\alpha_i(T) &= \frac{ (\lambda - 2 \log I_i)}{1 + T/S}, \notag \\
S &= \sum_{i \in \mathcal{K}} I_i, \notag \\
\lambda &=\left\{\begin{array}{ll} \frac{- q + \sqrt{q^2 - 4 p r}}{2 p} & \quad \text{if} \ w^*_b \neq \frac{1}{2} \\ \frac{4\sum_{i \in \mathcal{K}^{\prime}} I_i \log I_i + T + S}{2\sum_{i \in \mathcal{K}^{\prime}} I_i} & \quad \text{if} \ w^*_b = \frac{1}{2} \\ \end{array}\right. \notag \\
p &= S(2 I_b - S), \notag \\
q &= - 4 \sigma_b^2 \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{I_i^2 \log I_i }{\sigma_i^2 } \notag \\
& \qquad + 2 (S- I_b) \left(2 \sum\limits_{i \in \mathcal{K}^{\prime}} I_i \log I_i + T+S \right), \notag\\
r &= 4 \sigma_b^2 \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{I_i^2 \log^2 I_i }{\sigma_i^2 } - \left( 2 \sum\limits_{i \in \mathcal{K}^{\prime}} I_i \log I_i + T+S \right)^2. \notag \end{align}
\textit{Proof:} See Appendix A.2.
The solution $W(T)$ is an analytical function of the simulation budget $T$ and is asymptotically optimal. For a certain simulation budget $T$, a non-best design $i$, for $i \in \mathcal{K}^{\prime}$, tends to be allocated more simulation budget by $W_i(T)$ than by $w^*_i$ if $\alpha_i(T) > 1$ and be allocated less simulation budget by $W_i(T)$ than by $w^*_i$ if $\alpha_i(T) < 1$. The relationship between the best design and non-best designs remains unchanged. In particular, when the simulation budget goes to infinity, i.e., $T \rightarrow \infty$, we have $\alpha_i(T) \rightarrow 1 $, then $W_i(T) \rightarrow w_i^*$, for $i \in \mathcal{K}^{\prime}$, and consequently, $W_b(T) \rightarrow w_b^*$, implying that the solution $W(T)$ achieves asymptotic optimality.
When the simulation budget $T$ is sufficiently large, the solution $W(T)$ in Lemma 2 is feasible due to the fact $\lim_{T \rightarrow \infty } W_i(T) \rightarrow w^*_i \geq 0$, for $i \in \mathcal{K}$. However, when the simulation budget $T$ is finite, some of the budget allocation ratios in Lemma 2 may violate the non-negativity constraints and become infeasible. Because $W(T)$ is derived by temporarily omitting the non-negativity constraints (condition $\mathcal{C}_4$). For non-best designs, let $\langle j \rangle$, $j = 1,2,\dots,k - 1$, be the ascending order statistics of $I_i$, for $i \in \mathcal{K}^{\prime}$, i.e., $I_{\langle 1 \rangle} \leq I_{\langle 2 \rangle} \leq \dots \leq I_{\langle k-1 \rangle}$. Lemma 3 gives the feasibility condition of $W(T)$.
\textit{Lemma 3}:
Suppose that the solution $W(T) = (W_1(T),\allowbreak W_2(T),\dots, W_k(T))$ solves conditions $\mathcal{C}_1$, $\widehat{\mathcal{C}}_2$, and $\mathcal{C}_3$. If $W(T)$ is feasible, the simulation budget $T$ satisfies \begin{equation*}
T \geq T_0 =
\left\{\begin{array}{ll} \max\{T_1,T_2\} & \text{if} \ w^*_b \neq \frac{1}{2} \\ 4 \sum_{i \in \mathcal{K}^{\prime}} I_i \log\frac{I_{\langle k-1 \rangle}}{I_i} - S & \text{if} \ w^*_b = \frac{1}{2} \\ \end{array}\right. \end{equation*} where \begin{align*}
T_1 &= 2 \sum\limits_{i \in \mathcal{K}^{\prime}} \left[\frac{\sigma^2_b I_i^2}{\sigma_i^2 (S - I_b)} - I_i \right] \log \frac{I_{\langle k - 1 \rangle}}{I_i} - S,\\
T_2 &= 2 \sum\limits_{i \in \mathcal{K}^{\prime}} I_i \log \frac{I_{\langle k - 1 \rangle}}{I_i} \\
& \qquad + 2 \sigma_b \sqrt{ \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{ I_i^2 }{\sigma_i^2 } \left( \log \frac{I_{\langle k - 1 \rangle}}{I_i} \right)^2 } - S. \end{align*}
\textit{Proof:} See Appendix A.3.
Lemma 3 shows that $W(T)$ is always feasible when the simulation budget is relatively large, i.e., $T \geq T_0$. However, when the simulation budget $T$ is small, i.e., $T < T_0$, there exists a factor $\alpha_i(T)$ could become negative and result in $W_i(T) = w_i^* \alpha_i(T) < 0$. This implies that $W_i(T)$ could be discounted too heavily to be feasible due to the effect of $\alpha_i(T)$. To address this issue, when $T < T_0$, we allocate $T$ simulation budget according to $W(T_0)$, which is always a feasible solution to Problem $\mathcal{P}1$. Let $\lceil T_0 \rceil$ denote the smallest integer that is larger or equal to $T_0$. For non-best designs $i \in \mathcal{K}^{\prime}$, define \begin{equation} \label{new allocation ratio for design i}
\widetilde{W}_i(T) =\left\{\begin{array}{ll} W_i(T) & \quad \text{if\ } T \geq T_0 \\ W_i(\lceil T_0 \rceil) & \quad \text{if\ } T \leq T_0 \\ \end{array}\right. \end{equation} and the best design $b$ \begin{equation} \label{new allocation ratio for design b}
\widetilde{W}_b(T) = \sigma_b \sqrt{\sum_{i \in \mathcal{K}^{\prime}}\frac{( \widetilde{W}_i(T))^{2} }{\sigma_i^2}}. \end{equation}
\textit{Theorem 2}:
When the number of designs $k$ is sufficiently large, the solution $\widetilde{W}(T) = (\widetilde{W}_1(T),\widetilde{W}_2(T),\dots,\widetilde{W}_k(T))$ defined in \eqref{new allocation ratio for design i} and \eqref{new allocation ratio for design b} solves Problem $\mathcal{P}1$ and approximately maximizes the $\text{APCS}$.
The approximation used in \eqref{first order taylor expansion} and asymptotic analyses yield an approximate solution $\widetilde{W}(T)$, which is in analytical form. This not only greatly facilitates the implementation of our results, but generates insights into the budget allocation strategy for identifying the best design. The factors $\alpha_i(T)$, for $i \in \mathcal{K}^{\prime}$, play key roles in influencing the behavior of the derived new budget-adaptive allocation ratios $\widetilde{W}_i(T)$. The following proposition gives an intuition on the behavior of $\alpha_i(T)$, for $i \in \mathcal{K}^{\prime}$.
\textit{Proposition 2}:
When $0 < T < \infty$, we have $\alpha_{\langle 1 \rangle}(T) \geq 1 $, $\alpha_{\langle k - 1 \rangle}(T) \leq 1$, and $\alpha_{\langle 1 \rangle}(T) \geq \alpha_{\langle 2 \rangle}(T) \geq \dots \geq \alpha_{\langle k - 1 \rangle}(T)$, where all the equalities hold if and only if $I_i$, for $i \in \mathcal{K}^{\prime}$, are all equal.
\textit{Proof:} See Appendix C.2.
\textit{Remark 3}:
Notice that a non-best design $i$ with a large $I_i$ has a small $\alpha_i(T)$ and is hard to distinguish from the best design. Recall the analyses in Remark 1, the asymptotically optimal solution $w^*$ tends to assign high budget allocation ratios to non-best designs with large $I_i$. Due to the effects of $\alpha_i(T)$, for $i \in \mathcal{K}^{\prime}$, $\widetilde{W}(T)$ discounts the budget allocation ratios of non-best designs with large $I_i$ (e.g., design $\langle k - 1 \rangle$) and increases the budget allocation ratios of non-best designs with small $I_i$ (e.g., design $\langle 1 \rangle$). These adjustments are based on a finite simulation budget $T$. More specifically, under a finite simulation budget $T$, compared with $w^*$, $\widetilde{W}(T)$ discounts the simulation budget allocated to non-best designs that are hard to distinguish from the best, while it increases the simulation budget allocated to those that are easy to distinguish. In particular, if all non-best designs are equally hard or easy to distinguish from the best, i.e., $I_i$, for $i \in \mathcal{K}^{\prime}$, are all equal, $\widetilde{W}(T)$ is identical to $w^*$.
The asymptotic optimality of $\widetilde{W}(T)$ is a theoretical evidence supporting the approximation used in \eqref{first order taylor expansion}. Furthermore, the finite-budget properties possessed by $\widetilde{W}(T)$ are even more important. Because $\widetilde{W}(T)$ considers the impact of budget size $T$ on the budget allocation ratios, and in practice, the simulation budget $T$ is often limited by its high expense. As we will see in the numerical experiments, the asymptotically optimal budget-adaptive allocation rule $\widetilde{W}(T)$ greatly improves the efficiency of selecting the best design.
\subsection{Budget Allocation Algorithm} \label{budget allocation procedure} Based on the preceding analyses, we develop two heuristic algorithms based on two approaches implementing the proposed budget-adaptive allocation rule. Without loss of generality, we consider a fully sequential setting, where only one replication is allocated at each step. To facilitate presentation, we introduce some additional notations. \begin{table}[H] \normalsize
\begin{tabular}{ll} $t$ & The number of simulation replications has been \\
& allocated so far; \\ $A_{t}$ & Design that is allocated the $t$-th replication; \\ $\hat{b}^{(t)}$ & Design with the smallest sample mean at step $t$; \\ $\hat{\mu}^{(t)}_i$ & Sample mean of design $i$ at step $t$; \\ $(\hat{\sigma}^{(t)}_i)^2$ & Sample variance of design $i$ at step $t$; \\ $N^{(t)}_i$ & The number of simulation replications has been \\
& allocated to design $i$ across $t$ steps; \\ $w^{*(t)}_i$ & OCBA allocation ratio of design $i$ at step $t$; \\ $\widetilde{W}^{(t)}_i(T)$ & The proportion of simulation budget allocated \\
& to design $i$ at step $t$ with total budget size $T$. \end{tabular} \end{table} To calculate the budget allocation ratios at each step $t$, we use every design's sample mean $\hat{\mu}^{(t)}_i$ and sample variance $(\hat{\sigma}^{(t)}_i)^2$ as plug-in estimates for its true mean $\mu_i$ and variance $\sigma^2_i$, respectively, for $i \in \mathcal{K}$. Chen \cite{chen1996lower} and Chick and Inoue \cite{chick2001new} describe the main superiority of fully sequential procedures is that it can improve each stage's sampling efficiency by incorporating information from all earlier stages.
\subsubsection{Final-Budget Anchorage Allocation}
We develop an efficient fully sequential allocation algorithm called final-budget anchorage allocation (FAA). First, each design is initially sampled $n_0$ replications. In the second stage, we run one more replication according to $\widetilde{W}^{(t)}(T) = (\widetilde{W}^{(t)}_1(T),\widetilde{W}^{(t)}_2(T),\dots,\widetilde{W}^{(t)}_k(T))$, observe the output sample, update sample estimates and allocation ratios, and repeat until exhausting simulation budget to further distinguish the performance of each design. Notice that in each iteration, the final simulation budget $T$ is anchored, and the goal is to maximize the PCS after $T$ simulation budget is depleted. The “most starving” technique introduced in \cite{chen2011stochastic} can be applied to define an allocation policy \begin{equation} \label{allocation policy: sequential}
A_{t+1}^{FAA} = \arg \max_{i \in \mathcal{K}} \left\{ (t+1) \times \widetilde{W}^{(t)}_i(T) - N^{(t)}_i \right\}, \end{equation} which allocates the $(t+1)$-th replication to a design that is the most starving for it at step $t$. After the simulation budget is exhausted, the design with the smallest overall mean performance is selected as the best. The fully sequential FAA procedure is described in Algorithm \ref{alg:FAA}. \begin{algorithm} \caption{FAA}\label{alg:FAA}
\renewcommand{\textbf{Input:}}{\textbf{Input:}} \renewcommand{\textbf{Output:}}{\textbf{Output:}} \begin{algorithmic} \Require Set of designs $\mathcal{K}$, initial sample size $n_0$, and simulation budget $T$. \State \textbf{Initialization}: Set $t = n_0 \times k$ and $N^{(t)}_1 = N^{(t)}_2 = \dots = N^{(t)}_k = n_0$. Perform $n_0$ replications for each design. \While{$t < T$} \State Update $\hat{\mu}^{(t)}_i$, $(\hat{\sigma}^{(t)}_i)^2$, for $i \in \mathcal{K}$, $\hat{b}^{(t)} = \arg \min_{i \in \mathcal{K}} \hat{\mu}_i^{(t)}$. \State Calculate $w^{*(t)}_i$, for $i \in \mathcal{K}$, according to (\ref{OCBA allocation rule}). \State Calculate $\widetilde{W}^{(t)}_i(T)$, for $i \in \mathcal{K}$, according to (\ref{new allocation ratio for design i}) and (\ref{new allocation ratio for design b}). \State Find $A_{t+1}^{FAA}$ according to (\ref{allocation policy: sequential}). \State Perform additional one replication for design $A_{t+1}^{FAA}$. \State Set $N^{(t+1)}_{A^{FAA}_{t+1}} = N^{(t)}_{A^{FAA}_{t+1}} + 1$, $N^{(t+1)}_i = N^{(t)}_i $, for $ i \in \mathcal{K}$ and $i \neq A^{FAA}_{t+1}$, and $t = t + 1$. \EndWhile \Ensure $\hat{b}^{(T)} = \arg \min_{i \in \mathcal{K}} \hat{\mu}_i^{(T)}$ \end{algorithmic} \end{algorithm}
\subsubsection{Dynamic Anchorage Allocation}
We extend the FAA procedure to a more flexible variant, named as dynamic anchorage allocation (DAA), by allowing dynamically changing the anchored final budget instead of fixing it during the procedure. At each step $t$, we anchor $t + 1$ as the final budget, run additional one replication according to $\widetilde{W}^{(t)}(t + 1)$, observe the output sample, update sample estimates and allocation ratios, and repeat until the simulation budget is exhausted. Notice that in each iteration, the next simulation replication is anchored, and the goal becomes maximizing the PCS after the additional allocation. Again, the “most starving” technique in \cite{chen2011stochastic} can be used to obtain a budget allocation policy, which is define as \begin{equation} \label{allocation policy: dynamic}
A_{t+1}^{DAA} = \arg \max_{i \in \mathcal{K}} \left\{ (t+1) \times \widetilde{W}^{(t)}_i(t+1) - N^{(t)}_i \right\}. \end{equation} And the fully sequential DAA procedure can be implemented by Algorithm \ref{alg:DAA}. \begin{breakablealgorithm} \caption{DAA} \label{alg:DAA}
\renewcommand{\textbf{Input:}}{\textbf{Input:}} \renewcommand{\textbf{Output:}}{\textbf{Output:}} \begin{algorithmic} \Require Set of designs $\mathcal{K}$, initial sample size $n_0$, and simulation budget $T$. \State \textbf{Initialization}: Set $t = n_0 \times k$ and $N^{(t)}_1 = N^{(t)}_2 = \dots = N^{(t)}_k = n_0$. Perform $n_0$ replications for each design. \While{$t < T$} \State Update $\hat{\mu}^{(t)}_i$, $(\hat{\sigma}^{(t)}_i)^2$, for $i \in \mathcal{K}$, $\hat{b}^{(t)} = \arg \min_{i \in \mathcal{K}} \hat{\mu}_i^{(t)}$. \State Calculate $w^{*(t)}_i$, for $i \in \mathcal{K}$, according to (\ref{OCBA allocation rule}). \State Calculate $\widetilde{W}^{(t)}_i(t+1)$, for $i \in \mathcal{K}$, according to (\ref{new allocation ratio for design i}) and (\ref{new allocation ratio for design b}). \State Find $A_{t+1}^{DAA}$ according to (\ref{allocation policy: dynamic}). \State Perform additional one replication for design $A_{t+1}^{DAA}$. \State Set $N^{(t+1)}_{A^{DAA}_{t+1}} = N^{(t)}_{A^{DAA}_{t+1}} + 1$, $N^{(t+1)}_i = N^{(t)}_i $, for $ i \in \mathcal{K}$ and $i \neq A^{DAA}_{t+1}$, and $t = t + 1$. \EndWhile \Ensure $\hat{b}^{(T)} = \arg \min_{i \in \mathcal{K}} \hat{\mu}_i^{(T)}$. \end{algorithmic} \end{breakablealgorithm}
\section{Numerical Experiments} \label{numerical experiments}
In this section, we conduct numerical experiments on synthetic examples and a facility location problem to show the superior performance and applicability of our proposed FAA and DAA procedures. The experiments are conducted in MATLAB R2022b on a computer with Intel Core i5-10400 CPU with 2.90 GHz, 16 GB memory, a 64-bit operating system, and 6 cores with 12 logical processors. We use three simulation budget allocation procedures for comparison.
\begin{itemize}
\item Equal allocation (EA). This is the simplest method to conduct experiments. The simulation budget is equally allocated to all designs, i.e., $N_i = T/k$ and $w_i = 1/k$, for $i \in \mathcal{K}$. The equal allocation is a good benchmark for performance comparison.
\item OCBA allocation (see \cite{chen2000simulation}). OCBA is guided by the asymptotically optimal allocation rule defined in \eqref{OCBA allocation rule}. We implement a fully sequential OCBA procedure, which allocates a single replication at each step according to the “most starving” technique in \cite{chen2011stochastic}. Similarly, at each step, sample means and variances are used as plug-in estimates for the true means and variances to calculate the OCBA allocation ratios.
\item AOAP allocation (see \cite{peng2018ranking}). AOAP is an efficient budget allocation procedure that achieves both one-step-head optimality and asymptotic optimality. AOAP requires the variances of designs to be known, and again, we use sample variances as plug-in estimates for the true variances. When the simulation budget goes to infinity, AOAP reaches the asymptotically optimal budget allocation ratios defined in \eqref{large deviation ratio}. \end{itemize}
\subsection{Test Problems}
\subsubsection{Synthetic Examples}
To demonstrate the efficiency of the proposed FAA and DAA, we consider four synthetic problem settings, which are described as follows:
\textit{Example 1}: There are 10 alternative designs with sampling distributions $N(i,6^2)$, for $i = 1,2,\dots,10$. The goal is to identify the best design via simulation samples, where the best is $b=1$ in this example.
\textit{Example 2}: This is a variant of Example 1. All settings are the same except that the variance is decreasing with respect to the indices. In this example, better designs are with larger variances. The designs' sampling distributions are $N(i,(11-i)^2)$, for $i = 1,2,\dots,10$. Again, the best design is $b = 1$.
\textit{Example 3}: This is another variant of Example 1 with larger number of designs and variances. The designs' sampling distributions are $N(i,10^2)$, for $i = 1,2,\dots,50$. Again, the best design is $b = 1$.
\textit{Example 4}: There are 500 normal alternative designs. The sampling distribution of the best design, design $1$, is $N(0,6^2)$. As for non-best designs $i$, for $i = 2,3,\dots,500$, their sampling distributions are $N(\mu_i, \sigma_i^2)$, where $\mu_i$ and $\sigma_i$ are generated from two uniform distributions $U(1,16)$ and $U(3,9)$, respectively.
In all experiments with synthetic examples, we set the initial number of simulation replications per design to be 3, i.e., $n_0 = 3$. Because we want to distinguish the performances of different allocation procedures especially when the simulation budget is relatively small; the simulation budget $T$ is 1,000, 3,000, 5,000, and 80,000 for Example 1, 2, 3, and 4, respectively; the empirical PCS of each procedure is estimated from 100,000 independent macro replications for Example 1-3 and from 10,000 independent macro replications for Example 4. The performance comparison of the five procedures with different simulation budgets are reported in Figure \ref{fig:sfig1}-\ref{fig:sfig4} and Table \ref{table: performance comparison}.
\subsubsection{Facility Location Problem}
The facility location problem is a practical test problem provided by the Simulation Optimization Library (\url{https://github.com/simopt-admin/simopt}) and has also been studied in \cite{gao2016new}. There is a company selling one product that will never be out of stock in a city. Without loss of generality, the city is assumed to be a unit square, i.e., $[0,1]^2$, and distances are measured in units of 30 km. Two warehouses are located in the city and each of them has 10 trucks delivering orders individually. Orders are generated from 8 AM to 5 PM by a stationary Poisson process with a rate parameter 1/3 per minutes and are located in the city according to a density function \begin{equation*}
f(x,y) = 1.6 - ( |x-0.8| + |y - 0.8| ), \quad x, y \in [0,1]. \end{equation*} When order arrives, it is dispatched to the nearest warehouse with available trucks. Otherwise, it is placed into a queue and satisfied by following the first-in-first-out pattern when trucks become idle. Then, the trucks pick the order up, travel to the delivery point, deliver the products and return to their assigned warehouses waiting for the next order, where the pick-up and deliver time are exponentially distributed with mean 5 and 10, respectively. All trucks travel in Manhattan fashion at a constant speed 30 km/hour, and orders must be delivered on the day when it is received.
The objective is to find the locations of the two warehouses that can maximize the proportion of orders which are delivered within 60 minutes. Let $(z_{i,1},z_{i,2})$ and $(z_{i,3},z_{i,4})$ be the two locations, respectively. We consider 10 alternatives $(z_{i,1},z_{i,2},z_{i,3},z_{i,4}) = (0.49+0.01i,0.59+0.01i,0.59+0.01i,0.79+0.01i)$, for $i = 1,2,\dots,10$. In this experiment, we run 30 days of simulation in each replication, and the proportion of orders satisfied within 60 minutes is the average proportion of satisfied orders during the 30 days. Thus, the proportion of orders satisfied within 60 minutes is approximately normally distributed. By comparing $100,000$ simulation samples of each design, the best design is determined, i.e., $(z_{1,1},z_{1,2},z_{1,3},z_{1,4}) = (0.5,0.6,0.6,0.8)$.
In the facility location problem, we set the initial number of simulation replication per design $n_0$, the simulation budget $T$, and the number of independent macro replications for estimating the empirical PCS as 3, 800, and 10,000, respectively. The performance of each tested budget allocation procedure is presented in Figure \ref{fig:FACLOC} and Table \ref{table: performance comparison}.
\begin{figure}
\caption{(Color online) Performance comparison of the five procedures on Example 1}
\label{fig:sfig1}
\end{figure}
\begin{figure}
\caption{(Color online) Performance comparison of the five procedures on Example 2}
\label{fig:sfig2}
\end{figure}
\begin{figure}
\caption{(Color online) Performance comparison of the five procedures on Example 3}
\label{fig:sfig3}
\end{figure}
\begin{figure}
\caption{(Color online) Performance comparison of the five procedures on Example 4}
\label{fig:sfig4}
\end{figure}
\begin{figure}
\caption{(Color online) Performance comparison of the five allocation procedures on the facility location problem}
\label{fig:FACLOC}
\end{figure}
\begin{table}[thp]
\caption{Performance comparison of the five allocation procedures on different simulation budgets for synthetic examples} \label{table: performance comparison} \centering \begin{tabular}{@{}lccccccc@{}} \midrule \multirow{2}{*}{Example 1} & \multicolumn{7}{c}{Simulation budget} \\ \cmidrule(l){2-8}
& 50 & 100 & 200 & 400 & 600 & 800 & 1000 \\ \midrule EA & 0.425 & 0.523 & 0.631 & 0.744 & 0.805 & 0.846 & 0.876 \\ OCBA & 0.466 & 0.623 & 0.749 & 0.856 & 0.906 & 0.934 & 0.950 \\ AOAP & 0.492 & 0.643 & 0.760 & 0.857 & 0.902 & 0.928 & 0.943 \\ FAA (ours) & 0.474 & 0.631 & 0.771 & 0.981 & 0.930 & 0.954 & 0.967 \\ DAA (ours) & 0.473 & 0.631 & 0.771 & 0.986 & 0.934 & 0.957 & 0.969 \\ \midrule \multirow{2}{*}{Example 2} & \multicolumn{7}{c}{Simulation budget} \\ \cmidrule(l){2-8}
& 50 & 150 & 500 & 1000 & 1500 & 2000 & 3000 \\ \midrule EA & 0.389 & 0.505 & 0.654 & 0.753 & 0.811 & 0.850 & 0.900 \\ OCBA & 0.388 & 0.571 & 0.760 & 0.858 & 0.906 & 0.933 & 0.959 \\ AOAP & 0.404 & 0.583 & 0.751 & 0.844 & 0.892 & 0.919 & 0.949 \\ FAA (ours) & 0.398 & 0.589 & 0.789 & 0.890 & 0.935 & 0.955 & 0.974 \\ DAA (ours) & 0.396 & 0.586 & 0.792 & 0.895 & 0.938 & 0.958 & 0.976 \\ \midrule \multirow{2}{*}{Example 3} & \multicolumn{7}{c}{Simulation budget} \\ \cmidrule(l){2-8}
& 200 & 500 & 800 & 1000 & 2000 & 3000 & 5000 \\ \midrule EA & 0.281 & 0.382 & 0.443 & 0.375 & 0.581 & 0.643 & 0.725 \\ OCBA & 0.356 & 0.635 & 0.724 & 0.762 & 0.864 & 0.907 & 0.947 \\ AOAP & 0.429 & 0.672 & 0.755 & 0.791 & 0.886 & 0.924 & 0.955 \\ FAA (ours) & 0.383 & 0.677 & 0.775 & 0.814 & 0.912 & 0.945 & 0.970 \\ DAA (ours) & 0.382 & 0.679 & 0.782 & 0.822 & 0.920 & 0.953 & 0.974 \\ \midrule \multirow{2}{*}{Example 4} & \multicolumn{7}{c}{Simulation budget ($ \times 10^3$)} \\ \cmidrule(l){2-8}
& 6 & 8 & 10 & 20 & 30 & 60 & 80 \\ \midrule EA & 0.089 & 0.115 & 0.145 & 0.290 & 0.414 & 0.675 & 0.780 \\ OCBA & 0.591 & 0.631 & 0.658 & 0.736 & 0.768 & 0.818 & 0.834 \\ AOAP & 0.734 & 0.772 & 0.791 & 0.841 & 0.862 & 0.892 & 0.902 \\ FAA (ours) & 0.733 & 0.789 & 0.813 & 0.877 & 0.906 & 0.931 & 0.940 \\ DAA (ours) & 0.729 & 0.788 & 0.817 & 0.879 & 0.910 & 0.947 & 0.954 \\ \midrule \multirow{2}{*}{\makecell[l]{Facility \\location \\problem}} & \multicolumn{7}{c}{Simulation budget} \\ \cmidrule(l){2-8}
& 40 & 120 & 200 & 300 & 500 &700 & 800 \\ \midrule EA & 0.458 & 0.600 & 0.673 & 0.731 & 0.805 & 0.853 & 0.872 \\ OCBA & 0.501 & 0.715 & 0.805 & 0.863 & 0.922 & 0.952 & 0.959 \\ AOAP & 0.509 & 0.740 & 0.813 & 0.869 & 0.922 & 0.947 & 0.957 \\ FAA (ours) & 0.504 & 0.731 & 0.827 & 0.880 & 0.942 & 0.962 & 0.971 \\ DAA (ours) & 0.502 & 0.743 & 0.830 & 0.889 & 0.943 & 0.968 & 0.976 \\ \bottomrule \end{tabular}
\end{table}
\subsection{Discussion on Experiment Results}
From Figure \ref{fig:sfig1}-\ref{fig:FACLOC} and Table \ref{table: performance comparison}, we can see that both FAA and DAA have the best overall performances across all tested examples. This observation verifies the benefits of the proposed budget-adaptive allocation rule. AOAP performs the best at the beginning, but it surpassed by both FAA and DAA when the simulation budget is relatively large. This observation corresponds to the fact that AOAP is a myopic procedure that maximizes one-step-ahead improvement but can not adapt to the simulation budget. When the problem scale is small (e.g., Example 1-3 and the facility location problem), AOAP and OCBA are compatible. However, when the problem scale is relatively large (e.g., Example 4), AOAP performs better than OCBA. It is interesting to see in Example 2 that competitive designs (that are hard to distinguish from the best design) are assigned large budget allocation ratios by OCBA due to their large variances. However, based on the finite budget size, the budget allocation ratios assigned to competitive designs and non-competitive designs should be discounted and increased, respectively. Therefore, both FAA and DAA, which consider the impact of simulation budget, have better performances than OCBA on Example 2 (see Figure \ref{fig:sfig2}). Since EA incorporates no sample information and is a pure random sampling procedure, it is dominated by the other four procedures. The reason why both FAA and DAA have superior performances is that they consider the impact of budget size on budget allocation strategies and possess not only desirable finite-budget properties but also asymptotic optimality.
In Table \ref{table: performance comparison}, we can see that FAA performs slightly better than DAA when the simulation budget is relatively small. This verifies the benefit of anchoring the final budget. However, FAA is outperformed by DAA after the simulation budget grows up to 400, 500, 500, 10,000, and 120 in Example 1, 2, 3, 4, and the facility location problem, respectively. These observations illustrate that, when the simulation budget $T$ or the problem scale is relatively small, FAA is preferred. Otherwise, DAA is recommended.
As shown in Table \ref{table: Average runtime of the five allocation procedures (in seconds)}, we can see that EA are the fastest procedure as it does not utilize any sample information. The average runtimes for both FAA and DAA are almost the same, and they are longer than EA and OCBA but on the same magnitude as OCBA. The reason why both FAA and DAA are more time-consuming than OCBA is that, at each iteration, they require additional computational time on $\alpha_i(T)$ to consider the impact of budget size on budget allocation ratios. Furthermore, the average runtimes of AOAP are compatible with both FAA and DAA when the problem scale is relatively small, e.g., in Example 1-2. However, as the problem scale becomes large, e.g., in Example 3-4, the average runtimes of AOAP drastically increases and is much larger than both FAA and DAA. At each iteration, AOAP requires $(k-1)^2$ pairwise comparisons, while both FAA and DAA require $k-1$ parwise comparisons, which is the same as OCBA. Since pairwise comparisons are very time-consuming, both FAA and DAA are more computationally efficient and are more suitable for solving large-scale problems than AOAP, even though AOAP performs better when the simulation is relatively small.
The average runtimes for the five competing procedures are very close in the experiments on the facility location problem. This result implies that in the case where the complexity of the real system is high and the simulation time of such system is relative long, the additional runtimes required for calculating $\alpha_i(T)$ in both FAA and DAA, as well as the computational burden caused by too many pairwise comparisons in AOAP, can be negligible. Note that the industrial systems in the real world can be much more complex than the logistic system considered in this paper. Therefore, both FAA and DAA, as well as AOAP, are competitive in real industrial applications. These results verify the effectiveness and applicability of our proposed FAA and DAA procedures.
\begin{table}[thp]
\caption{Average runtime of the five allocation procedures (in seconds)} \label{table: Average runtime of the five allocation procedures (in seconds)} \centering \begin{tabular}{@{}lccccc@{}} \midrule
& EA & OCBA & AOAP & FAA (ours) & DAA (ours) \\ \hline Example 1 & 0.001 & 0.006 & 0.015 & 0.012 & 0.012 \\ Example 2 & 0.001 & 0.003 & 0.007 & 0.006 & 0.006 \\ Example 3 & 0.005 & 0.030 & 0.335 & 0.057 & 0.058 \\ Example 4 & 0.009 & 2.780 & 107.061 & 7.802 & 7.842 \\
\makecell[l]{Facility \\location \\problem} & 83.344 & 85.332 & 85.377 & 85.440 & 85.347\\ \bottomrule \end{tabular} \end{table}
\section{Conclusion} \label{conclusion} In this paper, we consider a simulation-based R\&S problem of selecting the best system design from a finite set of alternatives under a fixed budget setting. We propose a budget allocation rule that is adaptive to the simulation budget and possesses both desirable finite-budget properties and asymptotic optimality. Based on the proposed budget-adaptive allocation rule, two heuristic algorithms FAA and DAA are developed. In the numerical experiments, both FAA and DAA clearly outperform other competing procedures.
We highlight the most important implication of our contributions: simulation budget significantly impacts the budget allocation strategy, and a desirable budget allocation rule should be adaptive to the simulation budget. The proposed budget-adaptive allocation rule indicates that, compared with the asymptotically optimal OCBA allocation rule, the budget allocation ratios of non-best designs that are hard to distinguish from the best design should be discounted, while the budget allocation ratios of those that are easy to distinguish from the best should be increased. These adjustments are based on the simulation budget, which is often limited and finite. Therefore, the budget allocation strategy should be sensitive to the simulation budget, and highlights the significant difference between finite budget and sufficiently large budget allocation strategies. We believe these findings can help and motivate researchers to develop more efficient finite-budget allocation rules in future studies.
The budget allocation problem can be essentially formulated as a stochastic dynamic program (DP) problem. Both FAA and DAA are efficient procedures that can adapt to the simulation budget, but they ignore the dynamic feedback of the final step while sampling at the current step. Recently, Qin, Hong, and Fan \cite{qin2022non} in their preliminary version of work formulate the problem as a DP and investigate a non-myopic knowledge gradient (KG) procedure, which can dynamically look multiple steps ahead and take the dynamic feedback mechanism into consideration. However, exactly solving the DP is intractable due to the extremely high computational cost caused by “curse of dimensionality”. As a result, how to derive a computationally tractable allocation rule that can incorporate the dynamic feedback mechanism, remains a critical future direction.
\section*{Appendix}
\begingroup \allowdisplaybreaks
\subsection{Proof of Lemmas}
\subsubsection{Proof of Lemma 1} \label{Proof of Lemma 1}
The constraints of Problem $\mathcal{P}1$ are all affine functions of $w_i$, for $i \in \mathcal{K}$. Furthermore, showing $\text{APCS}$ is concave is equivalent to showing \begin{equation*}
g(w) = \sum_{i \in \mathcal{K}^{\prime}} \Phi \left( - \frac{\delta_{i,b}}{\sigma_{i,b}} \right), \end{equation*} is a convex function of $w$. To verify the convexity of $g(w)$, we need to show its Hessian matrix is positive semi-definite. The Hessian matrix for $g(w)$ is \begin{equation*} \nabla^2 g(w) = \begin{pmatrix} \frac{\partial^{2} g(w)}{\partial^2 w_1} & \frac{\partial^{2} g(w)}{\partial w_1 \partial w_2} & \cdots & \frac{\partial^{2} g_(w)}{\partial w_1 \partial w_k} \\ \frac{\partial^{2} g(w)}{\partial w_2 \partial w_1} & \frac{\partial^{2} g(w)}{\partial^2 w_2} & \cdots & \frac{\partial^{2} g(w)}{\partial w_2 \partial w_k} \\ \vdots & \vdots & \ddots & \vdots \\ \frac{\partial^{2} g(w)}{\partial w_k \partial w_1} & \frac{\partial^{2} g(w)}{\partial w_k \partial w_2} & \cdots & \frac{\partial^{2} g(w)}{\partial^2 w_k} \end{pmatrix}, \end{equation*} where, for $i,j \in \mathcal{K}^{\prime}$ and $i \neq j$, and $b$ \begin{align*}
\frac{\partial^{2} g(w)}{\partial w_i^{2}}&= \frac{1}{2 \sqrt{2 \pi}} \exp \left(-\frac{\delta_{i,b}^{2}}{2 \sigma_{i,b}^{2}}\right)\\ & \quad \times\left[\frac{\delta_{i,b}}{\sigma_{i,b}^{5} T^{2}} \frac{\sigma_{i}^{4}}{w_i^{4}}\left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}-\frac{3}{2}\right)+2 \frac{\delta_{i,b}}{\sigma_{i,b}^{3} T} \frac{\sigma_{i}^{2}}{w_i^{3}}\right], \\ \frac{\partial^{2} g(w)}{\partial w_b^{2}}&= \frac{1}{2 \sqrt{2 \pi}} \sum\limits_{i \in \mathcal{K}^{\prime}} \exp \left(-\frac{\delta_{i,b}^{2}}{2 \sigma_{i,b}^{2}}\right) \\ &\quad\times\left[\frac{\delta_{i,b}}{\sigma_{i,b}^{5} T^{2}} \frac{\sigma_{b}^{4}}{w_b^{4}}\left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}-\frac{3}{2}\right)+2 \frac{\delta_{i,b}}{\sigma_{i,b}^{3} T} \frac{\sigma_{b}^{2}}{w_b^{3}}\right], \\ \frac{\partial^{2} g(w)}{\partial w_i \partial w_b}&= \frac{1}{2 \sqrt{2 \pi}} \exp \left(-\frac{\delta_{i,b}^{2}}{2 \sigma_{i,b}^{2}}\right) \\ &\quad\times\left[\frac{\delta_{i,b}}{\sigma_{i,b}^{5} T^{2}} \frac{\sigma_{i}^{2}}{w_i^{2}} \frac{\sigma_{b}^{2}}{w_b^{2}}\left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}-\frac{3}{2}\right)\right],\\ \frac{\partial^{2} g(w)}{\partial w_i \partial w_{j}} &= 0. \end{align*} Furthermore, for any non-zero vector $a\in \mathbb{R}^k$, we have \begin{align*}
&a^{T} \nabla^2 g(w) a \\
&= \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{\partial^{2} g(w)}{\partial w_i^2} a_i^2 + \frac{\partial^{2} g(w)}{\partial w_b^{2}} a_b^2 \\
& \quad + 2\sum\limits_{i \in \mathcal{K}^{\prime}} \frac{\partial^{2} g(w)}{\partial w_i \partial w_b} a_i a_b\\
&= \frac{1}{2 \sqrt{2 \pi}} \sum\limits_{i \in \mathcal{K}^{\prime}} \exp \left(-\frac{\delta_{i,b}^{2}}{2 \sigma_{i,b}^{2}}\right)\\ & \quad \times a_i^2 \left[\frac{\delta_{i,b}}{\sigma_{i,b}^{5} T^{2}} \frac{\sigma_{i}^{4}}{w_i^{4}}\left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}-\frac{3}{2}\right)+2 \frac{\delta_{i,b}}{\sigma_{i,b}^{3} T} \frac{\sigma_{i}^{2}}{w_i^{3}}\right]\\
&\quad + \frac{1}{2 \sqrt{2 \pi}} \sum\limits_{i \in \mathcal{K}^{\prime}} \exp \left(-\frac{\delta_{i,b}^{2}}{2 \sigma_{i,b}^{2}}\right) \\
&\quad\times a_b^2 \left[\frac{\delta_{i,b}}{\sigma_{i,b}^{5} T^{2}} \frac{\sigma_{b}^{4}}{w_b^{4}}\left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}-\frac{3}{2}\right)+2 \frac{\delta_{i,b}}{\sigma_{i,b}^{3} T} \frac{\sigma_{b}^{2}}{w_b^{3}}\right]\\
&\quad + \frac{1}{2 \sqrt{2 \pi}} \sum\limits_{i \in \mathcal{K}^{\prime}} \exp \left(-\frac{\delta_{i,b}^{2}}{2 \sigma_{i,b}^{2}}\right) \\
&\quad\times 2 a_i a_b \left[\frac{\delta_{i,b}}{\sigma_{i,b}^{5} T^{2}} \frac{\sigma_{i}^{2}}{w_i^{2}} \frac{\sigma_{b}^{2}}{w_b^{2}}\left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}-\frac{3}{2}\right)\right] \\
&= \frac{1}{2 \sqrt{2 \pi}} \sum\limits_{i \in \mathcal{K}^{\prime}} \exp \left(-\frac{\delta_{i,b}^{2}}{2 \sigma_{i,b}^{2}}\right) \\
& \quad \times \frac{\delta_{i,b}}{\sigma_{i,b}^{3} T} \left[\frac{1}{\sigma_{i,b}^{2} T} \left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}-\frac{3}{2}\right) \left(\frac{\sigma_{i}^{2}}{w_i^{2}} a_i + \frac{\sigma_{b}^{2}}{w_b^{2}} a_b \right)^2 \right.\\
&\left.\quad \quad + 2 \left( \frac{\sigma_{i}^{2}}{w_i^{3}} a_i^2 + \frac{\sigma_{b}^{2}}{w_b^{3}} a_b^2 \right) \right]. \\ \end{align*} Since \begin{align*}
2 \left( \frac{\sigma_{i}^{2}}{w_i^{3}} a_i^2 + \frac{\sigma_{b}^{2}}{w_b^{3}} a_b^2 \right) & = 2 \frac{\sigma_{i,b}^2}{\sigma_{i,b}^2} \left( \frac{\sigma_{i}^{2}}{w_i^{3}} a_i^2 + \frac{\sigma_{b}^{2}}{w_b^{3}} a_b^2 \right)\\
& = \frac{2}{\sigma_{i,b}^2 T} \left(\frac{\sigma_i^2}{w_i} + \frac{\sigma_b^2}{w_b} \right) \left( \frac{\sigma_{i}^{2}}{w_i^{3}} a_i^2 + \frac{\sigma_{b}^{2}}{w_b^{3}} a_b^2 \right)\\
& = \frac{2}{\sigma_{i,b}^2 T} \left[\left(\frac{\sigma_{i}^{2}}{w_i^{2}} a_i + \frac{\sigma_{b}^{2}}{w_b^{2}} a_b \right)^2 \right.\\
&\left. \quad \quad+ \frac{\sigma_i^2 \sigma_b^2}{w_i w_b} \left( \frac{a_i}{w_i} - \frac{a_b}{w_b} \right)^2\right], \\ \end{align*} we have \begin{align*}
&\frac{1}{\sigma_{i,b}^{2} T} \left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}-\frac{3}{2}\right) \left(\frac{\sigma_{i}^{2}}{w_i^{2}} a_i+ \frac{\sigma_{b}^{2}}{w_b^{2}} a_b \right)^2\\
& \quad + 2 \left( \frac{\sigma_{i}^{2}}{w_i^{3}} a_i^2 + \frac{\sigma_{b}^{2}}{w_b^{3}} a_b^2 \right) \\
= & \frac{1}{\sigma_{i,b}^{2} T} \left[ \left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}-\frac{3}{2}\right) \left(\frac{\sigma_{i}^{2}}{w_i^{2}} a_i + \frac{\sigma_{b}^{2}}{w_b^{2}} a_b \right)^2 \right.\\
&\left. \quad+ 2 \left(\frac{\sigma_{i}^{2}}{w_i^{2}} a_i + \frac{\sigma_{b}^{2}}{w_b^{2}} a_b \right)^2 + 2 \frac{\sigma_i^2 \sigma_b^2}{w_i w_b} \left( \frac{a_i}{w_i} - \frac{a_b}{w_b} \right)^2\right] \\
= & \frac{1}{\sigma_{i,b}^{2} T} \left[ \left(\frac{1}{2} \frac{\delta_{i,b}^{2}}{\sigma_{i,b}^{2}}+\frac{1}{2}\right) \left(\frac{\sigma_{i}^{2}}{w_i^{2}} a_i + \frac{\sigma_{b}^{2}}{w_b^{2}} a_b \right)^2 \right.\\
& \left. \quad + 2 \frac{\sigma_i^2 \sigma_b^2}{w_i w_b} \left( \frac{a_i}{w_i} - \frac{a_b}{w_b} \right)^2\right]\\
\geq & 0, \end{align*} and therefore, $\nabla^2 g(w) \succeq 0$, and $g(w)$ is a convex function of $w$. Due to the two constraints $\sum_{i \in \mathcal{K}} w_i = 1$ and $w_i \geq 0$, for $i \in \mathcal{K}$, forming a convex set, Problem $\mathcal{P}1$ is a convex optimization problem. This result concludes the proof.
\subsubsection{Proof of Lemma 2}
By condition $\widehat{\mathcal{C}}_2$, for $i \in \mathcal{K}^{\prime}$, we have \begin{equation} \label{proof p1 w_i} \begin{split}
w_i &= \frac{\lambda - 2 \log I_i}{\frac{T}{I_i} + \frac{1}{w^*_i}} \\
&= \frac{I_i (\lambda - 2 \log I_i)}{T+S} , \end{split} \end{equation} in which, $S = \sum_{i \in \mathcal{K}} I_i$ and $w^*_i = I_i / S$. We substitute $w_i$ provided by \eqref{proof p1 w_i} into condition $\mathcal{C}_1$, and obtain \begin{equation} \label{proof p1 w_b}
w_b = \sigma_b \sqrt{ \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{I_i^2(\lambda - 2 \log I_i )^2}{\sigma_i^2 (T+S)^2} }. \end{equation} For $i \in \mathcal{K}^{\prime}$, by condition $\mathcal{C}_3$, \eqref{proof p1 w_i}, and \eqref{proof p1 w_b} \begin{equation} \label{verify lambda feasibility}
\sum\limits_{i \in \mathcal{K}^{\prime}} \frac{I_i (\lambda - 2 \log I_i)}{T+S} + \sigma_b \sqrt{ \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{I_i^2(\lambda - 2 \log I_i )^2}{\sigma_i^2 \left(T+S\right)^2} } = 1. \end{equation} Then, we have \begin{equation*}
\lambda =\left\{\begin{array}{ll} \frac{- q + \sqrt{q^2 - 4 p r}}{2 p} & \quad \text{if} \ w^*_b \neq \frac{1}{2} \\ \frac{4\sum_{i \in \mathcal{K}^{\prime}} I_i \log I_i + T + S}{2\sum_{i \in \mathcal{K}^{\prime}} I_i} & \quad \text{if} \ w^*_b = \frac{1}{2} \\ \end{array}\right. \end{equation*} where \begin{equation*} \begin{split}
p &= S(2 I_b - S), \\
q &= - 4 \sigma_b^2 \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{I_i^2 \log I_i }{\sigma_i^2 } \\
&\quad\quad+ 2 (S- I_b) \left(2 \sum\limits_{i \in \mathcal{K}^{\prime}} I_i \log I_i + T+S \right), \\
r &= 4 \sigma_b^2 \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{I_i^2 \log^2 I_i }{\sigma_i^2 } - \left( 2 \sum\limits_{i \in \mathcal{K}^{\prime}} I_i \log I_i + T+S \right)^2. \end{split} \end{equation*} Let $\alpha_i(T) = \frac{ (\lambda - 2 \log I_i)}{1 + T/S} $, for $i \in \mathcal{K}^{\prime}$, and substituting $\alpha_i(T)$ into \eqref{proof p1 w_i}. Then, Lemma 2 is proved.
\subsubsection{Proof of Lemma 3}
We now consider condition $\mathcal{C}_4$. For the best design $b$, its budget allocation ratio $W_b(T)$ is always non-negative. As for non-best designs $i \in \mathcal{K}^{\prime}$, let $W_i(T) \geq 0$, and we have \begin{align}
\lambda & \geq 2 \log I_i \quad \forall i \in \mathcal{K}^{\prime}, \notag \\
& \geq 2 \log I_{\langle k - 1 \rangle}. \notag \end{align}
If $w^*_b = 1/2$ \begin{equation*}
\lambda = \frac{4\sum_{i \in \mathcal{K}^{\prime}} I_i \log I_i + T + S}{2\sum_{i \in \mathcal{K}^{\prime}} I_i} \geq 2 \log I_{\langle k - 1 \rangle}, \end{equation*} and it can be checked that $T \geq 4 \sum_{i \in \mathcal{K}^{\prime}} I_i \log(I_{\langle k - 1 \rangle}/I_i) - S $.
If $w^*_b \neq 1/2$ \begin{equation*}
\lambda = \frac{- q + \sqrt{q^2 - 4 p r}}{2 p} \geq 2 \log I_{\langle k - 1 \rangle}, \end{equation*} where $p$, $q$, and $r$ are given in Lemma 2.
Case 1: If $ p = 2w_b^* - 1 > 0$, i.e. $w_b^* > 1/2$, we need to solve the following inequality \begin{equation} \label{proof p2 case1}
\sqrt{q^2 - 4 p r} \geq 4 p \log I_{\langle k - 1 \rangle} + q. \end{equation} Additionally \begin{equation*} \begin{split}
&4 p \log I_{\langle k - 1 \rangle} + q \\
=& 4 \sum\limits_{i \in \mathcal{K}^{\prime}} \left(\frac{\sigma^2_b (w_i^*)^2}{\sigma_i^2} - (1-w_b^*)w_i^* \right) \log \frac{ I_{\langle k - 1 \rangle}}{I_i} \\
&\quad + 2(1-w_b^*)\frac{T+S}{S}.\\ \end{split} \end{equation*} Furthermore, we define \begin{equation*}
T_1 = 2 \sum\limits_{i \in \mathcal{K}^{\prime}} \left[\frac{\sigma^2_b I_i^2}{\sigma_i^2 (S - I_b)} - I_i \right] \log \frac{I_{\langle k - 1 \rangle}}{I_i} - S. \end{equation*} When $T < T_1 $, $4 p \log I_{\langle k - 1 \rangle} + q < 0$, then Inequality \eqref{proof p2 case1} always holds. Otherwise, when $T \geq T_1 $, we take square on both sides of \eqref{proof p2 case1} and obtain \begin{equation*}
T \in (- \infty, T_3] \cup [T_2,+\infty), \end{equation*} in which \begin{equation*} \begin{split}
T_2 &= 2 \left[\sum\limits_{i \in \mathcal{K}^{\prime}} I_i \log \frac{I_{\langle k - 1 \rangle}}{I_i} \right.\\
&\left.\quad\quad+ \sigma_b \sqrt{ \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{ I_i^2 }{\sigma_i^2 } \left( \log \frac{I_{\langle k - 1 \rangle}}{I_i} \right)^2 } \right]- S,\\
T_3 &= 2 \left[\sum\limits_{i \in \mathcal{K}^{\prime}} I_i \log \frac{I_{\langle k - 1 \rangle}}{I_i} \right.\\
&\left. \quad\quad - \sigma_b \sqrt{ \sum\limits_{i \in \mathcal{K}^{\prime}} \frac{ I_i^2 }{\sigma_i^2 } \left( \log \frac{I_{\langle k - 1 \rangle}}{I_i} \right)^2 } \right]- S.\\ \end{split} \end{equation*} It can be checked that $T_2$ is strictly positive. Due to the fact $\lim_{T \rightarrow \infty} W_i(T) \rightarrow w_i^* \geq 0$, for $i \in \mathcal{K}^{\prime}$, we expect that $T$ can be sufficiently large. Therefore, a sufficient condition for $W(T)$ being feasible, when $w_b^* > 1/2$, is $T\geq \max\{T_1,T_2\}$.
Case 2: If $ p = 2w_b^* - 1 < 0$, i.e. $w_b^* < 1/2$, similarly, we need to solve the following inequality \begin{equation} \label{proof p2 case2}
\sqrt{q^2 - 4 p r} \leq 4 p \log I_{\langle k - 1 \rangle} + q. \end{equation} When $T < T_1 $, Inequality \eqref{proof p2 case2} never holds. Otherwise, we take square on both sides of \eqref{proof p2 case2} and, similarly, obtain $T \in (- \infty, T_3] \cup [T_2,+\infty)$. Therefore, a sufficient condition for $W(T)$ being feasible, when $w_b^* < 1/2$, is $T\geq \max\{T_1,T_2\}$.
Hence, the solution $W(T)$ is always feasible if $T\geq T_0$, where \begin{equation*}
T_0 =\left\{\begin{array}{ll} \max\{T_1,T_2\} & \quad \text{if} \ w^*_b \neq \frac{1}{2} \\ 4 \sum_{i \in \mathcal{K}^{\prime}} I_i \log\frac{I_{\langle k - 1 \rangle}}{I_i} - S & \quad \text{if} \ w^*_b = \frac{1}{2} \\ \end{array}\right. \end{equation*} These results conclude the proof.
\subsection{Proof of Theorems}
\subsubsection{Proof of Theorem 1}
Let $\lambda$ and $\nu_i$, for $i \in \mathcal{K}$, be the Lagrange multipliers. The KKT conditions for Problem $\mathcal{P}1$ are as follows \begin{equation} \label{KKT-1}
-\frac{1}{2 \sqrt{2 \pi}} \exp \left(-\frac{\delta_{i,b}^2}{2 \sigma_{i,b}^2}\right) \frac{\delta_{i,b}}{\sigma_{i,b}^3} \frac{\sigma_i^2}{T w_i^2}-\lambda-\nu_i=0, \ i \in \mathcal{K}^{\prime}, \end{equation} \begin{equation} \label{KKT-2}
-\frac{1}{2 \sqrt{2 \pi}} \frac{\sigma_b^2}{T w_b^2}\left[\sum_{i \in \mathcal{K}^{\prime}} \exp \left(-\frac{\delta_{i,b}^2}{2 \sigma_{i,b}^2}\right) \frac{\delta_{i,b}}{\sigma_{i,b}^3}\right] -\lambda-\nu_b = 0, \end{equation} \begin{equation} \label{KKT-3}
\nu_i w_i = 0, \quad i \in \mathcal{K}, \end{equation} \begin{equation} \label{KKT-4}
\sum\limits_{i \in \mathcal{K}} w_i = 1, \end{equation} \begin{equation} \label{KKT-5}
w_i \geq 0, \quad i \in \mathcal{K}. \end{equation} Assuming $w_i$ is strictly positive, thus we have $\nu_i = 0$, for $i \in \mathcal{K}$. Condition $\mathcal{C}_1$ can be derived by substituting \eqref{KKT-1} into \eqref{KKT-2}. Take logarithm on both sides of \eqref{KKT-1}, then condition $\mathcal{C}_2$ follows. Equations \eqref{KKT-4} and \eqref{KKT-5} are conditions $\mathcal{C}_3$ and $\mathcal{C}_4$, respectively, both of which guarantee the feasibility of Problem $\mathcal{P}1$.
\subsection{Proof of Propositions}
\subsubsection{Proof of Proposition 1}
We first show that for any pair of non-best designs $i,j \in \mathcal{K}^{\prime}$ and $i \neq j$, there exists a positive constant $c_{i,j} > 0$ such that $w_i/w_j \leq c_{i,j}$. We prove this by contradiction. Assume that there exists a pair of designs $i,j \in \mathcal{K}^{\prime}$ and $i \neq j$ such that $w_i/w_j$ can not be upper bounded, i.e., $w_i/w_j = \infty$. Since $w_i, w_j \in [0,1]$, it can be checked that $w_j \rightarrow 0$. By condition $\mathcal{C}_2$ in Theorem 1 \begin{equation} \label{proof of proposition verifying: eqn 1} \begin{split}
&\left[\frac{\delta_{j,b}^2}{ (\sigma_j^2/w_j + \sigma_b^2/w_b) } - \frac{\delta_{i,b}^2}{ (\sigma_i^2/w_i + \sigma_b^2/w_b) } \right] \frac{T}{2} \\
& \quad \quad = \log \frac{\delta_{j,b} (\sigma_j^2/w^2_j) (\sigma_i^2/w_i + \sigma_b^2/w_b)^{\frac{3}{2}}}{\delta_{i,b} (\sigma_i^2/w^2_i) (\sigma_j^2/w_j + \sigma_b^2/w_b)^{\frac{3}{2}} }. \end{split} \end{equation} As $w_j \rightarrow 0$, the term $\frac{\delta_{j,b}^2}{ (\sigma_j^2/w_j + \sigma_b^2/w_b) }$ in \eqref{proof of proposition verifying: eqn 1} will vanish, and the right-hand side in \eqref{proof of proposition verifying: eqn 1} will approach infinity. Then, as $w_j \rightarrow 0$, we have \begin{equation*}
- \frac{\delta_{i,b}^2}{ 2(\sigma_i^2/w_i + \sigma_b^2/w_b) } T
= + \infty, \end{equation*} which is true if $\sigma_i^2/w_i + \sigma_b^2/w_b \rightarrow 0^-$. However, this contradicts with the fact $\sigma_i^2/w_i + \sigma_b^2/w_b > 0$. Thus, the assumption is false, and for any pair of designs $i,j \in \mathcal{K}^{\prime}$ and $i \neq j$, there must exist a positive constant $c_{i,j} > 0$, such that $w_i/w_j \leq c_{i,j}$.
Let $w_{\text{min}}$ be the minimum budget allocation ratio of non-best designs, i.e., $w_{\text{min}} = \min_{i \in \mathcal{K}^{\prime}} w_i$. Therefore, there exists a positive constant $c$ such that $w_i \leq c w_{\text{min}}$, for $i \in \mathcal{K}^{\prime}$. By condition $\mathcal{C}_1$ in Theorem 1 \begin{equation*}
w_b = \sigma_b \sqrt{\sum\limits_{i \in \mathcal{K}^{\prime}} \frac{w_i^{2}}{\sigma_i^2}} \geq \frac{ \sigma_b\sqrt{k-1} }{\bar{\sigma}} w_{\text{min}}. \end{equation*} Then, for any non-best design $i \in \mathcal{K}^{\prime}$, we have \begin{equation*}
\frac{w_i}{w_b} \leq \frac{w_i}{w_{\text{min}}} \frac{\bar{\sigma}}{\sigma_b \sqrt{k-1} } \leq \frac{\bar{\sigma}}{\underline{\sigma}} \frac{c}{\sqrt{k-1}}. \end{equation*} This result concludes the proof, and therefore, $w_i/w_b = \mathcal{O}(1/\sqrt{k})$, for $i \in \mathcal{K}^{\prime}$.
\subsubsection{Proof of Proposition 2}
According to Lemma 2, we have \begin{equation*}
\alpha_i(T) = \frac{ (\lambda - 2 \log I_i)}{1 + T/S}, \quad i \in \mathcal{K}^{\prime}, \end{equation*} which is monotone decreasing with respect to $I_i$. Since $I_{\langle 1 \rangle} \leq I_{\langle 2 \rangle} \leq \dots \leq I_{\langle k - 1 \rangle}$, we have $\alpha_{\langle 1 \rangle}(T) \geq \alpha_{\langle 2 \rangle}(T) \geq \dots \geq \alpha_{\langle k - 1 \rangle}(T)$. Furthermore, if $I_{i}$, for $i \in \mathcal{K}^{\prime}$, are all equal, then $\lambda = 1 + T/S + 2 \log I_i$ and $\alpha_{\langle 1 \rangle}(T) = \alpha_{\langle 2 \rangle}^(T) = \dots = \alpha_{\langle k - 1 \rangle}(T) = 1$. On the contrary, if $\alpha_{\langle 1 \rangle}(T) = \alpha_{\langle 2 \rangle}(T) = \dots = \alpha_{\langle k - 1 \rangle}(T)$, it can be checked that all $I_i$s are equal.
We show $\alpha_{\langle 1 \rangle}(T) \geq 1 $ by contradiction, and $\alpha_{\langle k - 1 \rangle}(T) \leq 1$ can be proved similarly. Without lose of generality, we assume $\alpha_{\langle 1 \rangle}(T) < 1$. Based on preceding analyses, we know $\alpha_{\langle k - 1 \rangle}(T) \leq \alpha_{[k-2]}(T) \leq \dots \leq \alpha_{\langle 1 \rangle}(T) < 1$. According to Lemma 2, we have $W_i(T) = w^*_i \alpha_{i}(T) < w^*_i $ for non-best designs $i \in \mathcal{K}^{\prime}$, and $W_b(T) = \sigma_b \sqrt{\sum_{i \in \mathcal{K}^{\prime}}\frac{(w_i^* \alpha_{i}(T) )^{2} }{\sigma_i^2}} < w^*_b$ for the best design $b$. Then we have $\sum_{i \in \mathcal{K}} W_b(T) < \sum_{i \in \mathcal{K}} w^*_b = 1$, which contradicts with the fact $\sum_{i \in \mathcal{K}} W_b(T) = 1$. Therefore, $\alpha_{\langle 1 \rangle}(T)$ must be larger or equal to 1. These results conclude the proof. \endgroup
\section*{References}
\end{document}
|
arXiv
|
{
"id": "2304.02377.tex",
"language_detection_score": 0.7411515116691589,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{The Weighted Barycenter Drawing Recognition Problem} \author{Peter Eades\inst{1} \and Patrick Healy\inst{2} \and Nikola S. Nikolov\inst{2}}
\institute{ University of Sydney, \email{[email protected]} \and University of Limerick \email{patrick.healy,[email protected]}}
\maketitle
\begin{abstract}
We consider the question of whether a given graph drawing $\Gamma$ of a triconnected planar graph $G$ is a weighted barycenter drawing. We answer the question with an elegant arithmetic characterisation using the faces of $\Gamma$. This leads to positive answers when the graph is a Halin graph,
and to a polynomial time recognition algorithm when the graph is cubic. \end{abstract}
\section{Introduction}\label{se:intro}
The \emph{barycenter algorithm} of Tutte~\cite{tutte60,tutte63} is one of the earliest and most elegant of all graph drawing methods. It takes as input a graph $G=(V,E)$, a subgraph $F_0 = (V_0, E_0) $ of $G$, and a position $\gamma_a$ for each $a \in V_0$. The algorithm simply places each vertex $v \in V-V_0$ at the barycenter of the positions of its neighbours. The algorithm can be seen as the grandfather of force-directed graph drawing algorithms, and can be implemented easily by solving a system of linear equations. If $G$ is a \iffalse Nik: replaced plane by planar \fi planar triconnected graph, $F_0$ is the outside face of $G$, and the positions $\gamma_a$ for $a \in V_0$ are chosen so that $F_0$ forms a convex polygon, then the drawing output by the barycenter algorithm is planar and each face is convex.
The barycenter algorithm can be generalised to \iffalse Nik: added planar \fi planar graphs with positive edge weights, placing each vertex $i$ of $V-V_0$ at the weighted barycenter of the neighbours of $i$. This generalisation preserves the property that the output is planar and convex~\cite{DBLP:journals/cagd/Floater97a}. Further, weighted barycenter methods have been used in a variety of theoretical and practical contexts~\cite{DBLP:conf/gd/FraysseixM03,DBLP:journals/comgeo/VerdierePV03,DBLP:journals/corr/abs-0708-0964,Thomassen83}. Examples of weighted barycenter drawings (the same graph with different weights) are in Fig.~\ref{fi:example}.
In this paper we investigate the following question: given a straight-line \iffalse Nik: added planar \fi planar drawing $\Gamma$ of a triconnected planar graph $G$, can we compute weights for the edges of $G$ so that $\Gamma$ is the weighted barycenter drawing of $G$? We answer the question with an elegant arithmetic characterisation, using the faces of $\Gamma$. This yields positive answers when the graph is a Halin graph,
and leads to a polynomial time algorithm when the graph is cubic.
Our motivation in examining this question partly lies in the elegance of the mathematics, but it was also posed to us by Veronika Irvine (see \cite{DBLP:conf/gd/BiedlI17,tesselace}), who needed the characterisation to to create and classify ``grounds'' for bobbin lace drawings; this paper is a first step in this direction. Further, we note that our result relates to the problem of morphing from one planar graph drawing to another (see \cite{DBLP:conf/gd/Barrera-CruzHL14,FloaterGotsman}). Previous work has characterised drawings that arise from the Schnyder algorithm (see~\cite{DBLP:conf/wg/BonichonGHI10}) in this context. Finally, we note that this paper is the first attempt to characterise drawings that are obtained from force-directed methods. \begin{figure}
\caption{ Weighted barycenter drawings of the same graph embedding with different weights.
}
\label{fi:example}
\end{figure}
\section{Preliminaries: the weighted barycenter algorithm}\label{se:prelim}
Suppose that $G = (V,E)$ denotes a triconnected \iffalse Nik: changed plane to planar \fi planar graph and $w$ is a
\emph{weight function} that assigns a non-negative real weight $w_{ij}$ to each edge $(i,j) \in E$. We assume that the weights are positive unless otherwise stated. We denote $|V|$ by $n$ and $|E|$ by $m$. In this paper we discuss planar straight-line drawings of such graphs; such a drawing $\Gamma$ is specified by a position $\gamma_i$ for each vertex $i \in V$. We say that $\Gamma$ is \emph{convex} if every face is a convex polygon.
Throughout this paper, $F_0$ denotes the \iffalse Nik: changed outside to outer \fi outer face of a plane graph $G$. Denote the number of vertices on $F_0$ by $f_0$. In a convex drawing, the edges of $F_0$ form a simple convex polygon $P_0$. Some terminology is convenient: we say that an edge or vertex on $F_0$ is \emph{external}; a vertex that is not external is \emph{internal}; a face $F$ (respectively edge, $e$) is \emph{internal} if $F$ (resp. $e$) is incident to an internal vertex, and \emph{strictly internal} if every vertex incident to $F$ (resp. $e$) is internal.
The \emph{weighted barycenter algorithm} takes as input a triconnected \iffalse Nik: changed plane to planar \fi planar graph $G = (V,E)$ with a weight function $w$, together with $F_0$ and $P_0$, and produces a straight-line drawing $\Gamma$ of $G$ with $F_0$ drawn as $P_0$. Specifically, it assigns a position $\gamma_i$ to each internal vertex $i$ such that $\gamma_i$ is the weighted barycenter of its neighbours in $G$. That is: \begin{equation} \label{eq:bary0} \gamma_i = \frac{1}{\sum_{j \in N(i)} w_{ij}} \sum_{j \in N(i)} w_{ij} \gamma_j \end{equation} for each internal vertex $i$. Here $N(i)$ denotes the set of neighbours of $i$. If $\gamma_i = (x_i, y_i)$ then (\ref{eq:bary0}) consists of $2(n-f_0)$ linear equations in the $2(n-f_0)$ unknowns $x_i , y_i$. The equations (\ref{eq:bary0}) are called the \emph{(weighted) barycenter equations} for $G$. Noting that the matrix involved is a submatrix of the Laplacian of $G$, one can show that the equations have a unique solution that can be found by traditional (see for example~\cite{Trefethen97}) or specialised (see for example~\cite{DBLP:journals/siamcomp/SpielmanT11}) methods.
The weighted barycenter algorithm, \iffalse Nik: added a phrase \fi which can be viewed as a force directed method, was defined by Tutte~\cite{tutte60,tutte63} and extended by Floater~\cite{DBLP:journals/cagd/Floater97a}; the classic theorem says that the output is planar and convex: \begin{theorem} (Tutte~\cite{tutte60,tutte63}, Floater~\cite{DBLP:journals/cagd/Floater97a}) The drawing output by the weighted barycenter algorithm is planar, and each face is convex. \label{th:tutte} \end{theorem}
\section{The Weighted Barycenter Recognition Problem} \label{se:TheProblemt}
This paper discusses the problem of finding weights $w_{ij}$ so that a given drawing is the weighted barycenter drawing with these weights. More precisely, we say that a drawing $\Gamma$ is a \emph{weighted barycenter drawing} if there is a positive weight $w_{ij}$ for each internal edge $(i,j)$ such that for each internal vertex $i$, equations (\ref{eq:bary0}) hold. \begin{description} \item [{\bf The Weighted Barycenter Recognition problem}] \item [{\em Input:}] A straight-line planar drawing $\Gamma$ of a triconnected plane graph $G = (V,E)$, such that the vertices on the convex hull of $\{ \gamma_i : i \in V \}$ form a face of $G$. \item [{\em Question:}] Is $\Gamma$ a weighted barycenter drawing? \end{description} Thus we are given the location $\gamma_i = (x_i , y_i)$ of each vertex, and we must compute a positive weight $w_{ij}$ for each edge so that the barycenter equations (\ref{eq:bary0}) hold for each internal vertex.
Theorem~\ref{th:tutte} implies that if $\Gamma$ is a weighted barycenter drawing, then each face of the drawing is convex; however, the converse is false, even for triangulations (see Appendix).
\section{Linear Equations for the Weighted Barycenter Recognition problem}
In this section we show that the weighted barycenter recognition problem can be expressed in terms of linear equations. The equations use \emph{asymmetric} weights $z_{ij}$ for each edge $(i,j)$; that is, $z_{ij}$ is not necessarily the same as $z_{ji}$. To model this asymmetry we replace each undirected edge $(i,j)$ of $G$ with two directed edges $(i,j)$ and $(j,i)$; this gives a directed graph $\overrightarrow{G} = (V,\overrightarrow{E})$. For each vertex $i$, let $N^+(i)$ denote the set of \emph{out-neighbours} of $i$; that is, $N^+(i) = \{j \in V : (i,j) \in \overrightarrow{E} \}.$
Since each face is convex, each internal vertex is inside the convex hull of its neighbours. Thus each internal vertex position is a convex linear combination of the vertex positions of its neighbours. That is, for each internal vertex $i$ there are non-negative weights $z_{ij}$ such that \begin{equation} \label{eq:bary3}
\sum_{j \in N^+(i)} z_{ij} = 1 \text{~~~~and~~~~} \gamma_i = \sum_{j \in N^+(i)} z_{ij} \gamma_j . \end{equation}
The values of $z_{ij}$ satisfying (\ref{eq:bary3}) can be determined in linear time. For a specific vertex $i$, the $z_{ij}$ for $j \in N^+(i)$ can be viewed as a kind of \emph{barycentric coordinates} for $i$. In the case that $|N^+(i)| = 3$, these coordinates are unique.
Although equations (\ref{eq:bary0}) and (\ref{eq:bary3}) seem similar, they are not the same: one is directed, the other is undirected. In general $z_{ij} \neq z_{ji}$ for directed edges $(i,j)$ and $(j,i)$, while the weights $w_{ij}$ satisfy $w_{ij} = w_{ji}$. However we can choose a ``scale factor'' $s_i > 0$ for each vertex $i$, and scale equations (\ref{eq:bary3}) by $s_i$. That is, for each internal vertex $i$, \begin{equation} \label{eq:baryScaled} \gamma_i = \frac{1}{ \sum_{j \in N^+(i)} s_i z_{ij} } \sum_{j \in N^+(i)} s_i z_{ij} \gamma_j . \end{equation} The effect of this scaling is that we replace $z_{ij}$ by $s_i z_{ij}$ for each edge $(i,j)$.
We would like to choose a scale factor $s_i > 0$ for each internal vertex $i$ such that for each strictly internal edge $(i,j) \in E$, $ s_i z_{ij} = s_j z_{ji}$; that is, we want to find a real positive $s_i$ for each internal vertex $i$ such that \begin{equation} \label{eq:scale} s_i z_{ij} - s_j z_{ji} = 0 \end{equation} for each strictly internal edge $(i,j)$.
It can be shown easily that the existence of any nontrivial solution to (\ref{eq:scale}) implies the existence of a positive solution (see Appendix).
We note that any solution of (\ref{eq:scale}) for strictly internal edges gives weights $w_{ij}$ such that the barycenter equations (\ref{eq:bary0}) hold. We choose $w_{ij} = s_i z_{ij}$ for each (directed) edge $(i,j)$ that is incident to an internal vertex $i$. Equations (\ref{eq:scale}) ensure that $w_{ij} = w_{ji}$ for each strictly internal edge. For edges which are internal but not strictly internal, we can simply choose $w_{ij} = s_i z_{ij}$ for any value of $s_i$, since $z_{ji}$ is undefined.
Thus if equations (\ref{eq:scale}) have a nontrivial solution, then the drawing is a weighted barycenter drawing.
\subsubsection{The main theorem.}
We characterise the solutions of equations (\ref{eq:scale}) with an arithmetic condition on the faces of $\Gamma$. This considers the product of the weights $z_{ij}$ around directed cycles in $G$: if the product around each strictly internal face in the clockwise direction is the same as the product in the counter-clockwise direction, then equations (\ref{eq:scale}) have a nontrivial solution. \begin{theorem} \label{th:cycleProduct} Equations (\ref{eq:scale}) have a nontrivial solution if and only if for each strictly internal face $C = (v_0, v_1, \ldots , v_{k-1}, v_k = v_0)$ in $G$, we have \begin{equation} \label{eq:cycle} \prod_{i=0}^{k-1} z_{v_i,v_{i+1}} = \prod_{i=1}^{k} z_{v_{i},v_{i-1}}. \end{equation} \end{theorem} \begin{proof}
For convenience we denote $\frac{z_{ji}}{z_{ij}}$ by $\zeta_{ij}$ for each directed edge $(i,j)$; note that $\zeta_{ij} = 1/\zeta_{ji}$. Equations (\ref{eq:scale}) can be re-stated as \begin{equation} \label{eq:zeta} s_i - \zeta_{ij} s_j = 0 \end{equation} for each strictly internal edge $(i,j)$, and the equations (\ref{eq:cycle}) for cycle $C$ can be re-stated as \begin{equation} \label{eq:cycleZeta} \prod_{i=0}^{k-1} \zeta_{v_i,v_{i+1}} = 1. \end{equation} First suppose that equations (\ref{eq:zeta}) have nontrivial solutions $s_i$ for all internal vertices $i$, and $C = (v_0, v_1, \ldots , v_{k-1}, v_k = v_0)$ is a strictly internal face in $G$. Now applying (\ref{eq:zeta}) around $C$ clockwise beginning at $v_0$, we can have: $$ s_{v_0} = \zeta_{v_0,v_1} s_{v_1} ~ = ~ \zeta_{v_0,v_1} \zeta_{v_1,v_2} s_{v_2} ~ = ~ \zeta_{v_0,v_1} \zeta_{v_1,v_2} \zeta_{v_2,v_3} s_{v_3} ~ = ~\ldots $$ We can deduce that \begin{equation*}
s_{v_0} = \left( \prod_{i=0}^{j-1} \zeta_{v_i,v_{i+1}} \right) s_{v_j} = \left( \prod_{i=0}^{k-1} \zeta_{v_i,v_{i+1}} \right) s_{v_k} = \left( \prod_{i=0}^{k-1} \zeta_{v_i,v_{i+1}} \right) s_{v_0} \end{equation*} and this yields equation (\ref{eq:cycleZeta}).
Now suppose that equation (\ref{eq:cycleZeta}) holds for every strictly internal facial cycle of $G$. We first show that equation (\ref{eq:cycleZeta}) holds for \emph{every} strictly internal cycle. Suppose that (\ref{eq:cycleZeta}) holds for two cycles $C_1$ and $C_2$ that share a single edge, $(u,v)$, and let $C_3$ be the sum of $C_1$ and $C_2$ (that is, $C_3 = ( C_1 \cup C_2 ) - \{ (u,v) \}$).
Now traversing $C_3$ in clockwise order gives the clockwise edges of $C_1$ (omitting $(u,v)$) followed by the clockwise edges of $C_2$ (omitting $(v,u)$). But from equation (\ref{eq:cycleZeta}), the product of the edge weights $\zeta_{ij}$ in the clockwise order around $C_1$ is one, and the product of the edge weights $\zeta_{i'j'}$ in the clockwise order around $C_2$ is one. Thus the product of the edge weights $\zeta_{ij}$ in clockwise order around $C_3$ is $\frac{1}{\zeta_{uv} \zeta_{vu} } = 1$.
That is, (\ref{eq:cycleZeta}) holds for $C_3$. Since the facial cycles form a cycle basis, it follows that (\ref{eq:cycleZeta}) holds for every cycle.
Now choose a reference vertex $r$, and consider a depth first search tree $T$ rooted at $r$.
Denote the set of directed edges on the directed path in $T$ from $i$ to $j$ by $E_{ij}$. Let $s_{r} = 1$, and for each internal vertex $i \neq r$, let \begin{equation} \label{eq:s} s_{i} = \prod_{(u,v) \in E_{ri}} \zeta_{uv}. \end{equation} Clearly equation~(\ref{eq:zeta}) holds for every edge of $T$. Now consider a back-edge $(i,j)$ for $T$ (that is, a strictly internal edge of $G$ that is not in $T$), and let $k$ denote the least common ancestor of $i$ and $j$ in $T$. Then from (\ref{eq:s}) we can deduce that \begin{equation} \label{eq:moo1} \frac{s_i}{s_j} = \frac{ \prod_{(u,v) \in E_{ri}} \zeta_{uv} }{ \prod_{(u',v')\in E_{rj}} \zeta_{u'v'} } = \frac{ \prod_{(u,v) \in E_{ki}} \zeta_{uv} }{ \prod_{(u',v')\in E_{kj}} \zeta_{u'v'} }. \end{equation} Now let $C$ be the cycle in $\Gamma$ that consists of the reverse of the directed path in $T$ from $k$ to $j$, followed by the directed path in $T$ from $k$ to $i$, followed by the edge $(i,j)$. Since equation~(\ref{eq:cycleZeta}) holds for $C$, we have: \begin{equation} \label{eq:moo2} 1 = \left( \prod_{(v',u') \in E_{jk}} \zeta_{v'u'} \right) \left( \prod_{(u,v) \in E_{ki}} \zeta_{uv} \right) \zeta_{ij}
= \left( \frac{ \prod_{(u,v) \in E_{ki}} \zeta_{uv} } { \prod_{(u'v') \in E_{kj}} {\zeta_{u'v'}} } \right) \zeta_{ij} \end{equation} Combining equations (\ref{eq:moo1}) and (\ref{eq:moo2}) we have $s_i = \zeta_{ij} s_j$ and so equation (\ref{eq:zeta}) holds for each back edge $(i,j)$. We can conclude that (\ref{eq:zeta}) holds for all strictly internal edges. \qed
\end{proof}
\section{Applications}
We list some implications of Theorem~\ref{th:cycleProduct} for cubic, Halin~\cite{DBLP:journals/jgaa/Eppstein16} and planar graphs with degree larger than three. Proofs of the corollaries below are straightforward.
\begin{corollary} \label{cor:cubic1} A drawing $\Gamma$ of a cubic graph is a weighted barycenter drawing if and only if equations (\ref{eq:scale}) have rank smaller than $n-f_0$. \qed \end{corollary}
\begin{corollary} \label{cor:cubic2} For cubic graphs, there is a linear time algorithm for the weighted barycenter recognition problem. \qed \end{corollary}
For cubic graphs, the weights $z_{ij}$ are unique, and thus equations (\ref{eq:scale}) give a complete characterisation of weighted barycenter drawings. One can use Theorem~\ref{th:cycleProduct} to test whether a solution of equations (\ref{eq:scale}) exists, checking equations~(\ref{eq:cycle}) in linear time.
\begin{corollary} \label{cor:halin} Suppose that $\Gamma$ is a convex drawing of a Halin graph such that the internal edges form a tree. Then $\Gamma$ is a weighted barycenter drawing. \qed \end{corollary}
\subsubsection{Graphs with degree larger than three.}
For a vertex $i$ of degree $d_i > 3$, solutions for equations (\ref{eq:bary3}) are not unique. Nevertheless, these equations are linear, and we have 3 equations in $d_i$ variables.
Thus, for each vertex $i$, the solution $z_{ij}, j \in N(i)$, form a linear space of dimension at most $d_i-3$. In this general case, we have: \begin{corollary} \label{co:general} A drawing $\Gamma$ of a graph $G$ is a weighted barycenter drawing if and only if there are solutions $z_{ij}$ to equations (\ref{eq:bary3}) such that the cycle equation (\ref{eq:cycle}) holds for every internal face. \qed \end{corollary}
Although Corollary~\ref{co:general} is quite elegant, it does not lead to an immediately practical algorithm because the equations (\ref{eq:cycle}) are not linear.
\section{Conclusion}\label{se:conclusions}
Force-directed algorithms are very common in practice, and drawings obtained from force-directed methods are instantly recognisable to most researchers in Graph Drawing. However, this paper represents the first attempt to give algorithms to recognise the output of a particular force-directed method, namely the weighted barycenter method. It would be interesting to know if the results of other force-directed methods can be automatically recognised.
\subsubsection*{Acknowledgements.} We wish to thank Veronika Irvine for motivating discussions.
\pagebreak
\section*{Appendix} \ \subsection*{A triangulation which is not a weighted barycenter drawing.}
The weighted barycenter algorithm can be viewed as a force directed method, as follows. We define the \emph{energy} $\eta(i,j)$ of an internal edge $(i,j)$ by \begin{equation} \label{eq:energyEdge}
\eta(i,j) = \frac{1}{2} w_{ij} \delta (\gamma_i,\gamma_j)^2 = \frac{1}{2} w_{ij} \left( (x_i - x_j)^2 + (y_i-y_j)^2 \right)\\ \end{equation} where $\delta(,)$ is the Euclidean distance and $\gamma_i = (x_i , y_i)$. The energy $\eta(\Gamma)$ in the whole drawing is the sum of the internal edge energies. Taking partial derivatives with respect to each variable $x_i$ and $y_i$ reveals that $\eta(\Gamma)$ is minimised precisely when the barycenter equations (\ref{eq:bary0}) hold.
\begin{figure}
\caption{
A triangulation which is not a weighted barycenter drawing.
}
\label{fi:triangles3}
\end{figure} \begin{lemma} The drawing in Fig.~\ref{fi:triangles3} is not a weighted barycenter drawing. \end{lemma} \begin{proof} Suppose that the drawing $\Gamma$ in Fig.~\ref{fi:triangles3} is a weighted barycenter drawing with weights $w_{ij}$. The total energy $\eta(\Gamma)$ in the drawing is given by summing equation (\ref{eq:energyEdge}) over all internal edges, and the drawing $\Gamma$ minimises $\eta(\Gamma)$. Further, the minimum energy drawing is unique. Consider the drawing $\Gamma'$ of this graph where the inner triangle is rotated clockwise by $\epsilon$, where $\epsilon$ is small. The strictly internal edges remain the same length, while the edges between the inner and outer triangles become shorter. Thus, since every $w_{ij} > 0$ for every such $(i,j)$, $\eta(\Gamma') <\eta(\Gamma)$. This contradicts the fact that $\eta$ is minimised at $\Gamma$. \qed \end{proof}
\subsection*{Positive solutions for equations (\ref{eq:scale}).}
\begin{lemma} If equations (\ref{eq:scale}) have a nontrivial solution, then they have a nontrivial solution in which every $s_i$ is positive. \end{lemma} \begin{proof} Suppose that the vector $s$ is a solution to equations (\ref{eq:scale}), and $s_i \leq 0$ for some internal vertex $i$. Since $z_{ij} > 0$ for each $i,j$, and $G$ is connected, it is easy to deduce from (\ref{eq:scale}) that $s_j \leq 0$ for every internal vertex $j$. Further if $s_i = 0$ for some internal vertex $i$ then $s_j = 0$ for every internal vertex. Noting that $s$ is a solution to (\ref{eq:scale}) if and only if $-s$ is a solution, the Lemma follows. \qed \end{proof}
\subsection*{Properties of the coefficient matrix of the equations for the scale factors.}
Equations (\ref{eq:scale}) form a set of $m-f_0$ equations in the $n-f_0$ unknowns $s_i$. We can write (\ref{eq:scale}) as \begin{equation} \label{eq:B} B^T s=0 \end{equation} where $s$ is an $(n-f_0) \times 1$ vector and $B$ is an $(n-f_0) \times (m-f_0)$ matrix; precisely: \[ B_{ie} = \begin{cases} z_{ij} & \text{~if~} e \text{~is the edge~} (i,j) \in \overrightarrow{E}\\ -z_{ij} & \text{~if~}e \text{~is the edge~} (j,i) \in \overrightarrow{E}\\ 0 & \text{otherwise.} \end{cases} \] Note that $B$ is a weighted version of the directed incidence matrix of the graph $\overrightarrow{G}$. Adapting a classical result for incidence matrices yields a lower bound on the rank of $B$: \begin{lemma} The rank of $B$ is at least $n - f_0 - 1$. \end{lemma} \begin{proof} Since $G$ is triconnected, the induced subgraph of the internal vertices is connected. Consider a submatrix of $B$ consisting of rows that correspond to the edges of a tree that spans the internal vertices. It is easy to see that this $(n-f_0-1) \times (n-f_0) $ submatrix has full row rank (note that it has a column with precisely one nonzero entry). The lemma follows. \qed \end{proof} In fact, one can show (using the same method as in the proof of Theorem~\ref{th:cycleProduct}) that $B$ has rank exactly $n - f_0 - 1$ as long as the equations (\ref{eq:cycle}) hold for every strictly internal cycle.
\subsection*{Proofs of the corollaries.}
{\bf Corollary}~\ref{cor:cubic1} \begin{proof} In the case of a cubic graph, the weights $z_{ij}$ are unique. Thus, if the only solution to (\ref{eq:scale}) is $s_i = 0$ for every $i$, then the drawing is not a weighted barycenter drawing. \qed \end{proof}
{\bf Corollary}~\ref{cor:halin} \begin{proof}
A \emph{Halin graph}~\cite{DBLP:journals/jgaa/Eppstein16} \iffalse Nik: added a phrase \fi is a triconnected graph that consists of a tree, none of the vertices of which has exactly two neighbours, together with a cycle connecting the leaves of the tree. The cycle connects the leaves in an order so that the resulting graph is planar. A Halin graph is typically drawn so that the outer face is the cycle.
This corollary can be deduced immediately from Theorem~\ref{th:cycleProduct} since such a graph has no strictly internal cycles. More directly, we can solve the equations (\ref{eq:scale}) starting by assigning $s_r = 1$ for the root $r$, and adding one edge at a time. \qed \end{proof}
\end{document}
|
arXiv
|
{
"id": "1809.00628.tex",
"language_detection_score": 0.787733793258667,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[Plurisubharmonic defining functions]{A note on plurisubharmonic defining functions in $\mathbb{C}^{n}$} \author{J. E. Forn\ae ss, A.-K. Herbig} \subjclass[2000]{32T35, 32U05, 32U10} \keywords{Plurisubharmonic defining functions, Stein neighborhood basis, DF exponent} \thanks{Research of the first author was partially supported by an NSF grant.} \thanks{Research of the second author was supported by FWF grant P19147} \address{Department of Mathematics, \newline University of Michigan, Ann Arbor, Michigan 48109, USA} \email{[email protected]} \address{Department of Mathematics, \newline University of Vienna, Vienna, A-1090, Austria} \email{[email protected]} \date{} \begin{abstract}
Let $\Omega\subset\subset\mathbb{C}^{n}$, $n\geq 3$, be a smoothly bounded domain. Suppose that
$\Omega$ admits
a smooth defining function which is plurisubharmonic on the boundary of $\Omega$. Then the
Diederich--Forn\ae ss exponent can be chosen arbitrarily close to $1$, and the closure of $\Omega$
admits a Stein neighborhood basis. \end{abstract} \maketitle
\section{Introduction}
Let $\Omega\subset\subset\mathbb{C}^{n}$ be a smoothly bounded domain. Throughout, we suppose
that $\Omega$ admits a $\mathcal{C}^{\infty}$-smooth defining function $\rho$ which is
plurisubharmonic on the boundary, $b\Omega$, of $\Omega$. That is,
\begin{align}\label{E:psh}
H_{\rho}(\xi,\xi)(z):=\sum_{j,k=1}^{n}\frac{\partial^{2}\rho}{\partial z_{j}\partial\overline{z}_{k}}(z)
\xi_{j}\overline{\xi}_{k}\geq 0\;\;\text{for all}\;z\in b\Omega,\;\xi\in\mathbb{C}^{n}.
\end{align}
The question we are concerned with is what condition \eqref{E:psh} tells us about the
behaviour of the complex Hessian of $\rho$ - or of some other defining function of $\Omega$ -
away from the boundary of $\Omega$.
That $\rho$ is not necessarily plurisubharmonic in any neighborhood of $b\Omega$ can be seen
easily, for an example see Section 2.3 in \cite{For-Her}. In \cite{For-Her}, we showed that if $n=2$,
then for any $\epsilon>0$, $K>0$ there exist smooth defining functions
$\rho_{i}$, $i=1,2$, and a neighborhood $U$ of $b\Omega$ such that
\begin{align}\label{E:dim2result}
H_{\rho_{i}}(\xi,\xi)(q_{i})\geq-\epsilon|\rho_{i}(q_{i})|\cdot|\xi|^{2}
+K|\langle\partial\rho_{i}(q_{i}),\xi\rangle|^{2}
\end{align}
for all $\xi\in\mathbb{C}^{2}$ and $q_{1}\in\overline{\Omega}\cap U$,
$q_{2}\in\Omega^{c}\cap U$.
The estimates \eqref{E:dim2result} imply the existence of particular exhaustion functions for
$\Omega$ and the complement of $\overline{\Omega}$,
which is not a direct consequence of \eqref{E:psh}.
A Diederich--Forn\ae ss
exponent of a domain
is a number $\tau\in(0,1]$ for which there exists a smooth defining function $s$ such
that $-(-s)^{\tau}$ is strictly plurisubharmonic in the domain.
It was shown in \cite{DF1,Ran} that all smoothly
bounded, pseudoconvex domains have a Diederich--Forn\ae ss exponent. However, it is also known
that there are smoothly bounded, pseudoconvex domains for which the largest
Diederich--Forn\ae ss exponent has to be chosen arbitrarily close to $0$ (see \cite{DF2}). In
\cite{For-Her}, we showed that \eqref{E:dim2result}, $i=1$, implies that the Diederich--Forn\ae ss
exponent can be chosen arbitrarily close to $1$. We also showed that
\eqref{E:dim2result}, $i=2$, yields that the complement of $\Omega$ can be exhausted by bounded,
strictly plurisubharmonic functions. In particular, the closure of $\Omega$ admits a Stein
neighborhood basis.
For $n\geq 3$ we obtain the following:
\begin{theorem}\label{T:Main}
Let $\Omega\subset\subset\mathbb{C}^{n}$ be a smoothly bounded domain. Suppose that
$\Omega$ has a smooth defining function which is plurisubharmonic on the boundary of
$\Omega$. Then for any $\epsilon>0$ there exist a neighborhood $U$ of $b\Omega$ and smooth
defining functions $r_{1}$ and $r_{2}$ such that
\begin{align}\label{E:Main1}
H_{r_{1}}(\xi,\xi)(q)\geq -\epsilon
\left[|r_{1}(q)|\cdot|\xi|^{2}+\frac{1}{|r_{1}(q)|}\cdot\left|\langle\partial r_{1}(q),\xi\rangle\right|^{2}\right]
\end{align}
holds for all $q\in\Omega\cap U$, $\xi\in\mathbb{C}^{n}$, and
\begin{align}\label{E:Main2}
H_{r_{2}}(\xi,\xi)(q)\geq
-\epsilon
\left[r_{2}(q)\cdot|\xi|^{2}+\frac{1}{r_{2}(q)}\cdot\left|\langle\partial r_{2}(q),\xi\rangle\right|^{2}\right]
\end{align}
holds for all $q\in(\overline{\Omega})^{c}\cap U$, $\xi\in\mathbb{C}^{n}$. \end{theorem} Let us remark that our proof of Theorem \ref{T:Main} also works when $n=2$. However, the results of Theorem \ref{T:Main} are weaker than \eqref{E:dim2result}. Nevertheless, they are still strong enough to obtain that the Diederich--Forn\ae ss exponent can be chosen arbitrarily close to 1 and that the closure of the domain admits a Stein neighborhood basis. In particular, we have the following:
\begin{corollary}\label{C:DF}
Assume the hypotheses of Theorem \ref{T:Main} hold. Then
\begin{enumerate}
\item for all $\eta\in (0,1)$ there exists a smooth defining function $\tilde{r}_{1}$ of $\Omega$ such that
$-(-\tilde{r}_{1})^{\eta}$ is strictly plurisubharmonic on $\Omega$,
\item for all $\eta>1$ there exist a smooth defining function $\tilde{r}_{2}$ of $\Omega$ and a
neighborhood $U$ of $\overline{\Omega}$ such that $\tilde{r}_{2}^{\eta}$ is strictly plurisubharmonic
on $(\overline{\Omega})^{c}\cap U$.
\end{enumerate} \end{corollary}
We note that in \cite{DF3} it was proved that (i) and (ii) of Corollary \ref{C:DF} hold for so-called regular domains. Furthermore, in \cite{DF4} it was shown that pseudoconvex domains with real-analytic boundary are regular domains.
This article is structured as follows. In Section \ref{S:prelim}, we give the setting and define our basic notions. Furthermore, we show in this section which piece of the complex Hessian of $\rho$ at a given point $p$ in $b\Omega$ constitutes an obstruction for inequality \eqref{E:Main1} to hold for a given $\epsilon>0$. In Section \ref{S:modification}, we construct a local defining function which does not possess this obstruction term to \eqref{E:Main1} at a given boundary point $p$. Since this fixes our problem with \eqref{E:Main1} only at this point $p$ (and at nearby boundary points at which the Levi form is of the same rank as at $p$), we will need to patch the newly constructed local defining functions without letting the obstruction term arise again. This is done in Section \ref{S:cutoff}. In Section \ref{S:proof}, we finally prove \eqref{E:Main1} and remark at the end how to obtain \eqref{E:Main2}. We conclude this paper with the proof of Corollary \ref{C:DF} in Section \ref{S:DF}.
We would like to thank J. D. McNeal for fruitful discussions on this project, in particular we are very grateful to him for providing us with Lemma \ref{L:McNeal} and its proof.
\section{Preliminaries and pointwise obstruction}\label{S:prelim}
Let $(z_{1},\dots,z_{n})$ denote the coordinates of $\mathbb{C}^{n}$. We shall identify the vector $\langle\xi_{1},\dots,\xi_{n}\rangle$ in $\mathbb{C}^{n}$ with $\sum_{i=1}^{n}\xi_{i}\frac{\partial}{\partial z_{i}}$ in the $(1,0)$-tangent bundle of $\mathbb{C}^{n}$ at any given point. This means in particular that if $X$, $Y$ are $(1,0)$-vector fields with $X(z)=\sum_{i=1}^{n}X_{i}(z)\frac{\partial}{\partial z_{i}}$ and $Y(z)=\sum_{i=1}^{n}Y_{i}(z)\frac{\partial}{\partial z_{i}}$, then \begin{align*}
H_{\rho}(X,Y)(z)=\sum_{j,k=1}^{n}\frac{\partial^{2}\rho}{\partial z_{j}\partial\overline{z}_{k}}(z)
X_{j}(z)\overline{Y}_{k}(z). \end{align*}
Suppose $Z$ is another $(1,0)$-vector field with $Z(z)=\sum_{l=1}^{n}Z_{l}(z)\frac{\partial}{\partial z_{l}}$. For notational convenience, and because of lack of a better notation, we shall write \begin{align*}
(ZH_{\rho})(X,Y)(z):=
\sum_{j,k,l=1}^{n}\frac{\partial^{3}\rho}{\partial z_{j}\partial\overline{z}_{k}\partial z_{l}}(z)
X_{j}(z)\overline{Y}_{k}(z) Z_{l}(z). \end{align*}
We use the pointwise hermitian inner product $\langle .,.\rangle$ defined by $\langle\frac{\partial}{\partial z_{j}},\frac{\partial}{\partial z_{k}}\rangle=\delta_{j}^{k}$. Hoping that it will not cause any confusion, we also write $\langle .,.\rangle$ for contractions of vector fields and forms.
We will employ the so-called (sc)-(lc) inequality: $|ab|\leq\tau|a|^{2}+\frac{1}{4\tau}|b|^{2}$ for
$\tau>0$. Furthermore, we shall write $|A|\lesssim|B|$ to mean $|A|\leq c|B|$ for some constant $c>0$ which does not depend on any of the relevant parameters. In particular, we will only use this notation when $c$ depends solely on absolute constants, e.g., dimension, quantities related to the given defining function $\rho$.
Let us now work on proving inequality \eqref{E:Main1}.
Since $b\Omega$ is smooth, there exists a neighborhood $U$ of $b\Omega$ and a smooth map
\begin{align*}
\pi:\overline{\Omega}\cap U&\longrightarrow b\Omega\\
q&\longmapsto\pi(q)=p
\end{align*}
such that $\pi(q)=p$ lies on the line normal to $b\Omega$ passing through $q$ and $|p-q|$ equals
the Euclidean distance, $d_{b\Omega}(q)$, of $q$ to $b\Omega$.
After possibly shrinking $U$, we can assume that $\partial\rho\neq 0$ on $U$. We set
$N(z)=\frac{1}{|\partial\rho(z)|}\sum_{j=1}^{n}\frac{\partial\rho}{\partial\overline{z}_{j}}(z)
\frac{\partial}{\partial z_{j}}$. If $f$ is a smooth function on $U$, then it follows from Taylor's theorem
that
\begin{align}\label{E:BasicTaylor}
f(q)=f(p)-d_{b\Omega}(q)\left(\operatorname{Re} N\right)(f)(p)+\mathcal{O}\left(d_{b\Omega}^{2}(q)\right)\;\;
\text{for}\;\;q\in \overline{\Omega}\cap U;
\end{align}
for details see for instance Section 2.1 in \cite{For-Her}\footnote{Equation \eqref{E:BasicTaylor} above differs from (2.1) in \cite{For-Her} by a factor of $2$ in the second term on the right hand side. This stems from mistakenly using that outward normal of length $1/2$ instead of the one of unit length in
\cite{For-Her}. However, this mistake is inconsequential for the results in \cite{For-Her}.}.
Let $p\in b\Omega\cap U$ be given. Let $W\in\mathbb{C}^{n}$ be a weak, complex tangential vector at
$p$, i.e., $\langle\partial\rho(p),W\rangle=0$ and $H_{\rho}(W,W)(p)=0$. If $q\in\Omega\cap U$ with
$\pi(q)=p$, then \eqref{E:BasicTaylor} implies
\begin{align}\label{E:BasicTaylorH}
H_{\rho}(W,W)(q)=H_{\rho}(W,W)(p)-d_{b\Omega}(q)\left(\operatorname{Re} N\right)
\left(H_{\rho}(W,W)\right)(p)+\mathcal{O}(d^{2}_{b\Omega}(q))|W|^{2}.
\end{align}
Since $H_{\rho}(W,W)$ is a real-valued function, we have
\begin{align*}
\left(\operatorname{Re} N\right)\left(H_{\rho}(W,W)\right)=\operatorname{Re}\left[N\left(H_{\rho}(W,W)\right)\right].
\end{align*}
Moreover,
$H_{\rho}(W,W)$ is non-negative on $b\Omega\cap U$ and equals $0$ at $p$. That is,
$H_{\rho}(W,W)_{|_{b\Omega\cap U}}$ attains a local minimum at $p$. Therefore, any tangential
derivative of $H_{\rho}(W,W)$ vanishes at $p$. Since $N-\overline{N}$ is tangential to $b\Omega$,
we obtain
\begin{align*}
\operatorname{Re}\left[N\left(H_{\rho}(W,W)\right)\right](p)=N\left(H_{\rho}(W,W)\right)(p)
=(NH_{\rho})(W,W)(p),
\end{align*}
where the last equality holds since $W$ is a fixed vector.
Hence, \eqref{E:BasicTaylorH} becomes
\begin{align}\label{E:BasicTaylorHW}
H_{\rho}(W,W)(q)=-d_{b\Omega}(q)(NH_{\rho})(W,W)(p)
+\mathcal{O}\left(d_{b\Omega}^{2}(q)\right)|W|^{2}.
\end{align}
Clearly, we have a problem with obtaining \eqref{E:Main1}
when $(NH_{\rho})(W,W)$ is strictly positive at $p$. That is, when $H_{\rho}(W,W)$
is strictly decreasing along the real inward normal to $b\Omega$ at $p$, i.e.,
$H_{\rho}(W,W)$ becomes negative there, then \eqref{E:Main1} can not hold for the complex
Hessian $\rho$ when $
\epsilon>0$ is sufficiently close to zero. The question is whether we can find another smooth defining function $r$ of
$\Omega$ such that $(NH_{r})(W,W)(p)$ is less than $(NH_{\rho})(W,W)(p)$. The construction of
such a function $r$ is relatively easy and straightforward when $n=2$ (see Section 2.3 in \cite{For-Her}
for a non-technical derivation of $r$). The difficulty in higher dimensions arises simply from the fact that
the Levi form of a defining function might vanish in more than one
complex tangential direction at a given boundary point.
\section{Pointwise Modification of $\rho$}\label{S:modification}
Let $\Sigma_{i}\subset b\Omega$ be the set of boundary points at which the Levi form of $\rho$
has rank $i$, $i\in\{0,\dots,n-1\}$. Note that $\cup_{i=0}^{j}\Sigma_{i}$ is closed in $b\Omega$ for any
$j\in\{0,\dots,n-1\}$. Moreover,
$\Sigma_{j}$ is relatively closed in $b\Omega\setminus \cup_{i=0} ^{j-1}\Sigma_{i}$ for
$j\in\{1,\dots,n-1\}$. Of course, $\Sigma_{n-1}$ is the set of strictly pseudoconvex
boundary points of $\Omega$.
Let $p\in b\Omega\cap\Sigma_{i}$ for some $i\in\{0,\dots, n-2\}$ be given. Then there exist a
neighborhood $V\subset U$ of $p$ and smooth, linearly independent $(1,0)$-vector fields $W^{\alpha}$,
$1\leq\alpha\leq n-1-i$, on $V$, which are complex tangential to $b\Omega$ on $b\Omega\cap V$ and
satisfy $H_{\rho}(W^{\alpha},W^{\alpha})=0$ on $\Sigma_{i}\cap V$.
We consider those points $q\in\Omega\cap V$ with $\pi(q)=p$.
We shall work with the smooth function
\begin{align*}
r(z)=\rho(z)\cdot e^{-C\sigma(z)},\;\text{where}\; \;
\sigma(z)=\sum_{\alpha=1}^{n-1-i}H_{\rho}(W^{\alpha},W^{\alpha})(z)
\end{align*}
for $z\in V$. Here, the constant $C>0$ is fixed and to be chosen later.
Note that $r$ defines $b\Omega$ on $b\Omega\cap V$.
Furthermore, $\sigma$ is a smooth function on $V$ which is non-negative on $b\Omega\cap V$ and
vanishes on the set $\Sigma_{i}\cap
V$. That means that $\sigma_{|_{b\Omega\cap V}}$ attains a local minimum at each point in $\Sigma_{i}
\cap V$. Therefore, any tangential derivative of $\sigma$ vanishes on $\Sigma_{i}\cap V$.
Moreover, if $z\in\Sigma_{i}\cap V$ and $T\in\mathbb{C}T_{z}b\Omega$ is such that $H_{\rho}(T,T)$
vanishes at $z$, then $H_{\sigma}(T,T)$ is non-negative at that point.
Let $W\in
\mathbb{C}^{n}$
be a vector contained in the span of the vectors
$\left\{W^{\alpha}(p)\right\}_{\alpha=1}^{n-1-i}$. Then, using \eqref{E:BasicTaylorHW}, it follows that
\begin{align}\label{E:explaincutoff}
H_{r}(W,W)(q)
=&e^{-C\sigma(q)}\left[H_{\rho}(W,W)
-C\operatorname{Re}\left(\langle\partial\rho,W\rangle\overline{\langle\partial\sigma,W\rangle
}\right)\right.\notag\\
&\hspace{3.5cm}+\left.\rho\left(C^{2}\left|\langle\partial\sigma,W\rangle\right|^{2}
-CH_{\sigma}(W,W)\right)\right](q)\notag\\
=&e^{-C\sigma(q)}
\Bigl[
-d_{b\Omega}(q)(NH_{\rho})(W,W)(p)-C\rho(q)H_{\sigma}(W,W)(q)\bigr.
+\mathcal{O}\left(d_{b\Omega}^{2}(q)\right)|W|^{2}\notag\\
&\hspace{1.5cm}+C^{2}\rho(q)\left|\langle\partial\sigma(q),W\rangle \right|^{2}\bigl.-2C\operatorname{Re}\left(
\langle\partial\rho,W\rangle\overline{\langle\partial\sigma,W\rangle
}
\right)(q)
\Bigr].
\end{align}
Since $\langle\partial\sigma(p),W\rangle=0=\langle\partial\rho(p),W\rangle$, Taylor's theorem gives
\begin{align*}
\langle\partial\sigma(q),W\rangle=\mathcal{O}\left(r(q)\right)|W|=\langle\partial\rho(q),W\rangle.
\end{align*}
Therefore, we obtain
\begin{align}\label{E:BasicTaylorr}
H_{r}(W,W)(q)\geq -d_{b\Omega}(q)e^{-C\sigma(q)}(NH_{\rho})(W,W)(p)&-Cr(q)H_{\sigma}(W,W)(q)\\
&+\mathcal{O}\left(r^{2}(q)\right)|W|^{2},\notag \end{align}
where the constant in the last term depends on the choice of the constant $C$. However, in view of our claim \eqref{E:Main1}, this is inconsequential. From here on, we will not point out such negligible dependencies.
We already know that $H_{\sigma}(W,W)(p)$ is non-negative, i.e., of the right sign to correct
$(NH_{\rho})(W,W)(p)$ when necessary. The question is whether the sizes of
$(NH_{\rho})(W,W)(p)$ and
$H_{\sigma}(W,W)(p)$ are comparable in some sense. The following proposition clarifies this.
\begin{proposition}\label{P:compare}
There exists a constant $K>0$ such that
\begin{align*}
\left|(NH_{\rho})(W,W)(z_{0})\right|^{2}
\leq K|W|^{2}\cdot H_{\sigma}(W,W)(z_{0})
\end{align*}
holds for all $z_{0}\in\Sigma_{i}\cap V$ and $W\in\mathbb{C}T_{z_{0}}b\Omega$
with $H_{\rho}(W,W)(z_{0})=0$.
\end{proposition}
In order to prove Proposition \ref{P:compare}, we need the following lemma:
\begin{lemma}\label{L:compare}
Let $z_{0}\in b\Omega$ and $U$ a neighborhood of $z_{0}$. Let $Z$ be a smooth $(1,0)$-vector field
defined
on $U$, which is complex tangential to $b\Omega$ on $b\Omega\cap U$, and let
$Y\in\mathbb{C}^{n}$ be a vector belonging to $
\mathbb{C}T_{z_{0}}b\Omega$. Suppose that $Y$ and $Z$ are such that
\begin{align*}
H_{\rho}(Y,Y)(z_{0})=0=H_{\rho}(Z,Z)(z_{0}).
\end{align*}
Set $X=\sum_{j=1}^{n}\overline{Y}(Z_{j})\frac{\partial}{\partial z_{j}}$. Then the following holds:
\begin{enumerate}
\item $X$ is complex tangential to $b\Omega$ at $z_{0}$,
\item $\left(YH_{\rho}\right)(X,Z)(z_{0})=0$,\
\item $H_{H_{\rho}(Z,Z)}(Y,Y)(z_{0})\geq H_{\rho}(X,X)(z_{0})$.
\end{enumerate}
\end{lemma}
\begin{proof}[Proof of Lemma \ref{L:compare}]
(1) That $X$ is complex tangential to $b\Omega$ at $z_{0}$ was shown in Lemma 3.4 of
\cite{For-Her}.
(2) The plurisubharmonicity of $\rho$ says that
both $H_{\rho}(Y,Y)_{|_{b\Omega\cap U}}$ and $H_{\rho}(Z,Z)_{|_{b\Omega\cap U}}$
attain a local minimum at $z_{0}$. In fact, the function $H_{\rho}(aY+bZ,aY+bZ)_{|_{b\Omega}}$,
$a,b\in\mathbb{C}$, attains a local minimum at $z_{0}$. This means that any tangential derivative of
either one of those three functions must vanish at that point. In particular, we have
\begin{align*}
0&=\langle\partial H_{\rho}(aY+bZ,aY+bZ),X\rangle(z_{0})\\
&=|a|^{2}\langle\partial H_{\rho}(Y,Y),X\rangle(z_{0})
+2\operatorname{Re}\left(a\overline{b}\langle\partial H_{\rho}(Y,Z),X\rangle
\right)(z_{0})+|b|^{2}\langle\partial H_{\rho}(Z,Z),X\rangle(z_{0})\\
&=2\operatorname{Re}\left(a\overline{b}\langle\partial H_{\rho}(Y,Z),X\rangle
\right)(z_{0}).
\end{align*}
Since this is true for all $a,b\in\mathbb{C}$, it follows that $\langle\partial H_{\rho}(Y,Z),X\rangle$
must vanish
at $z_{0}$. But the plurisubharmonicity of $\rho$ at $z_{0}$ yields
\begin{align*}
\langle\partial H_{\rho}(Y,Z),X\rangle=\left(XH_{\rho}\right)(Y,Z)(z_{0})=
\left(YH_{\rho}\right)(X,Z)(z_{0}),
\end{align*}
which proves the claim.
(3) Consider the function
\begin{align*}
f(z)=\left(
H_{\rho}(Z,Z)\cdot H_{\rho}(X,X)-\left|H_{\rho}(X,Z)\right|^{2}\right)(z)\;\;\text{for}\;\;z\in U.
\end{align*}
Note that $f_{|_{b\Omega\cap U}}$ attains a local minimum at $z_{0}$. Since $Y$ is a weak direction at
$z_{0}$, it follows that $H_{f}(Y,Y)(z_{0})$ is
non-negative. This implies that
\begin{align}\label{E:comparetemp}
\left(H_{H_{\rho}(Z,Z)}(Y,Y)\cdot H_{\rho}(X,X)\right)(z_{0})\geq
\left|
\langle \partial H_{\rho}(X,Z), Y\rangle(z_{0})
\right|^{2},
\end{align}
where we used that both $H_{\rho}(Z,Z)$ and any tangential derivative of $H_{\rho}(Z,Z)$ at $z_{0}$
are zero. We compute
\begin{align*}
\langle\partial H_{\rho}(X,Z),Y\rangle(z_{0})
=(YH_{\rho})(X,Z)(z_{0})+H_{\rho}(X,X)(z_{0})
+H_{\rho}\left(\sum_{j=1}^{n}Y(X_{j})\frac{\partial}{\partial z_{j}},Z \right)(z_{0}).
\end{align*}
The first term on the right hand side equals zero by part (2) of Lemma \ref{L:compare}, and the third
term is zero as well since $\rho$ plurisubharmonic at $z_{0}$ and $Z$ is a weak direction there. Therefore,
\eqref{E:comparetemp} becomes
\begin{align*}
H_{\rho}(X,X)(z_{0})\leq H_{H_{\rho}(Z,Z)}(Y,Y)(z_{0}).
\end{align*}
\end{proof}
Now we can proceed to show Proposition \ref{P:compare}.
\begin{proof}[Proof of Proposition \ref{P:compare}]
Recall that we are working with vectors $W\in\mathbb{C}^{n}$ contained in the span of
$W^{1}(z_{0}),\dots,W^{n-1-i}(z_{0})$. We consider the function
\begin{align*}
h(z)=\left(
\sigma\cdot H_{\rho}(N,N)-\sum_{\alpha=1}^{n-1-i}\left|
H_{\rho}(N,W^{\alpha})
\right|^{2}
\right)(z) \;\;\text{for}\;\;z\in U,
\end{align*}
where $\sigma=\sum_{\alpha=1}^{n-1-i}H_{\rho}(W^{\alpha},W^{\alpha})$. Again, since $\rho$ is
plurisubharmonic on $b\Omega$, $h_{|_{b\Omega\cap V}}$ has a local minimum at $z_{0}$.
This, together with $H_{\rho}(W,W)(z_{0})=0$, implies that
$H_{h}(W,W)(z_{0})$ is non-negative. Since both $\sigma$ and $\langle\partial\sigma,
W\rangle=0$ vanish at $z_{0}$,
it
follows that
\begin{align*}
&\left(
H_{\sigma}(W,W)\cdot H_{\rho}(N,N)
\right)(z_{0})\\
&\geq
\sum_{\alpha=1}^{n-1-i}\left|
(WH_{\rho})(N,W^{\alpha})+H_{\rho}\left(\sum_{j=1}^{n}W(N_{j})\frac{\partial}{\partial z_{j}},W^{\alpha}
\right)
+H_{\rho}\left(
N,\sum_{j=1}^{n}\overline{W}(W_{j}^{\alpha})\frac{\partial}{\partial z_{j}}
\right)
\right|^{2}(z_{0})\\
&=
\sum_{\alpha=1}^{n-1-i}\left|
(WH_{\rho})(N,W^{\alpha})+H_{\rho}\left(
N,\sum_{j=1}^{n}\overline{W}(W_{j}^{\alpha})\frac{\partial}{\partial z_{j}}
\right)
\right|^{2}(z_{0}),
\end{align*}
where the last step follows from $\rho$ being plurisubharmonic at $z_{0}$ and the
$W^{\alpha}$'s being
weak directions there.
Moreover, we have that $(WH_{\rho})(N,W^{\alpha})$ equals $(NH_{\rho})(W,W^{\alpha})$ at $z_{0}$. Writing
$X^{\alpha}=\sum_{j=1}^{n}\overline{W}(W_{j}^{\alpha})\frac{\partial}{\partial z_{j}}$, we obtain
\begin{align*}
\left(
H_{\sigma}(W,W)\cdot H_{\rho}(N,N)
\right)(z_{0})
&\geq
\sum_{\alpha=1}^{n-1-i}\left|
(NH_{\rho})(W,W^{\alpha})+H_{\rho}(N,X^{\alpha})
\right|^{2}(z_{0})\\
&\geq
\sum_{\alpha=1}^{n-1-i}\left(
\frac{1}{2}\left|
(NH_{\rho})(W,W^{\alpha})
\right|^{2}-3
\left|H_{\rho}(N,X^{\alpha})\right|^{2}
\right)(z_{0}).
\end{align*}
Here the last step follows from the (sc)-(lc) inequality.
Since $\rho$ is plurisubharmonic at $z_{0}$, we can apply the Cauchy--Schwarz inequality
\begin{align*}
\left|
H_{\rho}(N,X^{\alpha})(z_{0})
\right|^{2}
&\leq
\left(
H_{\rho}(N,N)\cdot H_{\rho}(X^{\alpha},X^{\alpha})
\right)(z_{0})\\
&\leq \left(
H_{\rho}(N,N)\cdot H_{H_{\rho}(W^{\alpha},W^{\alpha})}(W,W)
\right)(z_{0}),
\end{align*}
where the last estimate follows by part (3) of Lemma \ref{L:compare} with $W$ and $W^{\alpha}$
in place of $Y$ and $Z$, respectively. Thus we have
\begin{align*}
\sum_{\alpha=1}^{n-1-i}\left|H_{\rho}(N,X^{\alpha})(z_{0}) \right|^{2}
\leq
\left(H_{\rho}(N,N)\cdot H_{\sigma}(W,W)\right)(z_{0}),
\end{align*}
which implies that
\begin{align}\label{E:comparetemp2}
\sum_{\alpha=1}^{n-1-i}\left|
(NH_{\rho})(W,W^{\alpha})(z_{0})
\right|^{2}
\leq
8\left(H_{\sigma}(W,W)\cdot H_{\rho}(N,N)\right)(z_{0}).
\end{align}
Since $W$ is a linear combination of $\{W^{\alpha}(z_{0})\}_{\alpha=1}^{n-1-i}$, we can write
$W=\sum_{\alpha=1}^{n-1-i}a_{\alpha}W^{\alpha}(z_{0})$ for some scalars $a_{\alpha}\in\mathbb{C}$.
Because of the linear independence of the $W^{\alpha}$'s on $V$, there exists a constant $K_{1}>0$
such
that
\begin{align*}
\sum_{\alpha}^{n-1-i}|b_{\alpha}|^{2}\leq K_{1}\left|
\sum_{\alpha=1}^{n-1-i}b_{\alpha}W^{\alpha}(z)
\right|^{2}\;\text{for all}\;z\in b\Omega\cap V,\;b_{\alpha}\in\mathbb{C}.
\end{align*}
Thus it follows that
\begin{align*}
\sum_{\alpha=1}^{n-1-i}
\left|
(NH_{\rho})(W,W^{\alpha})(z_{0})
\right|^{2}
&\geq
\frac{1}{K_{1}|W|^{2}}
\sum_{\alpha=1}^{n-1-i}\left|(NH_{\rho})(W,a_{\alpha}W^{\alpha})(z_{0})\right|^{2}\\
&\geq
\frac{1}{K_{1}(n-1-i)|W|^{2}}
\left|(NH_{\rho})(W,W)(z_{0})\right|^{2}.
\end{align*}
Hence, \eqref{E:comparetemp2} becomes
\begin{align*}
\left|(NH_{\rho})(W,W)(z_{0})\right|^{2}\leq
8K_{1}(n-1-i)|W|^{2}\left(H_{\sigma}(W,W)\cdot H_{\rho}(N,N)\right)(z_{0}).
\end{align*}
Let $K_{2}>0$ be a constant such that
$H_{\rho}(N,N)_{|_{b\Omega}}\leq K_{2}$ holds. Setting $K=8K_{1}K_{2}(n-1-i)$,
it follows that
\begin{align*}
\left|(NH_{\rho})(W,W)(z_{0})\right|^{2}\leq K|W|^{2}H_{\sigma}(W,W)(z_{0}).
\end{align*}
\end{proof}
Recall that we are considering a fixed boundary point $p\in\Sigma_{i}$ and all $q\in\Omega\cap V$,
$\pi(q)=p$, for
some sufficiently small neighborhood of $p$. After possibly shrinking $V$ it follows by Taylor's theorem
that
\begin{align*}
H_{\sigma}(W,W)(q)=H_{\sigma}(W,W)(\pi(q))+\mathcal{O}\left(d_{b\Omega}(q)\right)|W|^{2}
\end{align*}
holds for all $q\in\Omega\cap V$ with $\pi(q)=p$.
Using this and Proposition \ref{P:compare}, we get for $q\in\Omega\cap V$ with
$\pi(q)=p$ that
\begin{align*}
\left|(NH_{\rho})(W,W)(p)\right|^{2}\leq K|W|^{2}\left[H_{\sigma}(W,W)(q)
+\mathcal{O}\left(d_{b\Omega}(q)\right)|W|^{2}\right].
\end{align*}
Therefore, our basic estimate \eqref{E:BasicTaylorr} of the complex Hessian of $r$ in direction $W$
becomes
\begin{align*}
H_{r}(W,W)(q)\geq
- d_{b\Omega}(q)e^{-C\sigma(q)}(NH_{\rho})(W,W)(p)
&-r(q)\frac{C}{K|W|^{2}}\left|NH_{\rho}(W,W)\right|^{2}(p)\\
&+\mathcal{O}\left(r^{2}(q)\right)|W|^{2}.
\end{align*}
Let $c_{1}>0$ be such that $d_{b\Omega}(z)\leq c_{1}|\rho(z)|$ for all $z$ in $V$.
Then, if we choose
\begin{align}\label{E:chooseC}
C\geq\max_{z\in b\Omega, T\in\mathbb{C}^{n}, |T|=1}
\left\{
0,\frac{c_{1}\operatorname{Re}\left[(NH_{\rho})(T,T)(z)\right]-\frac{\epsilon}{2}}{\left|(NH_{\rho})(T,T)(z)\right|^{2}}K
\right\},
\end{align}
we obtain, after possibly shrinking $V$,
\begin{align}\label{E:BasicTaylorrfinal}
H_{r}(W,W)(q)\geq -\epsilon |r(q)|\cdot|W|^{2}
\end{align}
for all $q\in\Omega\cap V$ with $\pi(q)=p$ and $W\in\mathbb{C}^{n}$ in the span of
$\{W^{\alpha}(p)\}_{\alpha=1}^{n-1-i}$. In fact, after possibly shrinking $V$, \eqref{E:BasicTaylorrfinal} holds with, say, $2\epsilon$ in place of $\epsilon$ for all
$q\in\Omega\cap V$ satisfying $\pi(q)\in\Sigma_{i}\cap V$ and $W\in\mathbb{C}^{n}$ belonging to
the span of $\{W^{\alpha}(\pi(q))\}_{\alpha=1}^{n-1-i}$.
A problem with this construction is that
$r$ is not necessarily plurisubharmonic at those weakly pseudoconvex boundary points which are not
in $\Sigma_{i}$. This possible loss of plurisubharmonicity occurs because the
$W^{\alpha}$'s are not necessarily weak directions at those points. This means,
that we can not simply copy this construction with $r$ in place of $\rho$ to get good estimates near, say,
$\Sigma_{i+1}$. Let us be a more explicit. Suppose $\tilde{p}\in\Sigma_{i+1}\cap V$ is such that
at least one of the $W^{\alpha}$'s is not a weak direction at $\tilde{p}$. That means, if
$T$ is a weak complex tangential direction at $\tilde{p}$, then neither does $|\langle\partial\sigma(\tilde{p}),T\rangle|^{2}$ have to be
zero nor does $H_{\sigma}(T,T)(\tilde{p})$ have to be non-negative. In view of \eqref{E:explaincutoff},
this says that it might actually happen that
$(NH_{r})(T,T)(\tilde{p})$ is greater than $(NH_{\rho})(T,T)(\tilde{p})$ for such a vector $T$. That is, by
removing the obstruction term at $p$ we might
have worsened the situation at $\tilde{p}$.
One might think that this does not cause any real problems since we
still need to introduce a correcting function $\tilde{\sigma}$ to remove the obstruction to
\eqref{E:Main1} on the set $\Sigma_{i+1}\cap V$. However, it might be the case that
$(NH_{\rho})(T,T)(\tilde{p})=0$. In this case we do not know whether $H_{\tilde{\sigma}}(T,T)$ is
strictly positive at $\tilde{p}$, i.e., we do not know whether $H_{\tilde{\sigma}}(T,T)(\tilde{p})$ can make
up for any obstructing terms at $\tilde{p}$ introduced by $\sigma$. This says that we need to smoothly
cut off $\sigma$ in a manner such that, away from $\Sigma_{i}\cap V$,
$|\langle\partial\sigma,T\rangle|^{2}$ stays close to zero and $H_{\sigma}(T,T)$ does not become too
negative (relative to $\epsilon|T|^{2}$). The construction of such a cut off function will be done in the
next section.
\section{The cutting off}\label{S:cutoff}
Let us recall our setting: we are considering a given boundary point $p\in\Sigma_{i}$,
$0\leq i \leq n-2$, $V$ a
neighborhood of $p$ and smooth, linearly independent $(1,0)$-vector fields
$\{W^{\alpha}\}_{\alpha=1}^{n-1-i}$, $\alpha\in\{1,\dots,n-1-i\}$ on $V$, which are complex tangential
to $b\Omega$ on $b\Omega\cap V$ and satisfy $H_{\rho}(W^{\alpha},W^{\alpha})=0$ on
$\Sigma_{i}\cap V$. From now on, we also suppose that $V$ and the $W^{\alpha}$'s are
chosen such that the
span of $\{W^{\alpha}(z)\}_{\alpha=1}^{n-1-i}$ contains the null space of the Levi form of $\rho$ at $z$
for all $z\in\Sigma_{j}\cap V$ for $j\in\{i+1,\dots,n-2\}$. This can be done by first selecting
smooth $(1,0)$-vector fields $\{S^{\beta}(z)\}_{\beta=1}^{i}$ which are complex tangential to
$b\Omega$ on $b\Omega\cap V$ for some neighborhood $V$ of $p$ and orthogonal to each other
with respect to the Levi form of $\rho$ such that $H_{\rho}(S^{\beta},S^{\beta})>0$ holds on
$b\Omega\cap V$ after possibly shrinking $V$. Then one completes the basis of the complex
tangent space with smooth $(1,0)$-vector fields $\{W^{\alpha}(z)\}_{\alpha=1}^{n-1-i}$ such that
the $W^{\alpha}$'s are orthogonal to the $S^{\beta}$'s with respect to the Levi form of $\rho$.
Let $V'\subset\subset V$ be another neighborhood of $p$. Let $\zeta\in C^{\infty}_{c}(V,[0,1])$ be a function
which equals $1$ on $V'$. For given $m>2$, let $\chi_{m}\in C^{\infty}(\mathbb{R})$ be
an increasing function with $\chi_{m}(x)=1$ for all $x\leq 1$ and $\chi_{m}(x)=e^{m}$ for all
$x\geq e^{m}$ such that
\begin{align*}
\frac{x}{\chi_{m}(x)}\leq 2,\;\;
\chi_{m}'(x)\leq 2,\;\text{and}\;\;
x\cdot\chi_{m}''(x)\leq 4\;\;\text{for all}\;\; x\in[1,e^{m}].
\end{align*}
Set $\chi_{m,\tau}(x)=\chi_{m}\left(\frac{x}{\tau}\right)$ for given $\tau>0$. The above properties then
become
\begin{align*}
\frac{x}{\chi_{m,\tau}(x)}\leq 2\tau,\;\;
\chi_{m,\tau}'(x)\leq \frac{2}{\tau},\;\text{and}\;\;
x\cdot\chi_{m,\tau}''(x)\leq \frac{4}{\tau}\;\;\text{for all}\;\; x\in[\tau,\tau e^{m}].
\end{align*}
Set $g_{m,\tau}(x)=1-\frac{\ln(\chi_{m,\tau}(x))}{m}$. It follows by a straightforward computation that
\begin{align}\label{E:propsg}
&g_{m,\tau}(x)=1\;\text{for}\;x\leq\tau,\;\;0\leq g_{m,\tau}(x)\leq 1\;\text{for all}\;x\in\mathbb{R},
\;\text{and}\notag\\
&|g_{m,\tau}'(x)|\leq\frac{4}{m}\cdot\frac{1}{x},\;\;g_{m,\tau}''(x)\geq-\frac{8}{m}\cdot\frac{1}{x^{2}}\;\;
\text{for}\;x\in(\tau, \tau e^{m}).
\end{align}
For given $m,\;\tau> 0$ we define
\begin{align*}
s_{m,\tau}(z)=\zeta(z)\cdot\sigma(z)\cdot g_{m,\tau}(\sigma(z))\;\;\text{for}\;z\in V
\end{align*}
and $s_{m,\tau}=0$ outside of $V$. This function has the properties described at the
end of Section \ref{S:modification} if $m,\tau$ are chosen appropriately:
\begin{lemma}\label{L:cutoff}
For all $\delta>0$, there exist $m,\;\tau>0$ such that
$s_{m,\tau}$ satisfies:
\begin{enumerate}
\item[(i)] $s_{m,\tau}=\zeta\sigma$ for $\sigma\in[0,\tau]$,
\item[(ii)] $0\leq s_{m,\tau}\leq\delta$ on $b\Omega$.
\end{enumerate}
Moreover, if $z\in \left(b\Omega\cap V\right)\setminus\Sigma_{i}$ and
$T\in\mathbb{C}T_{z}b\Omega$, then
\begin{enumerate}
\item[(iii)] $|\langle\partial s_{m,\tau}(z),T\rangle|\leq\delta|T|$,
\item[(iv)] $H_{s_{m,\tau}}(T,T)(z)\geq-\delta|T|^{2}$ if $T\in\operatorname{span}\{W^{\alpha}(z)\}$.
\end{enumerate}
\end{lemma}
Note that part (iv) of Lemma \ref{L:cutoff} in particular says that if
$z\in\cup_{j=i+1}^{n-2}\Sigma_{j}\cap V$, then $H_{s_{m,\tau}}(T,T)(z)\geq-\delta|T|^{2}$ for all $T$
which are weak complex tangential vectors at $z$.
To prove Lemma \ref{L:cutoff}, we will need $|\langle\partial\sigma,T\rangle|^{2}
\lesssim\sigma|T|^{2}$ on
$\overline{\operatorname{supp}(\zeta)}\cap b\Omega$. That this is in fact true we learned from J. D. McNeal.
\begin{lemma}[\cite{McN}]\label{L:McNeal}
Let $U\subset\subset\mathbb{R}^{n}$ be open. Let $f\in C^{2}(U)$ be a non-negative
function on $U$. Then for any compact set $K\subset\subset U$, there exists a constant $c>0$ such that
\begin{align}\label{E:McNeal}
|\nabla f(x)|^{2}\leq c f(x)\;\;\text{for all}\;\;x\in K.
\end{align}
\end{lemma}
Since the proof by McNeal is rather clever, and since we are not aware of it being published,
we shall give it here.
\begin{proof}
Let $F$ be a smooth, non-negative function such that $F=f$ on $K$ and $F=0$ on
$\mathbb{R}^{n}\setminus U$.
For a given $x\in K$, we have for all $h\in\mathbb{R}^{n}$ that
\begin{align}\label{E:McNealTaylor}
0\leq F(x+h)&=F(x)+\sum_{k=1}^{n}\frac{\partial F}{\partial x_{k}}(x)h_{k}
+\frac{1}{2}\sum_{k,l=1}^{n}\frac{\partial^{2} F}{\partial x_{k}\partial x_{l}}(\xi)h_{k}h_{l}\notag\\
&=f(x)+\sum_{k=1}^{n}\frac{\partial f}{\partial x_{k}}(x)h_{k}+
\frac{1}{2}\sum_{k,l=1}^{n}\frac{\partial^{2} F}{\partial x_{k}\partial x_{l}}(\xi)h_{k}h_{l}
\end{align}
holds for some $\xi\in U$.
Note that \eqref{E:McNeal} is true if $\left(\nabla f\right)(x)=0$. So assume now that
$\left(\nabla f\right)(x)\neq 0$ and choose
$h_{k}=\frac{\frac{\partial f}{\partial x_{k}}(x)}{\left|\left(\nabla f\right)(x)\right|}\cdot t$ for
$t\in\mathbb{R}$. Then
\eqref{E:McNealTaylor} becomes
\begin{align*}
0\leq f(x)+\left|\left(\nabla f\right)(x)\right|t
+nL\cdot\frac{\sum_{k=1}^{n}\left|\frac{\partial f}{\partial x_{k}}(x)\right|^{2}}{\left|\left(\nabla f\right)
(x) \right|^{2}}
\cdot t^{2}\;\;\text{for all}\;\;t\in\mathbb{R},
\end{align*}
where $L=\frac{1}{2}\sup
\left\{\left|\frac{\partial^{2}F}{\partial x_{k}\partial x_{l}}(\xi)
\right|\;|\;\xi\in U,\;1\leq k,l\leq n\right\}$.
Therefore, \eqref{E:McNealTaylor} becomes
\begin{align*}
0\leq f(x)+\left|\left(\nabla f\right)(x)\right|\cdot t+nL\cdot t^{2}\;\;\text{for all}\;\;t\in\mathbb{R}.
\end{align*}
In particular, the following must hold for all
$t\in\mathbb{R}$:
\begin{align*}
-\frac{f(x)}{nL}+\frac{\left|\left(\nabla f\right)(x)\right|^{2}}{(2nL)^{2}}
\leq
\left(
t+\frac{\left|\left(\nabla f\right)(x)\right|}{2nL}
\right)^{2},
\end{align*}
which implies that
\begin{align*}
\left|\left(\nabla f\right)(x)\right|^{2}\leq 4nL\cdot f(x).
\end{align*}
\end{proof}
We can assume that $V$ is such that there exists a diffeomorphism $\phi:V\cap b\Omega
\longrightarrow U$ for some open set $U\subset\subset\mathbb{R}^{2n-1}$.
Set $f:=\sigma\circ\phi^{-1}$. Then $f$ satisfies the
hypotheses of Lemma \ref{L:McNeal}. Hence we get that there exists a constant $c>0$ such that
$\left|\left(\nabla f\right)(x)\right|^{2}\leq cf(x)$ for all $x\in K=\phi(\overline{\operatorname{supp}(\zeta)})$. This implies
that there exists a constant $c_{1}>0$, depending on $\phi$, such that
\begin{align}\label{E:McNealapplied}
\left|\langle\partial\sigma(z),T\rangle\right|^{2}\leq c_{1}\sigma(z)|T|^{2}\;\;
\text{for all}\;\; z\in\overline{\operatorname{supp}(\zeta)}\;\text{and}\; T\in\mathbb{C}T_{z}b\Omega.
\end{align}
Now we can prove Lemma \ref{L:cutoff}.
\begin{proof}[Proof of Lemma \ref{L:cutoff}]
Note first that $s_{m,\tau}$ is identically zero
on $b\Omega\setminus V$ for any $m>2$ and $\tau>0$.
Now let $\delta>0$ be given, let $m$ be a large, positive number, fixed and to be chosen later
(that is, in the proof of (iv)).
Below we will show how to
choose $\tau>0$ once $m$ has been chosen.
Part (i) follows directly from the definition of $s_{m,\tau}$ for any choice of $m>2$ and $\tau>0$.
Part (ii) also follows
straightforwardly, if $\tau>0$ is such that $\tau e^{m}\leq\delta$.
Notice that for all $z\in b\Omega\cap V$ with $\sigma(z)>\tau e^{m}$, $s_{m,\tau}(z)=0$, and hence
(iii), (iv) hold trivially there. Thus, to
prove (iii) and (iv) we only need to consider the two sets
\begin{align*}
S_{1}=\{z\in b\Omega\cap V\;|\;\sigma(z)\in(0,\tau)\}\;\;\text{and}\;\;
S_{2}=\{z\in b\Omega\cap V\;|\;\sigma(z)\in[\tau,\tau e^{m}]\}.
\end{align*}
\emph{Proof of (iii):} If $z\in S_{1}$, then $s_{m,\tau}(z)=\zeta(z)\cdot\sigma(z)$ and
if $T\in\mathbb{C}T_{z}b\Omega$, we get
\begin{align*}
\left|\langle\partial s_{m,\tau}(z),T\rangle
\right|&=
\left|
\sigma\cdot\langle\partial\zeta,T\rangle+\zeta\cdot\langle\partial\sigma, T\rangle
\right|(z)\\
&\overset{\eqref{E:McNealapplied}}{\leq} \left(c_{2}\sigma(z)
+\left(c_{1}\sigma(z)\right)^{\frac{1}{2}}\right)|T|, \;
\text{where}\;\; c_{2}=\max_{z\in b\Omega\cap V}|\partial\zeta(z)|.
\end{align*}
Thus, if we choose $\tau>0$ such that, $c_{2}\tau+(c_{1}\tau)^{\frac{1}{2}}\leq\delta$, then
(iii) holds on the set $S_{1}$.\\
Now suppose that $z\in S_{2}, T\in\mathbb{C}T_{z}b\Omega$ and compute:
\begin{align*}
\left|
\langle\partial s_{m,\tau}(z),T\rangle
\right|
&=\left|
\sigma\cdot g_{m,\tau}(\sigma)\cdot\langle\partial\zeta,T\rangle
+\zeta\cdot\langle\partial\sigma,T\rangle
\left(
g_{m,\tau}(\sigma)+\sigma\cdot g_{m,\tau}'(\sigma)
\right)
\right|(z)\\
&\overset{\eqref{E:McNealapplied},\eqref{E:propsg}}{\leq}
\left(
c_{2}\sigma(z)+\left(c_{1}\sigma(z)\right)^{\frac{1}{2}}\cdot\left(1+\frac{4}{m}\right)
\right)|T|.
\end{align*}
Thus, if we choose $\tau>0$ such that $c_{2}\tau e^{m}+2(c_{1}\tau e^{m})^{\frac{1}{2}}
\leq\delta$, then (iii) also holds on the set $S_{2}$.
\emph{Proof of (iv):} Let us first consider the case when $z\in S_{1}$. Then, again,
$s_{m,\tau}(z)=\zeta(z)\cdot\sigma(z)$ and if $T$ is in the span of
$\{W^{\alpha}(z)\}_{\alpha=1}^{n-1-i}$, we obtain
\begin{align*}
H_{s_{m,\tau}}(T,T)(z)=\left[\sigma\cdot H_{\zeta}(T,T)
+2\operatorname{Re}\left(\langle\partial\zeta,T\rangle\cdot
\overline{\langle\partial\sigma,T\rangle}\right)
+\zeta\cdot H_{\sigma}(T,T)
\right](z).
\end{align*}
Let $c_{3}>0$ be a constant such that $H_{\zeta}(\xi,\xi)(z)\geq
-c_{3}|\xi|^{2}$ for all $z\in b\Omega\cap V$, $\xi\in\mathbb{C}^{n}$. Then it follows, using
\eqref{E:McNealapplied} again, that
\begin{align*}
H_{s_{m,\tau}}(T,T)(z)\geq
-\left(c_{3}\sigma(z)+2c_{2}\left(c_{1}\sigma(z) \right)^{\frac{1}{2}}\right) |T|^{2}
+\zeta(z)\cdot H_{\sigma}(T,T)(z).
\end{align*}
Note that for $z$ and $T$ as above, $H_{\sigma}(T,T)(z)\geq 0$ when $\sigma(z)=0$.
Furthermore, the set
\begin{align*}
\left\{(z,T)\;|\;z\in b\Omega\cap\overline{\operatorname{supp}\zeta},\;\sigma(z)=0,\;
T\in\operatorname{span}\{W^{1}(z),\dots,W^{n-1-i}(z)\}\right\}
\end{align*}
is a closed subset of the complex tangent bundle of $b\Omega$.
Thus there exists a neighborhood $U\subset V$
of $\{z\in b\Omega\cap\overline{\operatorname{supp}\zeta}\;|\;\sigma(z)=0\}$ such that
\begin{align*}
H_{\sigma}(T,T)(z)\geq-\frac{\delta}{2}|T|^{2}
\end{align*}
holds for all $z\in b\Omega\cap U$, $T$ in the span of $\{W^{\alpha}(z)\}_{\alpha=1}^{n-1-i}$.
Let $\nu_{1}$ be the maximum of $\sigma$ on the closure of $b\Omega\cap U$, and let
$\nu_{2}$ be the minimum of $\sigma$ on
$\left(b\Omega\cap\overline{\operatorname{supp}\zeta}\right)\setminus U$.
Now choose $\tau>0$ such that
$\tau \leq\min\{\nu_{1},\frac{\nu_{2}}{2}\}$. Then $z\in S_{1}$ implies that $z\in b\Omega\cap U$ and
therefore
$\zeta(z)\cdot H_{\sigma}(T,T)(z)\geq-\frac{\delta}{2}|T|^{2}$ for all $T$ in the span of
the $W^{\alpha}(z)$'s.
If we also make sure that $\tau>0$ is such that $c_{3}\tau+2c_{2}\left(c_{1}\tau\right)^{\frac{1}{2}}
\leq\frac{\delta}{2}$, then (iv) is true on $S_{1}$.
Now suppose that $z\in S_{2}$ and $T$ in the span of
$\{W^{\alpha}(z)\}_{\alpha=1}^{n-1-i}$. We compute
\begin{align*}
H_{s_{m,\tau}}&(T,T)(z)=
\Bigl[\sigma g_{m,\tau}(\sigma) H_{\zeta}(T,T)
+2\operatorname{Re}\left(\langle\partial\zeta,T\rangle\cdot
\overline{\langle\partial\sigma,T\rangle}
\right)
\left(
g_{m,\tau}(\sigma)+\sigma g_{m,\tau}'(\sigma)
\right)\Bigr.\\
&\Bigl.+
\zeta\left|\langle\partial\sigma,T\rangle\right|^{2}
\left(2g_{m,\tau}'(\sigma)+\sigma g_{m,\tau}''(\sigma)\right)
+\zeta H_{\sigma}(T,T)
(g_{m,\tau}(\sigma)+\sigma g_{m,\tau}'(\sigma))
\Bigr](z)\\
&=\operatorname{I}+\operatorname{II}+\operatorname{III}+\operatorname{IV}.
\end{align*}
If we choose $\tau>0$ such that $\tau e^{m}c_{3}\leq\frac{\delta}{4}$, then it follows that
$\operatorname{I}\geq-\frac{\delta}{4}|T|^{2}$. Estimating the term $\operatorname{II}$ we get
\begin{align*}
\operatorname{II}\overset{\eqref{E:propsg}}{\geq}-2c_{2}\left|\langle\partial\sigma(z),T\rangle
\right||T|\left(1+\frac{4}{m}\right)
\overset{\eqref{E:McNealapplied}}{\geq} -4c_{2}\left(c_{1}\sigma(z)\right)^{\frac{1}{2}}|T|^{2}
\geq-\frac{\delta}{4}|T|^{2},
\end{align*}
if we choose $\tau>0$ such that $4c_{2}\left(c_{1}\tau e^{m}\right)^{\frac{1}{2}}\leq\frac{\delta}{4}$.
To estimate term
$\operatorname{IV}$, we only need to make sure that $\tau>0$ is so small that $z\in S_{2}$ implies that
$2\zeta(z)\cdot H_{\sigma}(T,T)(z)\geq
-\frac{\delta}{4}|T^{2}$. This can be done similarly to the case when $z\in S_{1}$.
Note that up to this point the size of the parameter $m$ played no role. That is, we obtain above
results for any choice of $m$ as long as $\tau>0$ is sufficiently small. The size of $m$ only matters
for the estimates on term $\operatorname{III}$: \eqref{E:propsg} and \eqref{E:McNealapplied} yield
\begin{align*}
\operatorname{III}&\geq -\left|\langle\partial\sigma(z),T\rangle\right|^{2}
\left[2|g_{m,\tau}'(\sigma(z))|+\sigma(z)g_{m,\tau}''(\sigma(z))\right]\\
&\geq -\left|\langle\partial\sigma(z),T\rangle\right|^{2}\frac{16}{m\sigma(z)}
\geq-\frac{16 c_{1}}{m}|T|^{2}.
\end{align*}
We now choose $m>0$ such that $\frac{16 c_{1}}{m}\leq\frac{\delta}{4}$, and then we choose
$\tau>0$ according to our previous computations .
\end{proof}
\section{Proof of \eqref{E:Main1}}\label{S:proof}
We shall prove \eqref{E:Main1} by induction over the rank of the Levi form of $\rho$. To start the
induction we construct a smooth defining function $r_{0}$ of $\Omega$ which satisfies
\eqref{E:Main1} on $U_{0}\cap\Omega$ for some neighborhood $U_{0}$ of $\Sigma_{0}$.
Let $\{V_{j,0}\}_{j\in J_{0}}$, $\{V_{j,0}'\}_{j\in J_{0}}\subset\subset\mathbb{C}^{n}$ be
finite, open covers of $\Sigma_{0}$ with $V_{j,0}'\subset\subset V_{j,0}$ such that there exist smooth,
linearly independent $(1,0)$-vector fields $W_{j,0}^{\alpha}$, $\alpha\in\{1,\dots,n-1\}$,
defined on $V_{j,0}$, which
are complex tangential to $b\Omega$ on $b\Omega\cap V_{j,0}$ and satisfy:
\begin{enumerate}
\item $H_{\rho}(W_{j,0}^{\alpha},W_{j,0}^{\alpha})=0$ on $\Sigma_{0}\cap V_{j,0}$ for all
$j\in J_{0}$,
\item the span of the $W_{j,0}^{\alpha}(z)$'s contains the null space of the Levi form of
$\rho$ at $z$ for all boundary points $z$ in $\cup_{j=i+1}^{n-2}\Sigma_{j}\cap V_{j,0}$.
\end{enumerate}
We shall write $V_{0}=\cup_{j\in J_{0}}V_{j,0}$
and $V_{0}'=\cup_{j\in J_{0}}V_{j,0}'$.
Choose smooth, non-negative functions
$\zeta_{j,0}$, $j\in J_{0}$, such that
\begin{align*}
\sum_{j\in J_{0}}\zeta_{j,0}=1\;\text{on}\;V_{0}',\;\sum_{j\in J_{0}}\zeta_{j,0}\leq 1\;\text{on}\;V_{0},
\;\text{and}\;\overline{\operatorname{supp}\zeta_{j,0}}\subset V_{j,0}\;\text{for all}\;j\in J_{0}.
\end{align*}
Set $\sigma_{j,0}=\sum_{\alpha=1}^{n-1}H_{\rho}(W_{j,0}^{\alpha},W_{j,0}^{\alpha})$. For given
$\epsilon>0$, choose $C_{0}$ according to $\eqref{E:chooseC}$. Then choose $m_{j,0}$
and $\tau_{j,0}$ as in Lemma \ref{L:cutoff} such that
\begin{align*}
s_{m_{j,0},\tau_{j,0}}(z)=
\begin{cases}
\zeta_{j,0}(z)\cdot\sigma_{j,0}(z)\cdot g_{m_{j,0},\tau_{j,0}}(\sigma_{j,0}(z)) &\text{if}\;\;z\in V_{j,0}\\
0 &\text{if}\;\;z\in (V_{j,0})^{c}
\end{cases}
\end{align*}
satisfies (i)-(iv) of Lemma \ref{L:cutoff} for $\delta_{0}=\frac{\epsilon}{C_{0}|J_{0}|}$.
Finally, set $s_{0}=\sum_{j\in J_{0}}s_{m_{j,0},\tau_{j,0}}$ and
define the smooth defining function $r_{0}=\rho e^{-C_{0}s_{0}}$.
By our choice of $r_{0}$ we have for all
$q\in\Omega\cap V_{0}'$ with $\pi(q)\in\Sigma_{0}\cap V_{0}'$ that
\begin{align*}
H_{r_{0}}(W,W)(q)\geq H_{r_{0}}(W,W)(\pi(q))-\epsilon \left|r_{0}(q)\right|\cdot |W|^{2}
=-\epsilon \left|r_{0}(q)\right|\cdot |W|^{2}
\end{align*}
for all $W\in\mathbb{C}T_{\pi(q)}b\Omega$.
In fact, by continuity there exists a neighborhood $U_{0}\subset V_{0}'$ of $\Sigma_{0}$ such that
\begin{align*}
H_{r_{0}}(W,W)(q)\geq H_{r_{0}}(W,W)(\pi(q))-2\epsilon |r_{0}(q)||W|^{2}
\end{align*}
holds for all $q\in\Omega\cap U_{0}$ with $\pi(q)\in b\Omega\cap U_{0}$ and
$W\in\mathbb{C}T_{\pi(q)}b\Omega$.
Now let $\xi\in\mathbb{C}^{n}$. For each $q\in\Omega\cap U_{0}$
with $\pi(q)\in b\Omega\cap U_{0}$ we shall
write $\xi=W+M$, where $W\in\mathbb{C}T_{\pi(q)}b\Omega$ and $M$ in the span of $N(\pi(q))$.
Note that then $|\xi|^{2}=|W|^{2}+|M|^{2}$.
We get for the complex Hessian of $r_{0}$ at $q$:
\begin{align*}
H_{r_{0}}(\xi,\xi)(q)&=H_{r_{0}}(W,W)(q)+2\operatorname{Re}\left(
H_{r_{0}}(W,M)(q)
\right)
+H_{r_{0}}(M,M)(q)\\
&\geq H_{r_{0}}(W,W)(\pi(q))-2\epsilon\left|r_{0}(q)\right|\cdot |W|^{2}+
2\operatorname{Re}\left(
H_{r_{0}}(W,M)(q)
\right)
+H_{r_{0}}(M,M)(q).
\end{align*}
Note that Taylor's theorem yields
\begin{align*}
H_{r_{0}}(W,M)(q)&=H_{r_{0}}(W,M)(\pi(q))+\mathcal{O}(d_{b\Omega}(q))|W||M|\\
&=
e^{-C_{0}s_{0}(\pi(q))}\left(H_{\rho}(W,M)-C\overline{\langle\partial\rho,M\rangle}
\langle\partial s_{0},W\rangle\right)(\pi(q))+\mathcal{O}(d_{b\Omega}(q))|W||M|.
\end{align*}
It follows by property (iii) of Lemma \ref{L:cutoff} that
$|\langle\partial s_{0},W\rangle|\leq\frac{\epsilon}{C_{0}}|W|$ on $b\Omega$.
After possibly shrinking $U_{0}$ we get
\begin{align*}
2\operatorname{Re}\left(H_{r_{0}}(W,M)(\pi(q))\right)\geq-4\epsilon|\partial\rho| |W| |M|
+e^{-C_{0}s_{0}(\pi(q))}2\operatorname{Re}\left(H_{\rho}(W,M)\right)(\pi(q)).
\end{align*}
Putting the above estimates together, we have
\begin{align*}
H_{r_{0}}(\xi,\xi)(q)\geq
&-2\epsilon|r_{0}(q)||W|^{2}-4\epsilon|\partial\rho||W||M|+H_{r_{0}}(M,M)(q)\\
&+e^{-C_{0}s_{0}(\pi(q))}
\left[
H_{\rho}(W,W)(\pi(q))+2\operatorname{Re}\left(H_{\rho}(W,M)(\pi(q))\right)
\right].
\end{align*}
An application of the (sc)-(lc) inequality yields
\begin{align*}
-\epsilon |W||M|\geq-\epsilon\left(|r_{0}(q)||\xi|^{2}
+\frac{1}{|r_{0}(q)|}\left|\langle\partial r_{0}(\pi(q)),\xi\rangle\right|^{2}
\right),
\end{align*}
where we used that $|\xi|^{2}=|W|^{2}+|M|^{2}$. Taylor's
theorem also gives us that
\begin{align*}
\left|\langle\partial r_{0}(\pi(q)),\xi\rangle\right|
&=e^{-C_{0}s_{0}(\pi(q))}\cdot\left|
\langle\partial\rho(\pi(q)),\xi\rangle
\right|\\
&\leq e^{-C_{0}s_{0}(\pi(q))}\cdot\left(
\left|\langle\partial\rho(q),\xi\rangle\right|+
\mathcal{O}\left(\rho(q)\right)|\xi|
\right),
\end{align*}
where the constant in the last term is independent of $\epsilon$. After possibly shrinking $U_{0}$, we
obtain for all $q\in\Omega\cap U_{0}$
\begin{align*}
\left|\langle\partial r_{0}(\pi(q)),\xi\rangle\right|
&\leq 2e^{-C_{0}s_{0}(q)}\cdot\left|\langle\partial\rho(q),\xi\rangle\right|
+\mathcal{O}\left(r_{0}(q)\right)|\xi|\\
&\leq 2\left|\langle\partial r_{0}(q),\xi\rangle\right|+2\left|\rho(q)\right|
\cdot\left|\langle\partial e^{-C_{0}s_{0}(q)},\xi\rangle\right|
+\mathcal{O}\left(r_{0}(q)\right)|\xi|,
\end{align*}
where, again, the constant in the last term is independent of $\epsilon$. Using that
$\left|\langle\partial s_{0},W\rangle\right|\leq\frac{\epsilon}{C_{0}}|W|$ on $b\Omega$, we get
\begin{align*}
\left|\langle\partial e^{-C_{0}s_{0}(q)},\xi\rangle\right|
\leq 2\epsilon |W|+\mathcal{O}\left( \left|\langle\partial\rho(\pi(q)),\xi\rangle \right| \right),
\end{align*}
which implies that
$|\langle\partial r_{0}(\pi(q)),\xi\rangle|\lesssim|\langle\partial r_{0}(q),\xi\rangle|+
|r_{0}(q)||\xi|$. Thus we have
\begin{align*}
-\epsilon|W||M|\gtrsim-\epsilon\left(|r_{0}(q)||\xi|^{2}
+\frac{1}{|r_{0}(q)|}\left|\langle\partial r_{0}(q),\xi\rangle\right|^{2}
\right).
\end{align*}
Since
\begin{align*}
H_{\rho}(W,W)(\pi(q))+2\operatorname{Re}(H_{\rho}(W,M)(\pi(q))=H_{\rho}(\xi,\xi)(\pi(q))-H_{\rho}(M,M)(\pi(q)),
\end{align*}
and $H_{\rho}(\xi,\xi)(\pi(q))$ is non-negative, it follows that
\begin{align*}
H_{r_{0}}(\xi,\xi)(q)\gtrsim -\epsilon\left(|r_{0}(q)||\xi|^{2}+\frac{1}{|r_{0}(q)|}
\left|\langle\partial r_{0}(q),\xi\rangle\right|^{2}\right)
+\mu_{0}H_{\rho}(\xi,\xi)(\pi(q))
\end{align*}
for some positive constant $\mu_{0}$.
Since the constants in $\gtrsim$ do not depend on the choice of $\epsilon$, this
proves \eqref{E:Main1} in an open neighborhood $U_{0}$ of $\Sigma_{0}$.
Let $l\in\{0,\dots,n-3\}$ be fixed and suppose that there exist a smooth defining function
$r_{l}$ of $b\Omega$ and an open neighborhood $U_{l}\subset\mathbb{C}^{n}$ of
$\cup_{i=0}^{l}\Sigma_{i}$ such that
\begin{align}\label{E:ihypotheses1}
H_{r_{l}}(\xi,\xi)(q)\geq -\epsilon
\left(
|r_{l}(q)||\xi|^{2}+\frac{1}{|r_{l}(q)|}\left|\langle\partial r_{l}(q),\xi\rangle\right|^{2}
\right)+\mu_{l}H_{\rho}(\xi,\xi)(\pi(q))
\end{align}
for all $q\in\Omega\cap U_{l}$ with $\pi(q)\in b\Omega\cap U_{l}$ and $\xi\in\mathbb{C}^{n}$. Here,
$\mu_{l}$ is some positive constant.
Furthermore, we suppose that the function $\vartheta_{l}$ defined by $r_{l}=\rho e^{-\vartheta_{l}}$
satisfies
the following
\begin{align}
\left|
\langle\partial\vartheta_{l}(z),T\rangle\right|&\leq\epsilon|T|\;\;\text{for all}\;\;z\in b\Omega, T\in
\mathbb{C}T_{z}b\Omega\label{E:ihypothesis2}\\
H_{\vartheta_{l}}(T,T)(z)&\geq-\epsilon|T|^{2}\;\;\text{for all}\;\; z\in\cup_{j=l+1}^{n-2}\Sigma_{j},
T\in\mathbb{C}T_{z}b\Omega\;\;\text{with}\;\;H_{\rho}(T,T)(z)=0\label{E:ihypothesis3}.
\end{align}
Let $k=l+1$. We shall now show that there exist a smooth defining function $r_{k}$ and
a neighborhood $U_{k}$ of $\cup_{i=0}^{k}\Sigma_{i}$ such that for some positive constant
$\mu_{k}$
\begin{align}\label{E:claimistep}
H_{r_{k}}(\xi,\xi)(q)\geq -\epsilon\left(
|r_{k}(q)||\xi|^{2}+\frac{1}{|r_{k}(q)|}\left|
\langle\partial r_{k}(q),\xi\rangle
\right|^{2}
\right)+\mu_{k}H_{\rho}(\xi,\xi)(\pi(q))
\end{align}
holds for all $q\in \Omega\cap U_{k}$ with $\pi(q)\in b\Omega\cap U_{k}$ and $\xi\in\mathbb{C}^{n}$.
Let $\{V_{j,k}\}_{j\in J_{k}}$, $\{V_{j,k}'\}_{j\in J_{k}}$ be finite, open covers of $\Sigma_{k}\setminus
U_{k-1}$ such that
\begin{enumerate}
\item[(1)] $V_{j,k}'\subset\subset V_{j,k}$ and $\overline{V}_{j,k}
\cap\left(\cup_{i=0}^{k-1}\Sigma_{i}\right)=\emptyset$
for all $j\in J_{k}$, and
\item[(2)] there exist smooth, linearly
independent $(1,0)$-vector fields $W_{j,k}^{\alpha}$, $\alpha\in\{1,\dots,n-1-k\}$,
and $S_{j,k}^{\beta}$, $\beta\in\{1,\dots,k\}$, defined on $V_{j,k}$, which are
complex tangential to $b\Omega$ on $b\Omega\cap V_{j,k}$ and satisfy the following:
\begin{enumerate}
\item[(a)] $H_{\rho}(W_{j,k}^{\alpha},W_{j,k}^{\alpha})=0$ on $\Sigma_{k}\cap V_{j,k},\;
\alpha\in\{1,\dots,n-1-k\},\;j\in J_{k}$,
\item[(b)] the span of $\{W_{j,k}^{1}(z),\dots,W_{j,k}^{k}(z)\}$ contains the null space of the Levi form
of $\rho$ at all boundary points $z$ belonging to $\cup_{i=k+1}^{n-2}\Sigma_{i}\cap V_{j,k}$,
\item[(c)] $H_{\rho}(S_{j,k}^{\beta}, S_{j,k}^{\beta})> 0$ on $b\Omega\cap \overline{V_{j,k}},\;
\beta\in\{1,\dots,k\},\;j\in J_{k}$,
\item[(d)] $H_{\rho}(S_{j,k}^{\beta},S_{j,k}^{\tilde{\beta}})=0$ for $\beta\neq\tilde{\beta}$ on
$b\Omega\cap V_{j,k}$, $\beta,\tilde{\beta}\in\{1,\dots,k\},\;j\in J_{k}$.
\end{enumerate}
\end{enumerate}
Note that above vector fields $\{W_{j,k}^{\alpha}\}$
always exist in some neighborhood of a given point in
$\Sigma_{k}$. However, we might not be able to cover $\Sigma_{k}$ with finitely many such
neighborhoods, when the closure of $\Sigma_{k}$ contains boundary points at which the Levi
form is of lower rank. Moreover, if the latter is the case, then (c) above is also impossible. These are the
reasons for proving \eqref{E:Main1} via induction over the rank of the Levi form of
$\rho$.
Suppose $S$ is in the span of $\{S_{j,k}^{\beta}(z)\}_{\beta=1}^{k}$ and
$W$ is in the span of $\{W_{j,k}^{\alpha}(z)\}_{\alpha=1}^{n-1-k}$
for some $z\in V_{j,k}$, then there is some constant
$\kappa_{k}>0$ such that $|S|^{2}+|W|^{2}\leq\kappa_{k}|S+W|^{2}$ for all $j\in J_{k}$.
We shall write $V_{k}=\cup_{j\in J_{k}}V_{j,k}$ and $V_{k}'=\cup_{j\in J_{k}}V_{j,k}'$.
Let $\zeta_{j,k}$ be non-negative, smooth functions such that
\begin{align*}
\sum_{j\in J_{k}}\zeta_{j,k}=1\;\text{on}\;V_{k}',\;\sum_{j\in J_{k}}\zeta_{j,k}\leq 1\;
\text{on}\;V_{k}\;\text{and}\;\overline{\operatorname{supp}\zeta_{j,k}}\subset V_{j,k}.
\end{align*}
Set $\sigma_{j,k}=\sum_{\alpha=1}^{n-k-1}H_{\rho}(W_{j,k}^{\alpha},W_{j,k}^{\alpha})$. Recall that
$\epsilon>0$ is given. Choose $C_{k}$ according to \eqref{E:chooseC} with
$\frac{\epsilon}{\kappa_{k}}$ in place of $\epsilon$ there.
We now choose $m_{j,k},\;\tau_{j,k}>0$ such that
\begin{align*}
s_{m_{j,k},\tau_{j,k}}(z)=
\begin{cases}
\zeta_{j,k}(z)\cdot\sigma_{j,k}(z)\cdot g_{m_{j,k}\tau_{j,k}}(\sigma_{j,k}(z))&\text{if}\;
z\in V_{j,k}\\
0&\text{if}\;z\in (V_{j,k})^{c}
\end{cases}
\end{align*}
satisfies (i)-(iv) of Lemma \ref{L:cutoff} with
$\delta_{k}=\frac{\epsilon}{C_{k}|J_{k}|\kappa_{k}}$. Set $s_{k}=\sum_{j\in J_{k}}s_{m_{j,k},\tau_{j,k}}$
and define the smooth defining function
$r_{k}=r_{k-1}e^{-Cs_{k}}$. We claim that this choice of $r_{k}$ satisfies
\eqref{E:claimistep}.
We shall first see that \eqref{E:claimistep} is true for all $q\in\Omega\cap U_{k-1}\cap V_{k}$
with $\pi(q)\in b\Omega\cap U_{k-1}\cap V_{k}$.
A straightforward computation yields
\begin{align}\label{E:istepint1}
H_{r_{k}}(\xi,\xi)(q)=
e^{-C_{k}s_{k}(q)}
\biggl[
H_{r_{k-1}}(\xi,\xi)\biggr.&+r_{k-1}\left(
C_{k}^{2}\left|\langle\partial s_{k},\xi\rangle\right|^{2}-C_{k}H_{s_{k}}(\xi,\xi)
\right)\\
\biggl.&-2C_{k}\operatorname{Re}\left(
\overline{\langle\partial r_{k-1},\xi\rangle}\langle\partial s_{k},\xi\rangle
\right)
\biggr](q).\notag
\end{align}
By induction hypothesis \eqref{E:ihypotheses1} we have good control over the first term in
\eqref{E:istepint1}:
\begin{align*}
e^{-C_{k}s_{k}(q)}H_{r_{k-1}}(\xi,\xi)(q)\geq
&-\epsilon\left(
|r_{k}(q)||\xi|^{2}+\frac{1}{|r_{k}(q)|}\left|\langle
e^{-C_{k}s_{k}(q)}\partial r_{k-1}(q),\xi\rangle
\right|^{2}
\right)\\
&+e^{-C_{k}s_{k}(q)}\mu_{k-1}H_{\rho}(\xi,\xi)(\pi(q)).
\end{align*}
Note that
\begin{align*}
\left|
\langle
e^{-C_{k}s_{k}(q)}\partial r_{k-1}(q),\xi\rangle
\right|^{2}
\leq
2\left|
\langle\partial r_{k}(q),\xi\rangle
\right|^{2}
+r_{k}^{2}(q)C_{k}^{2}\left|
\langle\partial s_{k}(q),\xi\rangle
\right|^{2}.
\end{align*}
Moreover, part (iii) of Lemma \ref{L:cutoff} implies that
\begin{align}\label{E:istepint2}
C_{k}^{2}\left|
\langle\partial s_{k}(q),\xi\rangle
\right|^{2}
\leq
2\epsilon|\xi|^{2}+\mathcal{O}\left(\left|
\langle\partial r_{k}(\pi(q)),\xi\rangle
\right|^{2}\right)
\leq 3\epsilon|\xi|^{2}+\mathcal{O}\left(\left|
\langle\partial r_{k}(q),\xi\rangle
\right|^{2}\right)
\end{align}
after possibly shrinking $U_{k-1}$ (in normal direction only). Thus
we have
\begin{align*}
e^{-C_{k}s_{k}(q)}H_{r_{k-1}}(\xi,\xi)(q)\gtrsim
-\epsilon\left(
|r_{k}(q)||\xi|^{2}+\frac{1}{|r_{k}(q)|}\left|\langle
\partial r_{k}(q),\xi\rangle
\right|^{2}
\right)+\mu H_{\rho}(\xi,\xi)(\pi(q)).
\end{align*}
For some positive constant $\mu\leq\mu_{k-1}$. Thus the first term on the right hand side of
\eqref{E:istepint1} is taken care of.
Now suppose $q\in\Omega\cap U_{k-1}\cap V_{k}$ is such that $\pi(q)\in b\Omega\cap V_{j,k}$ for
some $j\in J_{k}$. To be able to deal with the term $H_{s_{k}}(\xi,\xi)(q)$ in \eqref{E:istepint1}, we shall
write $\xi=S+W+M$, where
\begin{align*}
S\in\operatorname{span}\left(\{S_{j,k}^{\beta}(\pi(q))\}_{\beta=1}^{k} \right),\;\;
W\in\operatorname{span}\left(\{W_{j,k}^{\alpha}(\pi(q))\}_{\alpha=1}^{n-1-k}\right),\;\;\text{and}\;\;
M\in\operatorname{span}\left(N(\pi(q))\right).
\end{align*}
Then the (sc)-(lc) inequality gives
\begin{align*}
C_{k}H_{s_{k}}(\xi,\xi)(q)&\geq C_{k}H_{s_{k}}(W,W)(q)-\frac{\epsilon}{\kappa_{k}}|W|^{2}+
\mathcal{O}\left(|S|^{2}+|M|^{2}\right)\\
&\geq -2\frac{\epsilon}{\kappa_{k}}|W|^{2}+\mathcal{O}\left(|S|^{2}+|M|^{2}\right),
\end{align*}
where the last step holds since $s_{k}$ satisfies part (iv) of Lemma \ref{L:cutoff}.
The last inequality together with \eqref{E:istepint2} lets us estimate the second term
in \eqref{E:istepint1} as follows
\begin{align*}
e^{-C_{k}s_{k}(q)}r_{k-1}(q)&\left(
C_{k}^{2}\left|\langle\partial s_{k},\xi\rangle\right|^{2}-C_{k}H_{s_{k}}(\xi,\xi)
\right)(q)\\
&\gtrsim
-\epsilon
\left(
|r_{k}||\xi|^{2}+\frac{1}{|r_{k}|}\left|\langle
\partial r_{k},\xi\rangle
\right|^{2}
\right)(q)+\mathcal{O}(r_{k}(q))|S|^{2}.
\end{align*}
For the third term in \eqref{E:istepint1} we use \eqref{E:istepint2} again and obtain
\begin{align*}
-2C_{k}e^{-C_{k}s_{k}(q)}\operatorname{Re}\left(
\overline{\langle\partial r_{k-1},\xi\rangle}\langle \partial s_{k},\xi\rangle
\right)(q)
\gtrsim
-\epsilon\left(|r_{k}(q)||\xi|^{2}+\frac{1}{|r_{k}(q)|}\left|\langle\partial r_{k-1}(q),\xi\rangle
\right|^{2}\right).
\end{align*}
Collecting all these estimates, using $\frac{1}{\kappa_{k}}|W|^{2}\leq|\xi|^{2}$, we now have
for $q\in\Omega\cap U_{k-1}\cap V_{k}$ with $\pi(q)\in b\Omega\cap U_{k-1}\cap V_{k}$ and for some
$\mu>0$
\begin{align*}
H_{r_{k}}(\xi,\xi)(q)\gtrsim
&-\epsilon
\left(|r_{k}(q)||\xi|^{2}+\frac{1}{|r_{k}(q)|}\left|\langle\partial r_{k-1}(q),\xi\rangle
\right|^{2}\right)\\
&\hspace{4cm}+\mu H_{\rho}(\xi,\xi)(\pi(q))+\mathcal{O}(r_{k}(q))|S|^{2}\\
\gtrsim&-\epsilon
\left(|r_{k}(q)||\xi|^{2}+\frac{1}{|r_{k}(q)|}\left|\langle\partial r_{k-1}(q),\xi\rangle
\right|^{2}\right)
+\frac{\mu}{2}H_{\rho}(\xi,\xi)(\pi(q)).
\end{align*}
Here, the last estimate holds, after possibly shrinking $U_{k-1}\cap V_{k}$ (in normal direction only),
since $\rho$ is plurisubharmonic on $b\Omega$.
We still need to show that \eqref{E:claimistep} is true in some neighborhood of
$\Sigma_{k}\setminus U_{k-1}$. Let $U\subset V_{k}$ be a neighborhood of
$\Sigma_{k}\setminus U_{k-1}$, $q\in \Omega\cap U$ with $\pi(q)\in b\Omega\cap V_{j,k}$ for
some $j\in J_{k}$
and $\xi\in\mathbb{C}^{n}$. Writing $\vartheta_{k}=\vartheta_{k-1}+C_{k}s_{k}$, we get
\begin{align*}
H_{r_{k}}(\xi,\xi)(q)
=e^{-\vartheta_{k}(q)}
\Bigl[
H_{\rho}(\xi,\xi)-&2\operatorname{Re}\left(
\overline{\langle\partial\rho,\xi\rangle}
\langle\partial\vartheta_{k},\xi\rangle
\right)\Bigl.\\
&\Bigr.+\rho\left(
\left|\langle\partial\vartheta_{k},\xi\rangle \right|^{2}
-H_{\vartheta_{k-1}}(\xi,\xi)-C_{k}H_{s_{k}}(\xi,\xi)
\right)
\Bigr](q)\\
&=\operatorname{I}+\operatorname{II}+\operatorname{III}+\operatorname{IV}+\operatorname{V}.
\end{align*}
We write again $\xi=S+W+M$.
By construction of $s_{k}$ and by the induction hypotheses \eqref{E:ihypothesis2} and
\eqref{E:ihypothesis3} on
$\vartheta_{k-1}$, we can do estimates similar to the ones below \eqref{E:istepint1} to obtain
\begin{align*}
\operatorname{II}+\operatorname{III}+\operatorname{IV}\gtrsim
-\epsilon\left(
|r_{k}(q)||\xi|^{2}+\frac{1}{|r_{k}(q)|}\left|
\langle\partial r_{k}(q),\xi\rangle
\right|^{2}
\right)+\mathcal{O}(r_{k}(q))|S|^{2}\;\;\text{for}\;\;q\in U.
\end{align*}
So we are left with the terms $\operatorname{I}$ and $\operatorname{V}$.
Let us first consider the term $\operatorname{I}$. By Taylor's Theorem, we have
\begin{align*}
e^{-\vartheta_{k}(q)}H_{\rho}(\xi,\xi)(q)
=
e^{-\vartheta_{k}(q)}&\Bigl(H_{\rho}(\xi,\xi)(\pi(q))
-d_{b\Omega}(q)\operatorname{Re}\bigl[ (N
H_{\rho})(\xi,\xi)(\pi(q))\bigr]\Bigr)\\
&+\mathcal{O}\left(r_{k}^{2}(q)\right)|\xi|^{2}.
\end{align*}
By the (sc)-(lc) inequality we have
\begin{align*}
\operatorname{Re} \left[(NH_{\rho})(\xi,\xi)(\pi(q))\right]
\leq &\operatorname{Re}\left[(NH_{\rho})(W,W)(\pi(q))\right]\\
&+\frac{\epsilon}{c_{1}\kappa_{k}}|W|^{2}+
\mathcal{O}\left(\frac{c_{1}\kappa_{k}}{\epsilon}\right)(|S|^{2}+|M|^{2}),
\end{align*}
where $c_{1}>0$ is such that $d_{b\Omega}(q)\leq c_{1}|\rho(q)|$.
Therefore we obtain for some $\mu>0$
\begin{align*}
e^{-\vartheta_{k}(q)}H_{\rho}(\xi,\xi)(q)
\geq
&-d_{b\Omega}(q)e^{-\vartheta_{k}(q)}\operatorname{Re}\left[(NH_{\rho})(W,W)(\pi(q))\right]
+r_{k}(q)\frac{\epsilon}{\kappa_{k}}|W|^{2}\\
&+\mu H_{\rho}(\xi,\xi)(\pi(q))+\mathcal{O}\left(r_{k}(q)\right)(|S|^{2}+|M|^{2})
+\mathcal{O}\left(r_{k}^{2}(q)\right)|\xi|^{2}.
\end{align*}
To estimate term $\operatorname{V}$ we use (sc)-(lc) inequality again:
\begin{align*}
-r_{k}(q)C_{k}H_{s_{k}}(\xi,\xi)(q)
\geq -r_{k}(q)\left(C_{k}H_{s_{k}}(W,W)(q)-\frac{\epsilon}{\kappa_{k}}|W|^{2}\right)+
\mathcal{O}\left(r_{k}(q)\right)(|S|^{2}+|M|^{2}).
\end{align*}
After possibly shrinking $U$, we get for some $\mu>0$
\begin{align*}
\operatorname{I}+\operatorname{V}\geq
&
-d_{b\Omega}(q)e^{-\vartheta_{k}(q)}\operatorname{Re}\left[(NH_{\rho})(W,W)(\pi(q))\right]
+r_{k}(q)\left(-C_{k}H_{s_{k}}(W,W)(q)+\frac{2\epsilon}{\kappa_{k}}|W|^{2}
\right)\\
&+\mu H_{\rho}(\xi,\xi)(\pi(q))+\mathcal{O}\left(r_{k}(q)\right)|M|^{2}
+\mathcal{O}\left(r_{k}^{2}(q)\right)|\xi|^{2}.
\end{align*}
By our choice of $s_{k}$ and $C_{k}$, it follows that for all $q\in\Omega\cap U$ with
$\pi(q)\in b\Omega\cap U$ we have
\begin{align*}
-d_{b\Omega}e^{-\vartheta_{k}(q)}\operatorname{Re}\left[ (NH_{\rho})(W,W)(\pi(q))\right]
-r_{k}(q)C_{k}H_{s_{k}}(W,W)(q)\geq\frac{\epsilon}{\kappa_{k}}r_{k}(q)|W|^{2}.
\end{align*}
Putting our estimates for the terms $\operatorname{I}$--$\operatorname{V}$ together and letting
$U_{k}$ be the union of $U_{k-1}$ and
$U$, we obtain: for all $q\in\Omega\cap U_{k}$ with $\pi(q)\in b\Omega\cap U_{k}$, the
function $r_{k}$ satisfies
\begin{align*}
H_{r_{k}}(\xi,\xi)(q)\gtrsim -\epsilon\left(
|r_{k}(q)||\xi|^{2}+\frac{1}{|r_{k}(q)|}\left|
\langle\partial r_{k}(q),\xi\rangle
\right|^{2}
\right)+\mu_{k}H_{\rho}(\xi,\xi)(\pi(q))
\end{align*}
for some $\mu_{k}>0$.
Since the constants in $\gtrsim$ do
not depend on $\epsilon$ or on any other parameters which come
up in the construction of $r_{k}$, \eqref{E:claimistep} follows. Moreover, by construction, $\vartheta_{k}$
satisfies \eqref{E:ihypothesis2} and \eqref{E:ihypothesis3}.
Note that $\Sigma_{n-1}\setminus U_{n-2}$ is
a closed subset of the set of strictly pseudoconvex boundary points. Thus for
any smooth defining function $r$ there is some
neighborhood $U$ of $\Sigma_{n-1}\setminus U_{n-2}$ such that
\begin{align*}
H_{r}(\xi,\xi)\gtrsim |\xi|^{2}
+\mathcal{O}(|\langle\partial r(q),\xi\rangle|^{2})\;\;\text{for all}\;\;\xi\in\mathbb{C}^{n}
\end{align*}
holds on $U\cap\Omega$.
This concludes the proof of \eqref{E:Main1}.
The proof of \eqref{E:Main2} is essentially the same as the one of \eqref{E:Main1} except that a few
signs change. That is, the basic estimate \eqref{E:BasicTaylorH} for $q\in\overline{\Omega}^{c}\cap U$
becomes
\begin{align*}
H_{\rho}(W,W)(q)=2d_{b\Omega}(q)NH_{\rho}(W,W)(\pi(q))+\mathcal{O}\left(d_{b\Omega}^{2}(q)
\right)|W|^{2}
\end{align*}
for any vector $W\in\mathbb{C}^{n}$ which is a weak complex tangential direction at $\pi(q)$.
So an obstruction for \eqref{E:Main2} to hold at $q\in\overline{\Omega}^{c}\cap U$
occurs when $NH_{\rho}(W,W)$ is negative at $\pi(q)$ --
note that this happens exactly when we have no problem with \eqref{E:Main1}.
Since the obstruction terms
to \eqref{E:Main1} and \eqref{E:Main2} only differ by a sign, one would expect that the necessary
modifications of
$\rho$ also just differ by a sign. In fact, let $\vartheta_{n-2}$ be as in the proof of \eqref{E:Main1} --
that is, $r_{1}=\rho e^{-\vartheta_{n-2}}$ satisfies \eqref{E:Main1} for a given $\epsilon>0$. Then
$r_{2}=\rho e^{\vartheta_{n-2}}$ satisfies \eqref{E:Main2} for the same $\epsilon$.
\section{Proof of Corollary \ref{C:DF}}\label{S:DF}
We shall now prove Corollary \ref{C:DF}. We begin with part (i) by
showing first that for
any $\eta\in(0,1)$ there exist a $\delta>0$, a smooth defining function $r$ of $\Omega$ and a
neighborhood $U$ of $b\Omega$ such that
$h=-(-re^{-\delta|z|^{2}})^{\eta}$ is strictly plurisubharmonic on $\Omega\cap U$.
Let $\eta\in(0,1)$ be fixed, and $r$ be a smooth defining function of $\Omega$. For notational ease
we write $\phi(z)=\delta|z|^{2}$ for $\delta>0$. Here, $r$ and $\delta$ are fixed and to be chosen later.
Let us compute the complex Hessian of $h$ on $\Omega\cap U$:
\begin{align*}
H_{h}(\xi,\xi)=&\eta(-r)^{\eta-2}e^{-\phi\eta}
\Bigl[
(1-\eta)\Bigr.\left|\langle\partial r,\xi\rangle\right|^{2}-rH_{r}(\xi,\xi)\\
&+2r\eta \operatorname{Re}\left(\langle\partial r,\xi\rangle\langle\overline{\partial\phi,\xi}\rangle\right)
\Bigl.
-r^{2}\eta\left|\langle\partial\phi,\xi\rangle\right|^{2}
+r^{2}H_{\phi}(\xi,\xi)\Bigr].
\end{align*}
An application of the (sc)-(lc) inequality gives
\begin{align*}
2r\eta \operatorname{Re}\left(\langle\partial r,\xi\rangle\langle\overline{\partial\phi,\xi}\rangle\right)
\geq -\frac{1-\eta}{2}\left|\langle\partial r,\xi\rangle\right|^{2}
-\frac{2r^{2}\eta^{2}}{1-\eta}\left|\langle\partial\phi,\xi\rangle\right|^{2}.
\end{align*}
Therefore, we obtain for the complex Hessian of $h$ on $\Omega$ the following:
\begin{align*}
H_{h}(\xi,\xi)
\geq
\eta(-r)^{\eta-2}e^{-\phi\eta}
\Biggl[\frac{1-\eta}{2}|\langle\partial r,\xi\rangle|^{2}
\Biggr.&-rH_{r}(\xi,\xi)\\
&+r^{2}
\left\{
-\frac{\eta(1+\eta)}{1-\eta}|\langle\partial\phi,\xi\rangle|^{2}+H_{\phi}(\xi,\xi)
\right\}
\Biggl.\Biggr].
\end{align*}
Set $\delta=\frac{1-\eta}{2\eta(1+\eta)D}$, where $D=\max_{z\in\overline{\Omega}}|z|^{2}$. Then we
get
\begin{align*}
H_{\phi}(\xi,\xi)-\frac{\eta(1+\eta)}{1-\eta}\left|\langle\partial\phi,\xi\rangle\right|^{2}
=
\delta\left(H_{|z|^{2}}(\xi,\xi)-\frac{\eta(1+\eta)}{1-\eta}\delta\left|\langle\overline{z},
\xi\rangle\right|^{2}
\right)
\geq
\frac{\delta}{2}|\xi|^{2}.
\end{align*}
This implies that
\begin{align}\label{E:generalDFest}
H_{h}(\xi,\xi)\geq \eta(-r)^{\eta-2}e^{-\phi\eta}
\left[
\frac{1-\eta}{2}|\langle \partial r,\xi\rangle|^{2}
-rH_{r}(\xi,\xi)+\frac{\delta}{2}r^{2}|\xi|^{2}
\right]
\end{align}
holds on $\Omega$.
Set $\epsilon=\min\{\frac{1-\eta}{4},\frac{1-\eta}{8\eta(1+\eta)D}\}$.
By \eqref{E:Main1}
there exist a neighborhood $U$ of $b\Omega$ and a smooth defining function $r_{1}$
of $\Omega$ such that
\begin{align*}
H_{r_{1}}(\xi,\xi)(q)
\geq
-\epsilon\left(|r_{1}(q)||\xi|^{2}+\frac{1}{|r_{1}(q)|}|\langle\partial r_{1}(q),\xi\rangle|^{2}\right)
\end{align*}
holds for all $q\in\Omega\cap U$, $\xi\in\mathbb{C}^{n}$.
Setting $r=r_{1}$ and using \eqref{E:generalDFest}, we obtain
\begin{align*}
H_{h}(\xi,\xi)(q)\geq
\eta(-h(q))\cdot\epsilon|\xi|^{2}\;\;\text{for}\;\;q\in\Omega\cap U,\;\xi\in\mathbb{C}^{n}.
\end{align*}
It follows by standard arguments that there exists a defining function $\widetilde{r}_{1}$ such that
$-(-\widetilde{r}_{1})^{\eta}$ is strictly
plurisubharmonic on $\Omega$; for details see pg.\ 133 in \cite{DF1}.
This proves part (i) of Corollary \ref{C:DF}.
A proof similar to the one of part (i), using \eqref{E:Main2},
shows that for each $\eta>1$
there exists a smooth
defining function $\tilde{r}_{2}$, a neighborhood $U$ of $b\Omega$ and $\delta>0$ such that
$(r_{2}e^{\delta|z|^{2}})^{\eta}$ is strictly plurisubharmonic on $\overline{\Omega}^{c}\cap U$.
\end{document}
|
arXiv
|
{
"id": "0710.4165.tex",
"language_detection_score": 0.5958973169326782,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{\LARGE \bf
A Structured Optimal Controller with Feed-Forward for Transportation
}
\begin{abstract} We study an optimal control problem for a simple transportation model on a \change{path} graph. We give a closed form solution for the optimal controller, which can also account for planned disturbances using feed-forward. The optimal controller is highly structured, which allows the controller to be implemented using only local communication, conducted through two sweeps through the graph. \end{abstract}
\section{Introduction} In this paper we study a simple Linear Quadratic control transportation problem on a network. Such problems have well known solutions based on the Riccati equation \cite{kalman1960}. This gives a static feedback law $$ u = Kx, $$ where $u$ is the input to the system, $x$ is the state of the system and $K$ is a matrix with real entries. This matrix is in general dense. This is undesirable in large-scale problems, since it implies that measurements from the entire network are required to compute the optimal inputs at every node. Furthermore a centralized coordinator with knowledge of the entire system is required to determine the matrix $K$, and a complete redesign will be required in response to any changes in the network.
These factors have led to the development of a range of general purpose methods for structured control system design. Some notable themes include the notion of Quadratic Invariance \cite{RL06,LL11}, System Level Synthesis \cite{WMD18}, and the use of large-scale optimization techniques (e.g. \cite{LFJ13}). A downside with these approaches is that they improve scalability at the expense of performance. That is they search over families of controllers that exclude the dense optimal controller for \eqref{eq:problem}. While in comparison with the alternative this may be an acceptable trade-off, it implicitly assumes that just because the feedback law is dense, it cannot be efficiently implemented.
The main result of this paper is to show that the simple structure in our problem allows the optimal control law to be computed and implemented in a simple and scalable manner. The resulting control actions are the same as those from a Riccati approach, and could in principle be calculated that way. However, there are extra structural features in the control law that are obscured by the resulting dense feedback matrix representation, and it is not obvious how to exploit these to give a scalable implementation from the gain matrix obtained from the Riccati equation.
\subsection{Problem Formulation} We consider the problem of transportation and production of goods on a directed path graph
with vertices $v_1,v_2,\ldots{},v_N$ and directed edges $(e_N,e_{N-1}),\ldots{},(e_2,e_1)$. The dynamics are given by \begin{equation}\label{eq:dynamics}
z_i[t+1] = z_i[t] - u_{i-1}[t] + u_{i}[t-\tau_{i}] + v_i[t] + d_i[t]. \end{equation} All the variables are considered to be defined relative to some equilibrium. In the above $z_i[t] \in \mathds{R}$ is the quantity in node $i$ at time $t$. The system can be controlled using the variables $u_i[t]\in \mathds{R}$ and $v_i[t]\in \mathds{R}$. The variable $u_i[t]$ denotes the amount of the quantity that is transported from node $i+1$ to node $i$ (again relative to some equilibrium flows), and the transportation takes $\tau_{i}$ time units. For the last node $N$ it is assumed that $u_N[t] = 0$ for all $t$. The variable $v_i[t]$ denotes the flexible production or consumption of the quantity at the \emph{i}th node. Finally $d_i[t]\in \mathds{R}$ is the fixed production/consumption at the \emph{i}th node. This will be treated like a forecast, or \emph{planned disturbance}, that is known to the designer, but cannot be changed. This model could for instance describe a water irrigation network \cite{cantoni2007control} or a simple supply chain system \cite{inventory}. A state-space representation for \eqref{eq:dynamics} can be obtained by setting $z_i[t]$, $u_i[t-\delta],\ 1\leq\delta\leq\tau_i$ to equal the system state.
The goal is to optimally operate this network around some equilibrium point. The performance is measured by the cost of deviating from the equilibrium levels $q_iz_i^2$ and the cost of the variable production $r_iv_i^2$, where $q$ and $r$ are strictly positive constants. We thus consider the following linear quadratic control problem on a graph with $N$ nodes, \begin{equation}\label{eq:problem}\begin{aligned}
\minimize_{z,u,v} \quad & \sum_{t=0}^\infty\sum_{i=1}^N \big(q_iz_i[t]^2 + r_iv_i[t]^2 \big)\\
\st \quad & \text{dynamics in \eqref{eq:dynamics}}\\
& z[0], d_i[t]. \end{aligned}\end{equation} Note that there is no penalty on the internal flows $u_i$. This can for example be motivated by the transportation costs already being covered by the costs of the nominal flows (or in the case of water irrigation networks that gravity does the moving). This problem is in effect a dynamic extension of the types of scheduling problems considered in transportation networks \cite{ahuja1995applications}, and could be used to compliment such approaches by optimally adjusting a nominal schedule in real time using the feedback principle.
A similar problem has been studied in a previous paper \cite{heyden2018structured}. However we give several important extensions, in that we allow for non-homogeneous delays, production in every node and optimal feed-forward for planned disturbances. Allowing for non homogeneous delays is important as that will be the case for almost all applications. Taking planned disturbances into account allows for increased performance whenever such disturbances can be forecast. Finally, allowing for some variation in the consumption $v_i$ for each node will also generally increase performance whenever such variation is possible. The effect the feed-forward of planned disturbances can have on the controller performance is illustrated in \cref{fig:example}, where we see that the controller with feed-forward anticipates the action of the disturbances, allowing the effect to be better spread through the graph and the node levels to be more tightly regulated. This results in a significant improvement in performance.
\begin{figure}
\caption{Example of the effect of feed forward. The graph has five nodes and transportation delay $\tau_1 = 3$, $\tau_2 = 2$, $\tau_3 = 5$ and $\tau_4 =4$. There is a disturbance in node three from time 10 to 13 and in node 2 from time 12 to 15. We can see that the feed-forward manages to handle the disturbances better by spreading out their effect throughout the graph.
To quantify the difference one can consider the cost in \eqref{eq:problem}, which is $3.11$ with feed-forward and 11.35 without feed-forward.}
\label{fig:example}
\end{figure}
\subsection{Result Preview} The key structural feature that we identify in the optimal control law for \eqref{eq:problem} is that optimal inputs can be computed recursively by two sweeps through the graph (even though the control law that would be obtained from the Riccati equation would be dense). More specifically, two intermediate variables local to the \emph{i}th node $\delta_i[t]$ and $\mu_i[t]$ can be computed recursively through relationships on the form \[ \begin{aligned}
\delta_i[t]&=f(\text{local\_variables},\delta_{i-1}),\\
\mu_i[t]&=g(\text{local\_variables},\mu_{i+1}), \end{aligned} \] from which the optimal inputs $u_i[t]$ and $v_i[t]$ can be calculated based only on local variables. Conceptually this step is rather similar to solving a sparse system of equations with the structure of a directed path graph using back substitution.
The details are given in \cref{alg:com_gen}, and this process is illustrated in \cref{fig:com_ill}. This allows the optimal inputs to be computed by sweeping once through the graph from the first node to the final node to compute the $\delta$'s, and once from the final node to the first to compute the $\mu$'s. Both sweeps can be conducted in parallel. This represents a sort of middle ground between centralised control and decentralised control, in which global optimality is preserved whilst only requiring distributed communication. The price for this is that the sweep through the whole graph must be completed before the inputs can be applied, but for systems with reasonably long sample times (which is likely true in transportation or irrigation systems) this seems a modest price to pay. Interestingly the controller parameters can be computed in a similar distributed manner, allowing the controller to also be synthesised in a simple and scalable manner. This is shown in \cref{alg:init}.
\section{Results}
In this section we present two algorithms that together allow for the solution of \eqref{eq:problem}. The first of these algorithms computes the parameters of a highly structured control law for solving \eqref{eq:problem}, whereas the second shows that the control law has a simple distributed implementation. These features will be discussed in \cref{sec:impl}. In this section we will demonstrate that under suitable assumptions on the planned disturbances $d_i[t]$, \cref{alg:init,alg:com_gen} give the optimal solution to \eqref{eq:problem} . This constitutes the main theoretical contribution of the paper.
In the absence of the planned disturbances (i.e. with $d_i[t]\equiv{}0$), \eqref{eq:problem} is an infinite horizon LQ problem in standard form. It is of course highly desirable in applications to be able to include information about upcoming disturbances in the synthesis of the control law. However if we are given an infinite horizon of disturbances, \eqref{eq:problem} is no longer tractable. For the theoretical perspective, it turns out that the suitable assumption on the horizon length is as follows: \begin{assumption}\label{assump:horizon} Let the aggregate delay $\sigma_k$ in \eqref{eq:dynamics} be \begin{equation*}
\sigma_k = \sum_{i=1}^{k-1} \tau_i. \end{equation*} Given a horizon length $H\geq0$, assume that $d_i[t] = 0$ for all $t> H+(\sigma_N-\sigma_i)$ and for all $1\leq i \leq N$. \end{assumption}
Observe that if $d_i[t] = 0$ for all $t>H$ then \cref{assump:horizon} holds. Thus the assumption captures the natural notion of having a finite horizon $H$ of information about the disturbances $d_i[t]$ available when constructing the control input. In \cref{sec:simulations} we will investigate how the length of the horizon affects the performance of the controller.
We will now state the main results of this paper. The following theorem shows that \eqref{eq:problem} can be solved by running two simple algorithms; one for calculating all the necessary parameters, and one for computing the optimal inputs. Both algorithms can be implemented using only local communication as discussed in \cref{sec:impl}. A graphical illustration of the implementation of \cref{alg:com_gen}, which is the algorithm used for the on-line implementation, can be found in \cref{fig:com_ill}.
\begin{theorem}\label{thm:gen} Let \[ D_i[t] = \sum_{j=1}^i d_j[t-\sigma_j]. \]
Assume that $H$ and $d_j[t]$ satisfy \cref{assump:horizon}. Then the optimal inputs $u_i[t]$ and $v_i[t]$ for the problem in \eqref{eq:problem} are given by running \cref{alg:com_gen} with the parameters calculated by \cref{alg:init}. \end{theorem} \begin{proof} See the appendix. A sketch of the proof can be found in \cref{sec:proof_idea}. \end{proof}
\begin{remark}
In most cases the choice of $H$ can be made without considering its effect on the controller implementation, and can instead be chosen based only on the nodes' ability to forecast their disturbances. In applications it would also be natural to incorporate new information on upcoming disturbances in a receding horizon fashion.
This will be further discussed in \cref{sec:recedinghorizon}.
$\lozenge$ \end{remark}
\begin{remark} There is an asymmetry in \cref{assump:horizon} in that $(\sigma_N-\sigma_i)$ grows as $i$ decreases from $N$ to $1$. This means that this assumption allows nodes further down the graph to have longer horizons of planned disturbances. Of course there is no reason to believe that these nodes are better at predicting their disturbances. It is just that the derived theory can handle those disturbances in a straightforward manner since the optimal controller lumps the disturbances into time shifted sums, with a time shift proportional to $\sigma_i$.
$\lozenge$ \end{remark}
\begin{algorithm}[] \caption{Computation of control parameters.} \label{alg:init} \DontPrintSemicolon \SetNoFillComment \SetAlgoLined \KwIn{$q_i$, $r_i$, $\tau_i$, $H$} \KwOut{$\gamma_i$, $g_i(j)$, $P_i(\tau_i,1)$, $h_i$, $\phi_i(\Delta)$, $a_i$, $c_i$}
\nonl \hrulefill \\
\tcc{First Sweep, upstream direction} $\gamma_1 = q_1, \quad \rho_1 = r_1$ \tcp*[r]{initialize first node} \textbf{send} $\gamma_1$ and $\rho_1$ to upstream neighbor\\ \For {node i = 2:N} {
$\gamma_i = \frac{\gamma_{i-1}q_i}{\gamma_{i-1} + q_i}, \quad \rho_i = \frac{\rho_{i-1}r_i}{\rho_{i-1} + r_i}$\\
\textbf{send} $\gamma_i$ and $\rho_i$ to upstream neighbor }
\nonl \dotfill\\
\tcc{Second Sweep Downstream direction} $X_N(H+2) = -\frac{\gamma_N}{2}+\sqrt{\gamma_N\rho_N+\frac{\gamma_N^2}{4}}$ \\ \For {node i = N:1} {
$X_i(\tau_i) = \frac{\rho_i(X_{i+1}(1)+\gamma_{i})}{X_{i+1}(1)+\gamma_{i} + \rho_i}$ \tcp*{Not for node N}
$X_i(t-1) = \frac{\rho_i(X_i(t)+\gamma_i)}{X_i(t) + \gamma_i+\rho_i}$, \tcp*{$1\leq t-1\leq\tau_i-1$ or for i =N, $1\leq t-1 \leq H+1$}
$g_i(i) = \frac{X_i(i)}{X_i(i)+\gamma_i},\quad 2\leq i \leq \tau_i$ \\
$g_{i+1}(1) = \frac{X_{i+1}(1)}{X_{i+1}(1)+\gamma_{i}}$\\
$b_i = g_{i+1}(1)\prod_{j=2}^{\tau_i}g_i(j)$ \\
\textbf{send} $X_i(\tau_i)$, $b_i$ to downstream neighbor.\\
\tcp{for $1\leq l,m \leq \tau_i$}
$P_i(1,m) = \frac{X_i(1)}{\rho_{i}}$ \\
$P_i(l,m) = (1-\frac{X_i(l)}{\rho_{i}})g_i(l)P_i(l-1,m) + \frac{X_i(l)}{\rho(i)}, \quad l\leq m$ \\
$P_i(l,m) = (1-\frac{X_i(l)}{\rho_{i}})P_i(l-1,m) + \frac{X_i(l)}{\rho(i)}, \quad l> m$ \\ }
\nonl \dotfill \\
\tcc{Third Sweep, upstream direction} $h_1 = P_1(\tau_1,\tau_1)g_2(1)$\\ \textbf{send} $h_1$ to upstream neighbor.\\ \For {node i= 2:N-1} {
$h_i = (1-P_i(\tau_i,1))b_ih_{i-1} + P_i(\tau_i,\tau_i)g_{i+1}(1)$ \\
\textbf{send} $h_i$ to upstream neighbor. }
\nonl\dotfill \\
\tcc{Some final local Calculations} \tcp{For $1\leq \Delta \leq \tau_i$, Empty Product, $\prod_{j=2}^1$ = 1} $\phi_i(\Delta) = \left(1-P_i(\tau_i,\Delta) -(1-P_i(\tau_i,1))h_{i-1}{\prod_{j=2}^{\Delta +1}}g_k(j) \right)$\\ $a_i = \frac{X_i(1)}{r_i}+\frac{\gamma_i}{q_i}(1-\frac{X_i(1)}{\rho_i})$\\ $c_i = -\Big(\frac{X_i(1)}{r_i} - \frac{\gamma_iX_i(1)}{q_i\rho_i} \Big)(1-h_{i-1}) + \frac{\gamma_i}{q_i}h_{i-1}$ \end{algorithm} \SetInd{0.5em}{0.5em}
\begin{algorithm}[h!] \caption{Distributed Controller Implementation.} \label{alg:com_gen} \SetNoFillComment
\KwIn{$z_i[t]$, $u_i[t-(\tau_i-\Delta)]$, $d_i[t]$, $D_i[t+\sigma_i +\Delta]$} \KwOut{$u_i[t]$,$v_i[t]$}
\nonl \hrulefill \\
\tcp{Let $\tau_N = H + 1$.} \tcc{Upstream sweep - Done in parallel with downstream sweep} \For{node i = 1:N} { $\Phi_i[t] = \phi_i(1)z_i[t] +$\\ \nonl$ \ \ \ \ \sum_{\Delta=0}^{\tau_i-1} \phi_i(\Delta+1)\Big(u_i[t-(\tau_i-\Delta)]
+ D_i[t+\sigma_i+\Delta]\Big)$\\ $\delta_i[t] = \Phi_i[t] + (1-P_{i}(\tau_{i},1))\delta_{i-1}[t]$\\ \textbf{send} $\delta_i[t]$ upstream }
\nonl \dotfill\\
\tcc{Downstream sweep - Done in parallel with upstream sweep} \For{node i = N:1} { \tcp{Empty Product, $\prod_{j=2}^1$ = 1} $\pi_i[t] = z_i[t] + \sum_{\Delta=0}^{\tau_i-1} \Big(u_i[t-(\tau_i-\Delta)] + D_i[t+\sigma_i+\Delta]\Big)\prod_{j=2}^{\Delta+1}g_i(j)$\\ $\mu_i[t] = \pi_i[t] + b_i\mu_{i+1}[t]$\\ \textbf{send} $\mu_i[t]$ downstream}
\nonl \dotfill\\
\tcc{Calculate outputs} $u_{i-1}[t] = (1-\frac{\gamma_i}{q_i})\big(z_i[t]+u_i[t-\tau_i]+D_{i}[t+\sigma_i]\big)$ \\ \nonl $\qquad \qquad \qquad \qquad \qquad -a_i\delta_{i-1}[t] + c_i\mu_i[t] + d_i[t] -D_{i}[t+\sigma_i]$\\ $v_i[t] = -\frac{X_i(1)}{r_i}\Big(\delta_{i-1}[t] + (1-h_{i-1})\mu_i[t]\Big)$ \end{algorithm}
\section{Implementation}\label{sec:impl}
\begin{figure*}
\caption{ Illustration of the structured approach for calculating the optimal inputs using \cref{alg:com_gen}, for a 5 node example with $\tau_1 = 2$, $\tau_2 = 1$, $\tau_3 = 1$, $\tau_4 = 2$. The graph at the right of the figure illustrates the underlying dynamics of the network as in \eqref{eq:dynamics}. The left part of the figure illustrates the structure of the computations required to compute the optimal control according to \cref{alg:com_gen}. The solid circles corresponds to node states and dashed circles the quantities in transit. The number in each dashed circle denotes the value of $\Delta$, which then maps to $u_k[t-(\tau_k-\Delta)]$. The rectangles indicates the different intermediate calculations needed to determine the variables required to compute the optimal inputs (lines 2--3 and 7--8 in \cref{alg:com_gen}). These are horizontally aligned with the location in the network where they could be locally performed. The arrows indicate information flow. Each intermediate can be calculated using only the quantities from the incoming arrows. An upstream sweep is performed (the red arrows) in order to calculate the variables $\delta_i[t]$. The local intermediate $\Phi_i[t]$ variables are calculated (line 2), and then aggregated into the $\delta_i[t]$'s (line 3), which are sequentially passed up the graph. For the downstream sweep the local $\pi_i[t]$ variables are calculated (line 7) and aggregated into the $\mu_i[t]$'s (line 8). Both sweeps can be conducted in parallel, and once they have completed, the optimal inputs for the \emph{i}th node can be determined using the variables at the \emph{i}th location according to lines 11 and 12 in \cref{alg:com_gen}. }
\label{fig:com_ill}
\end{figure*}
In this section we will discuss the structure in \cref{alg:init,alg:com_gen}, and explain how they can be used to implement an optimal feedback control law for solving \eqref{eq:problem} in a distributed
manner. In both cases the order in which the computations occur is highly structured. This is illustrated for \cref{alg:com_gen}, which is the algorithm that must be run to compute the control inputs, in \cref{fig:com_ill}. Matlab code for using these algorithms to calculate the optimal control inputs is available at github\footnote{\texttt{https://github.com/Martin-Heyden/cdc-letters-2021}}, as well as code to verify that \cref{thm:gen} holds numerically. We will also discuss how to calculate $D_i$ and incorporate updates to the planned disturbances in a receding horizon style in \cref{alg:com_D}.
\subsection{Algorithms 1 and 2, and the Optimal Control Law}
The problem in \eqref{eq:problem} is at its heart an LQ problem, and the optimal controller is given by a static feedback law. The corresponding feedback matrix is generally dense, and that is the case for \eqref{eq:problem} as well. However certain special structural features of the process \eqref{eq:dynamics} are inherited by the optimal control law. It is these features we exploit to give a scalable implementation in \cref{alg:init,alg:com_gen}, which we will now discuss.
In terms of the algorithm variables, the optimal node production $v_i[t]$ for \eqref{eq:problem} is given by \begin{equation}\label{eq:opt_v}
v_i[t] = -\frac{X_i(1)}{r_i}\Big(\delta_{i-1}[t] + (1-h_{i-1})\mu_i[t]\Big), \end{equation} and the optimal internal flows $u_i[t]$ are given by \begin{multline}\label{eq:opt_u} u_{i-1}[t] = (1-\frac{\gamma_i}{q_i})\big(z_i[t]+u_i[t-\tau_i]+D_{i}[t+\sigma_i]\big) - a_i\delta_{i-1}[t]\\ + c_i\mu_i[t] + d_i[t] -D_{i}[t+\sigma_i]. \end{multline}
The parameters in these control laws (the symbols without a time index, which includes $X_i(1)$) are calculated in a simple and structured manner by \cref{alg:init}. Of course having an efficient method for computing the control law is less critical than having an efficient real time implementation of the control law (which is performed by \cref{alg:com_gen}), since the control law can be computed ahead of time. However the fact that this step is also highly structured indicates that the approach is scalable, since it allows for the the control law to be simply and efficiently updated in response to changes to the dynamics in \eqref{eq:dynamics} (perhaps resulting from the introduction of more nodes).
\cref{alg:init} computes all the parameters needed to give a closed form solution for the problem in \eqref{eq:problem}. The origin of the parameters in \cref{alg:init} are discussed briefly in the proof idea in \cref{sec:proof_idea} and full details are found in the proof in the appendix. The algorithm consists of three serial sweeps. The first sweep starts at node 1 and calculates $\gamma_i$ and $\rho_i$. The second sweep starts at node $N$, and calculates $X_i(t)$ The calculation of $X_i(t)$ has both local steps (line 10) and steps that requires communication (line 9). Also during the second sweep, the parameters $g$, $b$ and $P$ are calculated locally. The third sweep starts at node 1 again, and calculates the parameter $h$, which is needed to calculate the optimal production. Finally, after the third sweep, the parameters $\phi_i(\Delta)$, $a_i$ and $c_i$ are calculated in each node independently.
The real time implementation of the optimal control law also has a simple distributed implementation. This is the role of \cref{alg:com_gen}, and the structure of the implementation is illustrated in \cref{fig:com_ill}. The algorithm proceeds through two sweeps through the graph. These sweeps are independent of one another, and can be conducted in parallel. In the upstream sweep (from node 1 to node $N$), a set of local variables ($\Phi_i[t]$ and $\delta_i[t]$) are computed according to lines 2--3. This is done sequentially, since the computation of $\delta_i[t]$ depends on $\delta_{i-1}[t]$. $\delta_{i-1}[t]$ then gives all the information node $i$ needs from downstream nodes. Similarly the downstream sweep sequentially computes the $\pi_i[t]$ and $\mu_i[t]$ variables. Here $\mu_i[t]$ gives all the information needed from nodes upstream of node $i$. Once these two sweeps are completed, the optimal inputs can be calculated locally using lines 11 and 12.
\subsection{Receding Horizon and Calculation of $D_i[t]$}\label{sec:recedinghorizon}
We will now discuss how to implement the controller in a receding horizon style to account for updates and new information about the planned disturbances $d_i[t]$. In terms of both the optimal control problem in \eqref{eq:problem} and the controller implementation in \cref{alg:com_gen}, the planned disturbances are treated as fixed quantities, that are known up to some horizon length $H$ into the future (and equal to zero thereafter, c.f. \cref{assump:horizon}). The idea is that $d_i[t]$ determines the anticipated consumption of the quantity at node $i$ and time $t$. Having this information available ahead of time allows the optimal control law to anticipate the predicated usage, and optimally 'schedule' the transportation of the quantity through the network. As we will see in the examples this can lead to a significant improvement in performance. However, in practice we would want to update the values of the $d_i[t]$'s as time passes, and more up to date information becomes available.
A natural way to do this is to use a receding horizon approach. In this setting we assume that at each point in time, we essentially have a fresh problem, with a new set of planned disturbances. \cref{alg:com_gen} can then be used to compute the first optimal input for this problem, after which the problem resets, and we get a new horizon of planned disturbances. This ensures we always make the best action available to us with a given horizon of information about the disturbances. The question is then, how to efficiently update the part of the control law that depends on the planned disturbances.
The changes required to accommodate this are rather minor. The planned disturbances do not affect the control parameters or the distributed structure of the implementation of the control law. To see this, observe from \cref{alg:init,alg:com_gen} that all of the information about the planned disturbances is handled through the variables $D_i[t]$ defined in \cref{thm:gen}. An illustration of the relationship between individual disturbance $d_i$ and shifted disturbance vectors $D_i[t]$ can be found in \cref{fig:D_illustration}. \begin{figure}
\caption{Illustration for the terms included in $D_i[t]$ for a three node graph with $\tau_1 = \tau_2 = 2$. The first row of dots corresponds to $z_3[t],\ z_2[t],\ z_1[t]$ and the second corresponds to $z_3[t+1],\ z_2[t+1],\ z_1[t+1]$, and so on. Due to lack of space only $d_i[t]$ and $d_i[t+1]$ has been drawn. However, the pattern follows through the graph. From the figure we can see that $D_1[t+3] = D_2[t+3]-d_2[t]$, corresponding to the first part of \cref{alg:com_D}.}
\label{fig:D_illustration}
\end{figure} Thus the structure of the implementation of the control law remains the same, and only $D_i[t]$ needs to be updated as new information become available. This can be done efficiently by a sweep starting at the bottom of the graph. After all, although in the receding horizon framework we assume we have a `new' set of planned disturbances at each point in time, these will share a large amount of information with the planned disturbances from the previous time step. This is the role of \cref{alg:com_D}, which we now explain. \SetInd{0.5em}{1em} \begin{algorithm}[] \SetNoFillComment
\KwIn{changed $d_i[s]$} \KwOut{updated $D_i$}
\nonl \hrulefill \\
\tcc{Update Disturbances in parallel, $\mathcal{O}(1)$. Necessary even if there are no new disturbances} \For{node i = N:1}{ \textbf{Send} $D_{i-1}[t+1+\sigma_{i}-1] = D_{i}[t+\sigma_{i}] - d_{i}[t]$ downstream.\\ \textbf{Discard} $D_i[t+\sigma_{i}]$ \\ }
\nonl \dotfill\\
\tcc{Update Disturbances due to new planned disturbances} \For{node i = 1:N} { \tcc{ For $t+\sigma_i\leq s <t+\sigma_N+H$}
\If{$d_i[s-\sigma_i]$ \emph{changed} or $D_{i-1}[s]$ \emph{received}}
{\textbf{send} $D_i[s] = D_{i-1}[s] + d_i[s-\sigma_i] $ upstream
} } \caption{Calculation of $D$} \label{alg:com_D} \end{algorithm}
All the $D_i[t]$'s where none of the underlying $d_j[t]$ were changed can easily be updated. For $1\leq \Delta \leq \tau_i-1$ the information in $D_i[t + \sigma_i + \Delta]$ will be useful in node $i$ at the next time step as $$D_i[t+1+\sigma_i + \Delta-1] = D_i[t + \sigma_i + \Delta].$$ When $\Delta = 0$ the information can be used at the downstream neighbor as $D_i$ satisfies \[ D_{i-1}[t+1+\sigma_{i}-1] = D_{i}[t+\sigma_{i}] - d_{i}[t]. \] These updates can be done in all nodes simultaneously and the time it takes is thus independent of the size of the graph.
However, the shifted sums $D_i$ has to be initialized at time zero, and also updated when new disturbances $d_i$ are planned for times $t>0$. $D_i[t]$ only requires information from downstream, and can thus be calculated by a sweep starting at node one and going upstream. Starting at the first node, any of the $d_1[s],\ t\leq s\leq t+\sigma_N+H$ that have changed are sent to node two. Then for every node $i$, the aggregate $D_i[s],\ t+\sigma_{i} \leq s\leq t+\sigma_N+H$ is sent upstream if it has changed. This will be the case if node $i$ received $D_{i-1}[s],\ t+\sigma_{i}\leq s \leq t+\sigma_N+H$ from its downstream neighbor, or if $d_i[s],\ t\leq s \leq \sigma_N-\sigma_i+H$ has changed.
The steps for updating the disturbances are summarized in \cref{alg:com_D}. While the algorithm is essentially a sweep through the graph in the upstream direction, it might be best to not implement it in the upstream sweep of \cref{alg:com_gen} as then the downstream sweep would have to be done after the upstream sweep, due to its need for the shifted disturbance vectors.
On the other hand, the calculation does not rely on measurement from the system, and can thus be carried out either before or after \cref{alg:com_gen}.
We are now ready to discuss how the choice of $H$ affect the implementation of the controller. Firstly, a larger $H$ will lead to a very slight increase in the synthesis time due to more iterations of $X_N(t)$ being required. Secondly increasing $H$ will increase the memory requirement in node $N$, in that it requires to store $D_N[t+\Delta]$ for $0\leq \Delta \leq H$. Finally the requirement for the communication bandwidth when updating $D_i[t]$ will depend on the number of new disturbances $d_j[t]$, but is upper bounded by $H$ if $d_i[t] = 0$ for $t>H$ and by $H+\sigma_N$ if $d_i[t] = 0$ for $t>H+\sigma_N-\sigma_i$. Thus if the bandwidth is limited, and a lot of new disturbances are expected to be planned, one might need to limit the size of $H$. Otherwise it can be freely chosen based on the nodes abilities to forecast disturbances.
\section{Simulations}\label{sec:simulations}
In this section we explore the effect the feed-forward horizon has on the controller performance through simulations. In \cref{fig:cost_vs_horizon} the performance for different horizon lengths is shown. Two random nodes are affected by disturbances of total size between minus one and zero and during a time interval of length between 1 and 5. The node level cost is given by $q_i = 1$. The production cost is given by $r_i = 10N$, where $N$ is the number of nodes. This is an attempt to keep the production cost similar for different values of $N$. There are 50 simulations done for each case with random disturbances as previously described. For all the cases when $N$ is the same, all the disturbances are the same for all the different horizons and delay values.
The horizon lengths are the same for all nodes, i.e it is assumed that $d_i[t+d] = 0$ for $d>H$.
We can see that a large part of the performance increase form having feed-forward can be achieved for short disturbance horizons. We can also see that for larger delays, and for more nodes, a longer horizon is needed to get the same effect. As a rule of thumb, at least for this example, it seems like a horizon longer than 2/3 of the total delay gives almost no effect, and even a horizon of 1/3 of the total delay gives most of the performance increase.
\begin{figure}
\caption{Simulations comparing the effect the planning horizon has on the performance. The data-points with highest cost for each configuration corresponds to not using the planned disturbances at all. While the the rest corresponds to using all planned disturbances announced up to $H$ time units ahead in every node.}
\label{fig:cost_vs_horizon}
\end{figure}
\section{Proof Idea}\label{sec:proof_idea} In this section we will describe the main idea behind the technique used to derive the results, which is to study a time shifted sum of the node-levels $z_i$, and a time shifted sum of the production $v_i$. This will allow the problem to be solved in terms of these shifted sums, essentially reducing it to a problem with scalar variables. Outside the disturbance horizon the problem can be solved by a Riccati equation in one variable. While inside the disturbance horizon the problem is solved using dynamic programming, where each step has scalar variables.
Now for the definitions of the shifted sums, let the sum of a shifted level $S_k$ and sum of a shifted production vector $V_k$ be defined as \[
S_k[t] = \sum_{i=1}^k z_i[t-\sigma_i], \qquad V_k[t] = \sum_{i=1}^k v_i[t-\sigma_i]. \] Also Let $\bar{V}_k[t] = V_k[t]+D_k[t]$ to shorten some expressions.
We illustrate the main idea by considering these shifted sums with a short example. Consider a path graph with $N>2$, $\tau_1 = 1$, and $\tau_2>1$. Then $S_2[3] = z_1[3]+z_2[2]$. It can be checked that \[S_2[3] = z_1[0]+z_2[0] + \bar V_1[0]+\bar V_2[1]+\bar V_2[2] + u_1[-1] + u_2[-\tau_2].\] Note that the internal transportation $u_1[0]$ and $u_1[1]$ have canceled, and the sum is thus independent of the internal transportation $u$ (except those with negative time index, which correspond to initial conditions). Also note that any values for $z_2[2]$ and $z_1[3]$ can be achieved as long as $z_1[3]+z_2[2] = S_2[2]$. This follows from that $z_2[2]$ can take any value by choosing the appropriate value for $u_1[1]$. Thus the cost of $q_2z_2[2]^2 + q_1z_1[3]^2$ only depends on the value of $S_2[3]$, which is independent of all internal transportation $u_i[t],\ t\geq 0$. This means that all inputs except $u_1[1]$ can ignore its effect on the terms in $S_2[3]$. Furthermore, $S_2[3]$, and thus the corresponding cost, only depends on the sums $V_2[1]$ and $V_2[2]$ and not the individual productions $v_1[1]$, $v_1[2]$, $v_2[0]$ and $v_2[1]$.
This idea can be generalized. The cost function can be rewritten in terms of shifted levels, where each shifted level sum is independent of the internal transportation. Each shifted level can thus be minimized independently with respect to the internal transportation $u$. Just as in the example, the only constraint is that the shifted sum has the correct value. The optimal cost for a shifted vector $S_k[t]$ is given by the solution to
\begin{equation}\label{eq:opt_zi}
\begin{aligned}
\minimize_{z_i} \quad & \sum_{i=1}^k q_iz_i[t+\sigma_k-\sigma_i]^2 \\
\textrm{subject to} \quad & S_k[t+\sigma_k] = c,
\end{aligned}\end{equation} where $c$ depends on the initial conditions, $V_i$, and $D_i$. The problem has the solution $z_i[t+\sigma_k-\sigma_i] = \gamma_k/q_i\cdot c$ and cost $\gamma_kc^2$, where $\gamma_k$ is as given in \cref{alg:init}. Once the optimal level $z_k[1]$ is calculated, the optimal value for $u_{k-1}[0]$ can be found from the dynamics, which gives \begin{multline*}
u_{k-1}[0] = (1-\frac{\gamma_{k}}{q_k})(z_{k}[0] + u_{k}[-\tau_{k}]) + v_k[0] +d_k[0]\\
-\frac{\gamma_k}{q_k}\bar{V}_k[\sigma_k]
-\frac{\gamma_k}{q_k}(m_{k-1}[0] + \sum_{i=1}^{k-1} \sum_{d=0}^{\tau_i-1}\bar{V}_i[\sigma_i+d]). \end{multline*} Where $m_k[t] = \sum_{i=1}^k(z_i[t] + \sum_{\delta = 1}^{\tau_i}u_i[t-\delta])$. After inserting the optimal values for $v_k$ and $V_i$ the expression in \eqref{eq:opt_u} is achieved. Note that all the terms with coefficient $1$ corresponds to what would be in the node $k$ at time $t=1$ if $u_{k-1}[0] = 0$, and all terms with coefficient $-\gamma_k/q_k$ gives the total quantity in $S_k[1]$.
Furthermore, each shifted level sum only depends on the shifted production sums $V_k[t]$, and not the individual productions $v_i[t],\ i\leq k$. The optimal way to produce a specific amount $V_k[t]$ with a shifted production vector can be found by solving a problem similar to \eqref{eq:opt_zi}, with the optimal $v_i$ given by $v_i[t-\sigma_i] = \rho_k/r_i\cdot V_k[t] \text{ for } i \leq k$, and the cost given by $\rho_kV_k[t]^2$, where $\rho_k$ is as given in \cref{alg:init}. So the calculations of $\gamma_i$ and $\rho_i$ in the first sweep in \cref{alg:init} thus corresponds to solving the optimal distribution for a shifted level vector $S_k$ and the optimal production for a shifted production vector $V_k$. This is covered in \cref{lem:shifted_inv}.
Assuming all $u_i$ are picked so that the shifted levels are optimized, the total level cost is given by
\begin{equation}\label{eq:cost_S}
\sum_{t=0}^\infty\sum_{i=1}^N q_i z_i[t]^2 = \sum_{i=0}^N q_i z_i[0]^2
+ \sum_{i=1}^{N-1}\ \ \ \ \mathclap{\sum_{t=\sigma_i+1}^{\sigma_{i+1}}}\ \ \gamma_i S_i[t]^2 + \ \ \mathclap{\sum_{t=\sigma_N+1}^\infty}\ \ \gamma_NS_N[t]^2. \end{equation} And assuming all $v_i$ are picked so that each shifted production vector is optimized, then the total production cost is given by \begin{equation}\label{eq:cost_V}
\sum_{t=0}^\infty\sum_{i=1}^N r_iv_i[t]^2 = \sum_{i=1}^{N-1}\sum_{t=\sigma_i}^{\sigma_{i+1}-1}\rho_i V_i[t]^2 + \sum_{t=\sigma_N}^\infty \rho_NV_N[t]^2. \end{equation} This allows the problem in \eqref{eq:problem} to be solved in terms of $V_i$ and $S_i$, reducing it to a problem in scalar variables. The scalar problem can be solved analytically, giving a closed form solution.
This is done by first solving for all $V_N[t]$ outside the disturbance horizon, that is for $t >\sigma_N + H$. Using that outside the disturbance horizon the dynamics for the shifted levels $S_N[t]$ are $S_N[t+1] = S_N[t] +V_N[t]$ gives that those $V_N[t]$ are given by the solution to
\[\begin{aligned}
\minimize_{V_N[t]} \quad & \sum_{\mathclap{t=\sigma_N+H+1}}^\infty \gamma_NS_N[t]^2 + \rho_NV_N[t]^2 \\
\text{subject to} \quad & S_N[t+1] = S_N[t] +V_N[t]. \end{aligned}\] This problem can be solved through a Riccati equation in one variable, giving expressions for $V_N[t],\ t >\sigma_N + H$ in terms of $S_N[t]$. And more importantly, a cost to go in terms of $S_N[\sigma_N+H+1]$, that is $X_N[H+2]$ in \cref{alg:init}.
Each shifted sum in \eqref{eq:cost_S} can be expressed in terms of initial conditions, shifted production vectors, and shifted disturbance vectors, \begin{equation*}
S_k[\sigma_k+\Delta] = m_{k-1}[0]+ z_k[0] + \sum_{i=1}^{k-1}\sum_{d=\sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d]
+ \sum_{d = \sigma_k}^{\mathclap{\sigma_k+\Delta-1}} \bar{V}_k[d] . \end{equation*}
Using the cost to go from the Riccati equation as the terminal cost allows the rest of the $V_i$'s to be found analytically using dynamic programming. When solving this problem the cost to go $X_i$ in \cref{alg:init} is used. The parameter $g$ also appears naturally in the solution to each dynamic programming step, and the upstream aggregate $\mu_i$ in \cref{alg:com_gen} is used to simplify the expressions. This is covered in detail in \cref{lem:V_expr}.
The resulting solution gives $V_i[t]$ in terms of initial conditions and the previous $V_k$'s in \eqref{eq:cost_V}. However, $V_1[0]$ is known, which gives $V_1[1]$ and so on. When rewriting $V_i[0]$ in terms of only initial conditions the expressions can be simplified by using $\delta$ as defined in \cref{alg:com_gen}. This in turn requires $P$, $h$, and $\phi$ which were defined in \cref{alg:init} and $\Phi$ which was defined in \cref{alg:com_gen}. For the details see \cref{lem:mVsum}.
\section{Conclusions and Future Work} In this paper we studied an optimal control problem on a simple transportation model. We showed that the optimal controller is highly structured, allowing for a distributed implementation consisting of two sweeps through the graph. The optimal controller can also handle planned disturbances in an efficient way.
We believe that the results presented here can be extended to more general graph structures. More specifically for any graph with the structure of a directed tree both the proof technique and the results could be extended. We plan to explore this in a future publication.
\appendix \section{Appendix}
The proof follows the structure of the proof idea. Before we start we restate the definition of $m_k$ which was mentioned in the proof idea. \[ m_k[t] = \sum_{i=1}^k\left( z_i[t] + \sum_{d=1}^{\tau_i} u_i[t-d]\right)\] Also, we let the product over an empty set be equal to one, e.g., $\prod_{i=2}^1 g_i = 1$.
The proof will derive the optimal inputs at time $t = 0$. As the problem has an infinite horizon, one can freely shift the time, and the results will thus holds for all $t \geq 0$. We begin by showing that each shifted level can be optimally distributed and find the corresponding internal flows. \begin{lemma}\label{lem:shifted_inv} The following holds \begin{enumerate}[(i)]
\item Every shifted level $S_k$ satisfies
\begin{multline*}S_k[t+\sigma_k+1] =\\ z_k[t] + u_k[t-\tau_k] +m_{k-1}[t]
+ \sum_{i=1}^{k-1} \sum_{d=0}^{\tau_i-1}\bar{V}_i[t+\sigma_i+d]
+\bar{V}_k[t+\sigma_k]
\end{multline*}
\item Let $\gamma_k$ be defined as in \cref{alg:init}. The optimization problem
\[\begin{aligned}
\minimize_{z_i} \quad & \sum_{i=1}^k q_iz_i[t+\sigma_k-\sigma_i]^2 \\
\textrm{subject to} \quad & S_k[t+\sigma_k] = m,
\end{aligned}\]
has the solution $z_i = \gamma_k/q_im$ and the optimum value is given by $\gamma_km^2$.
\item When $u$ is chosen optimally, the cost for \eqref{eq:problem} is given by
\begin{equation*}
\sum_{t=0}^\infty\sum_{i=1}^N q_iz_i[t]^2 = \sum_{i=0}^Nq_iz_i[0]^2
+ \sum_{i=1}^{N-1}\sum_{t=\sigma_i+1}^{\sigma_{i+1}}\gamma_iS_i[t]^2 + \sum_{t=\sigma_N+1}^\infty \gamma_NS_N[t]^2. \end{equation*}
Also, the optimal $u_k[0]$ is given by \begin{multline*}
u_{k-1}[0] = (1-\frac{\gamma_{k}}{q_k})(z_{k}[0] + u_{k}[-\tau_{k}]) + v_k[0] +d_k[0]\\
-\frac{\gamma_k}{q_k}\bar{V}_k[\sigma_k]
-\frac{\gamma_k}{q_k}(m_{k-1}[0] + \sum_{i=1}^{k-1} \sum_{d=0}^{\tau_i-1}\bar{V}_i[\sigma_i+d]). \end{multline*}
\end{enumerate} \end{lemma} \begin{proof} For $k=1$ (i) reduces to the dynamics. Now assume that (i) holds for $k-1$. It follows from the definition of $S_k$ that \begin{equation}\label{eq:S_ind_shift} S_k[t+\sigma_k+1] = z_k[t+1] + S_{k-1}[t+\sigma_k+1]. \end{equation} It holds that
\begin{equation}\label{eq:Sk_step}
S_k[t+1] = S_k[t]+\bar{V}_k[t] + u_k[t-\sigma_k-\tau_k],
\end{equation}
since $u_i[t-\sigma_i-\tau_i]$ will cancel out for $i<k$.
This allows $S_{k-1}[t+\sigma_k+1]$ to be rewritten as
\begin{equation}\label{eq:sk_big_step}
S_{k-1}[t+\sigma_k+1] = S_{k-1}[t+\sigma_{k-1}+1]+\ \ \mathclap{\sum_{\Delta=\sigma_{k-1}+1}^{\sigma_k}}\ \ \ \bar{V}_{k-1}[t+\Delta] + \sum_{\Delta =0}^{\tau_{k-1}-1} u_{k-1}[t-\Delta].
\end{equation}
Using the induction assumption that (i) holds for $k-1$, \eqref{eq:sk_big_step} and the dynamics,
\begin{equation}\label{eq:dyn_in_proof}
z_k[t+1] = z_k[t] -u_{k-1}[t]+ u_k[t-\tau_k]+v_k[t]+d_k[t],
\end{equation}
allows \eqref{eq:S_ind_shift} to be rewritten as
\begin{multline*}
S_k[t+\sigma_k+1] = z_k[t] - u_{k-1}[t] + u_k[t-\tau_k]+v_k[t]+d_k[t] \\
+z_{k-1}[t] + u_{k-1}[t-\tau_{k-1}] +m_{k-2}[t]
+ \sum_{i=1}^{k-2} \sum_{d=0}^{\tau_i-1}\bar{V}_i[t+\sigma_i+d]\\
+\bar{V}_{k-1}[t+\sigma_{k-1}]
+\ \ \mathclap{\sum_{\Delta=\sigma_{k-1}+1}^{\sigma_k}}\ \ \ \bar{V}_{k-1}[t+\Delta] + \sum_{\Delta =0}^{\tau_k-1} u_{k-1}[t-\Delta].
\end{multline*} In the above it holds that \[ z_{k-1}[t]+u_{k-1}[t-\tau_{k-1}] -u_{k-1}[t]+\sum_{\Delta =0}^{\mathclap{\tau_{k-1}-1}} u_{k-1}[t-\Delta] + m_{k-2}[t] = m_{k-1}[t] \]
and
\begin{multline*}
v_k[t]+d_k[t] + \sum_{i=1}^{k-2} \sum_{d=0}^{\tau_i-1}\bar{V}_i[t+\sigma_i+d] + \bar{V}_{k-1}[t+\sigma_{k-1}] + \ \ \mathclap{\sum_{\Delta=\sigma_{k-1}+1}^{\sigma_k}}\ \ \ \bar{V}_{k-1}[t+\Delta]\\
= \sum_{i=1}^{k-1} \sum_{d=0}^{\tau_i-1}\bar{V}_i[t+\sigma_i+d]
+\bar{V}_k[t+\sigma_k].
\end{multline*}
And thus (i) holds for $k$ as well.
For (ii) the proposed solution satisfies the constraint as \[
\sum_{i=1}^k \frac{1}{q_i} = \frac1{\gamma_k}. \] If the proposed solution was not optimal then it would be possible to improve it by increasing $z_i$ by epsilon and decreasing $z_j$ by epsilon for $i,j\leq K$ as the problem is convex. However \[
\frac{\partial}{\partial z_i} q_iz_i[t+\sigma_k-\sigma_i]^2 = 2\gamma_km \] for $z_i[t+\sigma_k-\sigma_i] = \gamma_k/q_i m$ and all $i$, and thus the proposed solution is optimal.
For (iii) note that the sum $\sum_{t=0}^\infty\sum_{i=1}^N q_iz_i[t]^2$ can be written in terms of shifted level vectors as follows, \begin{multline} \label{eq:cost_decomp}
\sum_{t=0}^\infty\sum_{i=1}^N q_iz_i[t]^2 = \\
\sum_{i=0}^Nq_iz_i[0]^2
+ \sum_{i=1}^{N-1}\sum_{t=\sigma_i+1}^{\sigma_{i+1}} \sum_{j=1}^iq_jz_j[t-\sigma_j]^2 + \ \ \mathclap{\sum_{t=\sigma_N+1}^\infty}\ \ \ \ \ \sum_{j=1}^N q_jz_j[t-\sigma_j]^2. \end{multline} The inner sums corresponds to the objective in (ii). From (i) it follows that $S_k[t]$, $t\leq \sigma_{i+1}$ is independent of $u_j[t], \ \forall t \geq 0, \forall j$ and that $S_N[t]$ is independent of $u_j[t],\ \forall t,j$. Thus each shifted level sum in \eqref{eq:cost_decomp} is independent of the internal flows. Now consider arbitrary, but fixed productions $V$ and disturbances $D$. Then by (i) the sum of all shifted levels are fixed. If there exists $u$ so that each sum over shifted levels in \eqref{eq:cost_decomp} is the optimal solution to the problem in (ii), then those inputs must be optimal for the given $V$ and $D$. By choosing $u_{j-1}[t-\sigma_j-1]$ so that $z_j[t-\sigma_j]$ is optimal for (ii) for $2\leq j \leq i$ gives that all $z_j[t-\sigma_j]$ are optimally for $2\leq j \leq i$. However, since the constraint will always be satisfied, $z_1[t]$ will be optimal as well. Using (i), the optimal $z_k[1]$ from (ii) is given by \begin{equation*}
z_k[1] = \frac{\gamma_k}{q_k}\Big(m_{k-1}[0] +z_k[0] + u_k[-\tau_k]
+ \sum_{i=1}^{k-1}\sum_{\Delta = 0}^{\tau_i-1}\bar{V}_i[\sigma_i +\Delta] + \bar{V}_k[\sigma_k] \Big). \end{equation*} Inserting the dynamics in \eqref{eq:dyn_in_proof} into the LHS and solving for $u_{k-1}[0]$ gives the expression in (iii). \end{proof}
Now we will give the solution to the optimization problem which will arise in the dynamic programming problem that will need to be solved in the next lemma. \begin{lemma}\label{lem:aux} Let $X_i$ and $g_i(j)$ be defined as in \cref{alg:init}. Then \begin{enumerate}[(i)]
\item Let $j\geq1$. The optimization problem \[
\minimize_x \quad X_i(j+1)(a+b + x)^2 + \gamma_i(a+x)^2 + \rho_ix^2 \] has minimizer \[x = -\frac{X_i(j)}{\rho_i}(a+g_i(j+1)b),\]
with optimum value $X_i(j)\cdot(a+g_i(j+1)b)^2 + f(b)$. \item The optimization problem \[
\minimize_x \quad X_{i+1}(1)(a+b+ x)^2 + \gamma_{i}(a+x)^2 + \rho_{i}x^2 \] has minimizer \[
x = -\frac{X_{i}(\tau_{i})}{\rho_{i}}(a+g_{i+1}(1)b), \]
with optimum value $X_{i}(\tau_{i})\cdot(a+g_{i+1}(1)b)^2+f(b)$. \end{enumerate} \end{lemma} \begin{proof} We will show that the optimization problem \[
\minimize_x \quad c_1(a+b+x)^2 + c_2(a+x)^2+c_3x^2 \] has the solution \[
x = -\frac{c_1+c_2}{c_1+c_2+c_3}\big(a + \frac{c_1}{c_1+c_2}b\big) \]
and the minimal value is on the form \[
\frac{c_3(c_1+c_2)}{c_1+c_2+c_3}\left(a + \frac{c_1}{c_1+c_2}b\right)^2 + f(b), \] where $f(b)$ is independent of $a$. The lemma then follows by applying the above and using the definition for $X_i$ and $g_i(j)$.
There exits a unique solution as the problem is strictly convex.
Differentiating the objective function with respect to $x$ gives that the optimal $x$ is given by \[
x = -\frac{1}{c_1+c_2+c_3}\big((c_1+c_2)a + c_1b\big) \] from which the proposed $x$ follows.
The objective function can be rewritten as
\[
c_1(a+b)^2 + c_2a^2 + 2[(c_1+c_2)a + c_1b)]x + (c_1 + c_2 + c_3)x^2.
\]
Inserting the minimizer gives \[
\frac{1}{c_1+c_2+c_3}\Big((c_1+c_2+c_3)(c_1(a+b)^2+c_2a^2) - [(c_1+c_2)a + c_1b]^2\Big). \] The first term can be written as \begin{multline*}
(c_1+c_2+c_3)(c_1(a+b)^2+c_2a^2) \\=(c_1+c_2+c_3)\big[(c_1+c_2)a^2 + 2c_1ab + c_1b^2 \big]\\
=(c_1+c_2)^2a^2 + 2(c_1+c_2)c_1ab + c_1^2b^2 + \\
c_3(c_1+c_2)a^2 + 2c_1c_3ab + (c_2+c_3)c_1b^2. \end{multline*} Which gives that the objective function has the minimum value \[
\frac{1}{c_1+c_2+c_3}\Big[c_3(c_1+c_2)a^2 + 2c_1c_3ab + (c_2+c_3)c_1b^2\Big]. \] The last term is independent of $a$. The dependence on a is thus given by \[
\frac{c_3(c_1+c_2)}{c_1+c_2+c_3}\left(a^2 + \frac{2c_1}{c_1+c_2}ab\right) = \frac{c_3(c_1+c_2)}{c_1+c_2+c_3}\left(a + \frac{c_1}{c_1+c_2}b\right)^2 + f(b) \] \end{proof}
Armed with the results from the previous lemma, we will now apply dynamic programming to \eqref{eq:problem}. We will show that the problem can be solved in terms of the shifted levels $S_k$, shifted productions $V_k$ and shifted disturbances $D_k$. Outside of the horizon the problem can be solved using the Riccati equation. Using the cost to go given by the Riccati equation as initialization we can apply dynamic programming using the results from the previous lemma. \begin{lemma}\label{lem:V_expr} Let $\gamma_k$, $\rho_k$, $X_k$, $g_k$, $\mu_k$ be defined as in \cref{alg:init,alg:com_gen}. Let for $1\leq k\leq N-1$ and $1\leq \Delta \leq\tau_k$, and for $k=N$ and $1 \leq \Delta \leq H+2$ \begin{multline} \label{eq:xi}
\xi_k[\Delta-1] = m_{k-1}[0] + z_k[0] + \sum_{i=1}^{k-1}\sum_{d = \sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d] \\
+ \sum_{d=0}^{\tau_k-1} \Big(u_k[-(\tau_k-d)]+D_k[\sigma_{k}+d]\Big)\prod_{j=\Delta+1}^{d+1}g_k(j)
\\
+\sum_{d = \sigma_k}^{\sigma_k + \Delta-2}V_k[d] + \mu_{k+1}[0]g_{k+1}(1)\prod_{j=\Delta+1}^{\tau_k} g_k(j). \end{multline} Then the optimal $V_k$ for \eqref{eq:problem} is given by \begin{equation*}
V_k[\sigma_k+(\Delta-1)] = -\frac{X_k(\Delta)}{\rho_k}\xi_k[\Delta-1]. \end{equation*} The optimal individual productions are given by \[ v_k[\Delta-1] = \frac{\rho_k}{r_k}V_k[\sigma_k+(\Delta-1)].\] \end{lemma} \begin{proof} By \cref{lem:shifted_inv}-(i) and \eqref{eq:Sk_step} each shifted inventory level $S_k[\sigma_k +\Delta]$ with $1\leq \delta \leq \tau_{k}$ satisfies \begin{equation}\label{eq:Sk_delta_shift}
S_k[\sigma_k+\Delta] = m_{k-1}[0]+ z_k[0] + \sum_{i=1}^{k-1}\sum_{d=\sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d]
+ \sum^{\tau_k}_{\mathclap{d = \tau_k-(\Delta-1)}} u_k[-d]
+ \sum_{\mathclap{d = \sigma_k}}^{\mathclap{\sigma_k+\Delta-1}} \bar{V}_k[d]. \end{equation}
By (iii) in \cref{lem:shifted_inv} the cost can be rewritten as \begin{equation*}
\sum_{t=0}^\infty\sum_{i=1}^N q_iz_i[t]^2 = \sum_{i=0}^Nq_iz_i[0]^2
+ \sum_{i=1}^{N-1}\sum_{t=\sigma_i+1}^{\sigma_{i+1}}\gamma_iS_i[t]^2 + \sum_{\mathclap{t=\sigma_N+1}}^\infty \gamma_NS_N[t]^2. \end{equation*} And similarly, by \cref{lem:shifted_inv}-(ii), the optimal cost for a shifted production $V_i[t]$ is given by $\rho_iV_i[t]^2$ and individual productions are given by $v_i[t] = \rho_i/r_i\cdot V_i[t+ \sigma_i]$. This gives the total production cost in terms of $V_i$ as \[
\sum_{t=0}^\infty\sum_{i=1}^N r_iv_i[t]^2 = \sum_{i=1}^{N-1}\sum_{t=\sigma_i}^{\sigma_{i+1}-1}\rho_iV_i[t]^2 + \sum_{t=\sigma_N}^\infty \rho_NV_N[t]^2. \]
We can thus solve the problem in terms of $S_i$ and $V_i$, and then recover the optimal $v_i$. To that end define the cost to go for $1\leq k \leq N-1$ and $1\leq \Delta \leq \tau_k$ \begin{multline*}
\Gamma_k[\Delta] = \sum_{t = \sigma_k+\Delta}^{\sigma_{k+1}}\Big(\gamma_kS_k[t]^2+\rho_kV_k[t-1]^2\Big)
+ \sum_{i=k+1}^{N-1}\sum_{\sigma_i+1}^{\sigma_{i+1}}\Big( \gamma_iS_i[t]^2+\rho_iV_i[t-1]^2\Big)\\
+ \sum_{t=\sigma_N+1}^\infty \Big( \gamma_NS_N[t]^2+\rho_NV_N[t-1]^2\Big). \end{multline*} And for $k = N$ and $\Delta \geq 1$ \[
\Gamma_N[\Delta] = \sum_{t=\sigma_N+\Delta}^\infty \Big( \gamma_NS_N[t]^2+\rho_NV_N[t-1]^2\Big). \] We will show for $1\leq k \leq N-1$ and $1 \leq \Delta \leq \tau_i$, and for $k = N$ and $1\leq \Delta \leq H+2$, that \begin{equation}\label{eq:cost_to_go}
\Gamma_k[\Delta] = X_{k}(\Delta)\xi_k[\Delta-1]^2 + f(b), \end{equation} where $f(b)$ is independent of $V_k[t]$. $f(b)$ can thus be ignored in the optimization of $V_k[t]$.
Using \cref{lem:shifted_inv}-(i) combined with \eqref{eq:Sk_step} and that all $D_N[t] =0$ for $t>H+\sigma_N$ it follows that the optimal $V_N[t]$ for $t>\sigma_N+H$ is given by the solution to the problem \[\begin{aligned}
\minimize_{V_N[t]} \quad & \sum_{t=\sigma_N+H+1}^\infty \gamma_NS_N[t]^2 + \rho_NV_N[t]^2 \\
\text{subject to} \quad & S_N[t+1] = S_N[t] +V_N[t] \\
&S_N[\sigma_N+H+1] = m_N[0] + \sum_{i=1}^{N-1}\ \ \ \ \mathclap{\sum_{\delta=\sigma_i}^{\sigma_{i+1}-1}}\ \ \ \bar{V_i}[\delta] +\ \mathclap{\sum_{\delta = \sigma_{N}}^{\sigma_N+H}}\ \ \ \ \bar{V}_N[\delta]. \end{aligned}\] This is a standard LQR problem and the solution can be found by solving the following Riccati equation \[
X = X-X^2/(\rho_N+ X)+\gamma_N \Rightarrow X = \frac{\gamma_N}{2}+\sqrt{\gamma_N\rho_N+\frac{\gamma_N^2}{4}}. \]
Now let $X_N(H+2) = X-\gamma_N$. Then $\Gamma_N[\sigma_N+H+2]$ is given by
\[\Gamma_N[\sigma_N+H+2] = S_N[\sigma_N+H+1]^2X_N(H+2).\] Note that the cost for $S_k[\sigma_N+H+1]$ is not part of $\Gamma_N[\sigma_N+H+2]$, but it is part of the cost to go given by the solution $X$ to the Riccati equation. Furthermore, the optimal $V_N[t]$ for $t = \sigma_N+H+1$ is given by \begin{equation*}
V_N[t] = -\frac{X}{X+\rho_N}S_N[t]
= -\frac{X_N(H+1)+\gamma_N}{X_N(H+1)+\gamma_N+\rho_N}S_N[t] = -\frac{X_N(H)}{\rho_N}S_N[t]. \end{equation*} For $\Delta = H+2$ and $k=N$ the expression for $\xi_k[\Delta-1]$ reduces to \[
\xi_N[H+1] = m_N[0] + \sum_{i=1}^{N-1}\sum_{d = \sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d] + \sum_{d = \sigma_N}^{\sigma_N + H}\bar{V}_N[d], \] as $\mu_{N+1} = 0$, $u_N=0$ and $D_N[t] = 0$ for $t>\sigma_N+H$. By \eqref{eq:Sk_delta_shift} $\xi_N[H+1] = S_N[\sigma_N+H+1]$ and thus the lemma and \eqref{eq:cost_to_go} holds for $k=N$ and $\Delta = H+2$.
Assume that \eqref{eq:cost_to_go} holds for $k+1$ and $\Delta = 1$. Then the optimal $V_{k}[\sigma_{k+1}-1]$ is given by the minimizer for \[ \Gamma_{k}[\tau_{k}] = \Gamma_{k+1}[1] + \gamma_{k}S_k[ \sigma_{k+1}]^2 + \rho_k V_k[\sigma_{k+1}-1]^2.\] Using the assumption for the cost to go in \eqref{eq:cost_to_go} gives that $\Gamma_{k+1}[1] =X_{k+1}(1)\xi_{k+1}[0]^2$ and thus the optimal $V_{k}[\sigma_{k+1}-1]$ is given by the optimal value for the problem \begin{equation*}
\minimize_{V_{k}[\sigma_{k+1}-1]}\quad X_{k+1}(1)\xi_{k+1}[0]^2+\gamma_{k}S_{k}[\sigma_{k+1}]^2
+ \rho_{k}V_{k}[\sigma_{k+1}-1]^2. \end{equation*} For $\Delta = 1$ \eqref{eq:xi} reduces to \begin{equation}\label{eq:delta1reduce}
\xi_k[0] = m_{k-1}[0] +\sum_{i=1}^{k-1}\sum_{d = \sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d] + \mu_k[0], \end{equation} as \begin{equation*} \mu_k[0]= \pi_k[0]
+\mu_{k+1}g_{k+1}(1)\prod_{j=2}^{\tau_k} g_k(j) \end{equation*} and \[
\pi_k[0] = z_k[0] + \sum_{d=0}^{\tau_k-1} \Big(u_k[-(\tau_k-d)]+D_k[\sigma_{k}+d]\Big)\prod_{j=2}^{d+1}g_k(j). \] We also note that by \eqref{eq:Sk_delta_shift}, as $\sigma_{k+1} = \sigma_k + \tau_k$, \[ S_k[\sigma_{k+1}] = m_k[0] + \sum_{i=1}^{k}\sum_{d=\sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d]. \] Applying \cref{lem:aux}-(ii) with \[ \begin{aligned}
a &= S_{k}[\sigma_{k+1}]-V_{k}[\sigma_{k+1}-1]\\
b &= \xi_{k+1}[0] - S_k[\sigma_{k+1}] = \mu_{k+1}[0] \\
x &= V_{k}[\sigma_{k+1}-1], \end{aligned} \]
gives that the lemma and \eqref{eq:cost_to_go} hold for $k$ and $\Delta = \tau_{k}$ as
\[
\xi_k[\tau_k-1] = m_k[0] + \sum_{i=1}^{k-1}\sum_{d=\sigma_i}^{\sigma_{i+1}-1}\bar V_i[d] + \sum_{d=\sigma_k}^{\mathclap{\sigma_k+\tau_k-2}}\bar{V}_k[d] + \mu_{k+1}[0]g_{k+1}(1).
\]
Assume that \eqref{eq:cost_to_go} holds for some $k$ and $\Delta+1$, where $1\leq \Delta\leq \tau_i-1$ if $k<N$ and $1\leq \Delta\leq H+1$ if $k = N$. Then $V_k[\sigma_k+\Delta-1]$ can be found as the minimizer for \begin{equation*} \minimize_{V_k[\sigma_k+\Delta-1]} \quad X_k(\Delta+1)\xi_k[\Delta]^2 + \gamma_k S_k[\sigma_k+\Delta]^2
+ \rho_k V_k[\sigma_k+\Delta-1]^2. \end{equation*} Using that two of the terms in \eqref{eq:Sk_delta_shift} can be rewritten as \begin{equation*}
\sum^{\tau_k}_{\mathclap{d = \tau_k-(\Delta-1)}} u_k[-d]
+ \sum_{d = \sigma_k}^{\mathclap{\sigma_k+\Delta-1}} \bar{V}_k[d]
=\sum_{d = 0}^{\Delta-1} \Big(u_k[-(\tau_k-d)]+D_k[\sigma_k+d]\Big) + \sum_{d = \sigma_k}^{\mathclap{\sigma_k+\Delta-1}} V_k[d] \end{equation*}
and with $x =V_k[\sigma_k+\Delta-1]$, $a = S_k[\sigma_k+\Delta]-V_k[\sigma_k+\Delta-1]$, which equals \begin{multline*}
a = m_{k-1}[0]+ z_k[0] + \sum_{i=1}^{k-1}\sum_{d=\sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d] \\
+ \sum_{d = 0}^{\Delta-1} \Big(u_k[-(\tau_k-d)]+D_k[+\sigma_k+d]\Big)
+ \sum_{d = \sigma_k}^{\mathclap{\sigma_k+\Delta-2}} V_k[d], \end{multline*} and $b = \xi_k[\Delta]-S_k[\sigma_k+\Delta]$, which gives \begin{equation*}
b = \ \ \mathclap{\sum_{d=\Delta}^{\tau_k-1}} \ \ \ \Big(u_k[-(\tau_k-d)]+D_k[\sigma_{k}+d]\Big)\ \ \ \ \mathclap{\prod_{j=\Delta+2}^{d+1}} \ \ \ g_k(j)
+ \mu_{k+1}[0]g_{k+1}(1)\prod_{j=\Delta+2}^{\tau_k} g_k(j). \end{equation*} By applying \cref{lem:aux}-(i) it follows that \eqref{eq:cost_to_go} and the lemma holds for $k$ and $\Delta$ as well.
Thus the lemma holds for all $1\leq \Delta \leq\tau_k$ for $1\leq k\leq N-1$ and $1 \leq \Delta \leq H+2$ for $k=N$. \end{proof}
All that remains now is to find expressions for $V_k[\sigma_k]$ in terms of the initial conditions. The following lemma allows us to do so, using the expressions for $V_k$ derived in the previous lemma. \begin{lemma}\label{lem:mVsum} Let $h_k$, $P_k(i,j)$, $\phi_k(\Delta)$, $\pi_k[0]$, $\mu_k[0]$, $\Phi_k[0]$, and $\delta_k[0]$ be defined as in \cref{alg:init,alg:com_gen}. Then for $k\leq N-1$ \begin{equation}\label{eq:m_k_rewrite}
m_k[0] + \sum_{i=1}^{k}\sum_{d = \sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d] = \delta_{k}[0] - h_k\mu_{k+1}[0] \end{equation}
\end{lemma} \begin{proof} Let $B_k[0] = z_k[0] + u_k[-\tau_k] + D_k[\sigma_k]$ and $B_k[i] = u_k[-(\tau_k-i)] + D_k[\sigma_{k}+i]$ for $1\leq i < \tau_k$. We will prove the lemma by showing that for $1\leq \Delta \leq \tau_k$ \begin{multline}\label{eq:m_k_partial}
m_{k-1}[0] + \sum_{i=1}^{k-1}\sum_{d = \sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d] + \sum_{d = \sigma_k}^{\sigma_k+\Delta-1}V_k[d] =\\
(1-P_k(\Delta,1))\delta_{k-1}[0]
-\Big[h_{k-1}b_k(1-P(\Delta,1))+P_k(\Delta,\Delta)g_{k+1}(1)\ \ \ \mathclap{\prod_{j = \Delta+1}^{\tau_k}}\ \ \ g_k(j)\Big]\mu_{k+1}[0] \\
- \sum_{d=0}^{\tau_k-1}B_k[d]\Big[
P_k(\Delta,\min(d+1,\Delta))\prod_{j=\Delta+1}^{d+1}g_k(j)
+(1-P_k(\Delta,1))h_{k-1}\prod_{j=2}^{d+1}g_k(j)
\Big] \end{multline} More specifically we will show that \begin{enumerate}
\item \eqref{eq:m_k_partial} holds for $k=1$ and $\Delta = 1$.
\item If \eqref{eq:m_k_partial} holds for some $k$ and $\Delta-1$ then it holds for $\Delta$ as well.
\item If \eqref{eq:m_k_rewrite} holds for $k-1$ then \eqref{eq:m_k_partial} holds for $k$ and $\Delta = 1$.
\item If \eqref{eq:m_k_partial} holds for $k$ and $\Delta = \tau_k$ then $\eqref{eq:m_k_rewrite}$ holds for $k$. \end{enumerate}
For $k=1$ and $\Delta = 1$ the LHS of \eqref{eq:m_k_partial} is just $V_1[t]$. The RHS of \eqref{eq:m_k_partial} equals $-X_1(1)/\rho_1\cdot\mu_k[0]$ as $P_1(1,m) = X_1(1)/\rho_1$, $\delta_0=0$, and $h_0 = 0$. And by \cref{lem:V_expr} the RHS is also equal to $V_1[t]$ since by \eqref{eq:delta1reduce} \[
\xi_1[0] = \mu_k[0]. \]
Thus \eqref{eq:m_k_partial} holds for $k=1$ and $\Delta = 1$.
Applying \cref{lem:V_expr} on $V_k[\sigma_k+\Delta]$ for the LHS of \eqref{eq:m_k_partial} gives: \begin{multline}\label{eq:ind_step}
m_{k-1}[0] + \sum_{i=1}^{k-1}\sum_{d = \sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d] + \sum_{d = \sigma_k}^{\sigma_k+\Delta-1}V_k[d] = \\
(1-\frac{X_k(\Delta)}{\rho_k})\Big( m_{k-1}[0] + \sum_{i=1}^{k-1}\sum_{d = \sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d] + \sum_{d = \sigma_k}^{\sigma_k+\Delta-2}V_k[d] \Big)\\
-\frac{X_k(\Delta)}{\rho_k}\Big[\sum_{d=0}^{\tau_k-1} B_k[d]\prod_{j=\Delta+1}^{d+1}g_k(j) +
\mu_{k+1}[0]g_{k+1}(1)\prod_{j=\Delta+1}^{\tau_k} g_k(j)\Big] \end{multline} Now assume that \eqref{eq:m_k_partial} holds for $k$ and $\Delta-1$, we then show that \eqref{eq:m_k_partial} holds for $k$ and $\Delta$. Using \eqref{eq:ind_step} gives for the coefficients for the different terms of the LHS for \eqref{eq:m_k_partial} as follows. For $\delta_{k-1}[0]$ we get \[
(1-\frac{X_k(\Delta)}{\rho_k})(1-P_k(\Delta-1,1)) = 1-P_k(\Delta,1). \] For the terms in front of $\mu_{k+1}[0]$ we get \begin{multline*}
-\Big(1-\frac{X_k(\Delta)}{\rho_k}\Big)\Big(h_{k-1}b_k \big(1-P(\Delta-1,1) \big)
+P_k(\Delta-1,\Delta-1)g_{k+1}(1)\prod_{j = \Delta}^{\tau_k} g_k[j]\Big) \\
- \frac{X_k(\Delta)}{\rho_k}g_{k+1}(1)\prod_{j=\Delta+1}^{\tau_k} g_k(j) \\
=-h_{k-1}b_k\big(1-P(\Delta,1)\big) - P_k(\Delta,\Delta)g_{k+1}(1)\prod_{j=\Delta+1}^{\tau_k} g_k(j), \end{multline*} and for the coefficient for $B[d]$, \begin{multline*} -(1-\frac{X_k(\Delta)}{\rho_k})\Big[P_k(\Delta-1,\min(d+1,\Delta-1))\prod_{j=\Delta}^{d+1}g_k(j)
+ \\
(1-P_k(\Delta-1,1))h_{k-1}\prod_{j=2}^{d+1}g_k(j)\Big] - \frac{X_k(\Delta)}{\rho_k}\prod_{j=\Delta+1}^{d+1}g_k(j)\\ = - P_k(\Delta,\min(d+1,\Delta))\sum_{j=\Delta+1}^{d+1}g_k(j) -(1-P_k(\Delta,1))h_{k-1}\prod_{j=2}^{d+1}g_k(j). \end{multline*} Thus \eqref{eq:m_k_partial} holds for $k$ and $\delta$ as well.
Assume that \eqref{eq:m_k_rewrite} holds for $k-1$. Then we can show that \eqref{eq:m_k_partial} holds for $k$ and $\Delta = 1$. Using that \begin{equation}\label{eq:pi_k}
\pi_k[0] = \sum_{d=0}^{\tau_k-1}\Big( B[d]\prod_{j=2}^{d+1}g_k(j)\Big), \end{equation} the RHS of \eqref{eq:m_k_partial} reduces to \begin{equation*} \big[1-P_k(1,1)\big]\delta_{k-1}[0]
-\big[h_{k-1}(1-P_k(1,1))+P_k(1,1)\big](\pi_k[0] + b_k\mu_{k+1}[0]) . \end{equation*} Using \eqref{eq:ind_step} with $\Delta = 1$, the definition for $P_k(1,1)$, and inserting \eqref{eq:m_k_rewrite} gives that the LHS of \eqref{eq:m_k_partial} is equal to \begin{equation*} (1-P_k(1,1))\Big[\delta_{k-1}[0] - h_{k-1}\mu_{k}[0] \Big]
- P_k(1,1)\Big[\sum_{d=0}^{\tau_k-1} B[d]\prod_{\mathclap{j=\Delta+1}}^{d+1}g_k(j) +
\mu_{k+1}[0]b_k\Big]. \end{equation*} Using \eqref{eq:pi_k} and the definition for $\mu_k[0] = \pi_k[0] + b_k \mu_{k+1}[0]$ shows that the RHS and LHS are equal. And thus \eqref{eq:m_k_partial} hold for $k$ and $\Delta = 1$ if \eqref{eq:m_k_rewrite} holds for $k-1$.
Finally, we will show that if \eqref{eq:m_k_partial} holds for $k$ and $\Delta = \tau_k$ then \eqref{eq:m_k_rewrite} holds for $k$. Using the definition for $h_k$ the RHS of \eqref{eq:m_k_partial} reduces to \begin{multline*}
(1-P_k(\tau_k,1))\delta_{k-1}[0]
+h_k\mu_{k+1}[0] \\
- \sum_{i=d}^{\tau_k}B_k[d]\Big[P_k(\tau_k,d+1) + \big(1-P_k(\tau_k,1)h_{k-1}\big)\prod_{j=2}^{d+1}g_k(j)\Big] \end{multline*} For $\Delta = \tau_k$ the LHS of \eqref{eq:m_k_rewrite} is equal to the LHS of \eqref{eq:m_k_partial} plus \[
z_k[0] +\sum_{d=1}^{\tau_k}u_k[-d] + \sum_{d=\sigma_k}^{\sigma_{k+1}-1} D_k[d] = \sum_{d=0}^{\tau_k-1}B_k[d]. \] Thus it holds that the LHS of \eqref{eq:m_k_rewrite} is equal to \begin{multline*}
(1-P_k(\tau_k,1))\delta_{k-1}[0]
+h_k\mu_{k+1}[0] \\
+ \sum_{i=d}^{\tau_k}B_k[d]\Big[(1-P_k(\tau_k,d+1)) - \big(1-P_k(\tau_k,1)h_{k-1}\big)\prod_{j=2}^{d+1}g_k(j)\Big]. \end{multline*} Using the definition for $\phi_i(\Delta)$ in \cref{alg:init} and $\Phi_i$ and $\delta_k$ in \cref{alg:com_gen} shows that \eqref{eq:m_k_partial} gives \eqref{eq:m_k_rewrite} for $\Delta=\tau_k$, as \[
\sum_{i=d}^{\tau_k}B_k[d]\Big[(1-P_k(\tau_k,d+1)) - \big(1-P_k(\tau_k,1)h_{k-1}\big)\prod_{j=2}^{d+1}g_k(j)\Big] = \Phi_k[0]. \]
\end{proof}
We are now finally ready to prove the theorem, which follows from the previous lemmas.
\emph{Proof of \cref{thm:gen}:} \Cref{lem:V_expr} with $\Delta = 1$ and \cref{lem:mVsum} gives that \begin{equation}\label{eq:opt_V}\begin{aligned}
V_k[\sigma_k] &= -\frac{X_k(1)}{\rho_k}\Big[m_{k-1}[0] + \sum_{i=1}^{k-1}\sum_{d = \sigma_i}^{\sigma_{i+1}-1}\bar{V}_i[d]
+ \mu_k[0] \Big]\\
& = -\frac{X_k(1)}{\rho_k}\Big[\delta_{k-1}[0] + (1-h_k)\mu_k[0] \Big] \end{aligned}\end{equation}
from which the optimal $v_k[0] = \rho_k/r_k\cdot V_k[\sigma_k]$ as in \cref{alg:com_gen} follows. Using \cref{lem:shifted_inv}-(iii), \cref{lem:mVsum}, and that $v_k[0] = \rho_k/r_k\cdot V_k[\sigma_k]$ gives that \begin{multline*}
u_{k-1}[0] = (1-\frac{\gamma_k}{q_k})(z_k[t]+u_k[-\tau_k]+D_k[0]) +d_k[0]-D_k[0]\\
+V_k[\sigma_k]\Big(\frac{\rho_k}{r_k}-\frac{\gamma_k}{q_k} \Big) - \frac{\gamma_k}{q_k}\Big(\delta_{k-1}[0]-h_{k-1}\mu_k[0] \Big). \end{multline*} Inserting \eqref{eq:opt_V} gives that the optimal $u$ is as in \cref{alg:com_gen}.
The results will hold for $t\neq 0$ as the problem has an infinite horizon and one can always change the variables so that current time is time zero. \qed
\end{document}
|
arXiv
|
{
"id": "2106.04282.tex",
"language_detection_score": 0.7687753438949585,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Controlling Directionality and Dimensionality of Wave Propagation through Separable Bound States in the Continuum} \author{Nicholas Rivera$^{1}$, Chia Wei Hsu $^{1,2}$, Bo Zhen $^{1,3}$, Hrvoje Buljan$^{4}$, John D. Joannopoulos$^{1}$ \& Marin Solja\v{c}i\'{c}$^{1}$}
\affiliation{$^{1}$Department of Physics, Massachusetts Institute of Technology, Cambridge, MA 02139, USA \\ $^{2}$Department of Applied Physics, Yale University, New Haven, CT 06520, USA. \\ $^{3}$Physics Department and Solid State Institute, Technion, Haifa 32000, Israel \\ $^{4}$Department of Physics, University of Zagreb, Zagreb 10000, Croatia. }
\begin{abstract} \noindent
A bound state in the continuum (BIC) is an unusual localized state that is embedded in a continuum of extended states. \st{BICs can arise from the separability of the wave equation.} Here, we present the general condition for \st{separable} BICs to \st{exist, which allows us to control} \hl{arise from wave equation separability and show that} the directionality and dimensionality of \hl{their} resonant radiation \hl{can be controlled} by exploiting \hl{perturbations of certain} symmetry. Using this general framework\st{ for separable BICs}, we construct new examples of \st{them} \hl{separable BICs} in realistic models of optical potentials for ultracold atoms, photonic systems, and systems described by tight binding. Such BICs with easily reconfigurable radiation patterns allow for applications such as the storage and release of waves at a controllable rate and direction, systems that switch between different dimensions of confinement, and experimental realizations in atomic, optical, and electronic systems. \end{abstract}
\maketitle
\section{Introduction} The usual frequency spectrum of waves in an inhomogeneous medium consists of localized waves whose frequencies lie outside a continuum of delocalized propagating waves. For some rare systems, the bound states can be at the same frequency as an extended state. \st{These states, called} \hl{Such} bound states in the continuum (BICs) represent a method of wave localization that does not forbid outgoing waves from propagating, unlike traditional methods of localizing waves such as potential wells in quantum mechanics, conducting mirrors in optics, band-gaps, and Anderson localization.
BICs were first \st{shown to exist} \hl{predicted} theoretically in quantum mechanics by von Neumann and Wigner~\cite{neumann1929}. However, their BIC-supporting potential was highly oscillatory and could not be implemented in reality. The BIC was seen as a mathematical curiosity for the following 50 years until Friedrich and Wintgen showed that BICs could exist in models of multi-electron atoms~\cite{friedrich1985}. Since then, many theoretical examples of BICs have been demonstrated in quantum mechanics~\cite{robnik1986,nockel1992,duclos2001,prodanovic2013, ladron2003,ordonez2006,moiseyev2009},
electromagnetism~\cite{ochiai2001,ctyroky2001,watts2001,kawakami2002,apalkov2003,shipman2005,porter2005,bulgakov2008,marinica2008,molina2012,2013_Hsu_LSA,bulgakov2014,monticone2014,2014_Bulgakov_OL}, \hl{acoustics~\cite{porter2005}, and water waves~\cite{cobelli2011},} \st{and more general wave frameworks. On the other hand, experimental realizations are much fewer in number} \hl{with some experimentally realized~\cite{plotnik2011,lee2012,2013_Hsu_Nature,corrielli2013,weimann2013,lepetit2014}.} A few mechanisms explain most examples of BICs that have been discovered. \st{One of the} \hl{The} simplest mechanism is \st{that of the} symmetry\st{-protected BICs} \hl{ protection}, in which the BIC is decoupled from its degenerate continuua by symmetry mismatch~\cite{ochiai2001,fan2002,ladron2003,plotnik2011,lee2012,shipman2005,moiseyev2009}. \st{Another type arises when two resonances couple, leading one of them to become a BIC \cite{friedrich1985,marinica2008,lepetit2014}. When two resonant states with non-zero imaginary part of their frequency cross in the real part of their frequency as a parameter in the Hamiltonian is varied, one of these resonances can become a BIC.}\hl{Or, when two resonances are coupled, they can form BICs through Fabry-Perot interference~\cite{ordonez2006,marinica2008, 2013_Hsu_LSA, weimann2013} or through radiative coupling with tuned resonance frequencies~\cite{friedrich1985,lepetit2014,2014_Bulgakov_OL}.} More generally, parametric BICs can occur for special values of parameters in the Hamiltonian~\cite{porter2005,bulgakov2008,2013_Hsu_Nature,yi2014,bulgakov2014,monticone2014}, which can in some cases be understood using topological arguments~\cite{zhen2014}.
There exists yet another class of BICs, which has thus far been barely explored, and literature covering it is rare and scattered; these are BICs \hl{due to separability}~\cite{robnik1986,nockel1992,prodanovic2013,ctyroky2001,watts2001,kawakami2002,apalkov2003,duclos2001}. In this paper, we develop the most general properties of separable BICs \st{-} by providing the general criteria for \st{the} \hl{their} existence \st{of separable BICs} and by characterizing the continuua of these BICs. We demonstrate that separable BICs enable control over both the directionality and the dimensionality of resonantly emitted waves by exploiting symmetry, which is not possible in other classes of BICs. Our findings \hl{may} lead to applications such as\st{: a new class of} tunable-$Q$ directional resonators and quantum systems which can be switched between quantum dot (0D), quantum wire (1D), and quantum well (2D) modes of operation. While previous works on separable BICs have been confined to somewhat artificial systems, we \hl{also} present readily experimentally observable examples of separable BICs in photonic systems with directionally resonant states, as well as examples in optical potentials for cold atoms.
\section{Results} \subsection{General Condition for Separable BICs}
We start by considering a simple example of a two-dimensional separable system - one where the Hamiltonian operator can be written as a sum of Hamiltonians that each act on distinct variables $x$ or $y$, i.e; \begin{equation} H = H_x(x) + H_y(y). \end{equation} Denoting the eigenstates of $H_x$ and $H_y$ as $\psi_{x}^i(x)$ and $\psi_{y}^j(y)$, with energies $E_{x}^i$ and $E_{y}^j$, it follows that $H$ is diagonalized by the basis of product states $\psi_{x}^i(x)\psi_{y}^j(y)$ with energies $E^{i,j}=E_{x}^i+E_{y}^j$. If $H_x$ and $H_y$ each have a continuum of extended states starting at zero energy, and these Hamiltonians each have at least one bound state, $\psi_x^0$ and $\psi_y^0$, respectively, then the continuum of $H$ starts at energy $\text{min}(E_{x}^0,E_{y}^0)$, where the $0$ subscript denotes ground states. Therefore, if there exists a bound state satisfying $E^{i,j}>\text{min}(E_{x}^0,E_{y}^0)$, then it is a bound state in the continuum. We schematically illustrate this condition being satisfied in Figure 1, where we illustrate a separable system in which $H_x$ has two bound states, $\psi_x^0$ and $\psi_x^1$, and $H_y$ has one bound state, $\psi_y^0$. The first excited state of $H_x$ combined with the ground state of $H_y$, $\psi_x^1\psi_y^0$, while spatially bounded, has larger energy, $E_x^1 + E_y^0$, than the lowest continuum energy, $E_x^0$, and is therefore a BIC.
To extend separability to a Hamiltonian with a larger number of separable degrees of freedom is then straightforward. The Hamiltonian can be expressed in tensor product notation as $H = \sum\limits_{i=1}^N H_i$, where $H_i \equiv I^{\otimes i-1}\otimes h_i \otimes I^{\otimes N-i}$. In this expression, $N$ is the number of separated degrees of freedom, $h_i$ is the operator acting on the $i$-th variable, and $I$ is the identity operator. The variables may refer to the different particles in a non-interacting multi-particle system, or the spatial and polarization degrees of freedom of a single-particle system. Denote the $n_j$th eigenstate of $h_i$ by $|\psi_{i}^{n_j}\rangle$ with energy $E_i^{n_j}$. Then, the overall Hamiltonian $H$ is trivially diagonalized by the product states $|n_1,n_2,\cdot\cdot\cdot ,n_N\rangle \equiv |\psi_1^{n_1}\rangle\otimes|\psi_2^{n_2}\rangle\otimes\cdot\cdot\cdot\otimes|\psi_N^{n_N}\rangle$ with corresponding energies $E = E_1^{n_1} + E_2^{n_2} + \cdot\cdot\cdot + E_N^{n_N}$. Denoting the ground state of $h_i$ by $E_i^{0}$ and defining the zeros of the $h_i$ such that their continuua of extended states start at energy zero, the continuum of the overall Hamiltonian starts at $\underset{{1\leq i \leq N}}{\text{min}}(E_{i}^{0})$. Then, if the separated operators $\{ h_i \}$ are such that there exists a combination of separated bound states satisfying $E = \sum\limits_{i=1}^N E_{i}^{n_i} > \underset{{1\leq i \leq N}}{\text{min}}(E_{i}^{0})$, this combined bound state is a BIC of the overall Hamiltonian. For such separable BICs, coupling to the continuum is forbidden by the separability of the Hamiltonian. Stating the most general criteria for separable BICs allows us to straightforwardly extend the known separable BICs to systems in three dimensions, multi-particle systems, and systems described by the tight-binding approximation.
\subsection{Properties of the Degenerate Continuua}
A unique property that holds for all separable BICs is that the delocalized modes degenerate to the BIC are always guided in at least one direction. In many cases, there are multiple degenerate delocalized modes and they are guided in different directions. In 2D, we can associate partial widths $\Gamma_x$ and $\Gamma_y$ to the resulting resonances of these systems when the separable system is perturbed. These partial widths are the decay rates associated with energy-conserving transitions between the BIC and states delocalized in the $x$ and $y$ directions under perturbation, respectively. When separability is broken, we generally can not decouple leakage in the $x$ and $y$ directions because the purely-$x$ and purely-$y$ delocalized continuum states mix. However, in this section, we show that one can control the radiation to be towards the x (or y) directions only, by exploiting the symmetry of the perturbation.
In Fig. 2(a), we show a two-dimensional potential in a Schrodinger-like equation which supports a separable BIC. This potential is a sum of Gaussian wells in the $x$ and $y$-directions, given by \begin{equation} V(x,y) = -V_xe^{-\frac{2x^2}{\sigma_x^2}}-V_ye^{-\frac{2y^2}{\sigma_y^2}}. \end{equation} This type of potential is representative of realistic optical potentials for ultracold atoms and potentials in photonic systems. Solving the time-independent Schrodinger equation for $\{V_x, V_y \}=\{1.4, 2.2\}$ and $\{\sigma_x,\sigma_y \} = \{5,4\}$ (in arbitrary units) gives the spectra shown in Fig. 2(b). Due to the $x$ and $y$ mirror symmetries of the system, the modes have either even or odd parity in both the $x$ and the $y$ directions. This system has several BICs. Here we focus on the BIC $|n_x,n_y\rangle = |2,1\rangle$ at energy $E^{2,1} =-1.04$, with the mode profile shown in Fig.~2(c); being the second excited state in $x$ and the first excited state in $y$, this BIC is even in $x$ and odd in $y$. It is only degenerate to continuum modes $|0,E^{2,1}-E_{x}^0\rangle$ extended in the $y$ direction (Fig. 2(d)), and $|E^{2,1}-E_{y}^0,0\rangle$ extended in the $x$ direction (Fig. 2(e)), where an $E$ label inside a ket denotes an extended state with energy $E$.
If we choose a perturbation that preserves the mirror symmetry in the $y$-direction, as shown in the inset of Fig.~2(f), then the perturbed system still exhibits mirror symmetry in $y$. Since the BIC $|2,1\rangle$ is odd in $y$ and yet the $x$-delocalized continuum states $|E^{2,1}-E_{y}^0,0\rangle$ are even in $y$, there is no coupling between the two. As a result, the perturbed state radiates only in the $y$ direction, as shown in the calculated mode profile in Fig.~2(f). This directional coupling is a result of symmetry, and so it holds for arbitrary perturbation strengths.
On the other hand, if we apply a perturbation that is odd in $x$ but not even in $y$, as shown in the inset of Fig.~2(g), then there is radiation in the $x$ direction only to first-order in time-dependent perturbation theory. Specifically, for weak perturbations of the Hamiltonian, $\delta V$, the first-order leakage rate is given by Fermi's Golden Rule for bound-to-continuum coupling, $\Gamma \sim \sum_c |\langle 2,1 |\delta V| \psi_c \rangle|^2\rho_c(E^{2,1})$, where $\rho_c(E)$ is the density of states of continuum $c$, and $c$ labels the distinct continuua which have states at the same energy as the BIC. Since the BIC and the $y$-delocalized continuum states $|0,E^{2,1}-E_{x}^0\rangle$ are both even in $x$, the odd-in-$x$ perturbation does not couple the two modes directly, and $\Gamma_y$ is zero to the first order. As a result, the perturbed state radiates only in the $x$ direction, as shown in Fig.~2(g). At the second order in time-dependent perturbation theory, the BIC can make transitions to intermediate states $k$ at any energy, and thus the second-order transition rate, proportional to $\sum_c |T_{ci}|^2 = \sum_c |\sum_k \frac{\langle c|\delta V|k \rangle\langle k|\delta V|i\rangle}{E_i - E_k + i0^+}|^2$ does not vanish because the intermediate state can have even parity in the $x$-direction.
Another unique aspect of separability is that by using separable BICs in 3D, the dimensionality of the confinement of a wave can be switched between one, two, and three by tailoring perturbations applied to a single BIC mode. The ability to do this allows for a device which can simultaneously act as a quantum well, a quantum wire, and a quantum dot. We demonstrate this degree of control using a separable potential generated by the sum of three Gaussian wells (in $x$, $y$, and $z$ directions) of the form in Equation 2 with strengths $\{0.4,0.4,1 \}$ and widths $\{ 12,12,3 \}$, all in arbitrary units. The identical $x$ and $y$ potentials have four bound states at energies $E_x = E_y = -0.33, -0.20, -0.10 \text{ and } -0.029$. The $z$ potential has two bound states at energies $E_z =-0.61 \text{ and } -0.059$. All of the continuua start at zero. The BIC state $|1,1,1\rangle$ at energy $E^{1,1,1}=-0.47$ is degenerate to $|0,0,E_z\rangle,|0,1,E_z\rangle,|1,0,E_z\rangle,|E_x,m,0\rangle,|n,E_y,0\rangle$ and $|E_x,E_y,0\rangle$, where $E_i$ denotes an energy above zero, and $m$ and $n$ denote bound states of the $x$ and $y$ wells. For perturbations which are even in the $z$ direction, the BIC $|1,1,1\rangle$ does not couple to states delocalized in $x$ and $y$ ($|E_x,m,0\rangle,|n,E_y,0\rangle,\text{ and }|E_x,E_y,0\rangle$), because they have opposite parity in $z$, so the BIC radiates in the $z$-direction only. On the other hand, for perturbations which are even in the $x$ and $y$ directions, the coupling to states delocalized in $z$ ($|0,0,E_z\rangle,|0,1,E_z\rangle,\text{ and }|1,0,E_z\rangle$) vanishes, meaning that the BIC radiates in the $xy$-plane.
\subsection{Proposals for the experimental realizations of separable BICs}
BICs are generally difficult to experimentally realize because they are fragile under perturbations of system parameters. On the other hand, separable BICs are straightforward to realize and also robust with respect to changes in parameters that preserve the separability of the system. This reduces the difficulty of experimentally realizing separable BICs to ensuring separability. And while this is still in general nontrivial, it can be straightforwardly achieved in systems that respond linearly to the intensity of electromagnetic fields. In the next two realistic examples of BICs that we demonstrate, we use detuned light sheets to generate separable potentials in photonic systems and also for ultracold atoms.
\textit{Paraxial optical systems $-$} As a first example of this, consider electromagnetic waves propagating paraxially along the $z$-direction, in an optical medium with spatially non-uniform index of refraction $n(x,y) = n_{0} + \delta n(x,y)$, where $n_0$ is the constant background index of refraction and $\delta n \ll n_{0}$. The slowly varying amplitude of the electric field $\psi(x,y,z)$ satisfies the two-dimensional Schr\"{o}dinger equation (see Ref. 34 and references therein),
\begin{equation} i\frac{\partial \psi}{\partial z}= -\frac{1}{2k}\nabla_{\perp}^2 \psi-\frac{k \delta n }{n_0}\psi. \label{paraxial} \end{equation}
Here, $\nabla_{\perp}^2=\partial^2/\partial x^2+\partial^2/\partial y^2$, and $k=2\pi n_0/\lambda$, where $\lambda$ is the wavelength in vacuum. The modes of the potential $\delta n(x,y)$ are of the form $\psi=A_j(x,y)e^{i\beta_j z}$, where $A_j$ is the profile, and $\beta_j$ the propagation constant of the $j$th mode. In the simulations we use $n_0=2.3$, and $\lambda=485$~nm.
Experimentally, there are several ways of producing potentials $\delta n(x,y)$ of different type, e.g., periodic\cite{fleischer2005}, random\cite{schwartz2007}, and quasicrystal\cite{freedman2006}. One of the very useful techniques is the so-called optical induction technique where the potential is generated in a photosensitive material (e.g., photorefractives) by employing light~\cite{fleischer2003}. We consider here a potential generated by two perpendicular light sheets, which are slightly detuned in frequency, such that the time-averaged interference vanishes and the total intensity is the sum of the intensities of the individual light sheets. The light sheets are much narrower in one dimension ($x$ for one, $y$ for the other) than the other two, and therefore each light sheet can be approximated as having an intensity that depends only on one coordinate, making the index contrast separable. This is schematically illustrated in Fig. 3(a).
If the sheets are Gaussian along the narrow dimension, then the potential is of the form $\delta n(x,y)=\delta n_0 [\exp(-2(x/\sigma)^2)+\exp(-2(y/\sigma)^2)]$. It is reasonable to use $\sigma=30$~$\mu$m and $\delta n_0 = 5.7 \times 10^{-4}$. For these parameters, the one dimensional Gaussian wells have four bound states, with $\beta$ values of (in mm$^{-1}$): $2.1,1.3,0.55,\text{ and } 0.11$.
There are eight BICs: $|1,2\rangle,|2,1\rangle,|1,3\rangle,|3,1\rangle$,
$|2,3\rangle$, $|3,2\rangle, |2,2\rangle$, and $|3,3\rangle$.
Among them, $|1,3\rangle$ and $|3,1\rangle$ are symmetry protected.
Additionally, the BICs $|1,3\rangle$ and $|3,1\rangle$ can be used to demonstrate directional resonance in the $x$ or $y$ directions, respectively, by applying perturbations even in $x$ or $y$, respectively. Therefore, this photonic system serves as a platform to demonstrate both separable BICs and directional resonances.
\textit{Optical potentials for ultracold atoms $-$} The next example that we consider can serve as a platform for the first experimental measurement of BICs in quantum mechanics. Consider a non-interacting neutral Bose gas in an optical potential. Optical potentials are created by employing light sufficiently detuned from the resonance frequencies of the atom, where the scattering due to spontaneous emission can be neglected, and the atoms are approximated as moving in a conservative potential. The macroscopic wavefunction of the system is then determined by solving the Schr\"{o}dinger equation with a potential that is proportional to the intensity of the light~\cite{bloch2005}.
As an explicit example, consider an ultracold Bose gas of $^{87}Rb$ atoms. An optical potential is made by three Gaussian light sheets with equal intensity $I_1 = I_2 = I_3$, widths $20 \text{ }\mu\text{m}$, and wavelengths centered at $\lambda=1064 \text { nm}$~\cite{henderson2009}. This is schematically illustrated in Fig. 3(b). The intensity is such that this potential has depth equal to ten times the recoil energy $E_r = \frac{h^2}{2m\lambda^2}$. By solving the Schrodinger equation numerically, we find many BICs; the continuum energy starts at a reduced energy $\xi_c = \frac{2mEx_0^2}{\hbar^2} = -296.24$, where $x_0$ is chosen to be $1 \text{ }\mu m$. The reduced depth of the trap is $-446.93$. Each one-dimensional Gaussian supports 138 bound states. There are very many BICs in such a system. For concreteness, an example of one is
$|30,96,96\rangle$, with reduced energy $-146.62$.
\textit{Tight Binding Models$-$} The final example that we consider here is an extension of the separable BIC formalism to systems which are well-approximated by a tight-binding Hamiltonian that is separable. Consider the following one-dimensional tight-binding Hamiltonian, $H_i$, which models a one-dimensional lattice of non-identical sites: \begin{equation} H_i = \sum\limits_{k}\varepsilon_i^{k} |k\rangle\langle k|+t_i \sum \limits_{\langle lm \rangle}(|l\rangle\langle m| + |m\rangle\langle l| ), \end{equation} where $\langle lm \rangle$ denotes nearest-neighbors, $\varepsilon_i^k $ is the on-site energy of site $k$, and $k,l,\text{ and }m$ run from $-\infty$ to $\infty$. Suppose $\varepsilon_i^k = -V$ for $|k| < N$, and zero otherwise. For two Hamiltonians of this form, $H_1$ and $H_2$, $H_1 \otimes I + I \otimes H_2$ describes the lattice in Fig. 3(c). If we take $H_1 = H_2$ with $\{ V,t,N \} = \{ -1,-0.3,2 \}$, in arbitrary units, the bound state energies of the 1D-lattices are numerically determined to be $-0.93,-0.74,-0.46,\text{ and } -0.16$. Therefore the states $|2,2\rangle, |2,3\rangle, |3,2\rangle, |3,3\rangle, |3,1\rangle \text{ and } |1,3\rangle$ are BICs. The last two of these are also symmetry-protected from the continuum as they are odd in $x$ and $y$ while the four degenerate continuum states are always even in at least one direction. Of course, many different physical systems can be adjusted to approximate the system from Eq. (4), so this opens a path for observing separable BICs in a wide variety of systems.
\section{Summary}
With the general criterion for separable BICs, we have extended the existing handful of examples to a wide variety of wave systems including: three-dimensional quantum mechanics, paraxial optics, and lattice models which can describe 2D waveguide arrays, quantum dot arrays, optical lattices, and solids. These BICs exist in other wave equations such as the 2D Maxwell equations, and the inhomogeneous scalar and vector Helmholtz equations, meaning that their physical domain of existence extends to 2D photonic crystals, microwave optics, and also acoustics. To our knowledge, they do not exist (except perhaps approximately) in isotropic media in the full 3D Maxwell equations $\nabla \times \nabla \times \bold{E} = \nabla(\nabla \cdot \bold{E}) - \nabla^2\bold{E} = \epsilon \frac{\omega^2}{c^2}\bold{E}$ because the operator $\nabla(\nabla \cdot \text{ })$ couples the different coordinates and renders the equation non-separable. We have demonstrated numerically the existence of separable BICs in models of trapped Bose-Einstein condensates and photonic potentials, showing that separable BICs can be found in realistic systems. These simple and realistic models can facilitate the observation of BICs in quantum mechanics, which to this date has not been conclusively done.
More importantly, we have demonstrated two new properties unique to separable BICs: the ability to control the direction of emission of the resonance using perturbations, and also the ability to control the dimensionality of the resulting resonance. This may lead to two applications. In the first, perturbations are used as a switch which can couple waves into a cavity, store them, and release them in a fixed direction. In the second, the number of dimensions of confinement of a wave can be switched between one, two, and three by exploiting perturbation parity. The property of dimensional and directional control of resonant radiation serves as another potential advantage of BICs over traditional methods of localization.
\section{Acknowledgments} The authors would like to acknowledge Prof. Steven G. Johnson, Prof. Marc Kastner, Prof. Silvija Grade\v{c}ak, Wujie Huang, and Dr. Wenlan Chen for useful discussions and advice. Work of M.S. was supported as part of S3TEC, an EFRC funded by the U.S. DOE, under Award Number DE-SC0001299 / DE-FG02-09ER46577. This work was also supported in part by the MRSEC Program of the National Science Foundation under award number DMR-1419807. This work was also supported in part by the U. S. Army Research Laboratory and the U. S. Army Research Office through the Institute for Soldier Nanotechnologies, under contract number W911NF-13-D-0001.
\section{Author Contributions} N.R. came up with the idea for separable BICs, and controlling directionality and dimensionality with them, in addition to doing the numerical simulations. C.W.H., B.Z., H.B., and M.S. helped come up with readily observable physical examples of separable BICs in addition to helping develop the concept of separable BICs. M.S. and J.D.J. provided critical supervision for the work. N.R. wrote the manuscript with critical reading and editing from B.Z., C.W.H., H.B., J.D.J., and M.S..
\begin{figure*}
\caption{A schematic illustration demonstrating the concept of a separable BIC in two dimensions.}
\end{figure*}
\begin{figure*}
\caption{(a) A separable potential which is a sum of a purely x-dependent Gaussian well and a purely y-dependent Gaussian well. (b) The relevant states of the spectrum of the $x$-potential,$y$-potential, and total potential. (c) A BIC supported by this double well. (d,e) Continuum states degenerate in energy to the BIC. (f) A $y$-delocalized continuum state resulting from an even-$y$-parity perturbation of the BIC supporting potential. (g). An $x$-delocalized continuum state resulting from an odd-$x$-parity perturbation.}
\end{figure*}
\begin{figure*}
\caption{Separable physical systems with BICs. (a) A photorefractive optical crystal whose index is weakly modified by two detuned intersecting light sheets with different intensities. (b) An optical potential formed by the intersection of three slightly detuned light sheets with different intensities. (c) A tight-binding lattice.}
\end{figure*}
\end{document}
|
arXiv
|
{
"id": "1507.00923.tex",
"language_detection_score": 0.8539109230041504,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\newcommand{\eta}{\eta} \newcommand{\bar{\eta}}{\bar{\eta}} \newcommand{\eta^{\prime}}{\eta^{\prime}} \newcommand{\bar{\eta}^{\prime}}{\bar{\eta}^{\prime}} \newcommand{{\bf k}}{{\bf k}} \newcommand{{\bf l}}{{\bf l}} \newcommand{{\bf q}}{{\bf q}} \newcommand{{\bf r}}{{\bf r}} \newcommand{z}{z} \newcommand{\bar{z}}{\bar{z}} \newcommand{z^{\prime}}{z^{\prime}} \newcommand{\bar{z}^{\prime}}{\bar{z}^{\prime}} \newcommand{\lambda}{\lambda} \newcommand{\bar{\lambda}}{\bar{\lambda}} \newcommand{\omega}{\omega} \newcommand{\epsilon}{\epsilon} \newcommand{\bar{\psi}}{\bar{\psi}} \newcommand{\psi^{\prime}}{\psi^{\prime}} \newcommand{\bar{\psi}^{\prime}}{\bar{\psi}^{\prime}} \newcommand{\bar{\phi}}{\bar{\phi}} \newcommand{\phi^{\prime}}{\phi^{\prime}} \newcommand{\bar{\phi}^{\prime}}{\bar{\phi}^{\prime}} \newcommand{\Gamma}{\Gamma} \newcommand{\sigma}{\sigma}
\title{Quantum Entanglement under Non-Markovian Dynamics of Two Qubits Interacting with a common Electromagnetic Field} \author{C.~Anastopoulos$^{1}$\footnote{Corresponding author. Email address: [email protected]}, S.~Shresta$^{2,3}$\footnote{Present address: MITRE Corporation 7515 Colshire Drive, MailStop N390 McLean, VA 22102. Email Address: sanjiv$\[email protected]}, and B.~L. Hu$^{2}$\footnote{Email address: [email protected]}} \affiliation{$^1$Department of Physics, University of Patras, 26500 Patras, Greece, \\$^2$Department of Physics, University of Maryland, College Park, Maryland 20742-4111 \\ $^3$NIST, Atomic Physics Division, Gaithersburg, MD 20899-8423}
\date{January 31, 2007}
\begin{abstract} We study the non-equilibrium dynamics of a pair of qubits made of two-level atoms separated in space with distance $r$ and interacting with one common electromagnetic field but not directly with each other. Our calculation makes a weak coupling assumption
but no Born or Markov approximation. We derived a non-Markovian master equation for the evolution of the reduced density matrix of the two-qubit system after integrating out the electromagnetic field modes. It contains a Markovian part with a Lindblad type operator and a nonMarkovian contribution, the physics of which is the main focus of this study. We use the concurrence function as a measure of quantum entanglement between the two qubits. Two classes of states are studied in detail: Class A is a one parameter family of states which are the superposition of the highest energy $|I \rangle \equiv
|11 \rangle$ and lowest energy $|O \rangle \equiv |00 \rangle$
states, {\it viz}, $|A \rangle \equiv \sqrt p|I \rangle + \sqrt
{(1-p)} |O \rangle$, with $ 0 \le p \le 1 $; and Class B states $|B
\rangle$ are linear combinations of the symmetric $|+ \rangle =
\frac{1}{\sqrt 2} (|01 \rangle + |10 \rangle)$ and the antisymmetric
$|- \rangle = \frac{1}{\sqrt 2} (|01 \rangle - |10 \rangle)$ Bell states. We obtain similar behavior for the Bell states as in earlier results derived by using the Born-Markov approximation
\cite{FicTan06} on the same model. However, in the Class $|A \rangle$ states the behavior is qualitatively different: under the non-Markovian evolution we do not see sudden death of quantum entanglement and subsequent revivals, except when the qubits are sufficiently far apart. (The existence of sudden death was first reported for two qubits in two disjoint cavity electromagnetic fields \cite{YuEbePRL}, and the dark period and revival were found from calculations using the Born-Markov approximation \cite{FicTan06}). For an initial Bell state, our findings based on non-Markovian dynamics agree with those obtained under the Born-Markov approximation. We provide explanations for such differences of behavior both between these two classes of states and between the predictions from the Markov and non-Markovian dynamics. We also study the decoherence of this two-qubit system and find that the decoherence rate in the case of one qubit initially in an excited state does not change significantly with the qubits separation whereas it does for the case when one qubit is initially in the ground state. Furthermore, when the two qubits are close together, the coherence of the whole system is preserved longer than it does in the single qubit case or when the two qubits are far apart. \end{abstract}
\maketitle
\section{Introduction}
Investigation of quantum entanglement is both of practical and theoretical significance: It is viewed as a basic resource for quantum information processing (QIP) \cite{NielsenChuang} and it is a basic issue in understanding the nature of nonlocality in quantum mechanics \cite{EPR,Bell,GHZ}. However, even its very definition and accurate characterization are by no means easy, especially for multi-partite states (see, e.g., \cite{PeresBook,KarolBook,AlickiBook,Wootters,Bennett,Wer,VidWer}.) Nonetheless there are useful criteria proposed for the separability of a bipartite state, pure and mixed (e.g., \cite{Peres,Horod,Simon,Duan,Barnum,Brennen,HHH}), theorems proven (e.g, \cite{Ruskai,AlickiHorod}), and new mathematical tools introduced (e.g., \cite{Brody,Levay}), which add to advances in the last decade of this new endeavor \cite{Knill}.
Realistic quantum systems designed for QIP cannot avoid interactions with their environments, which can alter their quantum coherence and entanglement properties. Thus quantum decoherence and disentanglement are two essential obstacles to overcome for the design of quantum computers and the implementation of QIP. Environment-induced decoherence in the context of QIP has been under study for over a decade \cite{PazZurRev} and studies of environment-induced disentanglement has seen a rapid increase in recent years. There are now experimental proposals to measure finite time disentanglement induced by a reservoir \cite{Franca}. The relation between decoherence and disentanglement is an interesting one because both are attributable to the decay of quantum interference in the system upon the interaction with an environment. (See, e.g., \cite{VKG,RajRendell,Diosi,Dodd,DoddHal})
In addition to the mathematical investigations mentioned above which could provide rather general characterizations of quantum entanglement, detailed studies of physical models targeting actual designs of quantum computer components can add precious insight into its behavior in concrete settings. Two classes of models relevant to condensed matter and atomic-optical QIP schemes are of particular interest to us. The first class consists of the quantum Brownian motion model (QBM) and the spin-boson model (SBM). Quantum decoherence has been studied in detail in both models, and results on quantum disentanglement are also appearing (See \cite{VKG,Diosi,Dodd,DoddHal} for QBM under high temperature, negligible dissipation, and \cite{CHY} for an attempt towards the full non-Markovian regimes.) The second class of models describes atoms or ions in a cavity with a quantum electromagnetic field at zero or finite temperature. The model consists of two two level atoms (2LA) in an electromagnetic field (EMF). For a primary source on this topic, read, e.g.. \cite{Agarwal}. For a more recent description of its dynamics under the Born-Markov approximation, see the review of \cite{FicTan}.
\subsection{Two-atom entanglement via an electromagnetic field}
Quantum decoherence and entanglement between one 2LA and an EMF has been treated by us and many other authors earlier \cite{AH,SADH} and by Cummings and Hu recently towards the strong coupling regime \cite{CumHu06}, which provide insight in how the atom-field interaction affects their entanglement. There is recent report of exact solutions found for a 2LA in an EMF using the underlying pseudo-supersymmetry of the system \cite{SSG}. In the 2 atom-EMF model, the two atoms can be assumed to interact only with its own cavity EMF, or with a common EMF, and they can also interact with each other. The noninteracting case of separate fields was first studied by Yu and Eberly \cite{YuEbePRL,YuEbePRB} where `sudden death' of quantum entanglement was sighted. The noninteracting case of a common field was studied recently by Ficek and Tanas \cite{FicTan06}, An {\it et al} \cite{AnWangLuo}. Quantum decoherence of N-qubits in an electromagnetic field was studied by Palma, Suominen and Ekert \cite{PSK}. For entanglement of ions in cavities, see, e.g., \cite{BPH}.
For the purpose of quantum information processing, we have emphasized earlier in our studies of quantum decoherence that it is absolutely essential to keep track fully of the mutual influence of, or the interplay between, the system and the environment. If one chooses to focus only on the system dynamics, one needs to take into consideration the back-action of the environment, and vice versa.
In our prior work \cite{AH,SADH}, we used the influence functional formalism with a Grassmannian algebra for the qubits (system) and a coherent state path integral representation for the EMF (environment). Here, we employ a more standard operator method through perturbation theory, because the assumption of an initial vacuum state for the EMF allows a full resummation of the perturbative series, thus leading to an exact and closed expression for the evolution of the reduced density matrix of the two qubits. This approach incorporates the back-action of the environment on the system self-consistently and in so doing generates non-Markovian dynamics. We shall see that these features make a fundamental difference in the depiction of evolution of quantum entanglement in the qubit dynamics.
\subsection{The importance of including back-action self-consistently}
Since quantum entanglement is a more delicate quantity to track down than decoherence, an accurate description is even more crucial. For this, one needs to pay extra attention to back-actions. For example, in the case of two 2LA (system) in a cavity EMF (environment), the two parties are equally important. This means that we should include both the back-action of the field on the atoms, and the atoms on the field. In a more complete treatment as attempted here, we obtain results qualitatively different from earlier treatments where the back-action is not fully included or properly treated \cite{FicTan}. Some special effects like `sudden death' \cite{YuEbePRL} can in this broader context be seen as consequences only of rather special arrangements: Each atom interacting with its own EMF precludes the fields from talking to each other and in turn cuts off the atoms' natural inclination (by the dictum of quantum mechanics) to be entangled. In effect, this is only a limiting case of the full dynamics we explored here for the two-qubit entanglement via the EMF. This limit corresponds to the qubits being separated by distances much larger than the correlation length characterizing the total system. For a wide range of spatial separations within the correlation length, entanglement is robust: Our results for the full atom-field dynamics reveal that there is no sudden death.
\subsection{Non-Markovian dynamics from back-action}
It is common knowledge in nonequilibrium statistical mechanics \cite{Zwanzig} that for two interacting subsystems the two ordinary differential equations governing each subsystem can be written as an integro-differential equation governing one such subsystem, thus rendering its dynamics non-Markovian, with the memory of the other subsystem's dynamics registered in the nonlocal kernels (which are responsible for the appearance of dissipation and noise should the other subsystem possess a much greater number of degrees of freedom and are coarse-grained in some way). Thus inclusion of back-action self-consistently in general engenders non-Markovian dynamics. Invoking the Markov approximation as is usually done may lead to the loss of valuable information, especially for quantum systems. These assumptions need to be scrutinized carefully with due consideration of the different time scales in the system and the specific physical quantities that are of interest in the study.
For monitoring the evolution of quantum entanglement which is usually a more delicate process than coherence, if one lacks detailed knowledge of how the different important processes interplay, our experience is that it is safer not to put in any ad hoc assumption at the beginning (e.g., Markovian approximation, high temperature, white noise) but to start with a first principles treatment of the dynamics (which is likely non-Markovian) involving all subsystems concerned and respecting full self-consistency. This is because entanglement can be artificially and unknowingly curtailed or removed in these ad hoc assumptions. What is described here is not a procedural, but a substantive issue, if one seeks to coherently follow or manipulate any quantum system, as in QIP, because doing it otherwise can generate quantitatively inaccurate or even qualitatively wrong results.
Thus the inclusion of backreaction (which depends on the type of coupling and the features of the environment) usually leads to nonMarkovian dynamics \footnote{This does not preclude the possibility that certain types of backreaction effects when included leads to the same type of open system dynamics as the original (test- field) dynamics, which, if already in the Lindblad form, remains Markovian (e.g., renormalization of the coupling constants, invoking a mean field approximation).}. Also, under extreme conditions such as imposing infinite cutoff frequency and at high temperatures, the dynamics of, say, a quantum harmonic oscillator bilinearly coupled to an Ohmic bath can become Markovian \cite{CalLeg83}. Other factors leading to or effecting nonMarkovian behavior include the choice of special initial conditions. For example, the factorizable initial condition introduces a fiducial or special choice of time into the dynamics which destroys time-homogeneity.
A word about terminology might be helpful here: One usually refers to Markovian dynamics as that governed by a master equation with constant-in-time coefficients, i.e., described by a Linblad operator, and non-Markovian for all other types of dynamics. A more restricted condition limits the definition of non-Markovian dynamics to cases with non-trivial (nonlocal in time) integral kernels appearing in the master equation. This more stringent definition would refer to dynamics (depicted by master equations containing coefficients which are) both time-homogeneous and non-time-homogeneous as Markovian. We use the first and more common convention of terminology, in which the master equation (\ref{Lindbladlike}) which is local in time but non-time-homogeneous would be nonMarkovian. The Markovian regime emerges in the limit when the two qubits are far separated. This feature is similar to the HPZ master equation \cite{HPZ} for quantum Brownian motion, where Markovian (time-homogeneous) dynamics appears only in specific limiting conditions (high temperature and ohmic distribution of environmental modes, as alluded to above).
Our present study of the two qubit (2qb)- EMF system is also aimed at addressing a common folklore, namely, that in quantum optics one does not need to worry about non-Markovian effects. We will see that there is memory effect residing in the off diagonal components of the reduced density matrix for the 2 qubit system which comes from virtual photon exchange processes mediated by the field and which depends on the qubit separation. Perhaps the simplest yet strongest reason for the need to take nonMarkovian effects seriously is that results from the Markovian approximation are incomplete and lead to qualitatively wrong predictions.
\subsection{Relation to prior work and organization of this paper}
In this paper we study the non-Markovian dynamics of a pair of qubits (2LA) separated in space by distance $r$ interacting with one common electromagnetic field (EMF) through a Jaynes-Cummings type interaction Hamiltonian. We use the concurrence function as a measure of quantum entanglement between the two qubits. The same model was studied before in detail by Ficek and Tanas \cite{FicTan} using the Born-Markov approximation.
In a more recent paper \cite{FicTan06} they show the existence of dark periods and revival of quantum entanglement in contrast to the findings of Yu and Eberly which assumes two qubits in disjoint EMFs.
Our calculation makes a weak coupling assumption and ignores the two- and higher- photon- exchanges between the qubits, but it makes no Born or Markov approximation. We derive a non-Markovian master equation, which differs from the usual one of the Lindblad type: it contains extra terms that correspond to off-diagonal elements of the density matrix propagator. We concentrate on two classes of states, superpositions of highest and lowest energy states and the usual antisymmetric $|- \rangle$ and symmetric $|+\rangle $ Bell states \cite{Bell} and observe very different behavior. These are described in detail in the Discussions section. The difference between our results and that of Ref. \cite{FicTan06} highlights the effect of non-Markovian (with memory) evolution of quantum entanglement. In short, we find similar behavior in the Class B (Bell) states but qualitative different behavior in the evolution of Class A states. Ref \cite{FicTan06} found that their evolution leads generically to sudden death of entanglement and a subsequent revival. In our more complete treatment of the atom-field dynamics we indeed see the former effect present for large values of the inter-qubit distances. However, sudden death is absent for short distances, while there is no regime in which a revival of entanglement can take place. This calls for caution.
Another set of papers close to our work reported here is that of \cite{JakJam} who considered two 2LA in an infinite temperature field bath. When the atoms are separated at large distance the authors assume that they are located inside two independent baths. ( The severance of the field is subject to the same criticism above: A small but finite quantum entanglement cannot be equated to zero because the small amount can later grow.) For these conditions and under the Markovian approximation, the time evolution of the two-atom system is given by the ergodic dynamical semi-group. They ignore without justification the effect of distance on the interaction between the qubits. A paper of interest not directly related to the present model but which does show the dependence of the disentanglement rate on distance, like ours reported here, is that by Roszak and Machinikowski \cite{RosMac}. They consider a system of excitons with different coupling, and with a very different infrared behavior of the bath modes. The latter seems not to be relevant to the two atoms' case here.
This paper is organized as follows: Section 2 contains the main derivation. We write down the Hamiltonian for two 2-level atoms (2LA) interacting with a common electromagnetic field (EMF) at zero temperature, and we compute the relevant matrix elements for the propagator of the total system by resummation of the perturbative series (Appendix A). We then determine the evolution of the reduced density matrix of the atoms, which is expressed in terms of seven functions of time. We compute these functions using an approximation that amounts to keeping the contribution of the lowest loop order for the exchange of photons between the qubits. In Section 3 we examine the evolution of the reduced density matrix for two classes of initial states. We then describe the time evolution of quantum entanglement with spatial separation dependence in these states via the concurrence plotted for some representative cases. In Section 4, we study the decoherence of this system when the two qubits are initially disentangled. We consider two cases that correspond to one of the qubits being initially in a vacuum state and in an excited state. We compare these results with the single qubit cases and highlight the lessening of decoherence due to the presence of other qubit(s). In Section 5 we discuss and compare our results on disentanglement with the work of Yu and Eberly for two 2LA in separate EMF baths, and with the work of Fizek and Tanas on two 2LA in a common EMF bath under the Born-Markov approximation. We identify the point of departure of quantum dynamics under the Markovian approximation from the full non-Markovian dynamics and thereby demonstrate the limitations of the Born-Markov approximation. Finally, we discuss the domain of validity of the rotating wave approximation in describing these systems. In Appendix C we sketch an alternative derivation through the Feynman-Vernon influence functional technique, in which Grassmann variables are employed for the study of the atomic degrees of freedom.
\section{Two-Atoms interacting via a common Electromagnetic Field}
\subsection{The Hamiltonian}
We consider two 2-level atoms (2LA), acting as two qubits, labeled 1 and 2, and an electromagnetic field described by the free Hamiltonian $H_0$ \begin{eqnarray} \hat{H}_0 = \hbar \sum_{\bf k} \omega_{\bf k}\hat{b}^{\dagger}_{\bf k} \hat{b}_{\bf k} +\hbar\omega_o \hat{S}_+^{(1)} \hat{S}_-^{(1)} +\hbar\omega_o \hat{S}_+^{(2)} \hat{S}_-^{(2)} \end{eqnarray} where $\omega_{\bf k}$ is the frequency of the ${\bf k}^{\mbox{th}}$ electromagnetic field mode and $\omega_o$ the atomic frequency between the two levels of the atom, assumed to be the same for the two atoms. The electromagnetic field creation~(annihilation) operator is $b_{\bf k}^+$~($b_{\bf k}$), while $S_+^{(n)}$~($S_-^{(n)}$) are the spin raising~(lowering) operators for the $n^{\mbox{th}}$ atom. We will define the pointing vector from $1$ to $2$ as ${\bf r} ={\bf r}_2 -{\bf r}_1$ and we will assume without loss of generality that ${\bf r}_1 + {\bf r}_2 = 0$.
The two 2LAs do not interact with each other directly but only through the common electromagnetic field via the interaction Hamiltonian \begin{eqnarray} \hat{H}_I = \hbar\sum_{\bf k} g_{\bf k}\left( \hat{b}^{\dagger}_{\bf k} ( e^{-i{\bf k}\cdot{\bf r}/2} \hat{S}_{-}^{(1)} + e^{i{\bf k}\cdot{\bf r}/2} \hat{S}_{-}^{(2)} ) + b_{\bf k} ( e^{i{\bf k}\cdot{\bf r}/2} \hat{S}_{+}^{(1)} + e^{-i{\bf k}\cdot{\bf r}/2} \hat{S}_{+}^{(2)} ) \right), \label{Hint} \end{eqnarray} where $g_{\bf k } = \frac{\lambda}{\sqrt{\omega_{\bf k}}}$, $\lambda$ being the coupling constant. We have assumed that the dipole moments of the atoms are parallel. The total Hamiltonian of the atom-field system is \begin{equation} \hat{H} = \hat{H}_0 + \hat{H}_I. \end{equation}
\subsection{Perturbative expansion and resummation}
We assume that at $t = 0$ the state of the combined system of atoms+field is factorized and that the initial state of the EMF is the vacuum $|O \rangle$. For this reason we need to identify the action of the evolution operator $e^{-i\hat{H}t}$ on vectors of the form $|O \rangle \otimes | \psi \rangle$, where $|\psi \rangle$ is a vector on the Hilbert space of the two 2LA's.
For this purpose, we use the resolvent expansion of the Hamiltonian \begin{eqnarray} e^{-i\hat{H}t} = \int \frac{dE e^{-iE t}}{E - \hat{H} + i \eta} \end{eqnarray} and we expand
\begin{eqnarray}
(E - \hat{H})^{-1} = (E - \hat{H}_0)^{-1} + (E - \hat{H}_0)^{-1} \hat{H}_I (E - \hat{H}_0)^{-1} \nonumber \\ + (E - \hat{H}_0)^{-1} \hat{H}_I (E - \hat{H}_0)^{-1} \hat{H}_I (E - \hat{H}_0)^{-1} + \ldots \label{expansion} \end{eqnarray}
Of relevance for the computation of the reduced density matrix of the two qubits are matrix elements of the form $\langle z; i',j'|
(E - \hat{H})^{-1}| O; i,j \rangle$, where $z$ refers to a coherent state of the EM
field and $i, j = 0,1$, the value $i = 0$ corresponding to the
ground state of a single qubit and $i = 1$ to the excited state.
We compute the matrix elements above through the perturbation expansion (\ref{expansion}).
It turns out that we can effect a resummation of the perturbative series and thus obtain an exact expression for the matrix elements--see Appendix A for details of the resummation.
The non-vanishing matrix elements are the following \begin{eqnarray}
\langle z; 0,0| (E - \hat{H})^{-1}| O; 0, 1 \rangle &=& \sum_{\bf k} \frac{g_{\bf k} z^*_{\bf k} e^{ i \frac{{\bf k} \cdot {\bf r}}{2}}}{(E - \omega_o - \alpha(E) - \beta(E,r)e^{i {\bf k} \cdot {\bf r}})(E - \omega_{\bf k})} \\
\langle z; 0,1| (E - \hat{H})^{-1}| O; 0, 1 \rangle &=& \frac{1}{2}\left[ \frac{1}{E - \omega_o - \alpha(E) - \beta(E,r)} \right. \nonumber \\ &+& \left. \frac{1}{E - \Omega - \alpha(E) + \beta(E,r)} \right]
\\
\langle z; 1,0| (E - \hat{H})^{-1}| O; 0, 1 \rangle &=& \frac{1}{2}\left[ \frac{1}{E - \omega_o - \alpha(E) - \beta(E,r)} \right. \nonumber \\
&-& \left. \frac{1}{E - \omega_o - \alpha(E) + \beta(E,r)} \right]
\\
\langle z; 0,0| (E - \hat{H})^{-1}| O; 0, 0 \rangle &=& E^{-1} \\
\langle z; 1,1| (E - \hat{H})^{-1}| O; 1, 1 \rangle &=& \frac{1}{E - 2\omega_o - 2 \alpha (E - \omega_o)- f(E,r)} \\
\langle z; 0,0| (E - \hat{H})^{-1}| O; 1, 1 \rangle &=& \sum_{{\bf k k'}} \frac{\hat{H}_{{\bf k k'}} z^*_{{\bf k}} z^*_{{\bf k}}}{E - 2\omega_o - 2 \alpha (E - \omega_o)- f(E,r)}\\
\left( \begin{array}{cc} \langle z; 0,1| (E - \hat{H})^{-1}| O; 1, 1 \rangle \\
\langle z; 1, 0| (E - \hat{H})^{-1}| O; 1, 1 \rangle \end{array} \right) &=& \sum_{\bf k k'} \frac{g_{\bf k'} z^*_{\bf k}}{(E - 2 \omega_o)(E - \omega_o - \omega_{\bf k'})} \left( \begin{array}{cc} e^{-i\frac{{\bf k} \cdot {\bf r}}{2}} \\ e^{i\frac{{\bf k} \cdot {\bf r}}{2}} \end{array} \right) (1 -L)^{-1}_{\bf kk'} \end{eqnarray}
In the equations above the functions $\alpha(E), \beta(E,r)$ are \begin{eqnarray} \alpha(E) :&=& \sum_{\bf k} \frac{g_{\bf k}^2}{E - \omega_{\bf k}} \\ \beta(E,r) :&=& \sum_{\bf k} \frac{g_{\bf k}^2}{E - \omega_{\bf k}} e^{i{\bf k} \cdot {\bf r} }. \end{eqnarray}
The definitions of the kernel $H_{\bf k k'}$ and of the function $f$ involve complicated expressions. However, the term involving $H_{\bf kk'}$ does not contribute to the evolution of the reduced density matrix, while the function $f(E,r)$ is of order $\lambda^4$ and it can be ignored in the approximation we effect in Sec. II.E. Thus for the purpose of this investigation, the explicit definitions of $H$ and $f$ are not needed, and hence not given here.
Finally, the matrix $L$ is defined as \begin{eqnarray} L := \left( \begin{array}{cc} \Xi & \Theta \\ \bar{\Theta} & \bar{\Xi} \end{array} \right), \end{eqnarray} where $\Xi$ and $\Theta$ are matrices on the space of momenta \begin{eqnarray} \Xi_{\bf k k'} = \frac{1}{E - \Omega - \omega_{\bf k}} \left( \alpha(E - \omega_{\bf k}) \delta_{\bf kk'} + g_{\bf k} g_{\bf k'} (\frac{1}{E-2 \Omega} + \frac{e^{i ({\bf k} - {\bf k'}) \cdot {\bf r}}}{E - \omega_{\bf k} - \omega_{\bf k'}} )\right) \label{Xi} \\ \Theta_{\bf k k'} = \frac{1}{E - \Omega - \omega_{\bf k}} \left( \beta(E-\omega_{\bf k}, r) \delta_{\bf kk'} + g_{\bf k} g_{\bf k'} (\frac{1}{E-2 \Omega} + \frac{ 1}{E - \omega_{\bf k} - \omega_{\bf k'}} )\right), \label{Theta} \end{eqnarray} and the overbar denotes complex conjugation.
\subsection{The matrix elements of the propagator}
The next step is to Fourier transform the matrix elements of the resolvent in order to obtain the matrix elements of the evolution operator. Explicitly, \begin{eqnarray}
\langle z; 0, 0| e^{-i\hat{H}t}| O; 0, 1 \rangle &=& \sum_{\bf k} e^{i {\bf k}\cdot {\bf r}/2} z^*_{\bf k} s_{\bf k}(t) \label{00}\\
\langle z; 0, 1| e^{-i\hat{H}t}| O; 0, 1 \rangle &=& \int \frac{dE e^{-iEt}}{2} \left[ \frac{1}{E - \omega_o - \alpha(E) - \beta(E,r)} \right. \nonumber \\ &+& \left.\frac{1}{E - \omega_o - \alpha(E) + \beta(E,r)} \right] = : v_+(t) \label{iv+}\\
\langle z; 1, 1| e^{-i\hat{H}t}| O; 0, 1 \rangle &=& \int \frac{dE e^{-iEt}}{2} \left[ \frac{1}{E - \omega_o - \alpha(E) - \beta(E,r)} \right. \nonumber \\ &-& \left. \frac{1}{E - \omega_o - \alpha(E) + \beta(E,r)} \right] =: v_-(t) \label{iv-}\\
\langle z; 0, 0| e^{-i\hat{H}t}| O; 0, 0 \rangle &=& 1 \\
\langle z; 1, 1| e^{-i\hat{H}t}| O; 1, 1 \rangle &=& \int \frac{dE
e^{-iEt}}{E - 2 \omega_o - 2 \alpha(E - \omega_o) -f(E, r)} =: u(t) \label{iu}\\
\langle z; 0, 0| e^{-i\hat{H}t}| O; 1, 1 \rangle &=& \int dE e^{-iEt} \sum_{\bf k k'}
\frac{ \hat{H}_{\bf k k'} z^*_{\bf k} z^*_{\bf k'} }{ E - 2 \omega_o - 2 \alpha(E - \omega_o)
-f(E, r)} \\
\left( \begin{array}{cc} \langle z; 0, 1| e^{-i\hat{H}t}| O; 1, 1 \rangle
\\ \langle z; 1, 0| e^{-i\hat{H}t}| O; 1, 1 \rangle \end{array} \right) &=& \sum_{\bf k} z^*_{\bf k} \left( \begin{array}{cc} e^{-i\frac{{\bf k} \cdot {\bf r}}{2}} \nu_{\bf k}(t) \\ e^{i\frac{{\bf k} \cdot {\bf r}}{2}} \nu'_{\bf k}(t) \label{11} \end{array} \right), \end{eqnarray} where we defined the functions $s_{\bf k}(t), \nu_{\bf k}(t), \nu_{\bf k}'(t)$ as \begin{eqnarray} s_{\bf k}(t) &=& \int \frac{dE e^{-iEt}}{(E-\omega_o - \alpha(E) - \beta(E,r) e^{i {\bf k} \cdot {\bf r}})(E - \omega_{\bf k})} \label{sk}\\ \left( \begin{array}{cc} \nu_{\bf k}(t) \\ \nu'_{\bf k}(t) \end{array} \right) &=& \int \frac{dE e^{-iEt}}{E - 2 \omega_o} \sum_{\bf k'} (1 - L)_{\bf kk'} \left( \begin{array}{cc} \frac{g_{\bf k'}}{E - \omega_o - \omega_{\bf k'}} \\ \frac{g_{\bf k'}}{E - \omega_o - \omega_{\bf k'}} \end{array} \right) \label{nu} \end{eqnarray}
\subsection{The reduced density matrix}
We next compute the elements of the reduced density matrix for the qubit system by integrating out the EM field degrees of freedom \begin{eqnarray} \rho^{ij}_{i'j'}(t) = \sum_{i_0,j_0,i'_0,j'_0}
\rho^{i_0,j_0}_{i'_0,j'_0} (0) \int [dz] [dz^*] \langle O; i'_0j'_0|e^{i\hat{H}t}| z; i', j' \rangle \langle z,i,j|e^{-i\hat{H}t} |O, i_0,j_0 \rangle, \label{rdm} \end{eqnarray} where $[dz]$ is the standard Gaussian integration measure for the coherent states of the EM field.
Substituting Eqs. (\ref{00}-\ref{11}) into (\ref{rdm}) we obtain through a tedious but straightforward calculation the elements of the reduced density matrix
\begin{eqnarray} \rho^{I}_{I}(t) &=& \rho^{I}_{I}(0) |u|^2(t) \label{1111} \\ \rho^{11}_{01}(t) &=& \rho^{11}_{01}(0) u(t) v^*_+(t) + \rho^{11}_{10}(0) u(t) v^*_-(t) \\ \rho^{11}_{10}(t) &=& \rho^{11}_{10}(0) u(t) v^*_+(t) + \rho^{11}_{01}(0) u(t) v^*_-(t) \\ \rho^{I}_{00}(t) &=& \rho^{11}_{00} u(t) \\
\rho^{01}_{00}(t) &=& \rho^{01}_{00}(0) v_+(t) + \rho^{10}_{00}(0) v_-(t) + \rho^{11}_{01}(0) \mu_1(t) + \rho^{11}_{10}(0) \mu_2(t) \\ \rho^{10}_{00}(t) &=& \rho^{10}_{00}(0) v_+(t) + \rho^{10}_{00}(0) v_-(t) + \rho^{11}_{01}(0) \mu^*_2(t) + \rho^{11}_{10}(0) \mu^*_1(t) \\
\rho^{01}_{01}(t) &=& \rho^{01}_{01}(0) |v_+|^2(t) +
\rho^{01}_{10}(0) v_+(t) v^*_-(t) + \rho^{10}_{10}(0) |v_-|^2(t) \nonumber \\ &+& \rho^{10}_{01}(0) v_-(t) v^*_+(t) + \rho^{11}_{11}(0) \kappa_1(t) \\
\rho^{01}_{10}(t) &=& \rho^{01}_{10}(0) |v_+|^2(t) +
\rho^{10}_{01}(0) |v_-|^2(t) + \rho^{01}_{01}(0) v_+(t) v_-^*(t)\nonumber \\ & +& \rho^{10}_{10}(0) v_-(t) v^*_+(t) + \rho^{11}_{11}(0) \kappa_2(t) \\ \rho^{00}_{00}(t) &=& 1 - \rho^{11}_{11}(t) - \rho^{01}_{ 01}(t) - \rho^{10}_{10}(t) \label{0000} \end{eqnarray} where \begin{eqnarray} \mu_1(t) &=& \sum_{\bf k} g_{\bf k} \nu_{\bf k}(t) s^*_{\bf k}(t) \label{defm1} \\ \mu_2(t) &=& \sum_{\bf k} g_{\bf k} \nu_{\bf k}(t) s^*_{\bf k}(t) e^{-i {\bf k} \cdot {\bf r}} \label{defm2}\\
\kappa_1(t) &=& \sum_{\bf k} |\nu_{\bf k}|^2(t) \label{defk1}\\ \kappa_2(t) &=& \sum_{\bf k} \nu_{\bf k}(t) \nu'^*_{\bf k}(t) e^{- i {\bf k} \cdot {\bf r}}, \label{defk2} \end{eqnarray} and the functions $u(t), v_{\pm}(t)$ were defined in Eqs. (\ref{iu}), (\ref{iv+}) and (\ref{iv-}).
\subsection{Explicit forms for the evolution functions}
Eqs. (\ref{1111}-\ref{0000}) provide an {\em exact} expression for the evolution of the reduced density matrix for the system of two qubits interacting with the EM field in the vacuum state. The evolution is determined by seven functions of time $u, v_{\pm}, \kappa_{1,2}, \mu_{1,2}$, for which we have provided the full definitions. To study the details of the qubits' evolution we must obtain explicit forms for the functions above. For analytic expressions, we recourse to an approximation: Assuming weak coupling ($\lambda^2 << 1$), we ignore the contribution of all processes that involve the exchange of two or more photons between the two qubits.
\subsubsection{The functions $u, v_{\pm}$}
With the approximation above, the contribution of the function $f$ drops out from the definition of $u$. Thus we obtain \begin{eqnarray}
u(t) &=& \int \frac{dE
e^{-iEt}}{E - 2 \omega_o - 2 \alpha(E - \omega_o) } \\
v_{\pm} &=& \int dE e^{-iEt} \frac{1}{2}\left[ \frac{1}{E - \omega_o - \alpha(E) - \beta(E,r)} \pm \frac{1}{E - \omega_o - \alpha(E) + \beta(E,r)} \right]. \end{eqnarray}
We evaluate these expressions using an additional approximation. In performing the Fourier transform, we only keep the contribution of the poles in the integral and ignore that coming from a branch-cut that appears due to the presence of a logarithm in the exact expression of $\alpha(E)$--see Ref. \cite{AH} for details. We then obtain \begin{eqnarray} u(t) &=& e^{ - 2 i \omega_o t - 2 \Gamma_0 t} \label{u} \\ v_{\pm}(t) &=& \frac{e^{- i \omega_o t - \Gamma_0 t}}{2} \left( e^{ - i \sigma t - \Gamma_r t} \pm e^{i \sigma t + \Gamma_r t} \right). \label{v+-} \end{eqnarray}
In the equations above, we effected a renormalization of the frequency $\omega_o$ by a constant divergent term-- see \cite{AH}. The parameters $\Gamma_0, \Gamma_r$ and $\sigma(r) $ are defined as \begin{eqnarray} \Gamma_0 := - Im \, \alpha(\omega_o) \\ -\sigma(r) + i \Gamma_r := \beta(\omega_o, r), \end{eqnarray} and they read explicitly \begin{eqnarray} \Gamma_0 &=& \frac{\lambda^2 \omega_o}{2 \pi} \\ \Gamma_r &=& \frac{\lambda^2 \sin \omega_o r }{2 \pi r} \\ \sigma(r) &=& \frac{\lambda^2}{2 \pi^2 r} \left[ - \cos \omega_0 r [\frac{\pi}{2} - Si(\omega_0 r)] \right. \nonumber \\
&+& \left. \sin \omega_0r [\log(e^{\gamma} \omega_0 r) + \int_0^{\omega_0r} dz \, \frac{1 - \cos z}{z}] \right], \end{eqnarray} where $Si$ is the sine-integral function.
The term $\sigma(r)$ is a frequency shift caused by the vacuum fluctuations. It breaks the degeneracy of the two-qubit system and generates an effective dipole coupling between the qubits. At the limit $r \rightarrow 0$, this term becomes infinite. One should recall however that the physical range of $r$ is always larger than $a_B$, the Bohr radius of the atoms. As $r \rightarrow \infty$, $\sigma(r) \rightarrow 0$.
The constant $\Gamma_0$ corresponds to the rate of emission from individual qubits. It coincides with the rate of emission obtained from the consideration of a single qubit interacting with the electromagnetic field. The function $\Gamma_r$ is specific to the two-qubit system. It arises from Feynman diagrams that involve an exchange of photons between the qubits. Heuristically, it expresses the number of virtual photons
per unit time exchanged between the qubits. As such, $\Gamma_r^{-1}$ is
the characteristic time-scale for the exchange of information
between the qubits. As $r \rightarrow 0$, $\Gamma_r \rightarrow
\Gamma_0$ and as $r \rightarrow \infty$, $\Gamma_r \rightarrow 0$. Note
that the ratio $\Gamma/\Gamma_0 =
\frac{\sin \omega_o r}{\omega_o r}$, while smaller than unity, is of
the order of unity as long as $r$ is not much larger than
$\omega_o^{-1}$.
It is interesting to note that $\Gamma_r = 0$ for $r = n \pi \omega_0^{-1}$, where $n$ an integer. This is a resonant behaviour, similar to that of a classical oscillating dipole when $r = n \lambda/2$, where $\lambda$ is the oscillation wavelength.
\subsubsection{The functions $\kappa_{1,2}(t)$} We first compute the functions $\nu_{\bf k}, \nu'_{\bf k}$ of Eq. (\ref{nu}) keeping terms up to second loop order \begin{eqnarray} \left( \begin{array}{cc} \nu_{\bf k}(t) \\ \nu'_{\bf k}(t) \end{array} \right) &=& \int \frac{dE e^{-iEt}}{E - 2 \omega_o} \sum_{\bf k'} \frac{g_{\bf k'}} {E - \omega_o - \omega_{\bf k'}}\left( \begin{array}{cc} \delta_{\bf kk'} + \Xi_{\bf kk'} + \Theta_{\bf kk'} \\\delta_{\bf kk'} +
\bar{\Xi}_{\bf kk'} + \bar{\Theta}_{\bf kk'}\\ \end{array} \right), \end{eqnarray} where $\Xi$ and $\Theta$ are given by Eqs. (\ref{Xi}) and (\ref{Theta}).
The summation over ${\bf k'}$ yields within an order of $\lambda^5$ \begin{eqnarray} \left( \begin{array}{cc} \nu_{\bf k}(t) \\ \nu'_{\bf k}(t) \end{array} \right) &=& \int \frac{dE e^{-iEt}}{E - 2 \omega_o} \frac{g_{\bf k}}{E - \omega_o - \omega_{\bf k}} \left[ 1 + \frac{ \alpha(E - \omega_{\bf k}) + \beta(E - \omega_{\bf k})}{E - \omega_o - \omega_{\bf k}}\right] \left[ 1 + 2 \frac{\alpha(E - \omega_o)}{E - 2 \omega_o} \right]
\nonumber \\ &\times& \left( \begin{array}{cc} 1 + \sum_{\bf k'} \frac{g_{\bf k'}^2 (1 + e^{i ({\bf k} - {\bf k'}) \cdot {\bf r}})} {(E - \omega_{\bf k} - \omega_{\bf k'}) (E - \omega_o - \omega_{\bf k'})} \\ 1 + \sum_{\bf k'} \frac{g_{\bf k'}^2 (1 + e^{-i ({\bf k} - {\bf k'}) \cdot {\bf r}})} {(E - \omega_{\bf k} - \omega_{\bf k'}) (E - \omega_o - \omega_{\bf k'})} \end{array} \right). \label{nuF} \end{eqnarray}
The terms in brackets in the first line of the equation above can be absorbed in the leading-order denominators--see Eq. (\ref{nnn}). The term in the second line, however, only gives rise to a (time-independent) multiplicative term of the form $1 + O(\lambda^2)$. Hence, if we keep the leading order terms in the expression of $\nu_{\bf k}$, we may ignore this term. Then $\nu_{\bf k} = \nu'_{\bf k}$, and within an error of order $\lambda^5$ \begin{eqnarray} \nu_{\bf k}(t) = g_{\bf k} \int \frac{dE e^{-iEt}}{[E - 2 \omega_o - 2 \alpha (E - \omega_o)][E - \omega_o - \omega_{\bf k} - \alpha(E - \omega_{\bf k}) - \beta(E - \omega_{\bf k}, r))]}. \label{nnn} \end{eqnarray}
Using the same approximation as in Sec. II.E.1 for the Fourier transform, we obtain \begin{eqnarray} \nu_{\bf k}(t) = g_{\bf k} \frac{e^{ - i \Omega t - \Gamma_0 t}}{\omega - \omega_{\bf k} - \sigma -i \Gamma_0 + i \Gamma_r} \left( e^{- i \omega_0 t - \Gamma_0t} - e^{ -i \omega_{\bf k} t - i \sigma t - \Gamma_r t} \right) \end{eqnarray}
We then substitute the expression above for $\nu_{\bf k}$ into Eqs. (\ref{defk1}) and (\ref{defk2}) to get \begin{eqnarray} \kappa_1(t) = \frac{\lambda^2}{2 \pi^2} e^{-2 \Gamma_0t} \int_0^{\infty} k dk \, \frac{e^{-2 \Gamma_0 t} + e^{ - 2\Gamma_r t} - 2 e^{ - (\Gamma_0 + \Gamma_r) t} \cos [ (\omega_o - k - \sigma)t]}{(k - \omega_o + \sigma)^2 + (\Gamma_0 - \Gamma_r)^2} \\ \kappa_2(t) = \frac{\lambda^2}{2 \pi^2 r} e^{-2 \Gamma_0t} \int_0^{\infty} dk \, \frac{e^{-2 \Gamma_0 t} + e^{ - 2\Gamma_r t} - 2 e^{ - (\Gamma_0 + \Gamma_r) t} \cos [ (\omega_o - k - \sigma)t]}{(k - \omega_o + \sigma)^2 + (\Gamma_0 - \Gamma_r)^2} \sin kr \end{eqnarray}
For $\omega_o t >> 1$, it is a reasonable approximation to substitute the Lorentzian with a delta function. Hence, \begin{eqnarray} \kappa_1(t) = \Gamma_0 \kappa(t) \label{kappa1} \\ \kappa_2(t) = \Gamma_r \kappa(t) \label{kappa2} \end{eqnarray} where \begin{eqnarray} \kappa(t) \simeq \frac{1}{\Gamma_0 - \Gamma_r} e^{-2 \Gamma_0t} (e^{-\Gamma_0 t} - e^{- \Gamma_rt})^2 \label{kappa}. \end{eqnarray}
\subsubsection{The functions $\mu_{1,2}$}
We first compute the functions $s_{\bf k}(t)$ of Eq. (\ref{sk}) \begin{eqnarray} s_{\bf k}(t) = \frac{ e^{-i \omega_o t - \Gamma_0 t -(\Gamma_r + i \sigma) e^{i {\bf k} \cdot{\bf r}} t} - e^{- i \omega_{\bf k}t}}{\omega_o - \omega_{\bf k} - i \Gamma_0 + (\sigma - i \Gamma_r) e^{i {\bf k} \cdot{\bf r}}} \end{eqnarray} Substituting into Eqs. (\ref{defm1}) and (\ref{defm2}) we obtain \begin{eqnarray} \mu_1(t) = \frac{\lambda^2}{4 \pi^2} \int_{-1}^{1} d \xi \; \int k dk \; e^{ - i \omega_o t - \Gamma_0 t}\frac{e^{- i \omega_0 t - \Gamma_0t} - e^{ -i \omega_{\bf k} t - i \sigma t - \Gamma_r t}}{\Omega - \omega_{\bf k} - \sigma -i \Gamma_0 + i \Gamma_r} \nonumber\\ \times \frac{ e^{i \omega_o t - \Gamma_0 t -(\Gamma_r - i \sigma) e^{-i {\bf k} \cdot{\bf r}} t} - e^{ i \omega_{\bf k}t}}{\omega_o - \omega_{\bf k} + i \Gamma_0 + (\sigma + i \Gamma_r) e^{-i {\bf k} \cdot{\bf r}}}\\ \mu_2(t) = \frac{\lambda^2}{4 \pi^2} \int_{-1}^{1} d \xi \; \int k dk e^{-i {\bf k} \cdot {\bf r}} \; e^{ - i \omega_o t - \Gamma_0 t}\frac{e^{- i \omega_0 t - \Gamma_0t} - e^{ -i \omega_{\bf k} t - i \sigma t - \Gamma_r t}}{\omega_o - \omega_{\bf k} - \sigma -i \Gamma_0 + i \Gamma_r} \nonumber\\ \times \frac{ e^{i \omega_o t - \Gamma_0 t -(\Gamma_r - i \sigma) e^{-i {\bf k} \cdot{\bf r}} t} - e^{ i \omega_{\bf k}t}}{\omega_o - \omega_{\bf k} + i \Gamma_0 + (\sigma + i \Gamma_r) e^{-i {\bf k} \cdot{\bf r}}}. \end{eqnarray} An approximate evaluation of the $\xi$-integral followed by the further approximation $\frac{1}{x+ i \epsilon} \simeq i \pi \delta(x)$, gives an estimation of the leading order contribution \begin{eqnarray} \mu_1(t) = \Gamma_0 [\mu(t) + i \nu(t)] \label{mu1}\\ \mu_2(t) = \Gamma_r [\mu(t) + i \nu(t)] \label{mu2} \end{eqnarray} where $\mu + i \nu$ is the complex-valued function \begin{eqnarray} \mu(t) + i \nu(t)= \simeq \frac{1 }{\Gamma_0 + \frac{ 2 \sin \omega_o r}{\omega_o r} \Gamma_r - i \sigma ( 1 + 2 \frac{\sin \omega_o r}{\omega_o r})} e^{- i \omega_o t - \Gamma_0 t}\nonumber \\ \times ( e^{ - \Gamma_0 t} - e^{- \Gamma_rt} )( e^{ - \Gamma_0 t - 2 \frac{\sin \omega_o r}{\omega_o r} [\Gamma_r - i \sigma] t} - e^{ i \sigma t} ). \label{munu} \end{eqnarray}
\subsection{The master equation}
Given the explicit form of the functions computed in Sec. II.E, we write the evolution equations in the following form \begin{eqnarray} \rho^{I}_{I} (t) &=& e^{-4 \Gamma_0 t} \rho^{I}_{I}(0) \label{dmp1}\\ \rho^{I}_{O}(t) &=& e^{ - 2 i \omega_o t - 2 \Gamma_0 t} \rho^{I}_{O}(0) \\
\rho^{I}_-(t) &=& e^{-i \omega_o t - 3 \Gamma_0 t + i \sigma t + \Gamma_r t} \rho^{I}_-(0) \\ \rho^{I}_+ (t) &=& e^{-i \omega_o t - 3 \Gamma_0 t - i \sigma t -\Gamma_r t} \rho^{I}_+(0) \\ \rho^-_O(t) &=& e^{-i \omega_ot -\Gamma_0t + i \sigma + \Gamma_r t} \rho^-_O(0) + i (\Gamma_0 + \Gamma_r) \nu(t) \rho^{I}_+(0) + (\Gamma_0 - \Gamma_r) \mu(t) \rho^{I}_-(0) \\ \rho^+_{O}(t) &=& e^{-i \omega_ot -\Gamma_0t - i \sigma - \Gamma_r t} \rho^+_{O}(0) + (\Gamma_0 + \Gamma_r) \mu(t) \rho^{I}_+(0) + i (\Gamma_0 - \Gamma_r) \nu(t) \rho^{I}_-(0) \\ \rho^+_+(t) &=& e^{-2 \Gamma_0 t - 2 \Gamma_r t} \rho^+_+(0) + (\Gamma_0 + \Gamma_r) \kappa(t) \rho^{I}_{I}(0) \\ \rho^-_-(t) &=& e^{-2 \Gamma_0 t - 2 \Gamma_r t} \rho^-_-(0) + (\Gamma_0 - \Gamma_r) \kappa(t) \rho^{I}_{I}(0) \label{rss}\\ \rho^+_-(t) &=& e^{2 i \sigma t - 2 \Gamma_0 t} \rho^+_-(0). \label{dmpf} \end{eqnarray}
In the equations above we wrote the density matrix in a basis defined by $|I \rangle, |O\rangle, |+\rangle, |-\rangle$, where
\begin{eqnarray} |+ \rangle = \frac{1}{\sqrt{2}} \left( |01 \rangle +
|10 \rangle \right) \\
|- \rangle = \frac{1}{\sqrt{2}} \left( |01 \rangle - |10 \rangle \right). \end{eqnarray}
We note that in this approximation, the diagonal elements of the density matrix propagator (i.e. the ones that map $\rho^a_b(0)$ to $\rho^a_b(t)$ where $a,b \in \{O, +, -, I\}$) decay exponentially (except for the $\rho^{O}_{O}$ which is functionally dependent on the other matrix elements due to normalization). We shall see in Sec. V that this behavior is in accordance with the Born-Markov approximation. The situation is different for the off-diagonal elements of the density matrix propagator, i.e. the ones that map $\rho^a_b(0)$ to $\rho^{a'}_{b'}(t)$ for $a \neq a'$ and $b \neq b'$. They are given by more complex functions of time and they differ from the ones predicted by the Born-Markov approximation.
As we have the solution $\rho(t) = M_t[\rho(0)]$, where $M_t$ is the density matrix propagator defined by Eqs. (\ref{dmp1}-\ref{dmpf}), we can identify the master equation through the relation $\dot{\rho}(t) = \dot{M}_t[ M^{-1}_t[\rho(t)]]$. We obtain \begin{eqnarray} \dot{\rho}^{I}_{I} &=& - 4 \Gamma_0 \rho^I_I \label{ME1}\\ \dot{\rho}^{I}_- &=& ( -i \omega_o - 3 \Gamma_0 + i \sigma + \Gamma_r) \rho^{I}_- \\ \dot{\rho}^{I}_+ &=& ( -i \omega_o - 3 \Gamma_0 - i \sigma - \Gamma_r) \rho^{I}_+ \\ \dot{\rho}^{I}_{O} &=& (-2 i \omega_o - 2 \Gamma_0) \rho^{I}_{O} \\ \dot{\rho}^-_{O} &=& = (-i \omega_o - \Gamma_0 +i \sigma + \Gamma_r) \rho^-_{O} + (\Gamma_0 + \Gamma_r) \alpha_1(t) \rho^{I}_+ + (\Gamma_0 - \Gamma_r) \alpha_2(t) \rho^{I}_- \\ \dot{\rho}^+_{O} &=& = (-i \omega_o - \Gamma_0 +i \sigma + \Gamma_r) \rho^+_{O} + (\Gamma_0 + \Gamma_r) \alpha_3(t) \rho^{I}_+ + (\Gamma_0 - \Gamma_r) \alpha_4(t) \rho^{I}_- \\ \dot{\rho}^+_{+} &=& -2 (\Gamma_0 + \Gamma_r) \rho^+_+ + (\Gamma_0 + \Gamma_r) \alpha_5(t) \rho^{I}_{I} \\ \dot{\rho}^-_- &=& -2 (\Gamma_0 - \Gamma_r) \rho^+_+ + (\Gamma_0 - \Gamma_r) \alpha_6(t) \rho^{I}_{I} \\ \dot{\rho}^+_- &=& 2(i\sigma - \Gamma_0) \rho^+_-. \label{MEf} \end{eqnarray} Explicit expressions for the functions of time $\alpha_i(t), i = 1, \ldots, 6$ appearing in Eqs. (\ref{ME1}--\ref{MEf}) are given in the Appendix B.
We see that the evolution equation for the reduced density matrix of the two qubits, while it is local-in-time, it does not have constant-in-time coefficients. Hence, it does not correspond to a Markov master equation of the Lindblad type. Again, we note that the non-Markovian behavior is solely found in the off-diagonal terms of the evolution law and that the diagonal ones involve constant coefficients.
To facilitate comparison with the expressions obtained from the Born-Markov approximations we cast equations (\ref{ME1}--\ref{MEf}) into an operator form. \begin{eqnarray} \dot{\hat{\rho}} &=& - i [\hat{H}_0 + \hat{H}_i, \hat{\rho}] + \sum_{i,j=1}^2 \Gamma_{ij} (\hat{S}_+^{(i)} \hat{S}_-^{(j)} \hat{\rho} + \hat{\rho} \hat{S}_+^{(i)} \hat{S}_-^{(j)} - 2 \hat{S}_-^{(i)} \hat{\rho} \hat{S}_+^{(j)}) \nonumber \\ &+& (\Gamma_0 + \Gamma_r) {\bf F_t}[\hat{\rho}] + (\Gamma_0 - \Gamma_r) {\bf G_t}[\hat{\rho}]. \label{Lindbladlike} \end{eqnarray}
The first term on the right-hand-side of Eq. (\ref{Lindbladlike}) corresponds to the usual Hamiltonian evolution: the total Hamiltonian is a sum of the free Hamiltonian $\hat{H}_0$ and of a dipole interaction Hamiltonian \begin{eqnarray} \hat{H}_i = - \sigma (\hat{S}^- \otimes \hat{S}_+ + \hat{S}^+ \otimes \hat{S}^-). \end{eqnarray}
The second term in the right-hand-side of Eq. (\ref{Lindbladlike}) is the usual Lindblad term (see e.g., \cite{FicTan}), where we defined $\Gamma_{11} = \Gamma_{22} = \Gamma_0$ and $\Gamma_{12} = \Gamma_{21} = \Gamma_r$.
The last two terms contain effects that pertain to the off-diagonal terms of the reduced density matrix propagator and they are non-Markovian: ${\bf F_t}$ and ${\bf G_t}$ are trace-preserving linear operators on the space of density matrices and their explicit form is given in Appendix B. Assuming that the Markovian regime corresponds to the vanishing of ${\bf F_t}$ and ${\bf G_t}$, we find from Eqs. (\ref{FF}--\ref{GG}) that in this regime the functions $\alpha_i(t)$ in Eqs. (\ref{ME1}--\ref{MEf}) should reduce to the following constants \begin{eqnarray} \alpha_1 = \alpha_4 = 0; \hspace{2cm} \alpha_3 = \alpha_5 = \alpha_6 = 2; \hspace{2cm} \alpha_2 = -2. \label{coefficients}. \end{eqnarray}
In Sec. V, we shall discuss in more detail the physical origin of non-Markovian behavior in this two-qubit system.
\section{Disentanglement of two qubits}
In this section, we employ the results obtained above to study the evolution of the two qubits initially in an entangled state. We shall focus on the process of disentanglement induced by their interaction with the field.
\subsection{Class A states: Initial superposition of $|00 \rangle$ and $|11\rangle$}
We first examine the class of initial states we call Class A of the following type \begin{eqnarray}
| \psi_o \rangle = \sqrt{1-p} |00\rangle + \sqrt{p} |11 \rangle, \label{gggg} \end{eqnarray}
where $0 \leq p \leq 1$. Recall our definition $|I \rangle = |11
\rangle$ and $|O \rangle = |00 \rangle$. From Eqs. (\ref{1111} - \ref{0000}) we obtain \begin{eqnarray}
\hat{\rho}(t) = p^2 e^{-4 \Gamma_0 t} |I \rangle \langle I| + e^{-2 \Gamma_0 t} \sqrt{p(1-p)} ( e^{2 i \omega_o t} |I\rangle \langle O| +
e^{-2 i \omega_0 t} \, |O \rangle \langle I|) \nonumber \\ + [ \kappa_1(t)
- \kappa_2(t)] |- \rangle \langle -| + [\kappa_1 (t) + \kappa_2(t)]
|+ \rangle \langle +| + [1 - p^2 e^{-4 \Gamma_0 t} - 2 \kappa_1(t)] |O
\rangle \langle O|, \end{eqnarray} where the functions $\kappa_1(t)$ and $\kappa_2(t)$ are given by Eqs. (\ref{kappa1}--\ref{kappa2}).
These results are quite different from those reported in Ref.
\cite{FicTan06}, which were obtained under the Born-Markov approximation. While the $|I \rangle \langle I|$ and $|I\rangle
\langle O|$ terms are essentially the same, the $|- \rangle \langle
-|$ and $|+ \rangle \langle +|$ ones are not, as they involve non-diagonal elements of the density matrix propagator. For comparison, we reproduce here the explicit form of these matrix elements in our calculation \begin{eqnarray} \rho^+_+ &=& p \frac{\Gamma_0 + \Gamma_r}{\Gamma_0 - \Gamma_r} e^{-2 \Gamma_0t} (e^{-\Gamma_0 t} - e^{- \Gamma_rt})^2 \\ \rho^-_- &=& p e^{-2 \Gamma_0t} (e^{-\Gamma_0 t} - e^{- \Gamma_rt})^2, \end{eqnarray} and in that of Ref. \cite{FicTan06}
(translated into our notation): \begin{eqnarray} \rho^+_+ (t) &=& p \frac{\Gamma_0 + \Gamma_r}{\Gamma_0 - \Gamma_r} e^{-2 \Gamma_0 t} ( e^{-2 \Gamma_rt} - e^{- 2 \Gamma_0 t}) \label{FTt} \\ \rho^-_-(t) &=& p \frac{\Gamma_0 - \Gamma_r}{\Gamma_0 + \Gamma_r} e^{-2 \Gamma_0 t} ( e^{2 \Gamma_rt} - e^{- 2 \Gamma_0 t}). \label{FTs} \end{eqnarray} For large values of $r$, $\Gamma_r << \Gamma_0$ and the expressions above coincide. However, for $\omega_0 r << 1$ their difference is substantial. In this regime, $\Gamma_0 \simeq \Gamma_r$ and at times
$\Gamma_0 t \sim 1$ we obtain $(\Gamma_0 - \Gamma_r)t <<1$. According to the Markovian results of \cite{FicTan06}, in this regime the $|+ \rangle
\langle +|$ term is of order $O(\lambda^0)$ and hence comparable in size to the other terms appearing in the evolution of the density matrix. However, according to our results, which are based on the full non-Markovian dynamics, the $|+ \rangle \langle +|$ term is of order $\frac{\Gamma_0 - \Gamma_r}{\Gamma_0}$ and hence much smaller.
In general, for $\omega_0 r << 1$, we find that the $|- \rangle \langle
-|$ and $|+ \rangle \langle +|$ terms contribute little to the evolution of the reduced density matrix and they can be ignored. Since these terms are responsible for the sudden death and subsequent revival of entanglement studied in \cite{FicTan06}, we conclude that these effects are absent in this regime. Indeed, this can be verified by the study of concurrence as a function of time as appearing in Figs. 1 and 2. For large inter-qubit separations and for specific initial states ($ p > 0.5$) there is sudden death of entanglement--but no subsequent revivals.
\begin{figure}
\caption{A plot of the concurrence as a function of time for initial Class A state (\ref{gggg}) with $p = 0.5$ for different values of $\omega_o r$.}
\label{ghzconc}
\end{figure}
\begin{figure}
\caption{A plot of the concurrence as a function of time for initial Class A state (\ref{gggg}) for different values of $p$ and fixed inter-qubit distance ($\omega_o r = 1$).}
\label{ghzconc2}
\end{figure}
\subsection{Initial antisymmetric $|- \rangle$ Bell state}
We next consider the case of an initial antisymmetric $|- \rangle$ state for the two qubits. We find \begin{eqnarray}
\hat{\rho}(t) = e^{ -2 [\Gamma_0 - \Gamma_r]t} |- \rangle \langle -| +
\left(1 -e^{ -2 [\Gamma_0 - \Gamma_r]t}\right)|O \rangle \langle O | \end{eqnarray}
We see that the decay rate is $\Gamma_0 - \Gamma_r$. The effect of photon exchange between the qubits essentially slows down the overall emission of photons from the two-qubit system (sub-radiant behavior). As the qubits are brought closer, the decay is slower. At the limit $r \rightarrow 0$, there is no decay.
The results agree qualitatively with those obtained in Ref.
\cite{FicTan} through the Born-Markov approximation. Fig. 3 shows the time evolution of concurrence for an initial antisymmetric $|- \rangle$ state and for different inter-qubit distances.
\begin{figure}
\caption{A plot of the concurrence as a function of time for the initial antisymmetric $|- \rangle$ Bell state and for different values of the inter-qubit distance $r$ (in units of $\omega_o^{-1}$). The decay of concurrence proceeds more slowly when the qubits are closer together.}
\label{singlet conc}
\end{figure}
\subsection{Initial symmetric $|+ \rangle$ Bell state}
For an initial symmetric $|+ \rangle$ Bell state the reduced density matrix of the two qubits is
\begin{eqnarray}
\hat{\rho}(t) = e^{ -2 [\Gamma_0 + \Gamma_r]t} |+ \rangle \langle +| +
\left(1 -e^{ -2 [\Gamma_0 + \Gamma_r]t}\right) |O \rangle \langle O|. \end{eqnarray}
Here we have a super-radiant decay with rate given by $\Gamma_0 +
\Gamma_r$. Note that for both the antisymmetric $|- \rangle $ and symmetric $|+ \rangle $ states, a resonance appears when $r = n \pi \omega_o^{-1}$, and the decay becomes radiant. Fig. 4 shows the behavior of concurrence, in qualitative agreement with the corresponding results of Ref. \cite{FicTan}.
\begin{figure}
\caption{Plot of the concurrence as a function of time for initial symmetric $|+ \rangle$ Bell state and for different values of the inter-qubit distance $r$ (in units of $\omega_o^{-1}$).}
\label{triplet conc}
\end{figure}
\section{Decoherence of qubits}
In this section we study the evolution of factorized initial states, in order to understand how the decoherence of one qubit is affected by the presence of another.
\subsection{One qubit in vacuum state}
We assume that the first qubit is prepared initially in a superposition of the $|0 \rangle$ and $|1 \rangle$ states, and that the second qubit lies on the ground state. Therefore, the initial state is of the form \begin{eqnarray}
\left(\sqrt{p} |1 \rangle + \sqrt{1-p} |0 \rangle \right) \otimes |0
\rangle = \sqrt{p} |10 \rangle + \sqrt{1-p} |00 \rangle. \end{eqnarray}
From Eqs. ( \ref{1111}--\ref{0000}), we obtain the density matrix of the combined qubit system \begin{eqnarray}
\hat{\rho}(t) = p \left( |v_+|^2 |10\rangle \langle 10| + |v_-|^2
|01 \rangle \langle 01 | + v_-^* v_+ |01 \rangle \langle 10 | +
v_- v_+^* |10 \rangle \langle 01| \right) \nonumber \\
+ \sqrt{p(1-p)} \left( v_+^* |10 \rangle \langle 00| + v_-^* |01
\rangle \langle 00| + v_+ |00 \rangle \langle 10| + v_- |00 \rangle
\langle 01| \right) \nonumber \\
+ \left(1 - p (|v_+|^2 + |v_-|^2) \right) |00 \rangle \langle 00|, \end{eqnarray} where the functions $v_{\pm}(t)$ are given by Eq. (\ref{v+-}).
The two qubits become entangled through their interaction via the EM field bath. To study the decoherence in the first qubit, we trace out the degrees of freedom of the second one, thus constructing the reduced density matrix $\hat{\tilde{\rho}}_1$ \begin{eqnarray}
\hat{\tilde{\rho}}_1(t) = p |v_+|^2 |1 \rangle \langle 1| +
\sqrt{p(1-p} \left( v_+ |0\rangle \langle 1| + v_+^* |1 \rangle 0|
\right) + \left( 1 - p|v_+|^2 \right) |0 \rangle \langle 0|. \end{eqnarray}
At the limit of large interqubit distances, $\Gamma_r = 0 = \sigma(r)$, whence $v_+ \simeq e^{- i \omega_o t - \Gamma_0 t}$, the off-diagonal elements decay within a characteristic time-scale of order $\Gamma_0^{-1}$. These results coincide with those for the single qubit case--see Refs. \cite{AH, SADH}. However, in the regime $\omega_o r << 1$, the results are substantially different. The entanglement with the second qubit leads to a departure from pure exponential decay. In this regime, $\Gamma_r \simeq \Gamma_0$. This implies for times longer than $\Gamma_0^{-1}$ a substantial fraction of the off-diagonal elements persists. This decays eventually to zero within a time-scale of order $[\Gamma_0 - \Gamma_r]^{-1} >> \Gamma_0^{-1}$. Hence, the qubit preserves its coherence longer. (At the limit $r \rightarrow 0$ there is no decoherence.)
The reduced density matrix of the second qubit is \begin{eqnarray}
\hat{\tilde{\rho}}_2(t) = p |v_-|^2 |1 \rangle \langle 1| +
\sqrt{p(1-p)} \left( v_- |0\rangle \langle 1| + v_-^* |1 \rangle 0|
\right) + \left( 1 - p |v_-|^2 \right) |0 \rangle \langle 0|. \end{eqnarray}
Note that at the limit of small inter-qubit distances the asymptotic behavior ( for $\Gamma_0 t >> 1$) of $\hat{\tilde{\rho}}_1(t)$ is identical to that of $\hat{\tilde{\rho}}_2(t)$. The second qubit (even though initially on its ground state) develops a persistent quantum coherence as a result of the interaction with the first one.
\subsection{One qubit in excited state}
We also consider the case of a factorized initial state, in which the second qubit is excited \begin{eqnarray}
\left(\sqrt{p} |1 \rangle + \sqrt{1-p} |0 \rangle \right) \otimes |1
\rangle = \sqrt{p} |I \rangle + \sqrt{1-p} |01 \rangle. \end{eqnarray} This system behaves differently from that of Sec. IV.A. The matrix elements of the reduced density matrix read \begin{eqnarray} \rho^{11}_{11} &=& p e^{ - 4 \Gamma_0 t} \\ \rho^{11}_{01} &=& \sqrt{p(1-p)} e^{ - 2 i \omega_o t - 2 \Gamma_0 t} v_-^* \\ \rho^{01}_{00} &=& \sqrt{p(1-p)} \mu_1(t) \\ \rho^{10}_{00} &=& \sqrt{p(1-p)} \mu_2^*(t)\\
\rho^{01}_{01} &=& (1-p) |v_+|^2 + p \kappa_1 \\ \rho^{01}_{10} &=& (1-p) v_+ v^*_- + p \kappa_2, \end{eqnarray}
where the functions $\mu_{1,2}$ are given by Eqs. (\ref{mu1}--\ref{munu}), the functions $\kappa_{1,2}$ by Eqs. (\ref{kappa1}) and (\ref{kappa2}) and the functions $v_{\pm}$ by Eq. (\ref{v+-}).
The reduced density matrix of the first qubit is \begin{eqnarray}
\hat{\tilde{\rho}}_1(t) = \left((1-p)|v_+|^2 + p \kappa_1 + p e^{- 4
\Gamma_0t} \right) |1 \rangle \langle 1| + \left(1 - (1-p)|v_+|^2 - p
\kappa_1 - p e^{- 4 \Gamma_0t} \right) |0 \rangle \langle 0| \nonumber\\ + \sqrt{p(1-p)} \left( (\mu_2^* + e^{ -2 i \omega_o t - 2 \Gamma_0 t} v_+^*)
|0 \rangle \langle 1| + (\mu_2 + e^{ 2 i \omega_o t - 2 \Gamma_0 t} v_+^*)
|1 \rangle \langle 0| \right). \end{eqnarray}
At times $t >> \Gamma_0^{-1}$, the decay of the off-diagonal elements at is dominated by the function $\mu_2$, which then reads \begin{eqnarray} \mu_2(t) \simeq \frac{\Gamma_r e^{- i (\omega_o - \sigma) t - \Gamma_0 t} }{\Gamma_0 + \frac{ 2 \sin \omega_o r}{\omega_o r} \Gamma_r - i \sigma ( 1 + 2 \frac{\sin \omega_o r}{\omega_o r})} ( e^{ - \Gamma_r t} - e^{- \Gamma_0t} ). \end{eqnarray}
In the limit of large inter-qubit distance $r$, the off-diagonal element falls like $r^{ - 2 \Gamma_0t}$, while in the small $r$ limit it falls like $e^{ - \Gamma_0 t}$. Comparing with the case of Sec. IV.A, we see that the initial excitation of the second qubit leads to a lesser degree of entanglement of the total system, as it cannot absorb any virtual photons emitted from the first one. For this reason, the asymptotic decoherence rate does not vary significantly with the interqubit distance.
\section{Discussion}
In this section, we first summarize our results for the evolution of various initial states, we then discuss the origin of the non-Markovian behavior and finally, a possible limitation of our results that is due to the restricted domain of validity of the Rotating-Wave approximation.
\subsection{Description of the results}
For an initial Class A state (\ref{gggg}), the $|I \rangle$
component decays to the vacuum, but it also evolves into a linear combination of antisymmetric $|- \rangle $ and symmetric $|+ \rangle $ Bell states. However, if the qubits are close together the evolution to Bell states is suppressed. This behavior is qualitatively different from that of Ref. \cite{FicTan06}, which was obtained through the Born-Markov approximation. The corresponding terms differ by many orders of magnitude at the physically relevant time-scales. As a consequence, we find that there is neither sudden death nor revival of entanglement in this regime.
In retrospect, this difference should not be considered surprising. The Born-Markov method involves two approximations: i) that the back-action of the field on the atoms is negligible and ii) that all memory effects in the system are insignificant. When the qubits are found within a distance much smaller than the one corresponding to their characteristic wavelength, it is not possible to ignore back-action. The virtual photons exchanged by the qubits (at a rate given by $\Gamma_r$) substantially alter the state of the electromagnetic field.
On the other hand, the effect of virtual photons exchange between qubits drops off quickly at large separations $r$ -- the two qubits decay almost independently one of the other. Hence, the Born-Markov approximation -- reliable for the case of two separate qubits each interacting with an individual field -- also gives reasonable results for the two qubits interacting with a common field. In this regime sudden death is possible, but not revival of entanglement. In this sense, our results effectively reduces to those of Ref. \cite{YuEbePRL}: when the distance between the qubits is much larger than any characteristic correlation length scale of the system it looks as though the two qubits are found in different reservoirs.
Therefore, our results for initial states of Bell type are qualitatively similar to those obtained in Ref. \cite{FicTan}
through the Born-Markov approximation. The symmetric $|+ \rangle $
state decays super-radiantly and the antisymmetric $|- \rangle $ one sub-radiantly. For small values of $\omega_o r$, the antisymmetric $|- \rangle $ state decays very slowly and entanglement is preserved.
Concerning decoherence, when the inter-qubit separation is large and the second qubit lies on its ground state, our two qubit calculation reproduces previous results on the single qubit case \cite{AH, SADH}. However, if the qubits are close together the coherence is preserved longer. These results seem to suggest that in a many-qubit system, the inter-qubit quantum coherence can be sustained for times larger than the decoherence time of the single qubit case. This may suggest some physical mechanisms to resist decoherence in multi-qubit realistic systems.
\subsection{The origin and significance of the non-Markovian behavior}
With the exact solutions we have obtained (under weak coupling but no Born or Markov approximation) for the two qubit - EMF system we want to elaborate on the origin and significance of the non-Markovian behavior started in the Introduction. In the evolution equations (\ref{dmp1}--\ref{dmpf}) we note that the diagonal terms of the reduced density matrix propagator all decay exponentially. Their decay rate is therefore constant. It is well known that this feature is a sign of Markovian behavior. In fact, it characterizes the domain of validity of Fermi's golden rule: one could obtain the decay rates by direct application of this rule. Hence, as far as this part of the evolution is concerned, our results are fully compatible with the Markovian predictions.
However, the behavior of the non-diagonal terms in the reduced density matrix propagator is different. A look at Eqs.
(\ref{dmp1}-\ref{dmpf}) will convince the reader that the only non-zero such terms are ones that describe the effect of successive decays, for example the $|11 \rangle $ state first decaying into $|-
\rangle$ and then $|- \rangle $ decaying into the ground state $|00
\rangle$. Hence, the $\rho^-_-(t)$ term consists of one component that contains the remaining of the $|- \rangle \langle -|$ part of the initial state and another component that incorporates the decay of the $|11 \rangle \langle 11|$ part of the initial state towards the state $|- \rangle$. In our calculation, the latter term is encoded into the functions $\kappa_{1,2}(t)$, which are obtained by squaring the amplitudes $\nu_{\bf k}(t)$ as in Eqs. (\ref{defk1}--\ref{defk2}). At second loop order, these amplitudes are obtained from the summation of two Feynman diagrams --see Eq.
(\ref{nuF}). The structure of the poles in Eq. (\ref{nuF}) reveals that the first Feynman diagram describes the decays of the $|11
\rangle$ state, while the second one corresponds to processes involving the $|01 \rangle$ and $|10 \rangle$ states. When we construct the evolution functions $\kappa_{1,2}(t)$, we obtain terms that are both diagonal and off-diagonal with respect to the two types of process. It is the presence of the off-diagonal terms that is primarily (but not solely) responsible for the deviation of our results from the Markovian prediction. In the Markov approximation, the corresponding term involves summation (subtraction) of probabilities rather than amplitudes.
To justify the last statement, we note from Eqs. (\ref{1111}--\ref{0000}) that the evolution decouples the diagonal from the off-diagonal elements of the density matrix (the Markovian equations also have this property). Hence, the probabilities $p_a(t) = \rho^a_a(t), a \in \{ I, +, -, O\}$ evolve autonomously. Time--homogeneity (i.e. Lindblad time evolution) implies that their evolution can be given by a transfer matrix \begin{eqnarray} \dot{p}_a(t) = \sum_b T_a^b p_b(t). \label{transf} \end{eqnarray} Here $T^a_b$ is the {\em constant} decay rate for the process $ b \rightarrow a$. Noting that Eqs. (\ref{1111}--\ref{0000}) contain no transitions between $+$ and $-$ and no transitions from $-$ to $I$, Eq. (\ref{transf}) yields \begin{eqnarray} \dot{p}_-(t) = - 2(\Gamma_0 - \Gamma_r) p_-(t) + w p_{I}(t), \label{ps} \end{eqnarray} where $w = T^-_{I}$. Since $p_{I}(t)$ is determined by Eq. (\ref{ME1}) as $p_{I}(0) e^{-4 \Gamma_0t}$, we obtain for the initial condition $p_-(0) = 0$, \begin{eqnarray} p_-(t) \sim p_{I}(0) (e^{-2 (\Gamma_0 - \Gamma_r)t} - e^{-4 \Gamma_0 t}), \label{ccc} \end{eqnarray} in full agreement with Eq. (\ref{FTs}) obtained from the Lindblad master equation.
The derivation of Eq. (\ref{ccc}) provides an example of a more general fact: the Markovian assumption forces the off-diagonal terms of the density matrix propagator (in this case the one mapping $\rho^{I}_{I}(0)$ to $\rho^-_-(t)$) to be subsumed by the diagonal ones. The off-diagonal elements can have no independent dynamics of their own, unlike what would be the case if they were derived from a full calculation. One should also note that the diagonal terms of the propagator are obtained from the first order perturbation theory. Since they determine the off-diagonal terms within the Markov approximation, the latter only contains the information obtained from first-order perturbation theory. However, in our calculation the off-diagonal terms involve second--order effects and hence, reveals dynamical correlations that are inaccessible within the context of the Born-Markov approximation.
For the reasons above, the matrix elements (\ref{FTt}--\ref{FTs})
given by \cite{FicTan06} obtained under the Markovian approximation--carry the characteristic superradiant and subradiant behavior of the decay of the $|+ \rangle$ and $|- \rangle$ states respectively, a property that does not arise from our calculation.
Moreover, we note that Eq. (\ref{ps}) obtained by making the Markov approximation is related to arguments from {\em classical probability}, i.e. it involves addition of the probabilities associated to different processes \footnote{To be precise, in an exponential decay, it is possible to define a probability density for the occupation number of any state as a function of time. This is in general not possible in quantum theory. Eq. (\ref{ccc}) is then essentially the Kolmogorov additivity condition for these probabilities. }. On the other hand, a proper quantum mechanical calculation involves the addition of amplitudes--e.g. the ones appearing in the definition of the functions $\kappa_{1,2}(t)$-- and as such it must also contain `interference' terms. In effect, the Markovian approximation introduces by hand a degree of partial `decoherence' (or classicality) and we think that this explains why in general it predicts a faster classicalization of the system than the full analysis does.
We believe that the feature discussed above is generic. In the single qubit system, it was not present because its structure was too simple. There could be no intermediate decays. However, this effect should in principle be present in any system that contains
intermediate states. The Markovian approximation would then be valid only if specific conditions hold that render the `interference' terms negligible--for example if there exists a sharp separation of the relevant timescales.
To summarize, the Markov approximation essentially ignores interference terms that are relevant to processes that involve successive decays. These processes appear through off-diagonal terms in the reduced density matrix propagator. The Markov approximation misrepresents the intrinsic dynamics for these terms and ties them --by forcing additivity of probabilities--to the evolution of the diagonal ones. As a results, the off-diagonal terms are subsumed by the diagonal ones.
\subsection{The use of the rotating wave approximation} Finally, we add a few words on the accuracy of our model for the interaction between the 2LA's and the EMF. The Hamiltonian (\ref{Hint}) is obtained from the study of the interaction of the atomic degrees of freedom with the electromagnetic field. Its derivation involves the dipole and the rotating wave (RW) approximations--see Appendix A in Ref. \cite{AH}. The RW approximation consists in ignoring rapidly oscillating terms in the interaction-picture Hamiltonian and keeping only the ones that correspond to resonant coupling. (One ignores processes during which a photon is emitted {\em and} the atom becomes excited.) The terms that are dropped out in the RW approximation oscillate with a frequency of order $\omega_o$.
For a single qubit system the RW approximation is self-consistent. However, in the two-qubit system we keep terms in the Hamiltonian that vary in space as $e^{i {\bf k} \cdot {\bf r}}$. For the RW approximation to be consistent, one has to assume that ${\bf k} \cdot {\bf r} << \omega_o t$. Since ${\bf k}$ is peaked around the resonance frequency, this condition is equivalent to $ r << t$. Since the physically interesting time-scale for the study of disentanglement and decoherence corresponds to $\Gamma_0^{-1} \sim [\lambda^{2} \omega_0]^{-1}$, we expect the RW approximation to be reliable in this context, as long as $\lambda^2 \omega_o r << 1$. This is sufficient for realistic situations, in which $r$ is bounded by the size of the cavity and $\lambda^2 << 1$. However, the condition above does not hold at the formal limit $r \rightarrow \infty$. In this regime the RW approximation may break down. Indeed, in section IV.B the reduced density matrix for the single qubit in the limit $r \rightarrow \infty$ does not coincide with the corresponding expression obtained in the study of the single qubit. The presence of an {\em excited} second qubit, even if it is situated far away, affects the time evolution significantly. This effect is also present in the Born-Markov approximation. This is arguably an unphysical behavior, and we believe that it arises as an artefact of the RW approximation. \\
\noindent{\bf Acknowledgement} It is a pleasure for one of us (BLH) to hear the lectures of Dr. Ting Yu and the seminar of Professor Joe Eberly explicating their work on disentanglement of two qubits. We also thank the referee for urging us to derive the master equation so the non-Markovian features can be demonstrated more explicitly. This work is supported by grants from the NSF-ITR program (PHY-0426696), NIST and NSA-LPS. CA was supported by a Pythagoras II grant (EPEAEK).
\begin{appendix} \section{Perturbative resummation}
We notice that the action of the operator $[(E-H_0)^{-1}H_I]^2$ on any linear combination of the vectors $| O;0,1 \rangle $ and $| O; 1, 0 \rangle$ yields another linear combination of these vectors. If we denote by $p_n $ and $q_n$ the coefficients of $| O;0,1 \rangle $
and $| O; 1,0 \rangle$ respectively after the n-th action of $[(E-H_0)^{-1}H_I]^2$ (2n-th order of perturbation theory) we obtain \begin{eqnarray} \left( \begin{array}{l} p_n \\ q_n \end{array} \right) = \frac{1}{E - \Omega}\left( \begin{array}{ll} \alpha(E) & \beta^*(E,r) \\ \beta(E,r) & \alpha(E) \end{array} \right) \left( \begin{array}{l} p_{n-1} \\ q_{n-1} \end{array} \right) = : A \left( \begin{array}{l} p_{n-1} \\ q_{n-1} \end{array} \right) \end{eqnarray} Consequently, \begin{eqnarray} \left( \begin{array}{l} p_n \\ q_n \end{array} \right) = A^n \left( \begin{array}{l} p_0 \\ q_0 \end{array} \right), \end{eqnarray} and summing over all $n$ we obtain \begin{eqnarray} \sum_{n=0}^{\infty} \left( \begin{array}{l} p_n \\ q_n \end{array} \right) = \left( \sum_{n=0}^{\infty} A^n \right) \left( \begin{array}{l} p_0 \\ q_0 \end{array} \right) = (1 -A)^{-1} \left( \begin{array}{l} p_0 \\ q_0 \end{array} \right) \end{eqnarray}
To compute $\langle z; 0,1|(E-H)^{-1}|O;01 \rangle$ and $\langle z; 1,0|(E-H)^{-1}|O;01 \rangle$ according to the perturbation series we have $p_0 = (E-\omega_o)^{-1}$, $q_0 = 0$. The results in the main text then follow.
The resummation for $\langle z; 0,0|(E-H)^{-1}|O;0,1 \rangle$ proceeds from the fact that the terms in the $2n+1$ order of the perturbative expansion are of the form \begin{eqnarray}
\sum_{\bf k} \frac{g_{\bf k} \zeta^n_{\bf k} e^{i {\bf k} \cdot {\bf r}/2} }{E - \omega_{\bf k}} b^{\dagger}_k | O, 0 , 0 \rangle \end{eqnarray} We then note that \begin{eqnarray} \zeta_{\bf k}^n = [\frac{ \alpha(E) + \beta(E,r)e^{i {\bf k} \cdot {\bf r}} }{E - \omega_o}] \zeta_{\bf k}^{n-1}, \end{eqnarray} where $\zeta^0_{\bf k} = (E- \omega_o)^{-1}$. This geometric series can be summed to give the result quoted in the main text.
A similar summation is achieved for the terms $\langle z; 0,1|(E-H)^{-1}|O; 1, 1 \rangle $ and $\langle z; 1, 0|(E-H)^{-1}|O; 1, 1 \rangle $. In the $2n$-th order of perturbation theory the action of the series on $O; 1, 1 \rangle$ yields \begin{eqnarray} \sum_{\bf k} \frac{g_{\bf k} }{(E - 2 \Omega) (E - \Omega -
\omega_{\bf k})} \left( e^{-i {\bf k}\cdot {\bf r}/2} \gamma^n_{\bf k} |O; 0, 1 \rangle + e^{i {\bf k}\cdot {\bf r}/2} \delta^n_{\bf k}
|O; 1, 0 \rangle \right) \end{eqnarray}
\section{The time-dependent terms of the master equation}
The time-dependent coefficients $\alpha_i, i = 1, \ldots 6$ appearing in the master equation (\ref{ME1}--\ref{MEf}) are defined as \begin{eqnarray} \alpha_1(t) &=& i \frac{d}{d t}\left(\nu(t) e^{i \omega_ot + \Gamma_0t - i \sigma t - \Gamma_rt}\right) e^{2 \Gamma_0 t + 2 i \sigma t + 2 \Gamma_rt} \label{alpha1} \\ \alpha_2(t) &=& \frac{d}{d t} \left( \mu(t) e^{i\omega_0t + \Gamma_0 t- i \sigma t- \Gamma_r t} \right) e^{2 \Gamma_0t} \\ \alpha_3(t) &=& \frac{d}{d t} \left( \mu(t) e^{i\omega_0t + \Gamma_0 t + i \sigma t + \Gamma_r t} \right) e^{2 \Gamma_0t} \\ \alpha_4(t) &=& i \frac{d}{d t}\left(\nu(t) e^{i \omega_ot + \Gamma_0t + i \sigma t + \Gamma_rt}\right) e^{2 \Gamma_0 t - 2 i \sigma t - 2 \Gamma_rt}\\ \alpha_5(t) &=& \frac{d}{d t}\left( \kappa(t) e^{2\Gamma_0t + 2 \Gamma_rt}\right) e^{2 \Gamma_0t- 2 \Gamma_r t} \\ \alpha_6(t) &=& \frac{d}{d t}\left( \kappa(t) e^{2\Gamma_0t - 2 \Gamma_rt}\right) e^{2 \Gamma_0t+ 2 \Gamma_r t}. \label{alpha6} \end{eqnarray}
Using the notation \begin{eqnarray}
\hat{A}_{+I} &=& |+ \rangle \langle I|, \\
\hat{A}_{OI} &=& |O \rangle \langle I| \\
\hat{A}_{++} &=& |+ \rangle \langle +| \\
\hat{A}_{+-} &=& |+ \rangle \langle -| \\
\hat{A}_{--} &=& |- \rangle \langle -|, \end{eqnarray} the linear functional ${\bf F}_t$ and ${\bf G}_t$ of Eq. (\ref{Lindbladlike}) are defined as \begin{eqnarray} {\bf F_t}[\hat{\rho}] &=& -[2 - \alpha_5(t)] (\hat{A}_{+I} \hat{\rho} \hat{A}^{\dagger}_{+I} - \hat{A}_{OI} \hat{\rho} \hat{A}_{OI}^{\dagger}) - [2 - \alpha_3(t)] \hat{A}_{OI} \hat{\rho} \hat{A}_{++} - [2 - \alpha_3^*(t)] \hat{A}_{++} \hat{\rho} \hat{A}_{OI}^{\dagger} \nonumber \\
&+& \alpha_1(t) \hat{A}_{OI} \hat{\rho} \hat{A}_{+-} + \alpha_1^*(t) \hat{A}_{+-}^{\dagger}
\hat{\rho} \hat{A}_{OI}^{\dagger} \label{FF}\\ {\bf G_t}[\hat{\rho}] &=& - [2 - \alpha_6(t)] (\hat{A}_{-I} \hat{\rho} \hat{A}^{\dagger}_{-I} - \hat{A}_{OI} \hat{\rho} \hat{A}_{OI}^{\dagger}) +[2 + \alpha_2(t)] \hat{A}_{OI} \hat{\rho} \hat{A}_{--} + [2 + \alpha_2^*(t)] \hat{A}_{--} \hat{\rho} \hat{A}_{OI}^{\dagger} \nonumber \\ &+& \alpha_4(t) \hat{A}_{OI} \hat{\rho} \hat{A}_{+-}^{\dagger} + \alpha_4^*(t) \hat{A}_{+-}^{\dagger} \hat{\rho} \hat{A}_{OI}^{\dagger} \label{GG}. \end{eqnarray}
\section{A Grassmann path-integral derivation}
In Refs. \cite{AH,SADH} the evolution of the reduced density matrix for a 2LA interacting with the EM field was determined with a version of the Feynman-Vernon path integral method, using Grassmann variables for the atomic degrees of freedom. The same method can be applied to the present problem, even though the perturbative approach turned out to be more convenient because the vacuum initial state allowed for a summation of the perturbative series. However, the path-integral method may allow for a simpler treatment of other systems, such as an N-qubit system or the EM field at a finite temperature. For this reason, we include a sketch of the path-integral treatment in this Appendix, noting that it leads to the same results as the ones obtained in Sec. II.
\subsection{Coherent state path integral} The coherent states for the EMF (bosonic) and 2LA (qubit) states are defined respectively by \begin{eqnarray}
\label{emcoherentstates} |z_{\bf k} \rangle &=& \exp(z_{\bf k} b_{\bf k}^\dagger)
|0_{\bf k} \rangle \\
\label{gmncoherentstates} |\eta^{(n)} \rangle &=& \exp(\eta S_+^{(n)})
|0\rangle_{(n)}. \end{eqnarray}
The states $|0_{\bf k} \rangle$ and $|0\rangle_{(n)}$ are ground states of the electromagnetic field's ${\bf k}^{\mbox{th}}$ mode in vacuum and the $n^{\mbox{th}}$ qubit in its lower state, repectively. The bosonic coherent states are labeled by a complex number, $z_{\bf k}$, and the qubit coherent states are labelled by an anti-commuting~(i.e.~Grassmannian) number, $\eta^{(n)}$. A non-interacting eigenstate of two qubits and the electromagnetic spectrum can be written as the direct product of coherent states, \begin{equation}
|\{z_{\bf k}\},\eta^{(1)},\eta^{(2)} \rangle =|\{z_{\bf k}\}\rangle \otimes
|\eta^{(1)}\rangle \otimes |\eta^{(2)} \rangle. \end{equation}
In a path integral approach the transition amplitude between some chosen initial and final states is divided into many infinitesimal time steps. The resulting path integral corresponds to the matrix element of the evolution operator. In the coherent-state path integral representation the Hamiltonian is given by, \begin{eqnarray}\label{hamiltonian2}
\frac{\langle \bar{\eta}^{(1)}, \bar{\eta}^{(2)}, \{\bar{z}_{k}\} | H | \eta^{\prime(1)}, \eta^{\prime(2)}, \{z^{\prime}_{k}\} \rangle}{\exp[\bar{\eta}^{(1)} \eta^{\prime(1)} +\bar{\eta}^{(2)} \eta^{\prime(2)} +\sum_{\bf k} \bar{z}_{k} z^{\prime}_{k}]}= \hbar \omega_o \bar{\eta} \eta^{\prime} + \hbar \sum_{\bf k} \left[ \omega_{{\bf k}} \bar{z}_{\bf k} z^{\prime}_{\bf k} + \sum_n \left(\lambda_{{\bf k}}^{(n)} \bar{\eta}^{(n)} z^{\prime}_{\bf k} +\bar{\lambda}_{{\bf k}}^{(n)} \bar{z}_{\bf k} \eta^{\prime(n)} \right) \right]. \end{eqnarray} Spatial dependence is carried in the coupling constants, \begin{eqnarray}\label{spatialdependence} \lambda_{\bf k}^{(n)} = \frac{\lambda^{(n)}}{\omega_{\bf k}} e^{-i{\bf k}\cdot{\bf r}_n} \\ \bar{\lambda}_{\bf k}^{(n)} = \frac{\bar{\lambda}^{(n)}}{\omega_{\bf k}} e^{+i{\bf k}\cdot{\bf r}_n}. \end{eqnarray} The transition amplitude for N atoms after the j-th infinitesimal time steps is, \begin{equation} \label{jstepprop} K(j\epsilon,0) = \exp\bigg\{ \sum_n \bar{\eta}_{j}^{(n)} \bigg[\sum_m \psi_j^{(nm)}\eta_{0}^{(m)} +\sum_{{\bf k}{\bf l}} g_{j{\bf k}{\bf l}}^{(n)}z_{0{\bf l}}\bigg] +\sum_{\bf k} \bar{z}_{j{\bf k}} \bigg[\sum_{\bf l} f_{j{\bf k}{\bf l}}z_{0{\bf l}} +\sum_m \phi_{j{\bf k}}^{(m)}\eta_{0}^{(m)}\bigg] \bigg\}. \end{equation} The functions in the path integral are determined by finite-difference equations \begin{equation} \label{inductive1} \begin{array}{lll} \psi_{j+1}^{(nm)} = (1-i \omega_o \epsilon) \psi_{j}^{(nm)} -i\epsilon\sum_{{\bf k}} \lambda_{j{\bf k}}^{(n)} \phi_{j{\bf k}}^{(m)} && \psi_0^{(nm)} =\delta_{nm} \\ \phi_{j+1,{\bf k}}^{(m)} = (1-i \omega_{{\bf k}}\epsilon)\phi_{j{\bf k}}^{(m)} -i\epsilon \sum_n \bar{\lambda}_{j{\bf k}}^{(n)} \psi_{j}^{(nm)}
&& \phi_{0{\bf k}}^{(m)} =0 \end{array} \end{equation} \begin{equation} \label{inductive2} \begin{array}{lll} g_{j+1,{\bf k}{\bf l}}^{(n)} = (1-i \omega_o\epsilon) g_{j{\bf k}{\bf l}}^{(n)} -i\epsilon \lambda_{j{\bf k}}^{(n)} f_{j{\bf k}{\bf l}} && g_{0{\bf k}{\bf l}} = 0 \\ f_{j+1,{\bf k}{\bf l}} = (1-i\omega_{{\bf k}}\epsilon) f_{j{\bf k}{\bf l}} -i\epsilon \sum_m \bar{\lambda}_{j{\bf k}}^{(m)} \sum_{\bf q} g_{j{\bf q}{\bf l}}^{(m)} && f_{0{\bf k}{\bf l}} =\delta_{{\bf k}{\bf l}}. \end{array} \end{equation}
\subsection{Reduced Density Matrix} The reduced density matrix of the two 2LAs is \begin{equation} \label{evolved1} \rho(t) = \int d\mu(\eta_0^{(2)})d\mu(\eta_0^{\prime(2)})d\mu(\eta_0^{(1)})d\mu(\eta_0^{\prime(1)}) \rho(0) J_R(t,0). \end{equation} The dynamics of this open system is described by the density matrix propagator \begin{equation}\label{reduced propagator} J_R(t,0)\vert_{T=0} = \exp\bigg\{ \sum_{nm} \bigg[\bar{\eta}_t^{(n)} \psi(t)^{(nm)} \eta_0^{(m)} +\bar{\eta}_0^{\prime(n)} \bar{\psi}^{\prime}(t)^{(nm)} \eta^{\prime(m)}_t +\bar{\eta}_0^{\prime(n)} \sum_{\bf k} \bar{\phi}^{\prime}_{k}(t) \phi_{k}(t) \eta_0^{(m)}\bigg] \bigg\} \end{equation}
A general initial two-qubit density matrix written in the Grassmann representation is: \begin{eqnarray} \rho(0)= \left( \begin{array}{ccccccc} 1 && \bar{\eta}_0^{(2)} && \bar{\eta}_0^{(1)} && \bar{\eta}_0^{(1)}\bar{\eta}_0^{(2)}\end{array} \right) \left( \begin{array}{lllllll} \rho^{00}_{00} && \rho^{00}_{01} && \rho^{00}_{10} && \rho^{00}_{11} \\ \\ \rho^{01}_{00} && \rho^{01}_{01} && \rho^{01}_{10} && \rho^{01}_{11} \\ \\ \rho^{10}_{00} && \rho^{10}_{01} && \rho^{10}_{10} && \rho^{10}_{11} \\ \\ \rho^{11}_{00} && \rho^{11}_{01} && \rho^{11}_{10} && \rho^{11}_{11} \end{array} \right) \left( \begin{array}{c} 1 \\ \\ \eta_0^{\prime(2)} \\ \\ \eta_0^{\prime(1)} \\ \\ \eta_0^{\prime(1)}\eta_0^{\prime(2)} \end{array} \right). \end{eqnarray}
In addition to Eqs. Eq.(\ref{inductive1}-\ref{inductive2}), we derive finite difference equations for the expanded functions appearing in the reduced density matrix elements \begin{eqnarray} \psi_{j+1}^{(ab)}\psi_{j+1}^{(nm)} &=& (1-2i\omega_o \epsilon)\psi_{j}^{(ab)}\psi_{j}^{(nm)} -i\epsilon\sum_{{\bf k}} \lambda_{j{\bf k}}^{(n)} \phi_{j{\bf k}}^{(m)}\psi_{j}^{(ab)} -i\epsilon\sum_{{\bf k}} \lambda_{j{\bf k}}^{(a)} \phi_{j{\bf k}}^{(b)}\psi_{j}^{(nm)} \\ \psi_{j+1}^{(ab)}\phi_{j+1,{\bf k}}^{(m)} &=& (1-i\omega_o\epsilon-i\omega_{\bf k})\psi_{j}^{(ab)}\phi_{j{\bf k}}^{(m)} -i\epsilon\sum_{n} \bar{\lambda}_{j{\bf k}}^{(n)} \psi_{j}^{(nm)}\psi_{j}^{(ab)} -i\epsilon\sum_{{\bf l}} \lambda_{j{\bf l}}^{(a)} \phi_{j{\bf l}}^{(b)}\phi_{j{\bf k}}^{(m)} \\ \phi_{j+1,{\bf l}}^{(b)}\phi_{j+1,{\bf k}}^{(m)} &=& (1-i\omega_{\bf l}\epsilon-i\omega_{\bf k})\phi_{j{\bf l}}^{(b)}\phi_{j{\bf k}}^{(m)} -i\epsilon\sum_{n} \bar{\lambda}_{j{\bf k}}^{(n)} \psi_{j}^{(nm)}\phi_{j{\bf l}}^{(b)} -i\epsilon\sum_{a} \bar{\lambda}_{j{\bf l}}^{(a)} \psi_{j}^{(ab)}\phi_{j{\bf k}}^{(m)} \end{eqnarray} For a two-atom system, it is convenient to represent the functions appearing in the propagator by column vectors, \begin{eqnarray} \label{columnvec1} [\psi] &=& \left( \begin{array}{c} \psi^{(22)} \\ \psi^{(21)} \\ \psi^{(12)} \\ \psi^{(11)} \end{array}\right) \;\;\;\;\;\;\;\; \label{columnvec2} [\phi_{\bf k}] = \left( \begin{array}{c} \phi_{\bf k}^{(2)} \\ \phi_{\bf k}^{(1)} \end{array}\right) \;\;\;\;\;\;\; \label{columnvec3} [\psi\psi] = \left( \begin{array}{c} \psi^{(11)}\psi^{(22)} \\ \psi^{(21)}\psi^{(12)} \end{array}\right) \\ \label{columnvec4} [\psi\phi_{{\bf k}}] &=& \left( \begin{array}{c} \psi^{(22)}\phi_{\bf k}^{(1)} \\ \psi^{(21)}\phi_{\bf k}^{(2)} \\ \psi^{(12)}\phi_{\bf k}^{(1)} \\ \psi^{(11)}\phi_{\bf k}^{(2)} \end{array}\right) \;\;\;\;\;\;\;\;\; \label{columnvec5} [\phi\phi_{{\bf k}{\bf l}}] = \left( \begin{array}{c} \phi_{\bf k}^{(1)}\phi_{\bf l}^{(2)} \\ \phi_{\bf k}^{(2)}\phi_{\bf l}^{(1)} \end{array}\right). \end{eqnarray} At the continuous limit the finite difference equations yield a set of differential equations for the column vectors above. \begin{eqnarray} \label{functional1} \frac{d}{dt} [\psi] &=& -i\omega_o [\psi] -i\sum_{\bf k} G^{\dagger}_{\bf k} [\phi_{{\bf k}}] \\ \label{functional2} \frac{d}{dt} [\phi_{\bf k}] &=& -i\omega_{\bf k} [\phi_{\bf k}] -iG_{\bf k} [\psi] \\ \label{functional3} \frac{d}{dt} [\psi\psi] &=& -2i\omega_o [\psi\psi] -i\sum_{\bf k} A_{\bf k} [\phi\psi_{{\bf k}}] \\ \label{functional4} \frac{d}{dt} [\phi\psi_{{\bf k}}] &=& -i(\omega_o +\omega_{\bf k}) [\phi\psi_{{\bf k}}] -i A^{\dagger}_{\bf k} [\psi\psi] -i\sum_{\bf l} L_{\bf l} [\phi\phi_{{\bf k}{\bf l}}]\\ \label{functional5} \frac{d}{dt} [\phi\phi_{{\bf k}{\bf l}}] &=& -i(\omega_{\bf l} +\omega_{\bf k}) [\phi\phi_{{\bf k}{\bf l}}] -i G_{{\bf l}} [\phi\psi_{{\bf k}}] -i D_{{\bf k}} [\phi\psi_{{\bf l}}], \end{eqnarray} in terms of the matrices \begin{eqnarray} G_{\bf k} &=& \left(\begin{array}{cccc} \bar{\lambda}_{\bf k}^{(2)} & 0 & \bar{\lambda}_{\bf k}^{(1)} & 0 \\ 0 & \bar{\lambda}_{\bf k}^{(2)} & 0 & \bar{\lambda}_{\bf k}^{(1)} \end{array}\right) \;\;\;\;\;\;\; A_{\bf k} = \left(\begin{array}{cccc} \lambda_{\bf k}^{(1)} & 0 & 0 & \lambda_{\bf k}^{(2)} \\ 0 & \lambda_{\bf k}^{(1)} & \lambda_{\bf k}^{(2)} & 0 \end{array}\right) \\
D_{{\bf k}} &=& \left(\begin{array}{cccc} 0 & \bar{\lambda}_{\bf k}^{(2)} & 0 & \bar{\lambda}_{\bf k}^{(1)}\\ \bar{\lambda}_{\bf k}^{(2)} & 0 & \bar{\lambda}_{\bf k}^{(1)} & 0 \end{array}\right) \;\;\;\;\;\;\; L_{\bf l} = \left(\begin{array}{cc} 0 & \lambda_{\bf l}^{(2)} \\ 0 & \lambda_{\bf l}^{(2)} \\ \lambda_{\bf l}^{(1)} & 0 \\ \lambda_{\bf l}^{(1)} & 0 \end{array}\right) \end{eqnarray}
The solution of Eqs. (\ref{functional1}--\ref{functional5}) allows one to fully reconstruct the propagating amplitude and from this the elements of the reduced density matrix for the qubits. These equations can be solved by implementing an approximation scheme similar to that employed for the operator method in the main text. We do not provide the details, but only note that the results of the two methods coincide.
\end{appendix}
\end{document}
|
arXiv
|
{
"id": "0610007.tex",
"language_detection_score": 0.7538899183273315,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\newcommand{\begin{equation}}{\begin{equation}} \newcommand{\end{equation}}{\end{equation}} \newcommand{\begin{eqnarray}}{\begin{eqnarray}} \newcommand{\end{eqnarray}}{\end{eqnarray}} \newcommand{\hfil \break \break}{\hfil \break \break} \newcommand{\hat{a}}{\hat{a}} \newcommand{\hat{a}^{\dagger}}{\hat{a}^{\dagger}} \newcommand{\hat{a}_g^{\dagger}}{\hat{a}_g^{\dagger}} \newcommand{\hat{b}}{\hat{b}} \newcommand{\hat{b}^{\dagger}}{\hat{b}^{\dagger}} \newcommand{\hat{b}_g^{\dagger}}{\hat{b}_g^{\dagger}} \newcommand{\hat{c}}{\hat{c}} \newcommand{\hat{c}^{\dagger}}{\hat{c}^{\dagger}} \newcommand{\hat{d}}{\hat{d}} \newcommand{\hat{n}}{\hat{n}} \newcommand{\hat{n}^{\dagger}}{\hat{n}^{\dagger}} \newcommand{\hat{\rho}}{\hat{\rho}} \newcommand{\hat{\phi}}{\hat{\phi}} \newcommand{\hat{A}}{\hat{A}} \newcommand{\hat{A}^{\dagger}}{\hat{A}^{\dagger}} \newcommand{\hat{B}}{\hat{B}} \newcommand{\hat{B}^{\dagger}}{\hat{B}^{\dagger}} \newcommand{\hat{C}}{\hat{C}} \newcommand{\hat{D}}{\hat{D}} \newcommand{\hat{E}}{\hat{E}} \newcommand{\hat{L}}{\hat{L}} \newcommand{\hat{N}}{\hat{N}} \newcommand{\hat{O}}{\hat{O}} \newcommand{\hat{O}^{\dagger}}{\hat{O}^{\dagger}} \newcommand{\hat{S}}{\hat{S}} \newcommand{\hat{U}}{\hat{U}} \newcommand{\hat{U}^{\dagger}}{\hat{U}^{\dagger}} \newcommand{\hat{X}}{\hat{X}} \newcommand{\hat{Z}}{\hat{Z}} \newcommand{\hat{X}^{\dagger}}{\hat{X}^{\dagger}} \newcommand{\hat{Y}^{\dagger}}{\hat{Y}^{\dagger}} \newcommand{\hat{Z}^{\dagger}}{\hat{Z}^{\dagger}} \newcommand{\hat{H}}{\hat{H}} \newcommand{{\prime \prime}}{{\prime \prime}} \newcommand{{\prime \prime \prime}}{{\prime \prime \prime}}
\newcommand{\ket}[1]{\mbox{$|#1\rangle$}}
\newcommand{\bra}[1]{\mbox{$\langle#1|$}}
\newcommand{\ketbra}[2]{\mbox{$|#1\rangle \langle#2|$}}
\newcommand{\braket}[2]{\mbox{$\langle#1|#2\rangle$}}
\newcommand{\bracket}[3]{\mbox{$\langle#1|#2|#3\rangle$}}
\newcommand{\mbox{\boldmath $\cdot$}}{\mbox{\boldmath $\cdot$}} \newcommand{\otimes}{\otimes}
\newcommand{\op}[2]{\mbox{$|#1\rangle\langle#2|$}} \newcommand{\hak}[1]{\left[ #1 \right]} \newcommand{\vin}[1]{\langle #1 \rangle}
\newcommand{\abs}[1]{\left| #1 \right|} \newcommand{\tes}[1]{\left( #1 \right)} \newcommand{\braces}[1]{\left\{ #1 \right\}} \newcommand{I}{I}
\newcommand{\initial}[2]{|\psi^#1_#2(0)\rangle} \newcommand{\initialbra}[2]{\langle\psi^#1_#2(0)}
\newcommand{\decohered}[2]{|\psi^#1_#2(\gamma)\rangle} \newcommand{\decoheredbra}[2]{\langle\psi^#1_#2(\gamma)}
\newcommand{\gendecohered}[1]{|\psi^G(#1)\rangle} \newcommand{\gendecoheredbra}[1]{\langle\psi^G(#1)}
\newcommand{\projector}[2]{|\ksi^#1_#2 \rangle}
\title[Non-monotonic projection probabilites]{Non-monotonic dependence of projection probabilities as a function of distinguishability} \author{Gunnar Bj\"{o}rk, Saroosh Shabbir} \address{Department of Applied Physics, Royal Institute of Technology (KTH)\\ AlbaNova University Center, SE - 106 91 Stockholm, Sweden} \ead{[email protected]}
\date{\today}
\begin{abstract} Typically, quantum superpositions, and thus measurement projections of quantum states involving interference, decrease (or increase) monotonically as a function of increased distinguishability. Distinguishability, in turn, can be a consequence of decoherence, for example caused by the (simultaneous) loss of excitation or due to inadequate mode matching (either deliberate or indeliberate). It is known that for some cases of multi-photon interference, non-monotonic decay of projection probabilities occurs, which has so far been attributed to interference between four or more two photons. We show that such a non-monotonic behaviour of projection probabilities is not unnatural, and can also occur for single-photon and even semiclassical states. Thus, while the effect traces its roots from indistinguishability and thus interference, the states for which this can be observed do not need to have particular quantum features. \end{abstract}
\maketitle
\section{Introduction} Interference is the manifestation of the quantum mechanical law stating that when a final state can be reached through two or more indistinguishable ``paths'', the path probability amplitudes should be added before any final state probability is calculated. Since probability amplitudes are represented by complex numbers and may change signs, path probability amplitudes can thus ``annihilate'' each other, resulting in destructive interference. On the contrary, if the paths to the final state are distinguishable, if even only in principle, then the probabilities and not the amplitudes of the paths should be added. Since probabilities are real and non-negative, adding probabilities cannot result in destructive interference. Perhaps the best known experiment to demonstrate the general validity of this law, i.e., interference due to indistinguishability is the Hanbury Brown and Twiss stellar interferometer \cite{HBT,HBT2}. The surprise for the scientific community at the time was that intensities interfered, rather than wave amplitudes. However, this experiment firmly established that path indistinguishability is what causes interference.\\
Thus, it is perhaps to be expected that constructive (destructive) interference probabilities should decrease (increase) monotonically with increasing distinguishability. In either case one could consider that when interference is probed as a function of distinguishability a monotonic behaviour would follow. The Hong-Ou-Mandel (HOM) experiment \cite{HOM} is a prototype for examining the interference of increasingly distinguishable particles, which indeed shows monotonic dependence as a function of distinguishability. However, recent extensions of the HOM interference to higher photon numbers \cite{Ra, Tichy} have shown that under certain circumstances the measured interference signal varies non-monotonically with increasing distinguishability of the interfering particles (photons in this case).\\
The quantitative relationship between path distinguishability and interference visibility has been studied in the past in different contexts \cite{Jaeger,Englert,Durr,Bjork,Gavenda}. Here, we are not going to concern ourselves with visibility, but only look at how certain (projective) measurements of interference depend on the degree of distinguishability, and specifically whether they vary monotonically with increasing distinguishability or not. We thus do not study interference as a function of the relative phase between the paths, but for a fixed phase. Moreover, while first order interference oscillates sinusoidally with a path relative-phase period of $2 \pi$, the higher (even) order interference functions we will study (like the Hong-Ou-Mandel dip) do not oscillate as a function of the path relative-phase.\\
The authors of Ref. \cite{Ra, Tichy} associate distinguishable events with classical ``particles'' (or ``paths'') and the indistinguishable ``particles'' with quantum trajectories. They subsequently attribute the non-monotonicity to a quantum-to-classical transition of multi-particle interference. However, our analysis shows that contrary to what one may believe, interference probabilities can be non-monotonic for single particles as well, whether they are bosons or fermions. Moreover, the measured interference probabilities analysed and measured in \cite{Ra} were sums of, in principle distinguishable, final events, some increasing and some decreasing with increasing distinguishability. Taking the analysis further, we shall demonstrate that more fundamentally, isolated probabilities (projection probabilities onto pure, final states) can also vary non-monotonically. Finally we shall discuss that this behaviour occurs also classically, specifically with a classical light-source and both with conventional photo detectors and single-photon detectors. Thus we find that the non-monotonic behaviour is natural for all interfering particles or fields and is thus, in general, not a manifestation of a classical-to-quantum transition, or ``exclusive to setups with four or more particles that interfere simultaneously" \cite{Ra}.
\section{Analysis of the Hong-Ou-Mandel experiment} The Hong-Ou-Mandel (HOM) dip \cite{HOM} demonstrates how projection probabilities change as a function of distinguishability, where distinguishability is introduced by delaying one input photon with respect to the other in the second input port of a beam splitter (Fig.~\ref{Fig: schematic}a). The delay causes the loss of interference of the photons at the beam splitter, since time resolution of the output can in principle convey ``which-path" information.
\begin{flushleft} \begin{figure}
\caption{\textbf{a.} Schematic of the Hong-Ou-Mandel experiment. \textbf{b.} Early (left) and late (right) photon modes for each degree of freedom (red and green). Note that the t-axis represents the time at which the photons impinge on the beam splitter. $\gamma$ is the distinguishability transformation which increases downwards, so the top (bottom) figure represents the case when the temporal modes are perfectly indistinguishable (distinguishable).}
\label{Fig: schematic}
\end{figure} \end{flushleft}
We model the HOM experiment with the four mode input state \begin{equation} \decohered{{\scriptsize \textrm{HOM}}}{2} = \cos(\gamma)\ket{1,1}\otimes \ket{0,0} + \sin(\gamma)\ket{1,0}\otimes \ket{0,1},\end{equation} where $\gamma$ parameterizes the delay, and hence distinguishability, between the two photons and ranges from $0 \leq \gamma \leq \pi/2$ and the subscript 2 indicates that it is a two-photon state. The two states to the left of the tensor multiplication sign refers to early photon modes, indistinguishable except for one degree of freedom such as path (illustrated in Fig.~\ref{Fig: schematic}b) or polarization, whereas the two states to the right refer to late photon modes, perfectly distinguishable to the early photons, but indistinguishable among themselves except for, e.g., the path. In this case, the time of arrival can fully determine which photon impinges on the beam splitter first, hence the designation early and late for the two photon modes. Thus when $\gamma=0$ the input photons are perfectly indistinguishable and can display maximal interference when the two modes interact in a 50:50 beam splitter described by the unitary operator $\hat{U}$, \emph{viz.}, $\hat{U}\ket{1,1}\otimes \ket{0,0} = (\ket{2,0}-\ket{0,2})\otimes \ket{0,0}/\sqrt{2}$.\\
Qualitatively, the indistinguishability of the input state can be defined as its projection probability onto the maximally interfering state, i.e., $I \equiv | \initialbra{{\scriptsize \textrm{HOM}}}{2}\decohered{{\scriptsize \textrm{HOM}}}{2}|^2=\cos^2(\gamma).$ Note that $I$ is not a good measure in quantitative terms: for different perfectly distinguishable modes $I$ may not be zero as we shall see below. However, for our purposes this simple measure suffices as it is a monotonically decreasing function of $\gamma$ for all our investigated methods of increasing distinguishability.
The output state after the beam splitter is
\begin{eqnarray}\fl \hat{U} \decohered{{\scriptsize \textrm{HOM}}}{2}=\frac{\cos(\gamma)}{\sqrt{2}}(\ket{2,0}-\ket{0,2})\otimes \ket{0,0} \nonumber \\ + \frac{\sin(\gamma)}{2}(\ket{1,0}+\ket{0,1})\otimes (\ket{1,0}-\ket{0,1}).\end{eqnarray} Unfortunately, it is rather difficult to measure the projection onto the state $\hat{U}\initial{{\scriptsize \textrm{HOM}}}{2}$. The remedy in a typical experiment is to instead project onto $\ket{1,0,0,1}\bra{1,0,0,1}$ and thus instead measure the \textit{distinguishability}. The state $\ket{1,0,0,1}$ represent the case when an early photon takes the first ``path'' and a late photon takes the second ``path''. If the photons are measured in coincidence, but the coherence length and the time delay of the modes are much shorter than the coincidence measurement window, they will still be recorded as coincident. This is a typical experimental situation. In this case the state $\ket{0,1,1,0}$ will also be detected as coincident, and although the two states are orthogonal they will both contribute to this measurement of interference. Mathematically, since the events are distinguishable in practice (by, e.g., noting in which mode the early photon arrived) the respective measurement probabilities should be added. Then, the output projection probability can be written as \begin{equation} \fl |\bra{1,0,0,1}\hat{U} \decohered{{\scriptsize \textrm{HOM}}}{2}|^2 + |\bra{0,1,1,0}\hat{U} \decohered{{\scriptsize \textrm{HOM}}}{2}|^2 = \frac{\sin^2(\gamma)}{2} \nonumber = \frac{1-I}{2}. \end{equation} Thus, if $I$ is monotonic, so is this measurement of interference.\\
More generally, consider a state $\gendecohered{\gamma}$ of one or more photons, where $\gamma$ parameterizes the degree of distinguishability. For $\gamma=0$ the input modes are perfectly indistinguishable (except for one degree of freedom) when they interfere in some device described by the unitary transformation $\hat{U}$. If the input modes are made partially distinguishable, e.g., by shifting the modes in time, in polarization or in frequency, so that a state with $\gamma \neq 0$ is prepared, then the indistinguishability can be defined as the states' overlap with the maximally interfering state, i.e., $I_{\scriptsize \textrm{HOM}} \equiv |\gendecoheredbra{0}\gendecohered{\gamma}|^2$. After the unitary (interference) transformation, it is very natural to define the degree of indistinguishability as the output state's projection probability onto the output state state exhibiting maximal interference, i.e., $\hat{U}\gendecohered{0}$. In the following we shall define $\hat{U}\gendecohered{0} \gendecoheredbra{0}|\hat{U}^\dagger$ as the \textit{proper} indistinguishability projector.\\
Hence, independent of the exact nature of the interference between the photons, the degree of indistinguishability at the output will be $I_{\scriptsize \textrm{out}} = |\gendecoheredbra{0}|\hat{U}^\dagger \hat{U}\gendecohered{\gamma}|^2 = I_{\scriptsize \textrm{in}} $. From this, almost trivial, consideration we conclude that if the indistinguishability of the input varies in a monotonic fashion, then the \textit{properly measured} interference, and thus indistinguishability of the output modes, also varies monotonically. However, if one chooses a different measurement to probe the interference propensity, then one may obtain a non-monotonic probability as a function of distinguishability, e.g., by taking a sum of projections such that one increases with distinguishability while the other decreases. We demonstrate this in the two photon-pair HOM experiment analyzed below.\\
\section{Two photon-pair HOM experiment} The case where a pair of photons is delayed with respect to another pair at the input of the beam splitter \cite{Ra,Tichy} can be similarly modelled as \begin{eqnarray} \fl \decohered{{\scriptsize \textrm{HOM}}}{4} = \cos^2(\gamma)\ket{2,2}\otimes \ket{0,0} + \sqrt{2}\cos(\gamma)\sin(\gamma)\ket{2,1}\otimes \ket{0,1} \nonumber \\+ \sin^2(\gamma)\ket{2,0}\otimes \ket{0,2}.\end{eqnarray} Using the ``recipe'' above, the indistinguishability between the two pairs of photons can be quantified by $I = \cos^4(\gamma)$. The indistinguishability thus decreases monotonically with $\gamma$ in the relevant interval $0\leq \gamma\leq \pi/2$. In the case of two photons in each output, the coincidence detection window can again be set so that the photons in the output states $\ket{2,2,0,0}$, $\ket{2,1,0,1}$, $\ket{1,2,1,0}$ $\ket{2,0,0,2}$, $\ket{1,1,1,1}$, and $\ket{0,2,2,0}$ all are defined to be coincident (i.e. two photons are detected in each of the two spatial or polarization modes within the coincidence window.) The individual probabilities for these projections are as follows:
\begin{eqnarray}
|\bra{2,2,0,0}\hat{U} \ket{\xi(\gamma)} |^2 &=& \frac{\cos^4(\gamma)}{4}, \nonumber \\
|\bra{2,1,0,1}\hat{U} \ket{\xi(\gamma)} |^2 &=& \frac{\cos^2(\gamma)\sin^2(\gamma)}{8}, \nonumber \\
|\bra{1,2,1,0}\hat{U} \ket{\xi(\gamma)} |^2 &=& \frac{\cos^2(\gamma)\sin^2(\gamma)}{8}, \label{Eq: projs 2 pair}\\
|\bra{2,0,0,2}\hat{U} \ket{\xi(\gamma)} |^2 &=& \frac{\sin^4(\gamma)}{16}, \nonumber \\
|\bra{1,1,1,1}\hat{U} \ket{\xi(\gamma)} |^2 &=& \frac{\sin^4(\gamma)}{4}, \nonumber \\
|\bra{0,2,2,0}\hat{U} \ket{\xi(\gamma)} |^2 &=& \frac{\sin^4(\gamma)}{16}. \nonumber \end{eqnarray}
\begin{figure}
\caption{Probabilities of the $(m,n,\tilde{m},\tilde{n})$ events for the two-pair input HOM experiment, where the latter two digits with tilde refer to late output modes of the beam splitter. \textbf{a.} Sum of (2,2,0,0), (2,1,0,1), (1,2,1,0), (2,0,0,2), (1,1,1,1) and (0,2,2,0) event probabilities. \textbf{b.} (2,2,0,0) event probability. \textbf{c.} Sum of (4,0,0,0), (3,0,1,0) and (2,0,2,0) event probabilities. \textbf{d.} (4,0,0,0) event probability.}
\label{Fig: projection probabilities}
\end{figure}
We see that only the first of these projection probabilities is proportional to the indistinguishability of the input state. Collecting the terms contributing to the coincidence signal (under the assumption that the detection coincidence window $>$ the maximal time delay between the modes $>$ the coherence time) we get the coincidence count probability \begin{equation} P_4^{\scriptsize \textrm{HOM}} = \frac{\cos^4(\gamma)}{4} + \frac{\cos^2(\gamma)\sin^2(\gamma)}{4} + \frac{3 \sin^4(\gamma)}{8}.\end{equation} The probability in this case is not monotonic with respect to the distinguishability parameter $\gamma$ as can be seen in Fig. \ref{Fig: projection probabilities} and was experimentally verified in \cite{Ra}. To see this more clearly, the equation can be rewritten as \begin{equation} P_4^{\scriptsize \textrm{HOM}} = \frac{1}{8}\left (3 I -4 \sqrt{I} + 3 \right ).\end{equation} The function is non-monotonic partially because it measures a sum of terms, three growing with increasing distinguishability and one decreasing. Moreover, the second and third probabilities in (\ref{Eq: projs 2 pair}) are non-monotonic by themselves. For the bunching event (e.g., when all photons exit in the lower path or in the same polarization mode), the measured coincident probability is the sum of projections on $\ket{4,0,0,0}$, $\ket{3,0,1,0}$, $\ket{2,0,2,0}$ states. However, in this case the coincidence probability fortuitously turns out to be monotonic nonetheless \cite{Xiang}, as shown in Fig.~\ref{Fig: projection probabilities}. \\
\section{Distinguishability transformations for single particle states} We now turn to single particle states. In this case there is no difference between bosons and fermions, and for the latter this analysis is general and sufficient since, by the Pauli exclusion principle, fermions can only interfere into final states containing one excitation. We show that analogously to the deliberate distinguishability transformation in the two-photon HOM experiment, one can also prepare a single-particle state in an increasingly distinguishable manner in various ways: one way would be to send a single photon into a variable reflectivity, loss-less beam splitter. The larger the probability amplitude for the state corresponding to finding the photon in one of the arms as compared to the other, the more distinguishable the photon paths become. A similar method would be to prepare the photon in a linearly polarized mode. If the photon's polarization is diagonal to the horizontal-vertical (HV) polarization basis, then the two paths expressed in this basis become indistinguishable, but as the polarization of the photon is rotated towards, e.g., the horizontal direction, the paths become increasingly distinguishable in the HV basis.
Such an initial state can be written as \begin{equation} \ket{\psi^D_1(\gamma)} = \cos(\pi/4 + \gamma/2) \ket{1,0} + \sin(\pi/4 + \gamma/2) \ket{0,1}. \label{Eq:Initial indistinguishability} \end{equation} For $\gamma=0$ it is not possible to distinguish or guess in which of the two modes one would ``find'' the particle, if one were to measure, while for $\gamma=\pi/2$ one can be confident, \textit{a priori} of any measurement, to find the particle in the second mode. This state is projected onto the projector state $ \ket{\xi_1} = \cos(\beta_n) \ket{1,0} + e^{-i \theta_n}\sin(\beta_n) \ket{0,1}, \label{Eq:Projector} $ where $\beta_n$ and $\theta_n$ are variables that can be used to implement any pure, two-mode, single-photon projector. Note that here, for simplicity, we have incorporated the beam splitter, or any other unitary transformation, into this projector, and so referring to Fig.~\ref{Fig: schematic}, this projector state represents $\hat{U} \ket{\xi}$. If the two modes are taken to be the horizontally and vertically, linearly polarized modes, the projector can be implemented through a variable birefringence plate implementing $\theta_n$, followed by polarizer oriented with the transmission polarization at the angle $\beta_n$, that in turn is followed by a single-photon detector. The projection probability $P^D_1$ for this input state is \begin{eqnarray} \fl P^D_1 = \cos^2(\beta_n) \left [1-\sin(\gamma) \right]/2 + \left[\cos(\theta_n)\sin(2\beta_n) \cos(\gamma)\right]/2 \nonumber \\ + \sin^2(\beta_n) \left [1+\sin(\gamma)\right ]/2. \label{Eq:Projector D} \end{eqnarray}
\begin{figure}
\caption{\textbf{a.} The Stokes vector evolution for the different distinguishability models. Dotted (red) line indicates a deliberate distinguishability transformation due to a polarization rotation, Eq. (\ref{Eq:Initial indistinguishability}). Dashed (blue) line, indicates the increase of distinguishability due to linear loss, Eq. (\ref{Eq:Initial loss}). Solid (black) line indicates increasing phase noise, Eq. (\ref{Eq:Initial phase noise}). \textbf{b.} Projection probabilities for $P_1^D$ (dotted, red) and $P_1^L$ (dashed, blue) as a function of distinguishability}
\label{Fig: Bloch}
\end{figure}
This projection probability is not monotonic with respect to $\gamma$ for certain values of the other parameters as can be inferred from Fig.~\ref{Fig: Bloch} \textbf{a}. For polarized photons the figure shows the equatorial plane of the Poincar\'{e} sphere. For the projector state $\ket{\xi_1^D}=\cos(\pi/8) \ket{1,0}-\sin(\pi/8)\ket{0,1}$, corresponding to $\beta_n=\pi/8$ and $\theta_n=\pi$ and marked by the black dot in Fig.~\ref{Fig: Bloch} \textbf{a}, the overlap with the partially distinguishable state on the dotted, red line is zero where the thin black line emanating from the black dot and crossing the origin intersects the dotted curve. (The states, being antipodal on the Poincar\'{e} sphere, are orthogonal.) For states with either more or less distinguishability the overlap is finite since the states are non-orthogonal. Hence, the projection probability will be non-monotonic as a function of $\gamma$, with zero as a minimum as can be seen in Fig.~\ref{Fig: Bloch} \textbf{b}.
The projection measurement $\ket{\xi_1^D}\bra{\xi_1^D}$ corresponds to the interference between two spatial modes in an unbalanced beam splitter followed by a photo detector in one of the arms. In a polarization setting it corresponds to a half-wave plate oriented with its axis parallel to the linear polarization basis, followed by a polarizer set at 22.5 degrees and a photo detector. However, if the \emph{proper} distinguishability projector for this basis $\ket{\xi^P_1}=(\ket{0,1} + \ket{1,0})/\sqrt{2}$ is chosen, then the measured probability equals the indistinguishability $I^D_1=\cos^2(\gamma/2)$ that goes from 1 to 0.5 monotonically.
Other models of distinguishability can be analysed in a similar fashion. In the case where one of the mode experiences linear loss, then if the loss is large enough we can be convinced that any detected photon must have resided in ``the other'' mode. In this case, to preserve unitarity and normalization, at least a three mode model is needed. The input state can thus be modeled \begin{equation} \ket{\psi^L_1(\gamma)} = \left [\cos(\gamma) \ket{1,0,0} + \ket{0,1,0}+ \sin(\gamma) \ket{0,0,1} \right ]/\sqrt{2}. \label{Eq:Initial loss} \end{equation} If we assume that any excitation in the rightmost (ancillary) loss mode remains undetected, the projection probability $P^L_1$ onto the single-photon projector $\ket{\xi_1}\bra{\xi_1}$ is \begin{equation} P^L_1 = \left [ \cos^2(\gamma)\cos^2(\beta_n) + \cos(\theta_n)\sin(2\beta_n)\cos(\gamma) + \sin^2(\beta_n) \right ]/2. \label{Eq:Projector B}\end{equation} This projection probability is also not monotonic with respect to the parameter $\gamma$ for certain values of the other parameters as can easily be seen from Fig. \ref{Fig: Bloch} \textbf{a} and \textbf{b}, blue, dashed line (which is also drawn for $\beta_n=\pi/8$ and $\theta_n=\pi$). The indistinguishability $I^L_1=[1+\cos(\gamma)]^2/2$, however, is monotonic in the relevant interval $0 \leq \gamma \leq \pi/2$.
As a third model, we shall analyse distinguishability brought about by phase noise, the initial state can be written as, \begin{equation} \ket{\psi^N_1} = \left( \ket{1,0} + e^{i \phi_n} \ket{0,1} \right)/\sqrt{2}, \label{Eq:Initial phase noise} \end{equation} where $\phi_n$ is a random variable representing the (relative-)phase fluctuations between the modes. The projection probability of this state onto $\ket{\xi_1}\bra{\xi_1}$ is derived to be, \begin{equation} P^N_1 = \left [ 1 + \langle \cos(\theta_n + \phi_n\rangle \sin(2\beta_n) \right] /2, \label{Eq:Projector with noise} \end{equation} where $\langle \ \rangle$ denotes an ensemble average. If we take $\phi_n$ to have a symmetric distribution around a zero mean, we have $\langle \cos(\theta_n + \phi_n)\rangle = \cos(\theta_n)\langle \cos(\phi_n)\rangle$. In the following we are not interested in modelling any particular physical source of phase noise, but can instead simply use the parametrization $\langle \cos(\phi_n)\rangle = \cos(\gamma)$, $\gamma \in \{0,\pi/2 \}$ to characterize the degree of distinguishability between the two modes due to the fluctuating relative phase. Thus we arrive at our final formula \begin{equation} P^N_1 = \left [ 1 + \cos(\theta_n)\sin(2\beta_n)\cos(\gamma) \right ]/2, \label{Eq:Projector A}\end{equation} From the formula we can see that since $\cos(\gamma)$ is a monotonic function in the interval $\{0,\pi/2 \}$, so is $P^N_1$, irrespective of the value of any of the other parameters. Thus, for this model any single-photon projection probability described by $\ket{\xi_1}$ will either increase or decrease monotonically, or be flat. This can be seen in Fig. \ref{Fig: Bloch} \textbf{a} where the density matrix is transformed along a radial ray in the Poincar\'{e} sphere as a function of increasing phase noise or distinguishability. The indistinguishability in this case is $I^N_1=[1+\cos(\gamma)]/2$ which is monotonic when $0 \leq \gamma \leq \pi/2$.\\
It is well known that when analysing linear interference one can get the classical results by extrapolating the case of single-photon input, i.e., replacing the input probability amplitudes with field amplitudes and interpreting the output probability as a mean output intensity. From this analogy it is clear that if classical fields are used with the mode distinguishability models above, and the corresponding classical interference measurement were done, the corresponding non-monotonic output intensity curves would be obtained. This shows that non-monotonic interference probability does not, in general, stem from a quantum-to-classical transition.
\section{Experiment} The analysis and the predictions of non-monotonicity for the single-photon states above are so simple that we find it pointless to prove them by experiments. Here we shall therefore discuss the more interesting case of distinguishability (mode) transformation for the two-photon case, experimentally implemented as a polarization rotation. We saw in the analysis of the two-photon Hong-Ou-Mandel experiment that a monotonic behavior followed. This is not the case for all two-photon experiments as we shall show. To this end, two photons prepared in a diagonal, linearly polarized mode will have the maximum indistinguishability with respect to the HV modes. Rotating the diagonally polarized state to become, e.g., horizontally polarized, the distinguishability will become perfect. The initial two mode state can be written as \begin{eqnarray} \fl \ket{\psi^D_2} = \sin^2(\pi/4 + \gamma/2) \ket{2,0} + \sqrt{2} \sin(\pi/4 + \gamma/2)\cos(\pi/4 + \gamma/2) \ket{1,1} \nonumber \\+ \cos^2(\pi/4 + \gamma/2) \ket{0,2}, \label{Eq:Initial two-photon indistinguishability} \end{eqnarray} where $\gamma$ is the relative angle between the input and output polarizations. The two-photon interference projector we shall use for this state is \begin{equation} \ket{\xi^D_2} = \frac{1}{\sqrt{3}}\left ( \sqrt{2} \ket{2,0} + \ket{1,1} \right ). \label{Eq:Two photon projector D} \end{equation} It is deliberately chosen to on the one hand involve interference between the two states $\ket{2,0}$ and $\ket{1,1}$, and on the other hand not be the proper indistinguishability projector. Computing the projection probability, one gets \begin{equation} P^D_2 = \frac{4}{3} \sin^2(\pi/4 + \gamma/2)\cos^2(\gamma/2). \label{Eq: prediction}\end{equation} For a diagonally polarized input state, $\gamma=0$, one gets $P^D_2=2/3$. \begin{figure}\label{Fig: setup}
\end{figure}
In \cite{Hofmann} it was shown that any $N$ photon, two-mode projector can be implemented probabilistically by writing the state, expressed in terms of creation operators of the two modes, as a product of single-photon creation operators. For the state $\xi^D_2$ above the decomposition becomes \begin{equation} \xi^D_2 = \sqrt{\frac{2}{3}} \frac{\left ( \hat{a}^{\dagger} + \hat{b}^{\dagger} \right ) }{\sqrt{2}} \hat{a}^{\dagger} \ket{0,0} \label{Eq: Implementation D} \end{equation} The consequence is that the projector can probabilistically be implemented by splitting the diagonally polarized, two-photon state in a non-polarizing 50:50 beam splitter, there is a 50 \% chance that one photon exits from each beam splitter output port. Then after one port a polarizer is inserted that transmits diagonally polarized photons followed by a photo detector. Each time a diagonally polarized photon exits this port an ideal detector will click. After the second port one inserts a polarizer that transmits horizontally polarized photons. Hence, if one diagonally polarized photon exits this port there is a 50 \% chance that an ideal detector clicks. If one looks at the probability $P^E_2$ that the two detectors click in coincidence one finds that (for ideal detectors) the probability is $1/4$. Hofmann proved that $P^E_2 \propto P^D_2$ \cite{Hofmann}. Comparing the two expressions for a diagonally polarized, two-photon, input state we got 1/4 and 2/3, respectively. Hence, the coincidence probability $P_2^E$ will be connected to the projection probability $P_2^D$ as \begin{equation} P^E_2 = \frac{3}{8} P^D_2.\end{equation} Taking non-unity photo detection into account one obtains $P^E_2 = 3 \eta^2 P^D_2/8$, where it has been assumed that the detector efficiency $\eta$ is the same for both detectors. Fig.~\ref{Fig: exp} shows the experimental results - implemented via post-selection of the two-photon component of a weakly excited coherent state input \cite{Shabbir} - along with the theoretical curve. As can be seen, and as predicted by (\ref{Eq: prediction}), the curve is non-monotonic. \begin{figure}
\caption{Two-photon distinguishability polarization transformation. The black squares represent experimental data and the red curve is $A P^D_2$ with $A$ chosen for best fit. The data is normalized with respect to the maximum number of counts. Error bars shows $\pm \sigma$.}
\label{Fig: exp}
\end{figure}
Using classical light, with field amplitude $E_0$, the output of the above experiment would be $(E_0/2)^4 \cos^2(\gamma-\theta_1)\cos^2(\gamma-\theta_1)$, with $\theta_1$ and $\theta_2$ as 0 and $\pi/4$ respectively. This is also a non-monotonic function of the input polarization $\gamma$.
\section{Discussion} In this paper we have discussed various ways by which states can be made increasingly distinguishable with respect to a given measurement basis and we have shown that by suitable choice of interference projectors, non-monotonic projection probabilities can be demonstrated for most states, including single particle states and classical states detected either classically or with single particle detectors. Hence, in general non-monotonicity can neither be regarded as a consequence of \emph{quantum} interference between multiple photons nor as a quantum-to-classical transition phenomenon \cite{Ra, Tichy}. However, given an initial state in a particular measurement basis, a suitable ``maximally indistinguishable'' projector can be defined. We have named such a projector the \textit{proper} indistinguishability projector. With respect to this choice, the interference projection probability always change monotonically as a function of increasing distinguishability. Thus we see that the choice of measurement projector critically determines how the measurement probability behaves as a function of the distinguishability between the interfering modes, particles, or states.
\ack We gratefully acknowledge discussions with Prof. Y.-H. Kim and Mr. Y.-S. Ra. This work was supported by the Swedish Research Council (VR) through grant 621-2011-4575 and through its support of the Linn\ae us Excellence Center ADOPT.
\section*{References}
\end{document}
|
arXiv
|
{
"id": "1312.0055.tex",
"language_detection_score": 0.8027147650718689,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Measuring the Casimir-Polder potential with Unruh-DeWitt detector excitations}
\author{Kacper D\k{e}bski}
\email{[email protected]}
\affiliation{Institute of Theoretical Physics, University of Warsaw, Pasteura 5, 02-093 Warsaw, Poland}
\author{Piotr T. Grochowski}
\email{[email protected]}
\affiliation{Center for Theoretical Physics, Polish Academy of Sciences, Aleja Lotnik\'ow 32/46, 02-668 Warsaw, Poland}
\author{Andrzej Dragan}
\email{[email protected]}
\affiliation{Institute of Theoretical Physics, University of Warsaw, Pasteura 5, 02-093 Warsaw, Poland}
\affiliation{Centre for Quantum Technologies, National University of Singapore, 3 Science Drive 2, 117543 Singapore, Singapore}
\date{\today}
\begin{abstract}
We propose a new method to measure Casimir-Polder force in an optical cavity by means of atomic excitations.
We utilize a framework in which Unruh-DeWitt Hamiltonian mimics the full matter-field interaction and find a connection between the excitation rate of the two-level atom and the Casimir-Polder energy, allowing to map one onto the other.
We argue that such a realization opens a route to study Casimir-Polder potentials through measurement of excited state's population in large, spatially compact ensembles of atoms.
\end{abstract}
\maketitle
\section{Introduction}
The idea of light quantization is one of the cornerstones of the quantum theory.
Among many of its proofs, experiments involving a weak electromagnetic field played a crucial role in establishing the consensus.
The seminal paper of Casimir and Polder shows that even in the limit of a photonless state -- the quantum vacuum -- a neutral atom near a dielectric wall feels a force mediated through the quantum fluctuations~\cite{Casimir1948}.
Not before 1990s was such a behavior experimentally demonstrated, providing another confirmation of the quantization of the light and opening perspectives on investigating the quantum vacuum~\cite{Sukenik1993,Lamoreaux1997,Mohideen1998}.
The presence of Casimir-Polder forces was demonstrated in various setups, involving different types of plates, from conducting to dielectric ones, and different types of probes, from atomic, through mechanical, to Bose-Einstein condensates~\cite{Schwinger1978, Balian1978, Plunien1986,Bordag2001,Harber2005a,Obrecht2007,Klimchitskaya2009}.
In most cases, Casimir forces are attractive, however it was proposed that a repulsive character of the interaction is also possible~\cite{Dzyaloshinskii1961,Boyer1974,Bordag2001,Kenneth2002,Levin2010}, as recently showed experimentally~\cite{Munday2009}.
The QED minimal coupling is usually a model of choice in describing Casimir phenomena, however there are others, providing an effective description of the electromagnetic field.
One of the minimal models allowing system to exhibit both attractive and repulsive character of Casimir force is the Unruh-DeWitt (UDW) model~\cite{DeWitt1979} in the optical cavity~\cite{Alhambra2014}.
Routinely used as a pointlike particle detector in the quantum field theory~\cite{Crispino2008,Birrell1982,Louko2008,Hodgkinson2013,Brown2013}, UDW model constitutes of a two-level atom coupled to the scalar field.
Despite being noticeably simpler than the full QED Hamiltonian, it was shown to be a reasonable approximation to the atom-field interaction when no orbital angular momentum is exchanged~\cite{Martin-Martinez2013a}.
Moreover, it was used to study Casimir-Polder forces~\cite{Passante2003, Rizzuto2004,Spagnolo2006,Rizzuto2007}.
In this work, we analyze UDW model in the optical cavity, focusing on the following two aspects.
Using the second order perturbation theory, we calculate the Casimir-Polder (CP) potential as a function of the atom's position in the cavity.
We find that under some assumptions, it is intrinsically connected to the excitation probability of the two-level system that models the atom.
As a consequence, we propose a simple experimental scenario of measuring the CP force by means of the atomic excitations in a Bose-Einstein condensate placed near the wall.
The work is structured as follows.
In Sec. II we introduce the model, underline its connection to the full QED interaction Hamiltonian, and compute both CP potential and the excitation rate of the UDW detector.
In Sec. III we find the connection between the two and present the toy experimental proposal.
The final Sec. IV concludes the manuscript with the recapitulation and the outlook.
\section{Model}
We will work in natural units, $\hbar=c=e=1$.
Let us consider a scalar field of a mass $m$ governed by the Klein-Gordon equation:
\begin{equation}
\left(\Box+m^2 \right) \hat{\phi}=0
\end{equation}
in the cavity with a length of $L$ fulfilling Dirichlet boundary conditions, $\hat{\phi}(x=0)=\hat{\phi}(x=L)=0$
with the following mode solutions:
\begin{equation}
u_n(x,t)=\frac{1}{\sqrt{\omega_n L}} \sin{\left(k_n x\right)}e^{-i\omega_n t}\equiv u_n(x)e^{-i\omega_n t},
\end{equation}
where $\omega_n=\sqrt{k_n^2+m^2}$, $k_n=\frac{n\pi}{L}$, $n \in \mathbb{Z}$.
Using these modes, the field $\hat{\phi}$ can be decomposed as:
\begin{equation}
\hat{\phi}(x)=\sum_{n}\left[\hat{a}_n^{\dagger} u_{n}^*(x)+\hat{a}_n u_n(x)\right],
\end{equation}
where $\hat{a}_n$ and $\hat{a}_n^{\dagger}$ are annihilation and creation bosonic operators satisfying the canonical commutation relations, $\left[\hat{a}_n,\hat{a}_k^{\dagger}\right]=\delta_{nk}$ and $\left[\hat{a}_n,\hat{a}_k\right]=\left[\hat{a}_n^{\dagger},\hat{a}_k^{\dagger}\right]=0$.
In the distance $d$ from the boundary let us place a two-level system corresponding to the simplest model of an atom with an energy gap $\Omega$.
Such a description approximates e.g. a hydrogen atom which is not placed in a strong classical background field and therefore no transition is resonantly coupled to the cavity.
As we are interested in working with the vacuum state of the cavity, this proves to be a valid approximation.
Then, the full Hamiltonian of the considered model includes free Hamiltonians of the scalar field and of the atom, and the term accounting for the interaction between both of them, $\hat{H}_{\mathrm{I}}$.
One of the simplest possible choices of the interaction between the scalar field and the two-level system is the pointlike Unruh-DeWitt Hamiltonian.
In the Schr\"odinger picture it takes the following form:
\begin{equation}
\hat{H}_{\mathrm{UDW}}=\lambda~\hat{\mu}_{\mathrm{S}}~\hat{\phi}(x),
\label{eqn:UdWhamiltonian}
\end{equation}
where $\lambda$ -- dimensionless coupling constant, $\hat{\mu}_{\mathrm{S}}$ -- the monopole moment of the detector, $\hat{\mu}_{\mathrm{S}}=\hat{\sigma}^{+}+\hat{\sigma}^{-}=\ket{g}\bra{e}+\ket{e}\bra{g}$, where $\ket{g}$ is the ground state of the two-level system and $\ket{e}$ is its excited state.
Moreover, $\hat{\phi}(x)$ is the scalar field operator evaluated at the point at which the pointlike detector is placed.
In the spirit of electromagnetic field considerations, this first order term would be called a paramagnetic one.
This simple model of interaction can be extended to a more realistic form including the second order term in the Hamiltonian corresponding to a diamagnetic, self-interaction term of the full QED Hamiltonian.
This quadratic term of the Unruh-DeWitt Hamiltonian has the form:
\begin{align}
\hat{H}_{\mathrm{UDW}}^2&=\left(\lambda\left(\ket{g}\bra{e}+\ket{e}\bra{g}\right)\hat{\phi}\right)^2 \nonumber \\
&=\lambda^2\left(\ket{g}\bra{g}+\ket{e}\bra{e}\right)\hat{\phi}^2=\lambda^2\hat{\phi}^2.
\end{align}
It is worth to notice that such a term does not change a detector state.
At this point, let us revoke some key results from Refs.~\cite{Martin-Martinez2013a} that compare QED Hamiltonian and UDW one.
The minimal electromagnetic coupling in the Coulomb gauge reads:
\begin{equation}
\hat{H}_{\mathrm{QED}} = -\frac{1}{m} \vec{A}(\vec{x}) \cdot \vec{p}+\frac{1}{2 m} \left[ \vec{A}(\vec{x})\right]^2.
\end{equation}
The main difference is the vector character of the EM interaction in contrast to the scalar one that we consider.
However, a scalar field can be readily utilized to describe electric and magnetic contributions separately, given appropriate boundary conditions.
Indeed, such a description has been used to analyze Casimir-Polder interaction in the past.
It has to be noted that such a scalar model does not allow any exchange of the orbital momentum, however we retreat to the simple case of atomic transitions that obey this rule.
As mentioned above, the QED Hamiltonian consists of two terms -- paramagnetic and diamagnetic ones.
The simplified light-matter interaction Hamiltonians often neglect the second term while working with weak fields.
The minimal coupling in the vacuum implies interaction only with the quantum fluctuations $<\vec{A}^2>$.
However, in the vacuum, the value of $<\vec{A}^2>$ depends on the region in which an atom resides -- or in the language of quantum field theory -- on the region these quantum fluctuations are smeared over.
It happens that while approaching a limit of a pointlike atom (detector), this variance diverges.
So, it introduces a necessity to allow for a finite size of the atom, unlike in the simple UDW model.
In the original Casimir and Polder paper, such a problem was also present -- it was taken care of by the means of introducing a regularizing factor $e^{-\gamma k}$ in the integrals over momentum space.
However, we follow a procedure used by \cite{Martin-Martinez2013a}, where an explicit spatial form of the ground state of the atom is assumed:
\begin{equation}\label{pro}
\Psi(x)=\frac{e^{-x/a_0}}{a_0},
\end{equation}
where $a_0$ is some characteristic length associated to the spherically symmetric atomic profile (meant to be of the order of magnitude of Bohr radius).
Such an approach modifies the UDW Hamiltonian by effectively coupling the detector to an effective field,
\begin{equation}
\hat{\phi}_{\mathrm{R}}(x)=\sum_{n}f_n\left[\hat{a}_n^{\dagger} u_{n}^*(x)+\hat{a}_n u_n(x)\right],
\end{equation}
where
\begin{equation}f_n=\frac{2}{\left(a_0 k_n\right)^2+1}
\end{equation}
are Fourier transforms of the spatial profile \eqref{pro} evaluated at momentum $k_n$.
Such a momentum-space profile is typical for zero angular momentum orbitals and can describe the simplest case of a hydrogen atom and its lowest transition, 1s $\rightarrow$ 2s.
The next simplification of the UDW model involves assuming equal contributions from both of the nondiagonal parts of the Hamiltonian acting on the space spanned by the internal states of the atom.
In a general case, their relative weight can be unequal, but in the case of a spherical symmetry of both the ground and the excited states, they happen to be equal.
The other difference between QED and UDW Hamiltonians come from the fact that in the former the relative strength of para- and diamagnetic terms is given explicitly.
It is not the case in the latter, as it has to be computed for specific profiles of the ground and the excited states.
It can be done, however we will take the advantage of our model by considering a general, dimensionless parameter quantifying this relative strength.
Combining all of these considerations, we finally get the extended version of the UDW Hamiltonian that mimics the QED one:
\begin{equation}\label{ham}
\hat{H}_{\mathrm{I}}=\lambda\left(\ket{g}\bra{e}+\ket{e}\bra{g}\right)\hat{\phi}_{\mathrm{R}}+\alpha\frac{\lambda^2}{\Omega}\hat{\phi}_{\mathrm{R}}^2,
\end{equation}
where $\alpha$ is a dimensionless constant and a free parameter to tune the relative strength between para- and diamagnetic terms.
Energy $\Omega$ is introduced here to provide the correct units.
Such a Hamiltonian can effectively mimic some forms of the full QED interaction~\cite{Martin-Martinez2013a}.
It has to be noted that such a Hamiltonian is only a one-dimensional toy model that is utilized to model the qualitative effects coming from the full electromagnetic one.
By keeping free parameters $\lambda$ and $\alpha$ explicitly in the calculations, we will show that some interesting conclusions stay the same for their arbitrary values.
\subsection{Casimir-Polder potential}
The first step is to find how the full energy of the system changes with the position of the atom in the cavity.
The difference between this full energy, $E$, and the sum of the ground state energies of noninteracting cavity and the atom, $E_0$, is called the Casimir-Polder potential, $E_{\mathrm{CP}}$.
The usual Casimir-Polder force acting on the atom in the fixed cavity is then understood as a spatial derivative of the Casimir-Polder potential, $F=-\nabla E_{\mathrm{CP}}$.
We consider a system prepared in the state $\ket{g,0}$ -- the scalar field is in the vacuum state and two-level system is in the ground state.
The system is then slightly perturbed by the extended UDW Hamiltonian~\eqref{ham} with $\lambda$ being the perturbation parameter.
We will calculate the following energy in the second order of the perturbation theory.
It takes form:
\begin{eqnarray}
E&=&E_0+E^{(1)}+E^{(2)}+\mathcal{O}(\lambda^4), \nonumber \\
E^{(1)}&=&\bra{g,0}\hat{H}_{\mathrm{I}}\ket{g,0}, \nonumber \\
E^{(2)}&=&\sum_{n=0}^{\infty}\sum_{s=\{g,e\}}\frac{|\bra{s,n}\hat{H}_{\mathrm{I}}\ket{g,0}|^2}{E_0-(E_0+\omega_n+\Omega_s)},
\end{eqnarray}
where state $\ket{s,n}$, $s\in\{g,e\}$ corresponds to the arbitrary final state of the atom and the scalar field in the state $\ket{n}$ of energy $\omega_n$.
Furthermore, $\Omega_s$ is the energy of the detector in the state $\ket{s}$, meaning that $\Omega_g=0$ and $\Omega_e=\Omega$.
Then, we have:
\begin{eqnarray}
E^{(1)}&=&\frac{\alpha\lambda^2}{\Omega}\bra{0}\hat{\phi}_{\mathrm{R}}^2\ket{0}=\frac{\alpha\lambda^2}{\Omega L}\sum_{n=1}^{\infty}\frac{f_n^2 \sin^2{\left(k_n x\right)}}{\omega_n},\nonumber\\
E^{(2)}&=&-\sum_{n=1}^{\infty}\frac{\lambda^2}{\omega_n+\Omega}|\bra{n}\hat{\phi}_{\mathrm{R}}\ket{0}|^2+\mathcal{O}(\lambda^4)\nonumber\\
&=&-\sum_{n=1}^{\infty}\frac{\lambda^2}{\omega_n+\Omega}\frac{f_n^2 \sin^2{\left(k_n x\right)}}{\omega_n L}+\mathcal{O}(\lambda^4).\nonumber
\end{eqnarray}
It is useful to note that the term $f_n$ makes $E^{(1)}$ convergent.
The whole second-order Casimir-Polder potential then reads
\begin{align}
E_{\mathrm{CP}}&=E^{(1)}+E^{(2)} \nonumber \\
&=\lambda^2\sum_{n=1}^{\infty}\frac{f_n^2\sin^2{\left(k_n x\right)}}{\omega_n L \left(\omega_n+\Omega\right)}\left[\left(\alpha-1\right)+\alpha \frac{\omega_n}{\Omega}\right].
\label{eqn:ECP}
\end{align}
One can immediately see that depending on the parameter $\alpha$, the Casimir-Polder potential, and consequently Casimir-Polder force can be either positive or negative.
It confirms the usual phenomenology in which Casimir forces can be either repulsive or attractive, depending on the physical scenario involved.
\subsection{Probability of excitation}
The next step is to assume the same physical model but now with the interaction lasting for some finite time $\sigma$.
As the electromagnetic interaction cannot be switched on or off, we choose to interpret the finite interaction time as a time between the creation of a setup and a destructive measurement.
We aim to find the probability of measurement of the excited state of the atom, as it was initially prepared unexcited in the cavity.
Therefore, we have to define the time-dependent Hamiltonian of interaction, allowing for a finite time interaction.
We can modify previously showed model by adding a time-dependent switching function $\chi(t)$.
The modified, time-dependent version of the extended UDW Hamiltonian in the Schr\"odinger picture has the following form:
\begin{equation}
\hat{H}_{\mathrm{UDW}}(t)
=\chi(t)\left[\lambda~\hat{\mu}_{S} (t)~\hat{\phi}_{\mathrm{R}}(x)
+
\alpha \frac{\lambda^2}{\Omega}\left(\hat{\phi}_{\mathrm{R}}(x)\right)^2
\right].
\end{equation}
We assume that the interaction starts and ends rapidly, so that $\chi(t)=1$ for $t\in(0,\sigma)$ and $\chi(t)=0$ for any other time.
As it was mentioned before, the full Hamiltonian includes also a time-independent free scalar field and a free two-level system part.
We proceed to use the Dirac picture, because the full Hamiltonian contains time-independent $\hat{H}_0=\sum_n \omega_n \hat{a}_n^{\dagger}\hat{a}_n\otimes\Omega\hat{\sigma}^{+}\hat{\sigma}^{-}$ and a time-dependent interaction component
coming from the Unruh-DeWitt interaction.
The evolution in such a scenario is given by operator in the form: $\hat{U}
=\mathcal{T}\exp{
-i\int_{-\infty}^{\infty}
\mathrm{d}t\hat{H}_{\mathrm{I}}^{(D)}(t)
}$, where $(D)$ represents operator in the Dirac picture.
As a result of the interaction, the state of the field can be changed, however we are interested only in finding the probability of the detector's excitation.
The final state after the interaction between the detector and the scalar field can be written as $\ket{e,l}$, where $l\in\mathbb{N}\cup\{0\}$.
Using the Born rule, we can write probability of excitation $p_{g\xrightarrow{}e}$ in the form:
\begin{equation}
p_{g\xrightarrow{}e}=\sum_{l\in\mathbb{N}\cup\{0\}}|\bra{e,l}
\int_{-\infty}^{\infty} \mathrm{d}t
\hat{H}_{\mathrm{I}}^{(D)}(t) \ket{g,0}|^2.
\label{eqn:p2}
\end{equation}
The extended UDW Hamiltonian in the Dirac representation reads:
\begin{eqnarray} \label{ham2}
\hat{H}_{\mathrm{I}}^{(D)}(t)
=\chi(t)\left[\lambda~\hat{\mu}^{(D)}~\hat{\phi}_{\mathrm{R}}^{(D)}
+
\frac{\alpha\lambda^2}{\Omega}\left(\hat{\phi}_{\mathrm{R}}^{(D)}\right)^2
\right],
\end{eqnarray}
where:
\begin{eqnarray}
\hat{\mu}^{(D)}&=&\left(e^{i\Omega t} \hat{\sigma}^{+}+e^{-i\Omega t} \hat{\sigma}^{-}\right),\\
\hat{\phi}_{\mathrm{R}}^{(D)}(x)&=&\sum_{n}f_n\left[\hat{a}_n^{\dagger} u_{n}(x)e^{i\omega_n t}+H.c.\right].
\end{eqnarray}
Only the first part, linear in the coupling constant $\lambda$ contains an operator changing the state of the detector.
The second-order term does not contribute to the probability of excitation given by the equation
\eqref{eqn:p2}, because $\bra{e} \frac{\alpha\lambda^2}{\Omega}\left(\hat{\phi}_{\mathrm{R}}^{(D)}\right)^2\ket{g}=0$.
After some direct calculation, by plugging~\eqref{ham2} in~\eqref{eqn:p2}, we get:
\begin{equation}
p_{g\xrightarrow{}e}=4\lambda^2\sum_{n=1}^{\infty}\frac{f_n^2\sin^2{\left(k_n x\right)}}{\omega_n L}\frac{\sin^2{\left[\frac{1}{2}\sigma\left(\omega_n+\Omega\right)\right]}}{\left(\omega_n+\Omega\right)^2}.
\label{eqn:skonczoneprawd}
\end{equation}
The above result corresponds to an arbitrarily chosen time of interaction $\sigma$, but if the measurement apparatus does not have a time resolution good enough to work withing the scale of an atomic transition $1/\Omega$ one has to consider a coarse grained version of the above formula.
Such an averaged out excitation probability reads:
\begin{align}
p_{g\xrightarrow{}e}^{\text{av}}&=4\lambda^2\int \mathrm{d}\sigma \sum_{n=1}^{\infty}\frac{f_n^2\sin^2{\left(k_n x\right)}}{\omega_n L}\frac{\sin^2{\left[\frac{1}{2}\sigma\left(\omega_n+\Omega\right)\right]}}{\left(\omega_n+\Omega\right)^2} \nonumber \\
&=
2\lambda^2 \sum_{n=1}^{\infty}\frac{f_n^2\sin^2{\left(k_n x\right)}}{\omega_n L \left(\omega_n+\Omega\right)^2}
.
\label{eqn:pav}
\end{align}
This new quantity is no longer dependent on the interaction time and is an explicit function of a distance from the wall.
In the next Section we will show the connection between this averaged probability and the Casimir-Polder force.
\section{Retrieving Casimir-Polder potential from the average excitation rate}
We now proceed to compare both results.
Both energy and probability given by Eqs. \eqref{eqn:ECP} and \eqref{eqn:pav} are represented by infinite series.
We find that contributing terms in both of these series occur only for small (in comparison to the energy gap $\Omega$) values of $n$ , $\omega_n \ll \Omega$.
For a numerical analysis of that fact, see Appendix.
Therefore, we can treat $\frac{\omega_n}{\Omega}$ as a small parameter and expand both \eqref{eqn:ECP} and \eqref{eqn:pav} up to the first subleading order:
\begin{equation}
\label{cpeneeny}
E_{\mathrm{CP}}\approx \Omega \sum_{n} p_n (x) \left[ \left( \alpha-1\right)+\frac{\omega_n}{\Omega} \right]
\end{equation}
and
\begin{equation}
p_{g\xrightarrow{}e}^{\text{av}}\approx 2 \sum_{n} p_n (x) \left[1-2\frac{\omega_n}{\Omega} \right]
\end{equation}
where
\begin{equation}
p_n (x) = \frac{\lambda^2 f_n^2 \sin^2\left( k_n x \right) }{\omega_n L \Omega^2}.
\end{equation}
The universal function
\begin{equation}
F(x)=\sum_{n} p_n (x)
\end{equation}
that reproduces the general shape of both Casimir-Polder potential and the averaged excitation probability is shown at Fig.~\ref{fig1} together with its derivative, corresponding to the Casimir-Polder force.
\begin{figure}
\caption{(Top) Universal function $F(x)$ proportional to both Casimir-Polder potential and the average excitation probability for the atom at position $x/L$ in the cavity.
The parameters taken for the plot read: $L=1$, $m=1 \cdot 10^{-3}$, $\lambda =1 \cdot 10^{-2} $, $\Omega=1$.
(Bottom) The derivative of the universal function $F(x)$, corresponding to the Casimir-Polder force in the optical cavity.
}
\label{fig1}
\end{figure}
Up to the leading order in $\omega_n/\Omega$, we have a proportionality between the Casimir-Polder energy and the averaged excitation probability of the atom:
\begin{equation} \label{res}
E_{\mathrm{CP}} \approx \frac{1}{2} \Omega \left(\alpha - 1 \right) p_{g\xrightarrow{}e}^{\text{av}}.
\end{equation}
The proportionality constant is a function of the energy gap and the internal properties of the atom, implicitly contained in the parameter $\alpha$.
Unless $\alpha$ is extremely fine tuned to be 1, the proportionality between $E_{\mathrm{CP}}$ and $p_{g\xrightarrow{}e}^{\text{av}}$ is preserved.
One has to note that here is no realistic value of $\alpha$ for the three-dimensional electromagnetic model of interaction between atom an the field, as the presented one-dimensional toy model is aimed only at grasping qualitative effects present in the system.
However, within this toy model, taking $a_0=10^{-2}$ in natural units, $\alpha$ can be calculated to be $\alpha \sim 1/400$~\cite{Alhambra2014a}.
It shows that at least for hydrogen-like atoms in 1D toy model, value $\alpha \sim 1$ is not a typical one.
The result~\eqref{res} shows that Casimir-Polder potential can be indirectly recovered through analyzing the rate of atomic excitation near the wall.
However, the probability of exciting a single atom through the interaction with the vacuum fluctuations is extremely small.
It can be in some way adjusted by a proper choice of atomic species and of a proper atomic transition, however the most straightforward way is to use a large number of atoms that are placed within the same region.
The natural candidate for such an ensemble is a Bose-Einstein condensate placed near the wall of the optical cavity. Modern ultracold experiments provide a clean environment to probe such settings.
The first perk is the localization of the atoms by the means of optical trapping -- an ultracold sample can be placed in highly localized region of space (given by the usual Thomas-Fermi radius of BEC).
The second one is a fine control over the parameters of neutral atoms involved -- the size of the atomic cloud is tuned with the trapping and with the two-body interaction between the atoms, that can be effectively turned off by the means of Fesbach resonances.
Indeed, Bose-Einstein condensates have already been utilized to probe Casimir forces through their effect on the collective modes excited in the atomic cloud.
In Ref.~\cite{Obrecht2007} a BEC of rubidium was placed near the dielectric surface and the frequency of dipole mode was probed.
It allowed to measure nonzero temperature Casimir force acting on individual atoms.
We, however, propose a different scheme in which the Casimir force is not directly measured through the change of the collective motion, but rather the potential itself is analyzed through the excitation of atoms.
First, the BEC is treated as a collection of a large number of individual atoms that is spatially tightly packed and well described by the Thomas-Fermi profile and not as interacting many-body system.
The hydrogen-like atom assumption we have is a shortcoming, however condensates of such atoms have been produced.
We stress that our aim is not to perform the full three-dimensional calculation of the realistic QED system, but rather to show that at the level of a toy model mimicking the QED, there exists a direct relation between the Casimir-Polder energy and the excitation probability of an atom involved.
It is worth checking how the averaging over the typical density of BEC changes this relation.
The Thomas-Fermi single-particle density of BEC with perpendicular degrees of freedom integrated out and situated at $x_0$ from the boundary read:
\begin{equation} \label{bec}
n(x;x_0)=\frac{15 N}{16 R_{TF}} \left[1-\left(\frac{x-x_0}{R_{TF}} \right)^2 \right]^2,
\end{equation}
where $N$ -- total number of atoms in the condensate and $R_{TF}$ -- Thomas-Fermi radius.
If we follow the excitation scheme in which multiple single-shot experiments measuring population of the excited state (e.g. via in situ imaging) are performed, the averaged population of excited state takes the form:
\begin{equation} \label{ave}
N_{\text{exc}} (x)=\int p_{g\xrightarrow{}e}^{\text{av}}(u) n(u;x) \dd u.
\end{equation}
The effect of such an averaging, for different sizes of BECs is shown at Fig.~\ref{fig2}.
It appears that averaging over the density of BEC does not introduce appreciable changes to the spatial dependence of population of excited states in comparison to the Casimir-Polder potential.
\begin{figure}
\caption{Comparison between time-averaged excitation probability averaged over the density of atomic Bose-Einstein condensate for different Thomas-Fermi radii (red, dotted line -- $R_{TF} =0.05$, black, dashed line -- $R_{TF} =0.01$, orange, solid line -- $R_{TF} \rightarrow 0$).
The spatial averaging does not introduce appreciable quantitative difference, yielding same curve for each radius, corresponding to the Casimir-Polder potential.
Note that each curve starts at $R_{TF}$ due to the finite size of the BEC used as a probe.
The parameters taken for the plot read: $L=1$, $m=1 \cdot 10^{-3}$, $\lambda =1 \cdot 10^{-2} $, $\Omega=1$.
}
\label{fig2}
\end{figure}
\section{Recapitulation and outlook}
To summarize, we have analyzed a system consisting of an atom (described by a two-dimensional Hilbert space) interacting with a scalar field within a one-dimensional cavity.
We have argued that an extended version of the Unruh-DeWitt Hamiltonian coupled to the scalar Klein-Gordon field provides a qualitatively reasonable approximation to the full light-matter interaction when the vacuum state of the cavity is involved.
Utilizing the second-order perturbation theory, we have calculated the Casimir-Polder energy of the system and excitation probability of the atom when placed in a fixed distance from the wall of the cavity.
We have shown that up to the leading order, both of these quantities coincide with each other up to multiplicative constant depending on the internal structure of the atom.
As a result, we have argued that a suitable experimental setup involving measurement of the population of the excited atomic state in e.g. ultracold system of a two-level Bose-Einstein condensate, can be used to measure Casimir-Polder phenomena.
Moreover, we have checked how averaging over the BEC density affects the above-mentioned relation between CP energy and the excitation probability.
As a future line of work, a natural consequence would be a consideration a full three-dimensional system in a realistic experimental scenario.
Another potential research could involve studying applicability of the Untuh-DeWitt Hamiltonian in describing neutral atoms in optical cavities.
\section{Acknowledgments}
P.T.G. appreciates discussions with K. Rz\k{a}\.{z}ewski.
\begin{thebibliography}{29}
\makeatletter
\providecommand \@ifxundefined [1]{
\@ifx{#1\undefined}
}
\providecommand \@ifnum [1]{
\ifnum #1\expandafter \@firstoftwo
\else \expandafter \@secondoftwo
\fi
}
\providecommand \@ifx [1]{
\ifx #1\expandafter \@firstoftwo
\else \expandafter \@secondoftwo
\fi
}
\providecommand \natexlab [1]{#1}
\providecommand \enquote [1]{``#1''}
\providecommand \bibnamefont [1]{#1}
\providecommand \bibfnamefont [1]{#1}
\providecommand \citenamefont [1]{#1}
\providecommand \href@noop [0]{\@secondoftwo}
\providecommand \href [0]{\begingroup \@sanitize@url \@href}
\providecommand \@href[1]{\@@startlink{#1}\@@href}
\providecommand \@@href[1]{\endgroup#1\@@endlink}
\providecommand \@sanitize@url [0]{\catcode `\\12\catcode `\$12\catcode
`\&12\catcode `\#12\catcode `\^12\catcode `\_12\catcode `\%12\relax}
\providecommand \@@startlink[1]{}
\providecommand \@@endlink[0]{}
\providecommand \url [0]{\begingroup\@sanitize@url \@url }
\providecommand \@url [1]{\endgroup\@href {#1}{\urlprefix }}
\providecommand \urlprefix [0]{URL }
\providecommand \Eprint [0]{\href }
\providecommand \doibase [0]{http://dx.doi.org/}
\providecommand \selectlanguage [0]{\@gobble}
\providecommand \bibinfo [0]{\@secondoftwo}
\providecommand \bibfield [0]{\@secondoftwo}
\providecommand \translation [1]{[#1]}
\providecommand \BibitemOpen [0]{}
\providecommand \bibitemStop [0]{}
\providecommand \bibitemNoStop [0]{.\EOS\space}
\providecommand \EOS [0]{\spacefactor3000\relax}
\providecommand \BibitemShut [1]{\csname bibitem#1\endcsname}
\let\auto@bib@innerbib\@empty
\bibitem [{\citenamefont {Casimir}\ and\ \citenamefont
{Polder}(1948)}]{Casimir1948}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {H.~B.~G.}\
\bibnamefont {Casimir}}\ and\ \bibinfo {author} {\bibfnamefont
{D.}~\bibnamefont {Polder}},\ }\href {\doibase 10.1103/PhysRev.73.360}
{\bibfield {journal} {\bibinfo {journal} {Phys. Rev.}\ }\textbf {\bibinfo
{volume} {73}},\ \bibinfo {pages} {360} (\bibinfo {year} {1948})}\BibitemShut
{NoStop}
\bibitem [{\citenamefont {Sukenik}\ \emph {et~al.}(1993)\citenamefont
{Sukenik}, \citenamefont {Boshier}, \citenamefont {Cho}, \citenamefont
{Sandoghdar},\ and\ \citenamefont {Hinds}}]{Sukenik1993}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {C.~I.}\ \bibnamefont
{Sukenik}}, \bibinfo {author} {\bibfnamefont {M.~G.}\ \bibnamefont
{Boshier}}, \bibinfo {author} {\bibfnamefont {D.}~\bibnamefont {Cho}},
\bibinfo {author} {\bibfnamefont {V.}~\bibnamefont {Sandoghdar}}, \ and\
\bibinfo {author} {\bibfnamefont {E.~A.}\ \bibnamefont {Hinds}},\ }\href
{\doibase 10.1103/PhysRevLett.70.560} {\bibfield {journal} {\bibinfo
{journal} {Phys. Rev. Lett.}\ }\textbf {\bibinfo {volume} {70}},\ \bibinfo
{pages} {560} (\bibinfo {year} {1993})}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Lamoreaux}(1997)}]{Lamoreaux1997}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {S.~K.}\ \bibnamefont
{Lamoreaux}},\ }\href {\doibase 10.1103/PhysRevLett.78.5} {\bibfield
{journal} {\bibinfo {journal} {Phys. Rev. Lett.}\ }\textbf {\bibinfo
{volume} {78}},\ \bibinfo {pages} {5} (\bibinfo {year} {1997})}\BibitemShut
{NoStop}
\bibitem [{\citenamefont {Mohideen}\ and\ \citenamefont
{Roy}(1998)}]{Mohideen1998}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {U.}~\bibnamefont
{Mohideen}}\ and\ \bibinfo {author} {\bibfnamefont {A.}~\bibnamefont {Roy}},\
}\href {\doibase 10.1103/PhysRevLett.81.4549} {\bibfield {journal} {\bibinfo
{journal} {Phys. Rev. Lett.}\ }\textbf {\bibinfo {volume} {81}},\ \bibinfo
{pages} {4549} (\bibinfo {year} {1998})}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Schwinger}\ \emph {et~al.}(1978)\citenamefont
{Schwinger}, \citenamefont {DeRaad},\ and\ \citenamefont
{Milton}}]{Schwinger1978}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {J.}~\bibnamefont
{Schwinger}}, \bibinfo {author} {\bibfnamefont {L.~L.}\ \bibnamefont
{DeRaad}}, \ and\ \bibinfo {author} {\bibfnamefont {K.~A.}\ \bibnamefont
{Milton}},\ }\href {\doibase 10.1016/0003-4916(78)90172-0} {\bibfield
{journal} {\bibinfo {journal} {Ann. Phys. (N. Y).}\ }\textbf {\bibinfo
{volume} {115}},\ \bibinfo {pages} {1} (\bibinfo {year} {1978})}\BibitemShut
{NoStop}
\bibitem [{\citenamefont {Balian}\ and\ \citenamefont
{Duplantier}(1978)}]{Balian1978}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {R.}~\bibnamefont
{Balian}}\ and\ \bibinfo {author} {\bibfnamefont {B.}~\bibnamefont
{Duplantier}},\ }\href {\doibase 10.1016/0003-4916(78)90083-0} {\bibfield
{journal} {\bibinfo {journal} {Ann. Phys. (N. Y).}\ }\textbf {\bibinfo
{volume} {112}},\ \bibinfo {pages} {165} (\bibinfo {year}
{1978})}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Plunien}\ \emph {et~al.}(1986)\citenamefont
{Plunien}, \citenamefont {M{\"{u}}ller},\ and\ \citenamefont
{Greiner}}]{Plunien1986}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {G.}~\bibnamefont
{Plunien}}, \bibinfo {author} {\bibfnamefont {B.}~\bibnamefont
{M{\"{u}}ller}}, \ and\ \bibinfo {author} {\bibfnamefont {W.}~\bibnamefont
{Greiner}},\ }\href {\doibase 10.1016/0370-1573(86)90020-7} {\enquote
{\bibinfo {title} {{The Casimir effect}},}\ } (\bibinfo {year}
{1986})\BibitemShut {NoStop}
\bibitem [{\citenamefont {Bordag}\ \emph {et~al.}(2001)\citenamefont {Bordag},
\citenamefont {Mohideen},\ and\ \citenamefont {Mostepanenko}}]{Bordag2001}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {M.}~\bibnamefont
{Bordag}}, \bibinfo {author} {\bibfnamefont {U.}~\bibnamefont {Mohideen}}, \
and\ \bibinfo {author} {\bibfnamefont {V.~M.}\ \bibnamefont {Mostepanenko}},\
}\href {\doibase 10.1016/S0370-1573(01)00015-1} {\enquote {\bibinfo {title}
{{New developments in the Casimir effect}},}\ } (\bibinfo {year} {2001}),\
\Eprint {http://arxiv.org/abs/0106045} {arXiv:0106045 [quant-ph]}
\BibitemShut {NoStop}
\bibitem [{\citenamefont {Harber}\ \emph {et~al.}(2005)\citenamefont {Harber},
\citenamefont {Obrecht}, \citenamefont {McGuirk},\ and\ \citenamefont
{Cornell}}]{Harber2005a}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {D.~M.}\ \bibnamefont
{Harber}}, \bibinfo {author} {\bibfnamefont {J.~M.}\ \bibnamefont {Obrecht}},
\bibinfo {author} {\bibfnamefont {J.~M.}\ \bibnamefont {McGuirk}}, \ and\
\bibinfo {author} {\bibfnamefont {E.~A.}\ \bibnamefont {Cornell}},\ }\href
{\doibase 10.1103/PhysRevA.72.033610} {\bibfield {journal} {\bibinfo
{journal} {Phys. Rev. A - At. Mol. Opt. Phys.}\ }\textbf {\bibinfo {volume}
{72}} (\bibinfo {year} {2005}),\ 10.1103/PhysRevA.72.033610},\ \Eprint
{http://arxiv.org/abs/0506208} {arXiv:0506208 [cond-mat]} \BibitemShut
{NoStop}
\bibitem [{\citenamefont {Obrecht}\ \emph {et~al.}(2007)\citenamefont
{Obrecht}, \citenamefont {Wild}, \citenamefont {Antezza}, \citenamefont
{Pitaevskii}, \citenamefont {Stringari},\ and\ \citenamefont
{Cornell}}]{Obrecht2007}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {J.~M.}\ \bibnamefont
{Obrecht}}, \bibinfo {author} {\bibfnamefont {R.~J.}\ \bibnamefont {Wild}},
\bibinfo {author} {\bibfnamefont {M.}~\bibnamefont {Antezza}}, \bibinfo
{author} {\bibfnamefont {L.~P.}\ \bibnamefont {Pitaevskii}}, \bibinfo
{author} {\bibfnamefont {S.}~\bibnamefont {Stringari}}, \ and\ \bibinfo
{author} {\bibfnamefont {E.~A.}\ \bibnamefont {Cornell}},\ }\href {\doibase
10.1103/PhysRevLett.98.063201} {\bibfield {journal} {\bibinfo {journal}
{Phys. Rev. Lett.}\ }\textbf {\bibinfo {volume} {98}} (\bibinfo {year}
{2007}),\ 10.1103/PhysRevLett.98.063201},\ \Eprint
{http://arxiv.org/abs/0608074} {arXiv:0608074 [physics]} \BibitemShut
{NoStop}
\bibitem [{\citenamefont {Klimchitskaya}\ \emph {et~al.}(2009)\citenamefont
{Klimchitskaya}, \citenamefont {Mohideen},\ and\ \citenamefont
{Mostepanenko}}]{Klimchitskaya2009}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {G.~L.}\ \bibnamefont
{Klimchitskaya}}, \bibinfo {author} {\bibfnamefont {U.}~\bibnamefont
{Mohideen}}, \ and\ \bibinfo {author} {\bibfnamefont {V.~M.}\ \bibnamefont
{Mostepanenko}},\ }\href {\doibase 10.1103/RevModPhys.81.1827} {\bibfield
{journal} {\bibinfo {journal} {Rev. Mod. Phys.}\ }\textbf {\bibinfo {volume}
{81}},\ \bibinfo {pages} {1827} (\bibinfo {year} {2009})},\ \Eprint
{http://arxiv.org/abs/0902.4022} {arXiv:0902.4022} \BibitemShut {NoStop}
\bibitem [{\citenamefont {Dzyaloshinskii}\ \emph {et~al.}(1961)\citenamefont
{Dzyaloshinskii}, \citenamefont {Lifshitz},\ and\ \citenamefont
{Pitaevskii}}]{Dzyaloshinskii1961}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {I.~E.}\ \bibnamefont
{Dzyaloshinskii}}, \bibinfo {author} {\bibfnamefont {E.~M.}\ \bibnamefont
{Lifshitz}}, \ and\ \bibinfo {author} {\bibfnamefont {L.~P.}\ \bibnamefont
{Pitaevskii}},\ }\href {\doibase 10.1070/pu1961v004n02abeh003330} {\bibfield
{journal} {\bibinfo {journal} {Sov. Phys. Uspekhi}\ }\textbf {\bibinfo
{volume} {4}},\ \bibinfo {pages} {153} (\bibinfo {year} {1961})}\BibitemShut
{NoStop}
\bibitem [{\citenamefont {Boyer}(1974)}]{Boyer1974}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {T.~H.}\ \bibnamefont
{Boyer}},\ }\href {\doibase 10.1103/PhysRevA.9.2078} {\bibfield {journal}
{\bibinfo {journal} {Phys. Rev. A}\ }\textbf {\bibinfo {volume} {9}},\
\bibinfo {pages} {2078} (\bibinfo {year} {1974})}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Kenneth}\ \emph {et~al.}(2002)\citenamefont
{Kenneth}, \citenamefont {Klich}, \citenamefont {Mann},\ and\ \citenamefont
{Revzen}}]{Kenneth2002}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {O.}~\bibnamefont
{Kenneth}}, \bibinfo {author} {\bibfnamefont {I.}~\bibnamefont {Klich}},
\bibinfo {author} {\bibfnamefont {A.}~\bibnamefont {Mann}}, \ and\ \bibinfo
{author} {\bibfnamefont {M.}~\bibnamefont {Revzen}},\ }\href {\doibase
10.1103/PhysRevLett.89.033001} {\bibfield {journal} {\bibinfo {journal}
{Phys. Rev. Lett.}\ }\textbf {\bibinfo {volume} {89}} (\bibinfo {year}
{2002}),\ 10.1103/PhysRevLett.89.033001},\ \Eprint
{http://arxiv.org/abs/0202114} {arXiv:0202114 [quant-ph]} \BibitemShut
{NoStop}
\bibitem [{\citenamefont {Levin}\ \emph {et~al.}(2010)\citenamefont {Levin},
\citenamefont {McCauley}, \citenamefont {Rodriguez}, \citenamefont {Reid},\
and\ \citenamefont {Johnson}}]{Levin2010}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {M.}~\bibnamefont
{Levin}}, \bibinfo {author} {\bibfnamefont {A.~P.}\ \bibnamefont {McCauley}},
\bibinfo {author} {\bibfnamefont {A.~W.}\ \bibnamefont {Rodriguez}}, \bibinfo
{author} {\bibfnamefont {M.~T.}\ \bibnamefont {Reid}}, \ and\ \bibinfo
{author} {\bibfnamefont {S.~G.}\ \bibnamefont {Johnson}},\ }\href {\doibase
10.1103/PhysRevLett.105.090403} {\bibfield {journal} {\bibinfo {journal}
{Phys. Rev. Lett.}\ }\textbf {\bibinfo {volume} {105}} (\bibinfo {year}
{2010}),\ 10.1103/PhysRevLett.105.090403},\ \Eprint
{http://arxiv.org/abs/1003.3487} {arXiv:1003.3487} \BibitemShut {NoStop}
\bibitem [{\citenamefont {Munday}\ \emph {et~al.}(2009)\citenamefont {Munday},
\citenamefont {Capasso},\ and\ \citenamefont {Parsegian}}]{Munday2009}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {J.~N.}\ \bibnamefont
{Munday}}, \bibinfo {author} {\bibfnamefont {F.}~\bibnamefont {Capasso}}, \
and\ \bibinfo {author} {\bibfnamefont {V.~A.}\ \bibnamefont {Parsegian}},\
}\href {\doibase 10.1038/nature07610} {\bibfield {journal} {\bibinfo
{journal} {Nature}\ }\textbf {\bibinfo {volume} {457}},\ \bibinfo {pages}
{170} (\bibinfo {year} {2009})}\BibitemShut {NoStop}
\bibitem [{\citenamefont {DeWitt}(1979)}]{DeWitt1979}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {B.~S.}\ \bibnamefont
{DeWitt}},\ }in\ \href@noop {} {\emph {\bibinfo {booktitle} {Gen. Relativ. An
Einstein Centen. Surv.}}},\ \bibinfo {editor} {edited by\ \bibinfo {editor}
{\bibfnamefont {S.~W.}\ \bibnamefont {Hawking}}\ and\ \bibinfo {editor}
{\bibfnamefont {W.}~\bibnamefont {Israel}}}\ (\bibinfo {publisher}
{Cambridge University Press},\ \bibinfo {address} {Cambridge},\ \bibinfo
{year} {1979})\ pp.\ \bibinfo {pages} {680--745}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Alhambra}\ \emph
{et~al.}(2014{\natexlab{a}})\citenamefont {Alhambra}, \citenamefont {Kempf},\
and\ \citenamefont {Mart{\'{i}}n-Mart{\'{i}}nez}}]{Alhambra2014}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {{\'{A}}.~M.}\
\bibnamefont {Alhambra}}, \bibinfo {author} {\bibfnamefont {A.}~\bibnamefont
{Kempf}}, \ and\ \bibinfo {author} {\bibfnamefont {E.}~\bibnamefont
{Mart{\'{i}}n-Mart{\'{i}}nez}},\ }\href {\doibase 10.1103/PhysRevA.89.033835}
{\bibfield {journal} {\bibinfo {journal} {Phys. Rev. A - At. Mol. Opt.
Phys.}\ }\textbf {\bibinfo {volume} {89}} (\bibinfo {year}
{2014}{\natexlab{a}}),\ 10.1103/PhysRevA.89.033835}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Crispino}\ \emph {et~al.}(2008)\citenamefont
{Crispino}, \citenamefont {Higuchi},\ and\ \citenamefont
{Matsas}}]{Crispino2008}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {L.~C.~B.}\
\bibnamefont {Crispino}}, \bibinfo {author} {\bibfnamefont {A.}~\bibnamefont
{Higuchi}}, \ and\ \bibinfo {author} {\bibfnamefont {G.~E.~A.}\ \bibnamefont
{Matsas}},\ }\href {\doibase 10.1103/RevModPhys.80.787} {\bibfield {journal}
{\bibinfo {journal} {Rev. Mod. Phys.}\ }\textbf {\bibinfo {volume} {80}},\
\bibinfo {pages} {787} (\bibinfo {year} {2008})},\ \Eprint
{http://arxiv.org/abs/0710.5373} {arXiv:0710.5373} \BibitemShut {NoStop}
\bibitem [{\citenamefont {Birrell}\ and\ \citenamefont
{Davies}(1982)}]{Birrell1982}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {N.~D.}\ \bibnamefont
{Birrell}}\ and\ \bibinfo {author} {\bibfnamefont {P.~C.~W.}\ \bibnamefont
{Davies}},\ }\href
{https://books.google.pl/books/about/Quantum{\_}Fields{\_}in{\_}Curved{\_}Space.html?id=SEnaUnrqzrUC{\&}redir{\_}esc=y}
{\emph {\bibinfo {title} {{Quantum fields in curved space}}}}\ (\bibinfo
{publisher} {Cambridge University Press},\ \bibinfo {year} {1982})\ p.\
\bibinfo {pages} {340}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Louko}\ and\ \citenamefont {Satz}(2008)}]{Louko2008}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {J.}~\bibnamefont
{Louko}}\ and\ \bibinfo {author} {\bibfnamefont {A.}~\bibnamefont {Satz}},\
}\href {\doibase 10.1088/0264-9381/25/5/055012} {\bibfield {journal}
{\bibinfo {journal} {Class. Quantum Gravity}\ }\textbf {\bibinfo {volume}
{25}} (\bibinfo {year} {2008}),\ 10.1088/0264-9381/25/5/055012},\ \Eprint
{http://arxiv.org/abs/0710.5671} {arXiv:0710.5671} \BibitemShut {NoStop}
\bibitem [{\citenamefont {Hodgkinson}(2013)}]{Hodgkinson2013}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {L.}~\bibnamefont
{Hodgkinson}},\ }\emph {\bibinfo {title} {{Particle detectors in curved
spacetime quantum field theory}}},\ \href@noop {} {Ph.D. thesis} (\bibinfo
{year} {2013}),\ \Eprint {http://arxiv.org/abs/1309.7281v2}
{arXiv:1309.7281v2} \BibitemShut {NoStop}
\bibitem [{\citenamefont {Brown}\ \emph {et~al.}(2013)\citenamefont {Brown},
\citenamefont {Mart{\'{i}}n-Mart{\'{i}}nez}, \citenamefont {Menicucci},\ and\
\citenamefont {Mann}}]{Brown2013}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {E.~G.}\ \bibnamefont
{Brown}}, \bibinfo {author} {\bibfnamefont {E.}~\bibnamefont
{Mart{\'{i}}n-Mart{\'{i}}nez}}, \bibinfo {author} {\bibfnamefont {N.~C.}\
\bibnamefont {Menicucci}}, \ and\ \bibinfo {author} {\bibfnamefont {R.~B.}\
\bibnamefont {Mann}},\ }\href {\doibase 10.1103/PhysRevD.87.084062}
{\bibfield {journal} {\bibinfo {journal} {Phys. Rev. D - Part. Fields,
Gravit. Cosmol.}\ }\textbf {\bibinfo {volume} {87}} (\bibinfo {year}
{2013}),\ 10.1103/PhysRevD.87.084062}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Mart{\'{i}}n-Mart{\'{i}}nez}\ \emph
{et~al.}(2013)\citenamefont {Mart{\'{i}}n-Mart{\'{i}}nez}, \citenamefont
{Montero},\ and\ \citenamefont {{Del Rey}}}]{Martin-Martinez2013a}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {E.}~\bibnamefont
{Mart{\'{i}}n-Mart{\'{i}}nez}}, \bibinfo {author} {\bibfnamefont
{M.}~\bibnamefont {Montero}}, \ and\ \bibinfo {author} {\bibfnamefont
{M.}~\bibnamefont {{Del Rey}}},\ }\href {\doibase 10.1103/PhysRevD.87.064038}
{\bibfield {journal} {\bibinfo {journal} {Phys. Rev. D - Part. Fields,
Gravit. Cosmol.}\ }\textbf {\bibinfo {volume} {87}} (\bibinfo {year}
{2013}),\ 10.1103/PhysRevD.87.064038},\ \Eprint
{http://arxiv.org/abs/1207.3248} {arXiv:1207.3248} \BibitemShut {NoStop}
\bibitem [{\citenamefont {Passante}\ \emph {et~al.}(2003)\citenamefont
{Passante}, \citenamefont {Persico},\ and\ \citenamefont
{Rizzuto}}]{Passante2003}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {R.}~\bibnamefont
{Passante}}, \bibinfo {author} {\bibfnamefont {F.}~\bibnamefont {Persico}}, \
and\ \bibinfo {author} {\bibfnamefont {L.}~\bibnamefont {Rizzuto}},\ }\href
{\doibase 10.1016/S0375-9601(03)01131-9} {\bibfield {journal} {\bibinfo
{journal} {Phys. Lett. Sect. A Gen. At. Solid State Phys.}\ }\textbf
{\bibinfo {volume} {316}},\ \bibinfo {pages} {29} (\bibinfo {year}
{2003})}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Rizzuto}\ \emph {et~al.}(2004)\citenamefont
{Rizzuto}, \citenamefont {Passante},\ and\ \citenamefont
{Persico}}]{Rizzuto2004}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {L.}~\bibnamefont
{Rizzuto}}, \bibinfo {author} {\bibfnamefont {R.}~\bibnamefont {Passante}}, \
and\ \bibinfo {author} {\bibfnamefont {F.}~\bibnamefont {Persico}},\ }\href
{\doibase 10.1103/PhysRevA.70.012107} {\enquote {\bibinfo {title} {{Dynamical
casimir-polder energy between an excited- and a ground-state atom}},}\ }
(\bibinfo {year} {2004})\BibitemShut {NoStop}
\bibitem [{\citenamefont {Spagnolo}\ \emph {et~al.}(2006)\citenamefont
{Spagnolo}, \citenamefont {Passante},\ and\ \citenamefont
{Rizzuto}}]{Spagnolo2006}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {S.}~\bibnamefont
{Spagnolo}}, \bibinfo {author} {\bibfnamefont {R.}~\bibnamefont {Passante}},
\ and\ \bibinfo {author} {\bibfnamefont {L.}~\bibnamefont {Rizzuto}},\ }\href
{\doibase 10.1103/PhysRevA.73.062117} {\bibfield {journal} {\bibinfo
{journal} {Phys. Rev. A - At. Mol. Opt. Phys.}\ }\textbf {\bibinfo {volume}
{73}} (\bibinfo {year} {2006}),\ 10.1103/PhysRevA.73.062117}\BibitemShut
{NoStop}
\bibitem [{\citenamefont {Rizzuto}(2007)}]{Rizzuto2007}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {L.}~\bibnamefont
{Rizzuto}},\ }\href {\doibase 10.1103/PhysRevA.76.062114} {\bibfield
{journal} {\bibinfo {journal} {Phys. Rev. A - At. Mol. Opt. Phys.}\ }\textbf
{\bibinfo {volume} {76}} (\bibinfo {year} {2007}),\
10.1103/PhysRevA.76.062114}\BibitemShut {NoStop}
\bibitem [{\citenamefont {Alhambra}\ \emph
{et~al.}(2014{\natexlab{b}})\citenamefont {Alhambra}, \citenamefont {Kempf},\
and\ \citenamefont {Mart{\'{i}}n-Mart{\'{i}}nez}}]{Alhambra2014a}
\BibitemOpen
\bibfield {author} {\bibinfo {author} {\bibfnamefont {{\'{A}}.~M.}\
\bibnamefont {Alhambra}}, \bibinfo {author} {\bibfnamefont {A.}~\bibnamefont
{Kempf}}, \ and\ \bibinfo {author} {\bibfnamefont {E.}~\bibnamefont
{Mart{\'{i}}n-Mart{\'{i}}nez}},\ }\href {\doibase 10.1103/PhysRevA.89.033835}
{\bibfield {journal} {\bibinfo {journal} {Phys. Rev. A - At. Mol. Opt.
Phys.}\ }\textbf {\bibinfo {volume} {89}} (\bibinfo {year}
{2014}{\natexlab{b}}),\ 10.1103/PhysRevA.89.033835}\BibitemShut {NoStop} \end{thebibliography}
\appendix
\section{Low frequency approximation}
In this Appendix we demonstrate that the infinite series given by the \eqref{eqn:ECP} and \eqref{eqn:pav}, can be well approximated by a sum of finite number of terms.
We will include only the lowest modes of the field to the sum, so it can be called a low frequency approximation.
Furthermore, we want to show that the frequency of the last mode included in the approximation sum always stays much smaller than the energy gap $\Omega$.
\begin{figure}
\caption{Fidelity of the finite elements approximation used to find a value of Casimir-Polder potential or probability of excitation for a detector standing in the middle of the cavity.
The parameters taken for the plot read: $x=\frac{L}{2},~L=1,~m=1\cdot10^{-3},~\lambda=1\cdot10^{-2}$ (in the case of finding Casimir-Polder potential there is also $\alpha$ coupling constant).
Fidelity of:
(Top) probability of excitation,
(Center) Casimir-Polder potential for $\alpha=1\cdot10^{-1}$,
(Bottom) Casimir-Polder potential for $\alpha=2$.
}
\label{fig3}
\end{figure}
Let us start from the most general case.
Let us cosider a converged infinite sum $S=\sum_{n=1}^{\infty}a_n$.
The value of $S$ can be approximated by the $S_N=\sum_{n=1}^{N}a_n$.
The bigger $N$ is, the better approximation of the infinite series $S$ we get.
We can ask how many elements of the series need to be summed up to achieve a given quality of the approximation.
To well define this problem we have to determine how to measure the quality of the approximation of the series
One of the possibility is to use a fidelity function defined as:
\begin{equation}\label{fid}
\mathcal{F}_{a_n}^{N}=\frac{a_{N}}{\sum_{n=1}^{N}a_n}.
\end{equation}
Such a function tells us how big contribution to finite sum $S_N$ coming from the last element is.
The smaller $\mathcal{F}_{a_n}^{N}$ is, the better the quality of $S$ approximation given by $S_N$ becomes.
In the case presented above, we want to verify whether series which define Casimir-Polder potential and probability of excitation can be approximated by a sum including just $N$ elements such that $\omega_N$ is still much smaller than $\Omega$.
Using fidelity function~\eqref{fid} we can find the value of the function $\mathcal{F}_{E_n(\Omega)}^{N}$, where $E_n$ is such that $E_{\mathrm{CP}}=\sum_{n=1}^{\infty}E_n$.
To answer our question we can plot $\mathcal{F}_{E_n(\Omega=\omega_K)}^{N}$ as a function of $N$ and $K$ such that $\omega_K=\Omega$.
Similarly, for the probability of excitation we will plot $\mathcal{F}_{(p_{g\xrightarrow{}e})_n(\Omega)}^{N}$, where $(p_{g\xrightarrow{}e})_n$ is such that $p_{g\xrightarrow{}e}=\sum_{n}(p_{g\xrightarrow{}e})_n$.
For simplicity, we will consider only a detector standing in the middle of the cavity.
As a result, only odd modes of the field have non zero contribution to the final value.
The figure \ref{fig3} shows fidelity of the approximated series-defined Casimir-Polder potential and probability of excitation.
We can see that the lines connecting the points of the same value of the fidelity function have a convex shape.
It turns out that for every parameter describing the quality of the approximation by a $N$ elements sum, we can choose $\Omega$ for which $\omega_N\ll\Omega$.
For instance, let us consider the line of constant value of fidelity shown on the top of figure \ref{fig3}.
The same value of a fidelity occurs for pair $(N,K)\approx(7,7)$ and for $(N,K)\approx(12,20)$.
It means that for $\Omega=\omega_{2\cdot7+1}$ one has to sum $N=2\cdot7+1$ modes of the field to achieve the same fidelity as for $\Omega=\omega_{2\cdot20+1}$ and only $N=2\cdot12+1$ modes.
For series-defined Casimir-Polder potential and coupling $\alpha>1$ it is even better, because lines of constant fidelity decrease with $K$, so the bigger $\Omega=\omega_{2K+1}$ is, the smaller number of modes that are needed to be summed up to achieve given fidelity is.
\end{document}
|
arXiv
|
{
"id": "1912.08522.tex",
"language_detection_score": 0.6780110597610474,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[Stationary Nonlocal Cahn-Hilliard-Navier-Stokes System]{On the Stationary Nonlocal Cahn-Hilliard-Navier-Stokes System: Existence, Uniqueness and Exponential Stability \Addresses }
\author[T. Biswas, S. Dharmatti, L. N. M. Perisetti and M. T. Mohan]
{Tania Biswas\textsuperscript{1},
Sheetal Dharmatti\textsuperscript{1*},
Perisetti Lakshmi Naga Mahendranath\textsuperscript{1}
and Manil T Mohan\textsuperscript{2}}
\begin{abstract}
Cahn-Hilliard-Navier-Stokes system describes the evolution of two isothermal, incompressible, immiscible fluids in a bounded domain. In this work, we consider the stationary nonlocal Cahn-Hilliard-Navier-Stokes system in two and three dimensions with singular potential. We prove the existence of a weak solution for the system using pseudo-monotonicity arguments and Browder's theorem. Further, we establish the uniqueness and regularity results for the weak solution of the stationary nonlocal Cahn-Hilliard-Navier-Stokes system for constant mobility parameter and viscosity. Finally, in two dimensions, we establish that the stationary solution is exponentially stable (for convex singular potentials) under suitable conditions on mobility parameter and viscosity.
\end{abstract}
\maketitle
\section{Introduction}\label{sec1}\setcounter{equation}{0}
We consider a mathematical model of two isothermal, incompressible, immiscible fluids evolving in two or three dimensional bounded domains. This system of equations is well known as \emph{Cahn-Hilliard-Navier-Stokes(CHNS) system} or is also known as $\mathrm{H}$-Model. Cahn-Hilliard-Navier-Stokes model describes the chemical interactions between the two phases at the interface, which is achieved using a Cahn-Hilliard approach, and also the hydrodynamic properties of the mixture which is obtained by using Navier-Stokes equations with surface tension terms acting at the interface (see \cite{MR2580516}). If the two fluids have the same constant density, then the temperature differences are negligible and the diffusive interface between the two phases has a small but non-zero thickness, and thus we have the well-known ``H-Model" (see \cite{MR1404829}). The equations for evolution of the Cahn-Hilliard-Navier-Stokes/H-model are given by
\begin{equation}\label{1.1}
\mathopen{}\mathclose\bgroup\originalleft\{
\begin{aligned}
\varphi_t + \mathbf{u}\cdot \nabla \varphi &=\text{ div} (m(\varphi) \nabla \mu), \ \text{ in } \ \Omega \times (0,T),\\
\mu &= a \varphi - \mathrm{J}\ast \varphi + \mathrm{F}'(\varphi),\\
\rho \mathbf{u}_t - 2 \text{div } ( \nu(\varphi) \mathrm{D}\mathbf{u} ) + (\mathbf{u}\cdot \nabla )\mathbf{u} + \nabla \uppi &= \mu \nabla \varphi + \mathbf{h}, \ \text{ in } \ \Omega \times (0,T),\\
\text{div }\mathbf{u}&= 0, \ \text{ in } \ \Omega \times (0,T), \\
\frac{\partial \mu}{\partial\mathbf{n}} &= 0 \ , \mathbf{u}=0 \ \text{ on } \ \partial \Omega \times [0,T],\\
\mathbf{u}(0) &= \mathbf{u}_0, \ \ \varphi(0) = \varphi _0 \ \text{ in } \ \Omega,
\end{aligned}
\aftergroup\egroup\originalright.
\end{equation}
where $\Omega\subset \mathbb{R}^n,\ n=2,3$ and $\mathbf{u}(x,t)$ and $\varphi(x,t)$ denote the average velocity of the fluid and the relative concentration respectively. These equations are of the nonlocal type because of the presence of the term $\mathrm{J}$, which is the \emph{spatial-dependent internal kernel} and $\mathrm{J} \ast \varphi$ denotes the spatial convolution over $\Omega$. The mobility parameter is denoted by
$m$, $\mu$ is the \emph{chemical potential}, $\uppi$ is the \emph{pressure}, $a$ is defined by $a(x) := \int _\Omega \mathrm{J}(x-y) \/\mathrm{d}\/ y$, $\mathrm{F}$ is the configuration potential, which accounts for the presence of two phases, $\nu$ is the \emph{kinematic viscosity} and $\mathbf{h}$ is the external forcing term acting in the mixture. The strain tensor $\mathrm{D}\mathbf{u}$ is the symmetric part of the gradient of the flow velocity vector, i.e., $\mathrm{D}\mathbf{u}$ is $\frac{1}{2}\mathopen{}\mathclose\bgroup\originalleft(\nabla\mathbf{u}+(\nabla\mathbf{u})^{\top}\aftergroup\egroup\originalright)$. The chemical potential $\mu$ is the first variation of the free energy functional:
\begin{align*}
\mathcal{E}(\varphi) := \frac{1}{4}\int_\Omega \int_\Omega \mathrm{J}(x-y)(\varphi(x)-\varphi(y))^2\/\mathrm{d}\/ x \/\mathrm{d}\/ y + \int_\Omega \mathrm{F}(\varphi(x))\/\mathrm{d}\/ x.
\end{align*}
Various simplified models of this system are studied by several mathematicians and physicists. The local version of the system (see \cite{MR1700669, MR2580516}) is obtained by replacing $\mu $ equation by
$ \mu =- \Delta \varphi + \mathrm{F}' (\varphi) $, which is the first variation of the free energy functional
$$\mathcal{E}(\varphi) := \int_{\Omega} \mathopen{}\mathclose\bgroup\originalleft( \frac{1}{2} | \nabla \varphi(x)|^2 + \mathrm{F}(\varphi (x))\aftergroup\egroup\originalright)\, \/\mathrm{d}\/ x.$$
Another simplification appeared in the literature is to assume the constant mobility parameter and/or constant viscosity. The solvability, uniqueness and regularity of the system (\ref{1.1}) and of other simplified Cahn-Hilliard-Navier-Stokes models is well studied in the literature though most of the works are recent ones. Typically two types of potentials are considered in the literature, regular potential as well as singular potential. In general, singular potentials are difficult to handle and in such cases, $\mathrm{F}(\cdot)$ is usually approximated by polynomials in order to make the mathematical analysis easier (see \cite{MR2347608}).
The nonlocal Cahn-Hilliard-Navier-Stokes system with regular potential has been analysed by M. Grasseli et al. in \cite{MR2834896, MR3518604, MR3090070}, etc. Taking advantage of the results for the regular potential, they have also studied in \cite{MR3019479}, the existence of weak solution of the system with singular potential. Furthermore, they proved the existence of the global attractor in 2D and trajectory attractor in 3D. Strong solutions for the nonlocal Cahn-Hilliard-Navier-Stokes system was discussed in \cite{MR3903266}. Uniqueness results for the same were established in \cite{MR3518604}. In \cite{MR3688414}, authors considered the nonlocal Cahn-Hilliard equation with singular potential and constant mobility and studied well-posedness and regularity results. Moreover, they established the strict separation property in dimension 2. Regularity results in case of degenerate mobility were studied in \cite{frigeri10regularity}. The local Cahn-Hilliard-Navier-Stokes system with singular free energies has been studied in \cite{MR2563636, MR1700669}. Further, along the application side, the optimal control of nonlocal Cahn-Hilliard-Navier-Stokes equations and robust control for local Cahn-Hilliard-Navier-Stokes equations have been addressed in the works \cite{MR3456388,MR4104524,MR3565933,MR3436705, MR4108622,MR4131779}, etc.
Solvability results for the stationary nonlocal Cahn-Hilliard equations with singular potential were discussed in \cite{MR2108884}, whereas authors in \cite{MR2347608} proved the convergence to the equilibrium solution of Cahn-Hilliard system with logarithmic free energy. The existence of the equilibrium solution for the steady state Navier-Stokes equation is well known in the literature and can be found in the book \cite{MR1318914}. In \cite{MR3524178}, the authors discussed the existence of a weak solution to the stationary local Cahn-Hilliard-Navier-Stokes equations. The author in \cite{MR3555135} studied a coupled Cahn-Hilliard-Navier-Stokes model with delays in a two-dimensional bounded domains and discussed the asymptotic behaviour of the weak solutions and the stability of the stationary solutions. In this work, our main aim is to study the well-posedness of nonlocal steady state system corresponding to the model described in \eqref{1.1} in dimensions $2$ and $3$ and to examine the stability properties of this solution in dimension $2$ (for convex singular potentials).
Throughout this paper, we consider $\mathrm{F}$ to be a singular potential. A typical example is the logarithmic potential:
\begin{align}\label{2}
\mathrm{F} (\varphi) = \frac{\theta}{2} ((1+ \varphi) \ln (1+\varphi) + (1-\varphi) \ln (1-\varphi)) - \frac{\theta_c}{2} \varphi^2 ,\quad \varphi \in (-1,1),
\end{align}
where $\theta,\theta_c>0$. The logarithmic terms are related to the entropy of the system and note that $\mathrm{F}$ is convex if and only if $\theta\geq \theta_c$. If $\theta\geq \theta_c$ the mixed phase is stable and if $0<\theta<\theta_c$, the mixed phase is unstable and phase separation occurs.
To the best of our knowledge, the solvability results for stationary nonlocal Cahn-Hilliard-Navier-Stokes equations is not available in the literature. In the current work, using techniques similar to the one developed in \cite{MR3524178}, we resolve this issue.
We prove the existence of a weak solution to the stationary nonlocal Cahn-Hilliard-Navier-Stokes system in dimensions 2 and 3 with variable mobility and viscosity. Further, we answer the questions of uniqueness and regularity of the solution for the equations with constant viscosity and mobility parameters. In dimensions $2$ and $3$, we show that the weak solution possesses higher regularity. The uniqueness of weak solutions is established under certain conditions on the viscosity and mobility parameters. Lastly, for constant viscosity and mobility parameters and convex singular potential, we establish that the strong solution of steady state equations in dimension $2$, stabilizes exponentially.
The main difficulty to tackle while obtaining these results was to handle the nonlocal term. The nonlocal term in the equation needs careful estimation. Moreover, to the best of our knowledge, apart from existence and regularity results for steady state equation, which we have obtained here is the first work to discuss exponential stability of nonlocal Cahn-Hilliard-Navier-Stokes model. These results can be useful to study stabilisation properties of control problems associated with the system.
Rest of the paper is organised as follows: In the next section, we explain functional setting for the solvability of stationary nonlocal Cahn-Hilliard-Navier-Stokes equations \eqref{steadysys} (given below). We define the weak formulation of our system in section \ref{sec3}. The existence of a weak solution to the nonlocal Cahn-Hilliard-Navier-Stokes equations \eqref{steadysys} is proved using pseudo-monotonicity arguments and Browder's theorem in this section (see Theorem \ref{mainthm}). In further studies, we assume the mobility parameter and viscosity to be constant. The section \ref{sec4} is devoted to study the uniqueness of a weak solution for the system \eqref{steadysys} and some regularity results. We establish the uniqueness of weak solutions under certain assumptions on mobility parameter and viscosity (see Theorem \ref{unique-steady}). Further, we derive some regularity results for the solution. Finally, in section \ref{se4}, we establish that the stationary solution in two dimensions is exponentially stable (see Theorem \ref{thmexp}) under certain restrictions on mobility parameter and viscosity.
\section{Stationary Nonlocal Cahn-Hilliard-Navier-Stokes System}\label{se3}\setcounter{equation}{0}
In this section, we consider the stationary nonlocal Cahn-Hilliard-Navier-Stokes system in two and three dimensional bounded domains. Here, we consider the case of the coefficient of kinematic viscosity and mobility parameter depending on $\varphi$. Let us consider the following steady state system associated with the equation \eqref{1.1}:
\begin{equation}\label{steadysys}
\mathopen{}\mathclose\bgroup\originalleft\{
\begin{aligned}
\mathbf{u}\cdot \nabla \varphi &= \text{ div} (m(\varphi) \nabla \mu), \ \text{ in }\ \Omega,\\
\mu &= a \varphi - \mathrm{J}\ast \varphi + \mathrm{F}'(\varphi), \ \text{ in }\ \Omega,\\
- 2 \text{div } ( \nu(\varphi) \mathrm{D}\mathbf{u} ) + (\mathbf{u}\cdot \nabla )\mathbf{u} + \nabla \uppi &= \mu \nabla \varphi + \mathbf{h}, \ \text{ in }\ \Omega,\\
\text{div }\mathbf{u}&= 0, \ \text{ in }\ \Omega, \\
\frac{\partial \mu}{\partial\mathbf{n}} &= 0 \ , \mathbf{u}=\mathbf{0} \ \text{ on } \ \partial \Omega,\\
\end{aligned}
\aftergroup\egroup\originalright.
\end{equation}
with average of $\varphi$ given by $$\frac{1}{|\Omega|}\int_\Omega \varphi (x) \/\mathrm{d}\/ x = k \in (-1,1),$$ where $|\Omega|$ is the Lebesgue measure of $\Omega$.
Our main aim in this work is to study the existence, uniqueness, regularity and stability of the system \eqref{steadysys}. For solvability, we formulate the problem in an abstract setup and use the well known Browder's theorem to establish the existence of a weak solution to the system \eqref{steadysys}. We further study regularity, uniqueness and exponential stability of the system with constant viscosity and mobility parameters by establishing a-priori estimates and under certain conditions on viscosity and mobility.
\subsection{Functional setting} We first explain the functional spaces needed to obtain our main results. Let us define
\begin{align*}
\mathbb{G}_{\mathrm{div}} &:= \Big\{ \mathbf{u} \in \mathrm{L}^2(\Omega;\mathbb{R}^n) : \text{div }\mathbf{u}=0,\ \mathbf{u}\cdot \mathbf{n} \big|_{\partial \Omega}=0 \Big\}, \\
\mathbb{V}_{\mathrm{div}} &:= \Big\{\mathbf{u} \in \mathrm{H}^1_0(\Omega;\mathbb{R}^n): \text{div }\mathbf{u}=0\Big\},\\ \mathrm{H}&:=\mathrm{L}^2(\Omega;\mathbb{R}),\ \mathrm{V}:=\mathrm{H}^1(\Omega;\mathbb{R}),
\end{align*}
where $n=2,3$. Let us denote $\| \cdot \|$ and $(\cdot, \cdot)$, the norm and the scalar product, respectively, on both $\mathrm{H}$ and $\mathbb{G}_{\mathrm{div}}$. The duality between any Hilbert space $\mathbb{X}$ and its topological dual $\mathbb{X}'$ is denoted by $_{\mathbb{X}'}\langle \cdot,\cdot\rangle_{\mathbb{X}}$. We know that $\mathbb{V}_{\mathrm{div}}$ is endowed with the scalar product
$$(\mathbf{u},\mathbf{v})_{\mathbb{V}_{\mathrm{div}} }= (\nabla \mathbf{u}, \nabla \mathbf{v})=2(\mathrm{D}\mathbf{u},\mathrm{D}\mathbf{v}),\ \text{ for all }\ \mathbf{u},\mathbf{v}\in\mathbb{V}_{\mathrm{div}}.$$ The norm on $\mathbb{V}_{\mathrm{div}}$ is given by $\|\mathbf{u}\|_{\mathbb{V}_{\mathrm{div}}}^2:=\int_{\Omega}|\nabla\mathbf{u}(x)|^2\/\mathrm{d}\/ x=\|\nabla\mathbf{u}\|^2$. In the sequel, we use the notations $\mathbb{H}^m(\Omega):=\mathrm{H}^m(\Omega;\mathbb{R}^n)=\mathrm{W}^{m,2}(\Omega;\mathbb{R}^n)$ and $\mathrm{H}^m(\Omega):=\mathrm{H}^m(\Omega;\mathbb{R})=\mathrm{W}^{m,2}(\Omega;\mathbb{R})$ for Sobolev spaces of order $m$.
Let us also define
\begin{align*}
\mathrm{L}^2_{(k)}(\Omega)& := \mathopen{}\mathclose\bgroup\originalleft\{ f \in \mathrm{L}^2(\Omega;\mathbb{R}) : \frac{1}{|\Omega |} \int_\Omega f(x)\/\mathrm{d}\/ x = k \aftergroup\egroup\originalright\},\\
\mathrm{H}^1_{(0)} (\Omega)&:= \mathrm{H}^1(\Omega;\mathbb{R}) \cap\mathrm{L}^2_{(0)}(\Omega)= \mathopen{}\mathclose\bgroup\originalleft\{ f \in \mathrm{H}^1(\Omega;\mathbb{R}) : \int_\Omega f(x)\/\mathrm{d}\/ x = 0 \aftergroup\egroup\originalright\}\ ,\\ \mathrm{H}^{-1}_{(0)}(\Omega)&:=\mathrm{H}_{(0)}^1(\Omega)'.
\end{align*}
Note that $\mathrm{L}^2_{(0)}(\Omega)$ is a Hilbert space equipped with the usual inner product in $\mathrm{L}^2(\Omega)$.
Since $\Omega$ is a bounded smooth domain and the average of $f$ is zero in $\mathrm{H}^1_{(0)} (\Omega)$, using the Poincar\'e-Wirtinger inequality (see Lemma \ref{poin} below), we have $\|f\|\leq C_{\Omega}\|\nabla f\|$, for all $f\in \mathrm{H}^1_{(0)} (\Omega)$. Using this fact, one can also show that $\mathrm{H}^1_{(0)} (\Omega)$ is a Hilbert space equipped with the inner product
\begin{align*}
(\varphi ,\psi)_{\mathrm{H}^1_{(0)}} = (\nabla \varphi, \nabla \psi), \ \text{ for all }\ \varphi, \psi \in \mathrm{H}^1_{(0)}(\Omega) .
\end{align*}
We can prove the following dense and continuous embedding:
\begin{align} \label{zeroembedding}
\mathrm{H}^1_{(0)} (\Omega)\hookrightarrow \mathrm{L}^2_{(0)}(\Omega)\equiv \mathrm{L}^2_{(0)}(\Omega)' \hookrightarrow \mathrm{H}^{-1}_{(0)}(\Omega).
\end{align}
Note that the embedding is compact (see for example, Theorem 1, Chapter 5, \cite{MR2597943}). The projection $\mathrm{P}_0 : \mathrm{L}^2(\Omega) \rightarrow \mathrm{L}^2_{(0)}(\Omega)$ onto $\mathrm{L}^2$-space with mean value zero is defined by
\begin{align} \label{P0}
\mathrm{P}_0 f := f - \frac{1}{|\Omega |} \int_\Omega f(x)\/\mathrm{d}\/ x
\end{align}
for every $f \in \mathrm{L}^2(\Omega)$.
For every $f \in \mathrm{V}'$, we denote $\overline{f}$ the average of $f$ over $\Omega$, that is, $\overline{f} := |\Omega|^{-1} {_{\mathrm{V}'}}\langle f, 1 \rangle_{\mathrm{V}}$.
Let us also introduce the spaces (see \cite{MR3518604} for more details)
\begin{align*}\mathrm{V}_0 &=\mathrm{H}_{(0)}^{1}(\Omega)= \{ v \in \mathrm{V} : \overline{v} = 0 \},\\
\mathrm{V}_0' &=\mathrm{H}_{(0)}^{-1}(\Omega)= \{ f \in \mathrm{V}' : \overline{f} = 0 \},\end{align*}
and the operator $\mathcal{A} : \mathrm{V} \rightarrow \mathrm{V}'$ is defined by
\begin{align*}\,_{\mathrm{V}'}\langle \mathcal{A} u ,v \rangle_{\mathrm{V}} := \int_\Omega \nabla u(x) \cdot \nabla v(x) \/\mathrm{d}\/ x, \ \text{for all } \ u,v \in \mathrm{V}.\end{align*}
Clearly $\mathcal{A}$ is linear and it maps $\mathrm{V}$ into $\mathrm{V}_0'$ and its restriction $\mathcal{B}$ to $\mathrm{V}_0$ onto $\mathrm{V}_0'$ is an isomorphism.
We know that for every $f \in \mathrm{V}_0'$, $\mathcal{B}^{-1}f$ is the unique solution with zero mean value of the \emph{Neumann problem}:
$$
\mathopen{}\mathclose\bgroup\originalleft\{
\begin{array}{ll}
- \Delta u = f, \ \mbox{ in } \ \Omega, \\
\frac{\partial u}{\partial\mathbf{n}} = 0, \ \mbox{ on } \ \partial \Omega.
\end{array}
\aftergroup\egroup\originalright.
$$
In addition, we have
\begin{align} \!_{\mathrm{V}'}\langle \mathcal{A}u , \mathcal{B}^{-1}f \rangle_{\mathrm{V}} &= \!_{\mathrm{V}'}\langle f ,u \rangle_{\mathrm{V}}, \ \text{ for all } \ u\in \mathrm{V}, \ f \in \mathrm{V}_0' , \label{bes}\\
\!_{\mathrm{V}'}\langle f , \mathcal{B}^{-1}g \rangle_{\mathrm{V}} &= \!_{\mathrm{V}'}\langle g ,\mathcal{B}^{-1}f \rangle_{\mathrm{V}} = \int_\Omega \nabla(\mathcal{B}^{-1}f)\cdot \nabla(\mathcal{B}^{-1}g)\/\mathrm{d}\/ x, \ \text{for all } \ f,g \in \mathrm{V}_0'.\label{bes1}
\end{align}
Note that $\mathcal{B}$ can be also viewed as an unbounded linear operator on $\mathrm{H}$ with
domain $\mathrm{D}(\mathcal{B}) = \mathopen{}\mathclose\bgroup\originalleft\{v \in \mathrm{H}^2(\Omega)\cap \mathrm{V}_0 : \frac{\partial v}{\partial\mathbf{n}}= 0\text{ on }\partial\Omega \aftergroup\egroup\originalright\}$.
Below, we give some facts about the elliptic regularity theory of Laplace operator $ B_N = -\Delta + \mathrm{I} $ with Neumann boundary conditions.
\begin{lemma}[$ L^p $ regularity for Neumann Laplacian, \cite{SJL_1961-1962____A6_0}, Theorem 9.26, \cite{MR2759829}] \label{Lp_reg}
Let us assume that $ u $ satisfies $ B_N u = f$ and $ \frac{\partial u}{\partial n} =0$ in weak sense. Then the following holds:
\begin{itemize}
\item [(i)] Let $ f \in (W^{1,p'}(\Omega))'$, with $ 1<p'<\infty $. Then $ u \in W^{1,p}(\Omega) $, where $ \frac{1}{p}+\frac{1}{p'}=1 $ and there exists a constant $ C>0 $ such that
\begin{align*}
\|u\|_{W^{1,p}(\Omega)} \leq C \|f\|_{(W^{1,p'}(\Omega))'}.
\end{align*}
\item [(ii)] Let $f \in L^p(\Omega)$, with $1 < p < \infty$. Then, $u \in W^{2,p} (\Omega), -\Delta u + u = f$ for a.e. $x \in \Omega, \frac{\partial u}{\partial n}= 0$ on $\partial \Omega$ in the sense of traces and there exists $C > 0$ such that
\begin{align*}
\|u\|_{W^{2,p}(\Omega)} \leq C \|f\|_{L^p(\Omega)}.
\end{align*}
\end{itemize}
\end{lemma}
\subsection{Linear and non-linear operators}
Let us define the Stokes operator $\mathrm{A} : \mathrm{D}(\mathrm{A})\cap \mathbb{G}_{\mathrm{div}} \to \mathbb{G}_{\mathrm{div}}$. In the case of no slip boundary condition
$$\mathrm{A}=-\mathrm{P}\Delta,\ \mathrm{D}(\mathrm{A})=\mathbb{H}^2(\Omega) \cap \mathbb{V}_{\mathrm{div}},$$ where $\mathrm{P} : \mathbb{L}^2(\Omega) \to \mathbb{G}_{\mathrm{div}}$ is the \emph{Helmholtz-Hodge orthogonal projection}. Note also that, we have
$$\!_{\mathbb{V}_{\mathrm{div}}'}\langle\mathrm{A}\mathbf{u}, \mathbf{v}\rangle_{\mathbb{V}_{\mathrm{div}}} = (\mathbf{u}, \mathbf{v})_{\mathbb{V}\text{div}} = (\nabla\mathbf{u}, \nabla\mathbf{v}), \text{ for all } \ \mathbf{u}, \mathbf{v} \in \mathbb{V}_{\mathrm{div}}.$$
It should also be noted that $\mathrm{A}^{-1} : \mathbb{G}_{\mathrm{div}} \to \mathbb{G}_{\text{div }}$ is a self-adjoint compact operator on $\mathbb{G}_{\mathrm{div}}$ and by
the classical \emph{spectral theorem}, there exists a sequence $\lambda_j$ with $0<\lambda_1\leq \lambda_2\leq \lambda_j\leq\cdots\to+\infty$
and a family of $\mathbf{e}_j \in \mathrm{D}(\mathrm{A}),$ which is orthonormal in $\mathbb{G}_\text{div}$ and such that $\mathrm{A}\mathbf{e}_j =\lambda_j\mathbf{e}_j$. We know that $\mathbf{u} \in\mathbb{G}_{\mathrm{div}}$ can be expressed as $\mathbf{u}=\sum\limits_{j=1}^{\infty}\langle \mathbf{u},\mathbf{e}_j\rangle \mathbf{e}_j,$ so that $\mathrm{A}\mathbf{u}=\sum\limits_{j=1}^{\infty}\lambda_j\langle \mathbf{u},\mathbf{e}_j\rangle \mathbf{e}_j,$ for all $\mathbf{u}\in \mathrm{D}(\mathrm{A})\subset \mathbb{G}_{\mathrm{div}}$. Thus, it is immediate that
\begin{align}
\|\nabla\mathbf{u}\|^2=\langle \mathrm{A}\mathbf{u},\mathbf{u}\rangle =\sum_{j=1}^{\infty}\lambda_j|\langle \mathbf{u},\mathbf{e}_j\rangle|^2\geq \lambda_1\sum_{j=1}^{\infty}|\langle \mathbf{u},\mathbf{e}_j\rangle|^2=\lambda_1\|\mathbf{u}\|^2.
\end{align}
For $\mathbf{u},\mathbf{v},\mathbf{w} \in \mathbb{V}_{\mathrm{div}}$, we define the trilinear operator $b(\cdot,\cdot,\cdot)$ as
$$b(\mathbf{u},\mathbf{v},\mathbf{w}) = \int_\Omega (\mathbf{u}(x) \cdot \nabla)\mathbf{v}(x) \cdot \mathbf{w}(x)\/\mathrm{d}\/ x=\sum_{i,j=1}^n\int_{\Omega}u_i(x)\frac{\partial v_j(x)}{\partial x_i}w_j(x)\/\mathrm{d}\/ x,$$
and the bilinear operator $\mathrm{B}$ from $\mathbb{V}_{\mathrm{div}} \times \mathbb{V}_{\mathrm{div}} $ into $\mathbb{V}_{\mathrm{div}}'$ is defined by,
$$ \!_{\mathbb{V}_{\mathrm{div}}'}\langle \mathrm{B}(\mathbf{u},\mathbf{v}),\mathbf{w} \rangle_{\mathbb{V}_{\mathrm{div}}} := b(\mathbf{u},\mathbf{v},\mathbf{w}), \ \text{ for all } \ \mathbf{u},\mathbf{v},\mathbf{w} \in \mathbb{V}_\text{{div}}.$$
An integration by parts yields,
\begin{equation}\label{2.7}
\mathopen{}\mathclose\bgroup\originalleft\{
\begin{aligned}
b(\mathbf{u},\mathbf{v},\mathbf{v}) &= 0, \ \text{ for all } \ \mathbf{u},\mathbf{v} \in\mathbb{V}_\text{{div}},\\
b(\mathbf{u},\mathbf{v},\mathbf{w}) &= -b(\mathbf{u},\mathbf{w},\mathbf{v}), \ \text{ for all } \ \mathbf{u},\mathbf{v},\mathbf{w}\in \mathbb{V}_\text{{div}}.
\end{aligned}
\aftergroup\egroup\originalright.\end{equation}
For more details about the linear and non-linear operators, we refer the readers to \cite{MR0609732}. Let us now provide some important inequalities, which are used frequently in the paper.
\begin{lemma}[Gagliardo-Nirenberg inequality, Theorem 2.1, \cite{MR1230384}] \label{gn}
Let $\Omega\subset\mathbb{R}^n$ and $\mathbf{u}\in\mathrm{W}^{1,p}_0(\Omega;\mathbb{R}^n)$, $p\geq 1$. Then for any fixed number $p,q\geq 1$, there exists a constant $C>0$ depending only on $n,p,q$ such that
\begin{align}\label{gn0}
\|\mathbf{u}\|_{\mathbb{L}^r}\leq C\|\nabla\mathbf{u}\|_{\mathbb{L}^p}^{\theta}\|\mathbf{u}\|_{\mathbb{L}^q}^{1-\theta},\;\theta\in[0,1],
\end{align}
where the numbers $p, q, r, n$ and $\theta$ satisfy the relation
$$\theta=\mathopen{}\mathclose\bgroup\originalleft(\frac{1}{q}-\frac{1}{r}\aftergroup\egroup\originalright)\mathopen{}\mathclose\bgroup\originalleft(\frac{1}{n}-\frac{1}{p}+\frac{1}{q}\aftergroup\egroup\originalright)^{-1}.$$
\end{lemma}
A particular case of Lemma \ref{gn} is the well known inequality due to Ladyzhenskaya (see Lemma 1 and 2, Chapter 1, \cite{MR0254401}), which is given below:
\begin{lemma}[Ladyzhenskaya's inequality]\label{lady}
For $\mathbf{u}\in\mathrm{C}_0^{\infty}(\Omega;\mathbb{R}^n), n = 2, 3$, there exists a constant $C$ such that
\begin{align}\label{lady1}
\|\mathbf{u}\|_{\mathbb{L}^4}\leq C^{1/4}\|\mathbf{u}\|^{1-\frac{n}{4}}\|\nabla\mathbf{u}\|^{\frac{n}{4}},\text{ for } n=2,3,
\end{align}
where $C=2,4,$ for $n=2,3$ respectively.
\end{lemma}
Note that the above inequality is true even in unbounded domains. For $n=3$, $r=6$, $p=q=2$, from \eqref{gn0}, we find $\theta=1$ and
\begin{align*}
\|\mathbf{u}\|_{\mathbb{L}^6}\leq C\|\nabla\mathbf{u}\|=C\|\mathbf{u}\|_{\mathbb{V}_{\mathrm{div}}}.
\end{align*}
For $n=2$, the following estimate holds:
\begin{align*}
|b(\mathbf{u},\mathbf{v},\mathbf{w})| \leq \sqrt{2}\|\mathbf{u}\|^{1/2}\| \nabla \mathbf{u}\|^{1/2}\|\mathbf{v}\|^{1/2}\| \nabla \mathbf{v}\|^{1/2}\| \nabla \mathbf{w}\|,
\end{align*}
for every $\mathbf{u},\mathbf{v},\mathbf{w} \in \mathbb{V}_{\mathrm{div}}$. Hence, for all $\mathbf{u}\in\mathbb{V}_{\mathrm{div}},$ we have
\begin{align}
\label{be}
\|\mathrm{B}(\mathbf{u},\mathbf{u})\|_{\mathbb{V}_{\mathrm{div}}'}\leq \sqrt{2}\|\mathbf{u}\|\|\nabla\mathbf{u}\|\leq \sqrt{\frac{2}{\lambda_1}}\|\mathbf{u}\|_{\mathbb{V}_{\mathrm{div}}}^2 ,
\end{align}
by using the Poincar\'e inequality. Similarly, for $n=3$, we have
\begin{align*}
|b(\mathbf{u},\mathbf{v},\mathbf{w})| \leq 2\|\mathbf{u}\|^{1/4}\| \nabla \mathbf{u}\|^{3/4}\|\mathbf{v}\|^{1/4}\| \nabla \mathbf{v}\|^{3/4}\| \nabla \mathbf{w}\|,
\end{align*}
for every $\mathbf{u},\mathbf{v},\mathbf{w} \in \mathbb{V}_{\mathrm{div}}$. Hence, for all $\mathbf{u}\in\mathbb{V}_{\mathrm{div}},$ using the Poincar\'e inequality, we have
\begin{align}
\|\mathrm{B}(\mathbf{u},\mathbf{u})\|_{\mathbb{V}_{\mathrm{div}}'}\leq 2\|\mathbf{u}\|^{1/2}\|\nabla\mathbf{u}\|^{3/2}\leq \frac{2}{\lambda_1^{1/4}}\|\mathbf{u}\|_{\mathbb{V}_{\mathrm{div}}}^2.
\end{align}
We also need the following general version of the Gagliardo-Nirenberg interpolation inequality to prove the regularity results. For functions $\mathbf{u}: \Omega\to\mathbb{R}^n$ defined on a bounded Lipschitz domain $\Omega\subset\mathbb{R}^n$, the Gagliardo-Nirenberg interpolation inequality is given by:
\begin{lemma}[Gagliardo-Nirenberg interpolation inequality, Theorem on Page125 , \cite{MR109940}]\label{GNI} Let $\Omega\subset\mathbb{R}^n$, $\mathbf{u}\in\mathrm{W}^{m,p}(\Omega;\mathbb{R}^n), p\geq 1$ and fix $1 \leq p,q \leq \infty$ and a natural number $m$. Suppose also that a real number $\theta$ and a natural number $j$ are such that
\begin{align}
\label{theta}
\theta=\mathopen{}\mathclose\bgroup\originalleft(\frac{j}{n}+\frac{1}{q}-\frac{1}{r}\aftergroup\egroup\originalright)\mathopen{}\mathclose\bgroup\originalleft(\frac{m}{n}-\frac{1}{p}+\frac{1}{q}\aftergroup\egroup\originalright)^{-1}\end{align}
and
$\frac{j}{m} \leq \theta \leq 1.$ Then for any $\mathbf{u}\in\mathrm{W}^{m,p}(\Omega;\mathbb{R}^n),$ we have
\begin{align}\label{gn1}
\|\nabla^j\mathbf{u}\|_{\mathbb{L}^r}\leq C\mathopen{}\mathclose\bgroup\originalleft(\|\nabla^m\mathbf{u}\|_{\mathbb{L}^p}^{\theta}\|\mathbf{u}\|_{\mathbb{L}^q}^{1-\theta}+\|\mathbf{u}\|_{\mathbb{L}^s}\aftergroup\egroup\originalright),
\end{align}
where $s > 0$ is arbitrary and the constant $C$ depends upon the domain $\Omega,m,n$.
\end{lemma}
If $1 < p < \infty$ and $m - j -\frac{n}{p}$ is a non-negative integer, then it is necessary to assume also that $\theta\neq 1$. Note that for $\mathbf{u}\in\mathrm{W}_0^{1,p}(\Omega;\mathbb{R}^n)$, Lemma \ref{gn} is a special case of the above inequality, since for $j=0$, $m=1$ and $\frac{1}{s}=\frac{\theta}{p}+\frac{1-\theta}{q}$ in \eqref{gn1}, and application of the Poincar\'e inequality yields (\ref{gn0}). It should also be noted that \eqref{gn1} can also be written as
\begin{align}\label{gn2}
\|\nabla^j\mathbf{u}\|_{\mathbb{L}^r}\leq C\|\mathbf{u}\|_{\mathbb{W}^{m,p}}^{\theta}\|\mathbf{u}\|_{\mathbb{L}^q}^{1-\theta},
\end{align}
By taking $j=1$, $r=4$, $n=m=p=q=s=2$ in (\ref{theta}), we get $\theta=\frac{3}{4},$ and from \eqref{gn1} we get
\begin{align}\label{gu}
\|\nabla\mathbf{u}\|_{\mathbb{L}^4}\leq C\mathopen{}\mathclose\bgroup\originalleft(\|\Delta\mathbf{u}\|^{3/4}\|\mathbf{u}\|^{1/4}+\|\mathbf{u}\|\aftergroup\egroup\originalright).
\end{align}
Also, taking $j=1$, $r=4$, $n=3$, $m=p=q=s=2$ in (\ref{theta}), we get $\theta=\frac{7}{8},$ and
\begin{align}\label{gua}
\|\nabla\mathbf{u}\|_{\mathbb{L}^4}\leq C\mathopen{}\mathclose\bgroup\originalleft(\|\Delta\mathbf{u}\|^{7/8}\|\mathbf{u}\|^{1/8}+\|\mathbf{u}\|\aftergroup\egroup\originalright).
\end{align}
\begin{lemma}
[Poincar\'e-Wirtinger inequality, Corollary 12.28, \cite{MR3726909}]\label{poin}
Assume that $1 \leq p < \infty$ and that $\Omega$ is a bounded connected open subset of the $n$-dimensional Euclidean space $\mathbb{R}^n$ whose boundary is of class $\mathrm{C}$. Then there exists a constant $C_{\Omega,p}>0$, such that for every function $\phi \in \mathrm{W}^{1,p}(\Omega)$,
$$\|\phi-\overline{\phi}\|_{\mathrm{L}^{p}(\Omega )}\leq C_{\Omega,p}\|\nabla \phi\|_{\mathrm{L}^{p}(\Omega )},$$
where
$\overline{\phi}={\frac {1}{|\Omega |}}\int _{\Omega }\phi(y)\,\mathrm {d} y$
is the average value of $\phi$ over $\Omega$.
\end{lemma}
\subsection{Basic assumptions}
Let us now make the following assumptions on $\mathrm{J}$ and $\mathrm{F}$ in order to establish the solvability results of the system \eqref{steadysys}. We suppose that the potential $\mathrm{F}$ can be written in the following form $$\mathrm{F}= \mathrm{F}_1 + \mathrm{F}_2$$ where $\mathrm{F}_1 \in \mathrm{C}^{(2+2q)}(-1,1)$ with $q \in \mathbb{N}$ fixed, and $\mathrm{F}_2 \in \mathrm{C}^2(-1,1)$.
Now, we list the assumptions on $\nu, \mathrm{J}, \mathrm{F}_1, \mathrm{F}_2 $ and mobility $m$ (cf. \cite{MR3019479}).
\begin{itemize}
\item[(A1)] $ \mathrm{J} \in \mathrm{W}^{1,1}(\mathbb{R}^2;\mathbb{R}), \ \mathrm{J}(x)= \mathrm{J}(-x) \; \text {and} \ a(x) = \int_\Omega \mathrm{J}(x-y)\/\mathrm{d}\/ y \geq 0,$ a.e., in $\Omega$.
\item[(A2)] The function $\nu$ is locally Lipschitz on $\mathbb{R}$ and there exists $\nu_1, \nu_2 >0$ such that
$$ \nu_1 \leq \nu(s) \leq \nu_2, \ \text{ for all }\ s \in \mathbb{R}.$$
\item[(A3)] There exist $C_1>0$ and $\epsilon_0 >0$ such that
$$\mathrm{F}_1^{(2+2q)} (s) \geq C_1, \ \text{ for all } \ s \in (-1,1+\epsilon_0] \cup [1-\epsilon_0,1).$$
\item[(A4)] There exists $\epsilon_0 >0$ such that, for each $k=0,1,...,2+2q$ and each $j=0,1,...,q$,
$$ \mathrm{F}_1^{(k)}(s) >0, \ \text{ for all }\ s \in [1-\epsilon_0,1)$$
$$ \mathrm{F}_1^{(2j+2)}(s) \geq 0, \ \mathrm{F}_1^{(2j+1)}(s) \leq 0, \ \text{ for all }\ s \in (-1,1+\epsilon_0].$$
\item[(A5)] There exists $\epsilon_0 >0$ such that $\mathrm{F}_1^{(2+2q)}$ is non-decreasing in $[1-\epsilon_0,1)$ and non-increasing in $(-1,-1+\epsilon_0]$.
\item[(A6)] There exists $\alpha , \beta \in \mathbb{R}$ with $\alpha + \beta > - \min\limits_{[-1,1]} \mathrm{F}''_2$ such that
$$ \mathrm{F}_1''(s) \geq \alpha, \ \text{ for all }\ s \in (-1,1), \quad a(x) \geq \beta, \ \text{ a.e. }\ x \in \Omega.$$
\item[(A7)] $\lim\limits_{s \rightarrow -1} \mathrm{F}_1'(s) = -\infty$ and $\lim\limits_{s \rightarrow 1} \mathrm{F}_1'(s) = \infty$.
\item[(A8)] The function $m$ is locally Lipschitz function on $\mathbb{R}$ and there exists $m_1 , m_2 >0$ such that
\begin{align*}
m_1 \leq m(s) \leq m_2 , \ \text{ for all }\ s \in \mathbb{R}.
\end{align*}
\end{itemize}
\begin{remark} \cite{MR3000606}
We can represent the potential $\mathrm{F}$ as a quadratic perturbation of a convex function. That is
\begin{align}\label{decomp of F}
\mathrm{F}(s)= \mathrm{G} (s) - \frac{\kappa}{2} s^2,
\end{align}
where $\mathrm{G}$ is strictly convex and $\kappa>0$.
\end{remark}
We further assume that
\begin{itemize}
\item[(A9)] there exists $C_0 > 0$ such that $\mathrm{F}''(s) + a(x) \geq C_0$ , for all $s \in (-1,1)$ a.e. $x \in \Omega$ and
$ \|\mathrm{J} \|_{\mathrm{L}^1} \leq {C_0} + \kappa$.
\end{itemize}
\begin{remark}\label{remark J}
Assumption $\mathrm{J} \in \mathrm{W}^{1,1}(\mathbb{R}^2;\mathbb{R})$ can be weakened. Indeed, it can be replaced by $\mathrm{J} \in \mathrm{W}^{1,1}(\mathrm{B}_\delta;\mathbb{R})$, where $\mathrm{B}_\delta := \{z \in \mathbb{R}^2 : |z| < \delta \}$ with $\delta := \emph{diam}(\Omega)=\sup\limits_{x,y\in \Omega}d(x,y)$, where $d(\cdot,\cdot)$ is the Euclidean metric on $\mathbb{R}^2$, or also by
\begin{eqnarray} \label{Estimate J}
\sup_{x\in \Omega} \int_\Omega \mathopen{}\mathclose\bgroup\originalleft( |\mathrm{J}(x-y)| + |\nabla \mathrm{J}(x-y)| \aftergroup\egroup\originalright) \/\mathrm{d}\/ y < +\infty.
\end{eqnarray}
\end{remark}
Note that \eqref{Estimate J} says that $\sup\limits_{x \in \Omega} \|\mathrm{J}\|_{\mathrm{W}^{1,1}} $ is finite and hence assumption (A1) is justified.
\begin{remark}\label{remark F}
Assumptions (A3)-(A6) are satisfied in the case of the physically relevant logarithmic double-well potential \eqref{2}, for any fixed positive integer $q$. In particular setting
\begin{align*}
\mathrm{F}_1 (s)= \frac{\theta}{2}((1+s)\ln(1+s)+(1-s)\ln(1-s)), \qquad \mathrm{F}_2(s) =- \frac{\theta_cs^2}{2}
\end{align*}
then Assumption (A6) is satisfied iff $\beta > \theta_c - \theta.$
\end{remark}
\section{Existence of Weak Solution}\label{sec3}\setcounter{equation}{0}
In this section, we establish the existence of a weak solution to the system \eqref{steadysys} using pseudo-monotonicity arguments and Browder's theorem. Let us first give the definition of \emph{weak solution} of the system \eqref{steadysys}.
\begin{definition} \label{weakdef}
Let $\mathbf{h} \in \mathbb{V}'_{{\mathrm{div}}}$ and $k\in (-1,1)$ be fixed. A triple $(\mathbf{u},\mu , \varphi) \in \mathbb{V}_{{\mathrm{div}}} \times \mathrm{V} \times (\mathrm{V} \cap \mathrm{L}^2_{(k)}(\Omega) ) $ is called a \emph{weak solution} of the problem \eqref{steadysys} if
\begin{align}
\label{weakphi} \int_\Omega (\mathbf{u} \cdot \nabla \varphi) \psi \, \/\mathrm{d}\/ x =& -\int_\Omega m(\varphi) \nabla \mu \cdot \nabla \psi \, \/\mathrm{d}\/ x, \\
\label{weakmu} \int_\Omega \mu \psi \, \/\mathrm{d}\/ x=& \int_\Omega (a \varphi -\mathrm{J} * \varphi) \psi \, \/\mathrm{d}\/ x +\int_\Omega \mathrm{F}'(\varphi) \psi \, \/\mathrm{d}\/ x, \\
\label{weakform nse}
\int_\Omega (\mathbf{u} \cdot \nabla) \mathbf{u} \cdot \mathbf{v} \, \/\mathrm{d}\/ x+ \int_\Omega 2\nu (\varphi) \mathrm{D}\mathbf{u} \cdot \mathrm{D}\mathbf{v}\, \/\mathrm{d}\/ x =& \int_\Omega\mu \nabla \varphi \cdot \mathbf{v} \, \/\mathrm{d}\/ x + \langle \mathbf{h} , \mathbf{v} \rangle,
\end{align}
for every $\psi \in \mathrm{V}$ and $\mathbf{v} \in \mathbb{V}_{{\mathrm{div}}}$.
\end{definition}
Our aim is to establish the existence of a weak solution of the system \eqref{steadysys} in the sense of Definition \ref{weakdef}. But, when working with the above definition, some difficulties arise in the analysis of our problem. The most important one is that $\mathrm{L}^2_{(k)}(\Omega)$ is not a vector space for $k\neq 0$. But, we can assume that $k=0$ with out loss of generality. Otherwise replace $\varphi$ by $\widetilde{\varphi}:= \varphi - k$ and $\mathrm{F}$ by $\mathrm{F}_k$ with $\mathrm{F}_k(x) := \mathrm{F}(x+k)$ for all $x \in \mathbb{R}$. Thus, in order to establish a weak solution of the system \eqref{steadysys}, we first reformulate the problem \eqref{weakphi}-\eqref{weakform nse}. We prove the existence of a solution to the reformulated problem \eqref{reformphi}-\eqref{nsreform} (see below) instead of \eqref{weakphi}-\eqref{weakform nse}. We establish the equivalence of these two problems in Lemma \ref{reformtoweak}.
We reduce $\mu$ to $\mu_0$, which has mean value $0$, and it would help in proving coercivity of an operator in the later part of this section. Let us fix $ \mu_0 =\mu - \frac{1}{|\Omega |} \int_\Omega \mathrm{F}'(\varphi)\/\mathrm{d}\/ x$. Then the reformulated problem of \eqref{weakphi}-\eqref{weakform nse} is given by
\begin{align}
\label{reformphi} \int_\Omega (\mathbf{u} \cdot \nabla \varphi) \psi \,\/\mathrm{d}\/ x =& -\int_\Omega m(\varphi) \nabla \mu_0 \cdot \nabla \psi\, \/\mathrm{d}\/ x, \\
\label{mu-reform} \int_\Omega \mu_0 \psi \,\/\mathrm{d}\/ x=& \int_\Omega (a \varphi -\mathrm{J} * \varphi) \psi \,\/\mathrm{d}\/ x +\int_\Omega \mathrm{P}_0 (\mathrm{F}'(\varphi)) \psi \,\/\mathrm{d}\/ x, \\
\label{nsreform}
\int_\Omega (\mathbf{u} \cdot \nabla)\mathbf{u} \cdot\mathbf{v} \,\/\mathrm{d}\/ x+ \int_\Omega 2\nu (\varphi) \mathrm{D}\mathbf{u} \cdot \mathrm{D}\mathbf{v}\, \/\mathrm{d}\/ x =& \int_\Omega\mu_0 \nabla \varphi \cdot \mathbf{v} \,\/\mathrm{d}\/ x + \langle \mathbf{h} , \mathbf{v} \rangle,
\end{align}
where $\mathbf{u} \in \mathbb{V}_{\mathrm{div}} , \mu_0 \in \mathrm{V}_0, \varphi \in \mathrm{V}_0$. Now we show that proving the existence of a solution to the equations \eqref{reformphi}-\eqref{nsreform} would also give a solution to \eqref{weakphi}-\eqref{weakform nse}.
\begin{lemma}\label{reformtoweak}
Let $(\mathbf{u},\mu_0,\varphi) \in\mathbb{V}_{{\mathrm{div}}}\times \mathrm{V}_0 \times \mathrm{V}_0$ be a solution to the system \eqref{reformphi}-\eqref{nsreform}. Then $(\mathbf{u},\mu,\varphi)$ is a solution to the weak formulation \eqref{weakphi}-\eqref{weakform nse}, where $ \mu_0 =\mu - \frac{1}{|\Omega |} \int_\Omega \mathrm{F}'(\varphi)\/\mathrm{d}\/ x$.
\end{lemma}
\begin{proof}
Let $(\mathbf{u},\mu_0, \varphi) \in \mathbb{V}_{\mathrm{div}}\times \mathrm{V}_0 \times \mathrm{V}_0$ be a solution of the system \eqref{reformphi}-\eqref{nsreform}. Let $\overline{\mu}= \frac{1}{|\Omega|} \int_\Omega \mathrm{F}'(\varphi)\/\mathrm{d}\/ x$. Since $\overline{\mu}$ is a scalar, from \eqref{nsreform}, using integration by parts, one can easily deduce that $ \int_\Omega \overline{\mu} \nabla \varphi \cdot \mathbf{v} \/\mathrm{d}\/ x=0 $. Then, we have
\begin{align*}
\int_\Omega (\mathbf{u} \cdot \nabla) \mathbf{u} \cdot \mathbf{v} \/\mathrm{d}\/ x+ \int_\Omega 2\nu (\varphi) \mathrm{D}\mathbf{u} \cdot \mathrm{D}\mathbf{v} \/\mathrm{d}\/ x =& \int_\Omega\mu_0 \nabla \varphi \cdot \mathbf{v} \/\mathrm{d}\/ x + \langle \mathbf{h} , \mathbf{v} \rangle+ \int_\Omega \overline{\mu} \nabla \varphi \cdot \mathbf{v} \/\mathrm{d}\/ x \\
=& \int_\Omega\mu \nabla \varphi \cdot \mathbf{v} \/\mathrm{d}\/ x + \langle \mathbf{h} , \mathbf{v} \rangle ,
\end{align*}
which gives the equation \eqref{weakform nse}. Once again, since $\overline{\mu}$ is a scalar quantity, we can clearly see \eqref{weakphi} follows from \eqref{reformphi}. Now it is left to prove \eqref{weakmu}. From \eqref{mu-reform}, we have
\begin{align} \label{3.3}
\int_\Omega \mu_0 \psi \/\mathrm{d}\/ x=& \int_\Omega (a \varphi -\mathrm{J} * \varphi) \psi \/\mathrm{d}\/ x +\int_\Omega \mathrm{P}_0 (\mathrm{F}'(\varphi)) \psi \/\mathrm{d}\/ x.
\end{align}
Using \eqref{P0} and substituting value of $\mu_0$ in \eqref{3.3} we get,
\begin{align*}
\int_\Omega \mu \psi \/\mathrm{d}\/ x=& \int_\Omega (a \varphi -\mathrm{J} * \varphi) \psi \/\mathrm{d}\/ x +\int_\Omega (\mathrm{P}_0 (\mathrm{F}'(\varphi))+ \overline{\mu} )\psi \/\mathrm{d}\/ x \\
=& \int_\Omega (a \varphi -\mathrm{J} * \varphi) \psi \/\mathrm{d}\/ x +\int_\Omega \mathrm{F}'(\varphi)\psi \/\mathrm{d}\/ x,
\end{align*}
which completes the proof.
\end{proof}
\subsection{Preliminaries}
In order to formulate the problem \eqref{reformphi}-\eqref{nsreform} in the framework of Browder's Theorem (see Theorem \ref{Browder} below), we need some preliminaries, which we state below.
Let $\mathbb{X}$ be a Banach space and $\mathbb{X}'$ be its topological dual. Let $\mathrm{T}$ be a function from $\mathbb{X}$ to $\mathbb{X}'$ with domain $\mathrm{D}=\mathrm{D}(\mathrm{T}) \subseteq \mathbb{X}.$
\begin{definition}[Definition 2.3, \cite{MR3014456}] The function $\mathrm{T}$ is said to be
\begin{itemize}
\item[(i)] \emph{demicontinuous} if for every sequence $u_k \in \mathrm{D}, u \in \mathrm{D}$ and $u_k \rightarrow u$ in $\mathbb{X}$ implies that $\mathrm{T}(u_k) \rightharpoonup \mathrm{T}(u)$ in $\mathbb{X}'$,
\item[(ii)] \emph{hemicontinuous} if for every $u \in \mathrm{D}, v \in \mathbb{X}$ and for every sequence of positive real numbers $t_k$ such that $u+t_k v \in \mathrm{D},$ it holds that $t_k \rightarrow 0$ implies $\mathrm{T}(u+t_k v) \rightharpoonup \mathrm{T}(u)$ in $\mathbb{X}'$,
\item[(iii)] \emph{locally bounded} if for every sequence $u_k \in \mathrm{D}$, $u \in \mathrm{D}$ and $u_k \rightarrow u $ in $\mathbb{X}$ imply that $\mathrm{T}(u_k)$ is bounded in $\mathbb{X}'$.
\end{itemize}
\end{definition}
From the above definition, it is clear that a demicontinuous function is hemicontinuous and locally bounded.
\begin{definition}[Definition 2.1 (iv), \cite{MR3014456}]
We say that $\mathrm{T}$ is \emph{pseudo-monotone} if, for every sequence $u_k$ in $\mathbb{X}$ such that $u_k \rightharpoonup u$ in $\mathbb{X}$ and
\begin{equation*}
\limsup_{k \rightarrow \infty} \,_{\mathbb{X}'}\langle \mathrm{T}(u_k), u_k - u\rangle_{\mathbb{X}} \leq 0
\end{equation*}
implies
\begin{equation*}
\liminf_{k \rightarrow \infty} \,_{\mathbb{X}'}\langle \mathrm{T}(u_k), u_k-v \rangle_{\mathbb{X}} \geq \,_{\mathbb{X}'}\langle \mathrm{T}(u),u-v \rangle_{\mathbb{X}}
\end{equation*}
for every $v \in \mathbb{X}$. Moreover $\mathrm{T}$ is said to be \emph{monotone} if $$ \,_{\mathbb{X}'}\langle \mathrm{T}(u)-\mathrm{T}(v),u-v\rangle_{\mathbb{X}} \geq 0, \ \text{ for every }\ u ,v\in \mathrm{D}.$$
\end{definition}
\begin{definition}
A mapping $\mathrm{T}: \mathbb{X} \rightarrow \mathbb{X}'$ is said to be maximal monotone if it is monotone and its graph
\begin{align*}
\mathrm{G}(\mathrm{T})= \mathopen{}\mathclose\bgroup\originalleft\{ (u,w) : w \in \mathrm{T}(u) \aftergroup\egroup\originalright\} \subset \mathbb{X} \times \mathbb{X}'
\end{align*}
is not properly contained in graph of any other monotone operator. In other words, for $u\in\mathbb{X}$ and $w\in\mathbb{X}'$, the inequality $\,_{\mathbb{X}'}\langle w-\mathrm{T}(v),u-v\rangle_{\mathbb{X}}\geq 0$, for all $v\in\mathbb{X}$ implies $w=\mathrm{T}(u)$.
\end{definition}
\begin{definition}[Definition 2.3, \cite{MR3014456}]
Let $\mathbb{X}$ and $\mathbb{Y}$ be Banach spaces. A bounded linear operator $\mathrm{T}:\mathbb{X} \rightarrow \mathbb{Y}$ is said to be a \emph{completely continuous operator} if for every sequence $u_k \rightharpoonup u$ in $\mathbb{X}$ implies $\mathrm{T}(u_k) \rightarrow \mathrm{T}(u)$ in $\mathbb{Y}$.
\end{definition}
We can see that complete continuity implies pseudo-monotonicity (Corollary 2.12, \cite{MR3014456}).
\begin{lemma}[Lemma 5.1, \cite{MR3524178} ] \label{pseudo-monotone}
Let $\mathbb{X}$ be a real, reflexive Banach space and $\widetilde{\mathrm{T}} : \mathbb{X} \times \mathbb{X} \rightarrow \mathbb{X}'$ be such that for all $u \in \mathbb{X}$:
\begin{enumerate}
\item $\widetilde{\mathrm{T}} (u, \cdot): \mathbb{X} \rightarrow \mathbb{X}'$ is monotone and hemicontinuous.
\item $\widetilde{\mathrm{T}} ( \cdot,u): \mathbb{X} \rightarrow \mathbb{X}'$ is completely continuous.
\end{enumerate}
Then the operator $\mathrm{T}:\mathbb{X} \rightarrow \mathbb{X}'$ defined by $\mathrm{T}(u) := \widetilde{\mathrm{T}}(u,u)$ is pseudo-monotone.
\end{lemma}
\begin{definition}
Let $\mathbb{X}$ be a real Banach space and $f:\mathbb{X} \rightarrow (-\infty, \infty]$ be a functional on $\mathbb{X}$. A linear functional $g \in \mathbb{X}'$ is called \emph{subgradient} of $f$ at $u$ if $f(u) \not\equiv \infty$ and
\begin{align*}
f(v) \geq f(u) + \!_{\mathbb{X}'}\langle g, v-u \rangle_{\mathbb{X}},
\end{align*}
holds for all $v \in \mathbb{X}$.
\end{definition}
We know that subgradient of a functional need not be unique. The set of all subgradients of $f$ at $u$ is called \emph{subdifferential} of $f$ at $u$ and is denoted by $\partial f (u)$.We say that $f$ is \emph{Gateaux differentiable} at $u$ in $\mathbb{X}$ if $\partial f(u)$ consists of exactly one element (see \cite{MR1195128}).
\begin{lemma}[Theorem A, \cite{MR262827}] \label{rockfellar}
If $f$ is a lower semicontinuous, proper convex function on $\mathbb{X}$ (that is, $f$ is a convex function and $f$ takes values in the extended real number line such that $f(u)<+\infty $
for at least one $u\in\mathbb{X}$ and $f(u)>-\infty $
for every $u\in\mathbb{X}$), then $\partial f$ is a maximal monotone operator from $\mathbb{X}$ to $\mathbb{X}'$.
\end{lemma}
Now we state Browder's theorem, which we use to prove the existence of solution to the problem \eqref{reformphi}-\eqref{nsreform}.
\begin{theorem}[Theorem 32.A. in \cite{MR1033498}, Browder] \label{Browder}
\begin{enumerate}
\item Let $\mathbb{Y}$ be a non-empty, closed and convex subset of a real and reflexive Banach space $\mathbb{X}$.
\item Let $\mathrm{T}: \mathbb{Y} \rightarrow \mathcal{P}(\mathbb{X}')$ be a maximal monotone operator, where $\mathcal{P}(\mathbb{X}')$ denotes the power set of $\mathbb{X}'$.
\item Let $\mathrm{S}: \mathbb{Y} \rightarrow \mathbb{X}'$ be a pseudo-monotone, bounded and demicontinuous mapping.
\item If the set $\mathbb{Y}$ is unbounded, then the operator $\mathrm{S}$ is $\mathrm{T}$-coercive with respect to the fixed element $b \in \mathbb{X}'$, that is, there exists an element $u_0 \in \mathbb{Y} \cap \mathrm{D} (\mathrm{T})$ and $R>0$ such that
\begin{align*}
\!_{\mathbb{X}'}\langle \mathrm{S}(u), u-u_0 \rangle_{\mathbb{X}} > \,_{\mathbb{X}'}\langle b, u-u_0 \rangle_{\mathbb{X}},
\end{align*}
for all $u \in \mathbb{Y}$ with $\| u\|_{\mathbb{X}}>R$.
\end{enumerate}Then the problem
\begin{align}
b\in \mathrm{T} (u) + \mathrm{S}(u)
\end{align}
has a solution $u \in \mathbb{Y} \cap \mathrm{D}(\mathrm{T})$.
\end{theorem}
\subsection{The functional $f$} We mainly follow the work of \cite{MR3524178} (local Cahn-Hilliard-Navier-Stokes equations) to establish the solvability results of the system \ref{steadysys}. Before we proceed to prove our main result, we first consider the following functional and study its properties. Let us define
\begin{align}
\label{functional} f(\varphi):= \frac{1}{4}\int_\Omega \int_\Omega \mathrm{J}(x-y)(\varphi(x)-\varphi(y))^2\/\mathrm{d}\/ x \/\mathrm{d}\/ y + \int_\Omega \mathrm{G}(\varphi(x))\/\mathrm{d}\/ x,
\end{align}
where $\varphi \in \mathrm{L}^2_{(0)}(\Omega)$ and $\mathrm{G}$ is given in \eqref{decomp of F}. Using assumption \ref{decomp of F}, we define $\mathrm{G}(x)= \infty$ for $x \notin [-1, 1]$. The domain of $f$ is given by
\begin{align} \label{domain0f_f}
\mathrm{D}(f) = \mathopen{}\mathclose\bgroup\originalleft\{ \varphi \in \mathrm{L}^2_{(0)}(\Omega) : \mathrm{G}(\varphi) \in \mathrm{L}^1(\Omega) \aftergroup\egroup\originalright\}.
\end{align}
For $\varphi \notin \mathrm{D}(f)$, we set $f(\varphi)= +\infty$.
Note that $ \mathrm{D}(f) \neq \varnothing$.
Given a functional $f:\mathrm{L}^2_{(0)}(\Omega) \rightarrow (-\infty,\infty]$, its subgradient maps from $\mathrm{L}^2_{(0)}(\Omega)$ to $ \mathcal{P}{(\mathrm{L}^2_{(0)}(\Omega)')}$. We write $\partial_{\mathrm{L}^2_{(0)}}f$ as the subgradient of the functional $f:\mathrm{L}^2_{(0)}(\Omega) \rightarrow (-\infty,\infty]$. Since $\mathrm{V}_0 \hookrightarrow\mathrm{L}^2_{(0)}(\Omega)$, we can also consider $f$ as a functional from $\mathrm{V}_0$
to $(-\infty,\infty]$, and hence we have to distinguish between its different subgradients. If we consider $f:\mathrm{V}_0 \rightarrow (-\infty,\infty]$, then the subgradient of $f$ is denoted by $\partial_{\mathrm{V}_0} f$.
\begin{proposition} \label{gatuexderivative}
Let $f$ be defined as in \eqref{functional}. Then $f$ is Gateaux differentiable on $\mathrm{L}^2_{(0)} (\Omega)$ and we have
\begin{align}\label{sub}
\partial_{\mathrm{L}^2_{(0)}} f(\varphi) = a \varphi -\mathrm{J}*\varphi + \mathrm{P}_0 \mathrm{G}'(\varphi).
\end{align}
and
\begin{align*}
\mathrm{D}(\partial_{\mathrm{L}^2_{(0)}}f) = \mathopen{}\mathclose\bgroup\originalleft\{ \varphi \in \mathrm{L}^2_{(0)}(\Omega) : \mathrm{G}'(\varphi) \in \mathrm{H} \aftergroup\egroup\originalright\}.
\end{align*}
Furthermore, it holds that
\begin{align}\label{3.10}
\|\partial_{\mathrm{L}^2_{(0)}}f(\varphi)\|\leq \mathopen{}\mathclose\bgroup\originalleft(\|a\|_{\mathrm{L}^{\infty}}+\|\mathrm{J}\|_{\mathrm{L}^1}\aftergroup\egroup\originalright)\|\varphi\|+\|\mathrm{G}'(\varphi)\|\leq 2a^*\|\varphi\|+\|\mathrm{G}'(\varphi)\|.
\end{align}
\end{proposition}
\begin{proof}
Let $h \in \mathrm{L}^2_{(0)}(\Omega)$. Then, we have
\begin{align*}
& \frac{f(\varphi + \epsilon h)-f(\varphi)}{\epsilon} \nonumber\\&= \frac{1}{\epsilon} \mathopen{}\mathclose\bgroup\originalleft[\frac{1}{4}\int_\Omega \int_\Omega \mathrm{J}(x-y)((\varphi+ \epsilon h)(x)-(\varphi+\epsilon h)(y))^2\/\mathrm{d}\/ x \/\mathrm{d}\/ y + \int_\Omega \mathrm{G}((\varphi+\epsilon h)(x))\/\mathrm{d}\/ x \aftergroup\egroup\originalright. \\
& \hspace{2cm}\mathopen{}\mathclose\bgroup\originalleft. -\mathopen{}\mathclose\bgroup\originalleft(\frac{1}{4}\int_\Omega \int_\Omega \mathrm{J}(x-y)((\varphi)(x)-\varphi(y))^2\/\mathrm{d}\/ x \/\mathrm{d}\/ y + \int_\Omega \mathrm{G}(\varphi(x))\/\mathrm{d}\/ x \aftergroup\egroup\originalright) \aftergroup\egroup\originalright] \\
&= \frac{1}{4}\int_\Omega \int_\Omega \mathrm{J}(x-y)(2(\varphi(x)-\varphi(y))(h(x)-h(y)) + \epsilon (h(x)-h(y))^2) \\
& \hspace{1cm}+ \frac{1}{\epsilon} \int_\Omega (\mathrm{G}((\varphi+\varepsilon h)(x)) - \mathrm{G}(\varphi(x)) )\/\mathrm{d}\/ x .
\end{align*}
Hence
\begin{align}
\lim_{\epsilon \rightarrow 0 }\frac{f(\varphi + \epsilon h)-f(\varphi)}{\epsilon} = (a \varphi- \mathrm{J}*\varphi + \mathrm{G}'(\varphi), h).
\end{align}
Since $\mathrm{P}_0$ is the orthogonal projection onto $\mathrm{L}^2_{(0)}(\Omega)$, we get \eqref{sub}. Note that \eqref{3.10} is an immediate consequence of \eqref{sub}.
\end{proof}
\begin{lemma} \label{lowersemi}
The functional $f: \mathrm{L}^2_{(0)}(\Omega) \rightarrow (-\infty,\infty]$ defined by \eqref{functional} is proper convex and lower semicontinuous with $\mathrm{D}(f) \neq \varnothing$.
\end{lemma}
\begin{proof}
\textbf{Claim (1).} \emph{$f$ is proper convex with $\mathrm{D}(f) \neq \varnothing$:} In order to prove convexity, we use the fact that $f$ is convex if and only if $$(\partial_{\mathrm{L}^2_{(0)}} f(\varphi)-\partial_{\mathrm{L}^2_{(0)}} f(\psi) , \varphi - \psi) \geq 0,$$ for all $\varphi,\psi\in\mathrm{L}^2_{(0)}(\Omega)$. Observe that for $0<\theta<1$, using Taylor's series expansion and the definition of $\mathrm{P}_0$, we find that
\begin{align*}
&\hspace*{-0.3 cm}(a \varphi - \mathrm{J}*\varphi+ \mathrm{P}_0 \mathrm{G}'(\varphi) -(a \psi - \mathrm{J}*\psi+ \mathrm{P}_0 \mathrm{G}'(\psi) ), \varphi-\psi) \qquad \qquad \\
&=(a \varphi - \mathrm{J}*\varphi+ \mathrm{F}'(\varphi) +\kappa \varphi-(a \psi - \mathrm{J}*\psi+ \mathrm{F}'(\psi)+\kappa\psi ), \varphi-\psi) \qquad \qquad \\
&= (a(\varphi-\psi)-\mathrm{J}*(\varphi-\psi)+ \mathrm{F}'(\varphi)- \mathrm{F}'(\psi)+ \kappa(\varphi- \psi) ,\varphi-\psi) \qquad \ \\
&= (a(\varphi-\psi)-\mathrm{J}*(\varphi-\psi)+ \mathrm{F}''(\varphi + \theta \psi)(\varphi- \psi)+ \kappa(\varphi- \psi) ,\varphi-\psi) \\
&\geq C_0 \| \varphi- \psi\|^2 - \|\mathrm{J}\|_{\mathrm{L}^1} \| \varphi- \psi\|^2 + \kappa \| \varphi- \psi\|^2 \quad \hspace{3.65cm} \\
&\geq (C_0 + \kappa - \|\mathrm{J}\|_{\mathrm{L}^1})\| \varphi- \psi\|^2 \hspace{7.78cm} \\
&\geq 0,
\end{align*}
using (A9). Hence $f$ is a convex functional on $\mathrm{L}^2_{(0)}(\Omega) $. Since the domain of $f$ is $\mathrm{D}(f) = \mathopen{}\mathclose\bgroup\originalleft\{ \varphi \in \mathrm{L}^2_{(0)}(\Omega) : \mathrm{G}(\varphi) \in \mathrm{L}^1(\Omega) \aftergroup\egroup\originalright\} \neq \varnothing,$ it is immediate that $f$ is proper convex.
\textbf{Claim (2).} \emph{$f$ is lower semicontinuous:} Let $\varphi_k \in \mathrm{L}^2_{(0)}(\Omega)$ and $\varphi_k \rightarrow \varphi $ in $\mathrm{L}^2_{(0)}(\Omega)$ as $k \rightarrow \infty$. Our aim is to establish that $ f(\varphi)\leq \liminf\limits_{k \rightarrow \infty} f(\varphi_k).$ It is enough to consider the case $\liminf\limits_{k \rightarrow \infty} f(\varphi_k) < +\infty$. Thus, for this sequence, we can assume that $f(\varphi_k) \leq \mathrm{M}$, for some $\mathrm{M}$. This implies that $\varphi_k \in \mathrm{D}(f)$, for all $k\in\mathbb{N}$. Since $\mathrm{G}:[-1,1] \rightarrow \mathbb{R}$ is a continuous function, by adding a suitable constant, we can assume without loss of generality that $\mathrm{G} \geq 0$ and we have
\begin{align*}
\mathrm{G}(\varphi) \leq \liminf_{k \rightarrow \infty} \mathrm{G}(\varphi_k),
\end{align*}
which gives from Fatou's lemma that
\begin{align} \label{lowerG}
\int_\Omega \mathrm{G}(\varphi(x)) \/\mathrm{d}\/ x \leq \liminf_{k \rightarrow \infty} \int_\Omega \mathrm{G}(\varphi_k (x)) \/\mathrm{d}\/ x .
\end{align}
Now consider the functional $\mathrm{I}(\cdot)$ defined by
\begin{align*}
\mathrm{I}(\varphi) :=\int_\Omega \int_\Omega \mathrm{J}(x-y)(\varphi(x)-\varphi(y))^2\/\mathrm{d}\/ x \/\mathrm{d}\/ y = (a \varphi, \varphi) -(\mathrm{J}*\varphi, \varphi).
\end{align*}
We show that the function $\mathrm{I}(\cdot)$ is continuous. We consider
\begin{align*}
|\mathrm{I}(\varphi_k)-\mathrm{I}(\varphi)|=&|(a \varphi_k, \varphi_k)-(\mathrm{J}*\varphi_k, \varphi_k) - (a \varphi, \varphi) -(\mathrm{J}*\varphi, \varphi)| \\
\leq & |(a(\varphi_k- \varphi), \varphi_k )|+|(a \varphi, \varphi_k-\varphi)| + |(\mathrm{J}*(\varphi_k-\varphi), \varphi_k)| + |(\mathrm{J}*\varphi,\varphi_k-\varphi)|\\
\leq & \|a\|_{\mathrm{L}^\infty} \,(\|\varphi_k\| + \| \varphi\|) \|\varphi_k-\varphi\| + \|\mathrm{J}\|_{\mathrm{L}^1}(\|\varphi_k\| + \| \varphi\|) \|\varphi_k-\varphi\|,
\end{align*}
where we used Young's inequality, Young's inequality for convolution and H\"older's inequality. Then, we have $|\mathrm{I}(\varphi_k)-\mathrm{I}(\varphi)|\to 0 \ \ \text{as } k \rightarrow \infty $, since $\varphi_k \rightarrow \varphi $ in $\mathrm{L}^2_{(0)}(\Omega)$ as $k \rightarrow \infty$. Since continuity implies lower semicontinuity, we have
\begin{align} \label{lowerJ}
\int_\Omega \int_\Omega \mathrm{J}(x-y)(\varphi(x)-\varphi(y))^2\/\mathrm{d}\/ x \/\mathrm{d}\/ y =\liminf_{k \rightarrow \infty} \int_\Omega \int_\Omega \mathrm{J}(x-y)(\varphi_k(x)-\varphi_k(y))^2\/\mathrm{d}\/ x \/\mathrm{d}\/ y.
\end{align}
Combining \eqref{lowerG} and \eqref{lowerJ}, we get
\begin{align*}
f(\varphi)\leq \liminf_{k \rightarrow \infty} f(\varphi_k).
\end{align*}
This proves $f$ is lower semicontinuous.
\end{proof}
\begin{remark} \label{lower_H1}
The proper convexity of $f: \mathrm{V}_0 \rightarrow (-\infty,\infty]$ is immediate, since $\mathrm{V}_0 \hookrightarrow\mathrm{L}^2_{(0)}(\Omega)$. Let $(\varphi_k)_{k \in \mathbb{R}} \in \mathrm{V}_0 $ be such that $\varphi_k \rightarrow \varphi$ in $\mathrm{V}_0$. Then from Lemma \ref{poin}, we can easily see that $\varphi_k \rightarrow \varphi $ in $\mathrm{L}^2_{(0)}(\Omega)$. Therefore using Lemma \ref{lowersemi}, we get $f: \mathrm{V}_0 \rightarrow (-\infty,\infty]$ is lower semicontinuous.
\end{remark}
\begin{proposition} \label{maximalmonotone}
The subgradients $\partial_{\mathrm{L}^2_{(0)}} f$ and $\partial_{\mathrm{V}_0} f$ are maximal monotone operators.
\end{proposition}
\begin{proof}
In Lemma \ref{lowersemi}, we have proved that $f:\mathrm{L}^2_{(0)}(\Omega) \rightarrow \mathbb{R}$ is proper convex and lower semicontinuous. By using Lemma \ref{rockfellar}, we obtain that the operator $\partial_{\mathrm{L}^2_{(0)}} f : \mathrm{L}^2_{(0)}(\Omega) \rightarrow \mathcal{P}((\mathrm{L}^2_{(0)}\Omega)')$ is maximal monotone. Now for the operator $\partial_{\mathrm{V}_0} f$, Remark \ref{lower_H1} gives that $f: \mathrm{V}_0 \rightarrow (-\infty,\infty]$ is proper convex and lower semicontinuous. Hence, once again using Lemma \ref{rockfellar}, we get that $\partial_{\mathrm{V}_0} f$ is also maximal monotone.
\end{proof}
\begin{lemma}[Lemma 3.7, \cite{MR3524178}]
Consider the functional $f$ as in \eqref{functional}. Then for every $\varphi \in \mathrm{D}(\partial_{\mathrm{L}^2_{(0)}}f) \cap \mathrm{V}_0$ we have that $\partial_{\mathrm{L}^2_{(0)}}f(\varphi) \subseteq \partial_{\mathrm{V}_0}f(\varphi)$.
\end{lemma}
\begin{lemma}[Lemma 3.8, \cite{MR3524178}] \label{subgrad H1_0}
Let $\varphi \in \mathrm{D}(\partial_{\mathrm{V}_0} f)$ and $w \in \partial_{\mathrm{V}_0}f(\varphi)$. Suppose $w \in \mathrm{L}^2_{(0)}(\Omega)$, then $\varphi \in \mathrm{D}(\partial_{\mathrm{L}^2_{(0)}}f)$ and
\begin{align*}
w=\partial_{\mathrm{L}^2_{(0)}}f(\varphi) = a \varphi - \mathrm{J}*\varphi + \mathrm{P}_0 \mathrm{G}'(\varphi).
\end{align*}
\end{lemma}
\subsection{Abstract Formulation}
In this subsection, we define the following spaces in order to set up the problem in Browder's theorem (see Theorem \ref{Browder}).
\begin{align*}
&\mathbb{X} :=\mathbb{V}_{\mathrm{div}}\times \mathrm{V}_0 \times \mathrm{V}_0.
\end{align*}
Let us define,
\begin{align} \label{Zdef}
\mathrm{Z}:=\mathopen{}\mathclose\bgroup\originalleft\{ \varphi \in \mathrm{V}_0 :\varphi (x) \in [-1,1],\ \text{ a.e.} \aftergroup\egroup\originalright\},
\end{align}
and
\begin{align*}
\mathbb{Y}:=\mathbb{V}_{\mathrm{div}}\times \mathrm{V}_0 \times \mathrm{Z}.
\end{align*}
Clearly $\mathbb{Y}$ is a closed subspace of $\mathbb{X}$. Let $\mathrm{D}(\mathrm{T}):= \mathbb{V}_{\mathrm{div}}\times \mathrm{V}_0 \times \mathrm{D}(\partial_{\mathrm{V}_0}f)$ and we define a mapping $\mathrm{T}:\mathbb{Y} \rightarrow \mathcal{P}(\mathbb{X}')$ by
\begin{equation} \label{operator T}
\mathrm{T}(\mathbf{u},\mu_0, \varphi) :=
\mathopen{}\mathclose\bgroup\originalleft\{\begin{array}{cl}
\mathopen{}\mathclose\bgroup\originalleft\{ \mathopen{}\mathclose\bgroup\originalleft( \begin{array}{c}
0\\
0\\
\partial_{\mathrm{V}_0} f(\varphi)
\end{array} \aftergroup\egroup\originalright) \aftergroup\egroup\originalright\}, & \text{if } (\mathbf{u},\mu_0, \varphi) \in \mathrm{D}(\mathrm{T}), \\
\varnothing, & \text{otherwise.}
\end{array}\aftergroup\egroup\originalright.
\end{equation}
We define the operator $\mathrm{S}: \mathbb{Y} \rightarrow \mathbb{X}'$ as
\begin{align} \label{operator S}
\nonumber \!_{\mathbb{X}'}\langle \mathrm{S}\mathbf{x},\mathbf{y} \rangle_{\mathbb{X}} &:= \int_\Omega (\mathbf{u} \cdot \nabla) \mathbf{u} \cdot \mathbf{v} \/\mathrm{d}\/ x+ \int_\Omega 2\nu (\varphi) \mathrm{D}\mathbf{u} \cdot \mathrm{D}\mathbf{v} \/\mathrm{d}\/ x - \int_\Omega\mu_0 \nabla \varphi \cdot \mathbf{v} \/\mathrm{d}\/ x + \int_\Omega (\mathbf{u} \cdot \nabla \varphi) \cdot \eta \/\mathrm{d}\/ x \\
&\quad+\int_\Omega m(\varphi) \nabla \mu_0 \cdot \nabla \eta \/\mathrm{d}\/ x -\int_\Omega \mu_0 \psi \/\mathrm{d}\/ x -\int_\Omega \mathrm{P}_0 (\kappa \varphi) \psi \/\mathrm{d}\/ x,
\end{align}
for all $\mathbf{x}=(\mathbf{u},\mu_0, \varphi) \in \mathbb{Y}$, $\mathbf{y}=(\mathbf{v},\eta, \psi) \in \mathbb{X}$ and $\mathbf{b} \in \mathbb{X}'$ is defined by
\begin{align*}
\!_{\mathbb{X}'}\langle \mathbf{b},\mathbf{y} \rangle_{\mathbb{X}} := \langle \mathbf{h} , \mathbf{v} \rangle,
\end{align*}
for all $\mathbf{y}\in \mathbb{X}$. From the relations \eqref{operator T} and \eqref{operator S}, the problem $b \in \mathrm{T}(\mathbf{u}, \mu_0,\varphi) +\mathrm{S}(\mathbf{u}, \mu_0,\varphi)$ with $(\mathbf{u}, \mu_0,\varphi) \in \mathbb{Y}\cap \mathrm{D}(\mathrm{T}) $ can be written as
\begin{equation} \label{operatorform}
\mathopen{}\mathclose\bgroup\originalleft( \begin{array}{c}
0 \\
0 \\
\partial_{\mathrm{V}_0} f(\varphi)
\end{array}
\aftergroup\egroup\originalright)
+
\mathopen{}\mathclose\bgroup\originalleft(
\begin{array}{c}
(\mathbf{u} \cdot \nabla)\mathbf{u} -\text{div}(2\nu(\varphi) \mathrm{D}\mathbf{u})-\mu_0 \nabla \varphi\\
-\text{div}(m(\varphi) \nabla \mu_0)+\mathbf{u} \cdot \nabla \varphi \\
-\mu_0-\mathrm{P}_0(\kappa \varphi)
\end{array}
\aftergroup\egroup\originalright)
=
\mathopen{}\mathclose\bgroup\originalleft(
\begin{array}{c}
\mathbf{h} \\
0\\
0
\end{array}
\aftergroup\egroup\originalright),
\end{equation}
in $\mathbb{X}'$.
If we can prove that \eqref{operatorform} has a solution, then this solution solves the reformulated problem \eqref{reformphi}-\eqref{nsreform}. This is the content of the next lemma. Later, we discuss the existence of solution for \eqref{operatorform}.
\begin{lemma} \label{oper_reform}
Let $(\mathbf{u}, \mu_0,\varphi)\in \mathbb{Y} \cap \mathrm{D}(T)$ satisfies $ \mathbf{b} \in \mathrm{T}(\mathbf{u}, \mu_0,\varphi) + \mathrm{S}(\mathbf{u}, \mu_0,\varphi)$. Then $(\mathbf{u}, \mu_0,\varphi)$ is a solution of the reformulated problem \eqref{reformphi}-\eqref{nsreform}.
\end{lemma}
\begin{proof}
Let $(\mathbf{u},\mu_0,\varphi)$ be such that $ \mathbf{b} \in \mathrm{T}(\mathbf{u}, \mu_0,\varphi) + \mathrm{S}(\mathbf{u}, \mu_0,\varphi)$. From first and second equations of \eqref{operatorform}, it clearly follows that \eqref{reformphi} and \eqref{nsreform} are satisfied for all $\mathbf{v} \in \mathbb{V}_{\mathrm{div}}$ and $\psi \in \mathrm{V}_0$. Now from the third equation of \eqref{operatorform}, there exists $w \in \partial_{\mathrm{V}_0}f(\varphi)$ such that
\begin{align*}
w = \mu_0 +\mathrm{P}_0(\kappa \varphi) \ \text{ in } \ \mathrm{V}_0'.
\end{align*}
Since $\mu_0 + \mathrm{P}_0(\kappa \varphi) \in \mathrm{L}^2_{(0)}(\Omega)$ and $\varphi \in \mathrm{D}(\partial_{\mathrm{V}_0}f)$, we can see from Lemma \ref{subgrad H1_0} that $w=a\varphi - \mathrm{J}*\varphi + \mathrm{P}_0(\mathrm{G}'(\varphi))$ in $\mathrm{V}_0'$. This gives for every $\psi \in \mathrm{V}_0$
\begin{align*}
\int_\Omega (\mu_0 + \mathrm{P}_0(\kappa \varphi )) \psi \/\mathrm{d}\/ x = \int_\Omega ( a\varphi - \mathrm{J}*\varphi ) \psi \/\mathrm{d}\/ x+ \int_\Omega \mathrm{P}_0(\mathrm{G}'(\varphi))\/\mathrm{d}\/ x .
\end{align*}
Hence
\begin{align*}
\int_\Omega \mu_0 \psi \/\mathrm{d}\/ x & = \int_\Omega (a\varphi - \mathrm{J}*\varphi ) \/\mathrm{d}\/ x+ \int_\Omega \mathrm{P}_0(\mathrm{G}'(\varphi)-\kappa \varphi) \/\mathrm{d}\/ x \\
&= \int_\Omega (a\varphi - \mathrm{J}*\varphi) \psi \/\mathrm{d}\/ x + \int_\Omega \mathrm{P}_0(\mathrm{F}'(\varphi)) \/\mathrm{d}\/ x.
\end{align*}
for all $\psi\in \mathrm{V}_0$.
\end{proof}
In Lemma \ref{oper_reform}, we showed that the existence of $(\mathbf{u}, \mu_0, \varphi)$ such that $\mathbf{b} \in \mathrm{T}(\mathbf{u}, \mu_0, \varphi)+\mathrm{S}(\mathbf{u}, \mu_0, \varphi)$ implies $(\mathbf{u}, \mu_0, \varphi)$ satisfies the reformulation \eqref{reformphi}-\eqref{nsreform}. Now we show that there exists a $(\mathbf{u}, \mu_0, \varphi) \in \mathbb{Y} \cap \mathrm{D}(\mathrm{T})$ which satisfies $\mathbf{b} \in \mathrm{T}(\mathbf{u}, \mu_0, \varphi)+\mathrm{S}(\mathbf{u}, \mu_0, \varphi)$. To this purpose, we use Browder's theorem (see Theorem \ref{Browder}).
\begin{lemma}\label{browderproof}
Let $\mathrm{T},\mathrm{S}$ be as defined in \eqref{operator T} and \eqref{operator S}. Given $\mathbf{b} \in \mathbb{X}'$, there exists a triple $\mathbf{x}=(\mathbf{u}, \mu_0, \varphi) \in \mathbb{Y} \cap \mathrm{D}(\mathrm{T})$ such that $\mathbf{b} \in \mathrm{T}(\mathbf{x}) + \mathrm{S}(\mathbf{x})$.
\end{lemma}
\begin{proof}
Let us prove that the operators $\mathrm{T}$ and $\mathrm{S}$, and the spaces $\mathbb{X}$ and $\mathbb{Y}$ satisfy the hypothesis of Browder's theorem (see Theorem \ref{Browder}) in the following steps.
\textbf{(1)} We can see that the set $\mathbb{Y}$ is non-empty, closed, convex subset of $\mathbb{X}$. Also $\mathbb{X}$ is a reflexive real Banach space, since $\mathbb{V}_{\mathrm{div}}$ and $\mathrm{V}_0$ are reflexive.
\textbf{(2)} Now we show that the operator $\mathrm{T}:\mathbb{Y} \rightarrow \mathcal{P}(\mathbb{X}') $ is maximal monotone. Let us first show that $\mathrm{D}(\partial_{\mathrm{V}_0}f) \subseteq \mathrm{Z}. $ In order to get this result, let us take $\varphi \in \mathrm{D}(\partial_{\mathrm{V}_0}f)$. Then we know that $f(\varphi) \neq \infty$, since $\mathrm{D}(\partial_{\mathrm{V}_0}f) \subseteq \mathrm{D}(f)$. This gives that $\varphi (x) \in [-1,1]$, since $\mathrm{G}(x)=+\infty $ for $x \notin [-1,1]$. Hence $\varphi \in \mathrm{Z}$ and $\mathrm{D}(\partial_{\mathrm{V}_0}f) \subseteq \mathrm{Z}. $ From Proposition \ref{maximalmonotone}, the operator $ \partial_{\mathrm{V}_0} f: \mathrm{V}_0 \rightarrow \mathcal{P}(\mathrm{V}_0')$ is maximal monotone. This implies that the operator $\mathrm{T} : \mathbb{X} \rightarrow \mathcal{P}(\mathbb{X}')$ is maximal monotone. Observe that,
\begin{align*}
\mathrm{D}(\mathrm{T})= \mathbb{V}_{\mathrm{div}}\times \mathrm{V}_0 \times \mathrm{D}(\partial_{\mathrm{V}_0}f) \subseteq \mathbb{V}_{\mathrm{div}}\times \mathrm{V}_0 \times \mathrm{Z} = \mathbb{Y} \subset \mathbb{X}.
\end{align*}
Moreover by the definition of $\mathrm{T}$, for $(\mathbf{u}, \mu_0, \varphi) \notin \mathrm{D}(\mathrm{T})$, $\mathrm{T}(\mathbf{u}, \mu_0, \varphi) = \varnothing$. Hence it follows that $\mathrm{T}: \mathbb{Y} \rightarrow \mathcal{P}(\mathbb{X}')$ is a maximal monotone operator.
\textbf{(3)} We write $\mathrm{S} = \mathrm{S}_1 + \ldots +\mathrm{S}_7$ and show that each $\mathrm{S}_i$, for $i=1,\ldots,7$, is pseudo-monotone. Let us define
\begin{align*}
& _{\mathbb{X}'}\langle \mathrm{S}_1(\mathbf{u}, \mu_0, \varphi), (\mathbf{v}, \eta , \psi) \rangle_{\mathbb{X}} := \int_\Omega (\mathbf{u} \cdot \nabla ) \mathbf{u} \cdot \mathbf{v} \/\mathrm{d}\/ x, \\
& _{\mathbb{X}'}\langle \mathrm{S}_2(\mathbf{u}, \mu_0, \varphi), (\mathbf{v}, \eta , \psi) \rangle_{\mathbb{X}} := \int_\Omega 2\nu (\varphi) \mathrm{D}\mathbf{u} \cdot \mathrm{D}\mathbf{v} \/\mathrm{d}\/ x,\\
& _{\mathbb{X}'}\langle \mathrm{S}_3(\mathbf{u}, \mu_0, \varphi), (\mathbf{v}, \eta , \psi) \rangle_{\mathbb{X}} := \int_\Omega \mu_0 \nabla \varphi \cdot \mathbf{v} \/\mathrm{d}\/ x, \\
& _{\mathbb{X}'}\langle \mathrm{S}_4(\mathbf{u}, \mu_0, \varphi), (\mathbf{v}, \eta , \psi) \rangle_{\mathbb{X}} := \int_\Omega m(\varphi) \nabla \mu_0 \cdot \nabla \eta \/\mathrm{d}\/ x, \\
& _{\mathbb{X}'}\langle \mathrm{S}_5(\mathbf{u}, \mu_0, \varphi), (\mathbf{v}, \eta , \psi) \rangle_{\mathbb{X}} := \int_\Omega (\mathbf{u} \cdot \nabla \varphi )\eta \/\mathrm{d}\/ x, \\
& _{\mathbb{X}'}\langle \mathrm{S}_6(\mathbf{u}, \mu_0, \varphi), (\mathbf{v}, \eta , \psi) \rangle_{\mathbb{X}} := \int_\Omega \mu_0 \psi \/\mathrm{d}\/ x, \\
& _{\mathbb{X}'}\langle \mathrm{S}_7(\mathbf{u}, \mu_0, \varphi), (\mathbf{v}, \eta , \psi) \rangle_{\mathbb{X}} := \int_\Omega \mathrm{P}_0(\kappa \varphi ) \psi \/\mathrm{d}\/ x.
\end{align*}
Since completely continuous implies pseudo-monotone, we show that $\mathrm{S}_1, \mathrm{S}_3, \mathrm{S}_5,\mathrm{S}_6$ and $\mathrm{S}_7$ are completely continuous operators. Let us denote $\mathbf{x}_n = (\mathbf{u}_n, \mu_{0n}, \varphi_n)$, $\mathbf{x} = (\mathbf{u}, \mu_0, \varphi)$ and $\mathbf{y}=(\mathbf{v}, \eta , \psi)$. Assume that $\mathbf{x}_n \rightharpoonup \mathbf{x}$ in $\mathbb{Y}$. This means $\mathbf{u}_n \rightharpoonup \mathbf{u} $ in $\mathbb{V}_{\mathrm{div}}$ , $\mu_{0n} \rightharpoonup \mu_0 $ in $\mathrm{V}_0$ and $\varphi_n \rightharpoonup \varphi $ in $\mathrm{V}_0$, which in turn gives $\mathbf{u}_n \rightarrow \mathbf{u} $ in $\mathbb{G}_{\mathrm{div}}$, $\mu_{0n} \rightarrow \mu_0 $ in $\mathrm{L}^2_{(0)}(\Omega)$ and $\varphi_n \rightarrow \varphi $ in $\mathrm{L}^2_{(0)}(\Omega)$, using the compact embeddings $\mathbb{V}_{\mathrm{div}}\hookrightarrow\mathbb{G}_{\text{div }}$ and $\mathrm{V}_0\hookrightarrow\mathrm{L}^2_{(0)}(\Omega)$. We have to show that $\mathrm{S}_1 \mathbf{x}_n$ converges strongly to $\mathrm{S}_1 \mathbf{x} $ in $\mathbb{X}'$-norm. Using \eqref{2.7}, H\"older's and Ladyzhenskaya's inequalities, for $n=2$, we get
\begin{align}
|\,_{\mathbb{X}'}\langle \mathrm{S}_1 \mathbf{x}_n, \mathbf{y} \rangle_{\mathbb{X}}-\,_{\mathbb{X}'}\langle \mathrm{S}_1 \mathbf{x}, \mathbf{y} \rangle_{\mathbb{X}}| \nonumber=& |b(\mathbf{u}_n, \mathbf{u}_n,\mathbf{v})- b(\mathbf{u},\mathbf{u},\mathbf{v})| \nonumber \\
\nonumber=& |-b(\mathbf{u}_n, \mathbf{v}, \mathbf{u}_n -\mathbf{u}) -b(\mathbf{u}_n-\mathbf{u},\mathbf{v},\mathbf{u})| \\
\nonumber\leq & \|\mathbf{u}_n \|_{\mathbb{L}^4} \| \nabla \mathbf{v}\| \ \|\mathbf{u}_n-\mathbf{u} \|_{\mathbb{L}^4} + \|\mathbf{u} \|_{\mathbb{L}^4} \| \nabla \mathbf{v}\| \ \|\mathbf{u}_n-\mathbf{u} \|_{\mathbb{L}^4} \\
\label{S12} \leq & 2^{1/4} \mathopen{}\mathclose\bgroup\originalleft(\|\mathbf{u}_n\|^{1/2} \| \nabla \mathbf{u}_n\|^{1/2} + \|\mathbf{u}\|^{1/2} \| \nabla \mathbf{u}\|^{1/2} \aftergroup\egroup\originalright) \|\mathbf{u}_n - \mathbf{u} \|_{\mathbb{L}^4} \| \mathbf{v}\|_{\mathbb{V}_{\mathrm{div}}}. \qquad
\end{align}
For $n=3$, using H\"older's and Ladyzhenskaya's inequalities,, we get
\begin{align} \label{S13}
|\,_{\mathbb{X}'}\langle \mathrm{S}_1 \mathbf{x}_n, \mathbf{y} \rangle_{\mathbb{X}} -\,_{\mathbb{X}'}\langle \mathrm{S}_1 \mathbf{x}, \mathbf{y} \rangle_{\mathbb{X}}| \nonumber \nonumber
\leq & \| \mathbf{u}_n \|_{\mathbb{L}^4} \| \nabla \mathbf{v} \| \ \| \mathbf{u}_n - \mathbf{u} \|_{\mathbb{L}^4} + \| \mathbf{u}_n - \mathbf{u} \|_{\mathbb{L}^4} \| \nabla \mathbf{v} \| \|\mathbf{u} \|_{\mathbb{L}^4} \\ \leq & 2^{1/2}\mathopen{}\mathclose\bgroup\originalleft(\|\mathbf{u}_n\|^{1/4}\|\nabla\mathbf{u}_n\|^{3/4}+\|\mathbf{u}\|^{1/4}\|\nabla\mathbf{u}\|^{3/4}\aftergroup\egroup\originalright)\|\mathbf{u}_n-\mathbf{u}\|_{\mathbb{L}^4}\|\mathbf{v}\|_{\mathbb{V}_{\mathrm{div}}}.
\end{align}
Let us now estimate $|\,_{\mathbb{X}'}\langle \mathrm{S}_3\mathbf{x}_n,\mathbf{y} \rangle - \langle \mathrm{S}_3 \mathbf{x}, \mathbf{y} \rangle_{\mathbb{X}} |$ and $|\,_{\mathbb{X}'}\langle \mathrm{S}_5\mathbf{x}_n,\mathbf{y} \rangle_{\mathbb{X}}- \,_{\mathbb{X}'}\langle \mathrm{S}_5 \mathbf{x}, \mathbf{y} \rangle_{\mathbb{X}}|$. We perform an integration by parts, use H\"older's inequality and the embedding $\mathbb{H}_0^1(\Omega)\hookrightarrow\mathbb{L}^4(\Omega)$ to estimate $|\,_{\mathbb{X}'}\langle \mathrm{S}_3\mathbf{x}_n,\mathbf{y} \rangle - \langle \mathrm{S}_3 \mathbf{x}, \mathbf{y} \rangle_{\mathbb{X}} |$ as
\begin{align} \nonumber
|\,_{\mathbb{X}'}\langle \mathrm{S}_3\mathbf{x}_n,\mathbf{y} \rangle - \langle \mathrm{S}_3 \mathbf{x}, \mathbf{y} \rangle_{\mathbb{X}} |&= \mathopen{}\mathclose\bgroup\originalleft|-\int_\Omega \mu_{0n} \nabla \varphi_n \cdot \mathbf{v} \/\mathrm{d}\/ x +\int_\Omega \mu_0 \nabla \varphi \cdot \mathbf{v} \/\mathrm{d}\/ x\aftergroup\egroup\originalright| \\ \nonumber
&=\mathopen{}\mathclose\bgroup\originalleft|\int_\Omega \mu_{0n} \nabla (\varphi - \varphi_n) \cdot \mathbf{v} \/\mathrm{d}\/ x + \int_\Omega (\mu_0-\mu_{0n})\nabla \varphi \cdot \mathbf{v} \/\mathrm{d}\/ x \aftergroup\egroup\originalright|\\ \nonumber
& \leq \| \nabla \mu_{0n}\| \ \| \mathbf{v}\|_{\mathbb{L}^4} \|\varphi - \varphi_n \|_{\mathrm{L}^4} + \|\mu_0-\mu_{0n} \|_{\mathrm{L}^4} \|\nabla \varphi \| \ \|\mathbf{v} \|_{\mathbb{L}^4} \\ \label{S3}
& \leq C\mathopen{}\mathclose\bgroup\originalleft( \| \nabla \mu_{0n}\| \|\varphi - \varphi_n \|_{\mathrm{L}^4} + \|\mu_0-\mu_{0n} \|_{\mathrm{L}^4} \|\nabla \varphi \|\aftergroup\egroup\originalright) \| \mathbf{v} \|_{\mathbb{V}_{\mathrm{div}}}.
\end{align}
Similarly, we have
\begin{align} \nonumber
|\,_{\mathbb{X}'}\langle \mathrm{S}_5\mathbf{x}_n,\mathbf{y} \rangle_{\mathbb{X}}- \,_{\mathbb{X}'}\langle \mathrm{S}_5 \mathbf{x}, \mathbf{y} \rangle_{\mathbb{X}}| &= \mathopen{}\mathclose\bgroup\originalleft|\int_\Omega (\mathbf{u}_n \cdot \nabla \varphi_n) \eta \/\mathrm{d}\/ x -\int_\Omega (\mathbf{u} \cdot \nabla \varphi) \eta \/\mathrm{d}\/ x \aftergroup\egroup\originalright|\\ \nonumber
&= \mathopen{}\mathclose\bgroup\originalleft|\int_\Omega (\mathbf{u}_n-\mathbf{u}) \cdot \nabla \varphi_n \eta \/\mathrm{d}\/ x+\int_\Omega \mathbf{u} \cdot \nabla (\varphi_n -\varphi)\eta \/\mathrm{d}\/ x\aftergroup\egroup\originalright| \\ \label{S5}
&\leq \mathopen{}\mathclose\bgroup\originalleft( \|\mathbf{u}_n-\mathbf{u} \|_{\mathbb{L}^4}\| \varphi_n\|_{\mathrm{L}^4} +\| \mathbf{u}\|_{\mathbb{L}^4} \ \|\varphi_n -\varphi \| _{\mathrm{L}^4} \aftergroup\egroup\originalright)\|\nabla \eta \|.
\end{align}
Using H\"older's inequality, we obtain
\begin{align}\label{S6}
|\,_{\mathbb{X}'}\langle \mathrm{S}_6 \mathbf{x}_n, \mathbf{y} \rangle_{\mathbb{X}}-\,_{\mathbb{X}'}\langle \mathrm{S}_6 \mathbf{x}, \mathbf{y} \rangle_{\mathbb{X}}| \leq & \int_\Omega |\mu_{0n} - \mu_0|\,|\psi| \/\mathrm{d}\/ x
\leq \|\mu_{0n} - \mu_0 \| \, \|\psi \|
\end{align}
and
\begin{align}\label{S7}
|\,_{\mathbb{X}'}\langle \mathrm{S}_7 \mathbf{x}_n, \mathbf{y} \rangle_{\mathbb{X}} -\,_{\mathbb{X}'}\langle \mathrm{S}_7 \mathbf{x}, \mathbf{y} \rangle_{\mathbb{X}}| &= \mathopen{}\mathclose\bgroup\originalleft|\int_\Omega\mathrm{P}_0(\kappa \varphi_n) \psi \/\mathrm{d}\/ x - \int_\Omega \mathrm{P}_0(\kappa \varphi) \psi \/\mathrm{d}\/ x \aftergroup\egroup\originalright|
\leq \kappa \int_{\Omega} |\varphi_n - \varphi | |\psi| \/\mathrm{d}\/ x \nonumber\\&
\leq \kappa \|\varphi_n -\varphi \| \|\psi \|.
\end{align}
From \eqref{S12}-\eqref{S13}, we get
\begin{align*}
& \| \mathrm{S}_1 \mathbf{x}_n -\mathrm{S}_1\mathbf{x}\|_{\mathbb{X}'} \\ &\leq 2^{1/2} \mathopen{}\mathclose\bgroup\originalleft(\|\mathbf{u}_n\|^{1/2} \| \nabla \mathbf{u}_n\|^{1/2} + \|\mathbf{u}\|^{1/2} \| \nabla \mathbf{u}\|^{1/2} \aftergroup\egroup\originalright)( \|\mathbf{u}_n \|^{1/2}+\|\mathbf{u} \|^{1/2}) \|\mathbf{u}_n - \mathbf{u} \|^{1/2}\ \ (n=2),
\end{align*}
and
\begin{align*}
& \| \mathrm{S}_1 \mathbf{x}_n -\mathrm{S}_1\mathbf{x}\|_{\mathbb{X}'} \\ &\leq 2\mathopen{}\mathclose\bgroup\originalleft(\|\mathbf{u}_n\|^{1/4}\|\nabla\mathbf{u}_n\|^{3/4}+\|\mathbf{u}\|^{1/4}\|\nabla\mathbf{u}\|^{3/4}\aftergroup\egroup\originalright)( \|\mathbf{u}_n \|^{3/4}+\|\mathbf{u} \|^{3/4})\|\mathbf{u}_n-\mathbf{u}\|^{1/4} \ \ (n=3),
\end{align*}
and both converges to $0$ as $n\to\infty$ using the compact embedding of $\mathbb{V}_{\mathrm{div}}\hookrightarrow\mathbb{G}_{\mathrm{div}}$. Using Ladyzhenskaya's inequality and the compact embedding $\mathrm{V}_0 \hookrightarrow \mathrm{L}^2_{(0)}(\Omega)$, from \eqref{S3}-\eqref{S7}, we have
\begin{align*}
\| \mathrm{S}_3\mathbf{x}_n- \mathrm{S}_3\mathbf{x}\|_{\mathbb{X}'} &\leq \| \nabla \mu_{0n}\| \|\varphi - \varphi_n \|_{\mathrm{L}^4} + \|\mu_0-\mu_{0n} \|_{\mathrm{L}^4} \|\nabla \varphi \| \rightarrow 0, \\
\| \mathrm{S}_5\mathbf{x}_n- \mathrm{S}_5\mathbf{x} \|_{\mathbb{X}'} &\leq \|\mathbf{u}_n-\mathbf{u} \|_{\mathbb{L}^4}\| \varphi_n\|_{\mathrm{L}^4} +\| \mathbf{u}\|_{\mathbb{L}^4} \ \|\varphi_n -\varphi \| _{\mathrm{L}^4} \rightarrow 0, \\
\|\mathrm{S}_6 \mathbf{x}_n- \mathrm{S}_6 \mathbf{x} \|_{\mathbb{X}'} & \leq \| \mu_{0n}- \mu_0\| \rightarrow 0, \\
\|\mathrm{S}_7 \mathbf{x}_n- \mathrm{S}_7 \mathbf{x} \|_{\mathbb{X}'} &\leq \kappa \| \varphi_n - \varphi \| \rightarrow 0,
\end{align*}
as $n\to\infty$. This proves that $\mathrm{S}_1, \mathrm{S}_3, \mathrm{S}_5, \mathrm{S}_6$ and $\mathrm{S}_7$ are completely continuous and hence pseudo-monotone.
In order to prove the pseudo-monotonicity of $\mathrm{S}_4$, we use Lemma \ref{pseudo-monotone}. Let us define the operator $\widetilde{\mathrm{S}}_4 : \mathbb{X} \times \mathbb{X} \rightarrow \mathbb{X}'$ by
\begin{align*}
\,_{\mathbb{X}'}\langle \widetilde{\mathrm{S}}_4 (\mathbf{x}_1,\mathbf{x}_2),\mathbf{y} \rangle_{\mathbb{X}} := \int_\Omega m (\varphi_1) \nabla \mu_{0_2} \cdot \nabla \eta\/\mathrm{d}\/ x,
\end{align*}
where $\mathbf{x}_i = (\mathbf{u}_i, \mu_{0_i},\varphi_i) \in \mathbb{X}, i=1,2,$ respectively and $\mathbf{y}=(\mathbf{v},\eta,\psi) \in \mathbb{X}.$ If $\mathbf{x}=(\mathbf{u},\mu_0,\varphi)\in\mathbb{X}$, then
\begin{align*}
\,_{\mathbb{X}'}\langle \widetilde{\mathrm{S}}_4 (\mathbf{x},\mathbf{x}_1 - \mathbf{x}_2) , \mathbf{x}_1 - \mathbf{x}_2 \rangle_{\mathbb{X}} &= \int_\Omega m(\varphi_1) \nabla (\mu_{0_1}- \mu_{0_2}) \cdot \nabla (\mu_{0_1}- \mu_{0_2})\/\mathrm{d}\/ x\\&=\int_\Omega m(\varphi_1) |\nabla (\mu_{0_1}- \mu_{0_2}) |^2\/\mathrm{d}\/ x \geq 0,
\end{align*}
since $m(\cdot)\geq 0$. This implies $\widetilde{\mathrm{S}}_4(\mathbf{x}, \cdot)$ is monotone. Now for each fixed $\mathbf{x}\in\mathbb{X}$, it can be easily seen that $\,_{\mathbb{X}'}\langle \widetilde{\mathrm{S}}_4(\mathbf{x},\mathbf{x}_1+t_k\mathbf{x}_2),\mathbf{y}\rangle_{\mathbb{X}}\to \,_{\mathbb{X}'}\langle \widetilde{\mathrm{S}}_4(\mathbf{x},\mathbf{x}_1),\mathbf{y}\rangle_{\mathbb{X}}$, as $t_k\to0$. Hence, the operator $\widetilde{\mathrm{S}}_4(\mathbf{x}, \cdot)$ is hemicontinuous for each fixed $\mathbf{x}\in\mathbb{X}$. In order to prove complete continuity of $\widetilde{\mathrm{S}}_4(\cdot ,\mathbf{x})$ for all $\mathbf{x} \in \mathbb{X}$, we consider a sequence $\widetilde{\mathbf{x}}_k \subseteq \mathbb{X}$ such that $\widetilde{\mathbf{x}}_k \rightharpoonup \widetilde{\mathbf{x}}$ in $\mathbb{X}$ and $\widetilde{\mathbf{x}}=(\widetilde{\mathbf{u}}, \widetilde{\mu}_0, \widetilde{\varphi}) \in \mathbb{X}$. Then
\begin{align*}
\,_{\mathbb{X}'}\langle \widetilde{\mathrm{S}}_4(\widetilde{\mathbf{x}}_k ,\mathbf{x}) - \widetilde{\mathrm{S}}_4(\widetilde{\mathbf{x}} ,\mathbf{x}) , \mathbf{y}\rangle_{\mathbb{X}} &= \int_\Omega (m(\widetilde{\varphi}_k) \nabla \mu_0 - m(\widetilde{\varphi}) \nabla \mu_0) \cdot \nabla \mathbf{v}\/\mathrm{d}\/ x \\
&\leq \| m(\widetilde{\varphi}_k) \nabla \mu_0 - m(\widetilde{\varphi}) \nabla \mu_0 \| \ \|\mathbf{v} \|_{\mathbb{V}_{\mathrm{div}}}.
\end{align*}
We know that $\widetilde{\varphi}_k \rightarrow \widetilde{\varphi}$ in $\mathrm{L}_{(0)}^2(\Omega)$ and $m$ is a continuous function, we have to show that $\| m(\widetilde{\varphi}_k) \nabla \mu_0 - m(\widetilde{\varphi}) \nabla \mu_0 \| \rightarrow 0$ as $k\to\infty$. Let us define
$$\mathrm{H}(\widetilde{\varphi})(x):=g(x,\widetilde{\varphi}(x))=m(\widetilde{\varphi}(x))\nabla \mu_0(x).$$ Then Lemma 1.19, \cite{ruzicka2006nichtlineare} yields that $\mathrm{H}:\mathrm{L}^2_{(0)}(\Omega)\to\mathrm{L}^2(\Omega)$ is continuous and bounded. Since $\widetilde{\varphi}_k \rightarrow \widetilde{\varphi}$ in $\mathrm{L}_{(0)}^2(\Omega)$, this implies that $\| m(\widetilde{\varphi}_k) \nabla \mu_0 - m(\widetilde{\varphi}) \nabla \mu_0 \| \rightarrow 0$ as $k\to\infty$ and hence $\widetilde{\mathrm{S}}_4(\cdot,\mathbf{x})$ is completely continuous for all $\mathbf{x}\in\mathbb{X}$.
Using Lemma \ref{pseudo-monotone}, we get that the operator $\mathrm{S}_4(\mathbf{x})=\widetilde{\mathrm{S}}_4(\mathbf{x},\mathbf{x})$ is pseudo-monotone . A similar proof follows for $\mathrm{S}_2$, since it is of the same form as $\mathrm{S}_4$ and $\nu(\cdot)$ is a continuously differentiable function.
Since each of the operators $\mathrm{S}_i$'s are bounded, the operator $\mathrm{S}$ is also bounded. Since pseudo-monotonicity and locally boundedness implies demicontinuity (see Lemma 2.4, \cite{MR3014456} or Proposition 27.7, \cite{MR1033498}), $\mathrm{S}$ is demicontinuous.
\textbf{(4)} Observe that $\mathbb{Y}$ is unbounded. Choose $\mathbf{x}_0 =(\mathbf{0},0,0) \in \mathbb{Y} \cap \mathrm{D}(\mathrm{T}).$ Then we have to show that there is $R>0$ such that
\begin{align*}
\,_{\mathbb{X}'}\langle \mathrm{S}(\mathbf{x})-\mathbf{b}, \mathbf{x} \rangle_{\mathbb{X}} >0, \ \text{ for all } \ \mathbf{x} \in \mathbb{Y} \ \text{ with } \ \|\mathbf{x}\|_{\mathbb{X}} >R.
\end{align*}
For $\mathbf{x}=(\mathbf{u},\mu_0,\varphi)\in\mathbb{Y}$, let us consider
\begin{align}\label{com}
_{\mathbb{X}'}\langle \mathrm{S}(\mathbf{x})-\mathbf{b}, \mathbf{x} \rangle_{\mathbb{X}} \nonumber&= \int_\Omega2 \nu(\varphi)\mathrm{D}\mathbf{u} \cdot \mathrm{D}\mathbf{u} \/\mathrm{d}\/ x + \int_\Omega m(\varphi) \nabla \mu_0 \cdot \nabla \mu_0 \/\mathrm{d}\/ x- \int_\Omega \mu_0 \varphi \/\mathrm{d}\/ x \\
\nonumber&\quad- \int_\Omega \mathrm{P}_0(\kappa \varphi) \varphi \/\mathrm{d}\/ x - \langle \mathbf{h} , \mathbf{u} \rangle\\
&=: \mathrm{I}_1+ \mathrm{I}_2+ \mathrm{I}_3 + \mathrm{I}_4 +\mathrm{I}_5.
\end{align}
Since $|\varphi(x)| \leq 1$, and $\nu(\cdot)$ is a continuously differentiable function, we have
\begin{align*}
\mathrm{I}_1 = \int_\Omega 2\nu(\varphi) \mathrm{D}\mathbf{u} \cdot\mathrm{D}\mathbf{u} \/\mathrm{d}\/ x\geq 2\int_{\Omega}\nu(\varphi)|\nabla\mathbf{u}|^2\/\mathrm{d}\/ x \geq \widetilde{C}_1 \|\mathbf{u} \| ^2 _{\mathbb{V}_{\mathrm{div}}}.
\end{align*}
Now, since mean value of $\mu_0$ is $0$ and $\mu(\cdot)$ is a continuous function, we also obtain
\begin{align*}
\mathrm{I}_2=\int_\Omega m(\varphi) \nabla \mu_0 \cdot \nabla \mu_0 \/\mathrm{d}\/ x=\int_\Omega m(\varphi) |\nabla \mu_0|^2 \/\mathrm{d}\/ x \geq \widetilde{C}_2 \| \mu_0\|^2_{\mathrm{V}_0}.
\end{align*}
Using H\"older's, Poincar\'e's and Young's inequalities, we get
\begin{align*}
\mathrm{I}_3&=\int_\Omega \mu_0 \varphi \/\mathrm{d}\/ x \leq \int_{\Omega}|\mu_0|\/\mathrm{d}\/ x\leq |\Omega|^{1/2}\|\mu_0\|_{\mathrm{L}^2}\leq \widetilde{C}_3 \| \mu_0 \|_{\mathrm{V}_0},\\
\mathrm{I}_4&=\int_\Omega \mathrm{P}_0(\kappa \varphi) \varphi \/\mathrm{d}\/ x = \kappa \| \varphi \|^2 \leq \widetilde{C}_4,\\
\mathrm{I}_5&= \langle \mathbf{h} , \mathbf{u} \rangle \leq \|\mathbf{h}\|_{\mathbb{V}_{\mathrm{div}}'}\|\mathbf{u}\|_{\mathbb{V}_{\mathrm{div}}} \leq \frac{1}{2\widetilde{C}_1} \| \mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}}^2 +\frac{ \widetilde{C}_1}{2} \|\mathbf{u} \|^2_{\mathbb{V}_{\mathrm{div}}},
\end{align*}
for some constants $\widetilde{C}_1, \widetilde{C}_2, \widetilde{C}_3, \widetilde{C}_4 >0$. Combining the above inequalities and substituting in \eqref{com} yields
\begin{align*}
_{\mathbb{X}'}\langle \mathrm{S}(\mathbf{x})-\mathbf{b}, \mathbf{x} \rangle_{\mathbb{X}} & \geq \frac{\widetilde{C}_1}{2} \|\mathbf{u} \| ^2_{\mathbb{V}_{\mathrm{div}}} + \underbrace{\| \mu_0 \|_{\mathrm{V}_0}\mathopen{}\mathclose\bgroup\originalleft(\widetilde{C}_2 \| \mu_0 \|_{\mathrm{V}_0} -\widetilde{C}_3\aftergroup\egroup\originalright)}_{:=g\mathopen{}\mathclose\bgroup\originalleft(\|\mu_0\|_{\mathrm{V}_0}\aftergroup\egroup\originalright)}-\widetilde{C}_4 - \frac{1}{2\widetilde{C}_1} \| \mathbf{h}\|_{\mathbb{V}_{\mathrm{div}}'}^2
\end{align*}
for $\mathbf{x}=(\mathbf{u}, \mu_0, \varphi) \in \mathbb{Y}$. Since $\mathbf{h}\in\mathbb{G}$ is fixed, we can choose a constant $K>0$ large enough such that
\begin{align*}
\widetilde{C}_4 + \frac{1}{2\widetilde{C}_1}\| \mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}}^2 \leq K.
\end{align*}Furthermore, it can be easily seen that
$$\lim_{x\to+\infty}g(x)=+\infty.$$
Since $K$ is chosen and $g (\cdot)$ grows to infinity, we can choose $R>0$ such that
\begin{align*}
_{\mathbb{X}'}\langle \mathrm{S}(\mathbf{x})-\mathbf{b}, \mathbf{x} \rangle_{\mathbb{X} }>0 \ \text{ for all } \ \mathbf{x}=(\mathbf{u},\mu_0,\varphi)\in\mathbb{Y} \ \text{ such that } \ \|\mathbf{x}\|_{\mathbb{X}} >R.
\end{align*}
Hence using Theorem \ref{Browder}, there exists $(\mathbf{u}, \mu_0,\varphi) \in \mathbb{Y} \cap \mathrm{D}(\mathrm{T})$ such that $\mathbf{b} \in \mathrm{T}(\mathbf{u}, \mu_0,\varphi) +\mathrm{S}(\mathbf{u}, \mu_0,\varphi)$, which completes the proof.
\end{proof}
\begin{theorem} \label{mainthm}
Let $\Omega \subset \mathbb{R}^n$, $n=2,3$ be a bounded subset. Let $\mathbf{h} \in \mathbb{V}'_{{\mathrm{div}}}$ and $k \in (-1,1),$ then under the Assumptions (A1)-(A9), there exists a weak solution $(\mathbf{u}, \mu,\varphi )$ to the system \eqref{steadysys} such that
\begin{align*}
\mathbf{u} \in \mathbb{V}_{{\mathrm{div}}}, \quad \quad \varphi \in \mathrm{V} \cap \mathrm{L}^2_{(k)}(\Omega) \quad \mathrm{and} \quad \mu \in \mathrm{V}.
\end{align*}
and satisfies the weak formulation \eqref{weakphi}-\eqref{weakform nse}.
\end{theorem}
\begin{proof}
Lemma \ref{browderproof} gives the existence of $(\mathbf{u}, \mu_0,\varphi) \in \mathbb{Y} \cap \mathrm{D}(\mathrm{T})$ such that $\mathbf{b} \in \mathrm{T}(\mathbf{u}, \mu_0,\varphi) +\mathrm{S}(\mathbf{u}, \mu_0,\varphi)$. From Lemma \ref{oper_reform} we get that $(\mathbf{u}, \mu_0,\varphi)$ satisfies the reformulated problem \eqref{reformphi}-\eqref{nsreform}. Hence from Lemma \ref{reformtoweak}, we know that $(\mathbf{u}, \mu,\varphi)$ satisfies the weak formulation \eqref{weakphi}-\eqref{weakform nse}, which completes the proof.
\end{proof}
\begin{remark}\label{F'phi}
Note that, since $\mu \in \mathrm{V}$ and $\varphi \in \mathrm{V}\cap \mathrm{L}^2_{(k)}(\Omega)$ and $\|a\|_{\mathrm{L}^\infty} \leq \|\mathrm{J}\|_{\mathrm{L}^1}$, from \eqref{weakmu}, we have
\begin{align*}
\mathopen{}\mathclose\bgroup\originalleft| \int_\Omega\mathrm{F}'(\varphi)\psi \/\mathrm{d}\/ x\aftergroup\egroup\originalright|= \mathopen{}\mathclose\bgroup\originalleft| \int_\Omega\mu \psi \/\mathrm{d}\/ x- \int_\Omega(a \varphi-\mathrm{J} * \varphi)\psi \/\mathrm{d}\/ x \aftergroup\egroup\originalright|\leq (\|\mu \| + \|a\|_{\mathrm{L}^{\infty}} \|\varphi\| + \|\mathrm{J}\|_{\mathrm{L}^1} \|\varphi\| ) \|\psi\|,
\end{align*}
for all $\psi \in \mathrm{H}$, so that $\|\mathrm{F}'(\varphi)\| \leq \|\mu\| + 2 \|\mathrm{J}\|_{\mathrm{L}^1}\|\varphi\| < \infty$.
\end{remark}
\section{Uniqueness of Weak Solutions and Regularity}\setcounter{equation}{0}\label{sec4} In this section, we prove that the weak solution to the system \eqref{steadysys} obtained in Theorem \ref{mainthm} is unique. In this section, we assume that the viscosity coefficient $\nu$ and the mobility parameter $m$ are positive constants. We also prove some regularity results for the solution.
\subsection{Uniqueness} Recall that our existence result ensures that $\varphi(\cdot) \in [-1, 1]$, a.e. (see \eqref{Zdef}). Then, we have the following uniqueness theorem.
\begin{theorem}\label{unique-steady}
Let $(\mathbf{u}_i,\mu_i,\varphi_i) \in \mathbb{V}_{{\mathrm{div}}} \times \mathrm{V} \times (\mathrm{V}\cap \mathrm{L}^2_{(k)}(\Omega))$ for $i=1,2$ be two \emph{weak solutions} of the following system (satisfied in the weak sense).
\begin{equation}
\mathopen{}\mathclose\bgroup\originalleft\{
\begin{aligned}\label{3.2}
\mathbf{u} \cdot \nabla \varphi &= m\Delta \mu, \ \emph{ in }\ \Omega, \\
\mu &= a \varphi - \mathrm{J}*\varphi + \mathrm{F}'(\varphi) \\
-\nu \Delta \mathbf{u} + (\mathbf{u} \cdot \nabla) \mathbf{u} + \nabla \uppi &= \mu \nabla \varphi + \mathbf{h}, \ \emph{ in }\ \Omega,\\
\emph{div }\mathbf{u} &= 0, \ \emph{ in }\ \Omega, \\
\frac{\partial \mu}{ \partial \mathbf{n}}=0, \ \mathbf{u}&=0, \ \emph{ on }\ \partial\Omega,
\end{aligned}
\aftergroup\egroup\originalright.
\end{equation}
where $\mathbf{h}\in\mathbb{V}_{{\mathrm{div}}}'$. For $n=2$, we have $\mathbf{u}_1=\mathbf{u}_2$ and $\varphi_1=\varphi_2$, provided
\begin{align*}
& (i) \ \nu^2 > \frac{2\sqrt{2}}{\sqrt{\lambda_1}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}} + \frac{12 \nu }{ \lambda_1m C_0} \|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2,\\
&(ii) \ (\nu m)^2 \mathopen{}\mathclose\bgroup\originalleft( \frac{C_0}{4} -\frac{C}{C_0} \|\nabla \mathrm{J}\|^2_{\mathbb{L}^1} \aftergroup\egroup\originalright)> \nu m \mathopen{}\mathclose\bgroup\originalleft(\frac{C}{\lambda_1}\aftergroup\egroup\originalright) + \frac{2 C }{ C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{2}{\lambda_1} \aftergroup\egroup\originalright)^{\frac{1}{2}} \| \mathbf{h} \|^2 _{\mathbb{V}'_{\mathrm{div}}} \ \text{ with }\ \|\nabla\mathrm{J}\|^2_{\mathbb{L}^1}<\frac{C^2_0}{4C},
\end{align*}
Similarly, for $n=3$, we have $\mathbf{u}_1=\mathbf{u}_2$ and $\varphi_1=\varphi_2$, provided
\begin{align*}
&(i)\ \nu^2 > \mathopen{}\mathclose\bgroup\originalleft( \frac{16}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}} +\frac{12 \nu }{ \lambda_1m C_0} \|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2, \\
&(ii)\ (\nu m)^2 \mathopen{}\mathclose\bgroup\originalleft( \frac{C_0}{4} -\frac{C}{C_0} \|\nabla \mathrm{J}\|_{\mathbb{L}^1}^2 \aftergroup\egroup\originalright)> \nu m \mathopen{}\mathclose\bgroup\originalleft(\frac{C}{\lambda_1}\aftergroup\egroup\originalright) + \frac{2 C }{ C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}} \| \mathbf{h} \|^2 _{\mathbb{V}'_{\mathrm{div}}}, \ \text{ with }\ \|\nabla\mathrm{J}\|^2_{\mathbb{L}^1}<\frac{C^2_0}{4C},
\end{align*}
where $C$ is a generic constant depends on the embeddings.
\end{theorem}
\begin{proof}
Let us first find a simple bound for the velocity field. We take inner product of the first equation in \eqref{3.2} with $\mu$ and third equation with $\mathbf{u}$ to obtain
\begin{align} \label{phi}
(\mathbf{u} \cdot \nabla \varphi, \mu)= -m \| \nabla \mu \|^2
\end{align}
and
\begin{align} \label{u}
\nu \| \nabla \mathbf{u}\|^2 = (\mu \nabla \varphi, \mathbf{u}) +\langle\mathbf{h}, \mathbf{u}\rangle.
\end{align}
Adding \eqref{phi} and \eqref{u}, we obtain
\begin{align}\label{1p}
m\|\nabla\mu\|^2+\nu\|\nabla\mathbf{u}\|^2=\langle\mathbf{h}, \mathbf{u}\rangle,
\end{align}
where we used the fact that $(\mathbf{u} \cdot \nabla \varphi, \mu)=(\mu \nabla \varphi, \mathbf{u})$. From \eqref{1p}, we infer that
\begin{align*}
\nu\|\nabla\mathbf{u}\|^2 \leq m\|\nabla\mu\|^2 + \nu\|\nabla\mathbf{u}\|^2 =\langle\mathbf{h},\mathbf{u}\rangle \leq \| \mathbf{h} \|_{\mathbb{V}_{\mathrm{div}}'} \|\mathbf{u} \|_{\mathbb{V}_{\mathrm{div}}}.
\end{align*}
Finally, we have
\begin{align} \label{gradu estimate}
\nu \|\nabla\mathbf{u}\| \leq \| \mathbf{h} \|_{\mathbb{V}_{\mathrm{div}}'}.
\end{align}
Let $(\mathbf{u}_1, \mu_1,\varphi_1) $ and $(\mathbf{u}_2,\mu_2, \varphi_2)$ be two weak solutions of the system \eqref{3.2}. Note that the averages of $\varphi_1$ and $\varphi_2$ are same and are equal to $k$, which gives $\overline{\varphi}_1 - \overline{\varphi}_2=0$.
We can rewrite the third equation in \eqref{3.2} as
\begin{equation}\label{nonlin u rewritten}
-\nu \Delta \mathbf{u} + (\mathbf{u}\cdot \nabla )\mathbf{u} + \nabla \widetilde{\uppi}_{\mathbf{u}} = - \nabla a\frac{{\varphi}^2}{2} - (\mathrm{J}\ast \varphi)\nabla \varphi + \mathbf{h} \ \text{ in }\ \mathbb{V}_{{\mathrm{div}}}',
\end{equation}
where $\widetilde{\uppi}_{\mathbf{u}} :=\widetilde{\uppi} = \uppi -\mathopen{}\mathclose\bgroup\originalleft( \mathrm{F}(\varphi) + a\frac{{\varphi}^2}{2}\aftergroup\egroup\originalright)$. Let us define $\mathbf{u}^e:=\mathbf{u}_1-\mathbf{u}_2$, $\varphi^e:=\varphi_1-\varphi_2$ and $\widetilde{\uppi}^e:=\widetilde{\uppi}_{\mathbf{u}_1}-\widetilde{\uppi}_{\mathbf{u}_2}$. Then for every $\mathbf{v} \in \mathbb{V}_{\mathrm{div}}$ and $\psi \in \mathrm{V}, \, (\mathbf{u}^e,\varphi^e)$ satisfies the following:
\begin{align*}
(\mathbf{u}^e \cdot \nabla \varphi_1, \psi)+(\mathbf{u}_2\cdot\nabla\varphi^e,\psi )&= -m(\nabla \mu^e , \nabla \psi), \\
\nu (\nabla \mathbf{u}^e, \nabla \mathbf{v} )+ b(\mathbf{u}_1,\mathbf{u}^e,\mathbf{v})+ b(\mathbf{u}^e,\mathbf{u}_2,\mathbf{v})
& = -\frac{1}{2}\mathopen{}\mathclose\bgroup\originalleft(\varphi^e( \varphi_1+\varphi_2)\nabla a , \mathbf{v}\aftergroup\egroup\originalright)- ((\mathrm{J}*\varphi^e)\nabla\varphi_2 , \mathbf{v})\nonumber\\
&\quad- ((\mathrm{J}*\varphi_1)\nabla\varphi^e,\mathbf{v}),
\end{align*}
Let us choose $\mathbf{v} = \mathbf{u}^e$ and $\psi = \mathcal{B}^{-1} \varphi^e$ to get
\begin{align} \label{s48}
(\mathbf{u}^e \cdot \nabla \varphi_1, \mathcal{B}^{-1} \varphi^e)+(\mathbf{u}_2\cdot\nabla\varphi^e, \mathcal{B}^{-1} \varphi^e) &= m\langle \Delta \mu^e ,\mathcal{B}^{-1} \varphi^e \rangle, \\ \label{s49}
\nu\|\nabla\mathbf{u}^e\|^2&= -b(\mathbf{u}^e,\mathbf{u}_2,\mathbf{u}^e)-\frac{1}{2}(\nabla a\varphi^e( \varphi_1+\varphi_2),\mathbf{u}^e)\nonumber\\&\quad -( (\mathrm{J}*\varphi^e)\nabla\varphi_2,\mathbf{u}^e) - ((\mathrm{J}*\varphi_1)\nabla\varphi^e,\mathbf{u}^e),
\end{align}
where we used the fact that $b(\mathbf{u}_1, \mathbf{u}^e,\mathbf{u}^e)=0$.
Using \eqref{bes}, Taylor's series expansion and (A9), we estimate the term $(\nabla \mu^e ,\nabla \mathcal{B}^{-1} \varphi^e)$ from \eqref{s48} as
\begin{align}\label{3.13}
-\langle-\Delta \mu^e,\mathcal{B}^{-1} \varphi^e\rangle & =-(\mu^e, \varphi^e )
= -(a \varphi^e - \mathrm{J}*\varphi^e + \mathrm{F}'(\varphi_1)-\mathrm{F}'(\varphi_2),\varphi^e ) \nonumber \\
& = -((a + \mathrm{F}''(\varphi_1+ \theta \varphi_2)) \varphi^e, \varphi^e ) + (\mathrm{J}*\varphi^e, \varphi^e) \nonumber\\
& \leq -C_0 \| \varphi^e \|^2 + (\mathrm{J}*\varphi^e, \varphi^e).
\end{align}
Using \eqref{3.13} in \eqref{s48}, we obtain
\begin{align}\label{3.14}
mC_0\|\varphi^e\|^2 \leq -(\mathbf{u}^e\cdot\nabla\varphi_1,\mathcal{B}^{-1}\varphi^e)-(\mathbf{u}_2\cdot\nabla\varphi^e,\mathcal{B}^{-1}\varphi^e) + m(\mathrm{J}\ast \varphi^e , \varphi^e).
\end{align}
Let us now estimate the terms in the right hand side of \eqref{s49} and \eqref{3.14} one by one. We use the H\"older's, Ladyzhenskaya's, Young's and Poincar\'{e}'s inequalities to estimate $|b(\mathbf{u}^e,\mathbf{u}_2,\mathbf{u}^e)|$.
For $n=2$, we get
\begin{equation}\label{3.15a}
|b(\mathbf{u}^e,\mathbf{u}_2,\mathbf{u}^e)|\leq \|\mathbf{u}^e\|_{\mathbb{L}^4}^2\|\nabla\mathbf{u}_2\|\leq \sqrt{2}\|\mathbf{u}^e\|\|\nabla\mathbf{u}^e\|\|\nabla\mathbf{u}_2\|\leq \mathopen{}\mathclose\bgroup\originalleft( \frac{2}{\lambda_1} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\nabla\mathbf{u}^e\|^2 \|\nabla\mathbf{u}_2\|,
\end{equation}
and for $n=3$, we have
\begin{equation}\label{3.15b}
|b(\mathbf{u}^e,\mathbf{u}_2,\mathbf{u}^e)|\leq \|\mathbf{u}^e\|_{\mathbb{L}^4} ^2\|\nabla\mathbf{u}_2\| \leq 2\|\mathbf{u}^e\|^{\frac{1}{2}} \|\nabla\mathbf{u}^e\|^{\frac{3}{2}} \|\nabla\mathbf{u}_2\| \leq \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\nabla\mathbf{u}^e\|^2 \|\nabla\mathbf{u}_2\|.
\end{equation}
Since $(\mathbf{u}_i, \varphi_i)$ are weak solutions of \eqref{3.2} for $i=1,2,$ such that $\varphi_i \in [-1, 1] $, a.e., we have, $ |\varphi_i (x) | \leq 1$.
Using H\"older's, Ladyzhenskaya's and Young's inequalities and boundedness of $\varphi$, we obtain
\begin{align}\label{3.16a}
\mathopen{}\mathclose\bgroup\originalleft|\frac{1}{2}(\nabla a\varphi^e( \varphi_1+\varphi_2),\mathbf{u}^e)\aftergroup\egroup\originalright| &\leq \frac{1}{2}\|\nabla a\|_{\mathbb{L}^{\infty}}\|\varphi^e\|\|\varphi_1+\varphi_2\|_{\mathrm{L}^{\infty}}\|\mathbf{u}^e\| \nonumber\\
&\leq \frac{1}{2}\|\nabla a\|_{\mathbb{L}^{\infty}}\|\varphi^e\|(\|\varphi_1\|_{\mathrm{L}^{\infty}} +\|\varphi_2\|_{\mathrm{L}^{\infty}}) \|\mathbf{u}^e\| \nonumber\\
&\leq \frac{m C_0}{8}\|\varphi^e\|^2+\frac{2}{m C_0}\|\nabla a\|_{\mathbb{L}^{\infty}}^2\|\mathbf{u}^e\|^2.
\end{align}
In order to estimate $((\mathrm{J}*\varphi^e)\nabla\varphi_2,\mathbf{u}^e)$ and $((\mathrm{J}*\varphi_1)\nabla\varphi^e,\mathbf{u}^e)$, we write these terms using an integration by parts and the divergence free condition as
\begin{align*}
((\mathrm{J}*\varphi^e)\nabla\varphi_2,\mathbf{u}^e) = -((\nabla\mathrm{J}*\varphi^e)\varphi_2,\mathbf{u}^e), \\
((\mathrm{J}*\varphi_1)\nabla\varphi^e,\mathbf{u}^e) = -((\nabla\mathrm{J}*\varphi_1)\varphi^e,\mathbf{u}^e).
\end{align*}
Using H\"older's and Ladyzhenskaya's inequalities, and Young's inequality for convolution, we estimate $|((\nabla\mathrm{J}*\varphi^e)\varphi_2,\mathbf{u}^e)|$ as
\begin{align}\label{3.17a}
|((\nabla\mathrm{J}*\varphi^e)\varphi_2,\mathbf{u}^e)| &\leq \| \nabla \mathrm{J}*\varphi^e\|_{\mathbb{L}^2}\|\varphi_2\|_{\mathrm{L}^\infty} \|\mathbf{u}^e\|\leq \| \nabla \mathrm{J}\|_{\mathbb{L}^1} \|\varphi^e\| \|\mathbf{u}^e\| \nonumber \\
&\leq \frac{m C_0}{8}\|\varphi^e\|^2 + \frac{2}{m C_0}\| \nabla \mathrm{J}\|_{\mathbb{L}^1}^2 \|\mathbf{u}^e\|^2.
\end{align}
Similarly, we obtain
\begin{align}\label{3.18a}
\mathopen{}\mathclose\bgroup\originalleft|((\nabla\mathrm{J}*\varphi_1)\varphi^e,\mathbf{u}^e)\aftergroup\egroup\originalright| &\leq \frac{m C_0}{8}\|\varphi^e\|^2 + \frac{2}{m C_0}\| \nabla \mathrm{J}\|_{\mathbb{L}^1}^2 \|\mathbf{u}^e\|^2 .
\end{align}
Substituting \eqref{3.15a}, \eqref{3.16a}-\eqref{3.18a} in \eqref{s49}, then using \eqref{gradu estimate} and the fact that $\| \nabla a \|_{\mathrm{L}^\infty} \leq \|\nabla \mathrm{J} \|_{\mathbb{L}^1}$, for $n =2$, we obtain
\begin{subequations}
\begin{align}\label{3.19a}
\nu\|\nabla\mathbf{u}^e\|^2 & \leq \frac{3 m C_0}{8}\|\varphi^e\|^2 + \frac{1}{\nu} \mathopen{}\mathclose\bgroup\originalleft( \frac{2}{\lambda_1} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}} \|\nabla\mathbf{u}^e\|^2 +\frac{6}{ m C_0} \|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2 \|\mathbf{u}^e\|^2.
\end{align}
Combining \eqref{3.15b}, \eqref{3.16a}-\eqref{3.18a} and substituting it in \eqref{s49}, for $n=3$, we get
\begin{align} \label{3.19b}
\nu\|\nabla\mathbf{u}^e\|^2 & \leq \frac{3 m C_0}{8}\|\varphi^e\|^2 + \frac{1}{\nu} \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}} \|\nabla\mathbf{u}^e\|^2 +\frac{6}{ m C_0} \|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2 \|\mathbf{u}^e\|^2.
\end{align}
\end{subequations}
Now we estimate the terms in the right hand side of \eqref{3.14}. To estimate $(\mathbf{u}^e\cdot\nabla\varphi_1,\mathcal{B}^{-1}\varphi^e)$, we use an integration by parts, $\mathbf{u}^e\big|_{\partial\Omega}=0$ and the divergence free condition of $\mathbf{u}^e$ to obtain
\begin{align*}
(\mathbf{u}^e\cdot \nabla \varphi_1,\mathcal{B}^{-1}\varphi^e) = -(\mathbf{u}^e\cdot \nabla \mathcal{B}^{-1}\varphi^e,\varphi_1).
\end{align*}
Using H\"older's, Ladyzhenskaya's, Poincar\'e's and Young's inequalities, we estimate the above term as
\begin{align}\label{3.20a}
|(\mathbf{u}^e\cdot \nabla \mathcal{B}^{-1}\varphi^e,\varphi_1)| & \leq \|\mathbf{u}^e\| \, \|\nabla \mathcal{B}^{-1}\varphi^e\| \, \|\varphi_1\|_{\mathrm{L}^\infty}\nonumber\\
&\leq \|\mathbf{u}^e\| \, \|\nabla \mathcal{B}^{-1}\varphi^e \| \nonumber\\
&\leq \mathopen{}\mathclose\bgroup\originalleft( \frac{1}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright) \|\nabla\mathbf{u}^e\|\|\nabla \mathcal{B}^{-1}\varphi^e \|\nonumber\\
&\leq \frac{\nu}{2} \|\nabla\mathbf{u}^e\|^2 +\frac{1}{2\nu \lambda_1} \|\mathcal{B}^{-1/2}\varphi^e \|^2,
\end{align}
where $\lambda_1$ is the first eigenvalue of the Stokes operator $\mathrm{A}$.
Next we estimate $(\mathbf{u}_2\cdot\nabla\varphi^e,\mathcal{B}^{-1}\varphi^e)$ in the following way:
\begin{align}\label{3.21}
|(\mathbf{u}_2\cdot\nabla\varphi^e,\mathcal{B}^{-1}\varphi^e)| &\leq |(\mathbf{u}_2\cdot\nabla \mathcal{B}^{-1}\varphi^e,\varphi^e)| \nonumber \\
&\leq \|\mathbf{u}_2\|_{\mathbb{L}^4} \|\nabla \mathcal{B}^{-1} \varphi^e\|_{\mathbb{L}^4}\|\varphi^e\| \nonumber \\
&\leq \frac{m C_0}{8}\|\varphi^e\|^2 + \frac{2}{m C_0}\|\nabla \mathcal{B}^{-1}\varphi^e \|_{\mathbb{L}^4}^2\|\mathbf{u}_2\|_{\mathbb{L}^4}^2 \nonumber \\
& \leq \frac{m C_0}{8}\|\varphi^e\|^2 + \frac{2C}{m C_0} \|\nabla \mathcal{B}^{-1}\varphi^e\|_{\mathbb{H}^1}^2 \|\mathbf{u}_2\|_{\mathbb{L}^4}^2
\end{align}
where we used the embedding $\mathbb{H}^1 \hookrightarrow \mathbb{L}^4$. One can see that the $\mathrm{H}^2$-norm of $\zeta$ in $\mathrm{D}(\mathcal{B})$ is equivalent to the $\mathrm{L}^2$-norm of $(\mathcal{B} + \mathrm{I})\zeta $, that is,
$$\| \zeta \|_{\mathrm{H}^2} \cong \|(\mathcal{B} + \mathrm{I})\zeta\|.$$
Now since $ \mathcal{B}^{-1}\varphi^e \in \mathrm{D}(\mathcal{B})$ and $\mathcal{B}$ is a linear isomorphism, we have
\begin{align*}
\|\nabla \mathcal{B}^{-1}\varphi^e\|_{\mathbb{H}^1} \leq C\|\mathcal{B}^{-1}\varphi^e \|_{\mathrm{H}^2} \leq C \|(\mathcal{B}+\mathrm{I})\mathcal{B}^{-1}\varphi^e\| \leq C \|\varphi^e\|.
\end{align*}
Substituting in \eqref{3.21}, using Ladyzhenskaya's inequality and \eqref{gradu estimate} , for $n=2$,we obtain
\begin{subequations}
\begin{align}\label{3.22}
|(\mathbf{u}_2\cdot\nabla\varphi^e,\mathcal{B}^{-1}\varphi^e)|
&\leq \frac{m C_0}{8}\|\varphi^e\|^2 + \frac{2C}{m C_0} \|\varphi^e\|^2 \, \|\mathbf{u}_2\|_{\mathbb{L}^4}^2 \nonumber \\
&\leq \frac{m C_0}{8}\|\varphi^e\|^2 + \frac{2 C }{m C_0}\|\varphi^e\|^2 \mathopen{}\mathclose\bgroup\originalleft( \frac{\sqrt{2}}{\nu ^2\sqrt{\lambda_1}} \aftergroup\egroup\originalright) \| \mathbf{h} \|^2 _{\mathbb{V}'_{\mathrm{div}}},
\end{align}
and for $n=3$, we get
\begin{align} \label{s50}
|(\mathbf{u}_2\cdot\nabla\varphi^e,\mathcal{B}^{-1}\varphi^e )| \leq \frac{m C_0}{8}\|\varphi^e\|^2 + \frac{2C}{m C_0}\|\varphi^e\|^2 \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}} \frac{1}{\nu^2} \| \mathbf{h} \|^2 _{\mathbb{V}'_{\mathrm{div}}} .
\end{align}
\end{subequations}
It is only left to estimate $(\mathrm{J}\ast \varphi^e , \varphi^e)$.
Since $\bar{\varphi}=0$ we can write using the fact that $\|B_N^{1/2} \varphi\|^2 = (B_N \varphi,\varphi) = \|\nabla \varphi\|^2$ for all $\varphi \in D(B_N)$
\begin{align*}
|(\mathrm{J}*\varphi,\varphi)|&=|(B_N^{1/2}(\mathrm{J}*\varphi-\overline{\mathrm{J}*\varphi}), B_N^{-1/2} \varphi)| \\
&\leq \| B_N^{1/2}(\mathrm{J}*\varphi-\overline{\mathrm{J}*\varphi})\| \| B_N^{-1/2} \varphi\| \\
&\leq \|\nabla \mathrm{J} * \varphi\| \| B_N^{-1/2} \varphi\| \leq \|\nabla \mathrm{J}\|_{\mathrm{L}^1} \|\varphi\| \| B_N^{-1/2} \varphi\| \\
&\leq \frac{C_0}{4} \|\varphi\|^2 + \frac{1}{C_0}\|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2\| B_N^{-1/2} \varphi\|^2
\end{align*}
which implies
\begin{align}\label{3.23}
m|(\mathrm{J}*\varphi,\varphi)|\leq \frac{mC_0}{4} \|\varphi\|^2 + \frac{m}{C_0}\|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2\| B_N^{-1/2} \varphi\|^2
\end{align}
Combining \eqref{3.20a}, \eqref{3.22} and \eqref{3.23} and substituting it in \eqref{3.14}, for $n=2$, we infer
\begin{subequations}
\begin{align} \label{3.24a}
\frac{5m C_0}{8}\|\varphi^e\|^2 & \leq \frac{\nu}{2}\|\nabla\mathbf{u}^e\|^2 + \mathopen{}\mathclose\bgroup\originalleft( \frac{1}{2\nu \lambda_1}+\frac{m}{C_0}\|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2 \aftergroup\egroup\originalright)\|\mathcal{B}^{-1/2}\varphi^e \|^2 + \frac{2 C }{\nu ^2m C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{2}{\lambda_1} \aftergroup\egroup\originalright)^{\frac{1}{2}} \| \mathbf{h} \|^2 _{\mathbb{V}'_{\mathrm{div}}}\|\varphi^e \|^2 .
\end{align}
Combining \eqref{3.20a}, \eqref{s50} and \eqref{3.23} and substituting in \eqref{3.14}, for $n=3$, we find
\begin{align}\label{3.24b}
\frac{5m C_0}{8} \|\varphi^e\|^2 &\leq \frac{\nu}{2}\|\nabla\mathbf{u}^e\|^2 + \mathopen{}\mathclose\bgroup\originalleft( \frac{1}{2\nu \lambda_1}+\frac{m}{C_0}\|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2 \aftergroup\egroup\originalright) \|\mathcal{B}^{-1/2}\varphi^e\|^2 + \frac{2 C }{\nu^2 m C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}}\| \mathbf{h} \|^2_{\mathbb{V}'_{\mathrm{div}}} \|\varphi^e\|^2 .
\end{align}
\end{subequations}
Adding \eqref{3.19a} and \eqref{3.24a}, for $n=2$, we obtain
\begin{subequations}
\begin{align}
&\frac{\nu}{2} \|\nabla\mathbf{u}^e\|^2 + \frac{m C_0}{4}\|\varphi^e\|^2 \leq \frac{1}{\nu}\mathopen{}\mathclose\bgroup\originalleft( \frac{2}{\lambda_1} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}} \|\nabla\mathbf{u}^e\|^2 + \frac{6}{ m C_0} \| \nabla \mathrm{J}\|_{\mathbb{L}^1} ^2 \|\mathbf{u}^e\|^2 \nonumber \\
& \hspace{4cm}+ \mathopen{}\mathclose\bgroup\originalleft( \frac{1}{2 \nu \lambda_1}+\frac{m}{C_0}\|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2 \aftergroup\egroup\originalright) \|\mathcal{B}^{-1/2}\varphi^e \|^2 + \frac{2 C }{\nu ^2m C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{2}{\lambda_1} \aftergroup\egroup\originalright)^{\frac{1}{2}} \| \mathbf{h} \|^2 _{\mathbb{V}'_{\mathrm{div}}}\|\varphi^e\|^2.
\end{align}
Combining \eqref{3.19b} and \eqref{3.24b}, for $n =3$, we get
\begin{align}
& \frac{\nu}{2} \|\nabla\mathbf{u}^e\|^2 + \frac{m C_0}{4} \|\varphi^e\|^2 \leq \frac{1}{\nu} \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}}\|\nabla\mathbf{u}^e\|^2 +\frac{6}{m C_0} \| \nabla \mathrm{J}\|_{\mathbb{L}^1}^2 \|\mathbf{u}^e\|^2\nonumber \\
&\hspace{4cm}+ \mathopen{}\mathclose\bgroup\originalleft( \frac{1}{2 \nu\lambda_1} + \frac{m}{C_0}\|\nabla \mathrm{J}\|^2_{\mathbb{L}^1}\aftergroup\egroup\originalright) \|\mathcal{B}^{-1/2}\varphi^e \|^2 + \frac{2C }{\nu^2 m C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}}\| \mathbf{h} \|^2_{\mathbb{V}'_{\mathrm{div}}} \|\varphi^e\|^2 .
\end{align}
\end{subequations}
Now using the continuous embedding of $\mathrm{L}^2_{(0)} \hookrightarrow \mathrm{V}_0'$, that is, $ \|\mathcal{B}^{-1/2}\varphi^e \|^2 \leq C \| \varphi^e \|^2$ and Poincar\'e's inequality, we further obtain
\begin{align}\label{3.25}
&\mathopen{}\mathclose\bgroup\originalleft[ \lambda_1 \mathopen{}\mathclose\bgroup\originalleft( \frac{\nu}{2} - \frac{1}{\nu}\mathopen{}\mathclose\bgroup\originalleft( \frac{2}{\lambda_1} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}} \aftergroup\egroup\originalright) - \frac{6}{ m C_0} \|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2 \aftergroup\egroup\originalright] \| \mathbf{u}^e\|^2 \nonumber \\
& +\Bigg[\frac{mC_0}{4} -\frac{mC}{C_0}\|\nabla \mathrm{J}\|^2_{\mathbb{L}^1}- \frac{C}{2 \nu \lambda_1}
- \frac{2C }{\nu ^2m C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{2}{\lambda_1} \aftergroup\egroup\originalright)^{\frac{1}{2}} \| \mathbf{h} \|^2 _{\mathbb{V}'_{\mathrm{div}}} \Bigg]\|\varphi^e\|^2\leq 0.
\end{align}
From \eqref{3.25}, uniqueness for $n= 2$ follows provided quantities in both brackets in above inequality are strictly positive. Thus we conclude that for
\begin{align*}
& (i) \ \nu^2 > \frac{2\sqrt{2}}{\sqrt{\lambda_1}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}} + \frac{12 \nu }{ \lambda_1m C_0} \|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2,\\
&(ii) \ (\nu m)^2 \mathopen{}\mathclose\bgroup\originalleft( \frac{C_0}{4} -\frac{C}{C_0} \|\nabla \mathrm{J}\|^2_{\mathbb{L}^1} \aftergroup\egroup\originalright)> \nu m \mathopen{}\mathclose\bgroup\originalleft(\frac{C}{\lambda_1}\aftergroup\egroup\originalright) + \frac{2 C }{ C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{2}{\lambda_1} \aftergroup\egroup\originalright)^{\frac{1}{2}} \| \mathbf{h} \|^2 _{\mathbb{V}'_{\mathrm{div}}} \ \text{ with }\ \|\nabla\mathrm{J}\|^2_{\mathbb{L}^1}<\frac{C^2_0}{4C},
\end{align*}
uniqueness follows in two dimensions. Similarly for $n=3$, we obtain
\begin{align*}
& \mathopen{}\mathclose\bgroup\originalleft[ \lambda_1 \Bigg(\frac{\nu}{2 }-\frac{1}{\nu} \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}} \Bigg)-\frac{6}{ m C_0} \|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2 \aftergroup\egroup\originalright] \| \mathbf{u}^e\|^2 \nonumber \\
&+\Bigg[ \frac{mC_0}{4} - \frac{mC}{C_0} \|\nabla \mathrm{J}\|_{\mathbb{L}^1}^2 - \frac{C}{2\nu \lambda_1} - \frac{2 C }{\nu^2 m C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}}\| \mathbf{h} \|^2_{\mathbb{V}'_{\mathrm{div}}} \Bigg] \|\varphi^e \|^2 \leq 0.
\end{align*}
Hence, the uniqueness follows provided
\begin{align*}
&(i)\ \nu^2 > \mathopen{}\mathclose\bgroup\originalleft( \frac{16}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}} \|\mathbf{h}\|_{\mathbb{V}'_{\mathrm{div}}} +\frac{12 \nu }{ \lambda_1m C_0} \|\nabla \mathrm{J} \|_{\mathbb{L}^1}^2, \\
&(ii)\ (\nu m)^2 \mathopen{}\mathclose\bgroup\originalleft( \frac{C_0}{4} -\frac{C}{C_0} \|\nabla \mathrm{J}\|_{\mathbb{L}^1}^2 \aftergroup\egroup\originalright)> \nu m \mathopen{}\mathclose\bgroup\originalleft(\frac{C}{\lambda_1}\aftergroup\egroup\originalright) + \frac{2 C }{ C_0} \mathopen{}\mathclose\bgroup\originalleft( \frac{4}{\sqrt{\lambda_1}} \aftergroup\egroup\originalright)^{\frac{1}{2}} \| \mathbf{h} \|^2 _{\mathbb{V}'_{\mathrm{div}}}, \ \text{ with }\ \|\nabla\mathrm{J}\|^2_{\mathbb{L}^1}<\frac{C^2_0}{4C},
\end{align*}
which completes the proof.
\end{proof}
\subsection{Regularity of the weak solution}
In this subsection, we establish the regularity results for the weak solution to the system \eqref{3.2}. Let $(\mathbf{u}, \mu, \varphi ) \in \mathbb{V}_{\mathrm{div}} \times \mathrm{V} \times (\mathrm{V}\cap \mathrm{L}^2_{(k)}(\Omega)) $ and $\varphi(x) \in (-1,1) \ a.e.$ be the unique weak solution of the system \eqref{3.2}. In the next theorem, we prove the higher order regularity results for the system \eqref{3.2}.
\begin{theorem}\label{regularity}
Let $\mathbf{h}\in \mathbb{G}_{\mathrm{div}}$ and the assumptions of Theorem \ref{unique-steady} are satisfied. Then the weak solution $(\mathbf{u}, \mu, \varphi)$ of the system \eqref{3.2} satisfies:
\begin{align}\label{s3}
\mathbf{u}\in\mathbb{H}^2(\Omega),\ \ \mu \in \mathrm{H}^2(\Omega) \ \text{ and } \ \varphi \in \mathrm{W}^{1,p} (\Omega),
\end{align}
where $2\leq p < \infty$ if $n=2$ and $2 \leq p \leq 6$ if $n=3.$
\end{theorem}
\begin{proof}
\textbf{Step 1:} Let $ \mathbf{h} \in \mathbb{G}_{\mathrm{div}} $. Then from Theorem \ref{mainthm} and Theorem \ref{unique-steady}, we know that there exists a unique weak solution $ (\mathbf{u}, \varphi, \mu) $ satisfying
\begin{align}
\mathbf{u} \in \mathbb{V}_{\mathrm{div}}, \varphi \in \mathrm{V} \cap \mathrm{L}^2_{(k)} \ \mathrm{and} \ \mu \in \mathrm{V}. \label{reg1}
\end{align}
Now, we recall the form \eqref{nonlin u rewritten} that is
\begin{align} \label{reg2}
\nu (\nabla \mathbf{u}, \nabla \mathbf{v}) + b(\mathbf{u},\mathbf{u}, \mathbf{v}) = - \mathopen{}\mathclose\bgroup\originalleft< \nabla a\frac{{\varphi}^2}{2} - (\mathrm{J}\ast \varphi)\nabla \varphi , \mathbf{v} \aftergroup\egroup\originalright>+ \langle \mathbf{h},\mathbf{v} \rangle, \ \text{ for all }\ \mathbf{v} \in \mathbb{V}_{\mathrm{div}},
\end{align}
where $\nabla a(x) = \int_\Omega \nabla \mathrm{J}(x-y) \/\mathrm{d}\/ y$. Using the weak regularity \eqref{reg1}, $ |\varphi (x)| \leq 1 \ $ and Young's inequality for convolution, we observe that
\begin{align*}
\| (\nabla a) \varphi^2\|& \leq \|\nabla a\|_{\mathbb{L}^{\infty}}\|\varphi\|_{\mathrm{L}^4}^2\leq C ,\\
\|\mathrm{J}\ast \varphi \nabla \varphi\|& \leq \|\mathrm{J}\|_{\mathrm{L}^1}\|\varphi\|_{\mathrm{L}^\infty}\|\nabla \varphi\| \leq C.
\end{align*}
Hence, the right hand side of \eqref{reg2} belongs to $\mathbb{L}^2(\Omega)$. Then from the regularity of the steady state Navier-Stokes equations (see Chapter II, \cite{MR0609732}) we have that
\begin{align}
\mathbf{u} \in \mathbb{H}^2(\Omega). \label{reg4}
\end{align}
\vskip 0.1cm \noindent
\textbf{Step 2:} From the first equation of \eqref{3.2} we have
\begin{align}
m ( \nabla \mu, \nabla \psi) = (\mathbf{u} \cdot \nabla \varphi, \psi),\ \text{ for all }\ \psi \in \mathrm{V}. \label{reg3}
\end{align}
Using the Sobolev inequality, \eqref{reg4} and \eqref{reg1}, we observe that
\begin{align}
\|\mathbf{u} \cdot \nabla \varphi \| \leq C\|\mathbf{u}\|_{\mathbb{L}^\infty}\|\nabla \varphi\| \leq C \|\mathbf{u}\|_{\mathbb{H}^2} \|\nabla \varphi\| \leq C.
\end{align}
Since we have that $\frac{\partial \mu }{\partial \mathbf{n}}=0$, from the $L^p$ regularity of the Neumann Laplacian (see Lemma \ref{Lp_reg}) for $p=2$, we conclude that
\begin{align} \label{reg8}
\mu \in \mathrm{H}^2(\Omega).
\end{align}
\vskip 0.1cm \noindent
\textbf{Step 3:} An application of the Gagliardo-Nirenberg inequality together with \eqref{reg8} gives
\begin{align}\label{s9}
\|\nabla\mu\|_{\mathbb{L}^p}&\leq C\|\mu\|_{\mathrm{H}^2}^{1-\frac{1}{p}}\|\mu\|^{\frac{1}{p}} < \infty,
\end{align}
for $2\leq p < \infty$ in case of $n=2$ and
\begin{align} \label{gradmu_n3}
\|\nabla\mu\|_{\mathbb{L}^p} & \leq C\|\mu\|_{\mathrm{H}^2}^{\frac{5p-6}{4p}}\|\mu\|^{\frac{6-p}{4p}} < \infty,
\end{align}
for $2 \leq p \leq 6$ and $n=3.$
Let us take the gradient of $\mu= a\varphi- \mathrm{J}*\varphi + \mathrm{F}'(\varphi)$, multiply it by $\nabla\varphi|\nabla\varphi|^{p-2}$, integrate the resulting identity over $\Omega$, and use (A9), H\"older's and Young's inequalities and Young's inequality for convolution to obtain
\begin{align}\label{s5}
\int_{\Omega}\nabla\varphi|\nabla\varphi|^{p-2}\cdot \nabla\mu\/\mathrm{d}\/ x&=\int_{\Omega}\nabla\varphi|\nabla\varphi|^{p-2}\cdot\mathopen{}\mathclose\bgroup\originalleft(a\nabla\varphi+\nabla a\varphi-\nabla\mathrm{J}*\varphi+\mathrm{F}''(\varphi)\nabla\varphi\aftergroup\egroup\originalright)\/\mathrm{d}\/ x\nonumber\\
&=\int_{\Omega}(a+\mathrm{F}''(\varphi))|\nabla\varphi|^p\/\mathrm{d}\/ x+\int_{\Omega}\nabla\varphi|\nabla\varphi|^{p-2}\cdot\mathopen{}\mathclose\bgroup\originalleft(\nabla a\varphi-\nabla\mathrm{J}*\varphi\aftergroup\egroup\originalright)\/\mathrm{d}\/ x\nonumber\\
&\geq C_0\int_{\Omega}|\nabla\varphi|^p\/\mathrm{d}\/ x-\mathopen{}\mathclose\bgroup\originalleft(\|\nabla a\|_{\mathbb{L}^{\infty}}+\|\nabla\mathrm{J}\|_{\mathbb{L}^1}\aftergroup\egroup\originalright) \|\varphi \|_{\mathrm{L}^p} \| \nabla \varphi \|_{\mathbb{L}^p}^{p-1} \nonumber\\
&\geq \frac{C_0}{2}\|\nabla\varphi\|_{\mathbb{L}^p}^p -\frac{1}{p}\mathopen{}\mathclose\bgroup\originalleft(\frac{2(p-1)}{C_0p}\aftergroup\egroup\originalright)^{p-1}\mathopen{}\mathclose\bgroup\originalleft(\|\nabla a\|_{\mathbb{L}^{\infty}}+\|\nabla\mathrm{J}\|_{\mathbb{L}^1}\aftergroup\egroup\originalright)^p\|\varphi\|_{\mathbb{L}^p}^p.
\end{align}
Using Young's inequality, we also have
\begin{align}\label{s6}
\mathopen{}\mathclose\bgroup\originalleft| \int_{\Omega}\nabla\varphi|\nabla\varphi|^{p-2}\cdot \nabla\mu\/\mathrm{d}\/ x\aftergroup\egroup\originalright|&\leq \int_{\Omega}|\nabla\varphi|^{p-1}|\nabla\mu|\/\mathrm{d}\/ x\nonumber\\&\leq \frac{C_0}{4}\|\nabla\varphi\|_{\mathbb{L}^p}^p+\frac{1}{p}\mathopen{}\mathclose\bgroup\originalleft(\frac{4(p-1)}{C_0p}\aftergroup\egroup\originalright)^{p-1}\|\nabla\mu\|_{\mathbb{L}^p}^p.
\end{align}
Combining \eqref{s5} and \eqref{s6}, we get
\begin{align}
\label{s7}
\frac{C_0}{4}\|\nabla\varphi\|_{\mathbb{L}^p}^p\leq\frac{1}{p}\mathopen{}\mathclose\bgroup\originalleft(\frac{2(p-1)}{C_0p}\aftergroup\egroup\originalright)^{p-1}\mathopen{}\mathclose\bgroup\originalleft[2^{p-1}\|\nabla\mu\|_{\mathbb{L}^p}^p+ \mathopen{}\mathclose\bgroup\originalleft(\|\nabla a\|_{\mathbb{L}^{\infty}}+\|\nabla\mathrm{J}\|_{\mathbb{L}^1}\aftergroup\egroup\originalright)^p\|\varphi\|_{\mathrm{L}^p}^p\aftergroup\egroup\originalright].
\end{align}
Using \eqref{s9} and \eqref{gradmu_n3} we get
\begin{align*}
\|\nabla\varphi\|_{\mathbb{L}^p} < \infty
\end{align*}
for $2\leq p < \infty$ if $n=2$, and $2 \leq p \leq 6$ if $n=3.$
\end{proof}
\section{Exponential Stability}\label{se4}\setcounter{equation}{0}
The stability analysis of non-linear dynamical systems has a long history starting from the works of Lyapunov. For the solutions of ordinary or partial differential equations describing dynamical systems, different kinds of stability may be described. One of the most important types of stability is that concerning the stability of solutions near to a point of equilibrium (stationary solutions). In the qualitative theory of ordinary and partial differential equations, and control theory, Lyapunov’s notion of (global) asymptotic stability of an equilibrium is a key concept. It is important to note that the asymptotic stability do not quantify the rate of convergence. In fact, there is a strong form of stability which demands an exponential rate of convergence. The notion of exponential stability is far stronger and it assures a minimum rate of decay, that is, an estimate of how fast the solutions converge to its equilibrium. In particular, exponential stability implies uniform asymptotic stability. Stability analysis of fluid dynamic models has been one of the essential areas of applied mathematics with a good number of applications in engineering and physics (cf. \cite{MR0609732, MR3186318}, etc).
In this section, we consider the singular potential $\mathrm{F}$ in $(-1,1)$ of the form:
\begin{align*}
\mathrm{F} (\varphi) = \frac{\theta}{2} ((1+ \varphi) \log (1+\varphi) + (1-\varphi) \log (1-\varphi)) ,\quad \varphi \in (-1,1).
\end{align*}
Observe that in this case, the potential $\mathrm{F}$ is convex and $\kappa =0$ in \eqref{decomp of F}. For such potentials, we prove that the stationary solution $(\mathbf{u}^e,\varphi^e)$ of the system (\ref{steadysys}) with constant mobility parameter $m$ and constant coefficient of kinematic viscosity $\nu$ is exponentially stable in 2-D. That is, our aim is to establish that:
\begin{itemize}
\item
there exists constants $M>0$ and $\alpha>0$ such that
\begin{align*}
\|\mathbf{u}(t)-\mathbf{u}^e\|^2+\|\varphi (t)-\varphi^e\|^2\leq M e^{-\alpha t},
\end{align*}
for all $t\geq 0$.
\end{itemize}
\subsection{Global solvability of two dimensional CHNS system} We consider the following initial-boundary value problem:
\begin{equation}\label{4.1}
\mathopen{}\mathclose\bgroup\originalleft\{
\begin{aligned}
\varphi_t+ \mathbf{u} \cdot \nabla \varphi &= m\Delta \mu, \ \text{ in }\ \Omega\times(0,T), \\
\mu &= a \varphi - \mathrm{J}*\varphi + \mathrm{F}'(\varphi) \\
\mathbf{u}_t -\nu \Delta \mathbf{u} + (\mathbf{u} \cdot \nabla) \mathbf{u }+ \nabla \uppi &= \mu \nabla \varphi +\mathbf{ h}, \ \text{ in }\ \Omega\times(0,T),\\
\text{div }\mathbf{u}& = 0, \ \text{ in }\ \Omega\times(0,T), \\
\frac{\partial \mu}{ \partial \mathbf{n}}=0,& \ \mathbf{u}=0, \ \text{ on }\ \partial\Omega\times[0,T], \\
\mathbf{u}(0) = \mathbf{u}_0,& \quad \varphi(0)=\varphi_0, , \ \text{ in }\ \Omega.
\end{aligned}
\aftergroup\egroup\originalright.
\end{equation}
Let us now give the global solvability results available in the literature for the system \eqref{4.1}. We first give the definition of a weak solution for the system \eqref{4.1}
\begin{definition}[weak solution, \cite{MR3019479}]
Let $\mathbf{u}_0\in\mathbb{G}_{\mathrm{div}}$, $\varphi_0\in \mathrm{H}$ with $\mathrm{F}(\varphi_0)\in\mathrm{L}^1(\Omega)$ and $0<T<\infty$ be given. A couple $(\mathbf{u},\varphi)$ is a \emph{weak solution} to the system \eqref{4.1} on $[0,T]$ corresponding to $ [\mathbf{u}_0,\varphi_0]$ if
\begin{itemize}
\item [(i)] $\mathbf{u},\varphi$ and $\mu$ satisfy
\begin{equation}\label{sol}
\mathopen{}\mathclose\bgroup\originalleft\{
\begin{aligned}
& \mathbf{u} \in \mathrm{L}^{\infty}(0,T;\mathbb{G}_{\mathrm{div}}) \cap \mathrm{L}^2(0,T;\mathbb{V}_{\mathrm{div}}), \\
& \mathbf{u}_t \in \mathrm{L}^{4/3}(0,T;\mathbb{V}_{\mathrm{div}}'),\text{ if } d=3, \\
& \mathbf{u}_t \in \mathrm{L}^{2}(0,T;\mathbb{V}_{\mathrm{div}}'),\text{ if } d=2, \\
& \varphi \in \mathrm{L}^{\infty}(0,T;\mathrm{H}) \cap \mathrm{L}^2(0,T;\mathrm{V}), \\
& \varphi_t \in \mathrm{L}^{2}(0,T;\mathrm{V}'), \\
& \mu = a \varphi - \mathrm{J}*\varphi + \mathrm{F}'(\varphi) \in L^2(0,T;\mathrm{V}),
\end{aligned}
\aftergroup\egroup\originalright.
\end{equation}
and
$$\varphi \in \mathrm{L}^\infty (Q), \quad |\varphi (x,t)| <1 \ a.e. \ (x,t) \in Q:=\Omega \times (0,T);$$
\item [(ii)] For every $\psi\in\mathrm{V}$, every $\mathbf{v} \in \mathbb{V}_{\mathrm{div}}$ and for almost any $t\in(0,T)$, we have
\begin{align}
\langle \varphi_t,\psi\rangle + m(\nabla \mu, \nabla \psi) &=\int_{\Omega}(\mathbf{u}\cdot\nabla\psi)\varphi\/\mathrm{d}\/ x,\\
\langle \mathbf{u}_t,\mathbf{v}\rangle +\nu(\nabla\mathbf{u},\nabla\mathbf{v})+b(\mathbf{u},\mathbf{u},\mathbf{v})&=-\int_{\Omega}(\varphi\cdot\nabla\mu)\mathbf{v}\/\mathrm{d}\/ x+\langle\mathbf{h},\mathbf{v}\rangle.
\end{align}
\item [(iii)] The initial conditions
$\mathbf{u}(0)=\mathbf{u}_0,\ \varphi(0)=\varphi_{0}$ hold in the weak sense.
\end{itemize}
\end{definition}
Next, we discuss the existence and uniqueness of weak solution results available in the literature for the system \eqref{4.1}.
\begin{theorem}[Existence, Theorem 1 \cite{MR3019479}]\label{exist}
Assume that (A1)-(A7) are satisfied for some fixed positive integer $q$. Let $\mathbf{u}_0 \in \mathbb{G}_{\mathrm{div}}$, $\varphi_0 \in \mathrm{L}^\infty(\Omega)$ such that $\mathrm{F}(\varphi_0) \in \mathrm{L}^1(\Omega)$ and $\mathbf{h} \in \mathrm{L}^2_{\text{loc}}([0,\infty), \mathbb{V}_{\mathrm{div}}')$. In addition, assume that $|\overline{\varphi_0}|<1$. Then, for every given $T>0$, there exists a weak solution $(\mathbf{u},\varphi)$ to the equation \eqref{4.1} on $[0,T]$ such that $\overline{\varphi}(t) = \overline{\varphi_0}$ for all $t \in [0,T]$ and
\begin{align*}
\varphi \in \mathrm{L}^\infty (0,T; \mathrm{L}^{2+2q}(\Omega)).
\end{align*}
Furthermore, setting
\begin{align}\mathscr{E}(\mathbf{u}(t),\varphi(t)) = \frac{1}{2} \|\mathbf{u}(t)\|^2 + \frac{1}{4} \int_\Omega \int_\Omega \mathrm{J}(x-y) (\varphi(x,t) - \varphi(y,t))^2 \/\mathrm{d}\/ x \/\mathrm{d}\/ y + \int_\Omega \mathrm{F}(\varphi(t))\/\mathrm{d}\/ x,\end{align}
the following energy estimate holds
\begin{align}\label{energy}
\mathscr{E}(\mathbf{u}(t),\varphi(t)) + \int_s^t \mathopen{}\mathclose\bgroup\originalleft(2 \|\sqrt{\nu (\varphi)} D \mathbf{u}(s)\|^2 + m\| \nabla\mu(s) \|^2 \aftergroup\egroup\originalright)\/\mathrm{d}\/ s \leq \mathscr{E}(\mathbf{u}(s),\varphi(s)) + \int_s^t \langle \mathbf{h}(s), \mathbf{u}(s) \rangle\/\mathrm{d}\/ s,
\end{align}
for all $t \geq s$ and for a.a. $s \in (0,\infty)$.
If $d=2$, the weak solution $(\mathbf{u},\varphi)$ satisfies the following energy identity,
$$\frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t}\mathscr{E}(\mathbf{u}(t),\varphi(t)) + 2\|\sqrt{\nu (\varphi)} D \mathbf{u}(t) \|^2+ m\| \nabla \mu(t) \|^2 = \langle \mathbf{h}(t) , \mathbf{u}(t) \rangle,$$
that is, the equality in \eqref{energy} holds for every $t \geq 0.$
\end{theorem}
\begin{theorem}[Uniqueness, Theorem 3, \cite{MR3518604}]\label{unique}
Let $d=2$ and assume that (A1)-(A7) are satisfied for some fixed positive integer $q$. Let $\mathbf{u}_0 \in \mathbb{G}_{\mathrm{div}}$, $\varphi_0 \in \mathrm{L}^\infty (\Omega)$ with $\mathrm{F}(\varphi_0) \in \mathrm{L}^1(\Omega), |\overline{\varphi_0}|<1$ and $\mathbf{h} \in \mathrm{L}^2_{\text{loc}}([0,\infty);\mathbb{V}_{\mathrm{div}}')$. Then, the weak solution $(\mathbf{u},\varphi)$ corresponding to $(\mathbf{u}_0,\varphi_0)$ given by Theorem \ref{exist} is unique. Furthermore, for $i=1,2$, let $\mathbf{z}_i := (\mathbf{u}_i,\varphi_i)$ be two weak solutions corresponding to two initial data $\mathbf{z}_{0i} := (\mathbf{u}_{0i},\varphi_{0i})$ and external forces $\mathbf{h}_i$, with $\mathbf{u}_{0i} \in \mathbb{G}_{\mathrm{div}}$, $\varphi_{0i} \in \mathrm{L}^\infty (\Omega)$ such that $\mathrm{F}(\varphi_{0i}) \in \mathrm{L}^1(\Omega), |\overline{\varphi_{0i}}| < \eta ,$ for some constant $\eta \in [0,1)$ and $\mathbf{h}_i \in \mathrm{L}^2_{\text{loc}}([0,\infty);\mathbb{V}_{\mathrm{div}}')$. Then the following continuous dependence estimate holds:
\begin{align*}
&\| \mathbf{u}_2(t) - \mathbf{u}_1(t) \|^2 + \| \varphi_2(t) - \varphi_1(t) \|^2_{\mathrm{V}'} \\
&\quad + \int_0^t \mathopen{}\mathclose\bgroup\originalleft( \frac{C_0}{2} \| \varphi_2(\tau) - \varphi_1(\tau) \|^2 + \frac{\nu}{4} \|\nabla( \mathbf{u}_2(\tau) - \mathbf{u}_1(\tau)) \|^2 \aftergroup\egroup\originalright) d \tau \\
&\leq \mathopen{}\mathclose\bgroup\originalleft( \|\mathbf{u}_2(0) - \mathbf{u}_1(0) \|^2 + \| \varphi_2(0) - \varphi_1(0) \|^2_{\mathrm{V}'} \aftergroup\egroup\originalright) \Lambda_0(t) \\
&\quad + \| \overline{\varphi}_2(0) - \overline{\varphi}_1(0)\| \mathbb{Q} \mathopen{}\mathclose\bgroup\originalleft( \mathcal{E}(z_{01}),\mathcal{E}(z_{02}),\| \mathbf{h}_1 \|_{\mathrm{L}^2(0,t;\mathbb{V}_{div}')},\| \mathbf{h}_2 \|_{\mathrm{L}^2(0,t;\mathbb{V}_{div}')}, \eta \aftergroup\egroup\originalright) \Lambda_1(t) \\
&\quad + \| \mathbf{h}_2 - \mathbf{h}_1 \|^2_{\mathrm{L}^2(0,T;\mathbb{V}_{\mathrm{div}}')} \Lambda_2(t) ,
\end{align*}
for all $t \in [0,T]$, where $\Lambda_0(t)$, $\Lambda_1(t)$ and $\Lambda_2(t)$ are continuous functions, which depend on the norms of the two solutions. The functions $\Lambda_i(t)$ also depend on $\mathrm{F}$, $\mathrm{J}$ and $\Omega$, and $\mathbb{Q}$ depends on $\mathrm{F}$, $\mathrm{J}$ ,$\Omega$ and $\eta$.
\end{theorem}
\begin{remark}[Remark 3, \cite{MR3019479}]
The above theorems also imply $\mathbf{u}\in\mathrm{C}([0,T];\mathbb{G}_{\mathrm{div}})$ and $\varphi\in \mathrm{C}([0,T];\mathrm{H})$, for all $T>0$.
\end{remark}
Now, we prove that the stationary solution of \eqref{4.1} is exponentially stable in two dimensions. Let $(\mathbf{u}^e, \varphi^e)$ be the steady-state solution of the system \eqref{3.2}. From Theorem \ref{mainthm}, we know that there exists a unique weak solution for the system \eqref{3.2}. Remember that for $\mathbf{h}\in\mathbb{G}_{\mathrm{div}}$, the weak solution $(\mathbf{u}^e, \mu^e, \varphi^e)$ of the system \eqref{3.2} has the following regularity:
\begin{align*}
\nabla \varphi^e \in\mathrm{L}^p(\Omega),\ \mu^{e} \in \mathrm{H}^2(\Omega)\ \text{ and } \ \mathbf{u}^e\in\mathbb{H}^2(\Omega), \qquad 2 \leq p < \infty.
\end{align*}
Now, in the system \eqref{4.1}, we assume that $\mathbf{h}\in\mathbb{G}_{\text{div }}$ is independent of time and the initial data $\mathbf{u}_0\in\mathbb{G}_{\mathrm{div}}$, $\varphi_0\in\mathrm{L}^{\infty}(\Omega)$. Then, there exists a unique weak solution $(\mathbf{u},\varphi)$ of \eqref{4.1} such that the system has the regularity given in \eqref{sol}.
\begin{theorem}\label{thmexp} Let $\mathbf{u}_0\in\mathbb{G}_{\mathrm{div}}$, $\varphi_0\in\mathrm{L}^{\infty}(\Omega)$ and $\mathbf{h}\in\mathbb{G}_{\text{div }}$. Then under the Assumptions of Theorem \ref{unique-steady}, for
\begin{align*}
&(i)\ \nu^2>\frac{4}{\lambda_1 }\|\nabla\mathbf{u}^e\|^2,\\
&(ii)\ (C_0-\|\mathrm{J}\|_{\mathrm{L}^1})^2>2 C_{\Omega}^2\mathopen{}\mathclose\bgroup\originalleft(\frac{1}{2m}\|\mathbf{u}^e\|_{\mathbb{L}^{\infty}}^2+\frac{4}{\nu \sqrt{\lambda_1}}\|\nabla\mu^e\|_{\mathbb{L}^4}^2\aftergroup\egroup\originalright), \ \text{ with }\ C_0>\|\mathrm{J}\|_{\mathrm{L}^1}, \\
&(iii) \ \overline{\varphi}_0 = \overline{\varphi}^e,
\end{align*}
the stationary solution $(\mathbf{u}^e, \varphi^e)$ of \eqref{3.2}, with the regularity given in Theorem \ref{regularity}, is exponentially stable. That is, there exist constants $M>0$ and $\varrho>0$ such that
\begin{align*}
\|\mathbf{u}(t)-\mathbf{u}^e\|^2+\|\varphi(t)-\varphi^e\|^2\leq M e^{-\varrho t},\end{align*}
for all $t\geq 0$, where $(\mathbf{u}, \varphi)$ is the unique weak solution of the system \eqref{4.1},
$$M=\frac{\|\mathbf{y}_0\|^2 + 4\|\mathrm{J}\|_{\mathrm{L}^{1}}(\|\varphi_0\|^2+\|\varphi^{e}\|^2)+ \|\mathrm{F}(\varphi_0) \|_{\mathrm{L}^1} + \|\mathrm{F}(\varphi^e)\|_{\mathrm{L}^1} + \|\mathrm{F}'(\varphi^e)\| (\|\varphi_0\|+\|\varphi^{e}\|)}{\min\{(C_0-\|\mathrm{J}\|_{\mathrm{L}^1}),1\}}$$
and
\begin{align}
\varrho=\min\mathopen{}\mathclose\bgroup\originalleft\{\mathopen{}\mathclose\bgroup\originalleft(\lambda_1\nu-\frac{4}{\nu} \|\nabla\mathbf{u}^e\|^2\aftergroup\egroup\originalright) ,\mathopen{}\mathclose\bgroup\originalleft[\frac{m(C_0-\|\mathrm{J}\|_{\mathrm{L}^1})}{2C_{\Omega}^2}-\frac{1}{(C_0-\|\mathrm{J}\|_{\mathrm{L}^1})}\mathopen{}\mathclose\bgroup\originalleft(\|\mathbf{u}^e\|_{\mathbb{L}^{\infty}}^2+\frac{4}{\nu \sqrt{\lambda_1}}\|\nabla\mu^e\|_{\mathbb{L}^4}^2\aftergroup\egroup\originalright)\aftergroup\egroup\originalright]\aftergroup\egroup\originalright\}>0.
\end{align}
From the hypothesis of Theorems \ref{exist} and \ref{unique}, settings of Theorem \ref{mainthm} (see Lemma \ref{browderproof} also) and Remark \ref{F'phi}, note that the quantity $M$ is finite.
\end{theorem}
\begin{proof}
Let us define $\mathbf{y}:=\mathbf{u}-\mathbf{u}^e$, $\psi:=\varphi- \varphi^e$, $ \widetilde{\mu} = \mu - \mu^e $ and $\widetilde{\uppi} :=\uppi - \uppi^e$. Then we know that $(\mathbf{y},\psi )$ satisfies the following system in the weak sense:
\iffalse
\begin{align*}
y_t- \nu \Delta y + (y \cdot \nabla ) y + (y \cdot \nabla ) u^e+ (u^e \cdot \nabla ) y + \nabla \widetilde{\uppi} = \widetilde{\mu} \nabla \psi \\
\psi_t + y \cdot \nabla \psi + y \cdot \nabla \varphi^e + u^e \cdot \nabla \psi = \Delta \widetilde{\mu}
\end{align*}where $\widetilde{\mu}= a\psi - J*\psi + F'(\psi+ \varphi^e)- F'(\varphi^e)$.
So we consider the following system \fi
\begin{eqnarray}\label{4.3}
\mathopen{}\mathclose\bgroup\originalleft\{
\begin{aligned}
\psi_t + \mathbf{y} \cdot \nabla \psi + \mathbf{y} \cdot \nabla \varphi^e + \mathbf{u}^e \cdot \nabla \psi =& m\Delta \widetilde{\mu}, \ \text{ in }\ \Omega\times (0,T),\\
\widetilde{\mu}=& a\psi - \mathrm{J}*\psi + \mathrm{F}'(\psi+\varphi^e )- \mathrm{F}'(\varphi^e),\\
\mathbf{y}_t- \nu \Delta \mathbf{y} + (\mathbf{y }\cdot \nabla ) \mathbf{y} + (\mathbf{y} \cdot \nabla ) \mathbf{u}^e+ (\mathbf{u}^e \cdot \nabla ) \mathbf{y} + \nabla \widetilde{\uppi} =& \widetilde{\mu} \nabla \psi + \widetilde{\mu} \nabla \varphi^e + \mu^e \nabla \psi,\\&\qquad \qquad \qquad \ \text{ in }\ \Omega\times (0,T), \\
\text{div }\mathbf{y} = &0, \ \text{ in }\ \Omega\times (0,T),\\
\frac{\partial \widetilde{\mu}}{ \partial \mathbf{n}}=0, \ \mathbf{y}=&0, \ \text{ on }\ \partial \Omega\times(0,T),\\
\mathbf{y}(0) = \mathbf{y}_0, \ \psi(0)=&\psi_0, \ \text{ in }\ \Omega.
\end{aligned}
\aftergroup\egroup\originalright.
\end{eqnarray}
Now consider third equation of \eqref{4.3} and take inner product with $\mathbf{y}(\cdot)$ to obtain
\begin{align}
\label{4.4}
&\frac{1}{2}\frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t}\|\mathbf{y}(t)\|^2+ \nu \|\nabla \mathbf{y}(t)\|^2 + b(\mathbf{y}(t),\mathbf{u}^e,\mathbf{y}(t)) \nonumber\\&\quad= (\widetilde{\mu}(t)\nabla\psi(t),\mathbf{y}(t))+ (\widetilde{\mu}(t) \nabla \varphi^e,\mathbf{y}(t)) + (\mu^e \nabla \psi(t),\mathbf{y}(t)),
\end{align}
where we used the fact that $b(\mathbf{y},\mathbf{y},\mathbf{y})=b(\mathbf{u}^e,\mathbf{y},\mathbf{y})=0$ and $(\nabla\widetilde{\uppi},\mathbf{y})=(\widetilde{\uppi},\nabla\cdot\mathbf{y})=0$.
An integration by parts yields
\begin{align*}
(\widetilde{\mu}\nabla\psi,\mathbf{y})=-(\psi\nabla\widetilde{\mu},\mathbf{y})-(\psi\widetilde{\mu},\nabla\cdot\mathbf{y})=-(\psi\nabla\widetilde{\mu},\mathbf{y}),
\end{align*}
where we used the boundary data and divergence free condition of $\mathbf{y}$. Similarly, we have $(\widetilde{\mu} \nabla \varphi^e,\mathbf{y}) =-(\varphi^e\nabla \widetilde{\mu} ,\mathbf{y})$ and $ (\mu^e \nabla \psi,\mathbf{y})=-( \psi\nabla \mu^e ,\mathbf{y})$. Thus from \eqref{4.4}, we have
\begin{align}
\label{4.5}
&\frac{1}{2}\frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t}\|\mathbf{y}(t)\|^2+ \nu \|\nabla \mathbf{y}(t)\|^2 + b(\mathbf{y}(t),\mathbf{u}^e,\mathbf{y}(t)) \nonumber\\&\quad = -(\psi(t)\nabla\widetilde{\mu}(t),\mathbf{y}(t))-(\varphi^e\nabla \widetilde{\mu}(t) ,\mathbf{y}(t))-( \psi(t)\nabla \mu^e ,\mathbf{y}(t)).
\end{align}
Taking inner product of the third equation in \eqref{4.3} with $\widetilde{\mu}(\cdot)$, we obtain
\begin{align} \label{4.6}
(\psi_t(t),\widetilde{\mu}(t) )+ m\|\nabla \widetilde{\mu}(t)\|^2= -( \mathbf{y}(t) \cdot \nabla \psi (t), \widetilde{\mu}(t))- (\mathbf{y}(t) \cdot \nabla \varphi^e, \widetilde{\mu}(t)) - (\mathbf{u}^e \cdot \nabla \psi(t), \widetilde{\mu}(t)).
\end{align}
Using an integration by parts, divergence free condition and boundary value of $\mathbf{y}$, we get
\begin{align*}
( \mathbf{y} \cdot \nabla \psi ,\widetilde{\mu})=-(\psi\nabla\cdot\mathbf{y},\widetilde{\mu})-(\psi\nabla\widetilde{\mu},\mathbf{y})=-(\psi\nabla\widetilde{\mu},\mathbf{y}).
\end{align*}
Similarly, we have $(\mathbf{y} \cdot \nabla \varphi^e, \widetilde{\mu}) =-( \varphi^e \nabla \widetilde{\mu}, \mathbf{y}) $ and $ (\mathbf{u}^e \cdot \nabla \psi, \widetilde{\mu})= -( \psi \nabla \widetilde{\mu}, \mathbf{u}^e)$. Thus, from \eqref{4.6}, it is immediate that
\begin{align}
\label{4.7}
(\psi_t(t),\widetilde{\mu} (t))+ m\|\nabla \widetilde{\mu}(t)\|^2= (\psi(t)\nabla\widetilde{\mu}(t),\mathbf{y}(t))+( \varphi^e \nabla \widetilde{\mu}(t), \mathbf{y}(t)) + ( \psi (t) \nabla \widetilde{\mu}(t), \mathbf{u}^e).
\end{align}
Adding \eqref{4.5} and \eqref{4.7}, we infer that
\begin{align}\label{4.8}
& \frac{1}{2}\frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t}\|\mathbf{y}(t)\|^2 + \nu \|\nabla \mathbf{y}(t)\|^2 +(\psi_t(t),\widetilde{\mu}(t) )+ m\|\nabla \widetilde{\mu}(t)\|^2\nonumber\\
&\quad= - b(\mathbf{y}(t),\mathbf{u}^e,\mathbf{y}(t))-( \psi(t)\nabla \mu^e ,\mathbf{y}(t))+ ( \psi(t) \nabla \widetilde{\mu}(t), \mathbf{u}^e).
\end{align}
We estimate the term $(\psi_t,\widetilde{\mu} )$ from \eqref{4.8} as
\begin{align}\label{4.10}
(\psi_t, \widetilde{\mu})&= (\psi_t,a\psi - \mathrm{J}*\psi + \mathrm{F}'(\psi+\varphi^e )- \mathrm{F}'(\varphi^e) ) \\ \nonumber
&=\frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t} \mathopen{}\mathclose\bgroup\originalleft\{ \frac{1}{2} \|\sqrt{a} \psi \|^2 -\frac{1}{2} (\mathrm{J}*\psi,\psi) + \int_\Omega \mathrm{F}(\psi+\varphi^e) \/\mathrm{d}\/ x- (\mathrm{F}'(\varphi^e),\psi)\aftergroup\egroup\originalright\} \nonumber\\
&= \frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t} \mathopen{}\mathclose\bgroup\originalleft\{ \frac{1}{2} \|\sqrt{a} \psi \|^2 -\frac{1}{2} (\mathrm{J}*\psi,\psi) + \int_\Omega \mathrm{F}(\psi+\varphi^e) \/\mathrm{d}\/ x-\int_\Omega \mathrm{F}(\varphi^e) \/\mathrm{d}\/ x- \int_{\Omega}\mathrm{F}'(\varphi^e)\psi\/\mathrm{d}\/ x\aftergroup\egroup\originalright\}, \label{6.16}
\end{align}
since $\frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t}\mathopen{}\mathclose\bgroup\originalleft(\int_\Omega \mathrm{F}(\varphi^e) \/\mathrm{d}\/ x\aftergroup\egroup\originalright)=0$. Using H\"older's, Ladyzhenskaya's and Young's inequalities, we estimate $b(\mathbf{y},\mathbf{u}^e,\mathbf{y})$ as
\begin{align}\label{4.13}
|b(\mathbf{y},\mathbf{u}^e,\mathbf{y})|\leq \|\nabla\mathbf{u}^e\|\|\mathbf{y}\|_{\mathbb{L}^4}^2\leq \sqrt{2}\|\nabla\mathbf{u}^e\|\|\mathbf{y}\|\|\nabla\mathbf{y}\|\leq \frac{\nu}{4}\|\nabla\mathbf{y}\|^2+\frac{2}{\nu}\|\nabla\mathbf{u}^e\|^2\|\mathbf{y}\|^2.
\end{align}
We estimate the term $( \psi\nabla \mu^e ,\mathbf{y})$ from \eqref{4.8} using H\"older's, Ladyzhenskaya's and Young's inequalities as
\begin{align}\label{4.14}
|( \psi\nabla \mu^e ,\mathbf{y})|&\leq \|\psi\|\|\nabla\mu^e\|_{\mathbb{L}^4}\|\mathbf{y}\|_{\mathbb{L}^4}\leq \sqrt{2}\|\psi\|\|\nabla\mu^e\|_{\mathbb{L}^4}\|\mathbf{y}\|^{1/2}\|\nabla\mathbf{y}\|^{1/2}\nonumber\\
&\leq \sqrt{2} (\lambda_1)^{1/4}\|\psi\| \|\nabla\mu^e\|_{\mathbb{L}^4} \|\nabla\mathbf{y}\|\nonumber\\
&\leq \frac{\nu}{4}\|\nabla\mathbf{y}\|^2+ \frac{2}{\nu \sqrt{\lambda_1}}\|\nabla\mu^e\|_{\mathbb{L}^4}^2 \|\psi\|^2.
\end{align}
Similarly, we estimate the term $ ( \psi \nabla \widetilde{\mu}, \mathbf{u}^e)$ from \eqref{4.8} as
\begin{align}\label{4.15}
|( \psi \nabla \widetilde{\mu}, \mathbf{u}^e)|&\leq\|\psi\|\|\nabla \widetilde{\mu}\|\|\mathbf{u}^e\|_{\mathbb{L}^{\infty}}\leq \frac{m}{2}\|\nabla \widetilde{\mu}\|^2+\frac{1}{2m}\|\mathbf{u}^e\|_{\mathbb{L}^{\infty}}^2\|\psi\|^2.
\end{align}
Combining \eqref{6.16}-\eqref{4.15} and substituting it in \eqref{4.8}, we obtain
\begin{align}
\label{4.16}
& \frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t}\mathopen{}\mathclose\bgroup\originalleft\{\frac{1}{2}\|\mathbf{y}\|^2 + \frac{1}{2} \|\sqrt{a} \psi \|^2 -\frac{1}{2} (\mathrm{J}*\psi,\psi) + \int_\Omega \mathrm{F}(\psi+\varphi^e) \/\mathrm{d}\/ x-\int_\Omega \mathrm{F}(\varphi^e) \/\mathrm{d}\/ x- \int_{\Omega}\mathrm{F}'(\varphi^e)\psi\/\mathrm{d}\/ x\aftergroup\egroup\originalright\} \nonumber \\
&+ \frac{\nu}{2} \|\nabla \mathbf{y}\|^2 + \frac{m}{2}\|\nabla \widetilde{\mu}\|^2 \leq \mathopen{}\mathclose\bgroup\originalleft(\frac{1}{2m}\|\mathbf{u}^e\|_{\mathbb{L}^{\infty}}^2+\frac{2}{\nu \sqrt{\lambda_1}}\|\nabla\mu^e\|_{\mathbb{L}^4}^2\aftergroup\egroup\originalright)\|\psi\|^2+\frac{2}{\nu}\|\nabla\mathbf{u}^e\|^2\|\mathbf{y}\|^2.
\end{align}
Using Taylor's series expansion and Assumption (A9), we also obtain
\begin{align}\label{433}
(\widetilde{\mu},\psi)&=(a\psi - \mathrm{J}*\psi + \mathrm{F}'(\psi+\varphi^e )- \mathrm{F}'(\varphi^e), \psi)\nonumber\\
&=(a\psi+\mathrm{F}''(\varphi^e +\theta\psi)\psi,\psi)-(\mathrm{J}*\psi,\psi)\nonumber\\
&\geq (C_0-\|\mathrm{J}\|_{\mathrm{L}^1})\|\psi\|^2,
\end{align}
for some $0<\theta<1$. Observe that $(\overline{\tilde{\mu}}, \psi)=\overline{\tilde{\mu}}(1,\psi)=0$, since $\overline{\varphi}_0=\overline{\varphi}= \overline{\varphi}^e$. Thus an application of Poincar\'e inequality yields
\begin{align}\label{exp1}
(\tilde{\mu}, \psi)=(\tilde{\mu}-\overline{\tilde{\mu}}, \psi)\leq\|\tilde{\mu}-\overline{\tilde{\mu}}\|\|\psi\|\leq C_\Omega\|\psi\| \|\nabla \tilde{\mu}\|.
\end{align}
Hence, from \eqref{433} and \eqref{exp1}, we get
\begin{align} \label{exp3}
\|\psi\| \leq \frac{C_\Omega}{C_0 - \|\mathrm{J}\|_{\mathrm{L}^1}} \|\nabla \tilde{\mu}\|,
\end{align}
which implies from \eqref{exp1} that
\begin{align}
(\tilde{\mu}, \psi) \leq \frac{C_\Omega^2}{C_0 - \|\mathrm{J}\|_{\mathrm{L}^1}}\|\nabla \tilde{\mu}\|^2 \label{exp2}
\end{align}
Using Poincar\'e's inequality and \eqref{exp3} in \eqref{4.16}, we get
\begin{align}
& \frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t}\mathopen{}\mathclose\bgroup\originalleft\{\frac{1}{2}\|\mathbf{y}\|^2 + \frac{1}{2} \|\sqrt{a} \psi \|^2 -\frac{1}{2} (\mathrm{J}*\psi,\psi) + \int_\Omega \mathrm{F}(\psi+\varphi^e) \/\mathrm{d}\/ x-\int_\Omega \mathrm{F}(\varphi^e) \/\mathrm{d}\/ x- \int_{\Omega}\mathrm{F}'(\varphi^e)\psi\/\mathrm{d}\/ x\aftergroup\egroup\originalright\} \nonumber \\
&+ \mathopen{}\mathclose\bgroup\originalleft( \frac{\nu \lambda_1}{2} -\frac{2}{\nu}\|\nabla\mathbf{u}^e\|^2 \aftergroup\egroup\originalright)\|\mathbf{y}\|^2+ \mathopen{}\mathclose\bgroup\originalleft( \frac{m}{2} -\mathopen{}\mathclose\bgroup\originalleft( \frac{1}{2m}\|\mathbf{u}^e\|_{\mathbb{L}^{\infty}}^2-\frac{2}{\nu \sqrt{\lambda_1}}\|\nabla\mu^e\|_{\mathbb{L}^4}^2 \aftergroup\egroup\originalright) \frac{C_\Omega^2}{(C_0 - \|\mathrm{J}\|_{\mathrm{L}^1})^2}\aftergroup\egroup\originalright) \|\nabla \widetilde{\mu}\|^2 \leq 0.
\end{align}
Using Taylor's formula, we have the following identities:
\begin{align} \label{exp4}
&\frac{1}{2} \|\sqrt{a} \psi \|^2 -\frac{1}{2} (\mathrm{J}*\psi,\psi)+\int_\Omega \mathopen{}\mathclose\bgroup\originalleft[\mathrm{F}(\psi+\varphi^e)-\mathrm{F}(\varphi^e)- \mathrm{F}'(\varphi^e)\psi\aftergroup\egroup\originalright]\/\mathrm{d}\/ x \nonumber \\
& =(a\psi - \mathrm{J}*\psi, \psi)+ \frac{1}{2}\int_{\Omega}\mathrm{F}''(\varphi^{e}+\theta\psi)\psi^2\/\mathrm{d}\/ x, \nonumber \\
&=(a\psi - \mathrm{J}*\psi + \mathrm{F}'(\psi+\varphi^e )- \mathrm{F}'(\varphi^e), \psi) = (\tilde{\mu}, \psi)
\end{align}
for some $0<\theta<1$.
Using (A9) and Young's inequality for convolutions, we know that
\begin{align*}
\int_{\Omega}\mathopen{}\mathclose\bgroup\originalleft(a+\mathrm{F}''(\varphi^{e}+\theta\psi)\aftergroup\egroup\originalright)\psi^2\/\mathrm{d}\/ x\geq C_0\|\psi\|^2\geq \|\mathrm{J}\|_{\mathbb{L}^1}\|\psi\|^2
\geq (\mathrm{J}*\psi,\psi).
\end{align*}
Thus, we have
\begin{align*}
\int_{\Omega}\mathopen{}\mathclose\bgroup\originalleft(a+\mathrm{F}''(\varphi^{e}+\theta\psi)\aftergroup\egroup\originalright)\psi^2\/\mathrm{d}\/ x- (\mathrm{J}*\psi,\psi)\geq 0.
\end{align*}
Hence from \eqref{exp2}, we obtain
\begin{align}\label{436}
& \frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t}\mathopen{}\mathclose\bgroup\originalleft[\|\mathbf{y}\|^2 + (\widetilde{\mu},\psi)\aftergroup\egroup\originalright] + \mathopen{}\mathclose\bgroup\originalleft(\nu \lambda_1-\frac{4}{\nu}\|\nabla\mathbf{u}^e\|^2\aftergroup\egroup\originalright) \|\mathbf{y}\|^2 \nonumber\\&
\quad + \mathopen{}\mathclose\bgroup\originalleft[\frac{(C_0-\|\mathrm{J}\|_{\mathrm{L}^1})}{2C_{\Omega}^2}-\frac{1}{(C_0-\|\mathrm{J}\|_{\mathrm{L}^1})}\mathopen{}\mathclose\bgroup\originalleft(\frac{1}{2m}\|\mathbf{u}^e\|_{\mathbb{L}^{\infty}}^2+\frac{4}{\nu \sqrt{\lambda_1}}\|\nabla\mu^e\|_{\mathbb{L}^4}^2\aftergroup\egroup\originalright)\aftergroup\egroup\originalright] (\widetilde{\mu},\psi)\leq 0.
\end{align}
Now for $C_0>\|\mathrm{J}\|_{\mathrm{L}^1}$ and
\begin{align}
\nu^2&>\frac{4}{\lambda_1 }\|\nabla\mathbf{u}^e\|^2, \ \text {and }\
(C_0-\|\mathrm{J}\|_{\mathrm{L}^1})^2>2 C_{\Omega}^2\mathopen{}\mathclose\bgroup\originalleft(\frac{1}{2m}\|\mathbf{u}^e\|_{\mathbb{L}^{\infty}}^2+\frac{4}{\nu \sqrt{\lambda_1}}\|\nabla\mu^e\|_{\mathbb{L}^4}^2\aftergroup\egroup\originalright),
\end{align}
we use the variation of constants formula (see Lemma \ref{C3}) to find
\begin{align} \label{s51}
\|\mathbf{y}(t)\|^2 + (\widetilde{\mu}(t),\psi(t))\leq \mathopen{}\mathclose\bgroup\originalleft(\|\mathbf{y}(0)\|^2 + (\widetilde{\mu}(0),\psi(0))\aftergroup\egroup\originalright)e^{-\varrho t},
\end{align}
for all $t\geq 0$, where
\begin{align}
\varrho=\min\mathopen{}\mathclose\bgroup\originalleft\{\mathopen{}\mathclose\bgroup\originalleft(\lambda_1\nu-\frac{4}{\nu} \|\nabla\mathbf{u}^e\|^2\aftergroup\egroup\originalright) ,\mathopen{}\mathclose\bgroup\originalleft[\frac{(C_0-\|\mathrm{J}\|_{\mathrm{L}^1})}{2C_{\Omega}^2}-\frac{1}{(C_0-\|\mathrm{J}\|_{\mathrm{L}^1})}\mathopen{}\mathclose\bgroup\originalleft(\frac{1}{2m}\|\mathbf{u}^e\|_{\mathbb{L}^{\infty}}^2+\frac{4}{\nu \sqrt{\lambda_1}}\|\nabla\mu^e\|_{\mathbb{L}^4}^2\aftergroup\egroup\originalright)\aftergroup\egroup\originalright]\aftergroup\egroup\originalright\}>0.
\end{align}
From \eqref{433}, we also have $ (C_0-\|\mathrm{J}\|_{\mathrm{L}^1})\|\psi\|^2\leq (\widetilde{\mu},\psi),$ so that from \eqref{s51}, we infer that
\begin{align}\label{442}
\|\mathbf{y}(t)\|^2 + (C_0-\|\mathrm{J}\|_{\mathrm{L}^1})\|\psi\|^2\leq \mathopen{}\mathclose\bgroup\originalleft(\|\mathbf{y}(0)\|^2 + (\widetilde{\mu}(0),\psi(0))\aftergroup\egroup\originalright)e^{-\varrho t}.
\end{align}
Using H\"older's inequality, Young's inequality for convolutions and assumption (A1) in the identity \eqref{exp4}, we obtain
\begin{align}
(\widetilde{\mu}(0),\psi(0)) &= (a \psi_0, \psi_0)-(\mathrm{J}*\psi_0, \psi_0) + \int_\Omega \mathrm{F}(\varphi_0) \/\mathrm{d}\/ x+ \int_\Omega \mathrm{F}(\varphi^e)\/\mathrm{d}\/ x - \int_\Omega \mathrm{F}'(\varphi^e) \psi_0\/\mathrm{d}\/ x\nonumber \\
&\leq 2\|\mathrm{J}\|_{\mathrm{L}^{1}}\|\psi_0\|^2+ \|\mathrm{F}(\varphi_0) \|_{\mathrm{L}^1} + \|\mathrm{F}(\varphi^e)\|_{\mathrm{L}^1} + \|\mathrm{F}'(\varphi^e)\|_{\mathrm{L}^1} \|\psi_0\|_{\mathrm{L}^{\infty}} ,
\end{align}
where we have used the boundedness of $\psi_0$ and $\|a\|_{\mathrm{L}^{\infty}} \leq \|\mathrm{J}\|_{\mathrm{L}^{1}}$. Hence from \eqref{442}, we finally have
\begin{align}\label{444}
\|\mathbf{y}(t)\|^2 + \|\psi(t)\|^2 \leq \mathopen{}\mathclose\bgroup\originalleft( \frac{\|\mathbf{y}_0\|^2 + 2\|\mathrm{J}\|_{\mathrm{L}^{1}}\|\psi_0\|^2+ \|\mathrm{F}(\varphi_0) \|_{\mathrm{L}^1} + \|\mathrm{F}(\varphi^e)\|_{\mathrm{L}^1} + \|\mathrm{F}'(\varphi^e)\| \|\psi_0\| }{\min\{(C_0-\|\mathrm{J}\|_{\mathrm{L}^1}),1\}} \aftergroup\egroup\originalright) e^{-\varrho t},
\end{align}
which completes the proof.
\end{proof}
\appendix
\section{Variation of Constants Formula}
\renewcommand{\Alph{section}}{\Alph{section}}
\numberwithin{equation}{section}\setcounter{equation}{0}
In this appendix, we give a variant of variation of constants formula, which is useful when we have two or more differentiable functions with different constant coefficients.
\begin{lemma}[Lemma C.3, \cite{MR4120851}]\label{C3}
Assume that the differentiable functions $y(\cdot),z(\cdot):[0,T]\to[0,\infty)$ and the constants $a_1,a_2,k_1,k_2,k_3>0$ satisfy:
\begin{align}\label{c1}
\frac{\/\mathrm{d}\/ }{\/\mathrm{d}\/ t}(a_1 y(t)+a_2z(t))+k_1y(t)+k_2z(t)\leq 0,
\end{align}
for all $t\in[0,T]$. Then, we have
\begin{align}\label{c2}
y(t)+z(t)\leq C(y(0)+z(0))e^{-\uprho t},\ \text{ where }\ C=\frac{\max\{a_1,a_2\}}{\min\{a_1,a_2\}}\ \text{ and }\ \uprho=\min\mathopen{}\mathclose\bgroup\originalleft\{\frac{k_1}{a_1},\frac{k_2}{a_2}\aftergroup\egroup\originalright\}.
\end{align}
\end{lemma}
\begin{proof}
Since $a_1>0$, from \eqref{c1}, we have
\begin{align*}
\frac{\/\mathrm{d}\/ }{\/\mathrm{d}\/ t}\mathopen{}\mathclose\bgroup\originalleft(y(t)+\frac{a_2}{a_1}z(t)\aftergroup\egroup\originalright)+\frac{k_1}{a_1}\mathopen{}\mathclose\bgroup\originalleft(y(t)+\frac{k_2}{k_1}z(t)\aftergroup\egroup\originalright)\leq 0.
\end{align*}
Now, for $\frac{a_2}{a_1}\leq \frac{k_2}{k_1}$, from the above inequality, we also have
\begin{align*}
\frac{\/\mathrm{d}\/ }{\/\mathrm{d}\/ t}\mathopen{}\mathclose\bgroup\originalleft(y(t)+\frac{a_2}{a_1}z(t)\aftergroup\egroup\originalright)+\frac{k_1}{a_1}\mathopen{}\mathclose\bgroup\originalleft(y(t)+\frac{a_2}{a_1}z(t)\aftergroup\egroup\originalright)\leq 0.
\end{align*}
From the above relation, it is immediate that
\begin{align*}
\frac{\/\mathrm{d}\/}{\/\mathrm{d}\/ t}\mathopen{}\mathclose\bgroup\originalleft[e^{\frac{k_1}{a_1}t}\mathopen{}\mathclose\bgroup\originalleft(y(t)+\frac{a_2}{a_1}z(t)\aftergroup\egroup\originalright)\aftergroup\egroup\originalright]\leq 0,
\end{align*}
which easily implies
\begin{align}\label{c3}
a_1y(t)+a_2z(t)\leq (a_1y(0)+a_2z(0))e^{-\frac{k_1}{a_1}t}.
\end{align}
We can do a similar calculation by a division with $a_2>0$ and for $\frac{a_1}{a_2}\leq \frac{k_1}{k_2}$, we arrive at
\begin{align}\label{c4}
a_1y(t)+a_2z(t)\leq (a_1y(0)+a_2z(0))e^{-\frac{k_2}{a_2}t}.
\end{align}
Combining \eqref{c3} and \eqref{c4}, we finally obtain \eqref{c2}.
\end{proof}
\end{document}
|
arXiv
|
{
"id": "1811.00437.tex",
"language_detection_score": 0.4247763752937317,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{A non-abelian, non-Sidon, completely bounded $\Lambda(p)$ set} \author{Kathryn E. Hare} \address{Dept. of Pure Mathematics\\ University of Waterloo\\ Waterloo, Ont., \\ Canada} \email{[email protected]} \thanks{This research was supported in part by NSERC\ grant RGPIN 2016-03719} \author{Parasar Mohanty} \address{Dept. of Mathematics and Statistics\\ Indian Inst. of Tech.\\ Kanput, India, 208016} \email{[email protected]} \subjclass{Primary: 43A46, 43A30; Secondary: 42A55}
\begin{abstract} The purpose of this note is to construct an example of a discrete non-abelian group $G$ and a subset $E$ of $G$, not contained in any abelian subgroup, that is a completely bounded $\Lambda (p)$ set for all $p<\infty ,$ but is neither a Leinert set nor a weak Sidon set. \end{abstract}
\maketitle
\section{Introduction}
The study of lacunary sets, such as Sidon sets and $\Lambda (p)$ sets, constitutes an interesting theme in the theory of Fourier series on the circle group ${\mathbb{T}}$. It has many applications in harmonic analysis and in the theory of Banach spaces, and various combinatorial and arithmetic properties of these sets have been studied extensively. These concepts have also been investigated in the context of more general compact abelian groups (with their discrete dual groups) and compact non-abelian groups; see \cite {GH}, \cite{LR}, \cite{R} and the references cited therein. The study of these sets in the setting of discrete non-abelian groups was pioneered by Bozjeko \cite{B}, Fig\'{a}-Talamanca \cite{FP} and Picardello \cite{Pi}.
In abelian groups, there are various equivalent ways to define Sidon sets and these sets are plentiful. Indeed, every infinite subset of a discrete abelian group contains an infinite Sidon set. The natural analogues of these definitions in discrete non-abelian groups are known as strong Sidon, Sidon and weak Sidon sets. It was shown in \cite{Pi} that every weak Sidon set is $ \Lambda (p)$ for all $p<\infty $. In \cite{Le} Leinert introduced the concept of a $\Lambda (\infty )$ set, a notion only of interest in the non-abelian setting because in abelian groups such sets are necessarily finite. In striking contrast to the abelian situation, Leinert showed that the free group with two generators contains an infinite subset which is both weak Sidon and $\Lambda (\infty ),$ but does not contain any infinite Sidon subsets.
In \cite{Ha}, Harcharras studied the concept of completely bounded $\Lambda (p)$ sets, a property more restrictive than $\Lambda (p),$ but still possessed by Sidon sets. The converse is not true as every infinite discrete abelian group admits a completely bounded $\Lambda (p)$ set which is not Sidon; see \cite{HM}.
In this paper, we construct a non-amenable group $G$ and a set $E$ not contained in any abelian subgroup of $G,$ which is completely bounded $ \Lambda (p)$ for every $p<\infty ,$ but is neither $\Lambda (\infty )$ nor weak Sidon. It remains open if every infinite discrete group contains such a set $E$.
\section{Definitions}
Throughout this paper, $G$ will be an infinite discrete group. To define Sidon and $\Lambda (p)$ sets in this setting one requires the concepts of the Fourier algebra, $A(G)$, the von Neumann algebra, $VN(G),$ and the Fourier-Stieljies algebra, $B(G)$, as developed by P. Eymard in \cite{E} for locally compact groups. We also need the concept of a non-commutative $L^{p}$ -spaces introduced by I.E. Segal. We refer the reader to \cite{PX} for details on these latter spaces.
\begin{defn} (i) The set $E\subseteq $ $G$ is said to be a \textbf{strong (weak) Sidon set } if for all $f\in c_{0}(E)$ (resp., $l_{\infty }(E))$ there exists $g\in A(G)$ (resp., $B(G))$ such that $f(x)=g(x)\;$for all $x\in E$.
(ii) The set $E\subseteq G$ is said to be a \textbf{Sidon set }if there is a constant $C$ such that for all functions $f,$ compactly supported in $E,$ we have $\Vert f\Vert _{1}\leq C\Vert f\Vert _{VN(G)}$. The least such constant $C$ is known as the \textbf{Sidon constant} of $E$. \end{defn}
These definitions are well known to be equivalent in the commutative setting. For any discrete group it is the case that strong Sidon sets are Sidon and Sidon sets are weak Sidon. Finite groups are always strong Sidon sets. In \cite{Pi} it is shown that $E\subseteq G$ is Sidon if and only if for every $f\in $ $l_{\infty }(E)$ there is some $g\in B_{\rho }(G)$ that extends $f$, where $B_{\rho }(G)$ is the dual of the reduced $C^{\ast }$ algebra $C_{\rho }^{\ast }(G)$. Since in an amenable group $B_{\rho }(G)=B(G),$ weak Sidon sets are Sidon in this setting. Very recently, Wang \cite{Wa} showed that every Sidon set in any discrete group is a strong Sidon set. It remains open if every infinite amenable group contains an infinite Sidon subset.\textit{\ }
Picardello \cite{Pi} defined the notion of $\Lambda (p)$ sets in this setting and Harcharras \cite{Ha} introduced completely bounded $\Lambda (p)$ sets. For these, we require further notation. Let $\leftthreetimes $ denote the left regular representation of $G$ into $\mathcal{B}(l_{2}(G))$ and denote by $L^{p}(\tau _{0})$ the non-commutative $L^{p}$-space associated with the von Neumann algebra generated by $\leftthreetimes (G)$ with respect to the usual trace $\tau _{0}$. Let $L^{p}(\tau )$ denote the non-commutative $L^{p}$-space associated with the von Neumann algebra generated by $\leftthreetimes (G)\otimes \mathcal{B}(l_{2})$ with respect to the trace $\tau =\tau _{0}\otimes tr$, where $tr$ denotes the usual trace in $\mathcal{B}(l_{2})$. Observe that $L^{p}(\tau )$ has a cannonical operator space structure obtained from complex interpolation in the operator space category. We refer the reader to \cite{Pis} for more details.
\begin{defn} (i) Let $2<p<\infty $. The set $E\subseteq G$ is said to be a $\Lambda (p)$ \textbf{\ set} if there exists a constant $C_{1}>0$ such that for all finitely supported functions $f$ we have \begin{equation}
\left\Vert \sum\limits_{t\in E}f(t)\leftthreetimes (t)\right\Vert _{L^{p}(\tau _{0})}\leq C_{1}\left( |\sum\limits_{t\in E}|f(t)|^{2}\right) ^{ \frac{1}{2}}. \label{Lambdap} \end{equation}
(ii) The set $E\subseteq G$ is said to be a \textbf{completely bounded }$ \Lambda (p)$\textbf{\ set, }denoted $\Lambda ^{cb}(p),$ if there exists a constant $C_{2}>0$ such that \begin{equation} \Vert \sum\limits_{t\in E}\leftthreetimes (t)\otimes x_{t}\Vert _{L^{p}(\tau )}\leq C_{2}\max \left( \Vert (\sum_{t\in E}x_{t}^{\ast }x_{t})^{1/2}\Vert _{S_{p}},\Vert (\sum_{t\in E}x_{t}x_{t}^{\ast })^{1/2}\Vert _{S_{p}}\right) \label{CBLp} \end{equation} where $x_{t}$ are finitely supported families of operators in $S_{p}$, the $ p $-Schatten class on $l_{2}$.
The least such constants $C_{1}$ (or $C_{2})$ are known as the $\Lambda (p)$ \textbf{\ }(resp.,\textbf{\ }$\Lambda ^{cb}(p)$\textbf{) constants} of $E$. \end{defn}
It is known that every infinite set contains an infinite $\Lambda (p)$ set \cite{B} and that every weak Sidon set is a $\Lambda (p)$ set for each $ p<\infty $ \cite{Pi}. \ Completely bounded $\Lambda (p)$ sets are clearly $ \Lambda (p),$ but the converse is not true, as seen in \cite{Ha}.
Extending these notions to $p=\infty $ gives the Leinert and $L$-sets.
\begin{defn} (i) The set $E\subseteq $ $G$ is called a \textbf{Leinert or }$\Lambda (\infty )$\textbf{\ set} if there exists a constant $C>0$ such that for every function $f\in l_{2}(E)$ we have $\Vert f\Vert _{VN(G)}\leq C\Vert f\Vert _{2}$.
(ii) The sets of interpolation for the completely bounded multipliers of $ A(G)$ are called\textbf{\ }$L$\textbf{-sets}. \end{defn}
It is well known that the Leinert sets are the sets of interpolation for multipliers of $A(G),$ so any $L$-set is Leinert; see \cite{Po}. The set $E$ is said to satisfy the \textbf{Leinert condition} if every tuple $ (a_{1},...,a_{2s})\in E^{2s},$ with $a_{i}\neq a_{i+1},$ satisfies the independence-like relation \begin{equation} a_{1}a_{2}^{-1}a_{3}\dots a_{2s-1}a_{2s}^{-1}\neq e. \label{leinert} \end{equation} Here $e$ is the identity of $G$. It can be shown (\cite{Po}) that any set that satisfies the Leinert condition is an $L$-set.
It was seen in \cite{HM} that in abelian groups there are sets that are completely bounded $\Lambda (p)$ for all $p<\infty ,$ but not Sidon. Thus the inclusion, weak Sidon is $\Lambda ^{cb}(p),$ is strict for groups with infinite abelian subgroups. The purpose of this paper is to show the existence of sets not contained in \textit{any} abelian subgroup which also have this strict inclusion. In fact, we prove, more generally, the following result.
\begin{theorem} There is a discrete group $G$ that admits both infinite $L$-sets and weak Sidon sets, and an infinite subset $E$ of $G$ that is $\Lambda ^{cb}(p)$ for all $p<\infty ,$ but not a Leinert set, an $L$-set or a weak Sidon set. Moreover, any subset of $E$ consisting of commuting elements is finite. \label{mainthm} \end{theorem}
\section{Results and Proofs}
\subsection{Preliminary results}
To show that the set we will construct is not a Leinert or weak Sidon set, it is helpful to first establish some arithmetic properties of $\Lambda (p)$ and Leinert sets. We recall that a set $E\subseteq G$ is said to be \textbf{ quasi-independent} if all the sums \begin{equation*}
\left\{ \sum\limits_{x\in A}x:A\subset E,|A|<\infty \right\} \end{equation*} are distinct. Quasi-independent sets in abelian groups are the prototypical Sidon sets.
The first part of the following Lemma is well known for abelian groups.
\begin{lemma} Let $G$ be a discrete group. (i) Suppose $q>2$ and $E\subseteq G$ is a $ \Lambda (q)$ set with $\Lambda (q)$ constant $A$. If $a\in G$ has order $ p_{n}\geq $ $2n$, then \begin{equation*} \left\vert E\bigcap \{a,a^{2},...,a^{n}\}\right\vert \leq 10A^{2}n^{2/q} \text{.} \end{equation*}
(ii) Suppose $E\subseteq G$ is a Leinert set with Leinert constant $B$ and let $F\subseteq E$ be a finite commuting, quasi-independent subset. Then $ \left\vert F\right\vert \leq 6^{3}B^{2}$. \label{mainlem} \end{lemma}
\begin{proof} We will write $1_{X}$ for the characteristic function of a set $X$.
(i) Define the function $K_{n}$ on $G$ by \begin{equation*} K_{n}(x)=\sum_{j=-2n}^{2n}\left( 1-\frac{\left\vert j\right\vert }{n}\right) 1_{\{a^{j}\}}(x). \end{equation*} Let $J_{n}$ denote the function on $\mathbb{Z}_{p_{n}}$ (or $\mathbb{Z}$ if $ p_{n}=\infty $) defined in the analogous fashion. It is well known that the $ A(G)$ and $VN(G)$ norms for the function $K_{n}$ are dominated by the corresponding norms of the function $J_{n}$ on $\mathbb{Z}_{p_{n}}$.
As $L^{q^{\prime }}(\tau _{0})$ (for $q^{\prime }$ the dual index to $q$) is an interpolation space between $A(G)$ and $VN(G)$, it follows that \begin{eqnarray*} \left\Vert K_{n}\right\Vert _{L^{q^{\prime }}(\tau )} &\leq &\left\Vert K_{n}\right\Vert _{A(G)}^{1/q^{\prime }}\left\Vert K_{n}\right\Vert _{VN(G)}^{1/q} \\ &=&\left\Vert J_{n}\right\Vert _{A(\mathbb{Z}_{p_{n}})}^{1/q^{\prime }}\left\Vert J_{n}\right\Vert _{VN(\mathbb{Z}_{p_{n}})}^{1/q}\leq (4n+1)^{1/q}\text{.} \end{eqnarray*}
Suppose $E\bigcap \{a,a^{2},...,a^{n}\}$ consists of the $M$ elements $ \{a^{s_{j}}\}_{j=1}^{M}$ and put \begin{equation*} k_{n}(x)=\sum_{j=1}^{M}1_{\{a^{s_{j}}\}}(x). \end{equation*} Since $E$ has $\Lambda (q)$ constant $A,$ the generalized Holder's inequality implies \begin{eqnarray*} \frac{M}{2} &\leq &\sum_{j=1}^{M}K_{n}(a^{s_{j}})=\sum_{x\in G}K_{n}(x)k_{n}(x) \\ &\leq &\left\Vert K_{n}\right\Vert _{L^{q^{\prime }}(\tau _{0})}\left\Vert k_{n}\right\Vert _{L^{q}(\tau _{0})}\leq (4n+1)^{1/q}A\left\Vert k_{n}\right\Vert _{2} \\ &=&(4n+1)^{1/q}A\sqrt{M}. \end{eqnarray*} Consequently, $M\leq 2(4n+1)^{2/q}A^{2}\leq 10A^{2}n^{2/q}$, as claimed.
(ii) Let $H$ be the abelian group generated by $F$. Being quasi-independent, $F$ is a Sidon subset of $H$ with Sidon constant at most $6\sqrt{6}$ (\cite[ p.115]{GH}). Consider the function $h=1_{F}$ defined on $H$ and $g=1_{F}$ defined on $G$. The Sidon property, together with the fact that $\left\Vert h\right\Vert _{VN(H)}=\left\Vert g\right\Vert _{VN(G)},$ ensures that \begin{equation*} \left\vert F\right\vert =\left\Vert h\right\Vert _{\ell ^{1}}\leq 6\sqrt{6} \left\Vert h\right\Vert _{VN(H)}=6\sqrt{6}\left\Vert g\right\Vert _{VN(H)}. \end{equation*} Since $E$ has Leinert constant $B$, we have $\left\Vert f\right\Vert _{VN(G)}\leq B\left\Vert f\right\Vert _{2}$ for any function $f$ defined on $ G$ and supported on $E$. In particular, this is true for the function $g\,$, hence \begin{equation*} \left\vert F\right\vert \leq 6\sqrt{6}\left\Vert g\right\Vert _{VN(H)}\leq 6 \sqrt{6}B\sqrt{\left\vert F\right\vert }. \end{equation*} \end{proof}
\subsection{\noindent Proof of Theorem \protect\ref{mainthm}}
\begin{proof} We will let $G$ be the free product of the cyclic groups $Z_{p_{n}}$, $n\in N $, where $p_{n}>2^{n+1}$ are distinct odd primes. If $a_{n}$ is a generator of $Z_{p_{n}},$ then $\{a_{n}\}_{n=1}^{\infty }$ is both a weak Sidon and Leinert set, as shown in \cite{Pi}. The set $E$ will be the union of finite sets $E_{n}\subseteq Z_{p_{n}}$, where $\left\vert E_{n}\right\vert =n^{2}$ and $E_{n}\subset \{a_{n},...,a_{n}^{2^{n}}\}$. The fact that any commuting subset of $E$ is finite is obvious from the definition of $E$.\newline
We recall the following notation from \cite{Ha}: We say that a subset $ \Lambda \subseteq G$ has the $Z(p)$ property if $Z_{p}(\Lambda )<\infty $ where \begin{equation*} Z_{p}(\Lambda )=\sup_{x\in G}\left\vert \left\{ (x_{1},...,x_{p})\in \Lambda ^{p}:x_{i}\neq x_{j},x_{1}^{-1}x_{2}x_{3}^{-1}\cdot \cdot \cdot x_{p}^{(-1)^{p}}=x\right\} \right\vert . \end{equation*} In \cite{Ha}, Harcharras proved that if $2<p<\infty ,$ then every subset $ \Lambda $ of $G$ with the $Z(p)$ property is a $\Lambda ^{cb}(2p).$
We will construct the sets $E_{n}$ so that they have the property that for every even $s\geq 2$ there is an integer $n_{s}$ such that $Z_{s}\left( \bigcup_{n\geq n_{s}}E_{n}\right) \leq s!$. Consequently, $\bigcup_{n\geq n_{s}}E_{n}$ will be $\Lambda ^{cb}(2s)$ for all $s<\infty $. As finite sets are $\Lambda ^{cb}(p)$ for all $p<\infty ,$ and a finite union of $\Lambda ^{cb}(p)$ sets is again $\Lambda ^{cb}(p)$, it will follow that $E$ is $ \Lambda ^{cb}(p)$ for all $p<\infty $.
We now proceed to construct the sets $E_{n}$ by an iterative argument. Temporarily fix $n$ and take $g_{1}=a_{n}$. Inductively assume that for $ N<n^{2}$, $\{g_{i}\}_{i=1}^{N}$ $\subseteq \{a_{n},...,a_{n}^{2^{n}}\}$ have been chosen with the property that if \begin{equation}
\prod\limits_{j=1}^{N}g_{j}^{\varepsilon _{j}}=1\text{ for }\varepsilon _{j}=0,\pm 1,\pm 2, \sum_{j}|\varepsilon _{j}|\leq 2s,\text{ then all } \varepsilon _{j}=0. \tag{$\mathcal{P}_{N}$} \end{equation} Now choose \begin{equation*} g_{N+1}\neq \prod\limits_{j=1}^{N}g_{j}^{\varepsilon _{j}}\text{ for any }
\varepsilon _{j}=0,\pm 1,\pm 2\text{ and }\sum_{j}|\varepsilon _{j}|\leq 2s \end{equation*} and \begin{equation*} g_{N+1}^{2}\neq \prod\limits_{j=1}^{N}g_{j}^{\varepsilon _{j}}\text{ for any
}\varepsilon _{j}=0,\pm 1,\pm 2\text{ and }\sum_{j}|\varepsilon _{j}|\leq 2s. \end{equation*} There are at most $\binom{N}{2s}5^{2s}\leq C_{s}N^{2s}$ terms that $g_{N+1}$ must avoid and similarly for $g_{N+1}^{2}$ as the squares of elements of $ Z_{p_{n}}$ are all distinct. Provided $2C_{s}N^{2s}\leq 2^{n}$ then we can make such a choice of $g_{N+1}\in $ $\{a_{n},...,a_{n}^{2^{n}}\}$. Of course, it is immediate that property ($\mathcal{P}_{N+1}$) then holds. This can be done for every $N<n^{2}$ as long as $n$ is suitably large, say for $ n\geq n_{s}$. The set $E_{n}$ will be taken to be $\{g_{j}\}_{j=1}^{n^{2} \text{.}}.$
Now we need to check the claim that $Z_{s}(\bigcup\limits_{n\geq n_{s}}E_{n})\leq s!$. Towards this, suppose \begin{equation} x_{1}x_{2}^{-1}\cdot \cdot \cdot x_{s}^{-1}=y_{1}y_{2}^{-1}\cdot \cdot \cdot y_{s}^{-1} \label{P1} \end{equation} where $x_{i}$ are all distinct, $y_{j}$ are all distinct and all $ x_{i},y_{j}\in \bigcup\limits_{n\geq n_{s}}E_{n}$. The free product property guarantees that if this is true, then it must necessarily be the case that if we consider only the elements $x_{i_{k}}$ and $y_{j_{l}}$ which belong to a given $E_{n}$, we must have $\prod\limits_{k}x_{i_{k}}^{\delta _{k}}=$ $ \prod\limits_{l}y_{j_{l}}^{\varepsilon _{l}}$ for the appropriate choices of $\delta _{k},\varepsilon _{l}\in \{\pm 1\}$. As there at most $s$ choices for each of $x_{i_{k}}$ and $y_{i_{l}}$, our property ($\mathcal{P}_{N}$) ensures that this can happen only if $\{x_{i_{k}}:\delta _{k}=1\}$ $ =\{y_{j_{l}}:\varepsilon _{l}=1\}$ and similarly for the terms with $-1$ exponents. Hence we can only satisfy (\ref{P1}) if upon reordering, $ \{x_{1},x_{3},...,x_{s-1}\}=\{y_{1},y_{3},...,y_{s-1}\},$ and similarly for the terms with even labels. (We remark that for non-abelian groups, this is only a necessary but not, in general, sufficient condition for (\ref{P1}).) This suffices to establish that \begin{equation*} Z_{s}(\bigcup\limits_{n\geq n_{s}}E_{n})\leq ((s/2)!)^{2}\leq s! \end{equation*} and hence, as explained above, $E$ is a $\Lambda ^{cb}(p)$ set for all $ p\,<\infty $.
Next, we will verify that $E$ is not a weak Sidon set. We proceed by contradiction. According to \cite{Pi}, if it was, then $E$ would be a $ \Lambda (p)$ set for each $p>2,$ with $\Lambda (p)$ constant bounded by $C \sqrt{p}$ for a constant $C$ independent of $p$. Appealing to Lemma \ref {mainlem}(i), we have \begin{equation*} n^{2}=\left\vert E_{n}\right\vert =\left\vert E\bigcap \{a_{n},...,a_{n}^{2^{n}}\}\right\vert \leq 10C^{2}p2^{2n/p}. \end{equation*} Taking $p=2n$ for sufficiently large $n$ gives a contradiction.
Finally, to see that $E$ is not a Leinert set, we first observe that an easy combinatorial argument shows that any set of $N$ distinct elements contains a quasi-independent subset of cardinality at least $\log N/\log 3$. Thus we can obtain quasi-independent subsets $F_{n}\subseteq E_{n}$ with $\left\vert F_{n}\right\vert \rightarrow \infty $. But according to Lemma \ref{mainlem} (ii), this would be impossible if $E$ was a Leinert set. As $E$ is not Leinert, it is also not an $L$ set.
This concludes the proof. \end{proof}
\
\end{document}
|
arXiv
|
{
"id": "1910.08377.tex",
"language_detection_score": 0.7719739079475403,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{\bfseries \Large De-biased lasso for stratified Cox models with application to the national kidney transplant data}
\begin{abstract}
The Scientific Registry of Transplant Recipients (SRTR) system has become a rich resource for understanding the complex mechanisms of graft failure after kidney transplant, a crucial step for allocating organs effectively and implementing appropriate care.
As transplant centers that treated patients might strongly confound graft failures, Cox models stratified by centers can eliminate their confounding effects. Also,
since recipient age is a proven non-modifiable risk factor, a common practice is to fit models separately by recipient age groups. The moderate sample sizes, relative to the number of covariates, in some age groups may lead to biased maximum stratified partial likelihood estimates and unreliable confidence intervals even when samples still outnumber covariates.
To draw reliable inference on a comprehensive list of risk factors measured from both donors and recipients in SRTR, we propose a de-biased lasso approach via quadratic programming
for fitting stratified Cox models. We establish asymptotic properties and verify via simulations that our method produces consistent estimates and confidence intervals with nominal coverage probabilities.
Accounting for nearly 100 confounders in SRTR, the de-biased method detects that the graft failure hazard nonlinearly increases with donor's age among all recipient age groups, and that organs from older donors more adversely impact the younger recipients.
Our method also delineates the associations between graft failure and many risk factors such as recipients' primary diagnoses (e.g. polycystic disease, glomerular disease, and diabetes) and donor-recipient mismatches for human leukocyte antigen loci across recipient age groups. These results may inform the refinement of donor-recipient matching criteria for stakeholders.\\[0.5cm]
\textbf{Keywords:} Confidence intervals; diverging number of covariates; end-stage renal disease; graft failure free survival; statistical inference. \end{abstract}
\section{Introduction}
\label{sec:intro}
For patients with end-stage renal disease, one of the most lethal and
prevalent diseases in the U.S. \citep{saran2020us}, successful renal transplantation is effective for improving quality of life and prolonging survival \citep{wolfe1999comparison,kostro2016quality,ju2019patient}. The success of kidney transplantation hinges upon various factors related to the quality of transplant operations, the quality of donated kidneys, and the physical conditions of recipients \citep{rodger2012approach,legendre2014factors}, and it is crucial to evaluate and understand how these risk factors impact on renal graft failure in order to increase the chance of success \citep{hamidi2016identifying,legendre2014factors}. With the scarcity of organs and an increasing number of waitlisted candidates \citep{bastani2015worsening}, the results can inform more efficient strategies for kidney allocation \citep{rao2009alphabet,smith2012kidney} as well as evidence-based post-transplant care \citep{baker2017renal}.
Therefore, how to quantify the impacts of important factors associated with prognosis, particularly renal graft failure,
remains to be a central question in kidney transplantation.
The Scientific Registry of Transplant Recipients (SRTR) system, a federally funded organization that keeps records of transplant information from recipients and donors,
has become a rich resource for studying post-kidney transplantation prognosis \citep{dickinson2008srtr}.
Leveraging the SRTR data, one can develop a valid tool for characterizing the influences of risk factors on graft failure, a key step towards post-transplant prognosis. Most previous studies, which focused only on a small number of factors, i.e. kidney diagnosis, recipient age, recipient race, recipient gender, number of human leukocyte antigen (HLA) mismatches, donor age, donor race, donor gender, serum creatinine level and cold ischemia time \citep{alexander1994effect}, might have pre-excluded other important factors and not fully captured the complex mechanisms governing graft failure. The SRTR data contain comprehensive information on recipients and donors, such as recipient primary insurance and employment, procedure type, infection of multiple viruses, history of transplant, transfusion and drug abuse, and pre-transplant comorbidities. The data provide a unique opportunity for assessing the associations between graft failure and an extended list of variables simultaneously, which may reduce confounding \citep{wang2011gee}. Specifically, since donor age
is a major criterion for donor-recipient matching \citep{kasiske2002matching,rao2009comprehensive,veroux2012age}, the data enable us to examine its effect on graft failure by adjusting for confounders, including pre-existing comorbidities.
There are several statistical challenges.
On the one hand, as recipients received care in various transplant centers, the center-specific effects may confound the covariate effects of interest.
This motivates us
to consider Cox models stratified by transplant centers, a commonly used model in the relevant context \lu{without the need to explicitly model the potentially time-varying center effects \citep{he2021stratified}}.
On the other hand,
recipient age is a strong risk factor and there may exist
complicated interactions between recipients' age and other characteristics \citep{keith2004effect}.
For ease of interpretation and by convention \citep{morales2012risk,faravardeh2013predictors}, we have opted to divide our analyzable patient population (the adult recipients with
kidneys transplanted during 2000 and 2001) into $[18, 45]$, $(45, 60]$ and $60+$ years old groups (Table \ref{tab:character}), and
fit models separately for these three groups. Allowing model parameters to be age group-specific, we have avoided parametrically modeling the interactions between recipient age and the other risk factors.
When the number of covariates is relatively large
(94 in our data) compared to, though still less than, the sample size
(for example, 1448 patients with 1013 events in the $60+$ years old recipient group), the conventional maximum stratified partial likelihood estimation (MSPLE) may yield untrustworthy point estimates, confidence intervals and hypothesis testing results, as illustrated in our simulations.
\begin{table}[ht]
\centering
\caption{Study population characteristics by recipient age group}
\begin{tabular}{llll}
\hline
Recipient age group & $[18, 45]$ & $(45, 60]$ & $60+$ \\
Variable & \multicolumn{3}{c}{Mean (SD) / Count (\%) } \\ \hline
\# Centers & 84 (--) & 107 (--) & 43 (--) \\
\# Patients & 3388 (100\%) & 4359 (100\%) & 1448 (100\%) \\
\# Events & 1588 (46.9\%) & 2334 (53.5\%) & 1013 (70.0\%) \\
Recipient age & 35.7 (7.0) & 53.0 (4.2) & 66.6 (4.3) \\
Donor age (years) & & & \\
$\qquad$ $\le 10$ & 276 (8.1\%) & 223 (5.1\%) & 61 (4.2\%) \\
$\qquad$ $(10, 20]$ & 580 (17.1\%) & 611 (14.0\%) & 137 (9.5\%) \\
$\qquad$ $(20, 30]$ & 633 (18.7\%) & 683 (15.7\%) & 179 (12.4\%) \\
$\qquad$ $(30, 40]$ & 505 (14.9\%) & 599 (13.7\%) & 174 (12.0\%) \\
$\qquad$ $(40, 50]$ & 753 (22.2\%) & 947 (21.7\%) & 256 (17.7\%) \\
$\qquad$ $(50, 60]$ & 498 (14.7\%) & 893 (20.5\%) & 318 (22.0\%) \\
$\qquad$ $60+$ & 143 (4.2\%) & 403 (9.2\%) & 323 (22.3\%) \\
Recipient gender & & & \\
$\qquad$ Male & 1997 (58.9\%) & 2671 (61.3\%) & 913 (63.1\%) \\
$\qquad$ Female & 1391 (41.1\%) & 1688 (38.7\%) & 535 (36.9\%) \\
Donor gender & & & \\
$\qquad$ Male & 2039 (60.2\%) & 2563 (58.8\%) & 803 (55.5\%) \\
$\qquad$ Female & 1349 (39.8\%) & 1796 (41.2\%) & 645 (44.5\%) \\
\hline
\end{tabular}
\label{tab:character}
\end{table}
For proper inferences, we consider an asymptotic framework with a diverging number of covariates, wherein the number of covariates, though smaller than the sample size, can increase with the sample size \citep{he2000parameters,wang2011gee}.
\lu{Lasso provides a very popular tool for simultaneous variable selection and estimation with high-dimensional covariates \citep{tibshirani1997lasso}. For unstratified Cox models, \citet{huang2013oracle} and \citet{kong2014non} presented the oracle inequalities for the lasso estimator. However, with penalization, lasso estimates are biased towards zero \citep{van2014asymptotically}, and they do not possess regular limiting distributions even under linear regression with a fixed number of covariates \citep{fu2000asymptotics}. Conditional inference based on the selected model is invalid, either, due to the failure to account for uncertainty in model selection. Hence, lasso cannot be directly applied to draw statistical inference.}
There is literature \lu{on inference for} unstratified Cox proportional hazards models
under the related asymptotic framework.
For example, \citet{fang2017testing} proposed decorrelated score tests for a low-dimensional component in the regression parameters, and \citet{kong2018high}, \citet{yu2018confidence} and \citet{xia2021cox} proposed to correct asymptotic biases of the lasso estimator following the framework of \citet{van2014asymptotically}, \citet{zhang2014confidence} and \citet{javanmard2014confidence} that were originated from high-dimensional linear or generalized linear models. For Cox models, all of these methods, except \citet{xia2021cox} which considered the ``large $n$, diverging $p$'' scenario, assumed sparsity on the inverse information matrix.
This sparse matrix assumption, however, may not hold for models beyond linear regression, leading to insufficient bias correction and under-covered confidence intervals. Moreover, as these methods were not designed for modeling stratified data, they are not directly applicable to the analysis of the SRTR data. \lu{To our knowledge, the current literature lacks inferential methods with theoretical rigor for stratified Cox models with a diverging number of covariates.}
We propose a
de-biased lasso approach for Cox models
stratified by transplant centers,
which solves a series of quadratic programming problems to estimate the inverse information matrix, and corrects the biases from the lasso estimator for valid statistical inference. Our asymptotic results enable us to draw inference on any linear combinations of model parameters, including the low-dimensional targets in \citet{fang2017testing} and \citet{kong2018high} as special cases and fundamentally deviating from the stepwise regression adopted by \citet{rao2009comprehensive}.
\lu{When the number of covariates is relatively large compared to the sample size,
our approach yields less biased estimates and more properly covered confidence intervals than MSPLE as well as the methods of \citet{fang2017testing, kong2018high, yu2018confidence} adapted to the stratified setting.} Therefore, it is well-suited
for analyzing the SRTR data, especially among the oldest recipient group that has the smallest sample size.
Applications of our method to the SRTR data have generated reliable estimation and inference results for the effects of an expanded list of donor and recipient factors.
We find that receiving kidneys from older donors is associated with an increased hazard of graft failure after adjusting for many confounding factors, and that the dependence on donors' age is non-linear.
The results may inform more comprehensive assessments of post-transplant prognosis and kidney allocation.
The article is organized as follows. We introduce the proposed de-biased lasso approach in Section \ref{sec:method} and establish the asymptotic results in Section \ref{sec:theory}, which form the basis of inference for the SRTR data. We conduct simulations in Section \ref{sec:simul} and demonstrate that our method outperforms MSPLE in bias correction and confidence interval coverage. In Section \ref{sec:app}, we analyze the SRTR data by using the proposed de-biased approach.
Finally, we provide a few concluding remarks in Section \ref{sec:discuss}, the detailed list of covariates considered in the analysis of SRTR data in Appendix \ref{appA}, and regularity conditions in Appendix \ref{appB}. Technical details and proofs are deferred to the Supplementary Material.
\section{De-biased lasso for stratified Cox models via quadratic programming}
\label{sec:method}
We apply stratified Cox models to evaluate the impacts of risk factors on post-transplant graft failure using the SRTR data. For each recipient age group defined in the first row of Table \ref{tab:character}, let $K$ be the total number of transplant centers, and $n_k$ be the number of recipients in the $k$-th transplant center, $k=1, \cdots, K$. With $i$ indexing recipients within the $k$-th transplant center, let $T_{ki}$ denote the graft failure free survival time, i.e. the time from transplantation to graft failure or death, whichever comes first [a common endpoint in transplantation \citep{kasiske2011relationship}], $X_{ki}$ be a $p$-dimensional covariate vector,
and $C_{ki}$ be the censoring time. We assume random censoring, that is,
$T_{ki}$ and $C_{ki}$ are independent given $X_{ki}$. In the SRTR data, $p=94$ and $X_{ki}$ includes risk factors from both donors and recipients, such as gender, ABO blood type, history of diabetes and duration, angina/coronary artery disease, symptomatic peripheral vascular disease, drug treated systemic hypertension, drug treated COPD, and mismatch for each HLA locus between donors and recipients; see a full list of covariates in Appendix \ref{appA}.
Let $\delta_{ki} = 1(T_{ki} \le C_{ki})$ be the event indicator and $Y_{ki} = \min(T_{ki}, C_{ki})$ be the observed time.
With a center-specific baseline hazard function $\lambda_{0k}(t)$, a stratified Cox model for $T_{ki}$ stipulates that its conditional hazard at $t$ given $X_{ki}$ is
\[
\lambda_{ki}(t | X_{ki}) = \lambda_{0k}(t) \exp\{X_{ki}^T \beta^0\},
\]
where $\beta^0 = (\beta^0_1, \ldots, \beta^0_p)^T \in \mathbb{R}^p$ is the vector of common regression coefficients across all centers. \lu{It is reasonable to assume that the true regression coefficients $\beta^0$ are the same across strata \citep{kalbfleisch2002statistical}, while the center effects, though not of primary interest here, are accounted for via different baseline hazards $\lambda_{0k}(t)$'s.}
\subsection{Estimation method}
\label{subsec:dslasso}
The MSPLE of $\beta$ minimizes the following {\em negative} log stratified partial likelihood function
\begin{equation} \label{eq:negloglik1}
\ell(\beta) = -\displaystyle \frac{1}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \left[ \beta^T X_{ki} - \log \left\{ \frac{1}{n_k} \sum_{j=1}^{n_k} 1(Y_{kj} \ge Y_{ki}) \exp(\beta^T X_{kj}) \right\} \right] \delta_{ki},
\end{equation}
where $N = \sum_{k=1}^K n_k$.
In SRTR, the number of risk factors, though smaller than the sample size, is fairly large. In this case,
our numerical examination shows that MSPLEs are biased and their confidence intervals do not yield nominal coverage. We consider a de-biased approach that has been shown to yield valid inference in linear regression \citep{van2014asymptotically,zhang2014confidence,javanmard2014confidence}. Here we assume that $p<N$ but grows with $N$, which falls into the ``large $N$, diverging $p$'' framework. We extend the de-biased lasso to accommodate stratified Cox models.
For a vector $x = (x_1, \ldots, x_p)^T \in \mathbb{R}^p$, define $x^{\otimes 0} = 1$, $x^{\otimes 1} = x$ and $x^{\otimes 2} = x x^T$.
Let $\dot{\ell}(\beta)$ and $\ddot{\ell}(\beta)$ be the first and the second order derivatives of $\ell(\beta)$ with respect to $\beta$, i.e.
\begin{align*}
\dot{\ell}(\beta) &= - \displaystyle \frac{1}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \left\{ X_{ki} - \displaystyle \frac{\widehat{\mu}_{1k}(Y_{ki}; \beta) }{\widehat{\mu}_{0k} (Y_{ki}; \beta) } \right\} \delta_{ki}, \\
\ddot{\ell}(\beta) & = \displaystyle \frac{1}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \left\{ \displaystyle \frac{\widehat{\mu}_{2k}(Y_{ki}; \beta)}{\widehat{\mu}_{0k}(Y_{ki}; \beta)} - \left[ \displaystyle \frac{\widehat{\mu}_{1k}(Y_{ki}; \beta)}{\widehat{\mu}_{0k}(Y_{ki}; \beta)} \right]^{\otimes 2} \right\} \delta_{ki},
\end{align*}
where $\widehat{\mu}_{rk} (t; \beta) = {n_k}^{-1} \sum_{j=1}^{n_k} 1(Y_{kj} \ge t) X_{kj}^{\otimes r} \exp\{X_{kj}^T \beta\}, ~ r=0, 1, 2$.
The lasso estimate, $\widehat{\beta}$, minimizes the penalized negative log stratified partial likelihood,
\begin{equation} \label{eq:lasso}
\widehat{\beta} = {\arg\min}_{\beta \in \mathbb{R}^p} \{ \ell(\beta) + \lambda \| \beta \|_1 \},
\end{equation}
where $\lambda > 0$ is a tuning parameter that encourages sparse solutions. Here, $\| x \|_q = (\sum_{j=1}^p |x_j|^q)^{1/q}$ is the $\ell_q$-norm for $x \in \mathbb{R}^{p}$, $q \ge 1$.
As $\widehat{\beta}$ is typically biased, we can obtain the de-biased lasso estimator by a Taylor expansion of $\dot{\ell}(\beta^0)$ around $\widehat{\beta}$. To proceed,
let $\widehat{M}$ be a $p \times p$ matrix and $\widehat{M}_j$ its $j$th row. Pre-multiplying $\widehat{M}_j$ on both sides of the Taylor expansion and collecting terms, we have the following equality for the $j$th component of $\beta$:
\begin{equation} \label{eq:derive_bhat_M}
\widehat{\beta}_j - \beta^0_j + \overbrace{ \left( - \widehat{M}_j \dot{\ell}(\widehat{\beta}) \right) }^{I_j}
+ \overbrace{\left( - \widehat{M}_j {\Delta} \right) }^{II_j}
+ \overbrace{\left( \widehat{M}_j \ddot{\ell}(\widehat{\beta}) - {e}_j^T \right) \left( \widehat{\beta} - \beta^0 \right)}^{III_j} = - \widehat{M}_j \dot{\ell}(\beta^0),
\end{equation}
where the remainder $\Delta \in \mathbb{R}^p$ in $II_j$ can be shown asymptotically negligible given the convergence rate of the lasso estimator $\widehat{\beta}$, and so is $III_j$ if $\widehat{M}_j \ddot{\ell}(\widehat{\beta}) - e_j^T$ converges to zero with certain rate that will be discussed later in Section \ref{sec:theory}.
Hence, the de-biased lasso estimator
corrects the bias of $\widehat{\beta}_j$ with a one-step update of
\begin{equation} \label{eq:dslasso0}
\widehat{b}_j = \widehat{\beta}_j - \widehat{\Theta}_j \dot{\ell}(\widehat{\beta}),
\end{equation}
which replaces $\widehat{M}_j$ in \eqref{eq:derive_bhat_M} with the $j$-th row of $\widehat{\Theta}$, an estimate of the inverse information matrix $\Theta_{\beta^0}$, and $ - \widehat{\Theta}_j \dot{\ell}(\widehat{\beta})$ is the bias correction term
to $\widehat{\beta}_j$. Here, $\Theta_{\beta^0}$ is the inverse of the population version of $\widehat{\Sigma} $ given in the following \eqref{eq:hatsigma}; see the explicit definition of $\Theta_{\beta^0}$ underneath \eqref{popsig}.
Denote by
$\widehat{b} = (\widehat{b}_1, \ldots, \widehat{b}_p)^T$ the vector of the de-biased lasso estimates, and, for compactness, write (\ref{eq:dslasso0}) in a matrix form
\begin{equation} \label{eq:dslasso}
\widehat{b} = \widehat{\beta} - \widehat{\Theta} \dot{\ell}(\widehat{\beta}).
\end{equation}
Unlike $\widehat{\beta}$, the de-biased estimator $\widehat{b}$ in (\ref{eq:dslasso}) is no longer sparse. Motivated by \citet{javanmard2014confidence} on high-dimensional inference in linear regression, we propose to obtain $\widehat{\Theta}$ by solving a series of quadratic programming problems. First, we compute
\begin{equation} \label{eq:hatsigma}
\widehat{\Sigma} =
\displaystyle \frac{1}{N}\sum_{k=1}^K \sum_{i=1}^{n_k} \delta_{ki} \left[ X_{ki} - \widehat{\eta}_k(Y_{ki}; \widehat{\beta}) \right]^{\otimes 2},
\end{equation}
where $\widehat{\eta}_k(t; \beta) = \widehat{\mu}_{1k}(t; \beta) / \widehat{\mu}_{0k}(t; \beta)$ is the vector of weighted average covariates.
We use $\widehat{\Sigma}$, in lieu of $\ddot{\ell} (\widehat{\beta})$, for ease of proving theoretical properties.
Indeed, as shown in the Supplementary Material, $\| \widehat{\Sigma} - \ddot{\ell} (\widehat{\beta}) \|_{\infty} \stackrel{p}{\rightarrow}0 $ with a desirable rate under the conditions in Section \ref{sec:theory}. Next, for each $j = 1, \ldots, p$,
we solve a quadratic programming problem
\begin{equation} \label{eq:qp}
\min_{ m \in \mathbb{R}^p} \left\{ m^T \widehat{\Sigma} m: \| \widehat{\Sigma} m - e_j \|_{\infty} \le \gamma \right\},
\end{equation}
where $\gamma > 0$ is a tuning parameter that is different from the lasso tuning parameter $\lambda$, $e_j$ is a unit directional vector with only the $j$th element being one, and $\| \cdot \|_{\infty}$ is the matrix max norm, i.e. $\| A \|_{\infty} = \max_{i,j} |A_{ij}|$ for a real matrix $A $.
Denote by $m^{(j)}$ the column vector of solution to (\ref{eq:qp}). We
obtain a $p \times p$ matrix $\widehat{\Theta} = (m^{(1)}, \ldots, m^{(p)})^T$.
The constraint $\| \widehat{\Sigma} m - e_j \|_{\infty} \le \gamma$ in \eqref{eq:qp} controls deviations of the de-biased estimates from the lasso estimates. In an extreme case of $\gamma = 1$, an admissible solution is $m=0$, and therefore there is no bias correction in the de-biased estimator;
in another extreme case of $\gamma=0$, $m^{(j)}$ is the $j$th column of $\widehat{\Sigma} ^{-1}$.
We implement \eqref{eq:qp} by using R \texttt{solve.QP()}, which can be programmed in parallel for large $p$. We name the method {\sl de-biased lasso via quadratic programming} (hereafter, DBL-QP).
\subsection{Tuning parameter selection}
\label{subsec:tuning}
For the DBL-QP method, the lasso tuning parameter $\lambda$ can be selected via 5-fold cross-validation
as in \citet{simon2011regularization}.
The selection of $\gamma$ is crucial
as, for example, Figure \ref{fig:tuning} reveals that $\gamma$ should be selected within a specific range (shaded in figures) to achieve the most desirable bias correction and confidence interval coverage probability. It also shows the large bias and poor coverage resulting from MSPLE. Inappropriate tuning can yield even more biased estimates with poorer coverage than MSPLE. Results of lasso and oracle estimates are also provided as references, where oracle estimates are obtained from the reduced model that only contains truly nonzero coefficients.
\begin{figure}\label{fig:tuning}
\end{figure}
Intuitively, $\gamma$ should be chosen near zero, resulting in a de-biased estimator with estimates of large coefficients close to oracle estimates. We do not recommend evaluating cross-validation criteria by plugging in the de-biased estimates because of accumulative estimation errors. We opt for a hard-thresholding approach that more effectively removes noise from the de-biased estimates: we retain the de-biased lasso estimate for $\beta^0_j$ only if the null hypothesis $\beta_j^0 = 0$ is rejected; otherwise, we set it to zero (shown in Algorithm \ref{algo1}). The set $\widehat{A}$ in Step 2.2 of Algorithm \ref{algo1} is expected to estimate well the set of truly associated variables. Specifically, we set $\widehat{A}$ to be the index set of variables whose Wald statistic $ \sqrt{N} | \widehat{b}_j | /\widehat{\Theta}_{jj}^{1/2} > z_{\alpha/(2p)}$, where $z_{\alpha/(2p)}$ is the upper $\{ \alpha/(2p) \}$-th quantile of the standard normal distribution. The cutoff is determined by Theorem \ref{thm:main} and Bonferroni correction for multiple testing.
When implementing cross-validation, we can either take stratum as the sampling unit and randomly split strata, or randomly split observations within each stratum, to form training and testing subsets.
We find the former
improves stability of tuning parameter selection when there are a number of small-sized strata.
\begin{algorithm}[ht]
\caption{Selection of the tuning parameter $\gamma$ using cross-validation \label{algo1}}
\begin{algorithmic}
\item[\textbf{Step 1}] Pre-determine a grid of points for $\gamma$ in [0,1], denoted as $\gamma^{(g)}, g=1, \cdots, G$, and set each $cv_g = 0$.
\item[\textbf{Step 2}] Randomly assign the $K$ strata into $M$ folds,
leaving one fold for testing and the others for training.
Set $q = 1$.
\item[\quad \textbf{Step 2.1}] While $q \le M$, use the $q$th training set to compute the de-biased lasso estimator with $\gamma^{(g)}, g=1, \cdots, G$, denoted as $\widehat{b}^{(gq)}$, and define the active set $\widehat{A}^{(gq)}$.
\item[\quad \textbf{Step 2.2}] Define the thresholded de-biased lasso estimator $\widehat{b}^{(gq)}_{thres} = \widehat{b}^{(gq)} \cdot 1(j \in \widehat{A}^{(gq)})$, i.e. setting components of $\widehat{b}^{(gq)}$outside the active set $\widehat{A}^{(gq)}$ to 0.
\item[\quad \textbf{Step 2.3}] Compute the negative log partial likelihood on the $q$th testing set $\ell^{(q)} (\widehat{b}^{(gq)}_{thres})$.
\item[\quad \textbf{Step 2.4}] Set $cv_g \leftarrow cv_g + N^{(q)} \ell^{(q)} (\widehat{b}^{(gq)}_{thres})$, for $g = 1, \cdots, G$, where $N^{(q)}$ is the total number of observations in the $q$th testing set.
\item[\quad \textbf{Step 2.5}] Set $q \leftarrow q + 1$ and go to Step 2.1.
\item[\textbf{Step 3}] Let $\widehat{g} = \arg\min_{g} cv_g$. The final output tuning parameter value is $\gamma^{(\widehat{g})}$.
\end{algorithmic}
\end{algorithm}
\section{Valid statistical inference based on the de-biased lasso estimator}
\label{sec:theory}
This section presents asymptotic results, which lay the groundwork for using the de-biased lasso estimator described in Section \ref{sec:method} to infer on the risk factors of graft failure in the SRTR analysis. The pertaining large sample framework posits that the number of strata $K$ is fixed, the smallest stratum size $n_{min} = \min_{1 \le k \le K} n_k \rightarrow \infty$, and
$
{n_k} / {N} \rightarrow r_k > 0
$
as $n_{min} \rightarrow \infty$, $k=1, \cdots, K$.
\lu{This framework conforms to the real world setting of our concern, where the number of transplant centers nationwide is finite, and the number of patients or transplant events in each center increases over the years.}
We provide regularity conditions \lu{and their discussion} in Appendix \ref{appB}, \lu{and present all} the proofs in the Supplementary Material.
Let $\mu_{rk} (t; \beta) = {E} [1(Y_{k1} \ge t) X_{k1}^{\otimes r} \exp\{X_{k1}^T \beta\}]$ be the limit of $\widehat{\mu}_{rk} (t; \beta)$, $r = 0, 1, 2, ~ k=1, \cdots, K$. Then the limit of the weighted covariate process for $\widehat{\eta}_k(t; \beta) = \widehat{\mu}_{1k} (t; \beta) / \widehat{\mu}_{0k} (t; \beta)$ becomes $\eta_{k0}(t; \beta) = \mu_{1k} (t; \beta) / \mu_{0k} (t; \beta)$. Let
\[
\Sigma_{\beta^0, k} = {E} [\{ X_{ki} - \eta_{k0}(Y_{ki}; \beta^0) \}^{\otimes 2} \delta_{ki}]
\]
be the information matrix for the $k$-th stratum, $k=1, \cdots, K$. The overall information matrix across all strata then becomes the weighted average of the stratum-specific information matrices,
\begin{equation} \label{popsig}
\Sigma_{\beta^0} = \sum_{k=1}^K r_k \Sigma_{\beta^0, k}.
\end{equation}
The inverse information matrix is $\Theta_{\beta^0} = \Sigma_{\beta^0}^{-1}$, which is to be approximated by $\widehat{\Theta}$ obtained in Section \ref{subsec:dslasso}.
The following theorem establishes the asymptotic normality of any linear combination of the estimated regression parameters, {$c^T \widehat{b}$} for some loading vector $c \in \mathbb{R}^p$, obtained by the proposed DBL-QP method. For an $m \times r$ matrix $A = (a_{ij})$, define the $\ell_1$-induced matrix norm $\| A \|_{1,1} = \max_{1 \le j \le r} \sum_{i=1}^m |a_{ij}|$. For two positive sequences $\{a_n\}$ and $\{b_n\}$, we write $a_n \asymp b_n$ if there exist two constants $C$ and $C^{\prime}$ such that $0 < C \le a_n / b_n \le C^{\prime} < \infty$. Let $s_0$ be the number of nonzero elements of $\beta^0$.
\begin{theorem} \label{thm:main}
Assume that the tuning parameters $\lambda$ and $\gamma$ satisfy $\lambda \asymp \sqrt{\log(p)/n_{min}}$ and \\ $\gamma \asymp \| \Theta_{\beta^0} \|_{1,1} \{ \max_{1\le k \le K} | n_k/N - r_k| + s_0 \lambda \} $, and that $\| \Theta_{\beta^0} \|_{1,1}^2 \{ \max_{1 \le k \le K}|n_k/N - r_k| + s_0 \lambda \} p\sqrt{\log(p)} \rightarrow 0$ as $n_{min} \rightarrow \infty$.
Under Assumptions \ref{assump1}--\ref{assump5} given in Appendix \ref{appB}, for any $c \in \mathbb{R}^p$ such that $\| c \|_2 = 1$, $\| c \|_1 \le a_*$ with $a_* < \infty$ being an absolute positive constant, and $\{ c^T \Theta_{\beta^0} c \}^{-1} = \mathcal{O}(1)$, we have
\[
\displaystyle \frac{\sqrt{N} c^T (\widehat {b} - \beta^0)}{(c^T \widehat{\Theta} c)^{1/2} }
\overset{\mathcal{D}}{\rightarrow} \mathcal{N}(0,1).
\]
\end{theorem}
\lu{Note that, instead of listing it as a regularity condition in Appendix \ref{appB}, we assume $\{ c^T \Theta_{\beta^0} c \}^{-1} = \mathcal{O}(1)$ in the above theorem because the vector $c$ is also defined here. A similar condition is assumed in \cite{van2014asymptotically} [Theorem 3.3 (vi)] which is weaker than uniformly bounding the maximum eigenvalue of $\Sigma_{\beta^0}$. } The hypothesis testing with $H_{0}: c^T \beta^0 - a_0=0$ versus $H_{1}: c^T \beta^0- a_0 \ne 0$ for some constants $c \in \mathbb{R}^p$ and $a_0$
entails various applications.
For example, by setting $a_0=0$ and $c$ to be a basis vector with only one element being 1 and all the others 0, we can draw
inference on any covariate in the presence of all the other covariates.
In particular, we will draw inference on the pairwise differences in graft failure risk
among donor age groups, e.g. between $(10, 20]$ and $(20,30]$ (the reference level) years old, and among patients with different primary kidney diagnoses (diabetes is the reference level); see Section \ref{sec:app}.
Given an appropriately chosen $c$ and with $T = \sqrt{N} (c^T \widehat{b} - a_0) / (c^T \widehat{\Theta} c)^{1/2}$, we construct a two-sided test function
\[
\phi(T) = \left\{ \begin{array}{ll}
1 & \quad \mathrm{if} \ |T| > z_{\alpha/2} \\
0 & \quad \mathrm{if} \ |T| \le z_{\alpha/2}
\end{array} , \right.
\]
where $z_{\alpha/2}$ is the upper $(\alpha/2)$-th quantile of the standard normal distribution. Corollary \ref{coro:power} provides the asymptotic type I error and power of the test $\phi(T)$, and Corollary \ref{coro:ci} formalizes the construction of level $\alpha$ confidence intervals for $c^T \beta^0$ which ensures the nominal coverage probability asymptotically.
\begin{corollary} \label{coro:power}
Under the conditions specified in Theorem \ref{thm:main},
${P} (\phi(T)=1 | H_0) \rightarrow \alpha$ as $n_{min} \rightarrow \infty$. Moreover, under $H_1: a_0 - c^T \beta^0 \ne 0$,
${P} (\phi(T) = 1 | H_1) \rightarrow 1$.
\end{corollary}
\begin{corollary} \label{coro:ci}
Suppose that the conditions in Theorem \ref{thm:main} hold.
Construct the random confidence interval $\mathcal{R}(\alpha) = \left[ c^T \widehat{b} - z_{\alpha/2} (c^T \widehat{\Theta} c / N)^{1/2},~ c^T \widehat{b} + z_{\alpha/2} (c^T \widehat{\Theta} c / N)^{1/2}\right]$. Then ${P} (c^T \beta^0 \in \mathcal{R}(\alpha)) \rightarrow 1 - \alpha$ as $n_{min} \rightarrow \infty$, where the probability is taken
under the true $\beta^0$.
\end{corollary}
Our asymptotic results facilitate simultaneous inference on multiple contrasts in the context of post-transplant renal graft failure.
For example, an important question to address is whether donor age is associated with graft failure.
With categorized donor age in our data analysis, simultaneous comparisons among the seven categories, e.g. $\le 10, (10, 20],
(20,30], (30,40], (40,50], (50, 60]$ and $60+$, naturally form multiple null contrasts.
These contrasts can be formulated by $J \beta^0$, where $J$ is an $m\times p$ matrix, and $m$ represents the number of linear combinations or contrasts.
The following theorem and corollary summarize the results for inference on multiple contrasts, $J \beta^0$. See an application of the asymptotic results to the SRTR data with $(m,p)=(6,94)$ in Section \ref{sec:app}.
\begin{theorem} \label{thm:simul}
Suppose that $J$ is an $m\times p$ matrix with $rank(J) = m$, $\| J \|_{\infty, \infty} = \mathcal{O}(1)$ and $J \Theta_{\beta^0} J^T \rightarrow F$, where $F$ is a nonrandom $m \times m$ positive definite matrix. Assume that the tuning parameters $\lambda$ and $\gamma$ satisfy $\lambda \asymp \sqrt{\log(p)/n_{min}}$ and $\gamma \asymp \| \Theta_{\beta^0} \|_{1,1} \{ \max_{1\le k \le K} | n_k/N - r_k| + s_0 \lambda \} $, and that $\| \Theta_{\beta^0} \|_{1,1}^2 \{ \max_{1 \le k \le K}|n_k/N - r_k| + s_0 \lambda \} p\sqrt{\log(p)} \rightarrow 0$ as $n_{min} \rightarrow \infty$. Under Assumptions \ref{assump1}--\ref{assump3}, \ref{assump5} and \ref{assump6} given in Appendix \ref{appB}, we have
\[
\sqrt{N} J (\widehat{b} - \beta^0) \overset{\mathcal{D}}{\rightarrow} \mathcal{N}_m (0,F).
\]
\end{theorem}
Here, $\| A \|_{\infty,\infty} = \max_{1 \le i \le m} \sum_{j=1}^r |a_{ij}|$ is the $\ell_{\infty}$-induced matrix norm for an $m \times r$ matrix $A = (a_{ij})$.
The theorem implies the following corollary, which constructs test statistics and multi-dimensional confidence regions with proper asymptotic type I error rates and nominal coverage probabilities.
\begin{corollary} \label{coro:thm2}
Suppose the conditions in Theorem \ref{thm:simul} hold. For an $m \times p$ matrix $J$ as specified in Theorem \ref{thm:simul}, and under $H_0: J \beta^0 = a^0 \in \mathbb{R}^m$,
\[T^{\prime} = N (J \widehat{b} - a^0)^T \widehat{F}^{-1} (J \widehat{b} - a^0) \overset{\mathcal{D}}{\rightarrow} \chi^2_m,
\]
where $\widehat{F} = J \widehat{\Theta} J^T$. Moreover, for an $\alpha \in (0,1)$,
define the random set $\mathcal{R}^{\prime}(\alpha) = \{ a \in \mathbb{R}^m: N (J \widehat{b} - a)^T \widehat{F}^{-1} (J \widehat{b} - a) < \chi^2_{m, \alpha} \}$, where $\chi^2_{m, \alpha}$ is the upper $\alpha$-th quantile of $\chi^2_m$. Then ${P} (J\beta^0 \in \mathcal{R}^{\prime}(\alpha)) \rightarrow 1 - \alpha$ as $n_{min} \rightarrow \infty$, where the probability is taken under the true $\beta^0$.
\end{corollary}
\section{Simulation study}
\label{sec:simul}
We conduct simulations to examine the finite sample performance of the proposed DBL-QP approach in correcting estimation biases and maintaining nominal coverage probabilities of confidence intervals. For comparisons, we also perform MSPLE, the oracle estimation, \lu{and the three inference methods [``Nodewise'' for \citet{kong2018high}, ``CLIME'' for \citet{yu2018confidence}, and ``Decor'' for \citet{fang2017testing}] that are adapted to stratified Cox models}. The following scenarios pertain to four combinations of $(K, n_k, p)$, where $K, n_k$ and $p$ are the number of strata, stratum-specific sample size and the number of covariates, respectively. Specifically, Scenarios 1--3 refer to $(K, n_k, p)=(10, 100,10)$, $(10, 100,100)$, and $(5, 200,100)$, respectively.
In Scenario 4, $K=40$, $p = 100$, $n_k$'s are simulated from a Poisson distribution with mean 40 and then fixed in all of the replications. This scenario mimics the situation of the recipient group aged over 60, the smallest group in the SRTR data.
Covariates $X_{ki}$ are simulated from $N_p (0, \Sigma_x)$ and truncated at $\pm 3$, where $\Sigma_x$ has an AR(1) structure with the $(i,j)$-th entry being $0.5^{|i-j|}$. The true regression parameters $\beta^0$ are sparse. Its first element $\beta_1^0$ varies from 0 to 2 by an increment of 0.2,
four additional elements are assigned values of 1, 1, 0.3 and 0.3 with their positions randomly generated and then fixed for all of the simulations, and all other elements are zero. The underlying survival times $T_{ki}$ are simulated from an exponential distribution with hazard $\lambda(t|X_{ki}) = \lambda_{0k} \exp\{X_{ki}^T\beta^0\}$, where $\lambda_{0k}$ are generated from $\mathrm{Uniform}(0.5,1)$ and then fixed throughout.
As in \citet{fang2017testing} and \citet{fan2002variable}, the censoring times $C_{ki}$'s are simulated independently from an exponential distribution with hazard $\lambda_c(t|X_{ki}) = 0.2 \lambda_{0k} \exp\{X_{ki}^T\beta^0\}$, resulting in an overall censoring rate around 20\%.
For the lasso estimator, we use 5-fold within-stratum cross-validation to select $\lambda$.
In Scenarios 1--3 with small numbers of strata, each stratum serves as a cross-validation fold for the selection of $\gamma$; in Scenario 4 with 40 strata, we perform 10-fold cross-validation as described in Algorithm \ref{algo1} and randomly assign 4 strata to each fold. For each parameter configuration, we simulate 100 datasets, based on which we compare estimation biases of $\beta^0_1$, 95\% confidence interval coverage probabilities, model-based standard errors, and empirical standard errors across the \lu{six} methods.
\begin{landscape}
\begin{figure}
\caption{Estimation bias, 95\% coverage, model-based standard error, and empirical standard error for $\beta^0_1$ of six different methods. Black horizontal lines are references to 0 for bias or 95\% for coverage probability.}
\label{fig:4scenarios}
\end{figure}
\end{landscape}
Figure \ref{fig:4scenarios} shows that, in Scenario 1 that features a small number of covariates ($p=10$), all \lu{six} methods perform well \lu{and similarly}; in Scenarios 2--4 with a relatively large number of covariates ($p=100$), which is close to the number of covariates in the real data we will analyze, \lu{our proposed DBL-QP estimator well corrects the biases of the lasso estimates and maintains good confidence interval coverage (excluding the practically impossible ``Oracle'' estimator), but MSPLE, Nodewise, Decor and CLIME all present larger biases compared to DBL-QP as $\beta^0_1$ increases from 0 to 2. CLIME, Nodewise and MSPLE have worse confidence interval coverage in general.
As de-biased lasso methods, CLIME and Nodewise produce much smaller model-based standard error estimates, which also contribute to their poor covarage probabilities. This is likely due to that both methods (CLIME and Nodewise) use penalized estimators for inverse information matrix estimation, and such penalization induces biases towards zero.}
To recapitulate, the proposed DBL-QP provides less biased estimates and better confidence interval coverage than \lu{the conventional MSPLE and three other competitors (Nodewise, Decor and CLIME adapted to the stratified setup)} when the sample size is moderate relative to the number of covariates, although \lu{all methods} give almost identical results when $p$ is rather small. Hence, when $p < N$, our proposed DBL-QP approach is at least as good as \lu{all the other methods}, and should be recommended for use.
\section{Analysis of the SRTR kidney transplant data}
\label{sec:app}
\lu{The SRTR data set features 94 covariates from both donors and recipients, and the number of covariates is seen as relatively large for some recipient groups. With its reliable performance as demonstrated in simulations, we apply our DBL-QP approach to analyze the SRTR data, while using MSPLE as a benchmark.} The outcome is graft failure free survival, the time from transplant to graft failure or death, whichever comes first. Our primary goal is to investigate the joint associations of these covariates with graft failure for three recipient groups defined in Table \ref{tab:character} separately. By simultaneously considering all available donor and recipient covariates,
we aim to account for confounding and provide asymptotically valid inference for the covariate effects, which differs from post hoc inference that only focuses on a smaller set of covariates selected by stepwise selection. The effect of donor age, in the presence of other risk factors, is worth investigating, as the debatable ``one-size-fit-all'' practice of donor-recipient age matching unfortunately is not suited for the benefit of transplantation \citep{keith2004effect,veroux2012age,dayoub2018effects}.
\subsection{Data details}
Included in our analysis are 9,195 recipients who received kidney-only transplants from deceased donors, had no prior solid organ transplants, and were at least 18 years old at the time of transplantation during 2000 and 2001. We focus on those with these same cohort years in order to eliminate the cohort effect. Moreover, this group of patients had longer follow-up than those from the later cohort years. See Appendix \ref{appA} for a full list of included variables in the analysis.
In the three receipts' age groups, respectively, the sample sizes are 3388, 4359 and 1448, the censoring rates are 53.1\%, 46.5\% and 30.0\%, the median numbers of patients within each transplant center are 32, 31 and 27, and the restricted mean survival times by $13$ years are 9.1, 8.6 and 7.1 years.
To select the tuning parameters, we implement 5-fold cross-validation
by randomly selecting one fifth of transplant centers without replacement as testing data and the rest as training data.
\subsection{Results}
We begin with examining the overall effect of donors' age on graft failure and testing the null hypothesis that, within each recipient group and after adjusting for the other risk factors, all the donor age groups, i.e. $\le 10, (10, 20],
(20,30], (30,40], (40,50], (50, 60]$ and $60+$, have the same risk of graft failure. Based on Theorem \ref{thm:simul} and Corollary \ref{coro:thm2}, with $(m,p)=(6,94)$, we perform tests for the null contrasts, and the obtained statistics significantly reject the null hypotheses for all three recipient groups (within recipients aged 18-45: $\chi^2=40.4$, df=6, p-value=$3.9 \times 10^{-7}$; recipients aged 45-60: $\chi^2=34.5$, df=6, p-value=$5.3 \times 10^{-6}$; recipients aged over 60: $\chi^2=14.2$, df=6, p-value=$2.8 \times 10^{-2}$). Indeed, Figure \ref{fig:don_age}, which depicts the risk-adjusted effect of donors' age across the three recipient age groups,
shows a general trend of increasing hazards for those receiving kidneys from older donors, likely due to renal aging. The estimates and confidence intervals obtained by our
proposed DBL-QP differ from those obtained by MSPLE, and the differences are the most obvious in the $60+$ year recipient group, which has the smallest sample size. As presented in our simulations, MSPLE may produce biased estimates with improper confidence intervals, especially when the sample size is relatively small.
On the other hand, the proposed DBL-QP method may shed new light into the aging effect, which seems to be non-linear with respect to donors' age. First,
using the results of Theorem \ref{thm:main} and Corollary \ref{coro:ci}, our tests detect no significant differences in hazards between those receiving kidneys from donors aged under 10 or $(10, 20]$ and $(20, 30]$ (reference level) years old, within all the three recipient age groups.
Second, significantly increased hazards are observed as early as when donors' age reached 30-40, as compared to the reference level of $(20, 30]$, in the 18-45 years old recipient group, with an estimated hazard ratio (HR) of 1.16 (95\% CI: 1.01--1.34, p-value=$4.1\times 10^{-2}$). In contrast, there are no significant differences between receiving organs from $(30,40]$ years old donors and the reference level of $(20, 30]$, among the
45-60 years old recipients
(HR= 0.96, 95\% CI: 0.85--1.09, p-value=$5.1\times 10^{-1}$) and the $60+$ years old recipients (HR=1.07, 95\% CI: 0.88--1.30, p-value=$5.0\times 10^{-1}$).
Third, kidneys from $60+$ years old donors confer the highest hazards, with the estimated risk-adjusted HRs (compared to the reference level $(20, 30]$) being 1.83 (95\% CI: 1.48--2.28, p-value=$4.3 \times 10^{-8}$), 1.40 (95\% CI: 1.21--1.61, p-value=$4.1\times 10^{-6}$) and 1.37 (95\% CI: 1.14--1.63, p-value=$5.2 \times 10^{-4}$) among the three recipient age groups respectively.
This means that, compared to the older recipients, recipients of 18-45 years old tend to experience a greater hazard of graft failure when receiving kidneys from donors over 60 years old.
Caution needs to be exercised when allocating kidneys from older donors to young patients \citep{lim2010donor,kabore2017age,dayoub2018effects}.
\begin{figure}
\caption{Estimated hazard ratios and the corresponding $95\%$ confidence intervals of different donor age categories with reference to the $(20, 30]$ donor age category, after adjusting for all other variables, in three recipient groups.
}
\label{fig:don_age}
\end{figure}
Our method also delineates the associations of clinical indicators with graft failure, provides more reliable inference, and compares the relative strengths across recipient age groups. \lu{By naively applying lasso, 64, 44 and 27 covariates are selected with non-zero coefficients in the 18-45, 45-60, and 60$+$ years old recipient groups, respectively. In contrast, the proposed DBL-QP identifies 22, 22 and 14 significant covariates in these three recipient groups, respectively, from rigorous hypothesis tests with size 0.05 based on the asymptotic distribution.}
Figure \ref{fig:ci_three} shows the estimated coefficients and their 95\% confidence intervals for covariates that are significant at level 0.05 in at least one recipient group. We highlight several noteworthy results.
First, recipients' primary kidney diagnosis plays a critical role in kidney graft failure \citep{wolfe1991survival}. Compared to recipients with primary diagnosis of
diabetes (the reference level), those with polycystic kidneys (variable 2 in Figure \ref{fig:ci_three}) have a reduced risk of graft failure, with highly significant lower HRs of 0.54 (95\% CI: 0.42--0.70, p-value=$3.6 \times 10^{-6}$), 0.65 (95\% CI: 0.57--0.75, p-value=$4.4 \times 10^{-9}$) and 0.74 (95\% CI: 0.60--0.92, p-value=$5.3 \times 10^{-3}$) for the three age groups respectively.
Compared to diabetes,
primary diagnosis of
glomerular disease (variable 26 in Figure \ref{fig:ci_three}) is significantly associated with a reduced risk of graft failure only in the $60+$ years old recipient group (HR=0.79, 95\% CI: 0.66--0.96, p-value=$1.4 \times 10^{-2}$), and primary diagnosis of hypertensive nephrosclerosis (variable 29 in Figure \ref{fig:ci_three}) is
significantly associated with a higher hazard of graft failure only in the 45-60 years old recipient group (HR=1.12, 95\% CI: 1.01--1.23, p-value=$2.5 \times 10^{-2}$).
Second, since diabetes is the most prevalent among end-stage renal patients \citep{kovesdy2010glycemic}, we code recipients' diabetic status at transplant as non-diabetic (reference level), diabetic for 0-20 years (variable 13 in Figure \ref{fig:ci_three}), and 20+ years (variable 3 in Figure \ref{fig:ci_three}).
Our stratified analysis reveals that diabetics is a stronger risk factor for young recipients aged between 18 and 45 years old than for older recipients, regardless of duration of diabetes.
Third, instead of using the total number of mismatches as done in the literature, we consider the number of mismatches separately for each
HLA locus for more precisely pinpointing the effects of mismatching loci. Our results reveal that the HLA-DR mismatches (variable 9 in Figure \ref{fig:ci_three}) are more strongly associated with graft failure than the HLA-A (variable 18 in Figure \ref{fig:ci_three}) and HLA-B mismatches (non-significant in any recipient group), which are consistent with a meta-analysis based on 500,000 recipients \citep{shi2018impact}.
Finally, to study the granular impact of recipient age on graft failure \citep{karim2014recipient}, we treat recipient age (divided by 10) as a continuous variable (variable 4 in Figure \ref{fig:ci_three}) in the model within each recipient age group. Interestingly, we find that
increasing age is associated with a higher hazard
in the two older recipient groups (HR=1.31, 95\% CI: 1.19--1.44, p-value=$1.3\times 10^{-8}$, for recipients aged 45-60; HR=1.22, 95\% CI: 1.07--1.40, p-value=$3.6\times 10^{-3}$, for recipients aged 60+),
but with a lower hazard of graft failure in the 18-45 recipient age group (HR=0.89, 95\% CI: 0.83--0.95, p-value=$5.2 \times 10^{-4}$).
This is likely because that younger patients generally had poorer adherence to treatment, resulting in higher risks of graft loss \citep{kabore2017age}. The results also reinforce the necessity of separating analyses for different recipient age groups.
\begin{figure}
\caption{Estimated regression coefficients in the stratified Cox models using the proposed DBL-QP method, and the corresponding $95\%$ confidence intervals, presented by recipient age group. The covariates included are significant at level 0.05 in at least one recipient group, after adjusting for all other covariates.}
\label{fig:ci_three}
\end{figure}
As a side note, we compare DBL-QP and MSPLE in the estimated coefficients and standard errors. Figure \ref{fig:QPvsMPLE_all} shows that in the 45-60 age group with the largest number of subjects, the point estimates obtained by the two methods almost coincide with each other, whereas in the $60+$ age group with the smallest sample size, MSPLE tends to have larger absolute estimates than the de-biased lasso. Moreover, the standard errors estimated by MSPLE are likely to be larger than those by our method across all the age groups. These observations agree with the results of our simulations (Scenarios 2--4), which show that MSPLE yields large biases in estimated coefficients and standard errors, especially when the sample size is relatively small, whereas our proposed DBL-QP method draws more valid inferences by maintaining proper type I errors and coverage probabilities.
\begin{figure}
\caption{Comparison between the coefficient estimates (top) and the model-based standard errors (bottom) by the de-biased lasso (DBL-QP) and the maximum stratified partial likelihood estimation (MSPLE) in three recipient age groups. The solid red lines are 45-degree reference lines, and the dotted blue lines in the top figures represent the fitted linear regression of the DBL-QP estimates on the MSPLE estimates.}
\label{fig:QPvsMPLE_all}
\end{figure}
\section{Concluding remarks}
\label{sec:discuss}
The work is motivated by
an urgent call of better understanding the complex mechanisms behind post-kidney transplant graft failure. Our modeling framework is
Cox models stratified by transplant centers, due to their strong confounding effects on graft failure.
To adjust for confounders to the extent possible, we have included an extended list of 94 covariates from recipients and donors, which has not been done in the literature. A particular scientific question to address is the debatable donor-recipient age matching criterion in kidney transplantation. Fitting separate models by recipient age enables direct assessments of the donor age effects in different recipient age groups, which differs from using donor-recipient age difference as in \citet{ferrari2011effect}.
Specifically, we have followed a common practice of fitting separate models in age groups of 18-45, 45-60 and $60+$ years. The commonly used MSPLE yielded biased estimates and unreliable inference in some smaller age groups, though the samples outnumbered the covariates. In particular, the $60+$ years recipient group had only 1448 recipients in 43 different transplant centers, and MSPLE yielded more dramatic estimates for those donor age effects of over 30 years old (Figure \ref{fig:don_age}). Our simulation results also confirmed such a problematic phenomenon. Therefore, a statistical method that can guarantee reliable estimates and valid inference is much needed for delineating the associations of interest with graft failure when the number of covariates is relatively large in stratified Cox models.
Inspired by the de-biased lasso method for linear regression \citep{javanmard2014confidence}, we have developed a de-biased lasso approach via quadratic programming for stratified Cox models. \lu{Despite progress made in high-dimensional inference for Cox models, virtually no work has considered stratified settings, theoretically or empirically. We have shown that
in the ``large $N$, diverging $p$'' scenario, our approach possesses desirable
asymptotic properties and finite-sample performance, and is more suitable for the analysis of the SRTR data than the competing methods illustrated in our simulation studies.
Computationally, based on a previous work on Cox models without stratification \citep{xia2021cox}, for the estimation of $\Theta_{\beta^0}$, the computational speed using \texttt{solve.QP} in R was much faster than that using the R packages \texttt{clime} or \texttt{flare} adopted by \citet{yu2018confidence}. }
Applications of our method to the SRTR data generated new biological findings. After categorizing donors' age and controlling for other risk factors listed in Appendix \ref{appA}, we find that {organs from older donors} are associated with an increased hazard of graft failure and that the dependence on donors' age is non-linear: within the youngest recipient group (18-45 years), significant differences from the reference donor age category (20-30 years) were detected as early as when donors reached 30-40 years old, whereas significant differences were detected only when donors reached 50-60 or 60+ years
within the two older recipient groups, respectively; in other words, receiving kidneys from older or younger donors, such as $60+$ versus 20-30 years, presented larger differences than in the other two recipient groups. These results, which were not reported in the literature, may provide new empirical evidence to aid stake-holders, such as patients, families, physicians and policy makers, in decisions on donor-recipient age matching.
\lu{A few technical points are noteworthy. First,} our work deals with the ``large $N$, diverging $p$'' scenario, as embedded in the motivating data, and approximates $\Theta_{\beta^0}$ via quadratic programming without positing any sparsity conditions on $\Theta_{\beta^0}$. This distinguishes from the related literature \citep{fang2017testing,yu2018confidence,kong2018high} \lu{in the ``large $p$, small $N$'' scenario} that relies upon sparsity conditions on the numbers of non-zero elements in the rows of $\Theta_{\beta^0}$, which \lu{are hardly discussed in depth and may not hold nor have explicit interpretations for Cox models}. \lu{For example, when the rows of $\Theta_{\beta^0}$ are not sparse, our dimension requirement for $p$ is less stringent than in \citet{yu2018confidence}, by a factor of $\sqrt{\log(Np)}$. } \lu{Moreover, when $p > N$, the several de-biased methods aforementioned may not yield reliable inference results, as empirically $\Theta_{\beta^0}$ cannot be estimated well, and biases in the lasso estimator are often not sufficiently corrected for in this scenario for Cox models. New approaches, such as sample-splitting approaches \citep{fei2021estimation}
that bypass the estimation of $\Theta_{\beta^0}$, can be consulted.}
\lu{Second, tuning parameter selection is critical in high-dimensional inference. Our proposed method deploys a single tuning parameter $\gamma$ for de-biasing the estimates of all $\beta_j$'s. This is a computationally feasible and commonly adopted strategy, presenting a satisfactory performance in our numerical studies, and can be extended to adapt to the variability of individual coefficient estimation. For example, one may consider the following estimation procedure for the $j$th row of $\widehat{\Theta}$ along the line of adaptive CLIME \citep{cai2016estimating}: \[
\min_m \{ m^T \widehat{\Sigma} m: | (\widehat{\Sigma} m - e_j)_k | \le \gamma_{jk}, k = 1, \ldots, p \}. \] Here, $\gamma_{jk}$'s are supposed to be adaptively estimated through a carefully designed procedure. However, the design of such an appropriate procedure requires complicated theoretical analysis in Cox models, unstratified or stratified, to determine the desirable rates of $ \gamma_{jk}$'s, among other tasks. Given that such complexity is beyond the scope of this paper, we will not pursue this route here in details but will leave it for future research.}
\lu{ Third, though primarily focusing on the associations between the risk factors and survival (through Theorem \ref{thm:main}), the proposed method can be used for patient risk scoring and conditional survival probability estimation. For example, the de-biased estimates may be plugged into the Breslow's estimator \citep{kalbfleisch2002statistical} for stratum-specific baseline hazards. The conditional survival probability estimation may not go beyond the time point $\tau$ due to censoring. }
\lu{ Lastly, we use Cox models stratified by transplant centers to account for but avoid explicitly modeling the center effects. Alternatively, random effects models can be used for clustered survival data analysis; for example, \citet{vaida2000proportional} generalized the usual frailty model to allow multivariate random effects. However, in a random effects model, the distribution of random effects needs to be specified, and the coefficients only have conditional interpretations, given a cluster. We may pursue this elsewhere.
}
\lu{We have implemented the proposed DBL-QP method with cross-validation in R and Rcpp, which is available online at \href{https://github.com/luxia-bios/StratifiedCoxInference/}{https://github.com/luxia-bios/StratifiedCoxInference/} with simulated examples.}
\section*{Appendix}
\begin{appendix} \section{SRTR data}\label{appA} The SRTR dataset analyzed in this article can be accessed by applying through the OPTN website \href{https://optn.transplant.hrsa.gov}{https://optn.transplant.hrsa.gov}. The interpretation and reporting of the SRTR data results are solely the responsibility of the authors and should not be viewed as official opinions of the SRTR or the United States Government.
The 94 covariates, including dummy variables, are derived from the following factors. Donor factors include: ABO blood type, age, cytomegalovirus antibody, hepatitis C virus antibody, cause of death, cardiac arrest since event leading to declaration of death, serum creatinine, medication given to donor (DDAVP, dopamine and dobutamine), gender, height, history of cancer, cigarette smoking, history of drug abuse, hypertension, diabetes, inotropic support, inotropic agents at time of incision, non-heart beating donor, local or shared organ transplant, race, and weight.
Recipient factors include: ABO blood type, history of diabetes and duration, angina/coronary artery disease, symptomatic peripheral vascular disease, drug treated systemic hypertension, drug treated COPD, gender (and previous pregnancies for females), sensitization (whether peak and/or current panel-reactive antibodies exceed 20\%), previous malignancy, peptic ulcer disease, symptomatic cerebrovascular disease, race, total serum albumin, age at transplant, number of HLA mismatches (A, B and DR), cytomegalovirus status, total cold ischemic time, primary kidney diagnoses, pre-transplant dialysis and duration, the Epstein–Barr virus serology status, employment status, hepatitis B virus status, hepatitis C virus status, height, pre-implantation kidney biopsy, pre-transplant blood transfusions, transplant procedure type, warm ischemic time and weight.
\section{Regularity conditions}\label{appB} \renewcommand{B.\arabic{assumption}}{B.\arabic{assumption}}
Assumptions \ref{assump1}--\ref{assump5} below ensure that Theorem \ref{thm:main} hold.
\begin{assumption}\label{assump1}
Covariates are almost surely uniformly bounded, i.e. $\| X_{ki} \|_{\infty} \le M$ for some positive constant $M<\infty$ for all $k$ and $i$.
\end{assumption}
\begin{assumption}
\label{assump2}
$| X_{ki}^T \beta^0 | \le M_1$ uniformly for all $k$ and $i$ with some positive constant $M_1 < \infty$ almost surely.
\end{assumption}
\begin{assumption}
\label{assump3}
The follow-up time stops at a finite time point $\tau > 0$, with probability $\pi_0 = \min_{k} P (Y_{ki} \ge \tau) > 0$.
\end{assumption}
\begin{assumption}
\label{assump4}
\lu{ For any $t\in [0, \tau]$,
\[
\frac{c^T \Theta_{\beta^0}}{c^T \Theta_{\beta^0} c} \left[ \sum_{k=1}^K r_k \int_0^t \left\{ \mu_{2k}(u; \beta^0) - \frac{\mu_{1k}(u; \beta^0) \mu_{1k}(u; \beta^0)^T}{\mu_{0k}(u; \beta^0)} \right\} \lambda_{0k}(u) d u \right] \Theta_{\beta^0} c \rightarrow v(t; c)
\]
as $n \rightarrow \infty$ for some function $v(t; c) > 0$ of $t$ that also depends on the choice of $c$. }
\end{assumption}
\begin{assumption}
\label{assump5}
\lu{There exists a constant $\epsilon_0>0$ such that $\lambda_{\mathrm{min}}(\Sigma_{\beta^0}) \ge \epsilon_0$, where $\lambda_{\mathrm{min}} (\cdot)$ is the smallest eigenvalue of a matrix.
}
\end{assumption}
For inference on multiple linear combinations or contrasts as described in Theorem \ref{thm:simul}, Assumption \ref{assump4} needs to be replaced with the following Assumption \ref{assump6}, which is a multivariate version of Assumption \ref{assump4}.
\begin{assumption}
\label{assump6}
\lu{ For any $\omega \in \mathbb{R}^{m}$ and any $t\in [0, \tau]$,
\[
\frac{\omega^T J \Theta_{\beta^0} }{\omega^T J \Theta_{\beta^0} J^T \omega} \left[ ~ \sum_{k=1}^K r_k \int_0^t \left\{ \mu_{2k}(u; \beta^0) - \frac{\mu_{1k}(u; \beta^0) \mu_{1k}(u; \beta^0)^T}{\mu_{0k}(u; \beta^0)} \right\} d\Lambda_{0k}(u) \right] \Theta_{\beta^0} J^T \omega
\]
converges to $v^{\prime}(t; \omega, J)$ as $n \rightarrow \infty$, for some function $v^{\prime}(t; \omega, J) > 0$ of $t$, that also depends on the choice of $\omega$ and $J$. }
\end{assumption}
\lu{It is common in the literature of high-dimensional inference to assume bounded covariates as in Assumption B.1. \citet{fang2017testing} and \citet{kong2018high} also posed Assumption B.2 for Cox models, i.e. uniform boundedness on the multiplicative hazard. Under Assumption B.1, Assumption B.2 can be implied by the bounded overall signal strength $\| \beta^0 \|_1$. Assumption B.3 is a common assumption in survival analysis \citep{andersen1982cox}. Assumption B.4 and its multivariate version, Assumption B.6, ensure the convergence of the variation process, which is key in applying the martingale central limit theorem. They are less stringent comparing to the boundedness assumption on $\| \Theta_{\beta^0} X_{ki} \|_{\infty}$ that is equivalent to the assumptions for statistical inference in \citet{van2014asymptotically} on high-dimensional generalized linear models and in \citet{fang2017testing} on high-dimensional Cox models. The boundedness of the smallest eigenvalue of $\Sigma_{\beta^0}$ in Assumption B.5 is common in inference for high-dimensional models \citep{van2014asymptotically,kong2018high}.
Since we focus on random designs, unlike \citet{huang2013oracle}, \citet{yu2018confidence} and \citet{fang2017testing}, we do not directly assume the compatibility condition on $\ddot{\ell}(\beta^0)$; instead, we impose Assumption B.5 on the population-level matrix $\Sigma_{\beta^0}$, which leads to the compatibility condition for a given data set with probability going to one.}
\end{appendix}
\setcounter{section}{0} \renewcommand{S.\arabic{section}}{S.\arabic{section}} \renewcommand{S.\arabic{equation}}{S.\arabic{equation}} \renewcommand{S.\arabic{lemma}}{S.\arabic{lemma}}
\begin{center}
{\bf \Large
Supplement to ``De-biased lasso for stratified Cox models with application to the national kidney transplant data"} \\[1em]
{\large Lu Xia\textsuperscript{a}, Bin Nan\textsuperscript{b}, and Yi Li\textsuperscript{c} \\[1em] }
{\small
\textsuperscript{a}Department of Biostatistics, University of Washington, Seattle, WA \\ \textsuperscript{b}Department of Statistics, University of Californina, Irvine, CA \\ \textsuperscript{c}Department of Biostatistics, University of Michigan, Ann Arbor, MI }
\end{center}
For completeness of presentation, we first provide some useful lemmas and their proofs, and then give the proofs of the main theorems in this supplementary material. Since corollaries are direct results of the corresponding theorems, their proofs are straightforward and thus omitted.
\section{Technical Lemmas}
For the $i$th subject in the $k$th stratum, define the counting process $N_{ki}(t) = 1(Y_{ki} \le t, \delta_{ki} = 1)$. The corresponding intensity process $A_{ki}(t; \beta) = \int_0^t 1 (Y_{ki} \ge s) \exp(X_{ki}^T \beta) d\Lambda_{0k}(s)$, where $\Lambda_{0k}(t) = \int_{0}^{t} \lambda_{0k}(s) ds$ is the baseline cumulative hazard function for the $k$th stratum, $k=1, \cdots, K$, $i = 1, \cdots, n_k$. Let $M_{ki}(t; \beta) = N_{ki}(t) - A_{ki}(t; \beta)$, then $M_{ki}(t; \beta^0)$ is a martingale with respect to the filtration $\mathcal{F}_{ki}(t) = \sigma \{ N_{ki}(s), 1(Y_{ki} \ge s), X_{ki}: s \in (0, t] \}$.
Recall that the stratum-specific weighted covariate process $\widehat{\eta}_k(t; \beta) = \widehat{\mu}_{1k} (t; \beta) / \widehat{\mu}_{0k} (t; \beta)$, where $\widehat{\mu}_{rk} (t; \beta) = (1/ n_k) \sum_{i=1}^{n_k} 1(Y_{ki} \ge t) X_{ki}^{\otimes r} \exp\{X_{ki}^T \beta\}$. Their population-level counterparts are $\mu_{rk} (t; \beta) = {E} [1(Y_{k1} \ge t) X_{k1}^{\otimes r} \exp\{X_{k1}^T \beta\}]$ and $\eta_{k0}(t; \beta) = \mu_{1k} (t; \beta) / \mu_{0k} (t; \beta)$, $r = 0, 1, 2$, $k=1, \cdots, K$. It is easily seen that the process $\{X_{ki} - \widehat{\eta}_k(t; \beta^0)\}$ is predictable with respect to the filtration $\mathcal{F}(t) = \sigma \{ N_{ki}(s), 1(Y_{ki} \ge s), X_{ki}: s \in (0, t], k=1, \cdots, K, ~ i=1, \cdots, n_k \}$.
\begin{lemma} \label{chap4:lemma:mom}
Under Assumptions B.1--B.3, for $k = 1, \cdots, K$, we have
\begin{align*}
& \sup_{t \in [0, \tau]} | \widehat{\mu}_{0k}(t; \beta^0) - \mu_{0k}(t; \beta^0) | = \mathcal{O}_{P}(\sqrt{\log(p) / n_k}), \\
& \sup_{t \in [0, \tau]} \| \widehat{\mu}_{1k}(t; \beta^0) - \mu_{1k}(t; \beta^0) \|_{\infty} = \mathcal{O}_{P}(\sqrt{\log(p)/n_k}), \\
& \sup_{t \in [0, \tau]} \| \widehat{\eta}_k(t; \beta^0) - \eta_{k0}(t; \beta^0) \|_{\infty} = \mathcal{O}_{P}(\sqrt{\log(p)/n_k}).
\end{align*} \end{lemma}
Lemma \ref{chap4:lemma:mom} is the extension of the results of Lemma A1 in \citet{xia2021cox} to each of the $K$ strata. We omit its proof here.
\begin{lemma} \label{chap4:lemma:lead}
Assume $ \| \Theta_{\beta^0} \|_{1,1}^2 \{ \sqrt{\log(p) / n_{min}} + \max_k | n_k/N - r_k | \} \to 0$, and $\{ c^T \Theta_{\beta^0} c \}^{-1} = \mathcal{O}(1)$. Under Assumptions B.1--B.5, for any $c \in \mathbb{R}^p$ such that $\| c \|_2 = 1$ and $\| c \|_1 \le a_*$ with some absolute constant $a_* < \infty$, we have
\[
\displaystyle \frac{ \sqrt{N} c^T \Theta_{\beta^0} \dot{\ell}(\beta^0)}{\sqrt{c^T \Theta_{\beta^0} c}} \overset{\mathcal{D}}{\rightarrow} N(0,1).
\] \end{lemma}
\begin{proof}[\textbf{Proof of Lemma \ref{chap4:lemma:lead}}]
We rewrite
\begin{align} \label{chap4:eq:clt_decomp}
\displaystyle \frac{- \sqrt{N} c^T \Theta_{\beta^0} \dot{\ell}(\beta^0)}{\sqrt{c^T \Theta_{\beta^0} c}} & = \displaystyle \frac{1}{\sqrt{N}} \sum_{k=1}^K \sum_{i=1}^{n_k} \frac{c^T \Theta_{\beta^0}}{\sqrt{c^T \Theta_{\beta^0} c}} \left\{ X_{ki} - \frac{\widehat{\mu}_{1k}(Y_{ki}; \beta^0)}{\widehat{\mu}_{0k}(Y_{ki}; \beta^0)} \right\} \delta_{ki} \nonumber \\
& = \frac{1}{\sqrt{N}} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{\tau} \frac{c^T \Theta_{\beta^0}}{\sqrt{c^T \Theta_{\beta^0} c}} \left\{ X_{ki} - \frac{\widehat{\mu}_{1k}(t; \beta^0)}{\widehat{\mu}_{0k}(t; \beta^0)} \right\} dN_{ki}(t) \nonumber \\
& = \frac{1}{\sqrt{N}} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{\tau} \frac{c^T \Theta_{\beta^0}}{\sqrt{c^T \Theta_{\beta^0} c}} \left\{ X_{ki} - \frac{\widehat{\mu}_{1k}(t; \beta^0)}{\widehat{\mu}_{0k}(t; \beta^0)} \right\} dM_{ki}(t).
\end{align}
Denote $\displaystyle U(t) = \frac{1}{\sqrt{N}} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{t} \frac{c^T \Theta_{\beta^0}}{\sqrt{c^T \Theta_{\beta^0} c}} \left\{ X_{ki} - \frac{\widehat{\mu}_{1k}(s; \beta^0)}{\widehat{\mu}_{0k}(s; \beta^0)} \right\} dM_{ki}(s)$. Then the variation process for $U(t)$ is
\begin{align}
\langle U \rangle (t) & \displaystyle = \sum_{k=1}^K \sum_{i=1}^{n_k} \frac{1}{N} \int_0^t \frac{c^T \Theta_{\beta^0}}{c^T \Theta_{\beta^0} c} \left\{ X_{ki} - \widehat{\eta}_k(u; \beta^0) \right\}^{\otimes 2} 1(Y_{ki} \ge u) e^{X_{ki}^T \beta^0} d\Lambda_{0k}(u) \Theta_{\beta^0} c \nonumber \\
& = \displaystyle \frac{c^T \Theta_{\beta^0}}{c^T \Theta_{\beta^0} c} \left[ \sum_{k=1}^K \frac{n_k}{N} \int_0^t \left\{ \widehat{\mu}_{2k}(u; \beta^0) - \frac{\widehat{\mu}_{1k}(u; \beta^0) \widehat{\mu}_{1k}^T(u; \beta^0)}{\widehat{\mu}_{0k}(u; \beta^0)} \right\} d\Lambda_{0k}(u) \right] \Theta_{\beta^0} c.
\end{align}
Following the proof of Lemma A2 in \citet{xia2021cox}, we have
\begin{align*}
& \left\| \displaystyle \int_0^t \left\{ \widehat{\mu}_{2k} (u; \beta^0) - \frac{\widehat{\mu}_{1k}(u; \beta^0) \widehat{\mu}_{1k}(u; \beta^0)^T}{\widehat{\mu}_{0k}(u; \beta^0)} \right\} d \Lambda_{0k}(u) - \right. \\
& \qquad \displaystyle \left. \int_0^t \left\{ {\mu}_{2k} (u; \beta^0) - \frac{{\mu}_{1k}(u; \beta^0) {\mu}_{1k}(u; \beta^0)^T}{{\mu}_{0k}(u; \beta^0)} \right\} d \Lambda_{0k}(u) \right\|_{\infty} = \mathcal{O}_{P}(\sqrt{\log(p) / n_k})
\end{align*}
uniformly for all $t \in [0, \tau]$. Then,
\begin{align*}
& \displaystyle \frac{c^T \Theta_{\beta^0}}{c^T \Theta_{\beta^0} c} \left[ \sum_{k=1}^K \frac{n_k}{N} \int_0^t \left\{ \widehat{\mu}_{2k}(u; \beta^0) - \frac{\widehat{\mu}_{1k}(u; \beta^0) \widehat{\mu}_{1k}^T(u; \beta^0)}{\widehat{\mu}_{0k}(u; \beta^0)} \right\} d\Lambda_{0k}(u) \right] \Theta_{\beta^0} c \\
= & \frac{c^T \Theta_{\beta^0}}{c^T \Theta_{\beta^0} c} \left[ \sum_{k=1}^K r_k \int_0^t \left\{ \mu_{2k}(u; \beta^0) - \frac{\mu_{1k}(u; \beta^0) \mu_{1k}^T(u; \beta^0)}{\mu_{0k}(u; \beta^0)} \right\} d\Lambda_{0k}(u) \right] \Theta_{\beta^0} c ~ + ~ \\
& \qquad \mathcal{O}_{P} \left\{ \| c \|_1^2 \| \Theta_{\beta^0} \|_{1,1}^2 \left( \max_k |n_k/N - r_k| + \sqrt{\log(p)/ n_{min}} \right) \right\}.
\end{align*}
By Assumption B.4, $\langle U \rangle (t) \rightarrow_{P} v(t; c) > 0, ~ t \in [0, \tau]$.
For any $\epsilon > 0$, define $G_{ki}(u) = \displaystyle \frac{1}{\sqrt{N}} \frac{c^T \Theta_{\beta^0}}{\sqrt{c^T \Theta_{\beta^0} c}} \left\{ X_{ki} - \frac{\widehat{\mu}_{1k}(u; \beta^0)}{\widehat{\mu}_{0k}(u; \beta^0)} \right\}$ and the truncated process
$
\displaystyle U_{\epsilon}(t) = \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{t} G_{ki}(u) 1(|G_{ki}(u)| > \epsilon) dM_{ki}(u).
$
The variation process of $U_{\epsilon}(t)$ is
\[
\langle U_{\epsilon} \rangle (t) = \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{t} G^2_{ki}(u) 1(|G_{ki}(u)| > \epsilon) dA_{ki}(u),
\]
where $d A_{ki}(u) = 1(Y_{ki} \ge u) e^{X_{ki}^T \beta^0} d\Lambda_{0k}(u)$.
Since \[
| \sqrt{N} G_{ki}(u) | \le a_* \| \Theta_{\beta^0} \|_{1,1} 2M \{ c^T \Theta_{\beta^0} c \}^{-1/2} = \mathcal{O}(\| \Theta_{\beta^0} \|_{1,1}),
\]
then $1(|G_{ki}(u)| > \epsilon) = 0$ almost surely as $ \| \Theta_{\beta^0} \|_{1,1} / \sqrt{N} \asymp \| \Theta_{\beta^0} \|_{1,1} / \sqrt{n_{min}} \to 0$. So $\langle U_{\epsilon} \rangle (t) \rightarrow_{P} 0$. By the martingale central limit theorem, we obtain the desirable result. \end{proof}
\begin{lemma} \label{chap4:lemma:lasso}
Under Assumptions B.1--B.3 and B.5, for $\lambda \asymp \sqrt{\log(p)/n_{min}}$, the lasso estimator $\widehat{\beta}$ satisfies
\[
\| \widehat{\beta} - \beta^0 \|_1 = \mathcal{O}_{P}(s_0 \lambda), ~ \displaystyle \frac{1}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} |X_{ki}^T (\widehat{\beta} - \beta^0)|^2 = \mathcal{O}_{P}(s_0 \lambda^2).
\] \end{lemma}
\begin{proof}[\textbf{Proof of Lemma \ref{chap4:lemma:lasso}}]
This result is adapted from the proof in \citet{kong2014non}, with modifications as follows. An intermediate replacement for the negative log-likelihood in the $k$th stratum
\[
\ell^{(k)}(\beta) = \displaystyle - \frac{1}{n_k} \sum_{i=1}^{n_k} \left[ \beta^T X_{ki} - \log \left\{ \frac{1}{n_k} \sum_{j=1}^{n_k} 1(Y_{kj} \ge Y_{ki}) \exp(\beta^T X_{kj}) \right\} \right] \delta_{ki}
\]
can be defined as
\[
\widetilde{\ell}^{(k)}(\beta) = - \displaystyle \frac{1}{n_k} \sum_{i=1}^{n_k} \left\{ \beta^T X_{ki} - \log \mu_{0k}(Y_{ki}; \beta) \right\} \delta_{ki},
\]
which is a sum of $n_k$ independent and identically distributed terms.
The target parameter is $$\bar{\beta} = \arg\min_{\beta} \displaystyle {E} \left\{ \sum_{k=1}^K \frac{n_k}{N} \widetilde{\ell}^{(k)}(\beta) \right\}. $$ Then the excess risk for any given $\beta$ is
\[
\mathcal{E}(\beta) = \displaystyle {E} \left\{ \sum_{k=1}^K \frac{n_k}{N} \widetilde{\ell}^{(k)}(\beta) \right\} - \displaystyle {E} \left\{ \sum_{k=1}^K \frac{n_k}{N} \widetilde{\ell}^{(k)}(\bar\beta) \right\}.
\]
We refer remaining details to \citet{kong2014non}. \end{proof}
\begin{lemma} \label{chap4:lemma:const_sol}
Under Assumptions B.1--B.3 and B.5 and assume $\lambda \asymp \sqrt{\log(p)/n_{min}}$, it holds with probability going to one that $\| \Theta_{\beta^0} \widehat{\Sigma} - I_p \|_{\infty} \le \gamma$, with $\gamma \asymp \| \Theta_{\beta^0} \|_{1,1} \{ \max_{1\le k \le K} | n_k/N - r_k | + s_0 \lambda \}$. \end{lemma}
Lemma \ref{chap4:lemma:const_sol} shows that when $\gamma \asymp \| \Theta_{\beta^0} \|_{1,1} \{ \max_{1\le k \le K} | n_k/N - r_k | + s_0 \lambda \}$, the $j$th row of $\Theta_{\beta^0}$ ($j=1, \cdots, p$) will be feasible in the constraint of the corresponding quadratic programming problem with probability going to one.
\begin{proof}[\textbf{Proof of Lemma \ref{chap4:lemma:const_sol}}]
We first derive the rate for $\| \widehat{\Sigma} - \Sigma_{\beta^0} \|_{\infty}$. Note that
\begin{align*}
& \hspace{-0.4in} \| \widehat{\Sigma} - \Sigma_{\beta^0} \|_{\infty} \\
\le & \ \left\| \displaystyle \frac{1}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{\tau} \left[ \{ X_{ki} - \widehat{\eta}_k(t; \widehat{\beta}) \}^{\otimes 2} - \{ X_{ki} - \eta_{k0}(t; \beta^0) \}^{\otimes 2} \right] dN_{ki}(t) \right\|_{\infty} \\
& \quad + \left\| \displaystyle \frac{1}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{\tau} \{ X_{ki} - \eta_{k0}(t; \beta^0) \}^{\otimes 2} dN_{ki}(t) - \Sigma_{\beta^0} \right\|_{\infty} \\
\equiv & \ a_{N1} + a_{N2}.
\end{align*}
By the boundedness Assumption B.1, Lemma \ref{chap4:lemma:mom} and Lemma \ref{chap4:lemma:lasso},
\begin{align*}
a_{N1} \le & \left\| \displaystyle \frac{1}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{\tau} \{ X_{ki} - \widehat{\eta}_{k}(t; \widehat{\beta}) \} \{ \eta_{k0}(t; \beta^0) - \widehat{\eta}_{k}(t; \widehat{\beta}) \}^T dN_{ki}(t) \right\|_{\infty} \\
& + \left\| \displaystyle \frac{1}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{\tau} \{ \eta_{k0}(t; \beta^0) - \widehat{\eta}_{k}(t; \widehat{\beta}) \} \{ X_{ki} - \eta_{k0}(t; \beta^0) \}^T dN_{ki}(t) \right\|_{\infty} \\
\le & \displaystyle \frac{4M}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{\tau} \| \eta_{k0}(t; \beta^0) - \widehat{\eta}_{k}(t; \widehat{\beta}) \|_{\infty} dN_{ki}(t) \\
\le & \displaystyle \frac{4M}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{\tau} \| \eta_{k0}(t; \beta^0) - \widehat{\eta}_{k}(t; \beta^0) \|_{\infty} dN_{ki}(t) \\
& + \displaystyle \frac{4M}{N} \sum_{k=1}^K \sum_{i=1}^{n_k} \int_{0}^{\tau} \| \widehat{\eta}_{k}(t; \beta^0) - \widehat{\eta}_{k}(t; \widehat{\beta}) \|_{\infty} dN_{ki}(t) \\
\le & 4M \mathcal{O}_{P}(\sqrt{\log(p)/ n_{min}}) + 4M \mathcal{O}_{P}(s_0 \lambda) = \mathcal{O}_{P}(s_0 \lambda),
\end{align*}
where the last inequality is a result of Lemma \ref{chap4:lemma:mom} and the fact that $\sup_{t \in [0, \tau]} \| \widehat{\eta}_{k}(t; \beta^0) - \widehat{\eta}_{k}(t; \widehat{\beta}) \|_{\infty} = \mathcal{O}_{P}(\| \widehat{\beta} - \beta^0 \|_1) = \mathcal{O}_{P}(s_0\lambda)$; see the proof of Lemma A4 in \citet{xia2021cox}. Since $\Sigma_{\beta^0} = \sum_{k=1}^K r_k \Sigma_{\beta^0, k}$,
\begin{align*}
a_{N2} \le & \displaystyle \left\| \sum_{k=1}^K \frac{n_k}{N} \left[ \frac{1}{n_k} \sum_{i=1}^{n_k} \int_{0}^{\tau} \{ X_{ki} - \eta_{k0}(t; \beta^0) \}^{\otimes 2} dN_{ki}(t) - \Sigma_{\beta^0, k} \right] \right\|_{\infty} \\
& + \displaystyle \left\| \sum_{k=1}^K \left( \frac{n_k}{N} - r_k \right) \Sigma_{\beta^0,k} \right\|_{\infty} \\
\le & \displaystyle \sum_{k=1}^K \frac{n_k}{N} \left\| \frac{1}{n_k} \sum_{i=1}^{n_k} \int_{0}^{\tau} \{ X_{ki} - \eta_{k0}(t; \beta^0) \}^{\otimes 2} dN_{ki}(t) - \Sigma_{\beta^0, k} \right\|_{\infty} + \displaystyle \left\| \sum_{k=1}^K \left( \frac{n_k}{N} - r_k \right) \Sigma_{\beta^0,k} \right\|_{\infty}.
\end{align*}
The proof of Lemma A4 in \citet{xia2021cox} shows that, for $k=1, \cdots, K$,
\[
\displaystyle \left\| \frac{1}{n_k} \sum_{i=1}^{n_k} \int_{0}^{\tau} \{ X_{ki} - \eta_{k0}(t; \beta^0) \}^{\otimes 2} dN_{ki}(t) - \Sigma_{\beta^0, k} \right\|_{\infty} = \mathcal{O}_{P}(\sqrt{\log(p) / n_k})
\]
by Hoeffding's concentration inequality.
So $a_{N2} = \mathcal{O}_{P}(\sqrt{\log(p)/ n_{min}}) + \mathcal{O}(\max_{k} | n_k/N - r_k|)$. Then, combining the bounds on $a_{N1}$ and $a_{N2}$, $\| \widehat{\Sigma} - \Sigma_{\beta^0} \|_{\infty} = \mathcal{O}_{P}(s_0 \lambda + \max_k | n_k/N - r_k |)$.
Finally, it is easy to see that
\[
\| \Theta_{\beta^0} \widehat{\Sigma} - I_p \|_{\infty} = \| \Theta_{\beta^0} (\widehat{\Sigma} - \Sigma_{\beta^0} ) \|_{\infty} \le \| \Theta_{\beta^0} \|_{1,1} \| \widehat{\Sigma} - \Sigma_{\beta^0} \|_{\infty},
\]
and $\| \Theta_{\beta^0} \widehat{\Sigma} - I_p \|_{\infty} = \mathcal{O}_{P} ( \| \Theta_{\beta^0} \|_{1,1} \{ s_0 \lambda + \max_k | n_k/N - r_k | \})$. \end{proof}
\begin{lemma} \label{chap4:lemma:diff_theta}
Under the assumptions in Lemma \ref{chap4:lemma:const_sol}, if we further assume for some constant $\epsilon^{\prime} \in (0,1)$, $\limsup_{n_{min} \rightarrow \infty} p \gamma \le 1 - \epsilon^{\prime}$, then we have $\| \widehat{\Theta} - \Theta_{\beta^0} \|_{\infty} = \mathcal{O}_{P}(\gamma \| \Theta_{\beta^0} \|_{1,1})$. \end{lemma}
The proof of Lemma \ref{chap4:lemma:diff_theta} follows that of Lemma A5 in \citet{xia2021cox}, thus is omitted.
\begin{lemma} \label{chap4:lemma:max_score}
Under Assumptions B.1--B.3, for each $t > 0$,
\[
P (\| \dot{\ell}(\beta^0) \|_{\infty} > t) \le 2Kpe^{- C n_{min}t^2},
\]
where $C>0$ is an absolute constant. \end{lemma}
\begin{proof}[\textbf{Proof of Lemma \ref{chap4:lemma:max_score}}]
Since $\dot{\ell}(\beta^0) = \sum_{k=1}^K \frac{n_k}{N} \dot{\ell}^{(k)}(\beta^0)$, we have
\begin{align*}
P \left( \| \dot{\ell}(\beta^0) \|_{\infty} > t \right) & \le P \left( \sum_{k=1}^K \frac{n_k}{N} \| \dot{\ell}^{(k)}(\beta^0) \|_{\infty} > t \right) \\
& \le \sum_{k=1}^K P \left( \| \dot{\ell}^{(k)}(\beta^0) \|_{\infty} > t \right) \\
& \le \sum_{k=1}^K 2pe^{- C n_k t^2}.
\end{align*}
Note that $\| X_{ki} - \widehat{\eta}_k(t; \beta^0) \|_{\infty} \le 2M$ holds uniformly for all $k$ and $i$.
Then the last inequality is a direct result of Lemma 3.3(ii) in \citet{huang2013oracle} when applied to each of the $K$ strata. \end{proof}
\section{Proofs of Main Theorems}
\begin{proof}[\textbf{Proof of Theorem 3.1}]
Let $\dot{\ell}_j(\beta)$ be the $j$th element of the derivative $\dot{\ell}(\beta)$. By the mean value theorem, there exists $\widetilde{\beta}^{(j)}$ between $\widehat{\beta}$ and $\beta^0$ such that $\dot{\ell}_j(\widehat{\beta}) - \dot{\ell}_j(\beta^0) = \left. \frac{\partial \dot{\ell}_j(\beta)}{\partial \beta^T} \right|_{\beta=\widetilde{\beta}^{(j)}} (\widehat{\beta} - \beta^0)$. Denote the $p \times p$ matrix $D = \left( \left. \frac{\partial \dot{\ell}_1(\beta)}{\partial \beta} \right|_{\beta=\widetilde{\beta}^{(1)}}, \cdots, \left. \frac{\partial \dot{\ell}_p(\beta)}{\partial \beta} \right|_{\beta=\widetilde{\beta}^{(p)}} \right)^T$.
By the definition of the de-biased estimator $\widehat{b}$, we may decompose $c^T (\widehat{b} - \beta^0) $ as
\begin{align*}
c^T (\widehat{b} - \beta^0) & = - c^T \Theta_{\beta^0} \dot{\ell}(\beta^0) - c^T (\widehat{\Theta} - \Theta_{\beta^0}) \dot{\ell}(\beta^0) \nonumber \\
& \quad - c^T (\widehat{\Theta} \widehat{\Sigma} - I_p) (\widehat{\beta} - \beta^0) + c^T \widehat{\Theta} (\widehat{\Sigma} - D) (\widehat{\beta} - \beta^0) \\
& = - c^T \Theta_{\beta^0} \dot{\ell}(\beta^0) + (i) + (ii) + (iii),
\end{align*}
where $(i) = - c^T (\widehat{\Theta} - \Theta_{\beta^0}) \dot{\ell}(\beta^0), (ii) = - c^T (\widehat{\Theta} \widehat{\Sigma} - I_p) (\widehat{\beta} - \beta^0)$ and $(iii) = c^T \widehat{\Theta} (\widehat{\Sigma} - D) (\widehat{\beta} - \beta^0)$.
We first show $\sqrt{N}(i) = o_{P}(1)$ and $\sqrt{N}(ii) = o_{P}(1)$.
By Lemma \ref{chap4:lemma:diff_theta} and Lemma \ref{chap4:lemma:max_score},
\begin{align*}
| \sqrt{N}(i) | & \le \sqrt{N} \| c\|_1 \cdot \| \widehat{\Theta} - \Theta_{\beta^0} \|_{\infty, \infty} \cdot \| \dot{\ell}(\beta^0) \|_{\infty} \\
& \le \sqrt{N} a_{*} \mathcal{O}_{P}( p \gamma \| \Theta_{\beta^0} \|_{1,1}) \mathcal{O}_{P}(\sqrt{\log(p)/n_{min}}) \\
& = \mathcal{O}_{P}( \| \Theta_{\beta^0} \|_{1,1} p \gamma \sqrt{\log(p)} ) \\
& = o_{P}(1),
\end{align*}
where the last equality is a direct result of the assumption that $\| \Theta_{\beta^0} \|_{1,1}^2 \{ \max_k|n_k/N-r_k| + s_0\lambda \} p \sqrt{\log(p)} \to 0$ when $\lambda\asymp \sqrt{\log(p)/n_{min}}$.
By Lemma \ref{chap4:lemma:lasso},
\begin{align*}
| \sqrt{N}(ii) | & \le \sqrt{N} \| c\|_1 \| (\widehat{\Theta} \widehat{\Sigma} - I_p) (\widehat{\beta} - \beta^0) \|_{\infty} \\
& \le \sqrt{N} a_* \| \widehat{\Theta} \widehat{\Sigma} - I_p \|_{\infty} \| \widehat{\beta} - \beta^0 \|_1 \\
& \le \sqrt{N} a_* \gamma \| \widehat{\beta} - \beta^0 \|_1 \\
& = \mathcal{O}_{P}(\sqrt{N} \gamma s_0 \lambda) \\
& \le \mathcal{O}_{P}( \sqrt{N} \| \Theta_{\beta^0} \|_{1,1} \{ \max_k |n_k/N - r_k| + s_0 \lambda \} p \sqrt{\log(p)/n_{min}} ) \\
& = o_{P}(1).
\end{align*}
We then show that $\sqrt{N}(iii) = o_{P}(1)$.
Note that $\widehat{\Sigma} - D = (\widehat{\Sigma} - \Sigma_{\beta^0}) + (\Sigma_{\beta^0} - \ddot{\ell}(\beta^0)) + (\ddot{\ell}(\beta^0) - D)$. By the proof of Lemma \ref{chap4:lemma:const_sol}, we see that with $\lambda \asymp \sqrt{\log(p) / n_{min}}$, $\| \widehat{\Sigma} - \Sigma_{\beta^0} \|_{\infty} = \mathcal{O}_{P}(s_0 \lambda + \max_k | n_k/ N - r_k |)$.
Based on the proof of Theorem 1 in \citet{xia2021cox}, for each stratum, $\| \ddot{\ell}^{(k)}(\beta^0) - D^{(k)} \|_{\infty} = \mathcal{O}_{P}(\sqrt{\log(p)/n_k})$, where
$D^{(k)} = \left( \left. \frac{\partial \dot{\ell}^{(k)}_1(\beta)}{\partial \beta} \right|_{\beta=\widetilde{\beta}^{(1)}}, \cdots, \left. \frac{\partial \dot{\ell}^{(k)}_p(\beta)}{\partial \beta} \right|_{\beta=\widetilde{\beta}^{(p)}} \right)^T$. Since the overall negative log partial likelihood $\displaystyle \ell(\beta) = \sum_{k=1}^K \frac{n_k}{N} \ell^{(k)}(\beta)$, and $D= \displaystyle \sum_{k=1}^K \frac{n_k}{N} D^{(k)}$, then $\| \ddot{\ell}(\beta^0) - D \|_{\infty} = \mathcal{O}_{P}(\sqrt{\log(p)/n_{min}})$. Also, $\| \Sigma_{\beta^0, k} - \ddot{\ell}^{(k)}(\beta^0) \|_{\infty} = \mathcal{O}_{P}(\sqrt{\log(p)/n_k})$ by the proof of Theorem 1 in \citet{xia2021cox}. Then
\begin{align*}
\| \Sigma_{\beta^0} - \ddot{\ell}(\beta^0) \|_{\infty} & \le \left\| \sum_{k=1}^K r_k \Sigma_{\beta^0, k} - \sum_{k=1}^K \frac{n_k}{N} \Sigma_{\beta^0, k} \right\|_{\infty} + \left\| \sum_{k=1}^K \frac{n_k}{N} \Sigma_{\beta^0, k} - \sum_{k=1}^K \frac{n_k}{N} \ddot{\ell}^{(k)}(\beta^0) \right\|_{\infty} \\
& \le K \max_k(| n_k / N - r_k | \| \Sigma_{\beta^0, k} \|_{\infty}) + K \mathcal{O}_{P}(\sqrt{\log(p)/n_{min}}) \\
& = \mathcal{O}_{P}(\max_k |n_k/N - r_k| + \sqrt{\log(p)/n_{min}}).
\end{align*}
Therefore, for $\lambda \asymp \sqrt{\log(p)/n_{min}}$, $\| \widehat{\Sigma} - D \|_{\infty} = \mathcal{O}_{P}(s_0 \lambda + \max_k |n_k/N - r_k|)$, and
\begin{align*}
| \sqrt{N} (iii) | & \le \sqrt{N} \| c \|_1 \| \widehat{\Theta} \|_{\infty, \infty} \| \widehat{\Sigma} - D \|_{\infty} \| \widehat{\beta} - \beta^0 \|_1 \\
& \le \mathcal{O}_{P}\left(\sqrt{N} \| \Theta_{\beta^0} \|_{1,1} (s_0 \lambda + \max_k | n_k/N - r_k |) \right) s_0 \lambda \\
& \le \mathcal{O}_{P}\left( \sqrt{N /n_{min}} \| \Theta_{\beta^0} \|_{1,1} (s_0 \lambda + \max_k | n_k/N - r_k | ) p \sqrt{\log(p) } \right)\\
& = o_{P}(1).
\end{align*}
Finally, for the variance,
\begin{align*}
| c^T ( \widehat{\Theta} - \Theta_{\beta^0}) c | & \le \| c \|_1^2 \| \widehat{\Theta} - \Theta_{\beta^0} \|_{\infty} \\
& \le a_*^2 \mathcal{O}_{P}(\gamma \| \Theta_{\beta^0} \|_{1,1}) = o_{P}(1).
\end{align*}
By Slutsky's theorem and Lemma \ref{chap4:lemma:lead}, $\sqrt{n} c^T (\widehat{b} - \beta^0) / (c^T \widehat{\Theta} c)^{1/2} \overset{\mathcal{D}}{\rightarrow} N(0,1)$. \end{proof}
\begin{proof}[\textbf{Sketch proof of Theorem 3.4}] Theorem 3.4 can be easily proved using Cram\'{e}r-Wold device. For any $\omega \in \mathbb{R}^m$, since the dimension of $\omega$ is a fixed integer, we can invoke Theorem 3.1 by taking $c=J^T \omega$. Note that $\| J^T \omega \|_1 \le \| J^T \|_{1,1} \| \omega \|_1 = \| J \|_{\infty, \infty} \| \omega \|_1 = \mathcal{O}(1)$. \end{proof}
\end{document}
|
arXiv
|
{
"id": "2211.08868.tex",
"language_detection_score": 0.7257236242294312,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[$C^*$-algebras and Perron--Frobenius theory]{A note on lattice ordered $C^*$-algebras and Perron--Frobenius theory}
\author{Jochen Gl\"uck} \email{[email protected]} \address{Jochen Gl\"uck, Institute of Applied Analysis, Ulm University, 89069 Ulm, Germany} \keywords{Lattice ordered $C^*$-algebra; Perron--Frobenius theory; completely positive semigroup; commutativity from order} \subjclass[2010]{46L05; 46B40; 47B65}
\date{\today}
\begin{abstract}
A classical result of Sherman says that if the space of self-adjoint elements in a $C^*$-algebra $\mathcal{A}$ is a lattice with respect to its canonical order, then $\mathcal{A}$ is commutative. We give a new proof of this theorem which shows that it is intrinsically connected with the spectral theory of positive operator semigroups. Our methods also show that some important Perron--Frobenius like spectral results fail to hold in any non-commutative $C^*$-algebra. \end{abstract}
\maketitle
\section{Introduction}
Let us consider the space $\mathcal{A}_{\operatorname{sa}}$ of self-adjoint elements of a $C^*$-algebra $\mathcal{A}$. There is a canonical order on $\mathcal{A}_{\operatorname{sa}}$ which is given by $a \le b$ if and only if $b-a$ is positive semi-definite. With respect to this order $\mathcal{A}_{\operatorname{sa}}$ is an ordered real Banach space, meaning that the \emph{positive cone} $(\mathcal{A}_{\operatorname{sa}})_+ := \{a \in \mathcal{A}_{\operatorname{sa}}| \, a \ge 0\}$ is closed, convex and invariant with respect to multiplication by scalars $\alpha \ge 0$ and that it fulfils $(\mathcal{A}_{\operatorname{sa}})_+ \cap -(\mathcal{A}_{\operatorname{sa}})_+ = \{0\}$. If $\mathcal{A}$ is commutative, then it follows from Gelfand's representation theorem for commutative $C^*$-algebras that $\mathcal{A}_{\operatorname{sa}}$ is actually lattice ordered, i.e.\ that all elements $a,b \in \mathcal{A}_{\operatorname{sa}}$ have an \emph{infimum} (or \emph{greatest lower bound}) $a \land b$. In 1951 Sherman proved that the converse implication also holds, i.e.\ if $\mathcal{A}_{\operatorname{sa}}$ is lattice ordered, then $\mathcal{A}$ is commutative \cite[Theorems~1 and~2]{Sherman1951}.
Since then a great wealth of results has appeared which vary and generalise Sherman's theorem in several directions. We only quote a few of them: in \cite[Theorems~1 and~2]{Fukamiya1954} it is shown that if the ordered space $\mathcal{A}_{\operatorname{sa}}$ is only assumed to have the Riesz decomposition property, then $\mathcal{A}$ is commutative. In \cite[Theorem~2]{Archbold1974} it is proved that if for a fixed element $a \in \mathcal{A}_{\operatorname{sa}}$ and all $b \in \mathcal{A}_{\operatorname{sa}}$ the infimum $a \land b$ exists, then $\mathcal{A}$ is commutative. In \cite[Corollary~11 and Theorem~12]{Kadison1951}, \cite[Theorem~3.7]{Archbold1972}, \cite[Theorem~3.5]{Cho-Ho1973} and \cite[Corollary~2.4]{Green1977} it is shown that if $\mathcal{A}$ is in some sense too far from being commutative, then $\mathcal{A}_{\operatorname{sa}}$ is even a so-called \emph{anti-lattice}, meaning that two elements of $a,b\in \mathcal{A}_{\operatorname{sa}}$ have an infimum only if $a \le b$ or $b \le a$. In \cite{Curtis1958} Sherman's result is adapted to more general Banach algebras. The ``commutativity-from-order'' theme was also considered from a somewhat different viewpoint in \cite{Topping1965} and a more recent contribution to the theory can be found in \cite{Spain2001}.
It might not come as a surprise that the proofs for many of the above mentioned results use methods which are rather typical for the theory of operator algebras. The main goal of this note is to show that Sherman's theorem can also be proved by using only elementary properties of $C^*$-algebras if we instead employ the so-called \emph{Perron--Frobenius theory}; this notion is usually used to refer to the spectral theory of positive operators and semigroups, in particular on Banach lattices. We present our new proof of Sherman's result in Section~\ref{section:a-new-proof-of-shermans-theorem}.
Since our approach relies on non-trivial spectral theoretic results, we do not claim that our proof is simpler than the original one; our main point is that our proof employs completely different methods and thus sheds some new light on Sherman's theorem. In particular, the proof establishes a beautiful connection between commutativity of operator algebras and Perron--Frobenius theory. In Section~\ref{section:perron-frobenius-theory-on-c-star-algebras} we have a somewhat deeper look into this connection by proving that certain Perron--Frobenius type results can never hold on a non-commutative $C^*$-algebra.
\section{A new proof of Sherman's theorem} \label{section:a-new-proof-of-shermans-theorem}
In this section we give a new proof of Sherman's commutativity result by means of Perron--Frobenius spectral theory. For every Banach space $E$ we denote by $\mathcal{L}(E)$ the space of bounded linear operators on $E$.
\begin{theorem}[Sherman] \label{thm:sherman}
Let $\mathcal{A}$ be a $C^*$-algebra and let the space $\mathcal{A}_{\operatorname{sa}}$ of self-adjoint elements in $\mathcal{A}$ be endowed with the canonical order. Then $\mathcal{A}_{\operatorname{sa}}$ is lattice ordered if and only if $\mathcal{A}$ is commutative. \end{theorem} \begin{proof}
The implication ``$\Leftarrow$'' follows readily from Gelfand's representation theorem for commutative $C^*$-algebras. To prove the implication ``$\Rightarrow$'', assume that $\mathcal{A}_{\operatorname{sa}}$ is lattice ordered. Fix $a \in \mathcal{A}_{\operatorname{sa}}$. It suffices to show that $a$ commutes with every element $c \in \mathcal{A}_{\operatorname{sa}}$.
For every $t \in \mathbb{R}$ we define an operator $T_t \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ by $T_tc = e^{-ita}\,c\,e^{ita}$ (where the exponential function is computed in the unitization of $\mathcal{A}$ in case that $\mathcal{A}$ does not contain a unit). Obviously, the operator family $\mathcal{T} := (T_t)_{t \in \mathbb{R}} \subseteq \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ is a contractive positive $C_0$-group on $\mathcal{A}_{\operatorname{sa}}$. The orbit $t \mapsto T_tc$ is differentiable for every $c \in \mathcal{A}_{\operatorname{sa}}$ and its derivative at $t = 0$ is given by $-iac + ica$; hence, the $C_0$-group $\mathcal{T}$ has a bounded generator $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ which is given by $Lc = -iac + ica$ for all $c \in \mathcal{A}_{\operatorname{sa}}$.
Let $\sigma(L)$ denote the spectrum of (the complex extension of) $L$; then $\sigma(L)$ is contained in $i\mathbb{R}$ since the $C_0$-group $\mathcal{T}$ is bounded and $\sigma(L)$ is non-empty and bounded since $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$. Now we use that $\mathcal{A}_{\operatorname{sa}}$ is a lattice to show that $\sigma(L) = \{0\}$: since the norm is monotone on $\mathcal{A}_{\operatorname{sa}}$ (meaning that we have $\|a\| \le \|b\|$ whenever $0 \le a \le b$, see \cite[Theorem~2.2.5(3)]{Murphy1990}), we can find an equivalent norm on $\mathcal{A}_{\operatorname{sa}}$ which renders it a Banach lattice (see Section~\ref{section:perron-frobenius-theory-on-c-star-algebras} for a description of this norm). However, if an operator $L$ with spectral bound $0$ generates a bounded, eventually norm continuous and positive $C_0$-semigroup on a Banach lattice, then it follows from Perron--Frobenius theory that $0$ is a dominant spectral value of $L$, i.e.\ that we have $\sigma(L) \cap i\mathbb{R} = \{0\}$; see \cite[Corollary~C-III-2.13]{Arendt1986}. Hence, $\sigma(L) = \{0\}$.
Since $L$ generates a bounded $C_0$-group it now follows from the semigroup version of Gelfand's $T = \id$ theorem \cite[Corollary~4.4.12]{Arendt2011} that $\mathcal{T}$ is trivial, i.e.\ that $T_t$ is the identity operator on $\mathcal{A}_{\operatorname{sa}}$ for all $t \in \mathbb{R}$. Hence $L = 0$, so $0 = Lc = -iac + ica$ for all $c \in \mathcal{A}_{\operatorname{sa}}$. This shows that $a$ commutes with all elements of $\mathcal{A}_{\operatorname{sa}}$. \end{proof}
In the above proof we quoted a result of Perron--Frobenius theory for positive operator semigroups from \cite[Corollary~C-III-2.13]{Arendt1986}, and this result is in turn based on a rather deep theorem in \cite[Theorem~C-III-2.10]{Arendt1986}. However, we only needed a simple special case since the generator of our semigroup is bounded. Let us demonstrate how this special case can be treated without employing the entire machinery of Perron--Frobenius theory:
\begin{proposition} \label{prop:perron-frobenius-for-bounded-generator}
Let $E$ be a Banach lattice and let $L \in \mathcal{L}(E)$ be an operator with spectral bound $s(L) = 0$. If $e^{tL}$ is positive for every $t \ge 0$, then $\sigma(L) \cap i \mathbb{R} = \{0\}$. \end{proposition} \begin{proof}
Assume that $e^{tL} \ge 0$ for all $t \ge 0$. Since $E$ is a Banach lattice it follows that $L + \|L\| \ge 0$, see \cite[Theorem~C-II-1.11]{Arendt1986}. One can prove by a simple resolvent estimate that the spectral radius of a positive operator on a Banach lattice is contained in the spectrum, see e.g.~\cite[Proposition~V.4.1]{Schaefer1974}. Hence $r(L + \|L\|) \in \sigma(L + \|L\|)$, so we conclude that the spectral bound $\spb(L+\|L\|) = \|L\|$ coincides with the spectral radius $r(L + \|L\|)$. Thus, $\sigma(L + \|L\|) \cap (\|L\| + i\mathbb{R}) = \{\|L\|\}$, which proves the assertion. \end{proof}
\begin{remark} \label{rem:gelfands_t-id-theorem}
In our proof of Theorem~\ref{thm:sherman} we used another non-trivial result, namely the semigroup version of Gelfand's $T = \id$ theorem. Let us note that we can instead use Gelfand's $T = \id$ theorem for single operators: if we know that $\sigma(L) = \{0\}$ and that the group $(e^{tL})_{t \in \mathbb{R}}$ is bounded, then it follows from the spectral mapping theorem for the holomorphic functional calculus (or for eventually norm continuous semigroups) that $\sigma(e^{tL}) = \{1\}$ for every $t \in \mathbb{R}$; since every $e^{tL}$ is doubly power bounded, we conclude from Gelfand's $T = \id$ theorem for single operators (see e.g.~\cite[Theorem~B.17]{Engel2000}) that $e^{tL} = \id_{\mathcal{A}_{\operatorname{sa}}}$ for all $t \in \mathbb{R}$. \end{remark}
Despite what was said in Proposition~\ref{prop:perron-frobenius-for-bounded-generator} and Remark~\ref{rem:gelfands_t-id-theorem}, our proof of Sherman's theorem still relies on Gelfand's $T = \id$ theorem for single operators, which is a non-trivial result. Thus, our proof is not elementary; yet, all its non-elementary ingredients are essentially independent of $C^*$-algebra theory.
\section{Perron--Frobenius theory on $C^*$-algebras} \label{section:perron-frobenius-theory-on-c-star-algebras}
Our proof of Theorem~\ref{thm:sherman} suggests that certain Perron--Frobenius type spectral results can only be true on commutative $C^*$-algebras. Let us discuss this in a bit more detail: indeed, it was demonstrated in \cite[pp.\,387--388]{Arendt1986} and \cite[Section~4]{Luczak2010} by means of concrete examples that typical Perron--Frobenius results which are true on Banach lattices fail in general on non-commutative $C^*$-algebras. On the other hand, some results can even by shown in the non-commutative setting if one imposes additional assumptions (such as irreducibility and complete positivity) on the semigroup or the operator under consideration.
For single matrices and operators, such results can for instance be found in \cite{Evans1978, Groh1981, Groh1982, Groh1983}. For time-continuous operator semigroups we refer for example to \cite[Section~D-III]{Arendt1986}; the papers \cite{Albeverio1978, Luczak2010} contain results for both the single operator and the time-continuous case, and in \cite[Section~6.1]{Batkai2012} Perron--Frobenius type results on $W^*$-algebras are proved by a different approach, using a version of the so-called Jacobs--DeLeeuw--Glicksberg decomposition. An example for an application of Perron--Frobenius theory on $C^*$-algebras to the analysis of quantum systems can be found in \cite[Theorem~2.2]{Jaksic2014}.
In this section we are headed in a different direction: we use the idea of our proof of Theorem~\ref{thm:sherman} to show that certain Perron--Frobenius type results are \emph{never} true on non-commutative $C^*$-algebras.
To state our result we need a bit of notation. Some of the following notions have already been used tacitly above, but to avoid any ambiguity in the formulation of the next theorem, it is important to recall them explicitly here. If $E$ is a Banach space, then we denote by $\mathcal{L}(E)$ the space of all bounded linear operators on $E$; the dual space of $E$ is denoted by $E'$ and the adjoint of an operator $L \in \mathcal{L}(E)$ is denoted by $L' \in \mathcal{L}(E')$.
By an \emph{ordered Banach space} we mean a tuple $(E,E_+)$, often only denoted by $E$, where $E$ is a real Banach space, and $E_+ \subset E$ is a closed and pointed cone in $E$, meaning that $E_+$ is closed, that $\alpha E_+ + \beta E_+ \subseteq E_+$ for all $\alpha,\beta \in [0,\infty)$ and $E_+ \cap -E_+ = \{0\}$. On an ordered Banach space there is a canonical order relation $\le$ which is given by ``$x \le y$ iff $y - x \in E_+$''. The cone $E_+$ is called \emph{generating} if $E = E_+ - E_+$ and it is called \emph{normal} if there exists a constant $c > 0$ such that $\|x\| \le c\|y\|$ whenever $0 \le x \le y$.
If $\mathcal{A}$ is a $C^*$-algebra, then the space $\mathcal{A}_{\operatorname{sa}}$ of self-adjoint elements in $\mathcal{A}$ is usually endowed with the cone $(\mathcal{A}_{\operatorname{sa}})_+ := \{a \in \mathcal{A}_{\operatorname{sa}}: \; \sigma(a) \subseteq [0,\infty)\}$; thus one obtains the canonical order on $\mathcal{A}_{\operatorname{sa}}$ that we already considered in Theorem~\ref{thm:sherman}. Note that the cone $(\mathcal{A}_{\operatorname{sa}})_+$ is normal (since $\|a\| \le \|b\|$ whenever $0 \le a \le b$, see \cite[Theorem~2.2.5(3)]{Murphy1990}) and generating \cite[p.\,45]{Murphy1990}.
Let $E$ be an ordered Banach space. For all $x,y \in E$ the \emph{order interval} $[x,y]$ is given by $[x,y] := \{f \in E: \, x \le f \le y\}$. The space $E$ is said to have the \emph{Riesz decomposition property} if $[0,x] + [0,y] = [0,x+y]$ for every $x,y \in E_+$. If $E$ is lattice ordered, then the cone $E_+$ is generating and $E$ has the Riesz decomposition property \cite[Corollary~1.55]{Aliprantis2007}, but the converse is not in general true.
If $E$ is an ordered Banach space with generating and normal cone and if the induced order makes $E$ a vector lattice, then $\|x\|_{\operatorname{BL}} := \sup_{0 \le z \le |x|} \|z\|$ defines an equivalent norm on $E$ which renders it a Banach lattice (use the Riesz decomposition property of $E$ to see that $\|\cdot\|_{\operatorname{BL}}$ satisfies the triangle inequality and use e.g.\ \cite[Theorem~2.37(3)]{Aliprantis2007} to see that the norm $\|\cdot\|_{\operatorname{BL}}$ is indeed equivalent to the original norm).
Let $E$ be an ordered Banach space. An operator $L \in \mathcal{L}(E)$ is called \emph{positive} if $LE_+ \subseteq E_+$; this is denoted by $L \ge 0$. By $E'_+$ we denote the set of all positive functionals on $E$; here, a functional $x' \in E'$ is called \emph{positive} if $\langle x', x\rangle \ge 0$ for all $x \in E_+$. It follows from the Hahn--Banach separation theorem that a vector $x \in E$ is contained in $E_+$ if and only if $\langle x', x\rangle \ge 0$ for all $x' \in E'_+$; hence, an operator $L \in \mathcal{L}(E)$ is positive if and only if $L'E'_+ \subseteq E'_+$. If the positive cone in $E$ is generating, then $E'_+ \cap - E'_+ = \{0\}$ and thus, the dual space $E'$ is also an ordered Banach space.
Let $E$ be an ordered Banach space. We call an operator $L \in \mathcal{L}(E)$ \emph{quasi-positive} if $L + \alpha \id_E \ge 0$ for some $\alpha \ge 0$, we call it \emph{exponentially positive} if $e^{tL} \ge 0$ for all $t \in [0,\infty)$ and and we call it \emph{cross positive} if $\langle x', Lx \rangle \ge 0$ for all $x \in E_+$ and all $x' \in E'_+$ which fulfil $\langle x', x \rangle = 0$. We point out, however, that those properties are sometimes named differently in the literature.
If $A: E \supseteq D(A) \to E$ is a closed linear operator on a real Banach space $E$, then $\sigma(A)$ denotes the spectrum of the complex extension of $A$ to some complexification of $E$; similarly, all other notions from spectral theory are understood to be properties of the complex extension of $A$. The \emph{spectral bound} of $A$ is the number $s(A) := \sup\{\re \lambda: \, \lambda \in \sigma(A)\} \in [-\infty,\infty]$ and if $s(A) \in \mathbb{R}$, then the \emph{boundary spectrum} if $A$ is defined to be the set $\sigma_{{\operatorname{bnd}}}(A) := \{\lambda \in \sigma(A): \; \re \lambda = s(A)\}$. If $L \in \mathcal{L}(E)$, then the \emph{spectral radius} if $L$ is denoted by $r(L)$ and the \emph{peripheral spectrum} of $L$ is defined to be the set $\sigma_{\operatorname{per}}(L) := \{\lambda \in \sigma(L): \; |\lambda| = r(L)\}$.
Many results in Perron--Frobenius theory ensure the \emph{cyclicity} of parts of the spectrum; here, a set $S \subseteq \mathbb{C}$ is called \emph{cyclic} if $re^{i\theta} \in S$ ($r \ge 0$, $\theta \in \mathbb{R}$) implies that $re^{in\theta} \in S$ for all integers $n \in \mathbb{Z}$. The set $S$ is called \emph{additively cyclic} if $\alpha + i\beta \in S$ ($\alpha, \beta \in \mathbb{R}$) implies that $\alpha + in\beta \in S$ for all integers $n \in \mathbb{Z}$.
An important result in Perron--Frobenius theory states that on any given Banach lattice $E$, every positive power bounded operator with spectral radius $1$ has cyclic peripheral spectrum; in fact, an even somewhat stronger result is true, see \cite[Theorem~V.4.9]{Schaefer1974}. However, it is still an open problem whether \emph{every} positive operator on a Banach lattice has cyclic peripheral spectrum; see \cite{Glueck2015, Glueck2016} for a detailed discussion of this and for some recent progress in this question. An analogues result for $C_0$-semigroups says that if the generator of a positive and bounded $C_0$-semigroup has spectral bound $0$, then its boundary spectrum is additively cyclic; see \cite[Theorem~C-III-2.10]{Arendt1986} for a slightly stronger result.
Theorem~\ref{thm:perron-frobenius-on-c-star-algebras} below shows that the above mentioned results are \emph{never} true an a non-commutative $C^*$-algebra. In fact, they are not even true for \emph{completely positive} operators and semigroups on such spaces; recall that a linear operator $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$, whose complex extension to $\mathcal{A}$ is denoted by $L_\mathbb{C}$, is called \emph{completely positive} if the operator $L_\mathbb{C} \otimes \id_{\mathbb{C}^{k \times k}} \in \mathcal{L}(\mathcal{A} \otimes \mathbb{C}^{k \times k})$ (more precisely: its restriction to the self-adjoint part of the $C^*$-algebra $\mathcal{A} \otimes \mathbb{C}^{k \times k}$) is positive for every $k \in \mathbb{N}_0$; see \cite[Section~II.6.9]{Blackadar2006} for some details. We call a $C_0$-semigroup $(e^{tA})_{t \ge 0}$ on $\mathcal{A}_{\operatorname{sa}}$ \emph{completely positive} if the operator $e^{tA}$ is completely positive for every $t \ge 0$. Completely positive operators play an important role in quantum physics. Here we are going to show that the above mentioned Perron--Frobenius type results are not even true for completely positive operators and semigroups if the $C^*$-algebra $\mathcal{A}$ is non-commutative. To show this we only need the simple fact that for every $a \in \mathcal{A}$ the operator $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ given by $Lc = a^*ca$ is completely positive.
\begin{theorem} \label{thm:perron-frobenius-on-c-star-algebras}
Let $\mathcal{A}$ be a $C^*$-algebra and let the space $\mathcal{A}_{\operatorname{sa}}$ of self-adjoint elements in $\mathcal{A}$ be endowed with the canonical order. The following assertions are equivalent:
\begin{enumerate}[\upshape (i)]
\item $\mathcal{A}$ is commutative.
\item $\mathcal{A}_{\operatorname{sa}}$ is lattice ordered.
\item $\mathcal{A}_{\operatorname{sa}}$ has the Riesz decomposition property.
\item Every cross positive operator in $\mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ is quasi-positive.
\item Every exponentially positive operator in $\mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ is quasi-positive.
\item Every power bounded positive operator in $\mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ with spectral radius $1$ has cyclic peripheral spectrum.
\item Every power bounded completely positive operator in $\mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ with spectral radius $1$ has cyclic peripheral spectrum.
\item If the generator $A$ of a bounded positive $C_0$-semigroup on $\mathcal{A}_{\operatorname{sa}}$ has spectral bound $0$, then its boundary spectrum is additively cyclic.
\item If the generator $A$ of a bounded completely positive $C_0$-semigroup on $\mathcal{A}_{\operatorname{sa}}$ has spectral bound $0$, then its boundary spectrum is additively cyclic.
\item Every norm continuous, bounded and completely positive $C_0$-group $(e^{tL})_{t \in \mathbb{R}}$ on $\mathcal{A}_{\operatorname{sa}}$ is trivial, i.e.\ it fulfils $e^{tL} = \id_{\mathcal{A}_{\operatorname{sa}}}$ for all $t \in \mathbb{R}$.
\end{enumerate} \end{theorem} \begin{proof}
We show ``(i) $\Rightarrow$ (ii) $\Rightarrow$ (iv) $\Rightarrow$ (v) $\Rightarrow$ (x) $\Rightarrow$ (i)'' as well as ``(ii) $\Rightarrow$ (iii) $\Rightarrow$ (v)'', ``(ii) $\Rightarrow$ (vi) $\Rightarrow$ (vii) $\Rightarrow$ (x)'' and ``(ii) $\Rightarrow$ (viii) $\Rightarrow$ (ix) $\Rightarrow$ (x)''.
``(i) $\Rightarrow$ (ii)'' This follows from Gelfand's representation theorem for commutative $C^*$-algebras.
``(ii) $\Rightarrow$ (iv)''. Since $\mathcal{A}_{\operatorname{sa}}$ is lattice ordered and since the positive cone in $\mathcal{A}_{\operatorname{sa}}$ is normal, there is an equivalent norm on $\mathcal{A}_{\operatorname{sa}}$ which renders it a Banach lattice. Hence, it follows from \cite[Theorem~C-II-1.11]{Arendt1986} that every cross positive operator on $\mathcal{A}_{\operatorname{sa}}$ is quasi-positive.
``(iv) $\Rightarrow$ (v)'' Assume that (iv) is true and let $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ be exponentially positive. Due to (iv) it suffices to show that $L$ is cross positive, so let $0 \le a \in \mathcal{A}_{\operatorname{sa}}$ and $0 \le \varphi \in \mathcal{A}_{\operatorname{sa}}'$ such that $\langle \varphi, a \rangle = 0$. Then we obtain
\begin{align*}
\langle \varphi, La \rangle = \lim_{t \downarrow 0} \frac{\langle \varphi, (e^{tL} - \id_{\mathcal{A}_{\operatorname{sa}}})a \rangle}{t} = \lim_{t \downarrow 0} \frac{\langle \varphi, e^{tL}a \rangle}{t} \ge 0.
\end{align*}
Hence, $L$ is cross positive and it now follows from (iv) that $L$ is quasi-positive.
``(v) $\Rightarrow$ (x)'' Suppose that (v) is true, let $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ and assume that the group $(e^{tL})_{t \in \mathbb{R}}$ is bounded and completely positive. We argue as in Remark~\ref{prop:perron-frobenius-for-bounded-generator}: we have $\emptyset \not= \sigma(L) \subseteq i\mathbb{R}$ and we know from (v) that $L + \alpha \id_{\mathcal{A}_{\operatorname{sa}}}$ is positive for some $\alpha \ge 0$. Since the positive cone in $\mathcal{A}_{\operatorname{sa}}$ is generating and normal, it follows that the spectral radius of the operator $L + \alpha \id_{\mathcal{A}_{\operatorname{sa}}}$ is contained in its spectrum \cite[Section~2.2 in the Appendix]{Schaefer1999}; hence, the spectral radius equals the spectral bound $s(L + \alpha \id_{\mathcal{A}_{\operatorname{sa}}})$, which is in turn equal to $\alpha$. Thus, $\sigma(L + \alpha \id_{\mathcal{A}_{\operatorname{sa}}}) = \sigma(L + \alpha \id_{\mathcal{A}_{\operatorname{sa}}}) \cap (i\mathbb{R} + \alpha) = \{\alpha\}$, so we conclude that $\sigma(L) = \{0\}$. It now follows from the semigroup analogue of Gelfand's $T = \id$ theorem \cite[Corollary~4.4.12]{Arendt2011} that $e^{tL} = \id_{\mathcal{A}_{\operatorname{sa}}}$ for all $t \in \mathbb{R}$.
``(x) $\Rightarrow$ (i)'' Let us argue as in the proof of Theorem~\ref{thm:sherman}: fix $a \in \mathcal{A}_{\operatorname{sa}}$ and define $T_t \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ by $T_tc = e^{-ita}ce^{ita}$ for every $t \in \mathbb{R}$ (where the exponential function is computed in the unitization of $\mathcal{A}$ if $\mathcal{A}$ does not contain a unit itself). Then $(T_t)_{t \in \mathbb{R}}$ is a bounded, completely positive and operator norm continuous group on $\mathcal{A}_{\operatorname{sa}}$; its generator $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ is given by $Lc = -iac + ica$ for all $c \in \mathcal{A}_{\operatorname{sa}}$. It follows from (x) that $L = 0$ and hence, $a$ commutes with every $c \in \mathcal{A}_{\operatorname{sa}}$, which in turn proves that $\mathcal{A}$ is commutative.
``(ii) $\Rightarrow$ (iii)'' Every vector lattice has the Riesz decomposition property, see e.g.\ \cite[Corollary~1.55]{Aliprantis2007}.
``(iii) $\Rightarrow$ (v)'' Suppose that $\mathcal{A}_{\operatorname{sa}}$ has the Riesz decomposition property. Since the cone in $\mathcal{A}_{\operatorname{sa}}$ is normal, it follows that the dual space $\mathcal{A}_{\operatorname{sa}}'$ is a vector lattice \cite[Theorem~2.47]{Aliprantis2007}. Moreover, the cone in $\mathcal{A}_{\operatorname{sa}}'$ is normal, too \cite[Theorem~2.42]{Aliprantis2007}, so $\mathcal{A}_{\operatorname{sa}}'$ is a Banach lattice with respect to an equivalent norm. If $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ is exponentially positive, then so is its adjoint $L' \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}}')$ and hence, $L'$ is quasi-positive according to \cite[Theorem~C-II-1.11]{Arendt1986}. This in turn implies that $L$ is quasi-positive itself.
``(ii) $\Rightarrow$ (vi)'' If $\mathcal{A}_{\operatorname{sa}}$ is lattice ordered, then it is a Banach lattice with respect to some equivalent norm (since the cone in $\mathcal{A}_{\operatorname{sa}}$ is normal). Hence, assertion (vi) follows from Perron--Frobenius theory of single operators on Banach lattices, see \cite[Theorem~V.4.9]{Schaefer1974}.
``(vi) $\Rightarrow$ (vii)'' This is obvious.
``(vii) $\Rightarrow$ (x)'' Suppose that (vii) is true and that $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ generates a completely positive and bounded group $(e^{tL})_{t \in \mathbb{R}}$. Then the spectrum of $L$ is bounded and contained in the imaginary axis. Applying the spectral mapping theorem for small $t>0$ we conclude that the peripheral spectrum of $e^{tL}$ can only be cyclic if it is contained in $\{1\}$. Hence, $\sigma(L) = \{0\}$, so assertion (x) follows from the semigroup analogue of Gelfand's $T = \id$ theorem \cite[Corollary~4.4.12]{Arendt2011}.
``(ii) $\Rightarrow$ (viii)'' Assume that $\mathcal{A}_{\operatorname{sa}}$ is lattice ordered. We use again that $\mathcal{A}_{\operatorname{sa}}$ is then a Banach lattice for some equivalent norm. Hence, assertion (viii) follows from Perron--Frobenius theory for positive $C_0$-semigroups on Banach lattices, see \cite[Theorem~C-III-2.10]{Arendt1986}.
``(viii) $\Rightarrow$ (ix)'' This is obvious.
``(ix) $\Rightarrow$ (x)'' If (ix) holds and if $L \in \mathcal{L}(\mathcal{A}_{\operatorname{sa}})$ generates a completely positive and bounded group $(e^{tL})_{t \in \mathbb{R}}$, then it follows from (ix) that $\sigma(L) = \{0\}$. Hence we can again employ the semigroup analogue of Gelfand's $T = \id$ theorem \cite[Corollary~4.4.12]{Arendt2011} to conclude that (x) holds. \end{proof}
The above proof shows that many (not all) of the implications in Theorem~\ref{thm:perron-frobenius-on-c-star-algebras} are also true on general ordered Banach spaces with generating and normal cone. However, since we are mainly interested in $C^*$-algebras in this note, we omit a detailed discussion of this.
\end{document}
|
arXiv
|
{
"id": "1605.05190.tex",
"language_detection_score": 0.7714067697525024,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Congruences between modular forms and related modules } \author{Miriam Ciavarella} \date{ } \maketitle
\begin{abstract}
We fix $\ell$ a prime and let $M$ be an integer such that $\ell\not|M$; let $f\in S_2(\Gamma_1(M\ell^2))$ be a newform supercuspidal of fixed type related to the nebentypus, at $\ell$ and special at a finite set of primes. Let ${\bf T}^\psi$ be the local quaternionic Hecke algebra associated to $f$. The algebra ${\bf T}^\psi$ acts on a module $\mathcal M^\psi_f$ coming from the cohomology of a Shimura curve. Applying the Taylor-Wiles criterion and a recent Savitt's theorem, ${\bf T}^\psi$ is the universal deformation ring of a global Galois deformation problem associated to $\overline\rho_f$. Moreover $\mathcal M^\psi_f$ is free of rank 2 over ${\bf T}^\psi$. If $f$ occurs at minimal level, by a generalization of a Conrad, Diamond and Taylor's result and by the classical Ihara's lemma, we prove a theorem of raising the level and a result about congruence ideals. The extension of this results to the non minimal case is an open problem.
\end{abstract} Keywords: modular form, deformation ring, Hecke algebra, quaternion algebra, congruences. \\ 2000 AMS Mathematics Subject Classification: 11F80
\section*{Introduction}
The principal aim of this article is to detect some results about isomorphism of complete intersection between an universal deformation ring and a local Hecke algebra and about cohomological modules free over an Hecke algebra. From this results, it is possible deduce that there is an isomorphism between the quaternionic cohomological congruence module for a modular form and the classical congruence module.\\ Our work take place in a context of search which has its origin in the works of Wiles and Taylor-Wiles on the Shimura-Taniyama-Weil conjecture. Recall that the problem addressed by Wiles in \cite{wi} is to prove that a certain ring homomorphism $\phi:\mathcal R_\mathcal D\to{\bf T}_\mathcal D$ is an isomorphism (of complete intersection), where $\mathcal R_\mathcal D$ is the universal deformation ring for a mod $\ell$ Galois representation arising from a modular form and ${\bf T}_\mathcal D$ is a certain Hecke algebra. \\ Extending our results to a more general class of representations letting free the ramification to a finite set of prime we need a small generalization of a result of Conrad Diamond and Taylor. We will describe it, using a recent Savitt's theorem and we deduce from it two interesting results about congruences.\\ Our first result extends a work of Terracini \cite{Lea} to a more general class of types and allows us to work with modular forms having a non trivial nebentypus. Our arguments are largely identical to Terracini's in many places; the debt to Terracini's work will be clear throughout the paper. One important change is that since we will work with Galois representations which are not semistable at $\ell$ but only potentially semistable, we use a recent Savitt's theorem \cite{savitt}, that prove a conjecture of Conrad, Diamond and Taylor (\cite{CDT}, conjecture 1.2.2 and conjecture 1.2.3), on the size of certain deformation rings parametrizing potentially Barsotti-Tate Galois representations, extending results of Breuil and M\'{e}zard (conjecture 2.3.1.1 of \cite{BM}) (classifying Galois lattices in semistable representations in terms of \lq\lq strongly divisible modules \rq\rq) to the potentially crystalline case in Hodge-Tate weights $(0,1)$.
Given a prime $\ell$, we fix a newform $f\in S_2(\Gamma_0(N\Delta'\ell^2),\psi)$ with nebentypus $\psi$ of order prime to $\ell$, special at primes dividing $\Delta'$ and such that its local representation $\pi_{f,\ell}$ of $GL_2({\bf Q}_\ell)$, is associated to a fixed regular character $\chi$ of ${\bf F}_{\ell^2}^\times$ satisfying $\chi|_{{\bf Z}_\ell^\times}=\psi_\ell|_{{\bf Z}_\ell^\times}$. We consider the residual Galois representation $\overline\rho$ associated to $f$. We denote by ${\bf T}^\psi$ the ${\bf Z}_\ell$-subalgebra of $\prod_{h\in\mathcal B}\mathcal O_{h,\lambda}$ where $\mathcal B$ is the set of normalized newforms in $S_2(\Gamma_0(N\Delta'\ell^2),\psi)$ which are supercuspidal of type $\tau=\chi^\sigma\oplus\chi$ at $\ell$ and whose associated representation is a deformation of $\overline\rho$. By the Jacquet-Langlands correspondence and the Matsushima-Murakami-Shimura isomorphism, one can see such forms in a local component ${\mathcal M^\psi}$ of the $\ell$-adic cohomology of a Shimura curve. By imposing suitable conditions on the type $\tau$, we describe for each prime $p$ dividing the level, a local deformation condition of $\overline\rho_p$ and applying the Taylor-Wiles criterion in the version of Diamond \cite{D} and Fujiwara, we prove that the algebra ${\bf T}^\psi$ is characterized as the universal deformation ring ${\mathcal R^\psi}$ of our global Galois deformation problem. We point out that in order to prove the existence of a family of sets realizing simultaneously the conditions of a Taylor-Wiles system, we make large use of Savitt's theorem \cite{savitt}: assuming the existence of a newform $f$ as above, the tangent space of the deformation functor has dimension one over the residue field. Our first resul is the following: \begin{theorem} \begin{itemize} \item[a)] $\Phi:{\mathcal R^\psi}\to{\bf T}^\psi$ is an isomorphism of complete intersection; \item[b)] ${\mathcal M^\psi}$ is a free ${\bf T}^\psi$-module of rank 2. \end{itemize} \end{theorem}
\noindent We observe that in \cite{CDT} the authors assume that the type $\tau$ is strongly acceptable for $\overline\rho_\ell$. In this way they assure the existence of a modular form under their hyphotesis. Since we are interested to study the quaternionic cohomolgical module associated to the modular form $f$, as a general hyphotesis we suppose that there exist a modular form $f$ in our conditions, in other words we are assuming that uor cohomological module is not empty. \noindent Keeping the definitions as in \cite{CDT}, Savitt's result allows to suppress the assumption of acceptability in the definition of strongly acceptability, thus it is possible to extend Conrad, Diamond and Taylor's result \cite{CDT} relaxing the hypotheses on the residual representation.\\ \noindent Under the hypothesis that $f$ occurs with minimal level (i.e. the ramification at primes $p$ dividing the Artin conductor of the Galois representation $\rho_f$ is equal to the ramification of $\overline\rho_f$ at $p$) the module ${\mathcal M^\psi}$, used to construct the Taylor-Wiles system, can be also seen as a part of a module $\mathcal M^{\rm mod}$ coming from the cohomology of a modular curve, as described in \cite{CDT} \S5.3. Applying the extended Conrad, Diamond and Taylor's methods and by the Ihara's lemma for the cohomology of modular curves \cite{CDT}, the first part of our result can be extended by allowing the ramification on a set of primes $S$ disjoint from $N\Delta'\ell$. In this way it is possible to obtain results of the form: \begin{itemize} \item $\Phi_S:\mathcal R^\psi_S\to{\bf T}^\psi_S$ is an isomorphism of complete intersections, \end{itemize} where $\mathcal R^\psi_S$ is an universal deformation ring letting free the ramification at primes in $S$, ${\bf T}^\psi_S$ is a local Hecke algebra. We observe that, since there is not an analogous of the Ihara's lemma for the cohomology of the Shimura curve, we don't have any information about the correspondent module $\mathcal M^\psi_S$ coming from the cohomology of a Shimura curve. In particular, in the general case $S\not=\emptyset$, it is not possible to show that $\mathcal M^\psi_S$ is free over ${\bf T}^\psi_S$
We observe that as a consequence of the generalization of the Cornrad, Diamond and Taylor's result, two results about raising the level of modular forms and about the ideals of congruence follows. Let $S_1,S_2$ be two subsets of $\Delta_2$, we slightly modify the deformation problem, by imposing the condition sp at primes $p$ in $S_2$ and by allowing ramification at primes in $S_1$. Let we denote by $\eta_{S_1,S_2}$ the congruence ideal of a modular form relatively to the set $\mathcal B_{S_1,S_2}$ of the newforms of weight 2, Nebentypus $\psi$, level dividing $N\Delta_1\Delta_2\ell$ which are supercuspidal of type $\tau$ at $\ell$, special at primes in $\Delta_1S_2$. We prove that there is an isomorphism of complete intersections between the universal deformation ring $\mathcal R^\psi_{S_1,S_2}$ and the Hecke algebra ${\bf T}^\psi_{S_1,S_2}$ acting on the space $\mathcal B_{S_1,S_2}$ and $$\eta_{\Delta_2,\emptyset}=C\eta_{S_1,S_2}$$ where $C$ is a constant depending of the modular form. In particular we prove the following result: \begin{theorem}
Let $f=\sum a_nq^n$ be a normalized newform in $S_2(\Gamma_0(M\ell^2),\psi)$ supercuspidal of type $\tau=\chi\oplus\chi^\sigma$ at $\ell$, special at primes in a finite set $\Delta'$, there exist $g\in S_2(\Gamma_0(qM\ell^2),\psi)$ supercuspidal of type $\tau$ at $\ell$, special at every prime $p|\Delta'$ such that $f\equiv g\ {\rm mod}\ \lambda$ if and only if $$a_q^2\equiv\psi(q)(1+q)^2\ {\rm mod}\ \lambda$$ where $q$ is a prime such that $(q,M\ell^2)=1,$ $q\not\equiv-1\ {\rm mod}\ \ell$. \end{theorem}
We observe that our results concerning the cohomological modules, holds only at the minimal level since a quaternionic analogue of the Ihara's lemma is not available in this case. Let $S$ be a finite set of primes not dividing $M\ell$; we fix $f\in S_2(\Gamma_0(N\Delta'\ell^2 S),\psi)$ supercuspidal of type $\tau$ at $\ell$, special at primes $p|\Delta'$. If we modify our Galois deformation problem allowing ramification at primes in $S$, we obtain a new universal deformation ring $\mathcal R_S$ and a new Hecke algebra ${\bf T}^\psi_S$ acting on the newforms giving rise to such representation. We make the following conjecture: \begin{conjecture}\label{conj1} \begin{itemize} \item $\mathcal R_S\to{\bf T}^\psi_S$ is an isomorphism of complete intersection;
\item let $\mathcal M^\psi_S$ be the module $H^1({\bf X}_1(NS),\mathcal O)_{\mathfrak m_S}^{\widehat\psi}$ coming from the cohomology of the Shimura curve ${\bf X}_1(NS)$ associated to the open compact subgroup of $B_{\bf A}^{\times,\infty}$, $V_1(NS)=\prod_{p\not|NS\ell}R_p^\times\prod_{p|NS}K_p^1(N)\times(1+u_\ell R_\ell)$ where $K_p^1(N)$ is defined in section \ref{shi}, and $u_\ell$ is a uniformizer of $B_\ell^\times$. $\mathcal M^\psi_S$ is a free ${\bf T}^\psi_S$-module of rank 2. \end{itemize} \end{conjecture} \noindent Conjecture \ref{conj1} easily follows from the following conjecture:
\begin{conjecture}\label{noscon}
Let $q$ be a prime number such that $q\not|N\Delta'\ell^2$. We fix a maximal non Eisenstein ideal of the Hecke algebra ${\bf T}_0^{\widehat\psi}(N)$ acting on the group $H^1({\bf X}_1(N),\mathcal O)^{\widehat\psi}$. Let ${\bf X}_1(N)$ be the Shimura curve $${\bf X}_1(N)=B^\times\setminus B_{\bf A}^\times/K_\infty^+V_1(N)$$ where $$V_1(N)=\prod_{p\not|N\ell}R_p^\times\prod_{p|N}K_p^1(N)\times(1+u_\ell R_\ell)$$ where $K_p^1(N)$ is defined in section \ref{shi}, and $u_\ell$ is a uniformizer of $B_\ell^\times$. The map $$\alpha_\mathfrak m:H^1({\bf X}_1(N),\mathcal O)_\mathfrak m^{\widehat\psi}\times H^1({\bf X}_1(N),\mathcal O)_\mathfrak m^{\widehat\psi}\to H^1({\bf X}_1(Nq),\mathcal O)_{\mathfrak m^q}^{\widehat\psi}$$ is such that $\alpha\otimes_\mathcal O k$ is injective, where $\mathfrak m^q$ is the inverse image of the ideal $\mathfrak m$ under the natural map ${\bf T}_0^{\widehat\psi}(Nq)\to{\bf T}_0^{\widehat\psi}(N)$ and $k=\mathcal O/\lambda$.
\end{conjecture}
\noindent This conjecture would provide an analogue for the Shimura curves of the Ihara's Lemma in case $\ell|\Delta$. In \cite{DT} and in \cite{DTi}, Diamond and Taylor show that if $\ell$ not divides the discriminant of the indefinite quaternion algebra, then the analogue of conjecture \ref{noscon} holds.
\section{Notations}
\noindent For a rational prime $p$, ${\bf Z}_p$ and ${\bf Q}_p$ denote the ring of $p$-adic integers and the field of $p$-adic numbers, respectively. If $A$ is a ring, then $A^\times$ denotes the group of invertible elements of $A$. We will denote by ${\bf A}$ the ring of rational ad\'eles, and by ${\bf A}^\infty$ the finite ad\'eles.\\ Let $B$ be a quaternion algebra on ${\bf Q}$, we will denote by $B_{\bf A}$ the adelization of $B$, by $B_{\bf A}^\times$ the topological group of invertible elements in $B_{\bf A}$ and $B_{\bf A}^{\times,\infty}$ the subgroup of finite ad\'eles.\\ Let $R$ be a maximal order in $B$.For a rational place $v$ of ${\bf Q}$ we put $B_v=B\otimes_{\bf Q}{\bf Q}_v$; if $p$ is a finite place we put $R_p=R\otimes_{\bf Z}{\bf Z}_p$. \\
If $p$ is a prime not dividing the discriminant of $B$, included $p=\infty$, we fix an isomorphism $i_p:B_p\to M_2({\bf Q}_p)$ such that if $p\not=\infty$ we have $i_p(R_p)=M_2({\bf Z}_p)$.\\
We write $GL_2^+({\bf R})=\{g\in GL_2({\bf R})|\ det\ g>0\}$ and $K_\infty={\bf R}^\times O_2({\bf R}),$ $K_\infty^+={\bf R}^\times SO_2({\bf R}).$ If $K$ is a field, let $\overline K$ denote an algebraic closure of $K$; we put $G_K={\rm Gal}(\overline K/K)$. For a local field $K$, $K^{unr}$ denotes the maximal unramified extension of $K$ in $\overline K$; we put $I_K={\rm Gal}(\overline K/K^{unr})$, the inertia subgroup of $G_K$. For a prime $p$ we put $G_p=G_{{\bf Q}_p}$, $I_p=I_{{\bf Q}_p}$. If $\rho$ is a representation of $G_{\bf Q}$, we write $\rho_p$ for the restriction of $\rho$ to a decomposition group at $p$.
\section{The local Hecke algebra ${\bf T}^\psi$}\label{de}
We fix a prime $\ell>2$. Let ${\bf Z}_{\ell^2}$ denote the integer ring of ${\bf Q}_{\ell^2}$, the unramified quadratic extension of ${\bf Q}_\ell$. Let $M\not=1$ be a square-free integer not divisible by $\ell$. We fix $f$ an eigenform in $S_2(\Gamma_1(M\ell^2))$, then $f\in S_2(\Gamma_0(M\ell^2),\psi)$ for some Dirichlet character $\psi:({\bf Z}/M\ell^2{\bf Z})^\times\to\overline{\bf Q}^\times.$\\ For abuse of notation, let $\psi$ be the adelisation of the Dirichlet character $\psi$ and we denote by $\psi_p$ the composition of $\psi$ with the inclusion ${\bf Q}_p^\times\to {\bf A}^\times$.\\
We fix a regular character $\chi:{\bf Z}_{\ell^2}^\times\to \overline{\bf Q}^\times$ of conductor $\ell$ such that $\chi|_{{\bf Z}_\ell^\times}=\psi_\ell|_{{\bf Z}_\ell^\times}$ and we extend $\chi$ to ${\bf Q}_{\ell^2}^\times$ by putting $\chi(\ell)=-\psi_\ell(\ell)$. We observe that $\chi$ is not uniquely determined by $\psi$ and, if we fix an embedding of $\overline{\bf Q}$ in $\overline{\bf Q}_\ell$, we can ragard the values of $\chi$ in this field.\\ Since, by local classfield theory, $\chi$ can be regarded as a character of $I_\ell$ and we can consider the type $\tau=\chi\oplus\chi^\sigma:I_\ell\to GL_2(\overline{\bf Q}_\ell)$, where $\sigma$ denotes the complex conjugation.\\
We fix a decomposition $M=N\Delta'$ where $\Delta'$ is a product of an odd number of primes. If we shoose $f\in S_2(\Gamma_1(M\ell^2))$ such that the automorphic representation $\pi_f=\otimes_v\pi_{f,v}$ of $GL_2({\bf A})$ associated to $f$ is supercuspidal of type $\tau=\chi\oplus\chi^\sigma$ at $\ell$ and special at every primes $p|\Delta'$, then $\pi_{f,\ell}=\pi_\ell(\chi)$, where $\pi_\ell(\chi)$ is the representation of $GL_2({\bf Q}_\ell)$ associated to $\chi$, with central character $\psi_\ell$ and conductor $\ell^2$ (see \cite{Ge}, \S 2.8). Moreover, under our hypotheses, the nebentypus $\psi$ factors through $({\bf Z}/N\ell{\bf Z})^\times$. As a general hypothesis, we assume that $\psi$ has order prime to $\ell$.\\ Let $WD(\pi_\ell(\chi))$ be the $2$-dimensional representation of the Weil-Deligne group at $\ell$ associated to $\pi_\ell(\chi)$ by local Langlands correspondence. Since by local classfield theory, we can identify ${\bf Q}_{\ell^2}^\times$ with $W_{{\bf Q}_{\ell^2}}^{ab}$, we can see $\chi$ as a character of $W_{{\bf Q}_{\ell^2}}$; by (\cite{Ca} \S 11.3), we have \begin{equation}\label{ind}
WD(\pi_\ell(\chi))=Ind^{W_{{\bf Q}_{\ell}}}_{W_{{\bf Q}_{\ell^2}}}(\chi)\otimes|\ |_\ell^{-1/2}. \end{equation}
\noindent Let $\rho_f:G_{\bf Q}\to GL_2(\overline{\bf Q}_\ell)$ be the Galois representation associated to $f$ and $\overline\rho:G_{\bf Q}\to GL_2(\overline{\bf F}_\ell)$ be its reduction modulo $\ell$.\\ As in \cite{Lea}, we impose the following conditions on $\overline\rho$: \begin{equation}\label{con1} \overline\rho\ {\rm is\ absolutely\ irreducible}; \end{equation} \begin{equation}\label{cond2}
{\rm if}\ p|N\ {\rm then}\ \overline\rho(I_p)\not=1; \end{equation} \begin{equation} \label{rara}
{\rm if}\ p|\Delta'\ {\rm and}\ p^2\equiv 1 {\rm mod}\ \ell\ {\rm then}\ \overline\rho(I_p)\not=1; \end{equation} \begin{equation}\label{end} {\rm End}_{\overline{\bf F}_\ell[G_\ell]}(\overline\rho_\ell)=\overline{\bf F}_\ell. \end{equation} \begin{equation}\label{c3} {\rm if}\ \ell=3,\ \overline\rho\ \ {\rm is\ not\ induced\ from\ a\ character\ of\ }\ {\bf Q}(\sqrt{-3}). \end{equation} \noindent Let $K=K(f)$ be a finite extension of ${\bf Q}_\ell$ containing ${\bf Q}_{\ell^2}$, ${\rm Im}(\psi)$ and the eigenvalues for $f$ of all Hecke operators. Let $\mathcal O$ be the ring of integers of $K$, $\lambda$ be a uniformizer of $\mathcal O$, $k=\mathcal O/(\lambda)$ be the residue field.
\noindent Let $\mathcal B$ denote the set of normalized newforms $h$ in $S_2(\Gamma_0(M\ell^2),\psi)$ which are supercuspidal of type $\chi$ at $\ell$, special at primes dividing $\Delta'$ and whose associated representation $\rho_h$ is a deformation of $\overline\rho$. For $h\in\mathcal B$, let $h=\sum_{n=1}^\infty a_n(h)q^n$ be the $q$-expansion of $h$ and let $\mathcal O_h$ be the $\mathcal O$-algebra generated in ${\bf Q}_\ell$ by the Fourier coefficients of $h$. Let ${\bf T}^\psi$ denote the sub-$\mathcal O$-algebra of $\prod_{h\in\mathcal B}\mathcal O_h$ generated by the elements $\widetilde T_p=(a_p(h))_{h\in\mathcal B}$ for $p\not|M\ell$.\\
\section{Deformation problem}
Our next goal is to state a global Galois deformation condition of $\overline\rho$ which is a good candidate for having ${\bf T}^\psi$ as an universal deformation ring.
\subsection{The global deformation condition of type $(\rm{sp},\tau,\psi)_Q$}\label{univ}
First of all we observe that our local Galois representation $\rho_{f,\ell}=\rho_\ell$ is of type $\tau$ (\cite{CDT}).\\
We let $\Delta_1$ be the product of primes $p|\Delta'$ such that $\overline\rho(I_p)\not=1$, and $\Delta_2$ be the product of primes $p|\Delta'$ such that $\overline\rho(I_p)=1$.\\ We denote by $\mathcal C_\mathcal O$ the category of local complete noetherian $\mathcal O$-algebras with residue field $k$. Let $\epsilon:G_p:\to{\bf Z}_\ell^\times$ be the cyclotomic character and $\omega:G_p\to{\bf F}_\ell^\times$ be its reduction mod $\ell$. By analogy with \cite{Lea}, We define the global deformation condition of type $(\rm{sp},\tau,\psi)_Q$: \begin{definition}\label{def} Let $Q$ be a square-free integer, prime to $M\ell$. We consider the functor $\mathcal F_Q$ from $\mathcal C_\mathcal O$ to the category of sets which associate to an object $A\in\mathcal C_\mathcal O$ the set of strict equivalence classes of continuous homomorphisms $\rho:{G_{\bf Q}}\to GL_2(A)$ lifting $\overline\rho$ and satisfying the following conditions: \begin{itemize} \item[a$_Q$)] $\rho$ is unramified outside $MQ\ell$;
\item[b)] if $p|\Delta_1N$ then $\rho(I_p)\simeq\overline\rho(I_p)$ ;
\item[c)] if $p|\Delta_2$ then $\rho_p$ satisfies the sp-condition, that is ${\rm tr}(\rho(F))^2=(p\mu(p)+\mu(p))^2=\psi_p(p)(p+1)^2$ for a lift $F$ of ${\rm Frob}_p$ in $G_p$; \item[d)] $\rho_\ell$ is weakly of type $\tau$; \item[e)] ${\rm det}(\rho)=\epsilon\psi$, where $\epsilon:G_{\bf Q}\to{\bf Z}_\ell^\times$ is the cyclotomic character. \end{itemize} \end{definition}
\noindent It is easy to prove that the functor $\mathcal F_Q$ is representable.
\noindent Let $\mathcal R^\psi_Q$ be the universal ring associated to the functor $\mathcal F_Q$. We put $\mathcal F=\mathcal F_0$, ${\mathcal R^\psi}=\mathcal R^\psi_0$.\\ We observe that if $\overline\rho(I_p)=1$, by the Ramanujan-Petersson conjecture proved by Deligne, the sp-condition rules out thouse defomations of $ \overline\rho$ arising from modular forms which are not special at $p$. This space includes the restrictions to $G_p$ of representations coming from forms in $S_2(\Gamma_0(N\Delta'\ell^2),\psi)$ which are special at $p$, but it does not contain those coming from principal forms in $S_2(\Gamma_0(N\Delta'\ell^2),\psi)$.
\section{Cohomological modules coming from the Shimura curves}\label{shi}
We fix a prime $\ell>2$. Let $\Delta'$ be a product of an odd number of primes, different from $\ell$. We put $\Delta=\ell\Delta'$. Let $B$ be the indefinite quaternion algebra over ${\bf Q}$ of discriminant $\Delta$. Let $R$ be a maximal order in $B$. Let $N$ be an integer prime to $\Delta$. We put $$K_p^0(N)=i_p^{-1}\left\{\left( \begin{array} [c]{cc} a & b\\ c & d \end{array}
\right)\in GL_2({\bf Z}_p)\ |\ c\equiv 0\ {\rm mod}\ N\right\}$$
$$K_p^1(N)=i_p^{-1}\left\{\left( \begin{array} [c]{cc} a & b\\ c & d \end{array}
\right)\in \ GL_2({\bf Z}_p)\ |\ c\equiv 0\ {\rm mod}\ N,\ a\equiv 1\ {\rm mod}\ N\right\}.$$
Let $s$ be a prime $s\not|N\Delta$. We define $$V_0(N,s)=\prod_{p\not|Ns}R^\times_p\times \prod_{p|N}K_p^0(N)\times K_s^1(s^2)$$ and $$V_1(N,s)=\prod_{p\not|N\ell s}R^\times_p\times\prod_{p|N}K_p^1(N)\times K_s^1(s^2)\times (1+u_\ell R_\ell).$$
We observe that there is an isomorphism $$V_0(N,s)/V_1(N,s)\simeq({\bf Z}/N{\bf Z})^\times\times{\bf F}_{\ell^2}^\times.$$ We will consider the character $\widehat\psi$ of $V_0(N,s)$ with kernel $V_1(N,s)$ defined as follow: $$\widehat\psi:=\prod_{p|N}\psi_p\times\chi:({\bf Z}/N{\bf Z})^\times\times{\bf F}_{\ell^2}^\times\to{\bf C}^\times$$ and we shall consider the space $S_2(V_0(N,s),\widehat\psi)$ of quaternionic modular forms with nebentypus $\widehat\psi$.\\ For $i=0,1$ let $\Phi_i(N,s)=(GL_2^+({\bf R})\times V_i(N,s))\cap B_{\bf Q}^\times$ and we consider the Shimura curves: $${\bf X}_i(N,s)=B_{\bf Q}^\times\setminus B^\times_{\bf A}/K_\infty^+\times V_i(N,s).$$ The finite commutative group $\Omega=({\bf Z}/N{\bf Z})^\times\times{\bf F}_{\ell^2}^\times$ naturally acts on the $\mathcal O$-module $H^*({\bf X}_1(N,s),\mathcal O)$ via its action on ${\bf X}_1(N,s)$. Since there is an injection of $H^*({\bf X}_1(N,s),\mathcal O)$ in $H^*({\bf X}_1(N,s),K)$, by (\cite{H2}, \S 7) the cohomology group $H^1({\bf X}_1(N,s),\mathcal O)$ is also equipped with the action of Hecke operator $T_p$, for $p\not=\ell$ and diamond operators $\langle n\rangle$ for $n\in({\bf Z}/N{\bf Z})^\times$. The Hecke action commutes with the action of $\Omega$, since we do not have a $T_\ell$ operator. The two actions are $\mathcal O$-linear.\\ We can write $\Omega=\Omega_1\times\Omega_2$ where $\Omega_1$ is the $\ell$-Sylow subgroup of $\Omega$ and $\Omega_2$ is the subgroup of $\Omega$ with order prime to $\ell$. Since $\Omega_2\subseteq\Omega$, $\Omega_2$ acts on $H^*({\bf X}_1(N,s),\mathcal O)$ and so $H^*({\bf X}_1(N,s),\mathcal O)=\bigoplus_\varphi H^*({\bf X}_1(N,s),\mathcal O)^\varphi$ where $\varphi$ runs over the characters of $\Omega_2$ and $H^*({\bf X}_1(N,s),\mathcal O)^\varphi$ is the sub-Hecke-module of $H^*({\bf X}_1(N,s),\mathcal O)$ on which $\Omega_2$ acts by the character $\varphi$. Since, by hypothesis, $\psi$ has order prime to $\ell$, $H^*({\bf X}_1(N,s),\mathcal O)^{\widehat\psi}=H^*({\bf X}_1(N,s),\mathcal O)^\varphi$ for some character $\varphi$ of $\Omega_2$. So $H^*({\bf X}_1(N,s),\mathcal O)^{\widehat\psi}$ is a direct summand of $H^*({\bf X}_1(N,s),\mathcal O)$. {} \noindent It follows easily from the Hochschild-Serre spectral sequence that $$H^*({\bf X}_1(N,s),\mathcal O)^{\widehat\psi}\simeq H^*({\bf X}_0(N,s),\mathcal O({\widehat\psi}))$$ where $\mathcal O({\widehat\psi})$ is the sheaf $B^\times\setminus B_{\bf A}^\times\times\mathcal O/K_\infty^+\times V_0(N,s),$ $B^\times$ acts on $B^\times_{\bf A}\times\mathcal O$ on the left by $\alpha\cdot(g,m)=(\alpha g,m)$ and $K_\infty^+\times V_0(N,s)$ acts on the right by $(g,m)\cdot v=(g,m)\cdot(v_\infty,v^\infty)=(gv,{\widehat\psi}(v^\infty)m)$ where $v_\infty$ and $v^\infty$ are respectively the infinite and finite part of $v$ . By translating to the cohomology of groups we obtain (see \cite{sl}, Appendix) $H^1({\bf X}_1(N,s),\mathcal O)^{\widehat\psi}\simeq H^1(\Phi_0(N,s),\mathcal O(\widetilde\psi)),$ where $\widetilde\psi$ is the restriction of ${\widehat\psi}$ to $\Phi_0(N,s)/\Phi_1(N,s)$ and $\mathcal O(\widetilde\psi)$ is $\mathcal O$ with the action of $\Phi_0(N,s)$ given by $a\mapsto\widetilde\psi^{-1}(\gamma)a$.
\noindent It is well know the Hecke action on $H^1(\Phi_0(N,s),\mathcal O(\widetilde\psi))$ and the structure of the module $H^1({\bf X}_1(N,s),K)^{\widehat\psi}$ over the Hecke algebra. Let ${\bf T}^{\widehat\psi}_0(N,s)$ be the $\mathcal O$-algebra generated by the Hecke operators $T_p$, $p\not=\ell$ and the diamond operators, acting on $H^1({\bf X}_1(N,s),\mathcal O)^{\widehat\psi}$.
\begin{proposition}\label{ra} $H^1({\bf X}_1(N,s),K)^{\widehat\psi}$ is free of rank 2 over $${\bf T}^{\widehat\psi}_0(N,s)\otimes K={\bf T}^{\widehat\psi}_0(N,s)_K.$$ \end{proposition}
\noindent The proof of proposition \ref{ra}, easy follows from the following lemmas: \begin{lemma} Let $L\supseteq K$ be two Galois fields, let $V$ be a vector space on $K$ and let $T$ be a $K$-algebra. If $V\otimes L$ is free of rank $n$ on $T\otimes L$ then $V$ is free of rank $n$ on $T$. \end{lemma}
\begin{proof} Let $G={\rm Gal}(L/K)$. By the Galois theory, since $V\otimes L$ is free of rank $n$ on $T\otimes L$, we have $V\simeq(V\otimes L)^G\simeq((T\otimes L)^n)^G\simeq((T\otimes L)^G)^n\simeq T^n.$ \end{proof}
\begin{lemma} Let ${\bf T}_0^{\widehat\psi}(N,s)_{\bf C}$ denote the algebra generated over ${\bf C}$ by the operators $T_p$ for $p\not=\ell$ acting on $H^1({\bf X}_1(N,s),{\bf C})^{\widehat\psi}$. Then $H^1({\bf X}_1(N,s),{\bf C})^{\widehat\psi}$ is free of rank 2 over ${\bf T}^{\widehat\psi}_0(N,s)_{\bf C}$. \end{lemma}
\noindent We observe that the proof of this lemma follows by the same analysis explained in (\cite{Lea}, Proof of proposition 1.2), defining an homomorphism $$JL:S_2(V_1(N,s))\to S_2(\Gamma_0(s^2\Delta')\cap\Gamma_1(N\ell^2))$$ which is injective when restricted to the space $S_2(V_0(N,s),\widehat\psi)$ and equivariant by Hecke operators.
\section{The $\mathcal O$-module ${\mathcal M^\psi}$ }\label{tw}
Throughout this section, we largely mirror section 3 of \cite{Lea}, and we formulate a result that generalizes of a result of Terracini to the case of nontrivial nebentypus.\\ We set $\Delta=\Delta'\ell$; let $B$ be the indefinite quaternion algebra over ${\bf Q}$ of discriminant $\Delta$. Let $R$ be a maximal order in $B$.\\
It is convenient to choose an auxiliary prime $s\not|M\ell$, $s>3$ such that no lift of $\overline\rho$ can be ramified at $s$; such prime exists by \cite{DT}, Lemma 2. We consider the group $\Phi_0=\Phi_0(N,s)$; it easy to verify that the group $\Phi_0$ has not elliptic elements (\cite{Lea}).
\noindent There exists an eigenform $\widetilde f\in S_2(\Gamma_0(Ms^2\ell^2),\psi)$ such that $\rho_f=\rho_{\widetilde f}$ and $T_s\widetilde f=0$. By the Jacquet-Langlands correspondence, the form $\widetilde f$ determines a character ${\bf T}^{\widehat\psi}_0(N,s)\to k$ sending the operator $t$ in the class ${\rm mod}\ \lambda$ of the eigenvalue of $t$ for $\widetilde f$. The kernel of this character is a maximal ideal $\mathfrak m$ in ${\bf T}^{\widehat\psi}_0(N,s)$. We define ${\mathcal M^\psi}=H^1({\bf X}_1(N,s),\mathcal O)^{\widehat\psi}_\mathfrak m.$ By combining Proposition 4.7 of \cite{DDT} with the Jacquet-Langlands correspondence we see that there is a natural isomorphism ${\bf T}^\psi\simeq{\bf T}_0^{\widehat\psi}(N,s)_\mathfrak m.$ Therefore, by Proposition \ref{ra}, ${\mathcal M^\psi}\otimes_\mathcal O K\ {\rm is\ free\ of\ rank\ 2\ over}\ {\bf T}^\psi\otimes_\mathcal O K.$
\noindent Let $\mathcal B$ denote the set of newforms $h$ of weight two, nebentypus $\psi$, level dividing $M\ell^2s^2$, special at primes $p$ dividing $\Delta'$, supercuspidal of type $\tau$ at $\ell$ and such that $\overline\rho_h\sim\overline\rho$. For a newforms $h\in\mathcal B$, we let $K_h$ denote the field over ${\bf Q}_\ell$ generated by its coefficients $a_n(h)$, $\mathcal O_h$ denote the ring of integers of $K_h$ and let $\lambda$ be a uniformizer of $\mathcal O_h$. We let $A_h$ denote the subring of $\mathcal O_{h}$ consisting of those elements whose reduction mod $\lambda$ is in $k$. We know that with respect to some basis, we have $\rho_h:G_{\bf Q}\to GL_2(A_h)$ a deformation of $\overline\rho$ satisfying our global deformation problem.\\ The universal property of ${\mathcal R^\psi}$ furnishes a unique homomorphism $\pi_h:{\mathcal R^\psi}\to A_h$ such that the composite $G_{\bf Q}\to GL_2({\mathcal R^\psi})\to GL_2(A_h)$ is equivalent to $\rho_h$. Since ${\mathcal R^\psi}$ is topologically generated by the traces of $\rho^{\rm univ}({\rm Frob}_p)$ for $p\not=\ell$, (see \cite{Ma}, \S1.8), we conclude that the map ${\mathcal R^\psi}\to\prod_{h\in\mathcal B}\mathcal O_h$ such that $r\mapsto(\pi_h(r))_{h\in\mathcal B}$ has image ${\bf T}^\psi$. Thus there is a surjective homomorphism of $\mathcal O$-algebras $\Phi:{\mathcal R^\psi}\to{\bf T}^\psi.$ Our goal is to prove the following
\begin{theorem}\label{goal} \begin{itemize} \item[a)] ${\mathcal R^\psi}$ is complete intersection of dimension 1; \item[b)] $\Phi:{\mathcal R^\psi}\to{\bf T}^\psi$ is an isomorphism; \item[c)] ${\mathcal M^\psi}$ is a free ${\bf T}^\psi$-module of rank 2. \end{itemize} \end{theorem}
\subsection{Proof of theorem \ref{goal}}
\noindent In order to prove theorem \ref{goal}, we shall apply the Taylor-Wiles criterion in the version of Diamond and Fujiwara and we continue to follow section 3 of \cite{Lea} closely.\\ \noindent We shall prove the existence of a family $\mathcal Q$ of finite sets $Q$ of prime numbers, not dividing $M\ell$ and of a ${\mathcal R^\psi}_Q$-module ${\mathcal M^\psi}_Q$ for each $Q\in\mathcal Q$ such that the system $({\mathcal R^\psi}_Q,{\mathcal M^\psi}_Q)_{Q\in\mathcal Q}$ satisfies the conditions (TWS1), (TWS2), (TWS3), (TWS4), (TWS5) and (TWS6) of \cite{Lea}.
\noindent If this conditions are satisfy for the family $({\mathcal R^\psi}_Q,{\mathcal M^\psi}_Q)_{Q\in\mathcal Q}$, it will be called a {\bf Taylor-Wiles system} for $({\mathcal R^\psi},{\mathcal M^\psi})$. Then theorem \ref{goal} will follow from the isomorphism criterion (\cite{D}, theorem 2.1) developed by Wiles, Taylor-Wiles.\\ As in section 3.1 of \cite{Lea}, let $Q$ be a finite set of prime numbers not dividing $N\Delta$ and such that \begin{itemize} \item[(A)] $q\equiv 1\ {\rm mod}\ \ell,\ \ \forall q\in Q$; \item[(B)] if $q\in Q$, $\overline\rho({\rm Frob}_q)$ has distinct eigenvalues $\alpha_{1,q}$ and $\alpha_{2,q}$ contained in $k$. \end{itemize}
We will define the modules $\mathcal M_Q$. If $q\in Q$ we put $$K'_q=\left\{\alpha\in R_q^\times\ |\ i_q(\alpha)\in \left( \begin{array} [c]{cc} H_q & *\\ q{\bf Z}_q & * \end{array}
\right)\right\}$$ where $H_q$ is the subgroup of $({\bf Z}/q{\bf Z})^\times$ consisting of elements of order prime to $\ell$. By analogy with the definition of $V_Q$ in section 3.3 of \cite{Lea}, we define $$V_Q'(N,s)=\prod_{p\not|NQs}R_p^\times\times\prod_{p|N}K_p^0(N)\times K_s^1(s^2)\times\prod_{q|Q}K'_q$$
$$V_Q(N,s)=\prod_{p\not|NQs}R_p^\times\times\prod_{p|NQ}K_p^0(NQ)\times K_s^1(s^2)$$ $$\Phi_Q=(GL_2({\bf R}^+)\times V_Q(N,s))\cap B^\times,\ \ \Phi_Q'=(GL_2({\bf R}^+)\times V_Q'(N,s))\cap B^\times.$$
\noindent Then $\Phi_Q/\Phi_Q'\simeq\Delta_Q$ acts on $H^1(\Phi_Q',\mathcal O(\widetilde\psi))$. Let ${\bf T}_Q^{'\widehat\psi}(N,s)$ (resp. ${\bf T}_Q^{\widehat\psi}(N,s)$) be the Hecke $\mathcal O$-algebra generated by the Hecke operators $T_p$, $p\not=\ell$ and the diamond operators ( that are those Hecke operators coming from $\Delta_Q$) , acting on $H^1({\bf X}'_Q(N,s),\mathcal O)^{\widehat\psi}$ (resp. $H^1({\bf X}_Q(N,s),\mathcal O)^{\widehat\psi}$) where ${\bf X}'_Q(N,s)$ (resp. ${\bf X}_Q(N,s)$) is the Shimura curve associated to $V'_Q(N,s)$ (resp. $V_Q(N,s)$).\ There is a natural surjection $\sigma_Q:{\bf T}_Q^{'\widehat\psi}(N,s)\to {\bf T}_Q^{\widehat\psi}(N,s)$. Since the diamond operator $\langle n\rangle$ depends only on the image of $n$ in $\Delta_Q$, ${\bf T}_Q^{'\widehat\psi}(N,s)$ is naturally an $\mathcal O[\Delta_Q]$-algebra. Let $\widetilde\alpha_{i,q}$ for $i=1,2$ be the two roots in $\mathcal O$ of the polinomial $X^2-a_q(f)X+q$ reducing to $\alpha_{i,q}$ for $i=1,2$.
There is a unique eigenform $\widetilde f_Q\in S_2(\Gamma_0(MQs^2\ell^2),\widehat\psi)$ such that $\rho_{\widetilde f_Q}=\rho_f$, $a_s(\widetilde f_Q)=0$, $a_q(\widetilde f_Q)=\widetilde\alpha_{2,q}$ for $q|Q$, where $\widetilde\alpha_{2,q}$ is the lift of $\alpha_{2,q}$.\\ By the Jacquet-Langlands correspondence, the form $\widetilde f_Q$ determines a character $\theta_Q:{\bf T}_Q^{\widehat\psi}(N,s)\to k$ sending $T_p$ to $a_p(\widetilde f_Q)\ {\rm mod}\ \lambda$ and the diamond operators to 1. We define $\widetilde\mathfrak m_Q={\rm ker} \theta_Q,$ $\mathfrak m_Q=\sigma^{-1}_Q(\widetilde\mathfrak m_Q)$, and $\mathcal M_Q=H^1(\Phi'_Q,\mathcal O(\widetilde\psi))_{\mathfrak m_Q}.$ Then the map $\sigma_Q$ induce a surjective homomorphism ${\bf T}_Q^{'\widehat\psi}(N,s)_{\mathfrak m_Q}\to {\bf T}_Q^{\widehat\psi}(N,s)_{\widetilde\mathfrak m_Q}$ whose kernel contains $I_Q({\bf T}_Q^{'\widehat\psi}(N,s))_{\mathfrak m_Q}$. \\ If $\mathcal Q$ is a family of finite sets $Q$ of primes satisying conditions $(A)$ e $(B)$, conditions (TWS1) and (TWS2) holds, as proved in \cite{Lea}, proposition 3.2; by the same methods as in \S 6 of \cite{DDT1} and in \S 4, \S 5 of \cite{Des}, it is easy to prove that our system $({\mathcal R^\psi}_Q,\mathcal M_Q)_{Q\in\mathcal Q}$ realize simultaneously conditions (TWS3), (TWS4), (TWS5). \\ We put $\delta_q=\left( \begin{array} [c]{cc} q & 0\\ 0 & 1 \end{array} \right).$
Let $\eta_q$ be the id\`{e}le in $B_{\bf A}^\times$ defined by $\eta_{q,v}=1$ if $v\not|q$ and $\eta_{q,q}=i^{-1}_q\left( \begin{array} [c]{cc} q & 0\\ 0 & 1 \end{array} \right).$ By strong approximation, write $\eta_q=\delta_qg_\infty u$ with $\delta_q\in B^\times$, $g_\infty\in GL_2^+({\bf R})$, $u\in V_{Q'}(N,s)$. We define a map \begin{eqnarray} H^1(\Phi_Q,\mathcal O(\widetilde\psi))&\to& H^1(\Phi_{Q'},\mathcal O(\widetilde\psi))\\
x&\to& x|_{\eta_q} \end{eqnarray}
as follows: let $\xi$ be a cocycle representing the cohomology class $x$ in $H^1(\Phi_{Q},\mathcal O(\widetilde\psi))$; then $x|_{\eta_q}
$ is represented by the cocycle $$\xi|_{\eta_q}(\gamma)=\widehat\psi(\delta_q)\cdot\xi(\delta_q\gamma\delta_q^{-1}).$$ We observe that if $\mathcal Q$ is a family of finite sets $Q$ of primes satisying conditions $(A)$ e $(B)$, then condition (TWS6) hold for the system $({\mathcal R^\psi}_Q,\mathcal M_Q)_{Q\in\mathcal Q}$. The proof is essentially the same as in \cite{Lea}, using the following lemma: \begin{lemma}
$$T_p(x|_{\eta_q})=(T_p(x)|_{\eta_q})\ \ \ {\rm if}\ \ p\not|MQ'\ell,$$ \begin{equation}\label{r}
T_q(x|_{\eta_q})=q\widehat\psi(q)res_{\Phi_Q/\Phi_{Q'}}x, \end{equation}
$$T_q(res_{\Phi_Q/\Phi_{Q'}}x)=res_{\Phi_Q/\Phi_{Q'}}(T_q(x))-x|_{\eta_q}.$$ \end{lemma} \proof We prove (\ref{r}). We put $\widetilde\delta_q=\left( \begin{array} [c]{cc} 1 & 0\\ 0 & q \end{array} \right),$ and we decompose the double coset $\Phi_{Q'}\widetilde\delta_q\Phi_{Q'}=\coprod_{i=1}^q\Phi_{Q'}\widetilde\delta_qh_i$ with $h_i\in\Phi_{Q'}$; if $\gamma\in\Phi_{Q'}$, then: \begin{eqnarray}
T_q(\xi|_{\eta_q})(\gamma)&=&\sum_{i=1}^q\widehat\psi(h_i\widetilde\delta_q)\xi|_{\eta_q}(\widetilde\delta_qh_i\gamma h_{j(i)}^{-1}\widetilde\delta^{-1}_q)\nonumber\\ &=&\sum_{i=1}^q\widehat\psi(h_i)\widehat\psi(\widetilde\delta_q)\widehat\psi(\delta_q)\xi(\delta_q\widetilde\delta_qh_i\gamma h_{j(i)}^{-1}\widetilde\delta^{-1}_q\delta^{-1}_q)\nonumber\\ &=&\widehat\psi(q)\sum_{i=1}^q\widehat\psi(h_i)\xi(h_i\gamma h_{j(i)}^{-1}) \end{eqnarray} where the latter sum holds since $\delta_q\widetilde\delta_q=\left( \begin{array} [c]{cc} q & 0\\ 0 & q \end{array}
\right)$. From the cocycle relations we have $T_q(x|_{\eta_q})=q\widehat\psi(q)res_{\Phi_Q/\Phi_{Q'}}x.$\qed
\noindent The conditions defining the functor $\mathcal F_Q$, characterize a global Galois deformation problem with fixed determinant (\cite{M}, \S 26). We let $ad^0\overline\rho$ denote the subrepresentation of the adjoint representation of $\overline\rho$ over the space of the trace-0-endomorphisms and we let $ad^0\overline\rho(1)={\rm Hom}(ad^0\overline\rho,\mu_p)\simeq Symm^2(\overline\rho)$, with the action of $G_p$ given by $g\varphi)(v)=g\varphi(g^{-1}v).$ Local deformation conditions a$_Q$), b), c), d) allow one to define for each place $v$ of ${\bf Q}$, a subgroup $L_v$ of $H^1(G_v, ad^0\overline\rho)$, the tangent space of the deformation functor (see \cite{M}). We will describe the computation of the local terms of the dimension formula coming from the Poitou-Tate sequence:
\begin{itemize} \item ${\rm dim}_k(H^0(G_{\bf Q},ad^0\overline\rho)={\rm dim}_k(H^0(G_{\bf Q},ad^0\overline\rho(1))=0$, by the same argument as in \cite{Des} pp.441. \item ${\rm dim}_kL_\ell=1$. In fact let ${\bf R}^D_{\mathcal O,\ell}$ be the local universal deformation ring associated to a local deformation problem of being weakly of type $\tau$ \cite{CDT}. Since, in dimension 2, potentially Barsotti-Tate is equivalent to potentially crystalline (hence potentially semi stable) of Hodge-Tate weight $(0,1)$ (see \cite{FM}, theorem C2), this allow us to apply Savitt's result (\cite{savitt}, theorem 6.22). Since under our hypothesis ${\bf R}^D_{\mathcal O,\ell}\not=0$, we deduce that there is an isomorphism $ \mathcal O[[X]]\simeq{\bf R}^D_{\mathcal O,\ell}$. \item ${\rm dim}_kH^0(G_\ell,ad^0\overline\rho)=0$, because of hypothesis (\ref{end})
\item ${\rm dim}_kH^1(G_p/I_p,(ad^0\overline\rho)^{I_p})-{\rm dim}_kH^0(G_p,ad^0\overline\rho)=0$ for $p|N\Delta_1$\\ If we let $W=ad^0\overline\rho$, this follows from the exact exact sequence \begin{displaymath} 0\to H^0(G_p,W)\to H^0(I_p,W)\stackrel{\scriptstyle{\rm Frob}_p-1}{\to}H^0(I_p,W)\to H^1(G_p/I_p,W^{I_p})\to 0. \end{displaymath}
\item ${\rm dim}_kL_p=1$ for $p|\Delta_2$, infact the following lemma holds: \begin{lemma}\label{versal} The versal defomation ring of the local defomation problem of satisfying the sp-condition is $\mathcal O[[X,Y]]/(X,XY)=\mathcal O[[Y]].$ \end{lemma} \item ${\rm dim}_kH^0(G_p,ad^0\overline\rho)=1$, since the eigenvalues of $\overline\rho({\rm Frob}_p)$ are distinct.
\item ${\rm dim}_kH^1(G_q,ad^0\overline\rho)=2$ if $q|Q$, in fact $H^1(G_q/I_q,W)=W/({\rm Frob}_q-1)W$ has dimension 1, because in the base $\left\{\left( \begin{array} [c]{cc} 1 & 0\\ 0 & -1 \end{array} \right), \left( \begin{array} [c]{cc} 0 & 0\\ 0 & 1 \end{array} \right), \left( \begin{array} [c]{cc} 0 & 0\\ 0 & 1 \end{array} \right)\right\}$ of $W$, ${\rm Frob}_q$ is the matrix $$\left( \begin{array} [c]{ccc} 1 & 0 & 0\\ 0 & \alpha_{1,q}\alpha_{2,q}^{-1} & 0\\ 0 & 0 & \alpha_{1,q}^{-1}\alpha_{2,q} \end{array} \right)$$
and $\alpha_{1,q}\not=\alpha_{2,q}$ by hypothesis. We observe that: \begin{eqnarray}
H^1(I_q,W)^{G_q/I_q}&=&\{\alpha\in {\rm Hom}({\bf Z}_q^\times,W)\ |\ ({\rm Frob}_q-1)\alpha=0\}\nonumber\\ &\simeq& W[{\rm Frob}_q-1]\nonumber \end{eqnarray} is again one-dimensional, and $H^2(G_q/I_q,W)=0$ since $G_q/I_q\simeq\widehat{{\bf Z}}.$ The desidered result follows from the inflation-restriction exact sequence.
\item ${\rm dim}_kH^0(G_q,ad^0\overline\rho)=1$ if $q|Q$, since the eigenvalues of ${\rm Frob}_q$ are $1, \alpha_{1,q}^2, \alpha_{2,q}^2$ and $\alpha_{1,q}^2, \alpha_{2,q}^2\not=1$ for hypothesis.
\item ${\rm dim}_kH^1(G_\infty,ad^0\overline\rho)=0$, since $|G_\infty|=2\not=\ell$. \item ${\rm dim}_kH^0(G_\infty,ad^0\overline\rho)=1$, since the eigenvalues of complex conjugation on $ad^0\overline\rho$ are $\{1,-1,-1\}.$ \end{itemize}
\proof[lemma \ref{versal}]
\noindent We first observe that it is possible to characterize the deformations of $\overline\rho_p$ in the unramified case, if $p^2\not\equiv 1\ {\rm mod}\ \ell$ and, as in Lemma 2.1 in \cite{Lea}, it it is easy to prove that the versal deformation ring ${\mathcal R^\psi}'_p$ is generated by two elements $X,Y$ such that $XY=0$. \\ It is immediate to see that the sp-condition is equivalent to require that every homomorphism $\varphi:{\mathcal R^\psi}'_p\to A$ associated to the deformation $\rho$ of $\overline\rho_{G_p}$ over a $\mathcal O$-algebra $A\in\mathcal C_\mathcal O$, has $\varphi(X)=0$.\qed
\noindent The dimension formula allow us to obtain the following identity: \begin{eqnarray}\label{f}
{\rm dim}_kSel_Q(ad^0\overline\rho)-{\rm dim}_kSel^*_Q(ad^0\overline\rho(1))=|Q| \end{eqnarray} and, since the minimal number of topological generators of ${\mathcal R^\psi}_Q$ is equal to ${\rm dim}_kSel_Q(ad^0\overline\rho)$,
we obtain that the $\mathcal O$-algebra ${\mathcal R^\psi}_Q$ can be generated topologically by $|Q|+{\rm dim}_kSel^*_Q(ad^0\overline\rho(1))$ elements. Applying the same arguments as in $\cite{DDT}$ the proof of theorem \ref{goal} follows.
\section{Quaternionic congruence module}
In this section we will keep the notations as in the previous sections.\\
As remarked in section \ref{shi}, there is an injective map $$JL:S_2(V_0(N),\widehat\psi)\to S_2(\Gamma_0(\Delta')\cap\Gamma_1(N\ell^2))$$ where $V_0(N)=\prod_{p\not|N}R_p^\times\times\prod_{p|N}K_p^0(N)$ is an open compact subgroup of $B_{\bf A}^{\times,\infty}$. We put $V_{\widehat\psi}=JL(S_2(V_0(N),\widehat\psi))$ the subset of $S_2(\Gamma_0(\Delta')\cap\Gamma_1(N\ell^2))$ generated by thouse new eigenforms with nebentypus $\psi$ which are supercuspidal of type $\tau$ at $\ell$ and special at primes dividing $\Delta'$. Let $K\subset\overline{\bf Q}_\ell$ be a finite extension containing ${\bf Q}_{\ell^2}$; we consider $f\in S_2(\Gamma_0(N\Delta'\ell^2),\psi)$ a newform, supercuspidal of type $\tau$ at $\ell$ and special at primes dividing $\Delta'$, and let $X$ be the subspace of $V_{\widehat\psi}(K)$ spanned by $f$. We remark, that there is an isomorphism between the $K$-algebra ${\bf T}_0^{\widehat\psi}(K)={\bf T}_0^{\widehat\psi}(N)\otimes_\mathcal O K$ generated over $K$ by the operators $T_p$, $p\not=\ell$ acting on $H^1({\bf X}_1(N),K)^{\widehat\psi}$ and the Hecke algebra generated by all the Hecke operators acting on $V_{\widehat\psi}(K)$. Thus $V_{\widehat\psi}=X\oplus Y$ where $Y$ is the orthogonal complement of $X$ with respect to the Petersson product over $V_{\widehat\psi}$ and there is an isomorphism $${\bf T}_0^{\widehat\psi}(K)\simeq{\bf T}_{0}^{\widehat\psi}(K)_X\oplus{\bf T}_{0}^{\widehat\psi}(K)_Y$$ where ${\bf T}_{0}^{\widehat\psi}(K)_X={\bf T}_0^{\widehat\psi}(N)_X\otimes K$ and ${\bf T}_{0}^{\widehat\psi}(K)_Y={\bf T}_0^{\widehat\psi}(N)_Y\otimes K $ are the $K$ algebras generated by the Hecke operators acting on $X$ and $Y$ respectively.\\ As in classical case, it is possible to define the quaternionic congruence module for $f$ and, by the Jacquet-Langlands correspondence, it is easy to prove that $\widetilde M^{\rm quat}=\mathcal O/(\lambda^n)$ where $n$ is the smallest integer such that $\lambda^ne_f\in{\bf T}_0^{\widehat\psi}(N)$ where $e_f$ is the projector onto the coordinate corresponding to $f$. There are the isomorphisms $$\widetilde M^{\rm quat}\simeq\frac{{\bf T}_0^{\widehat\psi}(N)_X\oplus{\bf T}_0^{\widehat\psi}(N)_Y}{{\bf T}_0^{\widehat\psi}(N)}\simeq\frac{e_f{\bf T}_0^{\widehat\psi}(N)}{e_f{\bf T}_0^{\widehat\psi}(N)\cap{\bf T}_0^{\widehat\psi}(N)}$$ where the first one is an isomorphism of $\mathcal O$-modules and the second one is obtained considernig the projection map of ${\bf T}_0^{\widehat\psi}(N)_X\oplus{\bf T}_0^{\widehat\psi}(N)_Y$ onto the first component.\\ Now, let $\mathfrak m$ be the maximal ideal of ${\bf T}_0^{\widehat\psi}(N)$ definend in section \ref{tw}, since $e_f{\bf T}_0^{\widehat\psi}(N)_\mathfrak m=e_f{\bf T}_0^{\widehat\psi}(N)$ then, by the results in the previous sections $$\widetilde M^{\rm quat}=\mathcal O/(\lambda^n)=\frac{e_f{\bf T}^\psi}{e_f{\bf T}^\psi\cap{\bf T}^\psi}=(\widetilde L^{\rm quat})^2$$ where $\widetilde L^{\rm quat}$ is the quaternionic cohomological congruence module for $f$ $$\widetilde L^{\rm quat}=\frac{e_fH^1({\bf X}_1(N),\mathcal O)^{\widehat\psi}}{e_fH^1({\bf X}_1(N),\mathcal O)^{\widehat\psi}\cap H^1({\bf X}_1(N),\mathcal O)^{\widehat\psi}}.$$
\section{A generalization of the Conrad, Diamond and Taylor's result using Savitt's theorem}\label{gen}
In \cite{CDT}, Conrad, Diamond and Taylor assume that the type $\tau$ is strongly acceptable for $\overline\rho|_{G_\ell}$ and they consider the global Galois deformation problem of being of {\bf type $(S,\tau)$ }, where $S$ is a set of rational primes which does not contain $\ell$. Savitt's theorem allows to suppress the assumption of acceptability in the definition of strong acceptability and their result, theorem 5.4.2., still follows. They first suppose that $S=\emptyset$ and they prove their result using the improvement on the method of Taylor and Wiles \cite{taywi} found by Diamond \cite{D} and Fujiwara \cite{Fu}, then they prove their result for an arbitrary $S$ by induction on $S$, using Ihara's Lemma.\\ In particular, if $S$ is a set of rational prime not dividing $N\Delta'\ell$, we consider a newform $f$ of weight 2, level dividing $SM\ell^2$, with nebentypus $\psi$ (not trivial), supercuspidal of type $\tau=\chi\oplus\chi^\sigma$ at $\ell$ and such that $\overline\rho_f=\overline\rho$ satisfies the conditions (\ref{con1}),(\ref{cond2}),(\ref{end}) and (\ref{c3}) of section \ref{de}. As a general hypothesis, we assume that $f$ occurs with type $\tau$ and minimal level, that is $\overline\rho_f$ is ramified at every prime in $\Delta'$.\\ We consider deformations of type $(S,\tau)$ of $\overline\rho_f$ unramified outside the level of $f$ and such that ${\rm det}(\rho)=\epsilon\psi$; we will call this deformation problem of type $(S,\tau,\psi)$. Then Savitt'theorem assure that the tangent space of the deformation functor at $\ell$ is still one-dimensional and so it is possible to go on with the same construction as in \cite{CDT}. Let $\mathcal R_S^{\rm mod,\psi}$ be classical type $(S,\tau,\psi)$ universal deformation ring which pa\-ra\-me\-trizes representations of type $(S,\tau,\psi)$ with residual representation $\overline\rho$ and let ${\bf T}_S^{\rm mod,\psi}$ be the classical Hecke algebra acting on the space of the modular forms of type $(S,\tau,\psi)$. If we denote by $\mathcal M_S^{\rm mod}$ the cohomological module defined in \S 5.3 of \cite{CDT}, which is essentially the \lq\lq$\tau$-part\rq\rq of the first cohomology group of a modular curve of level depending on $S$, let $\mathcal M^{\rm mod,\psi}_S$ be the $\psi$-part of $\mathcal M_S^{\rm mod}$. Then the following proposition follows: \begin{proposition}\label{gcdt}
The map $$\Phi_S^{\rm mod,\psi}:\mathcal R_S^{\rm mod,\psi}\to{\bf T}_S^{\rm mod,\psi}$$ is a complete intersection isomorphism and $\mathcal M^{\rm mod,\psi}_S$ is a free ${\bf T}_S^{\rm mod,\psi}$-module of rank 2. \end{proposition} \noindent In particular we observe that: \begin{lemma}
There is an isomorphism of ${\bf T}^\psi$-modules between $H^1({\bf X}_1(N),\mathcal O)^{\widehat\psi}_\mathfrak m$ and $\mathcal M^{{\rm mod},\psi}_\emptyset$ where ${\bf X}_1(N)$ is the Shimura curve associated to $V_1(N)=\prod_{p\not|N\ell}R_p^\times\prod_{p|N}K_p^1(N)\times(1+u_\ell R_\ell)$. \end{lemma} \proof We observe that if $f$ occurs with type $\tau$ and minimal level ${\mathcal R^\psi}\simeq\mathcal R^{{\rm mod},\psi}_\emptyset$. By theorem \ref{goal} and proposition \ref{gcdt}, there is an isomorphism between the Hecke algebras ${\bf T}^\psi\simeq{\bf T}^{{\rm mod},\psi}_\emptyset$ thus ${\mathcal M^\psi}\simeq \mathcal M^{{\rm mod},\psi}_\emptyset$ as ${\bf T}^\psi$-modules.\qed
\noindent We will describe some consequences of this result.
\section{Congruence ideals}\label{con}
In this section we generalize the results about congruence ideals of Terracini \cite{Lea}, considering modular forms with nontrivial nebentypus.\\ Let $\Delta_1$ be a set of primes, disjoint from $\ell$. By an abuse of notation, we shall sometimes denote by $\Delta_1$ also the product of the prime in this set.\\ As before, let $f$ be a newform in $S_2(\Gamma_0(N\Delta_1\ell^2),\psi))$, supercuspidal of type $\tau$ at $\ell$ and as a general hypothesis we assume that the residual representation $\overline\rho$, associated to $f$ occurs with type $\tau$ and minimal level. \\ We observe that if $\overline\rho_\ell$ has the form as at pp.525 of \cite{CDT}, then $f$ satisfies the above hypothesis.\\ We assume that the character $\psi$ satisfies the condition as in the section \ref{de}, that $\overline\rho$ is absolutely irreducible and that $\overline\rho_\ell$ has trivial centralizer.\\ Let $\Delta_2$ be a finite set of primes $p$, not dividing $\Delta_1\ell$ such that $p^2\not\equiv 1\ {\rm mod}\ \ell$ and ${\rm tr}(\overline\rho({\rm Frob}_p))^2\equiv\psi(p)(p+1)^2\ {\rm mod}\ \ell$. We let $\mathcal B_{\Delta_2}$ denote the set of newforms $h$ of weight 2, character $\psi$ and level dividing $N\Delta_1\Delta_2\ell$ which are special at $\Delta_1$, supercuspidal of type $\chi$ at $\ell$ and such that $\overline\rho_h=\overline\rho$. We choose an $\ell$-adic ring $\mathcal O$ with residue field $k$, sufficiently large, so that every representation $\rho_h$ for $h\in\mathcal B_{\Delta_2}$ is defined over $\mathcal O$ and $Im(\psi)\subseteq\mathcal O$. For every pair of disjoint subset $S_1, S_2$ of $\Delta_2$ we denote by $\mathcal R^\psi_{S_1,S_2}$ the universal solution over $\mathcal O$ for the deformation problem of $\overline\rho$ consisting of the deformations $\rho$ satisfying: \begin{itemize} \item[a)] $\rho$ is unramified outside $N\Delta_1 S_1S_2\ell$;
\item[b)] if $p|\Delta_1N$ then $\rho(I_p)=\overline\rho(I_p)$;
\item[c)] if $p|S_2$ then $\rho_p$ satisfies the sp-condition; \item[d)] $\rho_\ell$ is weakly of type $\tau$; \item[e)] ${\rm det}(\rho)=\epsilon\psi$ where $\epsilon:G_{\bf Q}\to{\bf Z}^\times_\ell$ is the cyclotomic character. \end{itemize} Let $\mathcal B_{S_1,S_2}$ be the set of newforms in $\mathcal B_{\Delta_2}$ of level dividing $N\Delta_1S_1S_2\ell$ which are special at $S_2$ and let ${\bf T}^\psi_{S_1,S_2}$ be the sub-$\mathcal O$-algebra of $\prod_{h\in\mathcal B_{S_1,S_2}}\mathcal O$ generated by the elements $\widetilde T_p=(a(h))_{h\in\mathcal B_{S_1,S_2}}$ for $p$ not in $\Delta_1\cup S_1\cup S_2\cup\{\ell\}$. Since $\mathcal R^\psi_{S_1,S_2}$ is generated by traces, we know that there exist a surjective homomorphism of $\mathcal O$-algebras $\mathcal R^\psi_{S_1,S_2}\to{\bf T}^\psi_{S_1,S_2}$. Moreover by the results obtained in section \ref{gen}, we have that $\mathcal R^\psi_{S_1,\emptyset}\to{\bf T}^\psi_{S_1,\emptyset}$ is an isomorphism of complete intersections, for any subset $S_1$ of $\Delta_2$.\\ If $\Delta_1\not=1$ then each ${\bf T}^\psi_{\emptyset,S_2}$ acts on a local component of the cohomology of a suitable Shimura curve, obtained by taking an indefinite quaternion algebra of discriminant $S_2\ell$ or $S_2\ell p$ for a prime $p$ in $\Delta_1$. Therefore, theorem \ref{goal} gives the following: \begin{corollary} Suppose that $\Delta_1\not=1$ and that $\mathcal B_{\emptyset,S_2}\not=\emptyset;$ then the map $$\mathcal R^\psi_{\emptyset,S_2}\to{\bf T}^\psi_{\emptyset,S_2}$$ is an isomorphism of complete intersections. \end{corollary} If $p\in S_2$ there is a commutative diagram: \begin{eqnarray} &\mathcal R^\psi_{S_1p,S_2/p}& \to \mathcal R^\psi_{S_1,S_2}\nonumber\\ &\downarrow &\ \ \ \ \ \downarrow\nonumber\\ &{\bf T}^\psi_{S_1p,S_2/p}&\to {\bf T}^\psi_{S_1,S_2}\nonumber\\ \end{eqnarray} where all the arrows are surjections.\\
For every $p|\Delta_2$ the deformation over $\mathcal R^\psi_{\Delta_2,\emptyset}$ restricted to $G_p$ gives maps $$\mathcal R^{\psi}_p=\mathcal O[[X,Y]]/(XY)\to\mathcal R^\psi_{\Delta_2,\emptyset}.$$ The image $x_p$ of $X$ and the ideal $(y_p)$ generated by the image $y_p$ of $Y$ in $\mathcal R^\psi_{\Delta_2,\emptyset}$ do not depend on the choice of the map. By an abuse of notation, we shall call $x_p,y_p$ also the image of $x_p,y_p$ in every quotient of $\mathcal R^\psi_{\Delta_2,\emptyset}$. If $h$ is a form in $\mathcal B_{\Delta_2,\emptyset}$, we denote by $x_p(h),y_p(h)\in\mathcal O$ the images of $x_p,y_p$ by the map $\mathcal R^\psi_{\Delta_2,\emptyset}\to\mathcal O$ corresponding to $\rho_h$.\\ \begin{lemma}
If $h\in\mathcal B_{\Delta_2}$ and $p|\Delta_2$, then: \begin{itemize} \item[a)] $x_p(h)=0$ if and only if $h$ is special at $p$; \item[b)] if $h$ is unramified at $p$ then $(x_p(h))=(a_p(h)^2-\psi(p)(p+1)^2)$; \item[c)] $y_p(h)=0$ if and only if $h$ is unramified at $p$; \item[d)] if $h$ is special at $p$, the oreder at $(\lambda)$ of $y_p(h)$ is the greatest positive integer $n$ such that $\rho_h(I_p)\equiv\sqrt{\psi(p)}\otimes 1\ {\rm mod}\ \lambda^n.$ \end{itemize} \end{lemma}
\begin{proof} It is an immediate consequence of the definition of the sp-condition (proof of lemma \ref{versal}). Statement b), follows from the fact that $$a_p(h)={\rm tr}(\rho_h({\rm Frob}_p))$$ and $$\rho_h({\rm Frob}_h)=\left( \begin{array} [c]{cc} \pm p\sqrt{\psi(p)}+x_p(h) & 0\\ 0 & p\psi(p)/(\pm p\sqrt{\psi(p)}+x_p(h)) \end{array} \right).$$ \end{proof}
\noindent In particular $(y_p)$ is the kernel of the map $\mathcal R^\psi_{S_1,S_2}\to\mathcal R^\psi_{S_1/p,S_2}$.
\noindent If $h\in\mathcal B_{S_1,S_2}$ let $\theta_{h,S_1,S_2}:{\bf T}^\psi_{S_1,S_2}\to\mathcal O$ be the character corresponding to $h$.
\noindent We consider the congruence ideal of $h$ relatively to $\mathcal B_{S_1,S_2}$: $$\eta_{h,S_1,S_2}=\theta_{h,S_1,S_2}(Ann_{{\bf T}^\psi_{S_1,S_2}}({\rm ker}\ \theta_{h,S_1,S_2})).$$
\noindent It is know that $\eta_{h,S_1,S_2}$ controls congruences between $h$ and linear combinations of forms different from $h$ in $\mathcal B_{S_1,S_2}$.
\begin{theorem}\label{cong} Suppose $\Delta_1\not=1$ and $\Delta_2$ as above. Then \begin{itemize} \item[a)] $\mathcal B_{\emptyset,\Delta_2}\not=0$; \item[b)] for every subset $S\subseteq\Delta_2$, the map $\mathcal R^\psi_{S,\Delta_2/S}\to{\bf T}^\psi_{S,\Delta_2/S}$ is an isomorphism of complete intersection;
\item[c)] for every $h\in\mathcal B_{\emptyset,\Delta_2}$, $\eta_{h,S,\Delta_2/S}=(\prod_{p|S}y_p(h))\eta_{h,\emptyset,\Delta_2}.$ \end{itemize} \end{theorem}
\noindent The proof of this theorem is essentilly the same as in \cite{Lea}.\\ If we combine point c) of theorem \ref{cong} to the results in Section 5.5 of \cite{CDT}, we obtain: \begin{corollary} If $h\in\mathcal B_{S_1,S_2},$ then \begin{displaymath}
\eta_{h,\Delta_2,\emptyset}=\prod_{p|\frac{\Delta_2}{S_1S_2}}x_p(h)\prod_{p|S_2}y_p(h)\eta_{h,S_1,S_2}.
\end{displaymath} \end{corollary} In particular, from this corollary we deduce the following theorem: \begin{theorem}
Let $f=\sum a_nq^n$ be a normalized newform in $S_2(\Gamma_0(M\ell^2),\psi)$ supercuspidal of type $\tau=\chi\oplus\chi^\sigma$ at $\ell$, with minimal level, special at primes in a finite set $\Delta'$, there exist $g\in S_2(\Gamma_0(qM\ell^2),\psi)$ supercuspidal of type $\tau$ at $\ell$, special at every prime $p|\Delta'$ such that $f\equiv g\ {\rm mod}\ \lambda$ if and only if $$a_q^2\equiv\psi(q)(1+q)^2\ {\rm mod}\ \lambda$$ where $q$ is a prime such that $(q,M\ell^2)=1,$ $q\not\equiv-1\ {\rm mod}\ \ell$. \end{theorem}
\noindent The problem of remuve the hypothesis of minimal level, is still open and could be solved by proving conjecture \ref{noscon}
\section{Problem: extension of results to the non minimal case}
Let $\ell\geq 2$ be a prime number. Let $\Delta'$ be a product of an odd number of primes, different from $\ell$. We put $\Delta=\Delta'\ell$. Let $B$ be the indefinite quaternion algebra over ${\bf Q}$ of discriminant $\Delta$. Let $R$ be a maximal order in $B$.\\
Let $N$ be a positive rational number, we observe that in our deformation problem, in section \ref{de}, we have assumed that the representation $\overline\rho$ associated to $f\in S_2(\Gamma_0(N\Delta\ell),\psi)$, occurs with type $\tau$ and minimal level at $N$ ( not necessarily at $\Delta'$). Let now $S$ be a finite set of rational primes which does not divide $M\ell$, where $M=N\Delta'$, we fix $f\in S(\Gamma_0(N\Delta\ell S),\psi)$ a modular newform of weight 2, level $N\Delta\ell S$, supercuspidal of type $\tau$ at $\ell$, special at primes dividing $\Delta'$ and with Nebentypus $\psi$. Let $\rho$ be the Galois representation associated to $f$ and let $\overline\rho$ be its reduction modulo $\ell$; we suppose that conditions (\ref{con1}), (\ref{cond2}), (\ref{rara}), (\ref{end}) and (\ref{c3}) hold.
\noindent If we denote by $\Delta_1$ the product of primes $p|\Delta'$ such that $\overline\rho(I_p)\not=1$ and by $\Delta_2$ the product of primes $p|\Delta'$ such that $\overline\rho(I_p)=1$. As usual we assume that $p^2\not\equiv 1\ {\rm mod}\ \ell$ if $p|\Delta_2$. We say that the representation $\rho$ is of {\bf type $(S,{\rm sp},\tau,\psi)$} if conditions b), c), d), e) of definition \ref{def} hold and \begin{itemize} \item[a)] $\rho$ is unramified outside $M\ell S$. \end{itemize}
\noindent This is a deformation condition. Let $\mathcal R^\psi_S$ be the universal deformation ring which parametrizes representations of type $(S,{\rm sp},\tau,\psi)$ with residual representation $\overline\rho$ and let ${\bf T}^\psi_S$ be the Hecke algebra acting on the space of modular forms of type $(S,{\rm sp},\tau,\psi)$.\\ Since the dimension of the Selmer groups do not satisfy the control conditions, it is not possible to construct a Taylor-Wiles system by considering a deformation problem of type $(S,{\rm sp},\tau,\psi)$ with $S\not=\emptyset$, as in section \ref{tw}; a still open problem is to prove theorem \ref{goal} for $\mathcal R^\psi_S$, ${\bf T}^\psi_S$ and ${\mathcal M^\psi}_S$. If $S=0$ then we retrouve our theorem \ref{goal}. A possible approach to this problem is to use, as in the classical case, the results of De Smit, Rubin, Schoof \cite{DRS} and Diamond \cite{Dia}, by induction on the cardinality of $S$. If we make the inductive hypothesis assuming the result true for $S$, we have to verify that the result holds for $S'=S\cup\{q\}$ where $q$ is a prime number not dividing $M\ell S$. Following the litterature, to prove the inductive step the principal ingredients are: \begin{itemize} \item a duality result about $\mathcal M^\psi_S$ (to appeare), showing the existence of a perfect pairing $\mathcal M^\psi_S\times \mathcal M^\psi_S\to\mathcal O$ which induces an isomorphism $\mathcal M^\psi_S\to{\rm Hom}_\mathcal O(\mathcal M^\psi_S,\mathcal O)$ of ${\bf T}^\psi_S$-modules; \item conjecture \ref{noscon}, saying that the natural ${\bf T}^\psi_{S'}$-injection $\left(\mathcal M^\psi_S\right)^2\hookrightarrow \mathcal M^\psi_{S'}$ is injective when tensorized with $k$. \end{itemize} For semplicity, we shall assume that $q^2\not\equiv 1\ {\rm mod}\ \ell$ for every $q\in S$.
\noindent Then, as observed in the previous section, there is an isomorphism: $$\frac{\mathcal R^\psi_{S'}}{(y_q)}\ \widetilde\to\ \mathcal R^\psi_{S}$$ that induces an isomorphism on the Hecke algebras: $$\frac{{\bf T}^\psi_{S'}}{(y_q)_{{\bf T}^\psi_{S'}}}\ \widetilde\to\ {\bf T}^\psi_{S}$$ where $(y_q)_{{\bf T}^\psi_{S'}}$ is the image of $(y_q)$ in ${\bf T}^\psi_{S'}$; but, for a general $S$, we don't have any information about $\mathcal M^\psi_S=H^1({\bf X}_1(NS),\mathcal O)_{\mathfrak m_S}^{\widehat\psi}$ as a ${\bf T}_S^{\widehat\psi}$-module, where $\mathfrak m_S$ is the inverse image of $\mathfrak m$ with respect to the natural map ${\bf T}_S^{\widehat\psi}\to {\bf T}_\emptyset^{\widehat\psi}$.
\end{document}
|
arXiv
|
{
"id": "0710.4677.tex",
"language_detection_score": 0.6813715100288391,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Nonexistence of tight spherical design of harmonic index $4$} \author{Takayuki Okuda and Wei-Hsuan Yu} \subjclass[2010]{Primary 52C35; Secondary 14N20, 90C22, 90C05}
\keywords{spherical design, equiangular lines} \address{ Department of Mathematics, Graduate School of Science, Hiroshima University 1-3-1 Kagamiyama, Higashi-Hiroshima, 739-8526 Japan} \email{[email protected]} \address{Department of Mathematics, Michigan State University, 619 Red Cedar Road, East Lansing, MI 48824} \email{[email protected]} \thanks{The first author is supported by Grant-in-Aid for JSPS Fellow No.25-6095 and the second author is supported in part by NSF grants CCF1217894, DMS1101697} \date{} \maketitle
\begin{abstract} We give a new upper bound of the cardinality of a set of equiangular lines in $\mathbb{R}^n$ with a fixed angle $\theta$ for each $(n,\theta)$ satisfying certain conditions. Our techniques are based on semi-definite programming methods for spherical codes introduced by Bachoc--Vallentin [J.~Amer.~Math.~Soc.~2008]. As a corollary to our bound, we show the nonexistence of spherical tight designs of harmonic index $4$ on $S^{n-1}$ with $n \geq 3$. \end{abstract}
\section{Introduction}
The purpose of this paper is to give a new upper bound of the cardinality of a set of equiangular lines with certain angles (see Theorem \ref{thm:rel}). As a corollary to our bound, we show the nonexistence of tight designs of harmonic index $4$ on $S^{n-1}$ with $n \geq 3$ (see Theorem \ref{thm:nonex-tight}).
Throughout this paper,
$S^{n-1} := \{ x \in \mathbb{R}^n \mid \| x \| = 1 \}$ denotes the unit sphere in $\mathbb{R}^{n}$. By Bannai--Okuda--Tagami \cite{ban13}, a finite subset $X$ of $S^{n-1}$ is called \emph{a spherical design of harmonic index $t$ on $S^{n-1}$} (or shortly, \emph{a harmonic index $t$-design on $S^{n-1}$}) if $\sum_{{\bold x} \in X} f({\bold x}) = 0$ for any harmonic polynomial function $f$ on $\mathbb{R}^{n}$ of degree $t$.
Our concern in this paper is in tight harmonic index $4$-designs. A harmonic index $t$-design $X$ is said to be \emph{tight} if $X$ attains the lower bound given by \cite[Theorem 1.2]{ban13}. In particular, for $t = 4$, a harmonic index $4$-design on $S^{n-1}$ is tight if its cardinality is $(n+1)(n+2)/6$. For the cases where $n=2$, we can construct tight harmonic index $4$-designs as two points ${\bold x}$ and ${\bold y}$ on $S^{1}$ with the inner-product $\langle {\bold x},{\bold y} \rangle_{\mathbb{R}^2} = \pm \sqrt{1/2}$.
The paper \cite[Theorem 4.2]{ban13} showed that if tight harmonic index $4$-designs on $S^{n-1}$ exist, then $n$ must be $2$ or $3(2k-1)^2 -4$ for some integers $k \geq 3$. As a main result of this paper, we show that the later cases do not occur. That is, the following theorem holds: \begin{thm}\label{thm:nonex-tight} For each $n \geq 3$, spherical tight design of harmonic index $4$ on $S^{n-1}$ does not exist. \end{thm}
A set of lines in $\mathbb{R}^n$ is called \emph{an equiangular line system} if the angle between each pair of lines is constant. By definition, an equiangular line system can be considered as a spherical two-distance set with the inner product set $\{ \pm \cos \theta \}$ for some constant $\theta$. Such the constant $\theta$ is called \emph{the common angle} of the equiangular line system. The recent development of this topic can be found in \cite{barg14, grea14}.
By \cite[Proposition 4.2]{ban13}, any tight harmonic index $4$-design on $S^{n-1}$ can be considered as an equiangular line system with the common angle $\arccos \sqrt{3/(n+4)}$. The proof of Theorem \ref{thm:nonex-tight} will be reduced to a new relative upper bound (see Theorem \ref{thm:rel}) for the cardinalities of equiangular line systems with a fixed common angle. Note that in some cases, our relative bound is better than the Lemmens--Seidel relative bound (see Section \ref{sec:main} for more details).
The paper is organized as follows: In Section \ref{sec:main}, as a main theorem of this paper, we give a new relative bound for the cardinalities of equiangular line systems with a fixed common angle satisfying certain conditions. Theorem \ref{thm:nonex-tight} is followed as a corollary to our relative bound. In Section \ref{sec:proof}, our relative bound is proved based on the method by Bachoc--Vallentin \cite{bac08a}.
\section{Main results}\label{sec:main}
In this paper, we denote by $M(n)$ and $M_{\cos \theta}(n)$ the maximum number of equiangular lines in $\mathbb{R}^n$ and that with the fixed common angle $\theta$, respectively. By definition, \[ M(n) = \sup_{0 \leq \alpha < 1} M_\alpha(n). \] The important problems for equiangular lines are to give upper and lower estimates $M(n)$ or $M_\alpha(n)$ for fixed $\alpha$. One can find a summary of recent progress of this topic in \cite{barg14, grea14}.
Let us fix $0 \leq \alpha < 1$. Then for a finite subset $X$ of $S^{n-1}$ with $I(X) \subset \{ \pm \alpha \}$, we can easily find an equiangular line system with the common angle $\arccos \alpha$
and the cardinality $|X|$, where \[ I(X) := \{ \langle {\bold x},{\bold y} \rangle_{\mathbb{R}^n} \mid {\bold x},{\bold y} \in X \text{ with } {\bold x} \neq {\bold y} \} \] is the set of inner-product values of distinct vectors in $X \subset S^{n-1} \subset \mathbb{R}^n$. The converse is also true. In particular, we have \begin{align*}
M_{\alpha}(n) = \max \{ |X| \mid X \subset S^{n-1} \text{ with } I(X) \subset \{ \pm \alpha \} \}, \end{align*} and therefore, our problem can be considered as a problem in special kinds of spherical two-distance sets.
In this paper, we are interested in upper estimates of $M_\alpha(n)$. According to \cite{lem73}, Gerzon gave the upper bound on $M(n)$ as $M(n) \leq n(n+1)/2$ and therefore, we have \[ M_\alpha(n) \leq \frac{n(n+1)}{2} \quad \text{ for any } \alpha. \] This upper bound is called the Gerzon absolute bound. Lemmens and Seidel \cite{lem73} showed that
\begin{equation*}
M_\alpha(n) \le \frac {n(1-\alpha^2)}{1-n\alpha^2} \quad \text{ in the cases where } 1-n\alpha^2 > 0.
\end{equation*} This inequality is sometimes called the Lemmens--Seidel
relative bound as opposed to the Gerzon absolute bound.
As a main theorem of this paper, we give other upper estimates of $M_\alpha(n)$ as follows:
\begin{thm}\label{thm:rel} Let us take $n \geq 3$ and $\alpha \in (0,1)$ with \[ 2-\frac{6\alpha-3}{\alpha^2} < n < 2 + \frac{6\alpha+3}{\alpha^2}. \] Then \[ M_\alpha(n) \leq 2 + \frac{(n-2)}{\alpha} \max \left\{ \frac{(1-\alpha)^3}{(n-2)\alpha^2 +6\alpha-3}, \frac{(1+\alpha)^3}{-(n-2)\alpha^2 + 6\alpha+3} \right\}. \] In particular, for an integer $l \geq 2$, if \[ 3l^2-6l+2 < n < 3l^2 + 6l+2 \] then \[ M_{1/l}(n) \leq 2 + (n-2) \max \left\{ \frac{(l-1)^3}{-3l^2 + 6l + (n-2)}, \frac{(l+1)^3}{3l^2 + 6l -(n-2)} \right\}. \] \end{thm}
Recall that by \cite[Proposition 4.2, Theorem 4.2]{ban13} for $n \geq 3$, if there exists a tight harmonic index $4$-design $X$ on $S^{n-1}$, then $n = 3(2k-1)^2-4$ for some $k \geq 3$ and \begin{align*} M_{\sqrt{3/(n+4)}}(n) = (n+1)(n+2)/6. \end{align*}
However, as a corollary to Theorem \ref{thm:rel}, we have the following upper bound of $M_{\sqrt{3/(n+4)}}(n)$
and obtain Theorem \ref{thm:nonex-tight}.
\begin{cor}\label{cor:tight4-eq} Let us put $n_k := 3(2k-1)^2-4$ and $\alpha_k := \sqrt{3/(n_k+4)} = 1/(2k-1)$. Then for each integer $k \geq 2$, \begin{align*} M_{\alpha_k}(n_k) &\leq 2 (k-1) (4k^3-k-1) (< (n_k+1)(n_k+2)/6). \end{align*} \end{cor}
It should be remarked that in the setting of Corollary \ref{cor:tight4-eq}, the Lemmens--Seidel relative bound does not work since \[ 1-n_k \alpha_k^2 = -2(4 k^2-4k-1)/(2k-1)^2 < 0 \] and our bound is better than the Gerzon absolute bound.
The proof of Theorem \ref{thm:rel} is given in Section \ref{sec:proof} based on Bachoc--Vallentin's SDP method for spherical codes \cite{bac08a}. The origins of applications of the linear programming method in coding theory can be traced back to the work of Delsarte \cite{del73}. Applications of semidefinite programming (SDP) method in coding theory and distance geometry gained momentum after the pioneering work of Schrijver \cite{Schrijver05code} that derived SDP bounds on codes in the Hamming and Johnson spaces. Schrijver's approach was based on the so-called Terwilliger algebra of the association scheme. The similar idea for spherical codes were formulated by Bachoc and Vallentin \cite{bac08a} regarding for kissing number problems. Barg and Yu \cite{barg13} modified it to achieve maximum size of spherical two-distance sets in $\mathbb{R}^n$ for most values of $n \leq 93$. In our proof, we restricted the method to obtain upper bounds for equiangular line sets.
We can see in \cite{GST06upperbounds,Schrijver05code} and \cite{bac08a, bac09opti, bac09sdp, Musin08bounds} that the SDP method works well for studying binary codes and spherical codes, respectively. Especially, for equiangular lines, Barg and Yu \cite{barg14} give the best known upper bounds of $M(n)$ for some $n$ with $n \leq 136$ by the SDP method. Our bounds in Corollary \ref{cor:tight4-eq} are the same as \cite{barg14} in lower dimensions. However, in some cases, we need some softwares to complete the SDP method. It should be emphasized that our theorem offer upper bound of $M_{\alpha_k}(n_k)$ for arbitrary large $n_k$ and the proof can be followed by hand calculations without using any convex optimization software.
\section{Proof of our relative bound}\label{sec:proof}
To prove Theorem \ref{thm:rel}, we apply Bachoc--Vallentin's SDP method for spherical codes in \cite{bac08a} to spherical two-distance sets. The explicit statement of it was given by Barg--Yu \cite{barg13}.
We use symbols $P^{n}_l(u)$ and $S^n_l(u,v,t)$ in the sense of \cite{bac08a}. It should be noted that the definition of $S^{n}_l(u,v,t)$ is different from that of \cite{bac09opti} and \cite{barg13} (see also \cite[Remark 3.4]{bac08a} for such the differences).
In order to state it, we define \begin{align*} W(x)&:= \begin{pmatrix}1&0\\0&0\end{pmatrix} + \begin{pmatrix}0&1\\1&1\end{pmatrix} (x_1+x_2)/3 + \begin{pmatrix} 0&0\\0&1 \end{pmatrix} (x_3+x_4+x_5+x_6), \\ S^{n}_l(x;\alpha,\beta) &:= S^{n}_{l}(1,1,1)+S^{n}_l(\alpha,\alpha,1)x_1 + S^{n}_l(\beta,\beta,1) x_2 + S^{n}_l(\alpha,\alpha,\alpha) x_3 \\ & \quad \quad + S^{n}_l(\alpha,\alpha,\beta) x_4 + S^{n}_l(\alpha,\beta,\beta) x_5 + S^{n}_l(\beta,\beta,\beta) x_6 \end{align*} for each $x = (x_1,x_2,x_3,x_4,x_5,x_6) \in \mathbb{R}^6$ and $\alpha, \beta \in [-1,1)$. We remark that $W(x)$ is a symmetric matrix of size $2$ and $S^{n}_{l}(x;\alpha,\beta)$ is a symmetric matrix of infinite size indexed by $\{ (i,j) \mid i,j = 0,1,2,\dots, \}$.
\begin{fact}[Bachoc--Vallentin \cite{bac08a} and Barg--Yu \cite{barg13}]\label{fact:SDP-problem} Let us fix $\alpha, \beta \in [-1,1)$. Then any finite subset $X$ of $S^{n-1}$ with $I(X) \subset \{ \alpha,\beta \}$ satisfies \[
|X| \leq \max \{ 1 + (x_1 + x_2)/3 \mid x = (x_1,\dots,x_6) \in \Omega^{n}_{\alpha,\beta} \} \] where the subset $\Omega^{n}_{\alpha,\beta}$ of $\mathbb{R}^{6}$ is defined by \[ \Omega_{\alpha,\beta}^{n} := \{\, x = (x_1,\dots,x_6) \in \mathbb{R}^{6} \mid \text{ $x$ satisfies the following four conditions } \,\}. \] \begin{enumerate} \item $x_i \geq 0$ for each $i = 1,\dots,6$. \item $W(x)$ is positive semi-definite. \item $3 + P^{n}_l(\alpha) x_1 + P^{n}_l (\beta) x_2 \geq 0$ for each $l=1,2,\dots.$ \item Any finite principal minor of $S^{n}_l(x;\alpha,\beta)$ is positive semi-definite for each $l = 0,1,2,\dots.$ \end{enumerate} \end{fact}
To prove Theorem \ref{thm:rel}, we use the following ``linear version'' of Fact \ref{fact:SDP-problem}:
\begin{cor}\label{cor:triangleLP} In the same setting of Fact $\ref{fact:SDP-problem}$, \[
|X| \leq \max \{ 1 + (x_1 + x_2)/3 \mid x = (x_1,\dots,x_6) \in \widetilde{\Omega}^{n}_{\alpha,\beta} \} \] where the subset $\widetilde{\Omega}^{n}_{\alpha,\beta}$ of $\mathbb{R}^{6}$ is defined by \[ \widetilde{\Omega}_{\alpha,\beta}^{n} := \{\, x = (x_1,\dots,x_6) \in \mathbb{R}^{6} \mid \text{ $x$ satisfies the following three conditions } \,\}. \] \begin{enumerate} \item $x_i \geq 0$ for each $i = 1,\dots,6$. \item $\det W(x) \geq 0$. \item $(S^{n}_l)_{i,i}(x;\alpha,\beta) \geq 0$ for each $l,i = 0,1,2,\dots$, where $(S^{n}_l)_{i,i}(x;\alpha,\beta)$ is the $(i,i)$-entry of the matrix $S^{n}_l(x;\alpha,\beta)$. \end{enumerate} \end{cor}
By Corollary \ref{cor:triangleLP}, the proof of Theorem \ref{thm:rel} can be reduced to show the following proposition:
\begin{prop}\label{prop:SDPrel} Let $n \geq 3$ and $0 < \alpha <1$. Then the following holds: \begin{enumerate} \item \begin{align*} \max \{ 1+(x_1+x_2)/3 \mid x \in \widetilde{\Omega}^{n}_{\alpha,-\alpha} \} \leq 2 + 2 + (n-2)\frac{(1-\alpha)^3}{\alpha((n-2)\alpha^2 +6\alpha-3)} \end{align*} if $(1-\alpha)^3(-(n-2)\alpha^2 + 6\alpha+3) \geq (1+\alpha)^3((n-2)\alpha^2 +6\alpha-3) \geq 0$. \item \begin{align*} \max \{ 1+(x_1+x_2)/3 \mid x \in \widetilde{\Omega}^{n}_{\alpha,-\alpha} \} \leq 2 + (n-2) \frac{(1+\alpha)^3}{\alpha(-(n-2)\alpha^2 + 6\alpha+3)} \end{align*} if $(1+\alpha)^3((n-2)\alpha^2 +6\alpha-3) \geq (1-\alpha)^3(-(n-2)\alpha^2 + 6\alpha+3) \geq 0$. \end{enumerate} \end{prop}
For the proof of Proposition \ref{prop:SDPrel}, we need the next explicit formula of $(S^{n}_3)_{1,1}$ which are obtained by direct computations:
\begin{lem}\label{lem:S3explicite} For each $-1 < \alpha < 1$, \begin{align*} (S^{n}_3)_{1,1}(1,1,1) &= 0 \\ (S^{n}_3)_{1,1}(\alpha,\alpha,1) &= \frac{n(n+2)(n+4)(n+6)}{3(n-1)(n+1)(n+3)} \alpha^2 (1-\alpha^2)^3 \\ (S^{n}_3)_{1,1}(\alpha,\alpha,\alpha) &= -\frac{n(n+2)(n+4)(n+6)}{(n-2)(n-1)(n+1)(n+3)} (\alpha-1)^3 \alpha^3 ((n-2)\alpha^2-6\alpha-3) \\ (S^{n}_3)_{1,1}(\alpha,\alpha,-\alpha) &= -\frac{n(n+2)(n+4)(n+6)}{(n-2)(n-1)(n+1)(n+3)} \alpha^3 (\alpha+1)^3 ((n-2)\alpha^2 +6\alpha-3). \end{align*} \end{lem}
\begin{proof}[Proof of Proposition $\ref{prop:SDPrel}$] Fix $\alpha$ with $0 < \alpha < 1$ and take any $x \in \widetilde{\Omega}^{n}_{\alpha,-\alpha}$. For simplicity we put $X = (x_1+x_2)/3$, $Y = x_3+x_5$ and $Z = x_4+x_6$. By computing $\det W(x)$, we have \begin{align} -X(X-1)+Y+Z \geq 0. \label{eq:Wrel} \end{align} Furthermore, we have $(S^{n}_3)_{1,1}(x;\alpha,-\alpha) \geq 0$, and hence, by Lemma \ref{lem:S3explicite}, \begin{multline*} (n-2) \frac{(1-\alpha^2)^3}{\alpha} X - (1-\alpha)^3 (-(n-2)\alpha^2 + 6\alpha+3) Y \\ - (1+\alpha)^3 ((n-2)\alpha^2 +6\alpha-3) Z \geq 0 \end{multline*} Therefore, in the cases where \[ (1-\alpha)^3(-(n-2)\alpha^2 + 6\alpha+3) \geq (1+\alpha)^3((n-2)\alpha^2 +6\alpha-3) \geq 0, \] we obtain \begin{align*} (n-2) \frac{(1-\alpha^2)^3}{\alpha} X - (1+\alpha)^3 ((n-2)\alpha^2 +6\alpha-3) (Y+Z) \geq 0. \end{align*} By combining with \eqref{eq:Wrel}, \begin{align*} (n-2) \frac{(1-\alpha^2)^3}{\alpha} X - (1+\alpha)^3 ((n-2)\alpha^2 +6\alpha-3) X(X-1) \geq 0 \end{align*} Thus we have \[ 2 + (n-2) \frac{(1-\alpha)^3}{\alpha((n-2)\alpha^2 +6\alpha-3)} \geq X+1 = 1 + (x_1+x_2)/3. \] By the similar arguments, in the cases where \[ (1+\alpha)^3((n-2)\alpha^2 +6\alpha-3) \geq (1-\alpha)^3(-(n-2)\alpha^2 + 6\alpha+3) \geq 0, \] we have \[ 2 + (n-2) \frac{(1+\alpha)^3}{\alpha(-(n-2)\alpha^2 +6\alpha+3)} \geq X+1 = 1 + (x_1+x_2)/3. \] \end{proof}
\begin{rem} Harmonic index $4$-designs are defined by using the functional space $\mathop{\mathrm{Harm}}\nolimits_{4}(S^{n-1})$. Therefore, it seems to be natural to consider $\mathop{\mathrm{Harm}}\nolimits_4(S^{n-1})$ in Bachoc--Vallentin's SDP method. In our proof, the functional space \[ H^{n-1}_{3,4} \subset \bigoplus_{m=0}^{4} H^{n-1}_{m,4} = \mathop{\mathrm{Harm}}\nolimits_{4}(S^{n-1}) \] $($see \cite{bac08a} for the notation of $H^{n-1}_{m,l}$$)$ plays an important role to show the nonexistence of tight designs of harmonic index $4$ since $(S^n_3)_{1,1}$ comes from $H^{n-1}_{3,4}$. We checked that if we consider $H^{n-1}_{0,4} \oplus H^{n-1}_{1,4} \oplus H^{n-1}_{2,4} \oplus H^{n-1}_{4,4}$ instead of $H^{n-1}_{3,4}$, our upper bound can not be obtained for small $k$. However, we can not find any conceptional reason of the importance of $H^{n-1}_{3,4}$. \end{rem}
\section*{Acknowledgements.} The authors would like to give heartfelt thanks to Eiichi Bannai, Alexander Barg and Makoto Tagami whose suggestions and comments were of inestimable value for this paper. The authors also would like to thanks Akihiro Munemasa, Hajime Tanaka and Ferenc Sz{\"o}ll{\H o}si for their valuable comments.
\providecommand{\bysame}{\leavevmode\hbox to3em{\hrulefill}\thinspace} \providecommand{\href}[2]{#2}
\end{document}
|
arXiv
|
{
"id": "1409.6995.tex",
"language_detection_score": 0.6606778502464294,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Multi-Agent Submodular Optimization}
\author{Richard Santiago \and F. Bruce Shepherd}
\institute{Richard Santiago \at
McGill University, Montreal, Canada \\
\email{[email protected]}
\and
F. Bruce Shepherd \at
University of British Columbia, Vancouver, Canada \\
\email{[email protected]} }
\date{Received: date / Accepted: date}
\maketitle
\begin{abstract} Recent years have seen many algorithmic advances in the area of submodular optimization: (SO) $\min/\max~f(S): S \in \mathcal{F}$, where $\mathcal{F}$ is a given family of feasible sets over a ground set $V$ and $f:2^V \rightarrow \mathbb{R}$ is submodular. This progress has been coupled with a wealth of new applications for these models. Our focus is on a more general class of {\em multi-agent submodular optimization} (MASO) $\min/\max \sum_{i=1}^{k} f_i(S_i): S_1 \uplus S_2 \uplus \cdots \uplus S_k \in \mathcal{F}$. Here we use $\uplus$ to denote disjoint union and hence this model is attractive where resources are being allocated across $k$ agents, each with its own submodular cost function $f_i()$. This was introduced in the minimization setting by Goel et al. In this paper we explore the extent to which the approximability of the multi-agent problems are linked to their single-agent versions, referred to informally as the {\em multi-agent gap}.
We present different reductions that transform a multi-agent problem into a single-agent one. For minimization, we show that (MASO) has an $O(\alpha \cdot \min \{k, \log^2 (n)\})$-approximation whenever (SO) admits an $\alpha$-approximation over the convex formulation. In addition, we discuss the class of ``bounded blocker'' families where there is a provably tight $O(\log n)$ multi-agent gap between (MASO) and (SO).
For maximization, we show that monotone (resp. nonmonotone) (MASO) admits an $\alpha (1-1/e)$ (resp. $\alpha \cdot 0.385$) approximation whenever monotone (resp. nonmonotone) (SO) admits an $\alpha$-approximation over the multilinear formulation; and the $1-1/e$ multi-agent gap for monotone objectives is tight. We also discuss several families (such as spanning trees, matroids, and $p$-systems) that have an (optimal) multi-agent gap of 1. These results substantially expand the family of tractable models for submodular maximization.
\keywords{submodular optimization \and approximation algorithms \and multi-agent}
\end{abstract}
\section{Introduction} \label{sec:intro}
A function $f:2^V \to \mathbb{R}$ is \emph{submodular} if $f(S) + f(T) \geq f(S \cup T) + f(S \cap T)$ for any $S,T \subseteq V$. We say that $f$ is \emph{monotone} if $f(S) \leq f(T)$ whenever $S \subseteq T$. Throughout, all submodular functions are nonnegative, and we usually assume that $f(\emptyset)=0$. Our functions are given by a \emph{value oracle}, where for a given set $S$ an algorithm can query the oracle to find its value $f(S)$.
For a family of feasible sets $S\in\mathcal{F}$ on a finite ground set $V$ we consider the following broad class of submodular optimization (SO) problems: \begin{equation} \label{eqn:SA} \mbox{SO($\mathcal{F}$) ~~~~Min~ / ~Max ~$f(S):S\in\mathcal{F}$} \end{equation} where $f$ is a nonnegative submodular set function on $V$. There has been an impressive recent stream of activity around these problems for a variety of set families $\mathcal{F}$. We explore the connections between these (single-agent) problems and their multi-agent incarnations. In the {\em multi-agent (MA)} version, we have $k$ agents each of which has an associated nonnegative submodular set function $f_{i}$, $i \in [k]$. As before, we are looking for sets $S\in\mathcal{F}$, however, we now have a 2-phase task: the elements of $S$ must also be partitioned amongst the agents. Hence we have set variables $S_{i}$ and seek to optimize $\sum_{i}f_{i}(S_{i})$. This leads to the multi-agent submodular optimization (MASO) versions: \begin{equation} \label{eqn:MA} \mbox{MASO($\mathcal{F}$) ~~~~Min~ / ~Max ~$\sum_{i=1}^{k} f_{i}(S_{i}):S_{1}\uplus S_{2}\uplus \cdots\uplus S_{k}\in\mathcal{F}$.} \end{equation}
\iffalse We mention that while we focus on a partition constraint (the notation ``$+$'' indicates disjoint union), in some settings, the problem is of interest even under the standard union constraint. \fi
The special case when $\mathcal{F}=\{V\}$ has been previously examined both for minimization (the minimum submodular cost allocation problem \cite{hayrapetyan2005network,svitkina2010facility,ene2014hardness,chekuri2011submodular}) and maximization (submodular welfare problem \cite{lehmann2001combinatorial,vondrak2008optimal}). For general families $\mathcal{F}$, however, we are only aware of the development in Goel et al. \cite{goel2009approximability} for the minimization setting. A natural first question is whether any multi-agent problem could be directly reduced (or encoded) to a single-agent one over the same ground set $V$. Goel et al. give an explicit example where such a reduction does not exist. More emphatically, they show that when $\mathcal{F}$ consists of vertex covers in a graph, the {\em single-agent (SA)} version (i.e., (\ref{eqn:SA})) has a 2-approximation while the MA version has an inapproximability lower bound of $\Omega(\log n)$.
Our first main objective is to explain the extent to which approximability for multi-agent problems is intrinsically connected to their single-agent versions, which we also refer to as the {\em primitive} associated with $\mathcal{F}$. We refer to the {\em multi-agent (MA) gap} as the approximation-factor loss incurred
by moving to the MA setting.
\iffalse Goel et al. show this is not the case. The following simple submodular-cost problem has a MA cost function which cannot be realized by a single submodular cost function over the same ground set $V$.
There are three tasks $\{A,B,C\}$ and two agents $\{1,2\}$. Each agent can do one of the tasks at cost 1 each. In addition agent 1 can do tasks $\{A,B\}$ at cost of only 1, and $\{A,B,C\}$ at cost of only 2. Agent 2 can do $\{A,C\}$ at cost of only 1, and $\{A,B,C\}$ at cost of only 2. We can associate a cost $c(S)$ for any subset $S \subseteq \{A,B,C\}$ which corresponds to the best allocation of tasks in $S$ to the agents. This function $c(S)$ is not, however, submodular even though each agent's cost function was submodular. To see this, note that the marginal cost of adding $B$ to $\{A\}$ is 0 which is smaller than the marginal cost of adding it to $\{A,C\}$. \fi
Our second objective is to extend the multi-agent model and show that in some cases this larger class remains tractable. Specifically, we define the {\em capacitated multi-agent submodular optimization (CMASO) problem} as follows: \begin{equation} \label{ma} \mbox{CMASO$(\mathcal{F})$}~~~~~~~ \begin{array}{rc} \max / \min & \sum_{i=1}^{k}f_{i}(S_{i}) \\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{F}\\
& S_{i}\in\mathcal{F}_{i}\,,\,\forall i\in[k] \end{array} \end{equation}
\noindent where we are supplied with subfamilies $\mathcal{F}_i$. Many existing applications fit into this framework and some of these can be enriched through the added flexibility of the capacitated model. \iffalse
For instance, one may include set bounds on the variables: $L_{i}\subseteq S_{i}\subseteq U_{i}$
for each $i$, or simple cardinality constraints: $|S_{i}| \leq b_{i}$ for each $i$. A well-studied (\cite{fleischer2006sap,goundan2007revisiting,calinescu2011maximizing}) application of CMASO in the maximization setting is the Separable Assignment Problem (SAP) , which corresponds to the setting where the objectives are taken to be modular, the $\mathcal{F}_i$ downward closed (i.e. hereditary) families, and $\mathcal{F}$ to be the trivial $2^V$. We believe the CMASO framework has considerable potential for future applications and \fi We illustrate this with concrete examples in Section~\ref{sec:applications}.
Prior work in both the single and multi-agent settings is summarized in Section~\ref{sec:related work}. We present our main results next.
\subsection{Our contributions} \label{sec:reduce}
We first discuss the minimization side of MASO (i.e. (\ref{eqn:MA})). Here the work of \cite{ene2014hardness} showed that for general nonnegative submodular functions the problem is in fact inapproximable within any multiplicative factor even in the case where $\mathcal{F}=\{V\}$ and $k=3$ (since it is NP-Hard to decide whether the optimal value is zero).
Hence we focus almost completely on nonnegative monotone submodular objectives $f_i$. In fact, even in the single-agent setting with a nonnegative monotone submodular function $f$, there exist a number of polynomial hardness results over fairly simple set families $\mathcal{F}$; examples include minimizing a submodular function subject to a cardinality constraint \cite{svitkina2011submodular} or over the family of spanning trees \cite{goel2009approximability}.
We show, however, that if the SA primitive for a family $\mathcal{F}$ admits approximation via a natural convex relaxation (see Appendices~\ref{sec:blocking} and \ref{sec:relaxations}) based on the Lov\'asz extension, then we may extend this to its multi-agent version with a modest blow-up in the approximation factor.
\begin{theorem}
\label{thm:klog}
Suppose there is a (polytime) $\alpha(n)$-approximation for monotone SO($\mathcal{F}$) minimization
via the blocking convex relaxation. Then there is a (polytime) $O(\alpha(n) \cdot \min\{k, \log^2 (n)\})$-approximation
for monotone MASO($\mathcal{F}$) minimization. \end{theorem}
We remark that the $O(\log^2 (n))$ approximation loss due to having multiple agents (i.e the MA gap) is in the right ballpark, since the vertex cover problem has a factor $2$-approximation for single-agent and a tight $O(\log n)$-approximation for the MA version \cite{goel2009approximability}.
We also discuss how Goel et al's $O(\log n)$-approximation for MA vertex cover is a special case of a more general phenomenon. Their analysis only relies on the fact that the feasible family (or at least its upwards closure) has a {\em bounded blocker property}. Given a family $\mathcal{F}$, the {\em blocker} $\mathcal{B}(\mathcal{F})$ of $\mathcal{F}$
consists of the minimal sets $B$ such that $B \cap F \neq \emptyset$ for each $F \in \mathcal{F}$. We say that $\mathcal{B}(\mathcal{F})$ is {\em $\beta$-bounded} if $|B| \leq \beta$ for all $B \in \mathcal{B}(\mathcal{F})$.
Families with bounded blockers have been previously studied in the SA minimization setting, where the works \cite{koufogiannakis2013greedy,iyer2014monotone} show that $\beta$-approximations are always available. Our next result (combined with these) establishes an $O(\log n)$ MA gap for bounded blocker families, thus improving the $O(\log^2 (n))$ factor in Theorem \ref{thm:klog} for general families.
We remark that this $O(\log n)$ MA gap is tight due to examples like vertex covers ($2$-approximation for SA and a tight $O(\log n)$-approximation for MA) or submodular facility location ($1$-approximation for SA and a tight $O(\log n)$-approximation for MA).
\begin{theorem} \label{thm:beta} Let $\mathcal{F}$ be a family with a $\beta$-bounded blocker. Then there is a randomized $O(\beta \log n)$-approximation algorithm for monotone $MASO(\mathcal{F})$ minimization.
\end{theorem}
While our work focuses almost completely on monotone objectives, we show in Section \ref{sec:MAfromSA} that upwards closed families with a bounded blocker remain tractable under some special types of nonmonotone objectives introduced by Chekuri and Ene.
\iffalse While our work focuses almost completely on monotone objectives, for the case of upwards closed families with a bounded blocker we can show the following.
\begin{theorem}
Let $\mathcal{F}$ be an upwards closed family with a $\beta$-bounded
blocker. Let the objectives be of the form $f_i = g_i + h$ where each $g_i$ is nonnegative monotone submodular and $h$ is nonnegative submodular. Then there is a randomized $O(k \beta \log n)$-approximation algorithm for the associated multi-agent problem. Moreover, in the case where $h$ is also symmetric there is an $O(\beta \log n)$-approximation algorithm.
\end{theorem} \fi
We conclude our minimization work by discussing a class of families which behaves well for MA minimization despite not having a bounded blocker. More specifically, in Section~\ref{sec:rings} we observe that crossing (and ring) families have an MA gap of $O(\log n)$.
\begin{theorem} There is a tight $\ln (n)$-approximation for monotone MASO($\mathcal{F}$) minimization over crossing families $\mathcal{F}$. \end{theorem}
We now discuss our contributions for the maximization setting. Our main result here establishes that if the SA primitive for a family $\mathcal{F}$ admits approximation via its multilinear relaxation (see Section \ref{sec:max-SA-MA-formulations}), then we may extend this to its multi-agent version with a constant factor loss.
\begin{theorem}
\label{thm:max-MA-gap}
If there is a (polytime) $\alpha(n)$-approximation for monotone SO($\mathcal{F}$) maximization
via its multilinear relaxation, then there is a (polytime) $(1-1/e) \cdot \alpha(n)$-approximation
for monotone MASO($\mathcal{F}$) maximization. Furthermore, given a downwards closed family $\mathcal{F}$,
if there is a (polytime) $\alpha(n)$-approximation
for nonmonotone SO($\mathcal{F}$) maximization
via its multilinear relaxation, then there is a (polytime) $0.385 \cdot \alpha(n)$-approximation
for nonmonotone MASO($\mathcal{F}$) maximization. \end{theorem}
We remark that the $(1-1/e)$ MA gap in the monotone case is tight due to examples like $\mathcal{F}=\{V\}$, where there is a trivial $1$-approximation for the SA problem and a tight $(1-1/e)$-approximation for the MA version \cite{vondrak2008optimal}.
In Section \ref{sec:MASA} we describe a simple generic reduction that shows that for some families an (optimal) MA gap of $1$ holds.
\iffalse In Section \ref{sec:MASA} we also describe a generic reduction that transforms any CMASO($\mathcal{F}$) problem (i.e. any multi-agent problem (\ref{ma})) into a SO($\mathcal{F}'$) problem. We use the simple idea of viewing assignments of elements $v$ to agents $i$ in a {\em multi-agent bipartite graph}. This was first proposed by Lehmann et al in \cite{lehmann2001combinatorial} and also used in Vondrak's work \cite{vondrak2008optimal}. In those cases $\mathcal{F}=\{V\}$ and $\mathcal{F}_i = 2^V$. Here we discuss the impact of the reduction on more general families $\mathcal{F}$ and $\mathcal{F}_i$, where there is a priori no reason that this should be well-behaved. We show how several properties of the objective and certain family of feasible sets stay \emph{invariant} (i.e. preserved) under the reduction. These lead to several algorithmic consequences, the most important ones being for maximization where we show an (optimal) MA gap of 1 for several families. \fi
\begin{theorem}
\label{thm:max-invariance}
Let $\mathcal{F}$ be a matroid, a $p$-matroid intersection, or a $p$-system. Then, if there is a (polytime) $\alpha$-approximation algorithm for monotone (resp. nonmonotone) SO($\mathcal{F}$) maximization,
there is a (polytime) $\alpha$-approximation algorithm for monotone (resp. nonmonotone) MASO($\mathcal{F}$) maximization. \end{theorem}
In the setting of CMASO (i.e. (\ref{ma})) our results provide additional modelling flexibility. They imply that one maintains decent approximations even while adding interesting side constraints.
For instance, for a monotone maximization instance of CMASO where $\mathcal{F}$ corresponds to a $p$-matroid intersection and the $\mathcal{F}_i$ are all matroids, our results from Section \ref{sec:MASA} lead to a $(p+1+\epsilon$)-approximation algorithm. We believe that these, combined with other results from Section \ref{sec:MASA}, substantially expand the family of tractable models for maximization.
While the impact of this reduction is more for maximization, it also has some interesting consequences in the minimization setting. We discuss in Section \ref{sec:reduction-properties} how some of our results help explaining why for the family of spanning trees, perfect matchings, and $st$-paths, the approximations factors revealed in \cite{goel2009approximability} for the monotone minimization problem are the same for both the single-agent and multi-agent versions.
\iffalse We now discuss a generic reduction that transforms any multi-agent problem (\ref{ma}) into a single-agent problem. The reduction has positive algorithmic consequences in the maximization regime. We use the
simple idea of viewing assignments of elements $v$ to agents $i$ in a {\em multi-agent bipartite graph}. This was first proposed by Nisan et al in \cite{lehmann2001combinatorial} and also used in Vondrak's work \cite{vondrak2008optimal}. In those cases $\mathcal{F}=\{V\}$ and the associated bipartite graph is $G([k],V)=([k]+V,E)$
where there is an edge $iv$ for each $i\in[k]$ and each $v\in V$.
Any valid assignment of some
elements in $V$ to agents corresponds to a $b$-matching $Z$ in $G$,
where $b[i]=\infty$ for each $i\in[k]$ and $b(v)=1$ for each $v\in V$.
While the above auxiliary graph is fine when $\mathcal{F}=\{V\}$,
in more general multi-agent~optimization problems we may use a subset $F \subset V$ from some family $\mathcal{F}$. Hence we are seeking $b$-matchings $Z$ for which $cov(Z)\in\mathcal{F}$. Here we define $cov(Z)$ to be the set of nodes $v \in V$ which are saturated by $Z$.
There is no apriori reason that these {\em $\mathcal{F}$-constrained $b$-matchings} should have any nice structure, even for simple $\mathcal{F}$. We show (Section \ref{sec:MASA}) that this is indeed the case if $\mathcal{F}$ is a matroid.
\begin{theorem} \label{thm:mainlift}
If $\mathcal{F}$ induces a matroid over $V$, then so does the family of $\mathcal{F}$-constrained $b$-matchings over the edges of
the multi-agent bipartite graph $G$.
\end{theorem}
Moreover, nice structure is also maintained if we start with a family of matroid bases or an intersection of $p$ matroids (Corollaries~\ref{Corollary p-matroid intersection} and \ref{Corollary bases}). In addition, these reductions also behave well in the more general CMASO model. This, together with the recent advances in SA submodular optimization (Section~\ref{sec:related work}), yields the following algorithmic consequences. \fi
\iffalse
\begin{theorem} \label{thm:max cap} Consider the capacitated multi-agent {\bf maximization} problem (CMASO) \begin{equation*} \begin{array}{rc} \max & \sum_{i \in [k]} f_i (S_i) \\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{F}\\
& S_{i}\in\mathcal{F}_{i}\,,\,\forall i\in[k] \end{array} \end{equation*} where the $f_i$'s are nonnegative submodular functions and the families $\mathcal{F}_{i}$ are matroids. Then we have the following results.
\begin{enumerate} \item If $\mathcal{F}$ is a {\bf matroid}, then there is a $(\frac{1}{2}-\epsilon)$-approximation for monotone functions and a $(\frac{1}{2+O(1)})$-approximation for nonmonotone functions. Moreover, in the special case where $\mathcal{F}_i=2^V$ for all $i$, there is a $(1-1/e)$-approximation for monotone functions and a $1/e$-approximation for nonmonotone functions.
\item If $\mathcal{F}$ is a {\bf $p$-matroid intersection}, then there is a $(\frac{1}{p+1}-\epsilon)$ and a $(\frac{1}{p+1+O(1)})$-approximation for monotone and nonmonotone functions respectively. Moreover, in the case where $\mathcal{F}_i=2^V$ for all $i$, there is a $(\frac{1}{p}-\epsilon)$-approximation
for monotone functions.
\item If $\mathcal{F}$ is the set of {\bf matroid bases} and $\mathcal{F}_i=2^V$ for all $i$, then there is a $(1-1/e)$-approximation for monotone functions and a $\frac{1}{2}(1-\frac{1}{\nu})$-approximation for nonmonotone functions where $\nu$ denotes the fractional base packing number. \end{enumerate} \end{theorem}
The reduction also applies to a model for robust submodular maximization
\cite{orlin2016robust}, where robustness is with respect to the adversarial removal of up to $\tau$ elements (see Section~\ref{sec:lifting-reduction}). This leads to the following algorithmic consequences.
\begin{theorem}
\label{thm:max robust}
Consider the multi-agent {\bf robust maximization} problem
\begin{equation*}
\label{eq:robust}
\begin{array}{ccc}
\max & \min & \sum_{i \in [k]} f_i (S_i - A_i), \\
S_1 \uplus \cdots \uplus S_k \in \mathcal{F} & A_i \subseteq S_i & \\
S_i \in \mathcal{F}_i & \sum_i |A_i|\leq \tau &
\end{array}
\end{equation*}
where the $f_i$'s are nonnegative monotone submodular functions and the families $\mathcal{F}_i$ are matroids. Then we have the following.
\begin{enumerate}
\item There is a
$(\frac{1}{\tau+1})(\frac{1}{p+1}-\epsilon)$-approximation for any fixed $\epsilon>0$ in the case where $\mathcal{F}$ is a $p$-matroid intersection.
\item In the special case where $\mathcal{F}_i = 2^V$ for all $i$, there is a $(\frac{1-1/e}{\tau+1})$-approximation in the case where $\mathcal{F}$ is a matroid, and a $(\frac{1}{\tau+1})(\frac{1}{p}-\epsilon)$-approximation for any fixed $\epsilon>0$ in the case where $\mathcal{F}$ is a $p$-matroid intersection.
\end{enumerate} \end{theorem}
\fi
\subsection{Some applications of (capacitated) multi-agent optimization} \label{sec:applications}
In this section we present several problems in the literature which are special cases of Problem (\ref{eqn:MA}) and the more general Problem (\ref{ma}). We also indicate how the extra generality of {\sc CMASO} (i.e. (\ref{ma})) gives modelling advantages. We start with the \underline{maximization} setting.
\begin{example}[The Submodular Welfare Problem]
~ The most basic problem in the maximization setting arises when
we take the feasible space $\mathcal{F}=\{V\}$. This describes a well-known model (introduced in \cite{lehmann2001combinatorial}) for allocating goods to agents, each of which
has a monotone submodular valuation (utility) function over baskets of goods. This is formulated as (\ref{eqn:MA}) by considering
nonnegative monotone functions $f_{i}$ and $\mathcal{F}=\{V\}$.
The CMASO framework allows us to incorporate additional constraints into this problem by defining the families $\mathcal{F}_{i}$ appropriately. For instance, one can impose cardinality constraints on the number of elements that an agent can take, or to only allow agent $i$ to take a set $S_i$ of elements satisfying some bounds $L_i \subseteq S_i \subseteq U_i$. \end{example}
\begin{example}[The Separable Assignment Problem]
~ An instance of the Separable Assignment Problem (SAP) consists of $m$ items and $n$ bins. Each bin $j$ has an associated downwards closed collection of feasible sets $\mathcal{F}_j$, and a modular function $v_{j}(i)$ that denotes the value of placing item $i$ in bin $j$. The goal is to choose disjoint feasible sets $S_j \in \mathcal{F}_j$ so as to maximize $\sum_{j=1}^n v_j(S_j)$. This well-studied problem (\cite{fleischer2006sap,goundan2007revisiting,calinescu2011maximizing}) corresponds to a CMASO instance where all the objectives are modular, $\mathcal{F}=2^V$, and the families $\mathcal{F}_i$ are downwards closed.
\end{example}
We next discuss an example where using matroid-capacity constraints $\mathcal{F}_i$ in CMASO is beneficial.
\begin{example}[Recommendation Systems and Matroid Constraints] ~ This is a widely deployed class of problems that entails
the targeting of product ads to a mass of (largely unknown) buyers or ``channels''. In \cite{chen2016conflict} a ``meta'' problem
is considered where (known) prospective buyers are recommended to interested sellers. This type of recommendation system incurs additional constraints such as
(i) bounds on the size of the buyer list provided to each seller (e.g., constrained by a seller's budget) and
(ii) bounds on how often a buyer appears on a list (to not bombard buyers). These constraints are modelled as
a ``$b$-matching'' problem in a bipartite buyer-seller graph $G_B$. They also consider a more sophisticated model which incorporates ``conflict-aware'' constraints
on the buyer list for each seller, e.g., no more than one buyer from a household should be recommended to a seller. They model conflicts using extra edges amongst the buyer nodes and they specify
an upper bound on the number of allowed conflict edges induced by a seller's recommendation list. Heuristics for this (linear-objective) model \cite{chen2016conflict} are successfully developed on Ebay data, even though
the computational problem is shown to be NP-hard. In fact, subsequent work \cite{chen2016group} shows that conflict-aware $b$-matching suffers an inapproximability bound of $O(n^{1-\epsilon})$.
We now propose an alternative model which admits an $O(1)$-approximation. Moreover, we allow a more general submodular multi-agent objective $\sum_i f_i(B_i)$ where $B_i$ are the buyers recommended to seller $i$.
To formulate this in the CMASO model (\ref{ma}) we consider the same complete buyer-seller bipartite graph from previous work. We now represent a buyer list $B_i$ as a set of edges $S_i$. In order that each buyer $v$ is not recommended more than its allowed maximum $b(v)$, we add the constraint that the number of edges in $F=\cup S_i$ which are incident to buyer node $v$ is at most $b(v)$. The family $\mathcal{F}$ of such sets $F$ forms a partition matroid. Hence the problem can be formulated as:
\[
\begin{array}{cc}
\max & \sum_{i=1}^{k}f_{i}(S_{i})\\
\mbox{s.t.} & S_1 \uplus S_2 \cdots \uplus S_k \in \mathcal{F} \\
& S_i \in \mathcal{F}_i, \;\; \forall i \in [k]
\end{array}
\]
We now define $\mathcal{F}_i$ to enforce conflict constraints for the seller as follows. Let $V_i$ denote the edges from seller node $i$ to allowable buyers for $i$ (possibly $V_i$ is all buyers).
We may then partition $V_i$ into ``households'' $V_{ij}$. In order to model conflicts, we insist that $S_i$ is allowed to include at most $1$ element from each $V_{ij}$.
The resulting family $\mathcal{F}_i$ is a partition or laminar matroid.
Our results imply that this new version has a polytime $O(1)$-approximation (in the value oracle model).
\iffalse
FBS. Seems fine as is.
\textcolor{red}{I dont like that I switched to non-disjoint union for the first matroid M. We can avoid this but only by making copies of the buyer nodes which I also dont like. Thoughts?} \textcolor{green}{Taking copies of the nodes seem like the cleanest option in my opinion. However, (and we should point this out when working with upwards-closed families in section 5 and 6), notice that when working with monotone function the disjoint union constraint can be relaxed to a regular union constraint. Hence, even though we are working with normal unions and upwards-closed families, we are still solving the original problem with disjoint union as stated in MSFO or the CMASO framework}
\fi \end{example}
\begin{example}[Sensor Placement] ~ The problem of placing sensors and information gathering has been popular in the submodularity literature \cite{krause2007near,krause1973optimal,krause08efficient}. We are given a set of sensors $V$ and a set of possible locations $\{1,2,\ldots,k\}$ where the sensors can be placed. There is also a budget constraint restricting the number of sensors that can be deployed. The goal is to place sensors at some of the locations so as to maximize the total ``informativeness''. Consider a multi-agent objective function $\sum_{i \in [k]} f_i(S_i)$, where $f_i(S_i)$ measures the informativeness of placing sensors $S_i$ at location $i$. It is then natural to consider a diminishing return (i.e. submodularity) property for the $f_i$'s. We can then formulate the problem as MASO($\mathcal{F}$)
where $\mathcal{F}:=\{ S \subseteq V: |S| \leq b \}$ imposes the budget constraint. We can also use CMASO for additional modelling flexibility. For instance, we may define $\mathcal{F}_i=\{ S\subseteq V_i: |S| \leq b_i \}$ where $V_i$ are the allowed sensors for location $i$ and $b_i$ is an upper bound on the sensors located there.
\iffalse We can then formulate the problem as a CMASO instance where $\mathcal{F}:=\{ S \subseteq V: |S| \leq b \}$ imposes the budget constraint and $\mathcal{F}_i$ gives additional modelling flexibility. For instance, we may define $\mathcal{F}_i=\{ S \subseteq V: S\subseteq V_i, |S| \leq b_i \}$, where $V_i$ are the allowed sensors for location $i$ and $b_i$ is an upper bound on the sensors located there. \fi \end{example}
\iffalse \begin{example}[Public Supply Chain] Consider the problem of transporting resources from the hinterland to external markets served by a single port. As prices fluctuate the underlying value of resources changes and this motivates the expansion of public or private networks. These problems have a large number of stakeholders such as port authority, multiple train and trucking carriers, labor unions and overseeing government transportation bodies. It is not uncommon for one good (say coal) to be transported by several distinct organizations and even more than one firm involved in the same mode of transport (say rail). The design and expansion of such networks therefore represent examples of strong multi-agent behaviour. Moreover, feasible sets based on matroid constraints lend themselves to very natural transportation structures such as arborescences (which are a special case of intersection of two matroids). \iffalse \textcolor{red}{This is a bit vague. Can we come up with a truly formal definition? It seems complicated since a carrier which builds a link should both pay for construction and benefit from carrying resources over the link. I always wanted to think about this to make a nice detailed model. Should be doable with some time.} \fi \end{example} \fi
We now discuss Problem (\ref{eqn:MA}) and (\ref{ma}) in the \underline{minimization} setting.
\begin{example}[Minimum Submodular Cost Allocation] \label{ex:MSCA} ~ The most basic problem in the minimization setting arises when we simply take $\mathcal{F}=\{V\}$. This problem, $\min \sum_{i=1}^{k}f_{i}(S_{i}): S_{1}\uplus \cdots\uplus S_{k}=V$, has been widely considered in the literature for both monotone \cite{svitkina2010facility} and nonmonotone functions \cite{chekuri2011submodular,ene2014hardness}, and is referred to as the {\sc Minimum Submodular Cost Allocation (MSCA) problem}\footnote{Sometimes referred to as submodular procurement auctions.} (introduced in \cite{hayrapetyan2005network,svitkina2010facility} and further developed in \cite{chekuri2011submodular}). This is formulated as (\ref{eqn:MA}) by taking $\mathcal{F}=\{V\}$. The CMASO framework allows us to incorporate additional constraints into this problem. The most natural are to impose cardinality constraints on the number of elements that an agent can take, or to only allow agent $i$ to take a set $S_i$ of elements satisfying some bounds $L_i \subseteq S_i \subseteq U_i$. \end{example}
\begin{example}[Multi-agent Minimization] ~ Goel et al \cite{goel2009approximability} consider the special cases of MASO($\mathcal{F}$) where the objectives are nonnegative monotone submodular and $\mathcal{F}$ is either the family of vertex covers, spanning trees, perfect matchings, or shortest $st$ paths. \end{example}
\iffalse \textcolor{red}{THIS NEXT EXAMPLE SHOULD BE TURNED INTO (OR REPLACED BY) A MULTI-AGENT ONE}
IS THERE A GOOD EXAMPLE FOR CROSSING FAMILIES?
\begin{example}[Rings] It is known \cite{schrijver2000combinatorial} that arbitrary submodular functions can be minimized efficiently on a ring family, i.e. a family of sets closed under unions and intersections (see \cite{orlin2009faster} for faster implementations, and work on modified ring families \cite{goemans2010symmetric}). A natural extension is to minimization of multivariate submodular functions over a \emph{multivariate ring family, }where by the latter we mean a family of tuples closed under (component wise) unions and intersections. We provide the formal definitions in Section \ref{sec:multi-ring} and show that this more general problem can still be solved efficiently by applying the reduction from Section \ref{sec:MASA}. \end{example} \fi
\subsection{Related work} \label{sec:related work}
{\bf Single Agent Optimization.} The high level view of the tractability status for unconstrained (i.e., $\mathcal{F}=2^V$) submodular optimization is that both maximization and minimization generally behave well. Minimizing a submodular set function is a classical combinatorial optimization problem which can be solved in polytime \cite{grotschel2012geometric,schrijver2000combinatorial,iwata2001combinatorial}.
Unconstrained maximization on the other hand is known to be inapproximable for general submodular set functions
but admits a polytime constant-factor approximation algorithm when $f$ is nonnegative \cite{buchbinder2015tight,feige2011maximizing}.
In the constrained maximization setting, the classical work \cite{nemhauser1978analysis,nemhauser1978best,fisher1978analysis} already established an optimal $(1-1/e)$-approximation factor for maximizing a nonnegative monotone submodular function subject to a cardinality constraint, and a $(1/(k+1))$-approximation for maximizing a nonnegative monotone submodular function subject to $k$ matroid constraints. This approximation is almost tight in the sense that there is an (almost matching) factor $\Omega(\log(k)/k)$ inapproximability result \cite{hazan2006complexity}. \iffalse in classical work of Fisher, Nemhauser, and Wolsey \cite{nemhauser1978analysis,nemhauser1978best,fisher1978analysis} it is shown that the greedy algorithm achieves an optimal $(1-\frac{1}{e})$-approximation factor for maximizing a monotone submodular function subject to a cardinality constraint, and a $(k+1)$-approximation for maximizing a monotone submodular function subject to $k$ matroid constraints. This approximation is almost tight in the sense that there is an (almost matching) factor $\Omega(k/\log(k))$ inapproximability result. Some of these results have been refined or improved recently. \fi For nonnegative monotone functions, \cite{vondrak2008optimal,calinescu2011maximizing} give
an optimal $(1-1/e)$-approximation based on multilinear extensions when $\mathcal{F}$ is a matroid;
\cite{kulik2009maximizing} provides a $(1-1/e-\epsilon)$-approximation when $\mathcal{F}$ is given by a constant number of knapsack constraints, and \cite{lee2010submodular} gives a local-search algorithm that achieves a $(1/k-\epsilon)$-approximation (for any fixed $\epsilon>0$) when $\mathcal{F}$ is a $k$-matroid intersection. For nonnegative nonmonotone functions, a $0.385$-approximation is the best factor known \cite{buchbinder2016constrained} for maximization under a matroid constraint, in \cite{lee2009non} a $1/(k+O(1))$-approximation is given for $k$ matroid constraints with $k$ fixed. A simple ``multi-greedy'' algorithm \cite{gupta2010constrained} matches the approximation of Lee et al. but is polytime for any $k$. Vondrak \cite{vondrak2013symmetry} gives a $\frac{1}{2}(1-\frac{1}{\nu})$-approximation under a matroid base constraint where $\nu$ denotes the fractional base packing number. Finally, Chekuri et al \cite{vondrak2011submodular} introduce a general framework based on relaxation-and-rounding that allows for combining different types of constraints. This leads, for instance, to $0.38/k$ and $0.19/k$ approximations for maximizing nonnegative submodular monotone and nonmonotone functions respectively under the combination of $k$ matroids and $\ell=O(1)$ knapsacks constraints.
\iffalse Recently, Orlin et al \cite{orlin2016robust} considered a robust formulation for constrained monotone submodular maximization, previously introduced by Krause et al \cite{krause2008robust}. The robustness in this model is with respect to the adversarial removal of up to $\tau$ elements. The problem can be stated as $\max_{S\in \mathcal{F}} \; \min_{A\subseteq S, |A|\leq \tau} \; f(S-A)$, where $\mathcal{F}$ is an independence system. Assuming an $\alpha$-approximation is available for the special case $\tau=0$ (i.e. the non-robust version), \cite{orlin2016robust} gives an $\alpha / (\tau +1)$-approximation for the robust problem. \fi
For constrained minimization, the news is worse \cite{goel2009approximability,svitkina2011submodular,iwata2009submodular}. If $\mathcal{F}$ consists of spanning trees (bases of a graphic matroid) Goel et al \cite{goel2009approximability} show a lower bound
of $\Omega(n)$, while in the case where $\mathcal{F}$ corresponds to the cardinality constraint $\{S:|S|\geq k\}$ Svitkina and Fleischer \cite{svitkina2011submodular} show a lower bound of $\tilde{\Omega}(\sqrt{n})$. There are a few exceptions. The problem can be solved exactly when $\mathcal{F}$ is a ring family (\cite{schrijver2000combinatorial}), triple family (\cite{grotschel2012geometric}), or parity family (\cite{goemans1995minimizing}). In the context of NP-Hard problems, there are almost no cases where good (say $O(1)$ or $O(\log n)$) approximations exist.
We have that the submodular vertex cover admits a $2$-approximation (\cite{goel2009approximability,iwata2009submodular}), and the $k$-uniform hitting set has $O(k)$-approximation.
{\bf Multi-agent Problems.} In the maximization setting the main multi-agent problem studied is the Submodular Welfare Maximization ($\mathcal{F}=\{V\}$)
for which the initial $1/2$-approximation \cite{lehmann2001combinatorial}
was improved to $1-1/e$ by Vondrak \cite{vondrak2008optimal} who introduced the continuous greedy algorithm. This approximation is in fact optimal \cite{khot2005inapproximability,mirrokni2008tight}. We are not aware of maximization work for Problem (\ref{eqn:MA}) for a nontrivial family $\mathcal{F}$.
For the multi-agent minimization setting, MSCA (i.e. $\mathcal{F}=\{V\}$) is the most studied application of Problem (\ref{eqn:MA}). For nonnegative monotone functions, MSCA is equivalent to the Submodular Facility Location problem considered in \cite{svitkina2010facility}, where a tight $O(\log n)$ approximation is given. If the functions $f_{i}$ are nonnegative and nonmonotone, then no multiplicative factor approximation exists \cite{ene2014hardness}. If, however, the functions can be written as $f_{i}=g_{i}+h$ for some nonnegative monotone submodular $g_{i}$ and a nonnegative symmetric submodular function $h$, an $O(\log n)$ approximation is given in \cite{chekuri2011submodular}. In the more general case where $h$ is nonnegative submodular, an $O(k \log n)$ approximation is provided in \cite{ene2014hardness}, and this is tight \cite{mirzakhani2014sperner}.
\iffalse Chekuri and Ene \cite{chekuri2011approximation} consider a variant where all the agents have the same objective function $f$, and moreover, agent $i$ is forced to take some given fixed element $s_i \in V$. More specifically, given a set of elements $\{s_1, s_2, \ldots, s_k \} \subseteq V$ they consider the problem $\min \sum_{i=1}^k f(S_i)$ subject to $s_i \in S_i$ and the partition constraint $S_1 + S_2 + \cdots + S_k = V$. A $2$-approximation and a $(1.5-1/k)$-approximation are given for this problem in the cases where $f$ is nonnegative submodular and nonnegative symmetric submodular respectively. \fi
Goel et al \cite{goel2009approximability} consider the minimization case of (\ref{ma}) for nonnegative monotone submodular functions, in which $\mathcal{F}$ is a nontrivial collection of subsets of $V$ (i.e. $\mathcal{F}\subset2^{V}$) and there is no restriction on the $\mathcal{F}_{i}$ (i.e. $\mathcal{F}_{i}=2^{V}$ for all $i$). In particular, given a graph $G$ they consider the families of vertex covers, spanning trees, perfect matchings, and shortest $st$ paths. They provide a tight $O(\log n)$ approximation for the vertex cover problem, and show polynomial hardness for the other cases. To the best of our knowledge \cite{goel2009approximability} is the only work
on Problem (\ref{eqn:MA}) for nontrivial collections $\mathcal{F}$.
\section{Multi-agent submodular minimization} \label{sec:multimin}
In this section we seek generic reductions for multi-agent minimization problems to their single-agent primitives. We mainly focus on the case of nonnegative monotone submodular objective functions and we work with a natural convex relaxation that is obtained via the Lov\'asz extension of a set function (cf. Appendices \ref{sec:blocking} and \ref{sec:relaxations}). We show that if the SA primitive admits approximation via such relaxation, then we may extend this to its MA version up to an $O(\min\{k, \log^2 (n)\})$ factor loss.
\iffalse Our main result in the minimization setting is the following.
\begin{theorem}
Assume there is a (polytime) $\alpha(n)$-approximation for monotone SO($\mathcal{F}$) minimization based on rounding the blocking convex relaxation. Then there is a (polytime) $O(\alpha(n) \cdot \min\{k, \log (n) \log (\frac{n}{\log n})\})$-approximation for monotone MASO($\mathcal{F}$) minimization. \end{theorem} \fi
As noted already, the $O(\log^2 (n))$ approximation factor loss due to having multiple agents is in the right ballpark since for vertex covers there is a factor $2$-approximation for SA submodular minimization, and a tight $O(\log n)$-approximation for the multi-agent version \cite{goel2009approximability}. In Section \ref{sec:MAfromSA} we discuss an extension of this vertex cover result to a larger class of families with a MA gap of $O(\log n)$.
\subsection{The single-agent and multi-agent formulations} \label{sec:SA-MA-formulations}
Due to monotonicity, one may often assume that we are working with a family $\mathcal{F}$ which is {\em upwards-closed} (sometimes referred to as {\em blocking families}), i.e. if $F \subseteq F'$ and $F \in \mathcal{F}$, then $F' \in \mathcal{F}$. This can be done without loss of generality even if we seek polytime algorithms, since separation over a polytope with vertices $\{\chi^F: F \in \mathcal{F}\}$ implies separation over its dominant. We refer the reader to Appendix~\ref{sec:blocking} for details.
\iffalse Due to monotonicity, one may often assume that we are working with a family $\mathcal{F}$ which is {\em upwards-closed}, aka a {\em blocking family} (cf. \cite{iyer2014monotone}).
The advantage is that to certify whether $F \in \mathcal{F}$, we only need to check that $F \cap B \neq \emptyset$ for each element $B$ of the family $\mathcal{B}(\mathcal{F})$ of minimal blockers of $\mathcal{F}$. We discuss the details in Appendix \ref{sec:blocking}. \fi
For a set function $f:\{0,1\}^V \to \mathbb{R}$ with $f(\emptyset)=0$ one can define its {\em Lov\'asz extension} $f^L:\mathbb{R}_+^V \to \mathbb{R}$ (introduced in \cite{lovasz1983submodular}) as follows. Let $0 < v_1 < v_2 < ... < v_m$ be the distinct positive values taken in some vector $z \in \mathbb{R}_+^V$, and let $v_0=0$. For each $i \in \{0,1,...,m\}$ define the set $S_i:=\{ j: z_j > v_i\}$. In particular, $S_0$ is the support of $z$ and $S_m=\emptyset$. One then defines (see Appendix \ref{sec:LE-def} for equivalent definitions): \[ f^L(z) = \sum_{i=0}^{m-1} (v_{i+1}-v_i) f(S_i). \]
It follows from the definition that $f^L$ is positively homogeneous, that is $f^L(\alpha z)=\alpha f^L(z)$ for any $\alpha > 0$ and $z \in \mathbb{R}_+^V$. Moreover, it is also straightforward to see that $f^L$ is a monotone function if $f$ is.
We have the following result due to Lov\'asz. \begin{lemma}
[Lov\'asz \cite{lovasz1983submodular}]
The function $f^L$ is convex if and only if $f$ is submodular. \end{lemma}
This now gives rise to natural convex relaxations for the single-agent and multi-agent problems (see Appendix \ref{sec:relaxations}) based on some upwards closed relaxation $\{z \geq 0:Az\geq r\}$ of the integral polyhedron $conv(\{\chi^S:S\in \mathcal{F}\})$. In particular, let us denote $P(\mathcal{F}):=\{z \geq 0:Az\geq r\}$, and assume $A$ is a matrix with nonnegative integral entries and $r$ is a vector with positive integral components (if $r_i = 0$ then the $ith$ constraint is always satisfied and we can remove it). For simplicity, we also assume that the entries of $A$ are polynomially bounded in $n$.
The {\em single-agent Lov\'asz extension formulation} (used in \cite{iwata2009submodular,iyer2014monotone}) is: \begin{equation} \label{SA-LE} (\mbox{SA-LE}) ~~~~ \min f^L(z): z \in P(\mathcal{F}), \end{equation} and the {\em multi-agent Lov\'asz extension formulation} (used in \cite{chekuri2011submodular} for $\mathcal{F}=\{V\}$) is: \begin{equation} \label{MA-LE} (\mbox{MA-LE}) ~~~~ \min \sum_{i=1}^k f^L_i(z_i): z_1 + z_2 + \cdots + z_k \in P(\mathcal{F}). \end{equation}
\iffalse The relaxation (SA-LE) has been already considered (\cite{iwata2009submodular,iyer2014monotone}) for different types of families $\mathcal{F}$, while we are only aware of (MA-LE) being used (\cite{chekuri2011submodular}) in the case $\mathcal{F}=\{V\}$. \fi
By standard methods (see Appendix \ref{sec:relaxations}) one may solve these problems in polytime if one can separate over the relaxation $P(\mathcal{F})$. This is often the case for many natural families such as spanning trees, perfect matchings, $st$-paths, and vertex covers.
\subsection{Rounding the (MA-LE) formulation for upwards closed families $\mathcal{F}$} It is shown in \cite{chekuri2011submodular} that in the setting of monotone objectives and $\mathcal{F} = \{V\}$, a fractional solution of (MA-LE) can be rounded into an integral one at an $O(\log n)$ factor loss. Moreover, they show this still holds for some special types of nonmonotone objectives.
\begin{theorem}[\cite{chekuri2011submodular}]
\label{thm:monot-MSCA}
Let $z_1+z_2+\cdots+z_k$ be a feasible solution for (MA-LE) in the setting where $\mathcal{F}=\{V\}$ (i.e. $\sum_{i\in [k]} z_i = \chi^V$)
and $f_i = g_i + h$ where the $g_i$ are nonnegative monotone submodular and $h$ is nonnegative symmetric submodular.
Then there is a randomized rounding procedure that outputs an integral feasible solution
$\bar{z}_1+\bar{z}_2+\cdots+\bar{z}_k$ such that $\sum_{i\in [k]} f^L_i (\bar{z}_i) \leq O(\log n) \sum_{i \in [k]} f_i^L(z_i)$ on expectation.
That is, we get a partition $S_1,S_2,\ldots,S_k$ of $V$ such that $\sum_{i \in [k]}f_i(S_i) \leq O(\log n) \sum_{i \in [k]} f_i^L(z_i)$
on expectation. \end{theorem}
Our next result shows that the above rounding procedure can be adapted in a straightforward way to the setting where we have a general upwards closed family $\mathcal{F}$. We omit the proof to Appendix \ref{sec:nonmonotone}. \begin{theorem}
\label{thm:sym-MSCA}
Consider an instance of (MA-LE) where $\mathcal{F}$ is an upwards closed family and $f_i = g_i + h$ where the $g_i$ are nonnegative monotone submodular and $h$ is nonnegative symmetric submodular. Let $z_1+z_2+\cdots+z_k$ be a feasible solution such that $\sum_{i \in [k]} z_i \geq \chi^U$ for some $U \in \mathcal{F}$.
Then there is a randomized rounding procedure that outputs an integral feasible solution
$\bar{z}_1+\bar{z}_2+\cdots+\bar{z}_k$ such that $\sum_{i \in [k]} \bar{z}_i \geq \chi^U$ and $\sum_{i\in [k]} f^L_i (\bar{z}_i) \leq O(\log |U|) \sum_{i \in [k]} f_i^L(z_i)$ on expectation. That is, we get a subpartition $S_1,S_2,\ldots,S_k$ such that $\biguplus_{i \in [k]} S_i \supseteq U \in \mathcal{F}$ and $\sum_{i \in [k]}f_i(S_i) \leq O(\log |U|) \sum_{i \in [k]} f_i^L(z_i)$
on expectation. \end{theorem}
\subsection{A multi-agent gap of $O(\min\{k, \log^2 (n)\})$} \label{sec:generic-min-approx}
In this section we present the proof of Theorem \ref{thm:klog}. The main idea behind our reductions is the following. We start with an optimal solution $z^* = z_1^* + z_2^* + \cdots + z^*_k$ to the multi-agent relaxation (MA-LE) and build a new feasible solution $\hat{z} = \hat{z}_1 + \hat{z}_2 + \cdots + \hat{z}_k$ where the $\hat{z}_i$ have supports $V_i$ that are pairwise disjoint. We interpret the $V_i$ as the set of items associated (or pre-assigned) to agent $i$.
Once we have such a pre-assignment we consider the single-agent problem $\min g(S): S \in \mathcal{F}$ where \begin{equation} \label{g-function}
g(S)=\sum_{i=1}^k f_i(S \cap V_i). \end{equation} It is clear that $g$ is nonnegative monotone submodular since the $f_i$ are as well. Moreover, for any solution $S \in \mathcal{F}$ for this single-agent problem we obtain a MA solution of the same cost by setting $S_i = S \cap V_i$, since we then have $g(S) = \sum_{i\in [k]} f_i (S \cap V_i) = \sum_{i \in [k]} f_i(S_i)$.
For a set $S \subseteq V$ and a vector $z \in [0,1]^V $ we denote by $z|_S$ the truncation of $z$ to elements of $S$.
That is, we set $z|_S (v) = z(v)$ for each $v \in S$ and to zero otherwise. Then notice that by definition of $g$
we have that $g^L(z) = \sum_{i \in [k]} f^L_i(z|_{V_i})$. Moreover, if we also have that the $V_i$ are pairwise disjoint, then $\sum_{i \in [k]} f^L_i(z|_{V_i}) = \sum_{i \in [k]} f^L_i(z_i)$. We formalize this observation in the following result.
\begin{proposition}
\label{prop:g-function}
Let $z = z_1 + z_2 + \cdots + z_k$ be a feasible solution to (MA-LE) where the
vectors $z_i$ have pairwise disjoint supports $V_i$. Then
$g^L(z) = \sum_{i \in [k]} f^L_i(z|_{V_i}) = \sum_{i \in [k]} f^L_i(z_i).$ \end{proposition}
The next two results show how one can get a feasible solution $\hat{z} = \hat{z}_1 + \hat{z}_2 + \cdots + \hat{z}_k$ where the $\hat{z}_i$ have pairwise disjoint supports, by losing a factor of $O( \log^2 (n) )$ and $k$ respectively. We remark that these two results combined prove Theorem \ref{thm:klog}.
\begin{theorem}
\label{thm:min-log}
Suppose there is a (polytime) $\alpha(n)$-approximation for monotone SO($\mathcal{F}$) minimization
based on rounding (SA-LE). Then there is a (polytime) $O(\alpha(n) \log (n) \log (\frac{n}{\log n}) )$-approximation
for monotone MASO($\mathcal{F}$) minimization. \end{theorem} \begin{proof}
Let $z^* = z_1^* + z_2^* + \cdots + z^*_k$ denote an optimal solution to (MA-LE) with value $OPT_{frac}$.
In order to apply a black box single-agent rounding algorithm we must create a different multi-agent solution. This is done in several steps, the first few of which are standard. The key steps are the {\em fracture, expand and return} steps which arise later in the process.
Let $a_{max}$ denote the largest entry of the matrix $A$.
Call an element $v$ {\em small} if $z^*(v) \leq \frac{1}{2 n \cdot a_{max}}$. Then note that the total contribution of small elements in any given constraint is at most a half, i.e. for any row $a_i$ of the matrix $A$ we have $a_i \cdot z|_{small} \leq \frac{1}{2}$. We obtain a new feasible solution $z' = z'_1 + z'_2 + \cdots + z'_k$ by removing all small elements from the support of the $z^*_i$ and then doubling the resulting vectors. Notice that this is indeed feasible since $A z' \geq 2(r-\frac{1}{2} \cdot \mathbf{1}) = 2r - \mathbf{1} \geq r$, where $\mathbf{1}$ denotes the vector of all ones. Moreover, by monotonicity and homogeneity of the $f^L_i$, this at most doubles the cost of $OPT_{frac}$.
We now prune the solution $z' = z'_1 + z'_2 + \cdots + z'_k$ a bit more. Let $Z_j$ be the elements $v$ such that $z'(v) \in (2^{-(j+1)},2^{-j}]$ for $j=0,1,2,\ldots,L$. Since $z'(v) > \frac{1}{2n\cdot a_{max}}$ for any element in the support, and we assume that $a_{max}$ is polynomially bounded in $n$, we have that $L = O(\log n)$. We call $Z_j$ {\em bin $j$} and define $r_j = 2^j$. We round up each $v \in Z_j$ so that $z'(v)=2^{-j}$ by augmenting the $z'_i$ values by at most a factor of $2$. We may do this simultaneously for all $v$ by possibly ``truncating'' the values associated to some of the elements. As before, this is fine since the $f^L_i$ are monotone. In the end, we call this a {\em uniform solution} $z'' = z''_1 + z''_2 + \cdots + z''_k$ in the sense that each $z''(v)$ is some power of $2$. Note that its cost is at most $4 \cdot OPT_{frac}$.
{\sc Fracture.} We now {\em fracture} the vectors $z''_i$ by defining vectors $z''_{i,j} = z''_i |_{Z_j}$ for each $i \in [k]$ and each $j \in \{0,1,\ldots,L\}$,
where recall that the notation $z|_S$ denotes the truncation of $z$ to elements of $S$. Notice that $z''_i = \sum_{j=0}^L z''_{i,j}$.
{\sc Expand.} Now for each $j \in \{0,1,\ldots,L\}$ we blow up the vectors $z''_{i,j}$ by a factor $r_j$. (Don't worry, this scaling is temporary.) Since $z''(v) = \frac{1}{r_j}$ for each $v \in Z_j$, this means that the resulting values yield a (probably fractional)
cover of $Z_j$. We can then use the rounding procedure discussed in Theorem \ref{thm:monot-MSCA} (with ground set $Z_j$ and taking $h\equiv 0$) to get an integral solution $z'''_{i,j}$ such that $\sum_i f^L_i (z'''_{i,j}) \leq O(\log |Z_j|) \sum_i f^L_i (r_j \cdot z''_{i,j})$ on expectation.
{\sc Return.} Now we go back to get a new MA-LE solution $\hat{z} = \hat{z}_1 + \hat{z}_2 + \cdots + \hat{z}_k$ by setting $\hat{z}_i = \sum_{j=0}^L \frac{1}{r_j} z'''_{i,j}$. Note that $\hat{z}=z''$ and so this is indeed feasible (and again uniform). Moreover, we have that the cost of this new solution satisfies \begin{multline*}
\sum_{i=1}^k f^L_i (\hat{z}_i) = \sum_{i=1}^k f^L_i ( \sum_{j=0}^L \frac{1}{r_j} z'''_{i,j} )
\leq \sum_{i=1}^k \sum_{j=0}^L \frac{1}{r_j} f^L_i ( z'''_{i,j} )
= \sum_{j=0}^L \frac{1}{r_j} \sum_{i=1}^k f^L_i ( z'''_{i,j} ) \\
\leq O( \sum_{j=0}^L \sum_{i=1}^k \log (|Z_j|) f^L_i (z''_{i,j}) )
\leq O( \sum_{j=0}^L \log (|Z_j|) \sum_{i=1}^k f^L_i (z''_i) )
\leq O( L \cdot \log (\frac{n}{L})) \cdot OPT_{MA} , \end{multline*} where in the first inequality we use the convexity and homogeneity of the $f^L_i$, in the second inequality we use again the homogeneity together with the upper bound for $\sum_i f^L_i (z'''_{i,j})$, in the third inequality we use monotonicity and the fact that $z''_{i,j} \leq z''_i$ for all $j$, and in the last one we use that $\sum_{i=1}^k f^L_i (z''_i) \leq 4 \cdot OPT_{frac} \leq 4 \cdot OPT_{MA}$ and \begin{equation*}
\sum_{j=0}^L \log |Z_j| = \log (\prod_{j=0}^L |Z_j|) \leq \log (\frac{\sum_{j=0}^L |Z_j|}{L+1})^{L+1} = (L+1) \cdot \log (\frac{n}{L+1}) = O( L \cdot \log (\frac{n}{L})), \end{equation*} where the inequality follows from the AM-GM inequality.
\iffalse \begin{eqnarray*} \sum_i f^L_i (\hat{z}_i) & = & \sum_i f^L_i ( \sum_{j=0}^L \frac{1}{r_j} z'''_{i,j} ) \leq \sum_i \sum_{j=0}^L \frac{1}{r_j} f^L_i ( z'''_{i,j} ) = \sum_{j=0}^L \frac{1}{r_j} \sum_i f^L_i ( z'''_{i,j} ) \leq O(\log n) \sum_i \sum_{j=0}^L f^L_i (z''_{i,j}) \\ & \leq & O(\log n) (L+1) \sum_i f^L_i (z''_i) \leq O(\log^2 (n)) \sum_i f^L_i (z^*_i) \leq O(\log^2 (n)) \cdot OPT(\mbox{MA}) , \end{eqnarray*} where in the first inequality we use the convexity and homogeneity of the $f^L_i$, in the second inequality we use again the homogeneity together with the upper bound for $\sum_i f^L_i (z'''_{i,j})$, and in the third inequality we use monotonicity and the fact that $z''_{i,j} \leq z''_i$ for all $j$. \fi
{\sc Single-Agent Rounding.} In the last step we use the function $g$ defined in (\ref{g-function}), with sets $V_i$ corresponding to the support of the $\hat{z}_i$. Given our $\alpha$-approximation rounding assumption for (SA-LE), we can round $\hat{z}$ to find a set $\hat{S}$ such that $g(\hat{S})\leq \alpha g^L(\hat{z})$. Then, by setting $\hat{S}_i = \hat{S} \cap V_i$ we obtain a MA solution satisfying: \begin{equation*} \sum_{i=1}^k f_i(\hat{S}_i) = g(\hat{S}) \leq \alpha g^L(\hat{z}) = \alpha \sum_{i=1}^k f^L_i (\hat{z}_i) \leq \alpha \cdot O( L \cdot \log (\frac{n}{L})) \cdot OPT_{MA}, \end{equation*} where the second equality follows from Proposition \ref{prop:g-function}. Since $L = O( \log n )$, this completes the proof. \qed \end{proof}
We now give an approximation in terms of the number of agents, which becomes preferable when $k < \log^2 (n)$.
\begin{lemma}
\label{lem:min-k}
Suppose there is a (polytime) $\alpha(n)$-approximation for monotone SO($\mathcal{F}$) minimization
based on rounding (SA-LE). Then there is a (polytime) $k \alpha(n)$-approximation
for monotone MASO($\mathcal{F}$) minimization. \end{lemma} \begin{proof}
Let $z^* = z_1^* + z_2^* + \cdots + z^*_k$ denote an optimal solution to (MA-LE) with value $OPT_{frac}$.
We build a new feasible solution $\hat{z} = \hat{z}_1 + \hat{z}_2 + \cdots + \hat{z}_k$ as follows.
For each element $v \in V$ let $i' = \argmax_{i \in [k]} z^*_i(v)$, breaking ties arbitrarily.
Then set $\hat{z}_{i'}(v)=k z^*_i(v)$ and $\hat{z}_{i}(v)=0$ for each $i\neq i'$.
By construction we have $\hat{z} \geq z^*$, and hence this is indeed a feasible solution. Moreover, by construction we also
have that $\hat{z}_i \leq k z_i^*$ for each $i \in [k]$. Hence, given the monotonicity and homogeneity of the $f^L_i$ we have
\begin{equation*}
\sum_{i=1}^k f^L_i(\hat{z}_i) \leq \sum_{i=1}^k f^L_i(k z^*_i) = k \sum_{i=1}^k f^L_i(z^*_i) = k \cdot OPT_{frac} \leq k \cdot OPT_{MA}.
\end{equation*}
Since the $\hat{z}_i$ have disjoint supports $V_i$, we can now use the function $g$ defined in (\ref{g-function})
and do a single-rounding argument as in Theorem \ref{thm:min-log}. This completes the proof. \qed \end{proof}
\iffalse We note that the above result does not particularly depend on the blocking formulation $P^*(\mathcal{F})$. It still holds true for any other upwards closed relaxation of $conv(\{\chi^S:S\in \mathcal{F}\})$ that we decide to use in the (SA-LE) convex formulation. \fi
The above lemma has interesting consequences in the case where $\mathcal{F}=\{V\}$. This is the submodular facility location problem considered by Svitkina and Tardos in \cite{svitkina2010facility}. They give an $O(\log n)$-approximation where $n$ denotes the number of customers/clients/demands. Lemma \ref{lem:min-k} implies we also have a $k$-approximation which is preferable in facility location problems where the number of customers swamps the number of facility locations (for instance, for Amazon).
\begin{corollary}
\label{Cor facility-location}
There is a polytime $k$-approximation for submodular facility location, where
$k$ denotes the number of facilities. \end{corollary} \begin{proof}
The single-agent version of the problem is the trivial $\min f(S): S \in \{V\}$. Hence a polytime exact algorithm is available for the single-agent problem and thus by Lemma \ref{lem:min-k} a polytime $k$-approximation is available for the multi-agent version. \qed \end{proof}
\iffalse The above two corollaries are in fact special cases of using the reduction from Section \ref{sec:MASA} combined with the following result.
\begin{lemma}
\label{lemma partition}
Let $\mathcal{P}=(E,\mathcal{I})$ be a partition matroid over $E$ with $\mathcal{I}=\{S\subseteq E:|S\cap E_{i}|\leq k_{i}\}$,
and $f$ a nonnegative monotone submodular function. Consider the
problem $\min f(S):S\in\mathbb{B}(\mathcal{P})^{\uparrow}$, where $\mathbb{B}(\mathcal{P})$ denotes the set of all bases of the matroid $\mathcal{P}$, and $\mathbb{B}(\mathcal{P})^{\uparrow}$ its upwards closure. Then there is a $\max_{i}\{|E_{i}|-k_{i}+1\}$-approximation.
If $P^*(\mathbb{B}(\mathcal{P})^{\uparrow})$ has a polytime separation oracle, then this is a polytime algorithm. \end{lemma} \begin{proof}
The blockers in this case are given by $\mathcal{B}(\mathbb{B}(\mathcal{P})^{\uparrow})=\{B\subseteq E:|B\cap E_{i}|\geq|E_{i}|-k_{i}+1\}$.
Now the result follows from Theorem \ref{thm:device} and Corollary \ref{cor:sep-algorithmic}.
\qed \end{proof}
Notice that both Corollary \ref{Cor Non-separable} and Corollary \ref{Cor facility-location} follow from applying the reduction from Section \ref{sec:MASA} and then using Lemma \ref{lemma partition} with $E_{i}=\delta(v_{i})$ and $k_{i}=1$ for each $i\in[n]$. \fi
\subsection{A tight multi-agent gap of $O(\log n)$ for bounded blocker families} \label{sec:MAfromSA}
In Section \ref{sec:generic-min-approx} we established an $O( \log^2 (n) )$ MA gap whenever there is a SA approximation algorithm based on the (SA-LE) formulation. For the vertex cover problem, however, there is an improved MA gap of $O(\log n)$ due to Goel et al. In this section we generalize their result by describing a larger class of families wtih such MA gap.
Recall that due to monotonicity, one may often assume that we are working with a family $\mathcal{F}$ which is {\em upwards-closed}, aka a {\em blocking family} (cf. \cite{iyer2014monotone}).
The advantage is that to certify whether $F \in \mathcal{F}$, we only need to check that $F \cap B \neq \emptyset$ for each element $B$ of the family $\mathcal{B}(\mathcal{F})$ of minimal blockers of $\mathcal{F}$. We discuss the details in Appendix \ref{sec:blocking}. The blocking relaxation for a family $\mathcal{F}$ is then given by $P^*(\mathcal{F}):=\{z \geq 0: z(B) \geq 1 ~\textit{for all $B \in \mathcal{B}(\mathcal{F})$} \}$. In this section we consider the formulations (SA-LE) and (MA-LE) in the special case where the fractional relaxation of the integral polyhedron is given by $P^*(\mathcal{F})$.
The $2 \ln (n)$-approximation algorithm of Goel et al. for multi-agent vertex cover relies only on the fact that the feasible set family has the following {\em bounded blocker property}. We call a clutter (family of noncomparable sets) $\mathcal{F}$ {\em $\beta$-bounded} if $|F| \leq \beta$ for all $F \in \mathcal{F}$.
We then say that $\mathcal{F}$ has a $\beta$-bounded blocker if $|B|\leq \beta$ for each $B \in \mathcal{B}(\mathcal{F})$.
The main SA minimization result for such families is the following. \iffalse \cite{koufogiannakis2013greedy} gives a greedy algorithm that achieves a $\beta$-approximation for such families. Moreover, this is a polytime algorithm when $\mathcal{B}(\mathcal{F})$ is of polynomial size. Subsequently, \cite{iyer2014monotone} achieved the same $\beta$-approximation factor via a convex program. In addition, their algorithm is polytime whenever $P^*(\mathcal{F})$ has a polytime separation oracle (which is a weaker assumption than the one from \cite{koufogiannakis2013greedy}). \fi \begin{theorem}[\cite{iyer2014monotone,koufogiannakis2013greedy}] \label{thm:SABBagain}
Let $\mathcal{F}$ be a family with a $\beta$-bounded
blocker. Then there is a $\beta$-approximation algorithm for monotone $SO(\mathcal{F})$ minimization.
If $P^*(\mathcal{F})$ has a polytime separation oracle, then this is a polytime algorithm. \end{theorem}
Our next result establishes an $O(\log n)$ MA gap for families with a bounded blocker. In fact, while our work has focused on monotone objectives (due to the inapproximabiltiy results for general submodular $f_i$) the next result extends to some special types of nonmonotone objectives. These were introduced in \cite{chekuri2011submodular} and \cite{ene2014hardness}, where a tractable middle-ground is found for the minimum submodular cost allocation problem (where $\mathcal{F} = \{V\}$). They work with objectives $f_i=g_i+h$ where the $g_i$ are monotone submodular and $h$ is symmetric submodular (in \cite{chekuri2011submodular}) or just submodular (in \cite{ene2014hardness}).
\iffalse
we establish a $\beta \ln (n)$-approximation for the monotone multi-agent minimization problem associated to families with a $\beta$-bounded blocker. We remark that this $O(\ln n)$-gap between SA and MA instances for families with bounded blockers is tight, as one can see in the examples of vertex covers ($2$-approximation for SA and a tight $O(\ln n)$-approximation for MA) or submodular facility location ($1$-approximation for SA and a tight $\ln (n)$-approximation for MA). \fi
\iffalse Notice that this result indeed generalizes the $2\ln(n)$-approximation for multi-agent vertex cover, since the family of vertex cover of a graph is $2$-bounded. \fi
\iffalse For the single-agent nonnegative monotone problem it was shown in \cite Bilmes that families with $\beta$-bounded blockers always admit a $\beta$-approximation. \fi
\iffalse \begin{theorem}
\label{BB alpha log(n) approx}Let $\mathcal{F}$ be a family with a $\beta$-bounded
blocker. Then there is a $\beta\ln(n)$-approximation algorithm for the associated multi-agent nonnegative monotone problem.
If $P^*(\mathcal{F})$ has a polytime separation oracle, then this is a polytime algorithm. \end{theorem} \begin{proof}
Let $x^{*}$ be an optimal solution to (\ref{MA-LP}). Recall that if $P^*(\mathcal{F})$ has a polytime separation oracle
then we have that $x^*$ can be computed in polynomial time (see Appendix \ref{sec:ma extensions}).
Consider
$Q=\{v\in V:\sum_{S\ni v}\sum_{i\in[k]}x^{*}(S,i)\geq\frac{1}{\beta}\}$.
Since $\mathcal{F}$ has a $\beta$-bounded blocker it follows that $Q\in\mathcal{F}$,
and moreover, $\beta x^{*}$ is a fractional cover of $Q$. That
is, $\sum_{S}\sum_{i}\beta x^{*}(S,i)\chi^{S}\geq\chi^{Q}$.
We
now appeal to standard analysis of the greedy algorithm applied to the fractional set cover (of $Q$) with set costs given by $c((S,i))=f_{i}(S)$.
This results in an integral cover of $Q$ whose cost is augmented by a factor of at most $\ln n$ times the fractional cover. Denote this
integral cover by $(S^{1},1),...,(S^{m_{1}},1),(S^{1},2),\ldots,$ $(S^{m_{2}},2),...,(S^{m_{k}},k)$.
Since
$\sum_{S}\sum_{i}x^{*}(S,i)f_{i}(S)\leq OPT(MA)$ we have that the
integral cover cost satisfies
\begin{eqnarray*}
\sum_{i=1}^{k}\sum_{j=1}^{m_{i}}c(S^{j},i) & \leq & \ln(n)\sum_{S}\sum_{i}\beta x^{*}(S,i)f_{i}(S)\\
& = & \beta\ln(n)\sum_{S}\sum_{i}x^{*}(S,i)f_{i}(S)\\
& \leq & \beta \ln(n)OPT(MA).
\end{eqnarray*}
In addition, by letting $S_{i}=\cup_{j=1}^{m_{i}}(S^{j},i)$, by submodularity (or just subadditivity)
we have
\[
\sum_{i=1}^{k}f_{i}(S_{i})=\sum_{i=1}^{k}f_{i}(\cup_{j=1}^{m_{i}}(S^{j},i))\leq\sum_{i=1}^{k}\sum_{j=1}^{m_{i}}f_{i}((S^{j},i))=\sum_{i=1}^{k}\sum_{j=1}^{m_{i}}c(S^{j},i).
\]
Finally, if $S_{i}\cap S_{j}\neq\emptyset$ for some $i\neq j$ throw
away the common shared elements from one of them arbitrarily. Denote
by $S_{1}^{*}$,...,$S_{k}^{*}$ these new sets, and notice that $S_{i}^{*}\subseteq S_{i}$ and $S_{1}^{*}\uplus ...\uplus S_{k}^{*}=Q$. By monotonicity of the functions $f_{i}$'s we have
\[
\sum_{i\in[k]}f_{i}(S_{i}^{*})\leq\sum_{i\in[k]}f_{i}(S_{i})\leq\beta\ln(n)OPT(\mbox{MA})
\]
and the result follows.\qed \end{proof}
\fi
We remark that by taking $h\equiv 0$ (which is symmetric submodular), we obtain a result for monotone functions. We note that in this setting we do not need $\mathcal{F}$ to be upwards closed, since due to monotonicity we can work with the upwards closure of $\mathcal{F}$ without loss of generality as previously discussed on Section \ref{sec:SA-MA-formulations} (see Appendix \ref{sec:blocking} for further details). Moreover, as previously pointed out, this $O(\log n)$ MA gap is tight due to examples like vertex covers ($2$-approximation for SA and a tight $O(\log n)$-approximation for MA) or submodular facility location ($1$-approximation for SA and a tight $O(\log n)$-approximation for MA).
\begin{theorem}
\label{BB-nonmonotone}Let $\mathcal{F}$ be an upwards closed family with a $\beta$-bounded
blocker. Let the objectives be of the form $f_i = g_i + h$ where each $g_i$ is nonnegative monotone submodular and $h$ is nonnegative symmetric submodular. Then there is a randomized $O(\beta \log n)$-approximation algorithm for the associated $MASO(\mathcal{F})$ minimization problem.
If $P^*(\mathcal{F})$ has a polytime separation oracle, then this is a polytime algorithm. \end{theorem} \begin{proof}
Let $z^* = \sum_{i \in [k]} z^*_i$ be an optimal solution to (MA-LE) based on the blocking relaxation $P^*(\mathcal{F})$ with value $OPT_{frac}$.
Consider the new feasible solution given by $\beta z^* = \sum_{i \in [k]} \beta z^*_i$ and let $U=\{v\in V: \beta z^*(v) \geq 1\}$.
Since $\mathcal{F}$ has a $\beta$-bounded blocker it follows that $U\in\mathcal{F}$.
We now have that $\sum_{i \in [k]} \beta z^*_i$ is a feasible solution such that $\sum_{i \in [k]} \beta z^*_i \geq \chi^U$.
Thus, we can use Theorem \ref{thm:sym-MSCA} to get an integral feasible solution
$\sum_{i \in [k]} \bar{z}_i$ such that $\sum_{i \in [k]} \bar{z}_i \geq \chi^U$ and $\sum_{i\in [k]} f^L_i (\bar{z}_i) \leq O(\log |U|) \sum_{i \in [k]} f_i^L(\beta z^*_i) \leq \beta \cdot O(\log n) \cdot OPT_{frac}$ on expectation.
\qed \end{proof}
It is shown in \cite{ene2014hardness} (see their Proposition 10) that given any nonnegative submodular function $h$, one may define a nonnegative symmetric submodular function $h'$ such that for any partition $S_1,S_2,\ldots,S_k$ we have $\sum_i h'(S_i) \leq k \sum_i h(S_i)$. This, with our previous result, yields the following corollary. \iffalse They also show that the $k$ factor loss due to the non-symmetry is in fact tight. \fi \begin{corollary}
\label{BB-nonmonotone-again}Let $\mathcal{F}$ be an upwards closed family with a $\beta$-bounded
blocker. Let the objectives be of the form $f_i = g_i + h$ where each $g_i$ is nonnegative monotone submodular and $h$ is nonnegative submodular. Then there is a randomized $O(k \beta \log n)$-approximation algorithm for the associated $MASO(\mathcal{F})$ minimization problem. \end{corollary}
One may wish to view the above results through the lens of MA gaps, leading to the question of what is the associated SA primitive? For $g_i+h$ objectives, the SA version should be for general (or symmetric) nonnegative submodular objectives. Moreover, as we only use upwards closed families, one may deduce that these single-agent versions have $\beta$-appoximations via the concept of the monotone closure of a nonmonotone objective from \cite{iyer2014monotone}. Hence our results establish MA gaps of $O(\log n)$ (resp. $O(k \log n)$) in these nonmonotone settings (and the factor of $k$ is tight \cite{mirzakhani2014sperner}).
Before concluding this section, we note that Theorems~\ref{thm:klog} and \ref{thm:SABBagain} imply a $k \beta$-approximation for families with bounded blockers, which becomes preferable when $k < O(\log n)$.
\begin{corollary} \label{cor:kgapBB} Let $\mathcal{F}$ be a family with a $\beta$-bounded
blocker. Then there is a $k \beta$-approximation algorithm for the associated monotone $MASO(\mathcal{F})$ minimization problem. \end{corollary}
\iffalse Many natural classes $\mathcal{F}$ have been shown to behave poorly for (nonnegative, monotone) submodular minimization. For instance, polynomial inapproximability has been established in the case where $\mathcal{F}$ is the family of edge covers (\cite{iwata2009submodular}), spanning trees (\cite{goel2009approximability}), or perhaps most damning is $\tilde{\Omega}(\sqrt{n})$-inapproximability
for minimizing $f(S)$ subject to the simple cardinality constraint $|S| \geq k$\footnote{which is equivalent to the constraint $|S|=k$ since f is monotone} \cite{svitkina2011submodular}. These results all show that it is difficult to distinguish between two similar submodular functions, one of which ``hides'' the optimal set. At some level, the proofs leverage the fact that $f$ can have wild (so-called) curvature \cite{conforti1984submodular}.
A few cases exist where good approximations hold. For instance, for ring families the problem can be solved exactly by adapting any algorithm that works for unconstrained submodular minimization (\cite{orlin2009faster,schrijver2000combinatorial}). Gr\"{o}tschel, Lov\'asz, and Schrijver (\cite{grotschel2012geometric}) show that the problem can be also solved exactly when $\mathcal{F}$ is a \emph{triple family}. Examples of these kind of families include the family of all sets of odd cardinality, the family of all sets of even cardinality, or the family of sets having odd intersection with a fixed $T \subseteq V$. More generally, Goemans and Ramakrishnan (\cite{goemans1995minimizing}) extend this last result by showing that the problem can be solved exactly when $\mathcal{F}$ is a \emph{parity family} (a more general class of families). In the context of NP-Hard problems, the cases in which a good (say $O(1)$ or $O(\log(n))$) approximation exists are almost non-existent. We have that the submodular vertex cover admits a $2$-approximation (\cite{goel2009approximability,iwata2009submodular}), and the $k$-uniform hitting set has $O(k)$-approximation. It is natural to examine the core reasons why these approximations algorithms work in the hope of harnessing ingredients in other contexts of interest.
We emphasize that our work in this section focuses only on nonnegative normalized monotone functions. We also restrict attention to upwards-closed families $\mathcal{F}$, that is, if $F \subseteq F'$ and $F \in \mathcal{F}$, then $F' \in \mathcal{F}$. Due to monotonicity this is usually a benign assumption (see Appendix \ref{sec:blocking}). For instance, if we have a well-described formulation (or approximation) for the polytope $P(\mathcal{F})$ whose vertices are $\{\chi^F: F \in \mathcal{F}\}$, then we can also separate over its upwards closure $P(\mathcal{F})^\uparrow:=\{z \geq x: x \in P\}$. A second issue one must address when working with the upwards-closure of a family $\mathcal{F}$ is whether, given $F'$ in the closure, one may find a set $F \in \mathcal{F}$ with $F \subseteq F'$ in polytime. This is also the case if a polytime separation oracle is available for $P(\mathcal{F})$. \fi
\iffalse The algorithms used in each case differ, and work for different reasons. The VC algorithm works since we are minimizing over a blocking family $\mathcal{F}$ whose blocker has bounded size sets. Facility location is based on a greedy algorithm whose \textquotedbl{}bang - per -buck\textquotedbl{} subroutine is based on detecting whether a given submod function takes on negative values. The alg of C-E\dots{} no clue. \\
{\em TO DO:\\ \\ Look up: k-hitting set result from Vondrak's slides. The paper with the hardness result considers the problem: vertex cover for k-regular hypergraphs.\\ } \fi
\iffalse \subsubsection{Bounded Blockers and Single-Agent Minimization} \label{subsec:BB-gap}
Algorithms for vertex cover and $k$-uniform hitting sets rely only on the fact that the feasible space has a {\em bounded blocker property}. We call a clutter (set family) $\mathcal{F}$ {\em $\beta$-bounded} if $|F| \leq \beta$ for all $F \in \mathcal{F}$.
The next result is the main algorithmic device of this section and has
several natural applications including implications for the class of multi-agent problems over $\mathcal{F}$. These are discussed in the following sections. \begin{theorem} \label{thm:device} Consider the (nonnegative, normalized, monotone) submodular minimization primitive \[ \min f(S):S\in\mathcal{F} \] and its natural LP (\ref{eqn:X}). If $\mathcal{B}(\mathcal{F})$ is $\beta$-bounded, then (\ref{eqn:X}) has integrality gap of at most $\beta$. \end{theorem} \begin{proof} Now assume we have that $x^{*}$ is an optimal solution for the primal (\ref{eqn:X}) with value {\sc opt}. We claim that $Q=\{v\in V:\sum_{S:S\ni v}x^{*}(S)\geq\frac{1}{\beta}\}$ is an $\beta$-approximation. Recall that $Q\in\mathcal{F}$ if and only if $Q\cap B\neq\emptyset$ for each $B\in\mathcal{B}(\mathcal{F})$. Pick an arbitrary
$B\in\mathcal{B}(\mathcal{F})$, since $x^{*}$ is a feasible solution we must have that $\sum_{v\in B}\sum_{S\ni v}x^{*}(S)\geq1$. Hence, there must exist some $v_{0}\in B$ such that $\sum_{S\ni v_{0}}x^{*}(S)\geq\frac{1}{|B|}\geq\frac{1}{\beta}$, where the last inequality follows from the fact that $\mathcal{B}(\mathcal{F})$ is $\beta$-bounded. It follows that $v_{0}\in Q$ and thus $Q\cap B\supseteq\{v_{0}\}\neq\emptyset$. Since $\beta x^{*}$ is a fractional cover of $Q$, we may now apply Corollary~\ref{cor:convexbound} (the latter part in particular) to deduce that $f(Q) \leq \beta$ {\sc opt}. \qed \end{proof}
In many circumstances this result becomes algorithmic. For instance, as discussed in Lemma~\ref{lem:polybound} in Appendix \ref{sec:extended-lp} we have the following.
\begin{corollary} \label{cor:algorithmic}
If $\mathcal{B}(\mathcal{F})$ is $\beta$-bounded and $|\mathcal{B}(\mathcal{F})| \in poly(n)$, then there is a polytime $\beta$-approximation for (\ref{eqn:X}). This is the case in particular if $\beta=O(1)$. \end{corollary}
We note that it is not necessary to have a $O(1)$-bounded blocker to derive an $O(1)$-approximation algorithm; for instance we could have that $\mathcal{F}$ is polynomially bounded in some cases so an exact algorithm exists. The inapproximability result of Khot et al. \cite{khot2008vertex} could be seen as a type of converse to Theorem~\ref{thm:device}; it suggests that if the blockers are large, and suitably rich, then submodular minimization over $\mathcal{F}$ is doomed.
\begin{theorem}[Khot, Regev \cite{khot2008vertex}]
Assuming the Unique Games conjecture, there is a $(k-\epsilon)$-inapproximability factor
for the problem of finding a minimum size vertex cover in a $k$-uniform hypergraph. \end{theorem}
The restriction that $|\mathcal{B}(\mathcal{F})| \in poly(n)$ could be quite severe. In Appendix~\ref{sec:combined} we discuss how LP (\ref{eqn:X}) can still be solved in polytime under a much weaker assumption. In particular we show the following.
\begin{corollary}
\label{cor:sep-algorithmic}
Assume there is a polytime separation oracle for the blocking formulation $P^*(\mathcal{F})$. Then LP (\ref{eqn:X}) can be solved in polytime. \end{corollary} \fi
\iffalse
\subsection{Multi-agent minimization over $\mathcal{F}=\{V\}$} \label{sec:MA min F=V}
In this section we discuss the multi-agent framework (\ref{ma}) in the minimization setting and in the special case where the functions $f_{i}$ are nonnegative and $\mathcal{F}=\{V\}$. That is, we focus on the problem \[ \begin{array}{cc} \min & \sum_{i=1}^{k}f_{i}(S_{i})\\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}=V\\ & S_{i}\in \mathcal{F}_i\,,\,\forall i\in[k] \end{array} \] where the functions $f_{i}$ are nonnegative submodular.
As previously discussed in Section~\ref{sec:applications} (see Example \ref{ex:MSCA}), the special case of this problem where the agents are allowed to take any subset of $V$ (i.e. $\mathcal{F}_{i}=2^{V}$ for all $i$) has been widely considered in the literature and it has been referred to as the Minimum Submodular Cost Allocation (MSCA) problem. We summarize the results for MSCA here.
\begin{theorem} Consider MSCA for nonnegative functions $f_{i}$. We have the following. \begin{enumerate} \item There is a polytime (tight) $O(\ln (n))$-approx for the case where the functions are monotone (\cite{hayrapetyan2005network,svitkina2010facility}). \item There is a polytime (tight) $O(\ln (n))$-approx for the case where the functions $f_{i}$ can be written as $f_{i}=g_{i}+h$ for some nonnegative monotone submodular $g_{i}$ and a nonnegative symmetric submodular function $h$ (\cite{chekuri2011submodular}). \item There is a polytime (tight) $O(k\cdot\ln (n))$-approx for the case where the functions $f_{i}$ can be written as $f_{i}=g_{i}+h$ for some nonnegative monotone submodular $g_{i}$ and a nonnegative submodular function $h$ (\cite{ene2014hardness,mirzakhani2014sperner}). \item For general functions $f_i$ and $k\geq3$ this problem does not admit any finite multiplicative approximation factor (\cite{ene2014hardness}). \end{enumerate} \end{theorem}
We now discuss the monotone MSCA problem with additional constraints on the sets that the agents can take. Two of the most natural settings to consider are to impose cardinality constraints on the number of elements that an agent can take, or to allow agent $i$ to only take elements from some subset $V_{i}\subseteq V$, or both. The rest of this section will be dealing with multi-agent problems that incorporate one (or both) of such constraints.
The following result is an observation that derives from the work of Svitkina-Tardos \cite{svitkina2010facility} on the submodular facility location problem. \begin{corollary} The problem \[ \begin{array}{cc} \min & \sum_{i=1}^{k}f_{i}(S_{i})\\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}=V\\
& S_{i}\subseteq V_{i}, \end{array} \]
where the functions $f_{i}$ are nonnegative monotone submodular admits a $\ln(\max_{i\in[k]}|V_{i}|)$-approximation. \end{corollary} \begin{proof} Svitkina-Tardos \cite{svitkina2010facility} provide a reduction of the problem \[ \begin{array}{cc} \min & \sum_{i=1}^{k}f_{i}(S_{i})\\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}=V \end{array} \] where the functions $f_{i}$ are nonnegative monotone submodular to a set cover instance. Then, the greedy algorithm (using unconstrained submodular minimization as a subroutine) is used to obtain an optimal
$\ln(n)$-approximation. The same reduction applied in this setting gives a $t$-set cover instance (i.e. an instance where all the sets have size at most $t$) with $t=\max_{i\in[k]}|V_{i}|$. Now the result follows from the fact that greedy gives a $\ln(t)$-approximation for the $t$-set cover problem. \qed \end{proof}
We now consider the above problem with an additional cardinality constraint on the sets the agents can take. We use the techniques from the lifting reduction presented in Section \ref{sec:MASA} and a known result from matching theory in bipartite graphs to obtain the desired approximation. We first introduce some definitions that will be useful for the proof.
Let $G=(A+B,E)$ be a bipartite graph with $|A|=k$ and $A=\{a_1,...,a_k\}$. We call a subset of edges $M\subseteq E$ \emph{saturating} if $cov(M)=B$. Also, given some nonnegative integers $b_{1},...,b_{k}$, we say that $M$ is a \emph{$(b_{1},...,b_{k})$-matching}
if $|M\cap\delta(a_{i})|\leq b_{i}$ for each $i\in[k]$.
\begin{lemma} Consider the following multi-agent minimization problem \[ \begin{array}{cc} \min & \sum_{i\in[k]}g_{i}(S_{i})\\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}=V\\
& S_{i}\subseteq V_{i}\\
& |S_{i}|\leq b_{i}. \end{array} \] Then, if the functions $g_{i}$ are nonnegative monotone submodular there exists a polytime $(\max_{i}b_{i})$-factor approximation. \end{lemma} \begin{proof} Apply the generic reduction from Section \ref{sec:MASA}, and let $G=([k]+V,E)$ be the corresponding bipartite graph (i.e., with $E=\uplus_{i\in[k]}\{(i,v):v\in V_{i}\}$) and $f$ the corresponding function. Define weights $w:E\to\mathbb{R}_{+}$ such that $w_{e}=f(e)$ for each $e\in E$. Let $M\subseteq E$ be any saturating $(b_{1},...,b_{k})$-matching in $G$. By submodularity of $f$ we have that $w(M)\geq f(M)$. Moreover, since $g$ can be written as $g(S_{1},...,S_{k})=\sum_{i\in[k]}g_{i}(S_{i})$, we have that $f(M)=\sum_{i\in[k]}f_{i}(M_{i})$ for some monotone submodular functions $f_{i}$ where $M_{i}:=M\cap\delta(i)$ and $f_{i}(M_{i})=g_{i}(cov(M_{i}))$. Moreover, by monotonicity of the $f_{i}$ we have \[
f_{i}(M_{i})\geq\max_{e\in M_{i}}f_{i}(e)=\max_{e\in M_{i}}f(e)\geq\frac{1}{|M_{i}|}\sum_{e\in M_{i}}f(e)=\frac{1}{|M_{i}|}w(M_{i})\geq\frac{1}{b_{i}}w(M_{i}) \] where the first equality follows from the fact that $f(e)=f_{i}(e)$ for every $e\in\delta(i)$, and the last inequality from the fact that $M$ is a saturating $(b_{1},...,b_{k})$-matching and hence
$|M_{i}|\leq b_{i}$ for every $i\in[k]$. Hence, it follows that for each saturating $(b_{1},...,b_{k})$-matching $M$ we have \[ w(M)\geq f(M)=\sum_{i\in[k]}f_{i}(M_{i})\geq\sum_{i\in[k]}\frac{1}{b_{i}}w(M_{i})\geq\sum_{i\in[k]}(\frac{1}{\max_{i\in [k]}b_{i}})w(M_{i})=\frac{1}{\max_{i\in [k]}b_{i}}w(M). \] In particular, if $M^{*}$ is a minimum saturating $(b_{1},...,b_{k})$-matching for the weights $w$, we have \[ \begin{array}{cccccc} w(M^{*}) & \geq f(M^{*})\geq & \min & f(M) & \geq & \frac{1}{\max_{i \in [k]}b_{i}}w(M^{*}).\\
& & \mbox{s.t.} & M\mbox{ is a saturating }(b_{1},...,b_{k})\mbox{\mbox{-matching}} \end{array} \] This is, $f(M^{*})$ is a $(\max_{i}b_{i})$-factor approximation. We conclude the argument by noticing that $w$ is a modular function, and hence we can find a minimum saturating $(b_{1},...,b_{k})$-matching in polynomial time. \qed \end{proof}
\iffalse \textcolor{red}{NOT SURE WHETHER THE NEXT COROLLARY SHOULD STAY. IT WAS WRITTEN MAINLY TO COMPARE IT TO THE MV CASE, WHICH WAS VERY HARD}
\begin{corollary} \label{cor:b_i=1} Consider the above problem in the special case where $k=n$ and all the $b_{i}=1$. Then, an optimal solution to the problem is given by a minimum perfect matching with respect to the weights $w$ as defined in the proof above. Moreover, this solution is still optimal even in the case where we do not require anymore the original functions $g_{i}$ to be nonnegative or monotone. \end{corollary} \fi \iffalse Surprisingly, the above problem becomes very hard to approximate if we allow general multivariate submodular functions instead of decomposable ones. This is true even under the special conditions of Corollary \ref{cor:b_i=1} in which $k=n$ and $b_{i}=1$ for all the agents. \begin{lemma} \label{lem:b_i=1Hard} There is an information-theoretic hardness lower bound of $\Omega(n)$ for the multivariate problem \[ \begin{array}{cc} \min & g(S_{1},\ldots,S_{k})\\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}=V\\
& S_{i}\subseteq V_{i}\\
& |S_{i}| \leq 1 \end{array} \] where $g$ is a nonnegative monotone multivariate submodular function. \end{lemma} \begin{proof}
We prove this in the special case where $k=n$. Notice that in this case the constraint $|S_i|\leq 1$ becomes $|S_i| = 1$. We reduce an instance of the submodular perfect matching (SPM) problem to the problem above. In \cite{goel2009approximability}, the following is shown for any fixed
$\epsilon,\delta > 0$. Any randomized $(\frac{n^{1-3\epsilon}}{1+\delta})$-approximation algorithm for the submodular minimum cost perfect matching problem on bipartite graphs (with at least one perfect matching), requires exponentially many queries. So let us consider an instance of SPM consisting of a bipartite graph
$G=(A+B,E)$ with $|A|=|B|$ and a nonnegative monotone submodular function $f:2^{E}\to\mathbb{R}_{+}$. The goal is to find a perfect matching
$M\subseteq E$ of minimum submodular cost. We can transform this to the general multivariate problem above by applying the reduction from Section \ref{sec:MASA} in the opposite direction. More formally, let $k=|A|,\:V=B,\mbox{ and }V_{i}=cov(\delta(a_{i}))$ where $A=\{a_{1},...,a_{k}\}$. Also, define a multivariate function $g:2^{V_{1}}\times...\times2^{V_{k}}\to\mathbb{R}_{+}$ by $g(S_{1},...,S_{k})=f(\uplus_{i\in[k]}\{(a_{i},b):b\in S_{i}\})$ for any $(S_{1},...,S_{k})\in2^{V_{1}}\times \cdots \times2^{V_{k}}$ (i.e. $S_{i}\subseteq V_{i}$). We have $g$ is multivariate submodular by Claim~\ref{Claim f submodular}. It follows that a solution to \[ \begin{array}{cc} \min & g(S_{1},\ldots,S_{k})\\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}=V\\
& S_{i}\subseteq V_{i}\\
& |S_{i}|=1 \end{array} \]
(where here $|V|=k$) is a solution to the original SPM instance. \qed \end{proof}
\fi
We conclude this section by showing that monotone MSCA (which admits an $O(\ln(n))$-approximation) becomes very hard when additional cardinality constraints are added.
\begin{claim} The problem \[ \begin{array}{cc} \min & \sum_{i=1}^{k}f_{i}(S_{i})\\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}=V\\
& |S_{i}|\leq b_{i} \end{array} \] where the $f_{i}$ are nonnegative monotone submodular functions is $\Omega(\frac{\sqrt{n}}{\ln n})$-hard. This is even the case when $k=2$. \end{claim} \begin{proof} It is known (\cite{svitkina2011submodular}) that the problem
$\min f(S): |S|=m$ where $f$ is a nonnegative monotone submodular function is $\Omega(\frac{\sqrt{n}}{\ln n})$-hard. We reduce a general instance of this problem to the problem of the claim. To see this notice that by monotonicity of $f$ we have \[ \begin{array}{ccccccccccc} \min & f(S) & = & \min & f(S) & = & \min & f(S) & = & \min & f(S)+f'(S')\\
\mbox{s.t.} & |S|=m & & \mbox{s.t.} & |S|\geq m & & \mbox{s.t.} & S+S'=V & & \mbox{s.t.} & S+S'=V\\
& & & & & & & |S'|\leq|V|-m & & & |S'|\leq|V|-m \end{array} \] where $f'(S')=0$ for all $S'\subseteq V$. \qed \end{proof}
\fi
\subsection{A tight multi-agent gap of $O(\log n)$ for ring and crossing families} \label{sec:rings}
It is well known (\cite{schrijver2000combinatorial}) that submodular minimization can be solved exactly in polynomial time over a ring family. In this section we observe that the MA problem over this type of constraint admits a tight $\ln (n)$-approximation. More generally, we consider {\em crossing families}. A family $\mathcal{F}$ of subsets of $V$ forms a ring family (aka lattice family) if for each $A,B \in \mathcal{F}$ we also have $A \cap B, A \cup B \in \mathcal{F}$.
A crossing family is one where we only require it for sets where $A \setminus B, B \setminus A, A \cap B, V- (A \cup B)$ are all non-empty. Hence any ring family is a crossing family.
For any crossing family $\mathcal{F}$ and any $u,v \in V$, let $\mathcal{F}_{uv}=\{A \in \mathcal{F}: u \in A, v \not\in A\}$. It is easy to see that $\mathcal{F}_{uv}$ is a ring family. Moreover, we may solve the original MA problem by solving the associated MA problem for each non-empty $\mathcal{F}_{uv}$ and then selecting the best output solution.
So we assume now that we are given a ring family in such a way that we may compute its minimal
set $M$ (which is unique). This is a standard assumption when working with ring families (cf. submodular minimization algorithm described in \cite{schrijver2000combinatorial}). Then, due to monotonicity and the fact that $\mathcal{F}$ is closed under intersections, it is not hard to see that the original problem reduces to the facility location problem $$ \min \sum_{i=1}^k f_i(S_i): S_1 \uplus \cdots \uplus S_k = M \; , $$
which admits a tight $(\ln |M|)$-approximation (\cite{svitkina2010facility}). In particular, for the special case where we have the trivial ring family $\mathcal{F} = \{V\}$ we get a tight $\ln (n)$-approximation. The next result summarizes these observations.
\begin{theorem} There is a tight $\ln (n)$-approximation for monotone $MASO(\mathcal{F})$ minimization over crossing families $\mathcal{F}$. \end{theorem}
\section{Multi-agent submodular maximization} \label{sec:MASA}
In this section we describe two different reductions. The first one reduces the capacitated multi-agent problem (\ref{ma}) to a single-agent problem, and it is based on the simple idea of taking $k$ disjoint copies of the original ground set. We show that several properties of the objective and family of feasible sets stay \emph{invariant} (i.e. preserved) under the reduction. We use this to establish an (optimal) MA gap of 1 for several families. Examples of such families include spanning trees, matroids, and $p$-systems.
Our second reduction is based on the multilinear extension of a set function. We establish that if the SA primitive admits approximation via its multilinear relaxation (see Section \ref{sec:max-SA-MA-formulations}), then we may extend this to its MA version with a constant factor loss, in the monotone and nonmonotone settings. Moreover, for the monotone case our MA gap is tight.
\subsection{The lifting reduction} \label{sec:lifting-reduction}
In this section we describe a generic reduction of (\ref{ma}) to a single-agent problem $$\max/\min f(S): S\in\mathcal{L}.$$ The argument is based on the idea of viewing assignments of elements $v$ to agents $i$ in a {\em multi-agent bipartite graph}. This simple idea (which is equivalent to making $k$ disjoint copies of the ground set) already appeared in the classical work of Fisher et al \cite{fisher1978analysis}, and has since then been widely used \cite{lehmann2001combinatorial,vondrak2008optimal,calinescu2011maximizing,singh2012bisubmodular}. We review briefly the reduction here for completeness and to fix notation.
Consider the complete bipartite graph $G=([k]+V,E)$.
Every subset of edges $S \subseteq E$ can be written uniquely as $S = \uplus_{i \in [k]} (\{i\} \times S_i)$ for some sets $S_i \subseteq V$. This allows us to go from a multi-agent objective (such as the one in (\ref{ma})) to a univariate objective $f:2^{E}\to\mathbb{R}$ over the lifted space. Namely, for each set $S \subseteq E$ we define $f(S)=\sum_{i \in [k]} f_i(S_i)$. The function $f$ is well-defined because each subset $S\subseteq E$ can be uniquely written as $S = \uplus_{i \in [k]} (\{i\} \times S_i)$ for some $S_i \subseteq V$.
We consider two families of sets over $E$ that capture the original constraints: $$ \mathcal{F}':=\{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{F}\} \hspace{15pt} \mbox{and} \hspace{15pt} \mathcal{H}:=\{S\subseteq E:S_{i}\in\mathcal{F}_{i},\;\forall i\in[k]\}. $$
\noindent We now have: \[ \begin{array}{cccccccc} \max/\min & \sum_{i \in [k]} f_i(S_i) & = & \max/\min & f(S) & = & \max/\min & f(S)\\ \mbox{s.t.} & S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{F} & & \mbox{s.t.} & S\in\mathcal{F}' \cap \mathcal{H} & & \mbox{s.t.} & S\in\mathcal{L}\\
& S_{i}\in\mathcal{F}_{i}\,,\,\forall i\in[k] \end{array}, \] where in the last step we just let $\mathcal{L}:=\mathcal{F}' \cap\mathcal{H}$.
\iffalse Notice that for the robust problem discussed in Theorem \ref{thm:max robust} we get the SA problem \begin{equation*} \begin{array}{cccccccc} \max & \min & \sum_{i \in [k]} f_i(S_i - A_i) & = & \max & \min & f(S-A). \\ S_1 \uplus \cdots \uplus S_k \in \mathcal{F} & A_i \subseteq S_i & & & S \in \mathcal{L} & A \subseteq S& \\
S_{i} \in \mathcal{F}_{i} & \sum_i |A_i|\leq \tau & & & & |A| \leq \tau & \end{array} \end{equation*} \fi
This reduction is interesting if our new function $f$ and family of sets $\mathcal{L}$ have properties which allows us to handle them computationally. This will depend on the original structure of the functions $f_i$ and the set families $\mathcal{F}$ and $\mathcal{F}_{i}$. In terms of the objective, it is straightforward to check (as previously pointed out in \cite{fisher1978analysis}) that if the $f_i$ are (nonnegative, respectively monotone) submodular functions, then $f$ as defined above is also (nonnegative, respectively monotone) submodular. In Section \ref{sec:reduction-properties} we discuss several properties of the families $\mathcal{F}$ and $\mathcal{F}_i$ that are preserved under this reduction.
\subsection{The single-agent and multi-agent formulations} \label{sec:max-SA-MA-formulations}
For a set function $f:\{0,1\}^V \to \mathbb{R}$ we define its \emph{multilinear extension} $f^M:[0,1]^V \to \mathbb{R}$ (introduced in \cite{calinescu2007maximizing}) as \begin{equation*} f^M(z)=\sum_{S \subseteq V} f(S) \prod_{v \in S} z_v \prod_{v \notin S} (1-z_v). \end{equation*} An alternative way to define $f^M$ is in terms of expectations. Consider a vector $z \in [0,1]^V$ and let $R^z$ denote a random set that contains element $v_i$ independently with probability $z_{v_i}$. Then $f^M(z)= \mathbb{E}[f(R^z)]$, where the expectation is taken over random sets generated from the probability distribution induced by $z$.
This gives rise to natural single-agent and multi-agent relaxations. The {\em single-agent multilinear extension relaxation} is: \begin{equation} \label{SA-ME} (\mbox{SA-ME}) ~~~~ \max f^M(z): z \in P(\mathcal{F}), \end{equation} and the {\em multi-agent multilinear extension relaxation} is: \begin{equation} \label{MA-ME} (\mbox{MA-ME}) ~~~~ \max \sum_{i=1}^k f^M_i(z_i): z_1 + z_2 + \cdots + z_k \in P(\mathcal{F}), \end{equation} where $P(\mathcal{F})$ denotes some relaxation of $conv(\{\chi^S:S\in \mathcal{F}\})$. While the relaxation (SA-ME) has been used extensively \cite{calinescu2011maximizing,lee2009non,feldman2011unified,ene2016constrained,buchbinder2016constrained} in the submodular maximization literature, we are not aware of any previous work using the multi-agent relaxation (MA-ME).
The following result shows that when $f$ is nonnegative submodular and the formulation $P(\mathcal{F})$ is downwards closed and admits a polytime separation oracle, the relaxation (SA-ME) can be solved approximately in polytime.
\begin{theorem}[\cite{buchbinder2016constrained,vondrak2008optimal}]
\label{thm:multilinear-solve-monot}
Let $f:2^V \to \mathbb{R}_+$ be a nonnegative submodular function and $f^M:[0,1]^V \to \mathbb{R}_+$ its multilinear extension. Let $P \subseteq [0,1]^V$ be any downwards closed polytope that admits a polytime separation oracle, and denote $OPT = \max f^M(z): z\in P$. Then there is a polytime algorithm (\cite{buchbinder2016constrained}) that finds $z^* \in P$ such that $f^M(z^*) \geq 0.385 \cdot OPT$. Moreover, if $f$ is monotone there is a polytime algorithm (\cite{vondrak2008optimal}) that finds $z^* \in P$ such that $f^M(z^*) \geq (1-1/e) OPT$. \end{theorem}
For monotone objectives the assumption that $P$ is downwards closed is without loss of generality. This is not the case, however, when the objective is nonmonotone. Nonetheless, this restriction is unavoidable, as Vondr{\'a}k \cite{vondrak2013symmetry} showed that no algorithm can find $z^* \in P$ such that $f^M(z^*) \geq c \cdot OPT$ for any constant $c>0$ when $P$ admits a polytime separation oracle but it is not downwards closed.
We can solve (MA-ME) to the same approximation factor as (SA-ME). This follows from the fact that the MA problem has the form $\{ \max g(w) : w \in W \subseteq {\bf R}^{nk} \}$ where $g(w)=g(z_1,z_2,\ldots,z_k)=\sum_{i \in [k]} f^M_i(z_i)$ and $W$ is the downwards closed polytope $\{w=(z_1,...,z_k): \sum_i z_i \in P(\mathcal{F})\}$. Clearly we have a polytime separation oracle for $W$ given that we have one for $P(\mathcal{F})$. Moreover, it is straightforward to check (see Lemma \ref{lem:max-multilinear} on Appendix \ref{sec:Appendix-Invariance}) that $g(w)=f^M(w)$, where $f$ is the function on the lifted space after applying the lifting reduction from Section \ref{sec:lifting-reduction}. Thus, $g$ is the multilinear extension of a nonnegative submodular function, and we can now use Theorem \ref{thm:multilinear-solve-monot}.
\subsection{A tight multi-agent gap of $1-1/e$} \label{sec:max-MA-gap}
In this section we present the proof of Theorem \ref{thm:max-MA-gap}. The high-level idea behind our reduction is the same as in the minimization setting (see Section \ref{sec:generic-min-approx}). That is, we start with an (approximate) optimal solution $z^* = z_1^* + z_2^* + \cdots + z^*_k$ to the multi-agent (MA-ME) relaxation and build a new feasible solution $\hat{z} = \hat{z}_1 + \hat{z}_2 + \cdots + \hat{z}_k$ where the $\hat{z}_i$ have supports $V_i$ that are pairwise disjoint. We then use for the SA rounding step the single-agent problem (as previously defined in (\ref{g-function}) for the minimization setting) $\max g(S): S \in \mathcal{F}$ where $g(S)=\sum_{i \in [k]} f_i(S \cap V_i)$.
Similarly to Proposition \ref{prop:g-function} which dealt with the Lov\'asz extension, we have the following result for the multilinear extension.
\begin{proposition}
\label{prop:max-g-function}
Let $z = \sum_{i\in [k]} z_i$ be a feasible solution to (MA-ME) where the
vectors $z_i$ have pairwise disjoint supports $V_i$. Then
$g^M(z) = \sum_{i \in [k]} f^M_i(z|_{V_i}) = \sum_{i \in [k]} f^M_i(z_i).$ \end{proposition}
We now have all the ingredients to prove our main result in the maximization setting.
\begin{theorem}
If there is a (polytime) $\alpha(n)$-approximation for monotone SO($\mathcal{F}$) maximization
based on rounding (SA-ME), then there is a (polytime) $(1-1/e) \cdot \alpha(n)$-approximation
for monotone MASO($\mathcal{F}$) maximization. Furthermore, given a downwards closed family $\mathcal{F}$,
if there is a (polytime) $\alpha(n)$-approximation
for nonmonotone SO($\mathcal{F}$) maximization
based on rounding (SA-ME), then there is a (polytime) $0.385 \cdot \alpha(n)$-approximation
for nonmonotone MASO($\mathcal{F}$) maximization. \end{theorem} \begin{proof}
We discuss first the case of monotone objectives.
{\sc STEP 1.}
Let $z^* = z_1^* + z_2^* + \cdots + z^*_k$ denote an approximate solution to (MA-ME) obtained via Theorem \ref{thm:multilinear-solve-monot}, and let $OPT_{frac}$ be the value of an optimal solution. We then have that $\sum_{i \in [k]} f^M_i(z^*_i) \geq (1-1/e) OPT_{frac} \geq (1-1/e) OPT_{MA}$.
{\sc STEP 2.}
For an element $v \in V$ let $\bf{e_v}$ denote the characteristic vector of $\{v\}$, i.e. the vector in $\mathbb{R}^V$ which has value $1$ in the $v$-th component and zero elsewhere. Notice that by definition of the multilinear extension we have that the functions $f^M_i$ are linear along directions $\bf{e_v}$ for any $v \in V$. It then follows that the function
\begin{equation*}
h(t) = f^M_i (z^*_i + t \mathbf{e_v} ) + f^M_{i'} (z^*_{i'} - t \mathbf{e_v} ) + \sum_{j\in [k], j\neq i,i'} f^M_j(z^*_j)
\end{equation*}
is also linear for any $v\in V$ and $i \neq i' \in [k]$, since it is the sum of linear functions (on $t$). In particular, given any $v \in V$ such that there exist $i \neq i' \in [k]$ with $z^*_i(v),z^*_{i'}(v)>0$, there is always a choice so that increasing one component and decreasing the other by the same amount does not decrease the objective value. We use this as follows.
Let $v \in V$ be such that there exist $i \neq i' \in [k]$ with $z^*_i(v),z^*_{i'}(v)>0$. Then, we either set $z^*_i(v) = z^*_i(v) + z^*_{i'}(v)$ and $z^*_{i'}(v) = 0$, or $z^*_{i'}(v) = z^*_i(v) + z^*_{i'}(v)$ and $z^*_i(v) = 0$, whichever does not decrease the objective value. We repeat until the vectors $z^*_i$ have pairwise disjoint support. Let us denote these new vectors by $\hat{z}_i$ and let $\hat{z}= \sum_{i\in [k]} \hat{z}_i$. Then notice that the vector $z^* = \sum_{i\in [k]} z^*_i$ remains invariant after performing each of the above updates (i.e. $\hat{z} = z^*$), and hence the new vectors $\hat{z}_i$ remain feasible.
\iffalse
Let $\bar{z} = (z^*_1,z^*_2,\ldots,z^*_k)$. From Lemma \ref{lem:max-multilinear} we know that $f^M(\bar{z}) = \sum_{i \in [k]} f^M_i(z^*_i)$. Moreover, since $f^M$ is the multilinear extension of a submodular function $f$, it follows from Lemma \ref{lem:multilinear-convex} that $f^M$ is convex in the directions $\bf{e_i} - \bf{e_j}$. That is, given any two components $z^*_i(v_j)$ and $z^*_{i'}(v_{j'})$ of the vector $\bar{z}=(z^*_1,z^*_2,\ldots,z^*_k)$, there is always a choice so that increasing one component and decreasing the other by the same amount does not decrease the objective value. We use this as follows.
Let $v \in V$ be such that there exist $i \neq i' \in [k]$ with $z^*_i(v),z^*_{i'}(v)>0$. Then, we either set $z^*_i(v) = z^*_i(v) + z^*_{i'}(v)$ and $z^*_{i'}(v) = 0$, or $z^*_{i'}(v) = z^*_i(v) + z^*_{i'}(v)$ and $z^*_i(v) = 0$, whichever does not decrease the objective value. We repeat until the vectors $z^*_i$ have pairwise disjoint support. Notice that the vector $z^* = \sum_{i\in [k]} z^*_i$ remains invariant after performing each of these changes, and hence the new modified vectors $z^*_i$ remain feasible at all times. Let us denote the new vectors by $\hat{z}_i$ and let $\hat{z}= \sum_{i\in [k]} \hat{z}_i$.
\fi
{\sc STEP 3.}
In the last step we use the function $g$ defined in (\ref{g-function}), with sets $V_i$ corresponding to the support of the $\hat{z}_i$.
Given our $\alpha$-approximation rounding assumption for (SA-ME), we can round $\hat{z}$
to find a set $\hat{S}$ such that $g(\hat{S})\geq \alpha g^M(\hat{z})$.
Then, by setting $\hat{S}_i = \hat{S} \cap V_i$ we obtain a MA solution satisfying
\begin{equation*}
\sum_{i \in [k]} f_i(\hat{S}_i) = g(\hat{S}) \geq \alpha g^M(\hat{z}) = \alpha \sum_i f^M_i (\hat{z}_i) \geq \alpha \sum_i f^M_i (z^*_i) \geq \alpha (1-1/e) OPT_{MA},
\end{equation*}
where the second equality follows from Proposition \ref{prop:max-g-function}.
This completes the proof for monotone objectives.
In the nonmonotone case the proof is very similar.
Here we restrict our attention to downwards closed families, since then we can get a $0.385$-approximation at STEP 1 via Theorem \ref{thm:multilinear-solve-monot}. We then apply STEP 2 and 3 in the same fashion as we did for monotone objectives. This leads to a $0.385 \cdot \alpha(n)$-approximation for the multi-agent problem.
\iffalse
In the case of nonmonotone objectives the proof is very similar, the only differences being at STEP 1. Notice that in the nonmonotone case we can only approximate the solution to the multilinear extension in the case where the family $\mathcal{F}$ is downwards closed, i.e. when $P(\mathcal{F})$ is downwards closed. Thus, in the nonmonotone setting we only work with downwards closed families. We can then get a $0.385$-approximation at STEP 1 for the multilinear extension optimization problem via Theorem \ref{thm:multilinear-solve-monot}. We then apply STEP 2 and 3 in the same fashion as we did for monotone objectives. This leads to a $0.385 \cdot \alpha(n)$-approximation for the multi-agent problem.
\fi \qed \end{proof}
\subsection{Invariance under the lifting reduction} \label{sec:reduction-properties}
In Section \ref{sec:max-MA-gap} we established a MA gap of $(1-1/e)$ for monotone objectives and of $0.385$ for nonmonotone objectives and downwards closed families based on the multilinear formulations. In this section we describe several families with an (optimal) MA gap of $1$. Examples of these family classes include spanning trees, matroids, and $p$-systems. Moreover, the reduction in this case is completely black box, and hence do not depend on the multilinear (or some other particular) formulation.
We saw in Section \ref{sec:lifting-reduction} how if the original functions $f_i$ are all submodular, then the lifted function $f$ is also submodular. We now focus on the properties of the original families $\mathcal{F}_i$ and $\mathcal{F}$ that are also preserved under the lifting reduction. We show, for instance, that if the family $\mathcal{F}$ induces a matroid (or more generally a $p$-system) over the original ground set $V$, then so does the family $\mathcal{F}'$ over the lifted space $E$. We summarize these results in Table \ref{table:properties-preserved}, and present most of the proofs in this section. We next discuss some of the algorithmic consequences of these invariance results.
\begin{table}[H]
\caption{Invariant properties under the lifting reduction}
\label{table:properties-preserved}
\resizebox{\linewidth}{!}{
\begin{tabular}{|c|c|c|c|}
\hline
& \textbf{Multi-agent problem} & \textbf{Single-agent (i.e. reduced) problem} & Result\tabularnewline
\hline
1 & $f_i$ submodular & $f$ submodular & \cite{lehmann2001combinatorial} \tabularnewline
\hline
2 & $f_i$ monotone & $f$ monotone & \cite{lehmann2001combinatorial} \tabularnewline
\hline
3 & $(V,\mathcal{F})$ a $p$-system & $(E,\mathcal{F}')$ a $p$-system & Prop \ref{prop:p-systems} \tabularnewline
\hline
4 & $\mathcal{F}$ = bases of a $p$-system & $\mathcal{F}'$ = bases of a $p$-system & Corollary \ref{cor:p-systems bases} \tabularnewline
\hline
5 & $(V,\mathcal{F})$ a matroid & $(E,\mathcal{F}')$ a matroid & Corollary \ref{cor:matroid-invariant}\tabularnewline
\hline
6 & $\mathcal{F}$ = bases of a matroid & $\mathcal{F}'$ = bases of a matroid & Corollary \ref{cor:matroid-bases} \tabularnewline
\hline
7 & $(V,\mathcal{F})$ a $p$-matroid intersection & $(E,\mathcal{F}')$ a $p$-matroid intersection & Appendix \ref{sec:Appendix-Invariance} \tabularnewline
\hline
\iffalse
8 & $\mathcal{F}=\{V\}$ & $\mathcal{F}'=$ set of bases of a partition matroid & Appendix \ref{sec:Appendix-Invariance} \tabularnewline
\hline
9 & $\mathcal{F}=2^{V}$ & $(E,\mathcal{F}')$ a partition matroid & Appendix \ref{sec:Appendix-Invariance} \tabularnewline
\hline
\fi
8 & $(V,\mathcal{F}_{i})$ a matroid for all $i\in[k]$ & $(E,\mathcal{H})$ a matroid & Appendix \ref{sec:Appendix-Invariance} \tabularnewline
\hline
9 & $\mathcal{F}_{i}$ a ring family for all $i\in[k]$ & $\mathcal{H}$ a ring family & Appendix \ref{sec:Appendix-Invariance} \tabularnewline
\hline
10 & $\mathcal{F}=$ forests (resp. spanning trees) & $\mathcal{F}'=$ forests (resp. spanning trees) & Section \ref{sec:reduction-properties}\tabularnewline
\hline
11 & $\mathcal{F}=$ matchings (resp. perfect matchings) & $\mathcal{F}'=$ matchings (resp. perfect matchings) & Section \ref{sec:reduction-properties}\tabularnewline
\hline
12 & $\mathcal{F}=$ $st$-paths & $\mathcal{F}'=$ $st$-paths & Section \ref{sec:reduction-properties}\tabularnewline
\hline
\end{tabular}
} \end{table}
In the setting of MASO (i.e. (\ref{eqn:MA})) this invariance allows us to leverage several results from the single-agent to the multi-agent setting. These are based on the following result, which uses the fact that the size of the lifted space $E$ is $nk$. \begin{theorem}
\label{thm:max-invariance1}
Let $\mathcal{F}$ be a matroid, a $p$-matroid intersection, or a $p$-system. If there is a (polytime) $\alpha(n)$-approximation algorithm for monotone (resp. nonmonotone) SO($\mathcal{F}$) maximization (resp. minimization),
then there is a (polytime) $\alpha(nk)$-approximation algorithm for monotone (resp. nonmonotone) MASO($\mathcal{F}$) maximization (resp. minimization). \end{theorem}
In both the monotone and nonmonotone maximization settings, the approximation factors $\alpha(n)$ for the family classes described in the theorem above are independent of $n$. Hence, we immediately get that $\alpha(nk)=\alpha(n)$ for these cases, and thus approximation factors for the corresponding MA problems are the same as for the SA versions. In our MA gap terminology this implies a MA gap of 1 for such problems.
In the setting of CMASO (i.e. (\ref{ma})) the results described on entries $8$ and $9$ of Table \ref{table:properties-preserved} provide additional modelling flexibility. This allows us to combine several constraints while keeping approximation factors fairly good. For instance, for a monotone maximization instance of CMASO where $\mathcal{F}$ corresponds to a $p$-matroid intersection and the $\mathcal{F}_i$ are all matroids, the above invariance results lead to a $(p+1+\epsilon$)-approximation.
\iffalse Let $\mathcal{C}$ be some family class (e.g. the class of matroids) over the original ground set $V$ and $\mathcal{C}'$ be the same class of objects (e.g. matroids) over the new ground set $E$. Then we say that the class $\mathcal{C}$ is \emph{invariant} if for each instance of MASO($\mathcal{F}$) with a family $\mathcal{F} \in \mathcal{C}$, the reduction outputs an instance of SO($\mathcal{F}'$) with a family $\mathcal{F}' \in \mathcal{C}'$. We show, for example, that the class of matroids or the class of $p$-systems are invariant in the previous sense.
These results are summarized in Table \ref{table:properties-preserved}, and note that they prove Theorem \ref{thm:max-invariance}. Moreover, combining these invariance results with the recent advances in SA submodular optimization (Section~\ref{sec:related work}), now imply optimal approximations for many MA maximization (both monotone and nonmonotone) problems. \fi
We now prove some of the results from Table \ref{table:properties-preserved}. We start by presenting some definitions that will be useful. For a subset of edges $S \subseteq E$ we define its \emph{coverage} $cov(S)$ as the set of nodes $v \in V$ saturated by $S$. That is, $v\in cov(S)$ if there exists $i\in [k]$ such that $(i,v)\in S$. We then note that by definition of $\mathcal{F}'$ it is straightforward to see that for each $S \subseteq E$ we have that \begin{equation}
\label{def:family-H}
S \in \mathcal{F}' \iff cov(S) \in \mathcal{F} \mbox{ and } |S|=|cov(S)|. \end{equation}
For a set $S \subseteq E$, a set $B\subseteq S$ is called a \emph{basis} of $S$ if $B$ is an inclusion-wise maximal independent subset of $S$. Our next result describes how bases and their cardinalities behave under the lifting reduction.
\begin{lemma}
\label{lem:bases-invariance}
Let $S$ be an arbitrary subset of $E$. Then for any basis $B$ (over $\mathcal{F}'$) of $S$ there
exists a basis $B'$ (over $\mathcal{F}$) of $cov(S)$ such that $|B'|=|B|$.
Moreover, for any basis $B'$ of $cov(S)$ there exists a basis $B$ of $S$ such that $|B|=|B'|$. \end{lemma} \begin{proof}
For the first part, let $B$ be a basis of $S$ and take $B':=cov(B)$. Since $B \in \mathcal{F}'$
we have by (\ref{def:family-H}) that $B'\in \mathcal{F}$ and $|B'| = |B|$. Now, if $B'$ is not a basis of $cov(S)$ then we can
find an element $v \in cov(S)-B'$ such that $B'+v \in \mathcal{F}$.
Moreover, since $v \in cov(S)$ there exists $i \in [k]$ such that $(i,v)\in S$.
But then we have that $B+(i,v) \subseteq S$ and
$B+(i,v)\in \mathcal{F}'$, a contradiction with the fact that $B$ was a basis of $S$.
For the second part, let $B'$ be a basis of $cov(S)$. For each $v \in B'$ let $i_v$ be such that
$(i_v,v) \in S$, and take $B:=\uplus_{v \in B'} (i_v,v)$. It is clear by definition of $B$ that
$cov(B)=B'$ and $|B|=|B'|$. Hence $B \in \mathcal{F}'$ by (\ref{def:family-H}). If $B$ is not a basis of $S$
there exists an edge $(i,v)\in S-B$ such that
$B+(i,v)\in \mathcal{F}'$. But then by (\ref{def:family-H}) we have that $cov(B+(i,v))\in \mathcal{F}$
and $B'\subsetneq cov(B+(i,v))\subseteq cov(S)$, a contradiction
since $B'$ was a basis of $cov(S)$. \qed \end{proof}
We say that $(V,\mathcal{F})$ is a \emph{$p$-system} if for each $U \subseteq V$, the cardinality of the largest basis of $U$ is at most $p$ times the cardinality of the smallest basis of $U$. Our following result is a direct consequence of Lemma \ref{lem:bases-invariance}.
\begin{proposition}
\label{prop:p-systems}If $(V,\mathcal{F})$ is a $p$-system, then $(E,\mathcal{F}')$
is a $p$-system. \end{proposition} \iffalse \begin{proof}
It follows by (\ref{def:family-H}) that
given any set $S \subseteq E$, we have that $B$ is a basis (over $\mathcal{H}$) of $S$ if and
only if $|B|=|cov(B)|$ and $cov(B)$ is a basis (over $\mathcal{F}$) of $cov(S)$.
In particular, it follows that the maximum cardinality of a basis of $S$ equals
the maximum cardinality of a basis of $cov(S)$, and same for bases of minimum cardinality.
Now the result follows from the fact that $\mathcal{F}$ is a $p$-system. \qed \end{proof} \fi
\begin{corollary}
\label{cor:p-systems bases}
If $\mathcal{F}$ corresponds to the set of bases of a $p$-system $(V,\mathcal{I})$, then $\mathcal{F}'$ also corresponds to the set of bases of some $p$-system $(E,\mathcal{I}')$. \end{corollary} \begin{proof}
Consider $(E,\mathcal{I}')$ where $\mathcal{I}':=\{S\subseteq E: cov(S)\in \mathcal{I} \mbox{ and } |cov(S)|=|S|\}$.
Then by Proposition \ref{prop:p-systems} we have that $(E,\mathcal{I}')$ is a $p$-system.
It is now straightforward to check that $\mathcal{F}'$ corresponds precisely to the set of bases of $(E,\mathcal{I}')$. \qed \end{proof}
The following two results follow from Proposition \ref{prop:p-systems} and Corollary \ref{cor:p-systems bases} and the fact that matroids are precisely the class of 1-systems. \begin{corollary} \label{cor:matroid-invariant}If $(V,\mathcal{F})$ is a matroid, then $(E,\mathcal{F}')$ is a matroid. \end{corollary}
\begin{corollary}
\label{cor:matroid-bases}Assume $\mathcal{F}$ is the set of bases of some matroid
$\mathcal{M}=(V,\mathcal{I})$, then $\mathcal{F}'$ is the set of bases of some matroid $\mathcal{M}'=(E,\mathcal{I}')$. \end{corollary}
We now focus on families defined over the set of edges of some graph $G$. To be consistent with our previous notation we denote by $V$ the set of edges of $G$, since this is the ground set of the original problem.
The lifting reduction is based on the idea of making $k$ disjoint copies for each original element, and visualize the new ground set (or lifted space) as edges in a bipartite graph. However, when the original ground set corresponds to the set of edges of some graph $G$, we may just think of the lifted space as being the set of edges of the graph $G'$ obtained by taking $k$ disjoint copies of each original edge. We think of the edge that corresponds to the $ith$ copy of $v$ as assigning element $v$ to agent $i$. We can formalize this by defining a mapping $\pi:E\to E'$ that takes an edge $(i,v)\in E$ from the lifted space to the edge in $G'$ that corresponds to the $ith$ copy of the original edge $v$. It is clear that $\pi$ is a bijection. Moreover, notice that given any graph $G$ and family $\mathcal{F}$ of forests (as subset of edges) of $G$, the bijection $\pi : E \to E'$ also satisfies that $\bar{\mathcal{F}}:=\{\pi (S): S \in \mathcal{F}'\}$ is precisely the family of forests of $G'$. That is, there is a one-to-one correspondence between forests of $G'$ and assignments $S_1 \uplus S_2 \cdots \uplus S_k = S \in \mathcal{F}$ for the original MA problem where $S$ is a forest of $G$. The same holds for spanning trees, matchings, perfect matchings, and st-paths.
\iffalse We now focus on families defined over the set of edges of some graph $G$. Our results for these types of families have the following flavour. Consider MASO($\mathcal{F}$) where $\mathcal{F}$ corresponds to the family of forests of some graph $G$. Then SO($\mathcal{F}'$) can also be seen as defined over the family $\mathcal{F}'$ of forests of some graph $G'$. More formally, let $E$ and $\mathcal{F}'$ be as defined in the reduction. Then given any graph $G$ and family $\mathcal{F}$ of forests (as subset of edges) of $G$, there exists a graph $G'$ with set of edges $E'$ and a bijection $\pi : E \to E'$ satisfying that $\bar{\mathcal{F}}:=\{\pi (S): S \in \mathcal{F}'\}$ is the family of forests of $G'$. We show that forests (and spanning trees), matchings (and perfect matchings), and st-paths are examples of such families.
\begin{lemma}
\label{Lemma Dummy Copies} Let $(V,\mathcal{F})$ where $V$ corresponds to the set of edges of some graph
$G$, and $\mathcal{F}$ to the family of either forests, spanning trees, matchings, perfect matchings, or $st$-paths of $G$.
Then the reduced problem is given by $(E,\mathcal{F}')$ where $E$ corresponds to the set of edges of some graph $G'$,
and $\mathcal{F}'$ to the family of either forests, spanning trees, matchings, perfect matchings, or $st$-paths of $G'$. \end{lemma} \begin{proof}
First notice that to be consistent with our previous notation we denote
by $V$ the set of edges of $G$, since this is the ground set of
the original problem. Hence, in this case an element $v\in V$ denotes
an edge of the original graph $G$.
The lifting reduction is based
on the idea of making $k$ disjoint copies for each original element,
and visualize the new ground set (or lifted space) as edges in a bipartite
graph. However, when the original ground set corresponds to the set
of edges of some graph $G$, we may just think of the lifted
space in this case as being the set of edges of the graph $G'$ obtained by taking
$k$ disjoint copies of each original edge. We think of choosing the
edge that corresponds to the $ith$ copy of $v$ as assigning element
$v$ to agent $i$. We can formalize this by defining a mapping $\pi:E\to E'$
that takes an edge $(i,v)\in E$ from the lifted space to the edge in $G'$ that corresponds
to the $ith$ copy of the original edge $v$. It is clear that $\pi$
is a bijection.
\textcolor{red}{the rest of the proof needs some work}
Now, in the case of forests. It is clear that for any feasible assignment $S_1,S_2,\ldots,S_k$
there is a forest in $G'$ corresponding to picking . In addition, this is a one-to-one
correspondence.
Then, via the bijections $\pi$ and $\pi'$ we can
identify a tuple $(S_{1},...,S_{k})$ with the set of edges in $G'$
that takes the first copy of the edges in $S_{1}$, the second copies
of the edges in $S_{2}$, and so on. Now, the following three statements
are straightforward.
(1) Let $\mathcal{F}$ be the family of forests of $G$. Then: $(S_{1},...,S_{k})$
is a feasible tuple for the original problem (i.e. $S_{1}+\cdots+S_{k}\in\mathcal{F}$)
$\iff\pi(S_{1},...,S_{k})\in\mathcal{H}\iff\pi'(\pi(S_{1},...,S_{k}))$ induces a forest in $G'$.
(2) Let $\mathcal{F}$ be the family of matchings in $G$. Then: $(S_{1},...,S_{k})$
is a feasible tuple for the original problem $\iff\pi(S_{1},...,S_{k})\in\mathcal{H}\iff\pi'(\pi(S_{1},...,S_{k}))$ induces a matching in $G'$.
(3) Let $\mathcal{F}$ be the family of $st$-paths in $G$. Then: $(S_{1},...,S_{k})$
is a feasible tuple for the original problem $\iff\pi(S_{1},...,S_{k})\in\mathcal{H}\iff\pi'(\pi(S_{1},...,S_{k}))$ induces an $\ensuremath{st}$-path in $G'$.\qed \end{proof} \fi
This observation becomes algorithmically useful given that $G'$ has the same number of nodes of $G$. Thus, any approximation factor or hardness result for the above combinatorial structures that depend on the number of nodes (and not on the number of edges) of the underlying graph, will remain the same in the MA setting. We note that this explains why for the family of spanning trees, perfect matchings, and $st$-paths, the approximations factors revealed in \cite{goel2009approximability} for the monotone minimization problem are the same for both the SA and MA versions. Our next result summarizes this.
\begin{theorem}
\label{thm:max-invariance2}
Let $\mathcal{F}$ be the family (seen as edges) of forests, spanning trees, matchings, perfect matchings, or $st$-paths of a
graph $G$ with $m$ nodes and $n$ edges.
Then, if there is a (polytime) $\alpha(n)$-approximation algorithm for monotone (resp. nonmonotone) SO($\mathcal{F}$) maximization (resp. minimization), there is a (polytime) $\alpha(nk)$-approximation algorithm for monotone (resp. nonmonotone) MASO($\mathcal{F}$) maximization (resp. minimization). Moreover, if there is a (polytime) $\alpha(m)$-approximation algorithm for monotone (resp. nonmonotone) SO($\mathcal{F}$) maximization (resp. minimization),
then there is a (polytime) $\alpha(m)$-approximation algorithm for monotone (resp. nonmonotone) MASO($\mathcal{F}$) maximization (resp. minimization). \end{theorem}
\section{Conclusion} \label{sec:conclusions}
A number of interesting questions remain. Perhaps the main one being whether the $O(\log^2 (n))$ MA gap for minimization can be improved to $O(\log n)$? We have shown this is the case for bounded blocker and crossing families. Another question is whether the $\alpha \log^2 (n)$ and $ \alpha k $ approximations can be made truly black box? I.e., do not depend on the convex formulation.
On separate work (\cite{santiago2016multivariate}) we discuss multivariate submodular objectives. We show that our reductions for maximization remain well-behaved algorithmically and this opens up more tractable models. This is the topic of planned future work.
\appendix \begin{center}
\large \textbf{APPENDIX} \end{center} \section{Upwards-closed (aka blocking) families} \label{sec:blocking}
In this section, we give some background for blocking families. As our work for minimization is restricted to monotone functions, we can often convert an arbitrary set family into its upwards-closure (i.e., a blocking version of it) and work with it instead. We discuss this reduction as well. The technical details discussed in this section are fairly standard and we include them for completeness. Several of these results have already appeared in \cite{iyer2014monotone}.
\subsection{Blocking families and a natural relaxation for the integral polyhedron} \label{sec:block}
A set family $\mathcal{F}$, over a ground set $V$ is {\em upwards-closed} if $F \subseteq F'$ and $F \in \mathcal{F}$, implies that $F' \in \mathcal{F}$; these are sometimes referred to as {\em blocking families}. Examples of such families include vertex-covers or set covers more generally, whereas spanning trees are not.
For a blocking family $\mathcal{F}$ one often works with the induced sub-family $\mathcal{F}^{min}$ of minimal sets. Then $\mathcal{F}^{min}$ has the property that it is a {\em clutter}, that is, $\mathcal{F}^{min}$ does not contain a pair of comparable sets, i.e., sets $F \subset F'$.
If $\mathcal{F}$ is a clutter, then $\mathcal{F}=\mathcal{F}^{min}$ and there is an associated {\em blocking} clutter $\mathcal{B}(\mathcal{F})$, which consists of the minimal sets $B$ such that $B \cap F \neq \emptyset$ for each $F \in \mathcal{F}$. We refer to $\mathcal{B}(\mathcal{F})$ as the {\em blocker} of $\mathcal{F}$.
One also checks that for an arbitrary upwards-closed family $\mathcal{F}$, we have the following. \begin{claim}[Lehman]
\label{claim:blocker}
\begin{enumerate}
\item $F \in \mathcal{F}$ if and only if $F \cap B \neq \emptyset$ for all $B \in \mathcal{B}(\mathcal{F}^{min})$.
\item $\mathcal{B}(\mathcal{B}(\mathcal{F}^{min})) = \mathcal{F}^{min}$.
\end{enumerate} \end{claim}
\noindent Thus the significance of blockers is that one may assert membership in an upwards-closed family $\mathcal{F}$ by checking intersections on sets from the blocker $\mathcal{B}(\mathcal{F}^{min})$. If we define $\mathcal{B}(\mathcal{F})$ to be the minimal sets which intersect every element of $\mathcal{F}$, then one checks that $\mathcal{B}(\mathcal{F})=\mathcal{B}(\mathcal{F}^{min})$. These observations lead to a natural relaxation for minimization problems over the integral polyhedron $P(\mathcal{F}) := conv(\{\chi^F: \textit{$F \in \mathcal{F}$}\})$. The {\em blocking formulation} for $\mathcal{F}$ is: \begin{equation} \label{eqn:polyhedron} P^*(\mathcal{F}) = \{z \in \mathbb{R}^V_{\geq 0}: z(B) \geq 1~~\forall B \in \mathcal{B}(\mathcal{F}^{min})=\mathcal{B}(\mathcal{F})\}. \end{equation}
\noindent Clearly we have $P(\mathcal{F}) \subseteq P^*(\mathcal{F})$.
\subsection{Reducing to blocking families}
Now consider an arbitrary set family $\mathcal{F}$ over $V$. We may define its {\em upwards closure} by $\mathcal{F}^{\uparrow}=\{F': F \subseteq F' \textit{ for some $F \in \mathcal{F}$}\}$. In this section we argue that in order to solve a monotone optimization problem over sets in $\mathcal{F}$ it is often sufficient to work over its upwards-closure.
As already noted $\mathcal{B}(\mathcal{F})=\mathcal{B}(\mathcal{F}^{\uparrow}) = \mathcal{B} (\mathcal{F}^{min})$ and hence one approach
is via the blocking formulation $P^*(\mathcal{F})=P^*(\mathcal{F}^{\uparrow})$. This requires two ingredients. First, we need
a separation algorithm for the blocking relaxation, but indeed this is often available for many
natural families such as spanning trees, perfect matchings, $st$-paths, and vertex covers. The second ingredient needed is the ability to turn an integral solution $\chi^{F'}$ from $P^*(\mathcal{F}^{\uparrow})$ or $P(\mathcal{F}^\uparrow)$
into an integral solution $\chi^F \in P(\mathcal{F})$. We now argue that this is the case if a polytime separation algorithm is available for the blocking relaxation $P^*(\mathcal{F}^{\uparrow})$ or for the polytope $P(\mathcal{F}):= conv(\{\chi^F: \textit{$F \in \mathcal{F}$}\})$.
\iffalse
doable if we have separation over the polyhedron $P(\mathcal{F})=conv(\chi^F: F \in \mathcal{F})$.
\fi
For a polyhedron $P$, we denote its {\em dominant} by $P^{\uparrow} := \{z: z \geq x \textit{~for some~} x \in P \}$. The following observation is straightforward.
\begin{claim}
\label{claim:lattice-points}
Let $H$ be the set of vertices of the hypercube in $\mathbb{R}^V$. Then
$$H \cap P(\mathcal{F}^\uparrow) = H \cap P(\mathcal{F})^\uparrow = H \cap P^*(\mathcal{F}^\uparrow).$$
In particular we have that $\chi^S \in P(\mathcal{F})^\uparrow \iff \chi^S \in P^*(\mathcal{F}^\uparrow)$. \end{claim}
\iffalse \begin{claim}
If we can separate over $P$ in polytime, then we can separate over the dominant $P^\uparrow$. \end{claim} \begin{proof}
Given a vector $y$, we can decide whether $y \in P^\uparrow$ by solving
\begin{align*}
x + s = y \\
x \in P \\
s \geq 0.
\end{align*}
Since can we easily separate over the first and third constraints, and a separation oracle for $P$ is given (i.e. we can also separate over the set of constraints imposed by the second line), it follows that we can separate over the above set of constraints in polytime. \qed \end{proof}
\begin{corollary}
If we can separate over $P(\mathcal{F})$, then we can separate over the dominant $P(\mathcal{F})^\uparrow \subseteq P^*(\mathcal{F})$. \end{corollary} \fi
We can now use this observation to prove the following.
\iffalse This yields the following mechanism for turning feasible sets in $\mathcal{F}^{\uparrow}$ into feasible sets in $\mathcal{F}$. \fi
\begin{lemma}
\label{lem:dominant-reduction}
Assume we have a separation algorithm for $P^*(\mathcal{F}^\uparrow)$. Then for any $\chi^{S} \in P^*(\mathcal{F}^\uparrow)$ we can find in polytime $\chi^{M} \in P(\mathcal{F})$ such that $\chi^{M} \leq \chi^{S}$. \end{lemma} \begin{proof}
Let $S=\{1,2,\ldots,k\}$. We run the following routine until no more elements can be removed:
\vspace*{10pt}
For $i \in S$\\
\hspace*{20pt} If $\chi^{S-i} \in P^*(\mathcal{F}^\uparrow)$ then $S=S-i$
\vspace*{10pt}
Let $\chi^M$ be the output. We show that $\chi^M \in P(\mathcal{F})$. Since $\chi^M \in P^*(\mathcal{F}^\uparrow)$, by Claim \ref{claim:lattice-points} we know that $\chi^M \in P(\mathcal{F})^\uparrow$. Then by definition of dominant there exists $x\in P(\mathcal{F})$ such that $x\leq \chi^M \in P(\mathcal{F})^\uparrow$. It follows that the vector $x$ can be written as $x = \sum_{i} \lambda_{i} \chi^{U_i}$ for some $U_i \in \mathcal{F}$ and $\lambda_i \in (0,1]$ with $\sum_i \lambda_i =1$. Clearly we must have that $U_i \subseteq M$ for all $i$, otherwise $x$ would have a non-zero component outside $M$. In addition, if for some $i$ we have $U_i \subsetneq M$, then there must exist some $j \in M$ such that $U_i \subseteq M-j \subsetneq M$. Hence $M-j \in \mathcal{F}^{\uparrow}$, and thus $\chi^{M-j} \in P(\mathcal{F})^\uparrow$ and $\chi^{M-j} \in P^*(\mathcal{F}^\uparrow)$. But then when component $j$ was considered in the algorithm above, we would have had $S$ such that $M \subseteq S$ and so $\chi^{S-j} \in P^*(\mathcal{F}^\uparrow)$ (that is $\chi^{S-j} \in P(\mathcal{F})^\uparrow$), and so $j$ should have been removed from $S$, contradiction. \qed \end{proof}
We point out that for many natural set families $\mathcal{F}$ we can work with the relaxation $P^*(\mathcal{F}^\uparrow)$ assuming that it admits a separation algorithm. Then, if we have an algorithm which produces $\chi^{F'} \in P^*(\mathcal{F}^\uparrow)$ satisfying some approximation guarantee for a monotone problem, we can use Lemma \ref{lem:dominant-reduction} to construct in polytime $F \in \mathcal{F}$ which obeys the same guarantee.
Moreover, notice that for Lemma \ref{lem:dominant-reduction} to work we do not need an actual separation oracle for $P^*(\mathcal{F}^\uparrow)$, but rather all we need is to be able to separate over $0-1$ vectors only. Hence, since the polyhedra $P^*(\mathcal{F}^\uparrow), \, P(\mathcal{F}^\uparrow)$ and $P(\mathcal{F})^\uparrow$ have the same $0-1$ vectors (see Claim \ref{claim:lattice-points}), a separation oracle for either $P(\mathcal{F}^\uparrow)$ or $P(\mathcal{F})^\uparrow$ would be enough for the routine of Lemma \ref{lem:dominant-reduction} to work. We now show that this is the case if we have a polytime separation oracle for $P(\mathcal{F})$. The following result shows that if we can separate efficiently over $P(\mathcal{F})$ then we can also separate efficiently over the dominant $P(\mathcal{F})^\uparrow$.
\iffalse We now show that we can also turn in polytime an integral solution $\chi^{F'} \in P(\mathcal{F}^\uparrow)$ into an integral solution $\chi^F \in P(\mathcal{F})$. We work from the point of view of the dominant $P(\mathcal{F})^\uparrow$. Notice that this is fine since by Claim \ref{claim:lattice-points} we have that $\chi^{F'} \in P(\mathcal{F}^\uparrow) \iff \chi^{F'} \in P(\mathcal{F})^\uparrow$. The following result shows that if we can separate efficiently over $P(\mathcal{F})$ then we can also separate efficiently over the dominant $P(\mathcal{F})^\uparrow$. \fi
\begin{claim}
If we can separate over a polyhedron $P$ in polytime, then we can also separate over its dominant $P^\uparrow$ in polytime. \end{claim} \begin{proof}
Given a vector $y$, we can decide whether $y \in P^\uparrow$ by solving
\begin{align*}
x + s = y \\
x \in P \\
s \geq 0.
\end{align*}
Since can we easily separate over the first and third constraints, and a separation oracle for $P$ is given (i.e. we can also separate over the set of constraints imposed by the second line), it follows that we can separate over the above set of constraints in polytime. \qed \end{proof}
Now we can apply the same mechanism from Lemma \ref{lem:dominant-reduction} to turn feasible sets from $\mathcal{F}^{\uparrow}$ into feasible sets in $\mathcal{F}$.
\begin{corollary}
\label{cor:dominant-reduction2}
Assume we have a separation algorithm for $P(\mathcal{F})^\uparrow$. Then for any $\chi^{S} \in P(\mathcal{F})^\uparrow$ we can find in polytime $\chi^{M} \in P(\mathcal{F})$ such that $\chi^{M} \leq \chi^{S}$. \end{corollary}
We conclude this section by making the remark that if we have an algorithm which produces $\chi^{F'} \in P(\mathcal{F}^\uparrow)$ satisfying some approximation guarantee for a monotone problem, we can use Corollary \ref{cor:dominant-reduction2} to construct $F \in \mathcal{F}$ which obeys the same guarantee.
\iffalse \begin{corollary}
Assume we have a separation algorithm for $P(\mathcal{F})^\uparrow$. Then for any $\chi^{S} \in P(\mathcal{F})^\uparrow$ we can find in polytime $\chi^{M} \in P(\mathcal{F})$ such that $\chi^{M} \leq \chi^{S}$. \end{corollary} \begin{proof}
This follows directly from Lemma \ref{lem:dominant-reduction} and the fact that $\chi^S \in P(\mathcal{F})^\uparrow \iff \chi^S \in P^*(\mathcal{F}^\uparrow)$. \qed \end{proof} \fi
\iffalse \begin{lemma}
Our results (i.e. black box reduction and bounded blockers argument) for minimization hold for well-behaved families $\mathcal{F}$, and not only for upwards-closed families. \end{lemma} \begin{proof}
So far the results are stated for upwards-closed families where a separation oracle is available for $P(\mathcal{F})^\uparrow$. If $\mathcal{F}$ is well-behaved, we have seen (Claim 1) that this is the case. Hence, we can find $\chi^{F'} \in P(\mathcal{F})^\uparrow$ satisfying some approximation guarantees. We now use Claim 2 to recover (in polytime) some $F
\in \mathcal{F}$ such that $\chi^F \leq \chi^{F'}$. By monotonicity the same approximation guarantees hold for this set $F$. \qed \end{proof} \fi
\iffalse We conclude this section by pointing out that for many natural set families $\mathcal{F}$ we can work with the relaxation $P^*(\mathcal{F}^\uparrow)$ assuming that it admits a separation algorithm. Then, if we have an algorithm which produces $\chi^{F'} \in P^*(\mathcal{F}^\uparrow)$ satisfying some approximation guarantee for a monotone problem, we can use Lemma \ref{lem:dominant-reduction} to construct $F \in \mathcal{F}$ which obeys the same guarantee. \fi \iffalse Alternatively, the reduction also shows that if we have a polytime separation oracle for $P(\mathcal{F})$ and an algorithm which produces $\chi^{F'}\in P(\mathcal{F})^\uparrow$ satisfying some approximation guarantee for a monotone problem, we can use \ref{} to construct $F \in \mathcal{F}$ which obeys the same guarantee. \fi
\section{Relaxations for constrained submodular minimization} \label{sec:relaxations}
Submodular optimization
techniques for minimization on a set family have involved two standard relaxations, one being linear \cite{goel2009approximability} and one being convex \cite{chekuri2011submodular,iwata2009submodular,iyer2014monotone}.
We introduce the latter in this section.
\iffalse
\subsection{Blocking Families and a Natural Relaxation for $\mathcal{P}(\mathcal{F})$}
We will be working with upwards-closed set families $\mathcal{F}'$, over a ground set $V$. which are {\em upwards-closed}, that is, if $F \subseteq F'$ and $F \in \mathcal{F}'$, then $F' \in \mathcal{F}'$. These are sometimes referred to as {\em blocking families}. Two of the families just discussed are blocking families - vertex-covers and cardinality-constrained families - whereas spanning trees are not.
For blocking families one normally works with the induced sub-family $\mathcal{F}$ of minimal sets. Then $\mathcal{F}$ has the property that it is a {\em clutter}, that is, $\mathcal{F}$ does not contain a pair of sets $F \subset F'$. The upwards-closed family determined by a clutter $\mathcal{F}$ is denoted by $\mathcal{F}^{\uparrow} = \{F: F \supseteq F', F' \in \mathcal{F}\}$. For such a clutter there is an associated {\em blocking} clutter $\mathcal{B}(\mathcal{F})$, which consists of the minimal sets $B$ such that $B \cap F \neq \emptyset$ for each $F \in \mathcal{F}$. We refer to $\mathcal{B}(\mathcal{F})$ as the {\em blocker} of $\mathcal{F}$. The following is straightforward. \begin{claim} \label{claim:blocker} \begin{enumerate} \item $F \in \mathcal{F}^{\uparrow}$ if and only if $F \cap B \neq \emptyset$ for all $B \in \mathcal{B}(\mathcal{F})$. \item $\mathcal{B}(\mathcal{B}(\mathcal{F}))) = \mathcal{F}$. \end{enumerate} \end{claim}
\noindent The significance of this is that one may assert membership in $\mathcal{F}^{\uparrow}$ by checking intersections on sets from the blocker $\mathcal{B}(\mathcal{F})$. That is $F \in \mathcal{F}$ if and only if $F \cap B \neq \emptyset$ for each $B \in \mathcal{B}(\mathcal{F})$.
This simple observation gives a starting point for exploring the integral polyhedron $P(\mathcal{F}) = \{ z \in \mathbb{R}^V: z \geq \chi^F \textit{for some $F \in \mathcal{F}$}\}$. The blocking condition gives rise to the {\em natural linear relaxation} for $P(\mathcal{F})$: \begin{equation} \label{eqn:polyhedron} P^*(\mathcal{F}) = \{z \in \mathbb{R}^V_{\geq 0}: z(B) \geq 1~~\forall ~B \in \mathcal{B}(\mathcal{F})\} \end{equation}
\noindent and so $F \in \mathcal{F}$ if and only if $\chi^F \in P^*(\mathcal{F})$.
\fi
\subsection{A convex relaxation} \label{sec:LE-def}
We will be working with upwards-closed set families $\mathcal{F}$, and their blocking relaxations $P^*(\mathcal{F})$.
As we now work with arbitrary vectors $z \in [0,1]^n$, we must specify how our objective function $f(S)$ behaves on all points $z \in P^*(\mathcal{F})$. Formally, we call $g:[0,1]^V \rightarrow \mathbb{R}$ an {\em extension} of $f$ if $g(\chi^S) = f(S)$ for each $S \subseteq V$. For a submodular objective function $f(S)$ there can be many extensions of $f$ to $[0,1]^V$ (or to $\mathbb{R}^V$). The most popular one has been the so-called {\em Lov\'asz Extension} (introduced in \cite{lovasz1983submodular}) due to several of its desirable properties.
We present two of several equivalent definitions for the Lov\'asz Extension. Let $0 < v_1 < v_2 < ... < v_m \leq 1$ be the distinct positive values taken in some vector $z \in [0,1]^V$. We also define $v_0=0$ and $v_{m+1}=1$ (which may be equal to $v_m$). Define for each $i \in \{0,1,...,m\}$ the set $S_i=\{ j: z_j > v_i\}$. In particular, $S_0$ is the support of $z$ and $S_m=\emptyset$. One then defines: \[ f^L(z) = \sum_{i=0}^{m} (v_{i+1}-v_i) f(S_i). \] It is not hard to check that the following is an equivalent definition of $f^L$ (e.g. see \cite{chekuri2011submodular}). For a vector $z \in [0,1]^V$ and $\theta \in [0,1]$, let $z^\theta := \{v \in V: z(v)\geq \theta\}$. We then have that $$f^L(z) = \mathbb{E}_{\theta \in [0,1]} f(z^\theta) = \int_{0}^{1} f(z^\theta) d\theta,$$ where the expectation is taken over the uniform distribution in $[0,1]$.
\begin{lemma} [Lov\'asz \cite{lovasz1983submodular}] The function $f^L$ is convex if and only if $f$ is submodular. \end{lemma}
One could now attack constrained submodular minimization by solving the problem \begin{equation} \label{SA-LE-appendix} (\mbox{SA-LE}) ~~~~ \min f^L(z): z \in P^*(\mathcal{F}), \end{equation} and then seek rounding methods for the resulting solution. This is the approach used in \cite{chekuri2011submodular,iwata2009submodular,iyer2014monotone}. We refer to the above as the {\em single-agent Lov\'asz extension formulation}, abbreviated as (SA-LE).
\iffalse We will later use two additional properties of the Lov\'asz Extension. We say that a function $g$ is (positively) {\em homogeneous} if for each $c \geq 0$, we have $g(cz) = c g(z)$.
\begin{lemma} \label{lem:homogext} The Lov\'asz Extension of $f$ is homogeneous if $f$ is submodular and normalized (i.e. $f(\emptyset)=0$). That is for any $x \in [0,1]^n$ and $\alpha \in [0,1]$ $f^L(\alpha x)=\alpha f^L(x)$. In particular, for any set $Q$ we have $f^L((\alpha) \chi^Q) = \alpha f(Q)$. \end{lemma} \begin{proof}
This follows by construction. Recall that \[ f^L(x) = \sum_{i=0}^{m} (v_{i+1}-v_i) f(S_i) \] where $v_0 = 0$, $v_{m+1} = 1$, and $0 \leq v_1 < v_2 < ... < v_m \leq 1$ are the distinct values taken in the vector $x \in [0,1]^n$. The sets $S_i$ are defined as $S_i=\{ j: x_j > v_i\}$ for each $i \in \{0,1,...,m\}$. It follows directly that \[ f^L(\alpha x) = \sum_{i=0}^{m-1} (\alpha) (v_{i+1}-v_i) f(S_i) + (1-\alpha v_m) f(\emptyset) = \alpha f^L(x) \] \qed \end{proof}
\begin{lemma} \label{lem:monotoneext} If $f(S)$ is a monotone set function, then $f^L$ is a monotone function. \end{lemma} \fi
\iffalse This can be used to establish the following. \begin{corollary} \label{cor:convexbound} Let $g:[0,1]^V \rightarrow \mathbb{R}$ be a convex, monotonic, homogeneous function. Suppose that $z \leq \sum_S x(S) \chi^S$, then $g(z) \leq \sum_S x(S) g(\chi^S)$. In particular, if $f^L$ is the extension of a normalized, monotonic, submodular function $f(S)$ and $\chi^Z \leq \sum_S x(S) \chi^S$, then $f(Z) \leq \sum_S x(S) f(S)$. \end{corollary} \begin{proof} We prove the latter statement, the first being similar. Without loss of generality, $C:=\sum_{S\subseteq V} x(S)>0$. Hence: \[ \begin{array}{cclr} f(Q) & = & f^L(\chi^{Q}) & (\mbox{by definition})\\
& = & \frac{C}{C}f^L(\chi^{Q})\\
& = & Cf^L(\frac{1}{C}\chi^{Q}) & (\mbox{by homogeneity})\\
& \leq & Cf^L(\frac{1}{C}\sum_{S\subseteq V} x(S)\chi^{S}) & (\mbox{by monotonicity})\\
& \leq & C\sum_{S\subseteq V}(\frac{x(S)}{C})f^L(\chi^{S}) & (\mbox{by convexity})\\
& = & \sum_{S\subseteq V}x(S)f^L(\chi^{S})\\
& = & \sum_{S\subseteq V} x(S)f(S). & (\mbox{by definition}). \end{array} \] \qed \end{proof} \fi
\iffalse \subsection{An extended LP formulation} \label{sec:extended-lp}
By Claim~\ref{claim:blocker} (see Appendix \ref{sec:blocking}), an integer programming formulation for the single-agent problem is \begin{equation} \label{eqn:X1} \begin{array}{cc} \min & \sum_{S\subseteq V}f(S)x(S)\\ \mbox{s.t.} & \sum_{v\in B}\sum_{S\ni v}x(S)\geq1,\,\forall B\in\mathcal{B}(\mathcal{F})\\
& x(S)\geq0. \end{array} \end{equation} The corresponding dual is given by
\begin{equation} \label{eqn:Y1} \begin{array}{cc} \max & \sum_{B\in\mathcal{B}(\mathcal{F})}y_{B}\\ \mbox{s.t.} & \sum_{v\in S}\sum_{B\ni v}y_{B}\leq f(S),\,\forall S\subseteq V\\
& y_{B}\geq0. \end{array} \end{equation}
Using the ideas from \cite{goel2009approximability}, one may solve these LPs in polytime if the blocking family is not too large.
\begin{lemma} \label{lem:polybound}
If $f$ is submodular and $|\mathcal{B}(\mathcal{F})| \in poly(n)$, then the linear programs (\ref{eqn:X1},\ref{eqn:Y1}) can be solved in polytime. \end{lemma} \iffalse \begin{proof} It is well-known \cite{grotschel2012geometric} that a class of LPs can be solved in polytime if (i) it admits a polytime separation algorithm
and (ii) has a polynomially bounded number of variables. Moreover, if this holds, then the class of dual problems can also be solved in polytime. Hence we focus on the dual (\ref{eqn:Y1}) for which
condition (ii) is assured by our assumption on $|\mathcal{B}(\mathcal{F})|$. Let's now consider its separation problem.
For a fixed vector $y \geq 0$ we define $z_{y}(S):=\sum_{v\in S}\sum_{v\in B} y_{B}$ for any $S\subseteq V$. Notice that $z_{y}$ is a modular function. Hence, $f-z_{y}$ is submodular and so we can solve the problem $\min_{S\subseteq V}f(S)-z_{y}(S)$ exactly in polytime. It follows that $y$ is feasible if and only if this minimum is nonnegative. Thus, we can separate in the dual efficiently. \qed \end{proof} \fi
The restriction that $|\mathcal{B}(\mathcal{F})| \in poly(n)$ is quite severe and we now show how it may be weakened.
\fi
\subsection{Tractability of the single-agent formulation (SA-LE)} \label{sec:combined}
In this section we show that one may solve (SA-LE) approximately as long as a polytime separation algorithm for $P^*(\mathcal{F})$ is available. This is useful in several settings and in particular for our methods which rely on the multi-agent Lov\'asz extension formulation (discussed in the next section).
\iffalse Recall from Appendix~\ref{sec:block} that $P^*(\mathcal{F}) = \{z \in \mathbb{R}^V_{\geq 0}: z(B) \geq 1~~\forall B \in \mathcal{B}(\mathcal{F}^{min})\}$. To do this, we work with a combination of the LP and the convex relaxation from the prior section. For a feasible solution $x$ to (\ref{eqn:X1}), we define its {\em image} as \[ z = \sum_S x(S) \chi^S. \] We sometimes work with an extended formulation \[ Q = \{(x,z): \mbox{$z$ is the image of feasible $x$ for (\ref{eqn:X1})}\}. \] One establishes easily that $P^*(\mathcal{F})$ is just a projection of $Q$. In fact, the convex and LP relaxations are equivalent in the following sense, due to the pleasant properties of the Lov\'asz Extension.
\begin{lemma} \label{lem:equivopts}
\[ \min \{ f^L(z): z \in P^*(\mathcal{F})\} = \min \{ f^L(z): (x,z) \in Q \} = \min \{ \sum_S x(S) f(S): \mbox{$(x,z) \in Q$} \}. \]
Moreover, given a feasible solution $z^* \in P^*(\mathcal{F})$ we may compute $x^*$ such that $(x^*,z^*) \in Q$ and $\sum_S x^*(S) f(S) = f^L(z^*)$. This may be done in time polynomially bounded in $n$ and the encoding size of $z^*$. \end{lemma} \begin{proof} The first equality follows from the preceding lemma. For the second equality we consider $(x,z) \in Q$. By Corollary~\ref{cor:convexbound}, we have that $f^L(z) \leq \sum_S x(S) f(S)$. Hence $\leq$ holds. Conversely, suppose that $z^*$ is an optimal solution for the convex optimization problem and let $x^*$ be determined by the ``level sets'' associated with $z^*$ in the definition of $f^L$. By definition we have $\sum_S x^*(S) f(S) = f^L(z^*)$. Moreover, one easily checks that since $z^*(B) \geq 1$ for each blocking set $B$, we also have that $x^*$ is feasible for (\ref{eqn:X1}). Hence $\geq$ holds as well. \qed \end{proof}
We make essential use of the following theorem as follows. It shows that the optimal solution for the convex programme coincides with that of a linear program. The latter LP, however, has optimal solutions have polybounded encoding sizes. This means that a polytime weak optimization routine (as given by Ellipsoid) can be turned into an polytime exact optimization routine for the convex programme (and hence by the Lemma, also for the LP).
We discuss this in further detail below. \fi
{\bf Polytime Algorithms.} One may apply the Ellipsoid Method to obtain a polytime algorithm which approximately minimizes a convex function over a polyhedron $K$ as long as various technical conditions hold. For instance, one could require that there are two ellipsoids $E(a,A) \subseteq K \subseteq E(a,B)$ whose encoding descriptions are polynomially bounded in the input size for $K$. We should also have polytime (or oracle) access to the convex objective function defined over ${\bf R}^n$. In addition, one must be able to polytime solve the subgradient problem for $f$.\footnote{For a given $y$, find a subgradient of $f$ at $y$.} One may check that the subgradient problem is efficiently solvable for Lov\'asz extensions of polynomially encodable submodular functions. We call $f$ {\em polynomially encodable} if the values $f(S)$ have encoding size bounded by a polynomial in $n$ (we always assume this for our functions). If these conditions hold, then methods from \cite{grotschel2012geometric} imply that for any $\epsilon >0$ we may find
an approximately feasible solution for $K$ which is approximately optimal. By approximate here we mean
for instance that the objective value is within $\epsilon$ of the real optimum. This can be done in time polynomially bounded in $n$ (size of input say) and $\log \frac{1}{\epsilon}$. Let us give a few details for our application.
\iffalse In our setting, the rightmost LP from the previous lemma is essentially identical to the linear program (\ref{eqn:X1}); it just works over $Q$ which is like carrying some extra meaningless variables $z$. The preceding result says that we can work instead over the more compact (in terms of variables) space $P^*(\mathcal{F})$ at the price of using a convex objective function. \fi Our convex problem's feasible space is $P^*(\mathcal{F})$ and it is easy to verify that our optimal solutions will lie in the $0-1$ hypercube $H$. So we may define the feasible space to be $H$ and the objective function to be $g(z)=f^L(z)$ if $z \in H \cap P^*(\mathcal{F})$ and $=\infty$ otherwise. (Clearly $g$ is convex in ${\bf R}^n$ since it is a pointwise maximum of two convex functions; alternatively, one may define the Lov\'asz Extension on ${\bf R}^n$ which is also fine.) Note that $g$ can be evaluated in polytime by the definition of $f^L$ as long as $f$ is polynomially encodable. We can now easily find an ellipsoid inside $H$ and one containing $H$ each of which has poly encoding size. We may thus solve the convex problem to within $\pm \epsilon$-optimality in time bounded by a polynomial in $n$ and $\log \frac{1}{\epsilon}$. \iffalse Note that our lemmas guarantee to produce $(x^*,z^*) \in Q$ so we have exact feasibility. \fi
\begin{corollary}
\label{cor:ellipsoid}
Consider a class of problems $\mathcal{F}, f$ for which $f$'s are submodular and polynomially-encodable in $n=|V|$.
If there is a polytime separation algorithm for the family of polyhedra $P^*(\mathcal{F})$,
then the convex program (SA-LE) can be solved to accuracy of $\pm \epsilon$ in time bounded
by a polynomial in $n$ and $\log \frac{1}{\epsilon}$. \end{corollary}
\iffalse
One final comment about exact solvability of these problems.
\iffalse
Since our convex problem is equivalent to an LP (over $Q$ with nonzero coefficients on the $x$-variables), we know that an optimal solution $(x^*,z^*)$ will be determined at a vertex of the polyhedron $Q$.
\fi
\textcolor{red}{It is straightforward to check by contradiction that $z^*$ is in fact a vertex of $P^*(\mathcal{F})$.}
As any such vertex is defined by an $n \times n$ nonsingular system, and hence by Cramer's Rule, the encoding size of $z^*$ is polynomially bounded (in $n$).
\iffalse But then $x^*$ could also be viewed as a vertex to a polyhedron of the form: $\{x: A x = z^*, \, x \geq 0 \}$. The support of any such vertex is again determined by a nonsingular linear system with at most $n$ rows. As $z^*$ is polybounded, so is the encoding size of $x^*$. \fi Since $f$ and $z^*$ are {\em polynomially-encodable}, we have the the optimal solution to (SA-LE) is polynomially bounded. Hence
we can pre-select a polybounded $\epsilon > 0$ which guarantees us to find the exact optimum, not just an $\epsilon$-optimal one. Thus we have the following.
\begin{corollary}
Consider a class of problems $\mathcal{F}, f$ for which $f$'s are submodular and polynomially-encodable in $n=|V|$. If there is a polytime separation algorithm for the family of polyhedra $P^*(\mathcal{F})$,
then one may find an exact optimum $z^*$ for the convex program (SA-LE) in polytime. \end{corollary}
\fi
\iffalse FBS: did not get to this.
The function $f^L$ is indeed well-behaved and this actually follows from Lemma~\ref{lem:equivopts}. Since the lefthand optimum is equal to the righthand, we know that the encoding complexity of the lefthand optimal solution is bounded by that of the LP. ...{\bf not clear.} I thought it was standard Cramer's Rule, but this may be trickier than I thought. I was hoping we would get a lemma that could replace Lemma \ref{lem:polybound}. \fi
\subsection{The multi-agent formulation} \label{sec:ma extensions}
The single-agent formulation (SA-LE) discussed above has a natural extension to the multi-agent setting. This was already introduced in \cite{chekuri2011submodular} for the case $\mathcal{F}=\{V\}$. \begin{equation} \label{MA-LE-appendix} (\mbox{MA-LE}) ~~~~ \min \sum_{i=1}^k f^L_i(z_i): z_1 + z_2 + \cdots + z_k \in P^*(\mathcal{F}). \end{equation} We refer to the above as the {\em multi-agent Lov\'asz extension formulation}, abbreviated as (MA-LE).
We remark that we can solve (MA-LE) as long as we have polytime separation of $P^*(\mathcal{F})$. This follows the approach from the previous section (see Corollary \ref{cor:ellipsoid}) except our convex program now has $k$ vectors of variables $z_1,z_2, \ldots ,z_k$ (one for each agent) such that $z=\sum_i z_i$. \iffalse The extended formulation now forces $z \in P^*(\mathcal{F})$, and we also have constraints forcing that each $z_i$ is the image of an associated $(x_i(S))$ vector, i.e. $z_i = \sum_S x(S,i)\chi^S$. As in the SA case, for any feasible $z^*=z_1^* + \cdots + z_k^* \in P^*(\mathcal{F})$ we can find in polytime an $x^*$ that is feasible for (\ref{MA-LP1}) and such that $\sum_i f^L_i(z_i^*)=\sum_i \sum_S f_i(S) x^*(S,i)$. This uses the same plan as for Lemma \ref{lem:equivopts}. Namely, given a vector $z_i$ we can use its level sets to determine an associate ``pre-image'' vector $(x(S,i): S \subseteq V)$.
Our arguments for tractability of the multi-agent formulation also depend on having a polytime weak optimization routine for the problem $\min \sum_i f^L_i(z_i): z_1 + z_2 + \ldots + z_k \in P^*(\mathcal{F})$. \fi This problem has the form $\{ \min g(w) : w \in W \subseteq {\bf R}^{nk} \}$ where $g$ is convex and $W$ is the full-dimensional convex body $\{w=(z_1,...,z_k): \sum_i z_i \in P^*(\mathcal{F})\}$. Clearly we have a polytime separation routine for $W$, and hence we may apply Ellipsoid as in the single-agent case.
\begin{corollary}
\label{cor:sep-sol-LP}
Assume there is a polytime separation oracle for $P^*(\mathcal{F})$. Then we can solve the multi-agent formulation (MA-LE) in polytime. \end{corollary}
\begin{corollary}
\label{cor:SA-MA-LP}
Assume we can solve the single-agent formulation (SA-LE) in polytime. Then we can also solve the multi-agent formulation (MA-LE) in polytime. \end{corollary} \begin{proof}
If we can solve (SA-LE) in polytime then we can also separate over $P^*(\mathcal{F})$ in polynomial time. Now the statement follows from Corollary~\ref{cor:sep-sol-LP}. \qed \end{proof}
\section{Dealing with some special types of nonmonotone objectives} \label{sec:nonmonotone}
We present the proof of Theorem \ref{thm:sym-MSCA} discussed in Section \ref{sec:SA-MA-formulations}. \iffalse We present a detailed proof of the results discussed on Section \ref{sec:BBnonmonotone} for nonmonotone objectives for completeness. The proof follows closely the arguments presented in \cite{chekuri2011submodular} (for the symmetric case) and in \cite{ene2014hardness} (for the general case). Both of these papers worked in the setting where $\mathcal{F} = \{V\}$. \fi
\iffalse \begin{theorem}
\label{thm:sym-MSCA-appendix}
Consider an instance of (MA-LE) where $\mathcal{F}$ is an upwards closed family and $f_i = g_i + h$ where the $g_i$ are nonnegative monotone submodular and $h$ is nonnegative symmetric submodular. Let $z_1+z_2+\cdots+z_k$ be a feasible solution such that $\sum_{i \in [k]} z_i \geq \chi^U$ for some $U \in \mathcal{F}$.
Then there is a randomized rounding procedure that outputs an integral feasible solution
$\bar{z}_1+\bar{z}_2+\cdots+\bar{z}_k$ such that $\sum_{i \in [k]} \bar{z}_i \geq \chi^U$ and $\sum_{i\in [k]} f^L_i (\bar{z}_i) \leq O(\log n) \sum_{i \in [k]} f_i^L(z_i)$ on expectation. That is, we get a subpartition $S_1,S_2,\ldots,S_k$ such that $\biguplus_{i \in [k]} S_i \supseteq U$ and $\sum_{i \in [k]}f_i(S_i) \leq O(\log n) \sum_{i \in [k]} f_i^L(z_i)$
on expectation. \end{theorem} \fi \begin{proof}[Theorem \ref{thm:sym-MSCA}]
Let $z = \sum_{i \in [k]} z_i$ be a feasible solution to (MA-LE) and such that $\sum_{i \in [k]} z_i \geq \chi^U$ for some $U \in \mathcal{F}$.
Consider the below CE-Rounding procedure originally described in the work of \cite{chekuri2011submodular} for the case $\mathcal{F}=\{V\}$.
\begin{algorithm}[ht]
\caption{CE-Rounding}
\label{alg:CE-rounding}
$S \leftarrow \emptyset$ \quad /* set of assigned elements */\\
$S_i \leftarrow \emptyset$ for all $i \in [k]$ /* set of elements assigned to agent $i$ */\\
\While{$S \notin \mathcal{F}$}{
Pick $i\in [k]$ uniformily at random\\
Pick $\theta \in [0,1]$ uniformily at random \\
$S(i,\theta):= \{v \in V: z_i(v)\geq \theta\}$ \\
$S_i \leftarrow S_i \cup S(i,\theta) $\\
$S \leftarrow S \cup S(i,\theta) $\\
}
/* Uncross $S_1,S_2,\ldots,S_k$ */\\
$S'_i \leftarrow S_i$ for all $i \in [k]$\\
\While{there exist $i \neq j$ such that $S'_i \cap S'_j \neq \emptyset$}{
\eIf{$h(S'_i) + h(S'_j - S'_i) \leq h(S'_i) + h(S'_j)$}{
$S'_j \leftarrow S'_j - S'_i$\\}{
$S'_i \leftarrow S'_i - S'_j$\\}
}
Output $(S'_1,S'_2,\ldots,S'_k)$
\end{algorithm}
It is discussed in \cite{chekuri2011submodular} and not difficult to see
that the first while loop assigns all the elements from $U$ in $O(k \log |U|)$ iterations with high probability.
Since $\mathcal{F}$ is upwards closed and $U \in \mathcal{F}$, this implies that the first while loop terminates in
$O(k \log |U|)$ iterations with high probability.
Moreover, it is clear that the uncrossing step takes a polynomial number of iterations.
Let $S_1,S_2,\ldots,S_k$ be the output after the first while loop and $S'_1,S'_2,\ldots,S'_k$ the final output of the rounding. At each iteration of the first while loop, the expected cost associated to the random set $S(i,\theta)$ is given by $$\mathbb{E}_{i,\theta}[f_i(S(i,\theta))] = \frac{1}{k} \sum_{i=1}^k \mathbb{E}_{\theta}[f_i(S(i,\theta))]= \frac{1}{k} \sum_{i=1}^k f^L_i (z_i).$$ Hence, given the subadditivity of the objectives (since the functions are submodular and nonnegative), the expected cost increase at each iteration of the first while loop is upper bounded by $\frac{1}{k} \sum_{i=1}^k f^L_i (z_i)$. Since the first while loop terminates w.h.p. in $O(k \log |U|)$ iterations, it now follows by linearity of expectation that the total expected cost of $\sum_i f_i(S_i)$ is at most $O(\log |U|) \sum_{i=1}^k f^L_i (z_i)$.
Finally, we use a result (see Lemma 3.1) from \cite{chekuri2011submodular} that guarantees that if $h$ is symmetric submodular then the uncrossing step of the rounding satisfies $\sum_i h(S'_i) \leq \sum_i h(S_i)$. Moreover, by monotonicity of the $g_i$ it is also clear that $\sum_i g_i(S'_i) \leq \sum_i g_i(S_i)$. Thus, we have that $S'_1 \uplus S'_2 \uplus \cdots \uplus S'_k = S \in \mathcal{F}$ and $\sum_i f_i(S'_i) \leq \sum_i f_i(S_i) \leq O(\log |U|) \sum_{i=1}^k f^L_i (z_i)$ on expectation. This concludes the argument.
\qed \end{proof}
\iffalse \begin{theorem} Let $\mathcal{F}$ be an upwards closed family with a $\beta$-bounded blocker. Let the objectives be of the form $f_i = g_i + h$ where each $g_i$ is nonnegative monotone submodular. Then we have the following results for the associated multi-agent problem. \begin{itemize}
\item[\bfseries{(1)}] There is a randomized $O(\beta\log n)$-approximation algorithm in the case where
$h$ is nonnegative symmetric submodular.
\item[\bfseries{(2)}] There is a randomized $O(k \beta\log n)$-approximation algorithm in the case where
$h$ is nonnegative submodular. \end{itemize} If $P^*(\mathcal{F})$ has a polytime separation oracle, then these are polytime algorithms.
\end{theorem} \begin{proof}
\textbf{(1)}.
Let $z^* = z^*_1 + z^*_2 + \cdots + z^*_k$ be an optimal solution to (MA-LE) with value $OPT_{frac}$.
Consider the vector $\beta z^* = \sum_{i\in [k]} \beta z^*_i$ and let $U=\{v\in V: \beta z^*(v) \geq 1\}$.
Since $\mathcal{F}$ has a $\beta$-bounded blocker it follows that $U\in\mathcal{F}$.
We now use the following rounding procedure described in the work of \cite{chekuri2011submodular}.
\begin{algorithm}[ht]
\label{alg:CE-rounding}
$S \leftarrow \emptyset$ \quad /* set of assigned elements */\\
$S_i \leftarrow \emptyset$ for all $i \in [k]$ /* set of elements assigned to agent $i$ */\\
\While{$S \notin \mathcal{F}$}{
Pick $i\in [k]$ uniformily at random\\
Pick $\theta \in [0,1]$ uniformily at random \\
$S(i,\theta):= \{v \in V: \beta z^*_i(v)\geq \theta\}$ \\
$S_i \leftarrow S_i \cup S(i,\theta) $\\
$S \leftarrow S \cup S(i,\theta) $\\
}
/* Uncross $S_1,S_2,\ldots,S_k$ */\\
$S'_i \leftarrow S_i$ for all $i \in [k]$\\
\While{there exist $i \neq j$ such that $S'_i \cap S'_j \neq \emptyset$}{
\eIf{$h(S'_i) + h(S'_j - S'_i) \leq h(S'_i) + h(S'_j)$}{
$S'_j \leftarrow S'_j - S'_i$\\}{
$S'_i \leftarrow S'_i - S'_j$\\}
}
Output $(S'_1,S'_2,\ldots,S'_k)$
\end{algorithm}
It is discussed in \cite{chekuri2011submodular} and not difficult to see
that the first while loop assigns all the elements from $U$ in $O(k \log n)$ iterations with high probability.
Since $\mathcal{F}$ is upwards closed and $U \in \mathcal{F}$, this implies that the first while loop terminates in
$O(k \log n)$ iterations with high probability.
Moreover, it is clear that the uncrossing step takes a polynomial number of iterations.
Let $S_1,S_2,\ldots,S_k$ be the output after the first while loop and $S'_1,S'_2,\ldots,S'_k$ the final output of the rounding. At each iteration of the first while loop, the expected cost associated to the random set $S(i,\theta)$ is given by $$\mathbb{E}_{i,\theta}[f_i(S(i,\theta))] = \frac{1}{k} \sum_{i=1}^k \mathbb{E}_{\theta}[f_i(S(i,\theta))]= \frac{1}{k} \sum_{i=1}^k f^L_i (\beta z^*_i) = \frac{1}{k} \sum_{i=1}^k \beta f^L_i ( z^*_i) = \frac{\beta}{k} OPT_{frac}.$$ Hence, given the subadditivity of the objectives (since the functions are submodular and nonnegative), the expected cost increase at each iteration of the first while loop is upper bounded by $\frac{\beta}{k} OPT_{frac}$. Since the first while loop terminates w.h.p. in $O(k \log n)$ iterations, it now follows by linearity of expectation that the total expected cost of $\sum_i f_i(S_i)$ is at most $O(\beta \log n) OPT_{frac}$.
Finally, we use a result (see Lemma 3.1) from \cite{chekuri2011submodular} that guarantees that if $h$ is symmetric submodular, then the uncrossing step of the rounding satisfies $\sum_i h(S'_i) \leq \sum_i h(S_i)$. Moreover, by monotonicity of the $g_i$ it is also clear that $\sum_i g_i(S'_i) \leq \sum_i g_i(S_i)$. Thus, we have that $S'_1 \uplus S'_2 \uplus \cdots \uplus S'_k = S \in \mathcal{F}$ and $\sum_i f_i(S'_i) \leq \sum_i f_i(S_i) \leq O(\beta \log n) OPT_{frac}$. This concludes the argument.
\textbf{(2)}. For the case where $h$ is just nonnegative submodular we use a result (see Proposition 10) from \cite{ene2014hardness}. Given a nonnegative submodular function $h$, they consider the nonnegative symmetric submodular function defined as $h'(S) = h(S) + h(V-S)$. They then show that for any partition $S_1,S_2,\ldots,S_k$ the function $h'$ satisfies that $\sum_i h'(S_i) \leq k \sum_i h(S_i)$.
Now, given a nonnegative submodular function $h$, we can get a randomized $O(\beta \log n)$-approximation via \textbf{(1)} for the problem with $f_i = g_i + h'$. Moreover, by the above observation from \cite{ene2014hardness} it follows that then this is a randomized $O(k \beta \log n)$-approximation for the original problem with $f_i = g_i + h$.
We point out that in subsequent work (\cite{mirzakhani2014sperner}) it was shown that the $k$ factor loss due to the non-symmetry is in fact necessary. \qed \end{proof} \fi
\iffalse \section{Applications of bounded blockers \textcolor{red}{I WOULD STRONGLY REMOVE THIS FROM THIS SUMBISSION. WE AGREED THAT WE WANT TO TONE DOWN BOUNDED BLOCKERS AS MUCH AS POSSIBLE, SO I THINK THIS SHOULD GO. IT WAS ORIGINALLY ADDED TO THE PAPER IN ORDER TO ARGUE HOW OUR ORIGINAL SA BOUNDED BLOCKER RESULT COULD BE USEFUL, BUT NOW THAT WE KNOW OTHER PEOPLE HAVE DONE ALREADY AND THE BOUNDED BLOCKER NOTION HAS BEEN ALREADY DISCUSSED OUT THERE, I DO NOT SEE ANY REASON WHY THIS SHOULD STAY.}} \label{sec:BB-applications}
We discuss several problems which can be addressed by bounded blockers. We start with some of the simplest examples which also indicate how one might find uses for Theorem~\ref{thm:device}. We later discuss applications to multi-agent minimization.
\begin{example}[{\bf Edge Covers}] {\em Let $G=(V,E)$ be an undirected graph and $\mathcal{F}$ denote the family
of edge covers of $G$. It follows that $\mathcal{B}(\mathcal{F})$ consists of the $n$ sets $\delta(v)$ for each $v \in V$.
Theorem~\ref{thm:device} immediately implies that
that the minimum submodular edge cover problem is $\Delta$-approximable
for graphs of max degree $\Delta$. In particular, the $\Omega(n/log(n))$ inapproximability result
from \cite{iwata2009submodular} is pinpointed on graphs with at least one high-degree node.
} \end{example}
This example is not surprising but gives clarity to the message that whenever the blocker $\mathcal{B}(\mathcal{F})$ induces a bounded-degree hypergraph, Theorem~\ref{thm:device} applies. Let's explore a few questions which can be approached like this. In \cite{svitkina2010facility} an $O(\ln n)$-approximation is given for submodular facility location, but for instances where the number of customers $n$ dwarfs the number of facilities $k$ the following is relevant.
\begin{example}[{\bf Facility Location}]{\em Suppose we have
a submodular facility location problem with $k$ facilities and $n$ customers.
We may represent solutions as subsets of edges in
a $k \times n$ bipartite graph $G$ which include exactly one edge incident to each
customer node. As our submodular functions
are monotone, we can work with the associated blocking family $\mathcal{F}$ for
which $\mathcal{B}(\mathcal{F})$ consists of the edges $\delta_G(v)$, where $v$ is a customer node.
A $k$-approximation now follows from Theorem~\ref{thm:device}. } \end{example}
We consider the several examples from the broad task of finding {\em low-cost dense networks}.
\begin{example}[{\bf Lightly Pruned Networks}] {\em Suppose we seek low-cost subgraphs
of some graph $G$ where nodes can only have a limited number $\tau$ of their edges removed. In other words,
the degree of
each node $v$ should remain at least $deg(v)-\tau$.
The family of subgraphs (viewed as edge subsets) $\mathcal{F}$ has as its
blocker the set of all sub-stars of size $\tau+1$:
$\mathcal{B}(\mathcal{F})=\{E' \subseteq \delta(v): |E'| = \tau+1, v \in V(G)\}$.
When $\tau$ is small (constant or polylog) one obtains sensible
approximations.
This can be extended to notions of density which incorporate some tractable family of cuts $\delta(S) \in \mathcal{C}$. That is,
where the designed subgraph must retain all but a bounded number $\tau$ of edges from each $\delta(S)$.} \end{example}
\begin{example}[{\bf Girth-Widening}]{\em
For an undirected graph $G$ we seek to minimize $f(J)$ where $J$ is subset of edges
for which $G-J$ does not have any short cycles. For instance, given a threshold $\tau$ we want
$G-J$ to not have any cycles of length at most $\tau$. We seek $J \in \mathcal{F}$ whose
blocker consists of all cycles with at most $\tau$ edges. The separation can be solved in this setting since
constrained shortest paths has a polytime algorithm if one of the two metrics is unit-cost. } \end{example}
Decomposing (or clustering) graphs into lower diameter subgraphs is a scheme which occasionally arises (for instance, the tree-embedding work of \cite{fakcharoenphol2003tight} employs a diameter-reduction step). We close with an example in this general vein.
\begin{example}[{\bf Diameter-Reduction}]
{\em Suppose we would like to find a submodular-sparse cut in a graph $G$ which decomposes it into subgraphs
$G_1,G_2, \ldots$ each of which has a bounded diameter $\tau$.
We are then looking at a submodular minimization problem whose bounded blocker consists of paths
of length exactly $\tau+1$. \\} \end{example} \fi
\section{Invariance under the lifting reduction} \label{sec:Appendix-Invariance}
\begin{corollary}
\label{Corollary p-matroid intersection}If $(V,\mathcal{F})$ is a p-matroid
intersection, then so is $(E,\mathcal{F}')$.\end{corollary} \begin{proof}
Let $\mathcal{F}=\cap_{i=1}^{p}\mathcal{I}_{i}$ for some matroids $(V,\mathcal{I}_{i})$. Then
we have that
\begin{align*}
\mathcal{F}'= & \{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{F}\}\\
= & \{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\cap_{i=1}^{p}\mathcal{I}_{i}\}\\
= & \{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{I}_{i},\forall i\in[p]\}\\
= & \bigcap_{i\in[p]}\{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{I}_{i}\}\\
= & \bigcap_{i\in[p]}\mathcal{I}_{i}'.
\end{align*}
Moreover, from Corollary \ref{cor:matroid-invariant} we know that
$(E,\mathcal{I}_{i}')$ is a matroid for each $i\in[p]$, and the result follows. \qed \end{proof}
\iffalse \begin{corollary}
\label{Corollary bases-}Assume $\mathcal{F}$ is the set of bases of some matroid
$\mathcal{M}=(V,\mathcal{I})$, then $\mathcal{H}$ is the set of bases of some matroid $\mathcal{M}'=(E,\mathcal{I}')$. \end{corollary} \begin{proof}
Let $\mathcal{F}=\mathcal{B}(\mathcal{M})$ and define $\mathcal{I}'=\{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{I}\}.$
From Corollary \ref{cor:matroid-invariant} we know that $\mathcal{M}'=(E,\mathcal{I}')$
is a matroid. We show that $\mathcal{H}=\mathcal{B}(\mathcal{M}')$. Let $r_{\mathcal{M}}$ and $r_{\mathcal{M}'}$
denote the rank of $\mathcal{M}$ and $\mathcal{M}'$ respectively. Then notice that
\begin{align*}
\mathcal{H}= & \{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{F}\}\\
= & \{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{B}(\mathcal{M})\}\\
= & \{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}=I\in\mathcal{I},|I|=r_{\mathcal{M}}\}\\
= & \{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{I},|S|=r_{\mathcal{M}}\},
\end{align*}
where the last equality follows from the fact that $|I|=|S_{1}\uplus \cdots\uplus S_{k}|=\sum_{i\in[k]}|S_{i}|=|S|$.
On the other hand we have that
\begin{align*}
\mathcal{B}(\mathcal{M}')= & \{S\subseteq E:S\in\mathcal{I}',|S|=r_{\mathcal{M}'}\}\\
= & \{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in\mathcal{I},|S|=r_{\mathcal{M}'}\}.
\end{align*}
Hence, $\mathcal{H}=\mathcal{B}(\mathcal{M}')$ if and only if $r_{\mathcal{M}}=r_{\mathcal{M}'}$. But notice
that for any $S\in\mathcal{I}'$ we have that $|S|=|I|$ for some $I\in\mathcal{I}$,
and hence $r_{\mathcal{M}'}\leq r_{\mathcal{M}}$. Also, for any $I\in\mathcal{I}$ we can find
some $S\in\mathcal{I}'$ (e.g. set $S_{1}=I$ and $S_{i}=\emptyset$ for $i\neq1$)
such that $|S|=|I|$, and thus $r_{\mathcal{M}'}\geq r_{\mathcal{M}}$. It follows that
$r_{\mathcal{M}}=r_{\mathcal{M}'}$ and hence $\mathcal{H}=\mathcal{B}(\mathcal{M}')$ as we wanted to show.\qed \end{proof} \fi
\iffalse \begin{proposition}
\label{Claim F =00003D 2^V}Let $\mathcal{F}=2^{V}$, then $(E,\mathcal{H})$ is a partition matroid.\end{proposition} \begin{proof}
Assume $\mathcal{F}=2^{V}$, then
\[
\begin{array}{ccl}
\mathcal{H} & = & \{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}\in2^{V}\}\\
& = & \{S\subseteq E:S_{i}\cap S_{j}=\emptyset,\;\forall i\neq j\}\\
& = & \{S\subseteq E:\forall v\in V,\,|S\cap\delta(v)|\leq1\}\\
& = & \{S\subseteq E:\forall v\in V,\,|S\cap([k]\times\{v\})|\leq1\}.
\end{array}
\]
Since $\{[k]\times\{v\}:v\in V\}$ is a partition of $E$, it follows
that $(E,\mathcal{H})$ is a partition matroid over $E$.\qed \end{proof} \begin{proposition}
\label{Claim F =00003D V}If $\mathcal{F}=\{V\}$, then $\mathcal{H}$ corresponds to
the set of bases of a partition matroid.\end{proposition} \begin{proof}
Let $\mathcal{F}=\{V\}$, then $\mathcal{H}=\{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}=V\}$. From
Claim \ref{Claim F =00003D 2^V} we know that $\mathcal{M}=(E,\mathcal{I})$ is a partition
matroid over $E$, where
\[
\mathcal{I}:=\{S\subseteq E:\forall v\in V,\,|S\cap([k]\times\{v\})|\leq1\}.
\]
Moreover, it is easy to see that the set of bases of $\mathcal{M}$ corresponds
exactly to $\{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}=V\}$. Since $\mathcal{H}=\{S\subseteq E:S_{1}\uplus \cdots\uplus S_{k}=V\}$,
the claim follows.\qed \end{proof} \fi We now discuss some invariant properties with respect to the families $\mathcal{F}_i$.
\begin{proposition}
\label{Claim F_i matroids}If $(V,\mathcal{F}_{i})$ is a matroid for each
$i\in[k]$, then $(E,\mathcal{H})$ is also a matroid.\end{proposition} \begin{proof}
Let $\mathcal{M}_{i}:=(\{i\}\times V,\mathcal{I}_{i})$ for $i\in[k]$, where $\mathcal{I}_{i}:=\{\{i\}\times S:S\in\mathcal{F}_{i}\}$.
Since $(V,\mathcal{F}_{i})$ is a matroid, we have that $\mathcal{M}_{i}$ is also a
matroid. Moreover, by taking the matroid union of the $\mathcal{M}_{i}$'s
we get $(E,\mathcal{H})$. Hence, $(E,\mathcal{H})$ is a matroid.
\qed \end{proof}
\begin{proposition}
\label{claim ring families}
If $\mathcal{F}_{i}$ is a ring family over $V$ for each $i\in[k]$, then
$\mathcal{H}$ is a ring family over $E$.\end{proposition} \begin{proof}
Let $S,T\in\mathcal{H}$ and notice that $S\cup T=\biguplus_{i\in[k]}(\{i\}\times(S_{i}\cup T_{i})$)
and $S\cap T=\biguplus_{i\in[k]}(\{i\}\times(S_{i}\cap T_{i})$).
Since $\mathcal{F}_{i}$ is a ring family for each $i\in[k]$, it follows that
$S_{i}\cup T_{i}\in\mathcal{F}_{i}$ and $S_{i}\cap T_{i}\in\mathcal{F}_{i}$ for each
$i\in[k]$. Hence $S\cup T\in\mathcal{H}$ and $S\cap T\in\mathcal{H}$, and thus
$\mathcal{H}$ is a ring family over $E$. \qed \end{proof}
We saw in Section \ref{sec:lifting-reduction} that if the original functions $f_i$ are all submodular, then the lifted function $f$ is also submodular. Recall that for any set $S \subseteq E$ in the lifted space, there are unique sets $S_i \subseteq V$ such that $S = \uplus_{i \in [k]} (\{i\} \times S_i)$. We think of $S_i$ as the set of items assigned to agent $i$. In a similar way, given any vector $\bar{z} \in [0,1]^E$, there are unique vectors $z_i \in [0,1]^V$ such that $\bar{z} = (z_1,z_2,\ldots,z_k)$, where we think of $z_i$ as the vector associated to agent $i$. Our following result establishes the relationship between the values $f^M(\bar{z})$ and $\sum_{i \in [k]} f_i^M(z_i)$.
\begin{lemma}
\label{lem:max-multilinear}
Let the functions $f_i$ and $f$ be as described in the lifting reduction on Section \ref{sec:lifting-reduction}.
Then for any vector $\bar{z}=(z_1,z_2,\ldots,z_k) \in [0,1]^E$, where $z_i \in [0,1]^V$ is the vector associated with agent $i$, we have that $f^M(\bar{z})=f^M(z_1,z_2,\ldots,z_k)=\sum_{i \in [k]} f_i^M(z_i)$. \end{lemma} \begin{proof}
We use the definition of the multilinear extension in terms of expectations (see Section \ref{sec:max-SA-MA-formulations}).
Recall that for a vector $z \in [0,1]^V$, $R^z$ denotes a random set that contains element $v_i$ independently with probability $z_{v_i}$.
We use $\mathbb{P}_{z}(S)$ to denote $\mathbb{P}[R^z=S]$. We then have
\begin{eqnarray*}
f^M (\bar{z}) & = & \mathbb{E} [f(R^{\bar{z}})] = \sum_{S \subseteq E} f(S) \mathbb{P}_{\bar{z}}(S) \\
& = & \sum_{S_1 \subseteq V} \sum_{S_2 \subseteq V} \cdots \sum_{S_k \subseteq V} [\sum_{i=1}^k f_i (S_i)] \cdot \mathbb{P}_{(z_1,z_2,\ldots,z_k)}(S_1,S_2,\ldots,S_k) \\
& = & \sum_{i=1}^k \sum_{S_1 \subseteq V} \sum_{S_2 \subseteq V} \cdots \sum_{S_k \subseteq V} f_i (S_i) \cdot \mathbb{P}_{(z_1,z_2,\ldots,z_k)}(S_1,S_2,\ldots,S_k)\\
& = & \sum_{i=1}^k \sum_{S_i \subseteq V} f_i (S_i) \sum_{S_j \subseteq V, j\neq i} \mathbb{P}_{(z_1,z_2,\ldots,z_k)}(S_1,S_2,\ldots,S_k)\\
& = & \sum_{i=1}^k \sum_{S_i \subseteq V} f_i(S_i) \mathbb{P}_{z_i}(S_i)
= \sum_{i=1}^k \mathbb{E}[f_i(S_i^{z_i})]
= \sum_{i=1}^k f^M_i(z_i).
\end{eqnarray*}
\qed \end{proof}
\end{document}
|
arXiv
|
{
"id": "1803.03767.tex",
"language_detection_score": 0.7737969160079956,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\footnote{{\it 2000 Mathematics Subject Classification.} Primary:32F45.
{\it Key words and phrases.} Lempert function, Arakelian's theorem.}
\begin{abstract} We prove that the multipole Lempert function is monotone under inclusion of pole sets. \end{abstract}
\maketitle
Let $D$ be a domain in $\mathbb C^n$ and let $A=(a_j)_{j=1}^l$, $1\le l\le\infty$, be a countable (i.e. $l=\infty$) or a non--empty finite subset of $D$ (i.e. $l\in\mathbb N$). Moreover, fix a function $\boldsymbol p:D\longrightarrow\mathbb R_+$ with $$ |\bs{p}|:=\{a\in D:\boldsymbol p(a)>0\}=A. $$ $\boldsymbol p$ is called a {\it pole function for $A$} on $D$ and $|\bs{p}|$ its {\it pole set}. In case that $B\subset A$ is a non--empty subset we put $\boldsymbol p_B:=\boldsymbol p$ on $B$ and $\boldsymbol p_B:=0$ on $D\setminus B$. $\boldsymbol p_B$ is a pole function for $B$.
For $z\in D$ we set $$
l_D(\boldsymbol p,z)=\inf\{\prod_{j=1}^l|\lambda_j|^{\boldsymbol p(a_j)}\}, $$ where the infimum is taken over all subsets $(\lambda_j)_{j=1}^l$ of $\mathbb D$ (in this paper, $\mathbb D$ is the open unit disc in $\mathbb C$) for which there is an analytic disc $\varphi\in\mathcal O(\mathbb D,D)$ with $\varphi(0)=z$ and $\varphi(\lambda_j)=a_j$ for all $j$. Here we call $l_D(\boldsymbol p,\cdot)$ {\it the Lempert function with $\boldsymbol p$-weighted poles at $A$} \cite{Wik1,Wik2}; see also \cite{jarpfl}, where this function is called {\it the Coman function for $\boldsymbol p$}.
Recently, F.~Wikstr\"om \cite{Wik1} proved that if $A$ and $B$ are finite subsets of a convex domain $D\subset\mathbb C^n$ with $\varnothing\neq B\subset A$ and if $\boldsymbol p$ is a pole function for $A$, then $l_D(\boldsymbol p,\cdot)\le l_D(\boldsymbol p_B,\cdot)$ on $D$.
On the other hand, in \cite{Wik2} F.~Wikstr\"om gave an example of a complex space for which this inequality fails to hold and he asked whether it remains to be true for an arbitrary domain in $\mathbb C^n$.
The main purpose of this note is to present a positive answer to that question, even for countable pole sets (in particular, it follows that the infimum in the definition of the Lempert function is always taken over a non-empty set).
\begin{theorem} For any domain $D\subset\mathbb C^n$, any countable or non-empty finite subset $A$ of $D$, and any pole function $\boldsymbol p$ for $A$ we have $$ l_D(\boldsymbol p,\cdot)=\inf\{l_D(\boldsymbol p_B,\cdot):\varnothing\neq B \text{ a finite subset of } A\}. $$
Therefore, $$ l_D(\boldsymbol p,\cdot)=\inf\{l_D(\boldsymbol p_B,\cdot):\varnothing\neq B\subset A\}. $$ \end{theorem}
\vskip 0.5cm
The proof of this result will be based on the following
\
\begin{theorem*}[Arakelian's Theorem \cite{Ara}] Let $E\subset\Omega\subset\mathbb C$ be a relatively closed subset of the domain $\Omega$. Assume that $\Omega^*\setminus E$ is connected and locally connected. (Here $\Omega^*$ denotes the one--point compactification of $\Omega$.)
If $f$ is a complex-valued continuous function on $E$ that is holomorphic in the interior of $E$ and if $\varepsilon>0$, then there is a $g\in\mathcal O(\Omega)$ with $|g(z)-f(z)|<\varepsilon$ for any $z\in E$. \end{theorem*} \
\begin{proof} Fix a point $z\in D$. First, we shall verify the inequality $$ l_D(\boldsymbol p,z)\le\inf\{l_D(\boldsymbol p,z):\varnothing\neq B \text{ a finite subset of } A\}:\leqno{(1)} $$
Take a non--empty proper finite subset $B$ of $A$. Without loss of generality we may assume that $B=A_m:=(a_j)_{j=1}^m$ for a certain $m\in\mathbb N$, where $A=(a_j)_{j=1}^l$, $m<l\le\infty$.
Now, let $\varphi:\Bbb D\to D$ be an analytic disc with $\varphi(\lambda_j)=a_j,$ $1\le j\le m,$ where $\lambda_0:=0$ and
$a_0:=z.$ Fix $t\in[\max_{0\le j\le m}|\lambda_j|,1)$ and put $\lambda_j=1-\frac{1-t}{j^2}$,\; $j\in A(m)$, where $A(m):=\{m+1,\dots,l\}$ if $l<\infty$, respectively $A(m):=\{j\in\mathbb N:j>m\}$ if $l=\infty$. Consider a continuous curve $\varphi_1:[t,1)\to D$ such that $\varphi_1(t)=\varphi(t)$ and $\varphi_1(\lambda_j)=a_j$, $j\in A(m)$. Define $$ f=\begin{cases}
\varphi|_{\overline{t\Bbb D}}\\
\varphi_1|_{[t,1)} \end{cases} $$ on the set $F_t=\overline{t\Bbb D}\cup[t,1)\subset\mathbb D$. Observe that $F_t$ satisfies the geometric condition in Arakelian's Theorem.
Since $(\lambda_j)_{j=0}^l$ satisfy the Blaschke condition, for any $k$ we may find a Blaschke product $B_k$ with zero set $(\lambda_j)_{j=0,j\neq k}^l$. Moreover, we denote by $d$ the function $\operatorname{dist}(\partial D,f)$ on $F_t$, where the distance arises from the $l^\infty$--norm. Let $\eta_1$, $\eta_2$ be continuous real-valued functions on $F_t$ with \begin{gather*} \eta_1,\eta_2\le\log\frac{d}{9},\qquad \eta_1,\eta_2=\min_{\overline{t\mathbb D}} \log\frac{d}{9}\hbox{ on }(\overline{t\mathbb D}),\\ \text{ and }\quad
\eta_1(\lambda_j)=\eta_2(\lambda_j)+\log(2^{-j-1}|B_j(\lambda_j)|),\quad j\in A(m). \end{gather*}
Applying three times Arakelian's theorem, we may find functions $\zeta_1, \zeta_2\in\mathcal O(\mathbb D)$ and a holomorphic mapping $h$ on $\mathbb D$ such that $$
|\zeta_1-\eta_1|\le 1,\;|\zeta_2-\eta_2|\le 1,\hbox{ and
}|h-f|\le
\varepsilon|e^{\zeta_1-1}|\le \varepsilon e^{\eta_1}\text{ on } F_t, $$ where $\varepsilon:=
\min\{\tfrac{|B_j(\lambda_j)|}{2^{j+1}}:j=0,\dots,m\}<1$ (in the last case apply Arakelian's theorem componentwise to the mapping $e^{1-\zeta_1}f$).
In particular, we have \begin{gather*}
|h-f|\le\frac{d}{9}\quad \text{ on } F_t,\\
|\gamma_j|\leq e^{\eta_1(\lambda_j)}2^{-j-1}|B_j(\lambda_j)|
=e^{\eta_2(\lambda_j)}2^{-j-1}|B_j(\lambda_j)|,\quad j=0,\dots,m,\\
|\gamma_j|\le e^{\eta_1(\lambda_j)}=
e^{\eta_2(\lambda_j)}2^{-j-1}|B_j(\lambda_j)| ,\;j\in A(m), \end{gather*} where $\gamma_j:=h(\lambda_j)-f(\lambda_j)$, $j\in\mathbb Z_+$ if $l=\infty$, respectively $0\leq j\leq l$ if $l\in\mathbb N$. Then, in virtue of $e^{\eta_2(\lambda_j)} \le e^{1+\operatorname{Re} \zeta_2(\lambda_j)}$, the function $$ g:=e^{\zeta_2}\sum_{j=0}^l\frac{B_j }{e^{\zeta_2(\lambda_j)}B_j(\lambda_j)}\gamma_j $$ is holomorphic on $\mathbb D$ with $g(\lambda_j)=\gamma_j$ and $$
|g|\le e^{\operatorname{Re}\zeta_2+1}\le e^{\eta_2+2}\le\frac{e^2}{9}d\quad \text{ on } F_t. $$ For $q_t:=h-g$ it follows that $q_t(\lambda_j)=f(\lambda_j)$ and $$
|q_t-f|\le\frac{e^2+1}{9}d<d\quad\text{ on }F_t. $$
Thus we have found a holomorphic mapping $q_t$ on $\mathbb D$ with $q_t(\lambda_j)=a_j$ and $q_t(F_t)\subset D$. Hence there is a simply connected domain $E_t$ such that $F_t\subset E_t\subset\mathbb D$ and $q_t(E_t)\subset D$.
Let $\rho_t:\mathbb D\to E_t$ be the Riemann mapping with $\rho_t(0)=0,$ $\rho_t'(0)>0$ and $\rho_t(\lambda_j^t)=\lambda_j$. Considering the analytic disc $q_t\circ\rho_t:\mathbb D\to D$, we get that $$
l_D(\boldsymbol p,z)\le\prod_{j=1}^l|\lambda_j^t|^{\boldsymbol p(a_j)}\leq\prod_{j=1}^m|\lambda_j^t|^{\boldsymbol p(a_j)}. $$ Note that by the Carath\'eodory kernel theorem, $\rho_t$ tends, locally uniformly, to the identity map of $\mathbb D$ as $t\to 1$. This shows that the last product converges to
$\prod_{j=1}^m|\lambda_j|^{\boldsymbol p(a_j)}$. Since $\varphi$ was an arbitrary competitor for $l_G(\boldsymbol p_{A_m},z)$, the inequality (1) follows.
On the other hand, the existence of an analytic disc whose graph contains $A$ and $z$ implies that $$ l_D(\boldsymbol p,z)\ge\limsup_{m\to\infty}l_D(\boldsymbol p_{A_m},z), $$ which completes the proof. \end{proof}
\begin{remark}{\rm Looking at the above proof shows that we have proved an approximation and simultaneous interpolation result, i.e. the constructed function $q_t$ approximates and interpolates the given function $f$.
We first mention that the proof of Theorem 1 could be simplified using a non-trivial result on interpolation sequences due to L.~Carleson (see, for example, Chapter 7, Theorem 3.1 in \cite{and}).
Moreover, it is possible to prove the following general result which extends an approximation and simultaneous interpolation result by P.~M.~Gauthier, W.~Hengartner, and A.~A.~Nersesyan (see \cite{gauhen}, \cite{ner}):
Let $D\subset\mathbb C$ be a domain, $E\subset D$ a relatively closed subset satisfying the condition in Arakelian's theorem, $\Lambda\subset E$ such that $\Lambda$ has no accumulation point in $D$ and $\Lambda\cap\operatorname{int} E$ is a finite set. Then for given functions $f,h\in\mathcal C(E)\cap\mathcal O(\operatorname{int} E)$ there exists a $g\in\mathcal O(D)$ such that $$
|g-f|<e^{\operatorname{Re} h} \text { on } E,\quad
\text{ and } \quad f(\lambda)=g(\lambda),\;\lambda\in\Lambda. $$ It is even possible to prescribe a finite number of the derivatives for $g$ at all the points in $\Lambda$. }
\end{remark}
\ As a byproduct we get the following result.
\begin{corollary} Let $D\subset\mathbb C^n$ be a domain and let $\boldsymbol p, \boldsymbol q:D\to\mathbb R_+$ be two pole functions on $D$ with $\boldsymbol p\leq\boldsymbol q$, $|\bs{q}|$ at most countable. Then $l_D(\boldsymbol q,z)\leq l_D(\boldsymbol p,z)$, $z\in D$. \end{corollary}
Hence, the Lempert function is monotone with respect to pole functions with an at most countable support.
\begin{remark}{\rm The Lempert function is, in general, non strictly monotone under inclusion of pole sets. In fact, take $D:=\mathbb D\times\mathbb D$, $A:=\{a_1,a_2\}\subset\mathbb D$, where $a_1\neq a_2$, and observe that
$l_D(\boldsymbol p,(0,0))=|a_1|$, where $\boldsymbol p:=\chi|_{A\times\{a_1\}}$ (use the product property in \cite{dietra}, \cite{jarpfl}, \cite{nikzwo}). }\end{remark}
\begin{remark}{\rm Let $D\subset\mathbb C^n$ be a domain, $z\in D$, and let now $\boldsymbol p:D\longrightarrow\mathbb R_+$ be a "general" pole function, i.e. $|\bs{p}|$ is uncountable. Then there are two cases:
1) There is a $\varphi\in\mathcal O(\mathbb D,D)$ with $\varphi(\lambda_{\varphi,a})=a$, $\lambda_{\varphi,a}\in\mathbb D$, for all $a\in|\bs{p}|$ and $\varphi(0)=z$. Defining \begin{multline*}
l_D(\boldsymbol p,z):=\inf\{\prod|\lambda_{\psi,a}|^{\boldsymbol p(a)}:\psi\in\mathcal O(\mathbb D,D) \text{ with }\\ \psi(\lambda_{\psi,a})=a \text{ for all } a\in|\bs{p}|, \psi(0)=z\} \end{multline*} then $l_D(\boldsymbol p,z)=0$.
Observe that $l_D(\boldsymbol p,z)=\inf\{l_D(\boldsymbol p_B,z):\varnothing \neq B\text{ a finite subset of } A\}$.
2) There is no analytic disc as in 1). In that case we may define $^{1)}$\footnote{1) Compare the definition of the Coman function (for the second case) in \cite{jarpfl}.} $$ l_D(\boldsymbol p,z):=\inf\{l_D(\boldsymbol p_B,z):\varnothing\neq B\text { a finite subset of } A\}. $$ Example \ref{ex} below may show that the definition in 2) is more sensitive than the one used in \cite{jarpfl}. } \end{remark}
\vskip 0.5cm
Before giving the example we use the above definition of $l_D(\boldsymbol p,\cdot)$ for an arbitrary pole function $\boldsymbol p$ to conclude.
\begin{corollary} Let $D\subset\mathbb C^n$ be a domain and let $\boldsymbol p, \boldsymbol q :D\longrightarrow\mathbb R_+$ be arbitrary pole functions with $\boldsymbol p\leq\boldsymbol q$. Then $l_D(\boldsymbol q,\cdot)\leq l_D(\boldsymbol p,\cdot)$. \end{corollary}
\begin{example}\label{ex} {\rm Put $D:=\mathbb D\times\mathbb D$ and let $A\subset\mathbb D$ be uncountable, e.g. $A=\mathbb D$. Then there is no $\varphi\in\mathcal O(\mathbb D,D)$ passing through $A\times\{0\}$ and
$(0,w)$, $w\in\mathbb D_*$. Put $\boldsymbol p:=\chi|_{A\times\{0\}}$ on $D$ as a pole function.
Let $B\subset A$ be a non--empty finite subset. Then applying the product property (see \cite{dietra}, \cite{jarpfl}, \cite{nikzwo}), we get $$ l_D(\boldsymbol p_B,(0,w))=g_D(B\times\{0\},(0,w))=\max\{g_{\mathbb D}(B,0),g_{\mathbb D}(0,w)\}, $$ where $g_{\mathbb D}(A,\cdot)$ denotes the Green function in $\mathbb D$ with respect to the pole set $A$.
Therefore, $l_D(\boldsymbol p,(0,w))=|w|$ $^{1)}$.} \end{example}
To conclude this note we mention that the Lempert function is not holomorphically contractible, even if the holomorphic map is a proper covering.
\begin{example} {\rm Let $\pi:\mathbb D_*\longrightarrow\mathbb D_*$, $\pi(z):=z^2$. Obviously, $\pi$ is proper and a covering. Fix two different points $a_1, a_2\in\mathbb D_*$ with $a_1^2=a_2^2=:c$. For a point $z\in\mathbb D_*$ we know that $$
l_{\mathbb D_*}(\chi|_{\{c\}},z^2)=\min\{l_{\mathbb D_*}(\chi|_{\{a_1\}},z),l_{\mathbb D_*}(\chi|_{\{a_2\}},z)\}
\geq l_{\mathbb D_*}(\chi|_A,z), $$
where $A:=\{a_j:j=1,2\}$ and $\chi|_B$ is the characteristic function for the set $B\subset D_*$. Assume that $l_{\mathbb D_*}(\chi|_{\{a_1\}},z)\leq l_{\mathbb D_*}(\chi|_{\{a_2\}},z)$. Then this left side is nothing than the classical Lempert function for the pair $(a_1,z)$. Recalling how to calculate it via a covering map one easily concludes that $l_{\mathbb D_*}(\chi|_A,z)<
l_{\mathbb D_*}(\chi|_{\{a_1\}},z)$. Hence $l_{\mathbb D_*}(\boldsymbol p
,\pi(z))>l_{\mathbb D_*}(\boldsymbol p\circ\pi,z)$, where $\boldsymbol p:=\chi|_{\{c\}}$.
Therefore, the Lempert function with a multipole behaves worse than the Green function with a multipole.} \end{example}
{\bf Acknowledgments.} This note was written during the stay of the first named author at the University of Oldenburg supported by a grant from the DFG (September - October 2004). He likes to thank both institutions for their support.
We thank the referee for pointing out an error in the first version of this paper.
\end{document}
\end{document}
|
arXiv
|
{
"id": "0409358.tex",
"language_detection_score": 0.6514773964881897,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Construction of Antisymmetric Variational Quantum States \\ with Real-Space Representation}
\author{Takahiro Horiba} \email{[email protected]} \author{Soichi Shirai} \author{Hirotoshi Hirai} \affiliation{Toyota Central R\&D Labs., Inc., 41-1 Yokomichi, Nagakute, Aichi 480-1192, Japan}
\date{\today}
\begin{abstract}
Electronic state calculations using quantum computers are mostly based on second quantization, which is suitable for qubit representation. Another way to describe electronic states on a quantum computer is first quantization, which is expected to achieve smaller scaling with respect to the number of basis functions than second quantization. Among basis functions, a real-space basis is an attractive option for quantum dynamics simulations in the fault-tolerant quantum computation (FTQC) era. A major difficulty in first quantization with a real-space basis is state preparation for many-body electronic systems. This difficulty stems from of the antisymmetry of electrons, and it is not straightforward to construct antisymmetric quantum states on a quantum circuit.
In the present paper, we provide a design principle for constructing a variational quantum circuit to prepare an antisymmetric quantum state. The proposed circuit generates the superposition of exponentially many Slater determinants, that is, a multi-configuration state, which provides a systematic approach to approximating the exact ground state. We implemented the variational quantum eigensolver (VQE) to obtain the ground state of a one-dimensional hydrogen molecular system. As a result, the proposed circuit well reproduced the exact antisymmetric ground state and its energy, whereas the conventional variational circuit yielded neither an antisymmetric nor a symmetric state. Furthermore, we analyzed the many-body wave functions based on quantum information theory, which illustrated the relation between the electron correlation and the quantum entanglement.
\end{abstract}
\maketitle
\section{Introduction}
Quantum computers are currently attracting increasing attention as promising hardware for materials computations~\cite{mcardle2020quantum, bauer2020quantum, ma2020quantum}, and a number of studies on a variety of material systems have been conducted ~\cite{nam2020ground, shirai2022calculation, shirai2023computational, hirai2023excited, colless2018computation, kassal2009quantum, rall2020quantum,kassal2008polynomial, oftelie2020towards, ollitrault2021molecular}. Such materials computations are mostly based on second quantization~\cite{peruzzo2014variational, mcclean2016theory, cerezo2021variational}, which is suitable for describing electronic states on quantum computers.
An alternative way to describe electronic states on quantum computers is first quantization, in which a wave function is specified by the expansion coefficients of the basis functions. With quantum computers, it is possible to obtain a wave function represented by an exponential number of basis functions with a polynomial number of qubits~\cite{zalka1998efficient}. Therefore, first quantization offers the possibility of achieving smaller scaling with respect to the number of basis functions than second quantization~\cite{abrams1997simulation, berry2018improved, babbush2019quantum, chan2023grid, su2021fault}.
Among basis functions, a real-space basis is an attractive option because a systematic improvement of computational accuracy and a rapid convergence to the continuum limit can be expected by increasing the number of qubits~\cite{hirose2005first}. In addition, a real-space basis can be applied to systems with a variety of boundary conditions~\cite{ohba2012linear, hirose2005first}, and thus is suitable for quantum dynamics calculations~\cite{childs2022quantum}. Recently, Chan et al.\ proposed quantum circuits for computing real-space quantum dynamics based on first quantization~\cite{chan2023grid}, which represented a promising blueprint of quantum simulations in the fault-tolerant quantum computing (FTQC) era.
However, first quantization has a significant challenge, namely state preparation. Let us consider preparing the ground state of the system, which is a typical choice of initial state for dynamics calculations. As a state preparation method, we consider quantum phase estimation (QPE)~\cite{nielsen2002quantum} which has been employed in several studies on first quantization~\cite{berry2018improved, chan2023grid}. By using QPE, it is possible to distill the ground state from an input state which has sufficient overlap with the ground state~\cite{berry2018improved}. Thus, the problem is how to prepare such an input state.
In first quantization with a real-space basis, preparing an input state takes a tremendous number of gate operations because probability the amplitudes of the many-body wave function need to be encoded into a state vector of a quantum circuit. Although several amplitude-encoding methods have been proposed~\cite{holmes2020efficient, koppe2023amplitude, ollitrault2020nonadiabatic, nakaji2022approximate}, it is not at all straightforward to prepare an approximate ground state with any of them. This state preparation problem is often avoided by using oracle circuits. Compared to constructing such oracle circuits, variational methods such as the variational quantum eigensolver (VQE)~\cite{peruzzo2014variational, mcclean2016theory, cerezo2021variational} are considered to be relatively feasible approaches. In fact, a number of studies have proposed preparing an input state using the VQE in second quantization~\cite{halder2023iterative,dong2022ground}.
Unfortunately, it is also not straightforward to implement the VQE in first quantization. This is due to the antisymmetry of electrons, which are Fermi particles. In second quantization, antisymmetry does not need to be considered for variational quantum states, because antisymmetry is naturally introduced as the anticommutation relations of creation and annihilation operators. By contrast, in first quantization, antisymmetry is imposed on the many-body wave function itself, and thus variational quantum states must satisfy antisymmetry. Nevertheless, there is no guarantee that quantum states generated by conventional variational quantum circuits will satisfy antisymmetry, which raises the possibility that the VQE will yield non-antisymmetric states. Therefore, a variational quantum circuit that generates only antisymmetric quantum states is required in order to obtain the electronic ground state by the VQE.
In the present paper, we provide a design principle for constructing such an antisymmetrized variational quantum circuit. Our proposed circuit consists of two types of variational circuits that achieve antisymmetry-preserving transformations of a state vector on a Hilbert space. It is noteworthy that the proposed circuit generates a multi-configuration (MC) state. That is, it is possible to generate superpositions of an exponentially large number of Slater determinants by alternately layering these two types of variational circuits. This scheme provides a systematic approach to approximating the exact ground state. To verify the validity of our method, we performed the VQE calculation for a one-dimensional hydrogen molecular (1D-$\rm{H_2}$) system, and demonstrated that the proposed circuit well reproduced the exact antisymmetric, or fermionic, ground state and its energy. In addition to implementing the VQE, we analyzed the many-body wave functions based on quantum information theory. Such an analysis reveals the microscopic electronic structure of a many-body wave function represented in real space and illustrates the relation between the electron correlation and the quantum entanglement.
\section{Method}
\begin{figure*}
\caption{Quantum circuit to generate an antisymmetrized variational quantum state for one-dimensional two-electron systems}
\label{fig:overall_circuit}
\end{figure*}
In this section, we introduce the first-quantized VQE framework based on the real-space representation and describe our proposed variational quantum circuit. We also describe the setting of the numerical experiments for a 1D-$\rm{H_2}$ system.
\subsection{First-quantized VQE with a real-space basis} To begin with, we briefly describe the first-quantized formulation of the many-body electron problem. The many-body Schr\"{o}dinger equation for an $\eta$-electron molecular system is expressed by the following equation: \begin{equation} H\ket{\psi(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})} = E\ket{\psi(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})}, \label{eq:shrodinger}\\ \end{equation} where $\ket{\psi(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})}$ is the antisymmetric many-body wave function to be constructed on a quantum circuit. The molecular Hamiltonian $H$ with the Born-Oppenheimer approximation is expressed in atomic units as follows: \begin{equation}
H = \sum_{i=1}^{\eta}\left[-\frac{{\bm{\nabla_i}}^2}{2}-\sum_{p}\frac{Z_{p}}{|\bm{r_i}-\bm{R_p}|}\right]+\sum_{i<j}\frac{1}{|\bm{r_i}-\bm{r_j}|}, \label{eq:hamiltonian}\\ \end{equation} where $r_i$ is the $i$th electron coordinate, $R_p$ is the $p$th nuclear coordinate, and $Z_p$ is the atomic number of the $p$th nucleus. The VQE in this study is based on this first-quantized Hamiltonian and the many-body wave function represented in real space.
Next, we introduce the real-space basis. Let us consider the expansion of the many-body wave function by the real-space basis $\ket{\delta(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})}$ (delta functions located at grid points $\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta}$). The expansion coefficient of a many-body wave function is given as \begin{equation}
\psi(\bm{r'_1},\bm{r'_2},\cdots,\bm{r'_\eta}) = \braket{\delta(\bm{r'_1},\bm{r'_2},\cdots,\bm{r'_\eta})|\psi(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})}.\\ \end{equation} The real-space basis of an $\eta$-electron system consists of that of one-electron system $\ket{\delta(\bm{r_i})}$, as follows: \begin{equation} \ket{\delta(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})} = \ket{\delta(\bm{r_1})}\ket{\delta(\bm{r_2})}\cdots\ket{\delta(\bm{r_\eta})}.\\ \end{equation} $\ket{\delta(\bm{r_i})}$ consists of three-dimensional spatial coordinates $x, y, z$ and one spin coordinate $s$. Assuming that $L$ qubits are assigned to each spatial dimension and one qubit to spin, $\ket{\delta(\bm{r_i})}$ is expressed as \begin{equation} \ket{\delta(\bm{r_i})} = \ket{{{x_1}}^{(i)}\cdots{{x_L}}^{(i)}} \ket{{{y_1}}^{(i)}\cdots{{y_L}}^{(i)}} \ket{{{z_1}}^{(i)}\cdots{{z_L}}^{(i)}} \ket{s^{(i)}}, \\ \end{equation} where ${x_k}^{(i)},{y_k}^{(i)},{z_k}^{(i)},s^{(i)} \in \left\{0,1\right\}, \forall k \in[1,L]$. The number of qubits that constitute $\ket{\delta(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})}$ is $\eta(3L+1)$, and thus the many-body wave function is expanded in $2^{\eta(3L+1)}$ basis functions. Quantum computers are expected to realize such an exponentially large number of basis functions with a polynomial number of qubits, which is a significant advantage over classical computers.
In order to implement the VQE, it is necessary to measure the energy expectation value of a quantum state $\ket{\psi(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})}$. The energy expectation value of molecular Hamiltonian $H$ in Eq.~\ref{eq:hamiltonian} is expressed as follows: \begin{equation} \begin{split}
E &= \braket{\psi(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})|H|\psi(\bm{r_1},\bm{r_2},\cdots,\bm{r_\eta})} \\ &= E_K + E_V^{e-n} + E_V^{e-e}, \end{split} \end{equation} where $E_K$ is the electron kinetic energy, $E_V^{e-n}$ is the electron-nuclei Coulomb attraction energy, and $E_V^{e-e}$ is the electron-electron Coulomb repulsion energy. The kinetic and Coulomb energy operators of the Hamiltonian are diagonal in momentum space and real space, respectively. The momentum space basis is obtained by the quantum Fourier transformation (QFT) of the real-space basis. Letting $\ket{\psi(\bm{k_i})}=U_{\mathrm{QFT}}\ket{\psi(\bm{r_i})}$ be the one-body momentum space basis, the kinetic energy $E_K$ is expressed as \begin{equation}
E_K =\sum_{i=1}^{\eta} -\frac{{\bm{k_i}}^2}{2}|\psi(\bm{k_i})|^2. \end{equation} The Coulomb energies $E_V^{e-n}, E_V^{e-e}$ are expressed in terms of one-body and two-body real-space bases as follows: \begin{equation} \begin{split}
E_V^{e-n} &= -\sum_{i=1}^{\eta}\sum_{p}\frac{Z_{p}}{|\bm{r_i}-\bm{R_p}|}|\psi(\bm{r_i})|^2, \\
E_V^{e-e} &= \sum_{i<j}\frac{1}{|\bm{r_i}-\bm{r_j}|}|\psi(\bm{r_i},\bm{r_j})|^2. \end{split} \end{equation}
Probability distributions $|\psi(\bm{k_i})|^2$, $|\psi(\bm{r_i})|^2$, and $|\psi(\bm{r_i,r_j})|^2$ can be obtained by measuring the output of the (QFT applied) quantum circuit in the computational basis (Pauli $Z$ basis). Note that, as pointed out by Chen et al., this method is not efficient in terms of sampling cost and is sensitive to discretization errors of grids~\cite{chan2023grid}. According to their work, QPE should be used for measuring energy expectation values in future quantum calculations, but this naive method was adopted in the present study because of a lack of sufficient computational resources to simulate QPE on a classical computer. The remaining issue in the first-quantized VQE is how to construct a variational quantum circuit that generates antisymmetric ansatz states. In the following, we describe the design principle for constructing a variational quantum circuit to settle the above issue.
\subsection{Antisymmetrized variational quantum circuit}
To explain the design principles of circuits, we consider the one-dimensional two-electron system, which is the minimum model required to describe the antisymmetry of electrons. In the following part of this section , we discuss two-electron systems, but we would like to note that the proposed method is applicable not only to two-electron systems but more generally to systems with larger numbers of electrons.
The overall circuit architecture is shown in Fig.~\ref{fig:overall_circuit}. The proposed circuit consists of three parts: (1) seed state preparation, (2) one-body and two-body space variational circuits, and (3) measurement of energy expectations. Since the measurement procedure of energy expectations has already been described, we here describe only the parts of the circuit that generate antisymmetric ansatz states. In the first part of the circuit, some antisymmetric state, i.e., seed state, is prepared, at which the state vector of the circuit belongs to the antisymmetric subspace. The subsequent one-body and two-body space variational circuits transform the state vector into the ground state while keeping it in the antisymmetric subspace. Such a preparation of an antisymmetric state and antisymmetry-preserving transformations are the main design principle of the proposed circuit. In the following, we describe each part of the circuit.
\subsubsection{Seed state preparation} The simplest antisymmetric state of a two-electron system is the Greenberger-Horne-Zeilinger (GHZ) state, as follows: \begin{equation} \ket{\psi_{\mathrm{GHZ}}}=\frac{1}{\sqrt{2}}(\ket{0}_1\ket{1}_2-\ket{1}_1\ket{0}_2). \end{equation} This state can be generated by a series of an X (NOT) gate, an H (Hadamard) gate, and a CNOT (controlled NOT) gate as shown in Fig.~\ref{fig:overall_circuit}. In this circuit, the spin coordinates $\ket{s^{(i)}}$ are antisymmetrized, meaning that this state is a singlet state in which the two spins are antiparallel $\downarrow_1\uparrow_2 - \uparrow_1\downarrow_2$ (0 is assumed to be $\downarrow$ and 1 to be $\uparrow$). This state is expressed as follows, including the spatial coordinates (the normalization constant $1/\sqrt{2}$ is omitted): \begin{equation} \ket{\psi_{\mathrm{seed}}(\bm{r_1},\bm{r_2})} = \ket{\bm{0}}_1\ket{0}_1 \otimes \ket{\bm{0}}_2\ket{1}_2 - \ket{\bm{0}}_1\ket{1}_1 \otimes \ket{\bm{0}}_2\ket{0}_2, \end{equation} where the spatial coordinate of each electron is $\ket{\bm{0}}=\ket{00\cdots0}$. Obviously, this state is antisymmetric under the exchange of two electrons and can be used as a seed state.
For more-than-two-electron systems, seed states cannot be constructed in such a simple way. As an example, for three-electron systems, the following Slater determinant consisting of the three states $\ket{00}$, $\ket{01}$, and $\ket{10}$ is one of the antisymmetric states (the normalization constant $1/\sqrt{3!}$ is omitted). \begin{equation} \begin{split} \ket{\psi_{00,01,10}}&= \begin{vmatrix} \ket{00}_1 & \ket{01}_1 & \ket{10}_1 \\ \ket{00}_2 & \ket{01}_2 & \ket{10}_2 \\ \ket{00}_3 & \ket{01}_3 & \ket{10}_3 \\ \end{vmatrix}\\ &=\ket{00}_1\ket{01}_2\ket{10}_3+\ket{01}_1\ket{10}_2\ket{00}_3\\ &\quad+\ket{10}_1\ket{00}_2\ket{01}_3-\ket{10}_1\ket{01}_2\ket{00}_3\\ &\quad-\ket{01}_1\ket{00}_2\ket{10}_3-\ket{00}_1\ket{10}_2\ket{01}_3. \end{split} \end{equation} Previous research has proposed methods for constructing quantum circuits to prepare such a Slater determinant, though it is not as simple as that for the GHZ state. Thanks to the work of Berry et al., an implementation of the Fisher-Yates shuffle~\cite{durstenfeld1964algorithm} on a quantum circuit has been provided and can generate the superpositions of an exponential number of permutation elements with polynomial computational complexity~\cite{berry2018improved}. Following their method, it is possible to systematically generate seed states for more-than-two-electron systems.
Another way to prepare a seed state is a variational approach, which achieves a seed state using conventional variational circuits. In this approach, the Hadamard test (swap test) can be employed as an objective function. Letting $\mathrm{SWAP}_{ij}$ be the swap operator acting on the subspace of the $i$th and $j$th electrons, the output of the Hadamard test for the quantum state $\ket{\psi}$ becomes \begin{equation}
p_0 = \frac{1+\braket{\psi|\mathrm{SWAP}_{ij}|\psi}}{2}, p_1 = \frac{1-\braket{\psi|\mathrm{SWAP}_{ij}|\psi}}{2}, \end{equation} where $p_0$ and $p_1$ are the measurement probabilities of the $\ket{0}$ and $\ket{1}$ of an ancilla qubit. Suppose $\ket{\psi}$ is an antisymmetric wave function, $\mathrm{SWAP}_{ij}\ket{\psi} = -\ket{\psi}$; then the output of the Hadamard test becomes $p_0=0,p_1=1$ (for symmetric wave functions, $p_0=1,p_1=0$). Therefore, an approximated antisymmetric state can be obtained by updating variational parameters to maximize the output of the Hadamard tests for all two-electron swap operations. By repeatedly performing the Hadamard test on an approximated antisymmetric state, the pure antisymmetric states can eventually be distilled, which can then be used as a seed state.
\subsubsection{One-body space variational circuit} The proposed variational quantum circuit consists of two components: a one-body space variational circuit and a two-body ($\eta$-body) space variational circuit. We first explain the one-body space variational circuit. A one-body space variational circuit consists of unitary operators that act on the subspace of each electron (one-body space). Consider $\mathcal{U}^{(1)}(\theta) = \left[U^{(1)}(\theta)\right]^{\otimes\eta}$ as the one-body space variational circuit, where $U^{(1)}(\theta)$ is a unitary operator (variational circuit) acting on a one-body space. Since $U^{(1)}(\theta)$ transforms a state vector equally in each one-body space, $\mathcal{U}^{(1)}(\theta)$ does not destroy the antisymmetry of the ansatz state. For two-electron systems, $\mathcal{U}^{(1)}(\theta)=U^{(1)}(\theta)\otimes U^{(1)}(\theta)$ acting on the seed state (GHZ state) $\ket{\psi_{\mathrm{seed}}(\bm{r_1}, \bm{r_2})}$ yields \begin{equation} \begin{split} \ket{\psi(\bm{r_1}, \bm{r_2})} &= \mathcal{U}^{(1)}(\theta) \ket{\psi_{\mathrm{seed}}(\bm{r_1}, \bm{r_2})}\\ &=U^{(1)}(\theta) \ket{\bm{0}}_1\ket{0}_1 \otimes U^{(1)}(\theta) \ket{\bm{0}}_2\ket{1}_2 \\ &\quad- U^{(1)}(\theta) \ket{\bm{0}}_1\ket{1}_1 \otimes U^{(1)}(\theta) \ket{\bm{0}}_2\ket{0}_2\\ &=\ket{\alpha(\theta)}_1 \otimes \ket{\beta(\theta)}_2 - \ket{\beta(\theta)}_1 \otimes \ket{\alpha(\theta)}_2. \end{split} \end{equation} As can be seen, $\mathcal{U}^{(1)}(\theta)$ transforms $\ket{\bm{0}}\ket{0},\ket{\bm{0}}\ket{1}$ into $\ket{\alpha(\theta)}=U^{(1)}(\theta)\ket{\bm{0}}\ket{0}$, $\ket{\beta(\theta)}=U^{(1)}(\theta)\ket{\bm{0}}\ket{1}$ in each one-body space, and antisymmetry is preserved under this transformation. The resulting state can be expressed by a single Slater determinant consisting of $\ket{\alpha(\theta)}$ and $\ket{\beta(\theta)}$ as follows: \begin{equation} \begin{split} \ket{\psi_{\mathrm{SD}}(\bm{r_1}, \bm{r_2})} &=\ket{\alpha(\theta)}_1 \otimes \ket{\beta(\theta)}_2 - \ket{\beta(\theta)}_1 \otimes \ket{\alpha(\theta)}_2\\ &=
\begin{vmatrix}
\ket{\alpha(\theta)}_1 & \ket{\beta(\theta)}_1 \\
\ket{\alpha(\theta)}_2 & \ket{\beta(\theta)}_2
\end{vmatrix}. \end{split} \label{eq:hf_state} \end{equation} This indicates that a one-body space variational circuit can only explore quantum states within the HF approximation. By using the two-body ($\eta$-body) space variational circuit described in the following, we can explore quantum states beyond the HF approximation.
\subsubsection{Two-body space variational circuit} The usual strategy to go beyond the HF approximation is based on configuration interaction (CI) theory, in which a many-body wave function is approximated by a linear combination, or superposition, of multiple Slater determinants. As previously mentioned, a one-body space circuit generates an electronic state expressed by a single Slater determinant. Therefore, a superposition of multiple one-body space circuits is expected to generate a superposition of multiple Slater determinants, that is, an MC state.
For a two-electron system, consider two different operators $U_a\otimes U_a, U_b\otimes U_b$, where $U_a$ and $U_b$ are unitary operators acting on a one-body space . Their superposition is expressed as \begin{equation} U^{(2)}(\theta) = c_a(\theta) \cdot U_a \otimes U_a + c_b(\theta) \cdot U_b \otimes U_b , \end{equation} where $c_a(\theta)$, $c_b(\theta)$ are superposition coefficients parametrized by $\theta$. Operator $U^{(2)}(\theta)$ acting on a single Slater determinant $\ket{\psi_{\mathrm{SD}}}$ yields \begin{equation} \begin{split} &U^{(2)}(\theta)\ket{\psi_{\mathrm{SD}}} \\ &= c_a(\theta) \begin{vmatrix}
U_a\ket{\alpha}_1 & U_a\ket{\beta}_1 \\
U_a\ket{\alpha}_2 & U_a\ket{\beta}_2 \\
\end{vmatrix} + c_{b}(\theta) \begin{vmatrix}
U_b\ket{\alpha}_1 & U_b\ket{\beta}_1 \\
U_b\ket{\alpha}_2 & U_b\ket{\beta}_2 \\
\end{vmatrix}\\ &=c_a(\theta)\ket{\psi^{a}_{\mathrm{SD}}}+c_b(\theta)\ket{\psi^{b}_{\mathrm{SD}}}. \label{eq:mc_state} \end{split} \end{equation} As expected, a superposition of two different Slater determinants $\ket{\psi^{a}_{\mathrm{SD}}}$, $\ket{\psi^{b}_{\mathrm{SD}}}$, is generated by $U^{(2)}(\theta)$.
\begin{figure}
\caption{Implementation of $U^{(2)}(\theta)$. Top: implementations of Ising gate and real-valued symmetry-preserving (RSP) ansatz. Bottom: conceptual picture of generation of superposed HF states by an $R_{zz}(\theta)$ gate.}
\label{fig:multi_conf}
\end{figure}
One of the simplest implementations of such an operator is the Ising gate~\cite{jones2003robust}. Ising gates such as $R_{xx}(\theta)$, $R_{yy}(\theta)$, and $R_{zz}(\theta)$ are represented by a superposition of $I \otimes I$ and $P \otimes P$ ($I$ is the identity operator, $P$ is the Pauli operator $X$, $Y$, $Z$); thus, they can be employed as $U^{(2)}(\theta)$. For example, $R_{zz}(\theta)$ shown in Fig.~\ref{fig:multi_conf} is represented by \begin{equation} \begin{split} R_{zz}(\theta) &=
\begin{pmatrix}
e^{-i\theta/2} & 0 & 0 & 0 \\
0 & e^{i\theta/2} & 0 & 0 \\
0 & 0 & e^{i\theta/2} & 0 \\
0 & 0 & 0 & e^{-i\theta/2} \\
\end{pmatrix}\\ &= \cos\frac{\theta}{2} \cdot I \otimes I -i\sin\frac{\theta}{2} \cdot Z \otimes Z. \end{split} \end{equation} As can be seen from Fig.~\ref{fig:multi_conf}, the Ising gates act on a two-body space; thus, we refer to $U^{(2)}(\theta)$ as a two-body space variational circuit. For $\eta$-electron systems, an $\eta$-body space variational circuit $U^{(\eta)}(\theta)$ is represented by $U^{(\eta)}(\theta) = \sum_{i} c_i(\theta) {U_i}^{\otimes \eta}$. Such operators can be easily implemented using Ising gates with cascaded CNOT gates across the $\eta$-body space~\cite{whitfield2011simulation, kuhn2019accuracy}.
The Ising gate is not the only option as a two-body space variational circuit. For example, the real-valued symmetry-preserving (RSP) ansatz~\cite{ibe2022calculating} shown in Fig.~\ref{fig:multi_conf} can also be used as a two-body space variational circuit. The RSP ansatz is represented by the superposition of the following operators: \begin{equation} \begin{split} U_{\mathrm{RSP}}(\theta) &= \begin{pmatrix}
\cos \theta & 0 & 0 & -\sin \theta \\
0 & 0 & 1 & 0 \\
0 & 1 & 0 & 0 \\
\sin \theta & 0 & 0 & \cos \theta \\ \end{pmatrix}\\ &=\frac{1}{2} \left[ X \otimes X + Y \otimes Y + \cos\theta( I \otimes I + Z \otimes Z )\right. \\ &\qquad \left.- i\sin\theta(X \otimes Y + Y \otimes X) \right]. \label{eq:rsp} \end{split} \end{equation} Here, terms $X\otimes Y, Y\otimes X$ appear as tensor products of different operators $X$ and $Y$. Either of these terms alone destroys the antisymmetry of ansatz states, but pairs of them preserve antisymmetry as follows: \begin{equation} \begin{split} &(X \otimes Y + Y \otimes X)\ket{\psi_{\mathrm{SD}}} \\ &= \begin{vmatrix} X\ket{\alpha}_1 & Y\ket{\beta}_1 \\ X\ket{\alpha}_2 & Y\ket{\beta}_2 \\ \end{vmatrix} + \begin{vmatrix} Y\ket{\alpha}_1 & X\ket{\beta}_1 \\ Y\ket{\alpha}_2 & X\ket{\beta}_2 \\ \end{vmatrix}\\ &=\ket{\psi'_{\mathrm{SD}}}+\ket{\psi''_{\mathrm{SD}}}. \end{split} \end{equation} In Eq.~\ref{eq:rsp}, the RSP ansatz is represented by a real-valued unitary operator, which is convenient for describing real-valued eigenstates of one-dimensional systems. Unfortunately, however, it is not obvious how to extend this ansatz to a more-than-two-body space. Although the Ising gates are considered a more suitable option for generalizing to a system with a larger number of electrons, in the present study, the RSP ansatz was employed for convenience.
As illustrated in Fig.~\ref{fig:multi_conf}, a single Slater determinant is split each time $U^{(2)}(\theta)$ acts on it. Therefore, it is expected that an exponentially large number of Slater determinants will be generated by repeatedly applying the two-body space variational circuit followed by the one-body space variational circuit. However, such consecutive application of two-body variational circuits does not increase the number of superposed Slater determinants as one might expect. This is because a product of multiple Pauli operators reduces to a single Pauli operator due to the commutation relation among Pauli operators ($P_i^2=I,P_1P_2=iP_3$). This reduction of Pauli operators can be prevented by using the alternating layered structure of one-body and two-body space variational circuits as shown in Fig.~\ref{fig:overall_circuit}. If the one-body space variational circuit is noncommutative with the Pauli operators, then the reduction of Pauli operators between two-body space variational circuits is prevented. Therefore, by repeatedly applying one-body and two-body space variational circuits alternately to a seed state, the number of superposed Slater determinants can be increased exponentially, and so is a systematic approach to approximating the exact ground state.
During the optimization process of the VQE framework, variational parameters in one-body and two-body circuits are simultaneously optimized so as to minimize the energy expectation values of an ansatz state.
As can be seen from Eqs.~\ref{eq:hf_state}--\ref{eq:mc_state}, optimizing one-body and two-body circuits corresponds to optimizing the electronic states composing each Slater determinant (electronic orbitals) and the superposition coefficients of the Slater determinants (CI coefficients), respectively. Therefore, this framework can be regarded as the multi-configuration self-consistent field (MCSCF) method, which is one of the post-Hartree-Fock \textit{ab initio} quantum chemistry methods. A notable advantage of the proposed method is its ability to perform MCSCF calculations based on an exponential number of electron configurations with polynomial computational complexity.
\subsection{Settings for Numerical Simulations} To verify the validity of our method, we performed the VQE calculation with the proposed circuit to obtain the ground state of a 1D-$\rm{H_2}$ system. The Hamiltonian of this system is expressed as follows: \begin{equation} \begin{split}
H&=\sum_{i=1}^{2}\left[-\frac{1}{2}\frac{d^2}{dr_i^2}-\sum_{p=1}^{2}\frac{1}{\left|r_i-R_p+\epsilon\right|}\right]\\
&\qquad +\frac{1}{|r_1-r_2+\epsilon|}+\frac{1}{|R_1-R_2|}. \label{eq:hydrogen} \end{split} \end{equation} where $r_1, r_2$ and $R_1, R_2$ are the positions of electrons and protons, respectively. To avoid zero division in the Coulomb interaction terms, a small real value $\epsilon$ is added to the denominators of those terms (soft Coulomb potential).
\begin{figure}
\caption{Probability amplitudes $\ket{\psi(r_1,r_2)}$ of degenerated ground states (one fermionic state ($S^2=0, S_z=0$) and three bosonic states ($S^2=1, S_z=0,\pm1$)) at interatomic distance $|R_1-R_2| = $ 0.5 bohrs ($R_1, R_2$ are shown by pink dots).
Each many-body wave function is shown in the four subspaces corresponding to the spin configurations $\uparrow_1\uparrow_2, \uparrow_1\downarrow_2, \downarrow_1\uparrow_2, \downarrow_1\downarrow_2$.
Note that the correspondence between colors and values is different for each plot.}
\label{fig:exact}
\end{figure}
Note here that the ground states of fermions and bosons are completely degenerate. Figure~\ref{fig:exact} shows four degenerate ground states obtained by the exact diagonalization of Eq.~\ref{eq:hydrogen} at the interatomic distance $|R_1-R_2| = $ 0.5 bohrs. As shown in Fig.~\ref{fig:exact}, the ground states of the fermions are the singlet state ($S^2=0, S_z=0$) whose spin symmetry is antisymmetric and spatial symmetry is symmetric. On the other hand, the ground states of the bosons are the triplet state ($S^2=1, S_z=0,\pm1$) whose spin and spatial symmetry are symmetric. Since the Hamiltonian of the system depends only on the spatial coordinate, there is no energy difference between fermionic and bosonic ground states whose spatial symmetries are the same. Therefore, without taking into account the symmetry of the variational quantum circuit, the quantum states obtained by the VQE can be a mixture of fermionic and bosonic states, even though an accurate energy value is obtained. We demonstrate such convergence to a symmetry-neglected state in the numerical simulations.
The following describes the calculation conditions of the system and the quantum circuits used for numerical simulations. Six qubits (5 qubits for spatial coordinate and 1 qubit for spin coordinate) represent the one-body space, and thus 12 qubits were used for the two-electron system. Spatial coordinates of electrons $r_1, r_2$ ranged from -0.5 bohrs to 0.5 bohrs and the grid width $\varDelta r$ was $1/2^{5}$ bohrs = 0.03125 bohrs. $\epsilon$ was set to $\varDelta r/2$. The spatial grid employed in this study was limited to a very coarse one, due to the computational resources of our classical computer, but the realization of the FTQC is expected to allow calculations with a grid fine enough to reproduce the electronic structure of real materials. For variational circuits, the hardware efficient (HE) ansatz~\cite{kandala2017hardware} and the real-valued symmetry preserving ansatz (RSP)~\cite{ibe2022calculating} were employed. The HE ansatz consists of Ry gates and CNOT gates~\cite{shee2022qubit}, which in the present study had 6 layers.
\begin{figure}
\caption{Circuit architectures employed by VQE. (a) Symmetry-neglected (SN) architecture; (b) Hartree-Fock (HF) architecture; (c) Multi-configuration (MC) architecture.
HE in the figure denotes the hardware efficient ansatz with 6 layers and RSP denotes the real-valued symmetry preserving ansatz.}
\label{fig:circ_sn}
\label{fig:circ_hf}
\label{fig:circ_mc}
\label{fig:circ}
\end{figure}
In order to demonstrate the effects of the antisymmetrization and the multi-configuration, we implemented the VQE with three different circuits. The architectures of these circuits are shown in Fig.~\ref{fig:circ}. The first architecture is the symmetry-neglected (SN) architecture shown in Fig.~\ref{fig:circ}(a), consisting of consecutive HE ansatz blocks acting on the two-body space. In this architecture, the antisymmetry of the seed state is not considered to be preserved, and thus it is expected that the VQE will yield a neither antisymmetric (fermionic) nor symmetric (bosonic) state. The number of blocks of the HE ansatz was 6. The second architecture is the Hartree-Fock (HF) architecture shown in Fig.~\ref{fig:circ}(b), consisting of consecutive one-body space variational circuits. In this architecture, the antisymmetry of the seed state is preserved, but the VQE yields a ground state within the HF approximation. The third architecture is the multi-configuration (MC) architecture shown in Fig.~\ref{fig:circ}(c), consisting of alternately layered one-body and two-body space variational circuits. In this architecture, the VQE can achieve a ground state beyond the HF approximation, and its energy is expected to be lower than that obtained with the HF architecture.
To confirm this energy stabilization, potential energy curves were calculated for the HF and MC architectures,
at 16 points of interatomic distance $|R_1-R_2|$ ranging from $\varDelta r$ to $16\varDelta r$ = 0.5 bohrs. The number of one-body space variational circuits was 15 for both architectures and that of two-body space variational circuits was 14 for the MC architecture. For the optimization of variational parameters, the steepest descent method with the Adam optimizer~\cite{kingma2014adam} was employed and the number of optimization steps was set to 10000. All calculations for quantum circuits were performed by a noiseless state vector simulator.
\section{Result and discussion}
\begin{figure*}
\caption{Many-body wave functions obtained by the VQE
with (a) the SN architecture, (b) the HF architecture and (c) the MC architecture
at the interatomic distance $|R_1-R_2| = $ 0.5 bohrs ($R_1, R_2$ are shown by pink dots).}
\label{fig:wf_sn}
\label{fig:wf_hf}
\label{fig:wf_mc}
\label{fig:wave_func}
\end{figure*}
\subsection{Results of the VQE Calculations}
Figure~\ref{fig:wave_func} shows the many-body wave functions obtained by the VQE with the different architectures at an interatomic distance $|R_1-R_2| = $ 0.5 bohrs. In the case of the wave function obtained with the SN architecture shown in Fig.~\ref{fig:wave_func}(a), the spatial coordinate is symmetric under the exchange of two electrons, but its spin coordinate is neither antisymmetric nor symmetric ($\uparrow_1\uparrow_2+\uparrow_1\downarrow_2\neq\uparrow_1\uparrow_2+\downarrow_1\uparrow_2$). This indicates that unless the symmetry for the variational quantum circuit is not considered, the resulting state of the VQE can converge to such a symmetry-neglected state. This is the notable difference from the second-quantized VQE. In contrast, as shown in Figs.~\ref{fig:wave_func}(b)(c), the wave functions obtained with the HF and MC architectures are the antisymmetric singlet states, as expected. The shape of the wave function obtained with the MC architecture well reproduces that of the exact fermionic ground state shown in Fig.~\ref{fig:exact}, while that with the HF architecture is apparently different from it. As will be described later, this difference in shape of the wave functions indicates the difference in representability of the quantum states between the HF approximation and CI theory.
The difference between them is also reflected in their ground state energies. Figure~\ref{fig:pec_ee}(a) shows the potential energy curves obtained with the HF and MC architectures. It can be seen that the ground state energies obtained with the MC architecture reproduce the results of the exact diagonalization well, whereas those obtained with the HF architecture converge to higher values. This demonstrates the energy stabilization due to the multi-configuration character, in other words, the effect of electron correlation, which is the pillar of the quantum chemistry theories~\cite{jensen2017introduction, helgaker2013molecular}.
\begin{figure}
\caption{(a) Potential energy curves and (b) entanglement entropy
obtained by exact diagonalization (black dashed lines) and VQE with the MC architecture (blue solid lines) and the HF architecture (red solid lines).
The calculated potential energy curves are significantly different from those of the three-dimensional hydrogen molecular system,
which is attributed to the difference in dimensions between the systems~\cite{hirai2022molecular}.}
\label{fig:pec}
\label{fig:ee}
\label{fig:pec_ee}
\end{figure}
\subsection{Analysis of Many-Body States based on Quantum Information Theory}
Although the numerical experiments in this section confirmed that the proposed method can reproduce the exact fermionic ground state well, the obtained many-body wave function provides little insight into the microscopic electronic structure. The electronic structure is mostly understood from the electron orbital picture; however, since our method is based on a real-space basis, the orbital picture of the obtained state is not obvious. Furthermore, a lack of an orbital picture makes it impossible to quantitatively evaluate the multi-configuration character of the obtained state. In order to get a clear orbital picture and evaluate the multi-configuration character of obtained states, we analyzed many-body quantum states from the perspective of quantum information theory. Quantum information has been attracting increasing attention in recent years as a key to understanding electron dynamics in chemical reactions and strongly correlated materials ~\cite{molina2015quantum, brandejs2019quantum, esquivel2011quantum, rissler2006measuring, boguslawski2013orbital, lin2013spatial, soh2019long, zhu2020entanglement, lanata2014principle, chen2019incoherent}.
We will first quantify the multi-configuration character of many-body states. To evaluate the multi-configuration character, we employed the entanglement entropy, which is typically employed to measure the degree of entanglement between subsystems. We shall now briefly explain the definition of entanglement entropy. A many-body wave function of the two-electron system can be decomposed into multiple product states by the Schmidt decomposition, as follows: \begin{equation} \ket{\psi(r_1,r_2)} = \sum_i \lambda_i \ket{\mu_i(r_1)} \otimes \ket{\chi_i(r_2)}, \label{eq:Schmidt} \end{equation} where $\lambda_i$ are the Schmidt coefficients (superposition coefficients of product states)
and $\ket{\mu_i(r_1)}, \ket{\chi_i(r_2)}$ are the Schmidt basis on subsystems. The entanglement entropy $S$ is defined by the Shannon entropy of the probability distribution $|\lambda_i|^2$ as \begin{equation}
S = -\sum_i |\lambda_i|^2 \log_2|\lambda_i|^2. \label{eq:ee} \end{equation} Considering the fact that the Shannon entropy of a normal distribution is proportional to the logarithm of its variance,
the larger the variance of probability distribution $|\lambda_i|^2$, that is, the more product states are superposed, the larger the entanglement entropy. Since the entanglement entropy of a single Slater determinant $\ket{\psi_{\rm{HF}}}=1/\sqrt2(\ket{\alpha}_1 \otimes \ket{\beta}_2 - \ket{\beta}_1 \otimes \ket{\alpha}_2)$ is 1, that of an MC state is expected to be larger than 1.
Figure~\ref{fig:pec_ee}(b) shows the entanglement entropies calculated for the ground states obtained by the exact diagonalization and the MC architecture. As expected, the entanglement entropy of the MC state is larger than 1 and increases as interatomic distance increases, which is in good agreement with the exact result. This behavior of entanglement entropy indicates that the ground state becomes more multi-configurational as the interatomic distance approaches the dissociation limit, which is a well-known fact in the field of conventional quantum chemistry~\cite{jensen2017introduction, helgaker2013molecular}.
We can now extract the electron orbital picture from the many-body state represented in real space. From the expression of the Schmidt decomposition in Eq.~\ref{eq:Schmidt}, Schmidt basis $\ket{\mu_i(r_1)},\ket{\chi_i(r_2)}$ can be regarded as one-body wave functions, that is, the electron orbitals of the electrons, and the Schmidt coefficients as their contributions to the many-body state. Therefore, the Schmidt basis and the Schmidt coefficients provide insight into the electron orbital picture of many-body wave functions.
\begin{figure}\label{fig:svd}
\end{figure}
Figure~\ref{fig:svd} shows the results of the Schmidt decomposition performed for one anti-parallel spin subspace $\ket{\psi(r_{1\downarrow},r_{2\uparrow})}$ of the exact ground state (upper left part of Fig.~\ref{fig:exact}) and the MC state (Fig.~\ref{fig:wave_func}(c)). The left side of the figure shows the first three Schmidt coefficients $\lambda_0, \lambda_1, \lambda_2$ of each interatomic distance. As shown, the contribution of the zeroth product state $\lambda_0$ is nearly constant at all interatomic distances, whereas that of the first product state $\lambda_1$ increases as interatomic distance increases, which is consistent with the behavior of the entanglement entropy shown in Fig.~\ref{fig:pec_ee}(b). This suggests that the configuration interaction between the zeroth and first product states, that is, their superposition,
explains the energy stabilization due to the multi-configuration character. The right side of the figure shows the product states constituting the MC state at the interatomic distance $|R_1-R_2| = $ 0.5 bohrs. The electron orbitals constituting each product state are indicated by black lines in each figure. Using these electron orbitals, we will illustrate the energy stabilization due to the multi-configuration character in the following discussion.
The electron orbitals constituting the zeroth and first product states have peaks at the proton positions $R_1, R_2$ (pink dots in figure), and these orbitals correspond to bonding and antibonding orbitals. Following the typical explanation of the linear combination of atomic orbitals (LCAO) approximation, let $\ket{p_i(r)}$ be the electron orbitals distributed around the $i$th proton. Then the bonding and antibonding orbitals are represented as $\ket{\psi_{\mathrm{\sigma}}(r)}=\ket{p_1(r)}+\ket{p_2(r)}$ and $\ket{\psi_{\mathrm{\sigma^{*}}}(r)}=\ket{p_1(r)}-\ket{p_2(r)}$, respectively. The bonding state of two electrons $\ket{\psi_{\mathrm{\sigma\sigma}}(r_1,r_2)}$ is expressed as the tensor product of the bonding orbitals $\ket{\psi_{\mathrm{\sigma}}(r)}$ as follows: \begin{equation}
\begin{split}
&\ket{\psi_{\mathrm{\sigma\sigma}}(r_1,r_2)}\\
&=\ket{\psi_{\mathrm{\sigma}}(r_1)}\otimes\ket{\psi_{\mathrm{\sigma}}(r_2)} \\
&= \ket{p_1(r_1)} \otimes \ket{p_2(r_2)} + \ket{p_2(r_1)} \otimes \ket{p_1(r_2)}\\
&\qquad + \ket{p_1(r_1)} \otimes \ket{p_1(r_2)} + \ket{p_2(r_1)} \otimes \ket{p_2(r_2)},
\end{split} \end{equation} The four terms in the above expression correspond to the four peaks in the zeroth product state and in the HF ground state shown in Fig.~\ref{fig:wave_func}(b); thus, \begin{equation} \ket{\psi_{\mathrm{\sigma\sigma}}(r_1,r_2)} \sim \ket{\mu_0(r_1)}\otimes\ket{\chi_0(r_2)} \sim \ket{\psi_{\mathrm{HF}}(r_1,r_2)}. \end{equation} The antibonding state, or two-electron excited state, $\ket{\psi_{\mathrm{\sigma^{*}\sigma^{*}}}(r_1,r_2)}$ is expressed as the tensor product of the antibonding orbitals $\ket{\psi_{\mathrm{\sigma^{*}}}(r)}$ as follows: \begin{equation}
\begin{split}
&\ket{\psi_{\mathrm{\sigma^{*}\sigma^{*}}}(r_1,r_2)}\\
&=-\ket{\psi_{\mathrm{\sigma^{*}}}(r_1)}\otimes\ket{\psi_{\mathrm{\sigma^{*}}}(r_2)} \\
&= \ket{p_1(r_1)} \otimes \ket{p_2(r_2)} + \ket{p_2(r_1)} \otimes \ket{p_1(r_2)}\\
&\qquad - \ket{p_1(r_1)} \otimes \ket{p_1(r_2)} - \ket{p_2(r_1)} \otimes \ket{p_2(r_2)},
\end{split} \end{equation} where $-1$ is taken as a global phase. The four terms in the above expression correspond to the four peaks in the first product state; thus, \begin{equation} \ket{\psi_{\mathrm{\sigma^{*}\sigma^{*}}}(r_1,r_2)} \sim \ket{\mu_1(r_1)}\otimes\ket{\chi_1(r_2)}. \end{equation} Superposing bonding and antibonding states equally yields \begin{equation}
\begin{split}
&\ket{\psi_{\mathrm{\sigma\sigma+\sigma^{*}\sigma^{*}}}(r_1,r_2)}\\
&=\frac{1}{2}\ket{\psi_{\mathrm{\sigma\sigma}}(r_1,r_2)} + \frac{1}{2}\ket{\psi_{\mathrm{\sigma^{*}\sigma^{*}}}(r_1,r_2)}\\
&= \ket{p_1(r_1)} \otimes \ket{p_2(r_2)} + \ket{p_2(r_1)} \otimes \ket{p_1(r_2)}.
\end{split} \end{equation} It can be seen that energetically unfavorable electron configurations $\ket{p_1(r_1)} \otimes \ket{p_1(r_2)}$, $\ket{p_2(r_1)} \otimes \ket{p_2(r_2)}$, where both electrons are distributed around the same proton, are cancelled out and a lower energy state is achieved. The two terms in the above expression correspond to the two peaks in the MC state $\ket{\psi_{\mathrm{MC}}(r_1,r_2)}$ shown in Fig.~\ref{fig:wave_func}(c); thus, \begin{equation} \ket{\psi_{\mathrm{\sigma\sigma+\sigma^{*}\sigma^{*}}}(r_1,r_2)} \sim \ket{\psi_{\mathrm{MC}}(r_1,r_2)}. \end{equation} This state represents the electron configuration where electrons are distributed around different protons to avoid each other, reminiscent of correlated, or entangled, electrons. To summarize, a multi-configuration state is inherently an entangled state consisting of multiple product states, and electron correlation can be interpreted as quantum entanglement in the electron system.
Although this discussion on electron orbitals may be just a textbook subject~\cite{jensen2017introduction, helgaker2013molecular}, it is worth noting that such an orbital picture can be extracted from a real-space represented by many-body states by using the Schmidt decomposition. In addition, the minor orbital contributions, other than bonding and antibonding orbitals, can be considered in our method. Indeed, as shown in the left part of Fig.~\ref{fig:svd}, the exact ground state and the MC state include the contribution of the second product state $\lambda_2\ket{\mu_2(r_{1\downarrow})}\otimes\ket{\chi_2(r_{2\uparrow})}$ whose electron orbitals have nodes at proton positions and are distributed in the interstitial region. Since a real-space basis is a complete basis set of a discrete space, it can represent both localized atomic orbitals and such delocalized orbitals, freeing us from the need for careful selection of basis functions in typical quantum chemical calculations. From this point, a real-space basis is an attractive option for the quantum simulation in the FTQC era.
The results shown in this section demonstrate one possible framework of quantum chemistry in the FTQC era, where we understand electronic structures based on electron orbitals derived from many-body wave functions represented in real space. In this study, many-body wave functions were analyzed by a classical computer; however, since it is almost impossible to obtain a state vector of even a quantum circuit of a few dozen qubits, this analysis must be performed on a quantum computer. Fortunately, several methods have been proposed to perform singular value decomposition on quantum computers~\cite{rebentrost2018quantum, bravo2020quantum, wang2021variational}, which may allow such analysis to be performed on a quantum computer with reasonable computational resources.
The problem concerned in our method is the barren plateaus~\cite{mcclean2018barren, cerezo2021cost, wang2021noise} in the optimization process of the variational state. In fact, the disappearance of gradients of variational parameters was observed during the optimization process, which required a large number of optimization steps. Such barren plateaus are the most serious problem in most variational quantum algorithms. Nevertheless, as mentioned in the Introduction, since by using the QPE, the true ground state can be distilled from the approximated ground state prepared by the VQE, the incomplete convergence of the VQE due to the barren plateaus will be tolerated to some extent. The complementary combination of the VQE and the QPE will allow the state preparation for first-quantized methods.
\section{Summary and Conclusion} In this paper, we propose a state preparation method for the first-quantized quantum simulations based on the real-space representation. We employ a variational method for preparing ground states, and provide a design principle for constructing antisymmetrized variational quantum circuits. Our proposed circuits are capable of generating a superposition of exponentially large number of Slater determinants, that is, MC states with polynomial numbers of quantum gates. We performed VQE calculations for a 1D-$\rm{H_2}$ system and confirmed that the proposed circuit reproduces well the exact fermionic ground state. In addition to performing VQE, we quantitatively evaluated the multi-configuration character of many-body states by using the entanglement entropy between two electrons, and extracted the electronic orbital picture from many-body states represented in real space by the Schmidt decomposition.
Quantum computers, as demonstrated in this study, have great potential to simulate many-body electron systems and shed light on their quantum information. We believe that our proposed method will contribute to realizing the first-quantized simulation for electron dynamics and will bring a deeper understanding of electron systems to materials science in the FTQC era.
\nocite{*}
\end{document}
|
arXiv
|
{
"id": "2306.08434.tex",
"language_detection_score": 0.8183220624923706,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{abstract} The purpose of this article is to shed light on the question of how many and what kind of cusps a rational cuspidal curve on a Hirzebruch surface can have. We use birational transformations to construct rational cuspidal curves with four cusps on the Hirzebruch surfaces and find associated results for these curves. \end{abstract}
\title{Rational cuspidal curves with four cusps on Hirzebruch surfaces}
\setcounter{tocdepth}{1}
\tableofcontents
\section{Introduction}
Let $C$ be a reduced and irreducible curve on a smooth complex surface $X$. A point $p$ on $C$ is called a \emph{cusp} if it is singular and if the germ $(C,p)$ of $C$ at $p$ is irreducible. A curve $C$ is called \emph{cuspidal} if all of its singularities are cusps, and $j$-cuspidal if it has $j$ cusps $p_1, \ldots, p_j$. Let $g$ denote the geometric genus of the curve, and recall that the curve is called \emph{rational} if $g=0$.
Assume that two curves $C$ and $C'$, without common components, meet at a point $p$, then $p$ is called an \emph{intersection point} of $C$ and $C'$. If $f$ and $f'$ are local equations of $C$ and $C'$ at $p$, then by the \emph{intersection multiplicity} $(C \cdot C')_p$ of $C$ and $C'$ at $p$ we mean $$(C \cdot C')_p := \dim_{\ensuremath{\mathbb{C}}}\ensuremath{\mathscr{O}}_{X,p}/(f,f').$$
We sometimes view the intersection of $C$ and $C'$ as a $0$-cycle, and express this by the notation $C \cdot C'$, where $$C \cdot C' = \sum _{p \in C \cap C'}(C \cdot C')_pp.$$
For any two divisors $C$ and $C'$ on $X$, we calculate the \emph{intersection number} $C \ensuremath{\,.\,} C'$ using linear equivalence and the pairing ${\mathrm{Pic}(X) \times \mathrm{Pic}(X) \rightarrow \ensuremath{\mathbb{Z}}}$ \cite[Theorem V 1.1, pp.357--358]{Hart:1977}.
By \cite[Theorem V 3.9, p.391]{Hart:1977}, there exists for any curve $C$ on a surface $X$ a sequence of $t$ monoidal transformations, $$V=V_t \xrightarrow{\sigma_t} V_{t-1} \xrightarrow{} \cdots \xrightarrow{} V_1 \xrightarrow{\sigma_1} V_0=X,$$ such that the reduced total inverse image of $C$ under the composition $\sigma:V \rightarrow X$, $$D := \sigma^{-1}(C)_{\mathrm{red}},$$ is a \emph{simple normal crossing divisor} (SNC-divisor) on the smooth complete surface $V$ (see \cite{Iitaka}). The pair $(V,D)$ and the transformation $\sigma$ are referred to as an \emph{embedded resolution} of $C$, and it is called a \emph{minimal embedded resolution} of $C$ when $t$ is the smallest integer such that $D$ is an SNC-divisor.
Let $p$ be a cusp on a curve $C$, let $m$ denote the \emph{multiplicity} of $p$, and let $m_i$ denote the multiplicity of the infinitely near points $p_i$ of $p$. Then the \emph{multiplicity sequence} $\overline{m}$ of the cusp $p$ is defined to be the sequence of integers $$\overline{m}=[m,m_1,\ldots,m_{t-1}],$$ where $t$ is the number of monoidal tranformations in the local minimal embedded resolution of the cusp, and we have $m_{t-1}=1$ (see \cite{Brieskorn}). We follow the convention of compacting the notation by omitting the number of ending 1s in the sequence and indexing the number of repeated elements, for example, we write $$[6,6,3,3,3,2,1,1]=[6_2,3_3,2].$$ Note that there are further relations between the elements in the sequence (see \cite{FlZa95}).
The collection of multiplicity sequences of a cuspidal curve will be referred to as its \emph{cuspidal configuration}.
The question of how many and what kind of cusps a rational cuspidal curve on a given surface can have naturally arises, and in this article we give new answers to the last part of this question for rational cuspidal curves on Hirzebruch surfaces. The first part of the question is addressed in a subsequent article \cite{MOEONC}.
Our main results can be summarized in the following theorem. \begin{thm*} For all $e\geq0$ and $h \in \{0,1\}$, except the pairs $(e,k)=(h,k)=(0,0)$, the following rational cuspidal curves with four cusps exist on $\ensuremath{\mathbb{F}_e}$ and $\ensuremath{\mathbb{F}_h}$. \begin{table}[H]
\renewcommand\thesubtable{}
\setlength{\extrarowheight}{2pt} \centering
{\noindent \begin{tabular*}{1\linewidth}{@{}c@{\extracolsep{\fill}}l@{}l@{}c}
\toprule
{\textnormal{\textbf{Type}}} & {\textnormal{\textbf{Cuspidal configuration}}} & \textnormal{\textbf{For}} & \textnormal{\textbf{Surface}}\\ \toprule
{$(2k+1,4)$} & $[4_{k-1+e},2_3],[2],[2],[2]$& $k \geq 0$ & $\ensuremath{\mathbb{F}_e}$\\
{$(3k+1-h,5)$} & $[4_{2k-1+h},2_3],[2],[2],[2]$& $k \geq 0$ & $\ensuremath{\mathbb{F}_h}$ \\
{$(2k+2-h,4)$} & $[3_{2k-1+h},2],[2_3],[2],[2]$ & $k \geq 0$ &$\ensuremath{\mathbb{F}_h}$ \\
{$(k+1-h,3)$} & $[2_{n_1}],[2_{n_2}],[2_{n_3}],[2_{n_4}]$& $k \geq 2$ and $\sum n_i = 2k+h$ & $\ensuremath{\mathbb{F}_h}$\\
{$(0,3)$} & $[2],[2],[2],[2]$& & $\mathbb{F}_2$\\
\bottomrule
\end{tabular*}}
\label{tab:fourcusps}
\end{table} \end{thm*}
The lack of examples of curves with more than four cusps leads us to the following conjecture. \begin{conj*} A rational cuspidal curve on a Hirzebruch surface has at most four cusps. \end{conj*}
Associated to these curves, we find results regarding the Euler characteristic of the logarithmic tangent sheaf $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ and the maximal multiplicity of a cusp on a rational cuspidal curve. Moreover, we give an example of a real rational cuspidal curve with four real cusps on a Hirzebruch surface.
\subsection{Structure} In Section \ref{MN} we motivate the study of rational cuspidal curves on Hirzebruch surfaces by recalling the history of the study of such curves on the projective plane. In Section \ref{PR} we give the basic definitions and preliminary results for (rational) cuspidal curves on Hirzebruch surfaces. Section \ref{4C} contains the main results of this article. Here we construct several series of rational cuspidal curves with four cusps on Hirzebruch surfaces. In Section \ref{AR} we present associated results, and in particular we compute $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ for the curves we have found in Section \ref{4C}. Moreover, we find a lower bound for the multiplicity of the cusp with the highest multiplicity. Additionally, we construct a real rational cuspidal curve with four real cusps. In the appendix we construct some of the rational cuspidal curves with four cusps on Hirzebruch surfaces that were found in Section \ref{4C}.
\section{Motivation: the case of plane curves}\label{MN} Let $\ensuremath{\mathbb{P}^2}$ denote the projective plane with coordinates $(x:y:z)$ and coordinate ring $\ensuremath{\mathbb{C}}[x,y,z]$. A reduced and irreducible curve $C$ on $\ensuremath{\mathbb{P}^2}$ is given as the zero set $\ensuremath{\mathscr{V}}(F)$ of a homogeneous, reduced and irreducible polynomial $F(x,y,z) \in \ensuremath{\mathbb{C}}[x,y,z]_d$ for some $d$. In this case, the polynomial $F$ and the curve $C$ is said to have degree $d$. Let $q$ be a smooth point on a plane curve $C$ with tangent $T$. Recall that the point $q$ is called an \emph{inflection point} if the local intersection multiplicity $(C \cdot T)_q$ of $C$ and $T$ at $q$ is $\geq 3$.
Plane rational cuspidal curves have been studied quite intensively both classically and the last 20 years. Classically, the study was part of the process of classifying plane curves, and additionally bounds on the number of cusps were produced (see \cite{Clebsch, Lefschetz, Salmon, Telling, Veronese, Wieleitner}). In the modern context, rational cuspidal curves with many cusps play an important role in the study of open surfaces (see \cite{Fent, FlZa94, Wak}), and the study of these curves was further motivated in the mid 1990s by Sakai in \cite{Sakai}. The two tasks at hand were first to classify all rational and elliptic cuspidal plane curves, and second to find the maximal number of cusps on a rational cuspidal plane curve.
The first complete classification of rational cuspidal curves of degree $d \leq 5$ that we have found, up to cuspidal configuration, is by Namba in \cite{Namba}, and a list of the curves of degree $d=5$ can be found in Table \ref{tab:degree5} \cite[Theorem 2.3.10, pp.179--182]{Namba}. The subclassification of the curves is due to the appearance of inflection points. The remarkable thing about this list of curves is that there are fewer curves with many cusps than expected. In fact, for curves of degree $d \leq 5$ there are only three curves with three cusps and only one with four cusps. For curves in higher degrees, less is known, but there is a classification of curves of degree $d=6$ by Fenske in \cite{Fent}, and in that case the highest number of cusps is three, and up to cuspidal configuration there are only two such curves.
\begin{table}[htb]
\renewcommand\thesubtable{}
\setlength{\extrarowheight}{2pt} \centering {\small
{\noindent \begin{tabular*}{\linewidth}{@{}c@{\extracolsep{\fill}}l@{}c}
\toprule
{\textbf{Curve}} &{\textbf{Cuspidal conf.}}& {\textbf{Parametrization}}\\
\toprule
$C_{1A}$&$\quad[4]$& $(s^5:st^4:t^5)$\\
$C_{1B}$&$\quad[4]$&$(s^5-s^4t:st^4:t^5)$\\
$C_{1C}$&$\quad[4]$&$(s^5+as^4t-(1+a)s^2t^3:st^4:t^5), \; a\neq-1$\\
$C_2$&$\quad [2_6]$&$(s^4t:s^2t^3-s^5:t^5-2s^3t^2)$\\
\midrule
$C_{3A}$&$\quad [3,2],[2_2]$ &$(s^5:s^3t^2:t^5)$\\
$C_{3B}$&$\quad [3,2],[2_2]$ &$(s^5:s^3t^2:st^4+t^5)$\\
$C_4$&$\quad [3], [2_3]$&$(s^4t-\frac{1}{2}s^5:s^3t^2:\frac{1}{2}st^4+t^5)$\\
$C_5$&$\quad [2_4],[2_2]$&$(s^4t-s^5:s^2t^3-\frac{5}{32}s^5:-\frac{47}{128}s^5+\frac{11}{16}s^3t^2+st^4+t^5)$\\
\midrule
$C_6$&$\quad [3],[2_2],[2]$&$(s^4t-\frac{1}{2}s^5:s^3t^2:-\frac{3}{2}st^4+t^5)$\\
$C_7$&$\quad [2_2],[2_2],[2_2]$&$(s^4t-s^5:s^2t^3-\frac{5}{32}s^5:-\frac{125}{128}s^5-\frac{25}{16}s^3t^2-5st^4+t^5)$\\
\midrule
$C_8$&$\quad [2_3],[2],[2],[2]$&$(s^4t:s^2t^3-s^5:t^5+2s^3t^2)$\\
\bottomrule
\end{tabular*}} }
\caption[Plane rational cuspidal curves of degree five.]{Plane rational cuspidal curves of degree five.}
\label{tab:degree5}
\end{table}
In two articles from the 1990s Flenner and Zaidenberg construct two series of rational cuspidal curves with three cusps, where one cusp has a relatively high multiplicity \cite{FlZa95, FlZa97}. A third series was found and further exploration of such curves was done by Fenske in \cite{Fen99a,Fent}. The known rational cuspidal curves with three or more cusps can be listed as follows.
\begin{enumerate}[$(I)$] \item For $d=5$, a rational cuspidal curve with three or more cusps has one of these cuspidal configurations \cite[Theorem 2.3.10, pp.179--182]{Namba}: \begin{align*} &[3],[2_2],[2],\\ &[2_2],[2_2],[2_2],\\ &[2_3],[2],[2],[2]. \end{align*} \item For any $a\geq b \geq 1$ there exists a rational cuspidal plane curve $C$ of degree $d=a+b+2$ with three cusps \cite[Theorem 3.5, p.448]{FlZa95}: \begin{equation*} [d-2],[2_a],[2_b]. \end{equation*} \item For any $a \geq 1$, there exists a rational cuspidal plane curve $C$ of degree $d=2a+3$ with three cusps \cite[Theorem 1.1, p.94]{FlZa97}: \begin{equation*} [d-3,2_a],[3_a],[2]. \end{equation*} \item For any $a \geq 1$, there exists a rational cuspidal plane curve $C$ of degree $d=3a+4$ with three cusps \cite[Theorem 1.1, p.512]{Fen99a}: \begin{equation*} [d-4,3_a],[4_a,2_2],[2]. \end{equation*} \end{enumerate} All these curves are constructed explicitly by successive Cremona transformations of plane curves of low degree. In cases $(II)$ and $(III)$ it is proved by Flenner and Zaidenberg in \cite{FlZa95,FlZa97} that these are the only tricuspidal curves with maximal multiplicity of this kind. The same is proved by Fenske in \cite{Fen99a} for case $(IV)$ under the assumption that $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})=0$. Note that this result is originally proved with $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})\leq0$, but by a result by Tono in \cite{Tono05} only $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})=0$ is needed. The curves in $(II)$, $(III)$ and $(IV)$ constitute three series of cuspidal curves with three cusps, with infinitely many curves in each series.
The lack of examples of curves with more than four cusps, leads to a conjecture originally proposed by Orevkov. \begin{conj} A plane rational cuspidal curve can not have more than four cusps. \end{conj}
Further research by Fenske in \cite[Section 5]{Fent} and Piontkowski in \cite{Piontkowski} implies that there are no other tricuspidal curves, hence the above conjecture was extended \cite{Piontkowski}. \begin{conj} A plane rational cuspidal curve can not have more than four cusps. If it has three or more cusps, then it is given in the above list. \end{conj}
Associated to these results, in \cite{FlZa94} Flenner and Zaidenberg conjectured that $\ensuremath{\mathbb{Q}}$-acyclic affine surfaces $Y$ with logarithmic Kodaira dimension $\overline{\kappa}(Y)= 2$ are \emph{rigid} and \emph{unobstructed}. Complements to rational cuspidal curves with three or more cusps on the projective plane are examples of such surfaces. In that case, the conjecture says that for $(V,D)$ the minimal embedded resolution of the plane curve, the Euler characteristic of the logarithmic tangent sheaf vanishes, $$\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})=\mathrm{h}^2(V,\ensuremath{\Theta_V \left\langle D \right\rangle})-\mathrm{h}^1(V,\ensuremath{\Theta_V \left\langle D \right\rangle})+\mathrm{h}^0(V,\ensuremath{\Theta_V \left\langle D \right\rangle})=0,$$ and in particular, $\mathrm{h}^1(V,\ensuremath{\Theta_V \left\langle D \right\rangle})=\mathrm{h}^2(V,\ensuremath{\Theta_V \left\langle D \right\rangle})=0$. Note that it is known by \cite[Theorem 6]{Iit} that $\mathrm{h}^0(V,\ensuremath{\Theta_V \left\langle D \right\rangle})=0$. This conjecture is referred to as the \emph{rigidity conjecture}.
In the study of plane rational cuspidal curves, there has, moreover, been found results bounding the multiplicity of the cusp with the highest multiplicity. This gives restrictions on the possible cuspidal configurations on such a curve. The first result on this matter is by Matsuoka and Sakai in \cite{MatsuokaSakai}, where it is shown that $$d<3\hat{m},$$ where with $m_j$ the multiplicity of the cusp $p_j$, $\hat{m}$ is defined to be $\hat{m}:=\max\{m_j\}$. Improved inequalities were found by Orevkov in \cite{Ore}, where for $\alpha=\frac{3 + \sqrt{5}}{2}$, the bound is improved in general to $$d < \alpha (\hat{m}+1)+\frac{1}{\sqrt{5}},$$ for curves with $\ol{\kappa}(\Po \setminus C)=-\infty$, this is further improved to $d < \alpha \hat{m}$, and for curves with $\ol{\kappa}(\Po \setminus C)=2$, the bound is $$d < \alpha (\hat{m}+1)-\frac{1}{\sqrt{5}}.$$
Note also the result by Tono in \cite{Tono05} on the number of cusps on a plane cuspidal curve. For plane \emph{rational} cuspidal curves, the upper bound on the number of cusps is 8. We deal with this question for cuspidal curves on Hirzebruch surfaces in a subsequent article \cite{MOEONC}.
\section{Notation and preliminary results}\label{PR} In this section we recall what we mean by a curve on a Hirzebruch surface and state preliminary results that give restrictions on the cuspidal configurations of rational cuspidal curves in this case.
We first recall some basic facts about the Hirzebruch surfaces. Let $\mathbb{F}_e$ denote the Hirzebruch surface of type $e$ for any $e \geq 0$. Recall that $\ensuremath{\mathbb{F}_e}$ is a projective ruled surface, with $\ensuremath{\mathbb{F}_e}=\mathbb{P}(\ensuremath{\mathscr{O}} \oplus \ensuremath{\mathscr{O}}(-e))$ and morphism $\pi: \mathbb{F}_e \longrightarrow \ensuremath{\mathbb{P}^1}$. Moreover, $p_a(\mathbb{F}_e)=0$ and $p_g(\mathbb{F}_e)=0$ \cite[Corollary V 2.5, p.371]{Hart:1977}.
For any $e \geq 0$, the surface $\ensuremath{\mathbb{F}_e}$ can be considered as the toric variety associated to a fan $\Sigma_e \subset \ensuremath{\mathbb{Z}}^2$, where the rays of the fan $\Sigma_e$ are generated by the vectors \begin{displaymath} v_1=\begin{bmatrix}1\\0 \end{bmatrix}, \quad v_2=\begin{bmatrix}0\\1 \end{bmatrix}, \quad v_3=\begin{bmatrix}-1\\e \end{bmatrix}, \quad v_4=\begin{bmatrix}0 \\-1 \end{bmatrix}. \end{displaymath} The coordinate ring of $\ensuremath{\mathbb{F}_e}$ (see \cite{COX}) is denoted by $S_e := \mathbb{C}[x_{0},x_{1},y_{0},y_{1}]$, where the variables can be given a grading by $\mathbb{Z}^{2}$, \begin{eqnarray*} \deg x_0 & = & (1,0),\\ \deg x_1 & = & (1,0),\\ \deg y_0 & = & (0,1),\\ \deg y_1 & = & (-e,1). \end{eqnarray*}
Let $S_e(a,b)$ denote the $(a,b)$-graded part of $S_e$, \begin{displaymath} S_e(a,b):=\mathrm{H}^0(\ensuremath{\mathbb{F}_e},\ensuremath{\mathscr{O}}_{\ensuremath{\mathbb{F}_e}}(a,b)) = \bigoplus_{\substack{\alpha_{0}+\alpha_{1}-e\beta_{1} = a \\ \beta_{0}+\beta_{1} = b}} \mathbb{C}x_{0}^{\alpha_{0}}x_{1}^{\alpha_{1}}y_{0}^{\beta_{0}}y_{1}^{\beta_{1}}. \end{displaymath}
A reduced and irreducible curve $C$ on $\ensuremath{\mathbb{F}_e}$ is given as the zero set $\ensuremath{\mathscr{V}}(F)$ of a reduced and irreducible polynomial $F(x_0,x_1,y_0,y_1) \in S_e(a,b)$. In this case, the polynomial $F$ is said to have bidegree $(a,b)$ and the curve $C$ is said to be of type $(a,b)$.
In the language of divisors, let $L$ be a \emph{fiber} of $\pi: \mathbb{F}_e \longrightarrow \ensuremath{\mathbb{P}^1}$ and $M_0$ the \emph{special section} of $\pi$. The Picard group of $\ensuremath{\mathbb{F}_e}$, $\mathrm{Pic}(\mathbb{F}_e)$, is isomorphic to $\ensuremath{\mathbb{Z}} \oplus \ensuremath{\mathbb{Z}}$. We choose $L$ and $M \sim eL+M_0$ as a generating set of ${\rm{Pic}}(\ensuremath{\mathbb{F}_e})$, and then we have \cite[Theorem V 2.17, p.379]{Hart:1977} $$L^2=0, \qquad L \ensuremath{\,.\,} M=1, \qquad M^2=e.$$ The canonical divisor $K$ on $\ensuremath{\mathbb{F}_e}$ can be expressed as \cite[Corollary V 2.11, p.374]{Hart:1977} $$K \sim (e-2)L-2M \; \text{ and }\; K^2=8.$$ Any irreducible curve $C\neq L,M_0$ corresponds to a divisor on $\mathbb{F}_e$ given by \cite[Proposition V 2.20, p.382]{Hart:1977} $$C \sim aL+bM,\quad b>0, \,a\geq 0.$$
Recall that there are \emph{birational transformations} between these surfaces and the projective plane, and these can be given in a quite explicit way (see \cite{MOEPHD}). Using the birational transformations, we are able to construct curves on one surface from a curve on another surface by taking the strict transform.
A first result concerning cuspidal curves on $\mathbb{F}_e$ regards the genus $g$ of the curve. \begin{cor}[Genus formula]\label{genusfe} A cuspidal curve $C$ of type $(a,b)$ with cusps $p_j$, for $j=1,\ldots,s$, and multiplicity sequences $\overline{m}_j=[m_0,m_1, \ldots, m_{t_j-1}]$ on the Hirzebruch surface $\mathbb{F}_e$ has genus $g$, where $$g=\frac{(b-1)(2a-2+be)}{2}-\sum_{j=1}^s \sum_{i=0}^{t_j-1} \frac{m_i(m_i-1)}{2}.$$ \end{cor}
\begin{proof} Recall that $C \sim aL+bM$, $K \sim (e-2)L-2M$, $L^2=0$, $L \ensuremath{\,.\,} M = 1$ and $M^2=e$. By the general genus formula \cite[Example V 3.9.2, p.393]{Hart:1977}, we have $$g=\frac{(aL+bM) \ensuremath{\,.\,} (aL+bM+(e-2)L-2M)}{2}+1-\sum_{j=1}^s \delta_j,$$ where $\delta$ is the \emph{delta invariant}. This gives \begin{align*} g&=\frac{b^2e-2be+ab+be-2a+ab-2b}{2}+1-\sum_{j=1}^s \sum_{i=0}^{t_j-1} \frac{m_i(m_i-1)}{2}\\ &=\frac{(b-1)(2a-2+be)}{2}-\sum_{j=1}^s \sum_{i=0}^{t_j-1} \frac{m_i(m_i-1)}{2}. \end{align*} \end{proof}
Secondly, the structure of the Hirzebruch surfaces gives restrictions on the multiplicity sequence of a cusp on a curve on such a surface. \begin{thm} \label{thm:multb} Let $p$ be a cusp on a reduced and irreducible curve $C$ of type $(a,b)$ with multiplicity sequence $\overline{m}=[m,m_1,\ldots, m_{t-1}]$ on a Hirzebruch surface $\ensuremath{\mathbb{F}_e}$. Then $m \leq b$. \end{thm}
\begin{proof} The coordinates of the point $p$ determine a unique fiber $L$. By intersection theory, $m \leq (L \cdot C)_p$ \cite[Exercise I 5.4, p.36]{Hart:1977}. Moreover, $(L \cdot C)_p \leq L \ensuremath{\,.\,} C$. By intersection theory again, $L \ensuremath{\,.\,} C = b$. Hence, $m \leq (L \cdot C)_p \leq L \ensuremath{\,.\,} C = b$. \end{proof}
Further restrictions on the type of points on a curve on $\ensuremath{\mathbb{F}_e}$ can be found using Hurwitz's theorem \cite[Corollary IV 2.4, p.301]{Hart:1977}. First, the general result in this situation.
\begin{thm}[Hurwitz's theorem for $\ensuremath{\mathbb{F}_e}$]\label{RHFe} Let $C$ be a curve of genus $g$ and type $(a,b)$, where $b > 0$, on a Hirzebruch surface $\ensuremath{\mathbb{F}_e}$ with $e>0$. Let $\tilde{C}$ denote the normalization of $C$, and let $\nu$ be the composition of the normalization map $\tilde{C} \rightarrow C$ and the projection map $C \rightarrow \ensuremath{\mathbb{P}^1}$ of degree $b$. Let $e_p$ denote the ramification index of a point $p \in \tilde{C}$ with respect to $\nu$. Then the following equality holds, $$2b+2g-2 = \sum_{p \in \tilde{C}} (e_p-1).$$ When $e=0$, for curves $C$ of genus $g$ and type $(a,b)$, with $a,b>0$, we have $$2\min\{a,b\}+2g-2 = \sum_{p \in \tilde{C}} (e_p-1).$$ \end{thm}
\begin{proof} The result follows from \cite[Corollary IV 2.4, p.301]{Hart:1977}. With $\nu$ as above, we get $$2b+2g-2 = \sum_{p \in \tilde{C}} (e_p-1).$$ When $e=0$, use the projection map of lower degree, i.e., $\min\{a,b\}$. \end{proof}
An immediate corollary gives restrictions on the multiplicities of cusps on a curve. \begin{cor}\label{RHM} Let $C$ be a cuspidal curve of type $(a,b)$ and genus $g$ with $s>0$ cusps $p_j$ with multiplicities $m_j$ on a Hirzebruch surface $\ensuremath{\mathbb{F}_e}$ with $e>0$. Then the following inequality holds, $$2b+2g-2 \geq \sum_{j=1}^s (m_j-1).$$ When $e=0$, the same is true with $\min\{a,b\}$ instead of $b$. \end{cor}
\begin{proof} The cusps $p_j$ of $C$ gives branching points with ramification index bigger than or equal to the multiplicity $m_j$, so the result follows from Theorem \ref{RHFe}.
\end{proof}
\section{Rational cuspidal curves with four cusps}\label{4C} In this section we give examples of rational cuspidal curves with four cusps on Hirzebruch surfaces, and our aim is to shed light on the question of how many and what kind of cusps a rational cuspidal curve on a Hirzebruch surface can have.
We are not able to construct many rational cuspidal curves with four cusps on the Hirzebruch surfaces. Indeed, on each $\ensuremath{\mathbb{F}_e}$, with $e \geq 0$, we construct one infinite series of rational cuspidal curves with four cusps. On $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ and $\ensuremath{\mathbb{F}_1}$ we construct another three infinite series of rational cuspidal curves with four cusps, and on $\mathbb{F}_2$ we construct a single additional rational cuspidal curve with four cusps.
The following theorem presents the series of rational fourcuspidal curves that consists of curves on all the Hirzebruch surfaces. In Appendix \ref{Appendix:A} we construct some of the curves from a plane fourcuspidal curve using birational maps and \begin{verbatim}Maple\end{verbatim} \cite{Maple}. \begin{thm} \label{thm:4cusp} For all $e \geq 0$ and $k \geq 0$, except for the pair ${(e,k)=(0,0)}$, there exists on the Hirzebruch surface $\ensuremath{\mathbb{F}_e}$ a rational cuspidal curve $C_{e,k}$ of type $(2k+1,4)$ with four cusps and cuspidal configuration $$[4_{k-1+e},2_3],[2],[2],[2].$$ \end{thm}
\begin{proof} We will show that for each $e \geq 0$ there is an infinite series of curves on $\ensuremath{\mathbb{F}_e}$, and we show this by induction on $k$. The proof is split in two, and we treat the case of $k$ odd and even separately. We construct the series of curves $C_{e,0}$ for $e \geq 1$, and then we construct the initial series $C_{e,1}$ and $C_{e,2}$, with $e \geq0$. We only treat the induction to prove the existence of $C_{e,k}$ for odd values of $k$, as the proof for even values of $k$ is completely parallel.
Let $C$ be the rational cuspidal curve of degree 5 on $\ensuremath{\mathbb{P}^2}$ with cuspidal configuration $[2_3],[2],[2],[2]$. Let $p$ be the cusp with multiplicity sequence $[2_3]$, and let $T$ be the tangent line to $C$ at $p$. Then $T \cdot C=4 p+r$, with $r$ a smooth point on $C$. Blowing up at $r$, the strict transform of $C$ is a curve $C_{1,0}$ of type $(1,4)$ on $\ensuremath{\mathbb{F}_1}$ with cuspidal configuration $[2_3],[2],[2],[2]$. Letting $T_{1,0}$ denote the strict transform of $T$ and $p_{1,0}$ the strict transform of $p$, we have ${T_{1,0} \cdot C_{1,0}=4 p_{1,0}}$. We observe that $p_{1,0}$ is fiber tangential. Let $E_{1}$ denote the special section on $\ensuremath{\mathbb{F}_1}$, and let $s_{0,1}=E_{1} \cap T_{1,0}$.
From $C_{1,0}$ we can proceed with the construction of curves on Hirzebruch surfaces in three ways.
First, we show by induction on $e$ that the curves $C_{e,0}$ exist on the Hirzebruch surfaces $\ensuremath{\mathbb{F}_e}$, for all $e \geq 1$. We have already seen that $C_{1,0}$ exists on $\ensuremath{\mathbb{F}_1}$, and that there exists a fiber $T_{1,0}$ with the property that ${T_{1,0} \cdot C_{1,0}=4 p_{1,0}}$ for the first cusp $p_{1,0}$. Now assume $e \geq 2$ and that the curve $C_{e-1,0}$ of type $(1,4)$ exists on $\mathbb{F}_{e-1}$ with cuspidal configuration $[4_{e-2},2_3],[2],[2],[2]$, where $p_{e-1,0}$ denotes the first cusp and $T_{e-1,0}$ has the property that ${T_{e-1,0} \cdot C_{e-1,0}=4 p_{e-1,0}}$. Then, with $E_{e-1}$ the special section of $\mathbb{F}_{e-1}$, blowing up at the intersection $s_{e-1,0} \in E_{e-1} \cap T_{e-1,0}$ and contracting $T_{e-1,0}$, we get $C_{e,0}$ on $\ensuremath{\mathbb{F}_e}$ of type $(1,4)$ with cuspidal configuration $[4_{e-1},2_3],[2],[2],[2]$. Moreover, we note that there exists a fiber $T_{e,0}$ with ${T_{e,0} \cdot C_{e,0}=4 p_{e,0}}$. So the series exists on all $e \geq 1$ for $k=0$.
Second, note that from the curve $C_{1,0}$ on $\ensuremath{\mathbb{F}_1}$ it is possible to construct the curve $C_{0,1}$ on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ by blowing up at $p_{1,0}$ before contracting $T_{1,0}$. The curve $C_{0,1}$ is a curve of type $(3,4)$ with cuspidal configuration ${[2_3],[2],[2],[2]}$, and there is a fiber $T_{0,1}$ such that ${T_{0,1}\ensuremath{\,.\,} C_{0,1}=4p_{0,1}}$. Blowing up at a point ${s_{0,1} \in T_{0,1} \setminus \{p_{0,1}\}}$ and contracting $T_{0,1}$ result in the curve $C_{1,1}$ on $\ensuremath{\mathbb{F}_1}$ of type $(3,4)$ with cuspidal configuration $[4,2_3],[2],[2],[2]$. Moreover, there exists a fiber $T_{1,1}$ with ${T_{1,1} \cdot C_{1,1}=4 p_{1,1}}$ and $p_{1,1} \notin E_1$. The same induction on $e$ as above proves that the series exists for $k=1$.
Third, note that from the curve $C_{1,0}$ on $\ensuremath{\mathbb{F}_1}$ it is possible to construct the curve $C_{0,2}$ on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ by blowing up at a point ${t_{1,0} \in T_{1,0} \setminus \{p_{1,0}, s_{1,0}\}}$ before contracting $T_{1,0}$. The curve $C_{0,2}$ is a curve of type $(5,4)$ with cuspidal configuration $[4,2_3],[2],[2],[2]$, and there is a fiber $T_{0,2}$ such that ${T_{0,2}\cdot C_{0,2}=4p_{0,2}}$. Blowing up at a point ${s_{0,2} \in T_{0,2} \setminus \{p_{0,2}\}}$ and contracting $T_{0,2}$ give the curve $C_{1,2}$ on $\ensuremath{\mathbb{F}_1}$ of type $(5,4)$ with cuspidal configuration ${[4_2,2_3],[2],[2],[2]}$. Moreover, there exists a fiber $T_{1,2}$ with ${T_{1,2} \cdot C_{1,2}=4 p_{1,2}}$ and $p_{1,2} \notin E_1$. The same induction on $e$ as above proves that the series exists for $k=2$.
Next assume $k \geq 3$, with $k$ odd, and that there exists a series of curves $C_{e,k-2}$ of type $(2k-3,4)$ on $\ensuremath{\mathbb{F}_e}$ for all $e \geq 0$ with cuspidal configuration $[4_{e+k-3},2_3],[2],[2],[2]$. Then, in particular, the curve $C_{1,k-2}$ on $\ensuremath{\mathbb{F}_1}$ with cuspidal configuration \linebreak $[4_{k-2},2_3],[2],[2],[2]$ exists. Moreover, there exists a fiber $T_{1,k-2}$ on $\ensuremath{\mathbb{F}_1}$ such that ${T_{1,k-2} \cdot C_{1,k-2}=4 p_{1,k-2}}$, where $p_{1,k-2}$ denotes the cusp with multiplicity sequence $[4_{k-2},2_3]$. With $E_1$ the special section on $\ensuremath{\mathbb{F}_1}$, let ${s_{1,k-2} \in E_1 \cap T_{1,k-2}}$. We now blow up at a point $t_{1,k-2} \in T_{1,k-2} \setminus \{p_{1,k-2},s_{1,k-2}\}$ and subsequently contract $T_{1,k-2}$. This gives the curve $C_{0,k}$ on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ of type $(2k+1,4)$ with cuspidal configuration $[4_{k-1},2_3],[2],[2],[2]$. With $T_{0,k}$ the strict transform of the exceptional line of the latter blowing up, we have ${T_{0,k} \cdot C_{0,k}=4 p_{0,k}}$. Blowing up at a point $s_{0,k} \in T_{0,k} \setminus \{p_{0,k}\}$ and contracting $T_{0,k}$ gives the curve $C_{1,k}$ on $\ensuremath{\mathbb{F}_1}$ of type $(2k+1,4)$ with cuspidal configuration $[4_k,2_3],[2],[2],[2]$. Moreover, there is a fiber $T_{1,k}$ with the property that ${T_{1,k}\cdot C_{1,k}=4p_{1,k}}$. With the same induction on $e$ as above, we get the series of curves $C_{e,k}$. \end{proof}
There are three infinite series of rational fourcuspidal curves that can be found on the Hirzebruch surfaces $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ and $\ensuremath{\mathbb{F}_1}$.
Before we list these three series, we consider the rational cuspidal curves with four cusps on $\ensuremath{\mathbb{F}_1}$ that we can get by blowing up a single point on $\ensuremath{\mathbb{P}^2}$. These curves represent examples from the series. \begin{thm} Let $C$ be the rational cuspidal curve with four cusps of degree 5 on $\ensuremath{\mathbb{P}^2}$. The following rational cuspidal curves with four cusps on $\mathbb{F}_1$ can be constructed from $C$ by blowing up a single point on $\ensuremath{\mathbb{P}^2}$. \begin{table}[H]
\renewcommand\thesubtable{}
\setlength{\extrarowheight}{2pt} \centering
{\noindent \begin{tabular*}{0.75\linewidth}{@{}c@{\extracolsep{\fill}}c@{}c@{}l}
\toprule
\textnormal{\textbf{\# Cusps}} & \textnormal{\textbf{Curve}} & \textnormal{\textbf{Type}} & \textnormal{\textbf{Cuspidal configuration}} \\
\toprule
\multirow{3}{0pt}{4} & $C_1$ & $(0,5)$& $[2_3],[2],[2],[2]$ \\
&$C_2$ &$(1,4)$ & $[2_3],[2],[2],[2]$ \\
&$C_3$ &$(2,3)$ & $[2_2],[2],[2],[2]$ \\
\bottomrule
\end{tabular*}}
\caption {Rational cuspidal curves on $\mathbb{F}_1$ with four cusps.}
\label{tab:prccw4}
\end{table} \end{thm}
\begin{proof} The curve $C_1$ is constructed by blowing up any point $r$ on $\ensuremath{\mathbb{P}^2} \setminus C$. Note that if $r$ is on the tangent line to a cusp on $C$, then $C_1$ has cusps that are fiber tangential. If $r$ is only on tangent lines of smooth points on $C$, then $C_1$ has smooth fiber tangential points.
The curve $C_2$ is constructed by blowing up any smooth point $r$ on $C$. Again, if $r$ is on a tangent line of $C$, $C_2$ will have points that are fiber tangential.
The curve $C_3$ is constructed by blowing up the cusp with multiplicity sequence $[2_3]$. \end{proof}
\noindent The fact that we can construct curves where the curve have nontransversal intersection with some fiber(s) is crucial in the later constructions. Although we do not get new cuspidal configurations in this first step, cusps can sometimes be constructed later on.
We now give the three series of rational cuspidal curves with four cusps on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ and $\ensuremath{\mathbb{F}_1}$. For notational purposes we denote these surfaces by $\ensuremath{\mathbb{F}_h}$, with $h \in \{0,1\}$, in the theorems. \begin{thm} For $h\in\{0,1\}$ and all integers $k\geq 0$, except the pair $(h,k)=(0,0)$, there exists on the Hirzebruch surface $\ensuremath{\mathbb{F}_h}$ a rational cuspidal curve $C_{h,k}$ of type $(3k+1-h,5)$ with four cusps and cuspidal configuration $$[4_{2k-1+h},2_3],[2],[2],[2].$$ \end{thm}
\begin{proof} The proof is by construction and induction on $k$. Let $C$ be a rational cuspidal curve of degree 5 on $\ensuremath{\mathbb{P}^2}$ with cuspidal configuration ${[2_3],[2],[2],[2]}$. Let $p$ be the cusp with multiplicity sequence $[2_3]$, and let $T$ be the tangent line to $C$ at $p$. There is a smooth point $r \in C$, such that $T \cdot C=4 p+r$. Blowing up at any point $t \in T \setminus \{p,r\}$, we get the curve $C_{1,0}$ of type $(0,5)$ and cuspidal configuration ${[2_3],[2],[2],[2]}$ on $\ensuremath{\mathbb{F}_1}$. Moreover, with $T_{1,0}$ the strict transform of $T$ and $p_{1,0}$, $r_{1,0}$ the strict transforms of the points $p$ and $r$, we have ${T_{1,0} \cdot C_{1,0}=4 p_{1,0}+r_{1,0}}$.
Now assume that the curve $C_{1,k-1}$ of type $(3(k-1),5)$ exists on $\ensuremath{\mathbb{F}_1}$ with cuspidal configuration $[4_{2(k-1)},2_3],[2],[2],[2]$, and the intersection $T_{1,k-1} \cdot C_{1,k-1}=4 p_{1,k-1}+r_{1,k-1}$ for a fiber $T_{1,k-1}$ and points as above. Then blowing up at $r_{1,k-1}$ and contracting $T_{1,k-1}$, we get a curve $C_{0,k}$ on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ of type $(3k+1,5)$ and cuspidal configuration $[4_{2k-1},2_3],[2],[2],[2]$. Moreover, there is a fiber $T_{0,k}$ with the property that $T_{0,k} \cdot C_{0,k}=4 p_{0,k}+r_{0,k}$. Blowing up at $r_{0,k}$ and contracting $T_{0,k}$, we get a rational cuspidal curve $C_{1,k}$ of type $(3k,5)$ on $\ensuremath{\mathbb{F}_1}$ with cuspidal configuration $[4_{2k},2_3],[2],[2],[2]$. \end{proof}
\begin{thm} For $h\in\{0,1\}$ and all integers $k\geq 0$, except the pair $(h,k)=(0,0)$, there exists on the Hirzebruch surface $\ensuremath{\mathbb{F}_h}$ a rational cuspidal curve of type $(2k+2-h,4)$ with four cusps and cuspidal configuration $$[3_{2k-1+h},2],[2_3],[2],[2].$$ \end{thm}
\begin{proof} The proof is by construction and induction on $k$. Let $C$ be the rational cuspidal curve of degree 5 on $\ensuremath{\mathbb{P}^2}$ with cuspidal configuration ${[2_3],[2],[2],[2]}$. Let $q$ be one of the cusps with multiplicity sequence $[2]$, and let $T$ be the tangent line to $C$ at $q$. Then there are smooth points $r,s \in C$, such that ${T \cdot C=3 q+r+s}$. Blowing up at $s$, we get the curve $C_{1,0}$ of type $(1,4)$ and cuspidal configuration $[2_3],[2],[2],[2]$ on $\ensuremath{\mathbb{F}_1}$. Moreover, with $T_{1,0}$ the strict transform of $T$ and $p_{1,0}$, $r_{1,0}$ the strict transforms of the points $p$ and $r$, we have $T_{1,0} \cdot C_{1,0}=3 p_{1,0}+r_{1,0}$.
Now assume that the curve $C_{1,k-1}$ of type $(2k-1,4)$ exists on $\ensuremath{\mathbb{F}_1}$ with cuspidal configuration $[3_{2k-2},2],[2_3],[2],[2]$, and the intersection $T_{1,k-1} \cdot C_{1,k-1}=3 p_{1,k-1}+r_{1,k-1}$ for a fiber $T_{1,k-1}$ and points as above. Then blowing up at $r_{1,k-1}$ and contracting $T_{1,k-1}$, we get a curve $C_{0,k}$ on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ of type $(2k+2,4)$ and cuspidal configuration $[3_{2k-1},2],[2_3],[2],[2]$. Moreover, there is a fiber $T_{0,k}$ with the property that $T_{0,k} \cdot C_{0,k}=3 p_{0,k}+r_{0,k}$. Blowing up at $r_{0,k}$ and contracting $T_{0,k}$, we get a rational cuspidal curve $C_{1,k}$ of type $(2k+1,4)$ on $\ensuremath{\mathbb{F}_1}$ with cuspidal configuration $[3_{2k},2],[2_3],[2],[2]$. \end{proof}
\begin{thm}\label{2erne} For $h\in\{0,1\}$, all integers $k \geq 2$, and every choice of $n_j \in \mathbb{N}$, with $j=1,\ldots,4$, such that $\sum_{j=1}^4 n_j=2k+h$, there exists on the Hirzebruch surface $\ensuremath{\mathbb{F}_h}$ a rational cuspidal curve of type $(k+1-h,3)$ with four cusps and cuspidal configuration $$[2_{n_1}],[2_{n_2}],[2_{n_3}],[2_{n_4}].$$ \end{thm}
\begin{proof} We prove the existence of the curves on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ by induction on $k$. In the proof we show that any curve on $\ensuremath{\mathbb{F}_1}$ can be reached from a curve on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ by an elementary transformation, hence we prove the theorem for $h\in \{0,1\}$.
First we observe that a choice of $n_j$ such that the condition $\sum_{j=1}^4 n_j=2k$ means that either all four $n_j$ are odd, two are odd and two are even, or all four are even. We split the proof into these three cases, and prove only the first case completely. The other two can be dealt with in the same way once we have proved the existence of a first curve.
We now prove the theorem when all $n_j$ are odd. Let $C$ be a rational cuspidal curve on $\ensuremath{\mathbb{P}^2}$ of degree 4 with three cusps and cuspidal configuration ${[2],[2],[2]}$ for cusps $p_j$, $j=1,2,3$. Let $p_4$ be a general smooth point on $C$ and let $T$ be the tangent line to $C$ at $p_4$. Then ${T \cdot C = 2 p_4+t_1+t_2}$, where $t_1,t_2$ are two smooth points on $C$. Blowing up at $t_1$ and $t_2$ and contracting $T$, we get a rational cuspidal curve on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ of type $(3,3)$ with four ordinary cusps.
Fixing notation, we say that we have a curve $C_2$ on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ of type $(3,3)$ and four cusps $p^2_j$, $j=1,\ldots,4$, all with multiplicity sequence $[2]$. Since the choice of $p_4 \in \ensuremath{\mathbb{P}^2}$ was general, there are four $(1,0)$-curves $L^2_j$ such that $$L^2_j \cdot C_2=2 p^2_j+r^2_j,$$ for smooth points $r^2_j \in C_2$. Now assume that we have a curve $C_{k-1}$ on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ of type $((k-1)+1,3)$, with cuspidal configuration ${[2_{n_1-2}],[2_{n_2}],[2_{n_3}],[2_{n_4}]}$ such that all $n_j$ are odd, and such that there exist fibers $L^{k-1}_j$ with $$L^{k-1}_j \cdot C_{k-1}=2 p^{k-1}_j+r^{k-1}_j,$$ for smooth points $r^{k-1}_j$ on $C_{k-1}$.
We blow up at $r^{k-1}_1$, contract the corresponding $L^{k-1}_1$ and get a curve $C_{1,k-1}$ on $\ensuremath{\mathbb{F}_1}$ of type $(k-1,3)$ with cuspidal configuration ${[2_{n_1-1}],[2_{n_2}],[2_{n_3}],[2_{n_4}]}$. Moreover, since $r^{k-1}_1$ was not fiber tangential, we have that $r^{1,k-1}_1 \notin E_1$, and the strict transform of the exceptional fiber of the blowing up, $L^{1,k-1}_1$, has intersection with $C_{1,k-1}$, $$L^{1,k-1}_1 \cdot C_{1,k-1}=2 p^{1,k-1}_1+r^{1,k-1}_1.$$ Blowing up at $r^{1,k-1}_1$ and contracting $L^{1,k-1}_1$ bring us back to $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ and a curve $C_{k}$ of type $(k+1,3)$ and cuspidal configuration $[2_{n_1}],[2_{n_2}],[2_{n_3}],[2_{n_4}]$. This takes care of the case when all $n_j$ are odd.
To prove the theorem when two $n_j$ are even or all $n_j$ are even, we only show that there exist curves on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ of the right type and cuspidal configurations $[2_2],[2_2],[2],[2]$ and ${[2_2],[2_2],[2_2],[2_2]}$. The rest of the argument is then similar to the above. To get the first curve, we blow up $C_2$ in $r^2_1$ and $r^2_2$ and contract $L^2_1$ and $L^2_2$. This is a curve $C_3$ of type $(4,3)$ with cuspidal configuration ${[2_2],[2_2],[2],[2]}$. The curve is on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ since it can be shown with direct calculations in \verb+Maple+ that $r^2_1$ and $r^2_2$ are not on the same $(0,1)$-curve. To get the second curve, we blow up at the analogous $r^3_3$ and $r^3_4$ on the curve $C_3$, before contracting $L^3_3$ and $L^3_4$. We are again on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ by a similar argument to the above, and the curve $C_4$ is of type $(5,3)$ and has cuspidal configuration $[2_2],[2_2],[2_2],[2_2]$. \end{proof}
\noindent Note that the construction of the curves in Theorem \ref{2erne} can also be done from the rational cuspidal cubic on $\ensuremath{\mathbb{P}^2}$. \begin{proof}[Alternative proof of Theorem \ref{2erne}] Let $C$ be the rational cuspidal cubic on $\ensuremath{\mathbb{P}^2}$. Let $s$ be a general point on $\ensuremath{\mathbb{P}^2}$, where general here means that $s$ is neither on $C$, nor the tangent line to the cusp, nor the tangent line to the inflection point on $C$. For example we can choose ${y^2z-x^3}$ as the defining polynomial of $C$, and take ${s=(0:1:1)}$. Then the polar curve of $C$ with respect to the point $s$, given by the defining polynomial ${2yz+y^2}$, intersects $C$ in three smooth points, ${(2^{\frac{2}{3}}:-2:1)}$, ${(2^{-\frac{1}{3}}(-1+\sqrt{3}i):-2:1)}$ and ${(2^{-\frac{1}{3}}(-1-\sqrt{3}i):-2:1)}$. Blowing up at $s$ brings us to $\ensuremath{\mathbb{F}_1}$ and a curve of type $(0,3)$ with one ordinary cusp, say $p_4$. We additionally have three fibers $L_j$, $j=1,\ldots,3$, with the property that ${L_j \cdot C=2 p_j+r_j}$ for smooth points $p_j$ and $r_j$ on $C$. Blowing up at the $r_j$'s and contracting the $L_j$'s, we get the desired series of curves. \end{proof}
The series in Theorem \ref{2erne} can be extended to a series of rational cuspidal curves with less than four cusps in an obvious way. We state this as a corollary. \begin{cor} For $h\in\{0,1\}$, all integers $k \geq 0$, and every choice of $n_j \in \mathbb{N} \cup \{0\}$, with ${j=1,\ldots,4}$, such that ${\sum_{j=1}^4 n_j=2k+h}$, there exists on the Hirzebruch surface $\ensuremath{\mathbb{F}_h}$ a rational cuspidal curve of type $(k+1-h,3)$ with $s \in \{0,1,2,3,4\}$ cusps and cuspidal configuration $$[2_{n_1}],[2_{n_2}],[2_{n_3}],[2_{n_4}].$$ \end{cor}
\begin{proof} These curves can be constructed from the curves in Theorem \ref{2erne} by a similar construction. In order to construct the curves with less than four cusps we have to blow up cusps on the curves in the series from Theorem \ref{2erne}. \end{proof}
Last in this section we provide an example of a curve not represented in any of the above series. This is the only example we have found of such a curve, and in particular the only such curve on $\mathbb{F}_2$. \begin{thm} On $\mathbb{F}_2$ there exists a rational cuspidal curve of type $(0,3)$ with four cusps and cuspidal configuration $$[2],[2],[2],[2].$$ \end{thm}
\begin{proof} Let $C$ be the plane rational fourcuspidal curve of degree $5$. Let $p$ be the cusp $[2_3]$ and $p_i$, $i=1,2,3$, the cusps with multiplicity sequence $[2]$. Let $T$ be the tangent line to $C$ at $p$. Let $L_i$ denote the line through $p$ and $p_i$, with $i=1,2,3$. There are smooth points $s$ and $r_i$, $i=1,2,3$, on $C$, such that $$T \cdot C=4 p + s, \qquad L_i \cdot C=2 p+2 p_i + r_i.$$
Blowing up at $p$ gives a $(2,3)$-curve on $\ensuremath{\mathbb{F}_1}$ with cuspidal configuration ${[2_2],[2],[2],[2]}$. Let $C'$ denote the strict transform of $C$, $T'$ and $L_i'$ the strict transforms of $T$ and $L_i$, and let $E'$ be the special section on $\ensuremath{\mathbb{F}_1}$. Let $p'$ be the cusp $[2_2]$, $p_i'$ the other cusps, and $s'$ and $r_i'$ the strict transforms of the points $s$ and $r_i$. Then we have the following intersections, $$E' \cdot C'=2 p', \qquad T'\cdot C'=2 p'+s', \qquad L_i'\cdot C'=2 p_i' + r_i'.$$ Since $p' \in E'$, blowing up at $p'$ and contracting $T'$, we get a cuspidal curve on $\mathbb{F}_2$ of type $(0,3)$ and cuspidal configuration ${[2],[2],[2],[2]}$.
\end{proof}
\section{Associated results}\label{AR} The main result in this section is that the rigidity conjecture proposed by Flenner and Zaidenberg for plane rational cuspidal curves can \emph{not} be extended to the case of rational cuspidal curves on Hirzebruch surfaces. First in this section we state and prove two lemmas for rational cuspidal curves on Hirzebruch surfaces, the first analogous to a lemma by Flenner and Zaidenberg \cite[Lemma 1.3, p.148]{FlZa94}, and the other a lemma bounding the sum of the so-called $M$-numbers of the curve. Second, we use these lemmas to give an explicit formula for $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ in this case. We calculate this value for the curves constructed in Section \ref{4C}, and with that we provide examples of curves for which $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle}) \neq 0$. Third, we use the two mentioned lemmas and other results to find a lower bound on the highest multiplicity of a cusp on a rational cuspidal curve on a Hirzebruch surface. Last in this section, we investigate real cuspidal curves on Hirzebruch surfaces.
\subsection{Two lemmas} We now state and prove two lemmas for rational cuspidal curves on Hirzebruch surfaces. First, we prove a lemma that is a variant of \cite[Lemma 1.3, p.148]{FlZa94}.
\begin{lem}\label{lemflzadef}\label{lemflzadeffe} Let $C$ be a rational cuspidal curve on $\ensuremath{\mathbb{F}_e}$. Let $(V,D)$ be the minimal embedded resolution of $C$, and let $K_V$ be the canonical divisor on $V$. Moreover, let $D_1, \ldots, D_r$ be the irreducible components of $D$, $\Theta_V$ the tangent sheaf of $V$, $\mathscr{N}_{D/V}$ the normal sheaf of $D$ in $V$, and let $c_2$ be the second Chern class of $V$. Then the following hold. \begin{enumerate} \item[$(0)$] $D$ is a rational tree. \item[$(1)$] $\chi(\Theta_V)=8-2r$. \item[$(2)$] $K_V^2=9-r$. \item[$(3)$] $c_2:=c_2(V)=3+r$. \item[$(4)$] $\displaystyle \chi \left(\bigoplus \mathscr{N}_{D_i/V}\right)=r+\sum_{i=1}^rD_i^2$. \item[$(5)$] $\displaystyle \chi(\ensuremath{\Theta_V \left\langle D \right\rangle})=K_V \ensuremath{\,.\,} (K_V+D)-1.$ \end{enumerate} \end{lem}
\begin{proof} Note that the proof is very similar to the proof of \cite[Lemma 1.3, p.148]{FlZa94}, only small details are changed.
\begin{enumerate} \item[$(0)$] Since $(V,D)$ is the minimal embedded resolution of $C$, $D$ is an SNC-divisor. Since $C$ is a rational curve, $\tilde{C}$ is smooth, and all exceptional divisors are smooth and rational, then all components of $D$ are smooth and rational. The dual graph of $D$, say $\Gamma$, is necessarily a connected graph, and since $C$ is cuspidal, $\Gamma$ will not contain cycles. Thus, $D$ is by definition a rational tree.
\item[$(3)$] Since $V$ is obtained by $r-1$ blowing ups, we have that the Chern class $c_2:=c_2(V)$ is $$c_2(V)=c_2(\ensuremath{\mathbb{F}_e})+r-1.$$ Moreover, by \cite[Corollary V 2.5, p.371]{Hart:1977} we have that $\chi(\mathscr{O}_{\ensuremath{\mathbb{F}_e}})=1$. With $K$ the canonical divisor on $\ensuremath{\mathbb{F}_e}$, we apply the formula in \cite[Remark V 1.6.1, p.363]{Hart:1977}, \begin{align*} 12\chi(\mathscr{O}_{\ensuremath{\mathbb{F}_e}})&=K^2+c_2(\ensuremath{\mathbb{F}_e})\\ 12 &=8 + c_2(\ensuremath{\mathbb{F}_e}).\\ \end{align*} We get $c_2(\ensuremath{\mathbb{F}_e})=4$, hence, \begin{align*} c_2(V)&=4+r-1\\ &=3+r. \end{align*}
\item[$(2)$] We have by \cite[Proposition V 3.4, p.387]{Hart:1977} that $\chi(\mathscr{O}_V)=\chi(\mathscr{O}_{\ensuremath{\mathbb{F}_e}})=1$. By the formula in \cite[Remark V 1.6.1, p.363]{Hart:1977} again, we get \begin{align*} K_V^2&=12\chi(\mathscr{O}_V)-c_2\\ &=12-(3+r)\\ &=9-r. \end{align*}
\item[$(4)$] Since $D_i$ is a rational curve on the surface $V$ for all $i$, we have that $g(D_i)=0$. By \cite[Proof of Proposition II 8.20, p.182]{Hart:1977} we have that $$\mathscr{N}_{D_i/V}\cong \mathscr{L}(D_i) \otimes \ensuremath{\mathscr{O}}_{D_i}.$$ Hence, by the Riemann--Roch theorem for curves \cite[p.362]{Hart:1977}, \begin{align*} \displaystyle \chi \left(\bigoplus \mathscr{N}_{D_i/V}\right)&= \chi \left(\bigoplus \mathscr{L}(D_i) \otimes \ensuremath{\mathscr{O}}_{D_i}\right)\\ &=\sum_{i=1}^r\left(D_i^2+1\right)\\ &=r+\sum_{i=1}^rD_i^2. \end{align*}
\item[$(1)$] By the Hirzebruch--Riemann--Roch theorem for surfaces \cite[Theorem A 4.1, p.432]{Hart:1977}, we have that for any locally free sheaf $\mathscr{E}$ on $V$ of rank $s$ with Chern classes $c_i(\mathscr{E})$, $$\chi(\mathscr{E})=\frac{1}{2}c_1(\mathscr{E}) \ensuremath{\,.\,} \bigl(c_1(\mathscr{E})+c_1(\Theta_V)\bigr)-c_2(\mathscr{E})+s\cdot \chi\left(\ensuremath{\mathscr{O}}_V\right).$$
Moreover, by \cite[Example A 4.1.2, p.433]{Hart:1977}, $\Theta_V$ has rank $s=2$ and ${c_1(\Theta_V)=-K_V}$. We have by the previous results, \begin{align*} \chi(\Theta_V)&=\frac{1}{2}(-K_V) \ensuremath{\,.\,} (-2K_V)-c_2(\Theta_V)+2\chi\left(\ensuremath{\mathscr{O}}_V\right)\\ &=K_V^2-c_2+2\\ &=9-r-(3+r)+2\\ &=8-2r. \end{align*}
\item[$(5)$] Observe first that since $D$ is an SNC-divisor, we have by direct calculation \begin{align*}\displaystyle D^2&=\sum_{i=1}^rD_i^2 + \sum_{i\neq j}D_iD_j\\ &=\sum_{i=1}^rD_i^2+(1+2(r-2)+1)\\ &=\sum_{i=1}^rD_i^2+2r-2. \end{align*}
Since $D$ is an effective divisor, we have by definition that ${p_a(D)=1-\chi(\ensuremath{\mathscr{O}}_D)}$. Since $D$ is a rational tree, by \cite[Lemma 1.2, p.148]{FlZa94}, $p_a(D)=0$. So by the adjunction formula \cite[Exericise V 1.3, p.366]{Hart:1977}, $$K_V \ensuremath{\,.\,} D=-D^2-2.$$
Using the additivity of $\chi$ on the short exact sequence (see \cite[pp.147,162]{FlZa94}), $$0 \longrightarrow \ensuremath{\Theta_V \left\langle D \right\rangle} \longrightarrow \Theta_V \longrightarrow \bigoplus \mathscr{N}_{D_i/V} \longrightarrow 0,$$ and the above results and remarks, we get
\begin{align*} \chi(\ensuremath{\Theta_V \left\langle D \right\rangle})&=\chi(\Theta_V)-\chi \left(\bigoplus \mathscr{N}_{D_i/V}\right)\\ &=(8-2r)-\Bigl(r+\sum_{i=1}^rD_i^2 \Bigr)\\ &=8-2r-(r+D^2-2r+2)\\ &=6-r-D^2\\ &=K_V^2-D^2-3\\ &=K_V^2+2K_V\ensuremath{\,.\,} D-2(-D^2-2)-D^2-3\\ &=(K_V+D)^2+1\\ &=K_V \ensuremath{\,.\,} (K_V+D)-1. \end{align*}
\end{enumerate} \end{proof}
The second lemma bounds the sum of the so-called $M$-numbers by the type of the curve, and this work is inspired by Orevkov (see \cite{Ore}).
For a cusp $p$, the associated \emph{$M$-number} can be defined as \begin{equation*} {M} := \eta+\omega-1, \end{equation*} where \begin{equation*} \eta := \sum_{i=0}^{t-1}\left(m_i-1\right), \end{equation*} and $\omega$ is the number of blowing ups in the minimal embedded resolution which center is an intersection point of the strict transforms of two exceptional curves of the resolution, i.e., an \emph{inner} blowing up.
Moreover, the $M$-number can be expressed in terms of the multiplicity sequence $\overline{m}$, \begin{equation*} {M} = \sum_{i=0}^{t-1}\left(m_i-1\right)+\sum_{i=1}^{t-1}\left(\left\lceil\frac{m_{i-1}}{m_i}\right\rceil-1\right)-1, \end{equation*} where $\lceil a \rceil$ is the smallest integer $\geq a$. See \cite[Definition 1.5.23, p.44]{Fent} and \cite[p.659]{Ore} for more details.
Before we state and prove the lemma, recall that $\ol{\kappa}(\Fe \setminus C)$ denotes the logarithmic Kodaira dimension of the complement to $C$ in $\ensuremath{\mathbb{F}_e}$ (see \cite{Iitaka}).
\begin{lem}\label{kod0M} For a rational cuspidal curve $C$ of type $(a,b)$ on $\ensuremath{\mathbb{F}_e}$ with $s$ cusps and $\ol{\kappa}(\Fe \setminus C)\geq 0$, we have $$\sum_{j=1}^s M_j \leq 2(a+b)+be.$$ \end{lem}
\begin{proof} Let $(V,D)$ and ${\sigma=\sigma_1 \circ \ldots \circ \sigma_t}$ be the minimal embedded resolution of $C$. Write ${\sigma^*(C)=\tilde{C}+\sum_{i=1}^t m_{i-1}E_i}$, with $\tilde{C}$ the strict transform of $C$ under $\sigma$, $m_{i-1}$ the multiplicity of the center of $\sigma_i$ and $E_i$ the exceptional curve of $\sigma_i$. Then by induction and \cite[Proposition V 3.2, p.387]{Hart:1977} we find that \begin{align*} \tilde{C}^2 &=\Bigl(\sigma^*(C)-\sum_{i=1}^t m_{i-1}E_i\Bigr)^2 \notag\\
&=C^2-\sum_{i=0}^{t-1}m_i^2 \notag \\ &=b^2e+2ab-\sum_{i=0}^{t-1}m_i^2. \end{align*}
\noindent By the genus formula, we may rewrite this, \begin{align*} \tilde{C}^2&=b^2e+2ab-\sum_{i=0}^{t-1}m_i^2\\ &=be+2(a+b)-2-\sum_{i=0}^{t-1} m_i. \end{align*}
\noindent Moreover, we have that for ${D=\tilde{C}+\sum_{i=1}^t E_i'}$, with $E_i'$ the strict transform of $E_i$ under the composition $\sigma_{i+1} \circ \cdots \circ \sigma_t$, \begin{align*} D^2&=\tilde{C}^2+2\tilde{C} \ensuremath{\,.\,} \Bigl(\sum_{i=1}^t E_i'\Bigr)+\Bigl(\sum_{i=1}^t E_i'\Bigr)^2 \notag\\ &=\tilde{C}^2+2s+\Bigl(\sum_{i=1}^t E_i'\Bigr)^2. \end{align*}
\noindent Now we split the latter term in this sum into the sum of the strict transforms of the exceptional divisors for each cusp, $$\sum_{i=1}^t E_i'=\sum_{j=1}^s E_{p_j},$$ where $s$ denotes the number of cusps. By \cite[Lemma 2, p.235]{MatsuokaSakai}, we have $$\omega_j=-E_{p_j}^2-1.$$
\noindent Combining the above results, we get \begin{align*} D^2&=be+2(a+b)-2- \sum_{i=0}^{t-1}m_i+2s-\sum_{j=1}^s (\omega_j+1)\\ &=be+2(a+b)-2-\sum_{i=0}^{t-1} m_i-\sum_{j=1}^s (\omega_j-1). \end{align*}
\noindent By the proof of Lemma \ref{lemflzadef}, we have $$6-r-D^2=(K_V+D)^2+1.$$ Note that $r$ denotes the number of components of the divisor $D$. This number is equal to the total number of blowing ups needed to resolve the singularities, plus one component from the strict transform of the curve itself. Following the notation established, we have $r=t+1$. Moreover, by assumption, $\ol{\kappa}(\Fe \setminus C) \geq 0$. By the logarithmic Bogomolov--Miyaoka--Yau-inequality (B--M--Y-inequality) in the form given by Orevkov \cite[Theorem 2.1, p.660]{Ore} and the topological Euler characteristic of the complement to the curve (see \cite{MOEONC}), we then have that $$(K_V+D)^2 \leq 6.$$
\noindent So we get \begin{align*} 0&\leq 1+r+D^2\\ & \leq 1+r+be+2(a+b)-2-\sum_{i=0}^{t-1} m_i-\sum_{j=1}^s (\omega_j-1)\\ & \leq -1+1+be+2(a+b)-\sum_{i=0}^{t-1}(m_i-1)-\sum_{j=1}^s (\omega_j-1)\\ &\leq be+2(a+b)-\sum_{j=1}^s M_j. \end{align*}
\noindent Hence, $$\sum_{j=1}^s M_j \leq 2(a+b)+be.$$ \end{proof}
\subsection{An expression for $\chi\left(\ensuremath{\Theta_V \left\langle D \right\rangle}\right)$} In this section we give a formula for $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ for curves $C$ on $\ensuremath{\mathbb{F}_e}$. Complements to rational cuspidal curves with three or more cusps on Hirzebruch surfaces can be shown to have $\ol{\kappa}(\Fe \setminus C)=2$ (see \cite{MOEONC}), however, these open surfaces are no longer $\ensuremath{\mathbb{Q}}$-acyclic, so we do not expect the rigidity conjecture of Flenner and Zaidenberg to hold in this case. Indeed, we calculate the value of $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ for the curves provided in Section \ref{4C}, and observe that for these curves we do not necessarily have $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})=0$.
\begin{thm}\label{chithetafe} For an irreducible rational cuspidal curve $C$ of type $(a,b)$ on $\ensuremath{\mathbb{F}_e}$ with $s$ cusps $p_j$ with respective $M$-numbers $M_j$, we have $$\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})=7-2a-2b-be+\sum_{j=1}^s M_j.$$ \end{thm} \begin{proof} By \cite[Proposition 2.4, p.445]{FlZa94}, $$K_V \ensuremath{\,.\,} (K_V+D)=K_{\ensuremath{\mathbb{F}_e}} \ensuremath{\,.\,} (K_{\ensuremath{\mathbb{F}_e}}+C)+\sum_{j=1}^s M_j.$$
\noindent By Lemma \ref{lemflzadeffe}, we then get \begin{align*} \chi(\ensuremath{\Theta_V \left\langle D \right\rangle})&= K_V \ensuremath{\,.\,} (K_V+D)-1\\ &= K_{\ensuremath{\mathbb{F}_e}} \ensuremath{\,.\,} (K_{\ensuremath{\mathbb{F}_e}}+C)+\sum_{j=1}^s M_j-1\\ &= ((e-2)L-2M) \ensuremath{\,.\,} \Bigl((a+e-2)L+(b-2)M\Bigr)-1 + \sum_{j=1}^s M_j\\ &= 7-2a-2b-be+\sum_{j=1}^s M_j. \end{align*} \end{proof}
With the above result in mind, we investigate $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ further. Let $C$ be a rational cuspidal curve of type $(a,b)$ on $\ensuremath{\mathbb{F}_e}$, and let $(V,D)$ be as before. By the above, we have that \begin{align*} \chi(\ensuremath{\Theta_V \left\langle D \right\rangle}) &:= \mathrm{h}^0(V,\ensuremath{\Theta_V \left\langle D \right\rangle})-\mathrm{h}^1(V,\ensuremath{\Theta_V \left\langle D \right\rangle})+h^2(V,\ensuremath{\Theta_V \left\langle D \right\rangle})\\ &\, = K_V \ensuremath{\,.\,} (K_V+D)-1\\ &\, = 7-2(a+b)-be+\sum_{j=1}^sM_j. \end{align*}
Moreover, when $\ol{\kappa}(\Fe \setminus C) \geq 0$, we see from Lemma \ref{kod0M} that $$\chi(\ensuremath{\Theta_V \left\langle D \right\rangle}) \leq 7.$$
If $\ol{\kappa}(\Fe \setminus C)=2$, then it follows from a result by Iitaka in \cite[Theorem 6]{Iit} that $\mathrm{h}^0(V,\ensuremath{\Theta_V \left\langle D \right\rangle})=0$. Then we have $$\chi(\ensuremath{\Theta_V \left\langle D \right\rangle}) = h^2(V,\ensuremath{\Theta_V \left\langle D \right\rangle}) -\mathrm{h}^1(V,\ensuremath{\Theta_V \left\langle D \right\rangle}).$$
In \cite[Lemma 4.1, p.219]{Tono05} Tono shows that if first, $\ol{\kappa}(\Fe \setminus C)=2$, and second, the pair $(V,D)$ is \emph{almost minimal} (see \cite{Miyanishi}), then $K_V \ensuremath{\,.\,} (K_V+D) \geq 0$. For plane curves, this result by Tono and a result by Wakabayashi in \cite{Wak} implies that for rational cuspidal curves with three or more cusps we have that $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle}) \geq 0$. Similarly, a rational cuspidal curve on a Hirzebruch surface that fulfills the two prerequisites has $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle}) \geq -1$. While a smiliar result to Wakabayashi's result ensures that three or more cusps implies $\ol{\kappa}(\Fe \setminus C)=2$ \cite{MOEONC}, rationality, however, is no longer a guarantee for almost minimality (see \cite[p.98]{MOEPHD}). Therefore, for rational cuspidal curves with three or more cusps on a Hirzebruch surface, $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ is not necessarily bounded below.
\hide{If $\ol{\kappa}(\Fe \setminus C)=2$ \emph{and} $(V,D)$ is almost minimal, we can apply a lemma by Tono \cite[Lemma 4.1, p.219]{Tono05}. In this case $K_V \ensuremath{\,.\,} (K_V+D) \geq 0$, hence $$-1 \leq \chi(\ensuremath{\Theta_V \left\langle D \right\rangle}) \leq 7.$$
Going back to the proof of Theorem \ref{ONCfe}, we see that for curves of genus $g$ on $\ensuremath{\mathbb{F}_e}$, we have that $n <2g+2$. As before, $n$ is the number of exceptional curves not in $D$ that will be contracted by the minimalization morphism.
For rational cuspidal curves, we see that we have $n=\{0,1\}$. This means that we, in contrast to the situation on $\ensuremath{\mathbb{P}^2}$, are not directly in the situation that the resolution of a rational curve gives an almost minimal pair $(V,D)$. Therefore, $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ is not necessarily bounded below in this case.}
For rational cuspidal curves with four cusps on Hirzebruch surfaces $\ensuremath{\mathbb{F}_e}$ and $\ensuremath{\mathbb{F}_h}$, where $e \geq 0$ and $h \in \{0,1\}$, the values of $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ is given in Table \ref{tab:fourcuspchitheta}.
\begin{table}[H]
\renewcommand\thesubtable{}
\setlength{\extrarowheight}{2pt} \centering
{\noindent \begin{tabular*}{1\linewidth}{@{}c@{\extracolsep{\fill}}l@{}c@{}c}
\toprule
{\textbf{Type}} & {\textbf{Cuspidal configuration}} & $\mathbf{\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})}$ & {\textbf{Surface}}\\ \toprule
{$(2k+1,4)$} & $[4_{k-1+e},2_3],[2],[2],[2]$& $1-k-e$ & $\ensuremath{\mathbb{F}_e}$\\
{$(3k+1-h,5)$} & $[4_{2k-1+h},2_3],[2],[2],[2]$& $-1$ & $\ensuremath{\mathbb{F}_h}$ \\
{$(2k+2-h,4)$} & $[3_{2k-1+h},2],[2_3],[2],[2]$ & $0$ &$\ensuremath{\mathbb{F}_h}$ \\
{$(k+1-h,3)$} & $[2_{n_1}],[2_{n_2}],[2_{n_3}],[2_{n_4}]$& $-1$ & $\ensuremath{\mathbb{F}_h}$\\
{$(0,3)$} & $[2],[2],[2],[2]$& $-1$ & $\mathbb{F}_2$\\
\bottomrule
\end{tabular*}}
\caption[$\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ for rational cuspidal curves with four cusps on $\ensuremath{\mathbb{F}_e}$.]{$\chi(\ensuremath{\Theta_V \left\langle D \right\rangle})$ for rational cuspidal curves with four cusps on $\ensuremath{\mathbb{F}_e}$ and $\ensuremath{\mathbb{F}_h}$. For the three first series, $k \geq 0$. For the fourth series, $k \geq 2$ and $\sum_{j=1}^4 n_j=2k+h$.}
\label{tab:fourcuspchitheta}
\end{table}
An important observation from this list is the fact that $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle}) \leq 0$ for all these curves. We reformulate this observation in a conjecture (cf. \cite{Bobins, FlZa94}).
\begin{conj} Let $C$ be a rational cuspidal curve with four or more cusps on a Hirzebruch surface $\ensuremath{\mathbb{F}_e}$. Then $\chi(\ensuremath{\Theta_V \left\langle D \right\rangle}) \leq 0$. \end{conj}
\subsection{On the multiplicity} In the following we establish a result on the multiplicities of the cusps on a rational cuspidal curve on a Hirzebruch surface. Note that this work is inspired by Orevkov (see \cite{Ore}).
Assume that $C$ is a rational cuspidal curve on a Hirzebruch surface $\ensuremath{\mathbb{F}_e}$. Let $p_1,\ldots,p_s$ denote the cusps of $C$, and $m_{p_1},\ldots,m_{p_s}$ their multiplicities. Renumber the cusps such that $m_{p_1}\geq m_{p_2} \geq \ldots \geq m_{p_s}$. Then for curves with $\ol{\kappa}(\Fe \setminus C) \geq 0$ we are able to establish a lower bound on $m_{p_1}$.
\begin{thm} A rational cuspidal curve $C$ of type $(a,b)$ on $\ensuremath{\mathbb{F}_e}$, with $\ol{\kappa}(\Fe \setminus C) \geq 0$ and $s$ cusps, must have at least one cusp $p_1$ with multiplicity $m:= m_{p_1}$ that satisfies the below inequality, $$m > \frac{3}{2}+a+b-\frac{1}{2}\sqrt{1+20(a+b)+4(a^2+b^2)+4be(1-b)}.$$ \end{thm}
\begin{proof} Using Lemma \ref{kod0M} and a lemma by Orevkov \cite[Lemma 4.1 and Corollary 4.2, pp.663--664]{Ore}, we get \begin{align*} 2(a+b)+be&\geq \sum_{j=1}^s M_j\\ & \geq M_{1}+\sum_{j=2}^sM_{j}\\ &> \frac{\mu_{1}}{m}+m-3+\sum_{j=2}^s M_{j}\\ &=\frac{(b-1)(2a-2+be)}{m}+m-3+\sum_{j=2}^s\Bigr(M_{j}-\frac{\mu_{j}}{m}\Bigr)\\ &\geq \frac{2ab-2(a+b)+2+b^2e-be}{m}+m-3+\sum_{j=2}^s\Bigr(M_{j}-\frac{\mu_{j}}{m_{j}}\Bigr)\\ &\geq \frac{2ab-2(a+b)+2+b^2e-be}{m}+m-3. \end{align*} \noindent This means that $$0 > \frac{2ab-2(a+b)+2+b^2e-be}{m}+m-3-2(a+b)-be.$$
\noindent Let $$g(a,b,m)= \frac{2ab-2(a+b)+2+b^2e-be}{m}+m-3-2(a+b)-be.$$ Factoring $g$, we have that $g<0$ for $$m > \frac{3}{2}+a+b-\frac{1}{2}\sqrt{1+20(a+b)+4(a^2+b^2)+4be(1-b)}.$$ \end{proof}
\begin{cor} A rational cuspidal curve $C$ of type $(a,b)$ with two or more cusps on a Hirzebruch surface $\ensuremath{\mathbb{F}_e}$ must have at least one cusp $p_1$ with multiplicity $m$ that satisfies the below inequality, $$m > \frac{3}{2}+a+b-\frac{1}{2}\sqrt{1+20(a+b)+4(a^2+b^2)+4be(1-b)}.$$ \end{cor}
\begin{proof} The corollary follows directly from the result in \cite{MOEONC} on the logarithmic Kodaira dimension of the complement to a curve with two cusps. \end{proof}
\begin{rem} Note that this theorem will exclude some potential curves. For example, a rational cuspidal curve of type $(a,4)$ on $\mathbb{F}_1$ with two or more cusps (see \cite{MOEONC}) must have at least one cusp with multiplicity $m=3$ for any $a\geq6$. We also have that any rational cuspidal curve of type $(a,5)$ on $\mathbb{F}_1$ with two or more cusps must have at least one cusp with multiplicity $m=3$ for any $a\geq3$. Similarly, any rational cuspidal curve of type $(a,b)$ on $\mathbb{F}_1$ with two or more cusps and $b\geq 6$ must have at least one cusp with multiplicity $m=3$. \end{rem}
\subsection{Real cuspidal curves} It is well known that the known plane rational cuspidal curves with three cusps can be defined over $\ensuremath{\mathbb{R}}$ \cite{FlZa95,FlZa97,Fen99a}. That is not the case for the plane rational cuspidal quintic curve with cuspidal configuration ${[2_3],[2],[2],[2]}$ (see \cite{MOEPHD}).
On the Hirzebruch surfaces, the question whether all cusps on real cuspidal curves can have real coordinates is still hard to answer. Recall that we call $C=\ensuremath{\mathscr{V}}(F)$ a real curve if the polynomial $F$ has real coefficients. However, all known curves on $\ensuremath{\mathbb{F}_e}$ can be constructed from curves on $\ensuremath{\mathbb{P}^2}$. Since the birational transformations can be given as real transformations, if it is possible to arrange the curve on $\ensuremath{\mathbb{P}^2}$ such that the preimages of the cusps have real coordinates, then the cusps will have real coordinates on the curve on the Hirzebruch surface as well. Note the possibility that this arrangement is not always attainable.
Considering the rational cuspidal curves on $\ensuremath{\mathbb{F}_e}$ with four cusps, we see that most of them are constructed from the plane rational cuspidal quintic with cuspidal configuration ${[2_3],[2],[2],[2]}$. Hence, we expect that the cusps on these curves can not all have real coordinates when the curve is real. Contrary to this intuition, however, there are examples of fourcuspidal curves with this property.
\begin{prp} The series of rational cuspidal curves of type ${(k+1-h,3)}$, $k \geq 2$, with four cusps and cuspidal configuration ${[2_{n_1}],[2_{n_2}],[2_{n_3}],[2_{n_4}]}$, where the indices satisfy $\sum_{j=1}^4n_j=2k+h$, on the Hirzebruch surfaces $\ensuremath{\mathbb{F}_h}$, $h \in \{0,1\}$, has the property that all cusps can be given real coordinates on a real curve. \end{prp}
\begin{proof} We have seen that the series of curves can be constructed using the plane rational cuspidal quartic $C$ with three cusps. Let $$y^2z^2+x^2z^2+x^2y^2-2xyz(x+y+z)$$ be a real defining polynomial of $C$. Then it is possible to find a tangent line to $C$ that intersects $C$ in three real points. For example, choose the line $T$ defined by $$\frac{2048}{125}x+\frac{2048}{27}y-\frac{1048576}{3375}z=0.$$ This line is tangent to $C$ at the point ${(\frac{64}{9}: \frac{64}{25}: 1)}$, and it intersects $C$ transversally at the points ${(16: \frac{16}{25}: 1)}$ and ${(\frac{4}{9}: 4: 1)}$. With this configuration, there exists a birational transformation from $\ensuremath{\mathbb{P}^2}$ to $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ that preserves the real coordinates of the cusps on $C$ and constructs a fourth cusp with real coordinates on the strict transform of $C$ on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$. We blow up the two real points at the transversal intersections and contract the tangent line $T$, using the birational map from $\ensuremath{\mathbb{P}^2}$ to $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ (see \cite{MOEPHD}). In coordinates, the map is given by a composition of a (real) change of coordinates on $\ensuremath{\mathbb{P}^2}$ and the map $\phi$, $$\begin{array}{cccc} \phi: &\ensuremath{\mathbb{P}^2} &\dashrightarrow &\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1} \\ &(x:y:z) &\mapsto& (x:y;z:x), \end{array}$$ with inverse $$\begin{array}{cccc} \phi^{-1}:& \ensuremath{\mathbb{P}^1 \times \mathbb{P}^1} &\dashrightarrow &\ensuremath{\mathbb{P}^2} \\ &(x_0:x_1;y_0:y_1) &\mapsto &(x_0y_1:x_1y_1:x_0y_0). \end{array}$$ The strict transform of $C$ is a real curve $C'$ on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ of type $(3,3)$ and cuspidal configuration ${[2],[2],[2],[2]}$, and all the cusps have real coordinates.
On $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$, since the cusps $p_j$ have real coordinates, a fiber, say $L_j$, intersecting $C'$ at a cusp is real. Using the defining polynomial of $L_j$ to substitute one of the variables $x_0$ or $x_1$ in the defining polynomial of $C$ and removing the factor of $x_i^3$, we are left with a polynomial with real coefficients in $y_0$ and $y_1$ of degree $3$. This polynomial has a double real root, and one simple, hence real, root. The double root corresponds to the $y$-coordinates of the cusp $p_j$, and the simple root to the $y$-coordinates of a smooth intersection point $r_j$ of $C$ and $L_j$. Successively blowing up at any $r_j$ and contracting the corresponding $L_j$ lead to the desired series of curves. Since the points we blow up have real coordinates, the transformations preserve the real coordinates of the cusps. Hence, all the curves in the series can have four cusps with real coordinates. \end{proof}
An image of a real rational cuspidal curve of type $(3,3)$ with four ordinary cusps on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ is given in Figure \ref{4realcusps}. In the figure, the surface $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ is embedded in $\mathbb{P}^3$ using the Segre embedding, and we have chosen a suitable affine covering of $\mathbb{P}^3$. The image is created in cooperation with Georg Muntingh using \verb+surfex+ \cite{Surfex}.
\begin{center} \begin{figure}\label{4realcusps}
\end{figure} \end{center}
\appendix \section{Construction of curves}\label{Appendix:A} \noindent We show by two examples the explicit construction of some of the rational cuspidal curves in Theorem \ref{thm:4cusp}. This is done with the computer program \verb+Maple+ \cite{Maple} and the package \verb+algcurves+. See \cite{MOEPHD} for description of the maps.
\subsection{The $(1,4)$-curves of Theorem \ref{thm:4cusp}} \begin{ex} The curves of type $(1,4)$ on $\ensuremath{\mathbb{F}_e}$, $e \geq 1$, can be constructed and checked with the following code. Here we have done this up to $e=2$, but the procedure generalizes to any $e \geq 1$. We first take the defining polynomial of the plane rational cuspidal quintic with four cusps, and verify its cuspidal configuration with the command \verb+singularities+. The \verb+Maple+ output of the command \verb+singularities+ is a list with elements on the form $$[[x,y,z],m,\delta,b],$$ where $[x,y,z]$ denotes the plane projective coordinates, $m$ the multiplicity, $\delta$ the delta invariant, and $b$ the number of branches of the singularity. {\scriptsize \begin{verbatim}
> with(algcurves):
> F := y^4*z-2*x*y^2*z^2+x^2*z^3+2*x^2*y^3-18*x^3*y*z-27*x^5:
> singularities(F, x, y); {[[0, 0, 1], 2, 3, 1], [[RootOf(27*_Z^3+16*z^3), -(9/4)*RootOf(27*_Z^3+16*z^3)^2/z, 1], 2, 1, 1]}
\end{verbatim} }
\noindent We next move the curve such that we can blow up the appropriate point on $\ensuremath{\mathbb{P}^2}$. By inspection of the defining polynomial, we find the tangent line to the curve at the cusp with multiplicity sequence $[2_3]$ to be $\ensuremath{\mathscr{V}}(x)$, and its smooth intersection point with the curve is $(0:1:0)$. We then change coordinates and move $(0:1:0)$ to $(0:0:1)$. {\scriptsize \begin{verbatim}
> z := 1: sort(F); -27*x^5+2*x^2*y^3-18*x^3*y+y^4-2*x*y^2+x^2
> singularities(F*x, x, y); unassign('z'): {[[0, 0, 1], 3, 7, 2], [[0, 1, 0], 2, 1, 2], [[(4/31)*RootOf(27*_Z^3+108*_Z-92)-24/31-(9/31)*RootOf(27*_Z^3+108*_Z-92)^2,...
...-24/31-(27/31)*RootOf(27*_Z^3+108*_Z-92)-(9/31)*RootOf(27*_Z^3+108*_Z-92)^2, 1], 2, 1, 1]}
> x := xa: y := za: z := ya:
\end{verbatim} }
\noindent Now we blow up $(0:0:1)$, take the strict transform and check that we have the $(1,4)$-curve on $\ensuremath{\mathbb{F}_1}$ by finding its singularities in all four affine coverings. Note that \verb+Maple+ provides false singularities, since the \verb+algcurve+ package considers curves as objects on $\ensuremath{\mathbb{P}^2}$. The existing singularities have coordinates on the form $[\cdot,\cdot,1]$.
{\scriptsize \begin{verbatim}
> xa := x0*y1: ya := x1*y1: za := y0:
> factor(F); -y1*(-y0^4*x1+2*y1^2*x0*y0^2*x1^2-y1^4*x0^2*x1^3-2*y1*x0^2*y0^3+18*y1^3*x0^3*y0*x1+27*y1^4*x0^5)
> F := -y0^4*x1+2*y1^2*x0*y0^2*x1^2-y1^4*x0^2*x1^3-2*y1*x0^2*y0^3+18*y1^3*x0^3*y0*x1+27*y1^4*x0^5:
> x0 := 1: y0 := 1: singularities(F, x1, y1); unassign('x0', 'y0'): {[[0, 1, 0], 3, 3, 3], [[1, 0, 0], 4, 9, 1], [[RootOf(16*_Z^3+27), -(4/9)*RootOf(16*_Z^3+27), 1], 2, 1, 1]}
> x0 := 1: y1 := 1: singularities(F, x1, y0); unassign('x0', 'y1'): {[[1, 0, 0], 2, 3, 1], [[RootOf(16*_Z^3+27), (4/3)*RootOf(16*_Z^3+27)^2, 1], 2, 1, 1]}
> x1 := 1: y1 := 1: singularities(F, x0, y0); unassign('x1', 'y1'): {[[0, 0, 1], 2, 3, 1], [[RootOf(27*_Z^3+16), -(9/4)*RootOf(27*_Z^3+16)^2, 1], 2, 1, 1]}
> x1 := 1: y0 := 1: singularities(F, x0, y1); unassign('x1', 'y0'): {[[0, 1, 0], 5, 13, 4], [[1, 0, 0], 4, 12, 4], [[RootOf(27*_Z^3+16), (3/4)*RootOf(27*_Z^3+16), 1], 2, 1, 1]}
\end{verbatim} }
\noindent The curve on $\ensuremath{\mathbb{F}_1}$ is positioned in such a way that we can apply the Hirzebruch one up transformation repeatedly, and get the $(1,4)$-curve on $\ensuremath{\mathbb{F}_e}$ for any $e$. After one transformation we have a curve on $\mathbb{F}_2$ with the prescribed singularities. The following code verifies the latter claim. {\scriptsize \begin{verbatim}
> x0 := x0a: x1 := x1a: y0 := y0a: y1 := x0a*y1a:
> F; -y0a^4*x1a+2*x0a^3*y1a^2*y0a^2*x1a^2-x0a^6*y1a^4*x1a^3-2*x0a^3*y1a*y0a^3 +18*x0a^6*y1a^3*y0a*x1a+27*x0a^9*y1a^4
> x0a := 1: y0a := 1: singularities(F, x1a, y1a); unassign('x0a', 'y0a'): {[[0, 1, 0], 3, 3, 3], [[1, 0, 0], 4, 9, 1], [[RootOf(16*_Z^3+27), -(4/9)*RootOf(16*_Z^3+27), 1], 2, 1, 1]}
> x0a := 1: y1a := 1: singularities(F, x1a, y0a); unassign('x0a', 'y1a'): {[[1, 0, 0], 2, 3, 1], [[RootOf(16*_Z^3+27), (4/3)*RootOf(16*_Z^3+27)^2, 1], 2, 1, 1]}
> x1a := 1: y1a := 1: singularities(F, x0a, y0a); unassign('x1a', 'y1a'): {[[0, 0, 1], 4, 9, 1], [[0, 1, 0], 5, 16, 4], [[RootOf(27*_Z^3+16), 4/3, 1], 2, 1, 1]}
> x1a := 1: y0a := 1: singularities(F, x0a, y1a); unassign('x1a', 'y0a'): {[[0, 1, 0], 9, 45, 4], [[1, 0, 0], 4, 18, 4], [[RootOf(27*_Z^3+16), 3/4, 1], 2, 1, 1]}
\end{verbatim} } \end{ex}
\subsection{A $(5,4)$-curve of Theorem \ref{thm:4cusp}} \begin{ex} Instead of constructing the entire series of $(1,4)$ curves, we may take the $(1,4)$-curve on $\ensuremath{\mathbb{F}_1}$ and construct a $(5,4)$-curve on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$. We then have to change coordinates before we apply the Hirzebruch one down transformation. The curve on $\ensuremath{\mathbb{F}_1}$ has a cusp at $(0:1;0,1)$, which we must move, say to $(0:1;1,1)$, before we apply the transformation to $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$. Then we check that we have the $(5,4)$-curve on $\ensuremath{\mathbb{P}^1 \times \mathbb{P}^1}$ with the prescribed singularities. {\scriptsize \begin{verbatim}
> x0 := x0a: x1 := x1a: y0 := y0a-x1a*y1a: y1 := y1a:
> x0a := x0b: x1a := x1b: y0a := x0*y0b: y1a := y1b:
> F; -x1b*x0b^4*y0b^4+4*x0b^3*y0b^3*x1b^2*y1b-6*x0b^2*y0b^2*x1b^3*y1b^2 +4*x0b*y0b*x1b^4*y1b^3-x1b^5*y1b^4+2*y1b^2*x0b^3*x1b^2*y0b^2 -4*y1b^3*x0b^2*x1b^3*y0b+2*y1b^4*x0b*x1b^4+y1b^4*x0b^2*x1b^3 -2*y1b*x0b^5*y0b^3+6*y1b^2*x0b^4*y0b^2*x1b-6*y1b^3*x0b^3*y0b*x1b^2 +18*y1b^3*x0b^4*x1b*y0b-18*y1b^4*x0b^3*x1b^2+27*y1b^4*x0b^5
> x0b := 1: y0b := 1: singularities(F, x1b, y1b); unassign('x0b', 'y0b'): {[[0, 1, 0], 5, 10, 5], [[1, 0, 0], 4, 15, 1], [[RootOf(16*_Z^3+27), 4/9-(16/27)*RootOf(16*_Z^3+27)+(16/81)*RootOf(16*_Z^3+27)^2, 1], 2, 1, 1]}
> x0b := 1: y1b := 1: singularities(F, x1b, y0b); unassign('x0b', 'y1b'): {[[1, 1, 0], 2, 3, 1], [[RootOf(16*_Z^3+27), RootOf(16*_Z^3+27)+(4/3)*RootOf(16*_Z^3+27)^2, 1], 2, 1, 1]}
> x1b := 1: y1b := 1: singularities(F, x0b, y0b); unassign('x1b', 'y1b'): {[[0, 1, 0], 4, 15, 1], [[1, 0, 0], 3, 3, 3], [[RootOf(27*_Z^3+16), -(9/4)*RootOf(27*_Z^3+16)-(27/16)*RootOf(27*_Z^3+16)^2, 1], 2, 1, 1]}
> x1b := 1: y0b := 1: singularities(F, x0b, y1b); unassign('x1b', 'y0b'): {[[0, 0, 1], 4, 9, 1], [[0, 1, 0], 5, 10, 5], [[1, 0, 0], 4, 6, 4], [[RootOf(27*_Z^3+16), 4/9-(1/3)*RootOf(27*_Z^3+16)+RootOf(27*_Z^3+16)^2, 1], 2, 1, 1]}
\end{verbatim} } \end{ex}
\end{document}
|
arXiv
|
{
"id": "1303.4178.tex",
"language_detection_score": 0.6861544847488403,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{On twin prime distribution and associated biases}
\author{Shaon Sahoo} \address{Indian Institute of Technology, Tirupati 517506, India} \email{[email protected]}
\keywords{Twin primes, Arithmetic progression, Chebyshev's bias}
\begin{abstract} We study some new aspects of the twin prime distribution, focusing especially on how the prime pairs are distributed in arithmetic progressions when they are classified according to the residues of their first members. A modified totient function is seen to play a significant role in this study - we analyze this function and use it to construct a new heuristics for the twin prime conjecture. For the twin primes, we also discuss a sieve similar to Eratosthenes' sieve and a formula similar to the Legendre's formula for the prime counting function. We end our work with a discussion on three types of biases in the distribution of twin primes. Where possible, we compare our results with the corresponding results from the distribution of primes. \end{abstract}
\maketitle \tableofcontents
\section{Introduction} \label{sec1}
Important proven results are scarce in the study of twin primes. The statements about the infinitude of twin primes and corresponding distribution are the well known conjectures in this field. Although unproven so far, there are some important breakthroughs towards proving these conjectures. Chen's work \cite{chen73} and more recent work by Zhang \cite{zhang14} may be noted in this regard.
In this work we study some new aspects of the distribution of twin primes; we especially focus on studying the twin primes in arithmetic progressions. Many important results are known when primes are studied in residue classes. For example, Dirichlet's work (1837) shows that, if $(a,q) = 1$, there exists infinite number of primes $p$ so that $p\equiv a~(\text{mod}~q)$. In analogy with these studies, one can also study twin primes after classifying them according to the residues of their first members. In this study of twin primes, a modified totient function ($\phi_2$) is seen to play a role similar to that of the Euler's totient function in the study of primes. We analyze this modified totient function and an associated modified M\"{o}bius function ($\mu_2$).
For the twin primes, we also discuss in this work a sieve similar to the sieve of Eratosthenes and a formula similar to the Legendre's formula for the prime counting function. We also provide a new heuristics for the twin prime conjecture using the modified totient function $\phi_2$.
While studying primes in residue classes, one of the most interesting results that came out is the biases or, as popularly known, the prime number races \cite{granville06}. Chebyshev (1853) first noted that certain residue classes are surprisingly more preferred over others. For example, if $\pi(x;q;a)$ denotes the number of primes $\le x$ in the residue class $a$ (mod $q$), it is seen that $\pi(x;4;3)>\pi(x;4;1)$ in most occasions. A detailed explanation of this bias (Chebyshev's) is given by Rubinstein and Sarnak in their 1994 paper \cite{rubinstein94}. More recently (2016), a different type of bias was reported by Oliver and Soundararajan \cite{oliver16}. They found that
a prime of a certain residue class is more likely to be followed by a prime of another residue class. A conjectural explanation was presented to understand such bias.
In the last part of this paper we discuss similar biases in the distribution of the twin prime pairs. In addition, we also report a third type of bias: if $(\hat{p}_n-\hat{p}_{n-1})$ represents the difference between the first members of two consecutive twin prime pairs, then $(\hat{p}_n-\hat{p}_{n-1}\pm1)$ are more likely to be prime numbers than to be composites. We also find that the possibility of $(\hat{p}_n-\hat{p}_{n-1}-1)$ to be a prime is more than that of $(\hat{p}_n-\hat{p}_{n-1}+1)$ to be a prime. The numerical results presented here are based on the analysis of the first 500 million primes ($\pi(x_0)=5\times 10^8$) and corresponding about 30 million twin prime pairs ($\pi_2(x_0)=29,981,546$). Here $x_0$ is taken as the last prime we consider, i.e. 11,037,271,757.
Besides these studies, we also state a conjecture regarding the arithmetic progression of twin primes in the line with the Green-Tao theorem and give some numerical examples of this type of progression.
\section{Notations and definitions} \label{sec2}
In this paper the letters $i,j,a,b,q,n$ and $N$ represent positive integers. The letters $x,y,z$ and $c$ represent real numbers, and $p$ always represents a prime number with $p_n$ being the $n$-th prime number. Here $\hat{p}$ represents a twin prime pair whose first member is $p$; $\hat{p}_n$ denotes the $n$-th twin prime pair. Unless mentioned otherwise, any statement involving a twin prime pair will be interpreted with respected to the first member of the pair, e.g., $\hat{p} \le x$ means that the first prime of the pair is less than or equal to $x$, or $\hat{p} \equiv a$ (mod $q$) means that the first member of the pair is congruent to $a$ (mod $q$).
We denote gcd of $a$ and $b$ by $(a,b)$. In this paper we denote a reduced residue class of $q$ by a number $a$ if ($a,q$) = 1 and $0<a<q$. A pair of integers are called a {\it coprime pair} to a third integer if both the integers are individually coprimes to the third integer; e.g., 6 and 14 form a coprime pair to 25. A coprime pair to an integer is called a {\it twin coprime pair} if the integers in the pair are separated by 2.
The functions $\mu(n)$ and $\phi(n)$ respectively denote the M\"{o}bius function and Euler's totient function. The functions $\omega(n)$ and $\omega_o(n)$ denote respectively the number of prime divisors and odd prime divisors of $n$ (for odd $n$, $\omega(n) = \omega_o(n)$, and for even $n$, $\omega(n) = \omega_o(n)+1$). \\
\noindent {\it Definitions:}\\
\indent $\phi_2(q):=\#\{a\le q~|(a,q)=1 ~{\rm and}~ (a+2,q)=1\}$.\\ \indent $\mu_2(n):=\mu(n)2^{\omega_o(n)}$.\\ \indent $\pi(x):= \#\{p \le x \}$.\\
\indent $\pi(x;q;a):= \#\{p \le x ~|p\equiv a (\text{mod}~q)\}$.\\
\indent $\pi(x;q;a_i|a_j):= \#\{p_n \le x ~| p_n\equiv a_i (\text{mod}~q)~\text{and}~p_{n+1}\equiv a_j (\text{mod}~q)\}$.\\ \indent $\delta(x;q;a):= \frac{\pi(x;q;a)}{\pi(x)}$.\\
\indent $\delta(x;q;a_i|a_j):= \frac{\pi(x;q;a_i|a_j)}{\pi(x;q;a_i)}$.\\ \indent $\Delta(x;q;a_i,a_j):= \pi(x;q;a_i) - \pi(x;q;a_j)$.\\ \indent $\bar{\Delta}(x;q;a_i,a_j):= \Delta(x;q;a_i,a_j)\frac{ln(x)}{\sqrt{x}}$.\\
\indent $\pi^{\pm}(x):= \#\{p_n \le x~| (p_{n+1}-p_n \pm 1) ~\text{is a prime} \}$\\ \indent $\delta^{\pm}(x):=\frac{\pi^{\pm}(x)}{\pi(x)}$\\
\indent $\pi_2(x):=\#\{\hat{p} \le x\}$.\\
\indent $\pi_2(x;q;a):=\#\{\hat{p} \le x ~| \hat{p}\equiv a (\text{mod}~q)\}$.\\
\indent $\pi_2(x;q;a_i|a_j):= \#\{\hat{p}_n \le x ~| \hat{p}_n\equiv a_i (\text{mod}~q)~\text{and}~\hat{p}_{n+1}\equiv a_j (\text{mod}~q)\}$.\\ \indent $\delta_2(x;q;a):= \frac{\pi_2(x;q;a)}{\pi_2(x)}$.\\
\indent $\delta_2(x;q;a_i|a_j):= \frac{\pi_2(x;q;a_i|a_j)}{\pi_2(x;q;a_i)}$.\\ \indent $\Delta_2(x;q;a_i,a_j):= \pi_2(x;q;a_i) - \pi_2(x;q;a_j)$. \\ \indent $\bar{\Delta}_2(x;q;a_i,a_j):= \Delta_2(x;q;a_i,a_j)\frac{ln^2(x)}{10\sqrt{x}}$.\\
\indent $\pi^{\pm}_2(x):= \#\{\hat{p}_n \le x~| (\hat{p}_{n+1}-\hat{p}_n \pm 1) ~\text{is a prime} \}$\\ \indent $\delta^{\pm}_2(x):=\frac{\pi^{\pm}_2(x)}{\pi_2(x)}$\\
\section{Arithmetic progressions and twin primes} \label{sec3}
Unlike the case of prime numbers where each reduced residue class has infinite number of primes (Dirichlet's Theorem), in case of the twin primes, some reduced residue class may not have more than one prime pair. This is because the existence of the second prime adds extra constraint on the first one. For example, there is no twin prime pair in the residue class $a=1$ (mod 6). For $q=10$, the residue class $a=3$ has only one prime pair, namely 3 and 5. These observations can be concisely presented in the following theorem. \begin{theorem} The reduced residue class $a$ (mod q) can not have more than one twin prime pair when $(a+2,q)>1$. For such $a$, if both $a$ and $a+2$ are prime numbers, then they represent the only twin prime pair in the residue class $a$. \label{th3.1} \end{theorem}
This theorem can simply be proved by noting that for any prime $p\equiv a$ (mod $q$), $p+2$ will be forced to be a composite if $(a+2,q)\neq 1$. Here only exception may happen when the number $a+2$ is itself a prime number. In such a case, if $a$ is also a prime number, then the pair consisting of $a$ and $a+2$ is the only twin prime pair in the class $a$.
It may be mentioned in the passing that some non-reduced residue class may have one twin prime pair. For example, when $q=9$, residue class $a=3$ has one prime pair, namely 3 and 5.
Theorem \ref{th3.1} helps us to come up with a conjecture for twin primes in the line of the Dirichlet's Theorem for prime numbers. For that we first define below an admissible class in the present context of twin prime distribution. \begin{definition} A residue class $a$ (mod $q$) is said to be admissible if $(a,q)=1$ and $(a+2,q)=1$. \label{df3.1} \end{definition}
With this definition of admissible class, we now propose the following conjecture. \begin{conjecture} Each admissible residue class of a given modulus contains infinite number of twin prime pairs. \label{cj3.1} \end{conjecture}
A stronger version of this conjecture can be formulated with the definition of a {\it modified totient function} $\phi_2(q)$. This function gives the number of {\it admissible} residue classes modulo $q$. \begin{definition}
The function $\phi_2(q)$ denotes the number of admissible residue classes modulo $q$, i.e., $\phi_2(q)=\#\{a<q~|(a,q)=1 ~{\rm and}~ (a+2,q)=1\}$. \label{df3.2} \end{definition}
Clearly $\phi_2(q)\le \phi(q)$, where $\phi(q)$ is the Euler's totient function. While calculating $\phi_2(q)$, a reduce residue class $a$ is excluded if there exists a prime divisor ($p$) of $q$ so that $p|a+2$. This observation helps us to deduce that $\phi_2(2^n)=\phi(2^n)=2^{n-1}$ for $n\ge 1$, and for any odd prime $p$, $\phi_2(p^m)=\phi(p)-p^{m-1}=p^m(1-2/p)$ when $m\ge 1$. In Section \ref{sec7} we prove that \begin{equation} \begin{split}
\phi_2(q) &= q \prod_{p|q}(1-\frac 2 p) \text{~~for odd q, and}\\ \nonumber
\phi_2(q) &= q (1-\frac 1 2) \prod_{p>2,~p|q}(1-\frac 2 p) \text{~~for even q}. \end{split} \end{equation} If we take $\phi_2(1) = 1$ then the function $\phi_2(q)$ is multiplicative. Above formula for $\phi_2(q)$ is stated as a theorem in Section \ref{sec4}.
The function $\phi_2(q)$ can also be expressed in terms of the principal Dirichlet character ($\chi_1$) in the following way: $\phi_2(q)=\sum_{i=1}^q\chi_1(i)\chi_1(i+2)=\sum_{k=1}^{\phi(q)}\chi_1(a_k+2)$, where the second sum is over the reduced residues ($a_k$). The values of the function for some integers are as follows: $\phi_2(4)= \phi(4)=2$, $\phi_2(6)= \phi(6)-1=1$, $\phi_2(7)= \phi(7)-1=5$ and $\phi_2(9)= \phi(9)-3=3$.
Using the function $\phi_2(q)$, we now propose the following conjecture regarding the twin prime counting function for a given admissible class. \begin{conjecture} For a given $q$ and an admissible class $a$ (mod $q$), we have \begin{equation} {\displaystyle \lim_{x\rightarrow\infty}} \frac{\pi_2(x)}{\pi_2(x;q;a)} = \phi_2(q). \nonumber \end{equation} \label{cj3.2} \end{conjecture}
We end this section with a conjecture on the arithmetic progression of twin primes in line with the Green-Tao theorem \cite{greentao08}. For some positive integers $a$ and $b$, let the sequence $a+kb$ represents (first members of) twin primes for $k=0, 1, 2, \cdots, l-1$ ($l>2$). This sequence is said to represent an arithmetic progression of twin primes with $l$ terms. \begin{conjecture} For every natural number $l$ ($l>2$), there exist arithmetic progressions of twin primes with $l$ terms. \label{cj3.3} \end{conjecture}
Some examples for the arithmetic progressions of twin primes are as follows: 41+420$\cdot$k, for k = 0, 1, .., 5; 51341+16590$\cdot$k, for k = 0, 1, .., 6; 2823809+570570$\cdot$k, for k = 0, 1, .., 6. Each number in a progression represents the first member of a twin prime.
\section{Functions $\mu_2(n)$ and $\phi_2(n)$} \label{sec4}
We begin this section with the following definition of a {\it modified M\"{o}bius function} $\mu_2(n)$. \begin{definition} If $\omega_o(n)$ denotes the number of odd prime divisors of $n$, then the {\it modified M\"{o}bius function} is defined as \begin{equation} \mu_2(n)=\mu(n)2^{\omega_o(n)},\nonumber \end{equation} where $\mu(n)$ is the M\"{o}bius function. \label{df4.1} \end{definition} The function $\mu_2(n)$ is multiplicative. The values of $\omega_o(n)$ for some $n$ are as follows: $\omega_o(1)=0$, $\omega_o(2)=0$, $\omega_o(3)=1$ and $\omega_o(6)=1$.
The main reason we defined $\mu_2(n)$ is that it is possible to express $\phi_2(n)$ in terms of this modified M\"{o}bius function. The following theorem (\ref{th4.1}) gives two different expressions of $\phi_2(n)$. The proof of the theorem is given in Section \ref{sec7}.
\begin{theorem} The function $\phi_2(n)$, as defined in Definition \ref{df3.2}, takes the following product form \begin{equation}
\phi_2(n) = n (1-\frac{\theta_n}{2}) \prod_{p>2,~p|n}(1-\frac 2 p), \label{eq_ph2} \end{equation} where $\theta_n$ is either 0 or 1 depending on whether $n$ is odd or even respectively. Furthermore, using $\mu_2(n)$ function, one can express $\phi_2(n)$ as the following divisor sum: \begin{equation}
\phi_2(n) = n \sum_{d|n} \frac{\mu_2(d)}{d}. \label{eq_p2m2} \end{equation} \label{th4.1} \end{theorem}
In the following we discuss some properties of the functions $\mu_2(n)$ and $\phi_2(n)$.
\begin{lemma} The divisor sum of the function $\mu_2$ is given by \begin{equation}
\sum_{d|n}\mu_2(d)= \left\{\begin{array}{ll} 0 & if~~ n~is~ even \\ \nonumber (-1)^{\omega(n)} & if~~ n~is~ odd. \end{array}\right. \end{equation} \label{lm4.1} \end{lemma} \begin{proof} The formula is true for $n=1$ since $\mu_2(1)=\mu(1)2^{0} = 1$. Assume now that $n>1$ and is written $n=p_1^{a_1}p_2^{a_2}\cdots p_k^{a_k}$. First we take $n$ to be even with $p_1 = 2$ and $a_1\ge1$. We note that for any odd divisor $d$, there is an even divisor $2d$ so that $\mu_2(d)+\mu_2(2d) = 0$. It may
be noted that we have $\mu_2(2^a d)=0$ for $a>1$. This shows that $\sum_{d|n}\mu_2(d) = 0$ when $n$ is even.
We now take $n$ to be odd. Using Lemma \ref{lm8.1}, we get the following. \begin{equation}
\sum_{d|n}\mu_2(d) = \sum_{d|n} \mu(d) 2^{\omega_o(d)} = \prod_{p|d}(1-2)=(-1)^{\omega(n)}. \nonumber \end{equation} \end{proof}
\begin{lemma}
If $n=\prod_{p|n}p^{a_p}$, then the divisor sum of the function $\phi_2$ is given by \begin{equation}
\sum_{d|n}\phi_2(d) = n \prod_{p>2,~p|n}\frac{\phi_2(p)+p^{-a_p}}{\phi(p)}. \nonumber \end{equation} \label{lm4.2} \end{lemma} \begin{proof} For odd $n$, taking advantage of the multiplicative property of $\phi_2$, we write
$\sum_{d|n}\phi_2(d) = \prod_{p|n}\{1+\phi_2(p)+\phi_2(p^2)+\cdots+\phi_2(p^{a_p})\}$. Now since $\phi_2(p^m)=p^m(1-\frac 2p)$ for odd $p$ and $m\ge1$, we
get $\sum_{d|n}\phi_2(d) = \prod_{p|n}\frac{p^{a_p}(p-2)+1}{p-1} = n \prod_{p|n}\frac{\phi_2(p)+p^{-a_p}}{\phi(p)}$. For even $n$, there will be an additional multiplicative factor to this result; the factor is $(1+\phi_2(2)+\phi_2(2^2)+\cdots+\phi_2(2^{a_2})) = 2^{a_2}$. This factor is a part of $n$. \end{proof}
\begin{theorem} Let $D(\mu_2,s)$ denote the Dirichlet series for $\mu_2$, i.e., $D(\mu_2,s):=\sum_{n=1}^{\infty}\frac{\mu_2(n)}{n^s}$. We have \begin{equation} D(\mu_2,s) = (1-\frac{1}{2^s})\prod_{p>2}(1-\frac{2}{p^s}) {~~ for ~~} Re(s) >1.\nonumber \end{equation} \label{th4.2} \end{theorem} \begin{proof} First we note that $2^{\omega_o(n)}\le 2^{\omega(n)}\le d(n)$, where $d(n)$ counts the number of divisors of $n$. It is known that $\sum_{n=1}^{\infty}\frac{d(n)}{n^s}$ is an absolutely convergent series for $Re(s)>1$ (chapter 11, \cite{apostol76}); this implies that the series $\sum_{n=1}^{\infty}\frac{\mu_2(n)}{n^s}$ also converges absolutely for $Re(s)>1$.
Using Lemma \ref{lm8.2}, we have $D(\mu_2,s)=\prod_p\{1+\frac{\mu_2(p)}{p^s}\} = (1-\frac{1}{2^s})\prod_{p>2}(1-\frac{2}{p^s})$. \end{proof}
\begin{theorem} Let $D(\phi_2,s)$ denote the Dirichlet series for $\phi_2$, i.e., $D(\phi_2,s):=\sum_{n=1}^{\infty}\frac{\phi_2(n)}{n^s}$. We have \begin{equation} D(\phi_2,s)= \zeta(s-1) D(\mu_2,s) \text{~~ for ~~} Re(s) >2,\nonumber \end{equation} where $\zeta(s-1)$ is the Riemann zeta function. \label{th4.3} \end{theorem} \begin{proof} We first note that $\phi_2(n)\le n$. So the series $\sum_{n=1}^{\infty}\frac{\phi_2(n)}{n^s}$ is absolutely convergent for $Re(s)>2$. The theorem now can be simply proved by taking $f(n)=\mu_2(n)$ and $g(n)=n$ in the Theorem \ref{th8.1}. \end{proof}
\noindent {\bf Remark.} We here note that $\frac{1}{D(\mu_2,s)}= (1-\frac{1}{2^s})^{-1}\prod_{p>2}(1-\frac{2}{p^s})^{-1} = (1+\frac{1}{2^s}+\frac{1}{2^{2s}}+\cdots)\prod_{p>2}(1+\frac{2}{p^s}+\frac{2^2}{p^{2s}}+\cdots) = \sum_{n=1}^{\infty}\frac{2^{\Omega_o(n)}}{n^s}$, where $\Omega_o(n)$ counts the number of odd prime divisors of $n$ with multiplicity. Now in analogy with the relation $\left(\sum_{n=1}^{\infty}\frac{\mu(n)}{n^s}\right)^{-1} = \zeta(s)$, one may define $\zeta_2(s):=\sum_{n=1}^{\infty}\frac{2^{\Omega_o(n)}}{n^s}$ so that Theorems \ref{th4.2} and \ref{th4.3} can be written respectively as $D(\mu_2,s) = 1/\zeta_2(s)$ and $D(\phi_2,s)=\zeta(s-1)/\zeta_2(s)$.
\begin{theorem} We have the following weighted average of $\mu_2(n)$: \begin{equation} \sum_{n\le x} \mu_2(n)\left[\frac{x}{n}\right] = L(x) + L(\frac x2) + 2L(\frac x4) + 4L(\frac x8) + \cdots, \nonumber \end{equation} where $L(x) = \sum_{n\le x} (-1)^{\omega(n)}$. It is understood that $L(x)=0$ for $0 < x <1$. \label{th4.4} \end{theorem} \begin{proof} Using Theorem \ref{th8.3} with $f(n) = \mu_2(n)$ and $g(n) = 1$, we get \begin{align*}
\sum_{n\le x} \mu_2(n)\left[\frac{x}{n}\right] = & \sum_{n\le x} \sum_{d|n} \mu_2(d) & \\
= & \sum_{odd~ n \le x} (-1)^{\omega(n)} & \text{(using Lemma \ref{lm4.1})}\\
= & L(x) - \sum_{n \le x/2} (-1)^{\omega(2n)} & \text{(where $L(x) = \sum_{n \le x} (-1)^{\omega(n)}$)}\\ \end{align*} Since $\sum_{n \le x/2} (-1)^{\omega(2n)} = \sum_{odd~ n \le x/2} (-1)^{\omega(2n)}+\sum_{n \le x/4} (-1)^{\omega(4n)}$ and $\omega(2n) = \omega(n) + 1$ for odd $n$, we get $K(x) = L(x) + K(\frac x2) - \sum_{n \le x/4} (-1)^{\omega(4n)}$, where $K(x)=\sum_{n\le x} \mu_2(n)\left[\frac{x}{n}\right]$. Now manipulating the last term repeatedly we get the following series: $K(x) = L(x) + K(\frac x2) + K(\frac x4) + K(\frac x8) + \cdots$. This series can be further manipulated to find $K(x)$ in the following way: $K(x) = L(x) + L(\frac x2) + 2L(\frac x4) + 4L(\frac x8) + \cdots$. \end{proof}
\section{Function $\phi_2(n)$ and twin primes} \label{sec5}
We noted in Conjecture \ref{cj3.2} how $\phi_2(n)$ can be helpful in studying the twin prime distributions. In this section we will present some more observations on how $\phi_2(n)$ can be useful in the study of twin primes. First we mention a simple application of $\phi_2(n)$ in the form of the following theorem. \begin{theorem} For a given $l \ge 1$, consider a set $\mathcal{P}$ of any $l$ distinct primes. There always exists a number $m>1$ so that $(p,m(m+2))=1$ for every prime $p \in \mathcal{P}$. \label{th5.1} \end{theorem} \begin{proof} When $l=1$ and $p\ge 5$, from Theorem \ref{th4.1}, we get $\phi_2(p) \ge 2$. This confirms the existence of at least one $m>1$ so that $(p,m(m+2))=1$. It is also easy to explicitly check that the theorem is correct when $p=2$ or 3.\\ When $l=2$, and two primes are 2 and 3, we can explicitly check that the theorem is valid. For example, $m=5$ works here. When one or both primes are 2 or 3, it is easy to see that $\phi_2(p_1p_2)\ge 2$ for two primes $p_1$ and $p_2$.\\ When $l\ge 3$, we can define $P_l = {\displaystyle \prod_{p \in \mathcal{P}}}p$, and note that $\phi_2(P_l) \ge 2$. Which confirms the existence of at least one $m>1$ so that $(p,m(m+2))=1$ for every prime $p \in \mathcal{P}$. \end{proof}
\begin{corollary} For a given $n>2$, there always exists a number $m>n$ so that $(p,m(m+2))=1$ for every prime $p\le n$. \label{cor5.1} \end{corollary} \begin{proof} This corollary can be shown to be a consequence of Theorem \ref{th5.1}. We will give here a separate direct proof of the corollary. Let $P_n = {\displaystyle \prod_{p\le n}}p$. From Theorem \ref{th4.1}, we have $\phi_2(P_n)={\displaystyle \prod_{odd~p\le n}}(p-2)\ge 1$. We recall that, between 1 and $P_n$, there are $\phi_2(P_n)$ twin coprime pairs to $P_n$. In this case all these pairs (one pair when $2<n<5$) lie above $n$. This proves the theorem. \end{proof}
This theorem tells us that it is always possible to find a twin coprime pair to the primorial ($P_n$) upto a given number ($n>2$). It may be mentioned here that we can have a stronger statement than the Corollary \ref{th5.1} by noting that there are $\phi_2(P_n)$ number of twin coprime pairs in ($n,P_n$). It is also interesting to note that the Theorem \ref{th5.1} implies that there are infinite number of primes.
According to the Hardy-Littlewood conjecture, $\pi_2(x) \sim 2C_2\frac{x}{(ln x)^2}$, where $C_2$ is the so-called twin prime constant; this constant is given by the following product form: $C_2= \prod_{p>2}\frac{p(p-2)}{(p-1)^2}$. In terms of $\phi(n)$ and $\phi_2(n)$, this constant can be written in the following series forms,
\begin{equation} C_2 = \sum_{odd~n=1}^{\infty}\frac{\mu(n)}{\phi^2(n)},~~\text{and} \label{eq_c2} \end{equation} \begin{equation} \frac{1}{C_2} = \sum_{odd~n=1}^{\infty} \frac{\mu^2(n)}{n\phi_2(n)}. \label{eq_inc2} \end{equation} Both the equations can be proved by expressing their right hand sides as Euler products (see also Lemma \ref{lm8.2}; to use it in the present context, leave out the sum over the even numbers from the left side and drop $p=2$ from the right side of the equation). The Equation \ref{eq_c2} also appears in \cite{golomb60}.
Let $P_n$ denote the product of the first $n$ primes, i.e., $P_n=\Pi_{k=1}^{n}p_k$. It is easy to check that \begin{equation} 2C_2 = {\displaystyle \lim_{n\rightarrow\infty}} \frac{P_n \phi_2(P_n)}{\phi^2(P_n)} = {\displaystyle \lim_{n\rightarrow\infty}} \frac{\phi_2(P^2_n)}{\phi^2(P_n)}. \label{eq_c2ro} \end{equation}
The last expression of $C_2$, as appears in Equation \ref{eq_c2ro}, is especially interesting here. It helps us to view the Hardy-Littlewood twin prime conjecture from a different angle. We explain this in the following.
\paragraph{\bf Heuristics for twin prime conjecture.} We take $X = P_x = \prod_{p\le x}p$ for large $x$. The number of integers which are coprime to and smaller than $X$ is $\phi(X)$. It may be mentioned here that all the coprimes, except 1, lie in $[x,X]$. The total number of the coprime pairs of all possible gaps is $\frac 12 \phi(X) ( \phi(X)-1) \sim \frac 12 \phi^2(X)$. Among these pairs, the number of the twin coprime pairs (pairs with gap = 2) is $\phi_2(X)$. All these twin coprime pairs lie in $[x,X]$. So in $[x,X]$, the fraction of the coprime pairs which are the twin coprime pairs is $\frac{2\phi_2(X)}{\phi^2(X)}$. Now the number of primes in $[x,X]$ is $\pi(X)-\pi(x)\sim \pi(X)$. The number of the all possible prime pairs (with any non-zero gap) is $\frac 12 \pi(X)(\pi(X)-1) \sim \frac 12 \pi^2(X)$. Among these prime pairs, the number of the twin prime pairs in $[x,X]$ would then be: $\pi_2(X) -\pi_2(x) \sim \frac 12 \pi^2(X) \times \frac{\phi_2(X)}{\frac 12 \phi^2(X)}$. From the prime number theorem we know $\pi(X) \sim \frac{X}{\ln X}$. So for large $X$, we finally get $\pi_2(X) \sim \frac{X^2}{\ln^2 X} \times \frac{\phi_2(X)}{\phi^2(X)} = \frac{X}{\ln^2 X} \times \frac{X\phi_2(X)}{\phi^2(X)}$. Now $\frac{X\phi_2(X)}{\phi^2(X)}$ converges to $2C_2$ as noted in Equation \ref{eq_c2ro}; this implies: $\pi_2(X) \sim 2C_2 \frac{X}{\ln^2 X}$.
Above argument can also be given in the following ways. We note that $\frac {\pi^2(X)}{\phi^2(X)}$ denotes the fraction of the coprime pairs which are the prime pairs (of any possible gaps). Now there are $\phi_2(X)$ twin coprime pairs; among them, the number of twin prime pairs would then be $\frac {\pi^2(X)}{\phi^2(X)} \times \phi_2(X) \sim 2C_2 \frac{X}{\ln^2 X}$. We also note that $\frac {\pi(X)}{\phi(X)}$ is the probability of a coprime to be a prime. So $\left(\frac {\pi(X)}{\phi(X)}\right)^2$ is the probability that the two members of a coprime pair are both primes. This suggest that the number of the twin prime pairs among $\phi_2(X)$ twin coprime pairs is $\left(\frac {\pi(X)}{\phi(X)}\right)^2 \times \phi_2(X) \sim 2C_2 \frac{X}{\ln^2 X}$.
In the preceding heuristics, the reason we take $X$ to be a large primorial (product of primes upto some large number) is that the density of coprimes vanishes for this special case ($\phi(X)/X \rightarrow 0$ as $X \rightarrow \infty$; see Theorem \ref{th5.2}). This vanishing density is a desired property in our heuristics as we know that the density of primes has this property ($\pi(N)/N \rightarrow 0$ as $N \rightarrow \infty$). If we take any arbitrary form of $X$ then the density of coprimes may not vanish. We may also note here that, as $X \rightarrow \infty$, $\frac{X\phi_2(X)}{\phi^2(X)}$ goes to a limit when the special form of $X$ is taken. On the other hand if $N$ increases arbitrarily then the function $\frac{N\phi_2(N)}{\phi^2(N)}$ does not have any limit.
Next we discuss the estimated values of $\phi(X)/X$ and $\phi_2(X)/X$ when $X$ is the product of primes upto $x$. \begin{theorem}[Mertens' result] As $x\rightarrow \infty$, we have ${\displaystyle \prod_{p\le x}(1-\frac{1}{p}) = \frac{e^{-\gamma}}{\ln x} + O(\frac{1}{\ln^2 x})}$. Here $\gamma \approx 0.577216$ is the Euler-Mascheroni constant. \label{th5.2} \end{theorem} The proof of this result can be found in the most standard text books.
\begin{theorem} As $x\rightarrow \infty$, we have ${\displaystyle (1- \frac 12) \prod_{2<p\le x}(1-\frac{2}{p}) = 2C_2\frac{e^{-2\gamma}}{\ln^2 x} + O(\frac{1}{\ln^3 x})}$. Here $C_2$ is the twin prime constant and $\gamma$ is the Euler-Mascheroni constant. \label{th5.3} \end{theorem} \begin{proof} It is possible to prove Theorem \ref{th5.3} directly using the result from Theorem \ref{th5.2}. It can be done in the following way.\\ ${\displaystyle (1- \frac 12) \prod_{2<p\le x}(1-\frac{2}{p}) = \left(2\prod_{2<p\le x} \frac{1-\frac 2p}{(1- \frac 1p)^2}\right) \prod_{p\le x}(1- \frac 1p)^2}$. \\ Now as $x\rightarrow \infty$, ${\displaystyle 2\prod_{2<p\le x} \frac{1-\frac 2p}{(1- \frac 1p)^2} = 2C_2}$ and the last part can be estimated from Mertens' second result, ${\displaystyle \prod_{p\le x}(1- \frac 1p)^2 = \frac{e^{-2\gamma}}{\ln^2 x} + O(\frac{1}{\ln^3 x})}$. This finally gives us\\ ${\displaystyle (1- \frac 12) \prod_{2<p\le x}(1-\frac{2}{p}) = 2C_2\frac{e^{-2\gamma}}{\ln^2 x} + O(\frac{1}{\ln^3 x})}$. \end{proof}
The function $\phi_2(n)$ also gives a non-trivial upper bound for $\pi_2(n)$. We make a formal statement in terms of the following theorem. \begin{theorem} For $n>2$, $\pi_2(n)\le \phi_2(n)+\omega(n)$. \label{th5.4} \end{theorem} By taking $n$ to be a large primorial (product of primes upto about $\ln n$), we see that for $n \rightarrow \infty$, using Theorem \ref{th5.3}, $\pi_2(n) \le \phi_2(n)+\omega(n) \sim \frac {2C_2 n e^{-2\gamma}}{(\ln \ln n)^2}$. This upper bound, of-course, is not very tight; a much better bound can be found directly from $\pi(n)$. We note that, except 5, no other prime number is part of two twin prime pairs; accordingly: $\pi_2(n)\le \frac 12 \pi(n) \sim \frac{n}{2 \ln n}$.
\section{Eratosthenes sieve and Legendre's formula for twin primes} \label{sec6}
The Eratosthenes sieve gives us an algorithm to find the prime numbers in a list of integers. The Legendre's formula uses the algorithm to write the prime-counting function in the following mathematical form: $\pi(x) - \pi(\sqrt{x})+1 = \sum_{d|P(\sqrt{x})} \mu(d) \left[\frac{x}{d}\right]$, where $P(z) = \prod_{p\le z}p$. Here we first discuss a similar sieve to detect the twin primes in $[1,x]$. This sieve may be called the Eratosthenes sieve for the twin primes. We use this sieve then to write a mathematical formula for twin-prime counting function - which may be called the Legendre's formula for twin primes.
\paragraph{\bf Eratosthenes sieve for twin primes.} Assume that we are to find out the (first members) of twin primes in $[1,N]$ with $N>2$. To start the sieving process, we first take a list of integers from 2 to $N+2$. We then carry out the following steps. Take the lowest integer ($n$) available in the list (initially $n=2$), and then remove (slashes ``/" in the example below) the integers from the list which are multiples of but greater than $n$. Then remove (strike-outs ``---" in the example below) the integers from the list which are two less than the integers removed in the last step. Circle the lowest integer available in the list (3 in the first instance). Now repeat the process with the circled number (new $n$) and continue the process till we reach $N$ in the list. At the end of the process, the circled numbers are the (first members) of the twin primes in $[1,N]$.
\noindent {\bf Example.} To find the (first members) of twin primes below 30, we take the following list of numbers from 2 to 32. We then carry out the above mentioned steps to circle out the twin primes below 30.
\noindent \sout{2} \circled{$\left. 3\right.$} \sout{$\slashed{4}$} \circled{$\left. 5\right.$} \sout{$\slashed{6}$} \sout{7} \sout{$\slashed{8}$} $\slashed{9}$ \sout{$\slashed{10}$} \circled{11} \sout{$\slashed{12}$} \sout{13} \sout{$\slashed{14}$} $\slashed{15}$ \sout{$\slashed{16}$} \circled{17} \sout{$\slashed{18}$} \sout{19} \sout{$\slashed{20}$} $\slashed{21}$ \sout{$\slashed{22}$} \sout{23} \sout{$\slashed{24}$} \sout{$\slashed{25}$} \sout{$\slashed{26}$} $\slashed{27}$ \sout{$\slashed{28}$} \circled{29} \sout{$\slashed{30}$} 31 $\slashed{32}$
\noindent So the twin primes (first members) below 30 are 3, 5, 11, 17 and 29.
\paragraph{\bf Legendre's formula for twin primes.} The main idea behind the sieve above can be used to find a formula for $\pi_2(x)$. We write this formula as a theorem below. The proof of the theorem is give in Section \ref{sec7}. \begin{theorem} If $P(z) = {\displaystyle \prod_{p\le z}}p$ and $[x]$ denotes the greatest integer $\le x$, we have \begin{equation} \begin{split}
\pi_2(x) - \pi_2(\sqrt{x}) & = \sum_{ab|P(\sqrt{x})}\mu(ab) \left[\frac{x-l_{a,b}}{ab}\right], \end{split} \label{eq6.1} \end{equation} where $a$ takes only odd values and $b$ takes both odd and even values. Here $l_{a,b}$ is the lowest (positive) integer which is divisible by $a$ and simultaneously $l_{a,b}+2$ is divisible by $b$. If we consider $l_{a,b}=at_{a,b}$, then $t_{a,b}$ is the solution of the congruence $a t_{a,b}+2\equiv 0~(mod~b)$ with $0< t_{a,b} < b$ when $b>1$. For $b=1$, we take $t_{a,b}=1$. \label{th6.1} \end{theorem}
\section{Proofs of Theorems \ref{th4.1} and \ref{th6.1} } \label{sec7} We will prove Theorem \ref{th4.1} by two methods - first by the principle of cross-classification (inclusion-exclusion principle) and then by the Chinese remainder theorem. The first method is also used for proving Theorem \ref{th6.1}.
\noindent {\bf Proof of Theorem \ref{th4.1}}\\ {\it Method 1:} We first assume that $n$ is an odd number and $p_1, \cdots, p_r$ are the distinct prime divisors of $n$.
Let $S = \{1,2,\cdots,n\}$ and $S_k = \{m\le n : p_k|m(m+2)\}$. It is clear that \begin{equation} \phi_2(n) = N\left(S -\bigcup_{i=1}^r S_i\right), \nonumber \end{equation} where $N(T)$ denotes the number of elements in a subset $T$ of $S$. Now by the principle of cross-classification (see Theorem \ref{th8.2} for notations and formal statement), we have \begin{equation} \phi_2(n) = N(S) - \sum_{1\le i \le r}N(S_i) + \sum_{1\le i<j \le r}N(S_iS_j)- \cdots + (-1)^rN(S_1S_2\cdots S_r). \nonumber \end{equation} We have $N(S)=n$ and $N(S_i)=2\frac{n}{p_i}$.
In general, to calculate $N(S_{k_1}\cdots S_{k_l})$, we write $N(S_{k_1}\cdots S_{k_l}) = {\displaystyle \sum_{d|P_l}} N(A_{P_l}^{d})$ where $P_l$ is the product
of given $l$ primes corresponding to the sets $S_{k_1}, S_{k_2}, \cdots, S_{k_l}$, and $A_{P_l}^{d}=\{m\le n : d|m ~\text{and}~ d'|m+2\}$ with $dd' = P_l$. For $d'=1$, $N(A_{P_l}^{d})=\frac {n}{P_l}$. For $d'>1$, $N(A_{P_l}^{d})$ is given by the number of solutions for $a$ in the congruence $d a + 2 \equiv 0 ~(\text{mod}~d')$ so that $0<d a \le n$. This congruence has a unique solution modulo $d'$; let $t$ be the solution where $1\le t < d'$. All the solutions for $a$ would then be in the form $t+s d'$ where $s$ is so that $0<(t+s d') d \le n$. We have now $s \le \frac {n}{P_l} - \frac{t}{d'}$. Since $\frac {n}{P_l}$ is an integer and $\frac{t}{d'}<1$, the maximum value of $s$ can be $\frac {n}{P_l}-1$. On the other hand, the minimum value of $s$ is 0. This gives $N(A_{P_l}^{d})= \frac {n}{P_l}$. Since there are $2^l$ distinct divisors of $P_l$, we get $N(S_{k_1}\cdots S_{k_l}) = 2^l \frac {n}{P_l}$. Now using the principle of cross-classification, we get $\phi_2(n) = n - 2\sum_{1\le i \le r} \frac{n}{p_i} + 2^2 \sum_{1\le i<j \le r} \frac{n}{p_ip_j} - \cdots (-1)^r 2^r \frac{n}{p_1 \cdots p_r} =
n \sum_{d|n} \frac{\mu_2(d)}{d}$.
If $n$ is even, we can take $p_1=2$ and correspondingly $S_1=\{m\le n : p_1|m \}$. We now have $N(S_1)=\frac{n}{p_1}$. To find $N(S_1S_{k_1}\cdots S_{k_l})$, where the sets $S_{k_1}, S_{k_2}, \cdots, S_{k_l}$ correspond to $l$ different odd primes, we write
$N(S_1S_{k_1}\cdots S_{k_l}) = {\displaystyle \sum_{d|P_l}} N(A_{P_l}^{d})$. Here $P_l$ is the product of given $l$ odd primes, and
$A_{P_l}^{d}=\{m\le n : d|m ~\text{and}~ 2d'|m+2\}$ with $dd' = P_l$. For $d'=1$, $N(A_{P_l}^{d})=\frac {n}{2P_l}$. For $d'>1$, $N(A_{P_l}^{d})$ is given by the number of solutions for $a$ in the congruence $d a + 2 \equiv 0 ~(\text{mod}~2d')$ so that $0<da \le n$. Proceeding
as before, we find that $N(A_{P_l}^{d})= \frac {n}{2P_l}$ and $N(S_1S_{k_1}\cdots S_{k_l}) = 2^l \frac {n}{2P_l}$. Using these values, we finally get $\phi_2(n)=n \sum_{d|n} \frac{\mu_2(d)}{d}$. We may here note that the distinctiveness of even $n$ is taken care by $\mu_2(n)$ which is defined by $\mu_2(n)=\mu(n)2^{\omega_o(n)}$ where $\omega_o(n)$ is the number of odd prime divisors of $n$.
\noindent {\it Method 2:} Let $n=p_1^{a_1}p_2^{a_2}\cdots p_r^{a_r}$. Now consider the system of congruences with $i$-th congruence as $x\equiv b_i$ (mod $p_i^{a_i}$), where $(b_i,p_i)=(b_i+2,p_i)=1$ and $0<b_i<p_i^{a_i}$. In the congruence, $b_i$ can take $\phi_2(p_i^a)$ values, where, as discussed in Section \ref{sec3}, $\phi_2(p^a) = p^a(1-\frac 2p)$ for an odd prime $p$, and $\phi_2(p^a) = p^a(1-\frac 1p)$ for the even prime $p=2$. For a given set $(b_1,b_2,\cdots b_r)$, the system of congruences has a unique solution modulo $n$ according to the Chinese remainder theorem. So for all possible values of $(b_1,b_2,\cdots b_r)$, there are $\prod_r \phi_2(p_r^{a_r})$ unique solutions which lie above 0 and below $n$.
\noindent {\bf Proof of Theorem \ref{th6.1}}\\ To find the number of twin primes in $[\sqrt{x},x]$, we can suitably use the concept behind the Eratosthenes sieve for twin primes discussed in Section \ref{sec6}. To translate this concept in the mathematical form, we adopt here the approach which is used in the proof (Method 1) of Theorem \ref{th4.1}.
As before, we take $p_1=2$ and $S_1=\{m\le x : p_1|m \}$. For any odd prime $p_k$, we take $S_k = \{m\le x : p_k|m(m+2)\}$. This sets help us to write \begin{equation} \pi_2(x)-\pi_2(\sqrt{x}) = N\left(S -\bigcup_{1\le k \le r} S_k\right), \nonumber \end{equation} where $r = \pi(\sqrt{x})$ and $N(T)$ denotes the number of elements in a subset $T$ of $S$. Now by the principle of cross-classification (Theorem \ref{th8.2}), we have \begin{equation} \begin{split} \pi_2(x)-\pi_2(\sqrt{x}) = & N(S) - \sum_{1\le i \le r}N(S_i) + \sum_{1\le i<j \le r}N(S_iS_j)- \cdots \\
& + (-1)^rN(S_1S_2\cdots S_r). \label{pi2x} \end{split} \end{equation}
We have $N(S)=[x]$, $N(S_1)=[\frac{x}{p_1}]$ for $p_1=2$, and $N(S_i)=2[\frac{x}{p_i}]$ for odd primes. In general, to find $N(S_{k_1}\cdots S_{k_l})$, where the sets $S_{k_1}, S_{k_2}, \cdots, S_{k_l}$ correspond to $l$ different odd primes, we write $N(S_{k_1}\cdots S_{k_l}) = {\displaystyle \sum_{d|P_l}} N(A_{P_l}^{d})$ where $P_l$ is the product of given $l$ odd primes and $A_{P_l}^{d}=\{m\le x : d|m ~\text{and}~ d'|m+2\}$ with $dd' = P_l$. For $d'=1$, $N(A_{P_l}^{d})=[\frac {x}{P_l}]$. For $d'>1$, $N(A_{P_l}^{d})$ is given by the number of solutions for $a$ in the congruence $d a + 2 \equiv 0 ~(\text{mod}~d')$ so that $0<d a \le x$. This congruence has a unique solution modulo $d'$; let $t_{d,d'}$ be the solution where $1\le t_{d,d'} < d'$. All the solutions for $a$ would then be in the form $t_{d,d'}+s d'$ where $s$ is so that $0<d (t_{d,d'}+s d') \le x$. Since $t_{d,d'} < d'$, $s$ can not be a negative integer. In other way, when $d t_{d,d'}>x$ or $s \le \frac {x}{P_l} - \frac{t_{d,d'}}{d'} < 0$, then we do not have any solution for $a$; hence in such a case $N(A_{P_l}^{d})=0$. Since $\frac{t_{d,d'}}{d'}<1$, we have $-1<\frac {x}{P_l} - \frac{t_{d,d'}}{d'}$. So in the case when $d t_{d,d'}>x$, we can write $N(A_{P_l}^{d})=1+\left[ \frac {x}{P_l} - \frac{t_{d,d'}}{d'} \right]$. When $d t_{d,d'}\le x$ or $\frac {x}{P_l} - \frac{t_{d,d'}}{d'}\ge 0$, then $s$ can take any value from 0 to $\left[\frac {x}{P_l} - \frac{t_{d,d'}}{d'}\right]$. Therefore, in this case also $N(A_{P_l}^{d})=1+\left[ \frac {x}{P_l} - \frac{t_{d,d'}}{d'} \right]$.
To find $N(S_1S_{k_1}\cdots S_{k_l})$, where the sets $S_{k_1}, S_{k_2}, \cdots, S_{k_l}$ correspond to $l$ different odd primes, we write
$N(S_1S_{k_1}\cdots S_{k_l}) = {\displaystyle \sum_{d|P_l}} N(A_{P_l}^{d})$. Here $P_l$ is the product of given $l$ odd primes, and $A_{P_l}^{d}=\{m\le x : d|m ~\text{and}~ 2d'|m+2\}$ with $dd' = P_l$. For $d'=1$, $N(A_{P_l}^{d})=[\frac {x}{2P_l}]$. For $d'>1$, $N(A_{P_l}^{d})$ is given by the number of solutions for $a$ in the congruence $da + 2 \equiv 0 ~(\text{mod}~2d')$ so that $0<da \le x$. Let $t_{d,2d'}$ be the unique solution of the congruence, where $0 < t_{d,2d'} < 2d'$. All the solutions for $a$ would then be in the form $t_{d,2d'}+2s d'$ where $s$ is so that $0< d (t_{d,2d'}+2s d') \le x$. Proceeding as for the odd $n$ case, we find that $N(A_{P_l}^{d})=1+\left[ \frac {x}{2P_l} - \frac{t_{d,2d'}}{2d'} \right]$.
If we write $N(S) = N(1)$, $N(S_i) = N(p_i)$, $N(S_iS_j) = N(p_ip_j)$ and so on, we can rewrite Equation \ref{pi2x} in the following way, \begin{equation}
\pi_2(x)-\pi_2(\sqrt{x}) = \sum_{a|P(\sqrt{x})}\mu(a)N(a), \label{pi2x2} \end{equation}
where $P(\sqrt{x})= {\displaystyle \prod_{p\le\sqrt{x}}p}$. Now $N(a)=\sum_{d|a}' N(A_{P_l}^{d})$ where $\sum'$ indicates the sum over only the odd divisors ($d$) of $a$ and $N(A_{P_l}^{d})=1+\left[ \frac {x}{a} - \frac{t_{d,d'}}{d'} \right]$ with $d'=a/d$ for $d'>1$. Here $0<t_{d,d'}<d'$ and $d t_{d,d'} +2 \equiv 0 ~(\text{mod}~d')$. For $d'=1$, we take $t_{d,d'}=1$.
Using the result that $\sum_{d|n}\mu(d) = 0$ for $n>1$, we finally get from Equation \ref{pi2x2},
$\pi_2(x)-\pi_2(\sqrt{x}) = {\displaystyle \sum_{dd'|P(\sqrt{x})}\mu(dd')\left[ \frac {x}{dd'} - \frac{t_{d,d'}}{d'} \right]}$, where $d$ is always odd and $d'$ can take both odd and even values.
\section{Some standard results} \label{sec8}
In this section we will provide some useful standard results without their proofs.
\begin{theorem} Consider two functions $F(s)$ and $G(s)$ which are represented by the following two Dirichlet series, \begin{equation} \begin{split} F(s) &= \sum_{n=1}^{\infty}\frac{f(n)}{n^s} ~~for~~ Re(s)>a, ~~and \\ \nonumber G(s) & = \sum_{n=1}^{\infty}\frac{g(n)}{n^s} ~~for~~ Re(s)>b. \end{split} \end{equation} Then in the half-plane where both the series converge absolutely, we have \begin{equation} F(s)G(s)= \sum_{n=1}^{\infty}\frac{h(n)}{n^s}, \nonumber \end{equation} where $h=f\ast g$, the Dirichlet convolution of $f$ and $g$: \begin{equation}
h(n) = \sum_{d|n}f(d)g(\frac nd). \nonumber \end{equation} \label{th8.1} \end{theorem}
The series for $F(s)$ and $G(s)$ are absolutely convergent respectively for $Re(s)>a$ and $Re(s)>b$. The proof of the theorem can be found in any standard number theory text book (e.g. chapter 11, \cite{apostol76}).
\begin{lemma} For a multiplicative function $f$, we have \begin{equation} \begin{split}
\sum_{d|n}\mu(d)f(d) = &\prod_{p|n}(1-f(p)) ~~\text{and} \\\nonumber
\sum_{d|n}\mu^2(d)f(d) = &\prod_{p|n}(1+f(p)). \end{split} \end{equation} \label{lm8.1} \end{lemma}
\begin{lemma} Let $f$ be a multiplicative function so that the series $\sum_n f(n)$ is absolutely convergent. Then the sum of the series can be expressed as the following absolutely convergent infinite product over all primes, \begin{equation} \sum_{n=1}^{\infty} f(n) = \prod_{p} \{1+f(p)+f(p^2)+ \cdots\}. \nonumber \end{equation} \label{lm8.2} \end{lemma}
\begin{theorem}[Principle of cross-classification] Let $S$ be a non-empty finite set and $N(T)$ denotes the number of elements of any subset $T$ of $S$. If $S_1, S_2, ..., S_n$ are given subsets of $S$, then \begin{equation} \begin{split} N\left(S-\bigcup_{i=1}^r S_i\right) = & ~N(S) - \sum_{1\le i \le n}N(S_i) + \sum_{1\le i<j \le n}N(S_iS_j)\\ \nonumber
&-\sum_{1\le i<j<k \le n}N(S_iS_jS_k)+ \cdots + (-1)^nN(S_1S_2\cdots S_n), \end{split} \end{equation} where $S-T$ consists of those elements of $S$ which are not in $T$, and $S_iS_j$, $S_iS_jS_k$, etc. denote respectively $S_i\cap S_j$, $S_i\cap S_j\cap S_k$, etc. \label{th8.2} \end{theorem}
\begin{theorem}
If $h(n)=\sum_{d|n}f(d)g(\frac nd)$, let\\ \begin{equation} H(x) = \sum_{n\le x} h(n), ~~ F(x) = \sum_{n\le x} f(n), ~~ \text{and} ~~ G(x) = \sum_{n\le x} g(n). \nonumber \end{equation} We then have \\ \begin{equation} H(x) = \sum_{n\le x} f(n) G(\frac xn) = \sum_{n\le x} g(n)F(\frac xn). \nonumber \end{equation} \label{th8.3} \end{theorem} The proof of the theorem can be found in, for example, chapter 3 of Ref. \cite{apostol76}.
\section{Biases in twin prime distribution} \label{sec10}
Our analysis of the first 500 million prime numbers and corresponding about 30 million twin prime pairs ($\pi(x_0)=5\times 10^8$ and $\pi_2(x_0)=29,981,546$) shows three different types of biases both in the distributions of the prime numbers and the twin prime pairs. For plotting different arithmetic functions, we take a data point after every 50 primes ($x_i=p_{_{50i}}$) while presenting the results concerning prime numbers, and we take a data point after every 25 twin prime pairs ($x_i=\hat{p}_{_{25i}}$) while presenting the results concerning twin primes. In the following we present our findings.
\subsection{Type-I bias} If $\pi_2(x;4;1)$ and $\pi_2(x;4;3)$ represent the number of prime pairs $\le x$ in the residue class $a=1$ and $a=3$ respectively, we find that $\pi_2(x;4;1)>\pi_2(x;4;3)$ for the most values of $x$. This can be seen from the Table \ref{tab1}. The bias is also evident from Figure \ref{fig1} where we plot the functions $\delta_2(x;4;3) = \frac{\pi_2(x;4;3)}{\pi_2(x)}$ and $\bar{\Delta}_2(x;4;3,1) = (\pi_2(x;4;3) - \pi_2(x;4;1))\frac{ln(x)}{10\sqrt{x}}$. In the definition of $\bar{\Delta}_2(x;4;3,1)$, the factor 10 in the denominator is an overall scaling factor. Ideally, without bias, $\delta_2(x;4;3)$ should be 0.5 and $\bar{\Delta}_2(x;4;3,1)$ should be zero. We also plot the corresponding functions for the prime numbers; we see that, while the prime numbers are biased towards the residue class $a=3$, the twin prime pairs are in contrast biased towards the residue class $a=1$.
\begin{table}[th] \renewcommand \arraystretch{1.5} \noindent\[
\begin{array}{r|r r r r r}
x & \pi(x;4;3) & \Delta(x;4;3,1) & \pi_2(x;4;3) & \Delta_2(x;4;3,1)\\ \hline 10^{7} & 332398 & 218 & 29498 & 16 \\ 5\cdot10^{7} & 1500681 & 229 & 119330 & -441 \\ 10^{8} & 2880950 & 446 & 219893 & -526 \\ 5\cdot10^{8} & 13179058 & 2250 & 919192 & -1786 \\ 10^{9} & 25424042 & 551 & 1711775 & -956 \\ 5\cdot10^{9} & 117479241 & 4260 & 7308783 & -600 \\ 10^{10} & 227529235 & 5960 & 13706087 & -505 \\ \end{array} \] \caption{(Type-I bias) Here biases in case of primes and twin primes are respectively quantified by the functions $\Delta(x;4;3,1)=\pi(x;4;3)-\pi(x;4;1)$ and $\Delta_2(x;4;3,1)=\pi_2(x;4;3)-\pi_2(x;4;1)$. } \label{tab1} \end{table}
\begin{figure}
\caption{(Type-I bias) Plots of the functions $\delta_2(x;4;3)$ and $\bar{\Delta}_2(x;4;3,1)$. The corresponding plots for the prime numbers are also shown. The broken lines show expected results if there were no bias.}
\label{fig1}
\end{figure}
Although the overall bias is seen to be towards the residue class $a=1$ for the prime pairs, there is an interval in between where the class $a=3$ is preferred. In fact we numerically find that upto about $x\approx50,000$, the residue class $a=1$ is preferred, then upto about $x\approx10^7$, the residue class $a=3$ is preferred. After this for a very long interval (we check upto about $x\approx 1.1\times 10^{10}$), the residue class $a=1$ is again preferred for the twin primes.
We also calculate the Brun's constant separately for the two residue classes. Let $B_2(x;4;1) = (\frac15 +\frac17)+(\frac{1}{17} +\frac{1}{19}) + \cdots$ and $B_2(x;4;3) = (\frac13 + \frac15) + (\frac{1}{11}+\frac{1}{13})+\cdots$. Here the first series involves the twin pairs whose first members are congruent to 1 (mod 4), similarly, the second series involves the twin primes whose first members are congruent to 3 (mod 4). In both cases we only consider the prime pairs in $[1,x]$. We find that $B_2(x_0;4;1)\approx0.802233$ and $B_2(x_0;4;3)\approx0.985735$, where $x_0$ is given by $\pi_2(x_0)=29,981,546$. We note here that $B_2(x_0) = B_2(x_0;4;1) + B_2(x_0;4;3) \approx 1.787967$. The value of $B_2(x_0)$ is still somewhat far from its known value of $B_2\approx1.902161$; this is due to extremely slow convergence of the series of the reciprocals of twin primes. We find that $B_2(x_0;4;3)$ keeps a lead over $B_2(x_0;4;1)$ from the beginning (the first twin pair belongs to class $a=3$). Since $B_2-B_2(x_0)\approx0.114194$ and $B_2(x_0;4;3)-B_2(x_0;4;1)=0.183502$, it can be concluded that, irrespective of how large $x$ is, $B_2(x;4;3)$ will always be larger than $B_2(x;4;1)$. In future study, it will be interesting to find out whether $lim_{x\rightarrow \infty} B_2(x;4;1)$ and $lim_{x\rightarrow \infty} B_2(x;4;3)$ exist and, if so, what the corresponding limits are.
\begin{figure}
\caption{(Type-II bias) Plots of the functions $\delta_2(x;4;1|1)$ and $\delta_2(x;4;3|1)$ along with their counterparts for the prime numbers are shown here. Plots of the complementary functions, like $\delta_2(x;4;3|3) = 1 - \delta_2(x;4;3|1)$, are not provided. The broken line shows expected result if there were no bias.}
\label{fig2}
\end{figure}
\subsection{Type-II bias} A second type of bias can be found if we consider the consecutive twin prime pairs. Our numerical results show that after a prime pair of certain residue class, it is more probable to find the next prime pair to be from the different residue class. To quantify
this bias, we define the following functions: $\delta_2(x;4;1|1):= \frac{\pi_2(x;4;1|1)}{\pi_2(x;4;1)}$ and
$\delta_2(x;4;1|3):= \frac{\pi_2(x;4;1|3)}{\pi_2(x;4,1)}$. Here $\pi_2(x;4;a_i|a_j)$ denotes the number of twin prime pairs ($\le x$) belonging to the residue class $a_i$ provided that their next pairs belong to the class $a_j$.
It may be noted that $\delta_2(x;4;1|1)+\delta_2(x;4;1|3)=1$.
Functions $\delta_2(x;4;3|1)$ and $\delta_2(x;4;3|3)$ are defined similarly. The plots of function $\delta_2(x;4;1|1)$ and $\delta_2(x;4;3|1)$ can be found in Figure \ref{fig2}. The corresponding plots for the prime numbers are also shown in the figure. If we consider the twin prime pairs to be completely uncorrelated, the values of these functions should be 0.5. But we see that, for example,
$\delta_2(x;4;1|3)>0.5>\delta_2(x;4;1|1)$ for all values of $x$ that we investigated. Exactly same behavior is seen in case of prime numbers, although the bias is more here. The bias in consecutive primes was investigated earlier in Ref. \cite{oliver16}.
We here would like to point out that the functions like $\delta_2(x;4;1|1)$ and $\delta_2(x;4;3|1)$ are quite regular and smooth compared
to a function like $\delta_2(x;4,3)$ or $\bar{\Delta}_2(x;4;3,1)$. The values of the functions $\delta_2(x;4;1|1)$ and $\delta_2(x;4;3|1)$ for some values of $x$ can be found in Table \ref{tab2}. \begin{table}[] \renewcommand \arraystretch{1.5} \noindent\[
\begin{array}{c|c c c c}
x & \delta(x;4;1|1) & \delta(x;4;3|1) & \delta_2(x;4;1|1) & \delta_2(x;4;3|1)\\ \hline 10^{7} & 0.4350 & 0.5647 & 0.4769 & 0.5228 \\
10^{8} & 0.4434 & 0.5565 & 0.4815 & 0.5198 \\
10^{9} & 0.4489 & 0.5511 & 0.4836 & 0.5166 \\
10^{10} & 0.4534 & 0.5466 & 0.4864 & 0.5136 \\
\end{array} \]
\caption{(Type-II bias) Variation of functions $\delta(x;4;a_i|a_j)$ and $\delta_2(x;4;a_i|a_j)$ are very slow and smooth. Values of the functions are rounded off to the four decimal places.} \label{tab2} \end{table}
Although $\delta_2(x;4;a_i|a_j)$ varies very slowly, we expect all these functions to approach 0.5 as $x$ goes to infinity. Assuming that there are infinite number of twin prime pairs, we propose the following conjecture. \begin{conjecture}
For any two residue classes $a_i$ and $a_j$, we have $\delta_2(x;4;a_i|a_j)=0.5$ as $x\rightarrow\infty$. \label{cj10.1} \end{conjecture}
\begin{figure}
\caption{(Type-III bias) Plots of the functions $\delta^{+}_2(x)$ and $\delta^{-}_2(x)$. The corresponding results for the prime numbers are also provided.}
\label{fig3}
\end{figure}
\subsection{Type-III bias} A third type of bias can be found if we study the gap between the first members of consecutive twin prime pairs. To quantify the bias, we define the following functions: $\delta^{+}_2(x):=\frac{\pi^{+}_2(x)}{\pi_2(x)}$, where $\pi^{+}_2(x)$ is the number of prime pairs ($\le x$) for each of which the following twin prime gap plus 1 is a prime number, i.e.,
$\pi^{+}_2(x) = \#\{\hat{p}_n \le x~| (\hat{p}_{n+1}-\hat{p}_n + 1) ~\text{is a prime} \}$. Similarly, the function $\delta^{-}_2(x)$ is defined to quantify the bias in the twin prime gap minus 1. If the prime pairs were distributed in a completely random manner, both $\delta^{\pm}_2(x)$ would have been less than 0.5 since we have more odd composites than the primes. But what we find is that $\delta^{-}_2(x) > \delta^{+}_2(x) > 0.5$ for all the values of $x$ that we investigated. Exactly similar bias is seen for the prime numbers, but in this case, interestingly, $\delta^{+}(x) > \delta^{-}(x) > 0.5$. The results related to this particular type of bias can be seen in Fig. \ref{fig3}. The values of the functions $\delta^{\pm}_2(x)$ for some values of $x$ can be found in Table \ref{tab3}. \begin{table}[ht] \renewcommand \arraystretch{1.5} \noindent\[
\begin{array}{c| c c c c}
x & \delta^{+}(x) & \delta^{-}(x) & \delta^{+}_2(x) & \delta^{-}_2(x) \\ \hline 10^{7} & 0.7418 & 0.6739 & 0.6673 & 0.7103 \\
10^{8} & 0.7203 & 0.6664 & 0.6401 & 0.6803 \\
10^{9} & 0.7026 & 0.6585 & 0.6186 & 0.6560 \\
10^{10} & 0.6875 & 0.6506 & 0.6001 & 0.6349 \\
\end{array} \] \caption{(Type-III bias) Variation of functions $\delta^{\pm}(x)$ and $\delta^{\pm}_2(x)$ are very slow and smooth. Values of the functions are rounded off to the four decimal places.} \label{tab3} \end{table}
Although $\delta^{\pm}(x)$ and $\delta^{\pm}_2(x)$ change very slowly, we expect all these functions to eventually go to 0 since the number of odd composite numbers grow faster than the prime numbers. Assuming infinitude of twin primes, we propose the following conjecture for this type of bias. \begin{conjecture} As $x \rightarrow \infty$, $\delta^{+}(x) = \delta^{-}(x) = 0$ and $\delta^{+}_2(x) = \delta^{-}_2(x) = 0$. \label{cj10.2} \end{conjecture}
\end{document}
|
arXiv
|
{
"id": "2111.09053.tex",
"language_detection_score": 0.7123509645462036,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Generalized Euler classes, differential forms and commutative DGAs}
\begin{abstract}
In the context of commutative differential graded algebras over $\mathbb Q$, we show that an iteration of ``odd spherical fibration" creates a ``total space" commutative differential graded algebra with only odd degree cohomology. Then we show for such a commutative differential graded algebra that, for any of its ``fibrations" with ``fiber" of finite cohomological dimension, the induced map on cohomology is injective. \end{abstract}
\section{Introduction} In geometry, one would like to know which rational cohomology classes in a base space can be annihilated by pulling up to a fibration over the base with finite dimensional fiber. One knows that if $[x]$ is a $2n$-dimensional rational cohomology class on a finite dimensional CW complex $X$, there is a $(2n-1)$-sphere fibration over $X$ so that $[x]$ pulls up to zero in the cohomology groups of the total space. In fact there is a complex vector bundle $V$ over $X$ of rank $n$ whose Euler class is a multiple of $[x]$. Thus this multiple is the obstruction to a nonzero section of $V$, and vanishes when pulled up to the part of $V$ away from the zero section, which deformation retracts to the unit sphere bundle.
Rational homotopy theory provides a natural framework to study this type of questions, where topological spaces are replaced by commutative differential graded algebras (commutative DGAs) and topological fibrations replaced by algebraic fibrations. This will be the context in which we work throughout the paper. The reader can read more in \cite{MR1802847, MR2355774, MR0646078} about the topological meaning of the results of this paper from the perspective of rational homotopy theory of manifolds and general spaces.
The first theorem (Theorem $\ref{thm:itoddsphere}$) of the paper states that the above construction, when iterated, creates a ``total space" commutative DGA with only odd degree cohomology.
\begin{thma} For each commutative DGA $(A, d)$, there exists an iterated odd algebraic spherical fibration $(TA, d) $ over $(A, d)$ so that all even cohomology [except dimension zero] vanishes. \end{thma}
Our next theorem (Theorem $\ref{thm:oddsphere}$) then limits the odd degree classes that can be annihilated by fibrations whose fiber has finite cohomological dimension.
\begin{thmb} Let $(B, d)$ be a connected commutative DGA such that $H^{2k}(B)=0$ for all $0 < 2k \leq 2N$. If $\iota \colon (B, d) \to (B\otimes \Lambda V, d) $ is an algebraic fibration whose algebraic fiber has finite cohomological dimension, then the induced map \[ \iota_\ast \colon \bigoplus_{i\leq 2N} H^i(B)\to \bigoplus_{i\leq 2N} H^i(B\otimes \Lambda V)\] is injective. \end{thmb}
It follows from the two theorems above that the iterated odd spherical fibration construction is universal for cohomology classes that pull back to zero by any fibrations whose fiber has finite cohomological dimension.
The paper is organized as follows. In Section $\ref{sec:pre}$, we recall some definitions from rational homotopy theory. In Section $\ref{sec:gysin}$, we use iterated algebraic spherical fibrations to prove Theorem A. In Section $\ref{sec:algsphere}$, we define bouquets of algebraic spheres and analyze their minimal models. In Section $\ref{sec:main}$, we prove Theorem B.
The work in this paper started during a visit of the authors at IHES. The authors would like to thank IHES for its hospitality and for providing a great environment.
\section{Preliminaries} \label{sec:pre}
We recall some definitions related to commutative differential graded algebras. For more details, see \cite{MR1802847, MR2355774, MR0646078}.
\begin{definition}
A commutative differential graded algebra (commutative DGA) is a graded algebra $B = \oplus_{i\geq 0} B^{i}$ over $\mathbb Q$ together with a differential $d\colon B^i \to B^{i+1}$ such that $d^2 = 0$, $xy = (-1)^{ij} yx,$ and $d(xy) = (d x) y + (-1)^{i} x (d y)$, for all $x\in B^{i}$ and $y\in B^{j}$.
\end{definition}
\begin{definition}
\begin{enumerate}[(1)]
\item
A commutative DGA $(B, d)$ is called connected if $B^0 = \mathbb Q$.
\item A commutative DGA $(B, d) $ is called simply connected if $(B, d)$ is connected and $H^1(B) = 0$.
\item
A commutative DGA $(B, d)$ is of finite type if $H^k(B)$ is finite dimensional for all $k\geq 0$.
\item A commutative DGA $(B,d )$ has finite cohomological dimension $d$, if $d$ is the smallest integer such that $H^k(B) = 0 $ for all $k> d$.
\end{enumerate} \end{definition}
\begin{definition}
A connected commutative DGA $(B, d)$ is called a model algebra if as a commutative graded algebra it is free on a set of generators $\{x_\alpha\}_{\alpha\in \Lambda}$ in positive degrees, and these generators can be partially ordered so that $d x_\alpha$ is an element in the algebra generated by $x_\beta$ with $\beta < \alpha$.
\end{definition}
\begin{definition}
A model algebra $(B, d)$ is called minimal if for each generator $x_\alpha$, $d x_\alpha$ has no linear term, that is,
\[ d(B) \subset B^+ \cdot B^+, \textup{ where } B^+ = \oplus_{k>0} B^k.\]
\end{definition}
\begin{remark}
For every connected commutative DGA $(A, d_A)$, there exists a minimal model algebra $(\mathcal M(A), d)$ and a morphism $\varphi\colon (\mathcal M(A), d)\to (A, d_A)$ such that $\varphi$ induces an isomorphism on cohomology. $(\mathcal M(A), d)$ is called a minimal model of $(A, d)$, and is unique up to isomorphism. See page $288$ of \cite{MR0646078} for more details, cf. \cite{MR1802847, MR2355774}. \end{remark}
\begin{definition}\label{def:algfib}
\begin{enumerate}[(i)]
\item An algebraic fibration (also called \textit{relative model algebra}) is an inclusion of commutative DGAs $(B, d) \hookrightarrow (B\otimes \Lambda V, d)$ with $V = \oplus_{k\geq 1} V^k$ a graded vector space; moreover, $V = \bigcup_{n=0}V(n)$, where $V(0) \subseteq V(1) \subseteq V(2) \subseteq \cdots $ is an increasing sequence of graded subspaces of $V$ such that
\[ d : V(0) \to B \quad \textup{ and } \quad d: V(n) \to B\otimes \Lambda V(n-1), \quad n\geq 1,\]
where $\Lambda V$ is the free commutative DGA generated by $V$.
\item An algebraic fibration is called minimal if
\[ \textup{Im}(d) \subset B^+\otimes \Lambda V + B\otimes \Lambda^{\geq 2} V.\]
\end{enumerate}
\end{definition}
Let $\iota\colon (B, d) \hookrightarrow (B\otimes \Lambda V, d)$ be an algebraic fibration. Suppose $B$ is connected. Consider the canonical augmentation morphism $\varepsilon: (B, d) \to (\mathbb Q, 0)$ defined by $\varepsilon (B^+) = 0$. It naturally induces a commutative DGA: \[ (\Lambda V, \bar d) := \mathbb Q\otimes_B (B\otimes \Lambda V, d).\] We call $(\Lambda V, \bar d) $ the algebraic fiber of the given algebraic fibration.
\section{Iterated odd spherical algebraic fibrations}\label{sec:gysin}
In this section, we show that for each commutative DGA, there exists an iterated odd algebraic spherical fibration over it such that the total commutative DGA has only odd degree cohomology.
Let $(B, d)$ be a connected commutative DGA. An \emph{odd algebraic spherical fibration} over $(B, d)$ is an inclusion of commutative DGAs of the form \[ \varphi: (B, d) \to (B\otimes \Lambda (x), d), \] such that $d x \in B$, where $x$ has degree $2k-1$ and $\Lambda(x)$ is the free commutative graded algebra generated by $x$. The element $e = d x \in B$ is called the Euler class of this algebraic spherical fibration.
\begin{proposition}\label{prop:gysin}
Let $(B, d)$ be a commutative DGA. For every even dimensional class $\beta\in H^{2k}(B)$ with $k>0$, there exists an odd algebraic spherical fibration $\varphi\colon (B, d) \to (B\otimes \Lambda(x), d)$ such that its Euler class is equal to $\beta$ and the kernel of the map $\varphi_\ast\colon H^{i+2k}(B) \to H^{i+2k}(B\otimes \Lambda (x) )$ is $H^i(B)\cdot\beta = \{a\cdot \beta \mid a\in H^i(B)\}$.
\end{proposition} \begin{proof}
Let $(B\otimes \Lambda (x), d)$ be the commutative DGA obtained from $(B, d)$ by adding a generator $x$ of degree $2k-1$ and defining its differential to be $d x = \beta$. We have the following short exact sequence
\[ 0 \to (B, d) \to (B\otimes \Lambda (x), d) \to ( B \otimes (\mathbb Q\cdot x), d\otimes \textup{Id}) \to 0,\]
which induces a long exact sequence \[ \cdots \to H^{i-1}(B\otimes (\mathbb Q \cdot x)) \to H^i(B) \to H^i(B\otimes \Lambda (x)) \to H^i(B\otimes (\mathbb Q \cdot x)) \to \cdots.
\]
Applying the identification $H^{i + (2k-1)}(B\otimes (\mathbb Q \cdot x)) \cong H^i(B)$, we obtain the following Gysin sequence
\[ \cdots \to H^{i}(B) \xrightarrow{\cup e} H^{i+2k}(B) \xrightarrow{\varphi_\ast} H^{i+2k}(B\otimes \Lambda (x)) \xrightarrow{\partial_{i+1}} H^{i+1}(B) \to \cdots. \]
This finishes the proof. \end{proof}
\begin{definition}
An \emph{iterated odd algebraic spherical fibration} over $(B, d)$ is algebraic fibration $(B, d) \hookrightarrow (B\otimes \Lambda V, d)$
such that $V^k=0$ for $k$ even. This fibration is called \emph{finitely iterated odd algebraic spherical fibration} if $\dim V < \infty$.
\end{definition}
Now let us prove the main result of this section.
\begin{theorem}\label{thm:itoddsphere}
For each commutative DGA $(A, d)$, there exists an iterated odd algebraic spherical fibration $(TA, d) $ over $(A, d)$ such that all even cohomology [except dimension zero] vanishes. \end{theorem}
\begin{proof}
We will construct $TA$ by induction. In the following, for notational simplicity, we shall omit the differential $d$ from our notation.
Let $\mathcal A_0 = A$. Suppose we have defined the iterated odd algebraic spherical fibration $\mathcal A_{m-1}$ over $A$. Fix a basis of $H^{2k}(\mathcal A_{m-1})$ for each $k>0$. Denote the union of all these bases by $\{a_i\}_{i\in I}$. Define $W_{m-1}$ to be a $\mathbb Q$ vector space with basis $\{x_i\}_{i\in I}$, where $|x_i|= |a_i| -1$.
We define $\mathcal A_{m}$ to be the iterated odd algebraic spherical fibration $ \mathcal A_{m-1} \otimes \Lambda (W_{m-1})$ over $\mathcal A_{m-1}$ with $d x_i = a_i$ for all $i\in I$. The inclusion map $\iota\colon \mathcal A_{m-1} \hookrightarrow \mathcal A_{m}$ induces the zero map $\iota_\ast = 0 \colon H^{2k}(\mathcal A_{m-1}) \to H^{2k}(\mathcal A_{m})$ for all $k>0$. By construction, $\mathcal A_m$ is also an iterated odd algebraic spherical fibration.
Finally, we define $TA$ to be the direct limit of $\mathcal A_m$ under the inclusions $\mathcal A_m \hookrightarrow \mathcal A_{m+1}$. Clearly, $TA$ is an iterated odd algebraic spherical fibration over $A$. More precisely, let $V = \bigcup_{i=0}^\infty W_{i}$. We have $TA = A\otimes \Lambda V$ with the filtration of $V$ given by $V(n) = \bigcup_{i=0}^n W_{i}$. Moreover, we have $H^{2k}(TA) = 0$ for all $2k>0$. This completes the proof.
\end{proof}
\begin{remark}\label{rm:finsphere}
If an element $\alpha \in H^\bullet(A) $ maps to zero in $H^{\bullet}(TA)$, then there exists a subalgebra $S_\alpha$ of $TA$ such that $S_{\alpha}$ is a \emph{finitely} iterated odd algebraic spherical fibration over $A$ and $\alpha$ maps to zero in $H^\bullet(S_{\alpha})$. \end{remark}
\section{Bouquets of algebraic spheres}
\label{sec:algsphere} In this section, we introduce a notion of bouquets of algebraic spheres. It is an algebraic analogue of usual bouquets of spheres in topology.
\begin{definition}\label{def:algsphere}
For a given set of generators $X= \{x_i\}$ with $x_i$ having odd degree $|x_i|$, we define the bouquet of odd algebraic spheres labeled by $X$ to be the following commutative DGA
\[\mathcal S(X) = \left( \bigwedge_{x_i\in X} \mathbb Q[x_i]\right)\Big/\langle x_ix_j = 0 \mid \textup{all } i, j \rangle \]
with the differential $d = 0$. \end{definition}
\begin{proposition}\label{prop:algsphere}
Let $\mathcal S(X)$ be a bouquet of odd algebraic spheres, and $\mathcal M(X) = (\Lambda V, d)$ be its minimal model. Then $\mathcal M(X)$ satisfies the following properties:
\begin{enumerate}[(i)]
\item $\mathcal M$ has no even degree generators, that is, $V$ does not contain even degree elements;
\item each element in $H^{\geq 1}(\mathcal M(X))$ is represented by a generator, that is, an element in $V$.
\end{enumerate} \end{proposition} \begin{proof}
This is a special case of Koszul duality theory, cf. \cite[Chapter 3, 7 \& 13]{MR2954392}. Since $\mathcal S = \mathcal S(X)$ has zero differential, we may forget its differential and view it as a graded commutative algebra. An explicit construction of a minimal model of $\mathcal S$ is given as follows: first take the Koszul dual coalgebra $\mathcal S^{\text{\textexclamdown}}$ of $\mathcal S$; then apply the cobar construction to $\mathcal S^{\text{\textexclamdown}}$, and denote the resulting commutative DGA by $\Omega \mathcal S^{\text{\textexclamdown}}$. By Koszul duality, $\mathcal M(X) \coloneqq \Omega \mathcal S^{\text{\textexclamdown}}$ is a minimal model of $\mathcal S$.
More precisely, set $W = \bigoplus_{i\geq 0} W_{i}$ to be the graded vector space spanned by $X$. Let $sW$ (resp. $s^{-1}W$) be the suspension (resp. desuspension) of $W$, that is, $(sW)_{i-1} = W_i$ (resp. $(s^{-1}W)_{i} = W_{i-1}$). Let $\mathcal L^c = \mathcal L^c(sW)$ be the cofree Lie coalgebra generated by $sW$. More explicitly, let $T^c(sW) = \bigoplus_{n\geq 0}(sW)^{\otimes n}$ be the tensor coalgebra, and $T^c(sW)^+ = \bigoplus_{n\geq 1} (sW)^{\otimes n}$. The coproduct on $T^c(sW)$ naturally induces a Lie cobraket on $T^c(sW)$. Then we have $\mathcal L^c(sW) = T^c(sW)^+/T^c(sW)^+\ast T^c(sW)^+$, where $\ast$ denotes the shuffle multiplication. With the above notation, we have $\mathcal S^{\text{\textexclamdown}} \cong \mathcal L^c$. The cobar construction of $\mathcal L^c$ is given explicitly by \[ \mathbb Q\to s^{-1}\mathcal L^c \xrightarrow{\ d \ } \Lambda^2 (s^{-1}\mathcal L^c) \to \cdots \to \Lambda^n(s^{-1}\mathcal L^c) \to \cdots \] with the differential $d$ determined by the Lie cobraket of $\mathcal L^c$. Now the desired properties of $\mathcal M(X)$ follow from this explicit construction. \end{proof}
\begin{remark}
In the special case of a bouquet of odd algebraic spheres where the cohomology of a commutative DGA model is that of a circle or the first Betti number is zero, this was discussed by Baues \cite[Corollary 1.2]{MR0442922} and by Halperin and Stasheff \cite[Theorem 1.5]{MR539532}.
\end{remark}
\section{Main theorem} \label{sec:main}
In this section, we show that if a commutative DGA has cohomology, up to a certain degree, isomorphic to that of a bouquet of odd algebraic spheres, then its minimal model is isomorphic to that of the bouquet of odd algebraic spheres, up to that given degree. Then we apply it to prove that if a commutative DGA has only odd degree cohomology up to a certain degree, then all nonzero cohomology classes up to that degree will never pull back to zero by any algebraic fibration whose fiber has finite cohomological dimension.
Suppose $B$ is a connected commutative DGA of finite type such that $H^{2k}(B)=0$ for all $ 0 < 2k \le 2N$. Let $X_i$ be a basis of $H^{i}(B)$ and $X = \bigcup_{i=1}^{2N+1}X_i$. Let $M = \mathcal M(X)$ be the bouquet of odd algebraic spheres labeled by $X$ from Definition $\ref{def:algsphere}$. Then we have $H^{i}(M) \cong H^i(B)$ for all $0\leq i \leq 2N$. Let $M_k \subset M$ be the subalgebra generated by the generators of degree $\leq k$.
\begin{lemma}Let $k$ be an odd integer. Then $H^{k+2}(M_k)=H^{k+1}(M_k) =0$. \end{lemma} \begin{proof}
$H^{k+1}(M_k)=0$ as $H^{k+1}(M_k) \to H^{k+1}(M)=0$ is injective.
By Proposition $\ref{prop:algsphere}$ above, $M$ has no even-degree generators. In particular, we have $M_k=M_{k+1}$. Moreover,
$H^{\geq 1}(M)$ is spanned by odd-degree generators. From the first observation it
follows that the map $H^{k+2}(M_k) \to H^{k+2}(M)$ is injective, and from the second that its range is $0$.
\end{proof}
It follows that for an odd $k$, we have $M_{k+2} =M_k \otimes \Lambda (V[k+2])$ as an algebra, where the vector space $V = V_1 \oplus V_2 $ is placed at degree $(k+2)$, with $V_1 \cong H^{k+2}(M)$ and $V_2 = H^{k+3}(M_{k})$. The differential can be described as follows. It suffices to define $d \colon V \to M_{k}$. We define $d=0$ on $V_1$. To define $d$ on $V_2$, let us choose a basis $\{a_i\}_{i\in I}$ of $H^{k+3}(M_k)$. Let $\{\tilde a_i\}_{i\in I}$ be the corresponding basis of $V_2$. Then we define $d \tilde a_i =a_i$.
\begin{proposition} \label{prop:constr}
For each odd integer $k\le 2N$, there exists a morphism $\varphi_k \colon M_k \to B$ such that
the induced map on cohomology $H^i(M_k)\cong H^i(M) \to H^i(B)$ is an isomorphism for $i \le k$. \end{proposition} \begin{proof}We construct the maps $\varphi_k$ by induction. By the previous lemma and the fact that $M$ has no even degree generators, it suffices to define $\varphi_k$ for odd integers $k$ . The case where $k=1$ is clear.
Now assume that we have constructed $\varphi_n$, with $n$ an odd integer $\leq 2N-3$. We shall extend $\varphi_n$ to a morphism $\varphi_{n+2}$ on $M_{n+2} = M_n\otimes \Lambda (V[n+2])$, where the vector space $V = V_1 \oplus V_2 $ is placed at degree $(n+2)$, with $V_1 \cong H^{n+2}(M)$ and $V_2 = H^{n+3}(M_{n})$. It suffices to define $\varphi_{n+2}$ on $V$. Let $\{b_j\}_{j\in J}$ be a basis of $H^{n+2}(B)$. Since $H^{n+2}(M)\cong H^{n+2}(B)$, let $\{\tilde b_j\}_{j\in J}$ be the corresponding basis of $V_1$. We define $\varphi_{n+2}$ on $V_1$ by setting $\varphi_{n+2}(\tilde b_j) = b_j$. Similarly, choose a basis $\{c_\lambda\}_{\lambda \in K}$ of $H^{n+3}(M_n)$, and let $\{\tilde c_\lambda\}_{\lambda \in K}$ be the corresponding
basis of $V_2$. Since $H^{n+3}(B) = 0$, for each $c_\lambda \in M_n$, there exists $\theta_\lambda\in B$ such that $\varphi_n(c_\lambda) = d \theta_\lambda$. We define $\varphi_{n+2}$ on $V_2$ by setting $\varphi_{n+2}(\tilde c_\lambda) =\theta_\lambda$.
By construction, the induced map $(\varphi_{n+2})_\ast$ on $H^i$ agrees with $(\varphi_n)_\ast$ for $i\leq n+1$ and $(\varphi_{n+2})_\ast$ is an isomorphism on $H^{2n+2}$. This finishes the proof.
\end{proof}
Now let $\mathcal{M}_B$ be a minimal model of $B$ and $(\mathcal{M}_B)_k$ be the subalgebra generated by the generators of degree $\le k$. Combining the above results, we have proved the following proposition.
\begin{proposition} \label{prop:sw} The commutative DGAs $(\mathcal{M}_B)_{2N-1}$ and $M_{2N-1}$ are isomorphic. \end{proposition}
Moreover, we have the following result, which is an immediate consequence of the construction in Proposition $\ref{prop:constr}$. \begin{corollary}\label{cor:tosphere}
Let $B$ be a connected commutative DGA such that $H^{2i}(B)=0$ for all $ 0 < 2i \le 2N$. Let $\alpha$ be a nonzero class in $H^{2k+1}(\mathcal{M}_B)$ with $2k+1<2N$.
Then there exists a morphism $\psi \colon \mathcal{M}_B \to (\Lambda(\eta), 0)$ such that $\psi_*(\alpha)=[\eta]$, where $\eta$ has degree $2k+1$ and $\Lambda(\eta)$ is the free commutative graded algebra generated by $\eta$. \end{corollary} \begin{proof} From the description of the minimal model $\mathcal M_B$ of $B$, it follows that $\mathcal M_B$ has a set of generators such that
all the cohomology groups up to degree $(2N-1)$ is generated by the cohomology classes of these generators; moreover we can choose these generators
so that the given class $\alpha$ is represented by a generator, say, $a$. Then we define $\psi$ by mapping $a$ to $\eta$ and the other generators to $0$.
\end{proof}
An inductive application of the same argument above proves the following.
\begin{proposition}
Suppose $(C, d)$ is a connected commutative DGA with $H^{2k}(C) = 0 $ for all $2k>0$. Let $X_i$ be a basis of $H^i(C)$ and $X_C = \bigcup_{i=1}^\infty X_i$. Then the bouquet of odd algebraic spheres $\mathcal M(X_C)$ is a minimal model of $(C, d)$. \end{proposition}
Applying the above proposition to the commutative DGA $(TA, d)$ from Theorem $\ref{thm:itoddsphere}$ immediately gives us the following corollary. \begin{corollary}\label{cor:ta}
With the same notation as above, the minimal model of $(TA, d)$ is isomorphic to a bouquet of odd algebraic spheres. \end{corollary}
Before proving the main theorem of this section, we shall prove the following special case first.
\begin{theorem}\label{thm:oddsphere} Let $(\Lambda (x), d)$ be the commutative DGA generated by $x$ of degree $2k+1\geq 1$ such that $d x=0$. For any algebraic fibration $\varphi: (\Lambda(x), d) \to (\Lambda(x)\otimes \Lambda V, d)$ whose algebraic fiber $(\Lambda V, \bar d )$ has finite cohomological dimension, the map $\varphi_\ast: H^j(\Lambda (x)) \to H^j(\Lambda(x)\otimes \Lambda V)$ is injective for all $j$.
\end{theorem} \begin{proof}
The case where $2k+1 = 1$ is trivial. Let us assume $2k+1>1$ in the rest of the proof.
Let $\varphi\colon (\Lambda V, d) \hookrightarrow (\Lambda(x)\otimes\Lambda V, d)$ be any algebraic fibration whose algebraic fiber has finite cohomological dimension. It suffices to show that $\varphi_\ast\colon H^{2k+1}(\Lambda(x)) \to H^{2k+1}(\Lambda(x)\otimes \Lambda V)$ is injective, since the induced map $\varphi_\ast$ on $H^i$ is automatically injective for $i\neq 2k+1$.
Now suppose to the contrary that
\[\varphi_\ast(x) = 0 \textup{ in } H^{2k+1}(\Lambda(x)\otimes \Lambda V). \]
Then we have $ x = d (w\cdot x + v)$
for some $w, v\in \Lambda V$. By inspecting the degrees of the two sides, one sees that $w = 0$. Therefore, we have $x = d v$ for some $v\in \Lambda V$. It follows that $\bar d v = 0$.
Now let $n\in \mathbb N$ be the smallest integer such that $[v^n] = 0$ in $H^{\bullet}(\Lambda V, \bard)$. Such an integer exists since $(\Lambda V, \bard)$ has finite cohomological dimension. Then there exists $u\in \Lambda V$ such that $v^n = \bar d u$. Equivalently, we have
\[ v^n = u_0\cdot x + d u,\]
for some $u_0\in \Lambda V$. It follows that
\[ 0 = d^2 u = d( v^n - u_0\cdot x) = nv^{n-1}\cdot x - (du_0)\cdot x. \]
Therefore, $v^{n-1} = \frac{1}{n} d u_0$, which implies that $[v^{n-1}] = 0$ in $H^{\bullet}(\Lambda V, \bard)$. We arrive at a contradiction. This completes the proof.
\end{proof}
Now let us prove the main result of this section.
\begin{theorem} \label{thm:inj} Let $(B, d)$ be a connected commutative DGA such that $H^{2k}(B)=0$ for all $0 < 2k \leq 2N$. If $\iota \colon (B, d) \to (B\otimes \Lambda V, d) $ is an algebraic fibration whose algebraic fiber has finite cohomological dimension, then the induced map
\[ \iota_\ast \colon \bigoplus_{i< 2N} H^i(B)\to \bigoplus_{i< 2N} H^i(B\otimes \Lambda V)\]
is injective. \end{theorem} \begin{proof}
Let $f\colon (\mathcal M_B, d)\to (B, d)$ be a minimal model algebra of $B$. \begin{claim*}
For any algebraic fibration $\iota \colon (B, d) \to (B\otimes \Lambda V, d) $, there exist an algebraic fibration $\varphi\colon (\mathcal M_B, d) \to (\mathcal M_B\otimes \Lambda V, d)$ and a quasi-isomorphism $g\colon (\mathcal M_B\otimes \Lambda V, d) \to (B\otimes \Lambda V, d)$ such that the following diagram commutes:
\[ \xymatrix{ \mathcal M_B \ar@{^{(}->}[d]_{\varphi} \ar[r]^f & B \ar@{^{(}->}[d]^\iota \\
\mathcal M_B\otimes \Lambda V \ar[r]^-{g} & B\otimes \Lambda V.} \] \end{claim*} We construct $\varphi$ and $g$ inductively. Consider the filtration $V = \cup_{n=0}^\infty V(k)$ from Definition $\ref{def:algfib}$. Choose a basis $\{x_i\}_{i\in I_0}$ of $V(0)$. Let $x = x_i$ be a basis element. If $d x = a \in B$, then $d a = d^2 x = 0$. It follows that there exists $\tilde a \in \mathcal M_B$ such that $f(\tilde a) = a + d c$ for some $c\in B$. We define an algebraic fibration $\varphi_0\colon (\mathcal M_B, d) \hookrightarrow (\mathcal M_B\otimes \Lambda(x), d)$ by setting $d x = \tilde a$. Moreover, we extend $f\colon (\mathcal M_B, d) \to (B, d)$ to a morphism (of commutative DGAs) $g_0\colon (\mathcal M_B\otimes \Lambda (x), d) \to (B\otimes \Lambda (x), d)$ by setting $g(x) = x + c$. By the Gysin sequence from Section $\ref{prop:gysin}$, we see that $g_0$ is a quasi-isomorphism. Now apply the same construction to all basis elements $\{x_i\}_{i\in I_0}$. We still denote the resulting morphisms by $\varphi_0 \colon (\mathcal M_B, d) \to (\mathcal M_B\otimes \Lambda (V(0)), d)$ and $g_0\colon (\mathcal M_B\otimes \Lambda (V(0)), d) \to (B\otimes \Lambda (V(0)), d)$.
Now suppose we have constructed an algebraic fibration \[ \varphi_{k} \colon (\mathcal M_B\otimes \Lambda (V(k-1)), d) \to (\mathcal M_B\otimes \Lambda (V(k)), d)\] and a quasi-isomorphism $g_k\colon (\mathcal M_B\otimes \Lambda (V(k)), d) \to (B\otimes \Lambda (V(k)), d)$ such that the following diagram commutes: \[ \xymatrixcolsep{4pc}\xymatrix{ \mathcal M_B\otimes \Lambda (V(k-1)) \ar@{^{(}->}[d]_{\varphi_k} \ar[r]^{g_{k-1}} & B\otimes \Lambda(V(k-1)) \ar@{^{(}->}[d]^\iota \\
\mathcal M_B\otimes \Lambda (V(k)) \ar[r]^-{g_k} & B\otimes \Lambda (V(k)).} \] Let $\{y_i\}_{i\in I_{k+1}}$ be a basis of $V(k+1)$ that extends the basis $\{x_i\}_{i\in I_{k}}$ of $V(k)\subseteq V(k+1)$. Apply the same construction above to elements in $\{y_i\}_{i\in I_{k+1}} \backslash \{x_i\}_{i\in I_{k}}$, but with $B\otimes \Lambda (V(k))$ in place of $B$, and $\mathcal M_B\otimes \Lambda (V(k))$ in place of $\mathcal M_B$.
We define $(\mathcal M_B\otimes \Lambda V, d)$ to be the direct limit of $(\mathcal M_B\otimes \Lambda (V(k)), d)$ with respect to the morphisms $\varphi_{k} \colon (\mathcal M_B\otimes \Lambda (V(k-1)), d)$. We define $\varphi$ to be the natural inclusion morphism $ (\mathcal M_B, d) \hookrightarrow (\mathcal M_B\otimes \Lambda V, d)$. The morphisms $g_k$ together also induce a quasi-isomorphism $g\colon (\mathcal M_B\otimes \Lambda V, d) \to (B\otimes \Lambda V, d)$, which makes the diagram in the claim commutative. This finishes the proof of the claim.
Now assume to the contrary that there exists $0\neq \alpha\in H^{2k+1}(B)$ with $2k+1 < 2N$ such that $\iota_\ast(\alpha) = 0$. Let $\tilde \alpha \in H^{2k+1}(\mathcal M_B)$ be the class such that $f_\ast(\tilde \alpha) = \alpha$. In particular, we have $\varphi_{\ast}(\tilde \alpha) = 0$. By Corollary $\ref{cor:tosphere}$, there exists a morphism $\psi\colon (\mathcal M_B, d) \to (\Lambda (\eta), 0)$ such that $\psi_{\ast}(\tilde \alpha) = \eta$. Now let \[ \tau\colon (\Lambda(\eta), 0) \to (\Lambda(\eta) \otimes \Lambda V, d) = (\Lambda(\eta) \otimes_{\mathcal M_B} (\mathcal M_B\otimes \Lambda V), d)\] be the push-forward algebraic fibration of $\varphi\colon (\mathcal M_B, d) \to (\mathcal M_B\otimes \Lambda V, d)$. It follows that \[ \tau_\ast(\eta) = \tau_\ast \psi_{\ast}(\tilde \alpha) = (\psi\otimes 1)_\ast\varphi_{\ast}(\tilde \alpha) = 0 \] which contradicts Theorem $\ref{thm:oddsphere}$. This completes the proof.
\end{proof}
\end{document}
|
arXiv
|
{
"id": "1704.04308.tex",
"language_detection_score": 0.7575244903564453,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[The shapes of pure cubic fields]{T\MakeLowercase{he shapes of pure cubic fields}} \subjclass[2010]{11R16, 11R45, 11E12} \keywords{Pure cubic fields, lattices, equidistribution, carefree couples}
\author{R\MakeLowercase{obert} H\MakeLowercase{arron}} \address{ Department of Mathematics\\ Keller Hall\\ University of Hawai`i at M\={a}noa\\ Honolulu, HI 96822\\ USA } \email{[email protected]} \date{\today}
\begin{abstract} We determine the shapes of pure cubic fields and show that they fall into two families based on whether the field is wildly or tamely ramified (of Type I or Type II in the sense of Dedekind). We show that the shapes of Type I fields are rectangular and that they are equidistributed, in a regularized sense, when ordered by discriminant, in the one-dimensional space of all rectangular lattices. We do the same for Type II fields, which are however no longer rectangular. We obtain as a corollary of the determination of these shapes that the shape of a pure cubic field is a complete invariant determining the field within the family of all cubic fields. \end{abstract}
\maketitle
\tableofcontents
\section{Introduction} The shape of a number field is an invariant coming from the geometry of numbers that can be seen as a refinement of the discriminant, or as a positive definite relative of the trace-zero form. Specifically, a degree $n$ number field $K$ can be embedded into its Minkowski space as $j_\mathbf{R}:K\hookrightarrow K\otimes_\mathbf{Q}\mathbf{R}\cong\mathbf{R}^n$ yielding a rank $n$ lattice $j_\mathbf{R}(\mc{O}_K)\subseteq\mathbf{R}^n$, where $\mc{O}_K$ denotes the ring of integers of $K$. The \emph{shape} of $K$ is defined to be the equivalence class of the rank $n-1$ lattice given by the image $\mc{O}_K^\perp$ of $j_\mathbf{R}(\mc{O}_K)$ in $\mathbf{R}^n/\mathbf{R} j_\mathbf{R}(1)$ up to scaling, rotations, and reflections. Explicitly, one obtains a quadratic form on $K$ via \begin{equation}\label{eqn:quadform}
\alpha\longmapsto\!\!\!\sum_{\sigma\in\mathrm{Hom}(K,\mathbf{C})}|\sigma(\alpha)|^2 \end{equation} and the shape can be defined as the $\mathrm{GL}_{n-1}(\mathbf{Z})$-equivalence class, up to scaling by positive real numbers, of the restriction of this form to the image of $\mc{O}_K$ under the projection map $\alpha\mapsto\alpha-\tr_{K/\mathbf{Q}}(\alpha)/n$. The shape of $K$ may equivalently be considered as an element of the double-coset space $\mathrm{GL}_{n-1}(\mathbf{Z})\backslash\mathrm{GL}_{n-1}(\mathbf{R})/\mathrm{GO}_{n-1}(\mathbf{R})$, where $\mathrm{GO}$ denotes the group of orthogonal similitudes. This space, which we denote $\mc{L}_{n-1}$ and call the \emph{space of rank $(n-1)$ lattices}, carries a natural $\mathrm{GL}_{n-1}(\mathbf{R})$-invariant measure.
Little is known about the shapes of number fields. Their study was first taken up in the PhD thesis of David Terr \cite{Terr}, a student of Hendrik Lenstra's. In it, Terr shows that the shapes of both real and complex cubic fields are equidistributed in $\mc{L}_2$ when the fields are ordered by the absolute value of their discriminants. This result has been generalized to $S_4$-quartic fields and to quintic fields by Manjul Bhargava and Piper Harron in \cite{Manjul-Piper, PiperThesis} (where the cubic case is treated, as well). Aside from this, Bhargava and Ari Shnidman \cite{Manjul-Ari} have studied which real cubic fields have a given shape, and in particular which shapes are possible for fields of given quadratic resolvent. Finally, Guillermo Mantilla-Soler and Marina Monsurr\`{o} \cite{M-SM} have determined the shapes of cyclic Galois extensions of $\mathbf{Q}$ of prime degree. Note however that \cite{M-SM} uses a slightly different notion of shape: they instead restrict the quadratic form in \eqref{eqn:quadform} to the space of elements of trace zero in $\mc{O}_K$. It is possible to carry out their work with our definition and our work with theirs; the answers are slightly different as described below.\footnote{Also, when \cite{Manjul-Ari} refers to `shape' in the complex cubic case, they use the trace form instead of \eqref{eqn:quadform}, yielding an indefinite quadratic form, hence the need for the present article.}
This article grew out of the author's desire to explore an observation he made that, given a fixed quadratic resolvent in which $3$ ramifies, the shapes of real $S_3$-cubic fields sort themselves into two sets depending on whether $3$ is tamely or wildly ramified; a \emph{tame versus wild dichotomy}, if you will. Considering complex cubics, the pure cubics---that is, those of the form $\mathbf{Q}(m^{1/3})$---are in many ways the simplest. These were partitioned into two sets by Dedekind, who, in the typical flourish of those days, called them Type I and Type II. In our context, \emph{pure} cubics are exactly those whose quadratic resolvent is $\mathbf{Q}(\omega)$, $\omega$ being a primitive cube root of unity, and Type I (resp.\ Type~II) corresponds to $3$ being wildly (resp.\ tamely) ramified. The first theorem we prove (Theorem~\ref{thm:A}) computes the shape of a given pure cubic field and shows that, just as in the real case, pure cubic shapes exhibit a tame versus wild dichotomy.\footnote{We remark that this dichotomy also appears in \cite{M-SM} and in ongoing work of the author with Melanie Matchett Wood. In the former, it is shown that, for a given prime $\ell\geq5$, only two `shapes' (in the sense of \cite{M-SM}) are possible for cyclic extensions of degree $\ell$, and the `shape' is exactly determined by whether $\ell$ is tame or wild (for $\ell=3$, the two possibilities collapse to one, a result already appearing in \cite{Terr}, and for $\ell=2$, well, there is only one rank one lattice). We however remark that, with our definition of shape, the dichotomy in \cite{M-SM} disappears, while with their definition, \textit{all} pure cubics have rectangular `shape'. It would be very interesting to better understand the subtle difference between the two definitions. In the ongoing project with Wood, the tame versus wild dichotomy is shown to hold for totally real Galois quartic fields.} In the real cubic case, \cite{Manjul-Ari} shows that for fields of a fixed quadratic resolvent, there are only finitely many options for the shape. However, there are infinitely many possibilities for the shape of a pure cubic: indeed, Theorem~\ref{thm:B} shows that there is a bijection between pure cubic fields and their shapes! In fact, this theorem goes further and says that the shape of a pure cubic field uniquely determines it (up to isomorphism) within the collection of all cubic fields. Contrast this with Mantilla-Soler's result that says that complex cubic fields of the same discriminant have isometric integral trace forms \cite[Theorem~3.3]{M-S}.
A natural question to ask is: how are the shapes distributed? The results of \cite{Manjul-Ari}, together with the aforementioned observation, imply that, after splitting the shapes of real $S_3$-cubic fields as needed according to tame versus wild, the shapes are equidistributed amongst the finitely many options. In Theorem~\ref{thm:C}, we show that the shapes of Type~I and Type~II pure cubic fields are equidistributed within their respective one-dimensional family of shapes (viewing $\mc{L}_2$ as a subset of the upper-half plane, and noting that the natural measure on $\mc{L}_2$ is given by the hyperbolic metric on the upper-half plane). Unlike the other equidistribution results described above, the spaces of lattices under consideration have infinite measure. We are thus lead to a ``regularized'' notion of equidistribution taking into account the different growth rates that occur for fields whose shape lies in bounded versus unbounded subsets of the one-dimensional spaces of lattices.
\subsection{Statement of the main theorems} Let $K=\mathbf{Q}(m^{1/3})$ be a pure cubic field and write $m=ab^2$, where $a$ and $b$ are relatively prime, squarefree, positive integers. Note that the cube root of $m^\prime:=a^2b$ also generates $K$ (as $mm^\prime$ is a cube), so we may, and do, assume that $a>b$. We will refer to $r_K:=a/b$ as the \emph{ratio of $K$}. Then, $K$ is Type II if and only if $3\nmid m$ and $r_K\equiv\pm1\mod{9}$.
As described above, the shape of $K$ is a rank 2 lattice, up to orthogonal similitudes. For a lattice of rank $2$, it is common to consider it as living in $\mathbf{C}$ and to take a representative of its equivalence class that has basis $\{1,z\}$ with $z\in\mf{H}$, the upper-half plane. Changing the basis by an element of $\SL(2,\mathbf{Z})$ and rewriting the basis as $\{1,z^\prime\}$ corresponds to the action of $\SL(2,\mathbf{Z})$ by fractional linear transformations on $\mf{H}$. In fact, we allow changing the basis by $\mathrm{GL}(2,\mathbf{Z})$: defining \[
\vect{0&1\\1&0}\cdot z:=1/\ol{z} \] allows us to extend the action of $\SL(2,\mathbf{Z})$ to all of $\mathrm{GL}(2,\mathbf{Z})=\SL(2,\mathbf{Z})\sqcup\vect{0&1\\1&0}\SL(2,\mathbf{Z})$. Thus, the space of rank 2 lattices may be identified (up to some identification on the boundary) with a fundamental domain for the action of $\mathrm{GL}(2,\mathbf{Z})$ on $\mf{H}$ given by \[
\mc{F}:=\{z=x+iy\in\mf{H}:0\leq x\leq1/2, x^2+y^2\geq1\}. \]
Here is our first main theorem. \begin{theoremA}\label{thm:A}\mbox{}
\begin{itemize}
\item Pure cubic fields of Type I have shapes lying on the imaginary axis inside $\mc{F}$; specifically, the shape of $K$ is $ir_K^{1/3}\in\mc{F}$. Thus, their shapes are rectangular.
\item Pure cubic fields of Type II have shapes lying on the line segment $\mathrm{Re}(z)=1/3$, $\mathrm{Im}(z)>1/3$ in $\mf{H}$; specifically, the shape of $K$ is $(1+ir_K^{1/3})/3\in\mf{H}$. The shapes of Type II pure cubics are thus parallelograms with no extra symmetry.
\end{itemize} \end{theoremA} This implies that pure cubic fields do indeed exhibit a tame versus wild dichotomy. The proof of this theorem will be accomplished by explicitly writing down a basis for a lattice representing the shape.
\begin{remark}
For Type II fields, the given line segment does not lie in $\mc{F}$; instead, it lies in the union of $\mc{F}$ with its translates by $\gamma_1=W$, $\gamma_2=SU$, and $\gamma_3=SUSW$, where
\[
W=\vect{0&1\\1&0},\quad S=\vect{0&-1\\1&0},\quad\text{and}\quad U=\vect{1&-1\\0&1}.
\]
These 4 pieces correspond to $r_K^{1/3}$ in the intervals $I_0=(\sqrt{8},\infty)$, $I_1=(\sqrt{5},\sqrt{8})$, $I_2=(\sqrt{2},\sqrt{5})$, and $I_3=(1,\sqrt{2})$, respectively. To describe the location of the shapes within $\mc{F}$, one need simply act on the appropriate subsegments of the line segment by $\gamma_i^{-1}$. One then obtains the union of a line segment and three circular arcs: (the intersection of $\mc{F}$ with) the line $\mathrm{Re}(z)=1/3$ and the circles centred at $3/2$, $-1/2$, and $1/2$, all of radius $3/2$. \end{remark}
An important corollary to this theorem is that the shape provides a complete invariant for pure cubic fields. \begin{theoremA}\label{thm:B}
If the shape of a cubic field is one of the shapes occurring in Theorem~\ref{thm:A}, then the field is the uniquely determined pure cubic field of that shape. \end{theoremA} Once we know Theorem~\ref{thm:A}, the proof of this result is a simple argument involving the field of rationality of the coordinates of the shape in $\mf{H}$. Note that there are plenty of pure cubic fields sharing the same discriminant. For instance, $\mathbf{Q}(6^{1/3})$ and $\mathbf{Q}(12^{1/3})$ both have discriminant $-3^3\cdot 6^2$, but their shapes are $i6^{1/3}$ and $i(\frac{3}{2})^{1/3}$, respectively. This result illustrates how different the shape and the trace form can be. Indeed, Mantilla-Soler shows in \cite[Theorem~3.3]{M-S} that two complex cubic fields have isometric integral trace form if and only if they have the same discriminant. It is natural to ask whether the shape is a complete invariant for complex cubic fields in general.\footnote{The shape fails miserably to distinguish (real) Galois cubic fields: they all have hexagonal shape!} We will present a positive answer to this question in upcoming work using a rationality argument as in the proof of Theorem~\ref{thm:B} below.
As for equidistribution, we must introduce measures on $\mathscr{S}_{\mathrm{I}}:=\{iy:y\geq1\}$ and $\mathscr{S}_{\mathrm{II}}:={\{\frac{1+iy}{3}:y\geq1\}}$. As a natural choice, we pick those induced by the hyperbolic metric on $\mf{H}$, equivalently those invariant under a subgroup of $\SL_2(\mathbf{R})$. Recall the hyperbolic line element on $\mf{H}$ is \[
ds=\frac{\sqrt{dx^2+dy^2}}{y} \] and thus induces the measure $\frac{dy}{y}$ on both $\mathscr{S}_{\mathrm{I}}$ and $\mathscr{S}_{\mathrm{II}}$. Alternatively, note that the diagonal torus $\{\operatorname{diag}(y,y^{-1})\}\subseteq\SL_2(\mathbf{R})$ acting on $i$ yields a homeomorphism from itself to the upper imaginary axis mapping the elements with $y>1$ onto $\mathscr{S}_{\mathrm{I}}$. As such, it induces an invariant measure on $\mathscr{S}_{\mathrm{I}}$, namely $\frac{dy}{y}$.\footnote{To be precise, $\operatorname{diag}(y,y^{-1})$ sends $i$ to $y^2i$, so in terms of the coordinate $iy$, the induced measure is $\frac{dy}{2y}$. We will ignore this $1/2$.} By conjugating this torus, we obtain the same for $\mathscr{S}_{\mathrm{II}}$. We denote the induced measures on $\mathscr{S}_{\textrm{I}}$ and $\mathscr{S}_{\textrm{II}}$ by $\mu_{\textrm{I}}$ and $\mu_{\textrm{II}}$, respectively. A slight complication arises as equidistribution is typically considered for finite measures, whereas on both $\mathscr{S}_{\mathrm{I}}$ and $\mathscr{S}_{\mathrm{II}}$ the measure $\frac{dy}{y}$ is merely $\sigma$-finite. This is in fact reflected in the asymptotics for pure cubic fields! Indeed, the number of pure cubic fields of discriminant bounded by $X$ is on the order of $\sqrt{X}\log(X)$ (see \cite[Theorem~1.1]{Cohen-Morra} or \cite[Theorem~8]{Manjul-Ari}), whereas those of bounded discriminant with shape lying in a finite interval (i.e.\ with $r_K$ in a finite interval) only grow like $\sqrt{X}$ (see Theorem~\ref{thm:fieldasymptotics} below). We thus ``regularize'' our notion of equidistribution and obtain the following result. \begin{theoremA}\label{thm:C} Define \[
C_{\mathrm{I}}=\frac{2C\sqrt{3}}{15}\quad\text{and}\quad C_{\mathrm{II}}=\frac{C\sqrt{3}}{10}, \] where \[
C=\prod_p\left(1-\frac{3}{p^2}+\frac{2}{p^3}\right), \] the product being over all primes $p$. For $?=$ I, resp.~II, and real numbers $1\leq R_1<R_2$, let $[R_1,R_2)_?$ denote the ``interval'' $i[R_1,R_2)$, resp.~$(1+i[R_1,R_2))/3$, in $\mathscr{S}_?$. Then, for all $R_1, R_2$, \[
\lim_{X\rightarrow\infty}\frac{\#\left\{K\text{ of type ?}:|\Delta(K)|\leq X,\operatorname{sh}(K)\in[R_1,R_2)_?\right\}}{C_?\sqrt{X}}=\int_{[R_1,R_2)_?}d\mu_?, \] where $\Delta(K)$ is the discriminant of $K$ and $\operatorname{sh}(K)$ is the shape of $K$ (taken in $\mathscr{S}_?$). \end{theoremA}
\begin{remark}\mbox{}
\begin{enumerate}
\item For the usual, unregularized, notion of equidistribution, the left-hand side above would be
\[
\lim_{X\rightarrow\infty}\frac{\#\left\{K:|\Delta(K)|\leq X,\operatorname{sh}(K)\in[R_1,R_2)_?\right\}}{\#\left\{K:|\Delta(K)|\leq X\right\}},
\]
where the denominator is $C_?\sqrt{X}\log(X)+o(\sqrt{X})$. Given the different growth rates of the numerator and the denominator, this limit is always $0$.
\item We phrase this theorem in a more measure-theoretic way below (see Theorem~\ref{thm:C'}), defining a sequence of measures $\mu_{?,X}$ that converges weakly to $\mu_?$.
\end{enumerate} \end{remark}
\section{Determining the shape}\label{sec:proofofthmA} In this section, we provide integral bases for pure cubic fields (\S\ref{sec:int_bases}) and go on to use these bases to explicitly determine bases of $\mc{O}_K^\perp$ (\S\ref{sec:shape}). These calculations directly prove Theorem~\ref{thm:A}. In \S\ref{sec:thumb}, we prove Theorem~\ref{thm:B}.
\subsection{Some basic facts about pure cubic fields}\label{sec:int_bases} We briefly determine an integral basis and the discriminant of a pure cubic field, and explain how ramification at $3$ is what distinguishes Type~I from Type~II.
Every pure cubic field $K$ corresponds to exactly one pair $(m,m^\prime)$, where $m=ab^2$, $m^\prime=a^2b$, and $a$ and $b$ are two positive, relatively prime, squarefree integers whose product is at least $2$. Let $\alpha$ and $\beta$ be the roots in $K$ of $x^3-m$ and $x^3-m^\prime$, respectively, so that $\beta=\alpha^2/b$, $\alpha=\beta^2/a$, and $K=\mathbf{Q}(\alpha)=\mathbf{Q}(\beta)$. The discriminant of $x^3-m$ is $-3^3a^2b^4$, so that $\Delta(K)\mid\gcd(-3^3a^2b^4, 3^3a^4b^2)=-3^3a^2b^2$. The fact that $x^3-m$ (resp.\ $x^3-m^\prime$) is Eisenstein at all primes dividing $a$ (resp.\ $b$) implies that the index $[\mathcal{O}_K:\mathbf{Z}[\alpha,\beta]]$ is relatively prime to $ab$, and hence divides $3$. Thus, if $3|m$, then the index is $1$ and $\{1,\alpha,\beta\}$ is an integral basis of $\mc{O}_K$. Otherwise, since $[\mathcal{O}_K:\mathbf{Z}[\alpha,\beta]]=[\mathcal{O}_K:\mathbf{Z}[\alpha-m,\beta]]$, we may consider the minimal polynomial $(x+m)^3-m$ of $\alpha-m$. It is Eisenstein at $3$ if and only if $m\not\equiv\pm1\mod{9}$,\footnote{In fact, it is Eisenstein if and only if $m^3\not\equiv m\mod{9}$, but the only cubes modulo $9$ are $0,\pm1$, and $3\nmid m$, by assumption. Note also that the condition $m\equiv\pm1\mod{9}$ is equivalent to $r_K\equiv\pm1\mod{9}$; indeed, $r_K=m/b^3$.} in which case, again, $\{1,\alpha,\beta\}$ is an integral basis of $\mc{O}_K$. For $m\equiv\pm1\mod{9}$, (the ``Type II'' fields) one may check by hand that the element $\nu=(1\pm\alpha+\alpha^2)/3$ is integral. Clearly, $[\mathbf{Z}[\nu,\beta]:\mathbf{Z}[\alpha,\beta]]=3$, so $\{1,\nu,\beta\}$ is then an integral basis of $\mc{O}_K$. This shows that the discriminant of $K$ is $-3^3a^2b^2$ (resp.\ $-3a^2b^2$) for Type I (resp.\ Type II). We summarize these results in the following lemma.
\begin{lemma}\label{lem:intbasis}
With the above notation,
\begin{itemize}
\item the discriminant of a pure cubic of Type I is $\Delta(K)=-3^3a^2b^2$ and $\{1,\alpha,\beta\}$ is an integral basis;
\item the discriminant of a pure cubic of Type II is $\Delta(K)=-3a^2b^2$ and $\{1,\nu,\beta\}$ is an integral basis.
\end{itemize} \end{lemma}
The two possible factorizations of a ramified prime in a cubic field are $\mf{p}_1\mf{p}_2^2$ and $\mf{p}^3$. The prime $3$ being tamely ramified means that $3=\mf{p}_1\mf{p}_2^2$, where $\mf{p}_1$ is unramified and $\mf{p}_2$ is tamely ramified. Consequently, $\mf{p}_2=\mf{p}_2^{2-1}$ exactly divides the different of $K/\mathbf{Q}$. Since $\mf{p}_2$ has norm $3$, this implies $3$ exactly divides $\Delta(K)$. On the other hand, if $3$ is wild, then $3=\mf{p}^3$, in which case $\mf{p}^3$ divides the different. Again, $\mf{p}$ has norm $3$ so $3^3$ divides $\Delta(K)$.
\subsection{The shape}\label{sec:shape} Let $\omega:=e^{2\pi i/3}\in\mathbf{C}$ and let $m^{1/3}$ denote the real cube root of $m$. Let $j_\mathbf{R}:K\rightarrow\mathbf{R}^3$ denote the embedding into the ``Minkowski space'' of $K$ (following Neukirch's normalizations, see \cite[\S{I.5}]{Neukirch}): if $\sigma$ denotes the unique real embedding and $\tau$ denotes the complex embedding sending $\alpha$ to $\omega m^{1/3}$, then \[
j_\mathbf{R}(a):=\big(\sigma(a),\mathrm{Re}(\tau(a)),\mathrm{Im}(\tau(a))\big). \] We also let $j:K\rightarrow\mathbf{C}^3$ be given by $j(a)=(\sigma(a),\tau(a),\ol{\tau}(a))$. It is understood that the space $\mathbf{R}^3$ is equipped with the inner product $\langle\cdot,\!\cdot\rangle$ given by the diagonal matrix $\operatorname{diag}(1,2,2)$. This is the inner product obtained by restricting the standard Hermitian inner product on $\mathbf{C}^3$ to the image of $j$, identified with $\mathbf{R}^3$ as in \cite[\S{I.5}]{Neukirch}. As it is unlikely to cause confusion, we shall abuse notation and write $\alpha$ and $\beta$ to sometimes mean $\sigma(\alpha)=m^{1/3}$ and $\sigma(\beta)=m^{\prime1/3}$ depending on context.
Given a rank 2 lattice $\Lambda$ with Gram matrix\footnote{Recall that a Gram matrix for a lattice $\Lambda$ is the symmetric matrix whose $(i,j)$-entry is the inner product of the $i$th and $j$th basis vectors, for some choice of basis.} $B=(B_{ij})$ the associated point in $\mf{H}$ is $z=x+iy$, where \begin{equation}\label{eqn:xycoords}
x=\frac{B_{0,1}}{B_{0,0}},\quad y=\sqrt{\frac{B_{1,1}}{B_{0,0}}-x^2}. \end{equation}
The starting point is the following proposition which immediately proves Theorem~\ref{thm:A} for Type I fields.
\begin{proposition}\label{prop:shape_of_OK}
For any pure cubic field $K$, the image of the set $\{1,\alpha,\beta\}$ under $j_\mathbf{R}$ is an orthogonal set whose Gram matrix is
\[ \vect{3\\& 3\alpha^2\\&&3\beta^2}.
\] \end{proposition} \begin{proof}
First off,
\[
j(1)=(1,1,1),\quad j(\alpha)=(\alpha,\omega\alpha,\omega^2\alpha),\quad\text{and}\quad j(\beta)=(\beta,\omega^2\beta,\omega\beta).
\]
Thus, the proposition comes down to computing the Hermitian dot products of these vectors. \end{proof}
This yields Theorem~\ref{thm:A} for Type I fields as follows. The set $\{j_\mathbf{R}(\alpha),j_\mathbf{R}(\beta)\}$ is a basis of $\mc{O}_K^\perp$ whose Gram matrix is \[
\vect{3\alpha^2\\&3\beta^2}. \] Using the conversion to $z=x+iy\in\mf{H}$ in Equation~\eqref{eqn:xycoords} gives $x=0$ and $y=(\beta/\alpha)=(a/b)^{1/3}=r_K^{1/3}$.
To study Type II fields, we will need to first know the image of $\nu$ in $\mc{O}_K^\perp$. \begin{lemma}
For $a\in K$, the image $a^\perp$ of $a$ in $\mc{O}_K^\perp$ is
\[
a^\perp=j_\mathbf{R}(a)-\frac{\langle j_\mathbf{R}(a),j_\mathbf{R}(1)\rangle}{3}j_\mathbf{R}(1).
\] \end{lemma} This follows from linear algebra and the simple fact that $\langle j_\mathbf{R}(1),j_\mathbf{R}(1)\rangle=3$.
By Proposition~\ref{prop:shape_of_OK}, $\alpha^\perp=j_\mathbf{R}(\alpha)$ and $\beta^\perp=j_\mathbf{R}(\beta)$. Since $j_\mathbf{R}(\nu)=(j_\mathbf{R}(1)\pm j_\mathbf{R}(\alpha)+bj_\mathbf{R}(\beta))/3$, we get that \begin{equation}
\nu^\perp=j_\mathbf{R}(\nu)-\frac{1}{3}j_\mathbf{R}(1)=\frac{1}{3}(\pm\alpha^\perp+b\beta^\perp). \end{equation}
\begin{lemma}
The Gram matrix of the basis $\{\nu^\perp,\beta^\perp\}$ of $\mc{O}_K^\perp$ is
\[
\vect{\displaystyle\frac{\alpha^2(1+\alpha^2)}{3} & \displaystyle\frac{\alpha^4}{b} \\ \displaystyle\frac{\alpha^4}{b} & \displaystyle\frac{3\alpha^4}{b^2}}.
\] \end{lemma} \begin{proof}
Since $\langle\alpha^\perp,\beta^\perp\rangle=0$, we have
\[
\langle\nu^\perp,\nu^\perp\rangle=\frac{1}{9}(\langle\alpha^\perp,\alpha^\perp\rangle+b^2\langle\beta^\perp,\beta^\perp\rangle)=\frac{1}{9}(3\alpha^2+3b^2\beta^2).
\]
This is the claimed value since $b\beta=\alpha^2$. Furthermore,
\[
\langle\nu^\perp,\beta^\perp\rangle=\frac{b}{3}\langle\beta^\perp,\beta^\perp\rangle=b\beta^2=\alpha^4/b.
\]
Finally, we already know the inner product of $\beta^\perp$ with itself. \end{proof} This basis is not terribly pleasing from the point of view of shapes. Making the following change of basis yields a nicer Gram matrix and proves Theorem~\ref{thm:A} for Type II fields. \begin{lemma}
Let $\epsilon\in\{\pm1\}$ and $k\in\mathbf{Z}$ be such that $b=3k+\epsilon$. Define
\[
\gamma=\vect{3&-b\\1&-k}.
\]
The basis $\{v_1,v_2\}$ of $\mc{O}_K^\perp$ given by
\[
\vect{v_1\\v_2}=\gamma\vect{\nu^\perp\\\beta^\perp}
\]
has the Gram matrix
\[
\vect{3\alpha^2 & \alpha^2 \\ \alpha^2 & \displaystyle\frac{\alpha^2+\beta^2}{3}}.
\]
The associated point in $\mf{H}$ is $\frac{1}{3}+i\frac{r_K^{1/3}}{3}$. \end{lemma} \begin{proof}
We have that
\[
v_1=3\nu^\perp-b\beta^\perp=\pm\alpha^\perp+\alpha^{\perp2}-\alpha^{\perp2}=\pm\alpha^\perp
\]
and
\[
v_2=\nu^\perp-k\beta^\perp=\frac{\pm\alpha^\perp+(3k+\epsilon)\beta^\perp-3k\beta^\perp}{3}=\frac{1}{3}(\pm\alpha^\perp+\epsilon\beta^\perp).
\]
Thus,
\begin{align*}
\langle v_1,v_1\rangle=3\alpha^2,\quad\langle v_1,v_2\rangle=\frac{1}{3}\langle\alpha^\perp,\alpha^\perp\rangle=\alpha^2,\quad\\
\langle v_2,v_2\rangle=\frac{1}{9}(\langle\alpha^\perp,\alpha^\perp\rangle+\langle\beta^\perp,\beta^\perp\rangle)=\frac{\alpha^2+\beta^2}{3}.
\end{align*}
The associated point in $\mf{H}$ has $x=\alpha^2/(3\alpha^2)=1/3$ and
\[
y=\sqrt{\frac{\alpha^2+\beta^2}{9\alpha^2}-\frac{1}{9}}=\frac{\beta}{3\alpha}=\frac{r_K^{1/3}}{3}.
\] \end{proof}
This concludes the proof of Theorem~\ref{thm:A}.
\subsection{Proof of Theorem~\ref{thm:B}}\label{sec:thumb} Let $L$ be any `non-pure' cubic field and let $z_L=x_L+iy_L$ be any point in the upper-half plane corresponding to the lattice $\mc{O}_L^\perp$. By studying the fields of definition of $x_L$ and $y_L$, we will briefly show that $z_L$ cannot be the shape of a pure cubic field. Since Theorem~\ref{thm:A} already shows that non-isomorphic pure cubic fields have different shapes, Theorem~\ref{thm:B} will follow.
First, we note that the key result from Theorem~\ref{thm:A} needed here is that the $y$-coordinate of the shape of a pure cubic field $K$ is \textit{not} rational, rather it lies `properly' in the image of $K$ in $\mathbf{R}$. For this paragraph, we will call such a number `purely cubic'. Now, if $L$ is a real cubic field, then the inner product on $L\otimes_\mathbf{Q}\mathbf{R}$ is simply the trace form and so the Gram matrix of $\mc{O}_L^\perp$ has coefficients in $\mathbf{Q}$. The equations in \eqref{eqn:xycoords} giving $z_L$ in terms of the Gram matrix show that $x_L\in\mathbf{Q}$ and $y_L$ is in some quadratic extension of $\mathbf{Q}$. Thus, $y_L$ is not purely cubic. For a complex cubic field $L$, let $\sigma_L$ denote the unique real embedding of $L$. By definition of the inner product on $L\otimes_\mathbf{Q}\mathbf{R}$ as recalled above, the entries of the Gram matrix of any basis of $\mc{O}_L^\perp$ lie in the Galois closure $N_L$ of $\sigma_L(L)$. The equations in \eqref{eqn:xycoords} then show that $x_L\in N_L$ and that $y_L$ is in a quadratic extension $\wt{N}_L$ of $N_L$. Since cubic extensions have no non-trivial intermediate extensions, for any pure cubic field $K$, we have that $\wt{N}_L\cap K=\mathbf{Q}$. Thus, again, $y_L$ is not purely cubic.
\section{Equidistribution of shapes}\label{sec:equid} We have organized this section beginning with the most conceptual and ending with the most detailed. Specifically, \S\ref{sec:carefree} contains the real mathematical content: counting ``strongly carefree couples'' in a cone below a hyperbola, subject to congruence conditions. In \S\ref{sec:fieldcounting}, these results are translated into counts for pure cubic fields of bounded discriminant with ratio in a given interval. Finally, to begin with, \S\ref{sec:proofThmC} consists mostly of some basic measure theory in order to introduce and prove Theorem~\ref{thm:C} in a conceptually nicer way.
\subsection{Proof of Theorem~\ref{thm:C}}\label{sec:proofThmC} For $?=$ I or II, let $C_c(\mathscr{S}_?)$ denote the space of continuous functions on $\mathscr{S}_?$ with compact support, and for a positive integer $X$, define a positive linear functional $\varphi_{?,X}$ on $C_c(\mathscr{S}_?)$ by \[
\varphi_{?,X}(f)=\frac{1}{C_?\sqrt{X}}\sum_{\substack{K\text{ of type ?}\\ |\Delta(K)|\leq X}}f(\operatorname{sh}(K)), \] where the shape of $K$ is taken in $\mathscr{S}_?$. By the Riesz representation theorem, each $\varphi_{?,X}$ corresponds to a (regular Radon) measure $\mu_{?,X}$ on $\mathscr{S}_?$. Note that since these measures are finite sums of point measures, we have that \[
\mu_{?,X}([R_1,R_2)_?)=\frac{\#\left\{K\text{ of type ?}:|\Delta(K)|\leq X,\operatorname{sh}(K)\in[R_1,R_2)_?\right\}}{C_?\sqrt{X}}, \] for all $1\leq R_1<R_2$. \begin{theorem}\label{thm:C'} For $?=$ I or II, the sequence $\mu_{?,X}$ converges weakly to $\mu_?$. That is, for all $f\in C_c(\mathscr{S}_?)$, \[
\lim_{X\rightarrow\infty}\int fd\mu_{?,X}=\int fd\mu_?. \] \end{theorem} \begin{proof}
The first step will be to prove the statement given in Theorem~\ref{thm:C} in the introduction, i.e.\ we must obtain asymptotics for the number of pure cubic fields of bounded discriminant and ratio in a given interval. This is the subject of Theorem~\ref{thm:fieldasymptotics} below. We obtain, in the notation of that theorem,
\[
\mu_{?,X}([R_1,R_2)_?)=\frac{\mc{N}_?(X,R_1^3,R_2^3)}{C_?\sqrt{X}}=\log\left(\frac{R_2}{R_1}\right)+o(1).
\]
Thus, $\mu_{?,X}([R_1,R_2)_?)\rightarrow\mu_?([R_1,R_2)_?)$ as desired. Passing from this result to the full result is a standard argument. Indeed, let $f\in C_c(\mathscr{S}_?)$ be given and let $\epsilon>0$ be arbitrary. Then, there exists a pair $(f_1,f_2)$ of step functions (i.e.\ finite linear combinations of characteristic functions of intervals of the form $[R_1,R_2)_?$ above) with $f_1\leq f\leq f_2$ and $\int(f_2-f_1)d\mu_?<\epsilon$. Since we already know the theorem for such step functions, we obtain that
\begin{align*}
\int fd\mu_?-\epsilon&\leq\int f_1d\mu_?=\lim_{X\rightarrow\infty}\int f_1d\mu_{?,X}\\
&\leq\liminf_{X\rightarrow\infty}\int fd\mu_{?,X}\\
&\leq\limsup_{X\rightarrow\infty}\int fd\mu_{?,X}\\
&\leq\lim_{X\rightarrow\infty}\int f_2d\mu_{?,X}=\int f_2d\mu_?\\
&\leq\int fd\mu_?+\epsilon.
\end{align*}
As $\epsilon$ is arbitrary, the result follows. \end{proof}
\subsection{Counting pure cubic fields of bounded discriminant and ratio}\label{sec:fieldcounting} For real numbers $X, R_1,R_2\geq1$ with $R_1<R_2$, let $\mc{N}_?(X,R_1,R_2)$ denote the number of pure cubic fields (up to isomorphism) of type $?=$~I or II, of discriminant less than $X$ (in absolute value), and ratio in the interval $(R_1,R_2)$. When we drop the subscript, we count Type I and II together.
\begin{theorem}\label{thm:fieldasymptotics}
For fixed $R_1$ and $R_2$, we have the following asymptotics:
\begin{align}
\mc{N}_{\mathrm{I}}(X,R_1,R_2)&=\frac{2C}{15\sqrt{3}}\sqrt{X}\log\frac{R_2}{R_1}+o(\sqrt{X}),\\
\mc{N}_{\mathrm{II}}(X,R_1,R_2)&=\frac{C}{10\sqrt{3}}\sqrt{X}\log\frac{R_2}{R_1}+o(\sqrt{X}), \end{align} \begin{align}
\mc{N}(X,R_1,R_2)&=\frac{7C}{30\sqrt{3}}\sqrt{X}\log\frac{R_2}{R_1}+o(\sqrt{X}),
\end{align}
where $C$ is as in Theorem~\ref{thm:C}. \end{theorem}
\begin{proof}[Proof of Theorem~\ref{thm:fieldasymptotics}]
For Type~I fields, the discriminant is $-3^3a^2b^2$, so, in the notation of Theorem~\ref{prop:TypeIIabcount} below, we take $N=\frac{1}{3}\sqrt{X/3}$ and get that
\[
\mc{N}_{\mathrm{I}}(X,R_1,R_2)=\frac{1}{2}\left(\mc{S}_{\mathrm{I}}\left(\frac{1}{3}\sqrt{X/3}, R_2\right)-\mc{S}_{\mathrm{II}}\left(\frac{1}{3}\sqrt{X/3}, R_1\right)\right).
\]
The desired result follows from Theorem~\ref{prop:TypeIIabcount}.
Similarly, for Type~II fields, the discriminant is $-3a^2b^2$, so $N=\sqrt{X/3}$, and we have that
\[
\mc{N}_{\mathrm{II}}(X,R_1,R_2)=\frac{1}{2}\left(\mc{S}_{\mathrm{II}}(\sqrt{X/3}, R_2)-\mc{S}_{\mathrm{II}}(\sqrt{X/3}, R_1)\right),
\]
which gives the desired result by Theorem~\ref{prop:TypeIIabcount}. \end{proof}
The next section contains all the counting results used in the above proof.
\subsection{Counting strongly carefree couples in a hyperbolic pie slice, with congruence conditions}\label{sec:carefree} A pair $(a,b)$ of positive integers is called a \emph{strongly carefree couple} if $a$ and $b$ are relatively prime and squarefree. In \cite{Carefree}, Pieter Moree counts the number of strongly carefree couples in a box of side $N$ obtaining the asymptotic $CN^2+O(N^{3/2})$, where $C$ is as in Theorem~\ref{thm:C}. Counting pure cubic fields of bounded discriminant and ratio amounts to counting strongly carefree couples below the hyperbola $ab=N$ and within the cone $R^{-1}\leq a/b\leq R$, while counting only those of Type~II imposes a congruence condition modulo 9. In this section, we determine asymptotics for these counts following the methods of \cite{Carefree} and \cite[\S6.2]{Manjul-Ari}.
For $N,R\geq1$, let \[
\mc{S}(N,R)=\#\left\{(a,b)\in\mathbf{Z}_{\geq1}^2:(a,b)\text{ strongly carefree}, ab\leq N, \frac{1}{R}\leq\frac{a}{b}\leq R\right\}, \] \[
\mc{S}_{\mathrm{II}}(N,R)=\#\left\{(a,b)\in\mc{S}(N,R):3\nmid ab, a^2\equiv b^2\mod{9}\right\}, \] and \[
\mc{S}_{\mathrm{I}}(N,R)=\mc{S}(N,R)-\mc{S}_{\mathrm{II}}(N,R). \]
\begin{theorem}\label{prop:TypeIIabcount}
For fixed $R$, we have
\begin{align}
\mc{S}_{\mathrm{I}}(N,R)&=\frac{4C}{5}N\log R+o(N),\\
\mc{S}_{\mathrm{II}}(N,R)&=\frac{C}{5}N\log R+o(N),\\
\mc{S}(N,R)&=CN\log R+o(N).
\end{align} \end{theorem} \begin{proof}
For the two latter counts, we cover the hyperbolic pie slice with the following four pieces:
\begin{itemize}
\item[(i)] $1\leq a\leq\sqrt{RN}, 1\leq b\leq\frac{N}{a}$,
\item[(ii)] $1\leq a\leq\sqrt{\frac{N}{R}}, 1\leq b\leq\frac{N}{a}$,
\item[(iii)] $1\leq a\leq\sqrt{\frac{N}{R}}, 1\leq b\leq Ra$,
\item[(iv)] $1\leq a\leq\sqrt{RN}, 1\leq b\leq\frac{1}{R}a$.
\end{itemize}
Letting $\mc{S}^?(N,R)$ denote the number of points with condition $?\in\{\text{(i), (ii), (iii), (iv)}\}$ imposed, and similarly for $\mc{S}^?_{\mathrm{II}}(N,R)$, we have that
\[
\mc{S}_?(N,R)=\mc{S}_?^{(i)}(N,R)-\mc{S}_?^{(ii)}(N,R)+\mc{S}_?^{(iii)}(N,R)-\mc{S}_?^{(iv)}(N,R),
\]
for $?=\mathrm{II}$ or nothing. For $\mc{S}(N,R)$, we will apply Lemmas~\ref{lem:regionsi-iii} and \ref{lem:regionsii-iv} with $n=1$, whereas for $\mc{S}_{\mathrm{II}}(N,R)$ we take $n=9$ and sum over two $\psi$s, one sending $a$ to $a\mod{9}$, the other to $-a\mod{9}$. Note that all but one of the terms in Lemma~\ref{lem:regionsi-iii} cancel in $\mc{S}_?^{(i)}(N,R)-\mc{S}_?^{(ii)}(N,R)$ since they are independent of $R$. Applying Lemma~\ref{lem:regionsi-iii} with $\rho=R$, then $\rho=1/R$ and noting that Lemma~\ref{lem:regionsii-iv} implies the contribution of $\mc{S}_?^{(iii)}(N,R)-\mc{S}_?^{(iv)}(N,R)$ is negligible, we obtain the stated results. \end{proof}
The point counting in regions of the form (i)--(iv) above are contained in the following two lemmas. \begin{lemma}\label{lem:regionsi-iii} Let $n$ be a positive integer, let $\psi$ be any function from the positive integers prime to $n$ to $(\mathbf{Z}/n\mathbf{Z})^\times$, and let $\rho$ be a positive real number. Then, \begin{align*}
\sum_{\substack{a\leq\sqrt{\rho N}\\ (a,n)=1}}\mu^2(a)\!\!\!\!\!\sum_{\substack{b\leq N/a\\ (a,b)=1\\ b\equiv\psi(a)\mod{n}}}\!\!\!\!\!\mu^2(b)&=\frac{CN}{n}\prod_{p\mid n}\left(\frac{p^2}{(p-1)(p+2)}\right)\\
&\cdot\left(\frac{\log N}{2}+\frac{\log \rho}{2}+\gamma+3\kappa+\sum_{p\mid n}\frac{\log p}{p+2}\right)+O(N^{3/4+\epsilon}) \end{align*} for all $\epsilon>0$,where $\gamma$ is the Euler--Mascheroni constant, $C$ is as in Theorem~\ref{thm:C}, and \[
\kappa=\sum_p\frac{\log(p)}{p^2+p-2}. \] \end{lemma} \begin{proof}
Applying Lemma~\ref{lem:bsum} below to the inner sum with $a^\prime=\psi(a)$ yields
\[
\sum_{\substack{b\leq N/a\\ (a,b)=1\\ b\equiv\psi(a)\mod{n}}}\mu^2(b)=\frac{\varphi(a)N}{a^2n}\cdot\frac{1}{\zeta(2)}\cdot\prod_{p\mid an}\frac{p^2}{p^2-1}+O(2^{\omega(a)}\sqrt{N/a}).
\]
Estimating the error using that\footnote{Indeed, $2^{\omega(m)}\leq d(m)$ and $d(m)=o(m^{\epsilon})$ for all $\epsilon>0$ (see e.g. \cite[Exercise~13.13]{Apostol}).}
\[
\sum_{m\leq x}\mu^2(m)\frac{2^{\omega(m)}}{\sqrt{m}}=O(x^{1/2+\epsilon}),\text{ for all }\epsilon>0,
\]
we obtain, for all $\epsilon>0$, and up to $O(N^{3/4+\epsilon})$,
\begin{align*}
\sum_{\substack{a\leq\sqrt{\rho N}\\ (a,n)=1}}\mu^2(a)\!\!\!\!\!\sum_{\substack{b\leq N/a\\ (a,b)=1\\ b\equiv\psi(a)\mod{n}}}\!\!\!\!\!\mu^2(b)&=\frac{N}{n\zeta(2)}\prod_{p\mid n}\frac{p^2}{p^2-1}\sum_{\substack{a\leq\sqrt{\rho N}\\ (a,n)=1}}\mu^2(a)\frac{\varphi(a)}{a^2}\prod_{p\mid a}\frac{p^2}{p^2-1}\\
&=\frac{N}{n\zeta(2)}\prod_{p\mid n}\frac{p^2}{p^2-1}\sum_{\substack{a\leq\sqrt{\rho N}\\ (a,n)=1}}\mu^2(a)\prod_{p\mid a}\frac{1}{p+1}.
\end{align*}
Applying Perron's formula as in Lemma~\ref{lem:perron} below with $k=0$ and $x=\sqrt{\rho N}$ yields the desired result. \end{proof}
\begin{lemma}\label{lem:regionsii-iv}
Let $n,\psi$, and $\rho$ be as in Lemma~\ref{lem:regionsi-iii}. Then.
\[
\sum_{\substack{a\leq\sqrt{\rho N}\\ (a,n)=1}}\mu^2(a)\sum_{\substack{b\leq a/\rho\\ (a,b)=1\\ b\equiv\psi(a)\mod{n}}}\mu^2(b)=O(N^{3/4+\epsilon}),\text{ for all }\epsilon>0.
\] \end{lemma} \begin{proof}
This proof proceeds along the same lines as the previous one. Lemma~\ref{lem:bsum} gives
\[
\sum_{\substack{b\leq a/\rho\\ (a,b)=1\\ b\equiv\psi(a)\mod{n}}}\mu^2(b)=\frac{\varphi(a)}{\rho an}\cdot\frac{1}{\zeta(2)}\cdot\prod_{p\mid an}\frac{p^2}{p^2-1}+O(2^{\omega(a)}\sqrt{a/\rho}).
\]
Again using that $d(m)=o(m^\epsilon)$, for all $\epsilon>0$, we see that\footnote{Here, we simply replace the $\sqrt{a}$ in the error term above with $N^{1/4}$.}
\[
\sum_{m\leq x}2^{\omega(m)}=O(x^{1+\epsilon}),\text{ for all }\epsilon>0.
\]
This gives, for all $\epsilon>0$, and up to $O(N^{3/4+\epsilon})$,
\begin{align*}
\sum_{\substack{a\leq\sqrt{\rho N}\\ (a,n)=1}}\mu^2(a)\!\!\!\!\!\sum_{\substack{b\leq a/\rho\\ (a,b)=1\\ b\equiv\psi(a)\mod{n}}}\!\!\!\!\!\mu^2(b)&=\frac{1}{\rho n\zeta(2)}\prod_{p\mid n}\frac{p^2}{p^2-1}\sum_{\substack{a\leq\sqrt{\rho N}\\ (a,n)=1}}\mu^2(a)\frac{\varphi(a)}{a}\prod_{p\mid a}\frac{p^2}{p^2-1}\\
&=\frac{1}{\rho n\zeta(2)}\prod_{p\mid n}\frac{p^2}{p^2-1}\sum_{\substack{a\leq\sqrt{\rho N}\\ (a,n)=1}}\mu^2(a)\prod_{p\mid a}\frac{p}{p+1}.
\end{align*}
Since all the terms in the sum are less than 1, the sum itself is $O(\sqrt{N})$, and we are done. \end{proof}
In order to obtain the counts above with congruence conditions, we need the following improvements to \cite[\S2]{Carefree}. \begin{lemma}
Fix $n\in\mathbf{Z}\geq1$. For $a$ and $a^\prime$ relatively prime to $n$, let
\[
T_{a,a^\prime,n}(x)=\#\{b\leq x:(a,b)=1, b\equiv a^\prime\mod{n}\}.
\]
Then
\[
T_{a,a^\prime,n}(x)=\frac{\varphi(a)x}{an}+O(2^{\omega(a)}),
\]
where $\varphi$ is the Euler totient function and $\omega(a)$ is the number of distinct prime divisors of $a$. \end{lemma} Note that the right hand side is independent of $a^\prime$. \begin{proof} Using the identity \[
\sum_{d|m}\mu(d)=\begin{cases}
1&\text{if }m=1\\
0&\text{otherwise}
\end{cases} \] and the orthogonality of Dirichlet characters in the form \[
\frac{1}{\varphi(n)}\sum_{\chi\text{ mod }n}\chi(m)\ol{\chi}(m^\prime)=\begin{cases}
1&\text{if }m\equiv m^\prime\mod{n}\\
0&\text{otherwise}
\end{cases} \] (where the sum is over all Dirichlet characters modulo $n$), we obtain \begin{align*}
T_{a,a^\prime,n}(x)&=\sum_{\substack{b\leq x\\(a,b)=1\\b\equiv a^\prime\mod{n}}}1\\
&=\sum_{\substack{b\leq x\\b\equiv a^\prime\mod{n}}}\sum_{d\mid\gcd(a,b)}\mu(d)\\
&=\sum_{b\leq x}\left(\sum_{d\mid\gcd(a,b)}\mu(d)\frac{1}{\varphi(n)}\sum_{\chi\text{ mod }n}\chi(a^\prime)\ol{\chi}(b)\right)\\
&=\frac{1}{\varphi(n)}\sum_{d\mid a}\mu(d)\left(\sum_{c\text{ mod }n}\sum_{\chi\text{ mod }n}\chi(a^\prime)\ol{\chi}(c)\right)\\
&\phantom{=}\cdot\#\{\text{multiples }b\text{ of }d:b\leq x, b\equiv c\mod{n}\}\\
&=\sum_{d\mid a}\mu(d)\left(\frac{x}{dn}+O(1)\right)\\
&=\frac{\varphi(a)x}{an}+O(2^{\omega(a)}), \end{align*} where we've used that \[
\sum_{d\mid m}\mu(d)=\frac{\varphi(m)}{m}. \] \end{proof}
\begin{lemma}\label{lem:bsum}
Fix $n\in\mathbf{Z}\geq1$. For $a$ and $a^\prime$ relatively prime to $n$, let
\[
S_{a,a^\prime,n}(x)=\#\{b\leq x:(a,b)=1, b\equiv a^\prime\mod{n},b\text{ squarefree}\}.
\]
Then
\[
S_{a,a^\prime,n}(x)=\frac{\varphi(a)x}{an}\cdot\frac{1}{\zeta(2)}\cdot\prod_{p\mid an}\frac{p^2}{p^2-1}+O(2^{\omega(a)}\sqrt{x}),
\]
where $\zeta(2)=\pi^2/6$. \end{lemma} Once again, note that the right hand side is independent of $a^\prime$. \begin{proof}
By the inclusion-exclusion principle, \begin{align*}
S_{a,a^\prime,n}(x)&=\sum_{\substack{m\leq\sqrt{x}\\(m,an)=1}}\mu(m)T_{a,a^\prime m^{-2},n}(x/m^2)\\
&=\frac{\varphi(a)x}{an}\left(\sum_{\substack{m\leq\sqrt{x}\\(m,an)=1}}\frac{\mu(m)}{m^2}\right)+O(2^{\omega(a)}\sqrt{x})
\end{align*} \begin{align*}
&=\frac{\varphi(a)x}{an}\left(\sum^\infty_{\substack{m=1\\(m,an)=1}}\frac{\mu(m)}{m^2}\right)+O(2^{\omega(a)}\sqrt{x})\\
&=\frac{\varphi(a)x}{an}\cdot\frac{1}{\zeta^{(an)}(2)}+O(2^{\omega(a)}\sqrt{x})\\
&=\frac{\varphi(a)x}{an}\cdot\frac{1}{\zeta(2)}\cdot\prod_{p\mid an}\frac{p^2}{p^2-1}+O(2^{\omega(a)}\sqrt{x}). \end{align*} \end{proof}
Finally, our application of Perron's formula is dealt with in the following lemma. This approach appears in \cite[\S6.2]{Manjul-Ari}, though we include further details and a slightly more general result. \begin{lemma}\label{lem:perron}
Let $k$ be a nonnegative integer and let $n$ be a positive integer. Let
\[
A_k(a)=\mu^2(a)\prod_{p\mid a}\frac{p^k}{p+1}\quad\text{and}\quad f_k(s)=\sum_{\substack{a\geq1\\(a,n)=1}}\frac{A_k(a)}{a^s}.
\]
Then,
\[
\sum_{\substack{a\leq x\\(a,n)=1}}A_k(a)=\zeta(2)C\prod_{p\mid n}\left(1+\frac{1}{p+1}\right)^{\!\!-1}\cdot\begin{cases}
\displaystyle\log(x)+\gamma+3\kappa+\sum_{p\mid n}\frac{\log p}{p+2}&\text{if }k=0,\\
\displaystyle\frac{x^k}{k}&\text{if }k>0,
\end{cases}
\]
where $C$ is as in Theorem~\ref{thm:C} and $\kappa$ is as in Lemma~\ref{lem:regionsi-iii}. \end{lemma} \begin{proof}
Note that $A_k(a)=a^kA_0(k)$ so that $f_k(s)=f_0(s-k)$. Since $A_0(a)\leq 1/a$, $f_0(s)$ converges absolutely for $\mathrm{Re}(s)>0$. Perron's formula (see e.g.\ \cite[Theorem~11.18]{Apostol}) then states that for any $\epsilon>0$ and any $k\geq0$
\[
\sum_{a\leq x}\!{\vphantom{\sum}}^\ast A_k(a)=\frac{1}{2\pi i}\int_{k+\epsilon-i\infty}^{k+\epsilon+i\infty}f_k(s)\frac{x^s}{s}ds=\frac{1}{2\pi i}\int_{k+\epsilon-i\infty}^{k+\epsilon+i\infty}f_0(s-k)\frac{x^s}{s}ds,
\]
where the asterisk on the sum means that if $x$ is an integer, then the last term must be divided by 2. Following \cite[\S6.2]{Manjul-Ari}, we let $h(s)=f_0(s)/\zeta(s+1)$ and we compute
\[
\frac{1}{2\pi i}\int_{k+\epsilon-i\infty}^{k+\epsilon+i\infty}h(s-k)\zeta(s-k+1)\frac{x^s}{s}ds
\]
by shifting the integral just to the left of $\mathrm{Re}(s)=k$ picking up a pole of $\zeta(s)$ at $s=1$ and, when $k=0$, also the pole of $1/s$ at $s=0$. Since $h(s)$ converges for $\mathrm{Re}(s)>-1/2$, we may use Laurent series to determine the residues that occur. For $k>0$, near $s=k$
\[
h(s-k)\zeta(s-k+1)\frac{x^s}{s}=(h(0)+\cdots)\cdot(\frac{1}{s-k}+\cdots)\cdot(x^k+\cdots)\cdot(1/k+\cdots),
\]
so the residue is simply $h(0)x^k/k$. For $k=0$, near $s=0$
\[
h(s)\zeta(s+1)\frac{x^s}{s}=(h(0)+h^\prime(0)s+\cdots)\cdot(\frac{1}{s}+\gamma+\cdots)\cdot(1+s\log x+\cdots)\cdot1/s,
\]
so the residue is $h(0)\gamma+h(0)\log(x)+h^\prime(0)$.
In order to evaluate $h(0)$, we consider $h(s)/\zeta(s+2)=f(s)/(\zeta(s+1)\zeta(s+2))$ at $s=0$, where its Euler product converges. Note that
\[
\left(1+\frac{1}{p+1}\right)(1-p^{-1})(1-p^{-2})=1-\frac{3}{p^2}+\frac{2}{p^3}
\]
so
\[
\frac{h(0)}{\zeta(2)}=C\prod_{p\mid n}\left(1+\frac{1}{p+1}\right)^{-1}.
\]
In order to evaluate $h^\prime(0)$, we instead evaluate $h^\prime(0)/h(0)$. One may verify that
\[
h(s)=\prod_{p\mid n}\left(1+\frac{1}{p+1}p^{-s}\right)^{-1}\prod_p\left(1-\frac{1}{p(p+1)}(p^{-s}+p^{-2s})\right).
\]
Taking the logarithm, then the derivative, and evaluating at $s=0$ yields
\[
\frac{h^\prime(0)}{h(0)}=3\kappa+\sum_{p\mid n}\frac{\log(p)}{p+2}.
\] \end{proof}
\subsection*{Acknowledgments} The author would like to thank Piper Harron, Ari Shnidman, and Rufus Willett for some helpful conversations, as well as the referee for suggesting some points that should be clarified and asking some good questions.
\end{document}
|
arXiv
|
{
"id": "1509.01627.tex",
"language_detection_score": 0.6895366311073303,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\newcommand{\RN}[1]{
\textup{\uppercase\expandafter{\romannumeral#1}} }
\title{Cascaded High Dimensional Histograms: \\A Generative Approach to Density Estimation}
\author{\name Siong Thye Goh \email [email protected] \\
\AND
\name Cynthia Rudin \email [email protected] \\
\addr Massachusetts Institute of Technology \\ Cambridge, MA02139, USA.}
\editor{}
\maketitle
\begin{abstract} We present tree- and list- structured density estimation methods for high dimensional binary/categorical data. Our density estimation models are high dimensional analogies to variable bin width histograms. In each leaf of the tree (or list), the density is constant, similar to the flat density within the bin of a histogram. Histograms, however, cannot easily be visualized in higher dimensions, whereas our models can. The accuracy of histograms fades as dimensions increase, whereas our models have priors that help with generalization. Our models are sparse, unlike high-dimensional histograms. We present three generative models, where the first one allows the user to specify the number of desired leaves in the tree within a Bayesian prior. The second model allows the user to specify the desired number of branches within the prior. The third model returns lists (rather than trees) and allows the user to specify the desired number of rules and the length of rules within the prior. Our results indicate that the new approaches yield a better balance between sparsity and accuracy of density estimates than other methods for this task. \end{abstract}
\begin{keywords}
Density Estimation, Decision Trees, Histogram, Interpretable Modeling. \end{keywords}
\section{Introduction}
A histogram is a piecewise constant density estimation model. There are good reasons that the histogram is among the first techniques taught to any student dealing with data \citep{chakrabarti2006data}: (i) histograms are easy to visualize, (ii) they are accurate as long as there are enough data in each bin, and (iii) they have a logical structure that most people find interpretable. A downside of the conventional histogram is that all of these properties fail in high dimensions, particularly for binary or categorical data. One cannot easily visualize a conventional high dimensional histogram. For binary data this would require us to visualize a high dimensional hypercube. In terms of accuracy, there may not be enough data in each bin, so the estimates would cease to be accurate. In terms of interpretability, for a high dimensional histogram, a large set of logical conditions ceases to be an interpretable representation of the data, and can easily obscure important relationships between variables. Considering marginals is often useless for binary variables, since there are only two bins (0 and 1). The question is how to construct a piecewise constant density estimation model (like a histogram) that has the three properties mentioned above: (i) it can be visualized, (ii) it is accurate, (iii) it is interpretable.
In this paper we present three cascaded (tree- or list- structured) density estimation models. These are similar to variable bin-width histograms, (e.g., see \cite{wand1997data,scott1979optimal}), though our approaches use only a subset of the variables. A leaf (that is, a histogram bin) is defined by conditions on a subset of variables (e.g. ``the second component of $x$ is 0" and ``the first component of $x$ is 1"), and the density is estimated to be constant with each leaf.
Let us give an example to illustrate how each bin is modeled to be of constant density. Let us say we are modeling the population of burglaries (housebreaks) in a city. We might want to create a density model to understand how common or unusual the particular details of a crime might be (e.g., do we see crimes like this every month, or is this relatively uncommon?). A leaf (histogram bin) in our model might be the following: if \textit{premise} is \texttt{residence}, \textit{owner present} is \texttt{false}, \textit{location of entry} is \texttt{window}, then $p(\textrm{state})$ is 0.20. This means that the total density in the bin where these conditions hold is 0.20, that is, for 20\% of burglaries, the three conditions are met. Let us say we have an additional variable \textit{means of entry} with outcomes \texttt{pried}, \texttt{forced}, and \texttt{unlocked}, indicating how the criminal entered the premise. Each of these outcomes would be equally probably in the leaf, each with probability 0.20/3=.067. We described just one bin above, whereas a full tree could be that of Figure \ref{FigFarmSlide1}.
\begin{figure}
\caption{A sparse tree to represent the grain data set. Probability of belonging to the leaf, the densities ($f$) and volume (Vol) are specified in the sparse tree.}
\label{FigFarmSlide1}
\end{figure}
Bayesian priors control the shape of the tree. This helps with both generalization and interpretability. For the first method, the prior parameter controls the number of leaves. For the second method, the prior controls the desired number of branches for nodes of the tree. For the third method, which creates lists (one-sided trees), the prior controls the desired number of leaves and also the length of the leaf descriptions.
This generative structure aims to fix the three issues with conventional histograms: (i) visualization: we need only write down the conditions we used in the tree- or list- shaped cascade to visualize the model.
(ii) accuracy: the prior encourages the cascade to be smaller, which means the bins are larger, and generalize better. (iii) interpretability: the prior encourages sparsity, and encourages the cascade to obey a user-defined notion of interpretability.
Density estimation is a classic topic in statistics and machine learning.
Without using domain-specific generative assumptions, the most useful techniques have been nonparametric, mainly variants of kernel density estimation (KDE) \citep{akaike1954approximation,rosenblatt1956remarks,parzen1962estimation,cacoullos1966estimation,mahapatruni2011cake,nadaraya1970remarks,rejto1973density,wasserman2006all,silverman1986density,devroye1991exponential}. KDE is highly tunable, not domain dependent, and can generalize well, but does not have the interpretable logical structure of histograms. Similar alternatives include mixtures of Gaussians \citep{Li99mixturedensity,zhuang1996gaussian,ormoneit1995improved,ormoneit1998averaging,chen2006probability,seidl2009indexing}, forest density estimation \citep{Liu:2011:FDE:1953048.2021032}, RODEO \citep{AISTATS07_LiuLW} and other nonparametric Bayesian methods \citep{muller2004nonparametric}
which have been proposed for general purpose (not interpretable per se) density estimation. \citep{jebara2012bayesian} provides a Bayesian treatment of latent directed graph structure for non-iid data, but does not focus on sparsity. P\'olya trees are generated probabilistically for real valued features and could be used as priors \citep{wong2010optional}. The most similar paper to ours is on density estimation trees (DET) \citep{Ram:2011:DET:2020408.2020507}. DETs are constructed in a top-down greedy way. This gives them a disadvantage in optimization, often leading to lower quality trees. They also do not have a generative interpretation, and their parameters do not have a physical meaning in terms of the shape of the trees (unlike the methods defined in this work).
\section{Models}
For all the three models, we will need the following notation. There are $p$ features. We express the path to a leaf as the set of conditions on each feature along the path. For instance, for a particular leaf (leaf $t$ in Figure \ref{MyTree}), we might see conditions that require the first feature $x_{.1} \in \left\{ 4,5,6\right\}$ and the second feature $x_{.2} \in \left\{ 100,101\right\}$. Thus the leaf is defined by the set of all outcomes that obey these conditions, that is, the leaf could be $$x \in\left\{ x_{.1} \in \left\{4,5,6 \right\}, x_{.2} \in \left\{ 100, 101\right\}, x_{3.}, x_{.4}, \ldots, x_{.p} \textrm{ are any allowed values}\right\}.$$ This implies there is no restriction on $ x_{3.}, x_{.4}, \ldots, x_{.p}$ for observations within the leaf. Notationally, a condition on the $j^{th}$ feature is denoted $x_{.j} \in \sigma_j(l)$ where $\sigma_j(l)$ is the set of allowed values for feature $j$ along the path to leaf $l$. If there are no conditions on feature $j$ along the path to $l$, then $\sigma_j(l)$ includes all possible outcomes for feature $j$. Thus, leaf $l$ includes outcomes $x$ obeying: $$x \in \left\{ x_{.1} \in \sigma_1(l),x_{.2} \in \sigma_2(l), \ldots, x_{.p} \in \sigma_p(l)\right\}.$$
For categorical data, the volume of a leaf $l$ is defined to be $\mathbb{V}_l=\prod_{j=1}^p |\sigma_j(l)|$. We give an example of this computation next.
\subsubsection*{Volume Computation Example}
The data are categorical. Possible outcomes for $x_{.1}$ are $\left\{ 1,2,3,4,5,6,7\right\}$. Possible outcomes for $x_{.2}$ are $\left\{ 100,101,102,103\right\}$. Possible outcomes for $x_{.3}$ are $\left\{ 10,11,12,13,14,15\right\}$. Possible outcomes for $x_{.4}$ are $\left\{ 8,9,10\right\}$.
Consider the tree in Figure \ref{MyTree}. \begin{figure}
\caption{Example of computation of volume.}
\label{MyTree}
\end{figure} We compute the volume for leaf $l$. Here,
$\sigma_1(l)=\left\{ 4,5 \right\}$ since $l$ requires both $x_{.1} \in \left\{ 4,5,6\right\}$ and $x_{.1} \in \left\{ 4,5\right\}$. $\sigma_2(l)=\left\{ 100,101\right\},$ $\sigma_3(l)=\left\{ 10,15\right\},$ and $\sigma_4(l)=\left\{ 8,9,10\right\}$ because there is no restriction on $x_{.4}$. So $$\mathbb{V}_l=\prod_j |\sigma_j(l)|=2\cdot 2\cdot 2\cdot 3=24.$$
Our notation handles only categorical data for ease of exposition but can be extended to handle ordinal and continuous data. For ordinal data, the definition is the same as for categorical but $\sigma_j$ can (optionally) include only continguous values (e.g. $\left\{ 3,4,5\right\}$ but not $\left\{ 3,4,6\right\}$). For continuous variables, $\sigma_j$ is the ``volume" of the continuous variables, for example, for node condition $x_{.j} \in (0,0.5), \sigma_j=0.5-0$.
In the next three subsections, we present the leaf-based modeling approach, branch-based modeling approach, and an approach to construct density rule lists.
\subsection{Model \RN{1}: Leaf-based Cascade Model} We define prior and likelihood for the tree-based model. To create the tree we will optimize the posterior over possible trees.\\
\noindent \textbf{Prior:}\\
For this model, the main prior on tree $T$ is on the number of leaves $K_T$. This prior we choose to be Poisson with a particular scaling (which will make sense later on), where the Poisson is centered at a user-defined parameter $\lambda$. Notation $N_{K_T}$ is the number of trees with $K_T$ leaves. The prior is:
\begin{eqnarray*}P(\textrm{Number of leaves in }T= K_T| \lambda) &\propto& N_{K_T}\cdot \textrm{ Poisson}(K_T,\lambda)\\ &=& N_{K_T} e^{-\lambda}\frac{\lambda^{K_T}}{K_T!}.\end{eqnarray*} Thus $\lambda$ allows the user to control the number of leaves in the tree. The number of possible trees is finite, thus the distribution can be trivially normalized.
Among trees with $K_T$ leaves, tree $T$ is chosen uniformly, with probability $1/N_{K_T}$. This means the probability to choose a particular tree $T$ is Poisson: \begin{eqnarray*}
P(T|\lambda)\propto
P(T|K_T)P(K_T|\lambda)&\propto &\frac{1}{N_{K_T}}N_{K_T} e^{-\lambda}\frac{\lambda^{K_T}}{K_T!} = e^{-\lambda}\frac{\lambda^{K_T}}{K_T!} \\ &\propto& \textrm{Poisson}(K_T,\lambda). \end{eqnarray*}
We place a uniform prior over the probabilities for a data point to land in each of the leaves. To do this, we start from a Dirichlet distribution with equal parameters $\alpha_1=\ldots=\alpha_{K_T}=\alpha \in \mathbb{Z}^+$ where hyperparameter $\alpha>1$. We denote the vector with $K_T$ equal entries $[\alpha,...,\alpha]$ as $\bm{\alpha}_{K_T}$.
We draw multinomial parameters $\bm\theta=[{\theta}_1, \ldots, {\theta}_{K_T}]$ from Dir$(\bm{\alpha}_{K_T})$.
Thus, the first part of our model is as follows, given hyperparameters $\lambda$ and $\alpha$: \begin{eqnarray*} \textrm{Number of leaves in T: } K_T&\propto & \textrm{scaled Poisson}(\lambda), \textrm{ i.e., } N_{K_T}\cdot \textrm{ Poisson}(K_T,\lambda)\\ \textrm{Tree shape }\;:T& \propto & \textrm{Uniform over trees with }K_T\textrm{ leaves}\\ \textrm{Prior distribution over leaves: }\bm\theta &\propto & \textrm{Dir}(\bm{\alpha}_{K_T}). \end{eqnarray*} As usual, the prior can be overwhelmed given enough data.\\
\noindent \textbf{Likelihood:}\\
Let $n_l$ denote the number of points captured by the $l$-th leaf, and denote $\mathbb{V}_l$ to be the volume of that leaf, defined above. The probability to land at any specific value within leaf $l$ is $\frac{\theta_l}{\mathbb{V}_l}$. The likelihood for the full data set is thus $$P(X|\bm{\theta},T)=\prod_{l=1}^{K_T} \left( \frac{\theta_l}{\mathbb{V}_l} \right)^{n_l}.$$\\ \textbf{Posterior:} \\ The posterior can be written as follows, where we have substituted the distributions from the prior into the formula. Here, $B(\bm{\alpha}_{K_T})=\frac{\prod_{l=1}^{K_T} \Gamma(\alpha_l)}{\Gamma(\sum_{l=1}^{K_T} \alpha_l)}=\frac{(\Gamma(\alpha))^{K_T}}{\Gamma(K_T\alpha)}$ is the multinomial beta function which is also the normalizing constant for the Dirichlet distribution. \begin{align*}
&P(T|\lambda, \bm{\alpha},X) \\
& \propto \int_{\bm{\theta}: \textrm{simplex}} P(K_T|\lambda)\cdot P(T|K_T)\cdot P(\bm{\theta}|\bm{\alpha}_{K_T})\cdot P(X|\bm{\theta},T) d \bm{\theta}\\
&\propto\int_{\bm{\theta}: \textrm{simplex}} P(T|\lambda) \left[\frac{1}{B(\bm{\alpha}_{K_T})} \left(\prod_{l=1}^{K_T} {\theta}_l^{\alpha-1} \right)\right]\left[\prod_{l=1}^{K_T} \left(\frac{{\theta}_l}{\mathbb{V}_l} \right)^{n_l} \right] d \bm{\theta}\\
&\propto P(T|\lambda) \frac{1}{B(\bm{\alpha}_{K_T})} \left( \prod_{l=1}^{K_T}\left(\frac{1}{\mathbb{V}_l} \right)^{n_l}\right)\int_{\bm{\theta}: \textrm{simplex}} \prod_{l=1}^{K_T} {\theta}^{n_l+\alpha-1} d \bm{\theta} \\
&\propto P(T|\lambda)\frac{B(n_1+\alpha, \ldots, n_{K_T}+\alpha)}{B(\bm{\alpha}_{K_T})}\prod_{l=1}^{K_T}\frac{1}{\mathbb{V}_l^{n_l}} \\
&\propto P(T|\lambda) \frac{\Gamma (K_T \alpha)}{\Gamma (n+K_T\alpha)}\prod_{l=1}^{K_T} \frac{(n_l+\alpha-1)!}{(\alpha-1)!} {\mathbb{V}_l^{-n_l}}, \end{align*}
where $P(T|\lambda)$ is simply Poisson($K_T,\lambda$) as discussed earlier.
For numerical stability, we maximize the log-posterior which is equivalent to maximizing the posterior.
For the purposes of prediction, we are required to estimate the density that is being assigned to leaf $l$. This is calibrated to the data, simply as: $$\hat{f}=\frac{n_l}{n \mathbb{V}_l}$$ where $n$ is the total number of training data points and $n_l$ is the number of training data points that reside in leaf $l$. The formula implicitly states that the density in the leaf is uniformly distributed over the features whose values are undetermined within the leaf (features for which $\sigma_j$ contains all outcomes for feature $j$).
\subsection{Model \RN{2}: Branch-based Cascade Model}
In the previous model, a Dirichlet distribution is drawn only over the leaves. In this model, a Dirichlet distribution is drawn at every internal node to determine branching. Similar to the previous model, we choose the tree that optimizes the posterior. \\
\noindent \textbf{Prior:}\\ The prior is comprised of two pieces: the part that creates the tree structure, and the part that determines how data propagates through it.\\ \textbf{Tree Structure Prior}:
For tree $T$, we let $B_T= \left\{ b_i |i \in I \right\}$ be a multiset, where each element is the count of branches from a node of the tree. For instance, if in tree $T$, the three nodes have 3 branches, 2 branches, and 2 branches respectively, then $B_T=\left\{3,2,2\right\}$. We let $\NBT$ denote the number of trees with the same multiset $B_T$. Note that $B_T$ is unordered, so $\{3,2,2\}$ is the same multiset as $\{2,3,2\}$ or $\{2,2,3\}$.
Let $I$ denote the set of internal nodes of tree $T$ and let $L$ denote the set of leaves. We let $\mathbb{V}_l$ denote the volume at leaf $l$.
In the generative model, a Poisson distribution with parameter $\lambda$ is used at each internal node in a top down fashion to determine the number of branches.
Iteratively, for node $i$, the number of branches, $b_i$, obeys $b_i \sim \textrm{Poisson}(\lambda).$ Hence, at any node $i$, with probability $\exp(-\lambda) \frac{\lambda^{b_i}}{b_i!}$, there are $b_i$ branches from node $i$. This implies that with probability $\exp(-\lambda)$, the node is a leaf. In summary,
\begin{equation*} P(\text{Multiset of branches }= B|\lambda) \propto \NB \left[\prod_{i \in I}e^{-\lambda} \frac{\lambda^{b_i}}{b_i!}\right]\left[\prod_{l \in L}e^{-\lambda}\right].\end{equation*}
Among trees with multiset $B$, tree $T$ is chosen uniformly, with probability $\frac{1}{\NB}.$ This means the probability to choose a particular tree is:
\begin{equation} P(T|\lambda)\propto P(T|B_T)P(B_T|\lambda) \propto \frac{1}{\NBT} \NBT \left[\prod_{i \in I}e^{-\lambda} \frac{\lambda^{b_i}}{b_i!}\right]\left[\prod_{l \in L}e^{-\lambda}\right]=\left[\prod_{i \in I}e^{-\lambda} \frac{\lambda^{b_i}}{b_i!}\right]\left[\prod_{l \in L}e^{-\lambda}\right]. \label{branchtreeprior} \end{equation}
\noindent\textbf{Tree Propagation Prior:} After the tree structure is determined, we need a generative process for how the data propagate through each internal node. We denote $\theta_{l}$ as the probability to land in leaf $l$. We denote $\widetilde{\theta}_{ij}$ as the probability to traverse to node $j$ from internal node $i$. Notation $\bm{\theta}$ is the vector of leaf probabilities (the $\theta_l$'s), $\widetilde{\bm{\theta}}$ is the set of all $\widetilde\theta_{ij}$'s, and $\widetilde{\bm{\theta}}_i$ is the set of all internal node transition probabilities from node $i$ (the $\widetilde{\theta}_{ij}$'s).
We compute $P(\widetilde{\bm{\theta}_i}|\alpha, T)$ for all internal nodes $i$ of tree $T$. At each internal node, we draw a sample from a Dirichlet distribution with parameter $[\alpha,\ldots, \alpha]$ (of size equal to the number of branches $b_i$ of $i$) to determine the proportion of data, $\widetilde{\theta}_{i,j}$, that should go along the branch leading to each child node $j$ from the internal parent node $i$. Thus, $\widetilde{\bm\theta_i} \sim \textrm{Dir}(\bm\alpha)$ for each internal node $i$, that is:
$$P(\widetilde{\bm{\theta}}_i|\bm\alpha,T)=\frac{1}{B_{b_i}(\bm{\alpha})}\prod_{j \in C_i} \widetilde{\theta}_{ij}^{\alpha-1},$$ where $B_k(\bm{\alpha})$ is the normalizing constant for the Dirichlet distribution with parameter ${\alpha}$ and $k$ categories, and $C_i$ are the indices of the children of $i$. Thus, \begin{equation}\label{model2priorsecondterm}
P(\widetilde{\bm{\theta}}|\bm\alpha,T)=\prod_i P(\widetilde{\bm{\theta}}_i|\bm\alpha,T) = \prod_i \frac{1}{B_{b_i}(\bm{\alpha})}\prod_{j \in C_i} \widetilde{\theta}_{ij}^{\alpha-1}.\end{equation}
Thus, the prior is
$P(T|\lambda) \cdot P(\widetilde{\bm{\theta}}|\alpha,T)$, where $P(T|\lambda)$ is in (\ref{branchtreeprior}) and $P({\widetilde{\bm\theta}}|\alpha,T)$ is in (\ref{model2priorsecondterm}).
In summary, the prior of our model is as follows, given hyperparameters $\lambda$ and $\alpha$: \begin{eqnarray*}\label{priormodel} \textrm{Multiset of branches: } B_T &\propto & \NBT \left[\prod_{i \in I}e^{-\lambda} \frac{\lambda^{b_i}}{b_i!}\right]\left[\prod_{l \in L}e^{-\lambda}\right].\\ \textrm{Tree shape }\;:T& \sim & \textrm{Uniform over trees with branches } B_T.\\ \textrm{Prior distribution over each branch: }\widetilde{\bm\theta}_i &\sim & \textrm{Dir}(\alpha). \end{eqnarray*}
\noindent \textbf{Likelihood:}\\ The density within leaf $l$ is uniform and equal to
$$P(X=x|X \in \text{leaf } l)= \left\{ \begin{array}{cc} \frac{1}{\mathbb{V}_l}, & x \in \text{\text{leaf } l,} \\ 0, & \text{otherwise.} \end{array}\right.$$ We denote the set $P_l$ as the set of branches in the path from the root of the tree to the leaf $l$. The probability of $X$ taking on value $x$ (permitting that $x$ is an allowed outcome in leaf $l$) is thus:
\begin{align*}
P(X=x , X \in \text{leaf } l) &=P(X=x|X \in \text{leaf } l)\cdot P(X \in \text{leaf } l) \\
&= \frac{1}{\mathbb{V}_l} \cdot \theta_l = \frac{\prod_{(\hat{i},\hat{c}) \in P_l} \widetilde{\theta}_{\hat{i},\hat{c}}}{\mathbb{V}_l}.
\end{align*}
Denote the set of children of node $i$ by $C_i$ and the number of data points in node $c$ as $n_c$. It is true that: $$\prod_{l \in L} \left(\prod_{(\hat{i},\hat{c})\in P_l} \widetilde{\theta}_{\hat{i},\hat{c}}\right)^{n_l}=\prod_{i \in I}\prod_{c \in C_i} \widetilde{\theta}_{i,c}^{n_{c}}.$$ The equality stems from two distinct ways of counting the branches that a particular data point passes through from the root to the leaf. The first way of counting is to start from the leaf and count backwards from the leaf to the root (depth first). The second way of counting is by examining each internal node (breadth first).
Hence the likelihood of a particular data set can be written:
$$P(X|{\widetilde{\bm{\theta}}},T)=\prod_{l \in L} \left(\frac{\prod_{(\hat{i},\hat{c})\in P_l} \widetilde{\theta}_{\hat{i},\hat{c}}}{\mathbb{V}_l}\right)^{n_l}=\frac{\prod_{i \in I}\prod_{c \in C_i} \widetilde{\theta}_{i,c}^{n_{c}}}{\prod_{l \in L} \mathbb{V}_l^{n_l}}.$$
\noindent \textbf{Posterior:}\\
The posterior is proportional to the prior times the likelihood terms. Here we are integrating over the $\widetilde{\bm{\theta}}_i$ terms for each of the internal nodes $i$. \begin{align*}
&P(T|\lambda, \alpha, X) \\
&\propto \int P(B_T|\lambda) \cdot P(T|B_T) \cdot P(\widetilde{\bm{\theta}}| \alpha,T) \cdot P(X| \widetilde{\bm{\theta}},T) d\widetilde{\bm{\theta}} \\ &\propto \left[\prod_{l \in L} \left(\frac{ e^{-\lambda}}{\mathbb{V}_l^{n_l}}\right)\right]\left(\prod_{i \in I} e^{-\lambda} \frac{\lambda^{b_i}}{b_i!}\frac{1}{B_{b_i}(\alpha, \ldots, \alpha)} \int_{\widetilde{\bm{\theta}}_i \in \text{simplex}} \prod_{c \in C_i} \widetilde{\theta}_{i,c}^{\alpha-1} \widetilde{\theta}_{i,c}^{n_{c}}d \widetilde{\bm{\theta}}_i \right) \\
&= e^{-\lambda(|I|+|L|)} \left(\prod_{i \in I} \frac{\lambda^{b_i}}{b_i!}\frac{1}{B_{b_i}(\alpha, \ldots, \alpha)} \int_{\widetilde{\bm{\theta}}_i \in \text{simplex}} \prod_{c \in C_i} \widetilde{\theta}_{i,c}^{n_{c}+\alpha-1} d \widetilde{\bm{\theta}}_i \right) \prod_{l \in L} \left(\frac{ 1}{\mathbb{V}_l^{n_l}}\right)\\
&= e^{-\lambda(|I|+|L|)} \lambda^{\sum_{i \in I} b_i} \left(\prod_{i \in I} \frac{1}{b_i!}\frac{B_{b_i}(\alpha+n_{c_1}, \ldots, \alpha+n_{c_{b_i}})}{B_{b_i}(\alpha, \ldots, \alpha)} \right) \prod_{l \in L} \left(\frac{ 1}{\mathbb{V}_l^{n_l}}\right)\\
&= e^{-\lambda(|I|+|L|)} \lambda^{
|L|+|I|-1 } \left(\prod_{i \in I} \frac{1}{b_i!}\frac{B_{b_i}(\alpha+n_{c_1}, \ldots, \alpha+n_{c_{b_i}})}{B_{b_i}(\alpha, \ldots, \alpha)} \right) \prod_{l \in L} \left(\frac{ 1}{\mathbb{V}_l^{n_l}}\right) \end{align*}
where $c_1,\ldots,c_{b_i} \in C_i$ in the second last expression. We used the equation $\sum_{i \in I} b_i=|L|+|I|-1$ for a tree in the last line.
We use a specialized simulated annealing method to search for the maximum a posteriori tree.
Our algorithm moves among neighboring trees and records the best tree that has been found so far. The description of this method is in the appendix.\\
\noindent \textbf{Possible Extension:} We can include an upper layer of the hierarchical Bayesian Model to control (regularize) the number of features $d$ that are used in the cascade out of a total of $p$ dimensions. This would introduce an extra multiplicative factor within the posterior of
$\left( \begin{array}{c} p \\ d \end{array} \right)\gamma^d (1-\gamma)^{p-d}$, where $\gamma$ is a parameter between 0 and 1, where a smaller value favors a simpler model.
\begin{eqnarray*}
\lefteqn{\left( \begin{array}{c} p \\ d \end{array} \right)\gamma^d (1-\gamma)^{p-d} e^{-\lambda(|I|+|L|)} \lambda^{|I|+|L|-1}} \\ &&\left(\prod_{i \in I} \frac{1}{b_i!}\frac{B_{b_i}(\alpha+n_{c_1}, \ldots, \alpha+n_{c_{b_i}})}{B_{b_i}(\alpha, \ldots, \alpha)} \right) \prod_{l \in L} \left(\frac{ 1}{\mathbb{V}_l^{n_l}}\right). \end{eqnarray*}
\subsection{Model \RN{3}: Leaf-based Density Rule List}
Rather than producing a general tree, an alternative approach is to produce a rule list. A rule list is a one-sided tree. Rule lists are easier to optimize than trees. Each tree can be expressed as a rule list; however, some trees may be more complicated to express as a rule list. By using lists, we implicitly hypothesize that the full space of trees may not be necessary and that simpler rule lists may suffice.
An example of a density rule list is as follows:
\textbf{if} $x$ obeys $a_1$ \textbf{then} density$(x)=f_1$
\textbf{else if} $x$ obeys $a_2$ \textbf{then} density$(x)=f_2$
$\vdots$
\textbf{else if} $x$ obeys $a_m$ \textbf{then} density$(x)=f_m$
\textbf{else} density$(x)=f_0$.
The antecedents $a_1$,...,$a_m$ are chosen from a large pre-mined collection of possible antecedents, called $A$.
We define $A$ to be the set of all possible antecedents of size at most $H$, where the user chooses $H$. The size of $A$ is:
$$|A|=\sum_{j=0}^H A_j,$$ where $A_j$ is the number of antecedents of size $j$, $$A_j=\sum_{ \left[ \begin{array}{c} t_1, t_2, \ldots, t_j \in \left\{ 1,\ldots, p \right\} \\ \text{s.t. } t_1 > t_2 \ldots >t_j \end{array}\right]} \prod_{i=1}^j q_{t_i},$$ where feature $i$ consists of $q_i$ categories.
\noindent \textbf{Generative Process:}
We now sketch the generative model for the tree from the observations $x$ and antecedents $A$. Prior parameters $\lambda$ and $\eta$ are used to indicate preferences over the length of the density list and the number of conjunctions in each sub-rule $a_i$.
Define $a_{<j}$ as the antecedents before $j$ in the rule list if there are any. For example $a_{<3}=\left\{ a_1, a_2 \right\}$. Similarly, let $c_j$ be the cardinalities of the antecedents before $j$ in the rule list. Let $d$ denote the rule list. The generative model is as follows, following the exposition of \cite{LethamRuMcMa15}:
\begin{enumerate}
\item Sample a decision list length $m \sim P(m | A, \lambda)$. \item For decision list rule $j=1,\ldots, m:$ \\
Sample the cardinality of antecedent $a_j$ in $d$ as $c_j \sim P(c_j| c_{<j}, A, \eta)$. \\
Sample $a_j$ of cardinality $c_j$ from $P(a_j|a_{<j},c_j,A)$. \item For observation $i=1, \ldots, n$: Find the antecedent $a_j$ in $d$ that is the first that applies to $x_i$. If no antecedents in $d$ applies, set $j=0$. \item Sample parameter ${\bm{\theta}} \sim$ Dirichlet ($\bm{\alpha}$) for the probability to be in each of the leaves, where $\bm{\alpha}$ is a user-chosen vector of size $m+1$, usually where all elements are the same. $f_i=\frac{\theta_i}{\mathbb{V}_i}$, where $\mathbb{V}_i$ is the volume. \end{enumerate}
\noindent\textbf{Prior:}\\ The distribution of $m$ is the Poisson distribution, truncated at the total number of preselected antecedents:
$$P(m|A,\lambda)=\frac{\lambda^m/m!}{\sum_{j=0}^{|A|} (\lambda^j/j!)}, m=0, \ldots, |A|.$$
When $|A|$ is huge, we can use the approximation $P(m|A, \lambda) \approx \lambda^m/m!$, as the denominator of the previous term would be close to 1.
We let $R_j(c_1,\ldots, c_j,A)$ be the set of antecedent cardinalities that are available after drawing antecedent $j$, and we let $P(c_j|c_{<j},A, \eta)$ be a Poisson truncated to remove values for which no rules are available with that cardinality:
$$P(c_j|c_{<j},A,\eta)=\frac{(\eta^{c_j}/c_j!)}{\sum_{k \in R_{j-1}(c_{<j},A)}(\eta^k/k!)}, \hspace{5pt} c_j \in R_{j-1}(c_{<j},A).$$ We use a uniform distribution over antecedents in $A$ of size $c_j$ excluding those in $a_{j}$,
$$P(a_j|a_{<j},c_j,A) \propto 1, \hspace{20pt} a_j \in \left\{ a \in A \setminus a_{<k}: |a|=c_j\right\}.$$ The cascaded prior for the antecedent lists is thus:
$$P(d|A, \lambda, \eta)=P(m|A,\lambda) \cdot \prod_{j=1}^m P(c_j|c_{<j},A, \eta) \cdot P(a_j|a_{<j},c_j,A).$$ The prior distribution over the leaves $\bm{\theta}=[\theta_1,\ldots,\theta_m,\theta_0]$ is drawn from Dir($\bm{\alpha}_{m+1}$).
$$P(\bm{\theta}|\alpha)=\frac{1}{B_{m+1}(\alpha, \cdots,\alpha)}\prod_{l=0}^m \theta_l^{\alpha-1}$$ It is straightforward to sample an ordered antecedent list $d$ from the prior by following the generative model that we just specified, generating rules from the top down.
\noindent \textbf{Likelihood:}\\
\noindent Similar to the first model, the probability to land at any specific value within leaf $l$ is $\frac{\theta_l}{\mathbb{V}_l}$. Hence, the likelihood for the full data set is:
$$P(X|\bm{\theta},d)=\prod_{l=0}^{m}\left( \frac{\theta_l}{\mathbb{V}_l} \right)^{n_l}.$$
\noindent \textbf{Posterior:} \\
\noindent The posterior can be written as \begin{eqnarray*} \lefteqn{
P(d|A,\lambda, \eta, \alpha, X) } \\
&\propto &\int_{\bm{\theta} \in \textrm{simplex}} P(d|A,\lambda, \eta) \cdot P(\bm{\theta} | \alpha) \cdot P(X|\bm{\theta}, d) d\bm{\theta} \\
& = & P(d|A, \lambda, \eta) \int_{\bm{\theta} \in \textrm{ simplex}} \frac{1}{B_{m+1}(\alpha, \cdots,\alpha)} \prod_{l=0}^m \theta_l^{\alpha-1}\left(\frac{\theta_l}{\mathbb{V}_l}\right)^{n_l} d \bm{\theta}\\
&=& P(d|A, \lambda, \eta)\frac{\prod_{l=0}^m\Gamma{(n_l+\alpha)} \mathbb{V}_l^{-n_l}}{\Gamma(\sum_{l=0}^m (n_l+\alpha))}. \end{eqnarray*} where the last equality uses the standard Dirichlet-multinomial distribution derivation.
To search for optimal rule lists that fit the data, we use local moves (adding rules, removing rules, and swapping rules) and use the Gelman-Rubin convergence diagnostic applied to the log posterior function.
A technical challenge that we need to address in our problem is the computation of the volume of a leaf. Volume computation is not needed in the construction of a decision list classifier like that of Letham et al. \cite{LethamRuMcMa15} but it is needed in the computation of density list. There are multiple ways to compute the volume of a leaf of a rule list. The first set of approaches do not require the overhead of creating a complicated data structure, and thus might be better for smaller problems.\\ \textit{Approach 1}: create uniform data over the whole domain, and count the number of points that satisfy the antecedents. This approach would be expensive when the domain is huge but easy to implement for smaller problems. \\ \textit{Approach 2}: use an MCMC sampling approach to sample uniformly the whole domain space. This approach is again not practical when the domain size is huge as the number of samples required will increase exponentially due to curse of dimensionality.\\ \textit{Approach 3}: use the inclusion-exclusion principle to directly compute the volume of each leaf. Consider computing the volume of the $i$-th leaf. Let $V_{a_i}$ denote the volume induced by the rule $a_i$, that is the number of points in the domain that satisfy $a_i$. To belong to that leaf, a data point has to satisfy $a_i$ and not $a_{<i}$. Hence the volume of the $i$-th leaf is equal to the volume obeying $a_i$ alone, minus the volume that has been used by earlier rules. Hence, we have the following: \begin{align*} \mathbb{V}_i&=V_{a_i \wedge \bigwedge_{k=1}^{i-1} a_k^c}\\&=V_{a_i}-V_{a_i \wedge (\bigvee_{k=1}^{i-1} a_k)} \\&=V_{a_i}-V_{ (\bigvee_{k=1}^{i-1}a_i \wedge a_k)} \\&=V_{a_i}-\sum_{k=1}^{i-1}(-1)^{k+1}\sum_{1\leq j_1 \leq \ldots j_k \leq n} V_{a_i\wedge a_{j_1} \ldots \wedge a_{j_k}}, \end{align*} where the last expression is due to the inclusion-exclusion principle and it only involves the volume resulting from conjunctions. The volume resulting from conjunctions can be easily computed from data. Without loss of generality, suppose we want to compute the volume of $V_{a_1 \wedge \ldots \wedge a_k}$, for each feature that appears, we examine if there is any contradiction. For example if feature 1 is present in both $a_1$ and $a_2$ and they specify feature 1 to take different values, then we have found a contradiction and the volume should be 0. Suppose this is not the case, then the volume is equal to the product of the number of distinct categories of all the features that are not used. By using the inclusion-exclusion principle, we reduce the problem to just computing a volume of conjunctions, however, computing these volumes requires a clever data structure. This would be suitable for larger problems but might slow down computations for smaller problems.
\section{Experiments}
Our experimental setup is as follows. We considered five models: the leaf-based cascaded histograms, the branch-based cascaded histograms, the leaf-based density list, regular histograms and density estimation trees (DET) \citep{Ram:2011:DET:2020408.2020507}. To our knowledge, this essentially represents the full set of logical, high dimensional density estimation methods. To ascertain uncertainty, we split the data in half 5 times randomly and assessed test log-likelihood and sparsity of the trees for each method. A model with fewer bins and higher test likelihood is a better model.
For the histogram, we treated each possible configuration as a separate bin. DET was designed for continuous data, which meant that the computation of volume needed to be adapted -- it is the number of configurations in the bin (rather than the lengths of each bin multiplied together). The DET method has two parameters, the minimum allowable support in a leaf, and the maximum allowable support. We originally planned to use a minimum of 0 and a maximum of the size of the full dataset, but the algorithm often produced trivial models when we did this. We tried also values $\left\{ 0,3,5\right\}$ for the minimum values and $\left\{ 10,n,\lfloor\frac{n}{2} \rfloor \right\}$ where $n$ is the number of training data points, and reported results for the best of these. For the leaf-based cascade model, the mean of the Poisson prior was chosen from the set $\left\{ 5,8\right\}$ using nested cross validation. For the branch-based cascade model, the parameter to control the number of branches was chosen from the set $\left\{ 2,3\right\}$. $\gamma$ was fixed to be 0.5, and $\alpha$ was set to be 2 for the experiment. For the leaf-based density list model, the parameters $\lambda,\eta$ and $\alpha$ were chosen to be 3, 1 and 1 respectively.
\subsection{An Experiment on the Titanic Data Set}
The Titanic dataset has an observation for each of the 2201 people aboard the Titanic. There are 3 features: gender, whether someone is an adult, and the class of the passenger (first class, second class, third class, or crew member). A cascade would help us understand the set of people on board the Titanic.
\begin{figure}
\caption{The scatter plot for titanic. }
\label{Figtitanic}
\end{figure}
Figure \ref{Figtitanic} shows the results, both out-of-sample likelihood and sparsity, for each model, for each of the 5 folds. The histogram method had high likelihood, but also the most leaves (by design). The other methods performed similarly, arguably the leaf-based density list method performed slightly better in the likelihood-sparsity tradeoff. DET produced a trivial tree for one of the splits. In general, we will see similar results on other datasets: the histogram produces too many bins, the leaf-based density list model and leaf-based cascade perform well, and DET has inconsistent performance (possibly due to its top-down greedy nature, or the fact that DET approximately optimizes Hellinger distance rather than likelihood.) Figure \ref{Figtitanictree} shows one of the density cascades generated by the leaf-based method. The reason for the top split is clear: the distributions of the males and females were different, mainly due to the fact that the crew was mostly male. There were fewer children than adults, and the volume of crew members was very different than the volume of $1^{\textrm{st}}, 2^{\textrm{nd}}$, and $3^{\textrm{rd}}$ class passengers. Figure \ref{Figtitaniclist} shows one of the density lists generated by our model. It shows that male crew and third class male adults have higher density.
\begin{figure}
\caption{Tree representing titanic.}
\label{Figtitanictree}
\end{figure}
\begin{figure}
\caption{List representing titanic. Each arrow represents an ``else if" statement. This can be directly compared to the cascade in Figure \ref{Figtitanictree}. Slight differences in estimates between the two models occurred because we used different splits of data for the two figures. The estimates were robust to the change in data.}
\label{Figtitaniclist}
\end{figure}
\subsection{Crime Dataset} The housebreak data used for this experiment were obtained from the Cambridge Police Department, Cambridge, Massachusetts. The motivation is to understand the common types of modus operandi (M.O.) characterizing housebreaks, which is important in crime analysis. The data consist of 3739 separate housebreaks occurring in Cambridge between 1997 and 2012 inclusive. We used 6 categorical features.
\begin{enumerate} \item Location of entry: ``window," ``door," ``wall," and ``basement."
\item Means of entry: ``forceful" (cut, broke, cut screen, etc.), ``open area," ``picked lock," ``unlocked," and ``other."
\item Whether the resident is inside. \item Whether the premise is judged to be ransacked by the reporting officer. \item ``Weekday" or ``Weekend."
\item Type of premise. The first category is ``Residence" (including
apartment, residence/unk., dormitory, single-family house, two-family house, garage (personal), porch, apartment hallway, residence unknown, apartment basement, condominium). The second category is ``non-medical, non-religious work place" (commercial unknown, accounting firm, research, school). The third group consists of halfway houses, nursing homes, medical buildings, and assisted living. The fourth group consists of parking lots and parking garages, and the fifth group consists of YWCAs, YMCAs, and social clubs. The last groups are ``storage," ``construction site," ``street," and ``church" respectively. \end{enumerate} \begin{figure}
\caption{the scatter plot for Cambridge Police Department dataset.}
\end{figure}
The experiments show that DET and our approaches are competitive for the crime data set. (The histogram's results involve too many bins to fit on the figure.)
Let us discuss one of the trees obtained from the leaf-based cascade method where we have set the mean of the Poisson distribution to be chosen from the set $\left\{ 20,30\right\}$. The tree is in Figure \ref{fig:crimetree}. It states that most burglaries happen at residences -- the non-residential density has values less than $1 \times 10^{-4}.$ Given that a crime scene is a residence, most crimes happened when the resident was not present. If the premise is a residence and the resident was present for the housebreak, the burglary is more likely to happen on a weekday, in which case most burglaries involve forceful means of entry (density = $9.16 \times 10^{-3}$). When the premise is a residence and the resident was not present, the location of entry is usually either a window or a door. Given this setting: \begin{enumerate} \item If the means of entry is forceful, most crime happens on weekdays, and in that case it is almost twice as likely that the means of entry is through a door (density=0.15) compared to a window (density=0.078). If the crime happened on a weekend, it is more likely for the crime scene not to be ransacked (density=$5.50 \times 10^{-2}$) as compared to being ransacked (density=$2.41 \times 10^{-3}$). \item If the means of entry is either an open area or a lock is picked, it is more likely to be on a weekday (density=$2.48 \times 10^{-3}$) compared to a weekend (density=$7.30 \times 10^{-3}$). \item If the means of entry is none of the above, it is almost three times more likely to happen on a weekday (density=$3.60 \times 10^{-3}$) compared to a weekend (density=$1.07 \times 10^{-3}$). \end{enumerate} These types of results can be useful for crime analysts to assess whether a particular modus operandi is unusual. A density list for these data is presented in Figure \ref{fig:crimelist}.
\section{Empirical Analysis} Each subsection below is designed to provide insight into how the models operate.
\subsection{Sparse Tree Dataset} We generated a dataset that arises from a tree with 6 leaves, involving 3 features. The data consists of 1000 data points, where 100 points are tied at value (1,2,1), 100 points are at (1,2,2), 100 points are at (2,1,1), 400 points are at (2,1,2), and 300 points are at (2,2,2). The correct tree is in Figure \ref{Fig:simdata}.
\begin{figure}
\caption{Performance vs sparsity on sparse tree data set.}
\label{Fig:sparsetreescatter}
\end{figure}
\begin{figure}
\caption{Tree diagram for sparse tree data set.}
\label{Fig:simdata}
\end{figure}
\begin{figure}
\caption{Output for leaf-based model that recovers the data structure.}
\label{Fig:recoversimdata}
\end{figure}
\begin{figure}
\caption{List output for sparse tree data set.}
\label{Fig:listsimdata}
\end{figure}
We trained the models on half of the dataset and tested on the other half. Figure \ref{Fig:sparsetreescatter} shows the scatter plot of out-of-sample performance and sparsity. This is a case where the DET failed badly to recover the true model. It produced a model that was too sparse, with only 4 leaves. The leaf-based cascade method recovered the full tree from Figure \ref{Fig:simdata} and we present the tree in Figure \ref{Fig:recoversimdata}. The output for the corresponding density list is presented in Figure \ref{Fig:listsimdata}.
\subsection{Extreme Uniform Dataset}
We generated a 1-dimensional data set that consists of 100 data points. The data are simply all unique integers from 1 to 100. This is a case where the histogram badly fails to generalize. Figure \ref{Fig:extreme} shows the result.
\begin{figure}
\caption{Performance vs sparsity on uniform data. }
\label{Fig:extreme}
\end{figure}
The leaf-based and branch-based models both return the solution that consists of a single root node, implying that the data are in fact uniformly distributed, or at least that we do not have evidence to further split on the single node. The density list output is close to uniform as well. DET is competitive as well, though it does not return the trivial tree. The histogram totally fails, since the test data and training data do not overlap at all.
\section{Consistency}
A consistent model has estimates that converge to the real densities as the size of the training set grows. Consistency of conventional histograms is well studied for example, \cite{abou1976conditions,devroye1983distribution}. More generally, consistency for general rectangular partitions has been studied by \cite{zhao1991almost,lugosi1996consistency}. Typical consistency proofs, (e.g.,\cite{citeulike:3463149,Ram:2011:DET:2020408.2020507}) require the leaf diameters to become asymptotically smaller as the size of the data grows. In our case if the ground truth density is a tree, we do not want our models to asymptotically produce smaller and smaller bin sizes, we would rather they reproduce the ground truth tree. This means we require a new type of consistency proof. \\
\noindent\textbf{Definition 1}: \textit{Density trees} have a single root and there are conditions on each branch. A density value, $f_l$ is associated with each leaf $l$ of the tree. \\
\noindent\textbf{Definition 2}: Two trees, $T_1$ and $T_2$ are \textit{equivalent} with respect to density $f$ if they assign the same density values to every data point on the domain, $f_{T_1}(x)=f_{T_2}(x)$, for all $x$. We denote the class of trees that are equivalent to $T$ as $[T]_f$.\\
\noindent\textbf{Theorem 1}: Let $\Theta$ be the set of all density trees.
Consider these conditions: \begin{enumerate}
\item $T_n \in \argmax_T \textrm{Obj}(T)$. The objective function can be decomposed into $\textrm{Obj}(T)=\ln q_n(T|X)+\ln g_n(T|X)$ where \\ $\argmax_T \left[ \ln q_n(T|X)+ \ln g_n(T|X) \right]\equiv \argmax_T \ln g_n(T|X)$ as $n \rightarrow \infty$.
\item $\ln g_n(T|X)$ converges in probability, for any tree $T$, to the empirical log-likelihood that is obtained by the maximum likelihood principle, $\hat{l}_n(T|X)=\frac{1}{n}\sum_{i=1}^n \ln \hat{f}_n(x_i|T)$.
\item $\sup_{T \in \Theta}|\hat{l}_n(T|X)-l(T)| \xrightarrow{P} 0$ where $l(T)=\mathbb{E}_x(\ln(f(x|T)))$. \item $T^*_{\textrm{MLE}}\in \argmax_Tl(T)$ is unique up to equivalence among elements of $[T^*_{\textrm{MLE}}]_f$. \end{enumerate} If these conditions hold, then the trees $T_n$ that we learned, $T_n\in \argmax_T \textrm{Obj}(T)$, obey $T_n\in [T^*_{\textrm{MLE}}]_f$ for $n >M$ for some $M$.
The first condition and the second condition are true any time we use a Bayesian model. They are also true any time we use regularized empirical likelihood where the regularization term's effect fades with the number of observations. Note that the third condition is automatically true by the law of large numbers. The last condition is not automatically true, and requires regularity conditions for identifiability. The result states that our learned trees are equivalent to maximum likelihood trees when there are enough data. The proof is presented in Appendix C.
\section{Conclusion}
We have presented a Bayesian approach to density estimation using cascaded piecewise constant estimators. These estimators have nice properties: their prior encourages them to be sparse, which permits interpretability. They do not have the pitfalls of other nonparametric density estimation methods like density estimation trees, which are top-down greedy. They are consistent, without needing to asymptotically produce infinitesimally small leaves. Practically, the approaches presented here have given us insight into a real data set (the housebreak dataset from the Cambridge Police) that we could not have obtained reliably in any other way.
\acks{We would like to acknowledge support for this project from Accenture, Siemens, and Wistron. }
\begin{figure*}
\caption{Tree representing the crime data set.}
\label{fig:crimetree}
\end{figure*}
\begin{figure}
\caption{List representing the crime data set. Each arrow represent an ``else if" statement.}
\label{fig:crimelist}
\end{figure}
\appendix \section*{Appendix A. Simulated Annealing Algorithm} \label{app:theorem}
At each iteration we need to determine which neighboring tree to move to. To decide which neighbor to move to, we fix a parameter $\epsilon>0$ beforehand, where $\epsilon$ is small. At each time, we generate a number from the uniform distribution on (0,1), then:\\
1. If the number is smaller than $\frac{1-\epsilon}{4}$, we select uniformly at random a parent which has leaves as its children, and remove its children. This is always possible unless the tree is the root node itself in which case we cannot remove it and this step is skipped. \\ 2. If the random number is between $\frac{1-\epsilon}{4}$ and $\frac{1-\epsilon}{2}$, we pick a leaf randomly and a feature randomly. If it is possible to split on that feature, then we create children for that leaf. (If the feature has been used up by the leaf's ancestors, we cannot split.) \\ 3. If the random number is between $\frac{1-\epsilon}{2}$ and $\frac{3(1-\epsilon)}{4}$, we pick a node randomly, delete its descendants and split the node into two nodes containing subsets of the node's outcomes. Sometimes this is not possible, for example if we pick a node where all the features have been used up by the node's ancestors, or if the node has only one outcome. In that case we skip this step. \\
4. If the random number is between $\frac{3(1-\epsilon)}{4}$ and $(1-\epsilon)$, we choose two nodes that share a common parent, delete all their descendants and merge the two nodes.\\ 5. If the random number is more than $1-\epsilon$, we perform a structural change operation where we remove all the children of a randomly chosen node of the tree.
The last three actions avoid problems with local minima. The algorithms can be warm started using solutions from other algorithms, e.g., DET trees. We found it useful to occasionally reset to the best tree encountered so far or the trivial root node tree. Ordinal data are treated differently than binary categorical data in that merges and splits in Steps 3 and 4 are done on neighboring ordinal values.
\section*{Appendix B. Evaluation Metric}
We discuss evaluation metrics for use in out-of-sample testing and parameter tuning with nested cross-validation. The natural evaluation metric is the likelihood of the trained model calculated on the test data. Hellinger distance is an alternative \citep{hellinger1909neue}, however, we prefer likelihood for two reasons. (i) To compute the Hellinger distance, it is assumed that the real distribution is known, when in reality it is not known. (ii) Hellinger distance would be approximated by $2-\frac{2}{n}\sum_{i=1}^n \sqrt{\frac{\hat{f}(x_i)}{f(x_i)}}$ \citep{AISTATS07_LiuLW} where $f$ is not known in practice. The estimate of the Hellinger distance often comes out negative, which is nonsensical. An alternative is to use a least square criterion \citep[see][]{Load:1999}. We consider the risk for tree $h$: \begin{align*} L(h) &= \int (\hat{f}_n(x)-f(x))^2 dx \\ &= \int (\hat{f}(x))^2 dx -2 \int \hat{f}_n(x)f(x)dx+ \int f^2(x)dx. \end{align*} Since the true density is not known, the third term cannot be evaluated, is a constant, and thus can be ignored. The first two terms can be estimated using a leave one out estimator: $$\hat{J} (h) = \int \left( \hat{f}_n(x)\right)^2 dx-\frac{2}{n} \sum_{i=1}^{n} \hat{f}_{(-i)}(x_i)$$ where the second term's $f(x)$ vanishes because $x$ is drawn from density distribution $f$. We desire this value to be as negative as possible. We know that $\hat{f}(x_i)= \frac{n_l}{nV_l}$ if $i \in l$ where $l$ is a leaf. We approximate $\hat{f}_{(-i)}$ using the assumption that the tree does not change structurally when one point is removed: $$\hat{f}_{(-i)}(x_i) \approx \frac{n_l-1}{(n-1)V_l}=\frac{n (n_l-1)}{(n-1)n_l}\hat{f}(x_i)=\frac{n (n_l-1)}{(n-1)n_l}\hat{f}_l.$$ Hence, we simplify $\hat{J}(h)$ as follows: \begin{align*} \hat{J} (h) &= \int\left( \hat{f}_n(x)\right)^2 dx-\frac{2}{n} \sum_{l \in L} (n_l) \hat{f}_{(-i)}(x_i) \\ &\approx \int\left( \hat{f}_n(x)\right)^2 dx-2 \sum_{l \in L} \frac{n_l-1}{(n-1)}\hat{f}_l \\ &= \sum_{l \in L} \hat{f}_l^2V_l-2 \sum_{l \in L} \frac{n_l-1}{(n-1)}\hat{f}_l \\ &= \sum_{l \in L} \left(\hat{f}_lV_l- \frac{2(n_l-1)}{(n-1)}\right) \hat{f}_l \\ &=\sum_{l \in L} \left(\frac{n_l}{n}- \frac{2(n_l-1)}{(n-1)}\right) \hat{f}_l. \end{align*}
The formula can be used aas an alternative evaluation metric of which the more negative it is, the better the model fit the data. It can be viewed as a weighted sum of likelihood over the leaves.
\section*{Appendix C. Proof for Theorem 1}
From definition of $T_n$, $T_n$ is an optimal value of the log-objective function and hence it is also an optimal solution to $g_n$ as $n$ is sufficiently large due to the first condition. We have that
\[ \ln \textrm{Obj}(T_n|X)-\ln \textrm{Obj}(T^*_{\textrm{MLE}}|X) \geq 0 \] by definition of $T_n$ as the maximizer of \textrm{Obj}. Because \textrm{Obj} becomes close to $g_n$, we have that \begin{equation}\label{thmeqn1}
\ln g_n(T_n|X) - \ln g_n(T^*_{\textrm{MLE}}) \geq 0 \end{equation} as $n$ is sufficiently large.
From Condition 2, we know that $\ln g_n(T|X)-\hat{l}_n(T|X) \xrightarrow{P} 0$ and from Condition 3, we have $\hat{l}_n(T|X)-l(T) \xrightarrow{P} 0$. Adding this up using the fact that convergence in probability is preserved under addition, we know that $\ln g_n(T|X) \xrightarrow{P} l(T)$.
Hence by taking the limit of (\ref{thmeqn1}) as $n$ grows, we have that $\lim_{n \rightarrow \infty} l(T_n|X) \geq l(T^*_{\textrm{MLE}})$. Since $T^*_{\textrm{MLE}}$ is optimal for $l(T)$ by definition, and by Condition 4, we conclude that $T_n$ stays in $[T^*_{\textrm{MLE}}]$ when $n$ is sufficiently large.
\section*{Appendix D. Optimal Density for the Likelihood Function}
Denote the pointwise density estimate at $x$ to be $\hat{f}_{n,x}=\frac{n_{x}}{n}$. Denote the density estimate for all points within leaf $l$ similarly as $\hat{f}_{n,l}=\frac{\sum_{j\in \textrm{leaf }l} n_j}{n V_{l}}$.
The true density from which the data are assumed to be generated is denoted $D$. We assume that $D$ arises from a tree over the input space (otherwise we would be back to standard analysis, where the proof is well-known).
\noindent \textbf{Lemma 1}: Any tree achieving the maximum likelihood on the training data has pointwise density equal to $\hat{f_n}(x)$. This means for any $l$ in the tree and for any $x$, $\hat{f}_{n,l}(x)=\hat{f}_{n,x}(x)$.
\noindent\textbf{Proof:}
We will show that the pointwise histogram becomes better than using the tree if the tree is not correct. This is in some sense a version of a well-known result that the maximum likelihood is the pointwise maximum likelihood. We will show: \begin{equation*} \prod_{j\in \textrm{leaf } l} \hat{f}_{n,j}^{n_j} \geq \hat{f}_{n,l}^{\sum_{j\in \textrm{leaf } l} n_j}. \end{equation*} By taking logarithms, this reduces to
\begin{align} &\sum_{j\in \textrm{leaf } l} n_j \log \hat{f}_{n,j} \geq \sum_{j\in \textrm{leaf } l} n_j \log \hat{f}_{n,l} \label{aftertakinglog}\\ &\sum_{j\in \textrm{leaf } l} n_j \log \left( \frac{n_j}{n} \right) \geq \sum_{j\in \textrm{leaf } l} n_j \log \left( \frac{\sum_{m\in \textrm{leaf } l } n_m}{n V_l}\right) \nonumber\\ &\sum_{j\in \textrm{leaf } l} n_j \log n_j \geq \sum_{j\in \textrm{leaf } l} n_j \log \left( \frac{\sum_{m\in \textrm{leaf } l} n_m}{ V_l}\right) \nonumber \\ &\sum_ {j\in \textrm{leaf } l} n_j\log \left( \frac{n_j}{\sum_{m\in \textrm{leaf } l } n_m} \right) \nonumber \\ &\geq \sum_{m\in \textrm{leaf } l} n_j \log \left( \frac{1}{ \sum_{m\in \textrm{leaf } l} V_m} \right) \nonumber \\ &\sum_{j\in \textrm{leaf } l} \frac{n_j}{\sum_{m\in \textrm{leaf } l } n_m} \log \left( \frac{n_j}{\sum_{m\in \textrm{leaf } l } n_m}\right) \nonumber \\ &\geq \sum_{j\in \textrm{leaf } l } \frac{n_j}{\sum_{m\in \textrm{leaf } l} n_m}\log \left( \frac{1}{ \sum_{m\in \textrm{leaf } l } V_m}\right).\nonumber \end{align}
We know the last equation is true since this is just Gibb's inequality. Hence we have proven that the statement is true.
To avoid singularity, we separately consider the case when one of the $\hat{f}_{n,j}=0$. For a particular value of $q$, if $\hat{f}_{n,q}=0$, then $n_q=0$ by definition. Hence, if we include a new $x$ within the leaf that has no training examples, we will find that the left hand side term of (\ref{aftertakinglog}) remains the same but since the volume increases when we add the new point, the quantity on the right decreases. Hence the inequality still holds.
\end{document}
|
arXiv
|
{
"id": "1510.06779.tex",
"language_detection_score": 0.7857507467269897,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{An unconditionnally stable pressure correction scheme for compressible barotropic {N}avier-{S}tokes equations}
\author{T. Gallou\"et} \address{Universit\'e de Provence, France ([email protected])}
\author{L. Gastaldo} \address{Institut de Radioprotection et de S\^{u}ret\'{e} Nucl\'{e}aire (IRSN) ([email protected])}
\author{R. Herbin} \address{Universit\'e de Provence, France ([email protected])}
\author{J.C. Latch\'e} \address{Institut de Radioprotection et de S\^{u}ret\'{e} Nucl\'{e}aire (IRSN) ([email protected])}
\begin{abstract} We present in this paper a pressure correction scheme for barotropic compressible Navier-Stokes equations, which enjoys an unconditional stability property, in the sense that the energy and maximum-principle-based a priori estimates of the continuous problem also hold for the discrete solution. The stability proof is based on two independent results for general finite volume discretizations, both interesting for their own sake: the $L^2$-stability of the discrete advection operator provided it is consistent, in some sense, with the mass balance and the estimate of the pressure work by means of the time derivative of the elastic potential. The proposed scheme is built in order to match these theoretical results, and combines a fractional-step time discretization of pressure-correction type to a space discretization associating low order non-conforming mixed finite elements and finite volumes. Numerical tests with an exact smooth solution show the convergence of the scheme. \end{abstract}
\subjclass{35Q30,65N12,65N30,76M125}
\keywords{Compressible Navier-Stokes equations, pressure correction schemes}
\maketitle
\section{Introduction}
The problem addressed in this paper is the system of the so-called barotropic compressible Navier-Stokes equations, which reads: \begin{equation}
\left| \begin{array}{l} \displaystyle \frac{\partial\,\rho}{\partial t}+\nabla\cdot(\rho\, u)=0 \\[2ex] \displaystyle \frac{\partial}{\partial t}(\rho\, u)+\nabla\cdot(\rho\, u\otimes u)+\nabla p -\nabla\cdot\tau(u)=f \\[2ex] \displaystyle \rho=\varrho(p) \end{array} \right. \label{nsb}\end{equation} where $t$ stands for the time, $\rho$, $u$ and $p$ are the density, velocity and pressure in the flow, $f$ is a forcing term and $\tau(u)$ stands for the shear stress tensor. The function $\varrho(\cdot)$ is the equation of state used for the modelling of the particular flow at hand, which may be the actual equation of state of the fluid or may result from assumptions concerning the flow; typically, laws as $\rho=p^{1/\gamma}$, where $\gamma$ is a coefficient specific to the fluid considered, are obtained by making the assumption that the flow is isentropic. This system of equations is posed over $\Omega\times (0,T)$, where $\Omega$ is a domain of $\xR^d$, $d\leq 3$ supposed to be polygonal ($d=2$) or polyhedral ($d=3$), and the final time $T$ is finite. It must be supplemented by boundary conditions and by an initial condition for $\rho$ and $u$.
The development of pressure correction techniques for compressible Navier-Stokes equations may be traced back to the seminal work of Harlow and Amsden \cite{har-68-num, har-71-num} in the late sixties, who developped an iterative algorithm (the so-called ICE method) including an elliptic corrector step for the pressure. Later on, pressure correction equations appeared in numerical schemes proposed by several researchers, essentially in the finite-volume framework, using either a collocated \cite{pat-87-bar,dem-93-col,kob-96-cha,pol-97-pre,iss-98-pre,mou-01-hig} or a staggered arrangement \cite{cas-84-pre, iss-85-sol, iss-86-com, kar-89-pre, bij-98-uni, col-99-pro, van-01-sta, wal-02-sem, wen-02-mac, van-03-con, vid-06-sup} of unknowns; in the first case, some corrective actions are to be foreseen to avoid the usual odd-even decoupling of the pressure in the low Mach number regime. Some of these algorithms are essentially implicit, since the final stage of a time step involves the unknown at the end-of-step time level; the end-of-step solution is then obtained by SIMPLE-like iterative processes \cite{van-87-seg, kar-89-pre, dem-93-col, kob-96-cha, pol-97-pre, iss-98-pre, mou-01-hig}. The other schemes \cite{iss-85-sol, iss-86-com, pat-87-bar, bij-98-uni, col-99-pro, wes-01-pri, van-01-sta, wen-02-mac, van-03-con, vid-06-sup} are predictor-corrector methods, where basically two steps are performed in sequence: first a semi-explicit decoupled prediction of the momentum or velocity (and possibly energy, for non-barotropic flows) and, second, a correction step where the end-of step pressure is evaluated and the momentum and velocity are corrected, as in projection methods for incompressible flows (see \cite{cho-68-num, tem-69-sur} for the original papers, \cite{mar-98-nav} for a comprehensive introduction and \cite{gue-06-ove} for a review of most variants). The Characteristic-Based Split (CBS) scheme (see \cite{nit-06-cha} for a recent review or \cite{zie-95-gen} for the seminal paper), developped in the finite-element context, belongs to this latter class of methods.
Our aim in this paper is to propose and study a simple non-iterative pressure correction scheme for the solution of \eqref{nsb}. In addition, this method is designed so as to be stable in the low Mach number limit, since our final goal is to apply it to simulate by a drift-flux approach a class of bubbly flows encountered in nuclear safety studies, where pure liquid (incompressible) and pure gaseous (compressible) zones may coexist. To this purpose, we use a low order mixed finite element approximation, which meets the two following requirements: to allow a natural discretization of the viscous terms and to provide a spatial discretization that is intrinsically stable (\ie \ without the adjunction of stabilization terms to circumvent the so-called {\it inf-sup} or BB condition) in the incompressible limit.
In this work, a special attention is payed to stability issues. To be more specific, let us recall the {\it a priori} estimates associated to problem \eqref{nsb} with a zero forcing term, {\it i.e.} estimates which should be satisfied by any possible regular solution \cite{pll-98-mat, fei-04-dyn, nov-04-int}: \begin{equation}
\left| \quad \begin{array}{llll} (i) & \displaystyle \rho(x,t) > 0, && \displaystyle \forall x \in \Omega,\ \forall t \in (0,T) \\[2ex] (ii) & \displaystyle \int_\Omega \rho(x,t)\ {\rm d} x = \int_\Omega \rho(x,0)\ {\rm d} x, && \displaystyle \forall t \in (0,T) \\[3ex] (iii) \quad & \displaystyle \frac{1}{2}\ \frac{d}{dt}\int_{\Omega}\rho(x,t)\, u(x,t)^{2} \, {\rm d}x +\frac{d}{dt}\int_{\Omega} \rho(x,t) \,P(\rho(x,t)) \, {\rm d}x \\ &
\displaystyle +\int_{\Omega} \tau (u(x,t)):\nabla u(x,t) \, {\rm d}x = 0, & \qquad & \forall t \in (0,T) \end{array} \right. \label{apriori-e}\end{equation} In the latter relation, $P$, the "elastic potential", is a function derived from the equation of state, which satisfies: \begin{equation} P'(z)=\frac{\wp(z)}{z^2} \label{elasticpot_gene} \end{equation} where $\wp (\cdot)$ is the inverse function of $\varrho(\cdot)$, \ie\ the function giving the pressure as a function of the density. The usual choice is, provided that this expression makes sense: \begin{equation} P(z)=\int_0^z \frac{\wp(s)}{s^2}\ {\rm d}s \label{elasticpot} \end{equation} For these estimates to hold, the condition \eqref{apriori-e}-$(i)$ must be satisfied by the initial condition; note that a non-zero forcing term $f$ in the momentum balance would make an additional term appear at the right hand side of relation \eqref{apriori-e}-$(iii)$. This latter estimate is obtained from the second relation of \eqref{nsb} (\ie\ the momentum balance) by taking the scalar product by $u$ and integrating over $\Omega$. This computation then involves two main arguments which read: \begin{equation} \begin{array}{llll} (i) & \mbox{Stability of the advection operator:} \quad & \displaystyle \int_\Omega \left[ \frac{\partial}{\partial t}(\rho\, u)+\nabla\cdot(\rho\, u\otimes u) \right] \cdot u \, {\rm d}x= \frac{1}{2}\ \frac{d}{dt}\int_{\Omega}\rho\, u^2 \, {\rm d}x \\[3ex] (ii) & \mbox{Stability due to the pressure work:} \quad & \displaystyle \int_\Omega -p \, \nabla \cdot u \, {\rm d}x=\frac{d}{dt}\int_{\Omega} \rho(x,t) \,P(\rho(x,t)) \, {\rm d}x \end{array} \label{stab-arg}\end{equation} Note that the derivation of both relations make use of the fact that the mass balance equation holds in a crucial way.
This paper is organized as follows.
As a first step, we derive a bound similar to \eqref{stab-arg}-$(ii)$ for a given class of spatial discretizations; the latter are introduced in section \ref{subsec:disc} and the desired stability estimate (theorem \ref{VF2}) is stated and proven in section \ref{subsec:pwork}. We then show that this result allows to prove the existence of a solution for a fairly general class of discrete compressible flow problems. Section \ref{sec:darcy} gathers this whole study, and constitutes the first part of this paper.
In a second part (section \ref{sec:scheme}), we turn to the derivation of a pressure correction scheme the solution of which satisfies the whole set of {\it a priori} estimates \eqref{apriori-e}. To this purpose, besides theorem \ref{VF2}, we need as a second key ingredient a discrete version of the bound \eqref{stab-arg}-$(i)$ relative to the stability of the advection operator, which is stated and proven in section \ref{subsec:conv} (theorem \ref{VF1}). We then derive a fully discrete algorithm which is designed to meet the assumptions of these theoretical results, and establish its stability. Some numerical experiments show that this scheme seems in addition to present, when the solution is smooth, the convergence properties which can be expected, namely first order in time convergence for all the variables and second order in space in $\xLtwo$ and discrete $\xLtwo$ norm for the velocity and the pressure, respectively.
\section{Analysis of a class of discrete problems} \label{sec:darcy}
The class of problems addressed in this section can be seen as the class of discrete systems obtained by space discretization by low-order non-conforming finite elements of continuous problems of the following form: \begin{equation}
\left| \begin{array}{ll} \displaystyle A \, u + \nabla p = f & \mbox{ in } \Omega \\[2ex] \displaystyle \frac{\varrho(p)-\rho^\ast}{\delta t}+ \nabla \cdot \left( \varrho(p)\, u \right)=0 & \mbox{ in } \Omega \\[2ex] \displaystyle u=0 & \mbox{ on } \partial \Omega \end{array} \right . \label{pbdisc-cont}\end{equation} where the forcing term $f$ and the density field $\rho^\ast$ are known quantities, and $A$ stands for an abstract elliptic operator. The unknown of the problem are the velocity $u$ and the pressure $p$; the function $\varrho(\cdot)$ stands for the equation of state. The domain $\Omega$ is a polygonal ($d=2$) or polyhedral ($d=3$) open, bounded and connected subset of $\xR^d$. Of course, at the continuous level, this statement of the problem should be completed by a precise definition of the functional spaces in which the velocity and the pressure are searched for, together with regularity assumptions on the data. This is out of context in this section, as system \eqref{pbdisc-cont} is given only to fix ideas and we restrict ourselves here to prove some mathematical properties of the discrete problem, namely to establish some {\it a priori} estimates for its solution and to prove that this nonlinear problem admits some solution for fairly general equations of state.
This section is organized as follows. We begin by describing the considered discretization and precisely stating the discrete problem at hand. Then we prove, for the particular discretization at hand, a fundamental result which is a discrete analogue of the elastic potential identity. The next section is devoted to the proof of the existence of a solution, and we finally conclude by giving some practical examples of application of the abstract theory developped here.
\subsection{The discrete problem}\label{subsec:disc}
Let ${\cal M}$ be a decomposition of the domain $\Omega$ either into convex quadrilaterals ($d=2$) or hexahedrons ($d=3$) or in simplices.
By ${\cal E}$ and ${\cal E}(K)$ we denote the set of all $(d-1)$-edges $\sigma$ of the mesh and of the element $K \in {\cal M}$ respectively. The set of edges included in the boundary of $\Omega$ is denoted by ${\cal E}_{{\rm ext}}$ and the set of internal ones (\ie\ ${\cal E} \setminus {\cal E}_{{\rm ext}}$) is denoted by ${\cal E}_{{\rm int}}$. The decomposition ${\cal M}$ is supposed to be regular in the usual sense of the finite element literature (e.g. \cite{cia-91-bas}), and, in particular, ${\cal M}$ satisfies the following properties: $ \bar\Omega=\bigcup_{K\in {\cal M}} \bar K$; if $K,\,L \in {\cal M},$ then $\bar K \cap \bar L=\emptyset$ or $\bar K\cap \bar L$ is a common face of $K$ and $L$, which is denoted by $K|L$. For each internal edge of the mesh $\sigma=K|L$, $n_{KL}$ stands for the normal vector of $\sigma$, oriented from $K$ to $L$. By $|K|$ and $|\sigma|$ we denote the measure, respectively, of $K$ and of the edge $\sigma$.
The spatial discretization relies either on the so-called "rotated bilinear element"/$P_0$ introduced by Rannacher and Turek \cite{ran-92-sim} for quadrilateral of hexahedric meshes, or on the Crouzeix-Raviart element (see \cite{cro-73-con} for the seminal paper and, for instance, \cite[p. 83--85]{ern-05-aid} for a synthetic presentation) for simplicial meshes. The reference element $\widehat K$ for the rotated bilinear element is the unit $d$-cube (with edges parallel to the coordinate axes); the discrete functional space on $\widehat K$ is $\tilde{Q}_{1}(\widehat K)^d$, where $\tilde{Q}_{1}(\widehat K)$ is defined as follows: \[ \tilde{Q}_{1}(\widehat K)= {\rm span}\left\{1,\,(x_{i})_{i=1,\ldots,d},\,(x_{i}^{2}-x_{i+1}^{2})_{i=1,\ldots,d-1}\right\} \] The reference element for the Crouzeix-Raviart is the unit $d$-simplex and the discrete functional space is the space $P_1$ of affine polynomials. For both velocity elements used here, the degrees of freedom are determined by the following set of nodal functionals: \begin{equation} \displaystyle \left\{F_{\sigma,i},\ \sigma \in {\cal E}(K),\, i=1,\ldots,d\right\},
\qquad F_{\sigma,i}(v)=|\sigma|^{-1}\int_{\sigma} v_{i} \, {\rm d} \gamma \label{vdof}\end{equation} The mapping from the reference element to the actual one is, for the Rannacher-Turek element, the standard $Q_1$ mapping and, for the Crouzeix-Raviart element, the standard affine mapping. Finally, in both cases, the continuity of the average value of discrete velocities (\ie, for a discrete velocity field $v$, $F_{\sigma,i}(v)$) across each face of the mesh is required, thus the discrete space $W_{h}$ is defined as follows: \[ \begin{array}{ll} \displaystyle W_h = & \displaystyle \lbrace
\ v_h\in L^{2}(\Omega)\,:\, v_h|_K \in\tilde{Q}_{1}(K)^d,\,\forall K\in {\cal M};\ F_{\sigma,i}(v_h) \mbox{ continuous across each edge } \sigma \in {{\cal E}_{{\rm int}}},\ 1\leq i \leq d \,; \\[1ex] & \displaystyle \ \ F_{\sigma,i}(v_h)=0,\ \forall \sigma \in {\cal E}_{{\rm ext}},\ 1\leq i \leq d\ \rbrace \end{array} \] For both Rannacher-Turek and Crouzeix-Raviart discretizations, the pressure is approximated by the space $L_{h}$ of piecewise constant functions: \[
L_h=\left\{q_h\in L^{2}(\Omega)\,:\, q_h|_K=\mbox{ constant},\,\forall K\in {\cal M}\right\} \] Since only the continuity of the integral over each edge of the mesh is imposed, the velocities are discontinuous through each edge; the discretization is thus nonconforming in $H^1(\Omega)^d$. These pairs of approximation spaces for the velocity and the pressure are \textit{inf-sup} stable, in the usual sense for "piecewise $\xHone$" discrete velocities, \ie\ there exists $c_{\rm i}>0$ independent of the mesh such that: \[ \forall p \in L_h, \qquad \sup_{v \in W_h} \frac{\displaystyle \int_{\Omega,h} p \, \nabla \cdot v \, {\rm d}x}{\normHb{v}} \geq c_{\rm i} \normLd{p-\bar p} \] where $\bar p$ is the mean value of $p$ over $\Omega$, the symbol $\displaystyle \int_{\Omega,h}$ stands for $\displaystyle \sum_{K\in{\cal M}} \int_K$ and $\normHb{\cdot}$ stands for the broken Sobolev $\xHone$ semi-norm: \[
\normHbd{v}=\sum_{K\in {\cal M}} \int_K |\nabla v |^2 \, {\rm d}x=\int_{\Omega,h}| \nabla v |^2 \, {\rm d}x \]
From the definition \eqref{vdof}, each velocity degree of freedom can be univoquely associated to an element edge. We take benefit of this correspondence by using hereafter, somewhat improperly, the expression "velocity on the edge $\sigma$" to name the velocity vector defined by the degrees of freedom of the velocity components associated to $\sigma$. In addition, the velocity degrees of freedom are indexed by the number of the component and the associated edge, thus the set of velocity degrees of freedom reads: \[ \lbrace v_{\sigma, i},\ \sigma \in {\cal E}_{{\rm int}},\ 1 \leq i \leq d \rbrace \] We define $v_\sigma=\sum_{i=1}^d v_{\sigma, i} e_i $ where $e_i$ is the $i^{th}$ vector of the canonical basis of $\xR^d$. We denote by $\varphi_\sigma^{(i)}$ the vector shape function associated to $v_{\sigma, i}$, which, by the definition of the considered finite elements, reads: \[ \varphi_\sigma^{(i)}=\varphi_\sigma \, e_i \] where $\varphi_\sigma$ is a scalar function. Similarly, each degree of freedom for the pressure is associated to a mesh $K$, and the set of pressure degrees of freedom is denoted by $\lbrace p_K,\ K \in {\cal M} \rbrace$.
For any $K \in {\cal M}$, let $\rho^\ast_K$ be a quantity approximating a known density $\rho^\ast$ on $K$. The family of real numbers $(\rho^\ast_K)_{K\in{\cal M}}$ is supposed to be positive. The discrete problem considered in this section reads: \begin{equation}
\left| \begin{array}{ll} \displaystyle a(u,\varphi_\sigma^{(i)}) - \int_{\Omega,h} p\ \nabla \cdot \varphi_\sigma^{(i)} = \int_\Omega f \cdot \varphi_\sigma^{(i)} \, {\rm d}x \qquad & \displaystyle
\forall \sigma\in{\cal E}_{{\rm int}} \ (\sigma=K|L),\ 1 \leq i \leq d \\[2ex] \displaystyle
|K|\ \frac{\varrho(p_K)-\rho^\ast_K}{\delta t} + \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K}^+\, \varrho(p_K) - \mathrm{v}_{\sigma,K}^-\ \varrho(p_L)=0 \qquad & \displaystyle \forall K \in {\cal M} \end{array}\right . \label{pbdisc-disc}\end{equation}
where $\mathrm{v}_{\sigma,K}^+$ and $\mathrm{v}_{\sigma,K}^-$ stands respectively for $\mathrm{v}_{\sigma,K}^+ = \max (\mathrm{v}_{\sigma,K},\ 0)$ and $\mathrm{v}_{\sigma,K}^- = \max ( -\mathrm{v}_{\sigma,K},\ 0)$ with $\mathrm{v}_{\sigma,K}=|\sigma|\, u_\sigma \cdot n_{KL}=\mathrm{v}_{\sigma,K}^+- \mathrm{v}_{\sigma,K}^-$. The first equation is the standard finite element discretization of the first relation of \eqref{pbdisc-cont}, provided the following identity holds: \[ \forall v\in W_h,\ \forall w \in W_h,\qquad a(v,w)=\int_\Omega Av \cdot w \, {\rm d}x \] As the pressure is piecewise constant, the finite element discretization of the second relation of \eqref{pbdisc-cont}, \ie\ the mass balance, is similar to a finite volume formulation, in which we introduce the standard first-order upwinding. The bilinear form $a(\cdot,\cdot)$ is supposed to be elliptic on $W_h$, \ie\ to be such that the following property holds: \[ \exists c_{\rm a} >0 \mbox{ such that, } \forall v \in W_h,\qquad a(u,u) \geq c_{\rm a} \normsd{u} \] where $\norms{\cdot}$ is a norm over $W_h$. We denote by $\normS{\cdot}$ its dual norm with respect to the $\xLtwo(\Omega)^d$ inner product, defined by: \[ \forall v \in W_h, \qquad \normS{v}=\sup_{w \in W_h} \frac{\displaystyle \int_\Omega v \cdot w \, {\rm d}x}{\norms{w}} \]
\subsection{On the pressure control induced by the pressure forces work}\label{subsec:pwork}
The aim of this subsection is to prove that the discretization at hand satisfies a stability bound which can be seen as the discrete analogue of equation \eqref{stab-arg}-$(ii)$, which we recall here: \[ \int_\Omega p \, \nabla \cdot u \, {\rm d}x=\frac{d}{dt}\int_{\Omega} \rho \,P(\rho) \, {\rm d}x, \qquad \mbox{where } P(\cdot) \mbox{ is such that } P'(z)=\frac{\wp(z)}{z^2} \] The formal computation which allows to derive this estimate in the continuous setting is the following. The starting point is the mass balance, which is multiplied by $[\rho P(\rho)]'$: \[ [\rho P(\rho)]' \left[ \frac{\partial \rho}{\partial t} + \nabla \cdot (\rho u) \right]=0 \] This relation yields: \begin{equation} \frac{\partial [\rho P(\rho)]}{\partial t} + [\rho P(\rho)]' \left[ u \cdot \nabla \rho + \rho \nabla \cdot u \right]=0 \label{elP-step1}\end{equation} And thus: \[ \frac{\partial [\rho P(\rho)]}{\partial t} + u \cdot \nabla [\rho P(\rho)] + [\rho P(\rho)]' \rho \nabla \cdot u =0 \] Developping the derivative, we get: \begin{equation} \frac{\partial [\rho P(\rho)]}{\partial t} + \underbrace{u \cdot \nabla [\rho P(\rho)] + \rho P(\rho) \nabla \cdot u}_{\displaystyle \nabla \cdot (\rho P(\rho)\, u)} + \underbrace{\rho^2 [P(\rho)]' \nabla \cdot u}_{\displaystyle p \nabla \cdot u} =0 \label{elP-step2}\end{equation} and the result follows by integration in space. We are going here to reproduce this computation at the discrete level.
\begin{thrm}[Stability due to the pressure work]\label{VF2} Let the elastic potential $P$ defined by \eqref{elasticpot_gene} be such that the function $z \mapsto z\, P(z)$ is once continuously differentiable and strictly convex. Let $(\rho_K)_{K\in{\cal M}}$ and $(p_K)_{K\in{\cal M}}$ satisfy the second relation of \eqref{pbdisc-disc}. Then the following estimate holds: \begin{equation}
- \int_\Omega p \, \nabla \cdot u \, {\rm d}x=\sum_{K\in{\cal M}} - p_K \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K} \geq
\frac{1}{\delta t}\ \sum_{K\in{\cal M}} |K| \left[ \rho_K \, P(\rho_K) - \rho_K^\ast \, P(\rho_K^\ast)\right] \label{pot-el}\end{equation} \end{thrm}
\begin{proof} Let us write the divergence term in the discrete mass balance over $K$ (\ie\ the second relation of \eqref{pbdisc-disc}) under the following form: \[
\sum_{\sigma=K|L} \rho_\sigma \, \mathrm{v}_{\sigma,K} \] where $\rho_\sigma$ is either $\rho_K=\varrho(p_K)$ if $\mathrm{v}_{\sigma,K} \geq 0$ or $\rho_L=\varrho(P_L)$ if $\mathrm{v}_{\sigma,K} \leq 0$. Multiplying this term by the derivative with respect to $\rho$ of $\rho\, P(\rho)$ computed at $\rho_K$, denoted by $[\rho\, P(\rho)]'_{\rho_K}$, we obtain: \[
T_{{\rm div},K}= [\rho\, P(\rho)]'_{\rho_K} \sum_{\sigma=K|L} \rho_\sigma \, \mathrm{v}_{\sigma,K}
= [\rho\, P(\rho)]'_{\rho_K} \left[ \sum_{\sigma=K|L} \ (\rho_\sigma-\rho_K) \, \mathrm{v}_{\sigma,K}+
\rho_K \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K} \right] \]
This relation is a discrete equivalent to equation \eqref{elP-step1}: up to the multiplication by $1/|K|$, the first summation in the right hand side is the analogue of $u \cdot \nabla \rho$ and the second one to $\rho\,\nabla \cdot u$. Developping the derivative, we obtain the equivalent of relation \eqref{elP-step2}: \begin{equation} T_{{\rm div},K}=
[\rho\, P(\rho)]'_{\rho_K} \sum_{\sigma=K|L} (\rho_\sigma-\rho_K) \, \mathrm{v}_{\sigma,K}
+ \rho_K\, P(\rho_K) \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K}
+ \rho_K^2\, P'(\rho_K) \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K} \label{elPd1}\end{equation}
By definition \eqref{elasticpot_gene} of $P$, the last term reads $p_K \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K}$. The process will be completed if we put the first two terms in divergence form. To this end, let us sum up the $T_{{\rm div},K}$ and reorder the summation: \begin{equation}
\sum_{K\in{\cal M}} T_{{\rm div},K}= \sum_{K\in{\cal M}} p_K \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K} + \sum_{\sigma\in{\cal E}_{{\rm int}}} T_{{\rm div},\sigma} \label{elPd2}\end{equation}
where, if $\sigma=K|L$: \[ T_{{\rm div},\sigma}= \mathrm{v}_{\sigma,K}\ \left[ \rho_K\, P(\rho_K) + [\rho\, P(\rho)]'_{\rho_K} (\rho_\sigma-\rho_K) - \rho_L\, P(\rho_L) - [\rho\, P(\rho)]'_{\rho_L} (\rho_\sigma-\rho_L) \right] \]
In this relation, there are two possible choices for the orientation of $\sigma$, \ie\ $K|L$ or $L|K$, where $K$ and $L$ are two cells such that $\sigma=\bar K\cap \bar L$; we choose this orientation to have $\mathrm{v}_{\sigma,K} \geq 0$. Let $\bar \rho_\sigma$ be defined by: \begin{equation}
\left| \begin{array}{ll} \mbox{if }\rho_K \neq \rho_L \ :\qquad & \displaystyle \rho_K\, P(\rho_K) + [\rho\, P(\rho)]'_{\rho_K} (\bar \rho_\sigma-\rho_K) = \rho_L\, P(\rho_L) + [\rho\, P(\rho)]'_{\rho_L} (\bar \rho_\sigma-\rho_L) \\[2ex] \mbox{otherwise}: & \bar \rho_\sigma=\rho_K=\rho_L \end{array} \right. \label{defrhob}\end{equation} As the function $z \mapsto z\, P(z)$ is supposed to be once continuously differentiable and strictly convex, the technical lemma \ref{rhosigma} proven hereafter applies and $\bar \rho_\sigma$ is uniquely defined and satisfies $\bar\rho_{\sigma} \in [\min(\rho_{K},\rho_{L}), \, \max(\rho_{K},\rho_{L})]$. By definition, the choice $\rho_\sigma=\bar \rho_\sigma$ is such that the term $T_{{\rm div},\sigma}$ vanishes, which means that, by this definition, we would indeed have obtained that the first two terms of equation \eqref{elPd1} are a conservative approximation of the quantity $\nabla \cdot \rho P(\rho) u$ appearing in equation \eqref{elP-step2}, with the following expression for the flux: \[ \begin{array}{l} \displaystyle F_{\sigma,K}= [\rho P(\rho)]_\sigma \, \mathrm{v}_{\sigma,K}, \quad \mbox{with:} \\[2ex] \displaystyle \hspace{10ex} [\rho P(\rho)]_\sigma= \rho_K\, P(\rho_K) + [\rho\, P(\rho)]'_{\rho_K} (\bar \rho_\sigma-\rho_K) = \rho_L\, P(\rho_L) + [\rho\, P(\rho)]'_{\rho_L} (\bar \rho_\sigma-\rho_L) \end{array} \] Whatever the choice for $\rho_\sigma$ may be, we have: \[ T_{{\rm div},\sigma}= \mathrm{v}_{\sigma,K}\ (\rho_\sigma - \bar \rho_\sigma)\ ([\rho\, P(\rho)]'_{\rho_K}-[\rho\, P(\rho)]'_{\rho_L} ) \] With the orientation taken for $\sigma$, an upwind choice yields: \[ T_{{\rm div},\sigma}= \mathrm{v}_{\sigma,K}\ (\rho_K - \bar \rho_\sigma)\ ([\rho\, P(\rho)]'_{\rho_K}-[\rho\, P(\rho)]'_{\rho_L} ) \] and, using the fact that $z \mapsto [\rho\, P(\rho)]'_{z}$ is an increasing fuction since $z \mapsto z\, P(z)$ is convex and that $\bar\rho_{\sigma} \in [\min(\rho_{K},\rho_{L}), \, \max(\rho_{K},\rho_{L})]$, it is easily seen that $T_{{\rm div},\sigma}$ is non-negative.
Multiplying by $[\rho\, P(\rho)]'_{\rho_K}$ the mass balance over each cell $K$ and summing for $K\in{\cal M}$ thus yields, by equation \eqref{elPd2}: \begin{equation}
- \sum_{K\in{\cal M}} p_K \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K} =
R + \sum_{K\in{\cal M}} |K|\ [\rho\, P(\rho)]'_{\rho_K}\ \frac{\rho_K-\rho^\ast}{\delta t} \label{elPd3}\end{equation} where $R$ is non-negative, and the result follows invoking once again the convexity of $z \mapsto z\, P(z)$. \end{proof} \begin{rmrk}[On a non-dissipative scheme] The preceding proof shows that, for a scheme to conserve the energy (\ie\ to obtain a discrete equivalent of \eqref{apriori-e}-$(iii)$), besides other arguments, the choice of $\bar \rho_\sigma$ given by \eqref{defrhob} for the density at the face of the control volume in the discretization of the flux in the mass balance seems to be mandatory; any other choice leads to an artificial dissipation in the work of the pressure forces. Note however that, this discretization being essentially of centered type, the positivity of the density is not warranted in this case. \end{rmrk}
In the course of the preceding proof, we used the following technical lemma. \begin{lmm}\label{rhosigma} Let $g(\cdot)$ be a strictly convex and once continuously derivable real function. Let $\sigma$ be an internal edge of the mesh separating the cells $K$ and $L$. Then the following relations: \begin{equation}
\left| \begin{array}{ll} \mbox{if }\rho_K \neq \rho_L \ :\qquad & \displaystyle g(\rho_{K}) + g'(\rho_{K})(\bar\rho_\sigma-\rho_{K}) = g(\rho_{L}) + g'(\rho_{L})(\bar\rho_\sigma-\rho_{L}) \\[2ex] \mbox{otherwise}: & \bar \rho_\sigma=\rho_K=\rho_L \end{array} \right. \label{defrhob2}\end{equation} uniquely defines the real number $\bar\rho_\sigma$. In addition, we have $\bar\rho_{\sigma} \in [\min(\rho_{K},\rho_{L}),\,\max(\rho_{K},\rho_{L})]$. \end{lmm}
\begin{proof} If $\rho_K=\rho_L$, there is nothing to prove. In the contrary case, without loss of generality, let us choose $K$ and $L$ in such a way that $\rho_K < \rho_L$. By reordering equation \eqref{defrhob2}, we get: \[ g(\rho_K) + g'(\rho_K)\,(\rho_L-\rho_K) -g(\rho_L)= (\bar \rho_\sigma-\rho_L)\,\left[ g'(\rho_L)-g'(\rho_K)\right] \] As, because $g(\cdot)$ is strictly convex, $g'(\rho_L)-g'(\rho_K)$ does not vanish, this equation proves that $\bar\rho_\sigma$ is uniquely defined. In addition, for the same reason, the left hand side of this relation is negative and $g'(\rho_L)-g'(\rho_K)$ is positive, thus we have $\bar \rho_\sigma < \rho_L$. Reordering in another way equation \eqref{defrhob2} yields: \[ g(\rho_L) + g'(\rho_L)\,(\rho_K-\rho_L) -g(\rho_K)= (\bar \rho_\sigma-\rho_K)\,\left[ g'(\rho_K)-g'(\rho_L) \right] \] which, considering the signs of the left hand side and of $g'(\rho_K)-g'(\rho_L)$, in turn implies $\bar \rho_\sigma > \rho_K$. \end{proof}
\subsection{Existence of a solution}
The aim of this section is to prove the existence of a solution to the discrete problem under study. It follows from a topological degree argument, linking by a homotopy the problem at hand to a linear system.
This section begins by a lemma which is used further to prove the strict positivity of the pressure.
\begin{lmm}\label{max} Let us consider the following problem: \begin{equation} \forall K \in {\cal M}, \qquad
|K|\,\frac{\varphi_1(p_K)-\varphi_{1}(p^\ast_K)}{\delta t}+\sum_{\sigma=K|L}\,\mathrm{v}_{\sigma,K}^+\,\varphi_2(p_K)-\mathrm{v}_{\sigma,K}^-\,\varphi_2(p_L)=0 \label{eqk} \end{equation} where $\varphi_1$ is an increasing function and $\varphi_2$ is a non-decreasing and non-negative function. Suppose that there exists $\bar p$ such that: \begin{equation} \varphi_{1}(\bar p)
+ \delta t \, \varphi_2(\bar p) \,\max_{K\in{\cal M}} \left[ 0,\frac{1}{|K|} \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K} \right] =\min_{K\in{\cal M}}[\varphi_1(p^\ast_{K})] \label{eqbarp} \end{equation} Then, $\forall K \in {\cal M}$, $p_K$ satisfies $ p_K \geq \bar{p}$. \end{lmm}
\begin{proof}
Let us assume that there exists a cell $\bar K$ such that $p_{\bar K}=\min_{K \in {\cal M}} p_K <\bar p$. Multiplying by $\delta t/|\bar K|$ the relation \eqref{eqk} written for $K=\bar K$, we get: \begin{equation} \varphi_1(p_{\bar K })
+\frac{\delta t}{|\bar K|}\,\sum_{\sigma=\bar K |L} \,(\mathrm{v}_{\sigma,K}^+\,\varphi_2(p_{\bar K})-\mathrm{v}_{\sigma,K}^-\,\varphi_2(p_L)) = \varphi_1(p^\ast_{\bar K}) \label{eqkbar} \end{equation} Then substracting from \eqref{eqbarp} we have: \[ \begin{array}{ll} \displaystyle \varphi_1(p_{\bar K})-\varphi_1(\bar p)
+\frac{\delta t}{|\bar K|} \, \sum_{\sigma=\bar K |L} \mathrm{v}_{\sigma,K}^+ \,\varphi_2 (p_{\bar K})-\mathrm{v}_{\sigma,K}^-\,\varphi_2(p_L) & \\ \displaystyle
- \delta t\,\varphi_2 (\bar p)\,\max_{K \in {\cal M}}\left[0,\frac{1}{|K|} \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K} \right] & \displaystyle =\varphi_1(p^\ast_{\bar K})-\min_{K \in {\cal M}}[\varphi_1(p^\ast_K)]\geq 0 \end{array} \] The previous relation can be written as $T_1 + T_2 + T_3 \geq 0$ with: \[ \begin{array}{l} T_1= \varphi_1(p_{\bar K})-\varphi_1(\bar p)
\\[2ex] \displaystyle T_2= \delta t \, \varphi_2(p_{\bar K})\left[ \sum_{\sigma=\bar K|L} \frac{1}{|\bar K|} \mathrm{v}_{\sigma,K} \right]
- \delta t \,\varphi_{2}(\bar p)\,\max_{K \in {\cal M}}\left[0,\sum_{\sigma=K|L}\frac{1}{|K|} \mathrm{v}_{\sigma,K} \right]
\\[4ex] \displaystyle T_3= \frac{\delta t}{|\bar K|}\,\sum_{\sigma=\bar K |L} \mathrm{v}_{\sigma,K}^- \,(\varphi_2(p_{\bar K})-\varphi_{2}(p_L)) \end{array} \]
As $\varphi_1(\cdot)$ is an increasing function and, by assumption, $p_{\bar K}<\bar p$, we have $T_1 <0$. Similarly, $0 \leq \varphi_2(p_{\bar K}) \leq \varphi_2(\bar p)$ and the discrete divergence over $\bar K$ (\ie\ $1/|K|\ \sum_{\sigma=\bar K|L} u_\sigma$) is necessarily smaller than the maximum of this quantity over the cells of the mesh, thus $T_2 <0$. Finally, as, by assumption, $p_{\bar K} \leq p_L$ for any neighbouring cell $L$ of $\bar K$, $\varphi_2(\cdot)$ is a non-decreasing function and $\mathrm{v}_{\sigma,K}^- \geq 0$, $T_3 \leq 0$. We thus obtain a contradiction with the initial hypothesis, which proves $p_K \geq \bar p,\ \forall K \in {\cal M}$. \end{proof}
We now state the abstract theorem which will be used hereafter; this result follows from standard arguments of the topological degree theory (see \cite{deimling} for an exposition of the theory and \cite{eym-98-an} for another utilisation for the same objective as here, namely the proof of existence of a solution to a numerical scheme).
\begin{thrm}[A result from the topological degree theory]\label{degre} Let $N$ and $M$ be two positive integers and $V$ be defined as follows: \[ V=\{ (x,y) \in \xR^N \times \xR^M \mbox{ such that } y>0\} \] where, for any real number $c$, the notation $y>c$ means that each component of $y$ is greater than $c$. Let $b\in \xR^N \times \xR^M$ and $f$ and $F$ be two continuous functions respectively from $V$ and $V\times[0,1]$ to $\xR^N \times \xR^M$ satisfying: \begin{enumerate} \item[(i)]$F(\cdot,\,1)=f(\cdot)$; \item[(ii)] $\forall \alpha \in [0,1]$, if $v$ is such that $F(v,\alpha)=b$ then $v \in W$, where $W$ is defined as follows: \[ W=\{ (x,y) \in \xR^N \times \xR^M \mbox{ s.t. } \Vert x \Vert < C_1 \mbox{ and } \epsilon< y < C_2\} \] with $C_1$, $C_2$ and $\epsilon$ three positive constants and $\Vert \cdot \Vert$ is a norm defined over $\xR^N$; \item[(iii)] the topological degree of $F(\cdot,0)$ with respect to $b$ and $W$ is equal to $d_0 \neq 0$. \end{enumerate} Then the topological degree of $F(\cdot,1)$ with respect to $b$ and $W$ is also equal to $d_0 \neq 0$; consequently, there exists at least a solution $v\in W$ such that $f(v)=b$. \end{thrm}
We are now in position to prove the existence of a solution to the considered discrete Darcy system, for fairly general equations of states.
\begin{thrm}[Existence of a solution] Let us suppose that the equation of state $\varrho(\cdot)$ is such that: \begin{enumerate} \item $\varrho(\cdot)$ is increasing, $\varrho(0)=0$ and $\displaystyle \lim_{z \rightarrow +\infty} \varrho(z) = +\infty$, \item there exists an elastic potential $P$ such that (\ref{elasticpot_gene}) holds, the function $z \mapsto z\, P(z)$ is once continuously differentiable and strictly convex and $z\,P(z)\geq -C_P,\ \forall z \in (0,+\infty)$, where $C_P$ is a non-negative constant. \end{enumerate} Then, there exists a solution $(u_{\sigma},p_{K})_{\sigma\in{\cal E}_{{\rm int}},\, K \in {\cal M}}$ to the discrete problem at hand \eqref{pbdisc-disc}. \label{existence_gene}\end{thrm}
\begin{proof} This proof is obtained by applying theorem \ref{degre}. Let $N=d\ {\rm card}({\cal E}_{{\rm int}})$, $M=\ {\rm card}({\cal M})$ and the mapping $F:\ V \times [0,1] \rightarrow \xR^N \times \xR^M$ be given by: \[ F(u,p,\alpha)=
\left| \begin{array}{ll} \displaystyle v_{\sigma,i},\ & \sigma \in {\cal E}_{{\rm int}},\ 1 \leq i \leq d \\[1ex] q_K,\ & K \in {\cal M} \end{array} \right. \] with: \begin{equation}
\left| \begin{array}{l} \displaystyle v_{\sigma,i}= a(u,\varphi_\sigma^{(i)}) - \alpha \int_{\Omega,h} p\ \nabla \cdot \varphi_\sigma^{(i)} \, {\rm d}x - \int_\Omega f \cdot \varphi_\sigma^{(i)} \, {\rm d}x
\\[2ex] \displaystyle q_K=\frac{|K|}{\delta t} \left[\varrho(p_K)-\varrho(p_K^\ast) \right]
+ \alpha \sum_{\sigma=K|L} \mathrm{v}_{\sigma,K}^+\, \varrho_\alpha(p_K) - \mathrm{v}_{\sigma,K}^-\, \varrho(p_L) \end{array} \right . \label{eqalpha} \end{equation} where, $\forall K \in {\cal M}$, $p_K^\ast$ is chosen such that $\rho_K^\ast=\varrho(p_K^\ast)$; note that, by assumption, $\varrho(\cdot)$ is one to one from $(0,+\infty)$ to $(0,+\infty)$, so the preceding definition make sense. The problem $F(u,p,1)=0$ is exactly the system \eqref{pbdisc-disc}.
The present proof is built as follows: by applying theorem \ref{VF2}, we obtain a control on $u$ in the discrete norm, uniform with respect to $\alpha$. Since we work on a finite dimensional space, we then obtain a control on $p$ in $\xLinfty$ by using the conservativity of the system of equations. For the same reason, the control on $u$ yields a bound in $\xLinfty$ of the value of the discrete divergence, which is shown to allow, by lemma \ref{max}, to bound $p$ away from zero independently of $\alpha$. The proof finally ends by examining the properties of the system $F(u,p,0)=0$.
\noindent \underline{\textit{Step 1}}: $\alpha\in (0,1]$, $\norms{\cdot}$ estimate for the velocity. \\[1ex] Setting $v_\sigma=0$ in \eqref{eqalpha}, multiplying the corresponding equation by $u_{\sigma,i}$ and summing over $\sigma \in {\cal E}_{{\rm int}}$ and $1 \leq i\leq d$ yields the following equation: \[ a(u,u) - \alpha \int_{\Omega,h} p \, \nabla \cdot u \, {\rm d}x = \int_\Omega f \cdot u \, {\rm d}x \] By a computation very similar to the proof of theorem \ref{VF2}, we see that, from the second relation of \eqref{eqalpha} with $q_K=0$: \[ -\alpha \int_{\Omega,h} p \, \nabla \cdot u \, {\rm d}x \geq
\frac 1 \delta t \sum_{K\in{\cal M}} |K|\ \left[\rho_K\,P(\rho_K) - \rho^\ast_K\,P(\rho\ast_K)\right] \] where $\rho_K=\varrho(p_k)$. By the stability of the bilinear form $a(\cdot,\cdot)$ and Young's inequality, we thus get: \begin{equation} \underbrace{\frac{c_{\rm a}} 2 \normsd{u}}_{\displaystyle T_1}
+ \underbrace{\frac 1 \delta t \sum_{K\in{\cal M}} |K|\ \rho_K\,P(\rho_K)}_{\displaystyle T_2} \leq \frac 1 {2 c_{\rm a}}\normSd{f}
+ \frac 1 \delta t \sum_{K\in{\cal M}} |K|\ \rho^\ast_K\,P(\rho\ast_K) \label{e_est}\end{equation}
By assumption, $T_2 \geq -C_p |\Omega|$ and we thus get the following estimate on the discrete norm of the velocity: \begin{equation} \norms{u}\leq C_{1} \label{est_u}\end{equation} where $C_1$ only depends on the data of the problem, \ie\ $\delta t$, $f$ and $\rho^\ast$ and not on $\alpha$.
\noindent \underline{\textit{Step 2}}: $\alpha\in (0,1]$, $\xLinfty$ estimate for the pressure. \\[1ex] Let us now turn to the estimate of the pressure. By conservativity of the discrete mass balance, it is easily seen that: \[
\sum_{K\in {\cal M}} |K|\ \varrho(p_K)=\sum_{K\in {\cal M}} |K|\ \rho^\ast_K \] As each term in the sum on the left hand side is non-negative, we thus have: \[
\forall K \in {\cal M}, \qquad \varrho(p_K) \leq \frac 1 {\min_{K\in{\cal M}} |K|}\ \sum_{K\in {\cal M}} |K|\ \rho^\ast_K \] which, as, by assumption, $\displaystyle \lim_{z \rightarrow +\infty} \varrho(z) = +\infty$, yields: \begin{equation} \forall K \in {\cal M}, \qquad p_K \leq C_2 \label{est_p_max}\end{equation} where $C_2$ only depends on the data of the problem.
\noindent \underline{\textit{Step 3}}: $\alpha\in (0,1]$, $p$ bounded away from zero. \\[1ex] Applying lemma \ref{max} with $\varphi_1(\cdot)=\varphi_2(\cdot)=\varrho(\cdot)$, we get: \[ \forall K \in {\cal M}, \qquad p_K \geq \bar p_\alpha \] where $\bar p_\alpha$ is given by: \[ \varrho(\bar p_\alpha)=\frac{\displaystyle \min_{K\in{\cal M}} \varrho(p^\ast)}
{\displaystyle 1+\delta t \max_{K \in {\cal M}}\left[0,\alpha \sum_{\sigma=K|L}\frac{1}{|K|}\,\mathrm{v}_{\sigma,K} \right]} \] Note that $\bar p_\alpha$ is well defined since, by assumption, $\varrho(\cdot)$ is one to one from $(0,+\infty)$ to $(0,+\infty)$. As $\alpha \leq 1$, we get: \[ \varrho(\bar p_\alpha) \geq \frac{\displaystyle \min_{K\in{\cal M}} \varrho(p^\ast)}
{\displaystyle 1+\delta t \max_{K \in {\cal M}}\left[0,\sum_{\sigma=K|L}\frac{1}{|K|}\,\mathrm{v}_{\sigma,K} \right]} \] and, by equivalence of norms in a finite dimensional space, the bound \eqref{est_u} also yields a bound in the $\xLinfty$ norm and, finally, an upper bound for the right hand side of this relation. We thus get, still since $\varrho(\cdot)$ is increasing on $(0,+\infty)$ that, $\forall \alpha \in (0,1]$, $\bar p_\alpha \geq \epsilon_1$, and, finally: \begin{equation} \forall K \in {\cal M}, \qquad p_K \geq \epsilon_1 \label{est_p_min}\end{equation} where $\epsilon$ only depends on the data.
\noindent \underline{\textit{Step 4}}: conclusion. \\[1ex] For $\alpha=0$, the system $F(u,p,0)=0$ reads: \[
\left| \begin{array}{ll} \displaystyle a(u,\varphi_\sigma^{(i)}) = \int_\Omega f \cdot \varphi_\sigma^{(i)} \, {\rm d}x \hspace{10ex} & \forall \sigma \in {\cal E}_{{\rm int}},\ 1\leq i \leq d \\[3ex] \displaystyle \varrho(p_K)=\varrho(p_K^\ast) & \forall K \in {\cal M} \end{array} \right . \] Since $\varrho(\cdot)$ is one to one from $(0,+\infty)$ to $(0,+\infty)$ and thanks to the stability of the bilinear form $a(\cdot,\cdot)$, this system has one and only one solution (which, for the pressure, reads of course $p_K=p_K^\ast,\ \forall K \in {\cal M}$), which satisfies: \begin{equation} \norms{u}\leq C_{3},\qquad \epsilon_2=\min_{K\in{\cal M}} p_K^\ast \leq p \leq \max_{K\in{\cal M}} p_K^\ast=C_4 \label{est_up}\end{equation} In addition, the jacobian of this system is block diagonal: the first block, associated to the first relation, is constant (this part of the system is linear) and non-singular; the second one, associated to the second relation, is diagonal, and each diagonal entry is equal to the derivative of $\varrho(\cdot)$, taken at the considered point. As the function $\varrho(\cdot)$ is increasing, this jacobian matrix does not vanish for the solution of the system. Let $W$ be defined by: \[ W=\{(u,p) \in \xR^N \times \xR^M \mbox{ such that } \norms{\tilde\rho}{u} < 2 \max(C_1,C_3) \mbox{ and } \frac 1 2\,\min(\epsilon_1,\epsilon_2)< p < 2 \max(C_1,C_3) \} \] The topological degree of $F(\cdot,\cdot,0)$ with respect to $0$ and $W$ does not vanish, and by inequalities \eqref{est_u}, \eqref{est_p_min}, \eqref{est_p_max} and \eqref{est_up}, theorem \ref{degre} applies, which concludes the proof. \end{proof}
\subsection{Some cases of application}
First of all, let us give some examples for the bilinear form $a(\cdot,\cdot)$, for which the theory developped in this work holds. The first of them is: \[ a(u,v)=\int_\Omega u \cdot v \, {\rm d}x, \qquad \norms{u}=\normLdv{u}, \qquad \normS{f}=\normLdv{f} \] This choice for $a(\cdot,\cdot)$ yields a discrete Darcy-like problem which is, up to numerical integration technicalities, the projection step arising in the pressure correction scheme which is considered in the present paper (see section \ref{sec:scheme}). Note that, in this case, the boundary condition $u\in\xHone_0(\Omega)^d$ does not make sense at the continuous level; in addition, the considered discretization is known to be not consistent enough to yield convergence (see remark \ref{non-consistent!} hereafter) for the Darcy problem.
The bilinear form associated to the Stokes problem provides another example of application. It may read in this case: \[
a(u,v)=\int_{\Omega,h} \nabla u \cdot \nabla v \, {\rm d}x, \qquad \norms{u}=|u|_{\xHone(\Omega)^d} \] or, without additional theoretical difficulties: \[ a(u,v)= \mu \int_{\Omega,h} \nabla u \cdot \nabla v \, {\rm d}x + \frac \mu 3 \int_{\Omega,h} (\nabla \cdot u) \ (\nabla \cdot v) \, {\rm d}x \] this latter form, where the real number $\mu >0$ is the viscosity, corresponding to the physical shear stress tensor expression for a compressible flow of a constant viscosity Newtonian fluid.
In addition, consider the steady Navier-Stokes equations, or, more generally, a time step of a (semi-)implicit time discretization of the unsteady Navier-Stokes equations, in which case $a(\cdot,\cdot)$ and $f$ reads: \[ \begin{array}{l} \displaystyle a(u,v) = \frac 1 \delta t \int_\Omega \rho u \cdot v \, {\rm d}x + \int_{\Omega,h} (\nabla \cdot \rho w \otimes u) \cdot v \, {\rm d}x + \mu \int_{\Omega,h} \nabla u \cdot \nabla v \, {\rm d}x + \frac \mu 3 \int_{\Omega,h} (\nabla \cdot u) \ (\nabla \cdot v) \, {\rm d}x \\[3ex] \displaystyle f= \frac 1 \delta t \rho^\ast u^\ast + f_0 \end{array} \] where the steady case is obtained with $\delta t=+\infty$, $f_0$ is the physical forcing term, $\rho^\ast$ and $u^\ast$ stands for known density and velocity fields and $w$ is an advection field, which may be $u$ itself (and must be $u$ in the steady case) or be derived from the velocity obtained at the previous time steps. Let us suppose that the following identity holds: \[ \frac 1 \delta t \int_\Omega (\rho u - \rho^\ast u^\ast) \cdot u \, {\rm d}x + \int_{\Omega,h} (\nabla \cdot \rho w \otimes u) \cdot u \, {\rm d}x
\geq \frac 1 {2 \delta t} \left[ \int_\Omega \rho |u|^2 - \int_\Omega \rho^\ast |u^\ast|^2 \right] \] which is the discrete counterpart of equation \eqref{stab-arg}-$(i)$. The algorithm considered in this paper provides an example where this condition is verified (see section \ref{sec:scheme}). Then the present theory applies with little modifications: in the proof of existence theorem \ref{existence_gene}, the right hand side of the preceding equation must be multiplied by the homotopy parameter $\alpha$ (an thus this term vanishes at $\alpha=0$, which yields the problem considered in step $4$ above); the (uniform with respect to $\alpha$) stability in step 1 stems from the diffusion term, and steps $2$ and $3$ remain unchanged.
Note that, in the steady state case, an additional constraint is needed for the problem to be well posed, namely to impose the total mass $M$ of fluid in the computational domain to a given value. This constraint can be simply enforced by solving an approximate mass balance which reads: \[
c(h) \, \left[ \rho - \frac M {|\Omega|} \right] + \nabla \cdot \rho u =0 \]
where $|\Omega|$ stands for the measure of $\Omega$, $h$ is the spatial discretization step and $c(h)>0$ must tend to zero with $h$, fast enough to avoid any loss of consistency. With this form of the mass balance, the theory developped here directly applies.
Examining now the assumptions for the equation of state in theorem \ref{existence_gene}, we see that our results hold with equations of state of the form: \[ \varrho(p)= p^{1/\gamma} \qquad \mbox{or, equivalently} \qquad \rho=p^\gamma, \qquad \mbox{where } \gamma > 1 \] In this case, the elastic potential is given by equation \eqref{elasticpot}, which yields: \[ P(\rho)=\frac 1 {\gamma -1}\ \rho^{\gamma -1},\qquad \rho P(\rho)= \frac 1 {\gamma -1}\ \rho^\gamma \quad (=\frac 1 {\gamma -1}\ p) \] The same conclusion still holds with $\gamma=1$ (\ie \ $p=\rho$), with $P(\rho)=\log(\rho)$ satisfying equation \eqref{elasticpot_gene}. The case $\gamma > 1$ is for instance encountered for isentropic perfect gas flows, whereas $\gamma=1$ corresponds to the isothermal case. It is worth noting that this range of application is larger than what is known for the continuous case, for which the existence of a solution is known only in the case $\gamma > d/2$ \cite{pll-98-mat, fei-04-dyn, nov-04-int}.
\section{A pressure correction scheme} \label{sec:scheme}
In this section, we build a pressure correction numerical scheme for the solution of compressible barotropic Navier-Stokes equations \eqref{nsb}, based on the low order non-conforming finite element spaces used in the previous section, namely the Crouzeix-Raviart or Rannacher-Turek elements.
The presentation is organized as follows. First, we write the scheme in the time semi-discrete setting (section \ref{subsec:sd}). Then we prove a general stability estimate which applies to the discretization by a finite volume technique of the convection operator (section \ref{subsec:conv}). The proposed scheme is built in such a way that the assumptions of this stability result hold (section \ref{subsec:mom}); this implies first to perform a prediction of the density, as a non-standard first step of the algorithm and, second, to discretize the convection terms in the momentum balance equation by a finite volume technique specially designed to this purpose. The discretization of the projection step (section \ref{subsec:proj}) also combines the finite element and finite volume methods, in such a way that the theory developped in section \ref{sec:darcy} applies; in particular, the proposed discretization allows to take benefit of the pressure or density control induced by the pressure work, \ie\ to apply theorem \ref{VF2}. The remaining steps of the algorithm are described in section \ref{subsec:renorm} and an overview of the scheme is given in section \ref{subsec:overview}. The following section (section \ref{subsec:stab}) is devoted to the proof of the stability of the algorithm. Finally, we shortly address some implementation difficulties (section \ref{subsec:impl}), then provide some numerical tests (section \ref{subsec:num-exp}) performed to assess the time and space convergence of the scheme.
\subsection{Time semi-discrete formulation}\label{subsec:sd}
Let us consider a partition $0=t_0 < t_1 <\ldots < t_n=T$ of the time interval $(0,T)$, which, for the sake of simplicity, we suppose uniform. Let $\delta t$ be the constant time step $\delta t=t_{k+1}-t_k$ for $k=0,1,\ldots,n$. In a time semi-discrete setting, the scheme considered in this paper reads: \vskip-.6cm \begin{eqnarray} && \mbox{1 -- Solve for } \tilde \rho^{n+1}\mbox{\ : \quad} \frac{\tilde \rho^{n+1}-\rho^{n}}{\delta t}+\nabla \cdot (\tilde \rho^{n+1}\,u^{n})=0 \label{stab-scheme-1}\\ && \mbox{2 -- Solve for } \tilde p^{n+1}\mbox{\ : \quad} -\nabla \cdot \left(\frac{1}{\tilde \rho ^{n+1}} \nabla \tilde p^{n+1}\right) = -\nabla \cdot \left(\frac{1}{\sqrt{\tilde \rho^{n+1}}\sqrt{\tilde \rho^{n}}}\nabla p^n \right) \label{stab-scheme-2}\\ && \mbox{3 -- Solve for } \tilde u^{n+1}\mbox{\ : \quad} \nonumber \\ && \quad \frac{\tilde \rho^{n+1}\, \tilde u^{n+1}-\rho^{n}\,u^{n}}{\delta t} +\nabla \cdot (\tilde \rho^{n+1}\,u^n\otimes\tilde u^{n+1}) +\nabla \tilde p^{n+1} -\nabla \cdot \tau(\tilde u^{n+1}) = f^{n+1} \label{stab-scheme-3}\\ && \mbox{4 -- Solve for } \bar u^{n+1},\, p^{n+1},\, \rho^{n+1}\mbox{\ : \quad} \nonumber \\
&& \qquad \left| \begin{array}{l} \displaystyle \tilde \rho^{n+1}\,\frac{\bar u^{n+1}- \tilde u^{n+1}}{\delta t} + \nabla (p^{n+1}-\tilde p^{n+1})=0 \\ \displaystyle \frac{\varrho(p^{n+1})-\rho^n}{\delta t} + \nabla \cdot \left( \varrho(p^{n+1})\, \bar u^{n+1} \right)=0 \\ \displaystyle \rho^{n+1}=\varrho(p^{n+1}) \end{array} \right . \label{stab-scheme-4}\\ && \mbox{5 -- Compute } u^{n+1} \mbox{ given by: \quad} \sqrt{\rho^{n+1}}\,u^{n+1}=\sqrt{\tilde \rho^{n+1}}\bar u^{n+1} \label{stab-scheme-5} \end{eqnarray}
The first step is a prediction of the density, used for the discretization of the time derivative of the momentum. As remarked by Wesseling \textit{et al} \cite{bij-98-uni, wes-01-pri}, this step can be avoided when solving the Euler equations: in this case, the mass flowrate may be chosen as an unknown, using the explicit velocity as an advective field in the discretization of the convection term in the momentum balance; the velocity is then updated by dividing by the density at the end of the time step. For viscous flows, if the discretization of the diffusion term is chosen to be implicit, both the mass flowrate and the velocity appear as unknowns in the momentum balance; this seems to impeed the use of this trick. Let us emphasize that the way the step one is carried out (\textit{i.e.} solving a discretization of the mass balance instead as, for instance, performing a Richardson's extrapolation) is crucial for the stability.
Likewise, the second step is a renormalization of the pressure the interest of which is clarified only by the stability analysis. A similar technique has already been introduced by Guermond and Quartapelle for variable density incompressible flows \cite{gue-00-proj}.
Step 3 consists in a classical semi-implicit solution of the momentum equation to obtain a predicted velocity.
Step 4 is a nonlinear pressure correction step, which degenerates in the usual projection step as used in incompressible flow solvers when the density is constant (e.g. \cite{mar-98-nav}). Taking the divergence of the first relation of (\ref{stab-scheme-4}) and using the second one to eliminate the unknown velocity $\bar u^{n+1}$ yields a non-linear elliptic problem for the pressure. This computation is formal in the semi-discrete formulation, but, of course, is necessarily made clear at the algebraic level, as described in section \ref{subsec:impl}. Once the pressure is computed, the first relation yields the updated velocity and the third one gives the end-of-step density.
\subsection{Stability of the advection operator: a finite-volume result} \label{subsec:conv}
The aim of this section is to state and prove a discrete analogue to the stability identity \eqref{stab-arg}-$(i)$, which may be written for any sufficiently regular functions $\rho$, $z$ and $u$ as follows: \[ \int_\Omega \left[ \frac{\partial \rho z}{\partial t} + \nabla \cdot (\rho z u)\right ] z \, {\rm d}x = \frac{1}{2} \, \frac{d}{dt} \int_\Omega \rho z^2 \, {\rm d}x \label{VF1_continu} \] and holds provided that the following balance is satisfied by $\rho$ and $u$: \[ \frac{\partial \rho}{\partial t} + \nabla \cdot (\rho u) =0 \] As stated in introduction, applying this identity to each component of the velocity yields the central argument of the proof of the kinetic energy theorem.
The discrete analogue to this identity is the following. \begin{thrm}[Stability of the advection operator]\label{VF1}Let $(\rho_K^\ast)_{K\in {\cal M}}$ and $(\rho_K)_{K\in {\cal M}}$ be two families of positive real number satisfying the following set of equation: \begin{equation}
\forall K \in {\cal M},\qquad \frac{|K|}{\delta t} \ (\rho_K - \rho^\ast_K) + \sum_{\sigma=K|L} F_{\edge,K}=0 \label{mass_bal}\end{equation}
where $F_{\edge,K}$ is a quantity associated to the edge $\sigma$ and to the control volume $K$; we suppose that, for any internal edge $\sigma=K|L$, $F_{\edge,K}=-F_{\edge,L}$. Let $(z_K^\ast)_{K\in {\cal M}}$ and $(z_K)_{K\in {\cal M}}$ be two families of real numbers. For any internal edge $\sigma=K|L$, we define $z_\sigma$ either by $z_\sigma=\frac 1 2 (z_K +z_L)$, either by $z_\sigma=z_K$ if $F_{\edge,K} \geq 0$ and $z_\sigma=z_L$ otherwise. The first choice is referred to as an the "centered choice", the second one as "the upwind choice". In both cases, the following stability property holds: \begin{equation}
\sum_{K \in {\cal M}} z_K \left[ \frac{|K|}\delta t\ (\rho_K\, z_K -\rho_K^\ast\, z_K^\ast)+
\sum_{\sigma=K|L} F_{\edge,K} \, z_\sigma \right] \geq
\frac{1}{2}\ \sum_{K\in{\cal M}} \frac{|K|}\delta t\ \left[ \rho_K\, z_K^2 -\rho_K^\ast\, {z_K^\ast}^2\right] \label{ecinetique} \end{equation} \end{thrm}
\begin{proof} We write: \[
\sum_{K \in {\cal M}} z_K \left[ \frac{|K|}\delta t\ (\rho_K\, z_K -\rho_K^\ast\, z_K^\ast)+
\sum_{\sigma=K|L} F_{\edge,K} \, z_\sigma \right] =T_1 + T_2 \] where $T_1$ and $T_2$ reads: \[
T_1= \sum_{K \in {\cal M}} \frac{|K|}\delta t\ z_K\ (\rho_K\, z_K -\rho_K^\ast\, z_K^\ast), \hspace{10ex}
T_2=\sum_{K \in {\cal M}} z_K \left[ \sum_{\sigma=K|L} F_{\edge,K} \, z_\sigma \right] \] The first term reads: \[
T_1= \sum_{K \in {\cal M}} \frac{|K|}\delta t\ \left[z_K^2\, (\rho_K-\rho_K^\ast) + \rho_K^\ast \, z_K\, (z_K -z_K^\ast) \right] \] Developping the last term by the identity $a(a-b)=\frac 1 2 ( a^2 + (a-b)^2 -b^2)$, we get: \[
T_1= \underbrace{\sum_{K \in {\cal M}} \frac{|K|}\delta t\ z_K^2\, (\rho_K-\rho_K^\ast)}_{\displaystyle T_{1,1}}
+ \underbrace{\frac 1 2 \sum_{K \in {\cal M}} \frac{|K|}\delta t\ \rho_K^\ast\, (z_K^2 - {z_K^\ast}^2)}_{\displaystyle T_{1,2}}
+ \underbrace{\frac 1 2 \sum_{K \in {\cal M}} \frac{|K|}\delta t\ \rho_K^\ast\, (z_K - z_K^\ast)^2}_{\displaystyle T_{1,3}} \] The last term, namely $T_{1,3}$, is always non-negative and can be seen as a dissipation associated to the backward time discretization of equation \eqref{ecinetique}. We now turn to $T_2$: \[
T_2=\underbrace{\sum_{K \in {\cal M}} z_K^2 \left[ \sum_{\sigma=K|L} F_{\edge,K}\right]}_{\displaystyle T_{2,1}}+
\underbrace{\sum_{K \in {\cal M}} z_K \left[ \sum_{\sigma=K|L} F_{\edge,K} \, (z_\sigma-z_K) \right]}_{\displaystyle T_{2,2}} \] The first term, namely $T_{2,1}$, will cancel with $T_{1,1}$ by equation \eqref{mass_bal}. The second term reads, developping as previously the quantity $z_K\, (z_\sigma-z_K)$: \[
T_{2,2}= -\frac 1 2 \sum_{K \in {\cal M}} z_K^2 \left[ \sum_{\sigma=K|L} F_{\edge,K} \right]
- \underbrace{\frac 1 2 \sum_{K \in {\cal M}} \left[ \sum_{\sigma=K|L} F_{\edge,K}\ [(z_\sigma-z_K)^2 - z_\sigma^2] \right]}_{\displaystyle T_{2,3}} \] Reordering the sum in the last term, we have, as $F_{\edge,K}=-F_{\edge,L}$: \[
T_{2,3}=\frac 1 2 \sum_{\sigma \in {\cal E}_{{\rm int}}\ (\sigma=K|L)} F_{\edge,K}\ [(z_\sigma-z_K)^2-(z_\sigma-z_L)^2] \]
This expression can easily be seen to vanish with the centered choice. With the upwind choice, supposing without loss of generality that we have chosen for the edge $\sigma=K|L$ the orientation such that $F_{\edge,K} \geq 0$, we get, as $z_\sigma=z_K$: \[
T_{2,3}=- \frac 1 2 \sum_{\sigma \in {\cal E}_{{\rm int}}\ (\sigma=K|L)} F_{\edge,K}\ (z_K-z_L)^2 \leq 0 \] We thus have, by equation \eqref{mass_bal}: \[
T_{2,2} \geq -\frac 1 2 \sum_{K \in {\cal M}} z_K^2 \left[ \sum_{\sigma=K|L} F_{\edge,K} \right]
=\frac 1 2 \sum_{K \in {\cal M}} \frac{|K|}\delta t z_K^2\, (\rho_K - \rho_K^\ast) \] and thus: \[
T_1 + T_2 \geq \frac 1 2 \sum_{K \in {\cal M}} \frac{|K|}\delta t \left[z_K^2\, (\rho_K - \rho_K^\ast) + \rho_K^\ast\, (z_K^2 - {z_K^\ast}^2)\right] \] which concludes the proof. \end{proof}
\begin{rmrk} Equation \eqref{mass_bal} can be seen as a discrete mass balance, with $F_{\edge,K}$ standing for the mass flux across the edge $\sigma$, and the right hand side of \eqref{ecinetique} may be derived by the multiplication by $z_K$ and summation over the control volumes of the transport terms in a discrete balance equation for the quantity $\rho z$, reading: \[ \forall K \in {\cal M}, \qquad
\frac{|K|}\delta t\ (\rho_K\, z_K -\rho_K^\ast\, z_K^\ast)+ \sum_{\sigma=K|L} F_{\edge,K} \, z_\sigma +\cdots \mbox{[possible diffusion terms]} \cdots =0 \] In this context, the relation \eqref{mass_bal} is known to be exactly the compatibility condition which ensures a discrete maximum principle for the solution $z$ of this transport equation, provided that the upwind choice (or any monotone choice) is made for the expression of $z_\sigma$ \cite{lar-91-how}. We proved here that the same compatibility condition ensures a $\xLtwo$ stability for $\rho^{1/2} z$. \end{rmrk}
\subsection{Spatial discretization of the density prediction and the momentum balance equation}\label{subsec:mom}
The main difficulty in the discretization of the momentum balance equation is to build a discrete convection operator which enjoys the stability property \eqref{stab-arg}-$(i)$. To this purpose, we derive for this term a finite volume discretization which satisfies the assumptions of theorem \ref{VF1}.
The natural space discretization for the density is the same as for the pressure, \ie \ piecewise constant functions over each element. For the Rannacher-Turek element, this legitimates a standard mass lumping technique for the time derivative term, since no additional accuracy seems to have to be expected from a more precise integration. For the Crouzeix-Raviart element, the mass matrix is already diagonal. Let the quantity $|D_\sigma|$ be defined as follows: \begin{equation}
|D_\sigma| \eqdef \int_\Omega \varphi_\sigma \, {\rm d}x >0 \label{int-phis}\end{equation}
For the Crouzeix-Raviart element, $|D_\sigma|$ can be identified to the measure of the cone with basis $\sigma$ and with the mass center of the mesh as opposite vertex. The same property holds for the Rannacher-Turek element in the case of quandrangles ($d=2$) or cuboids ($d=3$), which are the only case considered here, even though extensions to non-perpendicular grids are probably possible.
For each internal edge $\sigma=K|L$, this conic volume is denoted by $D_{K,\sigma}$; the volume $D_\sigma= D_{K,\sigma}\cup D_{L,\sigma}$ is referred to as the "diamond cell" associated to $\sigma$, and $D_{K,\sigma}$ is the half-diamond cell associated to $\sigma$ and $K$ (see figure \ref{diamonds}). The measure of $D_{K,\sigma}$ is denoted by $|D_{K,\sigma}|$.
The discretization of the term $\rho^n \, u^n$ thus leads, in the equations associated to the velocity on $\sigma$, to an expression of the form $\rho^n_\sigma \, u^n_\sigma$, where $\rho^n_\sigma$ results from an average of the values taken by the density in the two elements adjacent to $\sigma$, weighted by the measure of the half-diamonds: \begin{equation}
\forall \sigma \in {\cal E}_{{\rm int}},\qquad |D_\sigma|\ \rho^n_\sigma= |D_{K,\sigma}|\ \rho^n_K + |D_{L,\sigma}|\ \rho^n_L \label{rho_faces}\end{equation}
\begin{figure}
\caption{Dual finite volume mesh: the so-called "diamond cells".}
\label{diamonds}
\end{figure}
In order to satisfy the compatibility condition introduced in the previous section, a prediction of the density is first performed, by a finite volume discretization of the mass balance equation, taking the diamond cells as control volumes: \begin{equation} \forall \sigma \in {\cal E}_{{\rm int}},\qquad
\frac{|D_\sigma|} \delta t\ (\tilde \rho_\sigma^{n+1}- \rho^n_\sigma) + \sum_{\varepsilon \in {\cal E}(D_\sigma)} F_{\edged,\edge}^{n+1}=0 \label{mass-diamond}\end{equation} where ${\cal E}(D_\sigma)$ is the set of the edges of $D_\sigma$ and $F_{\edged,\edge}$ stands for the mass flux across $\varepsilon$ outward $D_\sigma$. This latter quantity is expressed as follows: \[
F_{\varepsilon}^{n+1}=|\varepsilon|\ u_{\varepsilon}^n\cdot n_{\varepsilon,\sigma}\ \tilde \rho_{\varepsilon}^{n+1} \]
where $|\varepsilon|$ is the measure of $\varepsilon$, $n_{\varepsilon,\sigma}$ is the normal to $\varepsilon$ outward $D_\sigma$, the velocity $u_{\varepsilon}^n$ is obtained by interpolation at the center of $\varepsilon$ in $u^n$ (using the standard finite element expansion) and $\tilde \rho_{\varepsilon}^{n+1}$ is the density at the edge, calculated by the standard upwinding technique (\ie\ either $\tilde \rho_\sigma^{n+1}$ if $u_{\varepsilon}^n\cdot n_{\varepsilon,\sigma} \geq 0$ or $\tilde \rho_{\sigma'}^{n+1}$ otherwise, with $\sigma'$ such that $\varepsilon$ separates $D_\sigma$ and $D_{\sigma'}$, which we denote by $\varepsilon=D_\sigma | D_{\sigma'}$).
The discretization convection terms of the momentum balance equation are built from relation \eqref{mass-diamond}, according to the structure which is necessary to apply theorem \ref{VF1}. This yields the following discrete momentum balance equation: \begin{equation} \begin{array}{l} \displaystyle
\frac{|D_\sigma|} \delta t\ (\tilde \rho_\sigma^{n+1} \tilde u_{\sigma,i}^{n+1}- \rho^n_\sigma u_{\sigma,i}^n)
+ \sum_{\stackrel{\scriptstyle \varepsilon \in {\cal E}(D_\sigma),}{\scriptstyle \varepsilon=D_\sigma | D_{\sigma'}}}\frac 1 2\ F_{\edged,\edge}^{n+1}\ (\tilde u_{\sigma,i}^{n+1} + u_{\sigma',i}^{n+1}) \\ \hspace{8ex}\displaystyle + \int_{\Omega,h} \tau(\tilde u^{n+1}) : \nabla \varphi_\sigma^{(i)} \, {\rm d}x -\int_{\Omega,h} \tilde p^{n+1}\ \nabla \cdot \varphi_\sigma^{(i)} \, {\rm d}x = \int_\Omega f^{n+1} \cdot \varphi_\sigma^{(i)} \hspace{10ex} \sigma \in {\cal E}_{{\rm int}},\ 1 \leq i \leq d \end{array} \label{momentum}\end{equation} where we recall that $\varphi_\sigma^{(i)}$ the vector shape function associated to $v_{\sigma, i}$, which reads $\varphi_\sigma^{(i)}=\varphi_\sigma \, e_i$ with $e_i$ is the $i^{th}$ vector of the canonical basis of $\xR^d$ and $\varphi_\sigma$ the scalar shape function, and the notation $\int_{\Omega,h}$ means $\sum_{K\in{\cal M}} \int_K$.
Note that, for Crouzeix-Raviart elements, a combined finite volume/finite element method, similar to the technique employed here for the discretization of the momentum balance, has already been analysed for a transient non-linear convection-diffusion equation by Feistauer and co-workers \cite{ang-98-ana, dol-02-err, fei-03-mat}.
\subsection{Spatial discretization of the projection step}\label{subsec:proj}
The fully discrete projection step used in the described algorithm reads: \begin{equation}
\left| \begin{array}{ll} \displaystyle
|D_\sigma|\,\frac{\tilde \rho_\sigma^{n+1}}{\delta t} \, (\bar u_{\sigma,i}^{n+1} - \tilde u_{\sigma,i}^{n+1}) + \int_{\Omega,h} (p^{n+1}-\tilde p^{n+1})\ \nabla \cdot \varphi_\sigma^{(i)} \, {\rm d}x = 0 \hspace{12ex} & \displaystyle \sigma \in {\cal E}_{{\rm int}},\ 1 \leq i \leq d \\[2ex] \displaystyle
|K|\ \frac{\varrho(p^{n+1}_K)-\rho^n_K}{\delta t}
+ \sum_{\sigma=K|L} (\mathrm{v}_{\sigma,K}^+)^{n+1}\, \varrho(p_K^{n+1}) - (\mathrm{v}_{\sigma,K}^-)^{n+1}\, \varrho(p_L^{n+1})=0 & \displaystyle K \in {\cal M} \end{array}\right . \label{darcy-disc-un}\end{equation}
where $(\mathrm{v}_{\sigma,K}^+)^{n+1}$ and $(\mathrm{v}_{\sigma,K}^-)^{n+1}$ stands respectively for $\max (\mathrm{v}_{\sigma,K}^{n+1},\ 0)$ and $\max ( -\mathrm{v}_{\sigma,K}^{n+1},\ 0)$ with $\mathrm{v}_{\sigma,K}^{n+1}=|\sigma|\, \bar u_\sigma^{n+1} \cdot n_{KL}$. The first (vector) equation may be seen as the finite element discretization of the first relation of the projection step \eqref{stab-scheme-4}, with the same lumping of the mass matrix for the Rannacher-Turek element as in the prediction step. As the pressure is piecewise constant, the finite element discretization of the second relation of \eqref{stab-scheme-4}, \ie\ the mass balance, is equivalent to a finite volume formulation, in which we introduce the standard first-order upwinding. Exploiting the expression of the velocity and pressure shape functions, the first set of relations of this system can be alternatively written as follows: \begin{equation}
|D_\sigma|\,\frac{\tilde \rho_\sigma^{n+1}}{\delta t} \, (\bar u_\sigma^{n+1}-\tilde u_\sigma^{n+1})
+ |\sigma| \, \left[ (p_K^{n+1}-\tilde p_K^{n+1})-(p_L^{n+1}-\tilde p_L^{n+1}) \right]\, n_{KL} = 0 \qquad
\forall \sigma\in{\cal E}_{{\rm int}}, \ \sigma=K|L \label{darcy-disc-deux}\end{equation} or, in an algebraic setting: \begin{equation} \frac{1}{\delta t} \, {\rm M}_{\tilde \rho^{n+1}} \, (\bar u^{n+1} - \tilde u^{n+1}) + {\rm B}^t \, (p^{n+1}-\tilde p^{n+1}) = 0 \label{darcy-alg}\end{equation}
In this relation, ${\rm M}_{w}$ stands for the diagonal mass matrix weighted by $(w_\sigma)_{\sigma \in {\cal E}_{{\rm int}}}$ (so, for $1\leq i \leq d$ and $\sigma \in {\cal E}_{{\rm int}}$, the corresponding entry on the diagonal of ${\rm M}_{\tilde \rho^{n+1}}$ reads $({\rm M}_{\tilde \rho^{n+1}})_{\sigma,i}= (|D_\sigma|\, \tilde \rho_\sigma^{n+1})$), ${\rm B}^t$ is the matrix of $\xR^{(d\,N) \times M}$, where $N$ is the number of internal edges (\ie\ $N={\rm card}\,({\cal E}_{{\rm int}})$) and $M$ is the number of control volumes in the mesh (\ie\ $M={\rm card}\,({\cal M})$), associated to the gradient operator; consequently, the matrix ${\rm B}$ is associated to the opposite of the divergence operator. Throughout this section, we use the same notation for the discrete function (defined as usual in the finite element context by its expansion using the shape functions) and for the vector gathering the degrees of freedom; so, in relation \eqref{darcy-alg}, $\bar u$ stands for the vector of $\xR^{d\,N}$ components $\bar u_{\sigma,i},\ 1\leq i \leq d,\ \sigma \in {\cal E}_{{\rm int}}$ and $p$ stand for the vector of $\xR^M$ of components $p_K,\ K \in {\cal M}$. Both forms \eqref{darcy-disc-deux} and \eqref{darcy-alg} are used hereafter.
We have the following existence result. \begin{prpstn} Let the equation of state $\varrho(\cdot)$ be such that $\varrho(0)=0$, $\lim_{z \rightarrow + \infty} \varrho(z)= + \infty$ and there exists an elastic potential function $P(\cdot)$ such that the function $z \mapsto z\,P(z)$ is bounded from above in $(0, + \infty)$, once continuously differentiable and strictly convex. Then the nonlinear system \eqref{darcy-disc-un} admits at least one solution and any possible solution is such that $p_K >0$, $\forall K \in {\cal M}$ (or, equivalently, $\rho_K >0$, $\forall K \in {\cal M}$). \end{prpstn} \begin{proof} Let us suppose that $\tilde \rho_\sigma^{n+1} >0$, $\forall \sigma \in {\cal E}_{{\rm int}}$. Then the theory of section \ref{sec:darcy} applies, with: \[
\normsd{\bar u^{n+1}}=\sum_{\sigma \in {\cal E}_{{\rm int}}} |D_\sigma|\,\frac{\tilde \rho_\sigma^{n+1}}{\delta t} \, (\bar u_\sigma^{n+1})^2 \] This yields both the existence of a solution and the positivity of the pressure. In view of the form of the discrete density prediction \eqref{mass-diamond}, this latter property extends by induction to any time step of the computation (provided, of course, that the initial density is positive everywhere). \end{proof}
We finish this section by some remarks concerning the projection step at hand.
\begin{lmm}\label{bmbt}The following identity holds for each discrete pressure $q \in L_h$: \[ \forall K \in {\cal M}, \qquad
({\rm B}\,{\rm M}_{\tilde \rho^{n+1}}^{-1}\, {{\rm B}}^t\, q)_K = \sum_{\sigma=K|L} \frac{1}{\tilde \rho_\sigma^{n+1}}\ \frac{|\sigma|^2}{|D_\sigma|}\, (q_K-q_L) \] \end{lmm}
\begin{proof} Let $q \in L_h$ be given. By relation \eqref{darcy-disc-deux}, we have: \[
({\rm B}^t\,q)_{\sigma,i}=|\sigma|\ (q_K-q_L)\,n_{KL} \cdot e_i \] Let $1_K \in L_h$ be the characteristic function of $K$. Denoting by $(\cdot,\cdot )$ the standard Euclidian scalar product, by the previous relation and the definition of lumped velocity mass matrix, we obtain: \[ \begin{array}{ll} \displaystyle ({\rm B}\,{\rm M}_{\tilde \rho}^{-1}\, {\rm B}^t\, q,\ 1_K) & \displaystyle =({\rm M}_{\tilde \rho}^{-1}\, {\rm B}^t\, q, \ {\rm B}^{t}\,1_K) \\ & \displaystyle
=\sum_{\sigma \in {\cal E}_{{\rm int}}} \quad \sum_{i=1}^d \quad \frac{1}{\tilde\rho_{\sigma}^{n+1} |D_\sigma|} ({\rm B}^{t}\,q)_{\sigma,i} \ ({\rm B}^{t}\,1_K)_{\sigma,i} \\ & \displaystyle
= \sum_{\sigma \in {\cal E}_{{\rm int}},\ \sigma=K|L} \quad \sum_{i=1}^d \quad \frac{1}{\tilde\rho_{\sigma}^{n+1} |D_\sigma|}
\left[ |\sigma|\, (q_K-q_L)\ n_{KL}\cdot e_i \right] \left[ |\sigma|\, n_{KL}\cdot e_i\right] \end{array} \] which, remarking that $\sum_{i=1}^d (n_{KL}\cdot e_i)^2=1$, yields the result. \end{proof} \begin{rmrk}[On spurious pressure boundary conditions] In the context of projection methods for uncompressible flow, it is known that spurious boundary conditions are to be imposed to the pressure in the projection step, in order to make the definition of this step of the algorithm complete. These boundary conditions are explicit when the process to derive the projection step is first to pose the problem at the time semi-discrete level and then discretize it in space; for instance, with a constant density equal to one and prescribed velocity boundary conditions on $\partial \Omega$, the semi-discrete projection step would take the form: \[
\left| \begin{array}{ll} \displaystyle - \Delta\, (p^{n+1}-\tilde p^{n+1})=- \frac 1 \delta t \nabla \cdot \tilde u^{n+1} \hspace{5ex} & \displaystyle \mbox{in } \Omega \\[2ex] \displaystyle \nabla \, (p^{n+1}-\tilde p^{n+1}) \cdot n =0 & \displaystyle \mbox{on } \partial \Omega \end{array} \right . \] When the elliptic problem for the pressure is built at the algebraic level, the boundary conditions for the pressure are somehow hidden in the discrete operator ${\rm B}\,{\rm M}^{-1}\, {{\rm B}}^t$. Lemma \ref{bmbt} shows that this matrix takes the form of a finite-volume Laplace discrete operator, with homogeneous Neumann boundary conditions, \ie \ the same boundary conditions as in the time semi-discrete problem above stated. \label{spurious-BC}\end{rmrk}
\begin{rmrk}[On the non-consistency of the discretization at hand for the Darcy problem] Considering the semi-discrete problem \eqref{stab-scheme-4}, taking $\tilde \rho_\sigma^{n+1}=1,\ \forall \sigma \in {\cal E}_{{\rm int}}$, one could expect to recover a consistent discretization of a Poisson problem with homogeneous Neumann boundary conditions. The following example shows that this route is misleading.
Let us take for the mesh a uniform square grid of step $h$. The coefficient ${|\sigma|^2}/{|D_\sigma|}$ can be easily evaluated, and we obtain: \[
({\rm B}\,{\rm M}^{-1}\, {{\rm B}}^t\, q)_K = d \sum_{\sigma=K|L} \frac{|\sigma|}{h}\, (q_K-q_L) \] that is the usual finite volume Laplace operator, but multiplied by the space dimension $d$. This result is of course consistent with the wellknown non-consistency of the Rannacher-Turek element for the Darcy problem, and similar examples could also be given for simplicial grids, with the Crouzeix-Raviart element. \label{non-consistent!}\end{rmrk}
\subsection{Renormalization steps}\label{subsec:renorm}
The pressure renormalization (step 2 of the algorithm) is introduced for stability reasons, and its form can only be clarified by the stability study. It reads: \begin{equation} {\rm B}\,{\rm M}^{-1}_{\tilde \rho^{n+1}}\, {{\rm B}}^t \ \tilde p^{n+1} = {\rm B}\,{\rm M}^{-1}_{\sqrt{\tilde \rho^{n+1}}\,\sqrt{\rho^n}}\, {{\rm B}}^t \ p^n \label{p-renorm-mat}\end{equation} where the density at the face at time level $n$ necessary to the definition of ${\rm M}^{-1}_{\sqrt{\tilde \rho^{n+1}}\,\sqrt{\rho^n}}$ is obtained from the density field $\rho^n \in L_h$ by equation \eqref{rho_faces}. In view of the expression of these operators provided by lemma \ref{bmbt}, this relation equivalently reads: \begin{equation} \forall K \in {\cal M}, \qquad \qquad
\sum_{\sigma=K|L} \frac{1}{\tilde \rho_\sigma^{n+1}}\ \frac{|\sigma|^2}{|D_\sigma|}\, (\tilde p_K^{n+1}- \tilde p_L^{n+1})=
\sum_{\sigma=K|L} \frac{1}{\sqrt{\tilde \rho_\sigma^{n+1}}\sqrt{\rho_\sigma^n}}\ \frac{|\sigma|^2}{|D_\sigma|}\, (p_K^n- p_L^n) \label{p-renorm}\end{equation} As ${\rm B}^t$ and ${\rm B}$ stands for respectively the discrete gradient and (opposite of the) divergence operator, this system can be seen as a discretization of the semi-discrete expression of step 2; note however, as detailed in remark \ref{non-consistent!}, that this discretization is non-consistent.
The velocity renormalization (step 5 of the algorithm) simply reads: \begin{equation} \forall \sigma \in {\cal E}_{{\rm int}}, \quad \sqrt{\rho^{n+1}_\sigma} \, u^{n+1}_\sigma = \sqrt{\tilde \rho^{n+1}_\sigma} \, \bar u^{n+1}_\sigma \hspace{10ex} \mbox{or} \qquad {\rm M}_{\sqrt{\rho^{n+1}}}\ u^{n+1} = {\rm M}_{\sqrt{\tilde \rho^{n+1}}}\ \bar u^{n+1} \label{v-renorm}\end{equation}
\subsection{An overview of the algorithm}\label{subsec:overview}
To sum up, the algorithm considered in this section is the following one: \begin{enumerate} \item Prediction of the density -- The density on the edges at $t^n$, $(\rho_\sigma^n)_{\sigma \in {\cal E}_{{\rm int}}}$, being given by \eqref{rho_faces}, compute $(\tilde \rho^{n+1})_{\sigma \in {\cal E}_{{\rm int}}}$ by the upwind finite volume discretization of the mass balance over the diamond cells \eqref{mass-diamond}.
\item Renormalization of the pressure -- Compute a renormalized pressure $(\tilde p_K^{n+1})_{K\in{\cal M}}$ by equation \eqref{p-renorm}.
\item Prediction of the velocity -- Compute $(\tilde u^{n+1}_\sigma)_{\sigma \in {\cal E}_{{\rm int}}}$ by equation \eqref{momentum}, obtained by a finite volume discretization of the transport terms over the diamond cells and a finite element discretization of the other terms.
\item Projection step -- Compute $(\bar u^{n+1}_\sigma)_{\sigma \in {\cal E}_{{\rm int}}}$ and $(p_K^{n+1})_{K\in{\cal M}}$ from equation \eqref{darcy-disc-un}, obtained by a finite element discretization of the velocity correction equation and an upwind finite volume discretization of the mass balance (over the elements $K \in {\cal M}$).
\item Renormalization of the velocity -- Compute $(u^{n+1}_\sigma)_{\sigma \in {\cal E}_{{\rm int}}}$ from equation \eqref{v-renorm}. \end{enumerate}
The existence of a solution to step 4 is proven above; the other problems are linear, and their regularity follows from standard coercivity arguments, using the fact that the discrete densities (\ie\ $\rho^n$ and $\tilde \rho^{n+1}$) are positive, provided that this property is satisfied by the initial condition.
\subsection{Stability analysis} \label{subsec:stab}
In this section, we use the following discrete norm and semi-norm: \begin{equation} \begin{array}{ll} \displaystyle \forall v \in W_h, \qquad & \displaystyle
\normLdiscd{\tilde \rho}{v} = \sum_{\sigma \in {\cal E}_{{\rm int}}} |D_\sigma|\, \tilde \rho_\sigma \, v_\sigma^2 \\[2ex] \forall q \in L_h, \qquad & \displaystyle
\snormundiscd{\tilde \rho}{q}= \sum_{\sigma \in {\cal E}_{{\rm int}},\ \sigma=K|L} \frac{1}{\tilde \rho_\sigma}\ \frac{|\sigma|^2}{|D_\sigma|}\, (q_K-q_L)^2 \end{array} \end{equation} where $\tilde \rho=(\tilde \rho_\sigma)_{\sigma \in {\cal E}_{{\rm int}}}$ is a family of positive real numbers. The function $\normLdiscd{\tilde \rho}{\cdot}$ defines a norm over $W_h$, and $\snormundisc{\tilde \rho}{\cdot}$ can be seen as a weighted version of the $H^1$ semi-norm classical in the finite volume context \cite{eym-00-fin}. The links between this latter semi-norm and the problem at hand are clarified in the following lemma, which is a straightforward consequence of lemma \ref{bmbt}.
\begin{lmm}\label{bmbt_norm}The following identity holds for each discrete pressure $q \in L_h$: \[ ({\rm B}\,{\rm M}_{\tilde \rho}^{-1}\, {{\rm B}}^t\, q,q)=\snormundiscd{\tilde \rho}{q} \] \end{lmm}
We are now in position to state the stability of the scheme under consideration. \begin{thrm}[Stability of the scheme] Let the equation of state $\varrho(\cdot)$ be such that $\varrho(0)=0$, $\lim_{z \rightarrow + \infty} \varrho(z)= + \infty$ and there exists an elastic potential function $P(\cdot)$ satisfying \eqref{elasticpot_gene} such that the function $z \mapsto z\,P(z)$ is bounded from above in $(0, + \infty)$, once continuously differentiable and strictly convex. Let $(\tilde u^n)_{0\leq n \leq N}$, $(u^n)_{0\leq n \leq N}$, $(p^n)_{0\leq n \leq N}$ and $(\rho^n)_{0\leq n \leq N}$ be the solution to the considered scheme, whith a zero forcing term. Then the following bound holds for $n < N$: \begin{equation} \begin{array}{l} \displaystyle \frac 1 2 \normLdiscd{\rho^{n+1}}{u^{n+1}} + \int_\Omega \rho^{n+1} P(\rho^{n+1}) \, {\rm d}x + \delta t \sum_{k=1}^{n+1} \int_{\Omega,h} \nabla \tilde u^k:\tau(\tilde u^k)\, {\rm d}x + \frac{\delta t^2}{2} \snormundiscd{\tilde \rho^{n+1}}{p^{n+1}} \\ \displaystyle \hspace{50ex} \leq \frac 1 2 \normLdiscd{\rho^0}{u^0} + \int_\Omega \rho^0 P(\rho^0) \, {\rm d}x + \frac{\delta t^2}{2} \snormundiscd{\tilde \rho^0}{p^0}\qquad \label{stabres}\end{array} \end{equation} \label{stab}\end{thrm}
\begin{proof} Multiplying each equation of the step 3 of the scheme \eqref{momentum} by the corresponding unknown (\textit{i.e} the corresponding component of the velocity $\tilde u^{n+1}$ on the corresponding edge $\sigma$) and summing over the edges and the components yields, by virtue of the stability of the discrete advection operator (theorem \ref{VF1}): \begin{equation} \frac{1}{2\, \delta t} \normLdiscd{\tilde \rho^{n+1}}{\tilde u^{n+1}} - \frac{1}{2\, \delta t} \normLdiscd{\rho^n}{u^n} + \int_{\Omega,h} \tau(\tilde u^{n+1}): \nabla \tilde u^{n+1} \, {\rm d}x - \int_{\Omega,h} \tilde p^{n+1} \nabla \cdot \tilde u^{n+1} \, {\rm d}x \leq 0 \label{stab1} \end{equation} On the other hand, reordering equation \eqref{darcy-alg} and multiplying by ${\rm M}_{\tilde \rho^{n+1}}^{-1/2}$ (recall that ${\rm M}_{\tilde \rho^{n+1}}$ is diagonal), we obtain: \[ \frac{1}{\delta t}\, {\rm M}_{\tilde \rho^{n+1}}^{1/2} \bar u^{n+1} + {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, p^{n+1} = \frac{1}{\delta t}\, {\rm M}_{\tilde \rho^{n+1}}^{1/2} \tilde u^{n+1} + {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, \tilde p^{n+1} \] Squaring this relation gives: \[ \begin{array}{l}\displaystyle \left(\frac{1}{\delta t}\, {\rm M}_{\tilde \rho^{n+1}}^{1/2} \bar u^{n+1} + {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, p^{n+1}, \ \frac{1}{\delta t} \, {\rm M}_{\tilde \rho^{n+1}}^{1/2} \bar u^{n+1} + {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, p^{n+1}\right) = \\ \displaystyle \hspace{20ex} \left(\frac{1}{\delta t}\, {\rm M}_{\tilde \rho^{n+1}}^{1/2} \tilde u^{n+1} + {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, \tilde p^{n+1}, \ \frac{1}{\delta t}\, {\rm M}_{\tilde \rho^{n+1}}^{1/2} \tilde u^{n+1} + {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, \tilde p^{n+1}\right) \end{array} \] which reads: \[ \begin{array}{l} \displaystyle \frac{1}{\delta t^{2}}\, \left( {\rm M}_{\tilde \rho^{n+1}} \bar u^{n+1}, \ \bar u^{n+1}\right) + \left({\rm M}_{\tilde \rho^{n+1}}^{-1}\,{\rm B}^t\, p^{n+1}, \ {\rm B}^t\, p^{n+1}\right) +\frac{2}{\delta t} \left(\bar u^{n+1},\ {\rm B}^t\, p^{n+1}\right)= \\ \displaystyle \hspace{20ex} \frac{1}{\delta t^{2}}\, \left({\rm M}_{\tilde \rho^{n+1}} \tilde u^{n+1}, \ \tilde u^{n+1}\right) +\left({\rm M}_{\tilde \rho^{n+1}}^{-1}\,{\rm B}^t\, \tilde p^{n+1}, \ {\rm B}^t\, \tilde p^{n+1}\right) +\frac{2}{\delta t}\, \left(\tilde u^{n+1}, \ \tilde p^{n+1}\right) \end{array} \] Multiplying by $\delta t/2$, remarking that, $\forall v \in W_h,\ ({\rm M}_{\tilde \rho^{n+1}} v, \ v)= \normLdiscd{\tilde \rho^{n+1}}{v}$ and that, thanks to lemma \ref{bmbt_norm}, $\forall q \in L_h,\ ({\rm M}_{\tilde \rho^{n+1}}^{-1}\,{\rm B}^t\, q, \ {\rm B}^t\, q)=
({\rm B} \, {\rm M}_{\tilde \rho^{n+1}}^{-1}\,{\rm B}^t\, q,q)=\snormundiscd{\tilde \rho^{n+1}}{q}$, we get: \begin{equation} \begin{array}{l} \displaystyle \frac{1}{2 \delta t} \normLdiscd{\tilde \rho^{n+1}}{\bar u^{n+1}} +\frac \delta t 2 \snormundiscd{\tilde \rho^{n+1}}{p^{n+1}} +(\bar u^{n+1}, {\rm B}^t\, p^{n+1}) \\ \hspace{20ex} \displaystyle -\frac{1}{2 \delta t} \normLdiscd{\tilde \rho^{n+1}}{\tilde u^{n+1}} -\frac \delta t 2\,\snormundiscd{\tilde \rho^{n+1}}{\tilde p^{n+1}} - (\tilde u^{n+1}, {\rm B}^t\, \tilde p^{n+1}) =0 \end{array} \label{stab2}\end{equation} The quantity $-(\tilde u^{n+1}, {\rm B}^t\, \tilde p^{n+1})$ is nothing more than the opposite of the term $\displaystyle \int_{\Omega,h} \tilde p^{n+1} \nabla \cdot \tilde u^{n+1} \, {\rm d}x$ appearing in (\ref{stab1}), so summing (\ref{stab1}) and (\ref{stab2}) makes these terms disappear, leading to: \[ \begin{array}{l} \displaystyle \frac{1}{2 \delta t} \normLdiscd{\tilde \rho^{n+1}}{\bar u^{n+1}} - \frac{1}{2\, \delta t} \normLdiscd{\rho^n}{u^n} + \int_{\Omega,h} \tau(\tilde u^{n+1}) : \nabla \tilde u^{n+1} \, {\rm d}x \\[2ex] \hspace{20ex} \displaystyle +\frac \delta t 2 \, \snormundiscd{\tilde \rho^{n+1}}{p^{n+1}} -\frac \delta t 2 \, \snormundiscd{\tilde \rho^{n+1}}{\tilde p^{n+1}} + (\bar u^{n+1}, {\rm B}^t\, p^{n+1})\leq 0 \end{array} \] Finally, $(\bar u^{n+1}, {\rm B}^t\, p^{n+1})$ is precisely the pressure work which can be bounded by the time derivative of the elastic potential, as stated in theorem \ref{VF2}: \begin{equation} \begin{array}{l} \displaystyle \frac{1}{2 \delta t} \normLdiscd{\tilde \rho^{n+1}}{\bar u^{n+1}} + \int_{\Omega,h} \tau(\tilde u^{n+1}) : \nabla \tilde u^{n+1} \, {\rm d}x +\frac \delta t 2 \,\snormundiscd{\tilde \rho^{n+1}}{p^{n+1}} -\frac \delta t 2 \,\snormundiscd{\tilde \rho^{n+1}}{\tilde p^{n+1}} \\ \hspace{30ex} \displaystyle +\frac{1}{\delta t} \int_\Omega \rho^{n+1} P(\rho^{n+1}) \, {\rm d}x \leq \frac{1}{2\, \delta t} \normLdiscd{\rho^n}{u^n} +\frac{1}{\delta t} \int_\Omega \rho^{n} P(\rho^{n}) \, {\rm d}x \end{array}\label{eqstep2}\end{equation} The proof then ends by using the renormalization steps (step 2 and 5 of the algorithm). Step 2 reads in an algebric setting: \[ {\rm B}\, {\rm M}_{\tilde \rho^{n+1}}^{-1}\,{\rm B}^t\, \tilde p^{n+1} ={\rm B}\, {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\, {\rm M}_{\tilde \rho^{n}}^{-1/2}\,{\rm B}^t\, p^{n} \] Multiplying by $\tilde p^{n+1}$, we obtain: \[ \left({\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, \tilde p^{n+1},\ {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, \tilde p^{n+1}\right) = \left({\rm M}_{\tilde \rho^{n}}^{-1/2}\,{\rm B}^t\, p^{n}, \ {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\, {\rm B}^t\, \tilde p^{n+1}\right) \] and thus, by Cauchy-Schwartz inequality: \[ \left({\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, \tilde p^{n+1},\ {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, \tilde p^{n+1}\right) \leq \left({\rm M}_{\tilde \rho^{n}}^{-1/2}\,{\rm B}^t\, p^{n}, \ {\rm M}_{\tilde \rho^{n}}^{-1/2}\,{\rm B}^t\, p^{n}\right)^{1/2}\, \left({\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, \tilde p^{n+1},\ {\rm M}_{\tilde \rho^{n+1}}^{-1/2}\,{\rm B}^t\, \tilde p^{n+1} \right)^{1/2} \] This relation yields $\snormundiscd{\tilde \rho^{n+1}}{\tilde p^{n+1}} \leq \snormundiscd{\tilde \rho^{n}}{p^{n}}$. In addition, step 5 of the algorithm gives $\normLdiscd{\rho^{n+1}}{u^{n+1}}=\normLdiscd{\tilde \rho^{n+1}}{\bar u^{n+1}}$. Using these two relations in \eqref{eqstep2}, we get: \[ \begin{array}{l} \displaystyle \frac{1}{2 \delta t} \normLdiscd{\rho^{n+1}}{u^{n+1}} + \int_{\Omega,h} \tau(\tilde u^{n+1}): \nabla \tilde u^{n+1} \, {\rm d}x +\frac \delta t 2 \,\snormundiscd{\tilde \rho^{n+1}}{p^{n+1}} + \frac{1}{ \delta t} \int_\Omega \rho^{n+1} P(\rho^{n+1}) \, {\rm d}x \\[2ex] \hspace{50ex} \displaystyle \leq \frac{1}{2\, \delta t} \normLdiscd{\rho^n}{u^n} + \frac{1}{\delta t} \int_\Omega \rho^{n} P(\rho^{n}) \, {\rm d}x + \frac \delta t 2\,\snormundiscd{\tilde \rho^{n}}{p^n} \end{array} \] and the estimate of theorem \ref{stab} follows by summing over the time steps. \end{proof}
\begin{rmrk}[On the upwinding of the mass balance discretization, the \textit{inf-sup} stability of the discretization and the appearance of spurious pressure wiggles.] In the scheme considered in this section, the upwinding in the discretization of mass balance controls the onset of density oscillations. As long as the pressure and the density are linked by an increasing function, that is as long as the flow remains compressible with a reasonable equation of state, it will probably be sufficient to prevent from pressure oscillations. Besides, the fourth term of the left hand side of (\ref{stabres}), \textit{i.e.} the term involving $\snormundiscd{p^{n+1}}{\tilde \rho^{n+1}}$, provides a control on the discrete $H^1$ semi-norm of the pressure, at least for large time steps, and therefore also produces an additional pressure smearing. However, it comes up in the analysis as the composition of the discrete divergence with the discrete gradient; consequently, one will obtain such a smoothing effect only for \textit{inf-sup} stable discretizations. Note also that, even for steady state problems, some authors recommend the use of stable approximation space pairs to avoid pressure wiggles \cite{bris-90-com, for-93-fin}. \end{rmrk}
\begin{rmrk}[On a different projection step] Some authors propose a different projection step \cite{bij-98-uni,wes-01-pri}, which reads in time semi-discrete setting: \[
\left| \begin{array}{l} \displaystyle \frac{\varrho(p^{n+1})\, u^{n+1}- \tilde \rho^{n+1}\, \tilde u^{n+1}}{\delta t} + \nabla (p^{n+1}-\tilde p^{n+1})=0 \\[2ex] \displaystyle \frac{\varrho(p^{n+1})-\rho^n}{\delta t} + \nabla \cdot \left( \varrho(p^{n+1})\, u^{n+1} \right)=0 \end{array} \right . \]
Considering this system, one may be tempted by the following line of thought: choosing $q^{n+1}=\varrho(p^{n+1})\, u^{n+1}$ as variable, taking the discrete divergence of the first equation and using the second one will cause the convection term of the mass balance to disappear from the discrete elliptic problem for the pressure, whatever the discretization of this term may be. Consequently, the equation for the pressure will be free of the non-linearities induced by the upwinding and the dependency of the convected density on the pressure, while one still may hope to obtain a positive upwind (with respect to the density) scheme. In fact, this last point is incorrect. To be valid, it would necessitate that, from any solution $(q^{n+1}, p^{n+1})$, one be able to compute a velocity field $u^{n+1}$ by dividing $q^{n+1}$ by the density of the control volume located upstream with respect to $u^{n+1}$. Unfortunately, it is not always possible to obtain this upstream value; for instance, if for two neighbouring control volumes $K$ and $L$, $\rho_K<0$, $\rho_L>0$ and $q^{n+1} \cdot n_{K|L}>0$, neither the choice of $K$ nor $L$ for the upstream control volume is valid. Consequently, with this discretization, we are no longer able to guarantee neither the positivity of $\rho$ nor the absence of oscillations. However, as explained below, if the density remains positive, we will have a smearing of pressure or density wiggles due to the fact that the discretization is \textit{inf-sup} stable. \end{rmrk}
\subsection{Implementation}\label{subsec:impl}
The implementation of the first three steps (\ref{stab-scheme-1})-(\ref{stab-scheme-3}) of the numerical scheme is standard, and we therefore only describe here in details the fourth step, that is the projection step. The precise algebraic formulation of the system (\ref{stab-scheme-4}) reads: \begin{equation} \begin{array}{l}
\left| \begin{array}{l} \displaystyle \frac{1}{\delta t} {\rm M}_{\tilde \rho^{n+1}} \, (\bar u^{n+1} - \tilde u^{n+1}) + {\rm B}^t\, (p^{n+1} -\tilde p^{n+1})=0 \\[2ex] \displaystyle \frac{1}{\delta t} {\rm R}\, (\varrho(p^{n+1})-\rho^n) - {\rm B} {\rm Q}_{\rho^{n+1}_{\rm up}} \bar u^{n+1}=0 \end{array} \right . \end{array} \label{proj-alg}\end{equation}
where ${\rm M}_{\tilde \rho^{n+1}}$ and ${\rm Q}_{\rho^{n+1}_{\rm up}}$ are two diagonal matrices; for the first one, we recall that the entry corresponding to an edge $\sigma\in {\cal E}_{{\rm int}},\ \sigma=K|L$ is computed by multiplying the measure of the diamond associated to $\sigma$ by the predicted density (at the edge center) $\tilde \rho^{n+1}_\sigma$; in the second one, the same entry is obtained by just taking the density at $t^{n+1}$ in the element located upstream of $\sigma$ with respect to $\bar u^{n+1}$, {\it i.e.} either $\varrho\,(p_K^{n+1})$ or $\varrho\,(p_L^{n+1})$. Note that these definitions can be extended in a straightforward way for the boundary edges, if the velocity is not prescribed to zero on the boundary of the computational domain. The matrix ${{\rm R}}$ is diagonal and, for any $K\in{\cal M}$, its entry ${\rm R}_K$ is the measure of the element $K$. For the sake of simplicity, we suppose for the moment that the equation of state is linear: \[ \varrho(p^{n+1})=\frac{\partial \varrho}{\partial p}\ p^{n+1} \]
The elliptic problem for the pressure is obtained by multiplying the first relation of (\ref{proj-alg}) by ${\rm B}\ {\rm Q}_{\rho^{n+1}_{\rm up}}\ ({\rm M}_{\tilde \rho^{n+1}})^{-1}$ and using the second one. This equation reads: \begin{equation} {\rm L}\, p^{n+1}+\frac{\partial \varrho}{\partial p}\ \frac{1}{\delta t^{2}}{\rm R}\, p^{n+1} = {\rm L}\,\tilde p^{n+1}+\frac{1}{\delta t^{2}}{\rm R} \rho^n +\frac{1}{\delta t}{\rm B}\,{\rm Q}_{\rho^{n+1}_{\rm up}}\,\tilde u^{n+1} \label{step41}\end{equation} where ${\rm L}={\rm B}\,{\rm Q}_{\rho^{n+1}_{\rm up}}\,({\rm M}_{\tilde \rho^{n+1}})^{-1}\,{\rm B}^t$ can be viewed, for the discretization at hand, as a finite volume discrete approximation of the Laplace operator with Neumann boundary conditions (when the velocity is prescribed at the boundary), multiplied by the space dimension $d$ and the densities ratio (see remarks \ref{spurious-BC} and \ref{non-consistent!}). We recall that, by a calculation similar to the proof of lemma \ref{bmbt}, this matrix can be evaluated directly in the "finite volume way", by the following relation, valid for each element $K$: \[
({\rm L}\, p^{n+1})_K=\sum_{\sigma=K|L} \
d\ \frac{\displaystyle \rho_{{\rm up},\sigma}}{\tilde \rho^{n+1}_\sigma}\ \frac{|\sigma|^2}{|D_\sigma|}\, (p_K-p_L) \] where $\rho_{{\rm up},\sigma}$ stands for the upwind density associated to the edge $\sigma$. Provided that $p^{n+1}$ is known, the first relation of (\ref{proj-alg}) gives us the updated value of the velocity: \[ \bar u^{n+1}=\tilde u^{n+1}-\delta t\ ({\rm M}_{\tilde \rho^{n+1}})^{-1}\,{\rm B}^t\,(p^{n+1}-\tilde p^{n+1}) \label{step42} \] As, to preserve the positivity of the density, we want to use in the mass balance the value of the density upwinded with respect to $\bar u^{n+1}$, equations (\ref{step41}) and (\ref{step42}) are not decoupled, by contrast with what happens in usual projection methods. We thus implement the following iterative algorithm: \[ \begin{array}{ll} \displaystyle \mbox{Initialization: }\quad p_{0}^{n+1}= \tilde p^{n+1} \mbox{ and } \bar u_{0}^{n+1}=\tilde u^{n+1} \\[3ex]
\qquad \left| \begin{array}{l} \displaystyle \mbox{Step 4.1 -- Solve for } p_{k+1/2}^{n+1}\ : \\[3ex] \displaystyle \hspace{5ex} \left[{\rm L}+\frac{\partial \varrho}{\partial p}\ \frac{1}{\delta t^{2}} {\rm R} \right]\, p_{k+1/2}^{n+1}= {\rm L}\,\tilde p^{n+1} + \frac{1}{\delta t^{2}} {\rm R} \, \rho^n +\frac{1}{\delta t} {\rm B}\, {\rm Q}_{\rho^{n+1}_{\rm up}}\, \tilde u^{n+1} \\[3ex] \displaystyle \quad \mbox{where the density in ${\rm L}$ and ${\rm Q}_{\rho^{n+1}_{\rm up}}$ is evaluated at $p^{n+1}_k$ and the upwinding} \\ \displaystyle \quad \mbox{in ${\rm Q}_{\rho^{n+1}_{\rm up}}$ is performed with respect to $\bar u^{n+1}_k$} \\[3ex] \displaystyle \mbox{Step 4.2 -- Compute } p_{k+1}^{n+1}\mbox{ as } p_{k+1}^{n+1}=\alpha \, p_{k+1/2}^{n+1} + (1-\alpha)\, p_{k}^{n+1} \\[3ex] \displaystyle \mbox{Step 4.3 -- Compute } \bar u_{k+1}^{n+1}\mbox{ as }: \\[2ex] \displaystyle \hspace{5ex} \bar u_{k+1}^{n+1}=\tilde u^{n+1}-\delta t \ ({\rm M}_{\tilde \rho^{n+1}})^{-1}\,{\rm B}^t\,(p_{k+1}^{n+1}-\tilde p^{n+1}) \end{array} \right. \\[20ex] \displaystyle \mbox{Convergence criteria: }\quad \max \left[ \normz{p_{k+1}^{n+1}-p_{k}^{n+1}},\normz{u^{n+1}_{k+1}-u_{k}^{n+1}}\right] < \varepsilon \end{array} \] The second step of the previous algorithm is a relaxation step which can be performed to ensure convergence; however, in the tests presented hereafter, we use $\alpha=1$ and obtain convergence in few iterations (typically less than 5). When the equation of state is nonlinear, step 4.1 is replaced by one iteration of Newton's algorithm.
\subsection{Numerical experiments}\label{subsec:num-exp} In this section, we describe numerical experiments performed to assess the behaviour of the pressure correction scheme presented in this paper, in particular the convergence rate with respect to the space and time discretizations.
With $\Omega=(0,1)\times (-\frac 1 2,\frac 1 2)$, we choose for the momentum and density the following expressions: \[ \begin{array}{l} \displaystyle \rho\,u= -\frac{1}{4} \cos(\pi t)\left[\begin{array}{l} \sin(\pi x^{(1)}) \\ \cos(\pi x^{(2)}) \end{array}\right] \\[3ex] \displaystyle \rho=1+\frac{1}{4}\,\sin(\pi t)\,\left[\cos(\pi x^{(1)})-\sin(\pi x^{(2)})\right] \end{array} \] These functions satisfy the mass balance equation; for the momentum balance, we add the corresponding right hand side. In this latter equation, the divergence of the stress tensor is given by: \[ \nabla \cdot \tau(u)= \mu \Delta u + \frac \mu 3 \nabla\, \nabla \cdot u, \qquad \mu=10^{-2} \] The pressure is linked to the density by the following equation of state: \[ p=\varrho(\rho)=\frac{\rho - 1}{\gamma\ {\rm Ma}^2}, \qquad \gamma = 1.4, \ {\rm Ma}=0.5 \] where the parameter $\rm Ma$ corresponds to the characteristic Mach number.
\begin{figure}
\caption{velocity error as a function of the time step. }
\label{err_v}
\end{figure}
We use in these tests a special numerical integration of the forcing term of the momentum balance, which is designed to ensure that the discretization of a gradient is indeed a discrete gradient ({\it i.e.} if the forcing term $f$ can be recast under the form $f=\nabla g$, the discrete right hand side of the momentum balance belongs to the range of ${\rm B^t}$).
\begin{figure}
\caption{pressure error as a function of the time step. }
\label{err_p}
\end{figure}
Velocity and pressure errors obtained at $t=0.5$, in respectively $L^2$ and discrete $L^2$ norms and as a function of the time step, are drawn on respectively figure \ref{err_v} and figure \ref{err_p}, for $20 \times 20$, $40 \times 40$ and $80 \times 80$ uniform meshes. For large time steps, these curves show a decrease corresponding to approximately a first order convergence in time for the velocity and the pressure, until a plateau is reached, due to the fact that errors are bounded from below by the residual spatial discretization error. For both velocity and pressure, the value of the errors on this plateau show a spatial convergence order (in $L^2$ norm) close to 2.
\section{Conclusion}
We presented in this paper a numerical scheme for the barotropic Navier-Stokes compressible equations, based on a pressure-correction time stepping algorithm. For the spatial discretization, it combines low-order non-conforming mixed finite elements with finite volumes; in the incompressible limit, one recovers a classical projection scheme based on an \textit{inf-sup} stable pair of approximation spaces for the velocity and the pressure. This scheme is proven to enjoy an unconditional stability property: irrespectively of the time step, the discrete solution obeys the \textit{a priori} estimates associated to the continuous problem, \textit{i.e.} strict positivity of the density, bounds in $L^\infty$-in-time norm of the quantity $\int_\Omega \rho\, u^2\, {\rm d}x$ and $\int_\Omega \rho \, P(\rho)\, {\rm d}x$ and in $L^2$-in-time norm of the viscous dissipation $ \int_\Omega \tau(u): \nabla u \, {\rm d}x$. To our knowledge, this result is the first one of this type for barotropic compressible flows.
However, the scheme presented here is by no means "the ultimate scheme" for the solution to the compressible Navier-Stokes equations. It should rather be seen as an example of application (and probably one of the less sophisticated ones) of the mathematical arguments developped to obtain stability, namely theorems \ref{VF2} (discrete elastic potential identity) and \ref{VF1} (stability of the advection operator), and our hope is that these two ingredients could be used as such or adapted in the future to study other algorithms. For instance, a computation close to the proof of theorem \ref{stab} (and even simpler) would yield the stability of the fully implicit scheme; adding to this latter algorithm a prediction step for the density (as performed here) would also allow to linearize (once again as performed here) the convection operator without loss of stability. A stable pressure-correction scheme avoiding this prediction step can also be obtained, and is currently under tests at IRSN for the computation of compressible bubbly flows. Besides these variants, less diffusive schemes should certainly be searched for. Finally, the proposed scheme is currently the object of more in-depth numerical studies including, in particular, problems admitting less smooth solutions than the test presented here.
\end{document}
|
arXiv
|
{
"id": "0710.2987.tex",
"language_detection_score": 0.7088871598243713,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Boundary layer for a non-Newtonian flow \\ over a rough surface}
\author{David G\'erard-Varet \thanks{Institut de Math\'{e}matiques de Jussieu et Universit\'{e} Paris 7, 175 rue du Chevaleret, 75013 Paris France ({\tt [email protected]})} , Aneta Wr\'oblewska-Kami\'nska \thanks{ Institute of Mathematics, Polish Academy of Sciences, ul. \'Sniadeckich 8, 00-956 Warszawa, Poland ({\tt [email protected]})} }
\maketitle
\section{Introduction} The general concern of this paper is the effect of rough walls on fluids. This effect is important at various scales. For instance, in the area of microfluidics, recent experimental works have emphasized the role of hydrophobic rough walls in the improvement of slipping properties of microchannels. Also, in geophysics, as far as large scale motions are concerned, topography or shore variations can be assimilated to roughness. For high Reynolds number flows, an important issue is to understand how localized roughness triggers instabilities, and transition to turbulence. For laminar flows, the point is rather to understand how distributed roughness may have a macroscopic impact on the dynamics. More precisely, the hope is to be able to encode an averaged effect through an effective boundary condition at a smoothened wall. Such boundary condition, called {\em a wall law}, will avoid to simulate the small-scale dynamics that takes place in a boundary layer in the vicinity of the rough surface.
The derivation of wall laws for laminar Newtonian flows has been much studied, since the pioneering works of Achdou, Pironneau and Valentin \cite{Achdou:1995, Achdou:1998}, or J\"ager and Mikeli\'c \cite{Mikelic2001,Jager:2003}. See also \cite{Luchini:1995,Amirat:2001a,GV2003,Bresch,Mikelic2013}. A natural mathematical approach of this problem is by homogenization techniques, the roughness being modeled by a small amplitude/high frequency oscillation. Typically, one considers a Navier-Stokes flow in a channel $\Omega^\eps$ with a rough bottom: $$\Omega^\ep = \Omega \cup \Sigma_0 \cup R^\ep. $$ Precisely: \begin{itemize} \item $\Omega = (0,1)^2$ is the flat portion of the channel. \item $R^\eps$ is the rough portion of the channel: it reads $$ R^\eps = \{ x = (x_1,x_2), x_1 \in (0,1), 0 > x_2 > \eps \gamma(x_1/\eps) \}$$ with a bottom surface $\Gamma^\eps := \{x_2 = \eps \gamma(x_1/\eps) \}$ parametrized by $\eps \ll 1$. Function $\gamma = \gamma(y_1)$ is the {\em roughness pattern}. \item Eventually, $\Sigma_0 := (0,1) \times \{0\}$ is the interface between the rough and flat part. It is the artificial boundary at which the wall law is set. \end{itemize} Of course, within such model, the goal is to understand the asymptotic behavior of the Navier-Stokes solution $u^\eps$ as $\eps \rightarrow 0$. Therefore, the starting point is a formal approximation of $u^\eps$ under the form \begin{equation} \label{blexpansion}
u^\eps_{app}(x) = u^0(x) + \eps u^1(x) + \dots + u^0_{bl}(x/\eps) + \eps u^1_{bl}(x/\eps) + \dots .
\end{equation}
In this expansion, the $u^i = u^i(x)$ describe the large-scale part of the flow, whereas the $u^i_{bl} = u^i_{bl}(y)$ describe the boundary layer. The typical variable $y=x/\eps$ matches the small-scale variations induced by the roughness. In the case of homogeneous Dirichlet conditions at $\Gamma^\eps$, one can check formally that:
\begin{itemize} \item $u^0$ is the solution of the Navier-Stokes equation in $\Omega$, with Dirichlet condition at $\Sigma_0$. \item $u^0_{bl} = 0$, whereas $u^1_{bl}$ satisfies a Stokes equation in variable $y$ in the boundary layer domain
$$\Omega_{bl} := \{y = (y_1, y_2), y_1 \in \R, y_2 > \gamma(y_1)\}.$$
\end{itemize} The next step is to solve this boundary layer system, and show convergence of $u^1_{bl}$ as $y_2 \rightarrow +\infty$ to a constant field $u^\infty = (U^\infty, 0)$. This in turn determines the appropriate boundary condition for the large scale correction $u^1$. From there, considering the large scale part $u^0 + \eps u^1$, one can show that: \begin{itemize} \item The limit wall law is a homogeneous Dirichlet condition. Let us point out that this feature persists even starting from a microscopic pure slip condition, under some non-degeneracy of the roughness: \cite{Bucur:2008,Bucur:2012,BoDaGe}. \item The $O(\eps)$ correction to this wall law is a slip condition of Navier type, with $O(\eps)$ slip length. \end{itemize} All these steps were completed in aforementioned articles, in the case of periodic roughness pattern $\gamma$ : $\gamma(y_1 + 1) = \gamma(y_1)$. Over the last years, the first author has extended this analysis to general patterns of roughness, with ergodicity properties (random stationary distribution of roughness, {\it etc}). We refer to \cite{BaGeVa2008, DGV:2008,GeVaMa2010}. See also \cite{DaGeVa2011} for some recent work on the same topic.
The purpose of the present paper is to extend the former analysis to non-Newtonian flows. This may have various sources of interest. One can think of engineering applications, for instance lubricants to which polymeric additives confer a shear thinning behavior. One can also think of glaciology: as the interaction of glaciers with the underlying rocks is unavailable, wall laws can help. From a mathematical point of view, such examples may be described by a power-law model. Hence, we consider a system of the following form:
\begin{equation} \label{EQ1}
\left\{
\begin{aligned} -\dive S(Du) + \nabla p = e_1 & \quad \mbox{in} \: \Omega^\eps, \\ \dive u = 0 & \quad \mbox{in} \: \Omega^\eps, \\ u\vert_{\Gamma^\eps} = 0, \quad u\vert_{x_2 = 1} = 0&, \quad u \: \mbox{$1$-periodic in $x_1$}. \end{aligned} \right. \end{equation} As usual, $u = u(x) \in \R^2$ is the velocity field, $p = p(x) \in \R$ is the pressure. The source term $e_1$ at the right-hand side of the first equation corresponds to a constant pressure gradient $e_1 = (1,0)^t$ throughout the channel. Eventually, the left-hand side involves the stress tensor of the fluid. As mentioned above, it is taken of power-law type: $S : \R^{2\times 2}_{\rm sym} \to \R^{2\times 2}_{\rm sym}$ is given by \begin{equation} \label{defstress}
S : \R^{2\times 2}_{\rm sym} \to \R^{2\times 2}_{\rm sym}, \quad S(A) = \nu |A|^{p-2}A, \quad \nu > 0, \quad 1 < p < +\infty , \end{equation}
where $|A| = (\sum_{i,j} a_{i,j}^2)^{1/2}$ is the usual euclidean norm of the matrix $A$. {\em For simplicity, we shall take $\nu = 1$}. Hence, $S(Du) = |Du|^{p-2} Du$, where we recall that $Du = \frac{1}{2} (\nabla u + (\nabla u)^t)$ is the symmetric part of the jacobian. Following classical terminology, the case $p < 2$ resp. $p > 2$ corresponds to {\em shear thinning} fluids, resp. {\em shear thickening} fluids. The limit case $p=2$ describes a Newtonian flow. Note that we complete the equation in system \eqref{EQ1} by a standard no-slip condition at the top and bottom boundary of the channel. For the sake of simplicity, we assume periodicity in the large scale horizontal variable $x_1$. Finally, we also make a simplifying periodicity assumption on the roughness pattern $\gamma$: \begin{equation} \mbox{$\gamma$ is $C^{2,\alpha}$ for some $\alpha >0$, has values in $(-1,0)$, and is $1$-periodic in $y_1$}. \end{equation} For every $\eps > 0$ and any value of $p$, the generalized Stokes system \eqref{EQ1} has a unique solution $$ (u^\eps, p^\eps) \in W^{1,p}(\Omega^\ep) \times L^{p'}(\Omega^\ep)/\R .$$ The main point is to know about the asymptotic behavior of $u^\eps$, precisely to build some good approximate solution. With regards to the Newtonian case, we anticipate that this approximation will take a form close to \eqref{blexpansion}. Our plan is then: \begin{itemize} \item to derive the equations satisfied by the first terms of expansion \eqref{blexpansion}. \item to solve these equations, and show convergence of the boundary layer term to a constant field away from the boundary. \item to obtain error estimates for the difference between $u^\eps$ and $u^\eps_{app}$. \item to derive from there appropriate wall laws. \end{itemize} This program will be more difficult to achieve for non-Newtonian fluids, in particular for the shear thinning case $p < 2$, notably as regards the study of the boundary layer equations on $u_{bl} \: := u^1_{bl}$. In short, these equations will be seen to read $$ - \dive(S(A + D u_{bl})) + \nabla p = 0, \dive u = 0, \quad y \in \Omega_{bl} $$ for some explicit matrix $A$, together with periodicity condition in $y_1$ and a homogeneous Dirichlet condition at the bottom of $\Omega_{bl}$. Due to the nonlinearity of these equations and the fact that $A \neq 0$, the analysis will be much more difficult than in the Newtonian case, notably the proof of the so-called Saint-Venant estimates. We refer to section \ref{subsecstvenant} for all details.
Let us conclude this introduction by giving some references on related problems. In \cite{MarusicPaloka2000}; E. Maru\v{s}i\'c-Paloka considers power-law fluids with convective terms in infinite channels and pipes (the non-Newtonian analogue of the celebrated Leray's problem). After an appropriate change of unknown, the system studied in \cite{MarusicPaloka2000} bears some strong similarity to our boundary layer system. However, it is different at two levels : first, the analysis is restricted to the case $p > 2$. Second, our lateral periodicity condition in $y_1$ is replaced by a no-slip condition. This allows to use Poincar\'e's inequality in the transverse variable, and control zero order terms (in velocity $u$) by $\nabla u$, and then by $D u$ through the Korn inequality. It simplifies in this way the derivation of exponential convergence of the boundary layer solution (Saint-Venant estimates). The same simplification holds in the context of paper \cite{BoGiMa-Pa}, where the behaviour of a Carreau flow through a thin filter is analysed. The corrector describing the behaviour of the fluid near the filter is governed by a kind of boundary layer type system, in a slab that is infinite vertically in both directions. In this setting, one has $A = 0$, and the authors refer to \cite{MarusicPaloka2000} for well-posedness and qualitative behaviour. We also refer to the recent article \cite{Suarez-Grau_2015} dedicated to power-law fluids in thin domains, with Navier condition and anisotropic roughness (with a wavelength that is larger than the amplitude). In this setting, no boundary layer analysis is needed, and the author succeeds to describe the limit asymptotics by the unfolding method. Finally, we point out the very recent paper \cite{ChupinMartin2015}, where an Oldroyd fluid is considered in a rough channel. In this setting, no nonlinearity is associated to the boundary layer, which satisfies a Stokes problem.
\section{Boundary layer analysis} From the Newtonian case, we expect the solution $(u^\eps, p^\eps)$ of \eqref{EQ1} to be approximated by $$ u^\eps \approx u^0(x) + \eps u_{bl}(x/\eps), \quad p^\eps \approx p^0(x) + p_{bl}(x/\eps) ,$$ where \begin{itemize} \item $(u^0, p^0)$ describes the flow away from the boundary layer. We shall take $u^0 = 0$ and $p^0 = 0$ in the rough part $R^\eps$ of the channel. \item $(u_{bl}, p_{bl}) = (u_{bl}, p_{bl})(y)$ is a boundary layer corrector defined on the slab
$$\Omega_{bl} := \{y = (y_1, y_2), y_1 \in \T, y_2 > \gamma(y_1)\},$$ where $\T$ is the torus $\R/\Z$. This torus corresponds implicitly to a periodic boundary condition in $y_1$, which is inherited from the periodicity of the roughness pattern $\gamma$. We denote $$ \Omega_{bl}^{\pm} \: := \: \Omega_{bl} \cap \{ \pm y_2 > 0 \} $$ its upper and lower parts, and $$ \Gamma_{bl} := \{ y = (y_1, y_2), y_1 \in \T, y_2 = \gamma(y_1)\} $$ its bottom boundary. As the boundary layer corrector is supposed to be localized, we expect that $$ \nabla u_{bl} \rightarrow 0 \quad \mbox{as $y_2 \rightarrow +\infty$}. $$ \end{itemize} With this constraint in mind, we take $(u^0, p^0)$ to be the solution of
\begin{equation} \label{EQ2}
\left\{
\begin{aligned} -\dive S(Du^0) + \nabla p^0 = e_1 & \quad \mbox{in} \: \Omega, \\ \dive u^0 = 0 & \quad \mbox{in} \: \Omega, \\ u^0\vert_{\Sigma_0} = 0, \quad u^0\vert_{x_2 = 1} = 0&, \quad u^0 \mbox{ $1$-periodic in $x_1$}. \end{aligned} \right. \end{equation} The solution is explicit and generalizes the Poiseuille flow. A simple calculation yields: for all $x \in \Omega$, \begin{equation} \label{Poiseuille}
p^0(x) = 0, \quad u^0(x) = (U(x_2), 0), \quad U(x_2) = \frac{p-1}{p} \left(\sqrt{2}^{-\frac{p}{(p-1)}} - \sqrt{2}^{\frac{p}{(p-1)}} \left|x_2 - \frac{1}{2}\right|^{\frac{p}{p-1}} \right). \end{equation} We extend this solution to the whole rough channel by taking: $u^0 = 0, p^0 = 0$ in $R^\eps$. This zero order approximation is clearly continuous across the interface $\Sigma_0$, but the associated stress is not: denoting \begin{equation} \label{defA} A \: := D(u^0)\vert_{y_2 = 0^+}\: = \frac{1}{2}\begin{pmatrix} 0 & U'(0) \\ U'(0) & 0 \end{pmatrix}, \quad \mbox{with $U'(0) = \sqrt{2}^{\frac{p-2}{p-1}}$} \end{equation} we obtain
$$ [ S(Du^0) n - p^0 n ]\vert_{\Sigma_0} = |A|^{p-2}A n = \left( \begin{smallmatrix} - \sqrt{2}^{-p} U'(0)^{p-1} \\ 0 \end{smallmatrix}\right) = \left( \begin{smallmatrix} - \frac{1}{2} \\ 0 \end{smallmatrix}\right) $$ with $n = - e_2 = -(0,1)^t$ and $[f] := f\vert_{x_2 = 0^+} - f\vert_{x_2 = 0^-}$.
This jump should be corrected by $u_{bl}$, so that the total approximation $u^0(x) + \eps u_{bl}(x/\eps)$ has no jump. This explains the amplitude $\eps$ of the boundary layer term, as its gradient will then be $O(1)$. By Taylor expansion $U(x_2) = U(\eps y_2) = U(0) + \eps U'(0) y_2 + \dots$ we get formally $D(u^0 + \eps u_{bl}(\cdot/\eps)) \approx A + D u_{bl}$, where the last symmetric gradient is with respect to variable $y$. We then derive the following boundary layer system:
\begin{equation} \label{BL1}
\left\{
\begin{aligned}
- \dive S(A + D u_{bl}) + \nabla p_{bl} & = 0 \quad \mbox{in} \: \Omega_{bl}^+, \\
- \dive S(D u_{bl}) + \nabla p_{bl} & = 0 \quad \mbox{in} \: \Omega_{bl}^-, \\
\dive u_{bl} & = 0 \quad \mbox{in} \: \Omega_{bl}^+ \cup \Omega_{bl}^-, \\ u_{bl}\vert_{\Gamma_{bl}} & = 0, \\ u_{bl}\vert_{y_2 = 0^+} - u_{bl}\vert_{y_2 = 0^-} & = 0,
\end{aligned}
\right.
\end{equation}
together with the jump condition \begin{equation} \label{BL2}
\left( S( A + D u_{bl}) n - p_{bl} n \right) |_{y_2 = 0^+}
- \left(S(D u_{bl}) n - p_{bl} n\right) |_{y_2= 0^-} = 0, \quad n = (0,-1)^t.
\end{equation} Let us recall that the periodic boundary condition in $y_1$ is encoded in the definition of the boundary layer domain. The rest of this section will be devoted to the well-posedness and qualitative properties of \eqref{BL1}-\eqref{BL2}. We shall give detailed proofs only for the more difficult case $p < 2$, and comment briefly on the case $p \ge 2$ at the end of the section. Our main results will be the following: \begin{Theorem} {\bf (Well-posedness)} \label{thWP}
\noindent For all $1 < p < 2$, \eqref{BL1}-\eqref{BL2} has a unique solution $ (u_{bl},p_{bl}) \in W^{1,p}_{loc}(\Omega_{bl}) \times L^{p'}_{loc}(\Omega_{bl})/\R $
satisfying for any $M > |A|$:
$$D u_{bl} \, 1_{\{|D u_{bl}| \le M\}} \in L^2(\Omega_{bl}), \quad D u_{bl} \, 1_{\{|D u_{bl}| \ge M\}} \in L^p(\Omega_{bl}).$$
\noindent For all $p \ge 2$, \eqref{BL1}-\eqref{BL2} has a unique solution $ (u_{bl},p_{bl}) \in W^{1,p}_{loc}(\Omega_{bl}) \times L^{p'}_{loc}(\Omega_{bl})/\R $ s.t. $D u_{bl} \in L^p(\Omega_{bl}) \cap L^2(\Omega_{bl})$. \end{Theorem}
\begin{Theorem} {\bf (Exponential convergence)} \label{thEC}
\noindent For any $1 < p < +\infty$, the solution given by the previous theorem converges exponentially, in the sense that for some $C, \delta > 0$
$$ | u_{bl}(y) - u^\infty | \le C e^{-\delta y_2} \quad \forall \: y \in \Omega_{bl}^+,$$
where $u^\infty= (U^\infty, 0)$ is some constant horizontal vector field.
\end{Theorem}
\subsection{Well-posedness} \label{sectionWP}
\subsubsection*{A priori estimates}
We focus on the case $1 < p < 2$, and provide the {\it a priori} estimates on which the well-posedness is based. The easier case $p \ge 2$ is discussed at the end of the paragraph. As $A$ is a constant matrix, we have from \eqref{BL1}:
$$ - \dive S(A + D u_{bl}) + \dive S(A) + \nabla p_{bl} = 0 \quad \mbox{in} \: \Omega_{bl}^+, \quad - \dive S(D u_{bl}) + \nabla p_{bl} = 0 \quad \mbox{in} \: \Omega_{bl}^-. $$
We multiply the two equations by $D u_{bl}$ and integrate over $\Omega_{bl}^+$ and
$\Omega_{bl}^-$ respectively. After integrations by parts, accounting for the jump conditions at $y_2 = 0$, we get
\begin{equation} \label{variationalBL}
\int_{\Omega_{bl}^+} (S(A + Du_{bl}) - S(A)) : D u_{bl} \dy + \int_{\Omega_{bl}^-} S(D u_{bl}) : D u_{bl} \dy = -\int_{y_2 = 0} S(A)n \cdot u_{bl} {\rm\,d}S.
\end{equation}
The right-hand side is controlled using successively Poincar\'e and Korn inequalities (for the Korn inequality, see the appendix): \begin{equation}
|\int_{y_2 = 0} S(A)n \cdot u_{bl} \dy | \le C \| u_{bl} \|_{L^p(\{ y_2 = 0\})} \le C' \| \nabla u_{bl} \|_{L^p(\Omega_{bl}^-)} \le C'' \| D u_{bl} \|_{L^p(\Omega_{bl}^-)} .
\end{equation}
As regards the left-hand side, we rely on the following vector inequality, established in \cite[p74, eq. (VII)]{Lind}: for all $1 < p \le 2$, for all vectors $a,b$ \begin{equation} \label{vectorinequality}
( |b|^{p-2}b - |a|^{p-2} a \: | \: b-a) \: \ge \: (p-1) |b-a|^2 \int_0^1 |a + t(b-a)|^{p-2} dt. \end{equation}
In particular, for any $M > 0$, if $|b-a| \le M$, one has
\begin{equation} \label{ineq1}
( |b|^{p-2}b - |a|^{p-2} a \: | \: b-a) \: \ge \: \frac{p-1}{(|a| + M)^{2-p}} |b-a|^2 ,
\end{equation}
whereas if $|b-a| > M > |a|$, we get
\begin{equation} \label{ineq2}
( |b|^{p-2}b - |a|^{p-2} a \: | \: b-a) \:
\ge (p-1) |b-a|^2 \int_{\frac{|a|}{|b-a|}}^1 \left(2 t |b-a|\right)^{p-2} dt \ge 2^{p-3} \left(1 - (|a|/M)^{p-1}\right) |b-a|^p. \end{equation}
We then apply such inequalities to \eqref{variationalBL}, taking $a = A$, $b = A + Du_{bl}$. For $M > |A|$, there exists $c$ dependent on $M$ such that $$
\int_{\Omega_{bl}^+} (S(A + Du_{bl}) - S(A)) : D u_{bl} \dy \ge c \int_{\Omega_{bl}^+} \bbbone_{\{ |Du_{bl}| \le M \}} |Du_{bl}|^2 \dy \: + \: \int_{\Omega_{bl}^+} \bbbone_{\{ |Du_{bl}| > M \}} |D u_{bl}|^p \dy ,
$$
so that for some $C$ dependent on $M$ \begin{equation*}
\int_{\Omega_{bl}^+}| \bbbone_{\{ |Du_{bl}| \le M \}} Du_{bl}|^2 \dy \: + \: \int_{\Omega_{bl}^+} |\bbbone_{\{ |Du_{bl}| > M \}} D u_{bl}|^p \dy + \int_{\Omega_{bl}^-} | D u_{bl}|^p \dy \: \le \: C \, \| D u_{bl} \|_{L^p(\Omega_{bl}^-)} . \end{equation*} Hence, still for some $C$ dependent on $M$: \begin{equation} \label{aprioriestimate}
\int_{\Omega_{bl}^+} 1_{\{ |Du_{bl}| \le M \}} |Du_{bl}|^2 \dy \: + \: \int_{\Omega_{bl}^+} 1_{\{ |Du_{bl}| > M \}} |D u_{bl}|^p \dy
+ \int_{\Omega_{bl}^-} | D u_{bl}|^p \dy \: \le \: C . \end{equation} This is the {\it a priori} estimate on which Theorem \ref{thWP} can be established (for $p \in ]1,2]$). Note that this inequality implies that for any height $h$,
$$ \| Du_{bl} \|_{L^p(\Omega_{bl} \cap \{y_2 \le h \})} \le C_h $$ (bounding the $L^p$ norm by the $L^2$ norm on a bounded set). Combining with Poincar\'e and Korn inequalities, we obtain that $u_{bl}$ belongs to $W^{1,p}_{loc}(\Omega_{bl})$.
In the case $p\geq 2$, we can directly use the following inequality, which holds for all $a,\ b \in \R^n$:
\begin{equation}\label{abp_1}
| a - b |^p 2^{2-p} \leq 2^{-1} \left( |b|^{p-2} + |a|^{p-2}\right) |b-a|^{2} \leq\left\langle |a|^{p-2} a - |b|^{p-2}|b|, a-b \right\rangle .
\end{equation} It provides both an $L^2$ and $L^p$ control of the symmetric gradient of the solution. Indeed, taking $a = \tA + \tD_y \ub$, $b= \tA$ and using \eqref{variationalBL} we get the following a'priori estimates for $p\geq 2$
\begin{equation}\label{apesDup}
\begin{split}
& 2^{2-p} \int_{\Omega_{bl}^{+}} | \tD \ub |^p \dy + \int_{\Omega_{bl}^{-}} | \tD \ub |^p \dy +
2^{-1} |A|^{p-2} \int_{\Omega_{bl}^+} | \tD \ub |^2 \dy \: \\
& {\leq} \:
\int_{\Omega_{bl}^{+}} \left( S(A+ Du_{bl}) - S(A) \right) : D \ub \dy + \int_{\Omega_{bl}^{-}} S(D u_{bl} ) : D \ub \dy \\
& = \: - \int_{\Sigma_0} S(A) n \cdot u_{bl} {\rm \, d}S \\
& \leq \: c(\alpha) \| S(A) \|^{p'}_{L^{p'}(\Sigma_0)} + \alpha \| \ub \|^{p}_{L^p(\Sigma_0)}
\leq \: c(\alpha) \| S(A) \|^{p'}_{L^{p'}(\Sigma_0)}
+ \alpha C_\Gamma \| \nabla \ub \|^{p}_{L^p(\Omega_{bl}^{-})}
\\ &
\leq c(\alpha) \| S(A) \|^{p'}_{L^{p'}(\Sigma_0)}
+ \alpha C_\Gamma C_K\| \tD \ub \|^{p}_{L^p(\Omega_{bl}^{-})} ,
\end{split}
\end{equation} where the trace theorem, the Poincar\'e inequality and the Korn inequality were employed. By choosing the coefficient $\alpha$ small enough, and by the imbedding of $L^p(\Omega_{bl}^-)$ in $L^2(\Omega_{bl}^-)$, \eqref{apesDup} provides
\begin{equation}\label{es:Dup}
\int_{\Omega_{bl}} |\tD \ub|^{p} + |\tD \ub|^{2} \dy \leq C \| S(A) \|^{p'}_{L^{p'}(\Sigma_0)} < \infty .
\end{equation} Eventually, by Korn and Poincar\'e inequalities: $\ub \in W^{1,p}(\Omega_{bl})$ for $2\leq p < \infty$.
\subsubsection*{Construction scheme for the solution}
We briefly explain how to construct a solution satisfying the above estimates. We restrict to the most difficult case $p \in ]1,2]$. There are two steps:
{\em Step 1}: we solve the same equations, but {\em in the bounded domain $\Omega_{bl,n} = \Omega_{bl} \cap \{ y_2 < n \}$}, with a Dirichlet boundary condition at the top. As
$\Omega_{bl,n}$ is bounded, the imbedding of $W^{1,p}$ in $L^p$ is compact, so that a solution $u_{bl,n}$ can be built in a standard way. Namely, one can construct a sequence of Galerkin approximations $u_{bl,n,m}$ by Schauder's fixed point theorem. Then, as the estimate \eqref{aprioriestimate} holds for $u_{bl,n,m}$ uniformly in $m$ and $n$, the sequence $D u_{bl,n,m}$ is bounded in $L^p(\Omega_{bl,n})$ uniformly in $m$. Sending $m$ to infinity yields a solution $u_{bl,n}$, the convergence of the nonlinear stress tensor follows from Minty's trick. Note that one can then perform on $u_{bl,n}$ the manipulations of the previous paragraph, so that it satisfies \eqref{aprioriestimate} uniformly in $n$.
{\em Step 2}: we let $n$ go to infinity. We first extend $u_{bl,n}$ by $0$ for $y_2 > n$, and fix $M > |A|$. From the uniform estimate \eqref{aprioriestimate}, we get easily the following convergences (up to a subsequence):
\begin{equation}
\begin{aligned}
& u_{bl,n} \rightarrow u_{bl} \mbox{ weakly in } W^{1,p}_{loc}(\Omega_{bl}), \\
& D u_{bl,n} \rightarrow D u_{bl} \mbox{ weakly in } L^p(\Omega^-_{bl}), \\
& D u_{bl,n} {\bbbone}_{|Du_{bl,n}| < M}\rightarrow V_1 \mbox{ weakly in } L^2(\Omega^+_{bl}), \quad \mbox{ weakly-* in } L^\infty(\Omega_{bl}^+), \\
& D u_{bl,n} {\bbbone}_{|D u_{bl,n}| \ge M}\rightarrow V_2 \mbox{ weakly in } L^p(\Omega^+_{bl}). \end{aligned}
\end{equation}
Of course, $D u_{bl} = V_1 + V_2$ in $\Omega_{bl}^+$. A key point is that
$$\mbox{$S(A + D u_{bl,n}) - S(A)$ is bounded uniformly in $n$ in $(L^{p}(\Omega^+_{bl}))' = L^{p'}(\Omega^+_{bl})$ and in $\left(L^2(\Omega^+_{bl}) \cap L^\infty(\Omega_{bl}^+)\right)'$.}$$
and converges weakly-* to some $S^+$ in that space. To establish this uniform bound, we treat separately
$$ S_{n,1} \: := \: (S(A + D u_{bl,n}) - S(A)) {\bbbone}_{|Du_{bl,n}| < M}, \quad S_{n,2} \: := \: (S(A + D u_{bl,n}) - S(A)) {\bbbone}_{|Du_{bl,n}| \ge M}. $$ \begin{itemize}
\item For $S_{n,1}$, we use the inequality \eqref{ineq3}. It gives $|S_{n,1}| \le C |D u_{bl,n}| {\bbbone}_{|Du_{bl,n}| < M}$,
which provides a uniform bound in $L^{2} \cap L^\infty$, and so in particular in $L^{p'}$ and in $L^2$.
\item For $S_{n,2}$, we use first that $|S_{n,2}| \le C |D u_{bl,n}|^{p-1} {\bbbone}_{|D u_{bl,n}| \ge M}$, so that it is uniformly bounded in $L^{p'}$. We use then \eqref{ineq3}, so that
$ |S_{n,2}| \le C |D u_{bl,n}| {\bbbone}_{|D u_{bl,n}| \ge M}$, which yields a uniform bound in $L^p$, in particular in $(L^2 \cap L^\infty)'$ ($p \in ]1,2]$). \end{itemize} From there, standard manipulations give $$ \int_{\Omega_{bl}^+} (S(A+ D u_{bl,n}) - S(A)) : D u_{bl,n} \rightarrow \int_{\Omega_{bl}^+} S^+ : (V_1 + V_2) = \int_{\Omega_{bl}^+} S^+ : D u_{bl} $$ One has even more directly $$ \int_{\Omega_{bl}^-} S(D u_{bl,n}) : D u_{bl,n} \rightarrow \int_{\Omega_{bl}^-} S^- : D u_{bl} $$ and one concludes by Minty's trick that $S^+ = S(A+D u_{bl}) - S(A)$, $\: S^- = S(D u_{bl})$. It follows that $u_{bl}$ satisfies \eqref{BL1}-\eqref{BL2} in a weak sense. Finally, one can perform on $u_{bl}$ the manipulations of the previous paragraph, so that it satisfies \eqref{aprioriestimate}.
\subsubsection*{Uniqueness}
Let $u_{bl}^1$ and $u_{bl}^2$ be weak solutions of \eqref{BL1}-\eqref{BL2}, that is satisfying the variational formulation
\begin{equation} \label{VF}
\int_{\Omega_{bl}^+} S(A+ D u_{bl}^i) : D \varphi \: + \: \int_{\Omega_{bl}^-} S(D u_{bl}^i) : D \varphi = -\int_{y_2 = 0} S(A)n \cdot \varphi {\rm\,d}S, \quad i=1,2
\end{equation}
for all smooth divergence free fields $\varphi \in C^\infty_c(\Omega_{bl})$.
The point is then to replace $\varphi$ by $u_{bl}^1 - u_{bl}^2$, to obtain
\begin{equation} \label{zeroidentity}
\int_{\Omega_{bl}^+} \left( S(A+ D u_{bl}^1) - S(A + D u_{bl}^2) \right) : D (u_{bl}^1 - u_{bl}^2) \: + \: \int_{\Omega_{bl}^-} (S(D u_{bl}^1) - S(D u_{bl}^2) : D (u_{bl}^1 - u_{bl}^2) = 0 . \end{equation} Rigorously, one constructs by convolution a sequence $\varphi_n$ such that $D \varphi_n$ converges appropriately to $D u_{bl}^1 - D u_{bl}^2$. In the case $p < 2$, the convergence holds in $(L^2(\Omega_{bl}^+) \cap L^\infty(\Omega_{bl}^+)) + L^p(\Omega_{bl}^+)$, respectively in $L^p(\Omega_{bl}^-)$. One can pass to the limit as $n$ goes to infinity because $$ S(A+D u_{bl}^1) - S(A+D u_{bl}^2) = \left( S(A+D u_{bl}^1) - S(A) \right) + \left( S(A) - S(A+D u_{bl}^2) \right), $$ respectively $S(D u_{bl}^1) - S(D u_{bl}^2)$, belongs to the dual space: see the arguments of the construction scheme of section \ref{sectionWP}.
Eventually, by strict convexity of $M \rightarrow |M|^p$ ($p > 1$), \eqref{zeroidentity} implies that $D u_{bl}^1 = D u_{bl}^{2}$. This implies that $u_{bl}^1 - u_{bl}^2$ is a constant (dimension is $2$), and due to the zero boundary condition at $\pa \Omega_{bl}$, we get $u_{bl}^1 = u_{bl}^2$.
\subsection{Saint-Venant estimate} \label{subsecstvenant} We focus in this paragraph on the asymptotic behaviour of $u_{bl}$ as $y_2$ goes to infinity. The point is to show exponential convergence of $u_{bl}$ to a constant field. At first, we can use interior regularity results for the generalized Stokes equation in two dimensions. In particular, pondering on the results of \cite{Wolf} for $p < 2$, and \cite{Kaplicky2002} for $p \ge 2$, we have : \begin{Lemma} \label{lemma_unifbound} The solution built in Theorem \ref{thWP} satisfies: $u_{bl}$ has $C^{1,\alpha}$ regularity over $\Omega_{bl} \cap \{ y_2 > 1\}$ for some $0 < \alpha < 1$. In particular, $\nabla u_{bl}$ is bounded uniformly over $\Omega_{bl} \cap \{ y_2 > 1\}$. \end{Lemma}
{\em Proof}. Let $0 \le t < s$. We define $\Omega_{bl}^{t,s} \: := \: \Omega_{bl} \cap \{t < y_2 \le s\}$. Note that $\Omega_{bl} \cap \{y_2 > 1 \} = \cup_{t \in \N_*} \Omega_{bl}^{t,t+1}$. Moreover, from the {\it a priori estimate} \eqref{aprioriestimate} or \eqref{es:Dup}, we deduce easily that
\begin{equation} \label{uniformLp}
\| D u_{bl} \|_{L^p(\Omega_{bl}^{t,t+2})} \le C
\end{equation}
for all $t > 0$, for a constant $C$ that does not depend on $t$. We then introduce:
\begin{equation*}
v_t \: := \: 2 A \left(0,y_2 - (t+\frac{1}{2})\right) + u_{bl} \: - \frac{1}{2} \: \int_{\Omega_{bl}^{t-\frac{1}{2},t+\frac{3}{2}}} u_{bl} \dy, \quad y \in \Omega_{bl}^{t-\frac{1}{2},t+\frac{3}{2}}, \quad \forall t \in \N_*.
\end{equation*}
From \eqref{BL1}:
$$ -\dive(S(D v_t)) + \nabla p_{bl} = 0, \quad \dive v_t= 0 \quad \mbox{ in } \: \Omega_{bl}^{t-1/2,t+3/2}, \quad \forall t \in \N_*. $$
Moreover, we get for some $C$ independent of $t$:
\begin{equation} \label{uniformLpv}
\| v_t \|_{W^{1,p}(\Omega_{bl}^{t-1/2,t+3/2})} \le C \quad \forall t \in \N_*.
\end{equation} Note that this $W^{1,p}$ control follows from \eqref{uniformLp}: indeed, one can apply the Poincar\'e inequality for functions with zero mean, and then the Korn inequality. One can then ponder on the interior regularity results of articles \cite{Wolf} and \cite{Kaplicky2002}, depending on the value of $p$: $v_{t}$ has $C^{1,\alpha}$ regularity over $\Omega_{bl}^{t,t+1}$ for some $\alpha \in (0,1)$ (independent of $t$): for some $C'$,
$$\| v_t \|_{C^{1,\alpha}(\Omega_{bl}^{t,t+1})} \le C', \quad \mbox{and in particular} \quad \| \nabla v_t \|_{L^\infty(\Omega_{bl}^{t,t+1})} \le C' \quad \forall t \in \N_*.$$ Going back to $u_{bl}$ concludes the proof of the lemma.
We are now ready to establish a keypoint in the proof of Theorem \ref{thEC}, called a {\it Saint-Venant estimate}: namely, we show that the energy of the solution located above $y_2 = t$ decays exponentially with $t$. In our context, a good energy is
$$ E(t) \: := \: \int_{\{y_2 > t\}} | \nabla u_{bl} |^2 \dy $$
for $t > 1$. Indeed, from Lemma \ref{lemma_unifbound}, there exists $M$ such that $|D u_{bl}| \le M$ for all $y$ with $y_2 > 1$. In particular, in the case $p < 2$, when localized above $y_2 =1$, the energy functional that appears at the left hand-side of \eqref{aprioriestimate} only involves the $L^2$ norm of the symmetric gradient (or of the gradient by the homogeneous Korn inequality, {\it cf} the appendix). Hence, $\nabla u_{bl} \in L^2(\Omega_{bl} \cap \{ y_2 > 1 \})$. The same holds for $p \ge 2$, thanks to \eqref{es:Dup}. \begin{Proposition} \label{expdecay} There exists $C,\delta > 0$, such that $E(t) \le C \exp(-\delta t)$. \end{Proposition}
{\em Proof}. Let $t > 1$, $\Omega_{bl}^t := \Omega_{bl} \cap \{ y_2 > t \}$. Let $M$ such that $|D u_{bl}|$ is bounded by $M$ over $\Omega_{bl}^1$, which exists due to Lemma~\ref{lemma_unifbound}. As explained just above, one has
$ \int_{\Omega_{bl}^1 } |D u_{bl}|^2 < +\infty$, and by Korn inequality $E(1)$ is finite. In particular, $E(t)$ goes to zero as $t \rightarrow +\infty$ and the point is to quantify the speed of convergence. By the use of inequality \eqref{ineq1} (with $a = A$, $b = A + D u_{bl}$), we find \begin{equation} \label{boundE} \begin{aligned}
E(t) \le C \int_{\Omega_{bl}^t} |D u_{bl}|^2 \dy & \le
C' \int_{\Omega_{bl}^t}
\left( |A + D u_{bl}|^{p-2}(A + D u_{bl}) - | A |^{p-2} A \right): D u_{bl} \dy
\\ &
\le C' \lim\limits_{n \to \infty } \int_{\Omega_{bl}}
\left( |A + D u_{bl}|^{p-2}( A + D u_{bl}) - |A |^{p-2} A \right): D u_{bl} \, \chi_n (y_2) \dy
\end{aligned}
\end{equation}
for a smooth $\chi_n$ with values in $[0,1]$ such that $\chi_n = 1$ over $[t,t+n$], $\chi_n=0$ outside $[t-1, t+n+1]$, and $|\chi'_n| \le 2$. Then, we integrate by parts the right-hand side, taking into account the first equation in \eqref{BL1}. We write \begin{align} \nonumber & \int_{\Omega_{bl}}
\left( |A + D u_{bl}|^{p-2}( A + D u_{bl}) - |A |^{p-2} A \right): D u_{bl} \, \chi_n (y_2) \dy \\ \nonumber = & - \int_{\Omega_{bl}} \nabla p_{bl} \cdot u_{bl} \chi_n(y_2) \dy - \int_{\Omega_{bl}} \left( (S(A + D u_{bl}) - S(A)) \left( \begin{smallmatrix} 0 \\ \chi'_n \end{smallmatrix} \right) \right) \cdot u_{bl} \dy \\ \label{I1I2} = &\int_{\Omega_{bl}} \left(S(A) - S(A+D u_{bl})) \left( \begin{smallmatrix} 0 \\ \chi'_n \end{smallmatrix} \right) \right) \cdot u_{bl} \dy + \int_{\Omega_{bl}} p_{bl} \chi'_n u_{bl,2} \dy \: := \: I_1 + I_2. \end{align} To estimate $I_1$ and $I_2$, we shall make use of simple vector inequalities. Namely: \begin{equation} \label{ineq3}
\mbox{for all $p \in ]1,2]$, for all vectors $a,b$, $a\neq 0$, }, \quad | |b|^{p-2} b - |a|^{p-2} a | \: \le \: C_{p,a} \, |b-a| , \end{equation} whereas \begin{equation} \label{ineq3duo}
\mbox{for all $p > 2$, for all vectors $a,b$, $|b| \le M$}, \quad | |b|^{p-2} b - |a|^{p-2} a | \: \le \: C_{p,a,M} \, |b-a| . \end{equation} The latter is a simple application of the finite increment inequality. As regards the former, we distinguish between two cases: \begin{itemize}
\item If $|b-a| < \frac{|a|}{2}$, it follows from the finite increments inequality.
\item If $|b-a| \ge \frac{|a|}{2}$, we simply write \begin{align*}
| |b|^{p-2} b - |a|^{p-2} a | \: & \le \: |b|^{p-1} + |a|^{p-1} \: \le \: (3^{p-1} + 2^{p-1}) |b-a|^{p-1} \le \: (3^{p-1} + 2^{p-1}) (\frac{|a|}{2})^{p-2} |b-a| \end{align*}
using that $\left(\frac{2 |b-a|}{|a|}\right)^{p-1} \le \left(\frac{2 |b-a|}{|a|}\right)$ for $1 < p \le 2$. \end{itemize}
We shall also make use of the following: \begin{Lemma} \label{lem_averages} For any height $t > 0$ \begin{description} \item[i)] $\int_{\{y_2 =t\}} u_{bl,2} = 0$. \item[ii)] $\int_{\{y_2=t\}} (S(A+D u_{bl}) - S(A)) \left( \begin{smallmatrix} 0 \\ 1 \end{smallmatrix} \right) \cdot \left( \begin{smallmatrix} 1 \\ 0 \end{smallmatrix} \right) = 0$. \end{description} \end{Lemma} {\em Proof of the lemma}.
\noindent i) The integration of the divergence-free condition over $\Omega_{bl}^{0,t}$ leads to \begin{align*} 0 = \int_{\Omega_{bl}^{0,t}} \dive u_{bl} & = \int_{\{y_2 = t \}} u_{bl,2} - \int_{\{y_2 = 0^+ \}}u_{bl,2} = \int_{\{y_2 = t \}} u_{bl,2} - \int_{\{y_2 = 0^- \}} u_{bl,2} \\ & = \int_{\{y_2 = t \}} u_{bl,2} - \int_{\Omega_{bl^-}} \dive u_{bl} + \int_{\Gamma_{bl}} u_{bl} \cdot n = \int_{\{y_2 = t \}} u_{bl,2} , \end{align*} where the second and fourth inequalities come respectively from the no-jump condition of $u_{bl}$ at $y_2 = 0$ and the Dirichlet condition at $\Gamma_{bl}$.
\noindent ii) By integration of the first equation in \eqref{BL1} over $\Omega_{bl}^{0,t}$ we get: $$ \int_{y_2 = t} (S(A+D u_{bl}) - S(A) - p_{bl} Id) \left( \begin{smallmatrix} 0 \\ 1 \end{smallmatrix} \right) =
\int_{y_2 = 0^+} (S(A+D u_{bl}) - S(A) - p_{bl} Id) \left( \begin{smallmatrix} 0 \\ 1 \end{smallmatrix} \right). $$
In particular, the quantity $$ I := \int_{y_2 = t} (S(A+D u_{bl}) - S(A) - p_{bl} Id) \left( \begin{smallmatrix} 0 \\ 1 \end{smallmatrix} \right) \cdot \left( \begin{smallmatrix} 1 \\ 0 \end{smallmatrix} \right) =
\int_{y_2 = t} ( S(A+D u_{bl}) - S(A)) \left( \begin{smallmatrix} 0 \\ 1 \end{smallmatrix} \right) \cdot \left( \begin{smallmatrix} 1 \\ 0 \end{smallmatrix} \right) $$ is independent of the variable $t$. To show that it is zero, we apply inequality \eqref{ineq3} or \eqref{ineq3duo} with $a = A$ and $b = A + D u_{bl}$, so that
$$I^2 \: \le \: C \left( \int_{\{y_2=t\}} |D u_{bl}| \right)^2 \: \le \: C' \int_{\{y_2=t\}} |D u_{bl}|^2 $$ ($C$ is bounded by Lemma~\ref{lemma_unifbound}). As $D u_{bl}$ belongs to $L^2(\Omega^1_{bl})$, there exists a sequence $t_n$ such that
$ \int_{\{y_2=t_n\}} |D u_{bl}|^2 \rightarrow 0$ as $n \rightarrow +\infty$. It follows that $I = 0$. This concludes the proof of the Lemma.
We can now come back to the treatment of $I_1$ and $I_2$. \begin{itemize} \item Treatment of $I_1$. \end{itemize} We note that $\chi'_n$ is supported in $[t-1,t] \cup [t+n,t+n+1]$. By Lemma \ref{lem_averages} we can write \begin{align}\label{I1a}
I_1 & = \int_{\Omega_{bl}^{t-1,t}} \left(S(A) - S(A+D u_{bl})) \left( \begin{smallmatrix} 0 \\ \chi'_n \end{smallmatrix} \right) \right) \cdot (u_{bl} - \overline{c}) \\
& + \: \int_{\Omega_{bl}^{t+n,t+n+1}} \left(S(A) - S(A+D u_{bl})) \left( \begin{smallmatrix} 0 \\ \chi'_n \end{smallmatrix} \right) \right) \cdot (u_{bl} - \overline{c}_n) \: := I_{1,1} + I_{1,2} ,
\end{align} where \begin{equation}\label{mean} \overline{c} := \Xint-_{\Omega_{bl}^{t-1,t}} u_{bl} =
\left( \Xint-_{\Omega_{bl}^{t-1,t}} u_{bl,1} , 0 \right) \quad \mbox{ and } \quad \overline{c}_n := \Xint-_{\Omega_{bl}^{t+n,t+n+1}} u_{bl} =
\left( \Xint-_{\Omega_{bl}^{t+n,t+n+1}} u_{bl,1} , 0 \right).
\end{equation} Again, we apply inequality \eqref{ineq3} or \eqref{ineq3duo} to find
$$ I_{1,1} \le C \int_{\Omega_{bl}^{t-1,t}} | D u_{bl} | \, |u_{bl} - \overline{c}| $$ and by the Poincar\'e inequality for functions with zero mean, we easily deduce that
$$ I_{1,1} \le C' \int_{\Omega_{bl}^{t-1,t}} | \nabla u_{bl} |^2 = C' \left( E(t-1) - E(t) \right). $$ An upper bound on $I_{1,2}$ can be derived in the same way: $$ I_{1,2} \le C' (E(t+n) - E(t+n+1)) ,$$ where the right-hand side going to zero as $n \rightarrow +\infty$ since $E(t') \to 0$ as $t'\to \infty$. Eventually: \begin{equation} \label{estimI1} \limsup_{n \rightarrow +\infty} I_{1} \: \le \: C \left(E(t-1) - E(t) \right). \end{equation} \begin{itemize} \item Treatment of $I_2$. \end{itemize} We can again use the decomposition \begin{equation}\label{I2a}
I_2 = \int_{\Omega_{bl}^{t-1,t}} p_{bl} \chi'_n u_{bl,2} \: + \: \int_{\Omega_{bl}^{t+n,t+n+1}} p_{bl} \chi'_n u_{bl,2} \: := \: I_{2,1} + I_{2,2}.
\end{equation} From Lemma \ref{lem_averages} i), we infer that $$ \int_{\Omega_{bl}^{t-1,t}} \chi'_n(y_2) u_{bl,2}(y) \, \dy = 0. $$
By standard results, there exists $w \in H^1_0(\Omega_{bl}^{t-1,t})$ satisfying $ \dive w(y) = \chi'_n(y_2) u_{bl,2}(y), \quad y \in \Omega_{bl}^{t-1,t}$, and the estimate
$$\| w \|_{H^1(\Omega_{bl}^{t-1,t})} \le C \| \chi'_n(y_2) u_{bl,2}(y) \|_{L^2(\Omega_{bl}^{t-1,t})} \le C' \| u_{bl,2} \|_{L^2(\Omega_{bl}^{t-1,t})}, $$
for constants $C,C'$ that do not depend on $t$. As $w$ is zero at the boundary:
$$ I_{1,2} = \int_{\Omega_{bl}^{t-1,t}} p_{bl} \dive w = - \int_{\Omega_{bl}^{t-1,t}} \nabla p_{bl} \cdot w = \int_{\Omega_{bl}^{t-1,t}} (S(A+D u_{bl}) - S(A)) \cdot \nabla w $$ where the last equality comes from \eqref{BL1}. We find as before ({\it cf} \eqref{ineq3} or \eqref{ineq3duo}): \begin{align*}
|I_{1,2}| & \le C \int_{\Omega_{bl}^{t-1,t}} | D u_{bl} | |\nabla w| \le C \| D u_{bl} \|_{L^2(\Omega_{bl}^{t-1,t})} \| \nabla w \|_{L^2(\Omega_{bl}^{t-1,t})} \\
& \le C' \| D u_{bl} \|_{L^2(\Omega_{bl}^{t-1,t})} \, \|u_{bl,2} \|_{L^2(\Omega_{bl}^{t-1,t})} \le C'' \| \nabla u_{bl} \|_{L^2(\Omega_{bl}^{t-1,t})}^2 \end{align*} where we have controlled the $L^2$ norm of $u_{bl,2}$ by the $L^2$ norm of its gradient (we recall that $u_{bl,2}$ has zero mean). A similar treatment can be performed with $I_{2,2}$, so that $I_{2,1} \le C (E(t-1) - E(t))$, $\: I_{2,2} \le C ( E(t+n) - E(t+n+1))$ and
\begin{equation} \label{estimI2} \limsup_{n \rightarrow +\infty} I_{2} \: \le \: C \left(E(t-1) - E(t) \right). \end{equation}
Finally, combining \eqref{boundE}, \eqref{I1I2}, \eqref{estimI1} and \eqref{estimI2}, we get $$ E(t) \le C (E(t-1) - E(t)) $$ for some $C > 0$. It is well-known that this kind of differential inequality implies the exponential decay of Proposition \ref{expdecay} (see the appendix). The proof of the Proposition is therefore complete. We have now all the ingredients to show Theorem \ref{thEC}.
{\em Proof of Theorem \ref{thEC}}. Thanks to the regularity Lemma \ref{lemma_unifbound}, we know that $\nabla u_{bl}$ is uniformly bounded over $\Omega_{bl}^1$, and belongs to $L^2(\Omega_{bl}^1)$. Of course, this implies that $\nabla u_{bl}$ belongs to $L^q(\Omega_{bl}^1)$ for all $q \in [2,+\infty]$. More precisely, combining the $L^\infty$ bound with the $L^2$ exponential decay of Proposition \ref{expdecay}, we have that \begin{equation} \label{expdecayLq}
\| \nabla u_{bl} \|_{L^q(\Omega_{bl}^t)} \le C \exp(-\delta t)
\end{equation}
(for some $C$ and $\delta$ depending on $q$). This exponential decay extends straightforwardly to all $1 \le q < +\infty$. Let us now fix $q > 2$. To understand the behavior of $u$ itself, we write the Sobolev inequality: for all $y$ and $y' \in B(y,r)$,
\begin{equation} \label{Sobolevineq}
|u(y') - u(y)| \: \le \: C r^{1-\frac{2}{q}} \left( \int_{B(y,2r)} |\nabla u(z)|^q dz \right)^{1/q}.
\end{equation}
We deduce from there that: for all $y_2 > 2$, for all $s \ge 0$, \begin{align*}
& |u_{bl}(y_1,y_2+s) - u_{bl}(y_1,y_2)| \\
& \le |u_{bl}(y_1,y_2+s) - u_{bl}(y_1,y_2+ \lfloor s \rfloor) | + \sum_{k=0}^{\lfloor s \rfloor-1} |u_{bl}(y_1,y_2+k+1) - u_{bl}(y_1,y_2+k)| \\
& \le C \left( \| \nabla u_{bl} \|_{L^q(B((y_1,y_2+s)^t,1))} + \sum_{k=0}^{\lfloor s \rfloor-1} \| \nabla u_{bl} \|_{L^q(B((y_1,y_2+k)^t,1))} \right) \\
& \le C' \left( e^{-\delta(y_2+s)} + \sum_{k=0}^{\lfloor s \rfloor-1} e^{-\delta (y_2 + k)} \right)
\end{align*}
where the last inequality comes from \eqref{Sobolevineq}. This implies that $u_{bl}$ satisfies the Cauchy criterion uniformly in $y_1$, and thus converges uniformly in $y_1$ to some $u^\infty = u^\infty(y_1)$ as $y_2 \rightarrow +\infty$. To show that $u^\infty$ is a constant field, we rely again on \eqref{Sobolevineq}, which yields for all $|y_1 - y'_1| \le 1$:
$$ |u_{bl}(y_1,y_2) - u_{bl}(y'_1, y_2) | \le C |y_1 - y'_1|^{1 - \frac{2}{q}} \| \nabla u_{bl} \|_{L^q(B((y_1,y_2)^t,1))} \le C' e^{-\delta y_2}. $$ Sending $y_2$ to infinity gives: $u^\infty(y_1) = u^\infty(y'_1)$. Finally, the fact that $u^\infty$ is a horizontal vector field follows from Lemma \ref{lem_averages}, point i). This concludes the proof of the Theorem~\ref{thEC}.
Eventually, for later purposes, we state \begin{Corollary} {\bf (higher order exponential decay)} \label{higherorder} \begin{itemize} \item There exists $\alpha \in (0,1)$, such that for all $s \in [0,\alpha)$, for all $1 \leq q < +\infty$, one can find $C$ and $\delta > 0$ with
$$ \| u_{bl} - u^\infty \|_{W^{s+1,q}(\Omega_{bl}^t)} \le C \exp(-\delta t), \quad \forall t \ge 1. $$ \item There exists $\alpha \in (0,1)$, such that for all $s \in [0,\alpha)$, for all $1 \leq q < +\infty$, one can find $C$ and $\delta > 0$ with
$$ \| p_{bl} - p^t \|_{W^{s,q}(\Omega_{bl}^{t,t+1})} \: \le \: C \exp(-\delta t), \quad \forall t \ge 1, \quad \mbox{for some constant $p^t$}.$$ \end{itemize} \end{Corollary} {\em Proof of the corollary}. It was established above that
$$ |u(y_1,y_2+s) - u(y_1,y_2)| \le C' \left( e^{-\delta'(y_2+s)} + \sum_{k=0}^{\lfloor s \rfloor -1} e^{-\delta' (y_2 + k)}\right) . $$ for some $C'$ and $\delta' > 0$. From there, after sending $s$ to infinity, it is easily deduced that
$$ \| u_{bl} - u^\infty \|_{L^q(\Omega_{bl}^t)} \le C \exp(-\delta t) .$$ It then remains to control the $W^{s,q}$ norm of $\nabla u_{bl}$. This control comes from the $C^{0,\alpha}$ uniform bound on $\nabla u_{bl}$ over $\Omega_{bl}^1$, see Lemma \ref{lemma_unifbound}. By Sobolev imbedding, it follows that
$$ \| \nabla u_{bl} \|_{W^{s,q}(\Omega_{bl}^{t,t+1})} \le C, \quad \forall s \in [0,\alpha), \forall 1\leq q < +\infty $$ uniformly in $t$. Interpolating this bound with the bound
$\| \nabla u_{bl} \|_{L^q(\Omega_{bl}^{t,t+1})} \le C' \exp(-\delta' t)$ previously seen, we get
$$ \| \nabla u_{bl} \|_{W^{s,q}(\Omega_{bl}^{t,t+1})} \le C'' \exp(-\delta'' t), \quad \forall s \in [0,\alpha), \forall 1\leq q < +\infty .$$ The first inequality of the Lemma follows.
The second inequality, on the pressure $p_{bl}$, is derived from the one on $u_{bl}$. This derivation is somehow standard, and we do not detail it for the sake of brevity.
\section{Error estimates, wall Laws}
\subsection{Approximation by the Poiseuille flow.}
We now go back to our primitive system \eqref{EQ1}. A standard estimate on $u^\eps$ leads to
$$ \int_{\Omega^\eps} |D u^\eps|^p \: \le \: \int_{\Omega^\eps} e_1 \cdot u^\eps. $$ The Korn inequality implies that
$$ \int_{\Omega^\eps} |\nabla u^\eps|^p \: \le \: C \int_{\Omega^\eps} |D u^\eps|^p $$ for a constant $C$ independent of $\eps$: indeed, one can extend $u^\eps$ by $0$ for $x_2 < \eps \gamma(x_1/\eps)$ and apply the inequality on the square $\T \times [-1,1]$, {\it cf} the appendix. Also, by the Poincar\'e inequality:
$$ | \int_{\Omega^\eps} e_1 \cdot u^\eps | \le C \| u^\eps \|_{L^p(\Omega^\eps)} \le C' \| \nabla u^\eps \|_{L^p(\Omega^\eps)}. $$ We find that \begin{equation} \label{basic_estimate}
\| u^\eps \|_{W^{1,p}(\Omega^\eps)} \le C. \end{equation} In particular, it provides strong convergence of $u^\eps$ in $L^p$ by the Rellich theorem (up to a subsequence). As can be easily guessed, the limit of $u^0$ in $\Omega$ is the generalized Poiseuille flow $u^0$. One can even obtain an error estimate by a direct energy estimate of the difference (extending $u^0$ and $p^0$ by zero in $R^\eps$). We focus on the case $1 < p \le 2$, and comment briefly the easier case $p \ge 2$ afterwards. We write $\ue = \uz + \we$ and $p^\ep = p^0 + q^\ep$. We find, taking into account \eqref{EQ2}:
\begin{equation}\label{EQ4}
\begin{split}
- \Div \bS(\tD \uz + \tD \we) + \Div \bS(\tD \uz) + \nabla q^\ep & = {\bbbone}_{R^\eps} e_1 \quad \mbox{ in } \Omega^\ep \setminus \Sigma_0 ,\\
\Div \we & = 0 \quad \mbox{ in } \Omega^\ep , \\
\we & = 0 \quad \mbox{ on } \Gamma^\ep \cup \Sigma_1 , \\
\we & \mbox{ is periodic in } x_1 \mbox{ with period } 1 , \\
[\we]|_{\Sigma_0} = 0, \quad [\tS ( \tD \uz + \tD \we ) {n} - \tS ( \tD \uz ) {n} - q^\ep n ]|_{\Sigma_0} & = -\tS ( \tD \uz) {n}|_{x_2=0^+} .
\end{split}
\end{equation} In particular, performing an energy estimate and distinguishing between $\Omega$ and $R^\eps$, we find
\begin{equation}\label{weakEQ4}
\int_{\Omega} \left( \tS( \tD \uz + \tD \we) - \tS( \tD \uz) \right) : \tD \we + \int_{R^\ep} \tS( \tD \we ) : \tD \we
= - \int_{\Sigma_0} \tS (\tD \uz ){n}\vert_{x_2 = 0^+} \cdot \we {\rm d}S + \int_{R^\eps} e_1 \cdot \we
\end{equation}
Relying on inequalities \eqref{ineq1}-\eqref{ineq2}, we get for any $M > \| Du^0 \|_{L^\infty}$: \begin{multline}\label{zeroap1}
\| D \we \|_{L^p(\Omega \cap \{ |D \we| \ge M\})}^{p} + \| D \we \|_{L^2(\Omega \cap \{ |D \we| \le M\})}^{2} + \| D \we \|_{L^p(R^\eps)}^{p} \\
\le C \left( \left| \int_{\Sigma_0} \bS(\tD \uz) {n} \cdot \we {\rm\,d}S \right| + \left| \int_{R^\eps} e_1 \cdot \we \right| \right)
\end{multline} Then by the H\"older inequality and by Proposition~\ref{rescaledTracePoincare} in the appendix, we have that
\begin{equation}\label{es1}
| \int_{R^\ep} e_1 \cdot \we | \leq \eps^{\frac{p-1}{p}} \| \we \|_{L^p(R^\eps)}
\leq C \eps^{1 + \frac{p-1}{p}} \| \nabla \we \|_{L^p(R^\ep)} .
\end{equation} Next, since $D \uz$ is given explicitly and uniformly bounded, the Proposition~\ref{rescaledTracePoincare} provides
\begin{equation}\label{es2}
\begin{split}
& | \int_{\Sigma_0} ( \tS(\tD \uz) {n}\vert_{x_2 = 0^+} \cdot \we {\rm\,d}S |
\leq C \| \we \|_{L^{p}(\Sigma_0)}
\leq C' \ep^{\frac{p-1}{p}} \| \nabla \we \|_{L^{p}(R^\ep)} .
\end{split}
\end{equation} Note that, as $\we$ is zero at the lower boundary of the channel, we can extend it by $0$ below $R^\eps$ and apply Korn inequalities in a strip (see the appendix). We find
$$ \| \nabla \we \|_{L^p(R^\ep)} \le C \| D \we \|_{L^p(R^\ep)} $$ for some constant $C > 0$ independent of $\eps$. Summarising, we get
\begin{equation*}
\| D \we \|_{L^p(\Omega \cap \{ |D \we| \ge M\})}^{p} + \| D \we \|_{L^2(\Omega \cap \{ |D \we| \le M\})}^{2} + \| D \we \|_{L^p(R^\eps)}^{p}
\le C \ep^{\frac{p-1}{p}} \| D \we \|_{L^p(R^\ep)} \end{equation*} and consequently \begin{equation}\label{zeroap2}
\| D \we \|_{L^p(\Omega \cap \{ |D \we| \ge M\})}^{p} + \| D \we \|_{L^2(\Omega \cap \{ |D \we| \le M\})}^{2} + \| D \we \|_{L^p(R^\eps)}^{p}
\le C \ep \end{equation}
In the case $p \ge 2$ one needs to use \eqref{abp_1} instead of \eqref{ineq1}-\eqref{ineq2} what yields \begin{equation} \label{zeroap3}
\| D \we \|_{L^p(\Omega^\eps)} \le C \eps^{\frac{1}{p}}, \quad p \in [2, \infty). \end{equation} \subsection{Construction of a refined approximation} The aim of this section is to design a better approximation of the exact solution $u^\eps$ of \eqref{EQ1}. This approximation will of course involve the boundary layer profile $u_{bl}$ studied in the previous section. Consequences of this approximation in terms of wall laws will be discussed in paragraph \ref{parag_wall_laws}.
From the previous paragraph, we know that the Poiseuille flow $u^0$ is the limit of $u^\eps$ in $W^{1,p}(\Omega)$. However, the extension of $u^0$ by $0$ in the rough part of the channel was responsible for a jump of the stress tensor at $\Sigma_0$. This jump was the main limitation of the error estimates \eqref{zeroap2}-\eqref{zeroap3}, and the reason for the introduction of the boundary layer term $u_{bl}$. Hence, we hope to have a better approximation replacing $u^0$ by $u^0 + \eps u_{bl}(\cdot/\eps)$. Actually, one can still improve the approximation, accounting for the so-called boundary layer tail $u^\infty$. More precisely, {\em in the Newtonian case}, a good idea is to replace $u^0$ by the solution $u^{0,\eps}$ of the Couette problem: $$ -\Delta u^{0,\eps} + \nabla p^{0,\eps} = 0, \quad \dive u^{0,\eps} = 0, \quad u^{0,\eps}\vert_{\Sigma_0} = \eps u^\infty, \quad u^{0,\eps}\vert_{x_2 = 1} = 0. $$ One then defines: $$ u^{\eps} = u^{0,\eps} + \eps (u_{bl}(\cdot/\eps) - u_\infty) + r^\eps \: \mbox{ in $\Omega$}, \quad u^\eps = \eps u_{bl}(\cdot/\eps) \: \mbox{ in $R^\eps$},$$ where $r^\eps$ is a small divergence-free remainder correcting the $O(\exp(-\delta/\eps))$ trace of $u_{bl} - u^\infty$ at $\{x_2 = 1 \}$.
However, for technical reasons, the above approximation is not so successful in our context, so that we need to modify it a little. We proceed as follows. Let $N$ a large constant to be fixed later. We introduce:
$$ \Omega^{\eps}_N := \Omega^\eps \cap \{x_2 > N \eps |\ln \eps| \}, \quad \Omega^{\eps}_{0,N} = \Omega^\eps \cap \{ 0 < x_2 < N \eps |\ln \eps| \}, \quad \mbox{and} \quad \Sigma_N = \Pi \times \{ x_2 = N \eps |\ln \eps| \}. $$ First, we introduce the solution $u^{0,\eps}$ of \begin{equation} \label{u0eps} \left\{
\begin{aligned}
- \dive S(D u^{0,\eps}) + \nabla p^{0,\eps} & = e_1, \quad x \in \Omega^\eps_N, \\
\dive u^{0,\eps} & = 0, \quad x \in \Omega^\eps_N, \\
u^{0,\eps}\vert_{\Sigma_N} & = \left( x \rightarrow \left( \begin{smallmatrix} U'(0) x_2 \\ 0 \end{smallmatrix} \right) + \eps u^\infty \right)\vert_{\Sigma_N}, \\
u^{0,\eps}\vert_{\{ x_2 = 1 \}} & = 0.
\end{aligned}
\right.
\end{equation}
As for the generalized Poiseuille flow, the pressure $p^{0,\eps}$ is zero, and one has an explicit expression for $u^{0,\eps} = (U^\eps(x_2),0)$. In particular, one can check that \begin{equation} \label{explicit1}
U^\eps(x_2) = \beta(\eps) - \frac{(\sqrt{2})^{p'}}{p'} \left| \frac{1}{2} + \alpha(\eps) -x_2 \right|^{p'}, \end{equation}
where $\alpha(\eps)$ satisfies the equation ($x_{2,N} := N \eps |\ln \eps|$) \begin{equation} \label{explicit2}
-\frac{1}{p'}(\sqrt{2})^{p'} \left( \left| \frac{1}{2} + \alpha(\eps) - x_{2,N} \right|^{p'} - \left| \frac{1}{2} - \alpha(\eps) \right|^{p'} \right) = U'(0) x_{2,N} + \eps U^\infty \end{equation} and
$$ \beta(\eps) = \frac{(\sqrt{2})^{p'}}{p'} \left| \frac{1}{2} - \alpha(\eps)\right|^{p'} . $$
By the Taylor expansion, we find that
\begin{equation} \label{explicit3}
\alpha(\eps) = - \sqrt{2}^{p'-4} \eps U^\infty + O(\eps^2 |\ln \eps|^2).
\end{equation}
This will be used later.
Then, we consider the Bogovski problem
\begin{equation} \label{Bogov}
\left\{
\begin{aligned}
\dive r^\eps & = 0 \quad \mbox{in} \: \Omega^\eps_{N}, \\
r^\eps\vert_{\Sigma_N} & = \eps (u_{bl}(\cdot/\eps) - u^\infty)\vert_{\Sigma_N}, \\
r^\eps\vert_{\{x_2 = 1\}} & = 0.
\end{aligned}
\right.
\end{equation}
Since $u^\infty = (U^\infty,0)$, note that
$$ \int_{\Sigma_N} \eps (u_{bl}(\cdot/\eps) - u^\infty) \cdot e_2 = \int_{\Omega^\eps_N \cup \overline{R^\eps}} {\rm div}_y u_{bl}(\cdot/\eps) = 0. $$
Hence, the compatibility condition for solvability of \eqref{Bogov} is fulfilled: there exists a solution
$r^\eps$ satisfying
$$ \| r^\eps \|_{W^{1,p}(\Omega^\eps_N)} \le C \eps \| u_{bl}(\cdot/\eps) - u^\infty \|_{W^{1-\frac{1}{p},p}(\Sigma_N)}. $$ Using the first estimate of Corollary \ref{higherorder}, we find \begin{equation} \label{estimreps}
\| r^\eps \|_{W^{1,p}(\Omega^\eps_N)} \le C \eps^{\frac{1}{p}} \exp(-\delta N |\ln \eps|).
\end{equation}
Finally, we define the approximation $(u^\eps_{app}, p^\eps_{app})$ by the formula
\begin{equation} \label{uepsapp}
u^\eps_{app}(x) = \left\{ \begin{aligned} & u^{0,\eps}(x) + r^\eps(x) \quad x \in \Omega^\eps_N ,\\ & \left( \begin{smallmatrix} U'(0)x_2 \\ 0 \end{smallmatrix} \right) + \eps u_{bl}(x/\eps), \quad x \in \Omega^\eps_{0,N}, \\ & \eps u_{bl}(x/\eps), \quad x \in R^\eps , \end{aligned} \right. \end{equation} whereas
\begin{equation}
p^\eps_{app}(x) = \left\{ \begin{aligned} & 0 \quad x \in \Omega^\eps_N ,\\ & p_{bl}(x/\eps) \quad \quad x \in \Omega^\eps_{0,N} \cup R^\eps. \end{aligned} \right. \end{equation} With such a choice: $$u^\eps_{app}\vert_{\partial \Omega^\eps} = 0, \quad \dive u^\eps_{app} = 0 \quad \mbox{over $\Omega^\eps_N \cup \Omega^\eps_{0,N} \cup R^\eps$}.$$ Moreover, $u^\eps_{app}$ has zero jump at the interfaces $\Sigma_0$ and $\Sigma_N$: $$ [u^\eps_{app}]\vert_{\Sigma_0} = 0, \quad [u^\eps_{app}]\vert_{\Sigma_N} = 0. $$ Still, the stress tensor has a jump. More precisely, we find \begin{equation} \label{stressjump} \begin{aligned} \left[S(D u^\eps_{app})n - p^\eps_{app}n\right]\vert_{\Sigma_0} & = 0, \\
\left[S(D u^\eps_{app})n - p^\eps_{app}n\right]\vert_{\Sigma_N} & = \left( S(D u^0_\eps + D r^\eps)\vert_{\{x_2 = (N \eps |\ln \eps|)^+\}} - S(A + Du_{bl}(\cdot/\eps))\vert_{\{x_2 = (N \eps |\ln \eps|)^-\}}\right) e_2 \\
& - p_{bl}(\cdot/\eps) \vert_{\{x_2 = (N \eps |\ln \eps|)^-\}} e_2 . \end{aligned} \end{equation} The next step is to obtain error estimates on $u^\eps - u^\eps_{app}$. \subsection{Error estimates} We prove here: \begin{Theorem} {\bf (Error estimates)} \label{thmerror} \begin{itemize} \item For $1 < p \le 2$, there exists $C$ such that
$$ \| u^\eps - u^\eps_{app} \|_{W^{1,p}(\Omega^\eps)} \le C (\eps |\ln \eps|)^{1+\frac{1}{p'}} .$$ \item For $p \ge 2$, there exists $C$ such that
$$ \| u^\eps - u^\eps_{app} \|_{W^{1,p}(\Omega^\eps)} \le C (\eps |\ln \eps|)^{\frac{1}{p-1}+\frac{1}{p}} . $$ \end{itemize} \end{Theorem} \begin{Remark} A more careful treatment would allow to get rid of the $\ln$ factor in the last estimate ($p \ge 2$). We do not detail this point here, as we prefer to provide a unified treatment. Also, we recall that the shear thinning case ($1 < p \le 2$) has a much broader range of applications. More comments will be made on the estimates in the last paragraph \ref{parag_wall_laws}. \end{Remark}
{\em Proof of the theorem.} We write $v^\eps = u^\eps - u^\eps_{app}$, $q^\eps = p^\eps - p^\eps_{app}$. We start from the equation \begin{equation} \label{eq_error} -\dive S(D u^\eps) + \dive S(D u^\eps_{app}) + \nabla q^\eps = e_1 + \dive S(D u^\eps_{app}) + \nabla p^\eps_{app} := F^\eps \end{equation} satisfied in $\Omega^\eps \setminus (\Sigma_0 \cup \Sigma_N)$.
A quick computation shows that \begin{equation*} F^\eps = \left\{ \begin{aligned} & \dive S(D u^{0,\eps} + D r^\eps) - S(D u^{0,\eps}), \quad x \in \Omega^\eps_N, \\ & e_1, \quad x \in \Omega^\eps_{0,N} \cup R^\eps. \end{aligned} \right. \end{equation*} Defining $$ \langle F^\eps, v^\eps \rangle := \int_{\Omega^\eps_N} F^\eps \cdot v^\eps + \int_{\Omega^\eps_{0,N}} F^\eps \cdot v^\eps + \int_{R^\eps} F^\eps \cdot v^\eps $$ we get:
$$ | \langle F^\eps, v^\eps \rangle | \le \alpha_\eps \| \nabla v^\eps \|_{L^p(\Omega^\eps_N)} + \beta_\eps \| v^\eps \|_{L^p(\Sigma_N)} + \| v^\eps \|_{L^1(\Omega^\eps \setminus \Omega^\eps_N)} $$ where
$$ \alpha_\eps := \| S(D u^{0,\eps} + D r^\eps) - S(D u^{0,\eps}) \|_{L^{p'}(\Omega^\eps_N)}, \quad \beta_\eps := \| \left(S(D u^{0,\eps} + D r^\eps) - S(D u^{0,\eps})\right)e_2 \|_{L^{p'}(\Sigma_N)}. $$ We then use the inequalities \begin{equation} \label{poincarelike} \begin{aligned}
\| v^\eps \|_{L^p(\Sigma_N)} \: \le \: C (\eps |\ln \eps|)^{1/p'} \| \nabla v^\eps \|_{L^p(\Omega^\eps)}, \\
\| v^\eps \|_{L^1(\Omega^\eps \setminus \Omega^\eps_N)} \le C \eps^{\frac{1}{p'}} \| v^\eps \|_{L^p(\Omega^\eps \setminus \Omega^\eps_N)} \: \le \: C \eps^{\frac{1}{p'}} (\eps |\ln \eps|) \| \nabla v^\eps \|_{L^p(\Omega^\eps)}
\end{aligned}
\end{equation}
(see the appendix for similar ones). We end up with \begin{equation}
| \langle F^\eps, v^\eps \rangle | \le C \left( \alpha_\eps + \beta_\eps (\eps |\ln \eps|)^{1/p'} + \eps^{\frac{1}{p'}} (\eps |\ln \eps|) \right) \| \nabla v^\eps \|_{L^p(\Omega^\eps)} . \end{equation}
Back to \eqref{eq_error}, after multiplication by $v^\eps$ and integration over $\Omega^\eps$, we find: \begin{equation*} \begin{aligned}
& \int_{\Omega^\eps} \left( S(D u^\eps) - S(Du^\eps_{app}) \right) :\nabla v^\eps \\
& \le C \left( \alpha_\eps + \beta_\eps (\eps |\ln \eps|)^{1/p'} + \eps^{\frac{1}{p'}} (\eps |\ln \eps|) \right) \| \nabla v^\eps \|_{L^p(\Omega^\eps)} + \int_{\Sigma_N} \left( [S(D u^\eps_{app}) e_2]\vert_{\Sigma_N} \cdot v^\eps - [p^\eps_{app}]\vert_{\Sigma_N} v^\eps_2 \right). \end{aligned} \end{equation*} Let $p^{\eps,N}$ be a constant to be fixed later. As $v^\eps$ is divergence-free and zero at $\Gamma^\eps$, its flux through $\Sigma_N$ is zero: $\int_{\Sigma_N} v^\eps_2 = 0$. Hence, we can add $p^{\eps,N}$ to the pressure jump $[p^\eps_{app}]\vert_{\Sigma_N}$ without changing the surface integral. We get:
\begin{equation} \label{final_estimate} \begin{aligned}
& \int_{\Omega^\eps} \left( S(D u^\eps) - S(Du^\eps_{app}) \right) : \nabla v^\eps \\
& \le C \left( \alpha_\eps + \beta_\eps (\eps |\ln \eps|)^{1/p'} + \eps^{\frac{1}{p'}} (\eps |\ln \eps|) \right) \| \nabla v^\eps \|_{L^p(\Omega^\eps)} + \int_{\Sigma_N} \left( [S(D u^\eps_{app}) e_2]\vert_{\Sigma_N} \cdot v^\eps - ([p^\eps_{app}]\vert_{\Sigma_N} - p^{\eps,N}) v^\eps_2 \right) \\
& \le \left( \alpha_\eps + \beta_\eps (\eps |\ln \eps|)^{1/p'} + \eps^{\frac{1}{p'}} (\eps |\ln \eps|) \right) \| \nabla v^\eps \|_{L^p(\Omega^\eps)} + \gamma_\eps \| v^\eps \|_{L^p(\Sigma_N)} \\
& \le C \left( \alpha_\eps + (\beta_\eps + \gamma^\eps) (\eps |\ln \eps|)^{1/p'} + \eps^{\frac{1}{p'}} (\eps |\ln \eps|) \right) \| \nabla v^\eps \|_{L^p(\Omega^\eps)} , \end{aligned} \end{equation} where
$$ \gamma_\eps := \| [\left(S(D u^\eps_{app})]\vert_{\Sigma_N} - ([p^\eps_{app}]\vert_{\Sigma_N} - p^{\eps,N}\right) e_2) \|_{L^{p'}(\Sigma_N)}. $$ Note that we used again the first bound in \eqref{poincarelike} to go from the third to the fourth inequality. \begin{Lemma} \label{bounds} For $N$ large enough, and a good choice of $p^{\eps,N}$ there exists $C = C(N)$ such that
$$ \alpha_\eps \le C \eps^{10}, \quad \beta^\eps \le C \eps^{10}, \quad \gamma_\eps \le C \eps |\ln \eps|. $$ \end{Lemma}
Let us temporarily admit this lemma. Then, we can conclude the proof of the error estimates: \begin{itemize} \item In the case $1 \le p \le 2$, we rely on the inequality established in \cite[Proposition~5.2]{GM1975}: for all $p \in ]1,2]$, there exists $c$ such that for all $u,u' \in W_0^{1,p}(\Omega^\eps)$
$$ \int_{\Omega^\eps} \left( S(D u) - S(D u') \right) \cdot \nabla (u - u') \ge c \frac{\| Du - Du' \|^2_{L^p(\Omega^\eps)}}{(\| D u \|_{L^p(\Omega^\eps)} + \| D u' \|_{L^p(\Omega^\eps)})^{2-p}} $$ We use this inequality with $u = u^\eps$, $u' = u^\eps_{app}$. With the estimate \eqref{basic_estimate} and the Korn inequality in mind, we obtain
$$ \int_{\Omega^\eps} \left( S(D u^\eps) - S(Du^\eps_{app}) \right) \cdot \nabla v^\eps \ge c \| \nabla v^\eps \|_{L^p}^2. $$ Combining this lower bound with the upper bounds on $\alpha_\eps, \beta_\eps, \gamma_\eps$ given by the lemma, we deduce from \eqref{final_estimate} the first error estimate in Theorem \ref{thmerror}. \item In the case $2 \le p$, we use the easier inequality
$$ \int_{\Omega^\eps} \left( S(D u) - S(D u') \right) \cdot \nabla (u - u') \ge c \| D u - D u' \|_{L^p(\Omega^\eps)}^p, $$ so that
$$ \int_{\Omega^\eps} \left( S(D u^\eps) - S(Du^\eps_{app}) \right) \cdot \nabla v^\eps \ge c \| \nabla v^\eps \|_{L^p(\Omega^\eps)}^p. $$ The second error estimate from Theorem \ref{thmerror} follows. \end{itemize}
The final step is to establish the bounds of Lemma \ref{bounds}.
{\em Bound on $\alpha_\eps$ and $\beta_\eps$}. From Corollary \ref{higherorder} and the trace theorem, we deduce that \begin{equation} \label{traceubl}
\| u_{bl}(\cdot/\eps) - u^\infty \|_{W^{1+s-\frac{1}{q},q}(\{ x_2 = t \})} \le C \eps^{\frac{1}{q}-s-1}\exp(-\delta t/\eps) \end{equation} for some $s < \alpha$ (where $\alpha \in (0,1)$) and any $q > \frac{1}{s}$. Let $q > \max(p', \frac{2}{s})$. The solution $r^\eps$ of \eqref{Bogov} satisfies: $r^\eps \in W^{1+s,q}(\Omega^\eps_N)$ with \begin{equation*}
\| r^\eps \|_{W^{1+s,q}(\Omega^\eps_N)} \le C \eps^{\frac{1}{q}-s} \exp(-N\delta|\ln \eps|)
\end{equation*}
so that by Sobolev imbedding \begin{equation} \label{estimreps2}
\| D r^\eps \|_{L^\infty(\Sigma_N)} + \| D r^\eps \|_{L^q(\Sigma_N)} + \| D r^\eps \|_{L^\infty(\Omega^\eps_N)} \le C \| D r^\eps \|_{W^{s,q}(\Omega^\eps_N)} \le C \eps^{\frac{1}{q}-s} \exp(-N\delta|\ln \eps|) \end{equation}
This last inequality allows to evaluate $\beta_\eps$. Indeed, for $x \in \Sigma_N$, $C \ge |Du^{0,\eps}(x)| \ge c > 0$ uniformly in $x$. We can then use the upper bound \eqref{ineq3} for $p < 2$, or \eqref{ineq3duo} for $p \ge 2$, to obtain \begin{equation}\label{betaep}
\beta_\eps \le C \| D r^\eps \|_{L^{p'}(\Sigma_N)} \le C \| D r^\eps \|_{L^q(\Sigma_N)} \le C' \eps^{\frac{1}{q}-s} \exp(-N\delta|\ln \eps|) \le C' \eps^{10}\end{equation} for $N$ large enough.
To treat $\alpha_\eps$, we still have to pay attention to the cancellation of $D u^{0,\eps}$. Indeed, from the explicit expression of $u^{0,\eps}$, we know that there is some $x_2(\eps) \sim \frac{1}{2}$ at which $D u^{0,\eps}\vert_{x_2 = x_2(\eps)} = 0$. Namely, we write \begin{align*}
& \int_{\Omega^\eps_N} |S(D u^{0,\eps} + D r^\eps) - S(D u^{0,\eps})|^{p'} \\
& = \int_{\{x \in \Omega^\eps_N,\, | x_2 - x_2(\eps) | \le \eps^{10 p'}\}} |S(D u^{0,\eps} + D r^\eps) - S(D u^{0,\eps})|^{p'}
+ \int_{\{x\in \Omega^\eps_N,\, | x_2 - x_2(\eps) | \ge \eps^{10 p'}\}} |S(D u^{0,\eps} + D r^\eps) - S(D u^{0,\eps})|^{p'} \\ & := I_1 + I_2. \end{align*} The first integral is bounded by
$$ I_1 \le C \int_{\{x\in \Omega^\eps_N,\, | x_2 - x_2(\eps) | \le \eps^{10 p'}\}} | D u^{0,\eps} |^p + | D r^\eps |^p \le C \eps^{10 p'}, $$ where we have used the uniform bound satisfied by $D u^{0,\eps}$ and $D r^\eps$ over $\Omega^\eps_N$, see \eqref{estimreps2}. For the second integral, we can distinguish between $p < 2$ and $p \ge 2$. For $p < 2$, see \eqref{ineq3} and its proof, we get \begin{equation}\label{I_2}
I_2 \le C \int_{\{x\in \Omega^\eps_N,\, | x_2 - x_2(\eps) | \ge \eps^{10 p'}\}} |D u^{0,\eps}|^{(p-2)p'} |D r^\eps|^{p'}
\le C' \eps^{-M} \exp(-\delta' N|\ln \eps|) \end{equation} for some $M, C', \delta' > 0$, see \eqref{estimreps2}. In the case $p \ge 2$, as $D u^{0,\eps}$ and $D r^\eps$ are uniformly bounded, we derive a similar inequality by \eqref{ineq3duo}. In both cases, taking $N$ large enough, we obtain $I_2 \le C'' \eps^{10 p'}$, to end up with $\alpha_\eps \le C \eps^{10}$.
{\em Bound on $\gamma_\eps$}. We have \begin{align*}
\gamma_\eps &
\le \| \left(S(D u^{0,\eps} + D r^\eps) - S(D u^{0,\eps})\right) e_2 \|_{L^{p'}(\Sigma_N)}
+ \| \left(S(D u^{0,\eps}) - S(A)\right) e_2 \|_{L^{p'}(\Sigma_N)}
\\
& + \| (S(A) - S(A + D u_{bl}(\cdot/\eps))) e_2 \|_{L^{p'}(\Sigma_N)} + \| p_{bl}(\cdot/\eps) - p^{\eps,N} \|_{L^{p'}(\Sigma_N)} . \end{align*} The first term is $\beta_\eps$, so $O(\eps^{10})$ by previous calculations. The third term can be treated similarly to $\beta_\eps$. As $A \neq 0$, \eqref{ineq3} implies that \begin{equation}\label{I_3bis}
\| (S(A) - S(A + D u_{bl}(\cdot/\eps))) e_2 \|_{L^{p'}(\Sigma_N)} \le C \| D u_{bl}(\cdot/\eps) \|_{L^{p'}(\Sigma_N)} \le C' \exp(-\delta' N \ln \eps) ,
\end{equation} where the last inequality can be deduced from \eqref{traceubl}. It is again $O(\eps^{10})$ for $N$ large enough. For the second term of the right-hand side, we rely on the explicit expression of $u^{0,\eps}$. On the basis of \eqref{explicit1}-\eqref{explicit3}, we find that
$$ D(u^{0,\eps})\vert_{\Sigma_N} = A + O(\eps |\ln \eps|) $$ resulting in
$$ \| \left(S(D u^{0,\eps}) - S(A)\right) e_2 \|_{L^{p'}(\Sigma_N)} \le C \eps |\ln \eps|. $$ Finally, to handle the pressure term, we use the second term of Corollary \ref{higherorder}, which implies
$$ \| p_{bl} - p^t \|_{L^q(\{ y_2 = t\})} \: \le \: C \exp(-\delta t) \quad \mbox{for some constant $p^t$}.$$ We take $t = N \ln \eps$ and $p^{\eps,N} = p^t$ to get
$$ \| p_{bl}(\cdot/\eps) - p^{\eps,N} \|_{L^{p'}(\Sigma_N)} \le C' \exp(-\delta' N |\ln \eps|). $$
Taking $N$ large enough, we can make this term neglectible, say $O(\eps^{10})$. Gathering all contributions, we obtain $ \gamma_\eps \le C \eps |\ln \eps|$ as stated.
\subsection{Comment on possible wall laws} \label{parag_wall_laws} On the basis of the previous error estimates, we can now discuss the appropriate wall laws for a non-Newtonian flow above a rough wall. We focus here again on the shear thinning case ($1 < p \le 2$).
We first notice that the field $u^\eps_{app}$ (see \eqref{uepsapp}) involves in a crucial way the solution $u^ {0,\eps}$ of \eqref{u0eps}. Indeed, we know from \eqref{estimreps} that the contribution of $r^\eps$ in $W^{1,p}(\Omega^\eps_N)$ is very small for $N$ large enough. Hence, the error estimate of Theorem \ref{thmerror} implies that
$$ \| u^\eps - u^{0,\eps} \|_{W^{1,p}(\Omega^\eps_N)} = O((\eps |\ln \eps|)^{1+\frac{1}{p'}}) . $$ In other words, away from the boundary layer, $u^\eps$ is well approximated by $u^{0,\eps}$, with a power of $\eps$ strictly bigger than $1$. Although such estimate is unlikely to be optimal, it is enough to emphasize the role of the boundary layer tail $u^\infty$. Namely, the addition of the term $\eps u^\infty$ in the Dirichlet condition for $u^{0,\eps}$ (see the third line of \eqref{u0eps}) allows to go beyond a $O(\eps)$ error estimate. {\it A contrario}, the generalized Poiseuille flow $u^0$ leads to a $O(\eps)$ error only (away from the boundary layer). Notably, \begin{equation} \label{estimu0}
\| u^\eps - u^0 \|_{W^{1,p}(\Omega^\eps_N)} \ge \| u^{0,\eps} - u^0 \|_{W^{1,p}(\Omega^\eps_N)} - \| u^\eps - u^{0,\eps} \|_{W^{1,p}(\Omega^\eps_N)} \ge c \eps - o(\eps) \ge c' \eps , \end{equation} where the lower bound for $u^{0,\eps} - u^0$ is obtained using the explicit expressions.
Let us further notice that instead of considering $u^{0,\eps}$, we could consider the solution $u^0_\eps$ of \begin{equation} \label{u0epsbis} \left\{ \begin{aligned} - \dive S(u^0_\eps)) + \nabla p^0_\eps & = e_1, \quad x \in \Omega^\eps_N, \\ \dive u^0_\eps & = 0, \quad x \in \Omega^\eps_N, \\ u^0_\eps \vert_{\Sigma_0} & = \eps u^\infty, \\ u^0_\eps\vert_{\{ x_2 = 1 \}} & = 0. \end{aligned} \right. \end{equation} It reads $u^0_\eps = (U_\eps, 0)$ with
$$ U_\eps(x_2) = \beta'(\eps) - \frac{(\sqrt{2})^{p'}}{p'} \left| \frac{1}{2} + \alpha'(\eps) -x_2 \right|^{p'} $$ for $\alpha'$ and $\beta'$ satisfying $$
-\frac{1}{p'}(\sqrt{2})^{p'} \left( \left| \frac{1}{2} + \alpha'(\eps) \right|^{p'} - \left| \frac{1}{2} - \alpha'(\eps) \right|^{p'} \right) = \eps u^{\infty}_1 \quad \mbox{ and } \quad
\beta'(\eps) = \frac{(\sqrt{2})^{p'}}{p'} \left| \frac{1}{2} - \alpha'(\eps)\right|^{p'}. $$ We can compare directly these expressions to \eqref{explicit1}-\eqref{explicit2} and deduce that
$$ \| u^{0,\eps} - u^0_\eps \|_{W^{1,p}(\Omega^\eps_N)} = O(\eps |\ln \eps|), $$ which in turn implies that \begin{equation} \label{estimu0eps}
\| u^\eps - u^0_\eps \|_{W^{1,p}(\Omega^\eps_N)} = O(\eps |\ln \eps|). \end{equation} Hence, in view of \eqref{estimu0} and \eqref{estimu0eps}, we distinguish between two approximations (outside the boundary layer): \begin{itemize} \item A crude approximation, involving the generalized Poiseuille flow $u^0$. \item A refined approximation, involving $u^0_\eps$. \end{itemize} The first choice corresponds to the Dirichlet wall law $u\vert_{\Sigma_0} = 0$, and neglects the role of the roughness. The second choice takes it into account through the inhomogeneous Dirichlet condition: $u\vert_{\Sigma_0} = \eps u^\infty = \eps (U^\infty,0)$. Note that this last boundary condition can be expressed as a wall law, although slightly abstract. Indeed, $U^\infty$ can be seen as a function of the tangential shear $(D(u^0)n)_\tau\vert_{\Sigma_0} = \pa_2 u^0_1\vert_{\Sigma_0} = U'(0)$, through the mapping $$ U'(0) \: \rightarrow \: A := \left( \begin{smallmatrix} 0 & U'(0) \\ U'(0) & 0 \end{smallmatrix} \right) \: \rightarrow \: u_{bl} \:\: \mbox{solution of \eqref{BL1}-\eqref{BL2}} \: \rightarrow \: U^\infty = \lim_{y_2 \rightarrow +\infty} u_{bl,1}. $$ Denoting by ${\cal F}$ this application, we write $$(u^0_\eps)_\tau \vert_{\Sigma_0} = \eps {\cal F}((D(u^0)n)_\tau\vert_{\Sigma_0}) \approx \eps {\cal F}((D(u^0_\eps)n)_\tau\vert_{\Sigma_0})$$ whereas $ (u^0_\eps)_n = 0$. This provides the following refined wall law : $$ u_n\vert_{\Sigma_0} = 0, \quad u_\tau\vert_{\Sigma_0} = \eps {\cal F}\bigl((D(u) n)_\tau\vert_{\Sigma_0}\bigr). $$
This wall law generalizes the Navier wall law derived in the Newtonian case, where ${\cal F}$ is simply linear. Of course, it is not very explicit as it involves the nonlinear system \eqref{BL1}-\eqref{BL2}.
More studies will be necessary to obtain qualitative properties of the function ${\cal F}$, leading to a more effective boundary condition.
{\bf Acknowledgements: } The work of AWK is partially supported by Grant of National Science Center Sonata, No 2013/09/D/ST1/03692.
\section{Appendix : A few functional inequalities} \begin{Proposition} {\bf (Korn inequality)} Let $S_a := \T \times (a, a+1)$, $a \in \R$. For all $1 < p < +\infty$, there exists $C > 0$ such that: for all $a \in \R$, for all $u \in W^{1,p}(S_a)$, \begin{equation} \label{Korn1}
\| \nabla u \|_{L^p(S_a)} \: \le \: C \| D u \|_{L^p(S_a)}. \end{equation} \end{Proposition} {\em Proof.} Without loss of generality, we can show the inequality for $a = 0$: the independence of the constant $C$ with respect to $a$ follows from invariance by translation. Let us point out that the keypoint of the proposition is that the inequality is homogeneous. Indeed, it is well-known that the inhomogeneous Korn inequality \begin{equation} \label{Korn2}
\| \nabla u \|_{L^p(S_0)} \: \le \: C' \left( \| D u \|_{L^p(S_0)} + \| u \|_{L^p(S_0)}\right) \end{equation} holds. To prove the homogeneous one, we use reductio at absurdum : if \eqref{Korn1} is wrong, there exists a sequence $u_n$ in $W^{1,p}(S_0)$ such that \begin{equation} \label{absurd}
\| \nabla u_n \|_{L^p(S_0)} \: \ge \: n \| D u_n \|_{L^p(S_0)}.
\end{equation}
Up to replace $u_n$ by $u'_n := (u_n - \int_{S_0} u_n)/\| u_n\|_{L^p}$, we can further assume that
$$ \| u_n \|_{L^p} = 1, \quad \int_{S_0} u_n = 0.$$
Combining \eqref{Korn2} and \eqref{absurd}, we deduce that $1 \ge \frac{n - C'}{C'} \| D(u_n) \|_{L^p}$ which shows that $D(u_n)$ converges to zero in $L^p$. Using again \eqref{Korn2}, we infer that $(u_n)$ is bounded in $W^{1,p}$, so that up to a subsequence it converges weakly to some $u \in W^{1,p}$, with strong convergence in $L^p$ by Rellich Theorem. We have in particular \begin{equation} \label{contradiction}
\| u \|_{L^p} = \lim_n \| u_n \|_{L^p} = 1, \quad \int_{S_0} u = \lim_n \int_{S_0} u_n = 0. \end{equation} Moreover, as $D(u_n)$ goes to zero, we get $D(u) = 0$. This implies that $u$ must be a constant (dimension is 2), which makes the two statements of \eqref{contradiction} contradictory.
\begin{Corollary} Let $H_a := \T \times (a, + \infty)$. For all $1 < p < +\infty$, there exists $C > 0$ such that: for all $a \in \R$, for all $u \in W^{1,p}(H_a)$, \begin{equation*}
\| \nabla u \|_{L^p(H_a)} \: \le \: C \| D u \|_{L^p(H_a)}. \end{equation*} \end{Corollary} {\em Proof.} From the previous inequality, we get for all $n \in \N$:
$$ \int_{S_{a+n}} | \nabla u |^p \: \le C \: \int_{S_{a+n}} | D u |^p. $$ The result follows by summing over $n$. \begin{Corollary} Let $1 < p < +\infty$. There exists $C > 0$, such that for all $u \in W^{1,p}(\Omega_{bl}^-)$, resp. $u \in W^{1,p}(\Omega_{bl})$, satisfying $u\vert_{\Gamma_{bl}} = 0$, one has
$$ \| \nabla u \|_{L^p(\Omega_{bl}^-)} \le C \| D u \|_{L^p(\Omega_{bl}^-)}, \quad \mbox{resp.} \: \| \nabla u \|_{L^p(\Omega_{bl})} \le C \| D u \|_{L^p(\Omega_{bl})}. $$ \end{Corollary} {\em Proof}. One can extend $u$ by $0$ for all $y$ with $-1 < y_2 < \gamma(y_1)$, and apply the previous inequality on $S_{-1}$, resp. $H_{-1}$.
\begin{Proposition}[Rescaled trace and Poincar\'e inequalities]\label{rescaledTracePoincare} Let $\varphi \in W^{1,p}(R^\ep)$. We have
\begin{equation}\label{IQ1}
\| \varphi \|_{L^{p}(\Sigma)} \leq C \ep^{\frac{1}{p'}} \| \nabla_x \varphi \|_{L^p(R_\ep)} ,
\end{equation}
\begin{equation}\label{IQ2}
\| \varphi \|_{L^p(R_\ep)} \leq C \ep \| \nabla_x \varphi \|_{L^p(R_\ep)} .
\end{equation} \end{Proposition} {\em Proof.} Let $\tilde\varphi(y) = \varphi(\ep y)$, where $y\in S_k = S + (k,-1)$ (a rescaled single cell of rough layer). Then $\tilde \varphi \in W^{1,p}(S_k)$ for all $k\in \N$, and $\varphi = 0 $ on $\Gamma$. By the trace theorem and the Poincar\'e inequality: for all $p \in [1,\infty )$
$$\int_{S_k \cap \{y_2 = 0 \}} | \tilde\varphi(\bar{y},0)|^p {\rm\,d}\bar{y} \leq C \int_{S_k} |\nabla_y \tilde \varphi |^p {\rm\,d}y .$$ A change of variables provides
$$\int_{\ep S_k \cap \{x_2 = 0 \}} | \varphi(\bar{x},0)|^p \ep^{-1} {\rm\,d}\bar{x}
\leq C \int_{\ep S_k} \ep^p |\nabla_x \tilde\varphi(x) |^p \ep^{-2} {\rm\,d}x .$$ Summing over $k$ we obtain
$$\left( \int_{\Sigma} | \varphi(\tilde{x},0) |^p {\rm\, d} \tilde{x} \right)^{\frac{1}{p}}
\leq C \ep^{\frac{p-1}{p}} \left( \int_{R^\ep} | \nabla_{x} \varphi (x) |^p \dx \right)^{\frac{1}{p}}
$$ and \eqref{IQ1} is proved. The inequality \eqref{IQ2} is proved in the same way, as a consequence of the (one-dimensional) Poincar\'e inequality.
$\Box $
\end{document}
|
arXiv
|
{
"id": "1511.01252.tex",
"language_detection_score": 0.6173092126846313,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{ Extreme violation of local realism in quantum hypergraph states}
\date{\today}
\author{Mariami Gachechiladze} \author{Costantino Budroni} \author{Otfried G\"uhne} \affiliation{Naturwissenschaftlich-Technische Fakult\"at, Universit\"at Siegen, Walter-Flex-Str. 3, 57068 Siegen, Germany}
\begin{abstract} Hypergraph states form a family of multiparticle quantum states that generalizes the well-known concept of Greenberger-Horne-Zeilinger states, cluster states, and more broadly graph states. We study the nonlocal properties of quantum hypergraph states. We demonstrate that the correlations in hypergraph states can be used to derive various types of nonlocality proofs, including Hardy-type arguments and Bell inequalities for genuine multiparticle nonlocality. Moreover, we show that hypergraph states allow for an exponentially increasing violation of local realism which is robust against loss of particles. Our results suggest that certain classes of hypergraph states are novel resources for quantum metrology and measurement-based quantum computation. \end{abstract}
\pacs{03.65.Ta, 03.65.Ud}
\maketitle
{\it Introduction.---} Multiparticle entanglement is central for discussions about the foundations of quantum mechanics, protocols in quantum information processing, and experiments in quantum optics. Its characterization has, however, turned out to be difficult. One problem hindering the exploration of multiparticle entanglement is the exponentially increasing dimension of the Hilbert space. This implies that making statements about general quantum states is difficult. So, one has to concentrate on families of multiparticle states with an easier-to-handle description. In fact, symmetries and other kinds of simplifications seem to be essential for a state to be a useful resource. Random states can often be shown to be highly entangled, but useless for quantum information processing \cite{entangleduseful}.
An outstanding class of useful multiparticle quantum states is given by the family of graph states \cite{hein}, which includes the Greenberger-Horne-Zeilinger (GHZ) states and the cluster states as prominent examples. Physically, these states have turned out to be relevant resources for quantum metrology, quantum error correction, or measurement-based quantum computation \cite{hein}. Mathematically, these states are elegantly given by graphs, which describe the correlations and also a possible interaction structure leading to the graph state. In addition, graph states can be defined via a so-called stabilizer formalism: A graph state is the unique eigenstate of a set of commuting observables, which are local in the sense that they are tensor products of Pauli measurements. These stabilizer observables are important for easily computing correlations leading to violations of Bell inequalities \cite{Mermin, gthb}, as well as designing simple schemes to characterize graph states experimentally \cite{gtreview}.
Recently, this family of states has been generalized to hypergraph states \cite{Kruszynska2009, Qu2013_encoding, Rossi2013, Otfried, chenlei, lyons}. These states have been recognized as special cases of the so-called locally maximally entangleable (LME) states \cite{Kruszynska2009}. Mathematically, they are described by hypergraphs, a generalization of graphs, where a single hyperedge can connect more than two vertices. They can also be described by a stabilizer formalism, but this time, the stabilizing operators are not local. So far, hypergraph states have turned out to play a role for search algorithms in quantum computing \cite{scripta}, quantum fingerprinting protocols \cite{mora}, and they have been shown to be complex enough to serve as witnesses in all QMA problems \cite{qma}. They have recently been investigated in condensed matter physics as ground states of spin models with interesting topological properties \cite{Yoshida, Akimasa}. In addition, equivalence classes and further entanglement properties of hypergraph states have been studied \cite{Otfried}.
\begin{figure}\label{fig-hgbild}
\end{figure}
In this paper we show that hypergraph states violate local realism in an extreme manner, but in a way that is robust against loss of particles. We demonstrate that this leads to applications of these states in quantum metrology and quantum computation. We see that the stabilizer formalism describing hypergraph states, despite being nonlocal, can be used to derive Hardy-type nonlocality arguments \cite{Hardy92}, Bell inequalities for genuine multiparticle entanglement \cite{Svetlichny87}, or a violation of local realism with a strength exponentially increasing with the number of particles. Our approach starts precisely with the properties of the stabilizer, in order to identify the useful correlations provided by quantum mechanics. This is in contrast to previous approaches that were either too general, e.g. Bell inequalities for general multiparticle states \cite{popescurohrlich,wwzb}, or too restricted, considering only few specific examples of hypergraph states and leading to non robust criteria \cite{Otfried}.
The violation of local realism is the key to further applications in information processing: Indeed, it is well known that violation of a Bell inequality leads to advantages, in distributed computation scenarios \cite{brunnerreview, brukner}. In addition, we will explicitly show that certain classes of hypergraph states lead to Heisenberg scaling in quantum metrology and advantages in measurement-based quantum computation.
{\it Hypergraph states.---} A hypergraph $H=(V,E)$ consists of a set of vertices $V=\{1,...,N\}$ and a set of hyperedges $E\subset 2^V$, with $2^V$ the power set of $V$. While for graphs edges connect only two vertices, hyperedges can connect more than two vertices; examples of hypergraphs are depicted in Fig.~\ref{fig-hgbild}. For any hypergraph we define the corresponding hypergraph state $\ket{H}$ as the $N$-qubit state \begin{equation} \ket{H}=\prod_{e\in E} C_e\ket{+}^{\otimes N}, \label{eq-hg-creation} \end{equation} where $\ket{+}=(\ket{0}+\ket{1})/\sqrt{2}$, $e$ is a hyperedge and $C_e$ is a multi-qubit phase gate acting on the Hilbert space associated with the vertices $v\in e$, given by the matrix $C_e =\mathbbm {1} - 2\ket{1\dots 1}\bra{1\dots 1}$. The first nontrivial hypergraph state consists of $N=3$ qubits connected by a single hyperedge [see Fig.~\ref{fig-hgbild}(a)]. Hypergraph states have been recognized as special cases of LME states, generated via a fixed interaction phase of $\phi=\pi$ \cite{Kruszynska2009}.
Alternatively, we can define the hypergraph states using a stabilizer formalism \cite{Otfried}. For each qubit $i$ we define the operator \begin{equation} g_i=X_i\bigotimes_{e\in E} C_{e\backslash \{i\}}. \label{eq-hg-stabilizer} \end{equation} Here and in what follows, we denote by $X_i$ and $Z_i$ the Pauli matrices, acting on $i^{th}$ qubit. The hypergraph state can be defined as the unique eigenstate for all of them, $g_i \ket{H}=\ket{H}$ with the eigenvalue $+1$. Consequently, the hypergraph state is an eigenstate of the entire stabilizer, i.e., the commutative group formed by all the products of the $g_i$. It should be noted that the $g_i$ are, in general, non-local operators, as they are not tensor-products of operators acting on single parties. We say that a hyperedge has \textit{cardinality} $k$, if it circumscribes $k$ vertices and a hypergraph is \textit{$k$-uniform}, if all edges are $k$-edges. Finally, note that different hypergraphs may lead to equivalent hypergraph states, in the sense that the two states can be converted into one other by a local basis change. For small numbers of qubits, the resulting equivalence classes have been identified \cite{Otfried}.
{\it Local correlations from the nonlocal stabilizer.---} The key observation for the construction of our nonlocality arguments is that the stabilizer of hypergraph states, despite being nonlocal, predicts perfect correlations for some local measurements. In the following, we explain this for the three-qubit hypergraph state $\ket{H_3}$, but the method is general. The stabilizing operators for the three-qubit hypergraph state are \begin{equation}\label{1} g_1=X_1\otimes C_{23}, \quad g_2=X_2\otimes C_{13}, \quad g_3=X_3\otimes C_{12}. \end{equation} We can expand the controlled phase gate $C_{ij}$ on two qubits, leading to \begin{equation} \label{phasegate} g_1=X_1\otimes (\ketbra{00}+\ketbra{01}+\ketbra{10}-\ketbra{11}) \end{equation} and similar expressions for the other $g_i$. Since $g_1 \ket{H_3}=+\ket{H_3}$, the outcomes for $X$ measurements on the first qubit and $Z$ measurements on the second and third qubits are correlated: if one measures $"+"$ on the first qubit, then the other two parties cannot both measure $"-"$ in $Z$ direction, as this would produce $-1$ as the overall eigenvalue. So, we extract the first correlation from the stabilizer formalism: \begin{equation} \label{A1}
P(+--|XZZ)=0. \end{equation} The l.h.s of Eq.~(\ref{A1}) denotes the probability of measuring $+--$ in $XZZ$ on the qubit $1,2$, and $3$, respectively. Similarly, it follows that if one measures $"-"$ in $X$ direction on the first qubit, then the other parties, both have to measure $"-"$ in $Z$ direction. So we have: \begin{equation} \label{A}
P(-++|XZZ)+ P(-+-|XZZ)+P(--+|XZZ)=0, \end{equation} which implies, of course, that each of the probabilities is zero. Since the three-qubit hypergraph state is symmetric, the same correlations for measuring $X$ on other qubits can be obtained by considering $g_2$ and $g_3$, leading to permutations of correlations in Eq.~(\ref{A1} ,\ref{A}).
{\it The three-qubit hypergraph state $\ket{H_3}$.---} We start with the discussion of fully local hidden variable (HV) models. Such models assign for any value of the HV $\lambda$ results to all measurements of the parties in a local manner, meaning that the probabilities for a given HV factorize. If we denote by $r_i$ the result and by $s_i$ the measurement setting on the $i^{th}$ particle, respectively, then the probabilities coming from local models are of the form \begin{align}
P & (r_1,r_2,r_3 |s_1,s_2,s_3) = \\ = & \int d\lambda p(\lambda)
\chi^A(r_1|s_1,\lambda)
\chi^B(r_2|s_2,\lambda)
\chi^C(r_3|s_3,\lambda) \nonumber. \end{align} For probabilities of this form, it is well known that it suffices to consider models which are, for a given $\lambda$, deterministic. This means that $\chi^i$ takes only the values $0$ or $1$, and there is only a finite set of $\chi^i$ to consider.
\noindent {\bf Observation 1.} {\it If a fully local hidden variable model satisfies the conditions from Eq.~(\ref{A1}, \ref{A}) and their symmetric correlations coming from the permutations, then it must fulfill} \begin{equation}
P(+--|XXX)+P(-+-|XXX)+P(--+|XXX) = 0. \end{equation} The proof of this statement is done by exhausting all possible local deterministic assignments.
In contrast, for $\ket{H_3}$ we have \begin{align}
P(+--|XXX)=\frac{1}{16} \label{eq-h3corr} \end{align} and the same holds for the permutations of the qubits. The above is a so-called Hardy argument \cite{Hardy92}, namely, a set of joint probabilities equal to $0$ or, equivalently, logical implications that together implies that some other probability is equal to zero.
Our method shows how the correlations of the nonlocal stabilizer can be used for Hardy-type arguments. We recall that Hardy-type arguments have been obtained for all permutation-symmetric states \cite{Wang, Abramsky}. However, they involved different settings and have no direct connection with the stabilizer formalism, making a generalization complicated. In contrast, we will see that our measurements can even be used to prove genuine multiparticle nonlocality of the hypergraph state. First, we translate the Hardy-type argument into a Bell inequality:
\noindent {\bf Remark 2.} {\it Putting together all the null terms derived from the stabilizer formalism and subtracting the terms causing a Hardy-type argument, we obtain the Bell inequality \begin{align} \label{3correlations}
\langle \ensuremath{\mathcal{B}}_3^{(1)} \rangle&= \big[ P(+--|XZZ)+ P(-++|XZZ) \nonumber \\
+&P(-+-|XZZ)+P(--+|XZZ)+ \mbox{ permutat.} \big] \nonumber \\
-&[ P(+ --|XXX)+ \mbox{ permutations }\big] \geq 0, \end{align} where the permutations include all distinct terms that are obtained by permuting the qubits. The three-uniform hypergraph state violates the inequality (\ref{3correlations}) with the value of $\langle \ensuremath{\mathcal{B}}_3^{(1)} \rangle=-3 / 16$.}
This Bell inequality follows from the Hardy argument: If a deterministic local model predicts one of the results with the minus signs, it also has to predict at least one of the results corresponding to the terms with a plus sign, otherwise it contradicts with the Hardy argument. In addition, all the terms with a minus sign are exclusive, so a deterministic LHV model can predict only one of them.
The Hardy-type argument and the Bell inequality can be generalized to a higher number of qubits, if we consider $N$-qubit hypergraphs with the single hyperedge having a cardinality $N$:
\noindent {\bf Observation 3.} {\it Consider the $N$-qubit hypergraph state with a single hyperedge of cardinality $N$. Then, all the correlations coming from the stabilizer [as generalizations of Eqs.~(\ref{A1},\ref{A})] imply that for any possible set of results $\{r_i\}$ where one $r_{i_1}=+1$ and two $r_{i_2}= r_{i_3} = -1$ one has \begin{equation}
P(r_1, r_2, ..., r_N|X_1 X_2 ... X_N) = 0. \end{equation} For the hypergraph state, however, this probability equals $1/2^{(2N-2)}.$ This Hardy-type argument leads to a Bell inequality as in Eq.~(\ref{3correlations}) which is violated by the state with a value of $-(2^N-N-2)/2^{(2N-2)}.$}
Clearly, the violation of the Bell inequality is not strong, as it does not increase with the number of particles. Nevertheless, Observation~3 shows that the nonlocal stabilizer formalism allows one to easily obtain nonlocality proofs. In fact, one can directly derive similar arguments for other hypergraph states (e.g. states with one hyperedge of cardinality $N$ and one further arbitrary hyperedge), these results will be presented elsewhere. Note that these states are not symmetric, so the results of Refs.~\cite{Wang, Abramsky} do not apply.
So far, we only considered fully local models, where for a given HV all the probabilities factorise. Now we go beyond this restricted type of models to the so-called hybrid models \cite{Svetlichny87}.
We consider a bipartition of the three particles, say $A|BC$, and consider a model of the type $
P (r_1,r_2,r_3 |s_1,s_2,s_3) = \int d\lambda p(\lambda)
\chi^A(r_1|s_1,\lambda)
\chi^{BC}(r_2, r_3|s_2, s_3,\lambda). $ Here, Alice is separated from the rest, but $\chi^{BC}$ may contain correlations, e.g., coming from an entangled state between $B$ and $C$. In order to be physically reasonable, however, we still request $\chi^{BC}$ not to allow instantaneous signaling.
This kind of models, even if different bipartitions are mixed, cannot explain the correlations of the hypergraph state, meaning that the hypergraph state is genuine multiparticle nonlocal. First, one can see by direct inspection that the stabilizer conditions from Eqs.~(\ref{A1}, \ref{A})
are not compatible with the hypergraph correlations $P(---|XXX) = 1/16$
and $P(---|ZZZ) = 1/8$. Contrary to the correlations in Eq.~(\ref{eq-h3corr}) these are symmetric, and allow the construction of a Bell-Svetlichny inequality \cite{Svetlichny87} valid for all the different bipartitions:
\noindent {\bf Observation 4.} {\it Putting all the terms from the hypergraph stabilizer formalism
and the correlations $P(---|XXX)$ and $P(---|ZZZ)$ together, we obtain the following Bell-Svetlichny inequality for genuine multiparticle nonlocality, \begin{align} \label{3gen} \langle\ensuremath{\mathcal{B}}_3^{(2)}\rangle&=
\big[P(+--|XZZ)+ P(-++|XZZ) \nonumber \\
+&P(-+-|XZZ)+P(--+|XZZ)+ \mbox{ permutat.} \big] \nonumber \\
+&P(- --|XXX) - P(- - -|ZZZ) \geq 0, \end{align} which is violated by the state $\ket{H_3}$ with $\langle\ensuremath{\mathcal{B}}_3^{(2)}\rangle=-1 \backslash 16$.}
The proof is done by an exhaustive assignments of nonsignaling and local models.
To investigate the noise tolerance of Ineq.~\eqref{3gen}, we consider states of the type $ \varrho= (1-\varepsilon)\ket{H}\bra{H}+\varepsilon \mathbbm{1}/8 $ and ask how much noise can be added, while the inequality is still violated. The white noise tolerance of the Ineq.~(\ref{3gen}) is $\varepsilon = {1}/{13} \approx 7.69\%$ and is optimal in the sense that for larger values of $\varepsilon$ a hybrid model can be found which explains all possible measurements of $X$ and $Z$ (within numerical precision). The existence of such a model can be shown by linear programming (see Appendix A \cite{appremark}). With the same method we can also prove that the state becomes fully local with respect to $X$ and $Z$ measurements for $\varepsilon \geq 2/3 \approx 66.6\% $.
{\it Three uniform hypergraph states.---} Let us extend our analysis to hypergraph states with a larger number of particles. Here, it is interesting to ask whether the violation of Bell inequalities increases exponentially with the number of parties. Such a behaviour has previously been observed only for GHZ states \cite{Mermin} and some cluster states \cite{gthb}.
GHZ states are described by fully connected graphs (see Fig.~\ref{fig-hgbild}), i.e., fully connected two-uniform hypergraph states. It is thus natural to start with fully connected three-uniform hypergraph states. First, we observe that for such states on $N$ qubits and for even $m$ with $1<m<N$ \begin{equation}\label{lemma3} \langle \underset{m}{\underbrace{X\dots X}}Z\dots Z\rangle =\begin{cases} \begin{array}[t]{cc} +\frac{1}{2} & \mbox{\ensuremath{\mbox{if \ensuremath{m=2} mod \ensuremath{4}}}},\\ -\frac{1}{2} & \mbox{\ensuremath{\mbox{if \ensuremath{m=0} mod \ensuremath{4}}}}. \end{array}\end{cases} \end{equation} Moreover, if $m=N$, then the correlations are given by \begin{equation}\label{lemma33} \langle \underset{N}{\underbrace{XX\dots XX}}\rangle =\begin{cases} \begin{array}[t]{cc} 0 & \mbox{\ensuremath{\mbox{if \ensuremath{N=0} mod \ensuremath{4}}}},\\ 1 & \mbox{\ensuremath{\mbox{if \ensuremath{N=2} mod \ensuremath{4}}}}. \end{array}\end{cases} \end{equation} Finally, we always have $\langle{{ZZ\dots ZZ}}\rangle =0$ (see Appendix B for details \cite{appremark}).
We then consider the following Bell operator \begin{align} \label{bell2} \ensuremath{\mathcal{B}}_N & = - \big[AAA\dots AA\big] + \big[BBA\dots A + \;\mbox{permutat.}\big] - \nonumber\\
- & \big[BBBBA\dots A + \;\mbox{permutat.}\big] +
\big[\dots\big] - \dots \end{align} Note that this Bell operator is similar to $\langle \ensuremath{\mathcal{B}}_N^M \rangle $ of the original Mermin inequality \cite{Mermin}, but it differs in the number of $B$ (always even) considered. Using the correlations computed above, we can state:
\noindent {\bf Observation 5.} {\it If we fix in the Bell operator $\ensuremath{\mathcal{B}}_N$ in Eq.~(\ref{bell2}) the measurements to be $A=Z$ and $B=X$, then the $N$-qubit fully-connected three-uniform hypergraph state violates the classical bound, by an amount that grows exponentially with number of qubits, namely \begin{align} \left\langle \ensuremath{\mathcal{B}}_N \right\rangle_C &\leq 2^{\left\lfloor N/2\right\rfloor} \quad \mbox{ for local HV models, and} \nonumber \\ \left\langle \ensuremath{\mathcal{B}}_N \right\rangle_Q &\geq 2^{N-2}-\frac{1}{2} \quad \mbox{ for the hypergraph state.} \end{align} } The proof is given in Appendix C \cite{appremark}.
{\it Four-uniform hypergraph states.---} Finally, let us consider four-uniform complete hypergraph states. For them, the correlations of measurements as in Eq.~(\ref{lemma3}) are not so simple: They are not constant, and depend on $m$ as well as on $N$. Nevertheless, they can be explicitly computed, and detailed formulas are given in the Appendix D \cite{appremark}. {From} these correlations, we can state:
\noindent {\bf Observation 6.} {\it The $N$-qubit fully-connected four-uniform hypergraph state violates
local realism by an amount that grows exponentially with number of qubits. More precisely, one can find a Mermin-like Bell operator $\ensuremath{\mathcal{B}}_N$ such that \begin{align} \frac{\left\langle \ensuremath{\mathcal{B}}_N \right\rangle_Q} {\left\langle \ensuremath{\mathcal{B}}_N \right\rangle_C} \stackrel{N\rightarrow \infty}{\sim} \frac{\Big(1+\frac{1}{\sqrt{2}}\Big)^{N-1} }{\sqrt{2}^{N+3}} \approx \frac{1.20711^{N}}{2\sqrt{2}+2}. \end{align}} A detailed discussion is provided in Appendix E \cite{appremark}.
{\it Robustness.---} So far, we have shown that three- and four-uniform hypergraph states
violate local realism comparable to GHZ states. A striking difference is, however, that the entanglement and Bell inequality violation of hypergraph states is robust under particle loss. This is in stark contrast to GHZ states, which become fully separable if a particle is lost. We can state:
\noindent {\bf Observation 7.} {\it The $N$-qubit fully-connected four-uniform hypergraph state preserves the violation of the local realism even after loss of one particle. More precisely, for $N=8k+4$ we have ${\left\langle \ensuremath{\mathcal{B}}_{N-1}\right\rangle_Q}/ {\left\langle \ensuremath{\mathcal{B}}_N \right\rangle_Q} \stackrel{N\rightarrow \infty}{\sim} {1}/({\sqrt{2}+1}).$ This means that the reduced state shows the same exponential scaling of the Bell inequality violation as the original state. }
For the detailed discussions see Appendix F \cite{appremark}.
For three-uniform complete hypergraph states we can prove that the reduced states are highly entangled, as they violate inequalities testing for separability \cite{Roy} exponentially. This violation decreases with the number of traced out qubits, but persists even if several qubits are lost. This suggests that this class of hypergraph states is also more robust than GHZ states, details can be found in Appendix G \cite{appremark}. Despite the structural differences, this property resembles of the W state, which is itself less entangled but more robust than the GHZ state \citep{Buzek}. In addition, this may allow the lower detection efficiency in the experiments.
{\it Discussion and conclusion.---} A first application of our results is quantum metrology. In the standard scheme of quantum metrology one measures an observable $M_\theta$ and tries to determine the parameter $\theta$ which describes a phase in some basis \cite{giovanetti, weibo}. If one takes product states, one obtains a signal $\mean{M_\theta} \sim \cos{(\theta)}$ on a single particle, repeating it on $N$ particles allows to determine $\theta$ with an accuracy $\delta \theta \sim 1/\sqrt{N}$, the so-called standard quantum limit. Using an $N$-qubit GHZ state, however, one observes $\mean{(M_\theta)^{\otimes N}} \sim \cos{(N\theta)}$ and this phase super-resolution allows to reach the Heisenberg limit $\delta \theta \sim 1/{N}$. For a general state $\ensuremath{\varrho}$, it has been shown that the visibility of the phase super-resolution is given by the expectation value of the Mermin-type inequality, $V=Tr(\ensuremath{\mathcal{B}}_N \ensuremath{\varrho})/2^{N-1}$ \cite{weibo}. So, since the three-uniform hypergraph states violate this inequality with a value $\mean{\ensuremath{\mathcal{B}}_N}_Q \sim 2^{(N-2)}$ the visibility is $V \sim 1/2$, independently of the number of particles. This means that these states can be used for Heisenberg-limited metrology, and from our results they can be expected to have the advantage of being more robust to noise and particle losses.
The second application of exponential violation of Bell inequalities is \textit{nonadaptive measurement based quantum computation with linear side-processing} ($NMQC_{\oplus}$) \cite{Hoban11}. $NMQC_{\oplus}$ is a non-universal model of quantum computation where linear classical side-processing is combined with quantum measurements in a nonadaptive way, i.e., the choice of settings is independent of previous outcomes. In Ref.~\cite{Hoban11} the authors connect the expectation value of a full-correlation Bell expression \cite{wwzb} with the success probability of computing a Boolean function, specified as a function of the inequality coefficients, via $NMQC_{\oplus}$. In particular, the exponential violation of generalized Svetlichny inequalities \cite{Collins02} (equal to Mermin inequalities for even $N$), corresponds to a constant success probability $P_{\rm succ}$ of computing the pairwise $\mathrm{AND}$ on $N$ bits extracted from a uniform distribution, whereas in the classical case $P_{\rm succ}-1/2$ decrease exponentially with $N$. As a consequence, the exponential violation of the full-correlation Bell expression $\mathcal{B}_N$ can be directly related to an exponential advantage for computation tasks in the $NMQC_{\oplus}$ framework. Moreover, in several cases, e.g., $4$-uniform hypergraph states of $N = 6\ {\rm mod}\ 8$ qubits, also the Svetlichny inequality is violated exponentially, providing an advantage for computation of the pairwise $\mathrm{AND}$ discussed in Ref.~\cite{Hoban11}.
In summary, we have shown that hypergraph states violate local realism in many ways. This suggests that they are interesting resources for quantum information processing, moreover, this makes the observation of hypergraph states a promising task for experimentalists. In our work, we focused only on some classes of hypergraph states, but for future research, it would be desirable to identify classes of hypergraph states which allow for an all-versus-nothing violation of local realism or which are strongly genuine multiparticle nonlocal.
We thank D.~Nagaj, F.~Steinhoff, M.~Wie\'sniak, and B.~Yoshida for discussions. This work has been supported by the EU (Marie Curie CIG 293993/ENFOQI), the FQXi Fund (Silicon Valley Community Foundation), the DFG and the ERC.
\onecolumngrid
\section{Appendix A. Genuine multiparticle nonlocality with linear programming}\label{App:LP}
Fixed the number of measurement settings and outcomes, probabilities arising from a hybrid local-nonsignalling model, as the one described in the main text for the splitting $A|BC$, form a polytope whose extremal points are given by combination of deterministic local assignments for the party $A$ and extremal nonsignalling assignments, i.e., local deterministic and PR-boxes \cite{brunnerreview}, for the parties $BC$. In order to detect genuine multiparticle nonlocality, one has to consider all combinations of probabilities arising from the other local-nonsignalling splitting, namely $C|AB$ and $B|AC$. Geometrically, this corresponds to take the convex hull of the three polytopes associated with the three different splitting. Let us denote such polytopes as $\mathcal{P}_{A|BC}, \mathcal{P}_{C|AB} ,\mathcal{P}_{B|AC}$ and their convex hull as $\mathcal{P}_{L-NS}$. By definition of convex hull, every vector $\mathbf{p}\in \mathcal{P}_{L-NS}$ can be written as a convex combination of three vectors $\mathbf{p}_{
A|BC},
\mathbf{p}_{C|AB}$ and $\mathbf{p}_{B|AC}$, which, in turn, can be written as a convex combination of the vertices of the corresponding polytope.
To check whether a given point $\mathbf{p}$ belongs to $\mathcal{P}_{L-NS}$ it is, therefore, sufficient the description in terms of the extremal points of $\mathcal{P}_{A|BC}, \mathcal{P}_{C|AB} ,\mathcal{P}_{B|AC}$. Let us denote them as $\{\mathbf{v}_i\}$. The membership problem can then be formulated as a linear program (LP) \cite{brunnerreview}
\begin{equation}\begin{split}\label{e:lp} \text{maximize: } &\mathbf{\lambda} \cdot \mathbf{p} - C \\ \text{subject to: }&
\mathbf{\lambda} \cdot \mathbf{v}_i - C \leq 0\text{ , for all }\mathbf{v}_i \\
& \mathbf{\lambda} \cdot \mathbf{p} - C \leq 1. \end{split}\end{equation}
The variable of the LP are $\{\mathbf{\lambda},C\}$, where $\mathbf{\lambda}$ represents the coefficient of a Bell-Svetlichny inequality, detecting genuine multiparticle nonlocality, and $C$ the corresponding local-nonsignalling bound. The LP optimizes the coefficients $\mathbf{\lambda}$ to obtain the maximal value (at most $C+1$) for the quantum probabilities, while keeping the local-nonsignaling bound $C$. As a consequence, the vector $\mathbf{p}$ can be written as a convex combination of $\{\mathbf{v}_i\}$ if and only if the optimal value of the LP is 0.
The noise tolerance for $\ket{H_3}$ can then be computed by mixing it with white noise, i.e., $\ket{H_3}\bra{H_3}\mapsto \varrho= (1-\varepsilon)\ket{H}\bra{H}+\varepsilon \mathbbm{1}/8$, and compute for which values of $\varepsilon$ the LP \eqref{e:lp} gives optimal value $0$. Standard numerical techniques for LP give that up to $\varepsilon\approx {1}/{13} \approx 7.69\%$ the probabilities for $X$ and $Z$ measurements cannot be explained by a hybrid local-nonsignalling model.
\section{APPENDIX B: Correlations for three-uniform hypergraph states}
\subsection{B.1. Preliminary calculations}
Before starting with the actual calculations, we need to settle couple of identities and a look-up table, which we will refer to throughout the main proofs.
The first and probably the most important identity is a commutation relation between multi-qubit phase gates and Pauli X matrices \cite{Otfried}, \begin{equation}{\label{identity}} C_e\big(\bigotimes_{i \in K} X_k \big)=(\bigotimes_{i \in K} X_k \big)\big(\prod_{f\in \mathcal{P}(K)}C_{e\backslash\{f\}}\big). \end{equation} Here, $\mathcal{P}(K)$ denotes the power set of the index set $K$. Note that the product of the $C_{e\backslash\{f\}}$ may include the term $C_{\emptyset}$, which is defined to be $-\openone$ and leads to a global sign.
Furthermore, it turns out to be useful to recall some basic facts about binomial coefficients, as the appear frequently in the following calculations.
\begin{lemma} The following equalities hold: \begin{equation}\label{identity1} Re\Big[(1+i)^n\Big]=\sum_{k=0,4,\dots}^n\binom{n}{k}-\binom{n}{k+2}, \end{equation} \begin{equation}\label{identity2} Im\Big[(1+i)^n\Big]=\sum_{k=0,4,\dots}^n\binom{n}{k+1}-\binom{n}{k+3}. \end{equation} \end{lemma} \begin{proof} Here we derive (\ref{identity1}) and (\ref{identity2}) together: \begin{equation} s:=(1+i)^n=\sum^n_{k=0} \binom{n}{k}i^k =\sum^n_{k=0,4,\dots} \binom{n}{k}+ i\binom{n}{k+1} -\binom{n}{k+2}-i\binom{n}{k+3}. \end{equation} It is easy to spot that $Re[s]$ and $Im[s]$ indeed leads to the identities (\ref{identity1}) and (\ref{identity2}) respectively. \end{proof}
The following look-up table represents the values of (\ref{identity1}) and (\ref{identity2}) for different $n$. These values can be derived from the basic properties of complex numbers:
\begin{center}
\begin{tabular}{|c|c|c|c|c|c|} \hline \# & $n$ & $Re\Big[(1+i)^{n}\Big]$ & $Im\Big[(1+i)^{n}\Big]$ & $Re\Big[(1+i)^{n}\Big]$$+Im\Big[(1+i)^{n}\Big]$ & $Re\Big[(1+i)^{n}\Big]$$-Im\Big[(1+i)^{n}\Big]$\tabularnewline \hline \hline 1. & $n=0$ mod 8 & $+2^{\frac{n}{2}}$ & $0$ & $+2^{\frac{n}{2}}$ & $+2^{\frac{n}{2}}$\tabularnewline \hline 2. & $n=1$ mod 8 & $+2^{\frac{n-1}{2}}$ & $+2^{\frac{n-1}{2}}$ & $+2^{\frac{n+1}{2}}$ & $0$\tabularnewline \hline 3. & $n=2$ mod 8 & $0$ & $+2^{\frac{n}{2}}$ & $+2^{\frac{n}{2}}$ & $-2^{\frac{n}{2}}$\tabularnewline \hline 4. & $n=3$ mod 8 & $-2^{\frac{n-1}{2}}$ & $+2^{\frac{n-1}{2}}$ & $0$ & $-2^{\frac{n+1}{2}}$\tabularnewline \hline 5. & $n=4$ mod 8 & $-2^{\frac{n}{2}}$ & $0$ & $-2^{\frac{n}{2}}$ & $-2^{\frac{n}{2}}$\tabularnewline \hline 6. & $n=5$ mod 8 & $-2^{\frac{n-1}{2}}$ & $-2^{\frac{n-1}{2}}$ & $-2^{\frac{n+1}{2}}$ & $0$\tabularnewline \hline 7. & $n=6$ mod 8 & $0$ & $-2^{\frac{n}{2}}$ & $-2^{\frac{n}{2}}$ & $+2^{\frac{n}{2}}$\tabularnewline \hline 8. & $n=7$ mod 8 & $+2^{\frac{n-1}{2}}$ & $-2^{\frac{n-1}{2}}$ & $0$ & $+2^{\frac{n+1}{2}}$\tabularnewline \hline \end{tabular}\\ $\quad$\\
\textbf{Table 0.} Look-up table for the values of (\ref{identity1}) and (\ref{identity2}) . \end{center}
\subsection{B.2. Correlations for X and Z measurements on fully-connected three-uniform hypergraph states}
\begin{lemma} \label{lemma3full} Consider an arbitrary number of qubits $N$ and the three-uniform fully-connected hypergraph (HG) states. Then, if $m$ is even with $1<m<N$ the following equality holds \begin{equation} \langle \underset{m}{\underbrace{X\dots X}}Z\dots Z\rangle =\begin{cases} \begin{array}[t]{cc} +\frac{1}{2} & \mbox{\ensuremath{\mbox{if \ensuremath{m=2} mod \ensuremath{4}}}},\\ -\frac{1}{2} & \mbox{\ensuremath{\mbox{if \ensuremath{m=0} mod \ensuremath{4}}}}. \end{array}\end{cases} \end{equation} \end{lemma} \begin{proof} We can write:
\begin{equation}\label{appendixAeq1} K:=\bra{HG} X\dots XZ\dots Z\ket{HG}=\bra{+}^{\otimes N}\bigg(\prod_{e\in E} C_e\bigg) X\dots XZ\dots Z\bigg( \prod_{e\in E} C_e\bigg) \ket{+}^{\otimes N}. \end{equation} We can group all the controlled phase gates on the right hand side of the expression (\ref{appendixAeq1}). Note that the operators $C_e$ and $X_i$ do not commute, but we can use the identity ($\ref{identity}$). While regrouping we count the multiplicity of each phase gate. If each phase gate appears even times, we get an identity as $C^2=\mathbbm{1}$, if not, we keep these phase gates with the multiplicity one for the further calculations.
For the purposes which will become apparent shortly, we denote the parties which measure in $X$ direction by $\circledast$ and ones in $Z$ direction by $\bigtriangleup$, in a way that, for example, if an arbitrary phase gate acts on $XXXZZ$, it is represented as $\circledast\circledast\circledast\bigtriangleup\bigtriangleup$.
Without loss of generality, we fix one phase gate $C_e$ and consider all the possible scenarios of $\circledast$ and $\bigtriangleup$ it can be acting on. Since we work on three-uniform HG states, every phase gate acts on bashes of different three party systems. These parties can be either of type $\circledast$ or $\bigtriangleup$ and we have to consider all possible scenarios. Since we are working with symmetric states, we can sort the parties such that we have $m$ $\circledast$'s followed by $(N-m)$ $\bigtriangleup$'s:
\begin{equation} \label{appendixAeq2}
\underset{m}{\underbrace{\circledast\dots \circledast}} \underset{N-m}{\underbrace{\bigtriangleup\dots\bigtriangleup}} \end{equation}
and then we can enumerate all the scenarios of one phase gate acting on (\ref{appendixAeq2}): \\
\begin{enumerate}
\item $C_eZZZ $ corresponds to $\bigtriangleup\bigtriangleup\bigtriangleup \quad\quad\quad\quad$ $3. \; C_eXXZ$ corresponds to $\circledast\circledast\bigtriangleup$\\
\item $C_eXZZ$ corresponds to $\circledast\bigtriangleup\bigtriangleup\quad\quad\quad\quad$ $4.\; C_eXXX$ corresponds to $\circledast\circledast\circledast$\\ \end{enumerate}
We consider each case separately:\\
1. $C_eZZZ=ZZZC_e$ as $C_e$ and $Z$ commute. $C_e$ moves on the right side with the multiplicity one. To save us writing in the future, we will denote the multiplicity of the phase gate moving on the right side by $\# e$. In this particular case it is $ \# \bigtriangleup\bigtriangleup\bigtriangleup =1$. However, on the right side of the equation (\ref{appendixAeq1}) we have a product of all three-party phase gates. Therefore, we get $C_e$ with the multiplicity of two and $C_e^2=\mathbbm{1}$. Note that, as we have chosen an arbitrary three-qubit phase gate, the same result holds for every such phase gate. So, all three-qubit phase gates coming from the case 1, cancel out. We will see that with the same reasoning all three qubit phase gates cancel out (give an identity). \\
2. For $C_eXZZ$, we use the identity ($\ref{identity}$): \begin{equation} \label{appendixAeq3} C_eXZZ=XC_eC_{\{e\backslash X\}}ZZ. \end{equation} The three-qubit phase gate $C_e$, from (\ref{appendixAeq3}), appears with the multiplicity one $(\# \circledast\bigtriangleup\bigtriangleup=1)$ and like in the case 1, it gives an identity when being multiplied by the same three-qubit phase gate at the right side of the expression in (\ref{appendixAeq1}). It is more tricky to calculate the multiplicity of $C_{\{e\backslash X\}}$ ($\# \bigtriangleup\bigtriangleup$ as the $\circledast$ part (or equivalently, $X$ part) is removed from the set of vertices $e$.). For this we need to fix $\bigtriangleup\bigtriangleup$ and count all the scenarios when an arbitrary $C_{e}$ is reduced to $\bigtriangleup\bigtriangleup$. As we are working with the symmetric case, such scenario repeats $\binom{m}{1}=m$ times, where $m$ is the number of parties measuring in $X$ direction.We shortly denote this as $\# \bigtriangleup\bigtriangleup =\binom{m}{1}=m$. So, as $m$ is an even number, $(C_{\{e\backslash X\}})^m=(C_{\bigtriangleup\bigtriangleup})^m=\mathbbm{1}$.\\ Note that the gate $C_{\bigtriangleup\bigtriangleup}$ can only be generated from in case 2.\\
3. For $C_eXXZ$, we use the identity ($\ref{identity}$): \begin{equation} \label{appendixAeq4} C_eXXZ=XXC_eC_{\{e\backslash XX\}} \Big[\prod_{\forall X} C_{\{e\backslash X\}}\Big] Z. \end{equation} The three-qubit phase gate $C_e$, from (\ref{appendixAeq3}), appears with the multiplicity one $(\# \circledast\circledast\bigtriangleup=1)$; therefore, like in the two previous cases, it cancels out on the right side of the expression. A multiplicity of $C_{\{e\backslash XX\}}$ is calculated by fixing a concrete $\bigtriangleup$ and counting all possible appearance of arbitrary $\circledast\circledast$. As the number of parties measuring in direction is $X$ is $m$, this means that it is all combination of two parties with $X$ measurements out of total $m$ parties. So, \begin{equation}
\# \bigtriangleup=\binom{m}{2}=\frac{m(m-1)}{2}=\begin{cases} \begin{array}[t]{ccc} \mbox{even, if} & m=0 \; mod\; 4 & \Rightarrow C_{\triangle}\mbox{ cancels out.}\\ \mbox{odd, if } & m=2 \; mod \; 4 & \Rightarrow C_{\triangle}\mbox{ remains.} \end{array}\end{cases} \end{equation} The last one from this case is the multiplicity of $C_{\{e\backslash X\}}$ or $\# \circledast\bigtriangleup$. Here we fix one qubit from $m$ ($X$ direction) and one from $N-m$ ($Z$ direction) and count the number of such occurrences, when the third qubit is an arbitrary one of the type $\circledast$, which is exactly $\binom{m-1}{1}$. Therefore, \begin{equation} \#\circledast\bigtriangleup=\binom{m-1}{1}=m-1,\mbox{which is odd},\Rightarrow\; C_{\circledast\bigtriangleup}\mbox{remains}. \end{equation}\\
4. For $C_eXXX$, we use the identity ($\ref{identity}$): \begin{equation} \label{appendixAeq5} C_eXXX=XXXC_e \Big[\prod_{\forall X} C_{\{e\backslash X\}}\Big] \Big[\prod_{\forall X} C_{\{e\backslash XX\}}\Big]C_{\{\}}. \end{equation} $C_e$ occurs once and it gives an identity with the other one from the right side like in the previous cases. The multiplicity of $C_{\{e\backslash X\}}$ is $\# \circledast\circledast$. Here we fix two parties in $X$ direction and count the occurrence of this scenario by altering the third party, from the remaining $m-2$, in $X$ direction. Therefore, \begin{equation} \#\circledast\circledast=\binom{m-2}{1}=m-2,\mbox{which is even},\Rightarrow\; C_{\circledast\circledast}\mbox{ cancels out}. \end{equation} Similarly for $C_{\{e\backslash XX\}}$, we fix one party in $X$ direction and count all possibilities of choosing two parties out of remaining $m-1$. Therefore, \begin{equation}
\# \circledast=\binom{m-1}{2}=\frac{(m-1)(m-2)}{2}=\begin{cases} \begin{array}[t]{ccc} \mbox{odd, if} & m=0 \; mod\; 4 & \Rightarrow\; C_\circledast \mbox{ remains.}\\ \mbox{even if, } & m=2\; mod\; 4 & \Rightarrow\; C_\circledast \mbox{ cancels out.} \end{array}\end{cases} \end{equation} At last, we consider $C_{\{\}}$. This gate determines the global sign of the expectation value and it appears only when $C_e$ acts on systems which are all measured in $X$ direction. Therefore, \begin{equation}{\label{3unifGL}} \#\{\}=\binom{m}{3}=\frac{m(m-1)(m-2)}{2 \cdot 3},\mbox{which is even}\Rightarrow\; \mbox{ the global sign is positive }. \end{equation}\\
To go on, we need to consider two cases \#1: $m=0$ mod 4 and \#2: $m=2$ mod 4 and calculate the expectation value separately for both: \\
\textbf{Case \#1:} When $m=0 \; mod\; 4$, we write out all remaining phase gates and continue the derivation from equation (\ref{appendixAeq1}): \begin{equation} \langle K \rangle :=\bra{+}^{\otimes N} X\dots X Z\dots Z \prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} C_{\circledast}\ket{+}^{\otimes N}. \end{equation}\\
Using the fact that $X$ is an eigenstate of $\bra{+}$, we can get rid of all $X$s and then we can write $C_{\bigtriangleup}$ instead of $Z$:
\begin{align}\label{appendixAeq6}
\begin{split} \langle K \rangle= & \bra{+}^{\otimes N}\prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} C_{\circledast}C_{\bigtriangleup}\ket{+}^{\otimes N}=\frac{1}{\sqrt{2}^N}\sum_{i=00\dots 00}^{11\dots 11}\bra{i}\prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} C_{\circledast}C_{\bigtriangleup} \frac{1}{\sqrt{2}^N}\sum_{j=00\dots 00}^{11\dots 11}\ket{j}\\ = & \frac{1}{2^N}\bigg[\bra{00\dots 00}\prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} C_{\circledast}C_{\bigtriangleup}\ket{00\dots 00}+\dots +\bra{11\dots 11}\prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} C_{\circledast}C_{\bigtriangleup}\ket{11\dots 11}\bigg]\\ = &\frac{1}{2^N}Tr\bigg[\prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} C_{\circledast}C_{\bigtriangleup}\bigg]. \end{split} \end{align} In (\ref{appendixAeq6}), to get line two from the line one, note that $\prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} C_{\circledast}C_{\bigtriangleup}$ is a diagonal matrix. \\
To evaluate the trace of the given diagonal matrix, we need to find the difference between the number of $+1$ and $-1$ on the diagonal. We write every row in the computational basis by enumerating it with the binary notation. For each row, we denote by $\alpha$ the number of 1's in binary notation appearing in the first $m$ columns and by $\beta$, the same on the rest. For example, for $N=7$ , and $m=4$, the basis element $\ket{1101110}$ leads to $\alpha$=3 and $\beta=2$. Considering the phase gates in the equation (\ref{appendixAeq6}), the expression $(-1)^s$ defines whether in the given row the diagonal element is $+1$ or $-1$, where : \begin{equation} s:=\binom{\alpha}{1}\binom{\beta}{1}+\binom{\alpha}{1}+\binom{\beta}{1}=\alpha \beta +\alpha+\beta . \end{equation} In $s$, $\binom{\alpha}{1}\binom{\beta}{1}$ denotes how many $C_{\circledast\bigtriangleup} $ acts on the row. Also, $\binom{\alpha}{1}$ determines the number of $C_{\circledast}$ and $\binom{\beta}{1}$, number of $C_{\bigtriangleup} $. Every time when the phase gate acts, it changes the sign of the diagonal element on the row. Therefore, we need to determine the number s:\\
To see whether $s$ is even or odd, we have to consider the following cases exhaustively:\\ $\quad$\\ \begin{tabular}{lllll} \textbf{1. } & $\alpha$ is even \& $\beta$ is even & $(-1)^{s}=+1$ & & \multirow{2}{*}{$\Rightarrow$ These two cases sum up to zero.}\tabularnewline \textbf{2. } & $\alpha$ is even \& $\beta$ is odd & $(-1)^{s}=-1$ & & \tabularnewline \end{tabular}\\ $\quad$\\ \begin{tabular}{lllll} \textbf{3. } & $\alpha$ is odd \& $\beta$ is even & $(-1)^{s}=-1$ & & \multirow{2}{*}{$\Rightarrow$ These two cases contribute with the negative sign.}\tabularnewline \textbf{4.} & $\alpha$ is odd \& $\beta$ is odd & $(-1)^{s}=-1$ & & \tabularnewline \end{tabular}\\
From the cases 3 and 4, one can directly calculate the trace:
\begin{equation}\label{appendixAeq7} \langle K \rangle=\frac{1}{2^N}\bigg[-\sum_{\alpha=1,3,\dots }^m\binom{m}{\alpha}\sum_{\beta=0}^{N-m}\binom{N-m}{\beta}\bigg]=-\frac{2^{m-1}2^{N-m}}{2^N}=-\frac{1}{2}. \end{equation} So, we get that if $m$ is divisible by 4, \begin{equation} \langle \underset{m}{\underbrace{X\dots X}}Z\dots Z\rangle =-\frac{1}{2}. \end{equation}
\textbf{Case \#2:} We use the identical approach: when $m=2 \; mod \; 4 $, we write out all remaining phase gates and continue the derivation from equation (\ref{appendixAeq1}): \begin{equation}\label{appendixAeq8} \langle K \rangle=\bra{+}^{\otimes N} X\dots X Z\dots Z \prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} C_{\bigtriangleup}\ket{+}^{\otimes N}. \end{equation}\\ Again we use the fact that $X$ is an eigenstate of $\bra{+}$ and $Z=C_{\bigtriangleup}$. As in (\ref{appendixAeq8}), there is already one $(C_{\bigtriangleup})$, they cancel. Therefore, we are left with: \begin{equation}\label{appendixAeq8} \langle K \rangle=\bra{+}^{\otimes N} \prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} \ket{+}^{\otimes N}=\frac{1}{2^N}Tr\bigg[\prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\bigtriangleup} \bigg]. \end{equation}\\ We need to define the sign of the diagonal element by $(-1)^s$, where \begin{equation} s=\binom{\alpha}{1}\binom{\beta}{1}=\alpha\beta. \end{equation}
\begin{tabular}{lllll} \textbf{1. } & $\alpha$ is even & $(-1)^{s}=+1$ & & $\Rightarrow$ This case contributes with the positive sign in the trace \tabularnewline \end{tabular}
$ $
\begin{tabular}{lllll} \textbf{2. } & $\alpha$ is odd \& $\beta$ is even & $(-1)^{s}=+1$ & & \multirow{2}{*}{$\Rightarrow$ These two give zero contribution together.}\tabularnewline \textbf{3.} & $\alpha$ is odd \& $\beta$ is odd & $(-1)^{s}=-1$ & & \tabularnewline \end{tabular}
As the case 2 and 3 add up to zero, we only consider the case 1: \begin{equation} \langle K \rangle=\frac{1}{2^N}\sum_{\alpha=0,2,\dots }^m\binom{m}{\alpha}\sum_{\beta=0}^{N-m}\binom{N-m}{\beta}=\frac{2^{m-1}2^{N-m}}{2^N}=\frac{1}{2}. \end{equation} So, we get that if $m$ is NOT divisible by 4, \begin{equation} \langle \underset{m}{\underbrace{X\dots X}}Z\dots Z\rangle =\frac{1}{2}. \end{equation} This completes the proof. \end{proof}
\subsection{B.3. Correlations for X measurements on fully-connected three-uniform hypergraph states}
\begin{lemma} If every party makes a measurement in $X$ direction, then the expectations value is \begin{equation} \langle \underset{N}{\underbrace{XX\dots XX}}\rangle =\begin{cases} \begin{array}[t]{cc} 0 & \mbox{\ensuremath{\mbox{if \ensuremath{N=0} mod \ensuremath{4}}}},\\ 1 & \mbox{\ensuremath{\mbox{if \ensuremath{N=2} mod \ensuremath{4}}}}. \end{array}\end{cases} \end{equation} \end{lemma} \begin{proof} In this proof we employ the notation introduced in details in the proof of the lemma \ref{lemma3full}. \begin{equation} \label{allX3unif} \langle K \rangle:=\bra{H_3^N} XX \dots XX\ket{H_3^N}=\bra{+}^{\otimes N} \Big[ \prod_{e\in E} C_e \Big] XX\dots XX \Big[ \prod_{e\in E} C_e\Big] \ket{+}^{\otimes N}. \end{equation} We use the identity (\ref{identity}), to regroup the phase gates on the right hand side of the expression (\ref{allX3unif}). Therefore, we count the multiplicity of the remaining phase gates:\\
\begin{tabular}{crlcc}
& \#$\circledast\circledast\circledast$ & Each $C_{e},$ where $|e|=3$, occurs once and cancels with the one on right hand side. & & \tabularnewline
& \#$\circledast\circledast$ & $=\binom{N-2}{1}$ is even $\Rightarrow C_{\circledast\circledast}$ cancels& . & \tabularnewline
& \#$\circledast$ & $=\binom{N-1}{2}=\frac{(N-1)(N-2)}{2}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{odd, if }N=0\ mod\ 4 & \Rightarrow C_{\circledast}\mbox{ remains.}\\ \mbox{even, if \ensuremath{N=2\ }\ensuremath{mod\ 4}} &\Rightarrow C_{\circledast}\mbox{ cancels.} \end{array}\end{cases}$ & & \tabularnewline
& \#$\{\}$ & $=\binom{N}{3}=\frac{N(N-1)(N-2)}{2\cdot3}\ $ is even $\Rightarrow$ global sign $GS$ is positive. & & \tabularnewline \end{tabular}\\
Therefore, we need to consider two cases to continue the derivation of the expression (\ref{allX3unif}): \\
\textbf{Case \#1:} If $N=0\ mod \; 4$, then \begin{equation} \langle K \rangle =\bra{+}^{\otimes N}XX\dots XX \prod_{\circledast\in E} C_{\circledast} \ket{+}^{\otimes N}= \bra{+}^{\otimes N}\prod_{\circledast\in E} C_{\circledast} \ket{+}^{\otimes N} =0. \end{equation}
\textbf{Case \#2:} If $N=2\ mod \; 4$, then \begin{equation} \langle K \rangle = \bra{+}^{\otimes N}XX\dots XX \ket{+}^{\otimes N}=\Big[\braket{+}{+}\Big] ^N=1. \end{equation} \end{proof}
\subsection{B.4. Correlations for Z measurements on fully-connected three-uniform hypergraph states}
\begin{lemma} If every party makes a measurement in $Z$ direction, the expectation value is zero. Therefore, we need to show that \begin{equation} \langle K \rangle :=\langle ZZ\dots ZZ\rangle=0 \end{equation} \end{lemma} \begin{proof} \begin{equation} \langle K \rangle = \bra{+}^{\otimes N}\Big[\prod_{e \in E}C_e\Big]ZZ\dots ZZ\Big[\prod_{e \in E}C_e\Big] \ket{+}^{\otimes N}=\bra{+}^{\otimes N}ZZ\dots ZZ \ket{+}^{\otimes N}=\pm\frac{1}{2^N} Tr(Z\dots ZZ)=0. \end{equation} \end{proof}
\section{Appendix C: Proof of Observation 5}
\noindent {\bf Observation 5.} {\it If we fix in the Bell operator $\ensuremath{\mathcal{B}}_N$ in Eq.~(\ref{bell2}) the measurements to be $A=Z$ and $B=X$, then the $N$-qubit fully-connected three-uniform hypergraph state violates the classical bound, by an amount that grows exponentially with number of qubits, namely \begin{align} \left\langle \ensuremath{\mathcal{B}}_N \right\rangle_C &\leq 2^{\left\lfloor N/2\right\rfloor} \quad \mbox{ for local HV models, and} \nonumber \\ \left\langle \ensuremath{\mathcal{B}}_N \right\rangle_Q &\geq 2^{N-2}-\frac{1}{2} \quad \mbox{ for the hypergraph state.} \end{align} }
The classical bound can be computed using the bound for Mermin inequality, i.e., $\mean{\ensuremath{\mathcal{B}}^M_N}_C\leq 2^{\lfloor N/2 \rfloor}$. If $N$ is odd, our Bell inequality is exactly the Mermin inequality. If $N$ is even, $\mean{\ensuremath{\mathcal{B}}_N}_C = \mean{A \cdot \ensuremath{\mathcal{B}}^M_{N-1} +B \cdot \tilde\ensuremath{\mathcal{B}}^M_{N-1}}_C \leq 2 \mean{\ensuremath{\mathcal{B}}^M_{N-1}}_C $, where $\tilde\ensuremath{\mathcal{B}}$ denotes the inequality where $A$ and $B$ are exchanged, and the claim follows. The quantum value is computed directly. In particular, for odd $N$, we have $ \left\langle \ensuremath{\mathcal{B}}_N \right\rangle_Q=\sum_{k \;even}^N\binom{N}{k}\frac{1}{2}=2^{N-2}. $ If $N=0\ {\rm mod}\ 4$, then $ \left\langle \ensuremath{\mathcal{B}}_N \right\rangle_Q=\sum_{k\ even}^{N-1}\binom{N}{k}\frac{1}{2}-\left\langle X\dots X \right\rangle=2^{N-2}-{1}/{2} $ and for $N=2\ mod\ 4$, we have $ \left\langle \ensuremath{\mathcal{B}}_N \right\rangle_Q=\sum_{k\ even}^{N-1}\binom{N}{k}\frac{1}{2}-\left\langle X\dots X \right\rangle=2^{N-2}+{1}/{2}. $
Notice that, since Eq.~\eqref{bell2} in the main text is a full-correlation Bell inequality, it can be written as a sum of probabilities minus a constant equal to the number of terms. With the proper normalization, such probabilities exactly correspond to the success probability of computing a Boolean function (cf. conclusive discussion and Ref.~\cite{Hoban11}).
\section{Appendix D: Correlations for four-uniform hypergraph states}
\begin{lemma}\label{lemma4full} The following statements hold for N-qubit, four-uniform hypergraph states: \begin{enumerate} \item For the case $N=8k-2 \mbox{ or } \; N=8k-1,\; \mbox{ or } 8k $, we have:\\ $(i)$ \begin{equation}\label{third} \langle \underset{m}{\underbrace{X\dots X}}Z\dots Z\rangle =\begin{cases} \begin{array}[t]{cc} +\frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }} & \mbox{\ensuremath{\mbox{if \ensuremath{(m-1)=0} mod \ensuremath{4}}}},\\ -\frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }} & \mbox{\ensuremath{\mbox{if \ensuremath{(m-1)=2} mod \ensuremath{4}}}}. \end{array}\end{cases} \end{equation} $(ii)$ For $N=8k-1$, we have: \begin{equation} \langle \underset{N}{\underbrace{XX\dots XX}}\rangle =-1. \end{equation} For $N=8k-2$ or $N=8k$, these correlations will not be needed.\\
\item For $N=4k+1$, we have: \begin{equation} \langle \underset{m}{\underbrace{X\dots X}}Z\dots Z\rangle =\begin{cases} \begin{array}[t]{cc} +\frac{1}{2^{\left\lceil m/2\right\rceil }}, & \mbox{if \ensuremath{(m-1)=0 \ mod\ 4}},\\ -\frac{1}{2^{\left\lceil m/2\right\rceil }}, & \ \mbox{if \ensuremath{(m-1)=2 \ mod\ 4}},\\ \frac{1}{2^{\left\lfloor N/2\right\rfloor }}, & \mbox{if }\mbox{\ensuremath{m=N}}. \end{array}\end{cases} \end{equation}\\
\item For $N=8k+2,\; \mbox{ or } 8k+4 $, we have for even $m$:\\ (i) \begin{equation} \langle \underset{m}{\underbrace{X\dots X}}Z\dots Z\rangle =\begin{cases} \begin{array}[t]{cc} +\frac{2^{m/2-1}}{2^{N/2 }} & \mbox{if \ensuremath{(N-m)=0\ mod\ 4 }},\\ -\frac{2^{m/2-1}}{2^{N/2 }} & \mbox{if \ensuremath{(N-m)=2 \ mod\ 4 }}. \end{array}\end{cases} \end{equation}\\ (ii) \begin{equation}
\langle XX\dots XX\rangle =\frac{2^{\frac{N}{2}-1}+1}{2^{\frac{N}{2}}}. \end{equation} \\ \item For $N=8k+3$, $\langle \underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1\rangle$ for even $m$ gives the same exact result as the part 3, so we have: \begin{equation} \langle \underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1\rangle=\begin{cases} \begin{array}[t]{cc} +\frac{2^{m/2-1}}{2^{M/2 }} & \mbox{if \ensuremath{(M-m)=0\ mod\ 4 }},\\ -\frac{2^{m/2-1}}{2^{M/2 }} & \mbox{if \ensuremath{(M-m)=2 \ mod\ 4 }}, \end{array}\end{cases} \end{equation} where $M=N-1.$ \end{enumerate} \end{lemma}
\begin{proof} Each part of the theorem needs a separate consideration. note that the notation and machinery that is employed in the proof is based on the proof of the Lemma \ref{lemma3full}. Therefore, we advise the reader to become familiar with that one first. \\
\textbf{ Part 1. } We consider the cases when $N=8k-2, \; N=8k-1\; \mbox{ or } 8k $ and odd $m$ together. \\ We prove $(i)$ first:\\ \begin{equation}\label{part1_1} \langle G_1 \rangle = \bra{H_4^N} \underset{m}{\underbrace{X\dots X}}Z\dots Z\ket{H_4^N} =\bra{+}^{\otimes N}\Big(\prod_{e\in E}C_e \Big)XX\dots XZ\dots Z \Big(\prod_{e\in E}C_e \Big) \ket{+}^{\otimes N}. \end{equation} We need to use the identity (\ref{identity}) to regroup all the phase gates on the right hand side of the expression (\ref{part1_1}). If each phase gate occurs even number of times, they give an identity, otherwise, they are used in the further calculations. We consider each case separately in the following table:
\begin{tabular}{rrllll}
& & \multicolumn{4}{l}{}\tabularnewline \textbf{1.} & \#$\bigtriangleup\bigtriangleup\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m}{1}$ is odd $\Rightarrow C_{\bigtriangleup\bigtriangleup\bigtriangleup}$ remains. }\tabularnewline
& & & & & \tabularnewline \textbf{2.} & \#$\bigtriangleup\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m}{2}=\frac{m(m-1)}{2}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{even, if (\ensuremath{m-1)=0} mod 4 } & \Rightarrow C_{\bigtriangleup\bigtriangleup}\mbox{ cancels.}\\ \mbox{odd, if (\ensuremath{m-1)=2} mod 4 } & \Rightarrow C_{\bigtriangleup\bigtriangleup}\mbox{ remains.} \end{array}\end{cases}$ }\tabularnewline
& \#$\circledast\bigtriangleup\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m-1}{1}$ is even $\Rightarrow\; C_{\circledast\bigtriangleup\bigtriangleup}$ cancels.}\tabularnewline
& & & & & \tabularnewline \textbf{3.} & \#$\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m}{3}=\frac{m(m-1)(m-2)}{2\cdot3}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{even, if (\ensuremath{m-1)=0} mod 4 } &\Rightarrow C_{\bigtriangleup}\mbox{ cancels.}\\ \mbox{odd, if (\ensuremath{m-1)=2} mod 4 } & \Rightarrow C_{\bigtriangleup}\mbox{ remains.} \end{array}\end{cases}$}\tabularnewline
& \#$\circledast\circledast\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m-2}{1}$ is odd $\Rightarrow$ $C_{\circledast\circledast\bigtriangleup}$ remains.}\tabularnewline
& \#$\circledast\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m-1}{2}=\frac{(m-1)(m-2)}{2}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{even, if (\ensuremath{m-1)=0} mod 4 } & \Rightarrow C_{\circledast\bigtriangleup}\mbox{ cancels.}\\ \mbox{odd, if (\ensuremath{m-1)=2} mod 4 } & \Rightarrow C_{\circledast\bigtriangleup}\mbox{ remains.} \end{array}\end{cases}$}\tabularnewline
& & & & & \tabularnewline \textbf{4.}
& \#$\circledast\circledast\circledast$ & \multicolumn{4}{l}{$=\binom{m-3}{1}$ is even $\Rightarrow\; C_{\circledast\circledast\circledast}$ cancels.}\tabularnewline
& \#$\circledast\circledast$ & \multicolumn{4}{l}{$=\binom{m-2}{2}=\frac{(m-2)(m-3)}{2}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{odd, if (\ensuremath{m-1)=0} mod 4 } & \Rightarrow C_{\circledast\circledast}\mbox{ remains.}\\ \mbox{even, if (\ensuremath{m-1)=2} mod 4 } & \Rightarrow C_{\circledast\circledast}\mbox{ cancels.} \end{array}\end{cases}$}\tabularnewline
& \#$\circledast$ & \multicolumn{4}{l}{$=\binom{m-1}{3}=\frac{(m-1)(m-2)(m-3)}{2\cdot3}$ is even $\Rightarrow\; C_{\circledast}$ cancels. }\tabularnewline
& \#$\{\}$ & \multicolumn{4}{l}{$=\binom{m}{4}=\frac{(m-1)(m-2)(m-3)(m-4)}{2\cdot3\cdot4}\ $ affects global sign ($GL$) and will be discussed separately.}\tabularnewline \end{tabular} \begin{center} \textbf{Table 1.} Counting phase gates for a four-uniform case. $m$ is odd. \end{center}
$\quad $\\ \textbf{Remark:} All four-qubit phase gates move with the multiplicity one to the right side and therefore, cancel out with the same phase gate on the right. The detailed reasoning was discussed in proof of the Lemma (\ref{lemma3full}) in the Appendix A. So, we skipped such scenarios in the Table 1.
Now we consider two cases of $(m-1)$ separately and for each case we fix the global sign ($GL$) defined in the Table 1.\\
\textbf{Case \# 1:} $(m-1)=0$ mod 4: \begin{equation}\label{part1_2} \langle G_1 \rangle =\pm\bra{+}^{\otimes N} XX\dots XZ\dots Z \prod_{\forall \circledast, \forall\bigtriangleup } C_{\bigtriangleup \bigtriangleup \bigtriangleup} C_{\circledast \circledast \bigtriangleup}C_{\circledast\circledast} \ket{+}^{\otimes N}=\pm \frac{1}{2^N} Tr( \prod_{\forall \circledast, \forall\bigtriangleup } C_{\bigtriangleup \bigtriangleup \bigtriangleup} C_{\circledast \circledast \bigtriangleup}C_{\circledast\circledast}C_{\bigtriangleup}). \end{equation} \textbf{Remark:} We write the '$\pm$' sign as we have not fixed the global sign yet.\\
To evaluate the trace of the given diagonal matrix, we need to find the difference between the number of $+1$'s and $-1$'s on the diagonal. We write every row in the computational basis by enumerating it with the binary notation. Due to the symmetry of the problem, we assign the first $m$ columns to $X$ measurement ($\circledast$) and the rest to, $Z$($\bigtriangleup$). For each row, we denote by $\alpha$ the number of 1's in binary notation appearing in the first $m$ column and by $\beta$, the same on the rest. This notation is also adopted. See the proof of Lemma \ref{lemma3full} for more detailed explanation.\\
Considering the phase gates in (\ref{part1_2}), the expression $(-1)^s$ defines whether in the given row the diagonal element is $+1$ or $-1$, where : \begin{equation} s:=\binom{\beta}{ 3}+\binom{\alpha}{2}\binom{\beta}{1}+\binom{\alpha}{2}+\binom{\beta}{1}=\frac{\beta (\beta-1)(\beta-2)}{2\cdot 3}+\frac{\alpha(\alpha -1)}{2}(\beta+1)+\beta. \end{equation} The sign of the diagonal element is determined at follows:
\begin{tabular}{lllc} \textbf{1. } & $\alpha$ is even \& $\beta$ is even: & \multicolumn{2}{l}{if $\alpha=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2 $ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{2. } & $\alpha$ is odd \& $\beta$ is even: & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{3.} & (Any $\alpha$) \& $\beta$ is odd: & \multicolumn{2}{l}{if $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}\\
Having established the $\pm 1$ values for each row, we can sum them up to find the trace in (\ref{part1_2}). Here we use the identities (\ref{identity1}) and (\ref{identity2}) and afterwards the look-up Table 0 to insert the numerical values where necessary:
\begin{align} \label{part1_3} \begin{split} \langle G_1 \rangle = &\pm\frac{1}{2^N}\Bigg[\sum^{N-m}_{\beta=0,2,4\dots } \binom{N-m}{\beta}\bigg[\sum_{\alpha =0,4,\dots}^{m}\binom{m}{\alpha}+\binom{m}{\alpha +1}-\binom{m}{\alpha +2}-\binom{m}{\alpha +3}\bigg] \\ +&\sum_{\beta =1,5,\dots} \bigg[-\binom{N-m}{\beta}\sum_\alpha \binom{m}{\alpha}+\binom{N-m}{\beta+2}\sum_\alpha \binom{m}{\alpha}\bigg]\Bigg]\\ = &\pm\frac{1}{2^N}\Bigg[\bigg[Re\Big[(1+i)^m \Big] +Im\Big[(1+i)^m\Big]\bigg]2^{N-m-1} + 2^m \sum^{N-m}_{\beta =1,5,\dots} \bigg[-\binom{N-m}{\beta}+\binom{N-m}{\beta+2}\bigg]\Bigg]\\ =&\pm\frac{1}{2^N}\Bigg[\bigg[Re \Big[(1+i)^m \Big] +Im\Big[(1+i)^m\Big]\bigg]2^{N-m-1} - 2^m Im\Big[(1+i)^{N-m}\Big]\Bigg] \equiv\pm\frac{1}{2^N} E. \end{split} \end{align} We have to consider $N=8k-1$ and $N=8k$ or $N=8k-2$ separately to continue the derivation of (\ref{part1_3}):\\
1. For $N=8k-1$, using the values from the Table 0: \begin{equation} \langle G_1 \rangle=\pm \frac{1}{2^N} \Bigg[2^{\frac{m+1}{2}}2^{N-m-1}- 2^m Im\Big[(1+i)^{N-m}\Big]\Bigg] =\pm\frac{2^{\frac{m+1}{2}}2^{N-m-1}+2^m2^{\frac{N-m}{2}}}{2^N}= \pm \frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }}. \end{equation} \\
2. For $N=8k$ or$N=8k-2$ , using the values from the Table 0: \begin{equation} \langle G_1 \rangle=\pm \frac{1}{2^N} \Bigg[2^{\frac{m+1}{2}}2^{N-m-1}- 2^m Im\Big[(1+i)^{N-m}\Big]\Bigg] =\pm\frac{2^{\frac{m+1}{2}}2^{N-m-1}+2^m2^{\frac{N-m-1}{2}}}{2^N}= \pm \frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }}. \end{equation}
Therefore, \begin{equation}\label{part1_4}
\langle G_1 \rangle = \pm \frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }}. \end{equation} Concerning the sign in (\ref{part1_4}), it is affected by the product of two components: one from the case \textbf{4} from Table 1: $GL$ and the other by $E$ from (\ref{part1_3}). If $(m-1)=0$ mod 8, the equation $E$ has a positive sign and also $GL=+1$. And if $(m-1)=4$ mod 8, $E$ has a negative sign and $GL=-1$. Therefore, in both cases or equivalently, for $(m-1) =0\; mod\; 4$, \begin{equation} \langle G_1\rangle =+\frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }}. \end{equation}
\textbf{Case \# 2:} $(m-1)=2$ mod 4: \begin{align}\label{par2_1} \begin{split} \langle G_1 \rangle & =\pm\bra{+}^{\otimes N} XX\dots XZ\dots Z \prod_{\forall \circledast, \forall\bigtriangleup } C_{\bigtriangleup \bigtriangleup \bigtriangleup} C_{\bigtriangleup \bigtriangleup } C_{\circledast \circledast \bigtriangleup}C_{\circledast\bigtriangleup}C_{\bigtriangleup } \ket{+}^{\otimes N}\\ & =\pm \frac{1}{2^N} Tr( \prod_{\forall \circledast, \forall\bigtriangleup } C_{\bigtriangleup \bigtriangleup \bigtriangleup} C_{\bigtriangleup \bigtriangleup } C_{\circledast \circledast \bigtriangleup}C_{\circledast\bigtriangleup}). \end{split} \end{align} In this case, we apply the same technique to determine the sign:
\begin{equation} s:=\binom{\beta}{ 3}+\binom{\beta}{ 2}+\binom{\alpha}{2}\binom{\beta}{1}+\binom{\alpha}{1}\binom{\beta}{1}=\frac{\beta (\beta-1)(\beta-2)}{2\cdot 3}+\frac{\beta(\beta-1)}{2}+\frac{\alpha(\alpha -1)}{2}\beta+\alpha\beta. \end{equation} The sign of $s$, is determined at follows: \\
\begin{tabular}{lllc} \textbf{1. } & $\beta$ is even \& any $\alpha$ & \multicolumn{2}{l}{if $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{2. } & $\beta$ is odd \& $\alpha$ is even: & \multicolumn{2}{l}{if $\alpha=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{3.} & $\beta$ is odd \& $\alpha$ is odd: & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}\\
\begin{align} \label{part2_2} \begin{split} \langle G_1\rangle =& \pm\frac{1}{2^N} \Bigg[\sum_{\beta=0,4\dots } \bigg[\binom{N-m}{\beta}\sum_{\alpha =0}^{m}\binom{m}{\alpha}-\binom{N-m}{\beta +2}\sum_{\alpha =0}^{m}\binom{m}{\alpha}\bigg] \\ + &\sum_{\beta=1,3,\dots } \binom{N-m}{\beta}\bigg[\sum_{\alpha =0,4,\dots}^{m}\binom{m}{\alpha}-\binom{m}{\alpha +1}-\binom{m}{\alpha +2}+\binom{m}{\alpha +3}\bigg]\Bigg]\\ = & \pm \frac{1}{2^N}\Bigg[2^m \sum_{\beta =0,4,\dots} \bigg[\binom{N-m}{\beta}-\binom{N-m}{\beta+2}\bigg]+2^{N-m-1} \bigg[Re\Big[(1+i)^m \Big] -Im\Big[(1+i)^m\Big]\bigg]\Bigg]\\ = & \pm \frac{1}{2^N}\Bigg[2^m Re\Big[(1 +i)^{N-m}\Big] +2^{N-m-1} \bigg[Re\Big[(1+i)^m \Big] -Im\Big[(1+i)^m\Big]\bigg]\Bigg]\equiv \frac{1}{2^N}E. \end{split} \end{align}
We have to consider $N=8k-1$ and $N=8k$ or $N=8k-2$ separately to continue the derivation of (\ref{part1_3}):\\
1. For $N=8k-1$, using the values from the Table 0: \begin{equation} \langle G_1 \rangle=\pm \frac{1}{2^N} \Bigg[ 2^m Re\Big[(1+i)^{N-m}\Big]+2^{\frac{m+1}{2}}2^{N-m-1}\Bigg] =\pm\frac{2^m2^{\frac{N-m}{2}}+2^{\frac{m+1}{2}}2^{N-m-1}}{2^N}= \pm \frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }}. \end{equation} 2. For $N=8k$ or$N=8k-2$ , using the values from the Table 0: \begin{equation} \langle G_1 \rangle=\pm \frac{1}{2^N} \Bigg[ 2^m Re\Big[(1+i)^{N-m}\Big]+2^{\frac{m+1}{2}}2^{N-m-1}\Bigg] =\pm\frac{2^m2^{\frac{N-m-1}{2}}+2^{\frac{m+1}{2}}2^{N-m-1}}{2^N}= = \pm \frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }}. \end{equation}
Therefore, \begin{equation}\label{part2_3}
\langle G_1 \rangle = \pm \frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }}. \end{equation} Concerning the sign in (\ref{part2_3}), it is affected by the product of two components: one from the case \textbf{4} from Table 1: $GL$ and the other by $E$ in (\ref{part2_2}).\\ Hence, if $(m-3)=0$ mod 8, $E$ has a negative sign and $GL=+1$. And if $(m-3)=4$ mod 8, $E$ has a positive sign and $GL=-1$. Therefore, in both cases or equivalently, for $(m-1) =2\; mod\; 4$, \begin{equation}\label{part2_4}
\langle G_1 \rangle = - \frac{2^{\left\lfloor N/ 2\right\rfloor -m}+1}{2^{\left\lfloor N/ 2\right\rfloor - \left\lfloor m/ 2\right\rfloor }}. \end{equation} This completes the proof of part 1 (i).\\
\textbf{Part 1} $(ii)$ For $N=8k-1$, show that \begin{equation} \langle G_2\rangle=\langle \underset{N}{\underbrace{XX\dots XX}}\rangle =-1. \end{equation} Here as well we use the identity (\ref{identity}) to count the multiplicity of remaining phase gates. Since all the measurements are in $X$ direction, we need to make a new table with the same notations as in the previous case:\\
\begin{tabular}{lrl}
& & \tabularnewline $1.$& $\#\circledast\circledast\circledast\circledast$ & every gate occurs only once $\Rightarrow$ every $C_{e}$ cancels with the $C_{e}$ on the right hand side.\tabularnewline
& & \tabularnewline $2.$ & $\#\circledast\circledast\circledast$ & $\binom{8k-4}{1}$ is even $\Rightarrow \; C_{\circledast\circledast\circledast}$ cancels.\tabularnewline
& & \tabularnewline $3.$ & $\#\circledast\circledast$ &$ \binom{8k-3}{2}$ is even $\Rightarrow \; C_{\circledast\circledast}$ cancels.\tabularnewline
& & \tabularnewline $4.$ & $\#\circledast$ & $\binom{8k-2}{3}$ is even $\Rightarrow C_{\circledast}$ cancels.\tabularnewline
& & \tabularnewline $5.$ & $\#\{\}$ & $\binom{8k-1}{4}$ is odd $\Rightarrow$ we get a global negative sign. \tabularnewline \end{tabular} \begin{center} \textbf{Table 2.} Counting phase gates for a four-uniform HG when each system is measured in $X$ direction. \end{center}
$\quad $\\ Therefore, \begin{equation} \langle G_2 \rangle=-\frac{1}{2^N}Tr\big(\mathbbm{1}\big)=-1. \end{equation} This finishes the proof of part 1.\\
\textbf{Part 2:} We show that, for $N=4k+1$: \begin{equation} \langle G_3\rangle=\langle \underset{m}{\underbrace{X\dots X}}Z\dots Z\rangle =\begin{cases} \begin{array}[t]{cc} +\frac{1}{2^{\left\lceil m/2\right\rceil }}, & \mbox{if \ensuremath{(m-1)=0 \ mod\ 4}},\\ -\frac{1}{2^{\left\lceil m/2\right\rceil }}, & \ \mbox{if \ensuremath{(m-1)=2 \ mod\ 4}},\\ \frac{1}{2^{\left\lfloor N/2\right\rfloor }}, & \mbox{if }\mbox{\ensuremath{m=N}}. \end{array}\end{cases} \end{equation}
Since the number of systems measured in the $X$ direction is the same in this part as it was in part 1, we can use the results demonstrated in the Table 1. Therefore, we use equation (\ref{part1_3}) when $m-1=0$ mod 4 and (\ref{part2_2}), for $m-1=2$ mod 4. \\
\textbf{Case \# 1:} $(m-1)=0$ mod 4: \begin{equation}\label{part3_1} \langle G_3 \rangle=\pm\frac{1}{2^N}\Bigg[\bigg[Re \Big[(1+i)^m \Big] +Im\Big[(1+i)^m\Big]\bigg]2^{N-m-1} - 2^m Im\Big[(1+i)^{N-m}\Big]\Bigg]=\pm\frac{1}{2^N} E. \end{equation} As $N=4k+1$, we have that $\; N-m=4k+1-m=4k-(m-1)$, which is divisible by $4$. Therefore $Im\Big[(1+i)^{N-m}\Big]=0$. and equation (\ref{part3_1}) reduces to: \begin{equation}\label{part3_2} \langle G_3 \rangle=\pm\frac{2^{N-m-1}}{2^N}\bigg[Re \Big[(1+i)^m \Big] +Im\Big[(1+i)^m\Big]\bigg]=\pm \frac{1}{2^{\left\lceil m/2\right\rceil }}. \end{equation} We need to fix the global sign $GL$ from the Table 1. For this, we consider two cases. First, if $(m-1)=0$ mod 8, then $GL=+$ and so is the sign $E$ in equation (\ref{part3_1}): \begin{equation} \langle G_3 \rangle=+\frac{1}{2^{\left\lceil m/2\right\rceil }}. \end{equation} Second, if $(m-1)=4$ mod 8, then $GL=-$ and so is the sign of$E$ in equation (\ref{part3_1}): \begin{equation} \langle G_3 \rangle=+\frac{1}{2^{\left\lceil m/2\right\rceil }}. \end{equation}\\
\textbf{Case \# 2:} For $(m-1)=2$ mod 4: \begin{equation}\label{part3_3} \langle G_3 \rangle= \pm \frac{1}{2^N}\Bigg[2^m Re\Big[(1 +i)^{N-m}\Big] +2^{N-m-1} \bigg[Re\Big[(1+i)^m \Big] -Im\Big[(1+i)^m\Big]\bigg]\Bigg]. \end{equation}
As $N=4k+1$, we have $ N-m=4k+1-m=4k-(m-1)$, which is not divisible by $4$ but is an even number. Therefore, $Re\Big[(1+i)^{N-m}\Big]=0$. So, the equation (\ref{part3_3}) reduces to:
\begin{equation}\label{part3_4} \langle G_3 \rangle= \pm \frac{2^{N-m-1}}{2^N} \bigg[Re\Big[(1+i)^m \Big] -Im\Big[(1+i)^m\Big]\bigg]\equiv \pm \frac{2^{N-m-1}}{2^N} E. \end{equation} We need to fix the global sign $GL$ from the Table 1. For this, we consider two cases. First, if $(m-3)=0$ mod 8, then the global sign is positive but the sign of $E$ in (\ref{part3_4}) is negative. Therefore, \begin{equation} \langle G_3 \rangle=-\frac{1}{2^{\left\lceil m/2\right\rceil }}. \end{equation} Second, if $(m-3)=4$ mod 8, then the global sign is negative but the sign of $E$ in (\ref{part3_4}) is positive. Therefore, \begin{equation} \langle G_3 \rangle=-\frac{1}{2^{\left\lceil m/2\right\rceil }}. \end{equation}
\textbf{Case \# 3:} $m=N$ resembles part 1 $(ii)$. The only difference comes in with the number of qubits we are currently working with:
\begin{tabular}{lrl}
& & \tabularnewline $1$ & $\#\circledast\circledast\circledast\circledast$ & e every gate occurs only once $\Rightarrow$ every $C_{e}$ cancels with the $C_{e}$ on the right hand side.\tabularnewline
& & \tabularnewline $2.$ & $\#\circledast\circledast\circledast$ & $\binom{4k-2}{1}$ is even $\Rightarrow \; C_{\circledast\circledast\circledast}$ cancels.\tabularnewline
& & \tabularnewline $3.$ & $\#\circledast\circledast$ & $\binom{4k-1}{2}$ is odd $\Rightarrow C_{\circledast\circledast}$ remains. \tabularnewline
& & \tabularnewline $4.$ & $\#\circledast$ & $\binom{4k}{3}$ is even $\Rightarrow C_{\circledast}$ cancels.\tabularnewline
& & \tabularnewline $5.$ & $\#\{\}$ & the global sign depends on $k$, as $\binom{4k+1}{4}=\frac{(4k+1)4k(4k-1)(4k-2)}{2\cdot 3\cdot 4}$ \tabularnewline \end{tabular}\\ \begin{center} \textbf{Table 3}. Counting phase gates for a four-uniform HG when each system is measured in $X$ direction. \end{center}
$\quad $\\ Back to the expectation value, \begin{equation} \langle G_3 \rangle= \pm \frac{1}{2^N}Tr\bigg[\prod_{\forall \circledast}C_{\circledast\circledast}\bigg]. \end{equation} So, we have to count the difference between the amount of $+1$'s and $-1$'s on the diagonal. As we use exactly the same techniques before, we will skip the detailed explanation. The sign on the diagonal is: \begin{equation} (-1)^{\binom{\alpha}{2}}=(-1)^{\frac{\alpha(\alpha-1)}{2}}. \end{equation} and it is straightforward to evaluate it for each value of $\alpha$. \\ \begin{equation}\label{aaa} \langle G_3 \rangle=\pm\frac{1}{2^N}\Bigg [\sum_{\alpha =0,4,\dots}^{N}\binom{N}{\alpha}+\binom{N}{\alpha +1}-\binom{N}{\alpha +2}-\binom{N}{\alpha +3}\Bigg ] =\pm\frac{1}{2^N}\bigg[Re \Big[(1+i)^N \Big] +Im\Big[(1+i)^N\Big]\bigg]\equiv \pm\frac{1}{2^N}E. \end{equation}\\
Keeping in mind that $N=4k+1$, the global sign from the Table 3 is positive for even $k$ and negative for odd. The the sign of $E$ in (\ref{aaa}) is positive if $k$ is even and negative, otherwise. Therefore, \begin{equation} \langle G_3 \rangle=\frac{1}{2^{\left\lfloor N/2\right\rfloor }}. \end{equation}
This completes the proof of part 2.
\textbf{Part 3: } We start with (ii). We show that for $N=8k+2,\; \mbox{ or } 8k+4 $:\\ (ii) \begin{equation} \langle G_4 \rangle= \langle XX\dots XX\rangle =\frac{2^{\frac{N}{2}-1}+1}{2^{\frac{N}{2}}}. \end{equation}
Although the result seems identical, unfortunately, each case needs a separate treatment. The technique is similar to the previous proofs, though. We just mind the number of qubits we are working with:\\
For $N=8k+2$ we find the remaining phase gates as follows: \\
\begin{tabular}{lrl}
& & \tabularnewline $1$ & $\#\circledast\circledast\circledast\circledast$ & each gate only once; thus, every $C_{e}$ cancels with the $C_{e}$ on the right hand side.\tabularnewline
& & \tabularnewline $2.$ & $\#\circledast\circledast\circledast$ & $\binom{8k-1}{1}$ is odd $\Rightarrow \; C_{\circledast\circledast\circledast}$ remains.\tabularnewline
& & \tabularnewline $3.$ & $\#\circledast\circledast$ & $\binom{8k}{2}$ is even $\Rightarrow \; C_{\circledast\circledast}$ cancels.\tabularnewline
& & \tabularnewline $4.$ & $\#\circledast$ & $\binom{8k+1}{3}$ is even $\Rightarrow C_{\circledast }$ cancels.\tabularnewline
& & \tabularnewline $5.$ & $\#\{\}$ & $\binom{8k+2}{4}$ is even times $\Rightarrow$ we get a global positive sign. \tabularnewline
& & \tabularnewline \end{tabular} \begin{center} \textbf{Table 4.} Counting phase gates for a four-uniform HG when each system is measured in $X$ direction. \end{center}
$\quad $\\ Therefore, \begin{equation} \langle G_4 \rangle=\frac{1}{2^N}Tr\Big[ \prod_{\forall \circledast}C_{\circledast\circledast\circledast}\Big]. \end{equation}
We use $(-1)^s$ to define the sign of the diagonal element and $s=\binom{\alpha}{3}$. So, after considering all possible values of $\alpha$, it is directly obtained that \begin{align} \begin{split} \langle G_4 \rangle= \frac{1}{2^N} Tr(C_{\circledast\circledast\circledast}) & =\frac{1}{2^N}\bigg[\sum_{\alpha=0,2,\dots}^N\binom{N}{\alpha} + \sum_{\alpha=1,5,\dots}\Big[ \binom{N}{\alpha}-\binom{N}{\alpha+2}\Big] \bigg] \\ & =\frac{1}{2^N}\bigg[2^{N-1} +Im\big[(1+i)^N \big]\bigg] =\frac{2^{\frac{N}{2}-1}+1}{2^{\frac{N}{2}}}. \end{split} \end{align}\\
For $N=8k+4$ we find the remaining phase gates as follows: \\
\begin{tabular}{lrl}
& & \tabularnewline $1.$ & $\#\circledast\circledast\circledast\circledast$ & each gate occurs only once $\Rightarrow$ every $C_{e}$ cancels with the $C_{e}$ on the right hand side.\tabularnewline
& & \tabularnewline $2.$ & $\#\circledast\circledast\circledast$ & $ \binom{8k+1}{1}$ is odd $\Rightarrow \; C_{\circledast\circledast\circledast}$ remains.\tabularnewline
& & \tabularnewline $3.$ & $\#\circledast\circledast$ & $\binom{8k+2}{2}$ is odd $\Rightarrow\; C_{\circledast\circledast}$ remains.\tabularnewline
& & \tabularnewline $4. $ & $\#\circledast$ & $\binom{8k+3}{3}$ is odd $\Rightarrow \; =C_{\circledast}$ remains.\tabularnewline
& & \tabularnewline $5. $ & $\#\{\}$ & $\binom{8k+4}{4}$ is odd $\Rightarrow$ we get a global negative sign, $GL=-1$. \tabularnewline
& & \tabularnewline \end{tabular} \nopagebreak[9] \begin{center} \nopagebreak[9] \textbf{Table 5.} Counting phase gates for a four-uniform case in all $X$ direction. \end{center}
Therefore, \begin{equation} \langle G_4\rangle =-\frac{1}{2^N} Tr\Big[\prod_{\forall \circledast}C_{\circledast\circledast\circledast}C_{\circledast\circledast}C_{\circledast}\Big].
\end{equation} We use $(-1)^s$ to define the sign of the diagonal element and $s=\binom{\alpha}{3}+\binom{\alpha}{2}+\binom{\alpha}{1}$. So, after considering all possible values of $\alpha$, it is directly obtained that \begin{align} \begin{split} \langle G_4\rangle =-\frac{1}{2^N} Tr(C_{\circledast\circledast\circledast}C_{\circledast\circledast}C_{\circledast})& =-\frac{1}{2^N}\bigg[\sum_{\alpha=0,4,\dots}^N\Big[\binom{N}{\alpha}-\binom{N}{\alpha+2}\Big] -\sum_{\alpha=1,3,\dots}^N \binom{N}{\alpha}\bigg]\\ & =-\frac{1}{2^N}\bigg[-2^{N-1} +Re\big[(1+i)^N \big]\bigg] =\frac{2^{N-1}+2^{N/2}}{2^N} =\frac{2^{\frac{N}{2}-1}+1}{2^{\frac{N}{2}}}. \end{split} \end{align} This finishes the proof of part $(i)$.
(ii) We need to show that \begin{equation} \langle G_4\rangle =\langle \underset{m}{\underbrace{X\dots X}}Z\dots Z\rangle =\begin{cases} \begin{array}[t]{cc} +\frac{2^{m/2-1}}{2^{N/2 }} & \mbox{if \ensuremath{(N-m)=0\ mod\ 4 }},\\ -\frac{2^{m/2-1}}{2^{N/2 }} & \mbox{if \ensuremath{(N-m)=2 \ mod\ 4 }}. \end{array}\end{cases} \end{equation} Note that in this case $m$ is an even number. Therefore, we have to derive again from the scratch how phase gates can be moved to the right hand side of the expression and for this we use the identity ($\ref{identity}$).
\begin{tabular}{rrllll}
& & \multicolumn{4}{l}{}\tabularnewline \textbf{1.} & \#$\bigtriangleup\bigtriangleup\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m}{1}$ is even $\Rightarrow\; C_{\bigtriangleup\bigtriangleup\bigtriangleup}$ cancels.}\tabularnewline
& & & & & \tabularnewline \textbf{2.} & \#$\bigtriangleup\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m}{2}=\frac{m(m-1)}{2}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{even, if \ensuremath{m=0} mod 4} & \Rightarrow C_{\bigtriangleup\bigtriangleup}\mbox{ cancels.}\\ \mbox{odd, if \ensuremath{m=2} mod 4} & \Rightarrow C_{\bigtriangleup\bigtriangleup}\mbox{ remains.} \end{array}\end{cases}$ }\tabularnewline
& \#$\circledast\bigtriangleup\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m-1}{1}$ is odd $\Rightarrow C_{\circledast\bigtriangleup\bigtriangleup}$ remains.}\tabularnewline
& & & & & \tabularnewline \textbf{3.} & \#$\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m}{3}=\frac{m(m-1)(m-2)}{2\cdot3}$ is even $\Rightarrow\; C_{\bigtriangleup}$ cancels.}\tabularnewline
& \#$\circledast\circledast\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m-2}{1}$ is even $\Rightarrow\; C_{\circledast\circledast\bigtriangleup} $ cancels.}\tabularnewline
& \#$\circledast\bigtriangleup$ & \multicolumn{4}{l}{$=\binom{m-1}{2}=\frac{(m-1)(m-2)}{2}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{odd, if \ensuremath{m=0} mod 4} & \Rightarrow C_{\circledast\bigtriangleup}\mbox{ remains.}\\ \mbox{even, if \ensuremath{m=2} mod 4} & \Rightarrow C_{\circledast\bigtriangleup}\mbox{ cancels.} \end{array}\end{cases}$}\tabularnewline
& & & & & \tabularnewline \textbf{4.} & \#$\circledast\circledast\circledast$ & \multicolumn{4}{l}{$=\binom{m-3}{1}$ is odd $\Rightarrow \; C_{\circledast\circledast\circledast}$ remains. }\tabularnewline
& \#$\circledast\circledast$ & \multicolumn{4}{l}{$=\binom{m-2}{2}=\frac{(m-2)(m-3)}{2}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{odd, if \ensuremath{m=0} mod 4} & \Rightarrow C_{\circledast\circledast}\mbox{ remains.}\\ \mbox{even, if \ensuremath{m=2} mod 4} & \Rightarrow C_{\circledast\circledast}\mbox{ cancels.} \end{array}\end{cases}$}\tabularnewline
& \#$\circledast$ & \multicolumn{4}{l}{$=\binom{m-1}{3}=\frac{(m-1)(m-2)(m-3)}{2\cdot3}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{odd, if \ensuremath{m=0} mod 4} & \Rightarrow C_{\circledast}\mbox{ remains.}\\ \mbox{even, if \ensuremath{m=2} mod 4} & \Rightarrow C_{\circledast}\mbox{ cancels.}\end{array}\end{cases}$ }\tabularnewline
& \#$\{\}$ & \multicolumn{4}{l}{$=\binom{m}{4}=\frac{m(m-1)(m-2)(m-3)}{2\cdot3\cdot4}\ $ affects the global sign ($GL$) and will be discussed separately.}\tabularnewline \end{tabular} \nopagebreak[9] \begin{center} \textbf{Table 6.} Counting phase gates for a four-uniform HG state, for even $m$. \end{center}
$\;$\\ \textbf{Remark:} Similarly to previous proofs the four-qubit phase gates cancel out. Therefore, we directly skip the discussion about them.
We need to consider two cases, when $m=0$ mod 4 and $m=2$ mod 4 for each $N=8k+2$ and $8k+4$ separately:\\
\textbf{Case \# 1:} If $m=0$ mod 4:\\ \begin{align}\label{part4_1} \begin{split} \langle G_4 \rangle & =\pm\bra{+}^{\otimes N} XX\dots XZ\dots Z \prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast \bigtriangleup \bigtriangleup}C_{\circledast\bigtriangleup} C_{\circledast \circledast \circledast } C_{\circledast \circledast }C_{\circledast} \ket{+}^{\otimes N}\\ & =\pm \frac{1}{2^N} Tr\Big[ \prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast \bigtriangleup \bigtriangleup}C_{\circledast\bigtriangleup} C_{\circledast \circledast \circledast } C_{\circledast \circledast }C_{\circledast}C_{\bigtriangleup} \Big]. \end{split} \end{align}
We use $(-1)^s$ to define the sign of the diagonal element and $s=\binom{\alpha}{1}\binom{\beta}{2}+\binom{\alpha}{1}\binom{\beta}{1}+\binom{\alpha}{3}+\binom{\alpha}{2}+\binom{\alpha}{1}+\binom{\beta}{1}$. If $s$ is even, the value on the diagonal is $+1$ and $-1$, otherwise. We consider all possible values of $\alpha$ and $\beta$:\\
\begin{tabular}{lllc} \textbf{1. } & $\alpha$ is even \& $\beta$ is even & \multicolumn{2}{l}{if $\alpha=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{2. } & $\alpha$ is even \& $\beta$ is odd: & \multicolumn{2}{l}{if $\alpha=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}\\
From here one can easily spot that for even $\alpha$, there is equal number of $+1$ and $- 1$ on the diagonal. So, they do not contribute in the calculations. We now consider the odd $\alpha$:\\
\begin{tabular}{lllc} \textbf{3. } & $\alpha$ is odd \& $\beta$ is even & \multicolumn{2}{l}{if $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & & \tabularnewline \textbf{4. } & $\alpha$ is odd \& $\beta$ is odd: & \multicolumn{2}{l}{if $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}\\
We now continue calculation of the trace from ($\ref{part4_1}$): \begin{align}{\label{part4_2}} \begin{split} \langle G_4 \rangle & =\pm \frac{1}{2^N} \sum_{\alpha=1,3,5\dots } \binom{m}{\alpha} \bigg[\sum_{\beta =0,4,\dots}^{N-m}-\binom{N-m}{\beta}-\binom{N-m}{\beta +1}+\binom{N-m}{\beta +2}+\binom{N-m}{\beta +3}\bigg] \\ & = \pm \frac{2^{m-1}}{2^N} \bigg[-Re \Big[(1+i)^{N-m} \Big] - Im\Big[(1+i)^{N-m}\Big]\bigg]=\pm \frac{2^{m-1}}{2^N} \bigg( \mp2^{\frac{N-m}{2}} \bigg). \end{split} \end{align}
We have to take care of the sign which appears from the product of the sign of the sum of real and imaginary part in ($\ref{part4_2}$) and global sign $(GL)$, which we defined while deriving the remaining phase gates. If $m$ is divisible by $8$, $GL=+1$ and since we are in $N=8k+2$ case, $(N-m=8k+2-m)-2$ mod $8$ and therefore: \begin{equation} \langle G_4 \rangle= \frac{2^{m-1}}{2^N} \bigg( -2^{\frac{N-m}{2}} \bigg) = -\frac{2^{m/2-1}}{2^{N/2 }}. \end{equation} If $m$ is not divisible by $8$, the global sign $GL=-1$, and the sum of real and imaginary part also contribute with a negative sign. Thus, \begin{equation} \langle G_4 \rangle= -\frac{2^{m-1}}{2^N} \bigg( -(-2^{\frac{N-m}{2}}) \bigg) = -\frac{2^{m/2-1}}{2^{N/2 }}. \end{equation}\\
Since the $N=8k+4$ case is identical, we only have to mind the sign of the sum of the real and imaginary part. Here as well we consider two cases: if $m$ is divisible by $8$, then the global sign $GL=+1$, and the sign of the sum of real and imaginary part is $"-"$. Therefore, \begin{equation} \langle G_4 \rangle=\frac{2^{m-1}}{2^N} \bigg( -(-2^{\frac{N-m}{2}}) \bigg) = +\frac{2^{m/2-1}}{2^{N/2 }}. \end{equation} And if $m$ is not divisible by $8$, $GL=-1$, and the sign of real and imaginary part is $"+"$. Therefore, \begin{equation} \langle G_4 \rangle= -\frac{2^{m-1}}{2^N} \bigg( -(+2^{\frac{N-m}{2}}) \bigg) = +\frac{2^{m/2-1}}{2^{N/2 }}. \end{equation}\\
\textbf{Case \# 2:} If $m=2$ mod 4: \begin{align}\label{part2ii} \begin{split} \langle G_4\rangle & =\pm\bra{+}^{\otimes N} XX\dots XZ\dots Z \prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast \bigtriangleup \bigtriangleup} C_{\circledast \circledast \circledast } C_{\bigtriangleup \bigtriangleup } \ket{+}^{\otimes N}\\ & =\pm \frac{1}{2^N} Tr( \prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast \bigtriangleup \bigtriangleup} C_{\circledast \circledast \circledast } C_{\bigtriangleup \bigtriangleup } C_{ \bigtriangleup } ). \end{split} \end{align}\\ We use $(-1)^s$ to define the sign of the diagonal element and $s=\binom{\alpha}{1}\binom{\beta}{2}+\binom{\alpha}{3}+\binom{\beta}{2}+\binom{\beta}{1}$. If $s$ is even, the value on the diagonal is $+1$ and $-1$, otherwise. We consider all possible values of $\alpha$ and $\beta$:\\
Considering the terms from odd $\alpha$:\\
\begin{tabular}{lllc} \textbf{1. } & $\alpha$ is odd \& $\beta$ is even & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{2. } & $\alpha$ is odd \& $\beta$ is odd: & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}
It is easy to see that these cases adds up to $0$. \\
\begin{tabular}{lllc} \textbf{3. } & $\alpha$ is even \& $\beta$ is even & \multicolumn{2}{l}{if $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{4. } & $\alpha$ is even \& $\beta$ is odd: & \multicolumn{2}{l}{if $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}\\
Therefore, \begin{align} \begin{split} \langle G_4\rangle=\pm\frac{1}{2^N}\sum_{\alpha=0,2,4\dots } \binom{m}{\alpha} \bigg[\sum_{\beta =0,4,\dots}^{N-m}\binom{N-m}{\beta} & -\binom{N-m}{\beta +1}-\binom{N-m}{\beta +2}+\binom{N-m}{\beta +3}\bigg] \\ =\pm 2^{m-1} \bigg[Re \Big[(1+i)^{N-m} & \Big] - Im\Big[(1+i)^{N-m}\Big]\bigg]=\pm\frac{2^{m/2-1}}{2^{N/2 }}. \end{split} \end{align}
To fix the sign, we need to first consider $N=8k+2$, and $(m-2)=0$ mod $8$. Then the global sign $GL=+1$ and type of $N-m$ also yields a positive sign. But if $(m-2)=4$ mod $8$, global sign in negative and the $N-m$ also yields the negative sign. So, \\ \begin{equation} \langle G_4 \rangle=\frac{2^{m/2-1}}{2^{N/2 }}. \end{equation}
$N=8k+4$ case is identical to $N=8k+2$, therefore we will just state the result. For $N=8k+4$
\begin{equation} \langle G_4 \rangle=-\frac{2^{m/2-1}}{2^{N/2 }}. \end{equation} To sum up, \begin{equation} \langle G_4\rangle =\begin{cases} \begin{array}[t]{cc} +\frac{2^{m/2-1}}{2^{N/2 }} & \mbox{if \ensuremath{(N-m)=0\ mod\ 4 }},\\ -\frac{2^{m/2-1}}{2^{N/2 }} & \mbox{if \ensuremath{(N-m)=2 \ mod\ 4 }}. \end{array}\end{cases} \end{equation}\\ This finishes the proof of part 3.\\
\textbf{Part 4: } We show that for $N=8k+3$, $\langle \underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}\mathbbm1}\rangle$ for even $m$ gives the same exact result as the part 3 (i).\\
We tackle the problem as follows: We make a measurement on one of the qubits in $Z$ direction and depending on the measurement outcome, we obtain the new $\ket{H_{4_{new}}^{M}}$ state, where $M:=N-1$. Then we consider the expectation values for the all possible measurement outcomes and $\ket{H_{4_{new}}^{M}}$. From that we conclude the statement in the part 4. \\
Initial HG state $\ket{H_4^N}$ can be written in the following form as well: \begin{equation} \ket{H_4^N}=\prod_e C_e\ket{+}^{\otimes N}=\prod_{e',e''}C_{e'}C_{e''}\ket{+}^{\otimes N}=\prod_{e',e''}\big[\mathbbm{1} _{e'\backslash N}\ket{0}_N\bra{0}_N + C_{e'\backslash N}\ket{1}_N\bra{1}_N \big]C_{e''}\ket{+}^{\otimes N}, \end{equation} where $e'$ represents the gates containing the last qubit, $N$ and $e''$ represents the ones which do not contain $N^{th}$ qubit. And $e=e'+e''$. Then if one makes a measurement in $Z$ basis on the last qubit and obtains outcome $+$, \begin{align} \begin{split} \ket{H_{4_{new}}^{M+}}=\braket{0_N}{H_4^N}&=\prod_{e',e''}\big[\bra{0}_N\mathbbm{1} _{e'\backslash N}\ket{0}_N\bra{0}_N+\bra{0}_NC_{e'\backslash N}\ket{1}_N\bra{1}_N \big]C_{e''}\ket{+}^{\otimes N}\\ &=\prod_{e',e''}\mathbbm{1} _{e'\backslash N}\bra{0}_NC_{e''}\ket{+}^{\otimes N}=\bra{0}_N\prod_{e',e''}\mathbbm{1} _{e'\backslash N}C_{e''}\ket{+}^{\otimes M}\big(\ket{0}_N+\ket{1}_N\big)\\ &=\prod_{e',e''}\mathbbm{1} _{e'\backslash N}C_{e''}\ket{+}^{\otimes M}=\prod_{e''}C_{e''}\ket{+}^{\otimes M}. \end{split}. \end{align} So, $+$ outcome after measuring in $Z$ direction leaves us with $\ket{H_{4_{new}}^{M+}}$, which is precisely four uniform $M$-qubit HG state. Now, let us see what is the remaining state if one gets $-$ as an outcome result: \begin{align} \begin{split} \ket{H_{4_{new}}^{M-}}=\braket{1_N}{H_4^N}&=\prod_{e',e''}\big[\bra{1}_N\mathbbm{1} _{e'\backslash N}\ket{0}_N\bra{0}_N+\bra{1}_NC_{e'\backslash N}\ket{1}_N\bra{1}_N \big]C_{e''}\ket{+}^{\otimes N}\\ &=\prod_{e',e''}C_{e'\backslash N}\bra{1}_N C_{e''}\ket{+}^{\otimes N}= \bra{1}_N \prod_{e',e''}C_{e'\backslash N}C_{e''}\ket{+}^{\otimes M}\big(\ket{0}_N+\ket{1}_N\big)\\ &=\prod_{e',e''}C_{e'\backslash N}C_{e''}\ket{+}^{\otimes M}. \end{split} \end{align} So, $-$ outcome after measuring in $Z$ direction leaves us with $\ket{H_{4_{new}}^{M-}}$, which is precisely a symmetric $M$-qubit HG state with all possible edges of cardinality four and three. We will call such HG state a three- and four-uniform HG state.\\
Therefore, problem boils down to showing that,\\ $(i)$ If the measurement outcome is $+$, we get the $M=8k+2$ four-uniform HG state and the correlations are given in part 3.\\ $(ii)$ If the measurement outcome is $-$, we get $M=8k+2$ three- and fouruniform HG state and the following holds: \begin{equation} \langle G_5^-\rangle=\bra{H_{4_{new}}^{M-}} \underset{m}{\underbrace{X\dots X}}Z\dots Z\ket{H_{4_{new}}^{M-}}=\begin{cases} \begin{array}[t]{cc} -\frac{2^{m/2-1}}{2^{M/2 }} & \mbox{if \ensuremath{(M-m)=0\ mod\ 4 }},\\ +\frac{2^{m/2-1}}{2^{M/2 }} & \mbox{if \ensuremath{(M-m)=2 \ mod\ 4 }}. \end{array}\end{cases} \end{equation}
$ (i)$ $\ket{H_{4_{new}}^{M+}}$, where $M=8k+2$ was already considered in part 3.\\
(ii). For $\ket{H_{4_{new}}^{M-}}$, \begin{equation} \langle G_5^-\rangle=\bra{+}^{\otimes M}\Big[\prod_{e',e''\in E} C_{e' }C_{e''}\Big]\underset{m}{\underbrace{X\dots X}}Z\dots Z\Big[\prod_{e',e''\in E} C_{e }C_{e''}\Big]\ket{+}^{\otimes M}. \end{equation} Before, we treated three- and four-uniform cases separately. Now, we just need to put them together. \\
\textbf{Case \# 1:} If $m=0$ mod 4:\\ Then from equations (\ref{appendixAeq6}) and (\ref{part4_1}), we can directly write down that \begin{equation} \langle G_5^-\rangle=\frac{1}{2^M}Tr\bigg[\prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast \bigtriangleup \bigtriangleup} C_{\circledast \circledast \circledast } C_{\circledast \circledast } C_{\bigtriangleup}\bigg]. \end{equation}
We check the sign of each term on the diagonal by $(-1)^s$, where $s=\binom{\alpha}{1}\binom{\beta}{2}+\binom{\alpha}{3}+\binom{\alpha}{2}+\binom{\beta}{1}$. For this we need to consider each value of $\alpha$ and $\beta$ separately.\\
\begin{tabular}{lllll} \multirow{2}{*}{\textbf{1. }} & \multirow{2}{*}{$\alpha$ is even \& $\beta$ is even} & $\alpha=0$ mod 4 $\Rightarrow$ $+1$ & & \multirow{5}{*}{$\Rightarrow$ These two give zero contribution together.}\tabularnewline
& & $\alpha=2$ mod 4 $\Rightarrow$ $-1$ & & \tabularnewline
& & & & \tabularnewline \multirow{2}{*}{\textbf{2.}} & \multirow{2}{*}{$\alpha$ is even \& $\beta$ is odd} & $\alpha=0$ mod 4 $\Rightarrow$ $-1$ & & \tabularnewline
& & $\alpha=2$ mod 4 $\Rightarrow$ $+1$ & & \tabularnewline \end{tabular}\\ \\
\begin{tabular}{lllll} \multirow{2}{*}{\textbf{3. }} & \multirow{2}{*}{$\alpha$ is odd \& $\beta$ is even} & $\beta=0$ mod 4 $\Rightarrow$ $+1$ & & \multirow{5}{*}{}\tabularnewline
& & $\beta=2$ mod 4 $\Rightarrow$ $-1$ & & \tabularnewline
& & & & \tabularnewline \multirow{2}{*}{\textbf{4.}} & \multirow{2}{*}{$\alpha$ is odd \& $\beta$ is odd} & $\beta-1=0$ mod 4 $\Rightarrow$ $-1$& & \tabularnewline
& & $\beta-1=2$ mod 4 $\Rightarrow$ $+1$ & & \tabularnewline \end{tabular}\\
\begin{align}\label{part5_1} \begin{split} \langle G_5^- \rangle & =\pm\frac{1}{2^{M}}\sum_{\alpha\ odd}^{m}\binom{m}{\alpha}\bigg[\sum_{\beta=0,4}^{M-m}\binom{M-m}{\beta}-\binom{M-m}{\beta+1}-\binom{M-m}{\beta+2}+\binom{M-m}{\beta+3}\bigg]\\
&=\pm\frac{2^{m-1}}{2^{M}}\bigg[Re(1+i)^{M-m}-Im(1+i)^{M-m}\bigg]=\pm\frac{2^{\frac{m}{2}-1}}{2^{M/2}}. \end{split} \end{align}
If $m=0$ mod 8, real and imaginary part in (\ref{part5_1}) has a negative sign and the global sign coming from Table 6, $GL$ is positive. Note from equation (\ref{3unifGL}) that three uniform gate moving does not introduce any global signs. And if $m=4$ mod 8, real and imaginary part in (\ref{part5_1}) has a positive sign and the global sign coming from Table 6, $GL$ is negative.Therefore, \begin{equation} \langle G_5^+ \rangle =-\frac{2^{\frac{m}{2}-1}}{2^{M/2}}. \end{equation}
\textbf{Case \# 2:} If $m=2$ mod 4:\\ Then from equations (\ref{appendixAeq8}) and (\ref{part2ii}), we can directly write down that
\begin{equation}
\langle G_5^-\rangle=\pm\frac{1}{2^{M}}Tr\bigg[\prod_{\forall \circledast, \forall\bigtriangleup } C_{\circledast\triangle\triangle}C_{\circledast\triangle}C_{\circledast\circledast\circledast}C_{\triangle\triangle}\bigg].
\end{equation}
We check the sign of each term on the diagonal by $(-1)^s$, where $s=\binom{\alpha}{1}\binom{\beta}{2}+\binom{\alpha}{3}+\binom{\beta}{2}+\binom{\alpha}{1}\binom{\beta}{1}$. For this we need to consider each value of $\alpha$ and $\beta$ separately.\\
\begin{tabular}{lllll} \multirow{2}{*}{\textbf{1. }} & \multirow{2}{*}{$\alpha$ is even \& $\beta$ is even} & $\beta=0$ mod 4 $\Rightarrow$ $+1$ & & \multirow{5}{*}{}\tabularnewline
& & $\beta=2$ mod 4 $\Rightarrow$ $-1$ & & \tabularnewline
& & & & \tabularnewline \multirow{2}{*}{\textbf{2.}} & \multirow{2}{*}{$\alpha$ is even \& $\beta$ is odd} & $\beta-1=0$ mod 4 $\Rightarrow$ $+1$ & & \tabularnewline
& & $\beta-1=2$ mod 4 $\Rightarrow$ $-1$ & & \tabularnewline \end{tabular}\\ $\quad$\\
\begin{tabular}{lllll} \multirow{2}{*}{\textbf{3. }} & \multirow{2}{*}{$\alpha$ is odd \& $\beta$ is even} & & $\alpha-1=0$ mod 4 $\Rightarrow$ $+1$ & \multirow{4}{*}{$\Rightarrow$ These two give zero contribution together.}\tabularnewline
& & & $\alpha-1=2$ mod 4 $\Rightarrow$ $-1$ & \tabularnewline \multirow{2}{*}{\textbf{4.}} & \multirow{2}{*}{$\alpha$ is odd \& $\beta$ is odd} & & $\alpha-1=0$ mod 4 $\Rightarrow$ $-1$ & \tabularnewline
& & & $\alpha-1=2$ mod 4 $\Rightarrow$ $+1$ & \tabularnewline \end{tabular}
\begin{align}\label{part5_2} \begin{split} \langle G_{5}^-\rangle & = \pm\frac{1}{2^{M}}\sum_{\alpha\ even}^{m}\binom{m}{\alpha}\bigg[\sum_{\beta=0,4}^{M-m}\binom{M-m}{\beta}+\binom{M-m}{\beta+1}-\binom{M-m}{\beta+2}-\binom{M-m}{\beta+3}\bigg]\\ & = \pm\frac{1}{2^{M}} 2^{m-1}\bigg[Re(1+i)^{M-m}+Im(1+i)^{M-m}\bigg]=\pm\frac{2^{\frac{m}{2}-1}}{2^{M/2}}. \end{split} \end{align}
If $m-2=0$ mod 8, real and imaginary part in (\ref{part5_2}) has a positive sign and the global sign coming from Table 6, $GL$ is positive. Note from equation (\ref{3unifGL}) that the three uniform gate moving does not introduce any global signs. And if $m-2=4$ mod 8, real and imaginary part in (\ref{part5_2}) has a negative sign and the global sign coming from Table 6, $GL$ is negative.Therefore,
\begin{equation} \langle G_5^- \rangle =+\frac{2^{\frac{m}{2}-1}}{2^{M/2}}. \end{equation} Finally, we can put everything together. Since one can observe that $\langle G_5^- \rangle=-\langle G_5^+ \rangle$, \begin{equation} \ket{0}\bra{0}\langle G_5^+ \rangle-\ket{1}\bra{1}\langle G_5^- \rangle=\ket{0}\bra{0}\langle G_5^+ \rangle+\ket{1}\bra{1}\langle G_5^+\rangle =\mathbbm{1}\langle G_5^+\rangle= \langle \underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1\rangle. \end{equation} This completes the proof of part 4 and entire lemma.
\end{proof}
\section{Appendix E: Bell inequality violations for fully-connected four-uniform hypergraph states}
\begin{lemma}\label{theorem4unifviol} An arbitrary four-uniform HG state violates the classical bound by an amount that grows exponentially along with number of qubits. \end{lemma} \begin{proof}
To show this, either in Eq.~(\ref{bell2}) or in the original Mermin operator, $\langle \ensuremath{\mathcal{B}}_N^M \rangle $, we fix $A=Z$ and $B=X$. The choice of the Bell
operator depends on the number of qubits: From Lemma~(\ref{lemma4full})
for a given $N$ either the correlations for an even $m$ or an odd $m$ are
given. If $m$ is even, we choose Eq.~(\ref{bell2}) and
$\langle \ensuremath{\mathcal{B}}_N^M \rangle $, otherwise.
From Lemma \ref{lemma4full}, it is clear that we need to consider separate cases. However, we choose the one which encounters the smallest growth in violation and this is the $N=8k+3$ case. Other cases just encounter different factors in the growth or are greater. For the $N=8k+3$, the strategy consists of measuring
the Bell operator from Eq.~(\ref{bell2}) on $M=N-1$ qubits. Then we have: \begin{align} \begin{split} \langle \ensuremath{\mathcal{B}}_N \rangle _Q & \geq\sum_{m=2,4\dots}^{M}\binom{M}{m}\Big(\frac{1}{\sqrt{2}}\Big)^{M-m+2} = \frac{1}{2}\bigg[\sum_{m\ even}^{M} \binom{M}{m} \Big(\frac{1}{\sqrt{2}}\Big)^{M-m}\bigg]-\Big(\frac{1}{\sqrt{2}}\Big)^{M+2}\\ & =\frac{1}{4}\bigg[ \Big(1+\frac{1}{\sqrt{2}} \Big)^M -\Big(1-\frac{1}{\sqrt{2}} \Big)^M\bigg]-\Big(\frac{1}{\sqrt{2}}\Big)^{M+2}. \end{split} \end{align} Checking the ratio of the quantum and classical values, we have that \begin{align} \begin{split} \frac{\langle \ensuremath{\mathcal{B}}_N \rangle _Q}{\langle \ensuremath{\mathcal{B}}_N \rangle _C} \stackrel{N\rightarrow \infty}{\sim} \frac{\frac{1}{4}\Big(1+\frac{1}{\sqrt{2}} \Big)^{N-1} }{2^{\frac{N-1}{2}}} = \frac{\Big(1+\frac{1}{\sqrt{2}}\Big)^{N-1} }{\sqrt{2}^{N+3}} \approx \frac{1.20711^{N}}{2\sqrt{2} +2} . \end{split} \end{align} Looking at the expectation values from Lemma \ref{lemma4full}, it is straightforward to see that in all other cases of $N$, correlations are stronger than in the $N=8k+3$ case, so the quantum violation increases. \end{proof}
\section{Appendix F: Bell and separability inequality violations for fully-connected four-uniform hypergraph states after loosing one qubit} \begin{lemma} The following statement holds for $N=8k+4$ qubit, four-uniform complete hypergraph states: \begin{equation}\label{reducedlemma} \langle G_6 \rangle = \langle \underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1\rangle=\begin{cases} \begin{array}[t]{cc} -\Big(\frac{1}{\sqrt{2}}\Big)^{N-m+2} & \mbox{if \ensuremath{m=0\ mod\ 4 }},\\ +\Big(\frac{1}{\sqrt{2}}\Big)^{N-m+2} & \mbox{if \ensuremath{m=2 \ mod\ 4 }}. \end{array}\end{cases} \end{equation} \end{lemma} \begin{proof} The derivation of this result is very similar to the combinatorial calculations in the appendices D and E. Since $m$ is even, we refer to the Table 6 to see what gates remain after regrouping hyperedges on the right hand side of the expression (\ref{reducedlemma}):
\textbf{Case \# 1:} If $m=0$ mod 4:\\ \begin{align}\label{part4_1} \begin{split} \langle G_6 \rangle & =\pm\bra{+}^{\otimes N} \underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1 \prod_{\forall \circledast,\bigtriangleup,\diamondsuit } C_{\circledast \bigtriangleup \bigtriangleup}C_{\circledast\bigtriangleup} C_{\circledast \bigtriangleup \diamondsuit}C_{\circledast\diamondsuit} C_{\circledast \circledast \circledast } C_{\circledast \circledast }C_{\circledast} \ket{+}^{\otimes N}\\ & =\pm \frac{1}{2^N} Tr\Big[\prod_{\forall \circledast,\bigtriangleup,\diamondsuit } C_{\circledast \bigtriangleup \bigtriangleup}C_{\circledast\bigtriangleup} C_{\circledast \bigtriangleup \diamondsuit}C_{\circledast\diamondsuit} C_{\circledast \circledast \circledast }C_{\circledast \circledast }C_{\circledast}C_{\bigtriangleup} \Big]. \end{split} \end{align}
Here $\circledast$ again refers to X operator, $\bigtriangleup$ to Z and $\diamondsuit$ to $\mathbbm1$ and it is denoted as $\gamma$. The strategy now is similar to the previous proofs: count the number of +1's and -1's on the diagonal and then their difference divided by $2^N$ gives the trace.
We use $(-1)^s$ to define the sign of the diagonal element and $s=\binom{\alpha}{1}\binom{\beta}{2}+\binom{\alpha}{1}\binom{\beta}{1}+\binom{\alpha}{1}\binom{\beta}{1}\binom{\gamma}{1}+\binom{\alpha}{1}\binom{\gamma}{1}+\binom{\alpha}{3}+\binom{\alpha}{2}+\binom{\alpha}{1}+\binom{\beta}{1}$. If $s$ is even, the value on the diagonal is $+1$ and $-1$, otherwise. We consider all possible values of $\alpha$, $\beta$, and $\gamma$:\\
a) If $\gamma$ is even (that is $\gamma=0$):\\
\begin{tabular}{lllc} \textbf{1. } & $\alpha$ is even \& $\beta$ is even & \multicolumn{2}{l}{if $\alpha=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{2. } & $\alpha$ is even \& $\beta$ is odd: & \multicolumn{2}{l}{if $\alpha=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}\\
From here one can easily spot that for even $\alpha$, there is equal number of $+1$ and $- 1$ on the diagonal. So, they do not contribute in the calculations. We now consider the odd $\alpha$:\\
\begin{tabular}{lllc} \textbf{3. } & $\alpha$ is odd \& $\beta$ is even & \multicolumn{2}{l}{if $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & & \tabularnewline \textbf{4. } & $\alpha$ is odd \& $\beta$ is odd: & \multicolumn{2}{l}{if $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}\\
b) If $\gamma$ is odd (that is $\gamma=1$):\\
The cases \textbf{1} and \textbf{2} don't change. Therefore, they sum up to 0. \\
\begin{tabular}{lllc} \textbf{3. } & $\alpha$ is odd \& $\beta$ is even & \multicolumn{2}{l}{if $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \end{tabular}\\
\textbf{4. } Stays the same as in the previous case.
Therefore the result is: \begin{align}\label{F_1} \begin{split} \langle G_{6}\rangle & = \pm\frac{1}{2^{N}}\Big(\sum_{\gamma\; even}\binom{\gamma}{0}\Bigg[\sum_{\alpha\ odd}^{m}\binom{m}{\alpha}\bigg[\sum_{\beta=0,4\dots}^{N-m-1}-\binom{N-m-1}{\beta}-\binom{N-m-1}{\beta+1}+\binom{N-m-1}{\beta+2}+\binom{N-m-1}{\beta+3}\bigg]\Bigg]\\ &+\sum_{\gamma\; odd}\binom{\gamma}{1}\Bigg[\sum_{\alpha\ odd}^{m}\binom{m}{\alpha}\bigg[\sum_{\beta=0,4\dots}^{N-m-1}\binom{N-m-1}{\beta}-\binom{N-m-1}{\beta+1}-\binom{N-m-1}{\beta+2}+\binom{N-m-1}{\beta+3}\bigg]\Bigg]\Bigg)\\ &=\pm \frac{2^m}{2^N}\cdot (-Im[(1+i)^{N-m-1}])= \mp \bigg(\frac{1}{\sqrt{2}}\bigg)^{N-m+2} \end{split} \end{align}
It is time to fix a sign. One needs to keep in mind that the sign of the Eq. (\ref{F_1}) is negative: if $m=4$ mod $8$, the global sign from the Table 6 is negative, and $Im[(1+i)^{N-m-1}]=-2^\frac{N-m-2}{2}$. Therefore, an overall sign in negative. If $m=0$ mod $8$, global sign is positive and $Im[(1+i)^{N-m-1}]=2^\frac{N-m-2}{2}$. Therefore, an overall sign in negative.
\pagebreak \textbf{Case \# 2:} If $m=2$ mod 4:\\ \begin{align}\label{part2ii} \begin{split} \langle G_6\rangle & =\pm\bra{+}^{\otimes N} \underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1 \prod_{\forall \circledast,\bigtriangleup,\diamondsuit } C_{\circledast \bigtriangleup \bigtriangleup} C_{\circledast \circledast \circledast } C_{\bigtriangleup \bigtriangleup } C_{\circledast \bigtriangleup \diamondsuit} C_{\bigtriangleup \diamondsuit }\ket{+}^{\otimes N}\\ & =\pm \frac{1}{2^N} Tr( \prod_{\forall \circledast,\bigtriangleup,\diamondsuit } C_{\circledast \bigtriangleup \bigtriangleup} C_{\circledast \circledast \circledast } C_{\bigtriangleup \bigtriangleup } C_{\circledast \bigtriangleup \diamondsuit} C_{\bigtriangleup \diamondsuit } C_{ \bigtriangleup } ). \end{split} \end{align}\\
We use $(-1)^s$ to define the sign of the diagonal element and $s=\binom{\alpha}{1}\binom{\beta}{2}+\binom{\alpha}{1}\binom{\beta}{1}\binom{\gamma}{1}+\binom{\alpha}{3}+\binom{\beta}{2}+\binom{\beta}{1}\binom{\gamma}{1}+\binom{\beta}{1}$. If $s$ is even, the value on the diagonal is $+1$ and $-1$, otherwise. We consider all possible values of $\alpha$, $\beta$, and $\gamma$:\\
a) If $\gamma$ is even (that is $\gamma=0$):
Considering the terms from even $\alpha$:\\
\begin{tabular}{lllc} \textbf{1. } & $\alpha$ is even \& $\beta$ is even & \multicolumn{2}{l}{if $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{2. } & $\alpha$ is even \& $\beta$ is odd: & \multicolumn{2}{l}{if $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}
Considering odd $\alpha$:
\begin{tabular}{lllc} \textbf{3. } & $\alpha$ is odd \& $\beta$ is even & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{4. } & $\alpha$ is odd \& $\beta$ is odd: & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline \end{tabular}\\ It is easy to see that cases \textbf{3} and \textbf{4} adds up to $0$. \\
b) If $\gamma$ is odd (that is $\gamma=1$):\\
\textbf{1.} Stays the same as in the previous case. \\
\textbf{2.} Gets an opposite sign, therefore, will cancel out with the a) case \textbf{2} in the sum. \\
\textbf{3. } Stays the same as in the previous case.\\
\textbf{4. } Stays the same as in the previous case.
Therefore the result is: \begin{align}\label{part5_2} \begin{split} \langle G_{6}\rangle & = \pm\frac{1}{2^{N}} 2^m\bigg[\sum_{\beta=0,4..}^{N-m-1}\binom{N-m-1}{\beta}-\binom{N-m-1}{\beta+2}\bigg]=\pm\frac{2^m}{2^N}Re[(1+i)^{N-m-1}]. \end{split} \end{align}
It is time to fix a sign: if $m=2$ mod $8$, the global sign is positive from the Table 6, and $Re[(1+i)^{N-m-1}]=2^\frac{N-m-2}{2}$. Therefore, an overall sign is positive. If $m=6$ mod $8$, overall sign in negative and $Re[(1+i)^{N-m-1}]=-2^\frac{N-m-2}{2}$. Therefore, an overall sign in positive. \end{proof}
\begin{lemma}\label{theorem4unifviol} A $N$-qubit ($N=8k+4$) four-uniform HG state even after tracing out one party violates the classical bound by an amount that grows exponentially with number of qubits. Moreover, the violation only decreases with the constant factor. \end{lemma} \begin{proof}
Denote $M\equiv N-1$. Then \begin{align} \begin{split} \langle \ensuremath{\mathcal{B}}_N \rangle _Q & =\sum_{m=2,4\dots}^{M}\binom{M}{m}\Big(\frac{1}{\sqrt{2}}\Big)^{M-m+3}= \frac{1}{2\sqrt{2}}\bigg[\sum_{m\ even}^{M} \binom{M}{m} \Big(\frac{1}{\sqrt{2}}\Big)^{M-m}\bigg]-\Big(\frac{1}{\sqrt{2}}\Big)^{M+3}\\ & =\frac{1}{4\sqrt{2}}\bigg[ \Big(1+\frac{1}{\sqrt{2}} \Big)^M -\Big(1-\frac{1}{\sqrt{2}} \Big)^M\bigg]-\Big(\frac{1}{\sqrt{2}}\Big)^{M+3} \end{split} \end{align}
Checking the ratio of the quantum and classical values, we have that \begin{align} \begin{split} \frac{\langle \ensuremath{\mathcal{B}}_{N-1} \rangle _Q}{\langle \ensuremath{\mathcal{B}}_{N-1} \rangle _C} \stackrel{N\rightarrow \infty}{\sim} \frac{\frac{1}{4\sqrt{2}}\Big(1+\frac{1}{\sqrt{2}} \Big)^{N-1} }{2^{\frac{N-2}{2}}} = \frac{\Big(1+\frac{1}{\sqrt{2}}\Big)^{N-1} }{\sqrt{2}^{N+3}} \approx \frac{1.20711^{N}}{2\sqrt{2} +2} . \end{split} \end{align} While the same value for the $N=8k+4$ qubit four-uniform complete HG state is $\frac{\langle \ensuremath{\mathcal{B}}_{N} \rangle _Q}{\langle \ensuremath{\mathcal{B}}_{N} \rangle _C} \stackrel{N\rightarrow \infty}{\sim}\approx\frac{1.20711^{N}}{4}.$ Therefore, after tracing out a single qubit, the local realism violation decreases with the small constant factor. \end{proof}
It is important to note that the similar violation is maintained after tracing out more than one qubit. For example, numerical evidence confirms for $N=12$, that if one takes a Bell inequality with the odd number of $X$ measurements, instead of the even ones as we have chosen in the proof, an exponential violation is maintained after tracing out 2 qubits. But even if 5 qubit is traced out, the state is stilled entangled and this can be verified using the separability inequality \cite{Roy}. Exact violation is given in the following table:
\begin{center}
\begin{tabular}{|c|c|c|c|c|}
\hline \#k & Quantum Value & Classical Bound & Separability Bound & $\approx$Ratio\tabularnewline \hline \hline \textbf{0} & \textbf{153.141} & \textbf{64} & & \textbf{2.39283}\tabularnewline \hline \textcolor{red}{$1$} & \textcolor{red}{89.7188} & \textcolor{red}{32} & & \textcolor{red}{2.78125}\tabularnewline \hline $2$ & 37.1563 & 32 & & 1.16113\tabularnewline \hline \multicolumn{1}{c}{} & \multicolumn{1}{c}{} & \multicolumn{1}{c}{} & \multicolumn{1}{c}{} & \multicolumn{1}{c}{}\tabularnewline \hline $3$ & 15.4219 & - & $\sqrt{2}$ & 10.9049\tabularnewline \hline $4$ & 6.375 & - & $\sqrt{2}$ & 4.50781\tabularnewline \hline $5$ & 2.70313 & - & $\sqrt{2}$ & 1.9114\tabularnewline \hline \end{tabular}\\ $\quad$\\ \textbf{Table 7.} Violation of Bell (for odd $m$) and Separability inequalities in $N=12$ qubit 4-uniform HG state. Here k is the number of traced out qubits. Red line represents that when k = 1, or equivalently one qubit is traced out, the violation of Bell inequalities increases. This is caused by decrease in the classical bound \cite{Mermin}. \end{center}
\section{Appendix G: Separability inequality violations for fully-connected three-uniform hypergraph states after loosing one qubit} \begin{lemma} The following statements holds three-uniform complete hypergraph states:\\ (i) For $N=8k+5$, $N=8k+6$, or $N=8k+7$: \begin{equation}\label{G_1} \langle G_7 \rangle =\langle\underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1\rangle=\begin{cases} \begin{array}[t]{cc} -\frac{1}{2^{\left\lfloor\frac{N-1}{2}\right\rfloor}} & \mbox{if \ensuremath{(m-1)=0\ mod\ 4 }},\\ +\frac{1}{2^{\left\lfloor\frac{N-1}{2}\right\rfloor}} & \mbox{if \ensuremath{(m-1)=2 \ mod\ 4 }}. \end{array}\end{cases} \end{equation} (ii) For $N=8k+1$, $N=8k+2$, or $N=8k+3$: \begin{equation}\label{G_2} \langle G_7 \rangle =\langle\underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1\rangle=\begin{cases} \begin{array}[t]{cc} +\frac{1}{2^{\left\lfloor\frac{N-1}{2}\right\rfloor}} & \mbox{if \ensuremath{(m-1)=0\ mod\ 4 }},\\ -\frac{1}{2^{\left\lfloor\frac{N-1}{2}\right\rfloor}} & \mbox{if \ensuremath{(m-1)=2 \ mod\ 4 }}. \end{array}\end{cases} \end{equation} (iii) For $N=4k$: \begin{equation}\label{G_3} \langle G_7 \rangle =\langle\underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1\rangle=0 \end{equation} \end{lemma}
\begin{proof} As a starting point, we derive the remaining gates on the right hand side of the equations (\ref{G_1}), the same derivation turns out to be working for Eq. (\ref{G_2}) and (\ref{G_3}). This approach is analogous to the Appendix B, but now, $m$ is odd.
\begin{tabular}{lrl}
& & \tabularnewline $1.$ & $\# \circledast\circledast=$ & $\binom{m-2}{1}$ is odd $\Rightarrow \;C_{\circledast\circledast}$ remains.\tabularnewline
& & \tabularnewline $2.$ & $\#\circledast=$ & $\binom{m-1}{2}=\frac{(m-1)(m-2)}{2}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{even, if (\ensuremath{m-1)=0} mod 4 } &\Rightarrow C_{\circledast}\mbox{ cancels.}\\ \mbox{odd, if (\ensuremath{m-1)=2} mod 4 } & \Rightarrow C_{\circledast}\mbox{ remains.} \end{array}\end{cases}$ \tabularnewline
& & \tabularnewline $3.$ & $ \#\{\}=$ & $\binom{m}{3}=\frac{m(m-1)(m-2)}{2\cdot3}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{even, if (\ensuremath{m-1)=0} mod 4 } &\Rightarrow\mbox{ gives a positive sign.}\\ \mbox{odd, if (\ensuremath{m-1)=2} mod 4 } & \Rightarrow\mbox{ gives a negative sign.} \end{array}\end{cases}$ \tabularnewline
& & \tabularnewline $4.$ & $\#\bigtriangleup\bigtriangleup=$ & $\binom{m}{1}$ is odd $\Rightarrow \; C_{\bigtriangleup\bigtriangleup}$ remains. \tabularnewline $5.$ & $\#\circledast\bigtriangleup=$ & $\binom{m-1}{1}$ is even $\Rightarrow \; C_{\circledast\bigtriangleup}$ cancels. \tabularnewline $6.$ & $ \# \bigtriangleup=$ & $\binom{m}{2}=\frac{m(m-1)}{2\cdot3}$ is $\begin{cases} \begin{array}[t]{cc} \mbox{even, if (\ensuremath{m-1)=0} mod 4 } &\Rightarrow C_{\bigtriangleup}\mbox{ cancels.}\\ \mbox{odd, if (\ensuremath{m-1)=2} mod 4 } & \Rightarrow C_{\bigtriangleup}\mbox{ remains.} \end{array}\end{cases}$ \tabularnewline
& & \tabularnewline
\end{tabular} \begin{center} \textbf{Table 8.} Counting phase gates for a three-uniform HG when $m$ ($m$ is odd) systems are measured in $X$ direction. \end{center}
Consider two cases:\\
\textbf{1.} If $(m-1)=0$ mod $4$: \begin{align}\label{part4_1} \begin{split} \langle G_7 \rangle & =\pm\bra{+}^{\otimes N} \underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1 \prod_{\forall \circledast,\bigtriangleup,\diamondsuit } C_{\circledast\circledast}C_{\bigtriangleup\bigtriangleup} C_{\bigtriangleup\diamondsuit} \ket{+}^{\otimes N}\\ & =\pm \frac{1}{2^N} Tr\Big[C_{\circledast\circledast}C_{\bigtriangleup\bigtriangleup} C_{\bigtriangleup\diamondsuit} C_{\bigtriangleup} \Big]. \end{split} \end{align}
Here $\circledast$ again refers to X operator, $\bigtriangleup$ to Z and $\diamondsuit$ to $\mathbbm1$ and is denoted by $\gamma$. The strategy is similar to the previous case: count the number of +1's and -1's on the diagonal. Their difference divided by $2^N$, gives the trace. \\
We use $(-1)^s$ to define the sign of the diagonal element and $s=\binom{\alpha}{2}+\binom{\beta}{2}+\binom{\beta}{1}\binom{\gamma}{1}+\binom{\beta}{1}$. If $s$ is even, the value on the diagonal is $+1$ and $-1$, otherwise. We consider all possible values of $\alpha$, $\beta$ and $\gamma$:\\
a) If $\gamma$ is even (that is $\gamma=0$ ):\\
Considering the terms from even $\beta$:\\
\begin{tabular}{lllc} \textbf{1. } & $\beta$ is even \& $\alpha$ is even & \multicolumn{2}{l}{if $\alpha=0$ mod 4 and $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=0$ mod 4 and $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 and $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 and $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & & \tabularnewline \textbf{2. } & $\beta$ is even \& $\alpha$ is odd & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 and $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 and $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 and $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 and $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & & \tabularnewline \end{tabular}\\
Considering odd $\beta$:\\
\begin{tabular}{lllc} \textbf{3. } & $\beta$ is odd \& $\alpha$ is even & \multicolumn{2}{l}{if $\alpha=0$ mod 4 and $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=0$ mod 4 and $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 and $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 and $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \textbf{4. } & $\beta$ is odd \& $\alpha$ is odd & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 and $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 and $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 and $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 and $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \end{tabular}
b) If $\gamma$ is odd (that is $\gamma=1$):\\
Considering the terms from even $\beta$:\\
\textbf{1.} and \textbf{2.} Nothing changes in comparison to a) \textbf{1.} and \textbf{2.}\\
\textbf{3.} and \textbf{4.} These two terms have opposite sign from a) \textbf{3.} and \textbf{4.} Therefore, in the sum they cancel ($\gamma$ is always 1.)\\
Therefore, \begin{align} \begin{split} \langle G_{7}\rangle & = \frac{1}{2^{N-1}}\bigg[\sum_{\beta=0,4..}^{N-m-1}\binom{N-m-1}{\beta}-\binom{N-m-1}{\beta+2}\bigg]\bigg[\sum_{\alpha=0,4..}^{m}\binom{m}{\alpha}+\binom{m}{\alpha+1}-\binom{m}{\alpha+2}-\binom{m}{\alpha+3}\bigg]\\ &= \frac{1}{2^{N-1}}Re[(1+i)^{N-m-1}]\cdot\Big(Re[(1+i)^m]+Im[(1+i)^m]\Big). \end{split} \end{align}
Now we can consider each cases separately, for this we use the lookup \textbf{table 0}:\\
$(i)$ If $N=8k+5$ and if $(m-1)=4$ mod $8$, $Re[(1+i)^{N-m-1}]=-2^\frac{N-m-2}{2}$ and $Re[(1+i)^m]+Im[(1+i)^m]=2^\frac{m+1}{2}$. Therefore, \begin{equation} \langle G_{7}\rangle =-\frac{2^\frac{N-m-2}{2}\cdot 2^\frac{m+1}{2} }{2^{N-1}}=-\Big(\frac{1}{2}\Big)^{\frac{N-1}{2}}. \end{equation} And, if $(m-1)=0$ mod $8$, $Re[(1+i)^{N-m-1}]=+2^\frac{N-m-2}{2}$ and $Re[(1+i)^m]+Im[(1+i)^m]=-2^\frac{m+1}{2}$. Therefore, \begin{equation} \langle G_{7}\rangle =-\frac{2^\frac{N-m-2}{2}\cdot 2^\frac{m+1}{2} }{2^{N-1}}=-\Big(\frac{1}{2}\Big)^{\frac{N-1}{2}}. \end{equation} Exactly the same is true for $N=8k+7$. But for $N=8k+6$ and if $(m-1)=4$ mod $8$, $Re[(1+i)^{N-m-1}]=-2^\frac{N-m-1}{2}$ and $Re[(1+i)^m]+Im[(1+i)^m]=2^\frac{m+1}{2}$. Therefore, \begin{equation} \langle G_{7}\rangle =-\frac{2^\frac{N-m-1}{2}\cdot 2^\frac{m+1}{2} }{2^{N-1}}=-\Big(\frac{1}{2}\Big)^{\left\lfloor\frac{N-1}{2}\right\rfloor}. \end{equation} And, if $(m-1)=0$ mod $8$, $Re[(1+i)^{N-m-1}]=2^\frac{N-m-1}{2}$ and $Re[(1+i)^m]+Im[(1+i)^m]=-2^\frac{m+1}{2}$. Therefore, \begin{equation} \langle G_{7}\rangle =-\frac{2^\frac{N-m-1}{2}\cdot 2^\frac{m+1}{2} }{2^{N-1}}=-\Big(\frac{1}{2}\Big)^{\left\lfloor\frac{N-1}{2}\right\rfloor}. \end{equation}\\
$(ii)$ If $N=8k+1$ and if $(m-1)=4$ mod $8$, $Re[(1+i)^{N-m-1}]=-2^\frac{N-m-2}{2}$ and $Re[(1+i)^m]+Im[(1+i)^m]=-2^\frac{m+1}{2}$. Therefore, \begin{equation} \langle G_{7}\rangle =\frac{2^\frac{N-m-2}{2}\cdot 2^\frac{m+1}{2} }{2^{N-1}}=\Big(\frac{1}{2}\Big)^{\frac{N-1}{2}}. \end{equation} And, if $(m-1)=0$ mod $8$, $Re[(1+i)^{N-m-1}]=2^\frac{N-m-1}{2}$ and $Re[(1+i)^m]+Im[(1+i)^m]=2^\frac{m+1}{2}$. Therefore, \begin{equation} \langle G_{7}\rangle =\frac{2^\frac{N-m-1}{2}\cdot 2^\frac{m+1}{2} }{2^{N-1}}=\Big(\frac{1}{2}\Big)^{\left\lfloor\frac{N-1}{2}\right\rfloor}. \end{equation} It it analogous for other two cases as well.\\
$(iii)$ For $N=4k$, $Re[(1+i)^{N-m-1}]=0$. Therefore, $\langle G_{7}\rangle =0.$ \\
\textbf{2.} If $(m-1)=2$ mod $4$: \begin{align}\label{part4_1} \begin{split} \langle G_7 \rangle & =\bra{+}^{\otimes N} \underset{m}{\underbrace{X\dots X}}\underset{N-m-1}{\underbrace{Z\dots Z}}\mathbbm1 \prod_{\forall \circledast,\bigtriangleup,\diamondsuit } C_{\circledast\circledast}C_{\circledast}C_{\bigtriangleup\bigtriangleup} C_{\bigtriangleup\diamondsuit}C_{\bigtriangleup}C_{\diamondsuit} \ket{+}^{\otimes N}\\ & = \frac{1}{2^N} Tr\Big[C_{\circledast\circledast}C_{\circledast}C_{\bigtriangleup\bigtriangleup} C_{\bigtriangleup\diamondsuit}C_{\diamondsuit} \Big]. \end{split} \end{align}
Here $\circledast$ again refers to X operator, $\bigtriangleup$ to Z and $\diamondsuit$ to $\mathbbm1$ and is denoted by $\gamma$. The strategy is similar to the previous case: count the number of +1's and -1's on the diagonal and their difference divided by $2^N$, gives the trace. \\
We use $(-1)^s$ to define the sign of the diagonal element and $s=\binom{\alpha}{2}+\binom{\alpha}{1}+\binom{\beta}{2}+\binom{\beta}{1}\binom{\gamma}{1}+\binom{\gamma}{1}$. If $s$ is even, the value on the diagonal is $+1$ and $-1$, otherwise. We consider all possible values of $\alpha$, $\beta$ and $\gamma$:\\
a) If $\gamma$ is even (that is $\gamma=0$):\\
Considering the terms from even $\beta$:\\
\begin{tabular}{lllc} \textbf{1. } & $\beta$ is even \& $\alpha$ is even & \multicolumn{2}{l}{if $\alpha=0$ mod 4 and $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=0$ mod 4 and $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 and $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 and $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & & \tabularnewline \textbf{2. } & $\beta$ is even \& $\alpha$ is odd & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 and $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 and $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 and $\beta=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 and $\beta=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \end{tabular}\\
Considering odd $\beta$:\\
\begin{tabular}{lllc} \textbf{3. } & $\beta$ is odd \& $\alpha$ is even & \multicolumn{2}{l}{if $\alpha=0$ mod 4 and $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=0$ mod 4 and $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 and $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $\alpha=2$ mod 4 and $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & & \tabularnewline \textbf{4. } & $\beta$ is odd \& $\alpha$ is odd & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 and $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=0$ mod 4 and $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 and $(\beta-1)=0$ mod 4 $\Rightarrow$ $(-1)^{s}=+1$}\tabularnewline
& & \multicolumn{2}{l}{if $(\alpha-1)=2$ mod 4 and $(\beta-1)=2$ mod 4 $\Rightarrow$ $(-1)^{s}=-1$}\tabularnewline
& & & \tabularnewline \end{tabular}
b) If $\gamma$ is odd (that is $\gamma=1$)::\\
\textbf{1.} and \textbf{2.} These two terms have opposite sign from a) \textbf{1.} and \textbf{2.} Therefore, in the sum they cancel ($\gamma$ is always 1.)\\
\textbf{3.} and \textbf{4.} Nothing changes in comparison to a) \textbf{3.} and \textbf{4.}\\
Therefore, \begin{align} \begin{split} \langle G_{7}\rangle & = \frac{1}{2^{N-1}}\bigg[\sum_{\beta=0,4..}^{N-m-1}\binom{N-m-1}{\beta+1}-\binom{N-m-1}{\beta+3}\bigg]\bigg[\sum_{\alpha=0,4..}^{m}\binom{m}{\alpha}-\binom{m}{\alpha+1}-\binom{m}{\alpha+2}+\binom{m}{\alpha+3}\bigg]\\ &= \frac{1}{2^{N-1}}Im[(1+i)^{N-m-1}]\cdot\Big(Re[(1+i)^m]-Im[(1+i)^m]\Big). \end{split} \end{align} Now we can consider each cases separately, for this we use the lookup \textbf{table 0}:\\
$(i)$ If $N=8k+5$ and if $(m-1)=4$ mod $8$, $Im[(1+i)^{N-m-1}]=2^\frac{N-m-2}{2}$ and $Re[(1+i)^m]-Im[(1+i)^m]=-2^\frac{m+1}{2}$. And the overall sign from the Table 8 is negative. Therefore, \begin{equation} \langle G_{7}\rangle =-\frac{2^\frac{N-m-2}{2}\cdot (-2^\frac{m+1}{2}) }{2^{N-1}}=\Big(\frac{1}{2}\Big)^{\frac{N-1}{2}}. \end{equation} And, if $(m-1)=0$ mod $8$, $Im[(1+i)^{N-m-1}]=-2^\frac{N-m-2}{2}$ and $Re[(1+i)^m]-Im[(1+i)^m]=2^\frac{m+1}{2}$. Overall sign in negative. Therefore, \begin{equation} \langle G_{7}\rangle =-\frac{-2^\frac{N-m-2}{2}\cdot 2^\frac{m+1}{2} }{2^{N-1}}=\Big(\frac{1}{2}\Big)^{\frac{N-1}{2}}. \end{equation} The same holds for $N=8k+6$ and $N=8k+7$. Besides, $(ii)$ differs with the sign flip and is trivial to check.It is also trivial to prove $(iii)$, as when $N=4k$, $Im[(1+i)^{N-m-1}]=0$. \end{proof}
\begin{lemma}\label{theorem3unifviol} A $N$-qubit (except when $N=4k$) three-uniform HG state violates the separability inequality exponentially after tracing out a single qubit. \end{lemma} \begin{proof} We consider only one case, $N=8k+5$, as others are analogous. Here $M=N-1$:\\ \begin{align} \begin{split} \langle \ensuremath{\mathcal{B}}_N \rangle _Q & =\sum_{m \; odd}^{M}\binom{M}{m}\Big(\frac{1}{\sqrt{2}}\Big)^{M}= 2^{M-1}\cdot2^{-M/2}=\sqrt{2}^{N-3}. \end{split} \end{align} Separability bound is $\sqrt{2}$ \cite{Roy} and it does not depend on the number of qubits. \end{proof}
It is important to note that the similar violation is maintained after tracing out more than one qubit. Numerical evidence for $N=11$ is presented below.\\
\begin{center}
\begin{tabular}{|c|c|c|c|} \hline \#k & Quantum Value & Separability Bound & $\approx$Ratio\tabularnewline \hline \hline \textbf{0} & 511.5 & $\sqrt{2}$ & 361.69\tabularnewline \hline {$1$} & 16 & $\sqrt{2}$ & 11.31\tabularnewline \hline $2$ & 8 & $\sqrt{2}$ & 5.66\tabularnewline \hline $3$ & 4 & $\sqrt{2}$ & 2.83\tabularnewline \hline $4$ & 2 & $\sqrt{2}$ & 1.414\tabularnewline \hline \end{tabular}\\ $\quad$\\ \textbf{Table 9.} Violation of the separability inequalities \cite{Roy} in $N=11$-qubit 3-uniform HG states. Here $k$ is the number of traced out qubits. When $k =0$, the Mermin inequality is violated as expected. \end{center}
\twocolumngrid
\end{document}
|
arXiv
|
{
"id": "1507.03570.tex",
"language_detection_score": 0.5694345235824585,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\baselineskip=20pt \hoffset=-3cm \voffset=0cm \oddsidemargin=3.2cm \evensidemargin=3.2cm \thispagestyle{empty}
\hbadness=10000 \tolerance=10000 \hfuzz=150pt \baselineskip=20pt \hoffset=-3cm \voffset=0cm \oddsidemargin=3.2cm \evensidemargin=3.2cm \thispagestyle{empty}
\hbadness=10000 \tolerance=10000 \hfuzz=150pt
\title{\textbf{Relative Morse index theory and applications in wave equations}} \author{\Large Qi Wang $^{{\rm a,b}}$,$\quad$ Li Wu $^{{\rm b}}$} \date{} \maketitle \begin{center} \it\scriptsize ${}^{\rm a}$ School of Mathematics and Statistics, Henan University, Kaifeng 475000, PR China\\ ${}^{\rm b}$Department of Mathematics, Shandong University, Jinan, Shandong, 250100, PR China\\
\end{center}
\footnotetext[0]{$^a${\bf Corresponding author.} Supported by NNSF of China(11301148) and PSF of China(188576).} \footnotetext[0]{\footnotesize\;{\it E-mail address}: [email protected]. (Qi Wang), [email protected] (Li Wu).}
\noindent {\bf Abstract:} { We develop the relative Morse index theory for linear self-adjoint operator equation without compactness assumption and give the relationship between the index defined in \cite{Wang-Liu-2016} and \cite{Wang-Liu-2017}. Then we generalize the method of saddle point reduction and get some critical point theories by the index, topology degree and critical point theory. As applications, we consider the existence and multiplicity of periodic solutions of wave equations.}
\noindent{\bf Keywords:} {\small}Relative Morse index; Periodic solutions; Wave equations
\section{Introduction}\label{section-introduction} Many problems can be displayed as a self-adjoint operator equation \[
Au=F'(u),\;u\in D(A)\subset \mathbf H,\eqno{(O.E.)}, \] where $\mathbf H$ is an infinite-dimensional separable Hilbert space, $A$ is a self-adjoint operator on $\mathbf H$ with its domain $D(A)$, $F$ is a nonlinear functional on $\mathbf H$. Such as boundary value problem for Laplace's equation on bounded domain, periodic solutions of Hamiltonian systems, Schr\"{o}dinger equation, periodic solutions of wave equation and so on. By variational method, we know that the solutions of (O.E.) correspond to the critical points of a functional. So we can transform the problem of finding the solutions of (O.E.) into the problem of finding the critical points of the functional.
From 1980s, begin with Ambrosetti and Rabinowitz's famous work\cite{Ambrosetti-Rabinowitz-1973}(Mountain Pass Theorem), many crucial variational methods have been developed, such as Minimax-methods, Lusternik-Schnirelman theory, Galerkin approximation methods, saddle point reduction methods, dual variational methods, convex analysis theory, Morse theory and so on (see\cite{Amann-1976},\cite{Amann-Zehnder-1980},\cite{Aubin-Ekeland-1984},\cite{Chang-1993},\cite{Ekeland-1990},\cite{Ekeland-Temam-1976} and the reference therein).
We classified all of these variational problems into three kinds by the spectrum of $A$. For simplicity, denote by $\sigma(A)$, $\sigma_e(A)$ and $\sigma_d(A)$ the spectrum, the essential spectrum and the discrete finite dimensional point spectrum of $A$ respectively.
The first is $\sigma(A)=\sigma_d(A)$ and $\sigma(A)$ is bounded from below(or above), such as boundary value problem for Laplace's equation on bounded domain and periodic problem for second order Hamiltonian systems. Morse theory can be used directly in this kind and this is the simplest situation.
The second is $\sigma(A)=\sigma_d(A)$ and $\sigma(A)$ is unbounded from above and below, such as periodic problem for first order Hamiltonian systems. In this kind, Morse theory cannot be used directly because in this situation the functionals are strongly indefinite and the Morse indices at the critical points of the functional are infinite. In order to overcome this difficulty, the index theory is worth to note here. By the work \cite{Ekeland-1984} of Ekeland, an index theory for convex linear Hamiltonian systems was established. By the works \cite{Conley-Zehnder-1984,Long-1990,Long-1997,Long-Zehnder-1990} of Conley, Zehnder and Long, an index theory for symplectic paths was introduced. These index theories have important and extensive applications, e.g \cite{Dong-Long-1997,Ekeland-Hofer-1985,Ekeland-Hofer-1987,Liu-Long-Zhu-2002,Long-Zhu-2000}. In \cite{Zhu-Long-1999, Long-Zhu-2000-2} Long and Zhu defined spectral flows for paths of linear operators and redefined Maslov index for symplectic paths. Additionally, Abbondandolo defined the concept of relative Morse index theory for Fredholm operator with compact perturbation (see\cite{Abb-2001} and the references therein). In the study of the $L$-solutions (the solutions starting and ending at the same Lagrangian subspace $L$) of Hamiltonian systems, Liu in \cite{Liu-2007} introduced an index theory for symplectic paths using the algebraic methods and gave some applications in
\cite{ Liu-2007, Liu-2007-2}. This index had been generalized by Liu, Wang and Lin in \cite{Liu-Wang-Lin-2011}. In addition to the above index theories defined for specific forms, Dong in \cite{Dong-2010} developed an index theory for abstract operator equations (O.E.).
The third is $\sigma_e(A)\neq \emptyset$, the most complex situation. Since lack of compactness, many classical methods can not be used here. Specially, if $\sigma_e(A)\cap(-\infty,0)\neq \emptyset$ and $\sigma_e(A)\cap(0,\infty)\neq \emptyset$, Ding established a series of critical points theories and applications in homoclinic orbits in Hamiltonian systems, Dirac equation, Schr\"{o}dinger equation and so on, he named these problems???very strongly indefinite problems (see \cite{Ding-2007},\cite{Ding-2017}). Wang and Liu defined the index theory ($i_A(B),\nu_A(B)$) for this kind and gave some applications in wave equation, homoclinic orbits in Hamiltonian systems and Dirac equation, the methods include dual variation and saddle point reduction(see \cite{Wang-Liu-2016} and\cite{Wang-Liu-2017}). Additionally, Chen and Hu in \cite{Chen-Hu-2007} defined the index for homoclinic orbits of Hamiltonian systems. Recently, Hu and Portaluri in \cite{Hu-Portaluri-2017} defined the index theory for heteroclinic orbits of Hamiltonian systems.\\
In this paper, consider the kind of $\sigma_e(A)\neq \emptyset$. Firstly, we develop the relative Morse index theory. Compared with Abbondandolo's work(\cite{Abb-2001}), we generalize the concept of relative Morse index $i^*_A(B)$ for Fredholm operator without the compactness assumption on the perturbation term(see Section \ref{section-relative Morse index}). And we gave the relationship between the relative Morse index $i^*_A(B)$ and the index $i_A(B)$ defined in \cite{Wang-Liu-2016} and\cite{Wang-Liu-2017}. The bridge between them is the concept of spectral flow. As far as we know, the spectral flow is introduced by Atiyah-Patodi-Singer(see\cite{Atiyah-Patodi-Singer-1976}). Since then, many interesting properties and applications of spectral flow have been subsequently established(see\cite{Cappell-Lee-Miller-1994},\cite{Floer-1988},\cite{Robbin-Salamon-1993},\cite{Robbin-Salamon-1995} and \cite{Zhu-Long-1999}).
Secondly, we generalize the method of saddle point reduction and get some critical point theories. With the relative Morse index defined above, we will establish some new abstract critical point theorems by saddle point reduction, topology degree and Morse theory, where we do not need the nonlinear term to be $C^2$ continuous(see Section \ref{section-saddle point reduction}).
Lastly, as applications, we consider the existence and multiplicity of the periodic solutions for wave equation and give some new results(sec Section \ref{section-applications}). To the best of the authors' knowledge, the problem of finding periodic solutions of nonlinear wave equations has attracted much attention since 1960s. Recently, with critical point theory, there are many results on this problem. For example, Kryszewski and Szulkin in \cite{Kryszewski-Szulkin-1997} developed an infinite dimensional cohomology theory and the corresponding Morse theory, with these theories, they obtained the existence of nontrivial periodic solutions of one dimensional wave equation. Zeng, Liu and Guo in \cite{Zeng-Liu-Guo-2004}, Guo and Liu in \cite{Gou-Liu-2007} obtained the existence and multiplicity of nontrivial periodic solution of one dimensional wave equation and beam equation by their Morse index theory developed in \cite{Guo-Liu-Zeng-2004}. Tanaka in \cite{Tanaka-2006} obtained the existence of nontrivial periodic solution of one dimensional wave equation by linking methods. Ji and Li in \cite{Ji-Li-2006} considered the periodic solution of one dimensional wave equation with $x$-dependent coefficients.
By minimax principle, Chen and Zhang in \cite{Chen-Zhang-2014} and \cite{Chen-Zhang-2016} obtained infinitely many symmetric periodic solutions of $n$-dimensional wave equation. Ji in \cite{Ji-2018} considered the periodic solutions for one dimensional wave equation with bounded nonlinearity and $x$-dependent coefficients.
\section{Relative Morse Index $i^*_A(B)$ and the relationship with $i_A(B)$}\label{section-relative Morse index}
Let $\mathbf H$ be an infinite dimensional separable Hilbert space with inner product $(\cdot,\cdot)_\mathbf H$ and norm $\|\cdot\|_\mathbf H$. Denote by $\mathcal O(\mathbf H)$ the set of all linear self-adjoint operators on $\mathbf H$. For $A\in \mathcal O(\mathbf H)$, we denote by $\sigma(A)$ the spectrum of $A$ and $\sigma_e(A)$ the essential spectrum of $A$. We define a subset of $\mathcal O( \mathbf H)$ as follows \[
\mathcal O^0_e(a,b)=\{A\in \mathcal O(\mathbf H)|\;\sigma_e(A)\cap(a,b)=\emptyset \;{\rm and}\;\sigma(A)\cap (a, b)\ne \emptyset\}. \] Denote $\mathcal{L}_s(\mathbf H)$ the set of all linear bounded self-adjoint operators on $\mathbf H$ and a subset of $\mathcal{L}_s(\mathbf H)$ as follows
\begin{equation}\label{eq-L0}
\mathcal{L}_s(\mathbf H,a,b)=\{B\in \mathcal{L}_s(\mathbf H), \;a\cdot I<B< b\cdot I\}, \end{equation} where $I$ is the identity map on $\mathbf H$, $B< b\cdot I$ means that there exists $\delta>0$ such that $(b-\delta)\cdot I-B$ is positive define,
$B> a\cdot I$ has the similar meaning. For any $B\in\mathcal{L}_s(\mathbf H,a,b)$, we have the index pair ($i_A(B),\nu_A(B)$)(see \cite{Wang-Liu-2016,Wang-Liu-2017} for details).
In this section, we will define the relative Morse index $i^*_A(B)$ and give the relationship with $i_A(B)$.
\subsection{Relative Morse Index $i^*_A(B)$} As the beginning of this subsection, we will give a brief introduction of relative Morse index. The relative Morse index can be derived in different ways (see\cite{Abb-2001,Chang-Liu-Liu-1997,Fei-1995,Zhu-Long-1999}). Such kinds of indices have been extensively studied in dealing with periodic orbits of first order Hamiltonian systems. As far as authors known, the existing relative Morse index theory can be regarded as compact perturbation for Fredholm operator. Assume $A$ is a self-adjoint Fredholm operator on Hilbert space $\mathbf H$, with the orthogonal splitting \begin{equation}\label{eq-decomposition of space H} \mathbf H=\mathbf H^-_A\oplus \mathbf H^0_A\oplus \mathbf H^+_A, \end{equation} where $A$ is negative, zero and positive definite on $\mathbf H^-_A,\;\mathbf H^0_A$ and $\mathbf H^+_A$ respectively. Let $P_A$ denote the orthogonal projection from $\mathbf H$ to $\mathbf H^-_A$. If the perturbation term $F$ is a compact self-adjoint operator on $\mathbf H$, then we have $P_A-P_{A-F}$ is compact and $P_A:\mathbf H^-_{A-F}\to \mathbf H^-_A$ is a Fredholm operator and we can define the so called relative Morse index by the Fredholm index of $P_A:\mathbf H^-_{A-F}\to \mathbf H^-_A$.
Generally, if the operator $A$ is not Fredholm operator or the perturbation $F$ is not compact, $P_A:\mathbf H^-_{A-F}\to \mathbf H^-_A$ will not be Fredholm operator and the concept of relative Morse index will be meaningless, but if the perturbation lies in the gap of $\sigma_{e}(A)$, that is to say $A\in \mathcal O^0_e(\lambda_a,\lambda_b)$ for some $\lambda_a,\lambda_b\in\mathbb{R}$ and the perturbation $B\in \mathcal{L}_s(\mathbf H,\lambda_a,\lambda_b)$, we can also defined the relative Morse index $i^*_A(B)$ and give the relationship with the index $i_A(B)$ defined in \cite{Wang-Liu-2017}. Firstly, we need two abstract lemmas.
\begin{lem} \label{Fredholm projection}
Let $A:\mathbf H\rightarrow\mathbf H$ be a bounded self-adjoint operator.
Let $W,V$ be closed spaces of $\mathbf H$.
Denote the orthogonal projection $\mathbf H\rightarrow Y$ by $P_Y$ for any closed linear subspace $Y$ of $\mathbf H$. Assume that\\
\noindent (1). $(Ax,x)_\mathbf H<-\epsilon_1 \|x\|^2_\mathbf H,\;\forall x\in W\backslash\{0\}$, with some constant $\epsilon_1>0$,\\ \noindent (2). $(Ax,x)_\mathbf H> 0,\; \forall x\in V^{\bot}\backslash\{0\} $,\\ \noindent (3). $ (Ax,y)_\mathbf H=0, \forall x\in V,y\in V^\bot$.\\
Then $P_V|_{W}$ is an injection and $P_V(W)$ is a closed subspace of $\mathbf H$. Furthermore, if we assume \\ \noindent (4). $(Ax,x)_\mathbf H\leq 0,\; \forall x\in V\backslash\{0\}$,\\ and there is a closed subspace $U$ of $W^\bot$ such that\\ \noindent (5). $W^\bot/U$ is finite dimensional,\\ \noindent (6). $(Ax,x)_\mathbf H>0,\;\forall x\in U\setminus \{0\}$. \\ Then $P_V:W\rightarrow V$ and $P_W:V\rightarrow W$ are both Fredholm operators and \[ {\rm ind}(P_W:V\rightarrow W)=-{\rm ind}(P_V:W\rightarrow V). \] \end{lem} \begin{proof}
Note that $\ker P_V|_{W}= \ker P_V\cap W =V^\bot \cap W$. From condition (1) and (2), we have $V^\bot \cap W=\{0\}$, so $P_V|_{W}$ is an injection. For $x\in W$, from condition (2) and (3), we have \begin{align*}
-\|A\|\|P_Vx\|^2_\mathbf H&\leq(AP_V x,P_V x)_\mathbf H\\
&=(Ax,x)_\mathbf H-(A(I-P_V)x,(I-P_V)x)_\mathbf H\\
&\leq (Ax,x)_\mathbf H\\
&< -\epsilon_1 \|x\|^2_\mathbf H \end{align*} It follows that \begin{equation}\label{eq-relative Morse index-3}
\|P_V x\|_\mathbf H\ge \sqrt{\frac{\epsilon_1}{\|A\|}}\|x\|_\mathbf H,\;\forall x\in W, \end{equation} so $P_V(W)$ is a closed subspace of $\mathbf H$.
For any $x\in (P_V(W))^\bot \cap V $, that is to say $x\bot P_V(W)$ and $x\bot (I-P_V)(W)$, so we have $x\bot W$ and \begin{equation}\label{eq-relative Morse index-1} P_V(W)^\bot \cap V \subset W^\bot. \end{equation} From condition (4) and (6), \begin{align}\label{eq-relative Morse index-2} ((P_V(W))^\bot \cap V) \cap U&\subset V\cap U\nonumber\\
& =\{0\}. \end{align} From \eqref{eq-relative Morse index-1}, \eqref{eq-relative Morse index-2} and condition (5), $(P_V(W))^\bot \cap V$ is finite dimensional. It follows that $P_V:W\rightarrow V$ is a Fredholm operator. From \eqref{eq-relative Morse index-3}, we have \begin{align*}
\|(I-P_V)x\|^2&=\|x\|^2-\|P_V x\|^2\\
&\leq (1-\epsilon_1/\|A\|)\|x\|^2,\;\forall x\in W. \end{align*}
It follows that $\|I-P_V|_W\|<1$. So the operator $P_WP_V=P_W-P_W(I-P_V):W\rightarrow W$ is invertible. It follows that $P_W:V\rightarrow W$ is surjective, and \begin{equation}\label{eq-relative Morse index-4} \ker P_W\cap P_V(W)={0}. \end{equation} Note that $V$ has the following decomposition \[ V=P_V(W)\bigoplus ((P_V(W))^\bot \cap V), \] from \eqref{eq-relative Morse index-4} and $\dim((P_V(W))^\bot \cap V)<\infty$, we have $\ker P_W \cap V$ is finite dimensional. So the operator $P_W:V\rightarrow W$ is a Fredholm operator. Since $P_WP_V:W\rightarrow W$ is invertible, we have \begin{align*} 0&={\rm ind}(P_WP_V:W\rightarrow W)\\
&={\rm ind}(P_W:V\rightarrow W)+{\rm ind}(P_V:W\rightarrow V). \end{align*} Thus we have proved the lemma. \end{proof} \begin{lem}\label{finite_pertubation} Let $V_1\subset V_2$,$ W_1\subset W_2$ be linear closed subspaces of $\mathbf H$ such that $V_2/V_1$ and $W_2/W_1$ are finite dimensional linear spaces. Let $P_{V_i}$, $P_{W_j}$ be the orthogonal projections onto $V_i$ and $W_j$ and respectively, $i,j=1,2$. Assume that $P_{W_{j^*}}:V_{i^*}\rightarrow W_{j^*}$ is a Fredholm operator for some fixed $i^*,j^*\in \{1,2\}$. Then $P_{W_j}:V_i \rightarrow W_j$, $i,j=1,2$ are all Fredholm operators. Furthermore, we have \begin{align*} {\rm ind}(P_{W_j}:V_i\rightarrow W_j)=&{\rm ind}(P_{V_{i^*}}:V_i\rightarrow V_{i^*})+{\rm ind}(P_{W_{j^*}}:V_{i^*}\rightarrow W_{j^*})\\ &+{{\rm ind}}(P_{W_j}:W_{j^*}\rightarrow W_j). \end{align*} \end{lem} \begin{proof} Since $V_2/V_1$ and $W_2/W_1$ are finite dimensional linear spaces, $P_{W_j}-P_{W_{j^*}}$ and $P_{V_i}-P_{V_{i^*}}$ are both compact operator. So $P_{W_j}P_{V_i}-P_{W_{j^*}}P_{V_{i^*}}$ is also compact operator.
Note that on $V_i$,
\[
(P_{W_j}-P_{W_j}P_{W_{j^*}}P_{V_{i^*}})|_{V_i}=P_{W_j}(P_{W_j}P_{V_i}-P_{W_{j^*}}P_{V_{i^*}})|_{V_i}. \] It follows that $P_{W_j}-P_{W_j}P_{W_{j^*}}P_{V_{i^*}}:V_i\rightarrow W_j$ is compact. Then we can conclude that \begin{align*} {{\rm ind}}(P_{W_j}:V_i\rightarrow W_j)=&{\rm ind}(P_{W_j}P_{W_{j^*}}P_{V_{i^*}}:V_i\rightarrow W_j)\\
=&{\rm ind}(P_{V_{i^*}}:V_i\rightarrow V_{i^*})+{\rm ind}(P_{W_{j^*}}:V_{i^*}\rightarrow W_{j^*})\\
&+{{\rm ind}}(P_{W_j}:W_{j^*}\rightarrow W_j). \end{align*} We have proved the lemma. \end{proof}
With these two lemmas, we can define the relative Morse index. We consider a normal type that is $A\in \mathcal O^0_e(-1,1)$ for simplicity. Let $B\in \mathcal{L}_s(\mathbf H,-1,1)$ with its norm $\|B\|=c_B$, so we have $0\leq c_B<1$. Then $A-tB$ is a self-adjoint Fredholm operator for $t\in [0,1]$. We have $\sigma_{ess}(A-tB)\cap (-1+tc_B,1-tc_B)=0$. Let $E_{A-tB}(z)$ be the spectral measure of $A-tB$. Denote \begin{equation}\label{eq-projection splitting} P(A-tB,U)= \int_{U} dE_{A-tB}(z), \end{equation}
with $U\subset \mathbb{R}$, and rewrite it as $P(t,U) $ for simplicity. Let
\[
V(A-tB,U)={\rm im} P(t,U)
\]
and rewrite it as $V(t,U)$ for simplicity. For any $c_0\in\mathbb{R}$ satisfying $c_B<c_0<1$, we have
\[
((A-B)x,x)_\mathbf H>(c_0-c_B)\|x\|^2_\mathbf H,\;x\in V(0,(c_0,+\infty))\cap D(A).
\] So there is $\epsilon >0$, such that \begin{equation}\label{eq-relative Morse index-5}
((A-B)x,x)_\mathbf H>\epsilon ((|A-B|+I)x,x)_\mathbf H ,\forall x\in V(0,(c_0,+\infty))\cap D(A) \end{equation} Similarly, we have \begin{equation}\label{eq-relative Morse index-6}
((A-B)x,x)_\mathbf H<-\epsilon ((|A-B|+I)x,x)_\mathbf H ,\forall x\in V(0,(-\infty,-c_0))\cap D(A) \end{equation} Denote \[
P_{s,a}^{t,b}:=P(t,(-\infty,b))|_{V(s,(-\infty,a))},\forall t,s\in [0,1]\;{\rm and}\;a,b\in \mathbb{R}. \] Clearly, we have $P(t,(-\infty,b))=P_{s,+\infty}^{t,b}$ , $\forall s\in [0,1] $. \begin{lem}\label{inverse_relation} For any $a\in[-c_0,c_0]$, the map $P_{0,a}^{1,0}$ $P_{1,0}^{0,a}$ are both Fredholm operators.
Furthermore, we have ${\rm ind}(P_{0,a}^{1,0})=-{\rm ind}(P_{1,0}^{0,a})$. \end{lem} \begin{proof}
From \eqref{eq-relative Morse index-5} and \eqref{eq-relative Morse index-6}, there is $\epsilon>0$ such that
\[
((A-B)(|A-B|+I)^{-1}x,x)>\epsilon \|x\|^2,\;\forall x\in V(0,(c_0,+\infty)),
\]
and
\[
((A-B)(|A-B|+I)^{-1}x,x)<-\epsilon \|x\|^2,\;\forall x\in V(0,(-\infty,-c_0)).
\]
Now, let the operator $(A-B)(|A-B|+I)^{-1}$, the spaces $V(0,(-\infty,-c_0))$, $V(1,(-\infty,0])$ and $V(0,(c_0,+\infty))$ be the operator $A$ and the spaces $W,V$ and $U$ in Lemma \ref{Fredholm projection} correspondingly. It's easy to verify that condition (1), (2), (3), (4) and (6) are satisfied, and since $A\in \mathcal O^0_e(-1,1)$, $V(0,[-c_0,c_0])$ is finite dimensional, so condition (5) is satisfied. Then $P_{0,-c_0}^{1,0} $ and $P_{1,0}^{0,-c_0} $ are both Fredholm operators. We also have \[ {\rm ind}(P_{0,-c_0}^{1,0})=-{\rm ind}(P_{1,0}^{0,-c_0}).
\] By Lemma \ref{finite_pertubation}, $ P_{0,a}^{1,0}$ and $P_{1,0}^{0,a}$ are both Fredholm operators with $a\in [-c_0,c_0]$, and we have \[
{\rm ind}(P_{0,a}^{1,0})=-{\rm ind}(P_{1,0}^{0,a}), a\in [-c_0,c_0].
\] \end{proof}
\begin{rem}\label{inverse_relation(general result)}
Generally, we have $P_{s,a}^{t,b}\; and\; P_{t,b}^{s,a}$ are both Fredholm operators
with $a\in (-1+sc_B,1-sc_B)$, $b\in (-1+tc_B,1-tc_B)$ and we have
\[
{{\rm ind}}(P_{s,a}^{t,b})=-{{\rm ind}}(P_{t,b}^{s,a}).
\]
\end{rem} Here we replace $A,B$ by $A'=A-sB$ and $B'=(t-s)B$ respectively in Lemma \ref{inverse_relation}, then all the proof will be same, so we omit the proof here.
\begin{defi}\label{defi-index defined by relative Morse index} Define the relative Morse index by \[ i^*_A(B):={\rm ind}(P^{0,0}_{1,0}),\;\forall B\in\mathcal{L}_s(\mathbf H,-1,1). \] \end{defi}
\subsection{The relationship between $i^*_A(B)$ and $i_A(B)$} Now, we will prove that $i^*_A(B)=i_A(B)$ by the concept of spectral flow. We need some preparations. There are some equivalent definitions of spectral flow. We use the Definition 2.1, 2.2 and 2.6 in \cite{Zhu-Long-1999}. Let $A_s$ be a path of self-adjoint Fredholm operators. The APS projection of $A_s$ is defined by $Q_{A_s}=P(A_s,[0,+\infty))$ . Recall that locally, the spectral flow of $A_s$ is the s-flow of $Q_{A_s}$. Choose $\epsilon>0$ such that $V(A_{s_0},[0,+\infty))=V(A_{s_0},[-\epsilon,+\infty))$. Then $\epsilon \notin \sigma (A_{s_0})$. Let $P_{A_s}=P(A_s,[-\epsilon,+\infty))$. Then there is $\delta>0$ such that $P_{A_s}$ is continuous on $(s_0-\delta,s_0+\delta)$ and $P_{A_s}-Q_{A_s}$ is compact for $s\in(s_0-\delta,s_0+\delta)$ . The the s-flow of $Q_{A_s}$ on $[s_0,b]\subset (s_0-\delta,s_0+\delta)$ can be calculated as \begin{align*} sfl(Q_{A_s},[s_0,b])=& -{\rm ind}(P_{A_{s_0}}:V(A_{s_0},[0,\infty)\to V(A_{s_0},[0,\infty))\\ &+{\rm ind}(P_{A_{s_b}}:V(A_{s_b},[0,\infty)\to V(A_{s_b},[-\epsilon,\infty))\\ =&-{\rm dim}(V(A_{s_b},[-\epsilon,0)))\\ =&{\rm ind}({\rm Id}-P_{A_{s_b}}:V(A_{s_b},(-\infty,-\epsilon)\to V(A_{s_b},(-\infty,0)). \end{align*} If $A_s=A-sB$, with $\epsilon$ and $\delta$ chosen like above, we have $sf\{A-sB,[s_0,s_1]\}={\rm ind}P_{s_1,-\epsilon}^{s_1,0}$ for $[s_1,s_2]\subset [s_1,s_1+\delta]$.
\begin{lem}\label{continuity} Let $t_0\in [0,1]$. Let $a\in (-1+t_0c_B,1-t_0c_B)\backslash\sigma(A-t_0B) $. Then we have \[
\lim_{s\to t_0}\|P_{t_0,a}^{t,0}-P_{s,a}^{t,0}P_{t_0,a}^{s,a}\|=0,\forall t\in[0,1]. \] and \begin{align*} \lim_{s\to t_0}{{\rm ind}}(P_{s,a}^{t,0}))={{\rm ind}}(P_{t_0,a}^{t,0})\\ \lim_{s\to t_0}{{\rm ind}}(P_{t,0}^{s,a}))={{\rm ind}}(P_{t,0}^{t_0,a}). \end{align*} \end{lem} \begin{proof} Since $a\notin\sigma(A-t_0B) $, there is $\delta_1>0$ such that $P(\cdot,(-\infty,a))$ is a continuous path of operators on $(t_0-\delta_1,t_0+\delta_1)$, and
\[\|(P(s,(-\infty,a))-P(t_0,(-\infty,a)))\|<1\] with $s\in (t_0-\delta_1,t_0+\delta_1)$. Then $P_{t_0,a}^{s,a}$ and $ P_{s,a}^{t_0,a}$ are both homeomorphisms. Note that on $V(t_0,(-\infty,a))$, we have
\[
P_{t_0,a}^{t,0}-P_{s,a}^{t,0}P_{t_0,a}^{s,a}
=P(t,(-\infty,0))(P(t_0,(-\infty,a))-P(s,(-\infty,a)))|_{V(t_0,(-\infty,a))}.
\]
By the continuity of $P(s,(-\infty,a))$, it follows that
\[
\lim_{s\to t_0}\|P_{t_0,a}^{t,0}-P_{s,a}^{t,0}P_{t_0,a}^{s,a}\|=0. \] Then we have \begin{align*} {\rm ind}(P_{t_0,a}^{t,0})&=\lim_{s\to t_0}{\rm ind}(P_{s,a}^{t,0}P_{t_0,a}^{s,a})\\
&=\lim_{s\to t_0}{\rm ind}(P_{s,a}^{t,0})+{\rm ind}(P_{t_0,a}^{s,a}))\\
&=\lim_{s\to t_0}{{\rm ind}}(P_{s,a}^{t,0}). \end{align*} By remark \ref{inverse_relation(general result)}, we get \begin{align*} {\rm ind}(P_{t,0}^{t_0,a})&=-{\rm ind} (P_{t_0,a}^{t,0})\\
&=-\lim_{s\to t_0}{\rm ind}(P_{s,a}^{t,0})\\
&=\lim_{s\to t_0}{\rm ind}(P_{t,0}^{s,a}). \end{align*} \end{proof}
\begin{lem}\label{local_flow} For each $t_1\in [0,1]$, there is $\delta >0$ such that
\[{\rm ind} (P_{t_1,0}^{t_2,0})=sf\{A-t_1 B-s(t_2-t_1)B,[0,1]\}\] with $|t_2-t_1|<\delta$. \end{lem} \begin{proof} Since $A-t_1B$ is a Fredholm operator, there is $\epsilon>0$ such that $P(t_1,(-\infty,0))=P(t_1,(-\infty,-\epsilon))$. It follows that $\epsilon \notin \sigma (A-t_1B)$, and we have \[ P_{t_1,-\epsilon}^{t_2,0}=P_{t_1,0}^{t_2,0}. \] By lemma \ref{continuity} we have \[ \lim_{t_2\to t_1}{\rm ind}(P_{t_1,-\epsilon}^{t_2,-\epsilon})={\rm ind}(P_{t_1,-\epsilon}^{t_1,-\epsilon})=0. \] It follows that \begin{align*} \lim_{t_2\to t_1}{\rm ind}(P_{t_1,-\epsilon}^{t_2,0})&=\lim_{t_2\to t_1}{\rm ind}(P_{t_2,-\epsilon}^{t_2,0}P_{t_1,-\epsilon}^{t_2,-\epsilon})\\ &=\lim_{t_2\to t_1}{\rm ind}(P_{t_2,-\epsilon}^{t_2,0})+\lim_{t_2\to t_1}{\rm ind}(P_{t_1,-\epsilon}^{t_2,-\epsilon})\\ &=\lim_{t_2\to t_1}{\rm ind}(P_{t_2,-\epsilon}^{t_2,0}). \end{align*}
So there is $\delta>0$ such that ${{\rm ind}}(P_{t_1,-\epsilon}^{t_2,0})={\rm ind}(P_{t_2,-\epsilon}^{t_2,0}) $ with $|t_2-t_1|<\delta$, and $P(t,(-\infty,-\epsilon))$ is continuous on $(t_1-\delta,t_1+\delta)$. Note that $sf\{A-t_1 B-s(t_2-t_1)B,[0,1]\}={{\rm ind}} P_{t_2,-\epsilon}^{t_2,0}$ by continuation of $P(t,(-\infty,-\epsilon))$. Then the lemma follows. \end{proof} \begin{lem}\label{additional} ${{\rm ind}}(P_{0,0}^{1,0})={{\rm ind}}(P_{t,0}^{1,0})+{{\rm ind}}( P_{0,0}^{t,0})$ with $\forall t\in [0,1]$. \end{lem} \begin{proof} By Lemma \ref{finite_pertubation} and Lemma \ref{inverse_relation(general result)}, for any $t_0\in [0,1]$ \begin{align*} {\rm ind}(P_{t_0,0}^{1,0})+{\rm ind}( P_{0,0}^{t_0,0})&={\rm ind}(P_{t_0,a}^{1,0})+{\rm ind}( P_{t_0,0}^{t_0,a})+{\rm ind}( P_{t_0,a}^{t_0,0})+{\rm ind}( P_{0,0}^{t_0,a})\\
&={\rm ind}(P_{t_0,a}^{1,0})+{{\rm ind}}( P_{0,0}^{t_0,a}),\; \forall a\in (-1+sc_B,1-sc_B). \end{align*}
Choose $a_{t_0}\in (-1+t_0c_B,1-t_0c_B)$ and $a_{t_0}\notin \sigma(A-t_0B)$. By lemma \ref{continuity}, \[ f:t\to {\rm ind}(P_{t,a_{t_0}}^{1,0})+{\rm ind}( P_{0,0}^{t,a_{t_0}}) \]
is continuous at $t_0$. So the function $f:t\to {{\rm ind}}(P_{t,0}^{1,0})+{{\rm ind}}( P_{0,0}^{t,0}) $ is continuous on $[0,1]$. So it must be a constant function. It follows that \[ {{\rm ind}}(P_{0,0}^{1,0})=f(1)=f(t)={{\rm ind}}(P_{t,a}^{1,0})+{{\rm ind}}( P_{0,0}^{t,a}). \] \end{proof} \begin{rem} In fact, we have \[{{\rm ind}}(P_{a,0}^{b,0})={{\rm ind}}(P_{s,0}^{b,0})+{{\rm ind}}( P_{a,0}^{s,0})\] with $s\in [0,1]$. \end{rem} \begin{thm}\label{thm-relative morse index and spectral flow} We have \[ sf\{A-tB,[a,b]\}={{\rm ind}}(P_{a,0}^{b,0})=-{{\rm ind}}(P_{b,0}^{a,0}) \] with $[a,b]\subset [0,1]$. \end{thm} \begin{proof} It is a direct consequence of Lemma \ref{local_flow} and Lemma \ref{additional}. \end{proof}
Now by the property of $i_A(B)$(see \cite[Lemma 2.9]{Wang-Liu-2016},\cite[Lemma 2.3]{Wang-Liu-2017}) and Theorem \ref{thm-relative morse index and spectral flow}, we have the following result. \begin{prop}\label{prop-relations between indexes} $ i^*_A(B)=i_A(B), A\in\mathcal{O}^0_e(-1,1),B\in\mathcal{L}_s(\mathbf H,-1,1). $ \end{prop} Generally, with the same method we can define the relative Morse index $i^*_A(B)$ for $A\in\mathcal{O}^0_e(\lambda_a,\lambda_b)$, $B\in\mathcal{L}_s(\mathbf H,\lambda_a,\lambda_b)$ and we can prove the index $i^*_A(B)$ coincide with $i_A(B)$ by the concept of spectral flow, we omit them here.
\section{Saddle point reduction of (O.E.) and some abstract critical points Theorems}\label{section-saddle point reduction}
Now for simplicity, let $b>0$ and $a=-b$, for $A\in\mathcal O^0_e(-b,b)$, we consider the following operator equation \[ Az=F'(z),\;z\in D(A)\subset \mathbf H, \eqno(O.E.) \] where $F\in C^1(\mathbf H,\mathbb{R})$. Assume\\ \noindent($F_1$) $F\in C^1(\mathbf H,\mathbb{R})$, $F':\mathbf H\to\mathbf H$ is Lipschitz continuous
\begin{equation}\label{eq-the Lipschitz continuity of F'}
\|F'(z+h)-F'(z)\|_{\mathbf H}\leq l_F\|h\|_{\mathbf H},\;\forall z,h\in\mathbf H, \end{equation} with its Lipschitz constant $l_F<b$. \subsection{Saddle point reduction of (O.E.)}
In this part, assume $A\in\mathcal{O}^0_e(-b,b)$ and $F$ satisfies condition ($F_1$), we will consider the method of saddle point reduction without assuming the nonlinear term $F\in C^2(D(|A|^{1/2}))$, then we will give some abstract critical point theorems. Let $E_A(z)$ the spectrum measure of $A$, since $\sigma_e(A)\cap(-b,b)=\emptyset$, we can choose $l\in (l_F,b)$, such that
\[
-l, l\notin\sigma(A).
\] Different from the above section, in this section, consider projection map $P(A,U)$ defined in \eqref{eq-projection splitting} on $H$, for simplicity, we rewrite them as \begin{equation}\label{eq-projections} P^-_A:=P(A,(-\infty,-l)),\;P^+_A:=P(A,(l,\infty)),\;P^0_A:=P(A,(-l,l)), \end{equation} in this section.
Then we have the following decomposition which is different from \eqref{eq-decomposition of space H}, \[ \mathbf H=\widehat{\mathbf H}^-_A\oplus\widehat{\mathbf H}^+_A \oplus\widehat{\mathbf H}^0_A, \] where $\widehat{\mathbf H}^*_A:=P^*_A\mathbf H$($*=\pm, 0$) and $\widehat{\mathbf H}^0_A$ is finite dimensional subspace of $\mathbf H$, for simplicity we rewrite $\mathbf H^*:=\widehat{\mathbf H}^*_A$. Denote $A^*$ the restriction of $A$ on $\mathbf H^*$($*=\pm, 0$), thus we have $(A^\pm)^{-1}$ are bounded self-adjoint linear operators on $\mathbf H^\pm$ respectively and satisfying \begin{equation}\label{eq-the norm of the inverse of Apm}
\|(A^\pm)^{-1}\|\leq \frac{1}{l}. \end{equation}
Then (OE) can be rewritten as \begin{equation}\label{eq-decomposition 1 of OE} z^\pm =(A^\pm)^{-1}P^\pm_A F'(z^++z^-+z^0), \end{equation} and \begin{equation}\label{eq-decomposition 2 of OE} A^0z^0=P^0_AF'(z^++z^-+z^0), \end{equation} where $z^*=P^*_Az$($*=\pm, 0$), for simplicity, we rewrite $x:=z^0$. From \eqref{eq-the Lipschitz continuity of F'} and \eqref{eq-the norm of the inverse of Apm}, we have $(A^\pm)^{-1}P^\pm_A F'$ is contraction map on $\mathbf H^+\oplus \mathbf H^-$ for any $x\in\mathbf H^0$. So there is a map $z^\pm(x):\mathbf H^0\to\mathbf H^\pm$ satisfying \begin{equation}\label{eq-zpm(x)} z^\pm(x)=(A^\pm)^{-1}P^\pm F'(z^\pm(x)+x),\;\forall x\in\mathbf H, \end{equation} and the following properties. \begin{prop}\label{prop-continuous and property of saddle point reduction} (1) The map $z^\pm(x):\mathbf H^0\to \mathbf H^\pm$ is continuous, in fact we have \[
\|(z^++z^-)(x+h)-(z^++z^-)(x)\|_\mathbf H\leq\frac{l_F}{l-l_F}\|h\|_{\mathbf H},\;\;\forall x,h\in\mathbf H^0. \]
(2) $\|(z^++z^-)(x)\|_\mathbf H\displaystyle\leq \frac{l_F}{l-l_F}\|x\|_{\mathbf H}+\frac{1}{l-l_F}\|F'(0)\|_{\mathbf H}$. \end{prop} \noindent\textbf{Proof.}(1) For any $x,\;h\in \mathbf H^0$, here we write $z^\pm(x):=z^+(x)+z^-(x)$ and $(A^\pm)^{-1}P^\pm_A:=(A^+)^{-1}P^+_A+(A^-)^{-1}P^-_A$ for simplicity, we have \begin{align*}
\|z^\pm(x+h)-z^\pm(x)\|_{\mathbf H} &=\|(A^\pm)^{-1}P^\pm_A F'(z^\pm(x+h)+x+h)-(A^\pm)^{-1}P^\pm_A F'(z^\pm(x)+x)\|_{\mathbf H}\\
&\leq\frac{1}{l}\|F'(z^\pm(x+h)+x+h)-F'(z^\pm(x)+x)\|_{\mathbf H}\\
&\leq\frac{l_F}{l}\|z^\pm(x+h)-z^\pm(x)+h\|_{\mathbf H}\\
&\leq\frac{l_F}{l}\|z^\pm(x+h)-z^\pm(x)\|_{\mathbf H}+\frac{l_F}{l}\|h\|_{\mathbf H}. \end{align*}
So we have $\|z^\pm(x+h)-z^\pm(x)\|_\mathbf H\leq\frac{l_F}{l-l_F}\|h\|_{\mathbf H}$ and the map $z^\pm(x):{\mathbf H}^0\to {\mathbf H}^\pm$ is continuous.\\ (2)Similarly, \begin{align*}
\|z^\pm(x)\|_{\mathbf H}&=\|(A^\pm)^{-1}P^\pm_A F'(z^\pm(x)+x)\|_{\mathbf H}\\
&\leq\frac{1}{l}\|F'(z^\pm(x)+x)\|_{\mathbf H}\\
&\leq\frac{1}{l}\|F'(z^\pm(x)+x)-F'(0)\|_{\mathbf H}+\frac{1}{l}\|F'(0)\|_{\mathbf H}\\
&\leq\frac{l_F}{l}(\|z^\pm(x)\|_{\mathbf H}+\|x\|_H)+\frac{1}{l}\|F'(0)\|_{\mathbf H}. \end{align*}
So we have $\|z^\pm(x)\|_\mathbf H\leq \frac{l_F}{l-l_F}\|x\|_{\mathbf H}+\frac{1}{l-l_F}\|F'(0)\|_{\mathbf H}$.$
\Box$
\begin{rem}Denote $\mathbf E=D(|A|^{\frac{1}{2}})$, with its norm \[
\|z\|^2_\mathbf E:=\||A|^{\frac{1}{2}}(z^++z^-)\|^2_\mathbf H+\|x\|^2_\mathbf H,\;u\in \mathbf E. \] From \eqref{eq-zpm(x)}, we have $z^\pm(x)\in D(A)\subset\mathbf E$,
and we have\\
(1) The map $z^\pm(x):\mathbf H^0\to \mathbf E$ is continuous, and
\begin{equation}\label{eq-uniform continuous of z in E}
\|(z^++z^-)(x+h)-(z^++z^-)(x)\|_\mathbf E\leq\frac{l_F \cdot l^\frac{1}{2}}{l-l_F}\|h\|_{\mathbf H},\;\;\forall x,h\in\mathbf H^0.
\end{equation}
(2) $\|(z^++z^-)(x)\|_\mathbf E\displaystyle\leq \frac{l^\frac{1}{2}}{l-l_F}(l_F\cdot\|x\|_{\mathbf H}+\|F'(0)\|_{\mathbf H})$. \end{rem} \noindent{\bf Proof.} The proof is similar to Proposition \ref{prop-continuous and property of saddle point reduction}, we only prove (1). \begin{align*}
\|z^\pm(x+h)-z^\pm(x)\|_{\mathbf E} &=\|(|A|^{\frac{1}{2}})[z^\pm(x+h)-z^\pm(x)]\|_{\mathbf H}\\
&=\|(A^\pm)^{-\frac{1}{2}}[P^\pm_A F'(z^\pm(x+h)+x+h)-P^\pm_A F'(z^\pm(x)+x)]\|_{\mathbf H}\\
&\leq\frac{1}{l^\frac{1}{2}}\|F'(z^\pm(x+h)+x+h)-F'(z^\pm(x)+x)\|_{\mathbf H}\\
&\leq\frac{l_F}{l^\frac{1}{2}}\|z^\pm(x+h)-z^\pm(x)+h\|_{\mathbf H}\\
&\leq\frac{l_F}{l^\frac{1}{2}}\|z^\pm(x+h)-z^\pm(x)\|_{\mathbf H}+\frac{l_F}{l^\frac{1}{2}}\|h\|_{\mathbf H}\\
&\leq\frac{l_F}{l}\|z^\pm(x+h)-z^\pm(x)\|_{\mathbf E}+\frac{l_F}{l^\frac{1}{2}}\|h\|_{\mathbf H}, \end{align*}
where the last inequality depends on the fact that $\|z^\pm\|_\mathbf E\geq l^\frac{1}{2}\|z^\pm\|_\mathbf H$, so we have \eqref{eq-uniform continuous of z in E}.
Now, define the map $z:\mathbf H^0\to\mathbf H$ by \[ z(x)=x+z^+(x)+z^-(x). \]
Define the functional $a:\mathbf H^0\to \mathbb{R}$ by
\begin{equation}\label{eq-saddle point reduction}
a(x)=\frac{1}{2}(Az(x),z(x))_\mathbf H-F(z(x)),\;x\in \mathbf H^0.
\end{equation}
With standard discussion, the critical points of $a$ correspond to the solutions of (O.E.), and we have
\begin{lem}\label{lem-the smoothness of a}
Assume $F$ satisfies ($F_1$), then we have $a\in C^1(\mathbf H^0,\mathbb{R})$ and
\begin{equation}\label{eq-saddle point reduction-the derivative of a}
a'(x)=Az(x)-F'(z(x)),\;\;\forall x\in \mathbf H^0.
\end{equation}
Further more, if $F\in C^2(\mathbf H,\mathbb R)$, we have $a\in C^2(\mathbf H^0,\mathbb{R})$, for any critical point $x$ of $a$, $F''(z(x))\in \mathcal{L}_s(\mathbf H,-b,b)$ and the morse index $m^-_a(x)$ satisfies the following equality
\begin{equation}\label{eq-relation between morse index and our index}
m^-_a(x_2)-m^-_a(x_1)=i^*_A(F''(z(x_2)))-i^*_A(F''(z(x_1))),\;\;\forall x_1,x_2\in \mathbf H^0.
\end{equation}
\end{lem}
\noindent{\bf Proof.} For any $x,h\in\mathbf H^0$, write
\[
\eta(x,h):=z^+(x+h)+z^-(x+h)-z^+(x)-z^-(x)+h
\]
for simplicity, that is to say
\[
z(x+h)=z(x)+\eta(x,h),\;\;\forall x,h\in\mathbf H^0,
\]
and from \eqref{eq-uniform continuous of z in E}, we have
\begin{equation}\label{eq-uniform astimate of eta}
\|\eta(x,h)\|_\mathbf H\leq C\|h\|_{\mathbf H},\;\;\forall x,h\in\mathbf H^0,
\end{equation}
where $C=\displaystyle \frac{l+l_F}{l-l_F}$. Let $h\to 0$ in $\mathbf H^0$, and for any $x\in\mathbf H^0$, we have
\begin{align*}
a(x+h)-a(x)=&\frac{1}{2}[(Az(x+h),z(x+h))_\mathbf H-(Az(x),z(x))_\mathbf H]-[F(z(x+h))-F(z(x))]\\
=&(Az(x),\eta(x,h))_\mathbf H+\frac{1}{2}(A\eta(x,h),\eta(x,h))_\mathbf H\\
&-(F'(z(x)),\eta(x,h))_\mathbf H+o(\|\eta(x,h)\|_{\mathbf H}).
\end{align*}
From \eqref{eq-uniform astimate of eta} we have
\[
a(x+h)-a(x)=(Az(x)-F'(z(x)),\eta(x,h))_\mathbf H+o(\|h\|_\mathbf H),\;\;\forall x\in \mathbf H^0,\; {\rm and}\;\|h\|_\mathbf H\to 0.
\]
Since $z^\pm(x)$ is the solution of \eqref{eq-zpm(x)} and from the definition of $\eta(x,h)$, we have
\[
(Az(x)-F'(z(x)),\eta(x,h))_\mathbf H=(Az(x)-F'(z(x)),h)_\mathbf H,\;\;\forall x, h\in \mathbf H^0,
\]
so we have
\[
a(x+h)-a(x)=(Az(x)-F'(z(x)),h)_\mathbf H+o(\|h\|_\mathbf H),\;\;\forall x\in \mathbf H^0,\; {\rm and}\;\|h\|_\mathbf H\to 0,
\]
and we have proved \eqref{eq-saddle point reduction-the derivative of a}. If $F\in C^2(\mathbf H,\mathbb R)$, from \eqref{eq-zpm(x)} and by Implicit function theorem, we have $z^\pm \in C^1(\mathbf H^0,\mathbf H^\pm)$.
From \eqref{eq-zpm(x)} and\eqref{eq-saddle point reduction-the derivative of a}, we have
\[
a'(x)=Ax-P^0F(z(x))
\]
and
\[
a''(x)=A|_{\mathbf H_0}-P^0F''(z(x))z'(x),
\]
that is to say $a\in C^2(\mathbf H^0,\mathbb R)$. Finally, from Theorem \ref{thm-relative morse index and spectral flow} received above, Definition 2.8 and Lemma 2.9 in \cite{Wang-Liu-2016}, we have \eqref{eq-relation between morse index and our index}. $
\Box$ \subsection{Some abstract critical points Theorems} In this part, we will give some abstract critical points Theorems for (O.E.) by the method of saddle point reduction introduced above. Since we have Proposition \ref{prop-relations between indexes}, we will not distinguish $i^*_A(B)$ from $i_A(B)$. Beside condition ($F_1$), assume $F$ satisfying the following condition.\\
\noindent ($F_2$) There exist $B_1,B_2\in \mathcal{L}_s(\mathbf H,-b,b)$ and $B:\mathbf H\to \mathcal{L}_s(\mathbf H,-b,b)$ satisfying \[ B_1\leq B_2,\;i_A(B_1)=i_A(B_2),\; {\rm and}\; \nu_A(B_2)=0, \]
\[
B_1\leq B(z) \leq B_2,\forall z\in\mathbf H,
\]
such that \[
F'(z)-B(z)z=o(\|z\|_\mathbf H),\|z\|_\mathbf H\to\infty. \] Before the following Theorem, we need a Lemma. \begin{lem}\label{lem-0 has a positive distance from sigma(A-B)} Let $B_1,B_2\in \mathcal{L}_s(\mathbf H,-b,b)$ with $B_1\leq B_2,\;i_A(B_1)=i_A(B_2),\; {\rm and}\; \nu_A(B_2)=0$, then there exists $\varepsilon>0$, such that for all $B\in\mathcal{L}_s(\mathbf H)$ with \[
B_1\leq B \leq B_2, \] we have \[ \sigma(A-B)\cap (-\varepsilon,\varepsilon)=\emptyset. \] \end{lem} {\noindent}{\bf Proof.} For the property of $i_A(B)$, we have $\nu_A(B_1)=0$. So there is $\varepsilon>0$, such that \[ i_A(B_{1,\varepsilon})=i_A(B_1)=i_A(B_2)=i_A(B_{2,\varepsilon}), \] with $B_{*,\varepsilon}=B_*+\varepsilon\cdot I,(*=1,2)$. Since $B_{1,\varepsilon}\leq B-\varepsilon I<B+\varepsilon I\leq B_2'$. It follows that $i_A(B-\varepsilon I)=i_A(B+\varepsilon I)$. Note that \[ i_A(B+\varepsilon)-i_A(B-\varepsilon)=\sum_{-\varepsilon < t \le \varepsilon } \nu_A(B-t \cdot I). \] We have $0\notin \sigma(A-B-\eta),\;\forall \eta\in(-\varepsilon,\varepsilon)$, thus the proof is complete.$
\Box$
\begin{thm}\label{thm-abstract thm 1 for the existence of solution} Assume $A\in\mathcal O^0_e(-b,b)$. If $F$ satisfies conditions ($F_1$) and ($F_2$), then (O.E.) has at least one solution.
\end{thm}
\noindent{\bf Proof.} Firstly, for $\lambda\in[0,1]$, consider the following equation
\[
Az=(1-\lambda) B_1z+\lambda F'(z).\eqno(O.E.)_\lambda
\]
We claim that the set of all the solutions ($z,\lambda$) of (O.E.)$_\lambda$ are a priori bounded. If not, assume there exist $\{(z_n,\lambda_n)\}$ satisfying (O.E.)$_\lambda$ with $\|z_n\|_{\mathbf H}\to\infty$. Without lose of generality, assume $\lambda_n\to\lambda_0\in[0,1]$. Denote by \[ F_\lambda(z)=\frac{1-\lambda}{2}(B_1z,z)_{\mathbf H}+\lambda F(z),\;\forall z\in \mathbf H.
\]
Since $F$ satisfies condition ($F_1$) and $B_1\in \mathcal{L}_s(\mathbf H,-b,b)$, we have $F'_\lambda:\mathbf H\to\mathbf H$ is Lipschitz continuous with its Lipschitz constant less than $b$, that is to say there exists $\hat{l}\in [\l_F,b)$ such that
\[
\|F'_\lambda(z+h)-F'_\lambda(z)\|_{\mathbf H}\leq \hat{l}\|h\|_{\mathbf H},\;\forall z,h\in\mathbf H,\lambda\in[0,1].
\]
Now, consider the projections defined in \eqref{eq-projections}, choose $l\in (\hat{l},b)$ satisfying $-l,l\notin \sigma(A)$, from \eqref{eq-decomposition 1 of OE} and \eqref{eq-decomposition 2 of OE}, we decompose $z_n$ by
\[
z_n=z^+_n+z^-_n+x_n,
\]
with $z^*_n\in\mathbf H^*$($*=\pm,0$) and $z^{\pm}_n$ satisfies Proposition \ref{prop-continuous and property of saddle point reduction} with $l_F$ replaced by $\hat{l}$. So we have $\|x_n\|_\mathbf H\to\infty$. Denote by
\[
y_n=\frac{z_n}{\|z_n\|_\mathbf H},
\]
and $\bar{B}_n:=(1-\lambda_n)B_1+\lambda_nB(z_n)$, we have \begin{equation}\label{eq-the equation of yn}
Ay_n=\bar{B}_ny_n+\frac{o(\|z_n\|_\mathbf H)}{\|z_n\|_\mathbf H}.
\end{equation}
Decompose $y_n=y^{\pm}_n+y^0_n$ with $y^*_n=z^*_n/\|z_n\|_\mathbf H$, we have \begin{align*}
\|y^0_n\|_\mathbf H&=\frac{\|x_n\|_\mathbf H}{\|z_n\|_\mathbf H}\\
&\geq\frac{\|x_n\|_\mathbf H}{\|x_n\|_\mathbf H+\|z^+_n+z^-\|_\mathbf H}\\
&\geq\frac{(l-\hat{l})\|x_n\|_\mathbf H}{l\|x_n\|_\mathbf H+\|F'_\lambda(0)\|_\mathbf H}. \end{align*} That is to say
\begin{equation}\label{eq-y0n not to 0}
\|y^0_n\|_\mathbf H\geq c>0
\end{equation}
for some constant $c>0$ and $n$ large enough.
Since $B_1\leq B(z)\leq B_2$, we have $B_1\leq \bar{B}_n\leq B_2$.
Let $\mathbf H=\mathbf H^+_{A-\bar B_n}\bigoplus \mathbf H^-_{A-\bar B_n}$ with $A-\bar B_n$ is positive and negative define on $\mathbf H^+_{A-\bar B_n}$ and $\mathbf H^-_{A-\bar B_n}$ respectively. Re-decompose $y_n=\bar{y}^+_n+\bar{y}^-_n$ respect to $\mathbf H^+_{A-\bar B_n}$ and $\mathbf H^-_{A-\bar B_n}$. From Lemma \ref{lem-0 has a positive distance from sigma(A-B)} and \eqref{eq-the equation of yn}, we have
\begin{align}\label{eq-y0n to 0}
\| y^0_n\|^2_\mathbf H&\leq \| y_n\|^2_\mathbf H\nonumber\\
&\leq \frac{1}{\varepsilon} ((A-\bar{B}_n)y_n,\bar{y}^+_n+\bar{y}^-_n)_\mathbf H\nonumber\\
&\leq \frac{1}{\varepsilon} \frac{o(\|z_n\|_\mathbf H)}{\|z_n\|_\mathbf H}\|y_n\|_\mathbf H.
\end{align}
Since $\|z_n\|_\mathbf H\to \infty$ and $\|y_n\|=1$, we have $\|y^0_n\|_\mathbf H\to 0$ which contradicts to \eqref{eq-y0n not to 0}, so we have $\{z_n\}$ is bounded.
Secondly, we apply the topological degree theory to complete the proof. Since the solutions of (O.E.)$_\lambda$ are bounded, there is a number $R>0$ large eoungh,
such that all of the solutions $z_\lambda$ of (O.E.)$_\lambda$ are in the ball $B(0,R):=\{z\in \mathbf H| \|z\|_\mathbf H<R\}$. So we have the Brouwer degree
\[ deg (a'_1,B(0,R)\cap \mathbf H^0,0)= deg (a'_0,B(0,R)\cap \mathbf H^0,0)\neq 0,
\]
where $a_\lambda(x)=\frac{1}{2}(Az_\lambda(x),z_\lambda(x))_\mathbf H-F_\lambda(z_\lambda(x))$, $\lambda\in[0,1]$. That is to say (O.E.) has at least one solution.
$
\Box$
In Theorem \ref{thm-abstract thm 1 for the existence of solution}, the non-degeneracy condition of $B(z)$ is important to keep the boundedness of the solutions. The following theorem will not need this non-degeneracy condition, the idea is from \cite{Ji-2018}.
\begin{thm}\label{thm-abstract thm 3} Assume $A\in\mathcal O^0_e(-b,b)$. If $F$ satisfies conditions ($F_1$) and the following condition.\\ ($F^\pm_2$) There exists $M>0$, $B_\infty\in\mathcal{L}_s(\mathbf H, -b,b)$, such that \[ F'(z)=B_\infty z+r(z), \] with \[
\|r(z)\|_\mathbf H\leq M,\;\;\forall z\in\mathbf H, \] and \begin{equation}\label{eq-condition of r in ab-thm 3}
(r(z),z)_\mathbf H\to\pm\infty,\;\;\|z\|_\mathbf H\to\infty. \end{equation}
Then (O.E.) has at least one solution. \end{thm} \noindent{\bf Proof.} If $0\not\in \sigma(A-B_\infty)$, then with the similar method in Theorem \ref{thm-abstract thm 1 for the existence of solution}, we can prove the result. So we assume $0\in \sigma(A-B_\infty)$ and we only consider the case of ($F^-_2$). Since $0$ is an isolate eigenvalue of $A-B_\infty$ with finite dimensional eigenspace (see \cite{Wang-Liu-2016} for details), there exists $\eta>0$ such that \[ (-\eta,0)\cap\sigma(A-B_\infty)=\emptyset. \]
For any $\varepsilon\in (0,\eta)$, we have
$0\not\in \sigma(\varepsilon+A-B_\infty)$. Thus, with the similar method in Theorem \ref{thm-abstract thm 1 for the existence of solution}, we can prove that there exists $z_\varepsilon\in \mathbf H$ satisfying the following equation \begin{equation}\label{eq-equation of z-varepsilon} \varepsilon z_\varepsilon+(A-B_\infty)z_\varepsilon=r(z_\varepsilon). \end{equation} In what follows, We divide the following proof into two steps and $C$ denotes various constants independent of $\varepsilon$.
{\bf Step 1. We claim that $\|z_\varepsilon\|_\mathbf H\leq C$. } Since $z_\varepsilon$ satisfies the above equation, we have \begin{align*} \varepsilon (z_\varepsilon,z_\varepsilon)_\mathbf H&=-((A-B_\infty)z_\varepsilon,z_\varepsilon)_\mathbf H+(r(z_\varepsilon),z_\varepsilon)_\mathbf H\\
&\leq \frac{1}{\eta}\|(A-B_\infty)z_\varepsilon\|^2_\mathbf H+M\|z_\varepsilon\|_\mathbf H\\
&=\frac{1}{\eta}\|\varepsilon z_\varepsilon-r(z_\varepsilon)\|^2_\mathbf H+M\|z_\varepsilon\|_\mathbf H\\
&\leq \frac{\varepsilon^2}{\eta}\|z_\varepsilon\|^2_\mathbf H+C\|z_\varepsilon\|_\mathbf H +C. \end{align*}
So we have \[
\varepsilon\|z_\varepsilon\|_\mathbf H\leq C. \]
Therefore \begin{equation}\label{eq-boundedness of z-1}
\|(A-B_\infty)z_\varepsilon\|_\mathbf H=\|\varepsilon z_\varepsilon-r(z_\varepsilon)\|_\mathbf H\leq C. \end{equation} Now, consider the orthogonal splitting as defined in \eqref{eq-decomposition of space H}, \[ \mathbf H=\mathbf H^0_{A-B_\infty}\oplus\mathbf H^*_{A-B_\infty}, \] where $A-B_\infty$ is zero definite on $\mathbf H^0_{A-B_\infty}$, $\mathbf H^*_{A-B_\infty}$ is the orthonormal complement space of $\mathbf H^0_{A-B_\infty}$. Let $z_\varepsilon=u_\varepsilon+v_\varepsilon$ with $u_\varepsilon\in \mathbf H^0_{A-B_\infty}$ and $v_\varepsilon\in \mathbf H^*_{A-B_\infty}$. Since $0$ is an isolated point in $\sigma(A-B_\infty)$, from \eqref{eq-boundedness of z-1}, we have \begin{equation}\label{eq-boundedness of z-2}
\|v_\varepsilon\|_\mathbf H\leq C \end{equation} Additionally, since $r(z)$ and $v_\varepsilon$ are bounded, we have \begin{align}\label{eq-boundedness of z-3} (r(z_\varepsilon),z_\varepsilon)_\mathbf H&=(r(z_\varepsilon),v_\varepsilon)_\mathbf H+(r(z_\varepsilon),u_\varepsilon)_\mathbf H\nonumber\\
&=(r(z_\varepsilon),v_\varepsilon)_\mathbf H+(\varepsilon z_\varepsilon+(A-B_\infty)z_\varepsilon,u_\varepsilon)_\mathbf H\nonumber\\
&=(r(z_\varepsilon),v_\varepsilon)_\mathbf H+\varepsilon(u_\varepsilon,u_\varepsilon)_\mathbf H\nonumber\\
&\geq C. \end{align}
Therefor, from \eqref{eq-condition of r in ab-thm 3}, $\|u_\varepsilon\|_\mathbf H$ are bounded in $\mathbf H$ and we have proved the boundedness of $\|z_\varepsilon\|_\mathbf H$.
{\bf Step 2. Passing to a sequence of $\varepsilon_n\to 0$, there exists $z\in\mathbf H$ such that \[
\displaystyle\lim_{\varepsilon_n\to 0}\|z_{\varepsilon_n}-z\|_\mathbf H=0. \] } Different from the above splitting, now, we recall the projections $P^-_A,\;P^0_A$ and $P^+_A$ defined in \eqref{eq-projections} and the splitting $\mathbf H=\mathbf H^-\oplus\mathbf H^0\oplus\mathbf H^+$ with $\mathbf H^*=P^*_A$($*=\pm,0$). So $z_\varepsilon$ has the corresponding splitting \[ z_\varepsilon=z_\varepsilon^++z_\varepsilon^-+z_\varepsilon^0, \]
with $z_\varepsilon^*\in \mathbf H^*$ respectively. Since $\mathbf H^0$ is a finite dimensional space and $\|z_\varepsilon\|_\mathbf H\leq C$, there exists a sequence $\varepsilon_n\to 0$ and $z^0\in \mathbf H^0$, such that \[ \displaystyle\lim_{n\to\infty}z^0_{\varepsilon_n}=z^0. \]
For simplicity, we rewrite $z^*_n:=z^*_{\varepsilon_n}$, $A_n:=\varepsilon_n+A$ and $A^\pm_n:=A_n|_{\mathbf H^\pm} $. Since $z_\varepsilon$ satisfies \eqref{eq-equation of z-varepsilon}, we have \[ z^\pm_n=(A_n^\pm)^{-1}P^\pm_A F'(z^+_n+z^-_n +z^0_n). \]
Since $F$ satisfies ($F_1$), with the similar method used in Proposition \ref{prop-continuous and property of saddle point reduction}, for $n$ and $m$ large enough, we have
\begin{align*}
\|z^\pm_n-z^\pm_m\|_\mathbf H=&\|(A_n^\pm)^{-1}P^\pm_A F'(z_n)-(A_m^\pm)^{-1}P^\pm_A F'(z_m)\|_\mathbf H \\
\leq&\|(A_n^\pm)^{-1}P^\pm_A (F'(z_n)-F'(z_m))\|_\mathbf H+\|((A_n^\pm)^{-1}-(A_m^\pm)^{-1})P^\pm_A F'(z_m)\|_\mathbf H \\
\leq&\frac{l_F}{l}\|z_n-z_m\|_\mathbf H+\|((A_n^\pm)^{-1}-(A_m^\pm)^{-1})P^\pm_A F'(z_m)\|_\mathbf H. \end{align*} Since $(A_n^\pm)^{-1}-(A_m^\pm)^{-1}=(\varepsilon_m-\varepsilon_n)(A_n^\pm)^{-1}(A_m^\pm)^{-1}$ and $z_n$ are bounded in $\mathbf H$, we have \[
\|((A_n^\pm)^{-1}-(A_m^\pm)^{-1})P^\pm_A F'(z_m)\|_\mathbf H=o(1),\;\;n,m\to\infty. \] So we have \[
\|z^\pm_n-z^\pm_m\|_\mathbf H\leq \frac{l_F}{l-l_F}\|z^0_n-z^0_m\|_\mathbf H+o(1),\;\;n,m\to\infty, \]
therefor, there exists $z^\pm\in\mathbf H^\pm$, such that $\displaystyle\lim_{n\to\infty}\|z^\pm_n- z^\pm\|_\mathbf H=0$. Thus, we have \[
\displaystyle\lim_{n\to\infty}\|z_{\varepsilon_n}-z\|_\mathbf H=0, \] with $z=z^-+z^++z^0$. Last, let $n\to\infty$ in \eqref{eq-equation of z-varepsilon}, we have $z$ is a solution of (O.E.).$
\Box$ \begin{thm}\label{thm-abstract thm 2 for the multiplicity of solutions}
Assume $A\in\mathcal O^0_e(-b,b)$, $F$ satisfies ($F_1$) with $\pm l_F\not\in\sigma(A)$ and the following condition:\\ ($F^+_3$) There exist $B_3\in\mathcal{L}_s(\mathbf{H},-b,b)$ and $C\in\mathbb{R}$, such that \[
B_3>\beta:=\max\{\lambda|\lambda\in \sigma_A\cap(-\infty,l_F)\}, \] with \[ F(z)\geq\frac{1}{2}(B_3z,z)_\mathbf{H}-C,\;\;\forall z\in \mathbf H. \] Or ($F^-_3$) There exist $B_3\in\mathcal{L}_s(\mathbf{H},-b,b)$ and $C\in\mathbb{R}$, such that \[
B_3<\alpha:=\min\{\lambda|\lambda\in \sigma_A\cap(-l_F,\infty)\}, \] with \[ F(z)\leq\frac{1}{2}(B_3z,z)_\mathbf{H}+C,\;\;\forall z\in \mathbf H. \] Then (O.E.) has at least one solution. Further more, assume $F$ satisfies \\ ($F^\pm_4$) $F\in C^2(\mathbf H,\mathbb{R})$, $F'(0)=0$ and there exists $B_0\in\mathcal{L}_s(\mathbf{H},-b,b)$ with \begin{equation}\label{eq-twisted condition 1 in abstract thm 2} \pm(i_A(B_0)+\nu_A(B_0))<\pm i_A(B_3), \end{equation} such that \[
F'(z)=B_0z+o(\|z\|_\mathbf H),\;\;\|z\|_\mathbf H\to 0. \] Then (O.E.) has at least one nontrivial solution. Additionally, if \begin{equation}\label{eq-twisted condition 2 in abstract thm 2} \nu_A(B_0)=0 \end{equation} then (O.E.) has at least two nontrivial solutions.\end{thm} \noindent{\bf Proof.} We only consider the case of ($F^+_3$). According to the saddle point reduction, since $\pm l_F\not\in\sigma(A)$, we can choose $l\in(l_F,b)$ in \eqref{eq-projections} satisfying \[ [-l,-l_F]\cap\sigma(A)=\emptyset=[l_F,l]\cap\sigma(A). \] We turn to the function \[ a(x)=\frac{1}{2}(Az(x),z(x))-F(z(x)), \] where $z(x)=x+z^+(x)+z^-(x)$, $x\in \mathbf H^0$ and $z^\pm\in \mathbf H^\pm$. Denote by $w(x)=x+z^-(x)$ and write $z=z(x)$, $w=w(x)$ for simplicity. Since \begin{equation}\label{eq-eq 1 in abstract thm 2} a(x)=\left\{\frac{1}{2}(Aw,w)-F(w)\right\}+\left\{\frac{1}{2}[(Az,z)-(Aw,w)]-[F(z)-F(w)]\right\}. \end{equation}
By condition ($F^+_3$), we obtain \begin{equation}\label{eq-eq 2 in abstract thm 2} \frac{1}{2}(Aw,w)-F(w)\leq \frac{1}{2}((\beta-B_3)w,w)_\mathbf{H}+C, \end{equation} and the terms in the second bracket are equal to \begin{align}\label{eq-eq 3 in abstract thm 2} &\frac{1}{2}(Az^+,z^+)-\int^{1}_{0}(F'(sz^++w),z^+)ds\nonumber\\ =&\frac{1}{2}(Az^+,z^+)-(F'(z^++w),z^+)+\int^{1}_{0}(F'(z^++w)-F'(sz^++w),z^+)ds\nonumber\\ =&-\frac{1}{2}(Az^+,z^+)+\int^{1}_{0}(F'(z^++w)-F'(sz^++w),z^+)ds\nonumber\\
\leq&-\frac{1}{2}(Az^+,z^+)+\int^{1}_{0}(1-s)ds\cdot l_F\cdot\|z^+\|^2_\mathbf H\nonumber\\
\leq& -\frac{l-l_F}{2}\|z^+\|^2_\mathbf H, \end{align} where the last equality is from the fact that $Az^+=P^+F'(z^++w)$. From \eqref{eq-eq 1 in abstract thm 2},\eqref{eq-eq 2 in abstract thm 2} and \eqref{eq-eq 3 in abstract thm 2} we have
\begin{align*}
a(x)&\leq \frac{1}{2}((\beta-B_3)w,w)_\mathbf{H} -\frac{l-l_F}{2}\|z^+\|^2_\mathbf H+C\\
&\to-\infty,\; as \|x\|\to\infty. \end{align*}
Thus the function $-a(x)$ is bounded from below and satisfies the (PS) condition. So the maximum of $a$ exists and the maximum points are critical points of $a$.
In order to prove the second part, similarly, we only consider the case of ($F^+_3$) and ($F^+_4$). We only need to realize that $0$ is not a maximum point from \eqref{eq-twisted condition 1 in abstract thm 2}, so the maximum points discovered above are not $0$. In the last, if \eqref{eq-twisted condition 2 in abstract thm 2} is satisfied, we can use the classical three critical points theorem, since $0$ is neither a maximum nor degenerate and the proof is complete. $
\Box$
\begin{rem} (A). Theorem \ref{thm-abstract thm 2 for the multiplicity of solutions} is generalized from \cite[IV,Theorem 2.3]{Chang-1993}. In the first part of our Theorem, we do not need $F$ to be $C^2$ continuous.\\ (B). Theorem \ref{thm-abstract thm 2 for the multiplicity of solutions} is different from our former result in \cite[Theorem 3.6]{Wang-Liu-2016}. Here, we need the Lipschitz condition to keep the method of saddle point reduction valid, where, in \cite[Theorem 3.6]{Wang-Liu-2016}, in order to use the method of dual variation, we need the convex property. \end{rem}
\section{Applications in one dimensional wave equation}\label{section-applications} In this section, we will consider the following one dimensional wave equation \[ \left\{\begin{array}{ll}
\Box u\equiv u_{tt}-u_{xx}=f(x, t, u),\\
u(0,t)=u(\pi, t)= 0, \\
u(x, t+T)=u(x, t),\\
\end{array} \right.\forall (x, t)\in[0,\pi]\times S^1, \eqno(W.E.) \] where $T>0$, $S^1:=\mathbb{R}/T\mathbb{Z}$ and $f:[0,\pi]\times S^1\times\mathbb{R}\to\mathbb{R}$. In what follows we assume systematically that $T$ is a rational multiple of $\pi$. So, there exist coprime integers $(p,q)$, such that \[
T=\frac{2\pi q}{p}. \] Let \[
L^2:=\left\{u,u=\sum_{j\in\mathbb{N}^+,k\in\mathbb{Z}} u_{j,k}\sin jx\exp ik\frac{p}{q}t\right\}, \] where $i=\sqrt{-1}$ and $u_{j,k}\in \mathbb{C}$ with $u_{j,k}=\bar{u}_{j,-k}$, its inner product is \[ (u,v)_2=\sum_{j\in\mathbb{N}^+,k\in\mathbb{Z}}(u_{j,k},\bar{v}_{j,k}),\;u,v\in L^2, \] the corresponding norm is \[
\|u\|^2_2=\sum_{j\in\mathbb{N}^+,k\in\mathbb{Z}}|u_{j,k}|^2\;u,v\in L^2. \]
Consider $\Box$ as an unbounded self-adjoint operator on $L^2$. Its' spectrum set is \[
\sigma(\Box)=\{(p^2k^2-q^2j^2)/q^2|j\in\mathbb{N}^+,k\in\mathbb{Z}\}. \] It is easy to see $\Box$ has only one essential spectrum $\lambda_0=0$. Let $\Omega:=[0,\pi]\times S^1$, assume $f$ satisfying the following conditions.
\noindent($f_1$) $f\in C(\Omega\times\mathbb{R},\mathbb{R})$, there exist $ b\neq 0$ and $l_F\in(0,|b|)$, such that \[
|f_{ b}(x,t,u+v)-f_{ b}(x,t,u)|\leq l_F|v|,\;\;\forall (x,t)\in \Omega,\;u,v\in\mathbb{R}, \] where \[ f_{ b}(x,t,u):=f(x,t,u)-bu,\;\;\forall (x,t,u)\in\Omega\times \mathbb{R}. \]
Let the working space $\mathbf H:=L^2$ and the operator $A:=\Box-b\cdot I$, with $I$ the identity map on $\mathbf H$. Thus we have $A\in\mathcal O^0_e(-|b|,|b|)$. Denote $L^\infty:=L^\infty(\Omega, \mathbb{R})$ the set of all essentially bounded functions. For any $g\in L^\infty$, it is easy to see $g$ determines a bounded self-adjoint operator on $L^2$, by \[ u(x,t)\mapsto g(x,t)u(x,t),\;\;\forall u\in L^2, \]
without confusion, we still denote this operator by $g$, that is to say we have the continuous embedding $L^\infty\hookrightarrow \mathcal L_s(\mathbf H)$. Thus for any $g\in L^\infty\cap \mathcal L_s(\mathbf H,-|b|,|b|)$, we have the index pair ($i_A(g),\nu_A(g)$). Besides, for any $g_1,g_2\in L^\infty$, $g_1\leq g_2$ means that \[ g_1(x,t)\leq g_2(x,t),\;{\rm a.e.} (x,t)\in\Omega. \]
\noindent ($f_2$) There exist $g_1,g_2\in L^\infty\cap \mathcal L_s(\mathbf H,-|b|,|b|)$ and $g\in L^\infty(\Omega\times \mathbb{R},\mathbb{R})$, with \[ g_1\leq g_2,\;i_A(g_1)=i_A(g_2),\;\nu_A(g_2)=0, \] \[ g_1(x,t)\leq g(x,t,u)\leq g_2(x,t),\;\;\;{\rm a.e.} (x,t,z)\in\Omega\times\mathbb{R}, \] such that \[
f_{ b}(x,t,u)-g(x,t,u)u=o(|u|),\;|u|\to\infty,\;{\rm uniformly for }(x,t)\in \Omega. \]
We have the following results. \begin{thm}\label{thm-application-1} Assume $T$ is a rational multiple of $\pi$, $f$ satisfying ($f_1$) and ($f_2$), then (W.E.) has a weak solution. \end{thm} \noindent{\bf Proof of Theorem \ref{thm-application-1}.} Let \[ \mathcal{F}_b(x,t,u):=\int^u_0f_b(x,t,s)ds,\;\;\forall(x,t,u)\in\Omega\times\mathbb{R}, \] and \begin{equation}\label{eq-definition of F} F(u):=\int_\Omega \mathcal F_b(x,t,u(x,t))dxdt,\;\;\forall u\in \mathbf H. \end{equation}
It is easy to verify that $F$ will satisfies condition ($F_1$) and ($F_2$) if $f$ satisfies condition ($f_1$) and ($f_2$).
Thus, by Theorem \ref{thm-abstract thm 1 for the existence of solution}, the proof is complete.$
\Box$
Here, we give an example of Theorem \ref{thm-application-1}. \begin{eg}
For any $b\neq 0$, assume $\alpha,\beta\in (-|b|,|b|)$ and $ [\alpha,\beta]\cap \sigma(\Box-b)=\emptyset. $ Let \[
g(x,t,u):=\displaystyle\frac{\beta-\alpha}{2}\sin[\varepsilon_1\ln (|x|+|t|+|u|+1)]+\frac{\alpha+\beta}{2}, \] and $h\in C(\mathbb{R},\mathbb{R})$ is Lipschitz continuous with \[
h(u)=o(|u|),\;\;|u|\to\infty. \]
then \[ f(x,t,u):=bu+g(x,t,u)u+\varepsilon_2h(u) \] will satisfies condition ($f_1$) and ($f_2$) for $\varepsilon_1$ and $\varepsilon_2>0$ small enough. \end{eg}
\begin{thm}\label{thm-application-3} Assume $T$ is a rational multiple of $\pi$, $f$ satisfies ($f_1$) and the following condition,\\
($f^\pm_2$) There exists $g_\infty(x,t)\in L^\infty\cap\mathcal{L}_s(\mathbf H, -|b|,|b|)$ with \[
|f_b(x,t,u)-g_{\infty}(x,t)u|\leq M_{1}\;\;\forall (x,t,u)\in\Omega\times\mathbf R, \] and \begin{equation}\label{eq-condition of r}
\pm(f_b(x,t,u)-g_{\infty}(x,t)u,u)\geq c|u|,\;\;\forall (x,t,u)\in\Omega\times\mathbb R/[-M_{2},M_{2}], \end{equation} where $M_1,\;M_2,\;c>0$ are constants. Then (W.E.) has a weak solution. \end{thm}
\noindent{\bf Proof.} We only consider the case of $f^{-}_{2}$. Let $r(x,t,u):=f_b(x,t,u)-g_{\infty}(x,t)u$, then $r$ is bounded in $\mathbf H$. Generally speaking, from \eqref{eq-condition of r}, we cannot prove \eqref{eq-condition of r in ab-thm 3}, so we cannot use Theorem \ref{thm-abstract thm 3} directly. By checking the proof of Theorem \ref{thm-abstract thm 3}, in step 1, when we got \eqref{eq-boundedness of z-2}, \eqref{eq-condition of r in ab-thm 3} was only used to get the boundedness of $z^0_{\varepsilon}$. Now, with \eqref{eq-condition of r}, we can also get the boundedness of $z^0_{\varepsilon}$ from \eqref{eq-boundedness of z-2}. Recall that $\mathbf H=L^{2}(\Omega)$ in this section, from the boundedness of $z^{\pm}_{\varepsilon}$ in $\mathbf H$, we have the boundedness of $z^{\pm}_{\varepsilon}$ in $L^{1}(\Omega)$. On the other hand, since $\ker (A-g_{\infty})$ is a finite dimensional space, if $\|z^{0}_{\varepsilon}\|_{\mathbf H}\to\infty$, we have $\|z^{0}_{\varepsilon}\|_{L^{1}}\to\infty$, thus $\|z_{\varepsilon}\|_{L^{1}}\to\infty$. Therefor, we have the contradiction from \eqref{eq-boundedness of z-2} and \eqref{eq-condition of r}. So we have gotten the boundedness of $z^0_{\varepsilon}$. The rest part of the proof is similar to Theorem \ref{thm-abstract thm 3}, we omit it here.
\begin{eg} Here we give an example of Theorem \ref{thm-abstract thm 3}. For any $b\neq 0$, and $g_\infty\in C(\Omega)$ with \[
\|g_\infty\|_{C(\Omega)}<|b|. \] Let $r(u)=\varepsilon\arctan u$, then \[ f(x,t,u):=bu+g_\infty(x,t)u\pm r(u) \] will satisfies the conditions in Theorem \ref{thm-abstract thm 3} for $\varepsilon>0$ small enough. \end{eg}
Now, in order to use Theorem \ref{thm-abstract thm 2 for the multiplicity of solutions}, we assume $f$ satisfies the following conditions.
\noindent ($f^{\pm}_3$) There exists $g_3(x,t)\in L^\infty\cap\mathcal{L}_s(\mathbf H, -|b|,|b|)$, with \[
\pm g_3(x,t)>\max\{\lambda|\lambda\in \sigma_{(\pm A)}\cap (-\infty,l_F)\}, \] such that \[ \pm\mathcal{F}_b(x,t,u)\geq \frac{1}{2}(g_3(x,t)u,u)+c,\;\;\forall (x,t,u)\in\Omega\times\mathbb{R}, \] for some $c\in\mathbb{R}$.
\noindent ($f^\pm_4$) $f\in C^1(\Omega\times \mathbb{R},\mathbb{R})$, $f(x,t,0)\equiv 0,\;\forall(x,t)\in\Omega$ and \[ g_0(x,t):=f'_b(x,t,u),\;\;\forall (x,u)\in\Omega, \] with \[ \pm (i_A(g_0)+\nu_A(g_0))<\pm i_A(g_3). \] We have the following result. \begin{thm}\label{thm-application-2} Assume $T$ is a rational multiple of $\pi$. \\ (A.)If $f$ satisfies condition ($f_1$), ($f^+_3$)( or ($f^-_3$)), then (W.E.) has at least one solution.\\ (B.) Further more, if $f$ satisfies condition ($f^+_4$)( or ($f^-_4$)), then (W.E.) has at least one nontrivial solution. Additionally, if $\nu_A(g_0)=0$, then (W.E.) has at least two nontrivial solutions. \end{thm} The proof is to verify the conditions in Theorem \ref{thm-abstract thm 2 for the multiplicity of solutions}, we only verify the smoothness of $F(u)$ defined in \eqref{eq-definition of F}. From condition ($f_1$) and $f\in C^1(\Omega\times\mathbb{R})$, we have the derivative $f'_b(x,t,u)$ of $f_b$ with respect to $u$, satisfying \begin{equation}\label{eq-the boundedness of f'_b}
|f'_b(x,t,u)|\leq l_F,\;\;\forall (x,t,u)\in\Omega\times\mathbb{R}. \end{equation} For any $u,v\in\mathbf{H}$, \begin{align*} F'(u+v)-F'(u)&=f_b(x,t,u+v)-f_b(x,t,u)\\
&=f'_b(x,t,u)v+(f'_b(u+\xi v)-f'_b(u))v. \end{align*} From \eqref{eq-the boundedness of f'_b}, we have $f'_b(u+\xi v)-f'_b(u)\in \mathbf{H}$ and \[
\displaystyle\lim_{\|v\|_\mathbf{H}\to 0}\|f'_b(u+\xi v)-f'_b(u)\|_\mathbf{H}=0,\;\;\forall u\in\mathbf{H}. \] That is to say $F''(u)=f'_b(x,t,u)$ and $F\in C^2(\mathbf{H},\mathbb{R})$.
\begin{eg} In order to give an example for Theorem \ref{thm-application-2}, assume \begin{equation}\label{eq-spectrum of Box} \sigma(\Box)=\mathop{\cup}\limits_{n\in\mathbb{Z}}\{\lambda_n\}, \end{equation} with $\lambda_0=0$ and $\lambda_n<\lambda_{n+1}$ for all $n\in\mathbb{Z}$. Choose any $k\in\{2,3\cdots\}$. Let \[ g_0(x,t)\in C(\Omega,[\alpha,\beta]),\;{\rm with}\;[\alpha,\beta]\in(0,\lambda_k), \] and $h\in C(\mathbb{R},\mathbb{R})$ defined above. Define \[ g(x,t,u):=g_0(x,t)+\displaystyle(\lambda_k-g_0(x,t)-\varepsilon_1)\frac{2}{\pi}\arctan(\varepsilon_1u^2), \] then \[ f(x,t,u):=g(x,t,u)u+\varepsilon_2 h(u) \] will satisfies condition ($f_1$) and ($f^+_3$) with $b=\frac{\lambda_{k}}{2}$ and $\varepsilon_1,\varepsilon_2>0$ small enough. Further more, if $g_0,h$ are $ C^1$ continuous and $\beta<\lambda_{k-1}$, we have condition ($f^+_4$) is satisfied. Additionally, if $[\alpha,\beta]\cap\sigma(\Box)=\emptyset$, then $\nu_A(g_0)=0$. \end{eg}
\begin{rem} We can also use Theorem \ref{thm-abstract thm 1 for the existence of solution} , Theorem \ref{thm-abstract thm 3} and Theorem \ref{thm-abstract thm 2 for the multiplicity of solutions} to consider the radially symmetric solutions for the $n$-dimensional wave equation: \[ \left\{\begin{array}{ll}
\Box u\equiv u_{tt}-\vartriangle_x u=h(x,t,u), &t\in\mathbb{R},\;x\in B_R,\\
u(x,t)= 0, t\in\mathbb{R}, &t\in\mathbb{R},\;x\in \partial B_R,\\
u(x,t+T)=u(x,t), &t\in\mathbb{R},\;x\in B_R,\\
\end{array} \right. \eqno(n \textendash W.E.) \]
where $B_R=\{x\in\mathbb{R}^n,|x|<R\}$, $\partial B_R=\{x\in\mathbb{R}^n,|x|=R\}$, $n>1$ { and the nonlinear term $h$ is $T$-periodic in variable $t$}. Restriction of the radially symmetry allows us to know the nature of spectrum of the wave operator. Let $r=|x|$ and {$S^1:=\mathbb{R}/T$, if $h(x,t,u)=h(r,t,u)$} then the $n$-dimensional wave equation ($n$\textendash W.E.) can be transformed into: \[
\left\{\begin{array}{ll}
A_0u:=u_{tt}-u_{rr}-\frac{n-1}{r}u_r=h(r,t,u),\\
u(R,t)=0,\; \\
u(r,0)=u(r,T),\;u_t(r,0)=u_t(r,T),
\end{array} \right. \;\;\;(r,t)\in{\Omega:=[0,R]\times S^1}.\eqno(RS\textendash W.E.) \] $A_0$ is symmetric on $L^2(\Omega,\rho)$, where $\rho=r^{n-1}$ and \[
L^2(\Omega,\rho):=\left\{u|\|u\|^2_{L^2(\Omega,\rho)}:=\int_\Omega|u(t,r)|^2r^{n-1}dtdr<\infty\right\}. \]
By the asymptotic properties of the Bessel functions (see\cite{Watson-1952}), the spectrum of the wave operator can be characterized (see\cite[Theorem 2.1]{Schechter-1998}). Under some more assumption, the self-adjoint extension of $A_0$ has no essential spectrum, and we can get more solutions of (RS\textendash W.E.). \end{rem}
\end{document}
|
arXiv
|
{
"id": "1810.07933.tex",
"language_detection_score": 0.5839277505874634,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{center}
\large \textbf{On knot groups acting on trees}
\normalsize
\textsc{F.~A.~Dudkin$^1$, A.~S.~Mamontov$^1$}
\footnotetext[1]{The work was supported by the program of fundamental scientific researches of the SB RAS No I.1.1., project No 0314-2016-0001}
\end{center}
{\bf Abstract:} A finitely generated group $G$ acting on a tree with infinite cyclic edge and vertex stabilizers is called a generalized Baumslag--Solitar group {\it (GBS group)}. We prove that a 1-knot group $G$ is GBS group iff $G$ is a torus-knot group and describe all n-knot GBS groups for $n\geqslant 3$.
{\bf Keywords:} Knot group, GBS group, group acting on a tree, torus-knot group.
\section{Introduction}
An {\it $n$-knot group } is the fundamental group $\pi_1 (S^{n+2}-K^n,a)$ for an $n$-knot $K^n$ in $n+2$-dimenstional sphere $S^{n+2}$. Starting from a knot it is possible to construct Wirtinger presentation for its group with relations of the form $x_i^w=x^j$, where $x_i, x_j$ are letters, and $w$ is a word \cite{Kaw}. Now, let $G$ be a group represented by a set of generators and relations. When it is a knot group? Treating this just as a question of how to transform relations to a desired form is unfruitful. So to attempt the question we involve some known properties of knot groups and restrict the class of groups.
A finitely generated group $G$ acting on a tree with infinite cyclic edge and vertex stabilizers is called a generalized Baumslag--Solitar group {\it (GBS group)} \cite{forester2002}. By the Bass-Serre theorem, $G$ is representable as $\pi_1(\mathbb{B})$, the fundamental group of a graph of groups $\mathbb{B}$ (see \cite{Serr}) with infinite cyclic edge and vertex groups.
GBS groups are important examples of JSJ decompositions. JSJ decompositions appeared first in 3-dimensional topology with the theory of the characteristic submanifold by Jaco-Shalen and Johannson. These topological ideas were carried over to group theory by Kropholler for some Poincar\'{e} duality groups of dimension at least 3, and by Sela for torsion-free hyperbolic groups. In this group-theoretical context, one has a finitely generated group G and a class of subgroups $\mathcal{A}$ (such as cyclic groups, abelian groups, etc.), and one tries to understand splittings (i.e. graph of groups decompositions) of $G$ over groups in $\mathcal{A}$ (see \cite{GuLe}).
Given a $GBS$ group $G$, we can present the corresponding graph of groups by a labeled graph $\mathbb{A}=(A,\lambda)$, where $A$ is a finite connected graph (with endpoint functions $\partial_0,\partial_1\colon E(A)\to V(A)$) and $\lambda\colon E(A)\to \mathbb{Z}\setminus\{0\}$ labels the edges of $A$. The label $\lambda_e$ of an edge $e$ with the source vertex $v$ defines an embedding $\alpha_e\colon e\to v^{\lambda_e}$ of the cyclic edge group $\langle e \rangle$ into the cyclic vertex group $\langle v \rangle$ (for more details see \cite{ClayForII})
{\it The fundamental group} $\pi_1(\mathbb{A})$ of a {\it labeled graph} $\mathbb{A}=(A, \lambda)$ is given by generators and defining relations. Denote by $\overline{A}$ the graph obtained from $A$ by identifying $e$ with $\overline{e}$. A maximal subtree $T$ of the graph $\overline{A}$ defines the following presentation of the group $\pi_1(\mathbb{A})$
$$\left\langle \begin{array}{lcl}
g_v, v\in V(\overline{A}), &\|& g_{\partial_0(e)}^{\lambda(e)}=g_{\partial_1(e)}^{\lambda(\overline{e})}, e\in E(T),\\
t_e, e\in E(\overline{A})\setminus E(T) &\|&t_e^{-1}g_{\partial_0(e)}^{\lambda(e)}t_e=g_{\partial_1(e)}^{\lambda(\overline{e})}, e\in E(\overline{A})\setminus E(T) \end{array} \right\rangle$$ Generatos of first (second) type are called vertex (edge) elements. For different maximal subtrees, corresponding presentations define isomorphic groups.
It is sometimes useful to regard a GBS-group as a group obtained as follows: start with the group $\mathbb{Z}$, perform consecutive amalgamated products in accordance with the labels on the maximal subtree; finally, apply several times the construction of the HNN-extension (the number of times is equal to the number of the edges outside the maximal tree). In this approach, the standard theory of amalgamated products and HNN-extensions is applicable to the full extent. In particular, GBS-groups admit a normal form of an element and have no torsion.
If two labeled graphs $\mathbb{A}$ and $\mathbb{B}$ define isomorphic $GBS$ groups $\pi_1(\mathbb{A})\cong\pi_1(\mathbb{B})$ and $\pi_1(\mathbb{A})$ is not isomorphic to $\mathbb{Z}, \mathbb{Z}^2$ or Klein bottle group then there exists a finite sequence of {\it expansion} and {\it collapse} (see fig.1) moves connecting $\mathbb{A}$ and $\mathbb{B}$ \cite{forester2002}. A labeled graph is called {\it reduced} if it admits no collapse move (equivalently, the labeled graph contains no edges with distinct endpoints and labels $\pm 1$).
\begin{figure}
\caption{Expansion and collapse moves}
\end{figure}
An element $g$ from a GBS group $G$ is called elliptic if $g$ is conjugated with $a^k$, for some vertex generator $a$. The set consisting of all non-trivial elliptic elements is stable under conjugation, its elements have infinite order, and any two such elements are commensurable. These properties yield a homomorphism $\Delta$ from $G$ to the multiplicative group of non-zero rationals $\mathbb{Q}^\ast$, defined as follows.
Given $g\in G$, choose any non-trivial elliptic element $a$. There is a relation $g\cdot a^p\cdot g=a^q$, with $p, q$ non-zero, and define $\Delta(g)=\frac{p}{q}$. As pointed out in \cite{KropII} this definition is independent of the choices made ($a$ and the relation), and defines a {\it modular homomorphism}.
For different primes $p$ and $q$ let $T(p,q) = \langle x,y | x^p=y^q \rangle$ be a torus-knot group. It is easy to see that $T(p,q)$ is a $GBS$ group for any non-zero integers $p$ and $q$ (corresponding labeled graph has one edge with two different endpoints and labels $p$ and $q$).
A group is said to be {\it Hopfian} if any homomorphism of the group onto itself has trivial kernel, i.e. is an automorphism. Baumslag and Solitar \cite{BaSo} came up with a series of examples of two-generator one-relator non-Hopfian groups. In particular, such are the Baumslag--Solitar groups
$$BS(p,q)=\langle x,y |xy^{p}x^{-1}=y^{q} \rangle$$ where $p$ and $q$ are coprime integers, $p,q\neq 1$.
If a labeled graph $\mathbb{B}$ consists of one vertex and two inverse loops with labels $p$ and $q$, then $\pi_1(\mathbb{B})\cong BS(p,q)$. Therefore, every Baumslag--Solitar groups is a generalized Baumslag--Solitar group.
Main results of our work are listed below.
{\bf Theorem 1.} {\it Let $G$ be a GBS-group. Then $G$ is 1-knot group if and only if $G \simeq T(p,q)$.}
{\bf Theorem 2.} {\it Let $G$ be a GBS-group, $G\not\cong\mathbb{Z}$. Then $G$ is $n$-knot group for $n \geq 3$ if and only if $G$ is a homomorphic image of either $BS(m,m+1)$, where $m\geq 1$, or $T(p,q)$.}
All homomorphic images of $BS(m,m+1)$, where $m\geq 1$, and $T(p,q)$ described in terms of labeled graphs in lemmas 4 and 5 (see the proof of theorem 2).
M.Kervaire obtained the following necessary conditions for a group to be a knot-group.
{\bf Statement 1 \cite[14.1.1]{Kaw}}. {\it Let $\pi$ be $n$-knot group for $n \geq 1$. Then
1. $\pi$ is finitely generated
2. $\pi / \pi ' \simeq \mathbb{Z}$
3. $H_2(\pi)=0$
4. $\pi$ is a normal closure of a single element }
For $n \geq 3$ these conditions are also sufficient.
{\bf Statement 2 \cite[14.1.2]{Kaw}}. {\it If a group $G$ satisfies conditions 1-4 of lemma 1, then $G$ is $n$-knot group for $n\geq 3$.}
So our goal is to determine when these conditions are fulfilled in a GBS group.
\section{Preliminary results}
Given graph $A$, denote the number of edges out of maximal subtree by $\beta_1(A)$. Normal closure of an element $a$ in a group $G$ is denoted as $\langle \langle a \rangle \rangle _G$.
{\bf Lemma 1.} {\it Let $G$ be a GBS group such that $G/G' \simeq \mathbb{Z}$, then $\beta_1(A) \leqslant 1$.}
Proof. From a representation it is clear that $G/G' \simeq \mathbb{Z}^{\beta_1(A)}\times H$, where $H$ is a subgroup of $G/G'$ generated by images of vertex elements. Hence, $\beta_1(A) \leqslant 1$. Lemma is proved.
{\bf Lemma 2.} {\it Let $G=\pi_1(\mathbb{A})$ be a $GBS$-group such that $G= \langle \langle a \rangle \rangle _G$, and $\mathbb{A}$ be a labeled tree. Then $\mathbb{A}$ is a labeled segment.}
\begin{figure}
\caption{Labeled segment.}
\end{figure}
Proof. Assume the contrary. Then $\mathbb{A}$ has a trident subgraph.
\begin{figure}
\caption{Labeled trident.}
\end{figure}
Let $N$ be a normal closure of all vertex elements except end points $u,v,w$, then $G/N \simeq \mathbb{Z}_p * \mathbb{Z}_q * \mathbb{Z}_r$.
If $p,q,r$ are pairwise comprime, then for some normal subgroup $N_1$ containing $N$ and certain powers of $u,v,w$ we have $G/N_1 \simeq \mathbb{Z}_{p_1} * \mathbb{Z}_{q_1} * \mathbb{Z}_{r_1}$, where $p_1,q_1,r_1$ are different primes. By \cite{How} $G/N_1$ is not a normal closure of a single element, so neither is $G$, a contradiction.
Now let $(p,q)=d \not = 1$. GBS group $G= \langle \langle a \rangle \rangle _G$ has no torsion, therefore $G/G' \simeq \mathbb{Z}$. Thus we have $(G/G')/N \simeq Z_p \times Z_q \times Z_r$. With $d \not =1$ dividing $p$ and $q$ the later group cannot be a homomorphic image of $\mathbb{Z}$.
Using ideas of plateau from \cite{Levitt2015}, we prove.
{\bf Lemma 3.} {\it Keep notations of lemma 2. Then $l_i \bot k_j$ for all $i,j$.}
Proof. Suppose first that $j \leqslant i$.
Choose a pair $k_j, l_i$ so that $j \leqslant i$, $(k_j,l_i)=d>1$, $j$ is minimal and $i$ is maximal with these conditions. Let $p$ be a prime divisor of $d$. Then $k_x \perp p$ for $x < j$ and $l_y \perp p$ for $y > i$. Let $r$ be a maximal index such that $l_{r-1} \not \perp p$, $2 \leq r \leq j$. If there is no such index, then let $r=1$. In a similar way, choose minimal $f$, $f \geq i+1$ such that $p | k_f$.
\begin{figure}
\caption{Labeled segment, $j \leqslant i$.}
\end{figure}
Let $N$ be a normal closure of the following elements
$$N = \langle \langle a_1, \ldots a_{r-1}, a_{j+1}, \ldots a_i, a_{f+1} \ldots a_{s+1},a_r^p,\ldots a_j^p,a_{i+1}^p,\ldots , a_f^p \rangle \rangle$$
Then corresponding factor group $G_1=G / N$ is isomorphic to $\mathbb{Z}_p * \mathbb{Z}_p$, since relations
$$a_m^{k_{m+1}} = a_{m+1}^{l_{m+1}},$$
$$a_m^p = a_{m+1} ^p=1$$
and $p \perp k_{m+1}, p \perp l_{m+1}$ imply $a_m=a_{m+1}$, $m=r,\ldots j,i+1,\ldots , f$.
Hence $G_1 / G_1' \simeq \mathbb{Z}_p \times \mathbb{Z}_p$, which condtradicts $G/G' \simeq \mathbb{Z}$.
2. Now assume $i<j$ and $(k_j,l_i)=d>1$.
\begin{figure}
\caption{Labeled segment, $j>i$.}
\end{figure}
We may also assume that $j-i$ is minimal with this property. Then using 1 we obtain $k_r \perp l_f$, $r=i+1,\ldots j-1, f=i,\ldots , j-1$ and $k_j \perp l_{i+1}, \ldots l_{j-1}$. In particular, for a prime divisor $p$ of $d$ we have $p \perp k_{i+1}, \ldots k_{j-1}$ and $p \perp l_{i+1}, \ldots , l_{j-1}$. Let
$$N= \langle \langle a_2,\ldots , a_i, a_{j+1}, \ldots , a_s, a_{i+1}^p,\ldots , a_j^p \rangle \rangle.$$
Then as above we obtain
$$G/N \simeq \mathbb{Z}_{k_1} * \mathbb{Z}_p * \mathbb{Z}_{l_{s+1}}$$
Having $k_1 \perp l_i$ and $k_j \perp l_{s+1}$ we may assume that $k_1,p,l_{s+1}$ are different primes. By our assumption $G$, and hence $G/N$, is a normal closure of a single element, while $ \mathbb{Z}_{k_1} * \mathbb{Z}_p * \mathbb{Z}_{l_{s+1}}$ for different primes $k_1,p,l_{s+1}$ is not \cite{How}. This contradiction proves the lemma.
{\bf Lemma 4.} {\it Assume $G = \pi _1 (\mathbb{A})$, $\mathbb{A}$ is a labeled segment, and $l_i \perp k_j$ for all $i,j$. Then $G$ is a homomorphic image of some torus-knot group. }
Proof.
We prove by induction that $a_2 , \ldots a_{s-1}$ may be excluded from generators.
To check basis consider a group with generators $\{ a_1,a_2,a_3 \}$ and relations $a_1^{k_1}=a_2^{l_1}$, $a_2^{k_2}=a_3^{l_2}$. Choose $\alpha$ and $\beta$ such that $\alpha l_1+\beta k_2 =1$. Then $a_2=a_1^{\alpha k_1} a_3^{\beta l_2 }$.
For the induction step consider a group with generators $\{ a_1,a_2, \ldots, a_{s+1} \}$. By induction, $a_i = w_i (a_1,a_s)$, where $w_i$ is some word on letters $a_1,a_s$, and $i=2,\ldots s-1$. Choose $\alpha$ and $\beta$ so that $\alpha l_1\ldots l_{s-1}+\beta k_s =1$. Then $a_s=a_s^{\alpha l_1 \ldots l_{s-1}} a_s^{\beta k_s} = a_1^{\alpha k_1 \ldots k_{s-1}} a_{s+1}^{\beta l_s}$. So a group is generated by $a_1$ and $a_{s+1}$, which satisfy a relation $a_1^k=a_{s+1}^l$, with $k=k_1\ldots l_s \perp l=k_1\ldots l_s$. Therefore $G$ is a homomorphic image of some $\pi_1 (T_{k,l})$. The lemma is proved.
{\bf Lemma 5.} {\it If $G=\pi_1(\mathbb{A})$, $G=\langle \langle a \rangle \rangle_G$, and $\beta_1(A)=1$ then $\mathbb{A}$ is a labeled cycle with coprime labels $k_i \perp l_j$ for all $i,j$, $|{\displaystyle \prod_{i=1}^{s} k_i} - {\displaystyle \prod_{i=1}^{s} l_i}|=1$ (see fig.6) and $G= \langle \langle t \rangle \rangle_G$.}
\begin{figure}
\caption{Labeled cycle}
\end{figure}
Proof. If $\mathbb{A}$ is no a cycle, then there is a pendant vertex $u$. A graph $\mathbb{A}$ is reduced, and so the label $\lambda$ near $u$ is not $\pm 1$. Let $N$ be a normal closure of all vertices, except $u$. Then $G / G'N \simeq \langle t \rangle \times \langle u \rangle \simeq \mathbb{Z} \times \mathbb{Z}_{\lambda}$, a contradiction.
Let $d=(k_i,l_j)$ and $p$ be a prime divisor of $d$. In case $j <i$ let $N$ be a normal closure of $a_1,\ldots a_j,a_{j+1}^p,\ldots ,a_i^p,a_{i+1},\ldots , a_s$. Then $G / G'N \simeq \langle t \rangle \times \langle a_{j+1} \rangle \simeq \mathbb{Z} \times \mathbb{Z}_p$, as in lemma 3, a contradiction. The case $j \geq i$ is considered in a similar way.
Let $k= {\displaystyle \prod_{i=1}^{s} k_i}$, $l= {\displaystyle \prod_{i=1}^{s} l_i}$. If $k=1$, then $s=1$, because $\mathbb{A}$ is reduced. So $G \simeq \pi_1(\mathbb{A}) \simeq BS(1,l)$, $G /G' \simeq \mathbb{Z} \times \mathbb{Z}_{l-1}$. It follows that $l=2$, and $BS(1,2) = \langle \langle t \rangle \rangle$ satisfies the conclusion of the lemma.
If $k\not =1$ and $l\not =1$. Note that in $G$ we have an equality $t^{-1}a_i^kt=a_i^l$ for $1 \leq i \leq s$, and $l \perp k$.
A map $\phi _{[a_1,t]}: H=BS(k,l)=\langle a,r| r^{-1}a^kr=a^l \rangle \rightarrow G$
$$ \phi _{[a_1,t]}: \begin{cases} a \rightarrow a_1\\ r \rightarrow t[a_1,t] \end{cases}$$
is an embedding $BS(k,l) \hookrightarrow G$ \cite{DBS}, hence $|a_1|_{G/G'}=|a|_{H/H'}=|k-l|$. Therefore $|k-l|=1$. Moreover there is a subgroup $\langle t \rangle \times \langle a_1 \rangle$ in $G/G'$ isomorphic to $\mathbb{Z} \times \mathbb{Z}_{|k-l|}$. Note that $t^{-1}a_i^kta_i^{-k} \in \langle \langle t \rangle \rangle$, hence $a_i^{l-k} = a_i^{\pm 1} \in \langle \langle t \rangle \rangle$ for all $i$, and so $G = \langle \langle t \rangle \rangle$. The lemma is proved.
\section{Proof of theorems}
{\bf Proof of theorem 1}
Assume that $G$ is both a GBS-group and a 1-knot group.
1. By statement 1 $G/G' \simeq \mathbb{Z}$.
2. By lemma 1 $G\simeq \pi_1(\mathbb{A})$ and either $A$ is a tree or $\beta_1(A)=1$.
3. If $A$ is a tree, then $\Delta(G)=\{1\}$ and by \cite[Statement 2.5]{Levitt2007} $Z(G)=Z$.
4. By \cite[Corollary 6.3.6]{Kaw} $G\simeq T(p,q)$.
5. If $\beta_1(A)=1$, then by \cite[Theorem 6.3.9]{Kaw} $G=\pi_1 (\mathbb{A})$ is residually finite. By \cite[Corollary 7.7]{Levitt2015q} a GBS-group is residually finite iff either $G=BS(1,n)$ or $\Delta(G)=\{\pm 1\}$.
6. If $G=BS(1,n)$ then by Lemma 5 $n=2$ and $G=BS(1,2)$, and by \cite{Shalen} $|m|=|n|$, a contradiction.
7. If $\Delta(G)=\{\pm 1\}$, then there exist $a,g \in G$ such that $gag=a^{-1}$, and hence $G/G'$ has a torsion, a contradiction with $G/G' \simeq \mathbb{Z}$.
8. If $\Delta(G)=\{1\}$, then $Z(G)=\mathbb{Z}$ \cite[Statement 2.5]{Levitt2007}, and $G$ is isomorphic to $T(p,q)$ by \cite[Corollary 6.3.6]{Kaw}.
The Theorem 1 is proved.
{\bf Proof of theorem 2}
I. Assume that $G$ is a GBS-group, $G \not \simeq \mathbb{Z}$ and $G$ is $n$-knot group, $n\geq 3$.
1. By Statement 1 $G/G' \simeq \mathbb{Z}$, $G=\langle \langle a \rangle \rangle$.
2. By Lemmas 1-5 $G$ is a fundamental group of either some labeled segment (see fig.2) with $k_i \perp l_i$, or some labeled cycle (see fig.6) $k_i \perp l_i$.
3. If $G$ is a fundamental group of a segment on fig.2, then by Lemma 4 $G$ is a homomorphic image of $T(k,l)$.
4. If $G$ is a fundamental group of a cycle on fig.6 then by \cite[Theorem 1.1]{Levitt2015q} $G$ is a homomorphic image of a Baumslag Solitar group $BS(k,k+1)$.
II. Assume that $G$ is a homomorphic image of rank 2 of either $BS(m,m+1)$ or $T(p,q)$.
5. If $G$ is a homomorphic image of rank 2 of $BS(m,m+1)$, then by \cite[Theorem 1.1]{Levitt2015q}, using that $m \perp m+1$, we get $G=\pi_1(\mathbb{A})$, where $\mathbb{A}$ is a cycle, and ${\displaystyle \prod_{i=1}^{s} k_i} \perp {\displaystyle \prod_{i=1}^{s} l_i}$, i.e. $\mathbb{A}$ as in conclusion of Lemma 5. By \cite{Krop} we have $H_2(G)=0$. Then by Lemma 5 we obtain that properties 1-4 of Statement 1 hold, and hence by Lemma 2 $G$ is a group of $n$-knot, $n\geq 3$.
6. If $G$ is a homomorphic image of rank 2 of $T(p,q)$, then by \cite{Levitt2015} and \cite{Levitt2015q} $G=\pi_1(\mathbb{A})$, where $A$ is either a segment, or a circle or a lollipop. Moreover, from \cite{Levitt2015} it follows that $A$ is a segment from theorem 1.1 \cite{Levitt2015}. Then $A$ satisfies the conclusion of Lemma 4. Hence $G$ satisfies the conditions 1-4 of Statement 1, and by Statement 2 $G$ is $n$-knot group for $n\geq 3$.
The Theorem 2 is proved.
Acknowledgments. The authors are grateful to V.~G.~Bardakov and V.~A.~Churkin for valuable discussions.
\end{document}
|
arXiv
|
{
"id": "1807.06275.tex",
"language_detection_score": 0.6992961764335632,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[Upper central series for the group of unitriangular automorphisms...] {Upper central series for the group of unitriangular automorphisms of a free associative algebra}
\author{Valeriy G. Bardakov}
\address{Sobolev Institute of Mathematics, Novosibirsk State University, Novosibirsk 630090, Russia} \address{and}
\address{Laboratory of Quantum Topology, Chelyabinsk State University, Brat'ev Kashirinykh street 129,
Chelyabinsk 454001, Russia} \email{[email protected]}
\thanks{The first author thank the organizers of the Conference ``Groups, Geometry and Dynamics" (Almora, India, 2012) for this beautiful and interesting Conference.}
\author{Mikhail V. Neshchadim} \address{Sobolev Institute of Mathematics, Novosibirsk State University, Novosibirsk 630090, Russia} \email{[email protected]} \thanks{The authors was supported by Federal Target Grant ``Scientific and educational personnel of innovation Russia'' for 2009-2013 (government contract No. 02.740.11.0429). Also, this research was supported by the Indo-Russian DST-RFBR project grant DST/INT/RFBR/P-137 (No.~13-01-92697)}
\subjclass[2010]{Primary 16W20; Secondary 20E15, 20F14.}
\keywords{Free associative algebra, group of unitriangular automorphisms, upper central series}
\begin{abstract} We study some subgroups of the group of unitriangular automorphisms $U_n$ of a free associative algebra over a field of characteristic zero. We find the center of $U_n$ and describe the hypercenters of $U_2$ and $U_3$. In particular, we prove that the upper central series for $U_2$ has infinite length. As consequence, we prove that the groups $U_n$ are non-linear for all $n \geq 2$. \end{abstract}
\maketitle
\section{Introduction}
In this paper we consider a free associative algebra $A_n = K \langle x_1, x_2,$ $\ldots,$ $x_n \rangle$ over a field $K$ of characteristic zero. We assume that $A_n$ has unity. The group of $K$-automorphisms of this algebra, i.e. automorphisms that fix elements of $K$ is denoted by $\mathrm{Aut} \, A_n$. The group of tame automorphisms $\mathrm{TAut} \, A_n$ of $A_n$ is generated by the group of affine automorphism $\mathrm{Aff} \, A_n$ and the group of unitraingular automorphisms $U_n=U(A_n)$. From the result of Umirbaev \cite{U} follows that $\mathrm{Aut} \, A_3 \not= \mathrm{TAut} \, A_3$.
A question about linearity (i.e. about a faithful representation by finite dimensional matrices over some field) of $\mathrm{TAut} \, A_n$ was studied in the paper of Roman'kov, Chirkov, Shevelin \cite{R}, where it was proved that for $n \geq 4$ these groups are not linear. Sosnovskii \cite{S} proved that for $n \geq 3$ the group $\mathrm{Aut} \, P_n$ is not linear, where $P_n = K [x_1, x_2,$ $\ldots,$ $x_n ]$ is the polynomial algebra over a field $K$. His result follows from description of the upper central series for the group of unitriangular automorphisms $U(P_n)$ of $P_n$.
The structure of the present paper is the following. In Section 2 we introduce some notations, recall some facts on the automorphism group of free associative algebra and its subgroups. In the previous article \cite{B} we found the lower central series and the series of the commutator subgroups for $U_n$. In Section 3 we study the upper central series for $U_2$ and prove that the length of this series is infinite. Prove that $U_2$ is non-linear. In Section 4 we study the upper central series for $U_3$ and describe the hypercentral subgroups in the terms of some algebras. In Section 5 we find the center of $U_n$ for $n \geq 4$. Also, in Sections 4 and 5 we formulate some hypotheses and questions that are connected with the theory of non-commutative invariants in free associative algebra under the action of some subgroups of $U_n$.
\section{Some previous results and remark}
Let us recall definitions of some automorphisms and subgroups of $\mathrm{Aut} \, A_n$.
For any index $i\in \left\{1, \ldots, n \right\}$, a constant $\alpha \in K^* = K\backslash \{0\}$ and a polynomial $f = f(x_1, \ldots , \widehat{x_i}, \ldots ,x_n) \in A_n$ (where the symbol $\widehat{x_i}$ denotes that $x_i$ is not included in $f$) {\it the elementary automorphism } $\sigma (i, \alpha, f)$ is an automorphism in $\mathrm{Aut} \, A_n$ that acts on the variables $x_1, \ldots ,x_n$ by the rule: $$ \sigma (i, \alpha, f) :
\left\{ \begin{array}{lcl} x_i \longmapsto \alpha \, x_i + f, \,\, & \\ x_j \longmapsto x_j, \,\, & \mbox{if} & \,\, j \neq i. \\ \end{array} \right. $$ The group of tame automorphisms $\mathrm{TAut} \, A_n$ is generated by all elementary automorphisms.
The group of affine automorphisms $\mathrm{Aff} \, A_n$ is a subgroup of $\mathrm{TAut} \, A_n$ that consists of automorphisms $$ x_i\longmapsto a_{i1} x_1+ \ldots + a_{in} x_n + b_i,\,\, i=1, \ldots , n, $$ where $a_{ij}$, $b_i\in K$, $i,j=1, \ldots ,n$, and the matrix $(a_{ij})$ is a non-degenerate one. The group of affine automorphisms is the semidirect product $K^n \leftthreetimes \mathrm{GL}_n (K)$ and, in particular, embeds in the group of matrices $\mathrm{GL}_{n+1} (K)$.
The group of triangular automorphisms $T_n = T(A_n)$ of algebra $A_n$ is generated by automorphisms $$ x_i\longmapsto \alpha_i x_i+f_i(x_{i+1}, \ldots , x_n),\,\, i = 1,\ldots , n, $$ where $\alpha_i \in K^*$, $f_i\in A_n$ and $f_n\in K$. If all $\alpha_i =1$ then this automorphism is called {\it the unitriangular automorphism}. The group of unitriangular automorphisms is denoted by $U_n = U(A_n)$.
In the group $U_n$ let us define a subgroup $G_i$, $i = 1, 2, \ldots, n$ which is generated by automorphisms $$
\sigma (i, 1, f),~~ \, \mbox{where} \, f = f(x_{i+1}, \ldots , x_n) \in A_n. $$ Note that the subgroup $G_i$ is abelian and isomorphic to an additive subgroup of algebra $A_n$ that is generated by $x_{i+1}, \ldots , x_n$, $i = 1, \ldots, n-1$, and the subgroup $G_n $ is isomorphic to the additive group of the field $K$.
{\it The lower central series} of a group $G$ is the series $$ G = \gamma_1 G \geq \gamma_2 G \geq \ldots, $$ where $\gamma_{i+1} G = [\gamma_i G, G],$ $i = 1, 2, \ldots$. {\it The series of the commutator subgroups} of a group $G$ is the series $$ G = G^{(0)} \geq G^{(1)} \geq G^{(2)} \geq \ldots, $$ where $G^{(i+1)} = [G^{(i)}, G^{(i)}],$ $i = 0, 1, \ldots$. Here for subsets $H$, $K$ of $G$, $[H, K]$ denotes the subgroup of $G$ generated by the commutators $[h, k] = h^{-1} k^{-1} h k$ for $h \in H$ and $k \in K$.
Recall that the $k$-th hypercenter $Z_k = Z_k(G)$ of the upper central series of $G$ for the non-limited ordinal $k$ is defined by the rule $$ Z_k / Z_{k-1} = Z(G / Z_{k-1}) $$ or equivalently, $$
Z_k = \{ g \in G ~|~[g, h] \in Z_{k-1}~ \mbox{for all} ~h \in G \}, $$ and $Z_1(G) = Z(G)$ is the center of $G$. If $\alpha$ is a limit ordinal, then define $$ Z_{\alpha} = \bigcup_{\beta < \alpha} Z_{\beta}. $$
It was proved in \cite{B} that $U_n$ is a semidirect product of abelian groups: $$ U_n = (\ldots(G_1 \leftthreetimes G_2)\leftthreetimes \ldots ) \leftthreetimes G_n, $$ and the lower central series and the series of commutator subgroups of $U_n$ satisfy the following two properties respectvely:
1) For $n \geq 2$ $$ \gamma_2 U_n = \gamma_3 U_n = \ldots. $$ In particular, for $n \geq 2$ the group $U_n$ is not nilpotent.
2) The group $U_n$ is solvable of degree $n$ and the corresponding commutator subgroups have the form: $$
\begin{array}{l}
U_n= (\ldots(G_1 \leftthreetimes G_2)\leftthreetimes \ldots ) \leftthreetimes G_{n},\\
U_n^{(1)}= (\ldots(G_1 \leftthreetimes G_2)\leftthreetimes \ldots ) \leftthreetimes G_{n-1},\\
.........................................\\
U_n^{(n-1)}= G_1,\\
U_n^{(n)}= 1.
\end{array} $$
Yu. V. Sosnovskiy \cite{S} found the upper central series for the unitriangular group $U(P_n)$ of the polynomial algebra $P_n$. (Note, that he considered polynomials without free terms.) He proved that for $n \geq 3$ the group $U(P_n)$ has the upper central series of length $((n-1)(n-2)/2) \omega + 1$ for any field $K$, where $\omega$ is the first limit ordinal. If $\mathrm{char} \, K = 0$ then the hypercenters of $U(P_4)$ have the form\\
$Z_{k} = \{ (x_1 + f_1(x_3, x_4), x_2, x_3) ~|~ \mathrm{deg}_{x_3} f_1(x_3, x_4) \leq k -1 \}$,\\
$Z_{\omega} = \{ (x_1 + f_1(x_3, x_4), x_2, x_3) \}$,\\
$Z_{\omega + k} = \{ (x_1 + f_1(x_2, x_3, x_4), x_2, x_3) ~|~ \mathrm{deg}_{x_2} f_1(x_2, x_3, x_4) \leq k \}$,\\
$Z_{2 \omega} = \{ (x_1 + f_1(x_2, x_3, x_4), x_2, x_3) \}$,\\
$Z_{2\omega + k} = \{ (x_1 + f_1(x_2, x_3, x_4), x_2 + f_2(x_3, x_4), x_3) ~|~ \mathrm{deg}_{x_3} f_2(x_3, x_4) \leq k - 1 \}$,\\
$Z_{3 \omega} = \{ (x_1 + f_1(x_2, x_3, x_4), x_2 + f_2(x_3, x_4), x_3) \}$,\\
$Z_{3 \omega + 1} = U(P_4)$,\\ where $k = 1, 2, \ldots$ runs over the set of natural numbers and $f_1, f_2, f_3$ are arbitrary polynomials in $P_n$ which depend on the corresponding variables.
\section{Unitriangular group $U_2$}
Let $A_2 = K \langle x, y \rangle$ be the free associative algebra over a field $K$ of characteristic zero with the variables $x$ and $y$. Then $$ U_2= \left\{
\varphi= \left( x + f(y), y + b \right) ~|~ f(y) \in K\langle y \rangle,\,\,b \in K \right\} $$ is the group of unitriangular automorphisms of $A_2$.
It is not difficult to check the following Lemma
\begin{lemma}\label{l:form} 1) If $\varphi = \left( x + f(y), y + b \right) \in U_2$, then its inverse is equal to $$
\varphi^{-1}= \left( x - f(y - b), y - b \right); $$
2) if $ \varphi= \left( x + f(y), y + b \right)$ and $\psi= \left( x + h(y), y + c \right)\in U_2$, then the following formulas hold:
-- the formula of conjugation $$
\psi^{-1}\varphi \psi = \left( x + h(y) - h(y + b) + f(y + c), y + b \right), $$
-- the formula of commutation $$
\varphi^{-1}\psi^{-1}\varphi \psi = \left( x + h(y) - h(y + b) + f(y + c) - f(y), y \right). $$
\end{lemma}
Using this Lemma we can describe the center of $U_2$.
\begin{lemma} \label{l:c} The center of $U_2$ has the form
$Z(U_2) = \left\{ \varphi = \left( x + a, y \right) ~|~ a \in K \right\}$. \end{lemma}
\begin{proof} If $\varphi = \left( x + a, y \right)$, then from the formula of conjugation (see Lemma~\ref{l:form}) follows
that $\varphi \in Z(U_2)$. To prove the inverse inclusion suppose $ \varphi= \left( x+f(y), y+b \right) \in Z(U_2)$. Using the formula of conjugation we get $$
\varphi= \left( x + f(y), y + b \right) = \psi^{-1} \varphi \psi = \left( x + h(y) - h(y + b) + f(y + c)
y + b \right), $$ for any automorphism $\psi = \left( x + h(y), y + c \right) \in U_2$, i.e. $ f(y) = h(y) - h(y + b) + f(y + c)$. Taking $h = 0$ we get $f(y) = f(y + c)$ for every $c \in K$. Hence, $f(y) = a \in K$. We have only relation $0 = h(y) - h(y + b)$. Since $h(y)$ is arbitrary, it follows that $b=0$.
\end{proof}
\begin{lemma}\label{l:com} The following properties hold true in $U_2$.
1)
$[U_2, U_2] = \left\{\varphi = \left( x + f(y), y \right) ~|~ f(y) \in K \langle y \rangle \right\}$.
2) If $\varphi = \left( x + f(y), y \right)$, where $f(y) \in K \langle y \rangle \setminus K$, then $$
C_{U_2}(\varphi) = \left\{\left( x + h(y), y \right) ~|~ h(y)\in K \langle y \rangle \right\}, $$ where $C_{U_2}(\varphi)$ is the centralizer of $\varphi$ in $U_2$, i.e.
$C_{U_2}(\varphi) = \{ \psi \in U_2 ~|~ \psi \varphi = \varphi \psi\}$.
3) If $\varphi= \left( x,y+b \right)$, $b \in K$, then
$C_{U_2}(\varphi) = \left\{ \left( x + a, y + c \right) ~|~ a, c \in K \right\}$.
\end{lemma}
\begin{proof} 1) Let $\varphi = \left( x + f(y), y + b \right)$, $\psi = \left( x + h(y), y + c \right) \in U_2$. By the formula of commutation $$
\varphi^{-1} \psi^{-1} \varphi \psi = \left( x + h(y) - h(y + b) + f(y + c) - f(y), y \right). $$ It is easy to see that any element of $K \langle y \rangle$ can be represented in a form $r(y + d) - r(y)$ for some $r(y) \in K \langle y \rangle$ and $d\in K$. Hence, $$
K \langle y \rangle = \{ h(y) - h(y + b) + f(y + c) - f(y)~|~h, f \in K \langle y \rangle,~~b, c \in K\} $$ and 1) is true.
2) Let $ \varphi = \left( x+f(y), y \right)$, $f(y) \in K \langle y \rangle \setminus K$ and $ \psi = \left( x + h(y), y + c \right)$ be an arbitrary element of $C_{U_2}(\varphi)$. Using the formula of conjugation we get $$
\varphi= \left( x + f(y), y \right) = \psi^{-1} \varphi \psi = \left( x + f(y + c), y \right). $$ Hence $c = 0$.
3) Let $ \varphi = \left( x, y + b \right)$ and $ \psi = \left( x + h(y), y + c \right)$ be an arbitrary element of $C_{U_2}(\varphi)$. Using the formula of conjugation we get $$
\varphi = \left( x, y + b \right) = \psi^{-1} \varphi \psi = \left( x + h(y) - h(y + b), y + b \right). $$ Hence $h(y)=a \in K$.
\end{proof}
\begin{lemma}\label{l:co} If $s$ is a non-negative integer, then the $(s+1)th$ hypercenter of $U_2$ has the form $$
Z_{s+1}(U_2) = \left\{\varphi = \left( x+f(y),y \right)
~|~ f(y) \in K \langle y \rangle,\,\, \mathrm{deg} f(y) \leq s \right\}. $$ \end{lemma}
\begin{proof} If $s = 0$, then the assertion follows from Lemma \ref{l:c}. Suppose that for $s + 1$ the assertion holds true. We now prove it for $s + 2$. Let $$
\varphi = \left( x + f(y), y + b \right)\in Z_{s+2}(U_2). $$ Using the formula of commutation (see Lemma \ref{l:form}) for $\varphi$ and $\psi= \left( x + h(y), y + c \right)$ we get $$
\varphi^{-1}\psi^{-1}\varphi \psi = \left( x + h(y) - h(y + b) + f(y + c) - f(y), y \right). $$ If $b \neq 0$ and since $h(y)$ is an arbitrary polynomial of $K \langle y \rangle$, then $h(y) - h(y + b) + f(y + c) - f(y)$ represents arbitrary element of $K \langle y \rangle$. But it is not possible since for any automorphism $\varphi = (x + f(y), y) \in Z_{s+1}$ the degree of $f(y)$ is not bigger than $s$. Hence $b = 0$. Since $\mathrm{deg} f(y+c)-f(y) \leq s$ and $c$ is an arbitrary element of $K$, we have $\mathrm{deg} f(y) \leq s + 1$. So the inclusion from left to right is proved. The inverse inclusion is evident.
\end{proof}
Hence, from Lemma \ref{l:co} $$
Z_{\omega}(U_2)=\left\{\varphi= \left( x + f(y), y \right) ~|~ f(y) \in K\langle y \rangle \right\}, $$ and using Lemma \ref{l:com} we have $Z_{\omega}(U_2)=[U_2, U_2]$. Therefore $$
Z_{\omega+1}(U_2)=U_2. $$
\begin{corollary} The group $U_2$ is not linear. \end{corollary}
\begin{proof} We know (see, for example \cite{G}) that if a linear group does not contain torsion, then the length of the upper central series of this group is finite. But we proved that the length of the upper central series for $U_2$ is equal to $\omega + 1$. Hence, the group $U_2$ is not linear. \end{proof}
Since $U_2$ is a subgroup of $\mathrm{Aut} \, A_2$, we have proven that this group is not linear too. Using the fact that if $P_2$ is a polynomial algebra with unit over $K$, then $\mathrm{Aut} \, A_2 = \mathrm{Aut} \, P_2$ (see, for example \cite{C}).
\begin{corollary} Let $n \geq 2$. Then the groups $\mathrm{Aut} \, A_n$ and $\mathrm{Aut} \, P_n$ are not linear. \end{corollary}
It follows from the fact that $\mathrm{Aut} \, A_2 \leq \mathrm{Aut} \, A_n$ and $\mathrm{Aut} \, P_2 \leq \mathrm{Aut} \, P_n$ for all $n \geq 2$.
\begin{remark} In \cite{S} the author considered the polynomials without free terms and proved that $\mathrm{Aut} \, P_3$ is not linear. Using his method it is not difficult to prove that if $P_2$ contains free terms, then $\mathrm{Aut} \, P_2$ is not linear over any field of arbitrary characteristic. \end{remark}
\section{Unitriangular group $U_3$}
The group $U_3$ is equal to $$
U_3 = \{ (x_1 + f_1, x_2 + f_2, x_3 + f_3) ~|~f_1 = f_1(x_2, x_3) \in K \langle x_2, x_3 \rangle, $$ $$ f_2 = f_2(x_3) \in K \langle x_3 \rangle, f_3 \in K \}. $$
Define an algebra $S$ as subalgebra of $K \langle x_2, x_3 \rangle$ $$
S = \{ f(x_2, x_3)\in K \langle x_2, x_3 \rangle ~|~ f(x_2 + g(x_3), x_3 + h) = f (x_2, x_3)~ $$ $$ \mbox{for any}~g(x_3) \in K \langle x_3 \rangle,~~h \in K \} $$ Hence, $S$ is a subalgebra of fixed elements under the action of the group $$
\{ (x_2 + g, x_3 + h) ~|~g = g(x_3) \in K \langle x_3 \rangle, h \in K \} $$ which is isomorphic to $U_2$. The set $S$ is a subalgebra of $A_3$.
Define a set of commutators $$ c_1 = [x_2, x_3],~~c_{k+1} = [c_k, x_3],~~k = 1, 2, \ldots, $$ where $[a, b] = a b - b a$ is the ring commutator. Using induction on $k$, it is not difficult to check the following result.
\begin{lemma} The commutators $c_k$, $k = 1, 2, \ldots,$ lie in $S$. \end{lemma}
{\bf Hypothesis 1.} The algebra $K \langle c_1, c_2, \ldots \rangle$ is equal to $S$.
Note that the elements $c_1, c_2, \ldots$ are free generators of $K \langle c_1, c_2, \ldots \rangle$ (see \cite[p.~62]{Co}).
\begin{theorem} The center $Z(U_3)$ of the group $U_3$ is equal to $$
Z(U_3) = \{ (x_1 + f_1, x_2, x_3) ~|~f_1 = f_1(x_2, x_3) \in S \}. $$ \end{theorem}
\begin{proof} The inclusion $\supseteq$ is evident. Let $$ \varphi = (x_1 + f_1(x_2, x_3), x_2 + f_2(x_3), x_3 + f_3) $$ be some element in $Z(U_3)$ and $$ \psi = (x_1 + g_1(x_2, x_3), x_2 + g_2(x_3), x_3 + g_3) $$ be an arbitrary element of $U_3$. Since $\varphi \psi = \psi \varphi$ then we have the equalities\\
$x_1 + g_1(x_2, x_3) + f_1(x_2 + g_2(x_3), x_3 + g_3) = x_1 + f_1(x_2, x_3) + g_1(x_2 + f_2(x_3), x_3 + f_3),$\\
$x_2 + g_2(x_3) + f_2(x_3 + g_3) = x_2 + f_2(x_3) + g_2(x_3 + f_3),$\\
$x_3 + g_3 + f_3 = x_3 + f_3 + g_3.$ \\
\noindent The third equality holds for all $x_3$, $g_3$ and $f_3$. Rewrite the first and the second equality in the form\\
$g_1(x_2, x_3) + f_1(x_2 + g_2(x_3), x_3 + g_3) = f_1(x_2, x_3) + g_1(x_2 + f_2(x_3), x_3 + f_3),$\\
$g_2(x_3) + f_2(x_3 + g_3) = f_2(x_3) + g_2(x_3 + f_3).$\\
\noindent Let $g_1 = x_2$, $g_2 = g_3 = 0$. Then $$ x_2 + f_1(x_2, x_3) = f_1(x_2, x_3) + x_2 + f_2(x_3). $$ Hence $f_2(x_3) = 0$.
Let $g_1 = x_3$, $g_2 = g_3 = 0$. Then $$ x_3 + f_1(x_2, x_3) = f_1(x_2, x_3) + x_3 + f_3. $$ Hence $f_3 = 0.$ We have only one condition $$ f_1(x_2 + g_2(x_3), x_3 + g_3) = f_1(x_2, x_3), $$ i.e. $f_1 \in S$.
\end{proof}
Let us define the following subsets in the algebra $A_2 = K\langle x_2, x_3 \rangle$:\\
$S_1 = S$,\\
$S_{m+1} = \{ f \in A_2 ~|~f^{\varphi} - f \in S_m~ \mbox{for all}~ \varphi \in U_2 \}$, $m = 1, 2, \ldots$,\\
$S_{\omega} = \bigcup\limits_{m = 1}^{\infty} S_{m}$, \\
$S_{\omega+1} = \{ f \in A_2 ~|~f^{\varphi} - f \in S_\omega~ \mbox{for all}~ \varphi \in U_2 \}$,\\
$S_{\omega+m+1} = \{ f \in A_2 ~|~f^{\varphi} - f \in S_{\omega+m}~ \mbox{for all}~ \varphi \in U_2 \}$, $m = 1, 2, \ldots$,\\
$S_{2\omega} = \bigcup\limits_{m = 1}^{\infty} S_{\omega+m}$, \\
$R_{m} = \{ f = f(x_3) \in K \langle x_3 \rangle ~|~\mathrm{deg} \, f \leq m \}$, $m = 0, 1, \ldots ,$\\
$R_{\omega} = \bigcup\limits_{m = 0}^{\infty} R_{m}$. \\ It is not difficult to see that all $S_k$ are modules over $S$.
\begin{remark} If we consider the homomorphism $$ \pi : K \langle x_1, x_2, x_3 \rangle \longrightarrow K [x_1, x_2, x_3], $$ then\\
$S^{\pi} = K$,\\
$S_{m+1}^{\pi} = \{ f \in K[x_3] ~|~\mathrm{deg} \, f \leq m \}$, $m = 1, 2, \ldots$,\\
$S_{\omega}^{\pi} = K[x_3]$, \\
$S_{\omega+m+1}^{\pi} = \{ f \in K[x_2, x_3] ~|~\mathrm{deg}_{x_2} \, f \leq m \}$, $m = 0, 1, \ldots$,\\
$S_{2\omega}^{\pi} = K[x_2, x_3]$, \\
$R_{m}^{\pi} = \{ f \in K[x_3] ~|~\mathrm{deg} \, f \leq m \}$, $m = 1, 2, \ldots ,$\\
$R_{\omega}^{\pi} = K[x_3]$. \\
\end{remark}
\begin{theorem} The following equalities hold \begin{equation}\label{eq:m}
Z_{m} = \{ (x_1 + f_1(x_2, x_3), x_2, x_3)~|~ f_1 \in S_m \},~~ m = 1, 2, \ldots, 2\omega, \end{equation} \begin{multline}\label{eq:2m}
Z_{2\omega+m} = \{ (x_1 + f_1(x_2, x_3), x_2 + f_2(x_3), x_3)~|~ f_1 \in k\langle x_2, x_3 \rangle, f_2(x_3) \in R_m \},\\ m = 1, 2, \ldots, \omega, \end{multline} \begin{equation}\label{eq:3m} Z_{3\omega + 1} = U_3. \end{equation} \end{theorem}
\begin{proof} We use induction on $m$. To prove (\ref{eq:m}) for $m+1$, we assume that for all $m$ such that $1 \leq m < \omega$ equality (\ref{eq:m}) holds.
If $$ \varphi = (x_1 + f_1(x_2, x_3), x_2 + f_2(x_3), x_3 + f_3) \in Z_{m+1} $$ and $$ \psi = (x_1 + g_1(x_2, x_3), x_2 + g_2(x_3), x_3 + g_3) \in U_{3}, $$ then for some $$ \theta = (x_1 + h_1(x_2, x_3), x_2, x_3) \in Z_{m} $$ holds $\varphi \, \psi = \psi \, \varphi \, \theta$. Acting on the generators $x_1$, $x_2$, $x_3$ by $\varphi \, \psi$ and $\psi \, \varphi \, \theta$ we have two relations \begin{multline}\label{eq:1} f_1(x_2 + g_2(x_3), x_3 + g_3) - f_1(x_2, x_3) = h_1(x_2, x_3) + \\ + g_1(x_2 + f_2(x_3), x_3 + f_3) - g_1(x_2, x_3), \end{multline} \begin{equation}\label{eq:2} g_2(x_3) + f_2(x_2 + g_3) = f_2(x_3) + g_2(x_2 + f_3). \end{equation}
If $g_2 = 0$, then the relation (\ref{eq:2}) has the form $$ f_2(x_2 + g_3) = f_2(x_3). $$ Since $g_3$ is an arbitrary element of $K$, then $f_2 \in K$. But in this case (\ref{eq:2}) has the form $$ g_2(x_3 + f_3) = g_2(x_3). $$ Since $g_2(x_3)$ is an arbitrary element of $K\langle x_3 \rangle$, then $f_3 = 0$ and (\ref{eq:1}) has the form\\ \begin{equation}\label{eq:3} f_1(x_2 + g_2(x_3), x_3 + g_3) - f_1(x_2, x_3) = h_1(x_2, x_3) + g_1(x_2 + f_2, x_3) - g_1(x_2, x_3). \end{equation}\\ Let $g_1 = x_2^{N}$ for some natural number $N$. Using the homomorphism $$ \pi : K \langle x_1, x_2, x_3 \rangle \longrightarrow K[x_1, x_2, x_3] $$ and the equality $\mathrm{deg}_{x_2}\, h_1^{\pi} = 0$ we see that if $f_2 \not= 0$, then $$ \mathrm{deg}_{x_2}\, \left( f_1(x_2 + g_2(x_3), x_3 + g_3) - f_1(x_2, x_3) \right)^{\pi} = N - 1. $$ Since $N$ is any non-negative integer, then $f_2 = 0$ and $$ f_1(x_2 + g_2(x_3), x_3 + g_3) - f_1(x_2, x_3) \in S_m, $$ i.e. $f_1(x_2, x_3) \in S_{m+1}$ and we have proven the equality (\ref{eq:m}) for $m+1$: $$
Z_{m+1} = \{ (x_1 + f_1(x_2, x_3), x_2, x_3)~|~ f_1 \in S_{m+1} \}. $$
To prove (\ref{eq:m}) for $\omega+m+1$ assume that for all $\omega+m$ such that $1 \leq m < \omega$ equality (\ref{eq:m}) holds.
If $\varphi \in Z_{\omega+m+1}$ and $\psi \in U_3$ then for some $\theta \in Z_{\omega+m}$ we have $\varphi \, \psi = \psi\, \varphi \, \theta$ that give the relations (\ref{eq:1}) and (\ref{eq:2}). As in the previous case we can check that $f_2 \in K$, $f_3 = 0$ and (\ref{eq:1})--(\ref{eq:2}) are equivalent to (\ref{eq:3}). Let $g_1 = x_2^{N}$ for some natural number $N$. Using the homomorphism $$ \pi : K \langle x_1, x_2, x_3 \rangle \longrightarrow K [x_1, x_2, x_3] $$ and the inequality $\mathrm{deg}_{x_2}\, h_1^{\pi} \leq m - 1$ we see that if $f_2 \not= 0$, then $$ \mathrm{deg}_{x_2}\, \left( f_1(x_2 + g_2(x_3), x_3 + g_3) - f_1(x_2, x_3) \right)^{\pi} = N - 1 $$ for $N \geq m + 1$. But the degree of the left hand side is bounded. Hence $f_2 = 0$ and we have $$ f_1(x_2 + g_2(x_3), x_3 + g_3) - f_1(x_2, x_3) \in S_{\omega+m}, $$ i.e., $f_1(x_2, x_3) \in S_{m+1}$ and we have proven the equality (\ref{eq:m}) for $\omega+m+1$: $$
Z_{\omega+m+1} = \{ (x_1 + f_1(x_2, x_3), x_2, x_3)~|~ f_1 \in S_{\omega+m+1} \}. $$
To prove (\ref{eq:2m}) for $m+1$ assume that for all $m$ such that $1 \leq m < \omega$ equality (\ref{eq:2m}) holds.
If $\varphi \in Z_{2\omega+m+1}$, $\psi \in U_3$, then for some $\theta \in Z_{2\omega+m}$ we have $\varphi \, \psi = \psi\, \varphi \, \theta$. If $$ \varphi = (x_1 + f_1(x_2, x_3), x_2 + f_2(x_3), x_3 + f_3), $$ $$ \psi = (x_1 + g_1(x_2, x_3), x_2 + g_2(x_3), x_3 + g_3) $$ and $$ \theta = (x_1 + h_1(x_2, x_3), x_2 + h_2(x_3), x_3), $$ then we have the relations
\begin{equation*} \begin{split} x_1 + g_1(x_2, x_3) + f_1(x_2 + g_2(x_3), x_3 + g_3) & = x_1 + h_1(x_2, x_3) + f_1(x_2 + h_2(x_3), x_3) +\\ & + g_1(x_2 + h_2(x_3) + f_2(x_3), x_3 + f_3), \end{split} \end{equation*}
$$ x_2 + g_2(x_3) + f_2(x_3 + g_3) = x_2 + h_2(x_3) + f_2(x_3) + g_2(x_3 + f_3). $$ Since $h_1$ is an arbitrary element of $K \langle x_2, x_3 \rangle$, then we must consider only the second relation which is equal to $$ f_2(x_3 + g_3) - f_2(x_3) = h_2(x_3) + g_2(x_3 + f_3) - g_2(x_3). $$ Since $\mathrm{deg} \, h_2 \leq m$ and $g_2(x_3)$ is any element of $K \langle x_3 \rangle$, then $f_3 = 0$. Hence, $$ f_2(x_3 + g_3) - f_2(x_3) = h_2(x_3). $$ From this equality follows that $\mathrm{deg} \, f_2 \leq m + 1$. We have proven the equality (\ref{eq:2m}) for $m+1$: $$
Z_{2\omega+m+1} = \{ (x_1 + f_1(x_2, x_3), x_2 + f_2(x_3), x_3)~|~ f_1 \in K \langle x_2, x_3 \rangle, f_2(x_3) \in R_{m+1} \}. $$
To prove (\ref{eq:2m}) we note that $$ [U_3, U_3] \subseteq \{ (x_1 + f_1(x_2, x_3), x_2 + f_2(x_3), x_3) \}. $$
\end{proof}
We described the hypercenters of $U_3$ in the terms of the algebras $S_m$ and $S_{\omega+m}$. It is interesting to find sets of generators for these algebras. To do it we must give answers on the following questions.
{\bf Question 1} (see Hypothesis 1). Is it true that $$ S = K\langle c_1, c_2, \ldots \rangle? $$
{\bf Question 2.} Is it true that for all $m \geq 1$ the following equalities are true $$
S_{m+1} = \{ f \in K\langle S, x_3 \rangle ~|~\mathrm{deg}_{x_3} \, f \leq m \}? $$
{\bf Question 3.} Is it true that $$ \bigcup_{m=1}^{\infty} S_m = K\langle S, x_3 \rangle? $$
{\bf Question 4.} Is it true that for all $m \geq 1$ the following equalities are true $$
S_{\omega+m} = \{ f \in K \langle S, x_3, x_2 \rangle ~|~\mathrm{deg}_{x_2} \, f \leq m \}? $$
If $R$ is the Specht algebra of $A_2$, i.e. the subalgebra of $A_2$ that is generated by all commutators $$ [x_2, x_3],~~[[x_2, x_3], x_3],~~[[x_2, x_3], x_2], \ldots $$ then the following inclusions hold
\begin{equation}\label{eq:4}
S_{m+1} \subset \{ f \in K \langle R, x_3 \rangle ~|~\mathrm{deg}_{x_3} \, f \leq m \}, \end{equation}
\begin{equation}\label{eq:5}
S_{\omega+m} \subset \{ f \in K \langle R, x_3, x_2 \rangle ~|~\mathrm{deg}_{x_2} \, f \leq m \}, \end{equation} for all $m \geq 1$. It follows from the fact that $K \langle x_2, x_3 \rangle$ is a free left $R$-module
with the set of free generators $$ x_2^{\alpha} x_3^{\beta}, ~~~\alpha, \beta \geq 0. $$ Note that the inclusion (\ref{eq:4}) is strict. It follows from
\begin{proposition} The commutators $$ [x_2, \underbrace{x_3, \ldots, x_3}_k, x_2] = [c_k, x_2],~~k \geq 1, $$ do not lie in $S_{m}$, $m \geq 1$. \end{proposition}
\begin{proof} Indeed, for the automorphism $$ \varphi = (x_2 + g_2(x_3), x_3 + g_3) $$ of the algebra $K \langle x_2, x_3 \rangle$ we have $$ [c_k, x_2]^{\varphi} = [c_k^{\varphi}, x_2^{\varphi}] = [c_k, x_2 + g_2(x_3)] = [c_k, x_2] + [c_k, g_2(x_3)]. $$ If $g_2(x_3) = x_3^{N}$, then\\
\begin{align*} [c_k, x_3^{N}] & = c_k \, x_3^{N} - x_3^{N} c_k = (c_k x_3 - x_3 c_k) x_3^{N-1} + x_3 c_k x_3^{N-1} - x_3^{N} c_k \\
& = c_{k+1} x_3^{N-1} + x_3 (c_k x_3^{N-1} - x_3^{N-1} c_k) \\
& =c_{k+1} x_3^{N-1} + x_3 \left(( c_k x_3 - x_3 c_k) x_3^{N-2} + x_3 c_k x_3^{N-2} - x_3^{N-1} c_k \right)\\
& =c_{k+1} x_3^{N-1} + x_3 c_{k+1} x_3^{N-2} + x_3^{2} ( c_k x_3^{N-2} - x_3^{N-2} c_k )\\
& = \ldots \\
& = \sum_{p+q = N-1} x_3^p c_{k+1} x_3^q.\\ \end{align*}
Hence, for $\varphi = (x_2 + g_2(x_3), x_3 + g_3)$ we have $$ [c_k, x_2]^{\varphi} - [c_k, x_2] = \sum_{p+q = N-1} x_3^p c_{k+1} x_3^q. $$ If $$ g_2(x_3) = \sum_{n=0}^{N} a_n x_3^n, $$ then $$ [c_k, x_2]^{\varphi} - [c_k, x_2] = \sum_{n=1}^{N-1} a_n \sum_{p+q=n-1} x_3^p c_{k+1} x_3^q. $$ If $[c_k, x_2] \in S_m$ for some $m$, then $$ [[c_k, x_2], \varphi ] \equiv [c_k, x_2]^{\varphi} - [c_k, x_2] \in S_{m-1}. $$ Let $$ \psi = (x_2 + h_2(x_3), x_3 + h),~~~\varphi = (x_2 + x_3^N, x_3). $$ Then
\begin{align*} [[[c_k, x_2], \varphi ], \psi ] & = \left[\sum_{p+q=N-1} x_3^p c_{k+1} x_3^q, \psi \right] \\
& = \sum_{p+q=N-1} (x_3 + h)^p c_{k+1} (x_3 + h)^q - \sum_{p+q=N-1} x_3^p c_{k+1} x_3^q \\
& = \sum_{p+q=N-1} \left( \sum_{l=0}^{p-1} C_p^l x_3^l h^{p-l} \right) c_{k+1}
\left( \sum_{r=0}^{q} C_{q-1}^r x_3^r h^{q-r} \right)\\ \end{align*}
has the degree $N-2$ on the variable $x_3$. Continuing this process we are getting that if $\mathrm{deg} \, g_2(x_3) = N$, then $$ [c_k, x_2, \varphi, \psi_1, \ldots, \psi_{N-1}] \in S $$ for every $\psi_1, \ldots, \psi_{N-1} \in U_2$. Hence,\\
$[c_k, x_2, \varphi, \psi_1, \ldots, \psi_{N-1}] \in S,$\\
$...............................$ \\
$[[c_k, x_2, \varphi ] \in S_N,$\\
$[c_k, x_2] \in S_{N+1}.$\\
Using the similar ideas we can prove that $$ [c_k, x_2] \not\in S_{N}. $$
Since we can take arbitrary number $N > m$, then $$ [c_k, x_2] \not\in S_{m}, ~~m = 1, 2, \ldots $$ \end{proof}
\section{Center of the unitriangular group $U_n$, $n \geq 4$}
In this section we prove the following assertion
\begin{theorem} Any automorphism $\varphi$ in the center $Z(U_n)$ of $U_n$ has a form $$
\varphi=
\left( x_1 + f(x_2, \ldots, x_n), x_2, \ldots, x_n \right), $$ where the polynomial $f$ is such that $$ f(x_2 + g_2, \ldots, x_n + g_n) = f(x_2, \ldots, x_n) $$ for every $ g_2 \in K \langle x_3, \ldots, x_n \rangle$, $g_3\in K \langle x_4, \ldots, x_n \rangle, \ldots, g_n \in K$.
\end{theorem}
We will assume that $U_{n-1}$ includes in $U_n$ by the rule $$
U_{n-1} = \left\{ \varphi = \left( x_1, x_2 + g_2, \ldots, x_n + g_n \right) \in U_n ~|~
g_2 \in K \langle x_3, \ldots, x_n \rangle,\,\, \ldots, g_n \in K \right\}. $$ Hence we have the following sequence of inclusions for the subgroups $U_k$, $k=3,\ldots,n$ $$ U_n \geq U_{n-1} \geq \ldots \geq U_{3}. $$
In this assumption we can formulate Theorem by the following manner $$ Z(U_n) = \left\{
\varphi= \left( x_1 + f(x_2, \ldots, x_n), x_2, \ldots, x_n \right) ~|~ f^{U_{n-1}}=f \right\}, $$ where $$
f^{U_{n-1}} = \{ f^{\psi} ~|~ \psi \in U_{n-1} \}. $$
\begin{proof} Let $$ \varphi= \left( x_1 + f_1, x_2 + f_2, \ldots, x_n + f_n \right) \in Z(U_n) $$ and $$ \psi= \left( x_1 + g_1, x_2 + g_2, \ldots, x_n + g_n \right) $$ be an arbitrary element of $U_n$. Then $x_k^{\varphi\psi} = x_k^{\psi\varphi}$ for all $k = 1, 2, \ldots, n$. In particular, if $k = 1$, then
\begin{equation}\label{eq:11} ( x_1 + f_1)^\psi = ( x_1 + g_1)^\varphi. \end{equation}
Put $g_1 = x_2$, $g_2 = g_3 = \ldots = 0$. Then this relation has the form $$
x_1 + x_2 + f_1 = x_1 + f_1 + x_2 + f_2. $$ Hence, $f_2 = 0$.
Analogously, putting $g_1 = x_3$, $g_2 = g_3 = \ldots = 0$, we get $f_3 = 0$. Hence, $f_2 = f_3 = \ldots = f_n = 0$.
The relation (\ref{eq:11}) for arbitrary $\psi$ has the form $$ x + g_1(x_2, \ldots, x_n) + f_1( x_2 + g_2, \ldots, x_n + g_n) = x + f_1( x_2, \ldots, x_n) + g_1(x_2, \ldots, x_n). $$ Hence, $$ f_1( x_2+g_2, \ldots, x_n + g_n) = f_1( x_2, \ldots, x_n). $$
\end{proof}
Let us define the notations $$
\zeta U_n = \left\{ f(x_2, \ldots, x_n) \in K \langle x_2, \ldots, x_n \rangle ~|~ f^{U_{n-1}} = f \right\}, $$ $$
\zeta U_{n-1} = \left\{ f(x_3, \ldots, x_n) \in K \langle x_3, \ldots, x_n \rangle ~|~ f^{U_{n-2}} = f \right\}, $$ $$ ...................................................................................... $$ $$
\zeta U_{3} = \left\{ f(x_{n-2}, x_{n-1}, x_n) \in K \langle x_{n-2},x_{n-1},x_n \rangle ~|~ f^{U_{2}} = f \right\}. $$ Note that $\zeta U_{3} = S$.
We formulate the next hypothesis on the structure of the algebras $\zeta U_k$, $k=3, \ldots, n$.
{\bf Hypothesis 2.} The following inclusions hold $$ \zeta U_4 \subseteq K \langle \zeta U_3, x_{n-2} \rangle, $$ $$ \zeta U_5\subseteq K \langle \zeta U_4, x_{n-3} \rangle, $$ $$ ................................., $$ $$ \zeta U_n\subseteq K \langle \zeta U_{n-1}, x_{2} \rangle. $$
Recall that by Hypothesis 1 we have $$ \zeta U_3 = K \langle c_1, c_2, \ldots \rangle, $$ where $c_1 = [x_{n-1}, x_n]$, $c_{k+1} = [c_k, x_n]$, $k = 1, 2, \ldots $
\begin{proposition} If Hypotheses 1 and 2 are true, then the following equality holds $$ \zeta U_k = K \langle c_1, c_2, \ldots \rangle,\,\,\,k = 3, 4, \ldots, n. $$ \end{proposition}
\begin{proof} For $k = 4$ Hypothesis 2 has the form $$ \zeta U_4 \subseteq K \langle \zeta U_3, x_{n-2} \rangle, $$ i.e., every polynomial $f\in \zeta U_4$ can be represented in the form $$ f = F(x_{n-2}, c_1, c_2, \ldots, c_N) $$ for some non-negative integer $N$. Applying the automorphism $$ \psi = \left( x_1, x_2, \ldots, x_{n-2} + g_{n-2} , x_{n-1}, x_n \right), $$ we get $$ F(x_{n-2} + g_{n-2}, c_1, c_2, \ldots, c_N) = F(x_{n-2}, c_1, c_2, \ldots, c_N). $$ Here $g_{n-2} = g_{n-2}(x_{n-1}, x_n)$ is an arbitrary element of $K \langle x_{n-1}, x_n \rangle$. Putting in this equality $g_{n-2} = c_{N+1}$ and $x_{n-2} = 0$ we have $$ F(c_{N+1}, c_1, c_2, \ldots, c_N) = F(0, c_1, c_2, \ldots, c_N). $$ Since $c_1, c_2, \ldots $ are free generators,
$F$ does not contain the variable $x_{n-2}$. Hence $$ \zeta U_4 = K \langle c_1, c_2, \ldots \rangle. $$ Analogously, we can prove the equality $$ \zeta U_k = K \langle c_1, c_2, \ldots \rangle,\,\,\,k=3, 4, \ldots, n. $$ \end{proof}
We see that the description of the hypercenters of $U_n$ (see Hypotheses 1 and 2) is connected with the theory of non-commutative invariants in free associative algebra under the action of some subgroups of $U_n$. We will study these invariants in next papers.
\end{document}
|
arXiv
|
{
"id": "1405.6288.tex",
"language_detection_score": 0.5832586884498596,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\title{Rational $D(q)$-quadruples}
\author{Goran Dra\v zi\' c\fnref{fnote1}} \ead{[email protected]} \author{Matija Kazalicki\fnref{fnote2}} \ead{[email protected]} \fntext[fnote1]{Faculty of Food Technology and Biotechnology, University of Zagreb, Pierottijeva 6, 10000 Zagreb, Croatia.} \fntext[fnote2]{Department of Mathematics, University of Zagreb, Bijeni\v cka cesta 30, 10000 Zagreb, Croatia.}
\begin{abstract} For a rational number $q$, a \emph{rational $D(q)$-$n$-tuple} is a set of $n$ distinct nonzero rationals $\{a_1, a_2, \dots, a_n\}$ such that $a_ia_j+q$ is a square for all $1 \leqslant i < j \leqslant n$. For every $q$ we find all rational $m$ such that there exists a $D(q)$-quadruple with product $a_1a_2a_3a_4=m$. We describe all such quadruples using points on a specific elliptic curve depending on $(q,m).$ \end{abstract}
\begin{keyword} Diophantine $n$-tuples\sep Diophantine quadruples\sep Elliptic curves \sep Rational Diophantine $n$-tuples \end{keyword}
\end{frontmatter}
\linenumbers
\section{Introduction}
Let $q\in \mathbb{Q}$ be a nonzero rational number. A set of $n$ distinct nonzero rationals $\lbrace a_1, a_2, \dots, a_n\rbrace$ is called a rational $D(q)$-$n$-tuple if $a_ia_j+q$ is a square for all $1 \leqslant i<j\leqslant n.$ If $\lbrace a_1,a_2,\dots,a_n\rbrace$ is a rational $D(q)$-$n$-tuple, then for all $r\in \mathbb{Q},\lbrace ra_1,ra_2,\dots,ra_n\rbrace$ is a $D(qr^2)$-$n$-tuple, since $(ra_1)(ra_2)+qr^2=(a_1a_2+q)r^2$. With this in mind, we restrict to square-free integers $q.$ If we set $q=1$ then such sets are called rational Diophantine $n$-tuples.
The first example of a rational Diophantine quadruple was the set $$\left\lbrace\frac{1}{16}, \frac{33}{16}, \frac{17}{4}, \frac{105}{16}\right\rbrace$$ found by Diophantus, while the first example of an integer Diophantine quadruple, the set \[ \lbrace 1,3,8,120 \rbrace \] is due to Fermat.
In the case of integer Diophantine $n$-tuples, it is known that there are infinitely many Diophantine quadruples (e.g. $\{k-1, k+1, 4k, 16k^3-4k\},$ for $k\geq 2$). Dujella \cite{dujella2004there} showed there are no Diophantine sextuples and only finitely many Diophantine quintuples, while recently He, Togb\' e and Ziegler \cite{he2019there} proved there are no integer Diophantine quintuples, which was a long standing conjecture.
Gibbs \cite{gibbs2006some} found the first example of a rational Diophantine sextuple using a computer, and Dujella, Kazalicki, Miki\' c and Szikszai \cite{dujella2017there} constructed infinite families of rational Diophantine sextuples. Dujella and Kazalicki parametrized Diophantine quadruples with a fixed product of elements using triples of points on a specific elliptic curve, and used that parametrization for counting Diophantine quadruples over finite fields \cite{dujella2016diophantine} and for constructing rational sextuples \cite{dujella2017more}. There is no known rational Diophantine septuple.
Regarding rational $D(q)$-$n$-tuples, Dujella \cite{dujella2000note} has shown that there are infinitely many rational $D(q)$-quadruples for any $q\in \mathbb{Q}.$ Dujella and Fuchs in \cite{dujella2012problem} have shown that, assuming the Parity Conjecture, for infinitely squarefree integers $q\neq 1$ there exist infinitely many rational $D(q)$-quintuples. There is no known rational $D(q)$-sextuple for $q\neq a^2, a\in \mathbb{Q}.$
Our work uses a similar approach Dujella and Kazalicki had in \cite{dujella2016diophantine} and \cite{dujella2017more}.
Let $\{a,b,c,d\}$ be a rational $D(q)$-quadruple, for a fixed nonzero rational $q,$ such that $$ ab+q=t_{12}^2,\quad ac+q=t_{13}^2,\quad ad+q=t_{14}^2,$$ $$bc+q=t_{23}^2,\quad bd+q=t_{24}^2,\quad cd+q=t_{34}^2.$$
Then $(t_{12},t_{13},t_{14},t_{23},t_{24},t_{34},m=abcd)\in \mathbb{Q}^7$ defines a rational point on the algebraic variety $\mathcal{C}$ defined by the equations $$(t_{12}^2-q)(t_{34}^2-q)=m,$$ $$(t_{13}^2-q)(t_{24}^2-q)=m,$$ $$(t_{14}^2-q)(t_{23}^2-q)=m.$$
The rational points $(\pm t_{12}, \pm t_{13}, \pm t_{14}, \pm t_{23}, \pm t_{24}, \pm t_{34}, m)$ on $\mathcal{C}$ determine two rational $D(q)$ quadruples $\pm (a,b,c,d)$ $\Big($specifically, $a^2=\frac{(t_{12}^2-q)(t_{13}^2-q)}{t_{23}^2-q}\Big)$ if $a,b,c,d$ are rational, distinct and nonzero.
Any point $(t_{12},t_{13},t_{14},t_{23},t_{24},t_{34},m) \in\mathcal{C}$ corresponds to three points $Q'_1=(t_{12},t_{34}), Q'_2=(t_{13},t_{24})$ and $Q'_3=(t_{14},t_{23})$ on the curve\[ \mathcal{D}_m\colon (X^2-q)(Y^2-q)=m. \] If $\mathcal{D}_m(\mathbb{Q})=\emptyset,$ there are no rational $D(q)$-quadruples with product of elements equal to $m,$ so we assume there exists a point $P_1=(x_1,y_1) \in \mathcal{D}_m(\mathbb{Q}).$
The curve $\mathcal{D}_m$ is a curve of genus $1$ unless $m=0$ or $m=q^2,$ which we assume from now on. Since we also assumed a point $P_1 \in \mathcal{D}_m(\mathbb{Q}),$ the curve $\mathcal{D}_m$ is birationally equivalent to the elliptic curve \[ E_m\colon W^2=T^3+(4q^2-2m)T^2+m^2T \] via a rational map $f\colon \mathcal{D}_m \to E_m$ given by \begin{align*} T&=(y_1^2-q)\cdot\frac{2x_1(y^2-q)x+(x_1^2+q)y^2+x_1^2y_1^2-2x_1^2q-y_1^2q}{(y-y_1)^2}, \\ W&=T\cdot\frac{2y_1x(q-y^2)+2x_1y(q-y_1^2)}{y^2-y_1^2}. \end{align*} Note that $f$ maps $(x_1,y_1)$ to the point at infinity $\mathcal{O}\in E_m(\mathbb{Q})$, it maps $(-y_1,x_1)$ to a point of order four, $R=(m,2mq) \in E_m(\mathbb{Q})$, and maps $(-x_1,y_1)$ to \[ S=\left(\frac{y_1^2(x_1^2-q)^2}{x_1^2},\frac{qy_1(x_1^2+y_1^2)(x_1^2-q)^2}{x_1^3}\right)\in E_m(\mathbb{Q}), \] which is generically a point of infinite order.
We have the following associations \[ (a,b,c,d)\dashleftarrow \rightarrow \text{a point on }\mathcal{C}(\mathbb{Q})\longleftrightarrow (Q'_1,Q'_2,Q'_3) \in \mathcal{D}_m(\mathbb{Q})^3. \] In order to obtain a rational $D(q)$-quadruple from a triple of points on $\mathcal{D}_m(\mathbb{Q}),$ we must satisfy the previously mentioned conditions: $a,b,c,d$ must be rational, mutually disjoint and nonzero.
It is easy to see that if one of them is rational, then so are the other three (i.e. $b=\frac{t_{12}^2-q}{a}$), and that they will be nonzero when $m\neq 0,$ since $m=abcd.$
The elements of the quadruple $(a,b,c,d)$ corresponding to the triple of points $(Q'_1, Q'_2, Q'_3)$ are distinct, if no two of the points $Q'_1, Q'_2, Q'_3$ can be transformed from one to another via changing signs and/or switching coordinates. For example, the triple $(t_{12},t_{34}),(-t_{34},t_{12}),(t_{14},t_{23})$ would lead to $a=d.$ This condition on points in $\mathcal{D}_m$ is easily understood on points in $E_m.$
Assume $P\in E_m \leftrightarrow (x,y)\in \mathcal{D}_m,$ that is, $f(x,y)=P.$ Then \begin{equation}\label{ness}
S-P\leftrightarrow (-x,y),\quad P+R\leftrightarrow (-y,x). \end{equation} The maps $P\mapsto S-P$ and $P\mapsto P+R$ generate a group $G$ of translations on $E_m$, isomorphic to $\mathbb{Z}/2\mathbb{Z} \times \mathbb{Z}/4\mathbb{Z},$ and $G$ induces a group action on $E_m({\overline{\Q}}).$ In order to obtain a quadruple from the triple $(Q_1,Q_2,Q_3)\in E_m(\mathbb{Q})^3$, such that the elements of the quadruple are distinct, the orbits $G\cdot Q_1, G\cdot Q_2, G\cdot Q_3$ must be disjoint. This is because the set of points in $\mathcal{D}_m$ corresponding to $G\cdot P$ is $\{(\pm x,\pm y),(\pm y,\pm x)\}.$ We say that such a triple of points satisfies the non-degeneracy criteria.
Let $\overline{\mathcal{D}}_m$ denote the projective closure of the curve $\mathcal{D}_m$ defined by \[ \overline{\mathcal{D}}_m \colon (X^2-qZ^2)(Y^2-qZ^2)=mZ^4. \] The map $f^{-1} \colon E_m \to \overline{\mathcal{D}}_m$ is a rational map, and since the curve $E_m$ is smooth, the map is a morphism \cite[II.2.1]{silverman2009arithmetic}. The map $x\circ f^{-1}\colon E_m \to \mathbb{A}^1$ given by \[ x\circ f^{-1}(P)=\frac{X\circ f^{-1}(P)}{Z\circ f^{-1}(P)} \] has a pole in points $P_0$ such that $f^{-1}(P_0)=[1:0:0],$ and is regular elsewhere. The map $y\circ f^{-1}\colon E_m \to \mathbb{A}^1$ given by \[ y\circ f^{-1}(P)=\frac{Y\circ f^{-1}(P)}{Z\circ f^{-1}(P)} \] has a pole in points $P_2$ such that $f^{-1}(P_2)=[0:1:0],$ and is regular elsewhere. We define the rational map $g\colon E_m \to \mathbb{A}^1$ by \[ g(P)=(x_1^2-q)\cdot \left(\left(x\circ f^{-1}(P)\right)^2-q\right). \] The map $g$ has a pole in the same points as the map $x\circ f^{-1},$ and is regular elsewhere.
The maps $f$ and $g$ depend on a fixed point $P_1\in\mathcal{D}_m.$ We omit noting this dependency and simply denote these maps by $f$ and $g$. The motivation for the map $g$ is \cite[2.4, Proposition 4]{dujella2017more}. Dujella and Kazalicki use the $2$-descent homomorphism in the proof of Proposition 4, we will use $g$ for similar purposes.
\begin{theorem}\label{thm:1} Let $(x_1,y_1)\in \mathcal{D}_m(\mathbb{Q})$ be the point used to define the map $f\colon \mathcal{D}_m \to E_m.$ If $(Q_1,Q_2,Q_3)\in E_m(\mathbb{Q})^3$ is a triple satisfying the non-degeneracy criteria such that $(y_1^2-q)\cdot g(Q_1+Q_2+Q_3)$ is a square, then the numbers $$a=\pm\left(\frac{1}{m}\frac{g(Q_1)}{(x_1^2-q)}\frac{g(Q_2)}{(x_1^2-q)}\frac{g(Q_3)}{(x_1^2-q)}\right)^{1/2},$$ $$b=\frac{g(Q_1)}{a(x_1^2-q)}, c=\frac{g(Q_2)}{a(x_1^2-q)}, d=\frac{g(Q_3)}{a(x_1^2-q)}$$ are rational and form a \emph{rational $D(q)$-quadruple} such that $abcd=m$.
Conversely, assume $(a,b,c,d)$ is a \emph{rational $D(q)$-quadruple}, such that $m=abcd$. If the triple $(Q_1,Q_2,Q_3)\in E_m(\mathbb{Q})^3$ corresponds to $(a,b,c,d)$, then $(y_1^2-q)g(Q_1+Q_2+Q_3)$ is a square. \end{theorem}
It is not true that the existence of a rational point on $\mathcal{D}_m(\mathbb{Q})$ implies the existence of a rational $D(q)$-quadruple with product $m.$ Examples with further clarification are given in Section \ref{sec:4}. The following classification theorem holds: \begin{theorem}\label{thm:2} There exists a rational $D(q)$-quadruple with product $m$ if and only if \[m=(t^2-q)\left(\frac{u^2-q}{2u}\right)^2 \] for some rational parameters $(t,u).$ \end{theorem}
In Section \ref{sec:2} we study properties of the function $g$ which we then use in Section \ref{sec:3} to prove Theorems \ref{thm:1} and \ref{thm:2}. In Section \ref{sec:4}, we give an algorithm on how to determine whether a specific $m$ such that $\mathcal{D}_m(\mathbb{Q})\neq \emptyset$ admits a rational $D(q)$-quadruple with product $m.$ We conclude the section with an example of an infinite family.
\section{Properties of the function $g$}\label{sec:2}
In this section, we investigate the properties of the function $g$ which we will use to prove the main theorems. The following proposition describes the divisor of $g.$
\begin{proposition}\label{prop:3} The divisor of $g$ is $${\textrm{div}\:} g=2(S_1)+2(S_2)-2(R_1)-2(R_2),$$ where $S_1,R_1,S_2,R_2\in E_m(\mathbb{Q}(\sqrt{q}))$ with coordinates \begin{align*} S_1&=(~(y_1^2-q)(x_1-\sqrt{q})^2, \hspace{5.8pt}\quad 2y_1\sqrt{q}~(y_1^2-q)(x_1-\sqrt{q})^2~),\\ R_1&=(~(x_1^2-q)(y_1+\sqrt{q})^2, \hspace{5.8pt}\quad 2x_1\sqrt{q}~(x_1^2-q)(y_1+\sqrt{q})^2~),\\ S_2&=(~(y_1^2-q)(x_1+\sqrt{q})^2,~ -2y_1\sqrt{q}~(y_1^2-q)(x_1+\sqrt{q})^2~),\\ R_2&=(~(x_1^2-q)(y_1+\sqrt{q})^2,~ -2x_1\sqrt{q}~(x_1^2-q)(y_1-\sqrt{q})^2~). \end{align*}
The points $S_1, S_2, R_1$ and $R_2$ satisfy the following identities: \begin{align*} 2S_1&=2S_2=f(x_1,-y_1)=S+2R,\\ 2R_1&=2R_2=f(-x_1,y_1)=S,\\ S_1+R&=R_1,\quad R_1+R=S_2,\quad S_2+R=R_2,\quad R_2+R=S_1.\end{align*} \end{proposition}
\begin{proof} We seek zeros and poles of $g.$ The poles of $g$ are the same as the poles of $x\circ f^{-1}.$ To find zeros of $g$, notice that \[ (x\circ f^{-1}(P))^2-q=\frac{m}{(y\circ f^{-1}(P))^2-q}, \] so all we need to find are poles of $y \circ f^{-1}.$
The zeros of $x\circ f^{-1}$ are points on $E_m$ which map to affine points on $\overline{\mathcal{D}}_m$ that have zero $x$-coordinate. We can easily calculate such points. If $x=0,$ then $y^2=\frac{q^2-m}{q}.$ Denote $K=\sqrt{\frac{q^2-m}{q}}.$ We know $K\neq 0,$ since $m\neq q^2.$
The zeros of $x\circ f^{-1}$ are the points $f(0,K), f(0,-K)\in E_m({\overline{\Q}}),$ which are different since $K\neq 0.$ Since $x\circ f^{-1}$ is of degree two, both zeros are of order one. We conclude $x\circ f^{-1}$ has either one double pole, or two poles of order one.
Similarly, the zeros of $y\circ f^{-1}$ are the points $f(K,0), f(-K,0)\in E_m({\overline{\Q}}),$ both of order one. The map $y\circ f^{-1}$ also has either a double pole or two poles of order one.
Assume the point $P_0 \in E_m$ maps to a non-affine point in $\overline{\mathcal{D}}_m.$ This means that $Z\circ f^{-1}(P_0)=0,$ and at least one of the projective coordinate functions $X\circ f^{-1}, Y\circ f^{-1}$ is nonzero at $P_0.$ It follows that $P_0$ is a pole of at least one of the maps $x\circ f^{-1}, y\circ f^{-1}.$
Let $P_0 \in E_m$ be a pole of one of the maps $x\circ f^{-1}, y\circ f^{-1}.$ None of the points $ f^{-1}(P_0),$ $f^{-1}(P_0+R), f^{-1}(P_0+2R), f^{-1}(P_0+3R)$ are affine points on $\overline{\mathcal{D}}_m$ because if one of them is an affine point, then they all are, since the map $P\mapsto P+R$ viewed on $\overline{\mathcal{D}}_m$ maps affine points to affine points. We conclude that each of the points $P_0, P_0+R, P_0+2R, P_0+3R$ is a pole of one of the maps $x\circ f^{-1}, y\circ f^{-1}$ and with the previous claims we have that $x\circ f^{-1}, y\circ f^{-1}$ both have two poles of order one.
The map $P\mapsto S-P,$ viewed on $\overline{\mathcal{D}}_m$, also maps affine points to affine points. Similarly as above, the points $f^{-1}(S-P_0), f^{-1}(S-P_0+R), f^{-1}(S-P_0+2R), f^{-1}(S-P_0+3R)$ are not affine in $\overline{\mathcal{D}}_m,$ because the point $f^{-1}(P_0)$ would be affine as well. The sets $\lbrace P_0, P_0+R, P_0+2R, P_0+3R \rbrace$ and $\lbrace S-P_0, S-P_0+R, S-P_0+2R, S-P_0+3R \rbrace$ must be equal, otherwise the maps $x\circ f^{-1}, y\circ f^{-1}$ would have more than four different poles in total. This means that every pole satisfies the equality $2P_0=S+kR$ for some $k\in \lbrace 0, 1, 2, 3\rbrace.$ Equivalently, every pole $P_0$ is a fixed point of some involution $i_k$ of the form $P\mapsto S-P+kR.$ Each involution $i_k$ has four fixed points on $E_m({\overline{\Q}}),$ because any two fixed points differ by an element from the $[2]$-torsion.
The involution $i_0,$ viewed on $\overline{\mathcal{D}}_m,$ maps an affine point $(x,y)=f^{-1}(P)$ to $(-x,y)=f^{-1}(S-P).$ It has two affine fixed points which have $x$-coordinate equal to zero on $\overline{\mathcal{D}}_m$, as well as two fixed points which are not affine on $\overline{\mathcal{D}}_m.$ Such points are either poles of $x\circ f^{-1}$ or poles of $y \circ f^{-1}.$ Using Magma\cite{bosma1997magma} we calculate the coordinates explicitly to obtain $R_1$ and $R_2.$ Computationally, we confirm $R_1$ and $R_2$ are poles of $x \circ f^{-1},$ that is, poles of $g.$
The involution $i_2,$ viewed on $\overline{\mathcal{D}}_m,$ maps an affine point $(x,y)=f^{-1}(P)$ to $(x,-y)=f^{-1}(S-P+2R).$ It has two affine fixed points which have $y$-coordinate equal to zero on $\overline{\mathcal{D}}_m,$ as well as two fixed points which are not affine on $\overline{\mathcal{D}}_m.$ These points must be poles of the map $y\circ f^{-1},$ that is, zeros of $g.$ Again, using Magma, we calculate the coordinates to obtain $S_1$ and $S_2.$
Since the poles of $x\circ f^{-1}$ are of order one, then the poles of $g$ are of order two. The same is true for poles of $y\circ f^{-1},$ that is, for zeros of $g.$ The last row of identities in the statement of the theorem is checked by Magma. \end{proof}
\proposition \label{prop:4} There exists $h\in \mathbb{Q}(E_m)$ such that $g\circ[2]=h^2.$ \proof Let $\tilde{h}\in {\overline{\Q}}(E_m)$ such that \begin{align*} {\textrm{div}\:} \tilde{h}&=[2]^{\ast}((S_1)+(S_2)-(R_1)-(R_2))\\ &=\sum\limits_{T\in E_m[2]} (S'_1+T)+\sum\limits_{T\in E_m[2]} (S'_2+T)-\sum\limits_{T\in E_m[2]} (R'_1+T)-\sum\limits_{T\in E_m[2]} (R'_2+T), \end{align*} where $2S'_i=S_i,2R'_i=R_i$ and $[2]^{\ast}$ is the pullback of the doubling map on $E_m.$
Such $\tilde{h}$ exists because of Corollary 3.5 in Silverman\cite[III.3]{silverman2009arithmetic} stating that if $E$ is an elliptic curve and $D=\sum n_P(P)\in \text{Div}(E),$ then $D$ is principal if and only if $$\sum_{P\in E} n_P=0 \text{ \hspace{5pt}and } \sum_{P\in E} [n_P]P=0,$$ where the second sum is addition on $E.$
The first sum being equal to zero is immediate, and for the second one we have \[ \sum\limits_{T\in E_m[2]} (S'_1+T)+\sum\limits_{T\in E_m[2]} (S'_2+T)-\sum\limits_{T\in E_m[2]} (R'_1+T)-\sum\limits_{T\in E_m[2]} (R'_2+T)= \] \[ =[4](S'_1+S'_2-R'_1-R'_2)=[2](S_1+S_2-R_1-R_2)= \] \[ =[2](S_1-R_2+S_2-R_1)\stackrel{(\ast)}{=}[2](R+R)=\mathcal{O}, \] where $(\ast)$ follows from the last row of identities in Proposition \ref{prop:3}.
Easy calculations give us ${\textrm{div}\:} g\circ [2]={\textrm{div}\:} \tilde{h}^2$ which implies $C\tilde{h}^2=g\circ [2],$ for some $C\in {\overline{\Q}}.$ Let $h:=\tilde{h}\sqrt{C}\in {\overline{\Q}}(E_m)$ so that $h^2=g\circ [2].$ We will prove $h\in \mathbb{Q}(E_m).$
First, we show that every $\sigma\in \textrm{Gal}(\Qbar/\Q)$ permutes zeros and poles of $\tilde{h}$. Let us check what $\sigma$ does to $S_1$ and $S_2.$ Since $S_1$ and $S_2$ are conjugates over $\mathbb{Q}(\sqrt{q})$, the only possibilities for $S_1^\sigma$ are $S_1$ or $S_2.$ If $S_1^\sigma=S_1,$ then we must have $(S'_1)^\sigma=S'_1+T,$ where $T\in E_m[2],$ because $2((S'_1)^\sigma-S'_1)=(2S'_1)^\sigma-2S'_1=S_1^\sigma-S_1=\mathcal{O}.$ Thus $\sigma$ fixes $\sum\limits_{T\in E_m[2]} (S'_1+T)$. Since in this case we also know that $S_2^\sigma=S_2,$ we get that $\sigma$ fixes $\sum\limits_{T\in E_m[2]} (S'_2+T)$ as well.
If $S_1^\sigma=S_2$ it is easy to see that $$\left(\sum\limits_{T\in E_m[2]} (S'_1+T)\right)^\sigma=\sum\limits_{T\in E_m[2]} (S'_2+T) \text{ and } \left(\sum\limits_{T\in E_m[2]} (S'_2+T)\right)^\sigma=\sum\limits_{T\in E_m[2]} (S'_1+T).$$
Similar statements hold for $R_1$ and $R_2,$ so we conclude that $\tilde{h}$ is defined over $\mathbb{Q}.$ Both $h$ and $\tilde{h}$ have the same divisor so $h$ is also defined over $\mathbb{Q}$. Now we use the second statement from Theorem 7.8.3. in \cite{galbraith2012mathematics}: \begin{theorem} Let $C$ be a curve over a perfect field $k$ and let $f\in \overline{k}(C).$ \begin{enumerate} \item If $\sigma(f)=f, ~~$ for each $\sigma \in \textrm{Gal}(\overline{k}/k)$ then $f\in k(C).$ \item If ${\textrm{div}\:}(f)$ is defined over $k$ then $f=ch$ for some $c\in \overline{k}$ and $h\in k(C).$ \end{enumerate} \end{theorem} From the second statement of the previous theorem we conclude that $h=c\cdot h'$ where $c\in {\overline{\Q}}$ and $h'\in \mathbb{Q}(E_m).$ We know that $c^2(h')^2=h^2=g\circ [2]$, and that $g\circ [2] (\mathcal{O})=(x_1^2-q)^2$ is a rational square. It follows that $\displaystyle c^2=\frac{(x_1^2-q)^2}{h'(\mathcal{O})^2}$ is a rational square as well, hence $c$ is rational. Finally, we have $h\in \mathbb{Q}(E_m).$ \endproof We end this section with a theorem which will handle rationality issues in Theorem \ref{thm:1}. \begin{theorem} \label{thm:6} For all $P,Q\in E_m(\mathbb{Q})$ we have $g(P+Q)\equiv g(P)g(Q) \mod (\mathbb{Q}^*)^2.$\\ In particular, if $P\equiv Q \mod 2E_m(\mathbb{Q}) $ then $g(P)\equiv g(Q) \mod (\mathbb{Q}^*)^2.$ \end{theorem}
\proof Let $P',Q'\in E_m({\overline{\Q}})$ such that $2P'=P$ and $2Q'=Q.$ We prove that $$\frac{\sigma(h(P'+Q'))}{h(P'+Q')}=\frac{\sigma(h(P'))}{h(P')}\frac{\sigma(h(Q'))}{h(Q')}.$$
Following Silverman \cite[III.8]{silverman2009arithmetic}, assume $T\in E_m[2].$ From Proposition \ref{prop:4} it follows that $\displaystyle h^2(X+T)=g\circ [2] (X+T)=g\circ [2] (X)=h^2(X),$ for every $X\in E_m.$ This means that $\frac{h(X+T)}{h(X)}\in\{\pm 1\}.$ The morphism $$ E_m \rightarrow \mathbb{P}^1, \qquad X\mapsto \frac{h(X+T)}{h(X)}$$ is not surjective, so by \cite[II.2.3]{silverman2009arithmetic} it must be constant.
For $ \sigma \in \textrm{Gal}(\Qbar/\Q)$ we have $\sigma(P')-P' \in E_m[2],\sigma(Q')-Q' \in E_m[2]$ and $\sigma(P'+Q')-(P'+Q') \in E_m[2].$ This holds since $2P'=P\in E_m(\mathbb{Q})$ and $2Q'=Q\in E_m(\mathbb{Q}).$ Now we get $$\frac{\sigma(h(P'))}{h(P')}=\frac{h(\sigma(P'))}{h(P')}=\frac{h(P'+(\sigma(P')-P'))}{h(P')}=\frac{h(X+(\sigma(P')-P'))}{h(X)}.$$ Similarly $$\frac{\sigma(h(Q'))}{h(Q')}=\frac{h(X+(\sigma(Q')-Q'))}{h(X)},\quad \frac{\sigma(h(P'+Q'))}{h(P'+Q')}=\frac{h(X+(\sigma(P'+Q')-(P'+Q')))}{h(X)}.$$
Now \begin{align*} \frac{\sigma(h(P'+Q'))}{h(P'+Q')}&=\frac{h(X+(\sigma(P'+Q')-(P'+Q')))}{h(X)}\\ &=\frac{h(X+(\sigma(P'+Q')-(P'+Q')))}{h(X+\sigma(P')-P')}\frac{h(X+\sigma(P')-P')}{h(X)}\\ &=\frac{\sigma(h(Q'))}{h(Q')}\frac{\sigma(h(P'))}{h(P')} \end{align*} by plugging in $X=P'+Q'-\sigma(P')$ for the first $X$ and $X=P'$ for the second one. This leads to $$\frac{h(P'+Q')}{h(P')h(Q')}=\frac{\sigma(h(P'+Q'))}{\sigma(h(Q'))\sigma(h(P'))}=\sigma\left(\frac{h(P'+Q')}{h(P')h(Q')}\right)$$ for every $\sigma\in \textrm{Gal}(\Qbar/\Q).$ Now we conclude $$\frac{h(P'+Q')}{h(P')h(Q')}\in \mathbb{Q}\implies h^2(P'+Q')\equiv h^2(P')h^2(Q') \mod (\mathbb{Q}^*)^2.$$ Finally $$g(P+Q)=g\circ[2](P'+Q')= h^2(P'+Q')\equiv h^2(P')h^2(Q')=g(P)g(Q) \mod (\mathbb{Q}^*)^2.$$
The second statement of the theorem follows easily from the first.\linebreak If $P=Q+2S_3,$ with $S_3\in E_m(\mathbb{Q}),$ then $$g(P)=g(Q+2S_3)\equiv g(Q)g(S_3)^2\equiv g(Q) \mod (\mathbb{Q}^*)^2. $$ \endproof
Theorem \ref{thm:6} was more difficult to prove compared to a similar statement in \cite[2.4.]{dujella2017more}. Their version of the function $g$ had a very simple factorization$\mod (\mathbb{Q}^{\ast})^2,$ allowing them to use the $2-$descent homomorphism.
\section{Proofs of main theorems}\label{sec:3}
The main difficulty in the following proof is the issue of rationality of the quadruple. As we have mentioned, Theorem \ref{thm:6} will deal with this.
\emph{Proof of Theorem \ref{thm:1}:} From the assumptions on $(Q_1,Q_2,Q_3)$ we know \linebreak $(y_1^2-q)g(Q_1+Q_2+Q_3)$ is a square. We have \begin{align*} a^2&=\frac{g(Q_1)g(Q_2)g(Q_3)}{(x_1^2-q)^3m}=\frac{g(Q_1)g(Q_2)g(Q_3)(y_1^2-q)}{(x_1^2-q)^4(y_1^2-q)^2} \\ &\equiv g(Q_1+Q_2+Q_3)(y_1^2-q) \mod (\mathbb{Q}^*)^2. \end{align*}
The equivalence is a direct application of Theorem \ref{thm:6}. This implies $a^2$ is a rational square so $a$ is rational, which in turn implies $b,c$ and $d$ are rational numbers, as noted in the introduction. Since $abcd=m\neq 0,$ none of the numbers $ a,b,c,d$ are zero, and the non-degeneracy criteria of $(Q_1,Q_2,Q_3)$ ensure that $a,b,c,d$ are pairwise different. Lastly, $ab+q=(x\circ f^{-1}(Q_1))^2$ (with similar equalities holding for other pairs of the quadruple). The previous statements prove the quadruple $(a,b,c,d)$ is a rational $D(q)$-quadruple.
On the other hand, if $(a,b,c,d)$ is a rational $D(q)$-quadruple, then we can define the points $(Q_1,Q_2,Q_3)\in E_m(\mathbb{Q})^3$ in correspondence to $(a,b,c,d).$ Using the same identities$\mod (\mathbb{Q}^*)^2$ as above, we get that $$(y_1^2-q)g(Q_1+Q_2+Q_3)\equiv a^2~~ \mod (\mathbb{Q}^*)^2. $$ $
\square$
To prove Theorem \ref{thm:2} we use the following lemma: \begin{lemma}\label{lemma:7} Let $(a,b,c,d)$ be a rational $D(q)$-quadruple such that $abcd=m.$ There exists a point $(x_0,y_0)\in \mathcal{D}_m(\mathbb{Q}),$ such that $x_0^2-q$ is a rational square. \end{lemma} \begin{proof} From Theorem \ref{thm:1} we know that $(y_1^2-q)g(Q_1+Q_2+Q_3)$ is a square, where $(Q_1,Q_2,Q_3)\in E_m(\mathbb{Q})^3$ is the triple that corresponds to the quadruple $(a,b,c,d)$.
Let $Q=Q_1+Q_2+Q_3.$ We have \begin{align*} (y_1^2-q)g(Q)&=(y_1^2-q)(x_1^2-q)((x\circ f^{-1}(Q))^2-q)=m\cdot (x\circ f^{-1}(Q))^2-q)\\ &=m\cdot \frac{m}{(y\circ f^{-1}(Q))^2-q}=m^2\frac{1}{(y\circ f^{-1}(Q))^2-q}. \end{align*}
Since the left hand side is a square, we conclude $(y\circ f^{-1}(Q))^2-q$ is a square as well. Now define $(x_0,y_0):=f^{-1}(Q+R).$ We know that \[ (y\circ f^{-1}(Q))^2-q\stackrel{(\ref{ness})}{=}(x\circ f^{-1}(Q+R))^2-q=x_0^2-q \] so the claim follows. \end{proof}
\emph{Proof of Theorem \ref{thm:2}:} Assume we have a rational $D(q)$-quadruple. Using Lemma \ref{lemma:7} there exists a point $(x_0,y_0)\in \mathcal{D}_m(\mathbb{Q})$ such that $x_0^2-q$ is a rational square. Since $x_0^2-q=k^2,$ then $q=x_0^2-k^2=(x_0-k)(x_0+k).$ Denote $u=x_0-k,$ then $x_0+k=q/u$ and by adding the previous two equalities together to eliminate $k,$ we get $x_0=\frac{q+u^2}{2u}.$ Denoting $t=y_0$ we get $$m=(x_0^2-q)(y_0^2-q)=\left(\left(\frac{q+u^2}{2u}\right)^2-q\right)(t^2-q)=\left(\frac{q-u^2}{2u}\right)^2(t^2-q).$$
Now, let $m=\left(\frac{q-u^2}{2u}\right)^2(t^2-q)$ for some rational $(t,u).$ Denote $y_1=t, x_1=\frac{q+u^2}{2u}.$ It is easy to check that $(x_1^2-q)(y_1^2-q)=m,$ so there is a rational point $(x_1,y_1)\in\mathcal{D}_m(\mathbb{Q})$ such that $x_1^2-q=\left(\frac{u^2-q}{2u}\right)^2$ is a square. We use this point $(x_1,y_1)=\colon P_1$ to define the map $f:\mathcal{D}_m\to E_m.$ Let $Q_1=R+S, Q_2=2S$ and $Q_3=3S.$ The sets $G\cdot Q_i$ are disjoint and $g(Q_1+Q_2+Q_3)(y_1^2-q)=g(R+6S)(y_1^2-q)\equiv g(R)(y_1^2-q)=((x_1^2-q)(y_1^2-q))(y_1^2-q)$ mod $(\mathbb{Q}^*)^2$ is a rational square. The points $(Q_1,Q_2,Q_3)$ satisfy the conditions of Theorem \ref{thm:1} giving us a rational $D(q)$ quadruple.$
\square$
\begin{remark}\label{rem:8} The condition $m=\left(\frac{q-u^2}{2u}\right)^2(t^2-q)$ is equivalent to the fact that there exists $(x_0,y_0)\in\mathcal{D}_m(\mathbb{Q})$ such that $x_0^2-q$ is a square. This was proven in the preceding theorems. \end{remark}
\section{Examples}\label{sec:4}
There are plenty of examples where $m=(x_1^2-q)(y_1^2-q)$ for some rational $x_1$ and $y_1,$ such that there does not exist a rational $D(q)$-quadruple with product $m.$ Equivalently, $m$ cannot be written as $(x_0^2-q)(y_0^2-q)$ such that $x_0^2-q$ is a square.
According to Theorem \ref{thm:1}, to find out whether there is a rational $D(q)$-quadruple with product $m,$ one needs to check whether there is a point $T'\in E_m(\mathbb{Q})$ such that $g(T')(y_1^2-q)$ is a square. Theorem \ref{thm:6} tells us that we only need to check the points $T \in E_m(\mathbb{Q})/2E_m(\mathbb{Q}),$ which is a finite set. If for some explicit $q,m$ we know the generators of the group $E_m(\mathbb{Q}) / 2E_m(\mathbb{Q}),$ we can determine whether there exist rational $D(q)$-quadruples with product $m,$ and parametrize them using points on $E_m(\mathbb{Q}).$ For such computations we used Magma.
Let $q=3, x_1=5$ and $y_1=7$ making $m=(5^2-3)(7^2-3)=1012.$ The rank of $E_m$ is two, $E_m$ has one torsion point of order four, giving us in eight points in total to check. None of the points $T \in E_m(\mathbb{Q})/2E_m(\mathbb{Q})$ satisfy that $g(T)(y_1^2-q)$ is a square, so there are no $D(3)$-quadruples with product $1012.$
On the other hand, take $q=-3,x_1=1$ so that $x_1^2-q=4$ and let $y_1=t$ which makes $m=4\cdot(t^2+3).$ The point $S$ is a point of infinite order on $E_m(\mathbb{Q}(t)),$ and the triple $(Q_1,Q_2,Q_3)=(S+R,2S,3S)$ satisfies the conditions of Theorem \ref{thm:1}. We obtain the following family:\\ \begin{align*} a&=\frac{2\cdot (3 + 6 t^2 + 7 t^4)\cdot (27+162t^2+801t^4+1548t^6+1069t^8+306t^{10}+183t^{12})}{(3 + t^2)\cdot (1 + 3 t^2)\cdot (9+9t^2+19t^4+27t^6)\cdot(3+27t^2+33t^4+t^6)}, \\ b&=\frac{(3 + t^2)^2\cdot (1 + 3 t^2)\cdot (9+9t^2+19t^4+27t^6)\cdot(3+27t^2+33t^4+t^6)}{2\cdot (3 + 6 t^2 + 7 t^4)\cdot (27+162t^2+801t^4+1548t^6+1069t^8+306t^{10}+183t^{12})},\\ c&=\frac{2\cdot (3 + 6 t^2 + 7 t^4)\cdot(3 + 27 t^2 + 33 t^4 + t^6)\cdot (9 + 9 t^2 + 19 t^4 + 27 t^6)}{(3 + t^2)\cdot (1 + 3 t^2)\cdot(27 + 162 t^2 + 801 t^4 + 1548 t^6 + 1069 t^8 +
306 t^{10} + 183 t^{12})},\\ d&=\frac{2\cdot(3 + t^2)\cdot (1 + 3 t^2)\cdot(27 + 162 t^2 + 801 t^4 + 1548 t^6 + 1069 t^8 +
306 t^{10} + 183 t^{12})}{(3 + 6 t^2 + 7 t^4)\cdot(3 + 27 t^2 + 33 t^4 + t^6)\cdot (9 + 9 t^2 + 19 t^4 + 27 t^6)}. \end{align*} We can generalize the example above by setting $q=q, y_1=t, x_1=\frac{q+u^2}{2u}.$ The triple of points $(S+R,2S,3S)$ satisfies the conditions of Theorem \ref{thm:1} and we can calculate an explicit family of rational $D(q)$-quadruples with product $m,$ but this example is too large to print (the numerator of $a$ is a polynomial in the variables $(q,t,u)$ of degree forty).
All the computations in this paper were done in Magma \cite{bosma1997magma}.
\end{document}
|
arXiv
|
{
"id": "2002.02006.tex",
"language_detection_score": 0.6660143136978149,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\newtheorem{thm}{Theorem} \newtheorem{cor}{Corollary} \newproof{pf}{Proof}
\begin{frontmatter}
\title{Two stage design for estimating the product of means with cost in the case of the exponential family} \author[1]{Zohra BENKAMRA} \author[2]{Mekki TERBECHE} \author[1]{Mounir TLEMCANI\corref{*}}
\address[1]{Department of Mathematics. University of Oran, Algeria} \address[2]{Department of Physics, L.A.A.R University Mohamed Boudiaf, Oran, Algeria} \cortext[*]{Corresponding author : [email protected] (M.Tlemcani)}
\begin{abstract} We investigate the problem of estimating the product of means of independent populations from the one parameter exponential family in a Bayesian framework. We give a random design which allocates $m_{i}$ the number of observations from population $P_{i}$ such that the Bayes risk associated with squared error loss and cost per unit observation is as small as possible. The design is shown to be asymptotically optimal. \end{abstract}
\begin{keyword}
Two stage design, product of means, exponential family, Bayes risk, cost, asymptotic optimality.
\end{keyword}
\end{frontmatter}
\section{Introduction}
Assume that for $i=1,...,n$; a random variable $X_{i}$ whose distribution belongs to the one parameter exponential family is observable from population $P_{i}$ with cost $c_{i}$ per unit observation. The problem of estimating several means in the case of the exponential family distributions with linear combination of losses was addressed by \cite{cohen}. The problem of interest in this paper is to estimate the product of means using a Bayesian approach associated with squared error loss and cost. Since a Bayesian framework is considered; see, e.g., \cite{page,shapiro}, then typically optimal estimators are Bayesian estimators and the problem turns to design a sequential allocation scheme; see, e.g., \cite{woodroofe}, to select $m_{i}$ the number of observations from population $P_{i}$ such that the Bayes risk plus the corresponding budget $B=\sum_{i=1}^{n}c_{i}m_{i}$ is as small as possible. \cite{terbeche aas}, have defined a sequential design to estimate the difference between means of two populations from the exponential family with associated cost. The random allocation was shown to be the best from numerical considerations; see, e.g., \cite{terbeche phd}. Similarly, the problem of estimating the product of several means of independent populations, subject to the constraint of a total number of observations $M$ fixed, was addressed by \cite{rekab}, using a two stage approach. The allocation of $m_{i}$ was nonrandom and the first order optimality was shown for large $M$.
Suppose that $X_{i}$ has the distribution of the form \[ f_{\theta _{i}}(x_{i})\varpropto e^{\theta _{i}x_{i}-\psi (\theta _{i})},~x_{i}\in \mathbb{R} ,~\theta _{i}\in \Omega \] where $\Omega $ is a bounded open interval in $\mathbb{R}$. It follows that $E_{\theta _{i}}\left[ x_{i}\right] =\psi ^{\prime }\left( \theta _{i}\right) $ and $Var_{\theta _{i}}\left[ x_{i}\right] =\psi ^{\prime \prime }\left( \theta _{i}\right)$. One assumes that prior distribution for each $\theta _{i}$ is given by \[ \pi _{i}\left( \theta _{i}\right)\varpropto e^{r_{i}\left( \mu _{i}\theta _{i}-\psi \left( \theta _{i}\right)\right) } \] where $r_{i}$ and $\mu _{i}$ are reals and $r_{i}>0$, $i=1,...,n$. Here we treat $\theta _{i}$ as a realization of a random variable and assume that for each population, $x_{i1},...,x_{im_{i}}$ are conditionally independent and that $\theta _{1},...,\theta _{n}$ are a priori independent. Our aim is to estimate the product $\theta =\prod_{i=1}^{n}\psi ^{\prime }\left( \theta _{i}\right)$, subject to squared error loss and linear cost.
\section{The Bayes risk}
Let $\mathcal{F}_{m_{1},...,m_{n}}$ the $\sigma $-Field generated by $\left( X_{1},...,X_{n}\right) $ where $X_{i}=\left( x_{i1},...,x_{im_{i}}\right) $ and let $\mathcal{F}_{m_{i}}=\sigma \left( X_{i}\right) =\sigma \left( x_{i1},...,x_{im_{i}}\right) $. It was shown that \begin{eqnarray} E\left[ \psi ^{\prime }\left( \theta _{i}\right) /\mathcal{F}_{m_{i}} \right] &=&=\frac{\mu_{i}r_{i}+\sum_{j=1}^{m_{i}}x_{ij}}{m_{i}+r_{i}} \label{th 2.2.1-1} \\ Var\left[ \psi ^{\prime }\left( \theta _{i}\right) /\mathcal{F}_{m_{i}} \right] &=&E\left[ \frac{\psi ^{\prime \prime }\left( \theta _{i}\right) }{ m_{i}+r_{i}}/\mathcal{F}_{m_{i}}\right], \label{th 2.2.1-2} \end{eqnarray}
(see \cite{terbeche sort}). Using independence across populations, the Bayes estimator of $\theta $ is \[ \hat{\theta}=E\left[ \theta /\mathcal{F}_{m_{1},...,m_{n}}\right] =\prod\limits_{i=1}^{n}E\left[ \psi ^{\prime }\left( \theta _{i}\right) / \mathcal{F}_{m_{i}}\right] \] Assume that there exists $p\geq 1$ such that \begin{equation} E\left[ \left( \psi ^{\prime \prime }\left( \theta _{i}\right) \right) ^{p} \right] <+\infty ~ and ~ E\left[ \left( \psi ^{\prime }\left( \theta _{i}\right) \right) ^{2p}\right] <+\infty, \label{cond} \end{equation}
for all $i=1,...,n$; then the corresponding Bayes risk associated with quadratic loss and cost can be written as follows, \begin{equation} \label{r} R\left( m_{1},...,m_{n}\right) =E\left[ \sum\limits_{i=1}^{n}\frac{U_{im_{i}} }{m_{i}+r_{i}}+\sum\limits_{i=1}^{n}c_{i}m_{i}\right] +\sum \limits_{i=1}^{n}o\left( \frac{1}{m_{i}}\right) \end{equation} and by the way, it can be approximated for large samples by \begin{equation} \label{rtilde} \tilde{R}\left( P\right) =\tilde{R}\left( m_{1},...,m_{n}\right) =E\left[ \sum\limits_{i=1}^{n}\frac{U_{im_{i}}}{m_{i}+r_{i}}+\sum \limits_{i=1}^{n}c_{i}m_{i}\right] \end{equation} where $U_{im_{i}}=E\left[ V_{i}/\mathcal{F}_{m_{1},...,m_{n}}\right] $ and $V_{i}=\psi ^{\prime \prime }\left( \theta _{i}\right) \prod_{j\neq i}\psi ^{\prime 2}\left( \theta _{j}\right)$
\section{Lower bound for the scaled Bayes risk} From now on, the notation $c\rightarrow 0$ means that $c_{j}\rightarrow 0$, for all $j=1,...,n$. Assume that for all $i$, \begin{equation} \label{ci} \frac{c_{i}}{\sum\limits_{j=1}^{n}c_{j}}\rightarrow \lambda _{i}\in \left] 0,1\right[ ,\ as \ c\rightarrow 0. \end{equation} \begin{thm} \label{th 233}For any random design (P) satisfying \begin{equation} \label{ai} m_{i}\sqrt{c_{i}}\rightarrow a_{i}\neq 0~,~a.s., \ as\ c\rightarrow 0; \end{equation} then \begin{equation} \label{res} \liminf_{c\rightarrow 0}\frac{R(P)}{\sqrt{ \sum\limits_{j=1}^{n}c_{j}}}\geq 2E \left[ \sum\limits_{i=1}^{n}\sqrt{\lambda _{i}}\sqrt{V_{i}}\right] \end{equation} \end{thm}
\begin{pf} Expressions (\ref{r}) and (\ref{rtilde}) with the help of (\ref{ci}) and (\ref{ai}), give \begin{equation} \label{rrtilde} \liminf_{c\rightarrow 0 }\frac{R(P)}{\sqrt{ \sum\limits_{j=1}^{n}c_{j}}}=\liminf_{c\rightarrow 0}\frac{\tilde{R}(P)}{\sqrt{\sum\limits_{j=1}^{n}c_{j}}} \end{equation} and the scaled approximated Bayes risk satisfies the following inequality: \[ \frac{\tilde{R}(P)}{\sqrt{\sum\limits_{j=1}^{n}c_{j}}}\geq 2E\left[ \sum\limits_{i=1}^{n}\sqrt{\frac{c_{i}}{\sum\limits_{j=1}^{n}c_{j}}}\sqrt{ U_{im_{i}}}\right] -\sum\limits_{i=1}^{n}\sqrt{\frac{c_{i}}{ \sum\limits_{j=1}^{n}c_{j}}}\sqrt{c_{i}}r_{i}, \] since for all $i$, \begin{eqnarray*} \frac{U_{im_{i}}}{m_{i}+r_{i}}+c_{i}m_{i} &=&\left( \frac{\sqrt{U_{im_{i}}}}{\sqrt{m_{i}+r_{i}}}+\sqrt{c_{i}}\sqrt{ m_{i}+r_{i}}\right) ^{2}+2\sqrt{c_{i}}\sqrt{U_{im_{i}}}-c_{i}r_{i}\\ &\geq &2\sqrt{c_{i}}\sqrt{U_{im_{i}}}-c_{i}r_{i}. \end{eqnarray*}
Finally, Fatou's lemma and condition (\ref{ci}) give \[ \liminf_{c\rightarrow 0}\frac{\tilde{R}(P)}{ \sqrt{\sum\limits_{j=1}^{n}c_{j}}}\geq 2E\left[ \liminf_{c\rightarrow 0}\sum\limits_{i=1}^{n}\sqrt{\frac{c_{i}}{ \sum\limits_{j=1}^{n}c_{j}}}\sqrt{U_{im_{i}}}\right] =2E\left[ \sum\limits_{i=1}^{n}\sqrt{\lambda _{i}}\sqrt{V_{i}}\right] \] and the proof follows. \end{pf}
\section{First order optimal design} According to condition (\ref{ai}) and identity (\ref{rrtilde}) a first order optimal design with respect to $\sum_{i=1}^{n}m_{i}=m$ must satisfy \begin{equation} \label{conver} \frac{\tilde{R}(P)}{\sqrt{\sum\limits_{j=1}^{n}c_{j}}}-2E\left[ \sum\limits_{i=1}^{n}\sqrt{\lambda _{i}}\sqrt{V_{i}}\right] \rightarrow 0, \ as\ c\rightarrow 0, \end{equation} It should be pointed that condition (\ref{conver}) is actually similar to the first order efficiency property for A.P.O rules in Bayes sequential estimation; see, e.g., \cite{leng,hwang}, for one-parameter exponential families, which involves a sequential allocation procedure and a stopping time.
In our approach, condition (\ref{conver}) is handled by the following expansion. \begin{eqnarray*} \frac{\tilde{R}(P)}{\sqrt{\sum\limits_{j=1}^{n}c_{j}}} &=& \frac{2E\left[ \sum\limits_{i=1}^{n}\sqrt{c_{i}}\sqrt{U_{im_{i}}}\right] }{ \sqrt{\sum\limits_{j=1}^{n}c_{j}}} +\frac{E\left[ \sum\limits_{i=1}^{n}\frac{\left( \sqrt{U_{im_{i}}}-\left( m_{i}+r_{i}\right) \sqrt{c_{i}}\right) ^{2}}{m_{i}+r_{i}}\right] }{\sqrt{ \sum\limits_{j=1}^{n}c_{j}}} \\ &-&\sum\limits_{i=1}^{n}\sqrt{c_{i}}\sqrt{\frac{ c_{i}}{\sum\limits_{j=1}^{n}c_{j}}}r_{i} \end{eqnarray*} The last term goes to zero as $c\rightarrow 0$, thanks to condition (\ref{ci}). Hence, sufficient conditions for a design to satisfy (\ref{conver}) are \begin{eqnarray} E\left[ \sum\limits_{i=1}^{n}\sqrt{\frac{c_{i}}{\sum\limits_{j=1}^{n}c_{j}}} \sqrt{U_{im_{i}}}\right] -E\left[ \sum\limits_{i=1}^{n}\sqrt{\lambda _{i}} \sqrt{V_{i}}\right] &\rightarrow &0 \label{cs1} \\ E\left[ \frac{\left( \sqrt{U_{im_{i}}}-\left( m_{i}+r_{i}\right) \sqrt{c_{i}} \right) ^{2}}{\left( m_{i}+r_{i}\right) \sqrt{\sum\limits_{j=1}^{n}c_{j}}} \right] &\rightarrow &0,~\forall i \label{cs2} \end{eqnarray} as $c\rightarrow 0$.
\begin{thm} \label{th 241}Let $P$ a random policy satisfying $m_{i}\rightarrow +\infty ,~a.s.$, and suppose that condition (\ref{cond}) is true, then \[ E\left[ \sum\limits_{i=1}^{n}\sqrt{\frac{c_{i}}{\sum\limits_{j=1}^{n}c_{j}}} \sqrt{U_{im_{i}}}\right] -E\left[ \sum\limits_{i=1}^{n}\sqrt{\lambda _{i}} \sqrt{V_{i}}\right] \rightarrow 0, \ as\ c\rightarrow 0. \]
\end{thm}
\begin{pf} Remark that \begin{equation} \lim_{m_{1},...,m_{n}\rightarrow +\infty }\sqrt{U_{im_{i}}}=\sqrt{V_{i}} ,~a.s. \label{conv} \end{equation} Now \begin{eqnarray*} \sup_{m_{1},...,m_{n}}E\left[ \left( \sqrt{U_{im_{i}}}\right) ^{2}\right] &=&\sup_{m_{1},...,m_{n}}E\left[ U_{im_{i}}\right] \\ &=&E\left[ \psi ^{^{\prime \prime }}\left( \theta _{i}\right) \prod\limits_{j\neq i}\psi ^{\prime ^{2}}\left( \theta _{j}\right) \right] \\ &=&E\left[ \psi ^{^{\prime \prime }}\left( \theta _{i}\right) \right] \prod\limits_{j\neq i}E\left[ \psi ^{\prime ^{2}}\left( \theta _{j}\right) \right] <+\infty \end{eqnarray*} hence, the uniform integrability of $\sqrt{U_{im_{i}}}$ follows from condition (\ref{cond}) and martingales properties. Therefore, the convergence in (\ref{conv}) holds in $L^{1}$ and consequently : \[ \sqrt{\frac{c_{i}}{\sum\limits_{j=1}^{n}c_{j}}}\sqrt{U_{im_{i}}}\rightarrow \sqrt{\lambda _{i}}\sqrt{V_{i}} ~ in ~ L^{1}, \ as\ c\rightarrow 0, \] which achieves the proof. \end{pf}
\section{The two stage procedure}
Following the previous section, our strategy now is to satisfy condition ( \ref{cs2}). Then, we define the two stage sequential scheme as follows.
\begin{description} \item[Stage one] proceed for $k_{i}$ observation from population $P_{i}$ for $i=1,...,n$; such that $k_{i}\sqrt{c_{i}}\rightarrow 0$ and $ k_{i}\rightarrow +\infty $ as $c_{i}\rightarrow 0$.
\item[Stage two] for $i=1,...,n$; select $m_{i}$ integer as follows : \[ m_{i}=\max \left\{ k_{i},\left[ \frac{\sqrt{U_{ik_{i}}}}{\sqrt{c_{i}}}-r_{i}\right]\right\} \] where $\left[ x\right]$ denotes the integer part of $x$ and \[ U_{ik_{i}}=E\left[ \psi ^{^{\prime \prime }}\left( \theta _{i}\right) \prod\limits_{j\neq i}\psi ^{\prime ^{2}}\left( \theta _{j}\right) /\mathcal{ F}_{k_{1},...,k_{n}}\right] \] \end{description} We give now the main result of the paper. \begin{thm} \label{th 251}Assume condition (\ref{cond}) satisfied for a $p\geq 1$, then the two stage design is first order optimal. \end{thm}
\begin{pf} The $m_{i}$, as defined by the two stage, satisfies \[ \lim_{c_{i}\rightarrow 0}\left( m_{i}+r_{i}\right) \sqrt{c_{i}}=\sqrt{V_{i}} \] and since \[ \sqrt{\sum\limits_{j=1}^{n}c_{j}}\left( m_{i}+r_{i}\right) =\frac{\sqrt{c_{i} }\left( m_{i}+r_{i}\right) }{\sqrt{\frac{c_{i}}{\sum\limits_{j=1}^{n}c_{j}}}} \rightarrow \sqrt{\frac{V_{i}}{\lambda _{i}}}, \ as\ c\rightarrow 0, \] then \begin{equation} \frac{\left( \sqrt{U_{im_{i}}}-\left( m_{i}+r_{i}\right) \sqrt{c_{i}}\right) ^{2}}{\sqrt{\sum\limits_{j=1}^{n}c_{j}}\left( m_{i}+r_{i}\right) } \rightarrow 0,~a.s.,\ as\ c\rightarrow 0 \label{ae} \end{equation} To show the convergence in $L^{1}$, it will be sufficient to show the uniform integrability of the left hand side of (\ref{ae}). So, observe that \begin{eqnarray*} \frac{\left( \sqrt{U_{im_{i}}}-\left( m_{i}+r_{i}\right) \sqrt{c_{i}}\right) ^{2}}{\sqrt{\sum\limits_{j=1}^{n}c_{j}}\left( m_{i}+r_{i}\right) } &\leq & \frac{U_{im_{i}}+\left( m_{i}+r_{i}\right) ^{2}c_{i}}{\sqrt{ \sum\limits_{j=1}^{n}c_{j}}\left( m_{i}+r_{i}\right) } \\ &\leq &\frac{U_{im_{i}}}{\sqrt{U_{ik_{i}}}}+\sqrt{U_{ik_{i}}} \end{eqnarray*} $\sqrt{U_{ik_{i}}}$ is uniformly integrable, as a result of martingales L$^{p}$ convergence properties with $p=2$. Now, remark that \[ \frac{U_{im_{i}}}{\sqrt{U_{ik_{i}}}}\leq \max_{k^{\prime }}\sqrt{ U_{ik^{\prime }}} \] and for the remainder of the proof, we use Doob's inequality to show that $E\left[ \max_{k^{\prime }}\sqrt{U_{ik^{\prime }}}\right] <+\infty$. We have, \[ E\left[ \max_{k^{\prime }}\left( \sqrt{U_{ik^{\prime }}}\right) ^{2p}\right] \leq \left( \frac{2p}{2p-1}\right) ^{2p}E\left[ \left( \sqrt{V_{i}}\right) ^{2p}\right] <+\infty \] hence, since $p\geq 1,$ $\max_{k^{\prime }}\sqrt{U_{ik^{\prime }}}$ is integrable and the proof follows. \end{pf}
\section{Conclusion}
The proof of the first order asymptotic optimality for the two stage design has been obtained mainly through an adequate scaling of the approximated Bayes risk associated with squared error loss and cost, a lower bound for the scaled Bayes risk, martingales properties and Doob's inequality.
\section*{References}
\end{document}
|
arXiv
|
{
"id": "1202.5325.tex",
"language_detection_score": 0.5929726362228394,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{On some conjectures concerning Stern's sequence and its twist} \author{Michael Coons} \address{University of Waterloo, Dept.~of Pure Math., Waterloo, ON, N2L 3G1, Canada} \email{[email protected]} \thanks{Research supported by a Fields--Ontario Fellowship.}
\subjclass[2010]{Primary 11B37; Secondary 11B83} \keywords{Stern sequence, functional equations, binary expansion} \date{\today}
\begin{abstract} In a recent paper, Roland Bacher conjectured three identities concerning Stern's sequence and its twist. In this paper we prove Bacher's conjectures. Possibly of independent interest, we also give a way to compute the Stern value (or twisted Stern value) of a number based solely on its binary expansion. \end{abstract}
\maketitle
\section{Introduction}
We define the {\em Stern sequence} (also known as {\em Stern's diatomic sequence}) $\{s(n)\}_{n\geqslant 0}$ by $s(0)=0$, $s(1)=1$, and for all $n\geqslant 1$ by $$s(2n)=s(n),\qquad s(2n+1)=s(n)+s(n+1);$$ this is sequence A002487 in Sloane's list. We denote by $S(z)$ the generating function of the Stern sequence; that is, $$S(z):=\sum_{n\geqslant 0} s(n)z^n.$$ Stern's sequence has been well studied and has many interesting properties (see e.g., \cite{Dil1, Dil2, Leh1, Lin1} for details). One of the most interesting properties is that the sequence $\{s(n+1)/s(n)\}_{n\geqslant 1}$ is an enumeration of the positive reduced rationals without repeats.
Similarly Bacher \cite{B} introduced the {\em twisted Stern sequence} $\{t(n)\}_{n\geqslant 0}$ given by the recurrences $t(0)=0$, $t(1)=1$, and for $n\geqslant 1$ by $$t(2n)=-t(n),\qquad t(2n+1)=-t(n)-t(n+1).$$ We denote by $T(z)$ the generating function of the twisted Stern sequence; that is, $$T(z):=\sum_{n\geqslant 0} t(n)z^n.$$ Towards describing the relationship between the Stern sequence and its twist, Bacher \cite{B} gave many results, and two conjectures. As the main theorems of this article, we prove these conjectures, so we will state them as theorems (note that we have modified some of the notation).
\begin{theorem}\label{Bacherconj1} There exists an integral sequence $\{u(n)\}_{n\geqslant 0}$ such that for all $e\geqslant 0$ we have $$\sum_{n\geqslant 0} t(3\cdot 2^e+n)z^n=(-1)^eS(z)\sum_{n\geqslant 0}u(n)z^{n\cdot 2^e}.$$ \end{theorem}
Note that in this theorem (as in the original conjecture), it is implicit that the sequence $\{u(n)\}_{n\geqslant 0}$ is defined by the relationship $$U(z):=\sum_{n\geqslant 0}u(n)z^n =\frac{\sum_{n\geqslant 0}t(3+n)z^n}{S(z)}.$$
\begin{theorem}\label{Bacherconj2} (i) The series $$G(z):=\frac{\sum_{n\geqslant 0}(s(2+n)-s(1+n))z^n}{S(z)}$$ satisfies $$\sum_{n\geqslant 0}(s(2^{e+1}+n)-s(2^e+n))z^n=G(z^{2^e})S(z)$$ for all $e\in\mathbb{N}$.
Similarly, (ii) the series $$H(z):=-\frac{\sum_{n\geqslant 0}(t(2+n)+t(1+n))z^n}{S(z)}$$ satisfies $$(-1)^{e+1}\sum_{n\geqslant 0}(t(2^{e+1}+n)+t(2^e+n))z^n=H(z^{2^e})S(z)$$ for all $e\in\mathbb{N}$. \end{theorem}
These theorems were originally stated as Conjectures 1.3 and 3.2 in \cite{B}.
\section{Untwisting Bacher's First Conjecture}
In this section, we will prove Theorem \ref{Bacherconj1}, but first we note the following lemma which are a direct consequence of the definitions of the Stern sequence and its twist.
\begin{lemma}\label{AT} The generating series $S(z)=\sum_{n\geqslant 0}s(n)z^n$ and $T(z)=\sum_{n\geqslant 0}t(n)z^n$ satisfy the functional equations $$S(z^2)=\left(\frac{z}{1+z+z^2}\right)S(z)$$ and $$T(z^2)=\left(T(z)-2z\right)\left(\frac{-z}{1+z+z^2}\right),$$ respectively. \end{lemma}
We prove here only the functional equation for $T(z)$. The functional equation for the generating series of the Stern sequence is well--known; for details see, e.g., \cite{CoonsS1, Dil1}.
\begin{proof}[Proof of Lemma \ref{AT}] This is a straightforward calculation using the definition of $t(n)$. Note that \begin{align*} T(z)&= \sum_{n\geq 0}t(2n)z^{2n}+\sum_{n\geq 0}t(2n+1)z^{2n+1}\\ &=-\sum_{n\geq 0}t(n)z^{2n}+t(1)z+\sum_{n\geq 1}t(2n+1)z^{2n+1}\\ &=-T(z^2)+z-\sum_{n\geq 1}t(n)z^{2n+1}-\sum_{n\geq 1}t(n+1)z^{2n+1}\\ &=-T(z^2)+z-zT(z^2)-z^{-1}\sum_{n\geq 1}t(n+1)z^{2(n+1)}\\ &=-T(z^2)+2z-zT(z^2)-z^{-1}\sum_{n\geq 0}t(n+1)z^{2(n+1)}\\ &=-T(z^2)+2z-zT(z^2)-z^{-1}T(z^2). \end{align*} Solving for $T(z^2)$ gives $$T(z^2)=\left(T(z)-2z\right)\left(\frac{-z}{1+z+z^2}\right),$$ which is the desired result. \end{proof}
Since the proof of Theorem \ref{Bacherconj1} is easiest for the case $e=1$, and this case is indicative of the proof for the general case, we present it here separately.
\begin{proof}[Proof of Theorem \ref{Bacherconj1} for $e=1$] Recall that we define the sequence $\{u(n)\}_{n\geqslant 0}$ is by the relationship $$U(z):=\sum_{n\geqslant 0}u(n)z^n =\frac{\sum_{n\geqslant 0}t(3+n)z^n}{S(z)}.$$ Since $$\sum_{n\geqslant 0}t(3+n)z^n=\frac{1}{z^3}\left(T(z)+z^2-z\right),$$ we have that \begin{equation}\label{bUA}U(z)=\frac{T(z)+z^2-z}{z^3S(z)}=\frac{1}{z^3}\cdot\frac{T(z)}{S(z)}+\frac{z^2-z}{z^3}\cdot\frac{1}{S(z)}.\end{equation} Note that we are interested in a statement about the function $U(z^{2})$. We will use the functional equations for $S(z)$ and $T(z)$ to examine this quantity via \eqref{bUA}. Note that equation \eqref{bUA} gives, sending $z\mapsto z^{2}$ and using applying Lemma \ref{AT}, that $$U(z^2)=\frac{1}{z^6}\cdot\frac{T(z^2)}{S(z^2)}+\frac{z^4-z^2}{z^6}\cdot\frac{1}{S(z^2)}=\frac{1}{z^6S(z)}\left(2z-T(z)+(z^3-z)(1+z+z^2)\right).$$ Thus we have that \begin{align*} (-1)^1 S(z)U(z^2)&=\frac{-1}{z^6}\left(2z-T(z)-z-z^2-z^3+z^3+z^4+z^5\right)\\ &=\frac{1}{z^6}\left(T(z)-z+z^2-z^4-z^5\right)\\ &=\frac{1}{z^6}\sum_{n\geq 6}t(n)z^n\\ &=\sum_{n\geq 0}t(3\cdot 2+n)z^n, \end{align*} which is exactly what we wanted to show. \end{proof}
For the general case, complications arise in a few different places. The first is concerning $T(z^{2^e})$. We will build up the result with a sequence of lemmas to avoid a long and calculation--heavy proof of Theorem \ref{Bacherconj1}.
\begin{lemma}\label{T2e} For all $e\geq 1$ we have $$T(z^{2^e})=T(z)\prod_{i=0}^{e-1}\left(\frac{-z^{2^i}}{1+z^{2^i}+z^{2^{i+1}}}\right)-2\sum_{j=0}^{e-1}z^{2^j}\prod_{i=j}^{e-1}\left(\frac{-z^{2^i}}{1+z^{2^i}+z^{2^{i+1}}}\right).$$ \end{lemma}
\begin{proof} We give a proof by induction. Note that for $e=1$, the right--hand side of the desired equality is $$T(z)\left(\frac{-z}{1+z+z^{2}}\right)-2z\left(\frac{-z}{1+z+z^{2}}\right)=\left(T(z)-2z\right)\left(\frac{-z}{1+z+z^{2}}\right)=T(z^2)$$ where the last equality follows from Lemma \ref{AT}.
Now suppose the identity holds for $e-1$. Then, again using Lemma \ref{AT}, we have \begin{align*} T(z^{2^e}) = T((z^2)^{2^{e-1}})&=T(z^2)\prod_{i=0}^{e-2}\left(\frac{-z^{2^{i+1}}}{1+z^{2^{i+1}}+z^{2^{i+2}}}\right)-2\sum_{j=0}^{e-2}z^{2^{j+1}}\prod_{i=j}^{e-2}\left(\frac{-z^{2^{i+1}}}{1+z^{2^{i+1}}+z^{2^{i+2}}}\right)\\ &=\left(T(z)-2z\right)\left(\frac{-z}{1+z+z^{2}}\right)\prod_{i=1}^{e-1}\left(\frac{-z^{2^{i}}}{1+z^{2^{i}}+z^{2^{i+1}}}\right)\\ &\qquad\qquad-2\sum_{j=1}^{e-1}z^{2^{j}}\prod_{i=j}^{e-1}\left(\frac{-z^{2^{i}}}{1+z^{2^{i}}+z^{2^{i+1}}}\right)\\ &=\left(T(z)-2z\right)\prod_{i=0}^{e-1}\left(\frac{-z^{2^{i}}}{1+z^{2^{i}}+z^{2^{i+1}}}\right)-2\sum_{j=1}^{e-1}z^{2^{j}}\prod_{i=j}^{e-1}\left(\frac{-z^{2^{i}}}{1+z^{2^{i}}+z^{2^{i+1}}}\right)\\ &=T(z)\prod_{i=0}^{e-1}\left(\frac{-z^{2^{i}}}{1+z^{2^{i}}+z^{2^{i+1}}}\right)-2\sum_{j=0}^{e-1}z^{2^{j}}\prod_{i=j}^{e-1}\left(\frac{-z^{2^{i}}}{1+z^{2^{i}}+z^{2^{i+1}}}\right). \end{align*} Hence, by induction, the identity is true for all $e\geq 1$. \end{proof}
We will need the following result for our next lemma.
\begin{theorem}[Bacher \cite{B}]\label{B1.4} For all $e\geqslant 1$, we have $$\prod_{i=0}^{e-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)=\frac{(-1)^e}{z(1+z^{2^e})}\sum_{n=0}^{3\cdot 2^e}t(3\cdot 2^e+n)z^n.$$ \end{theorem}
The following lemma is similar to the comment made in Remark 1.5 of \cite{B}.
\begin{lemma} For all $e\geq 1$, we have that $$\sum_{n=0}^{3\cdot 2^e}t(n)z^n=z-z^2+\sum_{k=0}^{e-1}(-1)^kz^{3\cdot 2^k+1}(z^{2^k}+1)\prod_{i=0}^{k-1}(1+z^{2^i}+z^{2^{i+1}}).$$ \end{lemma}
\begin{proof} If $e\geq 1$, then we have \begin{align*} \sum_{n=0}^{3\cdot 2^e}t(n)z^n &= z-z^2+\sum_{k=0}^{e-1}\sum_{n=3\cdot 2^k}^{3\cdot 2^{k+1}}t(n)z^n\\ &= z-z^2+\sum_{k=0}^{e-1}\sum_{n=0}^{3\cdot 2^{k}}t(3\cdot 2^{k}+n)z^{n+3\cdot 2^{k}}\\ &= z-z^2+\sum_{k=0}^{e-1}z^{3\cdot 2^k}\sum_{n=0}^{3\cdot 2^{k}}t(3\cdot 2^{k}+n)z^{n}. \end{align*} Applying Theorem \ref{B1.4}, we have that $$\sum_{n=0}^{3\cdot 2^e}t(n)z^n=z-z^2+\sum_{k=0}^{e-1}z^{3\cdot 2^k}(-1)^kz(z^{2^k}+1)\prod_{i=0}^{k-1}(1+z^{2^i}+z^{2^{i+1}}),$$ which after some trivial term arrangement gives the result. \end{proof}
\begin{lemma}\label{tspp} For all $e\geq 1$, we have $$\sum_{n=0}^{3\cdot 2^e} t(n)z^n=2z\sum_{j=0}^{e-1}(-1)^j\prod_{i=0}^{j-1}(1+z^{2^i}+z^{2^{i+1}})-(-1)^ez(z^{2^e}-1)\prod_{i=0}^{e-1} (1+z^{2^i}+z^{2^{i+1}}).$$ \end{lemma}
\begin{proof} This lemma is again proved by induction, using the result of the previous lemma. Note that in view of the previous lemma, by subtracting the first term on the right--hand side of the desired equality, it is enough to show that for all $e\geq 1$, we have \begin{multline}\label{lrhs}z-z^2+\sum_{k=0}^{e-1}(-1)^k\left(z^{4\cdot 2^k}+z^{3\cdot 2^k}-2\right)z\prod_{i=0}^{k-1}(1+z^{2^i}+z^{2^{i+1}})\\ =-(-1)^ez(z^{2^e}-1)\prod_{i=0}^{e-1} (1+z^{2^i}+z^{2^{i+1}}).\end{multline}
If $e=1$, then the left--hand side of \eqref{lrhs} is $$z-z^2+(z^4+z^3-s)z=-z-z^2+z^4+z^5,$$ and the right--hand side of \eqref{lrhs} is $$-(-1)z(z^2-1)(1+z+z^2)=-z-z^2+z^4+z^5,$$ so that \eqref{lrhs} holds for $e=1$.
Now suppose that \eqref{lrhs} holds for $e-1$. Then \begin{align*} z-z^2+\sum_{k=0}^{e-1}(-1)^k&\left(z^{4\cdot 2^k}+z^{3\cdot 2^k}-2\right)z\prod_{i=0}^{k-1}(1+z^{2^i}+z^{2^{i+1}})\\ &= (-1)^{e-1}\left(z^{4\cdot 2^{e-1}}+z^{3\cdot 2^{e-1}}-2\right)z\prod_{i=0}^{e-2}(1+z^{2^i}+z^{2^{i+1}})\\ &\qquad\qquad+z-z^2+\sum_{k=0}^{e-2}(-1)^k\left(z^{4\cdot 2^k}+z^{3\cdot 2^k}-2\right)z\prod_{i=0}^{k-1}(1+z^{2^i}+z^{2^{i+1}})\\ &= (-1)^{e-1}\left(z^{4\cdot 2^{e-1}}+z^{3\cdot 2^{e-1}}-2\right)z\prod_{i=0}^{e-2}(1+z^{2^i}+z^{2^{i+1}})\\ &\qquad\qquad-(-1)^{e-1}z(z^{2^{e-1}}-1)\prod_{i=0}^{e-2} (1+z^{2^i}+z^{2^{i+1}})\end{align*} Factoring out the product we thus have that \begin{align*} z-z^2+\sum_{k=0}^{e-1}(-1)^k&\left(z^{4\cdot 2^k}+z^{3\cdot 2^k}-2\right)z\prod_{i=0}^{k-1}(1+z^{2^i}+z^{2^{i+1}})\\ &=(-1)^e\prod_{i=0}^{e-2} (1+z^{2^i}+z^{2^{i+1}})\cdot \left(-\left(z^{4\cdot 2^{e-1}}+z^{3\cdot 2^{e-1}}-2\right)z+z(z^{2^{e-1}}-1)\right)\\ &=-(-1)^ez\prod_{i=0}^{e-2} (1+z^{2^i}+z^{2^{i+1}})\cdot \left(z^{4\cdot 2^{e-1}}+z^{3\cdot 2^{e-1}}-z^{2\cdot 2^{e-1}}-1\right)\\ &=-(-1)^ez\prod_{i=0}^{e-2} (1+z^{2^i}+z^{2^{i+1}})\cdot (z^{2^e}-1)(1+z^{2^{e-1}}+z^{2^{e}})\\ &=-(-1)^ez(z^{2^e}-1)\prod_{i=0}^{e-1} (1+z^{2^i}+z^{2^{i+1}}), \end{align*} so that by induction, \eqref{lrhs} holds for all $e\geq 1$. \end{proof}
With these lemmas in place we are in position to prove Theorem \ref{Bacherconj1}.
\begin{proof}[Proof of Theorem \ref{Bacherconj1}] We start by restating \eqref{bUA}; that is $$U(z)=\frac{T(z)+z^2-z}{z^3S(z)}=\frac{1}{z^3}\cdot\frac{T(z)}{S(z)}+\frac{z^2-z}{z^3}\cdot\frac{1}{S(z)}.$$ Sending $z\mapsto z^{2^e},$ we have that $$U(z^{2^e})=\frac{1}{z^{3\cdot 2^e}}\cdot\frac{T(z^{2^e})}{S(z^{2^e})}+\frac{z^{2^{e+1}}-z^{2^e}}{z^{3\cdot 2^e}}\cdot\frac{1}{S(z^{2^e})}=\frac{1}{z^{3\cdot 2^e}}\cdot\frac{T(z^{2^e})}{S(z^{2^e})}+\frac{z^{2^{e+1}}-z^{2^e}}{z^{3\cdot 2^e}z^{2^e-1}S(z)}\cdot\prod_{i=0}^{e-1} (1+z^{2^i}+z^{2^{i+1}}),$$ where we have used the functional equation for $S(z)$ to give the last equality. Using Lemma \ref{T2e} and the functional equation for $S(z)$, we have that \begin{align}\nonumber\frac{T(z^{2^e})}{S(z^{2^e})}&=\frac{T(z)\prod_{i=0}^{e-1}\left(\frac{-z^{2^i}}{1+z^{2^i}+z^{2^{i+1}}}\right)-2\sum_{j=0}^{e-1}z^{2^j}\prod_{i=j}^{e-1}\left(\frac{-z^{2^i}}{1+z^{2^i}+z^{2^{i+1}}}\right)}{S(z)}\cdot\prod_{i=0}^{e-1} \left(\frac{1+z^{2^i}+z^{2^{i+1}}}{z^{2^i}}\right)\\ \label{ToverS}&=(-1)^e\frac{T(z)}{S(z)}-(-1)^e\frac{2z}{S(z)}\sum_{j=0}^{e-1}(-1)^{j}\prod_{i=0}^{j-1}\left({1+z^{2^i}+z^{2^{i+1}}}\right).\end{align} Applying this to the expression for $U(z^{2^e})$ we have, multiplying by $(-1)^eS(z)$, that \begin{multline*}(-1)^eS(z)U(z^{2^e})=\frac{1}{z^{3\cdot 2^e}}\left(T(z)-2z\sum_{j=0}^{e-1}(-1)^{j}\prod_{i=0}^{j-1}\left({1+z^{2^i}+z^{2^{i+1}}}\right)\right.\\ \left.+(-1)^ez(z^{2^e}-1)\prod_{i=0}^{e-1}\left({1+z^{2^i}+z^{2^{i+1}}}\right)\right).\end{multline*} Now by Lemma \ref{tspp}, this reduces to $$(-1)^eS(z)U(z^{2^e})=\frac{1}{z^{3\cdot 2^e}}\left(T(z)-\sum_{n=0}^{3\cdot 2^e} t(n)z^n\right)=\sum_{n\geq 0}t(3\cdot 2^3+n)z^n,$$ which proves the theorem. \end{proof}
\section{Untwisting Bacher's Second Conjecture}
In this section, we will prove Theorem \ref{Bacherconj2}. For ease of reading we have separated the proofs of the two parts of Theorem \ref{Bacherconj2}.
To prove Theorem \ref{Bacherconj2}(i) we will need the following lemma.
\begin{lemma}\label{pss} For all $k\geq 0$ we have that $$z\prod_{i=0}^{k-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)=\sum_{n=1}^{2^k}s(n)z^n+\sum_{n=1}^{2^k-1}s(2^k-n)z^{n+2^k}.$$ \end{lemma}
\begin{proof} Again, we prove by induction. Note that for $k=0$, the product and the left--most sum are both empty, thus they are equal to $1$ and $0$, respectively. Since $$z=s(1)z=\sum_{n=1}^{2^0}s(n)z^n$$ the theorem is true for $k=0$. To use some nonempty terms, we consider the case $k=1$. Then we have $$z\prod_{i=0}^{1-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)=z+z^2+z^3=\sum_{n=1}^{2^1}s(n)z^n+\sum_{n=1}^{2^1-1}s(2^1-n)z^{n+2^1},$$ so the theorem holds for $k=1$.
Now suppose the theorem holds for $k-1$. Then \begin{align*} z\prod_{i=0}^{k-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)&=\left(1+z^{2^{k-1}}+z^{2^{k}}\right)\cdot z\prod_{i=0}^{k-2}\left(1+z^{2^i}+z^{2^{i+1}}\right)\\ &=\left(1+z^{2^{k-1}}+z^{2^{k}}\right)\left(\sum_{n=1}^{2^{k-1}}s(n)z^n+\sum_{n=1}^{2^{k-1}-1}s(2^{k-1}-n)z^{n+2^{k-1}}\right)\\ &=\left(\sum_{n=1}^{2^{k-1}}s(n)z^n+\sum_{n=1}^{2^{k-1}-1}s(2^{k-1}-n)z^{n+2^{k-1}}+\sum_{n=1}^{2^{k-1}}s(n)z^{n+2^{k-1}}\right)\\ &\qquad +\left(\sum_{n=1}^{2^{k-1}-1}s(2^{k-1}-n)z^{n+2^{k}}+\sum_{n=1}^{2^{k-1}}s(n)z^{n+2^{k}}\right.\\ &\qquad\qquad\left.+\sum_{n=1}^{2^{k-1}-1}s(2^{k-1}-n)z^{n+3\cdot 2^{k-1}}\right)\\ &=\Sigma_1+\Sigma_2, \end{align*} where $\Sigma_1$ and $\Sigma_2$ represent the triplets of sums from the previous line (we have grouped the last sums in triplets since we will deal with them that way. Note that we have \begin{multline*}\Sigma_1=\sum_{n=1}^{2^{k-1}}s(n)z^n+s(2^{k-1})z^{2^k}+\sum_{n=1}^{2^{k-1}-1}\left(s(n)+s(2^{k-1}-n)\right)z^{n+2^{k-1}}\\ =\sum_{n=1}^{2^{k-1}}s(n)z^n+s(2^{k})z^{2^k}+\sum_{n=1}^{2^{k-1}-1}s(2^{k-1}+n)z^{n+2^{k-1}}=\sum_{n=1}^{2^{k}}s(n)z^n,\end{multline*} where we have used the fact that $s(2n)=s(n)$ and for $n\in[0,2^j]$ the identity $s(2^j+n)=s(2^j-n)+s(n)$ holds (see, e.g., \cite[Theorem 1.2(i)]{B} for details). Similarly, since $2^{k-1}-n=2^k-(n+2^{k-1})$ and $$s(2^{k-1}-n)+s(n)=s(2^{k-1}+n)=s(2^k+n)-s(n)=s(2^k-n)$$ (see Proposition 3.1(i) and Theorem 1.2(i) of \cite{B}), we have that \begin{align*}\Sigma_2&=\sum_{n=1}^{2^{k-1}-1}\left(s(2^{k-1}-n)+s(n)\right)z^{n+2^k}+s(2^{k-1})z^{3\cdot 2^{k-1}}+\sum_{n=1}^{2^{k-1}-1}s(2^{k-1}-n)z^{n+2^{k-1}+2^k}\\ &= \sum_{n=1}^{2^{k-1}-1}s(2^{k}-n)z^{n+2^k}+s(2^{k-1})z^{3\cdot 2^{k-1}}+\sum_{n=2^{k-1}-1}^{2^{k}-1}s(2^{k}-n)z^{n+2^k}\\ &=\sum_{n=1}^{2^k-1}s(2^k-n)z^{n+2^k}.\end{align*} Thus $$\Sigma_1+\Sigma_2=\sum_{n=1}^{2^{k}}s(n)z^n+\sum_{n=1}^{2^k-1}s(2^k-n)z^{n+2^k},$$ and by induction the lemma is proved. \end{proof}
\begin{proof}[Proof of Theorem \ref{Bacherconj2}(i)] We denote as before the generating series of the Stern sequence by $S(z)$. Splitting up the sum in the definition of $G(z)$ we see that $$G(z)=\frac{1}{z^{2}}(1-z)-\frac{1}{z S(z)},$$ so that using the functional equation for $S(z)$ we have \begin{align*}G(z^{2^e}) = \frac{1}{z^{2^{e+1}}}(1-z^{2^e})-\frac{1}{z^{2^e} S(z^{2^e})}= \frac{1}{z^{2^{e+1}}}(1-z^{2^e})-\frac{1}{z^{2^{e}}}\cdot\frac{\prod_{i=0}^{e-1}(1+z^{2^i}+z^{2^{i+1}})}{z^{2^{e}-1}S(z)}. \end{align*} This gives \begin{equation}\label{32rhsi} G(z^{2^e})S(z)=\frac{1}{z^{2^{e+1}}}\left((1-z^{2^e})S(z)-z\prod_{i=0}^{e-1}(1+z^{2^i}+z^{2^{i+1}})\right).\end{equation} We use the previous lemma to deal with the right--hand side of \eqref{32rhsi}; that is, the previous lemma gives that \begin{align*} (1-z^{2^e})S(z)-z\prod_{i=0}^{e-1}(1+z^{2^i}+z^{2^{i+1}})&= \sum_{n\geq 1}s(n)z^n-\sum_{n=1}^{2^e}s(n)z^n\\ &\qquad\qquad-\sum_{n\geq 1}s(n)z^{n+2^e}-\sum_{n=1}^{2^e-1}s(2^e-n)z^{n+2^e}\\ &=\sum_{n\geq 2^e+1} s(n)z^n-\sum_{n\geq 2^e}s(n)z^{n+2^e}\\ &\qquad\qquad-\sum_{n=1}^{2^e-1}(s(n)+s(2^e-n))z^{n+2^e}\\ &=\sum_{n\geq 1} s(2^e+n)z^{n+2^e}-\sum_{n\geq 2^e}s(n)z^{n+2^e}\\ &\qquad\qquad-\sum_{n=1}^{2^e-1}s(2^e+n)z^{n+2^e}\\ &=\sum_{n\geq 0} s(2^{e+1}+n)z^{n+2^{e+1}}-\sum_{n\geq 0}s(2^e+n)z^{n+2^{e+1}}. \end{align*} Dividing the last line by $z^{2^{e+1}}$ gives the desired result. This proves the theorem. \end{proof}
The proof of the second part of the theorem follows similarly. We will use the following lemma.
\begin{lemma}[Bacher \cite{B}] For $n$ satisfying $1\leq n\leq 2^e$ we have that \begin{enumerate} \item[(i)] $t(2^{e+1}+n)+t(2^{e}+n)=(-1)^{e+1}s(n),$ \item[(ii)] $t(2^{e}+n)=(-1)^e(s(2^e-n)-s(n))$, \item[(iii)] $t(2^{e+1}+n)=(-1)^{e+1}s(2^{e}-n)$. \end{enumerate} \end{lemma}
\begin{proof} Parts (i) and (ii) are given in Proposition 3.1 and Theorem 1.2 of \cite{B}, respectively. Part (iii) follows easily from (i) and (ii).
Note that (i) gives that $$t(2^{e+1}+n)=(-1)^{e+1}s(n)-t(2^{e}+n),$$ which by (ii) becomes \begin{equation*}t(2^{e+1}+n)=(-1)^{e+1}s(n)+(-1)^{e+1}s(2^e-n)-(-1)^{e+1}s(n)=(-1)^{e+1}s(2^e-n).\qedhere\end{equation*} \end{proof}
\begin{proof}[Proof of Theorem \ref{Bacherconj2} (ii)] We denote as before the generating series of the Stern sequence by $S(z)$. Splitting up the sum in the definition of $H(z)$ we see that $$H(z)=\frac{1}{z S(z)}-\frac{1+z}{z^{2}}\cdot\frac{T(z)}{S(z)}.$$ Since we will need to consider $H(z^{2^e})$, we will need to compute $\frac{T(z^{2^e})}{S(z^{2^e})}.$ Fortunately we have done this in the proof of Theorem \ref{Bacherconj1}, in \eqref{ToverS}, and so we use this expression here. Thus, applying the functional equation for $S(z)$, we have that \begin{multline*} H(z^{2^e})=\frac{\prod_{i=0}^{e-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)}{z^{2^{e+1}-1}S(z)}-(-1)^e\left(\frac{1+z^{2^e}}{z^{2^{e+1}}}\right)\frac{T(z)}{S(z)}\\ +(-1)^e\left(\frac{1+z^{2^e}}{z^{2^{e+1}}}\right)\frac{2z}{S(z)}\sum_{j=0}^{e-1}(-1)^j\prod_{i=0}^{j-1}\left(1+z^{2^i}+z^{2^{i+1}}\right),\end{multline*} so that \begin{multline*} (-1)^{e+1}H(z^{2^e})S(z)=\frac{1}{z^{2^{e+1}}}\left((1+z^{2^e})T(z)-(-1)^ez\prod_{i=0}^{e-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)\right.\\ \left.-2z(1+z^{2^e})\sum_{j=0}^{e-1}(-1)^j\prod_{i=0}^{j-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)\right).\end{multline*} An application of Lemma \ref{tspp} gives \begin{align*} (-1)^{e+1}H(z^{2^e})S(z)&=\frac{1}{z^{2^{e+1}}}\left((1+z^{2^e})T(z)-(-1)^ez\prod_{i=0}^{e-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)\right.\\ &\ \ \left.-(1+z^{2^e})\sum_{n=0}^{3\cdot 2^e}t(n)z^n-(-1)^ez(z^{2^e}-1)(1+z^{2^e})\prod_{i=0}^{e-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)\right)\\ &=\frac{1}{z^{2^{e+1}}}\left(\mathfrak{S}_1+\mathfrak{S}_2+\mathfrak{S}_3+\mathfrak{S}_3\right),\end{align*} where we have used the $\mathfrak{S}_i$ to indicate the terms in the previous line.
Note that \begin{align*}\mathfrak{S}_1+\mathfrak{S}_3&=(1+z^{2^e})\left(T(z)-\sum_{n=0}^{3\cdot 2^e}t(n)z^n\right)\\ &=(1+z^{2^e})\sum_{n\geq 1}t(3\cdot 2^e+n)z^{n+3\cdot2^e}\\ &=\sum_{n\geq 2^e+1}t(2^{e+1}+n)z^{n+2^{e+1}}+\sum_{n\geq 2^{e+1}+1}t(2^{e}+n)z^{n+2^{e+1}},\end{align*} so that $$\frac{\mathfrak{S}_1+\mathfrak{S}_3}{z^{2^e+1}}=\sum_{n\geq 2^e+1}t(2^{e+1}+n)z^{n}+\sum_{n\geq 2^{e+1}+1}t(2^{e}+n)z^{n}.$$ Using Lemma \ref{pss}, we have \begin{align*} \frac{\mathfrak{S}_2+\mathfrak{S}_4}{z^{2^{e+1}}}&=\frac{1}{z^{2^{e+1}}}\left(-(-1)^e-(-1)^e(z^{2^{e+1}}-1)\right)z\prod_{i=0}^{e-1}\left(1+z^{2^i}+z^{2^{i+1}}\right)\\ &=\frac{(-1)^{e+1}}{z^{2^{e+1}}}\cdot z^{2^{e+1}}\left(\sum_{n=1}^{2^e}s(n)z^n+\sum_{n=1}^{2^e-1}s(2^e-n)z^{n+2^e}\right)\\ &=(-1)^{e+1}\left(\sum_{n=1}^{2^e}s(n)z^n+\sum_{n=1}^{2^e-1}s(2^e-n)z^{n+2^e}\right). \end{align*} Using the proceeding lemma and the fact that $t(2^{e+1})+t(2^e)=0$ so that we can add in a zero term, we have that \begin{align*} \frac{\mathfrak{S}_2+\mathfrak{S}_4}{z^{2^{e+1}}}&=\sum_{n=1}^{2^e}(t(2^{e+1}+n)+t(2^e+n))z^n+\sum_{n=1}^{2^e-1}t(2^{e+1}+n)z^{n+2^e}\\ &=\sum_{n=1}^{2^e}(t(2^{e+1}+n)+t(2^e+n))z^n+\sum_{n=2^e}^{2^{e+1}-1}t(2^{e}+n)z^{n}\\ &=\sum_{n=0}^{2^e}(t(2^{e+1}+n)+t(2^e+n))z^n+\sum_{n=2^e}^{2^{e+1}-1}t(2^{e}+n)z^{n}\\ &=\sum_{n=0}^{2^e}t(2^{e+1}+n)z^n+\sum_{n=0}^{2^{e+1}-1}t(2^{e}+n)z^{n}. \end{align*} Putting together these results gives \begin{align*}(-1)^{e+1}H(z^{2^e})S(z)&=\frac{1}{z^{2^{e+1}}}\left(\mathfrak{S}_1+\mathfrak{S}_2+\mathfrak{S}_3+\mathfrak{S}_3\right)\\ &=\sum_{n\geq 0}t(2^{e+1}+n)z^n+\sum_{n\geq 0}t(2^{e}+n)z^{n},\end{align*} which proves the theorem. \end{proof}
\section{Computing with binary expansions}
To gain intuition regarding Bacher's conjectures mentioned in the first section, we found it very useful to understand what happens to the Stern sequence and its twist at sums of powers of $2$. Thus, in this section we prove the following theorem which removes the need to use the recurrences to give the values of the Stern sequence.
\begin{theorem} Let $n\geqslant 4$ and write $n=\sum_{i=0}^m 2^ib_i$, the binary expansion of $n$. Then $$s(n)=\left[\begin{matrix}1 & 1\end{matrix}\right]\left(\prod_{i=1}^{m-1}\left[\begin{matrix}1 & 1-b_i\\ b_i & 1 \end{matrix}\right]\right)\left[\begin{matrix}1\\ b_0 \end{matrix}\right]$$ \end{theorem}
\begin{proof} Note that from the definition of the Stern sequence we easily have that $$s(2a+x)=s(a)+x\cdot s(a+1)\qquad (x\in\{0,1\}).$$ It follows that for each $k\in\{0,1,\ldots,m-1\}$ we have that both $$s\left(\sum_{i=k}^m 2^{i-k}b_i\right)=s\left(\sum_{i=k+1}^m 2^{i-(k+1)}b_i\right)+b_k\cdot s\left(1+\sum_{i=k+1}^m 2^{i-(k+1)}b_i\right)$$ and $$s\left(1+\sum_{i=k}^m 2^{i-k}b_i\right)=(1-b_k)\cdot s\left(\sum_{i=k+1}^m 2^{i-(k+1)}b_i\right)+s\left(1+\sum_{i=k+1}^m 2^{i-(k+1)}b_i\right).$$ Starting with $k=0$ and applying the above equalities, and using the fact that $b_m=1$ so that $s(b_m)=s(b_m+1)=1$ gives the result. \end{proof}
We have a similar result for the twisted Stern sequence whose proof is only trivially different from the above and so we have omitted it.
\begin{theorem} Let $n\geqslant 4$ and write $n=\sum_{i=0}^m 2^ib_i$, the binary expansion of $n$. Then $$t(n)=(-1)^m\left[\begin{matrix}1 & -1\end{matrix}\right]\left(\prod_{i=1}^{m-1}\left[\begin{matrix}1 & 1-b_i\\ b_i & 1 \end{matrix}\right]\right)\left[\begin{matrix}1\\ b_0 \end{matrix}\right].$$ \end{theorem}
Indeed, both $s(n)$ and $t(n)$ are $2$--regular, and so the fact that $s(n)$ and $t(n)$ satisfy theorems like the two above is provided by Lemma 4.1 of \cite{AS} (note that while the existence is proven, the matrices are not explicitly given there).
\end{document}
|
arXiv
|
{
"id": "1008.0193.tex",
"language_detection_score": 0.5258661508560181,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{\centerline{\large \bf INVARIANTS OF LINKS OF CONWAY TYPE} \thispagestyle{firststyle} \section{Introduction}The purpose of this paper is to present a certain combinatorial method of constructing invariants of isotopy classes of oriented tame links. This arises as a generalization of the known polynomial invariants of Conway and Jones. These invariants have one striking common feature. If $L_+, L_-$ and $L_0$ are diagrams of oriented links which are identical, except near one crossing point, where they look like in Fig. 1.1\footnote{Added for e-print: we follow here the old Conway's convention \cite{C}; in modern literature the role of $L_+$ and $L_-$ is usually inverted. In \cite{Prz} the new convention is already used.}, then $w_{L_+}$ is uniquely determined by $w_{L_-}$ and $w_{L_0}$, and also $w_{L_-}$ is uniquely determined by $w_{L_+}$ and $w_{L_0}$. Here $w$ denotes the considered invariants (we will often take the liberty of speaking about the value of an invariant for a specific link or diagram of a link rather than for an isotopy class of links). In the above context, we agree to write $L_+^p, L_-^p$ and $L_0^p$ if we need the crossing point to be explicitly specified. \ \\ \ \\ \centerline{{\psfig{figure=SkeinTriplePT1.eps,height=3.5cm}}}\ \\
\centerline{\footnotesize{Fig. 1.1}} \indent In this paper we will consider the following general situation. Assume we are given an algebra $\A$ with a countable number of
$0$-argument operations $a_1, a_2,..., a_n,...$ and two 2-argument operations $|$ and $*$. We would like to construct invariants
satisfying the conditions \begin{align*}
w_{L_+}= &\ w_{L_-} | w_{L_0} \text{ and}\\
w_{L_-}= &\ w_{L_+} * w_{L_0} \text{ and} \\
w_{T_n}= &\ a_n \text{ for $T_n$ being a trivial link of n components.} \end{align*}
\indent We say that $(\A, a_1, a_2,..., |, *)$ is a Conway algebra if the following conditions are satisfied\footnote{Added for e-print: We were unaware, when writing this paper, that the condition 1.5, $(a\!*\!b)\!*\!(c\!*\!d) = (a\!*\!c)\!*\!(b\!*\!d)$, was used already for over 50 years, first time in \cite{Bu-Ma}, under various names, e.g. entropic condition (see for example \cite{N-P}).}: \\
${1.1} \quad a_n | a_{n+1} = a_n \\ {1.2} \quad a_n * a_{n+1} = a_n $\\ $\left. \begin{aligned}
{1.3} && &(a|b)|(c|d) = (a|c)|(b|d) \\
{1.4} && &(a|b)\!*\!(c|d) = (a\!*\!c)|(b\!*\!d) \\ {1.5} && &(a\!*\!b)\!*\!(c\!*\!d) = (a\!*\!c)\!*\!(b\!*\!d) \end{aligned} \right\} \quad \text{transposition properties}$ \\
${1.6} \quad (a|b)*b = a \\
{1.7} \quad (a*b)|b = a.$\\
\\ We will prove the following theorem: \\ \begin{theorem}\textbf{1.8.} For a given Conway algebra $\A$ there exists a uniquely determined invariant $w$ which attaches an element $w_L$ from $\A$ to every isotopy class of oriented links and satisfies the conditions\\ ${(1)} \quad w_{T_n} = a_n \hspace{2.5cm} \text{ initial conditions}$\\ $\left. \begin{aligned}
{(2)}&& &w_{L_+} = &\ w_{L_-} | w_{L_0}\\ {(3)}&& &w_{L_-} = &\ w_{L_+} * w_{L_0} \end{aligned} \right.\bigg\} \ Conway\ relations$ \end{theorem} It will be proved in \S{2}. \\ \indent Let us write here a few words about the geometrical meaning of the axioms 1.1-1.7 of Conway algebra. Relations 1.1 and 1.2 are introduced to reflect the following geometrical relations between the diagrams of trivial links of $n$ and $n+1$ components:
\centerline{\epsfig{figure=Figure12.eps,height=10cm}} \centerline{\footnotesize{Fig. 1.2}}
Relations 1.3, 1.4, and 1.5 arise when considering rearranging a link at two crossings of the diagram but in different order. It will become clear in \S{2}. Relations 1.6 and 1.7 reflect the fact that we need the operations $|$ and $*$ to be in some sense opposite one to another. \\
\begin{example}\textbf{1.9.} (Number of components). Set $\A=N-$ the set of natural numbers; $a_i=i$ and $i|j=i*j=i$. This algebra yields the number of components of a link.\end{example}
\begin{example}\textbf{1.10.} Set $\A=$ \big\{0, 1, 2\big\}; the operation $*$ is equal to $|$ and
$0|0\!=\!1,\ 1|0\!=\!0,\ 2|0\!=\!2,\ 0|1\!=\!0,\ 1|1\!=\!2,\ 2|1\!=\!1,\ 0|2\!=\!2,\ 1|2\!=\!1,\ 2|2\!=\!0$. Furthermore $a_i \equiv i \mod3.$ The invariant defined by this algebra distinguishes, for example, the trefoil knot from the trivial knot. \end{example} \begin{example}\textbf{1.11.}{(a)} $\A=Z [x^{\mp 1} , y^{\mp 1}, z]$; $a_1 = 1, a_2=x+y+z,\dots, a_i = (x+y)^{i-1} + z(x+y)^{i-2} + \dots+z(x+y)+z=(x+y)^{i-1}+z \big( \frac{(x+y)^{i-1}-1}{x+y-1} \big)$,{\small ...} .
We define $|$ and $*$ as follows: $w_2 | w_0 = w_1$ and $w_1 * w_0 = w_2$ where \begin{flalign*}
{1.12}&& xw_1+yw_2 &=w_0-z, \ \ w_1,w_2,w_0 \in \A.& \end{flalign*}
\end{example} \indent (b) $\A = Z$[ $x^{\mp1}$ , $y^{\mp1}$] is obtained from the algebra described in (a) by substitution $z=0$. In particular $a_i=(x+y)^{i-1}$ and 1.12 reduces to: \begin{flalign*}
{1.13}&& xw_1+yw_2 &= w_0.& \end{flalign*} We describe this algebra for two reasons: \\ \indent $-$first, the invariant of links defined by this algebra is the simplest generalization of the Conway polynomial \big(substitute $x=- \frac{1}{z}$, $y=\frac{1}{z}$, {[$K$-1]}\big) and the Jones polynomial \big(substitute $x= \frac{1}{t} \frac{1}{\sqrt{t}-\frac{1}{\sqrt{t}}}$, $y= \frac{-t}{\sqrt{t}-\frac{1}{\sqrt{t}}}$ {[$J$]}\big); \\ \indent$-$second, this invariant behaves well under disjoint and connected sums of links: \begin{center}
$\qquad P_{L_1 \sqcup L_2} (x,y)\!=\! (x+y)P_{L_1{\sharp} L_2}(x,y)\!=\!(x+y)P_{L_1}(x,y)\!\cdot\!P_{L_2}(x,y)$ \\
where $P_L\!(x,y)$ is a polynomial invariant of links yielded by $\A$.\\
\begin{example} \textbf{1.14.} (Linking number). Set $\A\!=\!N\!\times\!Z$, $a_i\!=\!(i,0)$ and\end{example} \begin{equation*}
(a,b)|(c,d)=\begin{cases}
(a,b-1) \ \textup{if} \ a > c \\
(a,b) \qquad \textup{if} \ a \leqslant c \end{cases} \end{equation*} \begin{equation*} (a,b)*(c,d)=\begin{cases}
(a,b+1)\ \textup{if} \ a > c \\
(a,b)\ \qquad \textup{if}\ a \leqslant c \end{cases} \end{equation*} \end{center} \indent The invariant associated to a link is a pair (number of components, linking number). \begin{remark}\textbf{1.15.} It may happen that for each pair $u,v\!\in\!\A$
there exists exactly one $w\!\in\!\A$ such that $v|w\!=\!u$ and $u\!*\!w\!=\!v$. Then we can introduce a new operation $\circ$: $\A \times \A \rightarrow \A$ putting $u \circ v=w$ (we have such a situation in Examples $1.10$ and $1.11$
but not in $1.9$ where $2|1\!=\!2\!*\!1\!=\!2\!=\!2|3\!=\!2\!*\!3$). Then $a_n\!=\!a_{n-1}\circ a_{n-1}$. If the operation $\circ$ is well defined we can find an easy formula for invariants of connected and disjoint sums of links. We can interpret $\circ$ as follows: if $w_1$ is the invariant of $L_+$ (Fig 1.1) and $w_2$ of $L_-$ then $w_1 \circ w_2$ is the invariant associated to $L_0$. \end{remark} \begin{remark}\textbf{1.16.} Our invariants often allow us to distinguish between a link and its mirror image. If $P_L (x, y, z)$ is an invariant of $L$ from Example 1.11 (a) and $\overline{L}$ is the mirror image of $L$ then \[P_{\overline{L}}(x,y,z) = P_L(y,x,z).\]
\end{remark} \indent We will call a crossing of the type \parbox{1.1cm}{\psfig{figure=PT-Lplus.eps,height=0.9cm}}
positive and crossing of the type
\parbox{1.1cm}{\psfig{figure=PT-Lmin.eps,height=0.9cm}} negative. This will be denoted by sgn $p= +$ or $-$. Let us consider now the following example. \begin{example}\textbf{1.17.} Let $L$ be the figure eight knot represented by the diagram \end{example} \centerline{\epsfig{figure=Figure13.eps,height=7.5cm}} \ \\ \centerline{\footnotesize{Fig. 1.3}} \ \\ To determine $w_L$ let us consider the following binary tree: \\ \centerline{\epsfig{figure=Fig14.eps,height=12cm}} \centerline{\footnotesize{Fig. 1.4}} As it is easily seen the leaves of the tree are trivial links and every branching reflects a certain operation on the diagram at the marked crossing point. To compute $w_L$ it is enough to have the following tree: \\ \ \\ \centerline{\psfig{figure=Skein-tree2.eps,height=3cm}}
Here the sign indicates the sign of the crossing point at which the operation was performed, and the leaf entries are the values of $w$ for the resulting trivial links. Now we may conclude that
\[w_L\!=\!a_1\!*\!(a_2|a_1).\] Such a binary tree of operations on the diagram resulting in trivial links at the leaves will be called the resolving tree of the diagram. \newline \indent There exists a standard procedure to obtain such a tree for every diagram. It will be described in the next paragraph and it will play an essential role in the proof of Theorem 1.8. It should be admitted that the idea is due to Ball and Metha {[$B$-$M$]} and we learned this from the Kauffman lecture notes {[$K$-3]}. \\
\section{Proof of the Main Theorem}\label{Section 2} \begin{definition}\textbf{2.1.} Let $L$ be an oriented diagram of $n$ components and let $b\!=$($b,\dots,b_n$) be base points of $L$, one point from each component of $L$, but not the crossing points. Then we say that $L$ is untangled with respect to $b$ if the following holds: if one travels along $L$ (according to the orientation of $L$) starting from $b_1$, then, after having returned to $b_1-$ from $b_2,\dots,$ finally from $b_n$, then each crossing which is met for the first time is crossed by a bridge.\end{definition} \indent It is easily seen that for every diagram $L$ of an oriented link there exists a resolving tree such that the leaf diagrams are untangled (with respect to appropriately chosen base points). This is obvious for diagrams with no crossings at all, and once it is known for diagrams with less than $n$ crossings we can use the following procedure for any diagram with $n$ crossings: choose base points arbitrarily and start walking along the diagram until the first ``bad" crossing $p$ is met, i.e. the first crossing which is crossed by a tunnel when first met. Then begin to construct the tree changing the diagram in this point. If, for example, sgn $p\!=\!+$ we get \\ \centerline{{\psfig{figure=Skein-tree.eps,height=2.1cm}}}\ \\
Then we can apply the inductive hypothesis to $L_0^p$ and we can continue the procedure with $L_-^p$ (walking further along the diagram and looking for the next bad point). \\
\\ \indent To prove Theorem 1.8 we will construct the function $w$ as defined on diagrams. In order to show that $w$ is an invariant of isotopy classes of oriented links we will verify that $w$ is preserved by the Reidemeister moves. \\
\\ \indent We use induction on the number $cr(L)$ of crossing points in the diagram. For each $k \geqslant 0$ we define a function $w_k$ assigning an element of $\A$ to each diagram of an oriented link with no more than $k$ crossings. Then $w$ will be defined for every diagram by $w_L = w_k(L)$ where $k \geqslant cr(L)$. Of course the functions $w_k$ must satisfy certain coherence conditions for this to work. Finally we will obtain the required properties of $w$ from the properties of $w_k$'s. \\ \indent We begin from the definition of $w_0$. For a diagram $L$ of $n$ components with $cr(L)=0$ we put \\
\begin{flalign*}
{2.2}&& w_0(L) &= a_n.& \end{flalign*} To define $w_{k+1}$ and prove its properties we will use the induction several times. To avoid misunderstandings the following will be called the ``Main Inductive Hypothesis": M.I.H. We assume that we have already defined a function $w_k$ attaching an element of $\A$ to each diagram $L$ for which $cr(L) \leqslant k$. We assume that $w_k$ has the following properties: \begin{flalign*} {2.3}&& w_k(U_n) &= a_n & \end{flalign*} for $U_n$ being an untangled diagram of $n$ components (with respect to some choice of base points). \begin{flalign*}
{2.4} && w_k(L_+) &= w_k(L_-)|w_k(L_0)&\\ {2.5} && w_k(L_-) &= w_k(L_+)*w_k(L_0) & \end{flalign*} for $L_+$, $L_-$ and $L_0$ being related as usually. \begin{flalign*} {2.6}&& w_k(L) &= w_k(R(L))& \end{flalign*} where $R$ is a Reidemeister move on $L$ such that $cr(R(L))$ is still at most $k$. \\ \indent Then, as the reader may expect, we want to make the Main Inductive Step to obtain the existence of a function $w_{k+1}$ with analogous properties defined on diagrams with at most $k$ +1 crossings. \\ \indent Before dealing with the task of making the M.I.S. let us explain that it will really end the proof of the theorem. It is clear that the function $w_k$ satisfying M.I.H. is uniquely determined by properties 2.3, 2.4, 2.5 and the fact that for every diagram there exists a resolving tree with untangled leaf diagrams. Thus the compatibility of the function $w_k$ is obvious and they define a function $w$ defined on diagrams. \\ \indent The function $w$ satisfies conditions (2) and (3) of the theorem because the function $w_k$ satisfy such conditions. \\ \indent If $R$ is a Reidemeister move on a diagram $L$, then $cr(R(L))$ equals at most $k=cr(L)+2$, whence \\ \indent $w_{R(L)}$=$w_k(R(L))$, $w_L$=$w_k(L)$ and by the properties of $w_k, w_k(L)$=$w_k(R(L))$ what implies $w_{R(L))}$=$w_L$. It follows that $w$ is an invariant of the isotopy class of oriented links. \\ \indent Now it is clear that $w$ has the required property (1) too, since there is an untangled diagram $U_n$ in the same isotopy class as $T_n$ and we have $w_k(U_n)=a_n$. \\ \indent The rest of this section will be occupied by the M.I.S. For a given diagram $D$ with $cr(D)\!\leqslant\!k+1$ we will denote by $\mathcal{D}$ the set of all diagrams which are obtained from $D$ by operations of the kind \parbox{3.1cm}{\psfig{figure=PTplustominus.eps,height=0.9cm}}
or \parbox{3.1cm}{\psfig{figure=PTplustozero.eps,height=0.9cm}}.
Of course, once base points $b$=($b_1,\dots, b_n$) are chosen on $D$, then the same points can be chosen as base points for any $L\!\in\!\mathcal{D}$, provided $L$ is obtained from $D$ by the operations of the first type only. \\ \indent Let us define a function $w_b$, for a given $D$ and $b$, assigning an element of $\A$ to each $L\!\in\!\mathcal{D}$. If $cr(L)\!<\!k+1$ we put \begin{flalign*} {2.7}&& w_b(L) &= w_k(L) & \end{flalign*} If $U_n$ is an untangled diagram with respect to $b$ we put \begin{flalign*} {2.8}&& w_b(U_n) &= a_{n} & \end{flalign*} ($n$ denotes the number of components). \\ Now we can proceed by induction on the number $b(L)$ of bad crossings in $L$ (in the symbol $b(L)\ b$ works simultaneously for ``bad" and for $b$=($b_1,\dots,b_n$). For a different choice of base points $b'$=($b'_1,\dots,b'_n$) we will write $b'(L))$. Assume that $w_b$ is defined for all $L\!\in\!\mathcal{D}$ such that $b(L)\!<\!t$. Then for $L$, $b(L)$=$t$, let $p$ be the first bad crossing for $L$(starting from $b_1$ and walking along the diagram). Depending on $p$ being positive or negative we have $L$=$L_+^p$ or $L$=$L_-^p$. We put \begin{flalign*} {2.9} & \ \ \ \ \ \ \ \ \ w_b(L)= \begin{cases}
w_b(L_-^p) | w_b(L_0^p), & \text{if sgn } p = + \\
w_b(L_+^p)*w_b(L_0^p), & \text{if sgn } p = -. \\ \end{cases}& \end{flalign*} We will show that $w_b$ is in fact independent of the choice of $b$ and that it has the properties required from $w_{k+1}$. \\
\\ \textbf{Conway Relations for $\boldsymbol{w_b}$} \\
\\ \indent Let us begin with the proof that $w_b$ has properties $2.4$ and $2.5$. We will denote by $p$ the considered crossing point. We restrict our attention to the case: $b(L_+^p)>b(L_-^p)$. The opposite situation is quite analogous. \\ \indent Now, we use induction on $b(L_-^p)$. If $b(L_-^p)$=$0$, then $b(L_+^p)$=$1$, $p$ is the only bad point of $L_+^p$, and by defining equalities $2.9$ we have
\[w_b(L_+^p)\!=\!w_b(L_-^p)|w_b(L_0^p)\] and using 1.6 we obtain \[w_b(L_-^p)\!=\!w_b(L_+^p)\!*\!w_b(L_0^p).\] Assume now that the formulae $2.4$ and $2.5$ for $w_b$ are satisfied for every diagram $L$ such that $b(L_-^p)\!<\!t$, $t\!\geqslant\!1$. Let us consider the case $b(L_-^p)$=$t$. \\ \indent By the assumption $b(L_+^p)\!\geqslant\!2$. Let $q$ be the first bad point on $L_+^p$. Assume that $q$=$p$. Then by $2.9$ we have
\[w_b(L_+^p)\!=\!w_b(L_-^p)| w_b(L_0^p).\] Assume that $q \neq p$. Let sgn $q=+$, for example. Then by 2.9 we have
\[w_b(L_+^p) = w_b(L{_+^p}{_+^q}) = w_b(L{_+^p}{_-^q}) | w_b(L{_+^p}{_ 0^q}).\] But $b(L{_-^p}{_-^q})\!<\!t$ and $cr(L{_+^p}{_ 0^q})\!\leqslant\!k$, whence by the inductive hypothesis and M.I.H. we have
\[w_b(L{_+^p}{_-^q})\!=\!w_b(L{_-^p}{_-^q})|w_b(L{_0^p}{_ -^q})\] and
\[w_b(L{_+^p}{_0^q})\!=\!w_b(L{_-^p}{_0^q})|w_b(L{_0^p}{_ 0^q})\] whence
\[w_b(L_+^p) = (w_b(L{_-^p}{_-^q}) | w_b(L{_0^p}{_-^q}))|( w_b(L{_-^p}{_ 0^q})| w_b(L{_0^p}{_0^q)})\] and by the transposition property 1.3 \begin{flalign*}
{2.10} && w_b(L_+^p) &=(w_b(L{_-^p}{_-^q})|w_b(L{_-^p}{_0^q})) |(w_b(L{_0^p}{_ -^q})|w_b(L{_0^p}{_0^q})).& \end{flalign*} On the other hand $b(L{_-^p}{_-^q})\!<\!t$ and $cr(L_ 0^p)\!\leqslant\!k$, so using once more the inductive hypothesis and M.I.H. we obtain \[ {2.11} \begin{split}
w_b(L_-^p) &= w_b(L{_-^p}{_+^q}) = w_b(L{_-^p}{_-^q}) | w_b(L{_-^p}{_ 0^q}) \\
w_b(L_0^p) &= w_b(L{_0^p}{_+^q}) = w_b(L{_0^p}{_-^q}) | w_b(L{_0^p}{_ 0^q}) \end{split} \] Putting 2.10 and 2.11 together we obtain
\[w_b(L_+^p) = w_b(L_-^p) | w_b(L_0^p)\] as required. If sgn $q=-$ we use $1.4$ instead of $1.3$. This completes the proof of Conway Relations for $w_b$. \\
\\ \textbf{Changing Base Points} \\
\\ \indent We will show now that $w_b$ does not depend on the choice of $b$, provided the order of components is not changed. It amounts to the verification that we may replace $b_i$ by $b'_i$ taken from the same component in such a way that $b'_i$ lies after $b_i$ and there is exactly one crossing point, say $p$, between $b_i$ and $b'_i$. Let $b'$=($b_1, \dots,b'_i,\dots,b_n$). We want to show that $w_b(L)\!=\!w_{b'}(L)$ for every diagram with $k+1$ crossings belonging to $\mathcal{D}$. We will only consider the case sgn $p=+$; the case sgn $p=-$ is quite analogous. \\ \indent We use induction on $B(L)=$max$(b(L),b'(L))$. We consider three cases. \\
\\ \indent \textsc{Cbp 1.} Assume $B(L)\!=\!0$. Then $L$ is untangled with respect to both choices of base points and by 2.8 \[w_b(L) = a_n = w_{b'}(L).\] \indent \textsc{Cbp 2.} Assume that $B(L)\!=\!1$ and $b(L)\!\neq\!b'(L)$. This is possible only when $p$ is a self-crossing point of the $i$-th component of $L$. There are two subcases to be considered. \\
\\ \indent \textsc{Cbp 2} (a): $b(L)\!=\!1$ and $b'(L)\!=\!0$. Then $L$ is untangled with respect to $b'$ and by 2.8 $$ w_{b'}(L)=a_n $$
$$ \ \ \ \ w_b(L)=w_b(L_+^p)= w_b(L_-^p)|w_b(L_0^p)$$ \indent Again we have restricted our attention to the case sgn $p\!=\!+$. Now, $w_b(L_-^p)\!=\!a_n$ since $b(L_-^p)\!=\!0$, and $L_0^p$ is untangled with respect to a proper choice of base points. Of course $L_0^p$ has $n+1$ components, so $w_b(L_0^p)\!=\!a_{n+1}$ by 2.8.
It follows that $w_b(L)\!=\!a_n|a_{n+1}$ and $a_n|a_{n+1}=a_n$ by $1.1$. \\
\\ \indent \textsc{Cbp 2}(b): $b(L)\!=\!0$ and $b'(L)\!=\!1$. This case can be dealt with like \textsc{Cbp 2}(a).\\
\\ \indent \textsc{Cbp 3.} $B(L)\!=\!t\!>\!1$ or $B(L)\!=\!1\!=\!b(L)\!=\!b'(L)$. We assume by induction $w_b(K)\!=\!w_{b'}(K)$ for $B(K)\!<\!B(L)$. Let $q$ be a crossing point which is bad with respect to $b$ and $b'$ as well. We will consider this time the case sgn $q=-$. The case sgn $q=+$ is analogous. \\ \indent Using the already proven Conway relations for $w_b$ and $w_{b'}$ we obtain \[w_b(L)= w_b(L_-^q)=w_b(L_+^q)\!*\!w_b(L_0^q) \] \[w_{b'}(L)= w_{b'}(L_-^q)= w_{b'}(L_+^q)\!*\!w_{b'}(L_0^q)\] But $B(L_+^q)\!<\!B(L)$ and $cr(L_0^q)\!\leqslant\!k$, whence by the inductive hypothesis and M.I.H. hold \[w_b(L_+^q) = w_{b'}(L_+^q)\] \[w_b(L_0^q) = w_{b'}(L_0^q)\] which imply $w_b(L)\!=\!w_{b'}(L)$. This completes the proof of this step (C.B.P.). \\ \indent Since $w_b$ turned out to be independent of base point changes which preserve the order of components we can now consider a function $w^0$ to be defined in such a way that it attaches an element of $\A$ to every diagram $L$, $cr(L) \leqslant k+1$ with a fixed ordering of components. \\
\\ \textbf{Independence of $\boldsymbol{w^0}$ of Reidemeister Moves} (I.R.M) \\
\\ \indent When $L$ is a diagram with fixed order of components and $R$ is a Reidemeister move on $L$, then we have a natural ordering of components on $R(L)$. We will show now that $w^0(L)=w^0(R(L))$. Of course we assume that $cr(L)$, $cr(R(L)) \leqslant k+1$. \\ \indent We use induction on $b(L)$ with respect to properly chosen base points $b=(b_1, \dots , b_n)$. Of course the choice must be compatible with the given ordering of components. We choose the base points to lie outside the part of the diagram involved in the considered Reidemeister move $R$, so that the same points may work for the diagram $R(L)$ as well. We have to consider the three standard types of Reidemeister moves (Fig. 2.1). \\ \ \\ \centerline{\psfig{figure=R123-PT.eps,height=2.5cm}}
\centerline{\footnotesize{Fig. 2.1}} \indent Assume that $b(L)=0$. Then it is easily seen that also $b(R(L))=0$, and the number of components is not changed. Thus \[w^0(L)=w^0(R(L))\ \textup{ by\ 2.8.}\] \indent We assume now by induction that $w^0(L)=w^0(R(L))$ for $b(L)<t$. Let us consider the case $b(L)=t$. Assume that there is a bad crossing $p$ in $L$ which is different from all the crossings involved in the considered Reidemeister move. Assume, for example, that sgn $p=+$. Then, by the inductive hypothesis, we have \begin{flalign*}
{2.12} && w^0(L_-^p) &= w^0(R(L_-^p))& \end{flalign*} and by M.I.H. \begin{flalign*}
{2.13} && w^0(L_0^p) &= w^0(R(L_0^p))& \end{flalign*} Now, by the Conway relation 2.4, which was already verified for $w^0$ we have
\[w^0(L)=w^0(L_+^p)=w^0(L_-^p)|w^0(L_0^p) \]
\[w^0(R(L))=w^0(R(L)_+^p)=w^0(R(L)_-^p)|w^0(R(L)_0^p) \] whence by 2.12 and 2.13 \[w^0(L)=w^0(R(L))\] Obviously $R(L_-^p)=R(L)_-^p$ and $R(L_0^p)=R(L)_0^p$. \\ \indent It remains to consider the case when $L$ has no bad points, except those involved in the considered Reidemeister move. We will consider the three types of moves separately. The most complicated is the case of a Reidemeister move of the third type. To deal with it let us formulate the following observation: \\ \indent Whatever the choice of base points is, the crossing point of the top arc and the bottom arc cannot be the only bad point of the diagram. \\
\centerline{\psfig{figure=TriangleB-PT.eps,height=4.1cm}} \centerline{\footnotesize{Fig. 2.2}} The proof of the above observation amounts to an easy case by case checking and we omit it. The observation makes possible the following induction: we can assume that we have a bad point at the crossing between the middle arc and the lower or the upper arc. Let us consider for example the first possibility; thus $p$ from Fig. 2.2 is assumed to be a bad point. We consider two subcases, according to sgn $p$ being $+$ or $-$. \\ \indent Assume sgn $p=+$. Then by Conway relations
\[w^0(L)=w^0(L_+^p)=w^0(L_-^p)|w^0(L_0^p)\]
\[w^0(R(L))=w^0(R(L)_+^p)=w^0(R(L)_-^p)|w^0(R(L)_0^p)\] But $R(L)_-^p=R(L_-^p)$ and by the inductive hypothesis \[w^0(L_-^p)=w^0(R(L_-^p))\] Also $R(L)_0^p$ is obtained from $L_0^p$ by two subsequent Reidemeister moves of type two (see Fig. 2.3), whence by M.I.H. \[w^0(R(L)_0^p)=w^0(L_0^p)\] and the equality \[w^0(L)=w^0(R(L))\ \textup{follows.}\]
\centerline{\epsfig{figure=Figure23.eps,height=3.5cm}} \centerline{\footnotesize{Fig.2.3}} Assume now that sgn $p=-$. Then by Conway relations \[w^0(L)=w^0(L_-^p)=w^0(L_+^p)*w^0(L_0^p) \] \[w^0(R(L))=w^0(R(L)_-^p)=w^0(R(L)_+^p)*w^0(R(L)_0^p)\] But $R(L)_+^p=R(L_+^p)$ and by the inductive hypothesis \[w^0(L_+^p)=w^0(R(L_+^p))\] Now, $L_0^p$ and $R(L)_0^p$ are essentially the same diagrams (see Fig. 2.4), whence $w^0(L_0^p)=w^0(R(L)_0^p)$ and the equality \[w^0(L)=w^0(R(L))\ \textup{follows.}\] \centerline{\epsfig{figure=Fig24.eps, height=3.5cm}} \centerline{\footnotesize{Fig. 2.4}} \ \\ \ \\ Reidemeister moves of the first type. The base points can always be chosen so that the crossing point involved in the move is good. \\ Reidemeister moves of the second type. There is only one case, when we cannot choose base points to guarantee the points involved in the move to be good. It happens when the involved arcs are parts of different components and the lower arc is a part of the earlier component. In this case the both crossing points involved are bad and they are of different signs, of course. Let us consider the situation shown in Fig. 2.5. \\ \centerline{\epsfig{figure=Fig25.eps, height=3.5cm}} \centerline{\footnotesize{Fig. 2.5}} We want to show that $w^0(R(L))=w^0(L)$. But by the inductive hypothesis we have \[w^0(L')=w^0(R'(L'))=w^0(R(L)).\] Using the already proven Conway relations, formulae 1.6 and 1.7 and M.I.H. if necessary, it can be proved that $w^0(L)=w^0(L')$. Let us discuss in detail the case involving M.I.H. It occurs when sgn $p=-$.
Then we have \[w^0(L)=w^0(L_+^q)=w^0(L_-^q)|w^0(L_0^q)=(w^0(L{_-^q}{_+^p})*w^0(L{_-^q}{_0^p}))|w^0(L_0^q)\] But $L{_-^q}{_+^p}= L'$ and by M.I.H. $w^0(L{_-^q}{_0^p})=w^0(L_0^q)$ (see Fig. 2.6, where $L{_-^q}{_0^p}$ and $L_0^q$ are both obtained from $K$ by a Reidemeister move of the first type). \\ \ \\ \centerline{\epsfig{figure=Fig26.eps, height=3.2cm}} \centerline{\footnotesize{Fig. 2.6}} Thus by 1.7: \[w^0(L) = w^0(L')\ \textup{whence}\] \[w^0(L)=w^0(R(L)). \] The case: sgn $p=+$ is even simpler and we omit it. This completes the proof of the independence of $w^0$ of Reidemeister moves. To complete the Main Inductive Step it is enough to prove the independence of $w^0$ of the order of components. Then we set $w_k = w^0$. The required properties have been already checked. \\
\\ \textbf {Independence of the Order of Components} (I.O.C.) \\
\\ \indent It is enough to verify that for a given diagram $L\ (cr(L) \leqslant k+1)$ and fixed base points $b=(b_1, \dots , b_i, b_{i+1}, \dots , b_n)$ we have \[w_b(L)=w_{b'}(L)\] where $b'=(b_1, \dots , b_{i+1}, b_{i}, \dots , b_n)$. This is easily reduced by the usual induction on $b(L)$ to the case of an untangled diagram. To deal with this case we will choose $b$ in an appropriate way. \\ \indent Before we do it, let us formulate the following observation: If $L_i$ is a trivial component of $L$, i.e. $L_i$ has no crossing points, neither with itself, nor with other components, then the specific position of $L_i$ in the plane has no effect on $w^0(L)$; in particular we may assume that $L_i$ lies separately from the rest of the diagram: \\ \ \\ \centerline{\epsfig{figure=Fig27.eps, height=2.2cm}} \centerline{\footnotesize{Fig. 2.7}} This can be easily achieved by induction on $b(L)$, or better by saying that it is obvious. \\ \indent For an untangled diagram we will be done if we show that it can be transformed into another one with less crossings by a series of Reidemeister moves which do not increase the crossing number. We can then use I.R.M. and M.I.H. This is guaranteed by the following lemma. \begin{lemma}\textbf{2.14.} Let $L$ be a diagram with $k$ crossings and a given ordering of components $L_1, L_2, \dots , L_n$. Then either $L$ has a trivial circle as a component or there is a choice of base points $b=(b_1, \dots , b_n)$; $b_i \in L_i$ such that an untangled diagram $L^{u}$ associated with $L$ and $b$ (that is all the bad crossings of $L$ are changed to good ones) can be changed into a diagram with less than $k$ crossings by a sequence of Reidemeister moves not increasing the number of crossings. \end{lemma} This was probably known to Reidemeister already, however we prove it in the Appendix for the sake of completeness.\\ \indent With I.O.C. proven we have completed M.I.S. and the proof of Theorem 1.8.\\ \section{Quasi-algebras}\label{Section 3} \indent We shall now describe a certain generalization of Theorem 1.8. This is based on the observation that it was not necessary
to have the operations $|$ and $*$ defined on the whole product $\A \times \A$. Let us begin with the following definition. \begin{definition}\textbf{3.0.} A quasi Conway algebra is a triple $(\A, B_1, B_2), B_1, B_2$ being subsets of $\A \times \A$, together with 0-argument
operations $a_1,a_2, \dots ,a_n, \dots$ and two 2-argument operations $|$ and $*$ defined on $B_1$ and $B_2$ respectively satisfying the conditions: \end{definition} \begin{align*} \left. \begin{aligned}
3.1 && \qquad &a_n | a_{n+1} = a_n \\ 3.2 && \qquad &a_n * a_{n+1} = a_n \\
3.3&& \qquad &(a|b)|(c|d) = (a|c)|(b|d) \\
3.4 && \qquad &(a|b)*(c|d) = (a*c)|(b*d) \\ 3.5 && \qquad &(a*b)*(c*d) = (a*c)*(b*d) \\
3.6 && \qquad &(a|b)*b = a \\
3.7 && \qquad &(a*b)|b = a. \\ \end{aligned} \right\} \indent \text{whenever the both sides are defined} \end{align*} \indent We would like to construct invariants of Conway type using such quasi-algebras. As before $a_n$ will be the value of the invariant for the trivial link of $n$ components. \\ \indent We say that $\A$ is geometrically sufficient if and only if for every resolving tree of each diagram of an oriented link all the operations that are necessary to compute the root value are defined. \begin{theorem}\textbf{3.8.} Let $\A$ be a geometrically sufficient quasi Conway algebra. There exists a unique invariant $w$ attaching to each isotopy class of links an element of $\A$ and satisfying the conditions \end{theorem} \begin{enumerate} \item $w_{T_n}=a_n$ for $T_n$ being a trivial link of n components,\\ \item if $L_+$, $L_-$ and $L_0$ are diagrams from Fig. 1.1, then
\[w_{L_+}=w_{L_-}|w_{L_0} \text{ and } \] \[w_{L_-}=w_{L_+}*w_{L_0}.\] \end{enumerate} \noindent The proof is identical with the proof of Theorem 1.8.\\ \indent As an example we will now describe an invariant, whose values are polynomials in an infinite number of variables. \begin{example}\textbf{3.9.} $\A=N \times Z[ y_1^{\mp1},{x'}_2^{\mp 1},{z'}_2,\ x_1^{\mp1},z_1,\ x_2^{\mp1},z_2,x_3^{\mp1},z_3,\dots],
B_1=B_2=B=\{ ((n_1,\ w_1),(n_2,w_2))\in\A\times\A:|n_1-n_2|=1\}, a_1=(1,1),a_2=(2, x_1+y_1+z_1),\dots, a_n=(n,\Pi_{i=1}^{n-1}(x_i+y_i)+z_1\Pi_{i=2}^{n-1} (x_i+y_i)+\dots+z_{n-2}(x_{n-1}+y_{n-1})+z_{n-1}),\dots$
where $y_i=x_i\frac{y_1}{x_1}$. To define the operations $|$ and $*$ consider the following system of equations:\end{example} \begin{flalign*} {(1)} && &x_1w_1+y_1w_2=w_0-z_1& \\ {(2)} && &x_2w_1+y_2w_2=w_0-z_2& \\ {(2^\prime)}&& &x'_2w_1+y'_2w_2=w_0-z'_2& \\ {(3)} && &x_3w_1+y_3w_2=w_0 -z_3 \\ {(3^\prime)}&& &x'_3w_1+y'_3w_2=w_0-z'_3 \\ && &\dots \\ {(i)}&& &x_iw_1+y_iw_2=w_0-z_i \\ {\text({i}^\prime})&& &x'_iw_1+y'_iw_2=w_0-z'_i \\ && &\ldots \end{flalign*} where $y'_i=\frac{x'_iy_1}{x_i},x'_i=\frac{x'_2x_1}{x_{i-1}}$ and $z'_i$ are defined inductively to satisfy \[\frac{z'_{i+1}-z_{i-1}}{x_1x'_2}= \Big(1+\frac{y_1}{x_1}\Big)\Big(\frac{z'_i}{x'_i}-\frac{z_i}{x_i}\Big).\]
We define $(n,w)=(n_1,w_1)|(n_2,w_2)$ (resp.$(n,w)=(n_1,w_1)*(n_2, w_2))$ as follows: $n=n_1$ and if $n_1=n_2-1$ then we use the equations $(n)$ to get $w$; namely $x_nw+y_nw_1=w_2-z_n$ (resp. $x_nw_1+y_nw=w_2-z_n$). If $n_1=n_2+1$ then we use the equation ($n'$) to get $w$; namely $x'_nw+y'_nw_1=w_2-z'_n$ (resp. $x'_nw_1+y'_nw=w_2-z'_n$). We can think of Example 1.11 as being a special case of Example 3.9. \\ \indent Now we will show that the quasi-algebra $\A$ for Example 3.9 satisfies the relations 1.1-1.7.\\ \indent It is an easy task to check that the first coordinate of elements from $\A$ satisfies the relations 1.1-1.7 (compare with Example 1.9) and to check the relations 1.1, 1.2, 1.6 and 1.7 so we will concentrate our attention on relations 1.3, 1.4, and 1.5.\\ \indent It is convenient to use the following notation: if $w\in\A$ then $w=(\lvert{w}\rvert,F)$ and for
\[w_1|w_2=(\lvert{w_1}\rvert,F_1)|(\lvert{w_2}\rvert, F_2)=(\lvert{w}\rvert,F)=w \] to use the notation \begin{equation*} F=\begin{cases}
F_1|_nF_2 \ \textup{if} \ n=\lvert{w_1}\rvert=\lvert{w_2}\rvert-1 \\
F_1|_{n'}F_2 \ \textup{if} \ n=\lvert{w_1}\rvert=\lvert{w_2}\rvert+1. \end{cases} \end{equation*} Similar notation we use for the operation $*$. \\
\indent In order to verify relations 1.3-1.5 we have to consider three main cases:\\ $1. \quad \lvert{a}\rvert=\lvert{c}\rvert-1=\lvert{b}\rvert+1=n$ \\ Relations 1.3-1.5 make sense iff $\lvert{d}\rvert=n$. The relation 1.3 has the form:
\[(F_a|_{n'}F_b)|_n(F_c|_{(n+1)'}F_d)=(F_a|_{n}F_c)|_{n'}(F_b|_{(n-1)}F_d).\] From this we get: \[\frac{1}{x_nx'_{n+1}}F_d-\frac{y'_{n+1}}{x_nx'_{n+1}}F_c-\frac{y_n}{x_nx'_n}F_b + \frac{y_ny'_n}{x_nx'_n} F_a-\frac{z'_{n+1}}{x_nx'_{n+1}}-\frac{z_n}{x_n}+ \frac{y_nz'_n}{x_nx'_n}= \] \[= \frac{1}{x'_nx_{n-1}} F_d-\frac{y_{n-1}}{x'_nx_{n-1}}F_b-\frac{y'_n}{x_nx'_n}F_c + \frac{y_ny'_n}{x_nx'_n}F_a-\frac{z_{n-1}}{x'_nx_{n-1}}-\frac{z'_n}{x'_n}+ \frac{y'_nz_n}{x_nx'_n}\] Therefore: \begin{align*} \text{(i)}&& &x_{n-1}x'_n=x_nx'_{n+1}\\ \text{(ii)}&& &\frac{y'_{n+1}}{x'_{n+1}}=\frac{y'_n}{x'_n} \\ \text{(iii)}&& &\frac{y_n}{x_n}=\frac{y_{n-1}}{x_{n-1}} \\ \text{(iv)}&& &\frac{z'_{n+1}}{x_nx'_{n+1}}+\frac{z_n}{x_n}-\frac{y_nz'_n}{x_nx'_n}=\frac{z_{n-1}}{x'_nx_{n-1}}+\frac{z'_n}{x'_n}-\frac{y'_nz_n}{x_nx'_n} \end{align*} When checking the relations 1.4 and 1.5 we get exactly the same conditions (i)-(iv).\\ $2. \quad \lvert a \rvert =\lvert b \rvert -1 = \lvert{c} \rvert -1 = n$.\\ \indent (I) $\lvert{d}\rvert=n.$\\ The relation 1.3 has the following form:
\[(F_a|_{n}F_b)|_n(F_c|_{(n+1)'}F_d)=(F_a|_{n}F_c)|_{n}(F_b|_{(n+1)'}F_d).\] We get after some calculations that it is equivalent to \begin{align*} \text{(v)}&& &\frac{y_n}{x_n}=\frac{y'_{n+1}}{x'_{n+1}}& \end{align*} The relations 1.4 and 1.5 reduce to the same condition (v). \\ \indent (II) $\lvert{d}\rvert=n+2.$ \\ Then the relations 1.3-1.5 reduce to the condition (iii). \\ $3. \quad \lvert{a}\rvert=\lvert{b}\rvert+1=\lvert{c}\rvert+1=n$ \\ \indent (I) $\lvert{d}\rvert = n-2$ \\ \indent (II) $\lvert{d}\rvert = n.$ \\ We get, after some computations, that relations 3 (I) and 3 (II) follow from the conditions (iii) and (v). \\ \indent Conditions (i)-(v) are equivalent to the conditions on $x'_i, y_i, y'_i$ and $z'_i$ described in Example 3.9. Therefore the quasi-algebra $\A$ from Example 3.9 satisfies the relations 1.1-1.7. Furthermore, if $L$ is a diagram and $p-$ its crossing, then the number of components of $L_0^p$ is always equal to the number of components of $L$ plus or minus one, so the set $B\subset\A\times\A$ is sufficient to define the link invariant associated with $\A$.\\
\section{Final remarks and problems}\label{Section 4} \begin{remark}\textbf{4.1.} Each invariant of links can be used to build a better invariant which will be called weighted simplex of the invariant. Namely, if $w$ is an invariant and $L$ is a link of $n$ components $L_1,\dots, L_n$ then we consider an $n-1$ dimensional simplex $\Delta^{n-1}=(q_1,\dots,q_n)$. We associate with each face ($q_{i_1},\dots,q_{i_k}$) of $\Delta^{n-1}$ the value $w_{L'}$ where $L' = L_{i_1}\cup \cdots \cup L_{i_k}$. \end{remark} \indent We say that two weighted simplices are equivalent if there exists a bijection of their vertices which preserves weights of faces. Of course, the weighted simplex of an invariant of isotopy classes of oriented links is also an invariant of isotopy classes of oriented links. \\ \indent Before we present some examples, we will introduce an equivalence relation $\thicksim_c$ (Conway equivalence relation) on isotopy classes of oriented links ($\mathcal{L}$). \begin{definition}\textbf{4.2.} $\thicksim_c$ is the smallest equivalence relation on $\mathcal{L}$ which satisfies the following condition: let $L'_1$ (resp. $L'_2$) be a diagram of a link $L_1$ (resp. $L_2$) with a given crossing $p_1$ (resp. $p_2$) such that $p_1$ and $p_2$ are crossings of the same sign and \[ \begin{split}
(L'_1)_-^{p_1} \thicksim_c (L'_2)_-^{p_2}\ &\text{and} \\
(L'_1)_0^{p_1} \thicksim_c (L'_2)_0^{p_2} \end{split} \] then $L_1\thicksim_c L_2$.\\ \end{definition} \indent It is obvious that an invariant given by a quasi Conway algebra is a Conway equivalence invariant. \begin{example}\textbf{4.3.}{(a)} Two links shown on Fig. 4.1 are Conway equivalent but they can be distinguished by weighted simplices of the linking numbers. \end{example} \centerline{\epsfig{figure= Fig41.eps,height=3.5cm}} \centerline{\footnotesize{Fig. 4.1}} {(b)} J. Birman has found three-braids (we use the notation of [M]); \[ \begin{split}
\gamma_1=\sigma_1^{-2}\sigma_2^3\sigma_1^{-1}\sigma_2^4\sigma_1^{-2}\sigma_2^{4}\sigma_1^{-1}\sigma_2 \\
\gamma_2=\sigma_1^{-2}\sigma_2^3\sigma_1^{-1}\sigma_2^4\sigma_1^{-1}\sigma_2\sigma_1^{-2}\sigma_2^4 \end{split} \] which closures have the same values of all invariants described in our examples and the same signature but which can be distinguished by weighted simplices of the linking numbers [B]. \\ \indent As the referee has kindly pointed out the polynomial invariants described in 1.11 (a) and 1.11 (b) are equivalent. Namely if we denote them by $w_L$ and $w'_L$ respectively, then we have \[w_L(x,\ y,\ z)=\Big(1-\frac{z}{1-x-y}\Big)w'_L(x,\ y)+\frac{z}{1-x-y}.\] \begin{problem}\textbf{4.4.} \begin{enumerate}
\item[(a)]
Is the invariant described in Example 3.9 better than the polynomial invariant from Example 1.11?\footnote{Added for e-print: Adam Sikora proved in his Warsaw master degree thesis written under direction of P.Traczyk, that the answer to Problem 4.4 (a) is negative, \cite{Si-1}.}
\item[(b)] Find an example of links which have the same polynomial invariant of Example 3.9 but which can be distinguished by some invariant given by a Conway algebra.\footnote{Added for e-print: Adam Sikora proved no invariant coming from a Conway algebra can distinguish links with the same
polynomial invariant of Example 1.11, \cite{Si-2}.}
\item[(c)] Do there exist two links $L_1$ and $L_2$ which are not Conway equivalent but which cannot be distinguished using any Conway algebra?
\item[(d)] Birman [$B$] described two closed 3-braid knots given by \[ \begin{split}
y_1=\sigma_1^{-3}\sigma_2^4\sigma_1^{-1}\sigma_2^5\sigma_1^{-3}\sigma_2^{5}\sigma_1^{-2}\sigma_2 \\
y_2=\sigma_1^{-3}\sigma_2^4\sigma_1^{-1}\sigma_2^5\sigma_1^{-2}\sigma_2\sigma_1^{-3}\sigma_2^5 \end{split} \] which are not distinguished by the invariants described in our examples and by the signature. Are they Conway equivalent? (they are not isotopic because their incompressible surfaces are different). \end{enumerate} \end{problem}
\begin{problem}\textbf{4.5.} Given two Conway equivalent links, do they necessarily have the same signature? \end{problem} \indent The examples of Birman and Lozano [$B$; Prop. 1 and 2] have different signature but the same polynomial invariant of Example 1.11 (b)(see [$B$]). \begin{problem}\textbf{4.6.} Let ($V_1$, $V_2$) be a Heegaard splitting of a closed 3-manifold $M$. Is it possible to modify the above approach using projections of links onto the Heegaard surface $\partial V_1$? \end{problem} \indent We have obtained the results of this paper in early December 1984 and we were not aware at the time that an important part of our results (the invariant described in Example 1.11 (b)) had been got three months before us by four groups of researchers: R. Lickorish and K. Millett, J. Hoste, A. Ocneanu, P. Freyd and D. Yetter and that the first two groups used arguments similar to ours. We have been informed about this by J. Birman (letter received on January 28, '85) and by J. Montesinos (letter received on February 11, '85; it included the paper by R. Lickorish and K. Millett and also a small joint paper by all above mentioned mathematicians). \\
\section{Appendix} Here we prove Lemma 2.14.\\ \indent A closed part cut out of the plane by arcs of $L$ is called an $i$-gon if it has $i$ vertices (see Fig. 5.1). \\
\centerline{{\psfig{figure=PT-2-2-9.eps,height=4.0cm}}}\ \\ \centerline{\footnotesize{Fig. 5.1}} Every $i$-gon with $i\leqslant2$ will be called $f$-gon ($f$ works for few). Now let $X$ be an innermost $f$-gon that is an $f$-gon which does not contain any other $f$-gon inside. \\ \indent If $X$ is 0-gon we are done because $\partial X$ is a trivial circle. If $X$ is 1-gon then we are done because int $X \cap L = \emptyset$ so we can perform on $L^u$ a Reidemeister move which decreases the number of crossings of $L^u$ (Fig. 5.2). \\ \centerline{{\psfig{figure=PT-2-2-10.eps,height=3.5cm}}}\ \\
\centerline{\footnotesize{Fig. 5.2}}\\ Therefore we assume that $X$ is a 2-gon. Each arc which cuts int $X$ goes from one edge to another. Furthermore, no component of $L$ lies fully in $X$ so we can choose base points $b=(b_1,\dots,b_n)$ lying outside $X$. This has important consequences: if $L^u$ is an untangled diagram associated with $L$ and $b$ then each 3-gon in $X$ supports a Reidemeister move of the third type (i.e. the situation of the Fig. 5.3 is impossible).\\
\centerline{{\psfig{figure=PT-2-2-11.eps,height=4.5cm}}}\ \\ \centerline{\footnotesize{Fig. 5.3}} \indent Now we will prove Lemma 2.14 by induction on the number of crossings of $L$ contained in the 2-gon $X$ (we denote this number by $c$). \\ \indent If $c=2$ then int $X \cap L=\emptyset$ and we are done by the previous remark (2-gon $X$ can be used to make the Reidemeister move of the second type on $L^u$ and to reduce the number of crossings in $L^u$). \\ \indent Assume that $L$ has $c>2$ crossings in $X$ and that Lemma 2.14 is proved for less than $c$ crossings in $X$. In order to make the inductive step we need the following fact. \begin{proposition}\textbf{5.1.} If $X$ is an innermost 2-gon with int $X\cap L \neq \emptyset$ then there is a 3-gon $\Delta \subset X$ such that $\Delta \cap \partial X \neq \emptyset$, int $\Delta \cap L = \emptyset$. \end{proposition} Before we prove Proposition 5.1 we will show how Lemma 2.14 follows from it. \\ \indent We can perform the Reidemeister move of the third type using the 3-gon $\Delta$ and reduce the number of crossings of $L^u$ in $X$ (compare Fig. 5.4).\\ \centerline{\epsfig{figure= Fig54.eps, height=3.5cm}}
\centerline{\footnotesize{Fig. 5.4}}\\ Now either $X$ is an innermost $f$-gon with less than $c$ crossings in $X$ or it contains an innermost $f$-gon with less that $c$ crossings in it. In both cases we can use the inductive hypothesis. \\ \indent Instead of proving Proposition 5.1 we will show a more general fact, which has Proposition 5.1 as a special case. \begin{proposition}\textbf{5.2.} Consider a 3-gon $Y=$($a$, $b$, $c$) such that each arc which cuts it goes from the edge $\overline{ab}$ to the edge $\overline{ac}$ without self-intersections (we allow $Y$ to be a 2-gon considered as a degenerated 3-gon with $\overline{bc}$ collapsed to a point). Furthermore let int $Y$ be cut by some arc. Then there is a 3-gon $\Delta \subset Y$ such that $\Delta \cap \overline{ab} \neq\emptyset$ and int $\Delta$ is not cut by any arc. \end{proposition} \indent \textsc{Proof of Proposition 5.2:} We proceed by induction on the number of arcs in int $Y\cap L$ (each such an arc cuts $\overline{ab}$ and $\overline{ac}$). For one arc it is obvious (Fig. 5.5). Assume it is true for $k$ arcs ($k\geqslant 1$) and consider ($k+1)$-th arc $\gamma$. Let $\Delta_0=$($a_1$, $b_1$, $c_1$) be a $3$-gon from the inductive hypothesis with an edge $\overline{a_1b_1}\subset \overline{ab}$ (Fig. 5.6). \\
\centerline{{\psfig{figure=PT-2-2-13.eps,height=5.2cm}}}\ \\ \centerline{\footnotesize{Fig. 5.5}}\\ If $\gamma$ does not cut $\Delta_0$ or it cuts $\overline{a_1b_1}$ we are done (Fig 5.6). Therefore let us assume that $\gamma$ cuts $\overline{a_1c_1}$ (in $u_1$) and $\overline{b_1c_1}$ (in $w_1$). Let $\gamma$ cut $\overline{ab}$ in $u$ and $\overline{ac}$ in $w$ (Fig. 5.7). We have to consider two cases: \\ \indent (a) $\quad \overline{uu_1}\ \cap$ int $\Delta_0=\emptyset$ (so $\overline{ww_1}\ \cap$ int $\Delta_0=\emptyset$); Fig. 5.7.\\ \centerline{\epsfig{figure=Fig56.eps, height=3.3in}}\\
\centerline{\footnotesize{Fig. 5.6}}\\ \ \\
\centerline{{\psfig{figure=PT-2-2-15.eps,height=5.9cm}}}\ \\ \centerline{\footnotesize{Fig 5.7}} Consider the 3-gon $ua_1u_1$. No arc can cut the edge $\overline{a_1u_1}$ so each arc which cuts the 3-gon $ua_1u_1$ cuts the edges $\overline{ua_1}$ and $\overline{uu_1}$. Furthermore this 3-gon is cut by less than $k+1$ arcs so by the inductive hypothesis there is a 3-gon $\Delta$ in $ua_1u_1$ with an edge on $\overline{ua_1}$ the interior of which is not cut by any arc. Then $\Delta$ satisfies the thesis of Proposition 5.2.\\ \indent (b) $\quad \overline{uw_1} \ \cap$ int$\Delta_0=\emptyset$ (so $\overline{wu_1}\ \cap$ int$\Delta_0 =\emptyset$). In this case we proceed like in case (a). \\ This completes the proof of Proposition 5.2 and hence the proof of Lemma 2.14.
\ \\ Department of Mathematics\\ Warsaw University\\ 00-901 Warszawa, Poland\\ \ \\ \\ Added for e-print:\\ New address of J.~H.~Przytycki:\\ Department of Mathematics\\ The George Washington University\\ Washington, DC\\ {\tt [email protected]}\\ and University of Gda\'nsk
\end{document}
|
arXiv
|
{
"id": "1610.06679.tex",
"language_detection_score": 0.7561485767364502,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Some Identities Involving Three Kinds of Counting Numbers}
\begin{center} L. C. Hsu
Mathematics Institute, Dalian University of Technology, Dalian 116024, China\\[5pt]
\end{center}\vskip0.5cm
\subsection*{Abstract} In this note, we present several identities involving binomial coefficients and the two kind of Stirling numbers.
\vskip0.5cm
{\large\bf 1. Introduction}
Adopting Knuth's notation, let us denote by $\left[\begin{array}{c}n\\k\end{array}\right]$ and $\left\{\begin{array}{c}n\\k\end{array}\right\}$ the unsigned (absolute) Stirling number of the first kind and the ordinary Stirling number of the second kind, respectively. Here in particular, $\left[\begin{array}{c}0\\0\end{array}\right]=\left\{\begin{array}{c}0\\0\end{array}\right\}=1$, $\left[\begin{array}{c}n\\0\end{array}\right]=\left\{\begin{array}{c}n\\0\end{array}\right\}=0~(n>0)$, and $\left[\begin{array}{c}n\\k\end{array}\right]=\left\{\begin{array}{c}n\\k\end{array}\right\}=0~(0\le n<k)$. Generally, $\left[\begin{array}{c}n\\k\end{array}\right]$, $\left\{\begin{array}{c}n\\k\end{array}\right\}$ and the binomial coefficients $\left(\begin{array}{c}n\\k\end{array}\right)$ may be regarded as the most important counting numbers in combinatorics. The object of this short note is to propose some combinatorial identities each consisting of these three kinds of counting numbers, namely the following $$ \sum_k\left[\begin{array}{c}k\\p\end{array}\right]\left\{\begin{array}{c}n+1\\k+1\end{array}\right\}(-1)^k= \left(\begin{array}{c}n\\p\end{array}\right)(-1)^p $$ $$ \sum_k\left[\begin{array}{c}k+1\\p+1\end{array}\right]\left\{\begin{array}{c}n\\k\end{array}\right\}(-1)^k= \left(\begin{array}{c}n\\p\end{array}\right)(-1)^n $$ $$ \sum_{j,k}\left[\begin{array}{c}n\\k\end{array}\right]\left\{\begin{array}{c}k\\j\end{array}\right\} \left(\begin{array}{c}n\\j\end{array}\right)(-1)^k=(-1)^n $$ $$ \sum_{j,k}\left\{\begin{array}{c}n\\k\end{array}\right\}\left[\begin{array}{c}k\\j\end{array}\right] \left(\begin{array}{c}n\\j\end{array}\right)(-1)^k=(-1)^n $$ $$ \sum_{j,k}\left(\begin{array}{c}n\\k\end{array}\right)\left\{\begin{array}{c}k\\j\end{array}\right\} \left[\begin{array}{c}j+1\\p\end{array}\right](-1)^j=\left\{\begin{array}{cl}0, &(n+1>p)\\[6pt] (-1)^n, &(n+1=p)\end{array}\right. $$ $$ \sum_{j,k}\left[\begin{array}{c}n\\k\end{array}\right]\left(\begin{array}{c}k\\j\end{array}\right) \left\{\begin{array}{c}j+1\\p\end{array}\right\}(-1)^j=\left\{\begin{array}{cl}0, &(n+1>p)\\[6pt] (-1)^n, &(n+1=p)\end{array}\right. $$ Here each of the summations in (1) and (2) extends over all $k$ such that $0\le k\le n$ or $p\le k\le n$, and all the double summations within (3)--(6) are taken over all possible integers $j$ and $k$ such that $0\le j\le k\le n$.
Note that (1) is a well-known identity that has appeared in the Table 6.4 of Graham-Knuth-Patashnik's book [1] (cf. formula (6.24)). It is quite believable that (1) and (2) may be the most simple identities each connecting with the three kinds of counting numbers. \\[10pt] {\large\bf 2. Proof of the identities}
In order to verify (2)--(6), let us recall that the orthogonality relations $$ \sum_{k}\left[\begin{array}{c}n\\k\end{array}\right]\left\{\begin{array}{c}k\\p\end{array}\right\} (-1)^{n-k}=\sum_k\left\{\begin{array}{c}n\\k\end{array}\right\}\left[\begin{array}{c}k\\p\end{array}\right] (-1)^{n-k}=\delta_{np}=\left\{\begin{array}{ll}0, &(n\neq p)\\[8pt] 1, &(n=p)\end{array}\right. $$ are equivalent to the inverse relations $$ a_n=\sum_{k}\left[\begin{array}{c}n\\k\end{array}\right](-1)^{n-k}b_k \Leftrightarrow b_n=\sum_k\left\{\begin{array}{c}n\\k\end{array}\right\}a_k. $$ Also, we shall make use of two known identities displayed in the Table 6.4 of [1], viz. $$ \left\{\begin{array}{c}n+1\\p+1\end{array}\right\}=\sum_k\left(\begin{array}{c}n\\k\end{array}\right) \left\{\begin{array}{c}k\\p\end{array}\right\} $$ $$ \left[\begin{array}{c}n+1\\p+1\end{array}\right]=\sum_k\left[\begin{array}{c}n\\k\end{array}\right] \left(\begin{array}{c}k\\p\end{array}\right). $$ Now, take $a_n=(-1)^n\left[\begin{array}{c}n+1\\p+1\end{array}\right]$ and $b_k=(-1)^k\left(\begin{array}{c}k\\p\end{array}\right)$, so that (10) can be embedded in the first equation of (8). Thus it is seen that (10) can be inverted via (8) to get the identity (2).
(3) and (4) are trivial consequences of (7). Indeed rewriting (7) in the form
~~~~~~~~~~~~~~~~~$\displaystyle \sum_k\left[\begin{array}{c}n\\k\end{array}\right]\left\{\begin{array}{c}k\\j\end{array}\right\}(-1)^k= \sum_k\left\{\begin{array}{c}n\\k\end{array}\right\}\left[\begin{array}{c}k\\j\end{array}\right](-1)^k= (-1)^n\delta_{nj}$
$(7)'$ \\[8pt] and noticing that $\sum_j\left(\begin{array}{c}n\\j\end{array}\right)\delta_{nj}= \left(\begin{array}{c}n\\n\end{array}\right)=1$, we see that (3)--(4) follow at once from $(7)'$.
For proving (5), let us make use of (9) and (7) with $p$ being replaced by $j$. We find $$ \sum_j\left\{\begin{array}{c}n+1\\j+1\end{array}\right\}\left[\begin{array}{c}j+1\\p\end{array}\right](-1)^j= \sum_j\sum_k\left(\begin{array}{c}n\\k\end{array}\right)\left\{\begin{array}{c}k\\j\end{array}\right\} \left[\begin{array}{c}j+1\\p\end{array}\right](-1)^j=(-1)^n\delta_{n+1,p}. $$ Hence (5) is obtained.
Similarly, (6) is easily derived from (10) and (7). \\[10pt] {\large\bf 3. Questions}
It may be a question of certain interest to ask whether some of the identities (1)--(6) could be given some combinatorial interpretations with the aid of the inclusion-exclusion principle or the method of bijections. Also, we have not yet decided whether (1)--(6) could be proved by the method of generating functions (cf. [2]).
{\bf AMS Classification Numbers}: 05A10, 05A15, 05A19.
\end{document}
|
arXiv
|
{
"id": "0911.0279.tex",
"language_detection_score": 0.6556994318962097,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\def
{}
\title{Understanding Trainable Sparse Coding\via matrix factorization}
\begin{abstract} Sparse coding is a core building block in many data analysis and machine learning pipelines. Typically it is solved by relying on generic optimization techniques, such as the Iterative Soft Thresholding Algorithm and its accelerated version (ISTA, FISTA). These methods are optimal in the class of first-order methods for non-smooth, convex functions. However, they do not exploit the particular structure of the problem at hand nor the input data distribution. An acceleration using neural networks, coined LISTA, was proposed in \cite{Gregor10}, which showed empirically that one could achieve high quality estimates with few iterations by modifying the parameters of the proximal splitting appropriately.
In this paper we study the reasons for such acceleration. Our mathematical analysis reveals that it is related to a specific matrix factorization of the Gram kernel of the dictionary, which attempts to nearly diagonalise the kernel with a basis that produces a small perturbation of the $\ell_1$ ball. When this factorization succeeds, we prove that the resulting splitting algorithm enjoys an improved convergence bound with respect to the non-adaptive version. Moreover, our analysis also shows that conditions for acceleration occur mostly at the beginning of the iterative process, consistent with numerical experiments. We further validate our analysis by showing that on dictionaries where this factorization does not exist, adaptive acceleration fails.
\end{abstract}
\section{Introduction}
Feature selection is a crucial point in high dimensional data analysis. Different techniques have been developed to tackle this problem efficiently, and amongst them sparsity has emerged as a leading paradigm. In statistics, the LASSO estimator \citep{Tibshirani1996} provides a reliable way to select features and has been extensively studied in the last two decades (\cite{Hastie2015} and references therein). In machine learning and signal processing, sparse coding has made its way into several modern architectures, including large scale computer vision \citep{coates2011importance} and biologically inspired models \citep{cadieu2012learning}. Also, Dictionary learning is a generic unsupervised learning method to perform nonlinear dimensionality reduction with efficient computational complexity \citep{Mairal2009a}. All these techniques heavily rely on the resolution of $\ell_1$-regularized least squares.
The $\ell_1$-sparse coding problem is defined as solving, for a given input $x \in {\mathbb{R}}^n$ and dictionary $D \in {\mathbb{R}}^{n \times m}$, the following problem: \begin{equation} \label{eq:sparsecoding}
z^*(x) = \arg\min_z F_x(z) \overset{\Delta}{=}\frac{1}{2} \| x - Dz \|^2 + \lambda \|z \|_1~. \end{equation} This problem is convex and can therefore be solved using convex optimization machinery. Proximal splitting methods \citep{Beck2009} alternate between the minimization of the smooth and differentiable part using the gradient information and the minimization of the non-differentiable part using a proximal operator \citep{Combettes2011}. These methods can also be accelerated by considering a momentum term, as it is done in FISTA \citep{Beck2009, Nesterov2005}.
Coordinate descent \citep{Friedman2007,Osher2009} leverages the closed formula that can be derived for optimizing the problem \autoref{eq:sparsecoding} for one coordinate $z_i$ given that all the other are fixed. At each step of the algorithm, one coordinate is updated to its optimal value, which yields an inexpensive scheme to perform each step. The choice of the coordinate to update at each step is critical for the performance of the optimization procedure. Least Angle Regression (LARS) \citep{Hesterberg2008} is another method that computes the whole LASSO regularization path. These algorithms all provide an optimization procedure that leverages the local properties of the cost function iteratively. They can be shown to be optimal among the class of first-order methods for generic convex, non-smooth functions \citep{bubeck2014theory}.
But all these results are given in the worst case and do not use the distribution of the considered problem. One can thus wonder whether a more efficient algorithm to solve \autoref{eq:sparsecoding} exists for a fixed dictionary $D$ and generic input $x$ drawn from a certain input data distribution. In \cite{Gregor10}, the authors introduced LISTA, a trained version of ISTA that adapts the parameters of the proximal splitting algorithm to approximate the solution of the LASSO using a finite number of steps. This method exploits the common structure of the problem to learn a better transform than the generic ISTA step. As ISTA is composed of a succession of linear operations and piecewise non linearities, the authors use the neural network framework and the backpropagation to derive an efficient procedure solving the LASSO problem. In \cite{Sprechmann2012}, the authors extended LISTA to more generic sparse coding scenarios and showed that adaptive acceleration is possible under general input distributions and sparsity conditions.
In this paper, we are interested in the following question: Given a finite computational budget, what is the optimum estimator of the sparse coding? This question belongs to the general topic of computational tradeoffs in statistical inference. Randomized sketches \citep{alaoui2015fast, yang2015randomized} reduce the size of convex problems by projecting expensive kernel operators into random subspaces, and reveal a tradeoff between computational efficiency and statistical accuracy. \cite{agarwal2012computational} provides several theoretical results on perfoming inference under various computational constraints, and \cite{chandrasekaran2013computational} considers a hierarchy of convex relaxations that provide practical tradeoffs between accuracy and computational cost. More recently, \cite{oymak2015sharp} provides sharp time-data tradeoffs in the context of linear inverse problems, showing the existence of a phase transition between the number of measurements and the convergence rate of the resulting recovery optimization algorithm. \cite{giryes2016tradeoffs} builds on this result to produce an analysis of LISTA that describes acceleration in conditions where the iterative procedure has linear convergence rate. Finally, \cite{xin2016maximal} also studies the capabilities of Deep Neural networks at approximating sparse inference. The authors show that unrolled iterations lead to better approximation if one allows the weights to vary at each layer, contrary to standard splitting algorithms. Whereas their focus is on relaxing the convergence hypothesis of iterative thresholding algorithms, we study a complementary question, namely when is speedup possible, without assuming strongly convex optimization. Their results are consistent with ours, since our analysis also shows that learning shared layer weights is less effective.
Inspired by the LISTA architecture, our mathematical analysis reveals that adaptive acceleration is related to a specific matrix factorization of the Gram matrix of the dictionary $B = D^\mathsf{T} D$ as $ B = A^\mathsf{T} S A - R~,$where $A$ is unitary, $S$ is diagonal and the residual is positive semidefinite: $R \succeq 0$. Our factorization balances between near diagonalization by asking that $\|R \|$ is small and small perturbation of the $\ell_1$ norm, \emph{i.e.} $\| A z \|_1 - \|z \|_1$ is small. When this factorization succeeds, we prove that the resulting splitting algorithm enjoys a convergence rate with improved constants with respect to the non-adaptive version. Moreover, our analysis also shows that acceleration is mostly possible at the beginning of the iterative process, when the current estimate is far from the optimal solution, which is consistent with numerical experiments. We also show that the existence of this factorization is not only sufficient for acceleration, but also necessary. This is shown by constructing dictionaries whose Gram matrix diagonalizes in a basis that is incoherent with the canonical basis, and verifying that LISTA fails in that case to accelerate with respect to ISTA.
In our numerical experiments, we design a specialized version of LISTA called FacNet, with more constrained parameters, which is then used as a tool to show that our theoretical analysis captures the acceleration mechanism of LISTA. Our theoretical results can be applied to FacNet and as LISTA is a generalization of this model, it always performs at least as well, showing that the existence of the factorization is a sufficient certificate for acceleration by LISTA. Reciprocally, we show that for cases where no acceleration is possible with FacNet, the LISTA model also fail to provide acceleration, linking the two speedup mechanisms. This numerical evidence suggest that the existence of our proposed factorization is sufficient and somewhat necessary for LISTA to show good results.
The rest of the paper is structured as follows.
\autoref{sec:math} presents our mathematical analysis and proves the convergence of the adaptive algorithm as a function of the quality of the matrix factorization. Finally, \autoref{sec:exp} presents the generic architectures that will enable the usage of such schemes and the numerical experiments, which validate our analysis over a range of different scenarios.
\section{Accelerating Sparse Coding with Sparse Matrix Factorizations} \label{sec:math}
\subsection{Unitary Proximal Splitting} In this section we describe our setup for accelerating sparse coding based on the Proximal Splitting method. Let $\Omega \subset {\mathbb{R}}^n$ be the set describing our input data, and $D \in {\mathbb{R}}^{n \times m}$ be a dictionary, with $m > n$. We wish to find fast and accurate approximations of the sparse coding $z^*(x)$ of any $x \in \Omega$, defined in \autoref{eq:sparsecoding}
For simplicity, we denote $B = D^\mathsf{T} D$ and $y = D^\dagger x$ to rewrite \autoref{eq:sparsecoding} as \begin{equation}
\label{sc2}
z^*(x) = \arg \min_z F_x(z) = \underbrace{\frac{1}{2}(y - z)^\mathsf{T} B (y-z)}_{E(z)} + \underbrace{\vphantom{\frac{1}{2}}\lambda \| z \|_1}_{G(z)} ~. \end{equation} For clarity, we will refer to $F_x$ as $F$ and to $z^*(x)$ as $z^*$. The classic proximal splitting technique finds $z^*$ as the limit of sequence $(z_k)_k$, obtained by successively constructing a surrogate loss $F_k(z)$ of the form \begin{equation}
F_k(z) = E(z_k) + (z_k-y)^\mathsf{T} B(z-z_k) + L_k \| z - z_k \|_2^2 + \lambda \| z \|_1~, \end{equation} satisfying $F_k(z) \geq F(z) $ for all $z \in {\mathbb{R}}^m$ . Since $F_k$ is separable in each coordinate of $z$, $z_{k+1} = \arg\min_z F_k(z)$ can be computed efficiently.
This scheme is based on a majoration of the quadratic form $(y-z)^\mathsf{T} B (y-z)$ with an isotropic quadratic form $L_k \| z_k - z \|_2^2$. The convergence rate of the splitting algorithm is optimized by choosing $L_k$ as the smallest constant satisfying $F_k(z) \geq F(z)$, which corresponds to the largest singular value of $B$.
The computation of $z_{k+1}$ remains separable by replacing the quadratic form $L_k \textbf{I}$ by any diagonal form. However, the Gram matrix $B = D^\mathsf{T} D$ might be poorly approximated via diagonal forms for general dictionaries. Our objective is to accelerate the convergence of this algorithm by finding appropriate factorizations of the matrix $B$ such that
\[B \approx A^\mathsf{T} S A~,~\text{ and } ~\| A z \|_1 \approx \|z \|_1~,\] where $A$ is unitary and $S$ is diagonal positive definite. Given a point $z_k$ at iteration $k$, we can rewrite $F(z)$ as \begin{equation} F(z) = E(z_k) + ( z_k - y )^\mathsf{T} B(z - z_k) + Q_B(z, z_k)~, \end{equation}
with $\displaystyle Q_B(v, w ) := \frac{1}{2}(v - w)^\mathsf{T} B (v - w) + \lambda \|v \|_1 ~$.
For any diagonal positive definite matrix $S$ and unitary matrix $A$, the surrogate loss \mbox{$\widetilde{F}(z,z_k) := E(z_k) + ( z_k - y)^\mathsf{T} B(z - z_k) + Q_S(A z, Az_k)$} can be explicitly minimized, since \begin{eqnarray} \label{eq:prox} \arg\min_z \widetilde{F}(z,z_k) & = & A^\mathsf{T} \arg\min_u \left( (z_k - y )^\mathsf{T} B A^\mathsf{T} (u - Az_k) + Q_S(u, Az_k)\right) \nonumber \\
&=& A^\mathsf{T} \arg\min_u Q_S\left(u, Az_k - S^{-1}AB(z_k - y)\right) \end{eqnarray}
where we use the variable change $u = Az$. As $S$ is diagonal positive definite, \autoref{eq:prox} is separable and can be computed easily, using a linear operation followed by a point-wise non linear soft-thresholding. Thus, any couple $(A, S)$ ensures an computationally cheap scheme. The question is then how to factorize $B$ using $S$ and $A$ in an optimal manner, that is, such that the resulting proximal splitting sequence converges as fast as possible to the sparse coding solution.
\subsection{Non-asymptotic Analysis} We will now establish convergence results based on the previous factorization. These bounds will inform us on how to best choose the factors $A_k$ and $S_k$ in each iteration.
For that purpose, let us define \begin{equation} \label{eq:eqdelta}
\delta_A(z) = \lambda \left(\| A z \|_1 - \| z\|_1 \right)~,~\text{and }~R = A^\mathsf{T} S A - B~. \end{equation} The quantity $\delta_A(z)$ thus measures how invariant the $\ell_1$ norm is to the unitary operator $A$, whereas $R$ corresponds to the residual of approximating the original Gram matrix $B$ by our factorization $A^\mathsf{T} SA$ . Given a current estimate $z_k$, we can rewrite \begin{equation} \label{eq:surro1} \widetilde{F}(z,z_k) = F(z) + \frac{1}{2}(z - z_k)^\mathsf{T} R (z - z_k) + \delta_A(z)~. \end{equation}
By imposing that $R$ is a positive semidefinite residual one immediately obtains the following bound. \begin{proposition}
\label{eq:bi1}
Suppose that $R = A^\mathsf{T} S A - B$ is positive definite, and define
\begin{flalign}
&& z_{k+1} = \arg\min_z \widetilde{F}(z, z_k)~&.
\label{eq:algo}\\
&\makebox[0pt][l]{Then } &F(z_{k+1}) - F(z^*) \leq \frac{1}{2} \| R \| \| z_k - z^* \|_2^2 +& \delta_A(z^*) - \delta_A(z_{k+1})~.&
\label{eq:bo1}
\end{flalign} \end{proposition} \begin{proof} By definition of $z_{k+1}$ and using the fact that $R \succ 0$ we have \begin{eqnarray*} F(z_{k+1}) - F(z^*) & \leq & F(z_{k+1}) - \widetilde{F}(z_{k+1},z_k) + \widetilde{F}(z^*,z_k) - F(z^*) \\ &=& - \frac{1}{2}(z_{k+1} - z_k)^\mathsf{T} R (z_{k+1} - z_k) - \delta_A(z_{k+1}) + \frac{1}{2}(z^* - z_k)^\mathsf{T} R (z^* - z_k) + \delta_A(z^*) \\ &\leq& \frac{1}{2}(z^* - z_k)^\mathsf{T} R (z^* - z_k) + \left(\delta_A(z^*) - \delta_A(z_{k+1})\right) ~. \end{eqnarray*} where the first line results from the definition of $z_{k+1}$ and the third line makes use of $R$ positiveness. \end{proof}
This simple bound reveals that to obtain fast approximations to the sparse coding
it is sufficient to find $S$ and $A$ such that $\| R \|$ is small and that the $\ell_1$ commutation term $\delta_A$ is small. These two conditions will be often in tension: one can always obtain $R\equiv 0$ by
using the Singular Value Decomposition of $B=A_0^\mathsf{T} S_0 A_0$ and setting $A = A_0$ and $S = S_0$. However, the resulting $A_0$ might introduce large commutation error $\delta_{A_0}$. Similarly, as the absolute value is non-expansive, \emph{i.e.} $\left| |a| - |b| \right| \leq \left| a - b \right| $, we have that \begin{eqnarray} \label{bo2}
|\delta_A(z)| = \lambda \left| \| A z \|_1 - \| z \|_1 \right| &\leq & \lambda \| (A - {\bf I}) z \|_1 \\
&\leq & \lambda \sqrt{2 \max( \| Az\|_0, \| z \|_0)}~\cdot\, \| A - {\bf I} \| ~\cdot\,\|z \|_2~, \nonumber \end{eqnarray}
where we have used the Cauchy-Schwartz inequality $\| x \|_1 \leq \sqrt{ \| x \|_0} \|x \|_2$ in the last equation. In particular, \autoref{bo2} shows that unitary matrices in the neighborhood of ${\bf I}$ with $\| A - {\bf I} \|$ small have small $\ell_1$ commutation error $\delta_A$ but can be inappropriate to approximate general $B$ matrix.
The commutation error also depends upon the sparsity of $z$ and $Az$~. If both $z$ and $Az$ are sparse then the commutation error is reduced, which can be achieved if $A$ is itself a sparse unitary matrix. Moreover, since
\[|\delta_A(z) - \delta_A(z')| \leq \lambda| \|z \|_1 - \| z' \|_1| + \lambda| \| A z \|_1 - \| A z'\|_1 |~\]
\[ \text{and }~ | \|z \|_1 - \| z'\|_1 | \leq \| z - z' \|_1 \leq \sqrt{\| z- z' \|_0 } \| z - z'\|_2\] it results that $\delta_A$ is Lipschitz with respect to the Euclidean norm; let us denote by $L_A(z)$ its local Lipschitz constant in z, which can be computed using the norm of the subgradient in $z$\footnote{ This quantity exists as $\delta_A$ is a difference of convex. See proof of \autoref{lemma1} in appendices for precisions.
}. An uniform upper bound for this constant is $(1+\|A\|_1)\lambda\sqrt{m}$, but it is typically much smaller when $z$ and $Az$ are both sparse.\\ Equation \autoref{eq:algo} defines an iterative procedure determined by the pairs $\{(A_k, S_k)\}_k$. The following theorem uses the previous results to compute an upper bound of the resulting sparse coding estimator. \begin{theorem}
\label{thm1}
Let $A_k, S_k$ be the pair of unitary and diagonal matrices corresponding to iteration $k$,
chosen such that $R_k = A_k^\mathsf{T} S_k A_k - B \succ 0$.
It results that
{\small
\begin{equation}
\label{eq:zo1}
F(z_{k}) - F(z^*) \leq \frac{(z^*- z_0)^\mathsf{T} R_0 (z^* - z_0) + 2 L_{A_0}(z_1) \| z^*-z_{1} \|_2 }{2k} + \frac{\alpha - \beta}{2k}~,
\end{equation}
\begin{eqnarray*}
\text{with } & \alpha = &\sum_{i=1}^{k-1} \left( 2L_{A_i}(z_{i+1})\|z^*-z_{i+1} \|_2 + (z^* - z_i)^\mathsf{T} ( R_{i-1} - R_{i}) (z^* - z_i) \right)~,\\
& \beta = &\sum_{i=0}^{k-1} (i+1)\left((z_{i+1}-z_i)^\mathsf{T} R_i (z_{i+1}-z_i) + 2\delta_{A_i}(z_{i+1}) - 2\delta_{A_i}(z_i) \right)~, \end{eqnarray*} }
where $L_A(z)$ denote the local lipschitz constant of $\delta_A$ at $z$. \end{theorem}
\textbf{Remarks:} If one sets $A_k={\bf I}$ and $S_k = \| B \| {\bf I}$ for all $k \ge 0$, \autoref{eq:zo1} corresponds to the bound of the ISTA algorithm \citep{Beck2009}.
We can specialize the theorem in the case when $A_0, S_0$ are chosen to minimize the bound \autoref{eq:bo1} and
$A_k = {\bf I}$, $S_k = \| B \| {\bf I}$ for $k \ge 1$. \begin{corollary}
If $A_k = {\bf I}$, $S_k = \| B \| {\bf I}$ for $k \ge 1$ then {\small \begin{equation} \label{zo22}
F(z_{k}) - F(z^*) \leq \frac{(z^*- z_0)^\mathsf{T} R_0 (z^* - z_0) + 2 L_{A_0}(z_1) (\| z^*-z_1\| + \| z_1 - z_0\|) + (z^* - z_1)^\mathsf{T} R_0 (z^* - z_1)^\mathsf{T}}{2k}~. \end{equation}} \end{corollary} This corollary shows that by simply replacing the first step of ISTA by the modified proximal step detailed in \autoref{eq:prox}, one can obtain an improved bound at fixed $k$ as soon as
\[2 \| R_0 \| \max(\| z^* - z_0 \|_2^2,\| z^* - z_1 \|_2^2) + 4 L_{A_0}(z_1) \max( \|z^* - z_0\|_2, \|z^* - z_1 \|_2) \leq \|B \| \| z^* - z_0\|_2^2~, \]
which, assuming $\| z^* - z_0 \|_2 \geq \| z^* - z_1 \|_2 $, translates into \begin{equation}
\| R_0 \| + 2 \frac{L_{A_0}(z_1)}{\| z^* - z_0\|_2} \leq \frac{\|B \|}{2}~. \end{equation} More generally, given a current estimate $z_k$, searching for a factorization $(A_k, S_k)$ will improve the upper bound when \begin{equation} \label{zo23}
\| R_k \| + 2 \frac{L_{A_k}(z_{k+1})}{\| z^* - z_k\|_2} \leq \frac{\|B \|}{2}~. \end{equation} We emphasize that this is not a guarantee of acceleration, since it is based on improving an upper bound. However, it provides a simple picture on the mechanism that makes non-asymptotic acceleration possible.
\subsection{Interpretation} \label{sub:interp}
In this section we analyze the consequences of \autoref{thm1} in the design of fast sparse coding approximations, and provide a possible explanation for the behavior observed numerically.
\subsubsection{`Phase Transition" and Law of Diminishing Returns} \label{sub:phase}
\autoref{zo23} reveals that the optimum matrix factorization in terms of minimizing the upper bound depends upon the current scale of the problem, that is, of the distance $\| z^* - z_k\|$. At the beginning of the optimization, when $\| z^* - z_k\|$ is large, the bound \autoref{zo23} makes it easier to explore the space of factorizations $(A, S)$ with $A$ further away from the identity. Indeed, the bound tolerates larger increases in $L_{A}(z_{k+1})$, which is dominated by \[
L_{A}(z_{k+1}) \leq \lambda (\sqrt{\|z_{k+1}\|_0} + \sqrt{\|Az_{k+1}\|_0})~, \]
\emph{i.e.} the sparsity of both $z_1$ and $A_0(z_1)$. On the other hand, when we reach intermediate solutions $z_k$ such that $\| z^* - z_k\|$ is small with respect to $L_A(z_{k+1})$, the upper bound is minimized by choosing factorizations where $A$ is closer and closer to the identity, leading to the non-adaptive regime of standard ISTA ($A=Id$).
This is consistent with the numerical experiments, which show that the gains provided by learned sparse coding methods are mostly concentrated in the first iterations. Once the estimates reach a certain energy level, section \ref{sec:exp} shows that LISTA enters a steady state in which the convergence rate matches that of standard ISTA.
The natural follow-up question is to determine how many layers of adaptive splitting are sufficient before entering the steady regime of convergence. A conservative estimate of this quantity would require an upper bound of $\| z^* - z_k\|$ from the energy bound $F(z_k) - F(z^*)$. Since in general $F$ is convex but not strongly convex, such bound does not exist unless one can assume that $F$ is locally strongly convex (for instance for sufficiently small values of $F$).
\subsubsection{Improving the factorization to particular input distributions} \label{sub:distrib} Given an input dataset $\mathcal{D}={(x_i, z^{(0)}_{i}, z^*_i)}_{i \leq N}$, containing examples $x_i \in {\mathbb{R}}^n$, initial estimates $z^{(0)}_{i}$ and sparse coding solutions $z^*_i$, the factorization adapted to $\mathcal{D}$ is defined as \begin{equation} \label{bty} \min_{A,S;~ A^\mathsf{T} A = {\bf I}, A^\mathsf{T} S A - B \succ 0} \frac{1}{N} \sum_{i \leq N} \frac{1}{2} ( z^{(0)}_{i} - z^*_i)^\mathsf{T} ( A^\mathsf{T} S A - B ) ( z^{(0)}_{i} - z^*_i) + \delta_A(z^*_i) - \delta_A(z_{1,i})~. \end{equation} Therefore, adapting the factorization to a particular dataset, as opposed to enforcing it uniformly over a given ball $B(z^*; R)$ (where the radius $R$ ensures that the initial value $z_0 \in B(z^*;R)$), will always improve the upper bound \autoref{eq:bo1}. Studying the gains resulting from the adaptation to the input distribution will be let for future work.
\section{Numerical Experiments} \label{sec:exp}
This section provides numerical arguments to analyse adaptive optimization algorithms and their performances, and relates them to the theoretical properties developed in the previous section. All the experiments were run using Python and Tensorflow. For all the experiments, the training is performed using Adagrad \citep{duchi2011adaptive}. The code to reproduce the figures is available online\footnote{The code can be found at ~\url{https://github.com/tomMoral/AdaptiveOptim}}.
\subsection{Adaptive Optimization Networks Architectures}
\begin{figure}
\caption{\textbf{ISTA} - Recurrent Neural Network }
\caption{\textbf{LISTA} - Unfolded network}
\caption{ Network architecture for ISTA/LISTA. The unfolded version (b) is trainable through backpropagation and permits to approximate the sparse coding solution efficiently.}
\label{fig:ista}
\label{fig:lista}
\end{figure}
\paragraph{LISTA/LFISTA} \label{par:lista} In \cite{Gregor10}, the authors introduced LISTA, a neural network constructed by considering ISTA as a recurrent neural net.
At each step, ISTA performs the following 2-step procedure : \begin{equation}
\left.\begin{aligned}
1.\hspace{1em} u_{k+1} &= z_k - \displaystyle\frac{1}{L} D^\mathsf{T}(Dz_k-x) = \underbrace{({\bf I} - \frac{1}{L}D^\mathsf{T} D)}_{W_g} z_k + \underbrace{\frac{1}{L}D^\mathsf{T}}_{W_e}x~,\\
2.\hspace{1em} z_{k+1} &= h_{\frac{\lambda}{L}}(u_{k+1}) \text{ where } h_\theta(u) = \text{sign}(u)(\lvert u\rvert-\theta)_+~,\\
\end{aligned}
\right\}
\quad \text{step $k$ of ISTA}
\label{eq:ista} \end{equation} This procedure combines a linear operation to compute $u_{k+1}$ with an element-wise non linearity. It can be summarized as a recurrent neural network, presented in \mbox{\autoref{fig:ista}.}, with tied weights. The autors in \cite{Gregor10} considered the architecture $\Phi_\Theta^K$ with parameters $\Theta = (W_g^{(k)}, W_e^{(k)}, \theta^{(k)})_{k=1,\dots K}$ obtained by unfolding $K$ times the recurrent network, as presented in \autoref{fig:lista}. The layers $\phi_\Theta^k$ are defined as \begin{equation} z_{k+1} = \phi_\Theta^k(z_k) := h_{\theta}( W_g z_k + W_e x)~. \end{equation} If $ W_g^{(k)} = {\bf I} - \frac{D^\mathsf{T} D}{L}$, $ W_e^{(k)} = \frac{D^\mathsf{T}}{L}$ and $ \theta^{(k)} = \frac{\lambda}{L}$ are fixed for all the $K$ layers, the output of this neural net is exactly the vector $z_K$ resulting from $K$ steps of ISTA. With LISTA, the parameters $\Theta$ are learned using back propagation to minimize the cost function: \mbox{$f(\Theta) = \mathbb E_x\left[F_x(\Phi^K_{\Theta}(x))\right]~.$}
A similar algorithm can be derived from FISTA, the accelerated version of ISTA to obtain LFISTA (see \autoref{fig:lfista} in \autoref{sec:lfista}~). The architecture is very similar to LISTA, now with two memory tapes:
\[z_{k+1} = h_{\theta}( W_g z_k + W_m z_{k-1} + W_e x)~.\]
\paragraph{Factorization network} \label{par:facnet}
Our analysis in \autoref{sec:math} suggests a refactorization of LISTA in more a structured class of parameters. Following the same basic architecture, and using \autoref{eq:prox}, the network FacNet, $\Psi_\Theta^K$ is formed using layers such that: \begin{equation}
z_{k+1} = \psi_\Theta^k(z_k) := A^\mathsf{T} h_{\lambda S^{-1}} (A z_{k} - S^{-1}A(D^\mathsf{T} D z_{k} - D^\mathsf{T} x) )~,
\label{eq:facnet} \end{equation} with $S$ diagonal and $A$ unitary, the parameters of the $k$-th layer. The parameters obtained after training such a network with back-propagation can be used with the theory developed in \autoref{sec:math}. Up to the last linear operation $A^\mathsf{T}$ of the network, this network is a re-parametrization of LISTA in a more constrained parameter space. Thus, LISTA is a generalization of this proposed network and should have performances at least as good as FacNet, for a fixed number of layers.
The optimization can also be performed using backpropagation. To enforce the unitary constraints on $A^{(k)}$, the cost function is modified with a penalty: \begin{equation} \label{eq:facnet}
f(\Theta) = \mathbb E_x\left[F_x(\Psi^K_{\Theta}(x))\right] + \frac{\mu}{K}\sum_{k=1}^K\left\|{\bf I} - \left(A^{(k)}\right)^ TA^{(k)}\right\|^2_2~, \end{equation} with $\Theta = (A^{(k)}, S^{(k)})_{k=1\dots K}$ the parameters of the K layers and $\mu$ a scaling factor for the regularization. The resulting matrix $A^{(k)}$ is then projected on the Stiefel Manifold using a SVD to obtain final parameters, coherent with the network structure.
\paragraph{Linear model} \label{par:linear}
Finally, it is important to distinguish the performance gain resulting from choosing a suitable starting point and the acceleration from our model. To highlights the gain obtain by changing the starting point, we considered a linear model with one layer such that $z_{out} = A^{(0)} x$. This model is learned using SGD with the convex cost function \mbox{$ f(A^{(0)}) = \|({\bf I} - DA^{(0)})x\|_2^2+\lambda \|A^{(0)}x\|_1~$}. It computes a tradeoff between starting from the sparsest point $\bf 0$ and a point with minimal reconstruction error $y~$. Then, we observe the performance of the classical iteration of ISTA using $z_{out}$ as a stating point instead of $\bf 0~$.
\subsection{Synthetic problems with known distributions} \label{sub:gen} \begin{figure}
\caption{
Evolution of the cost function $F(z_k)-F(z^*)$ with the number of layers or the number of iteration $k$ for different sparsity level.
\emph{(left)} $\rho=\nicefrac{1}{20}$ and \emph{(right)}$\rho=\nicefrac{1}{4}~.$
}
\label{fig:layers1}
\label{fig:layers2}
\label{fig:layers}
\end{figure}
\textbf{Gaussian dictionary} In order to disentangle the role of dictionary structure from the role of data distribution structure, the minimization problem is tested using a synthetic generative model with no structure in the weights distribution. First, $m$ atoms $d_i \in {\mathbb{R}}^n$ are drawn \emph{iid} from a multivariate Gaussian with mean $\textbf{0}$ and covariance \textbf I$_n$ and the dictionary $D$ is defined as
$\left(\nicefrac{d_i}{\|d_i\|_2}\right)_{i=1\dots m}~.$ The data points are generated from its sparse codes following a Bernoulli-Gaussian model. The coefficients $z=(z_1, \dots, z_m)$ are constructed with $z_i = b_i a_i$, where $b_i\sim \mathcal B(\rho)$ and $a_i \sim \mathcal N(0, \sigma \textbf I_m)~,$ where $\rho$ controls the sparsity of the data. The values are set to $ \medmuskip=0mu \thickmuskip=0mu \thinmuskip=0mu m=100,~\medmuskip=0mu \thickmuskip=0mu \thinmuskip=0mu n = 64$ for the dictionary dimension, $ \rho = \nicefrac{5}{m}$ for the sparsity level and $\medmuskip=0mu \thickmuskip=0mu \thinmuskip=0mu\sigma = 10$ for the activation coefficient generation parameters. The sparsity regularization is set to $\medmuskip=0mu \thickmuskip=0mu \thinmuskip=0mu\lambda = 0.01$. The batches used for the training are generated with the model at each step and the cost function is evaluated over a fixed test set, not used in the training.
\autoref{fig:layers} displays the cost performance for methods ISTA/FISTA/Linear relatively to their iterations and for methods LISTA/LFISTA/FacNet relatively to the number of layers used to solve our generated problem. Linear has performances comparable to learned methods with the first iteration but a gap appears as the number of layers increases, until a point where it achieves the same performances as non adaptive methods. This highlights that the adaptation is possible in the subsequent layers of the networks, going farther than choosing a suitable starting point for iterative methods. The first layers permit to achieve a large gain over the classical optimization strategy, by leveraging the structure of the problem. This appears even with no structure in the sparsity patterns of input data, in accordance with the results in the previous section. We also observe diminishing returns as the number of layers increases. This results from the phase transition described in \autoref{sub:phase}, as the last layers behave as ISTA steps and do not speed up the convergence. The 3 learned algorithms are always performing at least as well as their classical counterpart, as it was stated in \autoref{thm1}. We also explored the effect of the sparsity level in the training and learning of adaptive networks. In the denser setting, the arbitrage between the $\ell_1$-norm and the squared error is easier as the solution has a lot of non zero coefficients. Thus in this setting, the approximate method is more precise than in the very sparse setting where the approximation must perform a fine selection of the coefficients. But it also yield lower gain at the beggining as the sparser solution can move faster.
There is a small gap between LISTA and FacNet in this setup. This can be explained from the extra constraints on the weights that we impose in the FacNet, which effectively reduce the parameter space by half. Also, we implement the unitary constraints on the matrix $A$ by a soft regularization (see \autoref{eq:facnet}), involving an extra hyper-parameter $\mu$ that also contributes to the small performance gap. In any case, these experiments show that our analysis accounts for most of the acceleration provided by LISTA, as the performance of both methods are similar, up to optimization errors.
\begin{figure}
\caption{Evolution of the cost function $F(z_k)-F(z^*)$ with the number of layers or the number of iteration $k$ for a problem generated with an adversarial dictionary.}
\label{fig:adverse}
\end{figure}
\textbf{Adversarial dictionary}
The results from \autoref{sec:math} show that problems with a gram matrix composed of large eigenvalues associated to non sparse eigenvectors are harder to accelerate. Indeed, it is not possible in this case to find a quasi diagonalization of the matrix $B$ that does not distort the $\ell_1$ norm. It is possible to generate such a dictionary using Harmonic Analysis. The Discrete Fourier Transform (DFT) distorts a lot the $\ell_1$ ball, since a very sparse vector in the temporal space is transformed in widely spread spectrum in the Fourier domain. We can thus design a dictionary for which LISTA and FacNet performances should be degraded.
$ D = \left(\nicefrac{d_i}{\|d_i\|_2}\right)_{i=1\dots m}~$ is constructed such that $ d_{j,k} = e^{-2\pi i j \zeta_k}$, with $\left(\zeta_k\right)_{k \le n}$ randomly selected from $\left\{\nicefrac{1}{m},\dots, \nicefrac{\nicefrac{m}{2}}{m}\right\}$ without replacement.
The resulting performances are reported in \autoref{fig:adverse}. The first layer provides a big gain by changing the starting point of the iterative methods. It realizes an arbitrage of the tradeoff between starting from $\bf 0$ and starting from $y~.$ But the next layers do not yield any extra gain compared to the original ISTA algorithm. After $4$ layers, the cost performance of both adaptive methods and ISTA are equivalent. It is clear that in this case, FacNet does not accelerate efficiently the sparse coding, in accordance with our result from \autoref{sec:math}. LISTA also displays poor performances in this setting. This provides further evidence that FacNet and LISTA share the same acceleration mechanism as adversarial dictionaries for FacNet are also adversarial for LISTA.
\subsection{Sparse coding with over complete dictionary on images} \label{sub:img}
\textbf{Wavelet encoding for natural images} \label{par:wavelet} A highly structured dictionary composed of translation invariant Haar wavelets is used to encode 8x8 patches of images from the PASCAL VOC 2008 dataset. The network is used to learn an efficient sparse coder for natural images over this family. $500$ images are sampled from dataset to train the encoder. Training batches are obtained by uniformly sampling patches from the training image set to feed the stochastic optimization of the network. The encoder is then tested with $10000$ patches sampled from $100$ new images from the same dataset.
\textbf{Learned dictionary for MNIST} To evaluate the performance of LISTA for dictionary learning, LISTA was used to encode MNIST images over an unconstrained dictionary, learned {\it a priori} using classical dictionary learning techniques. The dictionary of $100$ atoms was learned from $10000$ MNIST images in grayscale rescaled to 17x17 using the implementation of \cite{Mairal2009a} proposed in scikit-learn, with $\lambda = 0.05$. Then, the networks were trained through backpropagation using all the $60000$ images from the training set of MNIST. Finally, the perfornance of these encoders were evaluated with the $10000$ images of the training set of MNIST.
\begin{figure}
\caption{Pascal VOC 2008}
\label{fig:pascal}
\caption{MNIST}
\label{fig:mnist}
\caption{
Evolution of the cost function $F(z_k)-F(z^*)$ with the number of layers or the number of iteration $k$ for two image datasets.}
\label{fig:images}
\end{figure}
The \autoref{fig:images} displays the cost performance of the adaptive procedures compared to non-adaptive algorithms. In both scenario, FacNet has performances comparable to the one of LISTA and their behavior are in accordance with the theory developed in \autoref{sec:math}. The gains become smaller for each added layer and the initial gain is achieved for dictionary either structured or unstructured. The MNIST case presents a much larger gain compare to the experiment with natural images. This results from the difference of structure of the input distribution, as the MNIST digits are much more constrained than patches from natural images and the network is able to leverage it to find a better encoder. In the MNIST case, a network composed of $12$ layers is sufficient to achieve performance comparable to ISTA with more than $1000$ iterations.
\section{Conclusions} \label{sec:conclusions}
In this paper we studied the problem of finite computational budget approximation of sparse coding. Inspired by the ability of neural networks to accelerate over splitting methods on the first few iterations, we have studied which properties of the dictionary matrix and the data distribution lead to such acceleration. Our analysis reveals that one can obtain acceleration by finding approximate matrix factorizations of the dictionary which nearly diagonalize its Gram matrix, but whose orthogonal transformations leave approximately invariant the $\ell_1$ ball. By appropriately balancing these two conditions, we show that the resulting rotated proximal splitting scheme has an upper bound which improves over the ISTA upper bound under appropriate sparsity.
In order to relate this specific factorization property to the actual LISTA algorithm, we have introduced a reparametrization of the neural network that specifically computes the factorization, and incidentally provides reduced learning complexity (less parameters) from the original LISTA. Numerical experiments of \autoref{sec:exp} show that such reparametrization recovers the same gains as the original neural network, providing evidence that our theoretical analysis is partially explaining the behavior of the LISTA neural network. Our acceleration scheme is inherently transient, in the sense that once the iterates are sufficiently close to the optimum, the factorization is not effective anymore. This transient effect is also consistent with the performance observed numerically, although the possibility remains open to find alternative models that further exploit the particular structure of the sparse coding. Finally, we provide evidence that successful matrix factorization is not only sufficient but also necessary for acceleration, by showing that Fourier dictionaries are not accelerated.
Despite these initial results, a lot remains to be understood on the general question of optimal tradeoffs between computational budget and statistical accuracy. Our analysis so far did not take into account any probabilistic consideration (e.g. obtain approximations that hold with high probability or in expectation).
Another area of further study is the extension of our analysis to the FISTA case, and more generally to other inference tasks that are currently solved via iterative procedures compatible with neural network parametrizations, such as inference in Graphical Models using Belief Propagation or other ill-posed inverse problems.
\appendix
\section{Learned Fista} \label{sec:lfista}
A similar algorithm can be derived from FISTA, the accelerated version of ISTA to obtain LFISTA (see \autoref{fig:lfista}~). The architecture is very similar to LISTA, now with two memory taps: It introduces a momentum term to improve the convergence rate of ISTA as follows: \begin{enumerate}
\item $\displaystyle y_k = z_{k} + \frac{\medmuskip=0mu \thickmuskip=0mu \thinmuskip=0mu t_{k-1} - 1}{t_k}(z_k - z_{k-1})$~,
\item $\displaystyle z_{k+1} = h_{\frac{\lambda}{L}}\left(y_k - \frac{1}{L}\nabla E(y_k)\right)= h_{\frac{\lambda}{L}}\left(({\bf I} - \frac{1}{L}B)y_k + \frac{1}{L}D^\mathsf{T} x\right)$~,
\item $ \displaystyle t_{k+1} = \frac{1 + \sqrt{1 + 4t_k^2}}{2}$~. \end{enumerate} By substituting the expression for $y_k$ into the first equation, we obtain a generic recurrent architecture very similar to LISTA, now with two memory taps, that we denote by LFISTA: \[z_{k+1} = h_{\theta}( W_g^{(k)} z_k + W_m^{(k)} z_{k-1} + W_e^{(k)} x)~.\] This model is equivalent to running $K$-steps of FISTA when its parameters are initialized with \begin{eqnarray}
W_g^{(k)} & = & \left(1 + \frac{t_{k-1} - 1}{t_{k}}\right)\left({\bf I} - \frac{1}{L}B\right)~, \nonumber \\
W_m^{(k)} & = & \left(\frac{1 - t_{k-1}}{t_{k}}\right)\left({\bf I} - \frac{1}{L}B\right)~, \nonumber \\
W_e^{(k)} & = & \frac{1}{L}D^\mathsf{T}~. \nonumber \end{eqnarray} The parameters of this new architecture, presented in \autoref{fig:lfista}~, are trained analogously as in the LISTA case.
\begin{figure}
\caption{ Network architecture for LFISTA. This network is trainable through backpropagation and permits to approximate the sparse coding solution efficiently.}
\label{fig:lfista}
\end{figure}
\section{Proofs}
\begin{lemma} \label{lemma1_a} Suppose that $R = A^\mathsf{T} S A - B$ is positive definite, and define \begin{equation} \label{eq:algo_a} z_{k+1} = \arg\min_z \widetilde{F}(z, z_k)~,\text{ and } \end{equation}
$\delta_A(z) = \|A z \|_1 - \|z \|_1$. Then we have \begin{equation} \label{eq:bo2_a} F(z_{k+1}) - F(z^*) \leq \frac{1}{2} \left( (z^* -z_{k})^\mathsf{T} R(z^* -z_{k}) - (z^* -z_{k+1})^\mathsf{T} R(z^* -z_{k+1}) \right) + \langle \partial \delta_A(z_{k+1}) , z_{k+1} - z^* \rangle ~. \end{equation} \end{lemma}
\begin{proof} We define $$f(t) = F\left( t z_{k+1} + (1-t) z^*\right)~,~t \in [0,1]~.$$ Since $F$ is convex, $f$ is also convex in $[0,1]$.
Since $f(0) = F(z^*)$ is the global minimum, it results that $f'(t)$ is increasing in $(0,1]$, and hence \begin{equation*} F(z_{k+1}) - F(z^*) = f(1) - f(0) = \int f'(t) dt \leq f'(1)~, \end{equation*} where $f'(1)$ is any element of $\partial f(1)$. Since $\delta_A(z)$ is a difference of convex functions, its subgradient can be defined as a limit of infimal convolutions \cite{Hiriart-Urruty1991}. We have $$\partial f(1) = \langle \partial F(z_{k+1}), z_{k+1} - z^* \rangle~,$$ and since $$\partial F(z) = \partial \widetilde{F}(z,z_k) - R (z - z_k) - \partial \delta_A(z)~\text{ and}~0 \in \partial \widetilde{F}(z_{k+1}, z_k)$$ it results that \begin{equation*} \partial F(z_{k+1}) = -R (z_{k+1} - z_k) - \partial \delta_A(z_{k+1})~, \end{equation*} and thus \begin{equation} F(z_{k+1}) - F(z^*) \leq (z^*-z_{k+1})^\mathsf{T} R ( z_{k+1} - z_k) + \langle \partial \delta_A(z_{k+1}) , (z^*-z_{k+1}) \rangle~. \end{equation} \autoref{eq:bo2_a} is obtained by observing that \begin{equation} \label{eq:bt1_a}
(z^*-z_{k+1})^\mathsf{T} R ( z_{k+1} - z_k) \leq \frac{1}{2} \left( (z^* -z_{k})^\mathsf{T} R(z^* -z_{k}) - (z^* -z_{k+1})^\mathsf{T} R(z^* -z_{k+1}) \right) ~, \end{equation} thanks to the fact that $R \succ 0$. \end{proof}
\begin{theorem} \label{eq:co1a_a} Let $A_k, S_k$ be the pair of unitary and diagonal matrices corresponding to iteration $k$, chosen such that $R_k = A_k^\mathsf{T} S_k A_k - B \succ 0$. It results that \begin{equation} \label{eq:zo1a_a} F(z_{k}) - F(z^*) \leq \frac{(z^*- z_0)^\mathsf{T} R_0 (z^* - z_0) + 2\langle \nabla \delta_{A_0}(z_{1}) , (z^*-z_{1}) \rangle }{2k} + \frac{\alpha - \beta}{2k}~,\text{ with} \end{equation} $$ \alpha = \sum_{n=1}^{k-1} \left( 2\langle \nabla \delta_{A_n}(z_{n+1}) , (z^*-z_{n+1}) \rangle + (z^* - z_n)^\mathsf{T} ( R_{n-1} - R_{n}) (z^* - z_n) \right)~,$$ $$\beta = \sum_{n=0}^{k-1} (n+1)\left((z_{n+1}-z_n)^\mathsf{T} R_n (z_{n+1}-z_n) + 2\delta_{A_n}(z_{n+1}) - 2\delta_{A_n}(z_n) \right)~.$$ \end{theorem} {\it Proof:} The proof is adapted from \citep{Beck2009}, Theorem 3.1. From \autoref{lemma1_a},
we start by using \autoref{eq:bo2_a} to bound terms of the form $F(z_n) - F(z^*)$: {\small \begin{equation*} F(z_n) - F(z^*) \leq \langle \nabla \delta_{A_n}(z_{n+1}) , (z^*-z_{n+1}) \rangle + \frac{1}{2} \left( (z^* -z_{n})^\mathsf{T} R_n (z^* -z_{n}) - (z^* -z_{n+1})^\mathsf{T} R_n(z^* -z_{n+1}) \right)~. \end{equation*} } Adding these inequalities for $n = 0 \dots k-1$ we obtain \begin{eqnarray} \label{eq:vg1_a} \left(\sum_{n=0}^{k-1} F(z_n) \right) - k F(z^*) &\leq& \sum_{n=0}^{k-1} \langle \nabla \delta_{A_n}(z_{n+1}) , (z^*-z_{n+1}) \rangle + \\ \nonumber && + \frac{1}{2} \left( (z^*- z_0)^\mathsf{T} R_0 (z^* - z_0) - (z^*- z_k)^\mathsf{T} R_{k-1} (z^* - z_k) \right) + \\ \nonumber && + \frac{1}{2} \sum_{n=1}^{k-1} (z^* - z_n)^\mathsf{T} ( R_{n-1} - R_{n}) (z^* - z_n)~. \end{eqnarray} On the other hand, we also have \begin{eqnarray*} \label{eq:vg2_a} F(z_n) - F(z_{n+1}) &\geq& F(z_n) - \tilde{F}(z_n,z_n) + \tilde{F}(z_{n+1},z_n) - F(z_{n+1}) \\ &=& - \delta_{A_n}(z_n) + \delta_{A_n}(z_{n+1}) + \frac{1}{2}(z_{n+1}-z_n)^\mathsf{T} R_n (z_{n+1}-z_n)~, \end{eqnarray*} which results in \begin{eqnarray} \label{eq:vg3_a} \sum_{n=0}^{k-1} (n+1)(F(z_n) - F(z_{n+1})) &\geq& \frac{1}{2}\sum_{n=0}^{k-1}(n+1) (z_{n+1}-z_n)^\mathsf{T} R_n (z_{n+1}-z_n) + \\ \nonumber && + \sum_{n=0}^{k-1}(n+1) \left( \delta_{A_n}(z_{n+1}) - \delta_{A_n}(z_n) \right) \\ \nonumber \left(\sum_{n=0}^{k-1} F(z_n) \right) - k F(z_k) &\geq& \sum_{n=0}^{k-1} (n+1)\left(\frac{1}{2}(z_{n+1}-z_n)^\mathsf{T} R_n (z_{n+1}-z_n) + \delta_{A_n}(z_{n+1}) - \delta_{A_n}(z_n) \right) ~. \end{eqnarray} Combining \autoref{eq:vg1_a} and \autoref{eq:vg3_a} we obtain \begin{eqnarray} F(z_k) - F(z^*) &\leq& \frac{(z^*- z_0)^\mathsf{T} R_0 (z^* - z_0) + 2\langle \nabla \delta_{A_0}(z_{1}) , (z^*-z_{1}) \rangle }{2k} + \frac{\alpha - \beta}{2k}\\ \nonumber \end{eqnarray} with $$ \alpha = \sum_{n=1}^{k-1} \left( 2\langle \nabla \delta_{A_n}(z_{n+1}) , (z^*-z_{n+1}) \rangle + (z^* - z_n)^\mathsf{T} ( R_{n-1} - R_{n}) (z^* - z_n) \right)~,$$ $$\beta = \sum_{n=0}^{k-1} (n+1)\left((z_{n+1}-z_n)^\mathsf{T} R_n (z_{n+1}-z_n) + 2\delta_{A_n}(z_{n+1}) - 2\delta_{A_n}(z_n) \right)~.$$ $\square$
\begin{corollary} \label{eq:coro2_a}
If $A_k = {\bf I}$, $S_k = \| B \| {\bf I}$ for $k>0$ then \begin{equation} \label{eq:zo22_a}
F(z_{k}) - F(z^*) \leq \frac{(z^*- z_0)^\mathsf{T} R_0 (z^* - z_0) + 2 L_{A_0}(z_1) (\| z^*-z_1\| + \| z_1 - z_0\|) + (z^* - z_1)^\mathsf{T} R_0 (z^* - z_1)^\mathsf{T}}{2k}~. \end{equation} \end{corollary} {\it Proof: } We verify that in that case, $R_{n-1} - R_n \equiv 0$ and for $n>1$ and $\delta_{A_n} \equiv 0$ for $n > 0$ $\square$.
\end{document}
|
arXiv
|
{
"id": "1609.00285.tex",
"language_detection_score": 0.7686305642127991,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Study of the transcendence of a family of generalized
continued fractions}
\begin{abstract} We study a family of generalized continued fractions, which are defined by a pair of substitution sequences in a finite alphabet. We prove that they are {\em stammering} sequences, in the sense of Adamczewski and Bugeaud. We also prove that this family consists of transcendental numbers which are not Liouvillian. We explore the partial quotients of their regular continued fraction expansions, arriving at no conclusion concerning their boundedness. \end{abstract}
\keywords{Continued fractions; Transcendence; Stammering Sequences.}
\section{Introduction}
The problem of characterizing continued fractions of numbers beyond rational and quadratic has received consistent attention over the years. One direction points to an attempt to understand properties of algebraic numbers of degree at least three, but at times even this line ends up in the realm of transcendental numbers.
Some investigations \cite{bruno,garrity,schwe} on algebraic numbers depart from generalizations of continued fractions. This line of investigation has been tried since Euler, see \cite{schwe,bruno} and references therein, with a view to generalize Lagrange's theorem on quadratic numbers, in search of proving a relationship between algebraic numbers and periodicity of multidimensional maps yielding a sequence of approximations to a irrational number. This theory has been further developed for instance to the study of ergodicity of the triangle map \cite{meno}. In fact, a considerable variety of algorithms may be called generalizations of continued fractions: for instance \cite{keane}, Jacobi-Perron's and Poincaré's algorithms in \cite{schwe}.
We report on a study of generalized continued fractions of the form: \begin{equation} \label{thab} \theta(a,b)\stackrel{{\rm def}}{=} a_0+\frac{b_0}{a_1+\frac{b_1}{a_2+\frac{b_2}{a_3+\frac{b_3}{ a_4+\frac { b_4 } { \ddots}}}}}
\ , \end{equation} (where $a_0\geq 0$, $a_n\in \mathbb N$, for $n\in \mathbb N$ and $b_n \in \mathbb N$, for $n\geq 0$), investigating a class with a regularity close, in a sense, to periodicity. This family of generalized continued fractions converges when $(a_n)$ and $(b_n)$ are finite valued sequences. They were considered {\em formally} in \cite{list}, in the context exemplified in Section \ref{exam1}. The {\em stammering} sequences \cite{acta}, and sequences generated by morphisms \cite{abd,aldavquef,allsha}, consist in a natural step away from periodic ones. Similarly to the results in \cite{abd,aldavquef}, on regular continued fractions, we prove that the family of numbers considered are transcendental.
\begin{teo} \label{teo1} Suppose $(a_n)$ and $(b_n)$, $n\geq 0$, are fixed points of primitive substitutions, then the number $\theta(a,b)$ is transcendental. \end{teo}
\def\mathsf S{\mathsf S} We also consider Mahler's classification within this family of transcendental numbers, proving that they cannot be Liouvillian. Let us call the numbers of the family of generalized continued fractions $\theta(a,b)$, with ${\mathbf a}$ and ${\mathbf b}$ stammering sequences coming from a primitive substitution number of {\em type} $\mathsf S_3$.
\begin{teo} Type $\mathsf S_3$ numbers are either $S$-numbers or $T$-numbers in Mahler's classification. \label{teo2} \end{teo}
The paper is organized as follows. In Section \ref{conv}, we prove the convergence of the generalized continued fraction expansions for type $\mathsf S_3$ numbers. In Section \ref{transc}, we prove the transcendence of type $\mathsf S_3$ numbers. In Section \ref{liou}, we use Baker's Theorem to prove that they are either $S$-numbers or $T$-numbers. In Section \ref{exam1}, we show some inconclusive calculations on the partial quotients of the regular continued fraction of an specific type $\mathsf S_3$ number.
\section{Convergence} \label{conv}
We start from the analytic theory of continued fractions \cite{wall} to prove the convergence of \eqref{thab} when $(a_n)$ and $(b_n)$, $n\geq 0$, are sequences in a finite alphabet.
Let $\overline{\R}^+=[0,\infty]$ denote the extended positive real axis, with the understanding that $a+\infty=\infty$, for any $a\in \overline{\R}^+$; $a\cdot \infty=\infty$, if $a>0$, $0\cdot \infty = 0$ and $a/\infty=0$, if $a\in \mathbb R$. We do not need to define $\infty/\infty$.
Given the sequences $(a_k)$ and $(b_k)$ of non-negative (positive) integers, consider the Möbius transforms $t_k:\overline{\R}^+\to \overline{\R}^+$ \[ t_k(w)= a_k+\frac{b_k}{w} \ ,\; k\in \mathbb N \ ,\] and their compositions \[ t_it_j(w)= t_i(a_j+b_j/w)=a_i+b_i/(a_j+b_j/w) \ .\] This set of Möbius transforms is closed under compositions and form a semigroup.
It is useful to consider the natural correspondence between Möbius transformations and $2\times 2$ matrices: \[ M_k=\begin{pmatrix} a_k & b_k \\ 1 & 0 \end{pmatrix} \ .\] Taking the
positive real cone ${\cal C}_2=\{(x,y) \ |\ x\geq 0\ , \ y\geq 0\ , x+y>0\}$ with the equivalence $(x,y) \sim \lambda (x,y)$ for every $\lambda>0$, we have an homomorfism between the semigroup of Möbius transforms, under composition, acting on $\overline{\R}^+$ and the algebra of matrices above (which are all invertible) acting on ${\cal C}_2/\sim$.
Assume the limit \[ \lim_{n\to \infty} t_0t_1t_2\cdots t_n(0) \] exists as a real positive number, then it is given once we know the sequences $(a_n)$ and $(b_n)$, $n\geq 0$. In this case, it is equal to $\lim_{n\to \infty} t_0t_1t_2\cdots t_{n-1}(\infty)$ as well, so that the initial point may be taken as $0$ or $\infty$ in the extended positive real axis.
In terms of matrices multiplication, we have \[ M_0M_1M_2\cdots M_n \begin{pmatrix} 0 \\ 1 \end{pmatrix} \sim M_0M_1M_2\cdots M_{n-1} \begin{pmatrix} 1 \\ 0 \end{pmatrix} \] in ${\cal C}_2$. Define $p_{-1}=1$, $q_{-1}=0$, $p_0=a_0$, $q_0=1$ and \[ \begin{pmatrix}
p_n & b_np_{n-1} \\
q_n & b_nq_{n-1}
\end{pmatrix} \stackrel{{\rm def}}{=} M_0M_1M_2 \cdots M_n \ ,\ n\geq 0 \ . \] We have the following second order recursive formulas for $(p_n,q_n)$: \begin{align} \label{recur} & p_{n+1}=a_{n+1}p_n+b_np_{n-1} \\ & q_{n+1}=a_{n+1}q_n+b_nq_{n-1} \nonumber \end{align} and the determinant formula \begin{equation} \label{detAn} p_nq_{n-1}-p_{n-1}q_n = (-1)^{n-1} b_0\cdots b_{n-1} \end{equation}
We recall the series associated with a continued fraction \cite{wall}: \begin{lema} Let $(q_n)$ denote the sequence of denominators given in \eqref{recur} for the continued fraction $\theta(a,b)$. \if 0 \begin{equation} \label{basecf} \frac{1}{1+\frac{b_1}{a_2+\frac{b_2}{a_3+\frac{b_3}{\ddots}}}} \ . \end{equation} \fi Let \begin{equation}
\rho_k=-\frac{b_k q_{k-1}}{q_{k+1}} \ ,\; k \in \mathbb N \ .\label{assure_conv} \end{equation} Then \[ a_0+\frac{b_0}{a_1}\left(1+\sum_{k=1}^{n-1} \rho_1\rho_2\ldots \rho_k \right) = \frac{p_n}{q_n} \ , n\geq 1 \ . \] \end{lema}
\begin{proof} For $n=1$, the sum is empty, $p_1=a_1a_0+b_0$ and $q_1=a_1$; the equality holds.
Consider the telescopic sum, for $n\geq 1$, \[ \frac{p_1}{q_1}+\sum_{k=1}^{n-1} \left( \frac{p_{k+1}}{q_{k+1}} - \frac{p_k}{q_k}\right) = \frac{p_n}{q_n} \ . \] From \eqref{detAn}, \[ \left( \frac{p_{k+1}}{q_{k+1}} - \frac{p_k}{q_k}\right) = (-1)^k \frac{b_0b_1\cdots b_k}{q_{k+1}q_k}\ .\] Now $\frac{b_0}{a_1}\rho_1=-\frac{b_0}{q_1}\frac{b_1q_0}{q_2} =-\frac{b_0b_1}{q_1q_2}=\frac{p_2}{q_2}-\frac{p_1}{q_1}$. Moreover \[ \frac{p_3}{q_3}-\frac{p_2}{q_2}= \frac{b_0b_1b_2}{q_3q_2} = \frac{b_0}{q_1} \left(-\frac{b_1q_0}{q_2}\right)\left(-\frac{b_2q_1}{ q_3 } \right) = \frac{b_0}{a_1} \rho_1\rho_2 \ . \] Multiplicative cancelling provides the argument to deduce the formula \[ \frac{p_{k+1}}{q_{k+1}}-\frac{p_k}{q_k}= \frac{b_0}{a_1} \rho_1\rho_2\ldots \rho_k \] by induction, finishing the proof. \end{proof}
Even though $b_n>1$ may occur in \eqref{thab}, note that we still have $q_{n+1}\geq 2^{(n-1)/2}$. Indeed $q_0=1$, $q_1=a_1$ and $q_2=a_2q_1+b_1q_0>1$. Finally, since $a_n\geq 1$ and $b_n\geq 1$ for all $n\in \mathbb N$, \[ q_{n+1}=a_{n+1}q_n+b_nq_{n-1} \geq q_n+q_{n-1}\geq 2^{n/2-1}+\frac{2^{(n-1)/2}}2> 2^{(n-1)/2} \ .\]
\begin{lema} If $(a_n)$ and $(b_n)$ are sequences on a finite alphabet $\mathscr A\subset [\alpha,\beta] \subset [1,\infty)$, then the generalized continued fraction \eqref{thab} converges. \label{dois} \end{lema}
\begin{proof} It follows from \eqref{recur} that $(q_n)$, $n\geq 1$, is increasing, and
\[ |\rho_k| = \left| \frac{b_kq_{k-1}}{q_{k+1}}\right| = \left(1+ \frac{a_{k+1}q_k}{b_kq_{k-1}}\right)^{-1}< (1+\alpha/\beta)^{-1} \ . \] Thus the series with general term $\rho_1\cdots \rho_k$ is bounded by a convergent geometric series. \end{proof}
\begin{lema} Let $(q_n)$ denote the sequence of denominators given in \eqref{recur} for a generalized continued fraction, with $(a_n)$ and $(b_n)$ sequences in a finite alphabet $\mathscr A \subset [\alpha,\beta]\subset [1,\infty)$. Then $q_n^{1/n}$ is bounded. \label{logq_nlim} \end{lema}
\begin{proof} From \eqref{recur}, $q_1=a_1$, and $q_{n+1}<(a_{n+1}+b_n)q_n \leq (2\beta)q_n$. Hence $q_n^{1/n}\leq 2\beta \sqrt[n]{a_1}\leq 2\beta^{3/2}$, where the last inequality is necessary only when $a_1=\beta$. \end{proof}
\section{Transcendence} \label{transc}
We now specialize the study of generalized continued fractions for sequences $(a_n)$ and $(b_n)$ which are generated by primitive substitutions. These sequences provide a wealth of examples of {\em stammering sequences}, defined below, following \cite{acta}.
Let us introduce some notation. The set ${\cal A}$ is called alphabet. A word $w$ on ${\cal A}$ is a finite or infinite sequence of
letters in ${\cal A}$. For finite $w$, $|w|$ denotes the number of letters composing $w$. Given a natural number $k$, $w^k$ is the word obtained by $k$ concatenated repetitions of $w$. Given a rational number $r>0$, which is not an integer, $w^r$ is the word $w^{\lfloor r\rfloor} w'$, where $\lfloor r\rfloor$ denotes the integer part of $r$ and $w'$ is a prefix of $w$ of length $\lceil (r-\lfloor
r\rfloor)|w|\rceil$, where $\lceil q\rceil=\lfloor q\rfloor +1$ is the upper integer part of $q$.
Note that if $(a_n)$ and $(b_n)$ are sequences on ${\cal A}$, then $(a_n,b_n)$ is a sequence in ${\cal A}\times {\cal A}$, which is also an alphabet. A sequence ${\bf a}=(a_n)$ has the {\em stammering property} if it is not a periodic sequence and, given $r>1$, there exists a sequence of finite words $(w_n)_{n\in \mathbb N}$, such that \begin{itemize} \item[a)] for every $n\in \mathbb N$, $w_n^r$ is a prefix of ${\bf a}$;
\item[b)] $(|w_n|)$ is increasing. \end{itemize} We say, more briefly, that $(a_n)$ is a stammering sequence with exponent $r$. It is clear that if $(a_n)$ and $(b_n)$ are both stammering with exponents $r$ and $s$ respectively, then $(a_n,b_n)$ is also stammering with exponent $\min\{r,s\}$.
\begin{lema} If $u$ is a substitution sequence on a finite alphabet $\mathscr A$, then $u$ is stammering. \label{fact1} \end{lema}
\begin{proof} Denote the substitution map by $\xi:\mathscr A\to \mathscr A^+$. Since $\mathscr A$ is a finite set, there is a $k\geq 1$ and $\alpha \in \mathscr A$ such that $\alpha$ is a prefix of $\xi^k(\alpha)$. $u=\lim_{n\to \infty} \xi^{kn}(\alpha)$. Moreover, there is a least finite $j$ such that $\alpha$ occurs a second time in $\xi^{jk}(\alpha)$. Therefore $u$ is stammering with $w\geq
1+\frac{1}{|\xi^{jk}(\alpha)|-1}$. \end{proof}
\begin{proof}[Proof of Theorem \ref{teo1}] From Lemma \ref{fact1} $(a_n)$ and $(b_n)$ stammering sequences with exponent $r>1$. Hence $\theta(a,b)$ has infintely many {\em good} quadratic approximations. Let $(w_n)\in ({\cal A}\times {\cal A})^*$ be a sequence of words of increasing length characterizing $(a_n,b_n)$ as a stammering sequence with exponent $r>1$. Consider $\psi_k(a,b)$ given by \[ \psi_k(a,b)= c_0+\frac{d_0}{c_1+\frac{d_1}{c_2+\frac{d_2}{c_3+\frac{d_3}{\ddots}}}}
\ , \] where $c_j=a_j$, $d_j=b_j$, for $0\leq j<k$, and $c_{j}=c_{j\pmod{k}}$ and $d_{j}=d_{j\pmod{k}}$, for $j\geq k$. $\psi_k$ is a root of the quadratic equation \[ q_{k-1}x^2+(q_k-p_{k-1})x-p_k=0 \ , \] which might not be in lowest terms.
Arguing as in Theorem 1 from \cite{acta}, we choose $k$ from the
subsequence of natural numbers given by $|w_n^r|$. Lemma \ref{logq_nlim} allows us to conclude that the generalized continued fraction $\theta(a,b)$ is transcendental if both $(a_n)$ and $(b_n)$ are stammering sequences with exponent $r>1$. \end{proof}
\section{Quest on Liouville numbers} \label{liou}
We address the question of Mahler's classification of the numbers for type ${\mathcal S}_3$ numbers. The statement of Baker's Theorem we quote use a measure of transcendence introduced by Koksma, which is equivalent to Mahler's, and we explain briefly, following \cite{plms}, Section 2.
Let $d\geq 1$ and $\xi$ a real number. Denote by $P(X)$ an arbitrary polynomial with integer coefficients, and
$H(P)=\max_{0\leq k\leq j} \{|a_k|\ :\ P(X)=a_0+a_1X+\cdots+a_jX^j\}$ is the height of the polynomial $P$. Let $w_d(\xi)$ be the supremum of the real numbers such that the inequality
\[ 0< |P(\xi)|\leq H(P)^{-w} \] is true for infinitely many polynomials $P(X)$ with integer coefficients and degree at most $d$. Koksma introduced $w_d^*(\xi)$ as the supremum of the real numbers $w^*$ such that
\[ 0< |\xi-\alpha| \leq H(\alpha)^{-w^*-1} \] are true for infinitely many algebraic numbers $\alpha$ of degree at most $d$, where $H(\alpha)$ is the height of the minimal polynomial with integer coefficients which vanishes at $\alpha$.
Let $w(\xi) = \lim_{d\to \infty} \frac{w_d(\xi)}{d}$, then $\xi$ is called \begin{itemize} \item an $A$-number if $w(\xi)=0$; \item an $S$-number if $0<w(\xi)<\infty$; \item a $T$-number if $w(\xi)=\infty$, but $w_d(\xi)<\infty$ for every integer $d\geq 1$; \item an $U$-number if $w(\xi)=\infty$ and $w_d(\xi)=\infty$ for some $d\geq 1$. \end{itemize}
It was shown by Koksma that $w^*_d$ and $w^*$ provide the same classification of numbers. Liouville numbers are precisely those for which $w_1(\xi)=\infty$, they are $U$-numbers of type 1.
\begin{teob}[Baker] \label{baker} Let $\xi$ be a real number and $\epsilon >0$. Assume there is an infinite sequence of irreducible rational numbers $(p_n/q_n)_{n\in \mathbb N}$, $(p_n,q_n)=1$, ordered such that $2\leq q_1< q_2\leq \cdots $ satisfying
\[ \left| \xi - \frac{p_n}{q_n} \right| < \frac 1{q_n^{2+\epsilon} } \ . \] Additionally, suppose that \[ \limsup_{n\to \infty} \frac{\log q_{n+1}}{\log q_n}<\infty \ , \] then there is a real number $c$, depending only on $\xi$ and $\epsilon$ such that \[ w_d^* (\xi) \leq \exp\exp ( cd^2) \ . \] for every $d\in \mathbb N$. Consequently, $\xi$ is either an $S$-number or a $T$-number. \end{teob}
\begin{proof}[Proof of Theorem \ref{teo2}] We note that the hypothesis of irreducibility is lacking for type $\mathsf S_3$ numbers. Let us write $d_n=(p_n,q_n)$. By eq. \eqref{detAn}, $d_n=b_0\ldots b_{n-1}$. Recall that, for primitive substitutions in ${\cal A}=\{\alpha,\beta\}\subset \mathbb N$, there is a frequency $\nu$, which is uniform in the sequence $({\mathbf b})$ \cite{queffe}, for which $b_k=\beta$. Thus, $d_n \approx \beta^\nu \alpha^{1-\nu}$ for large $n$. If, for every $n\in \mathbb N$, there is a number $\theta$ such that $d_n < \left(\frac{q_n}{d_n}\right)^\theta$, then
\[ 0< \left| \xi -\frac{p_n/d_n}{q_n/d_n}\right| < \frac 1{q_n^{2+\epsilon}}= \frac 1{d_n^{2+\epsilon} (q_n/d_n)^{2+\epsilon}} \ .\] In this case, from the estimates of $q_n$ and $d_n$, the limit \[ \limsup_{n\to \infty} \frac{\log(q_{n+1}/d_{n+1})}{\log(q_n/d_n)} <\infty \ .\] We would conclude from Theorem \ref{baker} that type $\mathsf S_3$ contains either $S$-numbers or $T$-numbers and no Liouville numbers.
From the analysis of Section \ref{conv}, keeping its notations, \[ (-1)^n \frac{d_n}{q_nq_{n-1}} = \frac{b_0}{a_1} \rho_1 \ldots \rho_{n-1} \ . \] Now
$|\rho_k|=\left(1+\frac{a_{k+1}q_k}{b_kq_{k-1}}\right)^{-1}$, and since $q_k\leq 2\beta q_{k-1}$, we conclude that
\[ |\rho_k| > \left( 1+ \frac{2\beta^2}{\alpha}\right)^{-1} \ . \] Therefore, recalling that $q_{n-1}\geq 2^{(n-3)/2}$ \[ \frac{d_n}{q_n} > q_{n-1} (1+2\beta^2/\alpha)^{-n+1} \frac{b_0}{a_1} \quad \Rightarrow\quad \frac{q_n}{d_n}< (1+2\beta^2/\alpha)^{n-1} 2^{-(n-3)/2}\frac{\beta}{\alpha} \ . \]
Therefore, we want to determine the existence of a solution for $\theta$ for the inequality \[ d_n < \left(\frac{q_n}{d_n}\right)^\theta \] considering that $d_n \approx \beta^{\nu n} \alpha^{(1-\nu)n}$ we obtain the inequality \[ \beta^{\nu n} \alpha^{(1-\nu)n} < (1+2\beta^2/\alpha)^{\theta(n-1)}2^{-\theta(n-3)/2} \frac{\beta}{\alpha} \ .\] For large $n$, it is sufficient to solve \[ \beta^{\nu}\alpha^{1-\nu} < \frac 1{2^{\theta/2}}\left(1+\frac{2\beta^2}{\alpha}\right)^\theta \ ,\] which clearly has the solution $\theta=1$, since $\alpha<\beta$ and $0<\nu<1$, implying $\beta^\nu \alpha^{1-\nu} < \beta$. We conclude that type $\mathsf S_3$ consists only of $S$-numbers or $T$-numbers. \end{proof}
\section{Example: partial quotients of a corresponding regular continued fraction} \label{exam1}
We now examine one specific example: a generalized continued fraction associated with the period doubling sequence. The period doubling sequence, which we denote by $\omega$, is the fixed point of the substitution $\xi(\alpha)=\alpha\beta$ and $\xi(\beta)=\alpha\alpha$ on the two lettered alphabet $\{\alpha,\beta\}$. It is also the limit of a sequence of {\em foldings}, and called a {\em folded sequence} \cite{allsha}.
We make some observations and one question about the partial quotients of the corresponding regular continued fraction representing the real number that the generalized continued fraction given by \eqref{thab} when both sequences $({\mathbf a})$ and $({\mathbf b})$ are given by the period doubling sequence: $a_n=b_n=\omega_n$.
We choose to view the period doubling sequence as the limit of folding operations. The algebra of matrices with fixed determinant will play a role. A folding is a mapping \begin{align*} {\cal F}_p: &{\cal A}^*\to {\cal A}^*\\ & w\mapsto wp\tilde{w} \end{align*} where $\tilde{w}$ equals the word $w$ reversed: if $w=a_1\ldots a_n$, $a_i\in {\cal A}$, then $\tilde{w}=a_n\ldots a_1$, and $p\in {\cal A}^*$.
It is clear that \[ \omega= \lim_{n\to \infty} ({\cal F}_a\circ {\cal F}_b)^n (a) \ ,\] see also \cite{allsha}, where the limit is understood in the product topology (of the discrete topology) in ${\cal A}^\mathbb N\cup {\cal A}^*$.
Let $\theta$ denote the number whose generalized continued fraction is obtained from the substitution of the letters $\alpha$ and $\beta$ by \[ A=\begin{pmatrix} 1 & 1 \\ 1 & 0 \end{pmatrix} \ , \quad B= \begin{pmatrix} 3 & 3 \\
1 & 0 \end{pmatrix} \] respectively. It corresponds to the choice $\{1,3\}$ for the alphabet where the sequences $({\mathbf a})$ and $({\mathbf b})$ take values.
Now we use Raney transducers \cite{raney} to describe the computation of some partial quotients of the regular continued fraction converging to $\theta$. A transducer ${\mathscr T}=(Q,\Sigma,\delta,\lambda)$, or two-tape machine, is defined by a set of states $Q$, an alphabet $\Sigma$, a transition function $\delta: Q\times \sigma \to Q$, and an output function $\lambda: Q\times \sigma \to \Sigma^*$, where $\sigma\subset \Sigma$ (a more general definition is possible \cite{allsha}, but this is sufficient for our purposes).
The states of Raney's tranducers are column and row (or doubly) balanced matrices over the non-negative integers with a fixed determinant. A matrix $\begin{pmatrix} a & b \\ c & d\end{pmatrix}$ is column balanced if $(a-b)(c-d)<0$ \cite{raney}.
Figure 1 shows the Raney transducer for determinant 3 doubly balanced matrices. In the text, we use the abbreviations: $\beta_1=\begin{pmatrix} 3 & 0 \\ 0 & 1 \end{pmatrix}$, $\beta_2=\begin{pmatrix} 1 & 0 \\ 0 & 3 \end{pmatrix}$ and $\beta_3=\begin{pmatrix} 2 & 1 \\ 1 & 2 \end{pmatrix}$. Then $Q=\{\beta_1,\beta_2,\beta_3\}$, $\Sigma=\{L,R\}$, where \[ R=\begin{pmatrix}
1 & 1 \\ 0 & 1
\end{pmatrix} \ , \quad L=\begin{pmatrix} 1 & 0 \\ 1 & 1 \end{pmatrix} \ ,\] the transition function $\delta$ and the output function are indicated in the graph.
For instance, if $RL^2R$ is the input word on state $\beta_2$, the output word is $L^2R^4$ and the final state is $\beta_1$. Any infinite word in $\Sigma^\mathbb N$ can be read by ${\mathscr T}$, but not every finite word can be fully read by ${\mathscr T}$, for instance, $L^{11}$ in state $\beta_1$ will produce $L^3$, but $L^2$ will stay in the reading queue in state $\beta_1$. Algebraically, these two examples are written as \begin{align*} \beta_2 RL^2R & = L \beta_3 LR = L LR \beta_1R = L^2 RR^3 \beta_1 = L^2R^4 \beta_1 \\ \beta_1 L^{11} &= L^3 \beta_1 L^2 \end{align*}
As explained in \cite{vdP}, Theorem 1 in \cite{list} or even Theorem 5.1 in \cite{raney}, one may use the transducer ${\mathscr T}$ to commute the matrices $A$ and $B$ to get an approximation to the continued fraction of $\theta$.
\begin{figure}
\caption{Transducer $\mathscr T$.}
\label{trans1}
\end{figure}
\def\mathscr A{\mathscr A} \def\mathscr B{\mathscr B} Introducing the matrix $J=\begin{pmatrix} 0 & 1 \\ 1 & 0 \end{pmatrix}$, we note the following relations: $A=RJ$, $B=\beta_1 RJ$, $\beta_2 J=J\beta_1$.
The homomorfism between ${\cal A}^*$ and the semigroup of matrices generated by $\{A,B\}$, as Moebius transforms, is the basis for discovering some curious properties of the regular continued fraction of $\theta$.
The sequence of matrices \[ ABA,\ ABAAABA,\ ABAAABABABAAABA \ ,\] which corresponds to ${\cal F}_b(a)$, ${\cal F}_a\circ{\cal F}_b(a)$, ${\cal F}_b\circ {\cal F}_a\circ {\cal F}_b(a)$, yields the beginning of the regular continued fraction expansion of $\theta$.
A step by step calculation shows the basic features in the use of ${\mathscr T}$: \begin{align*} ABAAABA &= RJ(\beta_1 RJ) RJRJRJ(\beta_1 RJ)RJ \\ &= R \beta_2 (JRJ)R(JRJ)R\beta_2 (JRJ)RJ \\ &= R \beta_2 LRLR\beta_2 LRJ \\ & = R L^3 \beta_2 RLR L^3 \beta_2 RJ \\ & =RL^3 L \beta_3 RL^3 \beta_2 RJ \\ & = RL^4 RL \beta_2 L^3 \beta_2 RJ \\ &= RL^4 RL L^9 \beta_2^2 RJ = RL^4 RL^{10} \beta_2^2 RJ \ . \end{align*} This means that the continued fraction of $\theta$ begins as $[1;4,1,k,\cdots]$, with $k\geq 10$.
Writing $T=ABAAABA$, upon the next folding \begin{align*} TBT &= RL^4RL^{10} \beta_2^2 RJ (\beta_1RJ) RL^4RL^{10} \beta_2^2 RJ \\ &= RL^4 RL^{10} \beta_2^2 R\beta_2 LRL^4RL^{10} \beta_2^2 RJ \\ &= RL^4RL^{10} \beta_2^2RL^3L\beta_3 L^3 RL^{10} \beta_2^2 RJ \\ & = RL^4R L^{10} \beta_2^2 RL^4LR \beta_1 L^2R L^{10} \beta_2^2 RJ \\ &= RL^4R L^{10} \beta_2^2 RL^5 RRL^2 \beta_2 L^{10} \beta_2^2 RJ \\ &=RL^4RL^{10} \beta_2 L \beta_3L^4 R^2 L^{32} \beta_2^3 RJ \\ &= RL^4RL^{10} \beta_2 L \beta_3 L^4 R^2 L^{32}\beta_2^3 RJ \\ &= RL^4 R L^{13} \beta_2 LR \beta_1 L^3 R^2L^{32} \beta_2^3 RJ \\ & = RL^4 RL^{16} \beta_2 R L \beta_1 R^2 L^{32} \beta_2^3 RJ \\ &= RL^4 RL^{17} \beta_3 R^6L^{10} \beta_1 L^2 \beta_2^3 RJ \\ &= RL^4 RL^{17} RL \beta_2 R^5L^{10} \beta_1 L^2\beta_2^3 RJ \\ &= RL^4 RL^{17} RL R LR^2 \beta_1 L^9 \beta_1 L^2 \beta_2^3 RJ \\ &= RL^4 RL^{17} RLRLR^2L^3 \beta_1^2 L^2 \beta_2^3 RJ \end{align*} Now we have the knowledge that the beginning of the regular continued fraction expansion of $\theta=[1;4,1,17,1,1,1,1,2,a,\cdots]$, with $a\geq 3$.
Instructions for the transducer are stuck on the right of this factorization to be read for the next folding. Note that only states $\beta_1$ and $\beta_2$ will remain, since a transition from $\beta_3$ is always possible given any finite word in $\Sigma^*$.
An inductive prediction as to whether the high power in the beginning, $L^{17}$, will consistently increase upon (sufficient) repetitions of foldings is out of reach. This observation poses the question: are the partial quotients of regular continued fraction of $\theta$ bounded? Similar calculations have been done with a simpler choice of the alphabet, that is, $\{1,2\}$, where the transducer has only two states (doubly balanced matrices with determinant 2).
\end{document}
|
arXiv
|
{
"id": "2004.06350.tex",
"language_detection_score": 0.7268756031990051,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{center}
\textbf{\Large{Investigating spatial scan statistics for multivariate functional data}}
\\ Camille Frévent$^1$, Mohamed-Salem Ahmed$^{1}$ , Sophie Dabo-Niang$^2$ and Michaël Genin$^1$ \\ $^1$Univ. Lille, CHU Lille, ULR 2694 - METRICS: Évaluation des technologies de santé et des pratiques médicales, F-59000 Lille, France.\\ $^2$Laboratoire Paul Painvelé UMR CNRS 8524, INRIA‐MODAL, University of Lille.
\end{center}
\begin{center}
\rule{1\linewidth}{.9pt} \end{center} \noindent \textbf{\Large Abstract}
\\ \noindent This paper introduces new scan statistics for multivariate functional data indexed in space. The new methods are derivated from a MANOVA test statistic for functional data, an adaptation of the Hotelling $T^2$-test statistic, and a multivariate extension of the Wilcoxon rank-sum test statistic. In a simulation study, the latter two methods present very good performances and the adaptation of the functional MANOVA also shows good performances for a normal distribution. Our methods detect more accurate spatial clusters than an existing nonparametric functional scan statistic. Lastly we applied the methods on multivariate functional data to search for spatial clusters of abnormal daily concentrations of air pollutants in the north of France in May and June 2020.\\
\noindent \textbf{Keywords: } Cluster detection, multivariate functional data, spatial scan statistics \\ \begin{center} \rule{1\linewidth}{.9pt} \end{center}
\noindent \section{Introduction}
Spatial cluster detection has been studied for many years. The goal is usually to develop new tools capable of detecting the aggregation of spatial sites that behave ``differently'' from other sites. In particular, spatial scan statistics detect statistically significant spatial clusters with a scanning window and without any pre-selection bias. This approach was originally proposed by \cite{spatialdisease} and \cite{spatialscanstat} in the cases of Bernouilli and Poisson models. They present a method based on the likelihood ratio and Monte-Carlo testing to detect significant clusters of various sizes and shapes. Following on from Kulldorff's initial work, several researchers have adapted spatial scan statistics to other spatial data distributions, such as ordinal \citep{jung2007spatial}, normal \citep{normalkulldorff}, exponential \citep{huang2007spatial}, and Weibull models \citep{bhatt2014spatial}. These methods were applied in many different fields such as epidemiology \citep{epidemio1, epidemio2, genin2020fine}, environmental sciences \citep{social2, social1}, geology \citep{geology}. \\
\noindent Thanks to progress in sensing and data storage capacity, data are increasingly being measured continuously over time. This led to the introduction of functional data analysis (FDA) by \cite{ramsaylivre}. A considerable amount of work has gone to adapt classical statistical methods to the univariate functional framework such as principal component analysis \citep{fpc_boente, fpc_berrendero} or regression \citep{reg_cuevas, reg_ferraty, reg_chiou} but also to the multivariate functional one \citep{local_functional, clustering_multi}). \\
\noindent In some research fields, such as environmental surveillance, pollution sensors are deployed in a geographical area. In a context where these sensors measure simultaneously the concentrations of many pollutants at regular intervals over a long period of time, environmental experts may search for environmental black-spots , that can be defined as geographical areas characterized by elevated concentrations of pollutants. For this purpose three different approaches can be considered. The simplest one consists in summarizing the information by averaging each variable over the time and to apply a parametric multivariate spatial scan statistic \citep{a_multivariate_gaussian} or the nonparametric one proposed by \cite{nonparam_multi} but this could lead to a huge loss of information when the data is measured over a long time period. Another solution could be to apply a spatial scan statistic for univariate functional data on each variable \citep{wilco_cucala, notre_fonctionnel}. However this does not allow to take into account the correlations between the variables. A relevant solution consists in using spatial scan statistics for multivariate functional data. According to the authors, the nonparametric spatial scan statistic for functional data proposed by \cite{wilco_cucala} could be extended to multivariate processes although it has never been evaluated in this context. Moreover to our knowledge no parametric scan statistic for multivariate functional data has been proposed. Thus we will define new spatial scan statistics for multivariate functional data based on statistical tests for comparing multivariate functional samples. Recently \cite{manovafonctional} and \cite{two_s} developed respectively a MANOVA test statistic and a functional Hotelling $T^2$-test statistic for multivariate functional data. We also propose to consider the multivariate extension of the Wilcoxon rank-sum test developed by \cite{oja} as a pointwise test statistic. Using these statistics we will adapt to the multivariate functional framework the parametric and the distribution-free spatial scan statistics for functional data proposed by \cite{notre_fonctionnel} and we will also investigate a new multivariate functional method based on the ranks of the observations at each time. \\
\noindent This paper develops three new spatial scan statistics for multivariate functional data. Section \ref{sec:method} describes the parametric multivariate functional scan statistic, the multivariate version of the distribution-free functional spatial scan statistic proposed by \cite{notre_fonctionnel} and a new rank-based spatial scan statistic for multivariate functional data . In Section \ref{sec:simulation} the behaviours of our methods are investigated through a simulation study and compared to the one proposed by \cite{wilco_cucala}. The methods are applied on a real dataset in Section \ref{sec:realdata}. Finally the paper is concluded with a discussion in Section \ref{sec:discussion}.
\section{Methodology} \label{sec:method}
\subsection{General principle} \noindent Let $\{ X(t), \ t \in \mathcal{T} \}$ be a $p$-dimensional vector-valued stochastic process where $\mathcal{T}$ is an interval of $\mathbb{R}$. Let $s_1, \dots, s_n$ be $n$ non-overlapping locations of an observation domain $S \subset \mathbb{R}^2$ and $X_1, \dots, X_n$ be the observations of $X$ in $s_1, \dots, s_n$. Hereafter all observations are considered to be independent, which is a classical assumption in scan statistics. \noindent Spatial scan statistics aim at detecting spatial clusters and testing their significance. Hence, one tests a null hypothesis $\mathcal{H}_0$ (the absence of a cluster) against a composite alternative hypothesis $\mathcal{H}_1$ (the presence of at least one cluster $w \subset S$ presenting abnormal values of $X$).
\noindent \cite{notre_fonctionnel} defined the notion of cluster in the univariate functional framework. Their definitions can be easily extended to the multivariate functional context by defining a \textit{multivariate magnitude cluster} $w$ as follows: \begin{equation} \forall t \in \mathcal{T}, \ \mathbb{E}[X_i(t)\mid s_i\in w] = \mathbb{E}[X_i(t)\mid s_i\notin w] + \Delta(t), \end{equation} where $\Delta(t) = (\Delta_1(t), \dots, \Delta_p(t))^\top$, all $\Delta_i$ are of constant and identical signs, and exists $i \in \llbracket 1 ; p \rrbracket$ such that $\Delta_i$ is non-zero over at least one sub-interval of $\mathcal{T}$. In the same way a \textit{multivariate shape cluster} can be defined as follows: \begin{equation} \forall t \in \mathcal{T}, \ \mathbb{E}[X_i(t)\mid s_i\in w] = \mathbb{E}[X_i(t)\mid s_i\notin w] + \Delta(t) \end{equation} where $\Delta(t) = (\Delta_1(t), \dots, \Delta_p(t))^\top$ and exists $i \in \llbracket 1 ; p \rrbracket$ such that $\Delta_i$ is not constant almost everywhere. \\
\noindent Since the article of \cite{cressie}, a scan statistic is defined by the maximum of a concentration index over a set of potential clusters $\mathcal{W}$.
In the following and without loss a generality, we focus on variable-size circular clusters \citep[as introduced by][]{spatialscanstat}. The set of potential clusters $\mathcal{W}$ is the set of discs centered on a location and passing through another one, with $|w|$ the number of sites in $w$: \begin{equation}
\mathcal{W} = \{ w_{i,j} \ / \ 1 \le |w_{i,j}| \le \frac{n}{2}, \ 1 \le i,j \le n \}, \end{equation} where $w_{i,j}$ is the disc centered on $s_i$ that passes through $s_j$. Thus, a cluster cannot cover more than 50\% of the studied region which is the recommended approach of \cite{spatialdisease}. Remark that in the literature other possibilities have been proposed such as elliptical clusters \citep{elliptic}, rectangular clusters \citep{rectangular} or graph-based clusters \citep{cucala_graph}. \\
\noindent We proposed a parametric scan statistic in subsection \ref{subsec:param}, a distribution-free one is detailed in subsection \ref{subsec:dfree} and a new rank-based scan statistic for multivariate functional data is developed in subsection \ref{subsec:max_wmw}.
\subsection{A parametric spatial scan statistic for multivariate functional data} \label{subsec:param}
\noindent In this subsection the process $X$ is supposed to take values in the Hilbert space $L^2(\mathcal{T}, \mathbb{R}^p)$ of $p$-dimensional vector-valued square-integrable functions on $\mathcal{T}$, equipped with the inner product $\langle X, Y \rangle = \int_{\mathcal{T}} X(t)^\top Y(t) \ \text{d}t$.
\noindent \cite{notre_fonctionnel} proposed a parametric scan statistic for univariate functional data based on a functional ANOVA. A multivariate version of the ANOVA is the classical MANOVA Lawley–Hotelling trace test \citep{oja}. It was adapted by \cite{manovafonctional} for $L^2(\mathcal{T}, \mathbb{R}^p)$ processes: considering two groups $g_1$ and $g_2$ of independent random observations of two $p$-dimensional stochastic processes $X_{g_1}$ and $X_{g_2}$ taking values in $L^2(\mathcal{T}, \mathbb{R}^p)$, it tests the equality of the two mean vector-valued functions $\mu_{g_1}$ and $\mu_{g_2}$ where $\mu_{g_i}(t) = \mathbb{E}[X_{g_i}(t)] \in \mathbb{R}^p$, $i=1,2, \ t \in \mathcal{T}$. \\
\noindent For the cluster detection problem, the null hypothesis $\mathcal{H}_0$ (the absence of a cluster) can be defined by $\mathcal{H}_0: \forall w \in \mathcal{W}, \ \mu_{w} = \mu_{w^\mathsf{c}} = \mu_S$, where $\mu_w$, $\mu_{w^\mathsf{c}}$ and $\mu_S$ stand for the mean functions in $w$, outside $w$ and over $S$, respectively. And the alternative hypothesis $\mathcal{H}_1^{(w)}$ associated with a potential cluster $w$ can be defined as $\mathcal{H}_1^{(w)}: \mu_{w} \neq \mu_{w^\mathsf{c}}$. Thus we can use the functional MANOVA to compare the mean functions in $w$ and $w^\mathsf{c}$. \\ Actually \cite{manovafonctional} presented the adaptation of different MANOVA tests to the functional framework. However the Wilks lambda test statistic, the Lawley-Hotelling trace test statistic and the Pillai trace test statistic presented in the article showed similar performances. In addition, they often outperformed in terms of power the tests proposed in the same article that use projections. Thus we decide to study the Lawley-Hotelling trace test for the cluster detection problem by using the following statistic: \begin{equation} \mathrm{LH}^{(w)} = \text{Trace}(H_w E_w^{-1}) \end{equation}
where $$H_w = |w| \int_{\mathcal{T}} [\bar{X}_w(t) - \bar{X}(t)][\bar{X}_w(t) - \bar{X}(t)]^\top \ \text{d}t + |w^\mathsf{c}| \int_{\mathcal{T}} [\bar{X}_{w^\mathsf{c}}(t) - \bar{X}(t)][\bar{X}_{w^\mathsf{c}}(t) - \bar{X}(t)]^\top \ \text{d}t$$ \text{ and } $$E_w = \sum_{j, s_j \in w} \int_{\mathcal{T}} [X_j(t) - \bar{X}_w(t)][X_j(t) - \bar{X}_w(t)]^\top \text{d}t + \sum_{j, s_j \in w^\mathsf{c}} \int_{\mathcal{T}} [X_j(t) - \bar{X}_{w^\mathsf{c}}(t)][X_j(t) - \bar{X}_{w^\mathsf{c}}(t)]^\top \text{d}t$$
where $\bar{X}_{g}(t) = \frac{1}{|g|} \sum_{i, s_i \in g} X_i(t)$ are empirical estimators of $\mu_g(t)$ ($g \in \{w, w^\mathsf{c}\}$), $\bar{X}(t) = \frac{1}{n} \sum_{i = 1}^n X_i(t)$ is the empirical estimator of $\mu_S(t)$. \\
\noindent Now, $\mathrm{LH}^{(w)}$ can be considered as a concentration index and maximized over the set of potential clusters $\mathcal{W}$, which results in the following definition of the parametric multivariate functional spatial scan statistic (PMFSS): \begin{equation} \Lambda_{\text{PMFSS}} = \underset{w \in \mathcal{W}}{\max} \ \mathrm{LH}^{(w)}. \end{equation} The potential cluster for which this maximum is obtained, namely the most likely cluster (MLC) is \begin{equation} \text{MLC} = \underset{w \in \mathcal{W}}{\arg \max} \ \mathrm{LH}^{(w)}. \end{equation}
\subsection{A distribution-free spatial scan statistic for multivariate functional data} \label{subsec:dfree}
\cite{notre_fonctionnel} proposed a distribution-free spatial scan statistic for univariate functional data based on the combination of the distribution-free scan statistic for non-functional data proposed by \cite{a_distribution_free} which relies on a Student's t-test, and the globalization of a pointwise test over the time \citep{hd_manova}. \\ Very recently, \cite{two_s} proposed a version of this pointwise test for $p$-dimensional functional data ($p \ge 2$) to compare the mean functions of $X$ in two groups. \\
\noindent We suppose that for each time $t$, $\mathbb{V}[X_i(t)] = \Sigma(t,t)$ for all $i \in \llbracket 1 ; n \rrbracket$, where $\Sigma$ is a $p \times p$ covariance matrix function. \\
\noindent Thus, as previously, in the context of cluster detection, the null hypothesis $\mathcal{H}_0$ can be defined as follows: $\mathcal{H}_0: \forall w \in \mathcal{W}, \ \mu_{w} = \mu_{w^\mathsf{c}} = \mu_S$, where $\mu_w$, $\mu_{w^\mathsf{c}}$ and $\mu_S$ stand for the mean functions in $w$, outside $w$ and over $S$, respectively. And the alternative hypothesis $\mathcal{H}_1^{(w)}$ associated with a potential cluster $w$ can be defined as follows: $\mathcal{H}_1^{(w)}: \mu_{w} \neq \mu_{w^\mathsf{c}}$. Next, \cite{two_s} proposed to compare the mean function $\mu_w$ in $w$ with the mean function $\mu_{w^\mathsf{c}}$ in $w^\mathsf{c}$ by using the following statistic: $$ T_{n,\text{max}}^{(w)} = \underset{t \in \mathcal{T}}{\sup} \ T_n(t)^{(w)}$$ where $T_n(t)$ is a pointwise statistic defined by the Hotelling $T^2$-test statistic
$$ T_n(t)^{(w)} = \frac{|w| |w^\mathsf{c}|}{n} (\bar{X}_w(t) - \bar{X}_{w^\mathsf{c}}(t))^\top \hat{\Sigma}(t,t)^{-1} (\bar{X}_w(t) - \bar{X}_{w^\mathsf{c}}(t)).$$ $\bar{X}_w(t)$ and $\bar{X}_{w^\mathsf{c}}(t)$ are the empirical estimators of the mean functions defined in subsection \ref{subsec:param}, and $$\hat{\Sigma}(s,t) = \frac{1}{n-2} \left[ \sum_{i, s_i \in w} (X_i(s) - \bar{X}_w(s)) (X_i(t) - \bar{X}_w(t))^\top + \sum_{i, s_i \in w^\mathsf{c}} (X_i(s) - \bar{X}_{w^\mathsf{c}}(s)) (X_i(t) - \bar{X}_{w^\mathsf{c}}(t))^\top \right]$$ is the pooled sample covariance matrix function. \\
\noindent Then $T_{n,\text{max}}^{(w)}$ is considered as a concentration index and maximized over the set of potential clusters $\mathcal{W}$, yielding to the following multivariate distribution-free functional spatial scan statistic (MDFFSS):
\begin{equation} \Lambda_{\text{MDFFSS}} = \underset{w \in \mathcal{W}}{\max} \ T_{n,\text{max}}^{(w)}. \end{equation} The most likely cluster is therefore \begin{equation} \text{MLC} = \underset{w \in \mathcal{W}}{\arg \max} \ T_{n,\text{max}}^{(w)}. \end{equation}
\subsection{A new rank-based spatial scan statistic for multivariate functional data} \label{subsec:max_wmw}
\cite{oja} developed a $p$-dimensional ($p \ge 2$) extension of the classical Wilcoxon rank-sum test using multivariate ranks. Following on \cite{oja}'s definitions, we can define the notion of ``pointwise multivariate ranks'' as following: \\ For each time $t \in \mathcal{T}$, the pointwise multivariate ranks are defined by $$R_i(t) = \frac{1}{n} \sum_{j=1}^{n} \text{sgn}(A_X(t)(X_i(t) - X_j(t)))$$ where $\text{sgn}(\cdot)$ is the spatial sign function defined as $$\begin{array}{ccccl} \text{sgn} &: & \mathbb{R}^p & \to & \mathbb{R}^p \\
& & x & \mapsto & \left\{ \begin{array}{cl}
||x||_2^{-1} x & \text{ if } x \neq 0 \\
0 & \text{ otherwise} \end{array} \right. \\ \end{array}$$ and $A_X(t)$ is a pointwise data-based transformation matrix that makes the pointwise multivariate ranks behave as though they are spherically distributed in the unit $p$-sphere: $$ \frac{p}{n}\sum_{i=1}^{n} R_i(t) R_i(t)^\top = \frac{1}{n} \sum_{i=1}^{n} R_i(t)^\top R_i(t) I_p.$$ Note that this matrix can be easily computed using an iterative procedure. \\
\noindent Without loss of generality, \cite{oja} compared the cumulative distribution functions of real multivariate observations in two groups. In the context of multivariate functional data, their statistic can be considered as a pointwise test statistic for each time $t$: the pointwise multivariate extension of the Wilcoxon rank-sum test statistic is defined as
$$W(t)^{(w)} = \frac{pn}{\sum_{i=1}^{n} R_i(t)^\top R_i(t)} \left[ \ |w| \ ||\bar{R}_w(t) ||_2^2 + |w^\mathsf{c}| \ ||\bar{R}_{w^\mathsf{c}}(t) ||_2^2 \ \right]$$ where $\bar{R}_{g}(t) = \frac{1}{|g|} \sum_{i, s_i \in g} R_i(t) $ for $g \in \{ w, w^\mathsf{c} \}.$ \\
\noindent Now we propose as previously to globalize the information over the time with $$W^{(w)} = \underset{t \in \mathcal{T}}{\sup} \ W(t)^{(w)}.$$
\noindent Then in the context of cluster detection, the null hypothesis is defined as $\mathcal{H}_0$: $\forall w \in \mathcal{W}, \ \forall t, \ F_{w,t} = F_{w^\mathsf{c},t}$ where $F_{w,t}$ and $F_{w^\mathsf{c},t}$ correspond respectively to the cumulative distribution functions of $X(t)$ in $w$ and outside $w$. The alternative hypothesis $\mathcal{H}_1^{(w)}$ associated with a potential cluster $w$ is $\mathcal{H}_1^{(w)}$: $\exists t$, \ $F_{w,t}(x) = F_{w^\mathsf{c},t}(x-\Delta_t)$, $\Delta_t \neq 0$. \\
\noindent Then $W^{(w)}$ can be considered as a concentration index and maximized over the set of potential clusters $\mathcal{W}$ so that the multivariate rank-based functional spatial scan statistic (MRBFSS) is defined as follows:
\begin{equation} \Lambda_{\text{MRBFSS}} = \underset{w \in \mathcal{W}}{\max} \ W^{(w)}. \end{equation} Thus the most likely cluster is \begin{equation} \text{MLC} = \underset{w \in \mathcal{W}}{\arg \max} \ W^{(w)}. \end{equation}
\subsection{Computing the significance of the MLC} \label{subsec:significance}
\noindent Once the most likely cluster has been detected, its significance must be evaluated. The distribution of the scan statistic $\Lambda$ ($\Lambda_{\text{PMFSS}}$, $\Lambda_{\text{MDFFSS}}$ or $\Lambda_{\text{MRBFSS}}$) is untractable under $\mathcal{H}_0$ due to the dependence between $\mathcal{S}^{(w)}$ and $\mathcal{S}^{(w')}$ if $w \cap w' \neq \emptyset$ ($\mathcal{S} = \mathrm{LH}, T_{n,\text{max}}$ or $W$). Then we chose to obtain a large set of simulated datasets by randomly permuting the observations $X_i$ in the spatial locations. This technique called ``random labelling'' was already used in spatial scan statistics \citep{normalkulldorff, a_multivariate_gaussian, notre_fonctionnel}.
\noindent Let $M$ denote the number of random permutations of the original dataset and $\Lambda^{(1)},\dots,\Lambda^{(M)}$ be the observed scan statistics on the simulated datasets. According to \cite{dwass} the p-value for $\Lambda$ observed in the real data is estimated by \begin{equation} \hat{p} = \frac{1 + \sum_{m=1}^M \mathds{1}_{\Lambda^{(m)} \ge \Lambda}}{M+1}. \end{equation} Finally, the MLC is considered to be statistically significant if the associated $\hat{p}$ is less than the type I error. \\
\section{A simulation study} \label{sec:simulation} \noindent \noindent A simulation study was conducted to compare the performances of the parametric multivariate functional spatial scan statistic (PMFSS) $\Lambda_{\text{PMFSS}}$, the multivariate distribution-free functional spatial scan statistic (MDFFSS) $\Lambda_{\text{MDFFSS}}$ and the new multivariate rank-based functional spatial scan statistic (MRBFSS) $\Lambda_{\text{MRBFSS}}$. \cite{wilco_cucala} proposed a nonparametric scan statistic for univariate functional data (NPFSS) $\Lambda_{\text{NPFSS}}$. However according to the authors it can be extended to the multivariate functional framework although it has not been studied in this context. Thus we decided to include their approach in the simulation, using the computation improvement proposed by \cite{notre_fonctionnel}.
\subsection{Design of simulation study}
\noindent Artificial datasets were generated by using the geographic locations of the 94 French \textit{départements} (county-type administrative areas) as shown in Figure \ref{fig:sitesgrid}. The location of each \textit{département} was defined by its administrative capital. For each artificial dataset, a spatial cluster $w$ (composed of eight \textit{départements} in the Paris region ; the red area, see Figure \ref{fig:sitesgrid} in Supplementary materials) was simulated.
\subsubsection{Generation of the artificial datasets}
\noindent The $X_i$ were simulated according to the following model with $p = 2$ \citep[see][for more details]{two_s, article_milano}: \begin{center} $\text{for each }i \in \llbracket 1 ; 94 \rrbracket,\ X_i(t) = (\sin{[2\pi t^2]}^5 ; 1 + 2.3 t + 3.4 t^2 + 1.5 t^3)^\top + \Delta(t) \mathds{1}_{s_i \in w} + \varepsilon_{i}(t), \ t \in [0;1]. $ \end{center} where $\varepsilon_{i}(t) = \sum_{k=1}^{100} Z_{i,k} \sqrt{1.5 \times 0.2^k} \theta_k(t)$, with $\theta_k(t) = \left\{ \begin{array}{ll} 1 & \text{ if } k = 1 \\ \sqrt{2} \sin{[k\pi t]} & \text{ if } k \text{ even} \\ \sqrt{2} \cos{[(k-1)\pi t]} & \text{ if } k \text{ odd and } k > 1 \end{array} \right.$. \\
\noindent The functions $X_i$ were measured at 101 equally spaced times on $[0;1]$. \\
\noindent Giving $\Sigma = \begin{pmatrix} 1 & \rho \\ \rho & 1 \end{pmatrix}$ the covariance matrix of the $Z_{i,k}$, three distributions for the $Z_{i,k}$ were considered: (i) a normal distribution: $Z_{i,k} \sim \mathcal{N}(0,\Sigma)$, (ii) a standardized Student distribution: $Z_{i,k} = U_{i,k} \left( \frac{V_{i,4}}{4} \right)^{-0.5}$ where the $U_{i,k}$ are independent $\mathcal{N}(0, \Sigma/2)$ variables and the $V_{i,4}$ are independent $\chi_2(4)$ variables and (iii) a standardized chi-square distribution: $Z_{i,k} = \left[ U_{i,k} - \begin{pmatrix} 4 \\ 4 \end{pmatrix} \right] / (2\sqrt{2})$ where the $U_{i,k}$ are independent and $U_{i,k} \sim \chi_2(4,\Sigma) = \Gamma(2,1/2, \Sigma)$ (rate parameterization). Remark that $\rho$ is also the correlation of the two components of $X(t)$ for each time. \\
\noindent Three values of correlation $\rho$ were tested: $\rho = 0.2$, $0.5$ and $0.8$, and three types of clusters with intensity controlled by some parameter $\alpha > 0$ were studied: $\Delta_1(t) = \alpha (t ; t)^\top$, $\Delta_2(t) = \alpha ( t (1-t) ; t (1-t) )^\top$ and $\Delta_3(t) = \alpha (\exp{[ - 100 (t-0.5)^2 ]}/3 ; \exp{[ - 100 (t-0.5)^2 ]}/3)^\top$. Since they vary over time and are positive and non-zero on $\mathcal{T} = [0;1]$ (except possibly in $t = 0$ or $t = 1$), they correspond to both multivariate magnitude and multivariate shape clusters. \\ Different values of the parameter $\alpha$ were considered for each $\Delta$: $\alpha \in \{0 ; \ 0.375 ; \ 0.75 ; \ 1.125 ; \ 1.5 \}$ for $\Delta_1$, $\alpha \in \{0 ; \ 1 ; \ 2 ; \ 3 ; \ 4 \}$ for $\Delta_2$ and $\alpha \in \{ 0 ; \ 1.25 ; \ 2.5 ; \ 3.75 ; \ 5 \}$ for $\Delta_3$. Note that $\alpha = 0$ was also tested in order to evaluate the maintenance of the nominal type I error. An example of the data for $\rho = 0.2$ and for the Gaussian distribution for the $Z_{i,k}$ is given in the Appendix (Figure \ref{fig:examplesimu}).
\subsubsection{Comparison of the methods}
\noindent For each distribution of the $Z_{i,k}$, each type of $\Delta$, each level of correlation $\rho$, and each value of $\alpha$, 1000 artificial datasets were simulated. The type I error was set to 5\% and 999 samples were generated by random permutations of the data to evaluate the p-value associated with each MLC. The performances of the methods were compared through four criteria: the power, the true positive rate, the false positive rate and the F-measure. \\
\noindent The power was estimated by the proportion of simulations yielding to the rejection of $\mathcal{H}_0$ according to the type I error. Among the simulated datasets yielding to the rejection of $\mathcal{H}_0$, the true positive rate is the average proportion of sites correctly detected among the sites in $w$, the false positive rate is the average proportion of sites in $w^\mathsf{c}$ that were included in the detected cluster and the F-measure corresponds to the average harmonic mean of the proportion of sites in $w$ within the detected cluster (positive predictive value) and the true positive rate.
\subsubsection{Results of the simulation study}
\noindent The results of the simulation are presented in Figures \ref{fig:resultssimu1}, \ref{fig:resultssimu2} and \ref{fig:resultssimu3}. \\ \noindent For $\alpha = 0$, all methods seem to maintain the correct type I error of 0.05 regardless of the type of process, the type of $\Delta$ and the level of correlation $\rho$ (see the power curves in Figures \ref{fig:resultssimu1}, \ref{fig:resultssimu2} and \ref{fig:resultssimu3}). \\
\noindent For all methods the performances slightly decrease when the correlation $\rho$ increases. \\ \noindent The NPFSS and the PMFSS show similar powers for the Gaussian distribution for the shifts $\Delta_1$ and $\Delta_2$. However for non-Gaussian distributions of the $Z_{i,k}$ or the shift $\Delta_3$, the NPFSS presents higher powers than the PMFSS. The MDFFSS presents the highest powers in the Gaussian case. However its performances also decrease when the data are not distributed normally: in that case the MRBFSS shows the highest powers (except for $\Delta_2$ even if they are still very high). In the Gaussian case it also presents better powers than the NPFSS (except for $\Delta_2$) and the PMFSS. \\ \noindent The MRBFSS almost always shows the highest true positive rates (except sometimes for $\Delta_2$). The true positive rates for the MDFFSS are also very high for normal data but they decrease for non-normal data. The PMFSS presents the lowest true positive rates, however it presents very low false positive rates. In terms of false positives, the MDFFSS always shows the better performances. The MRBFSS often shows higher false positive rates, however they are lower than the ones of the NPFSS (except for $\Delta_2$ although both are very close). As a result the MDFFSS shows the highest F-measures, followed by the MRBFSS. For $\Delta_1$, in the Gaussian case the PMFSS and the NPFSS present similar F-measures whereas the F-measures are lower for the PMFSS for non-normal distributions. The NPFSS, the PMFSS and the MRBFSS shows very close F-measures for the shift $\Delta_2$ and finally the F-measures for the NPFSS and the PMFSS are strongly lower than the ones of the MDFFSS and the MRBFSS for the local shift $\Delta_3$.
\begin{figure}
\caption{The simulation study: comparison of the NPFSS, MDFFSS, MRBFSS and PMFSS methods for the shift $\Delta_1(t) = (\alpha t ; \alpha t)^\top$. For each method and each level of correlation $\rho$, the power curves, the true positive and false positive rates, and the F-measure values for detection of the spatial cluster as the MLC are shown. $\alpha$ is the parameter that controls the cluster intensity.}
\label{fig:resultssimu1}
\end{figure}
\begin{figure}
\caption{The simulation study: comparison of the NPFSS, MDFFSS, MRBFSS and PMFSS methods for the shift $\Delta_2(t) = (\alpha t (1-t) ; \alpha t (1-t))^\top$. For each method and each level of correlation $\rho$, the power curves, the true positive and false positive rates, and the F-measure values for detection of the spatial cluster as the MLC are shown. $\alpha$ is the parameter that controls the cluster intensity.}
\label{fig:resultssimu2}
\end{figure}
\begin{figure}
\caption{The simulation study: comparison of the NPFSS, MDFFSS, MRBFSS and PMFSS methods for the shift $\Delta_3(t) = (\alpha \exp{[ - 100 (t-0.5)^2 ]}/3 ; \alpha \exp{[ - 100 (t-0.5)^2 ]}/3)^\top$. For each method and each level of correlation $\rho$, the power curves, the true positive and false positive rates, and the F-measure values for detection of the spatial cluster as the MLC are shown. $\alpha$ is the parameter that controls the cluster intensity.}
\label{fig:resultssimu3}
\end{figure}
\section{Application on real data} \label{sec:realdata}
\subsection{Air pollution in \textit{Nord-Pas-de-Calais}}
\noindent The data considered is the concentration in $\mu g.m^{-3}$ of four pollutants: ozone ($\text{O}_3$), nitrogen dioxide ($\text{NO}_2$) and fine particles $\text{PM}_{10}$ and $\text{PM}_{2.5}$ corresponding respectively to particles whose diameter is less than $10 \mu m$ and $2.5 \mu m$. Note that the $\text{PM}_{2.5}$ particles are included in the $\text{PM}_{10}$ particles. The data provided by the French national air quality forecasting platform PREV'AIR consists in the daily average of these variables from May 1, 2020 to June 25, 2020 (56 values for each variable) aggregated at the \textit{canton} (administrative subdivisions of \textit{départements}) level for each of the 169 \textit{cantons} of the \textit{Nord-Pas-de-Calais} (a region in northern France) located by their center of gravity. The pollutants daily concentration curves in each \textit{canton} are presented in Figure \ref{fig:courbes_pollutions} (left panels) and the spatial distributions of the average concentrations for each pollutant over the studied time period are presented in Figure \ref{fig:courbes_pollutions} (right panels).
\begin{figure}
\caption{Daily concentration curves of $\text{NO}_2$, $\text{O}_3$, $\text{PM}_{10}$ and $\text{PM}_{2.5}$ (from May 1, 2020 to June 25, 2020) in each of the 169 \textit{cantons} of \textit{Nord-Pas-de-Calais} (a region in northern France) (left panels), and the spatial distributions of the average concentrations for each pollutant over period from from May 1, 2020 to June 25, 2020 (right panels).}
\label{fig:courbes_pollutions}
\end{figure}
\noindent The maps in Figure \ref{fig:courbes_pollutions} show a spatial heterogeneity of the average concentration for each pollutant. High concentrations of $\text{O}_3$ tend to aggregate in the rural areas of Montreuil and Avesnes-sur-Helpe, and high concentrations of the other pollutants tend to aggregate in the urban areas of Calais, Dunkerque and Lille. Moreover the daily concentration curves present a marked temporal variability during the period from May 1, 2020 to June 25, 2020. Thus functional spatial scan statistics seem relevant to highlight the presence of \textit{cantons}-level spatial clusters of pollutants concentrations.
\subsection{Spatial clusters detection} \noindent For the sake of concision, we have decided to present here only one method based on the results of the simulation. With regard to the latter, we have chosen the MRBFSS because it presents stable performances whatever the correlation and the distribution of the variables.
\noindent We considered a round-shaped scanning window of maximum radius 10 km since small clusters of pollution are more relevant for interpretation because the sources of the pollutants are very localized: the main source of $\text{NO}_2$ is road traffic, and $\text{PM}_{2.5}$ is mainly emitted in urban (heating, road traffic) or industrial areas. The statistical significance of the MLC was evaluated through 999 Monte-Carlo permutations and the MLC is said to be statistically significant if its p-value is less than 0.05. Cluster detection was also performed for the other three methods, the results are presented in Figure \ref{fig:othermethods} and Table \ref{table:noconstraint} in Supplementary materials.
\subsection{Results}
\noindent The MRBFSS detected a significant most likely cluster (15 \textit{cantons}, $308\text{km}^2$, $\hat{p} = 0.001$) in the area of Lille. This cluster, corresponding to high values of $\text{NO}_2$, $\text{PM}_{10}$ and $\text{PM}_{2.5}$ concentrations, is presented in Figure \ref{fig:noconstraint}. For these three pollutants all the curves in the MLC are above the average concentrations in the \textit{Nord-Pas-de-Calais}. In environmental science it is well-known that those pollutants are more frequent in urban areas. Therefore this is consistent with the cluster observed here.
\begin{figure}
\caption{Most likely cluster of pollutants ($\text{NO}_2$, $\text{O}_3$, $\text{PM}_{10}$ and $\text{PM}_{2.5}$) concentrations detected by the MRBFSS. The daily concentration curves of the pollutants (from May 1, 2020 to June 25, 2020) in each \textit{canton} are presented with colored lines. The black curves are the daily average concentration curves in the \textit{Nord-Pas-de-Calais} (a region in northern France).}
\label{fig:noconstraint}
\end{figure}
\section{Discussion} \label{sec:discussion}
\noindent Here we developed a parametric multivariate functional scan statistic (PMFSS), a multivariate distribution-free functional spatial scan statistic (MDFFSS) and a multivariate rank-based functional spatial scan statistic (MRBFSS) which allow to detect clusters of abnormal values on multivariate functional data indexed in space. The goal of such methods is to alert the scientists if abnormal values are detected. Typically in the environmental-surveillance context they will generate an alarm if they detect areas where populations are multi-exposed to environmental pollutants. The new methods appear to be more relevant for multivariate functional data than a multivariate spatial scan statistic approach since the latter would face huge losses of information by summarizing each variable of the data by its average over the time period. Furthermore they also appear to be more relevant than using a univariate functional spatial scan statistic for each variable since this does not allow to take into account the correlations between the variables. \\
\noindent Although they only studied their approach in the univariate functional framework, \cite{wilco_cucala} suggest that the NPFSS can be extended to the multivariate case. Thus the MDFFSS, the PMFSS and the MRBFSS were compared with the NPFSS in a simulation study. The simulation study highlighted that the performances of all methods decreased with increasing correlation between the variables. The MRBFSS and the MDFFSS presented higher powers than the PMFSS and the NPFSS whatever the distribution and the correlation between the variables. The PMFSS and the MDFFSS showed the lowest false positive rates. However the MRBFSS presented the highest true positive rates and the PMFSS showed the lowest ones which results in very high F-measures for the MRBFSS and the MDFFSS, which improves the confidence in the clusters detected by these approaches compared to the ones detected by the NPFSS and the PMFSS. Moreover the ones detected by the NPFSS tended to contain more false positives which is less relevant in practice. Indeed in the case of the application of scan statistics to environmental surveillance, having fewer false positives rates is an advantage since the detection of spatial clusters is the starting point for future investigation by environmental experts within the cluster. \\ When the data were far from being normally distributed the performances of the PMFSS decreased as well as the ones of the MDFFSS. However they still maintain very low false positive rates. \\
\noindent For the sake of brevity we have chosen to apply only the MRBFSS to detect clusters of abnormal values of pollutants concentrations in the \textit{cantons} of \textit{Nord-Pas-de-Calais}, based on the results of the simulation since it shows stable performances whatever the distribution of the variables and the correlation. The method detected a significant most likely cluster in the area of Lille which presents high values of $\text{NO}_2$, $\text{PM}_{10}$ and $\text{PM}_{2.5}$ concentrations. \\
\noindent Remark that we only focused on circular clusters. In the application the maps of pollutants present elongated shapes of high average concentrations especially for $\text{PM}_{10}$ on the coastline, which suggests that other forms of clusters may be relevant to consider in the analysis. As an example \cite{tango} proposed to consider irregularly shaped clusters by considering all sets of sites connected to each other, with a predetermined maximum size. However it should be noted that using this approach generates many more potential clusters than the approach proposed by \cite{spatialscanstat} which drastically increases the computing time. The same disadvantage can be found with the elliptic clusters approach of \citep{elliptic}. However these problems can be overcome with the graph-based clusters proposed by \citep{cucala_graph}. Another possible approach was proposed by \cite{lin} who suggests to regroup the estimated circular clusters to form clusters with arbitrary shapes. \\
\noindent It should also be noted that the application on real data only considered the MLC. It may also be interesting to detect secondary clusters, which can be done by following the approach of \cite{spatialscanstat} who considers also clusters that had a high value of the concentration index ($\mathrm{LH}^{(w)}$ for the PMFSS, $T_{n, \text{max}}^{(w)}$ for the MDFFSS, $W^{(w)}$ for the MRBFSS, and $U(w)$ for NPFSS \citep[see][for details]{wilco_cucala}) and did not cover the MLC. \\
\noindent Finally in the context of spatial epidemiology, one could imagine case count data collected monthly on spatial units over a long period of time. In this context, to detect spatial clusters of diseases with the already existing methods, we often use cumulative incidences, which implies to get only one data per spatial unit. This induces a huge loss of information, particularly when the incidence curves show high temporal variability. However we should underline that the NPFSS, the PMFSS, the MDFFSS and the MRBFSS could be applied to count data including the possibility to adjust the analysis for the underlying population.
\appendix
\section{Supplementary materials} \begin{figure}
\caption{The 94 French \textit{départements} and the spatial cluster (in red) simulated for each artificial dataset.}
\label{fig:sitesgrid}
\end{figure}
\noindent Figure \ref{fig:examplesimu} shows an example of the data generated when the $Z_{i,k}$ are Gaussian and $\rho = 0.2$.
\begin{figure}
\caption{The simulation study: an example of the two components of the data generated for the Gaussian process and $\rho = 0.2$, with $\Delta(t) = \Delta_1(t) = 1.5(t ; t)^\top$ (left panel), $\Delta(t) = \Delta_2(t) = 4(t(1-t) ; t(1-t))^\top$ (middle panel) and $\Delta(t) = \Delta_3(t) = 5( \exp{[ - 100 (t-0.5)^2 ]}/3 ; \exp{[ - 100 (t-0.5)^2 ]}/3)^\top$ (right panel).
The red curves correspond to the observations in the cluster.}
\label{fig:examplesimu}
\end{figure}
\begin{figure}
\caption{Most likely clusters of pollutants ($\text{NO}_2$, $\text{O}_3$, $\text{PM}_{10}$ and $\text{PM}_{2.5}$) concentrations detected by the NPFSS, the PMFSS and the MDFFSS. The daily concentration curves of the pollutants (from May 1, 2020 to June 25, 2020) in each \textit{canton} are presented with colored lines. The black curves are the daily average concentration curves in the \textit{Nord-Pas-de-Calais} (a region in northern France).}
\label{fig:othermethods}
\end{figure}
\begin{table}[htbp] \centering \caption{Description of the most likely cluster of pollutants concentrations detected for the NPFSS, the PMFSS and the MDFFSS . }
\begin{tabular}{lccc} \hline & \multirow{2}{*}{\# \textit{cantons}} & \multirow{2}{*}{Surface} & \multirow{2}{*}{p-value} \\ & & & \\ \hline \multirow{2}{*}{\textbf{$\Lambda_{\text{NPFSS}}$}} & \multirow{2}{*}{15} & \multirow{2}{*}{308 km$^2$} & \multirow{2}{*}{0.001} \\
\multirow{2}{*}{\textbf{$\Lambda_{\text{PMFSS}}$}} & \multirow{2}{*}{13} & \multirow{2}{*}{264 km$^2$} & \multirow{2}{*}{0.001} \\
\multirow{2}{*}{\textbf{$\Lambda_{\text{MDFFSS}}$}} & \multirow{2}{*}{7} & \multirow{2}{*}{284 km$^2$} & \multirow{2}{*}{0.001} \\
& & & \\ \hline \end{tabular}
\label{table:noconstraint} \end{table}
\noindent Figure \ref{fig:othermethods} presents the most likely clusters obtained with the three other methods. The NPFSS detects exactly the same cluster as the MRBFSS and the most likely cluster for the PMFSS is quite similar to it. The result obtained with the MDFFSS seems at first sight quite surprising. However we only focussed here on the most likely clusters and it was found that this cluster is significant ($\hat{p} = 0.001$) for all methods and that the secondary cluster for the MDFFSS ($\hat{p} = 0.001$) is exactly the MLC of the NPFSS and the MRBFSS. Some characteristics of the detected MLCs are presented in Table \ref{table:noconstraint}.
\end{document}
|
arXiv
|
{
"id": "2103.14401.tex",
"language_detection_score": 0.795237123966217,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{On the local fluctuations of last-passage percolation models} \author{Eric Cator and Leandro P. R. Pimentel} \date{\today} \maketitle
\begin{abstract} Using the fact that the Airy process describes the limiting fluctuations of the Hammersley last-passage percolation model, we prove that it behaves locally like a Brownian motion. Our method is quite straightforward, and it is based on a certain monotonicity and good control over the equilibrium measures of the Hammersley model (local comparison). \end{abstract}
\section{Introduction and results}
In recent years there has been a lot of research on the Airy process and related processes such as the Airy sheet \cite{CQ}. Most papers use analytic methods and exact formulas given by Fredholm determinants to prove properties of these processes, but some papers use the fact that these processes are limiting processes of last-passage percolation models or random polymer models, and they use properties of these well studied models to prove the corresponding property of the limiting process. A nice example of these different approaches can be found in two recent papers, one by H\"agg \cite{H} and one by Corwin and Hammond \cite{CH}. H\"agg proves in his paper that the Airy process behaves locally like a Brownian motion, at least in terms of convergence of finite dimensional distributions. He uses the Fredholm determinant description of the Airy process to obtain his result. Corwin and Hammond on the other hand, use the fact that the Airy line process, of which the top line corresponds to the Airy process, can be seen as a limit of Brownian bridges, conditioned to be non-intersecting. They show that a particular resampling procedure, which they call the Brownian Gibbs property, holds for the system of Brownian bridges and also in the limit for the Airy line process. As a consequence, it follows that the local Brownian behavior of the Airy process holds in a stronger functional limit sense. Our paper will prove the same theorem, also using the fact that the Airy process is a limiting process, but in a much more direct way: we will consider the Hammersley last-passage percolation model, and show that we can control local fluctuations of this model by precisely chosen equilibrium versions of this model, which are simply Poisson processes. Then we show that in the limit this control suffices to prove the local Brownian motion behavior of the Airy process, as well as tightness of the models approaching the Airy process. We also extend the control of local fluctuations of the Hammersley process to scales smaller than the typical cube-root scale.
Our method is quite straightforward, yet rather powerful, mainly because we have a certain monotonicity and good control over the equilibrium measures. In fact, we think that we can extend our result to the more illustrious Airy sheet, the two dimensional version of the Airy process. We address the reader to \cite{CQ}, for a description of this process in terms of the renormalization fixed point of the KPZ universality class. However, here we run in to much more technical problems, and this will still require a lot more work, beyond the scope of this paper.
We will continue the introduction by developing notation, introducing all relevant processes and stating the three main theorems. In Section 2 we introduce the local comparison technique and in each of the following three sections one theorem is proved.
\subsection{The Hammersley Last Passage Percolation model}
The Hammersley last-passage percolation model \cite{AD} is constructed from a two-dimensional homogeneous Poisson point process of intensity $1$. Denote $[x]_t:=(x,t)\in{\mathbb R}^2$ and call a sequence $[x_1]_{t_1},[x_2]_{t_2},\dots,[x_k]_{t_k}$ of planar points increasing if $x_j<x_{j+1}$ and $t_j<t_{j+1}$ for all $j=1,\dots,k-1$. The last-passage time $L([x]_s,[y]_t)$ between $[x]_s<[y]_t$ is the maximal number of Poisson points among all increasing sequences of Poisson points lying in the rectangle $(x, y ]\times(s, t]$. Denote $L[x]_t:=L([0]_0,[x]_t)$ and define ${\mathcal A}_n$ by $$u\in{\mathbb R}\,\mapsto\,{\mathcal A}_n(u):=\frac{L[n+2un^{2/3}]_n-(2n+2un^{2/3})+u^2n^{1/3}}{n^{1/3}}\,.$$ Pr\"ahofer and Spohn \cite{PS} proved that \begin{equation}\label{eq:Airy} \lim_{n\to\infty}{\mathcal A}_n(\cdot)\stackrel{dist.}{=}{\mathcal A}(\cdot)\,, \end{equation} in the sense of finite dimensional distributions, where ${\mathcal A}\equiv({\mathcal A}(u))_{u\in{\mathbb R}}$ is the so-called Airy process. This process is a one-dimensional stationary process with continuous paths and finite dimensional distributions given by a Fredholm determinant \cite{J}: $${\mathbb P}\left({\mathcal A}(u_1)\leq \xi_1,\dots,{\mathcal A}(u_m)\leq \xi_m\right):=\det\left(I-f^{1/2}Af^{1/2}\right)_{L^2\left(\{u_1,\dots,u_m\}\times{\mathbb R}\right)}\,.$$ The function $A$ denotes the extended Airy kernel, which is defined as $$A_{s,t}(x,y):=\left\{\begin{array}{ll}\int_0^\infty e^{-z(s-t)}{\rm Ai}(x+z){\rm Ai}(y+z)dz\,,& \mbox{ if } s\geq t\,,\\&\\ -\int_{-\infty}^0 e^{-z(t-s)}{\rm Ai}(x+z){\rm Ai}(y+z)dz\,,& \mbox{ if } s< t\,,\end{array}\right.$$ where ${\rm Ai}$ is the Airy function, and for $\xi_1,\dots,\xi_m\in{\mathbb R}$ and $u_1<\dots<u_m$ in ${\mathbb R}$, \begin{eqnarray*} f\,:\,\{u_1,\dots,u_m\}\times{\mathbb R}&\to&{\mathbb R}\\ (u_i,x)&\mapsto&1_{(\xi_i,\infty)}(x)\,. \end{eqnarray*}
The main contribution of this paper is the development of a local comparison technique to study the local fluctuations of last-passage times and its scaling limit. The ideas parallel the work of Cator and Groeneboom \cite{CG1,CG2}, where they studied local convergence to equilibrium and the cube-root asymptotic behavior of $L$. This technique consists of bounding from below and from above the local differences of $L$ by the local differences of the equilibrium regime (Lemma \ref{lem:LocalComparison}), with suitable parameters that will allow us to handle the local fluctuations in the right scale (Lemma \ref{lem:ExitControl}). For the Hammersley model the equilibrium regime is given by a Poisson process. We have strong indications that the technique can be applied to a broad class of models, as soon as one has Gaussian fluctuations for the equilibrium regime. Although this is a very natural assumption, one can only check that for a few models. As a first application, we will prove tightness of ${\mathcal A}_n$.
\begin{thm}\label{thm:Tight} The collection $\{{\mathcal A}_n\}$ is tight in the space of cadlag functions on $[a,b]$. Furthermore, any weak limit of ${\mathcal A}_n$ lives on the space of continuous functions. \end{thm}
The local comparison technique can be used to study local fluctuations of last-passage times for lengths of size $n^{\gamma}$, with $\gamma\in(0,2/3)$ (so smaller than the typical scale $n^{2/3}$). Let ${\mathcal B}\equiv({\mathcal B}(u))_{u\geq 0}$ denote the standard two-sided Brownian motion process. \begin{thm}\label{thm:LocalFluct} Fix $\gamma\in(0,2/3)$ and $s>0$ and define $\Delta_n$ by $$u\in{\mathbb R}\,\mapsto\, \Delta_n(u):=\frac{L[sn+un^{\gamma}]_{n}-L[sn]_n-\mu un^{\gamma}}{\sigma n^{\gamma/2}}\,,$$ where $\mu:=s^{-1/2}$ and $\sigma:=s^{-1/4}$. Then $$\lim_{n\to\infty}\Delta_n(\cdot)\stackrel{dist.}{=}{\mathcal B}(\cdot)\,,$$ in the sense of weak convergence of probability measures in the space of cadlag functions. \end{thm}
As we mentioned in the previous section, the Airy process locally behaves like Brownian motion \cite{CH,H}. By applying the local comparison technique again, we will present an alternative proof of the functional limit theorem for this local behavior. \begin{thm}\label{thm:LocalAiry} Define ${\mathcal A}^\epsilon$ by $$u\in{\mathbb R}\,\mapsto\,{\mathcal A}^{\epsilon}(u):=\epsilon^{-1/2}\left({\mathcal A}(\epsilon u)-{\mathcal A}(0)\right)\,.$$ Then $$\lim_{\epsilon\to 0}{\mathcal A}^{\epsilon}(\cdot)\stackrel{dist.}{=}\sqrt{2}{\mathcal B}(\cdot)\,,$$ in the sense of weak convergence of probability measures in the space of continuous functions. \end{thm}
\subsection{The lattice model with exponential weights} Consider a collection $\{\omega_{[x]_t}\,:\,[x]_t\in{\mathbf Z}^2\}$ of i.i.d. non negative random variables with an exponential distribution of parameter one. Let $\Pi([x]_t,[y]_u)$ denote the collection of all lattice paths $\varpi=([z]_{v_j})_{j=1,\dots,k}$ such that: \begin{itemize} \item $[z]_{v_1}\in\{[x]_t+[1]_0,[x]_t+[0]_1\}$ and $[z]_{v_k}=[y]_u$; \item $[z]_{v_{j+1}}-[z]_{v_j}\in\{[1]_0,[0]_1\}$ for $j=0,1\dots,,k-1$. \end{itemize} The (lattice) last-passage percolation time between $[x]_t <[y]_u$ is defined by $$L^l([x]_t,[y]_u):=\max_{\varpi\in\Pi([x]_t,[y]_u)}\big\{\sum_{[z]_v\in\varpi}\omega_{[z]_v}\big\}\,.$$ Denote $L^l[x]_t:=L^l([0]_0,[x]_t)$ and define ${\mathcal A}^l_n$ by $$u\in{\mathbb R}\,\mapsto\,{\mathcal A}^l_n(u):=\frac{L^l[n+2^{5/3}un^{2/3}]_n-(4n+2^{8/3}un^{2/3})+2^{4/3}u^2n^{1/3}}{2^{4/3}n^{1/3}}\,.$$ Corwin, Ferrari and P\'ech\'e \cite{CFP} proved that \begin{equation}\label{eq:Airy} \lim_{n\to\infty}{\mathcal A}^l_n(\cdot)\stackrel{dist.}{=}{\mathcal A}(\cdot)\,, \end{equation} in the sense of finite dimensional distributions. The local comparison method can be used in this context as well. (The lattice version of Lemma \ref{lem:LocalComparison} is straightforward. For exponential weights, the analog to Lemma \ref{lem:ExitControl} was proved in \cite{BCS}.)
\begin{thm}\label{thm:LTight} The collection $\{{\mathcal A}^l_n\}$ is tight in the space of cadlag functions on $[a,b]$. Furthermore, any weak limit of ${\mathcal A}_n$ lives on the space of continuous functions. \end{thm}
\begin{thm}\label{thm:LLocalFluct} Fix $\gamma\in(0,2/3)$ and $s>0$ and define $\Delta_n$ by $$u\in{\mathbb R}\,\mapsto\, \Delta^l_n(u):=\frac{L^l[sn+un^{\gamma}]_{n}-L^l[sn]_n-\mu un^{\gamma}}{\sigma n^{\gamma/2}}\,,$$ where $\mu=\sigma:=s^{-1/2}(1+s^{1/2})$. Then $$\lim_{n\to\infty}\Delta^l_n(\cdot)\stackrel{dist.}{=}{\mathcal B}(\cdot)\,,$$ in the sense of weak convergence of probability measures in the space of cadlag functions. \end{thm}
To avoid repetitions, we will not present a proof of the lattice results. We hope that the reader can convince his (or her) self that the method that we will describe in detail for the Hammersley last-passage percolation model can be easily adapted to the lattice models with exponentials weights.
\begin{rem} We also expect that the local comparison method can be used in the log-gamma polymer model, introduced by Sepp\"al\"ainen \cite{T}. The polymer versions of Lemma \ref{lem:LocalComparison} and Lemma \ref{lem:ExitControl} were proved in \cite{T}. \end{rem}
\section{Local comparison and exit points}
The Hammersley last-passage percolation model has a representation as an interacting particle system, called the Hammersley process \cite{AD,CG1}. We will use notations used in \cite{CP}. In the positive time half plane we have the same planar Poisson point process as before. On the $x$-axis we take a Poisson process of intensity $\lambda>0$. The two Poisson process are assumed to be independent of each other. For $x\in {\mathbb R}$ and $t>0$ we define $$L_{{\lambda}}[x]_t\equiv L_{\nu_\lambda}[x]_t:=\sup_{z\in (-\infty,x]} \left\{ \nu_{\lambda}(z) + L([z]_0,[x]_t)\right\}\,,$$ where, for $z\leq x$, $$\nu_{\lambda}(z)=\left\{\begin{array}{ll}\mbox{ the number of Poisson points in }(0,z]\times\{0\}& \mbox{ for } z> 0\\ \mbox{ minus the number of Poisson points in }(z,0]\times \{0\} & \mbox{ for } z\leq0\,.\end{array}\right.$$ The process $(M^t_\lambda)_{t\geq 0}$, given by $$M^t_{\nu_{\lambda}}(x,y]\equiv M^t_{\lambda}(x,y]:=L_{\lambda}[y]_t-L_{\lambda}[x]_t\,\,\,\mbox{ for }x<y\,,$$ is a Markov process on the space of locally finite counting measures on ${\mathbb R}$. The Poisson process is the invariant measure of this particle system in the sense that \begin{equation}\label{eq:equilibrium} M^t_{\lambda}\stackrel{dist.}{=}\mbox{ Poisson process of intensity $\lambda$ for all $t\geq 0$}\,. \end{equation} Notice that the last-passage time $L=L_{\nu_0}$ can be recovered in the positive quadrant by choosing a measure $\nu_0$ on the axis that has no points to the right of $0$, and an infinite amount of points in every interval $(-{\varepsilon },0)$, $\forall {\varepsilon }>0$ (this could be called a ``wall'' of points). Thus $$L_{\lambda}\equiv L_{\nu_\lambda}({\cal P})\,\mbox{ and }\, L\equiv L_{\nu_0}({\cal P})\,$$ are coupled by the same two-dimensional Poisson point process $\cal P$, which corresponds to the basic coupling between $M^t_{\nu_\lambda}$ and $M^t_{\nu_0}$.
Define the exit points $$Z_{\lambda}[x]_t :=\sup\left\{z\in(-\infty,x]\,:\,L_{\lambda}[x]_t=\nu_\lambda(z)+L([z]_0,[x]_t)\right\}\,,$$ and $$Z'_{\lambda}[x]_t :=\inf\left\{z\in(-\infty,x]\,:\,L_{\lambda}[x]_t=\nu_\lambda(z)+L([z]_0,[x]_t)\right\}\,.$$ By using translation invariance and invariance under the map $(x,t)\mapsto(\lambda x,t/\lambda)$, we have that \begin{equation}\label{eq:sym} Z_\lambda[x+h]_t\stackrel{dist.}{=}Z_\lambda[x]_t+h\,\,\,\mbox{ and }\,\,\, Z_{\lambda}[x]_t\stackrel{dist.}{=}\lambda Z_1[\lambda x]_{t/\lambda}. \end{equation} We need to use one more symmetry. In \cite{CG1}, the Hammersley process was set up as a process in the first quadrant with sources on the $x$-axis and sinks on the $t$-axis. In our notation this means that the process $t\mapsto L_\lambda[0]_t$ is a Poisson process with intensity $1/\lambda$ which is independent of the Poisson process $\nu_\lambda$ restricted to the positive $x$-axis and independent of the Poisson process in the positive quadrant. We can now use reflection in the diagonal to see that the following equality holds: \begin{equation}\label{eq:diagsym} {\mathbb P}\left(Z'_\lambda[x]_t < 0\right) = {\mathbb P}\left(Z_{1/\lambda}[t]_x > 0\right). \end{equation} We use that $Z'_\lambda[x]_t<0$ is equivalent to the fact that the maximizing path to the left-most exit point crosses the positive $t$-axis, and not the positive $x$-axis.
The local comparison technique consists of bounding from above and from below the local differences of $L$ by the local differences of $L_\lambda$. These bounds depend on the position of the exit points. It is precisely summarized by the following lemma. \begin{lem}\label{lem:LocalComparison} Let $0\leq x\leq y$ and $t\geq 0$. If $Z'_{\lambda}[x]_t\geq0$ then $$L[y]_t-L[x]_t\leq L_{\lambda}[y]_t-L_{\lambda}[x]_t\,,$$ and if $Z_{\lambda}[y]_t\leq 0$ then $$L[y]_t-L[x]_t\geq L_{\lambda}[y]_t-L_{\lambda}[x]_t\,.$$ \end{lem}
\noindent{\bf Proof\,\,} When we consider a path $\varpi$ from $[x]_s$ to $[y]_t$ consisting of increasing points, we will view $\varpi$ as the lowest increasing continuous path connecting all the points, starting at $[x]_s$ and ending at $[y]_t$. In this way we can talk about crossings with other paths or with lines. The geodesic between $[x]_s$ and $[y]_t$ is given by the lowest path (in the sense we just described) that attains the maximum in the definition of $L([x]_s,[y]_t)$. We will denote this geodesic by $\varpi([x]_s,[y]_t)$. Notice that $$L([x]_s,[y]_t)=L([x]_s,[z]_r)+L([z]_r,[y]_t)\,,$$ for any $[z]_r\in\varpi([x]_s,[y]_t)$.
Assume that $Z'_{\lambda}[x]_t\geq 0$ and let ${\mathbf c}$ be a crossing between the two geodesics $\varpi([0]_0,[y]_t)$ and $\varpi([z']_0,[x]_t)$, where $z':=Z'_{\lambda}[x]_t$. Such a crossing always exists because $x\leq y$ and $z'=Z'_{\lambda}[x]_t\geq 0$. We remark that, by superaddivity, $$L_{\lambda}[y]_t \geq \nu_{\lambda}(z') + L([z']_0,[y]_t) \geq \nu_{\lambda}(z') + L([z']_0,{\mathbf c}) + L({\mathbf c},[y]_t)\,.$$ We use this, and that (since ${\mathbf c}\in\varpi([z']_0,[x]_t)$) $$ \nu_{\lambda}(z') + L([z']_0,{\mathbf c})-L_{\lambda} [x]_t= -L({\mathbf c},[x]_t)\,,$$ in the following inequality: \begin{eqnarray*} L_{\lambda}[y]_t - L_{\lambda}[x]_t & \geq & \nu_{\lambda}\big(z'\big)+L([z']_0,{\mathbf c}) + L({\mathbf c},[y]_t) - L_{\lambda}[x]_t\\ & = & L({\mathbf c},[y]_t) - L({\mathbf c},[x]_t)\,. \end{eqnarray*} By superaddivity, $$ - L({\mathbf c}\,,\,[x]_t)\geq L([0]_0,{\mathbf c})-L[x]_t\,,$$ and hence (since ${\mathbf c}\in\varpi([0]_0,[y]_t)$) \begin{eqnarray*} L_{\lambda}[y]_t - L_{\lambda}[x]_t & \geq & L({\mathbf c},[y]_t) - L({\mathbf c},[x]_t)\\ & \geq & L({\mathbf c},[y]_t) + L([0]_0,{\mathbf c})-L([0]_0,[x]_t)\\ & = & L[y]_t-L[x]_t\,. \end{eqnarray*} The proof of the second inequality is very similar. Indeed, denote $z:=Z_{\lambda}[y]_t$ and let ${\mathbf c}$ be a crossing between $\varpi([0]_0,[x]_t)$ and $\varpi([z]_0,[y]_t)$. By superaddivity, $$L_{\lambda}[x]_t \geq \nu_{\lambda}(z) + L([z]_0,[x]_t) \geq \nu_{\lambda}(z) + L([z]_0,{\mathbf c}) + L({\mathbf c},[x]_t)\,.$$ Since ${\mathbf c}\in\varpi([z]_0,[y]_t)$ we have that $$L_{\lambda}[y]_t-\nu_{\lambda}(z) - L([z]_0,{\mathbf c})=L({\mathbf c},[y]_t)\,,$$ which implies that \begin{eqnarray*} L_{\lambda}[y]_t - L_{\lambda}[x]_t & \leq &L_{\lambda}[y]_t- \nu_{\lambda}(z)-L([z]_0,{\mathbf c}) - L({\mathbf c},[x]_t)\\ & = & L({\mathbf c},[y]_t) - L({\mathbf c},[x]_t)\\ & \leq & L[y]_t-L([0]_0,{\mathbf c})- L({\mathbf c},[x]_t)\\ & = & L[y]_t-L[x]_t\,, \end{eqnarray*} where we have used that ${\mathbf c}\in\varpi([0]_0,[x]_t)$ in the last step.
$\Box$\\
\begin{rem} In fact the first statement of the lemma is also true when $Z_\lambda[x]_t\geq 0$ and the second statement is true when $Z'_\lambda[y]_t\leq 0$, both without any change to the given proof. This is a stronger statement, since $Z'_\lambda[x]_t\leq Z_\lambda[x]_t$, but we will only need the lemma as it is formulated. \end{rem}
In order to apply Lemma \ref{lem:LocalComparison} and extract good bounds for the local differences one needs to control the position of exit points. This is given by the next lemma. \begin{lem}\label{lem:ExitControl} There exist constant $C>0$ such that, $${\mathbb P}\left(Z_1[n]_n > r n^{2/3}\right)\leq \frac{C}{r^3}\,,$$ for all $r\geq 1$ and all $n\geq 1$. \end{lem}
\noindent{\bf Proof\,\,} See Corollary 4.4 in \cite{CG2}.
$\Box$\\
\section{Proof of Theorem \ref{thm:Tight}} For simple notation, and without loss of generality, we will restrict our proof to $[a,b]=[0,1]$. \begin{lem}\label{lem:tight} Fix $\beta\in(1/3,1)$ and for each $\delta\in(0,1)$ and $n\geq 1$ set $$\lambda_\pm=\lambda_\pm(n,\delta):=1\pm\frac{\delta^{-\beta}}{n^{1/3}}\,.$$ Define the event $$E_n(\delta):=\left\{Z'_{\lambda_{+}}[n]_n\geq 0\,\,\mbox{ and }\,\,Z_{\lambda_{-}}[n+2n^{2/3}]_n\leq 0\right\}\,.$$ Then there exists a constant $C>0$ such that, for sufficiently small $\delta>0$, $$\limsup_{n\to\infty}{\mathbb P}\left(E_n(\delta)^c\right)\leq C\delta^{3\beta}\,.$$ \end{lem}
\noindent{\bf Proof\,\,} Denote $r:=\delta^{-\beta}$ and let $$n_+:=\lambda_+ n< 2n\,\,\mbox{ and }\,\,h_{+}:=\left(\lambda_+-\frac{1}{\lambda_+}\right) n> rn^{2/3}> r n_+^{2/3}/2 \,$$ (for all sufficiently large $n$). By \eqref{eq:sym} and \eqref{eq:diagsym}, \begin{eqnarray*} {\mathbb P}\left(Z'_{\lambda_+}[n]_n<0\right)&=&{\mathbb P}\left(Z_{1/\lambda_+}[n]_n>0\right)\\ &=&{\mathbb P}\left(Z_1[n/\lambda_+]_{\lambda_+ n}>0\right)\\ &=&{\mathbb P}\left(Z_1[\lambda_+n-h_+]_{\lambda_+ n}>0\right)\\ &=&{\mathbb P}\left(Z_1[\lambda_+n]_{\lambda_+ n}>h_+\right)\\ &\leq &{\mathbb P}\left(Z_1[n_+]_{n_+}>r n_+^{2/3}/2\right)\,. \end{eqnarray*} Analogously, for $$n_-:= \frac{n}{\lambda_-}<2n\,\,\mbox{ and }\,\,h_{-}:=\left(\frac{1}{\lambda_-}-\lambda_-\right) n> rn^{2/3}>r n_-^{2/3}/2 \,,$$ we have that \begin{eqnarray*} {\mathbb P}\left(Z_{\lambda_-}[n+2 n^{2/3}]_n>0\right)&=&{\mathbb P}\left(Z_{\lambda_-}[n]_n>-2 n^{2/3}\right)\\ &=&{\mathbb P}\left(\lambda_-Z_{1}[\lambda_-n]_{n/\lambda_-}>-2 n^{2/3}\right)\\ &\leq&{\mathbb P}\left(Z_{1}[n_- - h_-]_{n/\lambda_-}>-2 n^{2/3}\right)\\ &=&{\mathbb P}\left(Z_{1}[n_-]_{n_-}>h_- -2 n^{2/3}\right)\\ &\leq& {\mathbb P}\left(Z_1[n_-]_{n_-}>(r-4) n_-^{2/3}/2\right)\,. \end{eqnarray*} Now one can use Lemma \ref{lem:ExitControl} to finish the proof.
$\Box$\\
\begin{lem}\label{lem:compa} Let $\delta\in(0,1)$ and $u\in[0,1-\delta)$. Then, on the event $E_n(\delta)$, for all $v\in[u,u+\delta]$ we have that $$ {\mathcal B}_{n,-}(v)-{\mathcal B}_{n,-}(u)-2\delta^{1-\beta}\leq{\mathcal A}_n(v)-{\mathcal A}_n(u)\,\leq\, {\mathcal B}_{n,+}(v)-{\mathcal B}_{n,+}(u)+4\delta^{1-\beta}\,,$$ where $${\mathcal B}_{n,\pm}(u):\frac{L_{\lambda\pm}[n+2un^{2/3}]_n-L_{\lambda\pm}[n]_n-\lambda_\pm 2un^{2/3}}{n^{1/3}}\,.$$ \end{lem}
\noindent{\bf Proof\,\,} For fixed $t$, $Z_\lambda'[x]_t$ and $Z_\lambda[x]_t$ are non-decreasing functions of $x$. Thus, on the event $E_n(\delta)$, $$Z'_{\lambda_+}[n+2un^{2/3}]_n\geq 0\,\mbox{ and }\,Z_{\lambda_-}[n+2(u+\delta)n^{2/3}]_n\leq 0\,.$$ By Lemma \ref{lem:LocalComparison}, this implies that, for all $v\in[u,u+\delta]$, $$L[n+vn^{2/3}]_{n}-L[n+un^{2/3}]_n\leq L_{\lambda_+}[n+vn^{2/3}]_{n}-L_{\lambda_+}[n+un^{2/3}]_n\,,$$ and $$L[n+vn^{2/3}]_{n}-L[n+un^{2/3}]_n\geq L_{\lambda_-}[n+vn^{2/3}]_{n}-L_{\lambda_-}[n+un^{2/3}]_n\,.$$ Since $$(\lambda_+-1)(2v-2u)n^{1/3}+v^2-u^2\leq 2\delta^{1-\beta}+2\delta\leq 4\delta^{1-\beta}\,,$$ and $$(\lambda_- -1)(2v-2u)n^{1/3}+v^2-u^2\geq -2\delta^{1-\beta}\,,$$ we have that, on the event $E_n(\delta)$, $${\mathcal A}_n(v)-{\mathcal A}_n(u)\,\leq\, {\mathcal B}_{n,+}(v)-{\mathcal B}_{n,+}(u)+4\delta^{1-\beta}\,.$$ and $${\mathcal A}_n(v)-{\mathcal A}_n(u)\,\geq\, {\mathcal B}_{n,-}(v)-{\mathcal B}_{n,-}(u)-2\delta^{1-\beta}\,,$$ for all $v\in[u,u+\delta]$.
$\Box$\\
\noindent{\bf Proof of Theorem \ref{thm:Tight}\,\,} For fixed $u\in[0,1)$ take $\delta>0$ such that $u+\delta\leq1$. By Lemma \ref{lem:compa},
$$\sup_{v\in[u,u+\delta]}|{\mathcal A}_n(v)-{\mathcal A}_n(u)|\leq \max\left\{\sup_{v\in[u,u+\delta]}|{\mathcal B}_{n,\pm}(v)-{\mathcal B}_{n,\pm}(u)| \right\}+4\delta^{1-\beta}\,,$$ on the event $E_n(\delta)$. Hence, for any $\eta>0$, \begin{eqnarray*}
{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal A}_n(v)-{\mathcal A}_n(u)|>\eta\right)&\leq&{\mathbb P}\left(E_n(\delta)^c\right)\\
&+&{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal B}_{n,+}(v)-{\mathcal B}_{n,+}(u)|>\eta-4\delta^{1-\beta}\right)\\
&+&{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal B}_{n,-}(v)-{\mathcal B}_{n,-}(u)|>\eta-4\delta^{1-\beta}\right)\,. \end{eqnarray*} By \eqref{eq:equilibrium}, $$P_n(x):=L_\lambda\big([n+x]_n\big)-L_\lambda\big([n]_n\big)\,,\,\mbox{ for }\,\,\,x\geq 0\,,$$ is a Poisson process of intensity $\lambda$. Since $\lambda^{\pm}\to1$ as $n\to\infty$, ${\mathcal B}_{n,-}(u/2)$ and ${\mathcal B}_{n,+}(u/2)$ converge in distribution to a standard Brownian motion ${\mathcal B}$. Thus, by Lemma \ref{lem:tight}, for $\delta<(\eta/8)^{1/(1-\beta)}$, \begin{eqnarray*}
\limsup_{n\to\infty}{\mathbb P}\left(\sup_{v\in[u,u+\delta]}|{\mathcal A}_n(v)-{\mathcal A}_n(u)|>\eta\right)&\leq& C\delta^{3\beta}+2{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal B}(2v)-{\mathcal B}(2u)|>\eta-4\delta^{1-\beta}\right)\\
&\leq&C\delta^{3\beta}+2{\mathbb P}\left( \sup_{v\in[0,1]}|{\mathcal B}(v)|>\frac{\eta}{2\sqrt{2\delta}}\right)\,, \end{eqnarray*} which implies that (recall that $\beta\in(1/3,1)$) \begin{equation}\label{tight}
\limsup_{\delta\to 0^+}\frac{1}{\delta}\left(\limsup_{n\to\infty}{\mathbb P}\left(\sup_{v\in[u,u+\delta]}|{\mathcal A}_n(v)-{\mathcal A}_n(u)|>\eta\right)\right)=0\,. \end{equation} Since \cite{CG2}
$$\limsup_{n\to\infty}\frac{{\mathbb E} |L[n]_n-2n|}{n^{1/3}}<\infty\,,$$ we have that $\{{\mathcal A}_n(0)\,,\,n\geq 1\}$ is tight. Together with \eqref{tight}, this shows tightness of the collection $\{{\mathcal A}_n\,,\,n\geq 1\}$ in the space of cadlag functions on $[0,1]$, and also that every weak limit lives in the space of continuous functions \cite{Bi}.
$\Box$\\
\section{Proof of Theorem \ref{thm:LocalFluct}} For simple notation, we will prove the statement for $s=1$ and restrict our selves to $[0,1]$. The reader can then check that rescaling gives the result for general $s>0$, since \[ L[sn]_n \stackrel{dist.}{=} L[s^{1/2}n]_{s^{1/2}n}.\] \begin{lem}\label{lem:PoissonTight} Fix $\gamma'\in(\gamma,2/3)$ and let $${\lambda}_{\pm}={\lambda}_{\pm}(n):=1\pm\frac{1}{n^{\gamma'/2}}\,.$$ Define the event $$E_n:=\left\{Z'_{{\lambda}_+}[n]_n\geq0\,\,\mbox{ and }\,\,Z_{{\lambda}_-}[n+n^{\gamma}]_n\leq0\right\}\,.$$ There exists a constant $C>0$ such that $${\mathbb P}\left(E_n^c\right)\leq \frac{C}{n^{1-3\gamma'/2}}\,$$ for all sufficiently large $n$. \end{lem}
\noindent{\bf Proof\,\,} Denote $r:=n^{1/3-\gamma'/2}$ and let $$n_+:=\lambda_+ n< 2n\,\,\mbox{ and }\,\,h_{+}:=\left(\lambda_+-\frac{1}{\lambda_+}\right) n> rn^{2/3}> r n_+^{2/3}/2 \,,$$ (for all sufficiently large $n$). By \eqref{eq:sym} and \eqref{eq:diagsym}, \begin{eqnarray*} {\mathbb P}\left(Z'_{\lambda_+}[n]_n<0\right)&=&{\mathbb P}\left(Z_{1/\lambda_+}[n]_n>0\right)\\ &=&{\mathbb P}\left(Z_1[n/\lambda_+]_{\lambda_+ n}>0\right)\\ &=&{\mathbb P}\left(Z_1[\lambda_+n-h_n]_{\lambda_+ n}>0\right)\\ &=&{\mathbb P}\left(Z_1[\lambda_+n]_{\lambda_+ n}>h_+\right)\\ &\leq&{\mathbb P}\left(Z_1[n_+]_{n_+}>r n_+^{2/3}/2\right)\,. \end{eqnarray*} Analogously, for $$n_-:= \frac{n}{\lambda_-}<2n\,\,\mbox{ and }\,\,h_{-}:=\left(\frac{1}{\lambda_-}-\lambda_-\right) n> rn^{2/3}> r n_-^{2/3}/2 \,,$$ we have that \begin{eqnarray*} {\mathbb P}\left(Z_{\lambda_-}[n+n^{\gamma}]_n>0\right)&=&{\mathbb P}\left(Z_{\lambda_-}[n]_n>-n^{\gamma}\right)\\ &=&{\mathbb P}\left(\lambda_-Z_{1}[\lambda_-n]_{n/\lambda_-}>-n^{\gamma}\right)\\ &\leq&{\mathbb P}\left(Z_{1}[n_-]_{n_-}>h_- - n^{\gamma}\right)\\ &\leq& {\mathbb P}\left(Z_1[n_-]_{n_-}>(r-n^{\gamma-2/3}) n_-^{2/3}/2\right)\,, \end{eqnarray*} Now one can use Lemma \ref{lem:ExitControl} to finish the proof.
$\Box$\\
\begin{lem}\label{lem:PoissonComparison} On the event $E_n$, for all $u<v$ in $[0,1]$, $$\Gamma_n^{-}(v)-\Gamma_n^{-}(u)-\frac{1}{n^{(\gamma'-\gamma)/2}}\leq \Delta_n(v)-\Delta_n(u)\leq \Gamma_n^+(v)-\Gamma_n^+(u)+\frac{1}{n^{(\gamma'-\gamma)/2}}\,,$$ where $$\Gamma_n^{\pm}(u):=\frac{L_{{\lambda}_\pm}[n+un^{\gamma}]_{n}-L_{{\lambda}_\pm}[n]_n-\lambda_\pm un^{\gamma}}{n^{\gamma/2}}\,.$$
\end{lem}
\noindent{\bf Proof\,\,} By Lemma \ref{lem:LocalComparison}, if $Z'_{\lambda_+}[n]_n\geq 0$ then $$L[n+vn^{\gamma}]_{n}-L[n+un^{\gamma}]_n\leq L_{\lambda_+}[n+vn^{\gamma}]_{n}-L_{\lambda_+}[n+un^{\gamma}]_n\,,$$ and if $Z_{\lambda_-}[n+n^{\gamma}]_n\leq 0$ then $$L[n+vn^{\gamma}]_{n}-L[n+un^{\gamma}]_n\geq L_{\lambda_-}[n+vn^{\gamma}]_{n}-L_{\lambda_-}[n+un^{\gamma}]_n\,.$$ Using that $\lambda_{\pm}:=1\pm n^{-\gamma'/2}$, one can finish the proof of the lemma.
$\Box$\\
\noindent{\bf Proof of Theorem \ref{thm:LocalFluct}\,\,} By Lemma \ref{lem:PoissonComparison}, on the event $E_n^c$,
$$| \Delta_n(v)-\Delta_n(u)|\leq\max \left\{|\Gamma_n^{\pm}(v)-\Gamma_n^{\pm}(u)|\right\}+\frac{1}{n^{(\gamma'-\gamma)/2}}\,.$$ Thus, by Lemma \ref{lem:PoissonTight}, \begin{eqnarray*}
{\mathbb P}\left(\sup_{v\in[u,u+\delta]} | \Delta_n(v)-\Delta_n(u)|>\eta\right)&\leq&{\mathbb P}\left(\sup_{v\in[u,u+\delta]} |\Gamma^+_n(v)-\Gamma^+_n(u)|+\frac{1}{n^{(\gamma'-\gamma)/2}}>\eta\right)\\
&+&{\mathbb P}\left(\sup_{v\in[u,u+\delta]} |\Gamma^-_n(v)-\Gamma^-_n(u)|+\frac{1}{n^{(\gamma'-\gamma)/2}}>\eta\right)\\ &+&{\mathbb P}\left(E_n^c\right)\,. \end{eqnarray*} As before, $\lambda^{\pm}\to1$ as $n\to\infty$, which implies that
$$\limsup_{n\to\infty}{\mathbb P}\left(\sup_{v\in[u,u+\delta]} | \Delta_n(v)-\Delta_n(u)|>\eta\right)\leq 2{\mathbb P}\left(\sup_{v\in[0,\delta]} |B(v)|>\eta\right)=2{\mathbb P}\left(\sup_{v\in[0,1]} |B(v)|>\frac{\eta}{\sqrt{\delta}}\right)\,,$$ and hence \begin{equation}\label{eq:tight}
\limsup_{\delta\to0^+}\frac{1}{\delta}\left(\limsup_{n\to\infty}{\mathbb P}\left(\sup_{v\in[u,u+\delta]} | \Delta_n(v)-\Delta_n(u)|>\eta\right)\right)=0\,. \end{equation} Since $\Delta_n(0)=0$, \eqref{eq:tight} implies tightness of the collection $\{\Delta_n\,,\,n\geq 1\}$ in the space of cadlag functions on $[0,1]$, and also that every weak limit lives in the space of continuous functions \cite{Bi}.
The finite dimensional distributions of the limiting process can be obtained in the same way. Indeed, by Lemma \ref{lem:PoissonTight} and Lemma \ref{lem:PoissonComparison}, for $u_1,\dots,u_k\in[0,1]$ and $a_1,\dots,a_k,\in{\mathbb R}$, $${\mathbb P}\left(\cap_{i=1}^k\left\{\Delta_n(u_i)\leq a_i\right\}\right)\geq {\mathbb P}\left(\cap_{i=1}^k\left\{\Gamma^+_n(u_i)\leq a_i -\frac{1}{n^{(\gamma'-\gamma)/2}}\right\}\right)-{\mathbb P}\left(E_n^c\right)\,,$$ and $${\mathbb P}\left(\cap_{i=1}^k\left\{\Delta_n(u_i)\leq a_i\right\}\right)\leq{\mathbb P}\left(\cap_{i=1}^k\left\{\Gamma^-_n(u_i)\leq a_i+\frac{1}{n^{(\gamma'-\gamma)/2}}\right\}\right)+{\mathbb P}\left(E_n^c\right)\,,$$ which shows that the finite dimensional distributions of $\Delta_n$ converge to the finite dimensional distributions of the standard Brownian motion process.
$\Box$\\
\section{Proof of Theorem \ref{thm:LocalAiry}}
\begin{lem}\label{lem:local} Fix $\beta\in(0,1/2)$ and for $\epsilon\in(0,1)$ let $$\lambda_\pm=\lambda_\pm(n,\epsilon):=1\pm\frac{\epsilon^{-\beta}}{n^{1/3}}\,.$$ Define the event
$$E_n(\epsilon):=\left\{Z'_{\lambda_{+}}([n]_n)\geq 0\,\,\mbox{ and }\,\,Z_{\lambda_{-}}([n+n^{2/3}]_n)\leq 0\right\}\,.$$ There exists a constant $C>0$ such that, for all sufficiently small $\epsilon>0$, $$\limsup_{n\to\infty}{\mathbb P}\left(E_n(\epsilon)^c\right)\leq C\epsilon^{3\beta}\,.$$ \end{lem}
\noindent{\bf Proof\,\,} The same proof as in Lemma \ref{lem:tight} applies.
$\Box$\\
\begin{lem}\label{lem:localcompa} On the event $E_n(\epsilon)$, for all $u\in[0,1-\delta)$ and $v\in[u,u+\delta]$, we have that $$ {\mathcal B}_{n,-}(\epsilon v)-{\mathcal B}_{n,-}(\epsilon u)-2\delta\epsilon^{1-\beta}\leq{\mathcal A}_n(\epsilon v)-{\mathcal A}_n(\epsilon u)\,\leq\, {\mathcal B}_{n,+}(\epsilon v)-{\mathcal B}_{n,+}(\epsilon u)+4\delta\epsilon^{1-\beta}\,.$$ \end{lem}
\noindent{\bf Proof\,\,} The same proof as in Lemma \ref{lem:compa} applies. Note that in this case we have \[ (\lambda_+ - 1)(2{\varepsilon } v- 2{\varepsilon } u) \leq 2{\varepsilon }^{1-\beta}\delta.\]
$\Box$\\
\noindent{\bf Proof of Theorem \ref{thm:LocalAiry}\,\,} For $u\in[0,1]$, let $${\mathcal A}_n^\epsilon(u):=\epsilon^{-1/2}\left({\mathcal A}_n(\epsilon u)-{\mathcal A}_n(0)\right)\,\mbox{ and }\,{\mathcal B}^\epsilon_{n,\pm}(u):=\epsilon^{-1/2}{\mathcal B}_{n,\pm}(\epsilon u)\,.$$ By Lemma \ref{lem:localcompa}, on the event $E_n(\epsilon)$, for all $v\in[u,u+\delta]$, $$ {\mathcal B}^\epsilon_{n,-}(v)-{\mathcal B}^\epsilon_{n,-}(u)-2\delta\epsilon^{1/2-\beta}\leq{\mathcal A}^\epsilon_n(v)-{\mathcal A}^\epsilon_n(u)\,\leq\, {\mathcal B}^\epsilon_{n,+}(v)-{\mathcal B}^\epsilon_{n,+}(u)+4\delta\epsilon^{1/2-\beta}\,,$$ which shows that
$$\sup_{v\in[u,u+\delta]}|{\mathcal A}^\epsilon_n(v)-{\mathcal A}^\epsilon_n(u)|\leq \max\left\{\sup_{v\in[u,u+\delta]}|{\mathcal B}_{n,\pm}^\epsilon(v)-{\mathcal B}_{n,\pm}^\epsilon(u)| \right\}+4\delta\epsilon^{1/2-\beta}\,.$$ Therefore, \begin{eqnarray*}
{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal A}^\epsilon_n(v)-{\mathcal A}^\epsilon_n(u)|>\eta\right)&\leq&{\mathbb P}\left(E_n(\epsilon)^c\right)\\
&+&{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal B}_{n,+}^\epsilon(v)-{\mathcal B}_{n,+}^\epsilon(u)|>\eta-4\delta\epsilon^{1/2-\beta}\right)\\
&+&{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal B}_{n,-}^\epsilon(v)-{\mathcal B}_{n,-}^\epsilon(u)|>\eta-4\delta\epsilon^{1/2-\beta}\right)\,. \end{eqnarray*} Since ${\mathcal A}_n^\epsilon$ is converging to $\cal A^\epsilon$, and ${\mathcal B}_{n,\pm}^\epsilon$ is converging to a Brownian motion ${\mathcal B}$, the preceding inequality implies that
$${\mathbb P}\left(\sup_{v\in[u,u+\delta]}|{\mathcal A}^\epsilon(v)-{\mathcal A}^\epsilon(u)|>\eta\right)\leq C\epsilon^{3\beta}+2{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal B}(2v)-{\mathcal B}(2u)|>\eta-4\delta\epsilon^{1/2-\beta}\right)\,.$$ Hence
$$\limsup_{\epsilon\to 0^+}{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal A}^\epsilon(v)-{\mathcal A}^\epsilon(u)|>\eta\right)\leq 2{\mathbb P}\left( \sup_{v\in[0,1]}|{\mathcal B}(v)|>\frac{\eta}{\sqrt{2\delta}}\right)\,,$$ which shows that \begin{equation}\label{eq:BlowTight}
\limsup_{\delta\to 0^+}\frac{1}{\delta}\left(\limsup_{\epsilon\to 0^+}{\mathbb P}\left( \sup_{v\in[u,u+\delta]}|{\mathcal A}^\epsilon(v)-{\mathcal A}^\epsilon(u)|>\eta\right)\right)=0\,. \end{equation} Since ${\mathcal A}^{\epsilon}(0)=0$, by \eqref{eq:BlowTight} we have that $\left\{{\mathcal A}^\epsilon\,,\,\epsilon\in(0,1]\right\}$ is tight \cite{Bi}.
The finite dimensional distributions of the limiting process can be obtained in the same way. Indeed, by Lemma \ref{lem:localcompa}, for $u_1,\dots,u_k\in[0,1]$ and $a_1,\dots,a_k,\in{\mathbb R}$, $${\mathbb P}\left(\cap_{i=1}^k\left\{{\mathcal A}^\epsilon_n(u_i)\leq a_i\right\}\right)\leq{\mathbb P}\left(\cap_{i=1}^k\left\{{\mathcal B}^\epsilon_{-,n}(u_i)\leq a_i+4\epsilon^{1/2-\beta}\right\}\right)+{\mathbb P}\left(E_n(\epsilon)^c\right)\,,$$ and $${\mathbb P}\left(\cap_{i=1}^k\left\{{\mathcal A}^\epsilon_n(u_i)\leq a_i\right\}\right)\geq {\mathbb P}\left(\cap_{i=1}^k\left\{{\mathcal B}^\epsilon_{+,n}(u_i)\leq a_i-4\epsilon^{1/2-\beta}\right\}\right)-{\mathbb P}\left(E_n(\epsilon)^c\right)\,.$$ Thus, by Lemma \ref{lem:local}, $$ {\mathbb P}\left(\cap_{i=1}^k\left\{{\mathcal A}^\epsilon(u_i)\leq a_i\right\}\right)\leq {\mathbb P}\left(\cap_{i=1}^k\left\{{\mathcal B}(2u_i)\leq a_i+4\epsilon^{1/2-\beta}\right\}\right)+C\epsilon^{3\beta}\,,$$ and $$ {\mathbb P}\left(\cap_{i=1}^k\left\{{\mathcal A}^\epsilon(u_i)\leq a_i\right\}\right)\geq {\mathbb P}\left(\cap_{i=1}^k\left\{{\mathcal B}(2u_i)\leq a_i-4\epsilon^{1/2-\beta}\right\}\right)-C\epsilon^{3\beta}\,,$$ which proves that, \begin{equation}\label{eq:BlowDist} \lim_{\epsilon\to 0^+}{\mathbb P}\left(\cap_{i=1}^k\left\{{\mathcal A}^\epsilon(u_i)\leq a_i\right\}\right)={\mathbb P}\left(\cap_{i=1}^k\left\{\sqrt{2}{\mathcal B}(u_i)\leq a_i\right\}\right)\,. \end{equation}
$\Box$\\
\end{document}
|
arXiv
|
{
"id": "1311.1349.tex",
"language_detection_score": 0.6048654317855835,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Attacking quantum key distribution with single-photon two-qubit quantum logic}
\author{Jeffrey H. Shapiro} \email[Electronic address: ]{[email protected]} \author{Franco N. C. Wong} \affiliation{Massachusetts Institute of Technology, Research Laboratory of Electronics, Cambridge, Massachusetts 02139 USA}
\date{\today}
\begin{abstract} The Fuchs-Peres-Brandt (FPB) probe realizes the most powerful individual attack on Bennett-Brassard 1984 quantum key distribution (BB84 QKD) by means of a single controlled-NOT (CNOT) gate. This paper describes a complete physical simulation of the FPB-probe attack on polarization-based BB84 QKD using a deterministic CNOT constructed from single-photon two-qubit quantum logic. Adding polarization-preserving quantum nondemolition measurements of photon number to this configuration converts the physical simulation into a true deterministic realization of the FPB attack. \end{abstract}
\pacs{03.67.Dd, 03.67.Lx, 42.50.Dv, 42.40.Lm}
\maketitle
\section{Introduction}
Bennett-Brassard 1984 quantum key distribution (BB84 QKD) using single-photon polarization states works as follows \cite{BB84}. In each time interval allotted for a bit, Alice transmits a single photon in a randomly selected polarization, chosen from horizontal ($H$), vertical ($V$), $+$45$^\circ$, or $-$45$^\circ$, while Bob randomly chooses to detect photons in either the $H$/$V$ or $\pm$45$^\circ$ bases. Bob discloses to Alice the sequence of bit intervals and associated measurement bases for which he has detections. Alice then informs Bob which detections occurred in bases coincident with the ones that she used. These are the \em sift\/\rm\ events, i.e., bit intervals in which Bob has a detection \em and\/\rm\ his count has occurred in the same basis that Alice used. An \em error\/\rm\ event is a sift event in which Bob decodes the incorrect bit value. Alice and Bob employ a prescribed set of operations to identify errors in their sifted bits, correct these errors, and apply sufficient privacy amplification to deny useful key information to any potential eavesdropper (Eve). At the end of the full QKD procedure, Alice and Bob have a shared one-time pad with which they can communicate in complete security.
In long-distance QKD systems, most of Alice's photons will go undetected, owing to propagation loss and detector inefficiencies. Dark counts and, for atmospheric QKD systems, background counts can cause error events in these systems, as can intrusion by Eve. Employing an attenuated laser source, in lieu of a true single-photon source, further reduces QKD performance as such sources are typically run at less than one photon on average per bit interval, and the occurrence of multi-photon events, although rare at low average photon number, opens up additional vulnerability. Security proofs have been published for ideal BB84 \cite{security1}, as have security analyses that incorporate a variety of non-idealities \cite{security2}. Our attention, however, will be directed toward attacking BB84 QKD, as to our knowledge no such experiments have been performed, although a variety of potentially practical approaches have been discussed \cite{attacks}. Our particular objective will be to show that current technology permits physical simulation of the Fuchs-Peres-Brandt (FPB) probe \cite{FPB}, i.e., the most powerful individual attack on single-photon BB84, and that developments underway in quantum nondemolition (QND) detection may soon turn this physical simulation into a full implementation of the attack. Thus we believe it is of interest to construct the physical simulation and put BB84's security to the test: how much information can Eve really derive about the key that Alice and Bob have distilled while keeping Alice and Bob oblivious to her presence.
The remainder of this paper is organized as follows. In Sec.~II we review the FPB probe and its theoretical performance. In Sec.~III we describe a complete physical simulation of this probe constructed from single-photon two-qubit (SPTQ) quantum logic. We conclude, in Sec.~IV, by showing how the addition of polarization-preserving QND measurements of photon number can convert this physical simulation into a true deterministic realization of the FPB attack on polarization-based BB84.
\section{The Fuchs-Peres-Brandt Probe} In an individual attack on single-photon BB84 QKD, Eve probes Alice's photons one at a time. In a collective attack, Eve's measurements probe groups of Alice's photons. Less is known about collective attacks \cite{collective}, so we will limit our consideration to individual attacks. Fuchs and Peres \cite{FP} described the most general way in which an individual attack could be performed. Eve supplies a probe photon and lets it interact with Alice's photon in a unitary manner. Eve then sends Alice's photon to Bob, and performs a probability operator-valued measurement (POVM) on the probe photon she has retained. Slutsky {\em et al.} \cite{Slutsky} demonstrated that the Fuchs-Peres construct---with the appropriate choice of probe state, interaction, and measurement---affords Eve the maximum amount of R\'{e}nyi information about the error-free sifted bits that Bob receives for a given level of disturbance, i.e., for a given probability that a sifted bit will be received in error. Brandt \cite{FPB} extended the Slutsky {\em et al.} treatment by showing that the optimal probe could be realized with a single CNOT gate. Figure~1 shows an abstract diagram of the resulting Fuchs-Peres-Brandt probe. In what follows we give a brief review of its structure and performance---see \cite{FPB} for a more detailed treatment---where, for simplicity, we assume ideal conditions in which Alice transmits a single photon per bit interval, there is no propagation loss and no extraneous (background) light collection, and both Eve and Bob have unity quantum efficiency photodetectors with no dark counts. These ideal conditions imply there will not be any errors on sifted bits in the absence of eavesdropping; the case of more realistic conditions will be discussed briefly in Sec.~IV. \begin{figure}
\caption{(Color online) Block diagram of the Fuchs-Peres-Brandt probe for attacking BB84 QKD.}
\end{figure}
In each bit interval Alice transmits, at random, a single photon in one of the four BB84 polarization states. Eve uses this photon as the control-qubit input to a CNOT gate whose computational basis---relative to the BB84 polarization states---is shown in Fig.~2, namely \begin{eqnarray}
|0\rangle &\equiv& \cos(\pi/8)|H\rangle + \sin(\pi/8)|V\rangle \\
|1\rangle &\equiv& -\sin(\pi/8)|H\rangle + \cos(\pi/8)|V\rangle, \end{eqnarray} in terms of the $H/V$ basis. Eve supplies her own probe photon, as the target-qubit input to this CNOT gate, in the state \begin{equation}
|T_{\rm in}\rangle \equiv C|+\rangle + S|-\rangle, \label{probeinput} \end{equation}
where $C = \sqrt{1-2P_E}$, $S = \sqrt{2P_E}$, $|\pm\rangle = (|0\rangle \pm |1\rangle)/\sqrt{2}$, and $0\le P_E\le 1/2$ will turn out to be the error probability that Eve's probe creates on Bob's sifted bits \cite{footnote1}. So, as $P_E$ increases from 0 to 1/2, $|T_{\rm in}\rangle$ goes from $|+\rangle$ to $|-\rangle$. The (unnormalized) output states that may occur for this target qubit are \begin{eqnarray}
|T_\pm\rangle &\equiv& C|+\rangle \pm \frac{S}{\sqrt{2}}|-\rangle \\
|T_E\rangle &\equiv& \frac{S}{\sqrt{2}}|-\rangle. \end{eqnarray} \begin{figure}
\caption{(Color online) Computational basis for Eve's CNOT gate referenced to the BB84 polarization states.}
\end{figure}
Here is how the FPB probe works. When Alice uses the $H/V$ basis for her photon transmission, Eve's CNOT gate effects the following transformation, \begin{eqnarray}
|H\rangle|T_{\rm in}\rangle &\longrightarrow&
|H\rangle|T_-\rangle + |V\rangle|T_E\rangle \label{Hin_out} \\
|V\rangle|T_{\rm in}\rangle &\longrightarrow&
|V\rangle|T_+\rangle +|H\rangle|T_E\rangle,\label{Vin_out} \end{eqnarray} where the kets on the left-hand side denote the Alice\,$\otimes$\,Eve state of the control and target qubits at the CNOT's input and the kets on the right-hand side denote the Bob\,$\otimes$\,Eve state of the control and target qubits at the CNOT's output. Similarly, when Alice uses the $\pm 45^\circ$ basis, Eve's CNOT gate has the following behavior, \begin{eqnarray}
|\mbox{$+$}45^\circ\rangle|T_{\rm in}\rangle &\longrightarrow&
|\mbox{$+$}45^\circ\rangle|T_+\rangle + |\mbox{$-$}45^\circ\rangle|T_E\rangle \label{plus_in_out}\\
|\mbox{$-$}45^\circ\rangle|T_{\rm in}\rangle &\longrightarrow&
|\mbox{$-$}45^\circ\rangle|T_-\rangle +|\mbox{$+$}45^\circ\rangle|T_E\rangle. \label{minus_in_out} \end{eqnarray}
Suppose that Bob measures in the basis that Alice has employed \em and\/\rm\ his outcome matches what Alice sent. Then Eve can learn their shared bit value, once Bob discloses his measurement basis, by distinguishing between the $|T_+\rangle$ and $|T_-\rangle$ output states for her target qubit. Of course, this knowledge comes at a cost: Eve has caused an error event whenever Alice and Bob choose a common basis and her target qubit's output state is $|T_E\rangle$. To maximize the information she derives from this intrusion, Eve applies the minimum error probability receiver for distinguishing between the single-photon polarization states $|T_+\rangle$ and $|T_-\rangle$. This is a projective measurement onto the polarization basis $\{|d_+\rangle,|d_-\rangle\}$, shown in Fig.~3 and given by \begin{eqnarray}
|d_+\rangle &=& \frac{|+\rangle + |-\rangle}{\sqrt{2}} = |0\rangle \\
|d_-\rangle &=& \frac{|+\rangle - |-\rangle}{\sqrt{2}} = |1\rangle. \end{eqnarray} \begin{figure}
\caption{(Color online) Measurement basis for Eve's minimum-error-probability discrimination between $|T_+\rangle$ and $|T_-\rangle$.}
\end{figure}
Two straightforward calculations will now complete our review of the FPB probe. First, we find the error probability that is created by Eve's presence. Suppose Alice and Bob use the $H/V$ basis and Alice has sent $|H\rangle$. Alice and Bob will incur an error if the control\,$\otimes$\,target output from Eve's CNOT gate is $|V\rangle|T_E\rangle$. The probability that this occurs is $\langle T_E| T_E\rangle = S^2/2 = P_E$. The same conditional error probability ensues for the other three error events, e.g., when Alice and Bob use the $\pm 45^\circ$ basis, Alice sends $|$$+45^\circ\rangle$, and the CNOT output is $|$$-45^\circ\rangle|T_E\rangle$. It follows that the unconditional error probability incurred by Alice and Bob on their sift events is $P_E$.
Now we shall determine the R\'{e}nyi information that Eve derives about the sift events for which Alice and Bob do not suffer errors. Let $B = \{0,1\}$ and $E = \{0,1\}$ denote the ensembles of possible bit values that Bob and Eve receive on a sift event in which Bob's bit value agrees with Alice's. The R\'{e}nyi information (in bits) that Eve learns about each Alice/Bob error-free sift event is \begin{eqnarray} I_R &\equiv& -\log_2\!\left(\sum_{b= 0}^1P^2(b)\right) \nonumber \\ &+&\sum_{e = 0}^1P(e)\log_2\!\left(\sum_{b = 0}^1 P^2(b\mid e)\right), \end{eqnarray} where $\{P(b), P(e)\}$ are the prior probabilities for Bob's and Eve's bit values, and $P(b\mid e)$ is the conditional probability for Bob's bit value to be $b$ given that Eve's is $e$. Alice's bits are equally likely to be 0 or 1, and Eve's conditional error probabilities satisfy \cite{Helstrom} \begin{eqnarray} \lefteqn{P(e = 1\mid b = 0) = P(e = 0\mid b = 1)} \\
&=& \frac{1}{2}\!\left(1 - \sqrt{1 - \frac{|\langle T_+|T_-\rangle|^2}{\langle T_+|T_+\rangle \langle T_-|T_-\rangle}}\right) \\ &=& \frac{1}{2}\!\left(1- \frac{\sqrt{4P_E(1-2P_E)}}{1-P_E}\right). \end{eqnarray} These results imply that $b$ is also equally likely to be 0 or 1, and that $P(b\mid e) = P(e\mid b)$, whence \begin{equation} I_R = \log_2\!\left(1 + \frac{4P_E(1-2P_E)}{(1-P_E)^2}\right), \end{equation} which we have plotted in Fig.~4. \begin{figure}
\caption{(Color online) Eve's R\'{e}nyi information about Bob's error-free sifted bits as a function of the error probability that her eavesdropping creates.}
\end{figure}
Figure~4 reveals several noteworthy performance points for the FPB probe. The $I_R = 0, P_E = 0$ point in this figure corresponds to Eve's operating her CNOT gate with $|T_{\rm in}\rangle = |+\rangle$ for its target qubit input. It is well known that such an input is unaffected by and does not affect the control qubit. Thus Bob suffers no errors but Eve gets no R\'{e}nyi information. The $I_R = 1, P_E = 1/3$ point in this figure corresponds to Eve's operating her CNOT gate with $|T_{\rm in}\rangle = \sqrt{1/3}|+\rangle + \sqrt{2/3}|-\rangle$, which leads to $|T_\pm\rangle \propto |d_\pm\rangle$. In this case Eve's Fig.~3 receiver makes no errors, so she obtains the maximum (1 bit) R\'{e}nyi information about each of Bob's error-free bits. The $I_R = 0, P_E = 1/2$ point in this figure corresponds to Eve's operating her CNOT gate with $|T_{\rm in}\rangle = |-\rangle$, which gives $|T_+\rangle = |T_-\rangle = |T_E\rangle = \sqrt{1/2}|-\rangle$. Here it is clear that Eve gains no information about Bob's error-free bits, but his error probability is 1/2 because of the action of the $|-\rangle$ target qubit on the control qubit.
\section{Physical Simulation in SPTQ Logic}
In single-photon two-qubit quantum logic, each photon encodes two independently controllable qubits \cite{SPTQ1}. One of these is the familiar polarization qubit, with basis $\{|H\rangle,|V\rangle\}$. The other we shall term the momentum qubit---because our physical simulation of the FPB probe will rely on the polarization-momentum hyperentangled photon pairs produced by type-II phase matched spontaneous parametric downconversion (SPDC)---although in the collimated configuration in which SPTQ is implemented its basis states are single-photon kets for right and left beam positions (spatial modes), denoted $\{|R\rangle, |L\rangle\}$. Unlike the gates proposed for linear optics quantum computing \cite{KLM}, which are scalable but non-deterministic, SPTQ quantum logic is deterministic but not scalable. Nevertheless, SPTQ quantum logic suffices for a complete physical simulation of polarization-based BB84 being attacked with the FPB probe, as we shall show. Before doing so, however, we need to comment on the gates that have been demonstrated in SPTQ logic.
It is well known that single qubit rotations and CNOT gates form a universal set for quantum computation. In SPTQ quantum logic, polarization-qubit rotations are easily accomplished with wave plates, just as is done in linear optics quantum computing. Momentum-qubit rotations are realized by first performing a SWAP operation, to exchange the polarization and momentum qubits, then rotating the polarization qubit, and finally performing another SWAP. The SWAP operation is a cascade of three CNOTs, as shown in Fig.~5. For its implementation in SPTQ quantum logic the left and right CNOTs in Fig.~5 are momentum-controlled NOT gates (M-CNOTs) and the middle CNOT is a polarization-controlled NOT gate (P-CNOT). (An M-CNOT uses the momentum qubit of a single photon to perform the controlled-NOT operation on the polarization qubit of that same photon, and vice versa for the P-CNOT gate.) Experimental demonstrations of deterministic M-CNOT, P-CNOT, and SWAP gates are reported in \cite{SPTQ1,SPTQ2}. \begin{figure}
\caption{Quantum circuit diagram for a SWAP gate realized as a cascade of three CNOTs. In SPTQ quantum logic the upper rail is the momentum qubit and the lower rail is the polarization qubit of the same photon.}
\end{figure}
Figure~6 shows a physical simulation of polarization-based BB84 under FPB attack when Alice has a single-photon source and Bob employs active basis selection; Fig.~7 shows the modification needed to accommodate Bob's using passive basis selection. In either case, Alice uses a polarizing beam splitter and an electro-optic modulator, as a controllable half-wave plate (HWP), to set the randomly-selected BB84 polarization state for each photon she transmits. Moreover, she employs a single spatial mode, which we assume coincides with the $R$ beam position in Eve's apparatus. Eve then begins her attack by imposing the probe state $|T_{\rm in}\rangle$ on the momentum qubit. She does this by applying a SWAP gate, to exchange the momentum and polarization qubits of Alice's photon, rotating the resulting polarization qubit (with the HWP in Fig.~6) to the $|T_{\rm in}\rangle$ state, and then using another SWAP to switch this state into the momentum qubit. This procedure leaves Alice's BB84 polarization state unaffected, although her photon, which will ultimately propagate on to Bob, is no longer in a single spatial mode. Eve completes the first stage of her attack by sending Alice's photon through a P-CNOT gate, which will accomplish the state transformations given in Eqs.~(\ref{Hin_out})--(\ref{minus_in_out}), and then routing it to Bob. If Bob employs active basis selection (Fig.~6), then in each bit interval he will use an electro-optic modulator---as a controllable HWP---plus a polarizing beam splitter to set the randomly-selected polarization basis for his measurement. The functioning of this basis-selection setup is unaffected by Alice's photon no longer being in a single spatial mode. The reason that we call Fig.~6 a physical simulation, rather than a true attack, lies in the measurement box. Here, Eve has invaded Bob's turf, and inserted SWAP gates, half-wave plates, polarizing beam splitters, and additional photodetectors, so that she can forward to Bob measurement results corresponding to photon counting on the polarization basis that he has selected while she retains the photon counting results corresponding to her $\{|d_+\rangle, |d_-\rangle\}$ measurement. Clearly Bob would never knowingly permit Eve to intrude into his receiver box in this manner. Moreover, if Eve could do so, she would not bother with an FPB probe as she could directly observe Bob's bit values. \begin{figure}
\caption{(Color online) Physical simulation of polarization-based BB84 QKD and the FPB-probe attack.}
\end{figure} \begin{figure}
\caption{(Color online) Modification of Fig.~6 to accommodate Bob's using passive basis selection.}
\end{figure}
If Bob employs passive basis selection (Fig.~7), then he uses a 50/50 beam splitter followed by static-HWP analysis in the $H$-$V$ and $\pm 45^\circ$ bases, with only the former being explicitly shown in Fig.~7. The rest of Eve's attack mimics what was seen in Fig.~6, i.e., she gets inside Bob's measurement boxes with SWAP gates, half-wave plates, and additional detectors so that she can perform her probe measurement while providing Bob with his BB84 polarization-measurement data. Because the Fig.~7 arrangement requires that twice as many SWAP gates, twice as many half-wave plates, and twice as many single-photon detectors be inserted into Bob's receiver system, as compared to what is needed in the Fig.~6 setup, we shall limit the rest of our discussion to the case of active basis selection as it leads to a more parsimonious physical simulation of the Fuchs-Peres-Brandt attack. We recognize, of course, that the decision to use active basis selection is Bob's to make, not Eve's. More importantly, however, in Sec.~IV we will show how the availability of polarization-preserving QND photon-number measurements can be used to turn Fig.~6 into a true, deterministic implementation of the FPB attack. The same conversion can be accomplished for passive basis selection. Before turning to the true-attack implementation, let us flesh out some details of the measurement box in Fig.~6 and show how SPDC can be used, in lieu of the single-photon source, to perform this physical simulation.
Let $|\psi_{\rm out}\rangle$ denote the polarization\,$\otimes$\,momentum state at the output of Eve's P-CNOT gate in Fig.~6. Bob's polarization analysis box splits this state, according to the basis he has chosen, so that one basis state goes to the upper branch of the measurement box while the other goes to the lower branch of that box. This polarization sorting does nothing to the momentum qubit, so the SWAP gates, half-wave plates, and polarizing beam splitters that Eve has inserted into the measurement box accomplish her $\{|d_+\rangle, |d_-\rangle\}$ projective measurement, i.e., the horizontal paths into photodetectors in Fig.~6 are projecting the momentum qubit of $|\psi_{\rm out}\rangle$ onto $|d_-\rangle$ and the vertical paths into photodetectors in Fig.~6 are projecting this state onto
$|d_+\rangle$. Eve records the combined results of the two $|d_+\rangle$ versus $|d_-\rangle$ detections, whereas Bob, who only sees the combined photodetections for the upper and lower branches entering the measurement box, gets his BB84 polarization data. Bob's data is impaired, of course, by the effect of Eve's P-CNOT.
Single-photon on-demand sources are now under development at several institutions \cite{single}, and their use in BB84 QKD has been demonstrated \cite{singleBB84}. At present, however, it is much more practical to use SPDC as a heralded source of single photons \cite{herald}. In SPDC, signal and idler photons are emitted in pairs, thus detection of the signal photon heralds the presence of the idler photon. Moreover, with appropriate configurations \cite{bidirectional}, SPDC will produce photons that are simultaneously entangled in polarization and in momentum. This hyperentanglement leads us to propose the Fig.~8 configuration for physically simulating the FPB-probe attack on BB84. Here, a pump laser drives SPDC in a type-II phase matched $\chi^{(2)}$ crystal, such as periodically-poled potassium titanyl phosphate (PPKTP), producing pairs of orthogonally-polarized, frequency-degenerate photons that are entangled in both polarization and momentum. The first polarizing beam splitter transmits a horizontally-polarized photon and reflects a vertically-polarized photon while preserving their momentum entanglement. Eve uses a SWAP gate and (half-wave plate plus polarizing beam splitter) polarization rotation so that her photodetector's clicking will, by virtue of the momentum entanglement, herald the setting of the desired $|T_{\rm in}\rangle$ momentum-qubit state on the horizontally-polarized photon emerging from the first polarizing beam splitter. Alice's electronically controllable half-wave plate sets the BB84 polarization qubit on this photon, and the rest of the Fig.~8 configuration is identical to that shown and explained in Fig.~6. Inasmuch as the SPDC source and SPTQ gates needed to realize the Fig.~8 setup have been demonstrated, we propose that such an experiment be performed. Simultaneous recording of Alice's polarization choices, Bob's polarization measurements and Eve's $|d_+\rangle$ versus $|d_-\rangle$ results can then be processed through the BB84 protocol stack to study the degree to which the security proofs and eavesdropping analyses stand up to experimental scrutiny. \begin{figure}
\caption{(Color online) Proposed configuration for a complete physical simulation of the FPB attack on BB84 that is based on hyperentangled photon pairs from type-II phase matched SPDC and gates built from SPTQ quantum logic.}
\end{figure}
\section{The Complete Attack}
Although the FPB attack's physical simulation, as described in the preceding section, is both experimentally feasible and technically informative, any vulnerabilities it might reveal would only be of academic interest were there no practical means to turn it into a true deterministic implementation in which Eve did \em not\/\rm\ need to invade Bob's receiver. Quantum nondemolition measurement technology provides the key to creating this complete attack. As shown in the appendix, it is possible, in principle, to use cross-phase modulation between a strong coherent-state probe beam and an arbitrarily polarized signal beam to make a QND measurement of the signal beam's total photon number while preserving its polarization state. Cross-phase modulation QND measurement of photon number has long been a topic of interest in quantum optics \cite{Imoto}, and recent theory has shown that it provides an excellent new route to photonic quantum computation \cite{Nemoto}. Thus it is not unwarranted to presume that polarization-preserving QND measurement of total photon number may be developed. With such technology in hand, the FPB-probe attack shown in Fig.~9 becomes viable. Here, Eve imposes a momentum qubit on Alice's polarization-encoded photon and performs a P-CNOT operation exactly as discussed in conjunction with Figs.~6 and 8. Now, however, Eve uses a SWAP-gate half-wave plate combination so that the $|d_+\rangle$ and $|d_-\rangle$ momentum qubit states emerging from her P-CNOT become $|V\rangle$ and $|H\rangle$ states entering the polarizing beam splitter that follows the half-wave plate. This beam splitter routes these polarizations into its transmitted and reflected output ports, respectively, where, in each arm, Eve employs a SWAP gate, a polarization-preserving QND measurement of total photon number, and another SWAP gate. The first of these SWAPs returns Alice's BB84 qubit to polarization, so that a click on Eve's polarization-preserving QND apparatus completes her $\{|d_+\rangle, |d_-\rangle\}$ measurement without further scrambling Alice's BB84 qubit beyond what has already occurred in Eve's P-CNOT gate. The SWAP gates that follow the QND boxes then restore definite ($V$ and $H$) polarizations to the light in the upper and lower branches so that they may be recombined on a polarizing beam splitter. The SWAP gate that follows this recombination then returns the BB84 qubit riding on Alice's photon to polarization for transmission to and measurement by Bob. This photon is no longer in the single spatial mode emitted by Alice's transmitter, hence Bob could use spatial-mode discrimination to infer the presence of Eve, regardless of the $P_E$ value she had chosen to impose. Eve, however, can preclude that possibility. Because the result of her $\{|d_+\rangle,|d_-\rangle\}$ measurement tells her the value of the momentum qubit on the photon being sent to Bob, she can employ an additional stage of qubit rotation to restore this momentum qubit to the $|R\rangle$ state corresponding to Alice's transmission. Also, should Alice try to defeat Eve's FPB probe by augmenting her BB84 polarization qubit with a randomly-chosen momentum qubit, Eve can use a QND measurement setup like that shown in Fig.~9 to collapse the value of that momentum qubit to $|R\rangle$ or $|L\rangle$, and then rotate that momentum qubit into the $|R\rangle$-state spatial mode before applying the FPB-probe attack. At the conclusion of her attack, she can then randomize the momentum qubit on the photon that will be routed on to Bob without further impact---beyond that imposed by her P-CNOT gate---on that photon's polarization qubit. So, unless Alice and Bob generalize their polarization-based BB84 protocol to include cooperative examination of the momentum qubit, Alice's randomization of that qubit will neither affect Eve's FPB attack, nor provide Alice and Bob with any additional evidence, beyond that obtained from the occurrence of errors on sifted bits, of Eve's presence. \begin{figure}
\caption{(Color online) Deterministic FPB-probe attack on polarization-based BB84 that is realized with polarization-preserving QND measurements and SPTQ quantum logic.}
\end{figure}
Some concluding remarks are now in order. We have shown that a physical simulation of the Fuchs-Peres-Brandt attack on polarization-based BB84 is feasible with currently available technology, and we have argued that the development of polarization-preserving QND technology for measuring total photon number will permit mounting of a true deterministic FBP-probe attack. Our analysis has presumed ideal conditions in which Alice employs a single-photon source, there is no propagation loss and no extraneous (background) light collection, and both Eve and Bob have unity quantum efficiency photodetectors with no dark counts. Because current QKD systems typically employ attenuated laser sources, and suffer from propagation loss, photodetector inefficiencies, and extraneous counts, it behooves us to at least comment on how such non-idealities could impact the FPB probe we have described.
The use of an attenuated laser source poses no problem for the configurations shown in Figs.~6--9. This is because the single-qubit rotations and the CNOT gates of SPTQ quantum logic effect the same transformations on coherent states as they do on single-photon states. For example, the same half-wave plate setting that rotates the single-photon $|H\rangle$ qubit into the single-photon $|V\rangle$ qubit will transform the horizontally-polarized coherent state $|\alpha\rangle_H$ into the vertically-polarized coherent state $|\alpha\rangle_V$. Likewise, the SPTQ P-CNOT gate that transforms a single photon carrying polarization ($|H\rangle = |0\rangle, |V\rangle = |1\rangle$) and momentum ($|R\rangle = |0\rangle, |L\rangle = |1\rangle$) qubits according to \begin{eqnarray}
\lefteqn{c_{HR}|HR\rangle + c_{HL}|HL\rangle + c_{VR}|VR\rangle + c_{VL}|VL\rangle \longrightarrow } \nonumber \\
&&c_{HR}|HR\rangle + c_{HL}|HL\rangle + c_{VR}|VL\rangle + c_{VL}|VR\rangle, \end{eqnarray} will transform the four-mode coherent-state input with eigenvalues \begin{eqnarray} \lefteqn{\hspace*{-.5in} \left[\begin{array}{cccc} \langle\hat{a}_{HR}\rangle & \langle\hat{a}_{HL}\rangle & \langle\hat{a}_{VR}\rangle & \langle\hat{a}_{VL}\rangle \end{array}\right] = }\nonumber \\ &&\hspace*{.25in}\left[\begin{array}{cccc} \alpha_{HR} & \alpha_{HL} & \alpha_{VR} & \alpha_{VL}\end{array}\right], \end{eqnarray} into a four-mode coherent-state output with eigenvalues \begin{eqnarray} \lefteqn{\hspace*{-.5in}\left[\begin{array}{cccc} \langle\hat{a}_{HR}\rangle & \langle\hat{a}_{HL}\rangle & \langle\hat{a}_{VR}\rangle & \langle\hat{a}_{VL}\rangle \end{array}\right] = }\nonumber \\ &&\hspace*{.25in} \left[\begin{array}{cccc} \alpha_{HR} & \alpha_{HL} & \alpha_{VL} & \alpha_{VR}\end{array}\right], \end{eqnarray} where the $\hat{a}$'s are annihilation operators for modes labeled by their polarization and beam positions. It follows that the coherent-state $P_E$ and $I_B$ calculations mimic the qubit derivations that we presented in Sec.~III, with coherent-state inner products taking the place of qubit-state inner products. At low average photon number, these coherent-state results reduce to the qubit expressions for events which give rise to clicks in the photodetectors shown in Figs.~6--9.
Finally, a word about propagation loss, detector inefficiencies, and extraneous counts from dark current or background light is in order. All of these non-idealities actually help our Eve, in that they lead to a non-zero quantum bit error rate between Alice and Bob in the absence of the FPB attack. If Eve's $P_E$ value is set below that baseline error rate, then her presence should be undetectable.
\vspace*{.2in} \begin{acknowledgments} The authors acknowledge useful technical discussions with Howard Brandt, Jonathan Smith and Stewart Personick. This work was supported by the Department of Defense Multidisciplinary University Research Initiative program under Army Research Office grant DAAD-19-00-1-0177 and by MIT Lincoln Laboratory. \end{acknowledgments}
\appendix* \section{QND Measurement} Here we show that it is possible, in principle, to use cross-phase modulation between a strong coherent-state probe beam and an arbitrarily-polarized signal beam to make a QND measurement of the signal beam's total photon number. Let $\{\hat{a}_H, \hat{a}_V, \hat{a}_P\}$ be the annihilation operators of the horizontal and vertical polarizations of the signal beam and the (single-polarization) probe beam, respectively at the input to a cross-phase modulation interaction. We shall take that interaction to transform these annihilation operators according to the following commutator-preserving unitary operation, \begin{eqnarray} \hat{a}_H &\longrightarrow& \hat{a}'_H \equiv \exp(i\kappa \hat{a}_P^\dagger\hat{a}_P)\hat{a}_H \\ \hat{a}_V &\longrightarrow& \hat{a}'_V \equiv \exp(i\kappa \hat{a}_P^\dagger\hat{a}_P)\hat{a}_V\\ \hat{a}_P &\longrightarrow&\hat{a}'_P \equiv \exp[i\kappa(\hat{a}_H^\dagger\hat{a}_H + \hat{a}_V^\dagger\hat{a}_V)]\hat{a}_P, \end{eqnarray}
where $0 < \kappa \ll 1$ is the cross-phase modulation coupling coefficient. When the probe beam is in a strong coherent state, $|\sqrt{N}_P\rangle$ with $N_P\gg 1/\kappa^2$, the total photon number in the signal beam can be inferred from a homodyne-detection measurement of the appropriate probe quadrature. In particular, the state of $\hat{a}'_P$ will be $|\sqrt{N}_P\rangle$ when the signal beam's total photon number is zero, and its state will be $|(1+i\kappa)\sqrt{N}_P\rangle$ when the signal beam's total photon number is one, where $\kappa \ll 1$ has been employed. Homodyne detection of the $\hat{a}'_{P2} \equiv {\rm Im}(\hat{a}'_P)$ quadrature thus yields a classical random-variable outcome $\alpha'_{P2}$ that is Gaussian distributed with mean zero and variance 1/4, in the absence of a signal-beam photon, and Gaussian distributed with mean $\kappa\sqrt{N}_P$ and variance 1/4 in the presence of a signal-beam photon. Note that these conditional distributions are independent of the polarization state of the signal-beam photon when it is present. Using the decision rule, ``declare signal-beam photon present if and only if $\alpha'_{P2} > \kappa\sqrt{N}_P/2$,'' it is easily shown that the QND error probability is bounded above by $\exp(-\kappa^2 N_P/2)/2 \ll 1$.
The preceding polarization independent, low error probability QND detection of the signal beam's total photon number does \em not\/\rm\ disturb the polarization state of that beam. This is so because the probe imposes the same nonlinear phase shift on both the $H$ and $V$ polarizations of the signal beam. Hence, if the signal-beam input is in the arbitrarily-polarized single-photon state, \begin{equation}
|\psi_S\rangle = c_H |1\rangle_{H}|0\rangle_{V} + c_V |0\rangle_{H}|1\rangle_{V}, \vspace*{.075in} \end{equation}
where $|c_H|^2 + |c_V|^2 = 1$, then, except for a physically unimportant absolute phase, the signal-beam output will also be in the state $|\psi_S\rangle$.
\end{document}
|
arXiv
|
{
"id": "0508051.tex",
"language_detection_score": 0.844533383846283,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Harbourne constants and arrangements of lines on smooth hypersurfaces in $\mathbb{P}^3_{\mathbb{C}}$} \author{Piotr Pokora} \date{\today} \maketitle \thispagestyle{empty} \begin{abstract} In this note we find a bound for the so-called linear Harbourne constants for smooth hypersurfaces in $\mathbb{P}^{3}_{\mathbb{C}}$. \keywords{line configurations, Miyaoka inequality, blow-ups, negative curves, the Bounded Negativity Conjecture} \subclass{14C20, 14J70} \end{abstract}
\section{Introduction}
In this short note we find a global estimate for Harbourne constants which were introduced in \cite{BdRHHLPSz} in order to capture and measure
the bounded negativity on various birational models of an algebraic surface. \begin{definition}
Let $X$ be a smooth projective surface. We say that $X$ \emph{has bounded negativity}
if there exists an integer $b(X)$ such that for every \emph{reduced} curve $C \subset X$ one has the bound
$$C^{2} \geq -b(X).$$ \end{definition}
The bounded negativity conjecture (BNC for short) is one of the most intriguing problems in the theory of projective surfaces and
attracts currently a lot of attention, see \cite{Duke, BdRHHLPSz, Harbourne1, XR}. It can be formulated as follows. \begin{conjecture}[BNC]
An arbitrary smooth \emph{complex} projective surface has bounded negativity. \end{conjecture} Some surfaces are known to have bounded negativity (see \cite{Duke, Harbourne1}). For example, surfaces with $\mathbb{Q}$-effective anticanonical divisor such as Del Pezzo surfaces, K3 surfaces and Enriques surfaces have bounded negativity. However, when we replace these surfaces by their blow ups, we do not know if bounded negativity is preserved. Specifically, it is not known whether the blow up of $\mathbb{P}^{2}$ at ten general points has bounded negativity or not.
Recently in \cite{BdRHHLPSz} the authors have showed the following theorem. \begin{theorem}(\cite[Theorem~3.3]{BdRHHLPSz})
Let $\mathcal{L}$ be a line configuration on $\mathbb{P}^{2}_{\mathbb{C}}$. Let $f: X_{s} \rightarrow \mathbb{P}^{2}_{\mathbb{C}}$ be the blowing up at $s$ distinct points on $\mathbb{P}^{2}_{\mathbb{C}}$ and let $\widetilde{\call}$ be the strict transform of $\mathcal{L}$. Then we have $\widetilde{\call}^{2} \geq -4\cdot s$. \end{theorem}
In this note, we generalize this result to the case of line configurations on smooth hypersurfaces $S_{n}$ of degree $n \geq 3$ in $\mathbb{P}^{3}_{\mathbb{C}}$.
A classical result tells us that every smooth hypersurface of degree $n=3$ contains $27$ lines. For smooth hypersurfaces of degree $n=4$ we know that the upper bound of the number of lines on quartic surfaces is 64 (claimed by Segre \cite{Segre} and correctly proved by Sch\"utt and Rams \cite{SR}). In general, for degree $n \geq 3$ hypersurfaces $S_{n}$ Boissi\'ere and Sarti (see \cite[Proposition~6.2]{BS}) showed that the number of lines on $S_{n}$ is less than or equal to $n(7n-12)$.
Using techniques similar to the one introduced in \cite{BdRHHLPSz} we prove the following result. \begin{theoremA} Let $S_{n}$ be a smooth hypersurface of degree $n \geq 4$ in $\mathbb{P}^{3}_{\mathbb{C}}$. Let $\mathcal{L} \subset S_{n}$ be a line configuration, with the singular locus ${\rm Sing}(\mathcal{L})$ consisting of $s$ distinct points. Let $f : X_{s} \rightarrow S_{n} $ be the blowing up at ${\rm Sing}(\mathcal{L})$ and denote by $\widetilde{\call}$ the strict transform of $\mathcal{L}$. Then we have $$\widetilde{\call}^2 > -4s -2n(n-1)^2.$$ \end{theoremA} In the last part we study some line configurations on smooth complex cubics and quartics in detail. Similar systematic studies on line configurations on the projective plane were initiated in \cite{Szpond}.
\section{Bounded Negativity viewed by Harbourne Constants} We start with introducing the Harbourne constants \cite{BdRHHLPSz}.
\begin{definition}\label{def:H-constants}
Let $X$ be a smooth projective surface and let
$\mathcal{P}=\{ P_{1},\ldots,P_{s} \}$ be a set of mutually
distinct $s \geq 1$ points in $X$. Then the \emph{local Harbourne constant of $X$ at $\mathcal{P}$}
is defined as
\begin{equation}\label{eq:H-const for calp}
H(X;\mathcal{P}):= \inf_{C} \frac{\left(f^*C-\sum_{i=1}^s \mult_{P_i}C\cdot E_i\right)^2}{s},
\end{equation}
where $f: Y \to X$ is the blow-up of $X$ at the set $\mathcal{P}$
with exceptional divisors $E_{1},\ldots,E_s$ and the infimum is taken
over all \emph{reduced} curves $C\subset X$.\\
Similarly, we define the \emph{$s$--tuple Harbourne constant of $X$}
as
$$H(X;s):=\inf_{\mathcal{P}}H(X;\mathcal{P}),$$
where the infimum now is taken over all $s$--tuples of mutually
distinct points in $X$. \\
Finally, we define the \emph{global Harbourne constant of $X$} as
$$H(X):=\inf_{s \geq 1}H(X;s).$$ \end{definition}
The relation between Harbourne constants and the BNC can be expressed in the following way.
Suppose that $H(X)$ is a finite real number.
Then for any $s \geq 1$ and any reduced curve $D$ on the blow-up of $X$
at $s$ points, we have
$$D^2 \geq sH(X).$$
Hence the BNC holds on all blow ups of $X$ at $s$ mutually distinct points with the constant $b(X) = sH(X)$.
On the other hand, even if $H(X)=-\infty$, the BNC might still be true.
It is very hard to compute Harbourne constants in general. Moreover, it is quite tricky to find these numbers even for the simplest types of reduced curves on a well-understood surface.
\section{Proof of the main result} Given a configurations of lines on $S_{n}$ we denote by $t_{r}$ the number of its $r$-ple points, at which exactly $r$ lines of the configuration meet. In the sequel we will repeatedly use two elementary equalities, namely $\sum_{i} {\rm mult}_{P_{i}}(C) = \sum_{k \geq 2}kt_{k}$ and $\sum_{k\geq 2} t_{k} = s$. In this section we will study \emph{linear Harbourne constants} $H_{L}$. We define only the local linear Harbourne constant for $S_{n}$ containing a line configuration $\mathcal{L}$ since this is the only one difference comparing to Definition \ref{def:H-constants}. \begin{definition} Let $S_{n}$ be a smooth hypersurface of degree $n \geq 2$ in $\mathbb{P}^{3}_{\mathbb{C}}$ containing at least one line and let
$\mathcal{P}=\{ P_{1},\ldots,P_{s} \}$ be a set of mutually
distinct $s$ points in $S_{n}$. Then the \emph{local linear Harbourne constant of $S_{n}$ at $\mathcal{P}$}
is defined as
\begin{equation}
H_{L}(S_{n}; \mathcal{P}):= \inf_{\mathcal{L}} \frac{\widetilde{\call}^{2}}{s},
\end{equation} where $\widetilde{\call}$ is the strict transform of $\mathcal{L}$ with respect to the blow up $f : X_{s} \rightarrow S_{n} $ at $\mathcal{P}$
and the infimum is taken over all \emph{reduced} line configurations $\mathcal{L} \subset S_{n}$. \end{definition}
Our proof is based on the following result due to Miyaoka \cite[Section~2.4]{Miyaoka}.
\begin{theorem}
Let $S_{n}$ be a smooth hypersurface in $\mathbb{P}^{3}_{\mathbb{C}}$ of degree $n \geq 4$ containing a configuration of $d$ lines. Then one has
$$nd -t_{2} + \sum_{k \geq 3}(k-4)t_{k} \leq 2n(n-1)^2.$$
\end{theorem} Now we are ready to give a proof of the Main Theorem. \begin{proof} Pick a number $n \geq 4$. Recall that using the adjunction formulae one can compute the self-intersection number of a line $l$ on $S_{n}$, which is equal to $$l^{2} = -2 - K_{S_{n}}.l = -2 - \mathcal{O}(n-4).l = 2-n.$$ Observe that the local linear Harbourne constant at ${\rm Sing}(\mathcal{L})$ has the following form \begin{equation} \label{Hconst} H_{L}(S_{n}; {\rm Sing}(\mathcal{L})) = \frac{ (2-n)d + I_{d} - \sum_{k\geq 2}k^{2}t_{k}}{\sum_{k\geq 2}t_{k}}, \end{equation} where $I_{d} = 2 \sum_{i < j} l_{i}l_{j}$ denotes the number of incidences of $d$ lines $l_{1}, ..., l_{d}$. It is easy to see that we have the combinatorial equality $$I_{d} = \sum_{k \geq 2}(k^{2} - k)t_{k},$$ hence we obtain $$I_{d} -\sum_{k \geq 2}k^{2}t_{k} = -\sum_{k \geq 2}kt_{k}.$$ Applying this to (\ref{Hconst}) we get $$H_{L}(S_{n}; {\rm Sing}(\mathcal{L})) = \frac{ (2-n)d - \sum_{k \geq 2} kt_{k}}{\sum_{k \geq 2}t_{k}}.$$ Simple manipulations on the Miyaoka inequality lead to $$nd + t_{2} -4\sum_{k \geq 2}t_{k} - 2n(n-1)^2 \leq -\sum_{k\geq 2} kt_{k},$$ and finally we obtain $$H_{L}(S_{n}; {\rm Sing}(\mathcal{L})) \geq -4 + \frac{ 2d + t_{2} -2n(n-1)^2}{s},$$ which completes the proof. \end{proof} It is an interesting question how the linear Harbourne constant behaves when degree $n$ of a hypersurface grows. We present two extreme examples. \begin{example} Let us consider the Fermat hypersurface of degree $n \geq 3$ in $\mathbb{P}^{3}_{\mathbb{C}}$, which is given by the equation $$F_{n} \,\, : \,\, x^{n} + y^{n} + z^{n} + w^{n} = 0.$$ It is a classical result that on $F_{n}$ there exists the line configuration $\mathcal{L}_{n}$ consisting of $3n^{2}$ lines and delivers $3n^{3}$ double points and $6n$ points of multiplicity $n$. It is easy to check that $${\rm lim}_{n \rightarrow \infty} H_{L}(F_{n}; {\rm Sing}(\mathcal{L}_{n})) = {\rm lim}_{n \rightarrow \infty} \frac{ 3 n^{2} \cdot (2-n) + 12n^{3} - 6n^{2} - 4 \cdot 3n^3 - n^{2} \cdot 6n} {3 n^3 + 6n} = -3.$$ On the other hand, the Main Theorem gives $$ {\rm lim}_{n \rightarrow \infty} H_{L}(F_{n}; {\rm Sing}(\mathcal{L}_{n})) \geq -4 + {\rm lim}_{n \rightarrow \infty} \frac{ 6n^{2} + 3n^3 - 2n(n-1)^2}{3n^3 + 6n} = -3 \frac{2}{3},$$ which shows that the estimate given there is quite efficient. \end{example} \begin{example} \label{Rams} This construction comes from \cite{R}. Let us consider Rams hypersurface $\mathbb{P}^{3}_{\mathbb{C}}$ of degree $n \geq 6$ given by the equation $$R_{n} \, : \, x^{n-1}\cdot y + y^{n-1} \cdot z + z^{n-1} \cdot w + w^{n-1}\cdot x = 0.$$
On $R_{n}$ there exists a configuration $\mathcal{L}_{n}$ of $n(n-2)+4$ lines, which delivers exactly $2n^2 - 4n + 4$ double points -- this configuration is the grid of $n(n-2)+2$ vertical disjoint lines intersected by two horizontal disjoint lines. The local linear Harbourne constant at ${\rm Sing}(\mathcal{L}_{n})$ is equal to $$H_{L}(R_{n}; {\rm Sing}(\mathcal{L}_{n})) = \frac{ (n^2-2n+4)\cdot(2-n) + 4n^2 - 8n + 8 - 4\cdot(2n^2 - 4n + 4)}{2n^2 - 4n + 4} = \frac{-n^3}{2n^2 -4n + 4}.$$ Then ${\rm lim}_{n \rightarrow \infty} H_{L}(R_{n}; {\rm Sing}(\mathcal{L}_{n})) = - \infty.$ \end{example} Example \ref{Rams} presents a quite interesting phenomenon since we can obtain very low linear Harbourne constants having singularities of minimal orders -- the whole game is made by the large number of (disjoint) lines. \section{Smooth cubics and quartics} We start with the case $n = 3$. As we mentioned in the first section every smooth cubic surface contains $27$ line, and the configuration of these lines have only double and triple points. These triple points are called \emph{Eckardt points}. Now we find a lower bound for the linear Harbourne constant for such hypersurfaces. \begin{proposition} Under the above notation one has $$H_{L}(S_{3}; {\rm Sing}(\mathcal{L})) \geq -2 \frac{5}{11}.$$ \end{proposition} \begin{proof} Recall that the combinatorial equality \cite[Example II.20.]{Urzua} for cubic surfaces has the form $$135 = t_{2} + 3t_{3}.$$
Moreover, another classical result asserts that the maximal number of \emph{Eckardt} points is equal to $18$ and this number is obtained on Fermat cubic. In order to get a sharp lower bound for $H_{L}$ we need to consider the case when the number of Eckardt points is the largest. To see this we show that the linear Harbourne constant for $t$ triple points is greater then for $t+1$ triple points. Simple computations show that $$H_{L}(S_{3};t) = \frac{-297+3t}{135-2t},$$ $$H_{L}(S_{3};t+1) = \frac{-294+3t}{133-2t},$$ and $H_{L}(S_{3};t+1) < H_{L}(S_{3};t)$ iff $(-297+3t)\cdot(133-2t) - (-294+3t)\cdot(135-2t) > 0$ for all $t \in \{0, ..., 18\}$, which is obvious to check. Having this information in hand we can calculate that for $18$ triple points and $81$ double points the local linear Harbourne constant at ${\rm Sing}(\mathcal{L})$ is equal to $$H_{L}(F_{3};{\rm Sing}(\mathcal{L})) = \frac{27\cdot(-1) + 270 - 4\cdot81 - 9 \cdot 18}{99} = -2 \frac{5}{11},$$ which ends the proof. \end{proof} \begin{example} Now we consider the case $n=4$ and we start with the configuration of $64$ lines on Schur quartic $Sch$. It is well-known that every line from this configuration intersects exactly $18$ other lines -- see for instance \cite[Proposition 7.1]{SR}. One can check that these $64$ lines deliver $8$ quadruple points, $64$ triple points and $336$ double points (in \cite{Urzua} we can find that the number of double points is equal to $192$, which is false). Then the local linear Harbourne constant at ${\rm Sing}(\mathcal{L})$ is equal to $$H_{L}(Sch; {\rm Sing}(\mathcal{L})) = \frac{(-2)\cdot64 + 1152 -16\cdot8 - 9\cdot 64 - 4\cdot 336}{336 + 64 + 8} = -2.509.$$ \end{example} Now we present an example of a line configuration on a smooth quartic which deliver the most negative (according to our knowledge) local linear Harbourne constant for this kind of surfaces. \begin{example}[Bauer configuration of lines] Let us consider Fermat quartic $F_{4}$. It is well-known that on $F_{4}$ there exists the configuration of $48$ lines. From this configuration one can extract a subconfiguration of $16$ lines which has only $8$ quadruple points. Then the local linear Harbourne constant at ${\rm Sing}(\mathcal{L})$ is equal to $$H_{L}(F_{4}; {\rm Sing}(\mathcal{L})) = \frac{ 16\cdot(-2) + 16\cdot6 - 16 \cdot8}{8} = -8.$$
Using Main Theorem we get $H_{L}(F_{4}; {\rm Sing}(\mathcal{L})) \geq -9$, which also shows efficiency of our result. \end{example} \paragraph*{\emph{Acknowledgement.}} The author like to express his gratitude to Thomas Bauer for sharing Example 4.3, to S\l awomir Rams for pointing out his construction in \cite{R} and to Tomasz Szemberg and Halszka Tutaj-Gasi\'nska for useful remarks. Finally, the author would like to thank the anonymous referee for many useful comments which allowed to improve the exposition of this note. The author is partially supported by National Science Centre Poland Grant 2014/15/N/ST1/02102.
Piotr Pokora,
Instytut Matematyki,
Pedagogical University of Cracow,
Podchor\c a\.zych 2,
PL-30-084 Krak\'ow, Poland.
\nopagebreak
\textit{E-mail address:} \texttt{[email protected]}
\end{document}
|
arXiv
|
{
"id": "1505.03822.tex",
"language_detection_score": 0.7617802619934082,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\title{Qualitative behaviour and numerical approximation of solutions to conservation laws with non-local point constraints on the flux and modeling of crowd dynamics at the bottlenecks}
\author[Besancon]{Boris Andreianov} \ead{[email protected]}
\author[Besancon]{Carlotta Donadello} \ead{[email protected]}
\author[Besancon]{Ulrich Razafison\corref{mycorrespondingauthor}} \cortext[mycorrespondingauthor]{Corresponding author} \ead{[email protected]}
\author[Warsaw]{Massimiliano D.~Rosini} \ead{[email protected]}
\address[Besancon]{Laboratoire de Math\'ematiques, Universit\'e de Franche-Comt\'e,\\ 16 route de Gray, 25030 Besan\c{c}on Cedex, France}
\address[Warsaw]{ ICM, University of Warsaw\\ ul.~Prosta 69, P.O. Box 00-838, Warsaw, Poland}
\begin{abstract} In this paper we investigate numerically the model for pedestrian traffic proposed in [B.~Andreianov, C.~Donadello, M.D.~Rosini, Crowd dynamics and conservation laws with nonlocal constraints and capacity drop, Mathematical Models and Methods in Applied Sciences 24 (13) (2014) 2685-2722] . We prove the convergence of a scheme based on a constraint finite volume method and validate it with an explicit solution obtained in the above reference. We then perform \emph{ad hoc} simulations to qualitatively validate the model under consideration by proving its ability to reproduce typical phenomena at the bottlenecks, such as Faster Is Slower effect and the Braess' paradox. \end{abstract}
\begin{keyword}
finite volume scheme \sep scalar conservation law \sep non-local point constraint \sep crowd dynamics \sep capacity drop \sep Braess' paradox \sep Faster Is Slower
\MSC 35L65 \sep 90B20 \sep 65M12 \sep 76M12 \end{keyword}
\end{frontmatter}
\section{Introduction}\label{sec:intro}
\noindent Andreianov, Donadello and Rosini developed in~\cite{BorisCarlottaMax-M3AS} a macroscopic model, called here ADR, aiming at describing the behaviour of pedestrians at bottlenecks. The model is given by the Cauchy problem for a scalar hyperbolic conservation law in one space dimension with non-local point constraint of the form \begin{subequations}\label{eq:constrianed} \begin{align}\label{eq:constrianed1}
&\partial_t\rho + \partial_xf(\rho) = 0 && (t,x) \in {\mathbb{R}}_+\times{\mathbb{R}},\\
\label{eq:constrianed3}
&\rho(0,x) =\bar\rho(x) && x \in {\mathbb{R}},\\
\label{eq:constrianed2}
&f\left(\rho(t,0\pm)\right) \le p\left( \int_{{\mathbb{R}}_-} w(x)~ \rho(t,x)~{{\rm{d}}} x\right) && t \in {\mathbb{R}}_+, \end{align} \end{subequations} where $\rho(t,x) \in \left[0,R\right]$ is the (mean) density of pedestrians in $x \in {\mathbb{R}}$ at time $t \in {\mathbb{R}}_+$ and $\bar\rho \colon {\mathbb{R}} \to [0,R]$ is the initial (mean) density, with $R >0$ being the maximal density. Then, $f \colon [0,R] \to {\mathbb{R}}_+$ is the flow considered to be bell-shaped, which is an assumption commonly used in crowd dynamics. A typical example of such flow is the so-called Lighthill-Whitham-Richards (LWR) flux~\cite{LWR1, LWR2, Greenshields_1934} defined by $$f(\rho)=\rho\,v_{\max}\left(1-\dfrac{\rho}{\rho_{\max}}\right),$$where $v_{\max}$ and $\rho_{\max}$ are the maximal velocity and the maximal density of pedestrians respectively. Throughout this paper the LWR flux will be used. Next $p \colon {\mathbb{R}}_+ \to {\mathbb{R}}_+$ prescribes the maximal flow allowed through a bottleneck located at $x=0$ as a function of the weighted average density in a left neighbourhood of the bottleneck and $w \colon {\mathbb{R}}_- \to {\mathbb{R}}_+$ is the weight function used to average the density.
Finally in \eqref{eq:constrianed2}, $\rho(t,0-)$ denotes the left measure theoretic trace along the constraint, implicitly defined by \begin{align*} &\lim_{\varepsilon\downarrow0}\frac{1}{\varepsilon} \int_0^{+\infty} \int_{-\varepsilon}^0 \modulo{\rho(t,x) - \rho(t,0-)}~\phi(t,x)~{{\rm{d}}} x~{{\rm{d}}} t=0& &\text{for all }\phi \in \Cc\infty({\mathbb{R}}^2;{\mathbb{R}}). \end{align*} The right measure theoretic trace, $\rho(t,0+)$, is defined analogously.\\
\noindent In the last few decades, the study of the pedestrian behaviour through bottlenecks, namely at locations with reduced capacity, such as doors, stairs or narrowings, drawn a considerable attention. The papers \cite{Schadschneider2008Evacuation, Cepolina2009532, Hoogendoorn01052005, Kopylow, Schreckenberg2006, Seyfried:59568, Zhang20132781} present results of empirical experiments. However, for safety reasons, experiments reproducing extremal conditions such as evacuation and stampede are not available. In fact, the unique experimental study of a crowd disaster is proposed in~\cite{Helbing_disaster}. The available data show that the capacity of the bottleneck (i.e.~the maximum number of pedestrians that can flow through the bottleneck in a given time interval) can drop when high-density conditions occur upstream of the bottleneck. This phenomenon is called \emph{capacity drop} and can lead to extremely serious consequences in escape situations. In fact, the crowd pressure before an exit can reach very high values, the efficiency of the exit dramatically reduces and accidents become more probable due to the overcrowding and the increase of the evacuation time (i.e.~the temporal gap between the times in which the first and the last pedestrian pass through the bottleneck). A linked phenomenon is the so-called \emph{Faster Is Slower} (FIS) effect, first described in~\cite{Helbing2000Simulating}. FIS effect refers to the jamming and clogging at the bottlenecks, that result in an increase of the evacuation time when the degree of hurry of a crowd is high. We recall that the capacity drop and the FIS effect are both experimentally reproduced in~\cite{Cepolina2009532,Soria20121584}. A further related (partly counter-intuitive) phenomenon is the so-called \emph{Braess' paradox} for pedestrian flows~\cite{hughes2003flow}. It is well known that placing a small obstacle before an exit door can mitigate the inter-pedestrian pressure and, under particular circumstances, it reduces the evacuation time by improving the outflow of people.
Note that as it happens for any first order model, see for instance~\cite[Part~III]{Rosinibook} and the references therein, ADR can not explain the capacity drop and collective behaviours at the bottlenecks. Therefore one of the difficulties we have to face is that the constraint $p$ has to be deduced together with the fundamental diagram from the empirical observations.
The aim of this paper is to validate ADR by performing simulations in order to show the ability of the model to reproduce the main effects described above and related to capacity drop that are FIS and Braess' paradox. To this end we propose a numerical scheme for the model and prove its convergence. The scheme is obtained by adapting the local constrained finite volume method introduced in~\cite{scontrainte} to the non-local case considered in ADR, using a splitting strategy.
The paper is organized as follows. In Section~\ref{sec:model} we briefly recall the main theoretical results for ADR. In Section~\ref{sec:NumericalMethod} we introduce the numerical scheme, prove its convergence and validate it with an explicit solution obtained in \cite{BorisCarlottaMax-M3AS}. In Section~\ref{sec:simulations} we perform simulations to show that ADR is able to reproduce the Braess' paradox and the FIS effect. In Subsection~\ref{sec:Braess+FIS} we combine local and non-local constraints to model a slow zone placed before the exit. Conclusions and perspectives are outlined in Section~\ref{sec:conclusions}.
\section{Well-posedness for the ADR model}\label{sec:model}
Existence, uniqueness and stability for the general Cauchy problem~\eqref{eq:constrianed} are established in~\cite{BorisCarlottaMax-M3AS} under the following assumptions: \begin{enumerate}[label=\bfseries{(W)}]
\item[\textbf{(F)}] $f$ belongs to $\mathbf{Lip}\left( [0,R]; \left[0, +\infty\right[ \right)$ and is supposed to be bell-shaped, that is $f(0) = 0 = f(R)$ and there exists $\sigma \in \left]0,R\right[$ such that $f'(\rho)~(\sigma-\rho)>0$ for a.e.~$\rho \in [0,R]$.
\item[\textbf{(W)}] $w$ belongs to $\L\infty({\mathbb{R}}_-;{\mathbb{R}}_+)$, is an increasing map, $\norma{w}_{\L1({\mathbb{R}}_-)} = 1$ and there exists $\textrm{i}_{w} >0$ such that $w(x) = 0$ for any $x \le -\textrm{i}_{w}$.
\item[\textbf{(P)}] $p$ belongs to $\mathbf{Lip} \left( \left[0,R\right]; \left]0,f(\sigma)\right] \right)$ and is a non-increasing map. \end{enumerate}
The regularity $w\in \L\infty({\mathbb{R}}_-;{\mathbb{R}}_+)$ is the minimal requirement needed in order to prove existence and uniqueness of~\eqref{eq:constrianed}. In this paper, we shall consider continuous $w$.
The existence of solutions for the Riemann problem for~\eqref{eq:constrianed} is proved in~\cite{AndreianovDonadelloRosiniRazafisonProc} for piecewise constant $p$. However, such hypothesis on $p$ is not sufficient to ensure uniqueness of solutions, unless the flux $f$ and the efficiency $p$ satisfy a simple geometric condition, see~\cite{AndreianovDonadelloRosiniRazafisonProc} for details. In the present paper, we consider either continuous nonlinear $p$ or a piecewise constant $p$ that satisfies such geometric condition.
The definition of entropy solution for a Cauchy problem~\eqref{eq:constrianed1}, \eqref{eq:constrianed3} with a fixed \emph{a priori} time dependent constraint condition \begin{align}\label{eq:constrianed2bis}
&f\left(\rho(t,0\pm)\right) \le q(t) && t \in {\mathbb{R}}_+ \end{align} was introduced in~\cite[Definition~3.2]{ColomboGoatinConstraint} and then reformulated in~\cite[Definition~2.1]{scontrainte}, see also~\cite[Proposition~2.6]{scontrainte} and \cite[Definition 2.2]{chalonsgoatinseguin}. Such definitions are obtained by adding a term that accounts for the constraint in the classical definition of entropy solution given by Kruzkov in~\cite[Definition~1]{Kruzkov}. The definition of entropy solution given in~\cite[Definition~2.1]{BorisCarlottaMax-M3AS} is obtained by extending these definitions to the framework of non-local constraints.
The following theorem on existence, uniqueness and stability of entropy solutions of the constrained Cauchy problem~\eqref{eq:constrianed} is achieved under the hypotheses~\textbf{(F)}, \textbf{(W)} and~\textbf{(P)}. \begin{thm}[Theorem~3.1 in~\cite{BorisCarlottaMax-M3AS}]\label{thm:1}
Let~\textbf{(F)}, \textbf{(W)}, \textbf{(P)} hold. Then, for any initial datum $\bar\rho \in \L\infty({\mathbb{R}};[0,R])$, the Cauchy problem~\eqref{eq:constrianed} admits a unique entropy solution $\rho$. Moreover, if $\rho' = \rho'(t,x)$ is the entropy solution corresponding to the initial datum $\bar\rho' \in \L\infty({\mathbb{R}};[0,R])$, then for all~$T>0$ and $L>{\rm{i}_w}$, the following inequality holds \begin{equation}\label{eq:lipdepen-localized} \norma{\rho(T) - \rho'(T)}_{\L1([-L,L])} \le e^{CT}\norma{\bar\rho - \bar\rho'}_{\L1(\{\modulo{x}\le L+MT\})}, \end{equation} where $M=\mathrm{Lip}(f)$ and $C=2 \mathrm{Lip}(p) \norma{w}_{\L\infty({\mathbb{R}}_-)}$. \end{thm}
The total variation of the solution may in general increase due to the presence of the constraint. In~\cite{BorisCarlottaMax-M3AS} the authors provide an invariant domain $\mathcal{D} \subset \L1 \left({\mathbb{R}} ; [0,R]\right)$ such that if $\bar\rho$ belongs to $\mathcal{D}$, then one obtains a Lipschitz estimate with respect to time of the $\L1$ norm and an a priori estimate of the total variation of \[ \Psi(\rho) = \mathrm{sign}(\rho-\sigma) [f(\sigma) - f(\rho)] = \int_\sigma^\rho \modulo{\dot{f}(r)} \, {\rm{d}} r. \]
\section{Numerical method for approximation of ADR}\label{sec:NumericalMethod} In this section we describe the numerical scheme based on finite volume method that we use to solve~\eqref{eq:constrianed}. Then we prove the convergence of our scheme and validate it by comparison with an explicit solution of~\eqref{eq:constrianed}. In what follows, we assume that \textbf{(F)}, \textbf{(W)} and~\textbf{(P)} hold.
\subsection{Non-local constrained finite volume method}
Let $\Delta x$ and $\Delta t$ be the constant space and time steps respectively. We define the points $x_{j+1/2}=j\Delta x$, the cells $K_j=[x_{j-1/2},x_{j+1/2}[$ and the cell centers $x_j=(j-1/2)\Delta x$ for $j\in{\mathbb{Z}}$. We define the time discretization $t^n=n\Delta t$. We introduce the index $j_c$ such that $x_{j_c+1/2}$ is the location of the constraint (a door or an obstacle). For $n\in{\mathbb{N}}$ and $j\in{\mathbb{Z}}$, we denote by $\rho_j^n$ the approximation of the average of $\rho(t^n,\cdot~)$ on the cell $K_j$, namely \begin{align*} &\rho_j^0=\frac{1}{\Delta x}\displaystyle\int_{x_{j-1/2}}^{x_{j+1/2}}\overline{\rho}(x)\,{\rm{d}} x& &\text{ and }& &\rho_j^n\simeq\frac{1}{\Delta x}\displaystyle\int_{x_{j-1/2}}^{x_{j+1/2}}\rho(t^n,x)\,{\rm{d}} x &\text{ if }n>0. \end{align*}
We recall that for the classical conservation law~\eqref{eq:constrianed1}-\eqref{eq:constrianed3}, a standard finite volume method can be written into the form \begin{equation} \label{schema} \rho_j^{n+1}=\rho_j^n-\frac{\Delta t}{\Delta x}\left(\mathcal{F}_{j+1/2}^n-\mathcal{F}_{j-1/2}^n\right), \end{equation} where $\mathcal{F}_{j+1/2}^n=F\left(\rho_j^n,\rho_{j+1}^n\right)$ is a monotone, consistent numerical flux, that is, $F$ satisfies the following assumptions: \begin{itemize} \item $F$ is Lipschitz continuous from $[0,R]^2$ to ${\mathbb{R}}$ with Lipschitz constant $\mathrm{Lip}(F)$, \item $F(a,a)=f(a)$ for any $a\in[0,R]$, \item $(a,b) \in [0,R]^2 \mapsto F(a,b) \in {\mathbb{R}}$ is non-decreasing with respect to $a$ and non-increasing with respect to $b$. \end{itemize} We also recall that in~\cite{scontrainte} the numerical flux for the time dependent constraint~\eqref{eq:constrianed2bis} is modified as follow in order to take into account the constraint condition \begin{align} \label{def.flux.num.contrainte}
&\mathcal{F}_{j+1/2}^n=\left\{\begin{array}{l@{\quad\text{if }}l}
F\left (\rho_j^n,\rho_{j+1}^n\right )&j\ne j_c,\\[6pt]
\min\left \{F\left(\rho_j^n,\rho_{j+1}^n\right),q^n\right \}&j=j_c,
\end{array}\right. \end{align} where $q^n$ is an approximation of $q(t^n)$. In the present paper, when dealing with a Cauchy problem subject to a non-local constraint of the form~\eqref{eq:constrianed2} we will use the approximation \begin{equation} \label{contrainte.approx} q^n=p\left(\Delta x\sum_{j\le j_c}w(x_j)\,\rho_j^n\right). \end{equation} Roughly speaking \begin{itemize} \item we apply the numerical scheme~\eqref{schema} for the problem~\eqref{eq:constrianed1}-\eqref{eq:constrianed3}, \item we apply the numerical scheme~\eqref{schema}-\eqref{def.flux.num.contrainte} for the problem~\eqref{eq:constrianed1}-\eqref{eq:constrianed3}-\eqref{eq:constrianed2bis}, \item we apply the numerical scheme~\eqref{schema}-\eqref{def.flux.num.contrainte}-\eqref{contrainte.approx} for the problem~\eqref{eq:constrianed}. \end{itemize}
\subsection{Convergence of the scheme} Let us introduce the finite volume approximate solution $\rho_{\Delta}$ defined by \begin{align}\label{def.rho.delta} &\rho_\Delta(t,x)=\rho_j^n& &\mbox{for }x\in K_j\mbox{ and }t\in [t^n,t^{n+1} [, \end{align} where the sequence $(\rho_{j}^n)_{j\in{\mathbb{Z}},\,n\in{\mathbb{N}}}$ is obtained by the numerical scheme~\eqref{schema}-\eqref{def.flux.num.contrainte}. Analogously, we also define the approximate constraint function \begin{align} \label{def.q.delta} &q_\Delta(t)=q^n& &\mbox{for }t\in [t^n,t^{n+1} [. \end{align}
First, we prove a discrete stability estimate valid for any domain $Q = [0,T] \times {\mathbb{R}}$ with $T>0$, for the scheme~\eqref{schema}-\eqref{def.flux.num.contrainte} applied to problem~\eqref{eq:constrianed1}-\eqref{eq:constrianed3}-\eqref{eq:constrianed2bis}. This estimate can be seen as the equivalent, in this framework, of the stability result established in~\cite[Proposition~2.10]{scontrainte}.
\begin{pro} \label{pro.diff.solutions} Let $\overline{\rho}$ be in $\L\infty({\mathbb{R}};[0,R])$ and $q_\Delta$, $\hat{q}_\Delta$ be piecewise constant functions of the form~\eqref{def.q.delta}. If $\rho_\Delta$ and $\hat{\rho}_\Delta$ are the approximate solutions of~\eqref{eq:constrianed1}-\eqref{eq:constrianed3}-\eqref{eq:constrianed2bis} corresponding, respectively, to $q_\Delta$ and $\hat{q}_\Delta$ and constructed by applying the scheme~\eqref{schema}-\eqref{def.flux.num.contrainte}, then we have $$\norma{\rho_\Delta-\hat{\rho}_\Delta}_{\L1(Q)}\le2 T \norma{q_\Delta-\hat{q}_\Delta}_{\L1([0,T])}.$$ \end{pro}
\begin{proof} For notational simplicity, let $N = \lfloor T/\Delta t\rfloor$. Let us also introduce $(\tilde{\rho}_j^n)_{j\in{\mathbb{Z}},\,n\in{\mathbb{N}}}$ defined by, $$\tilde{\rho}_j^{n+1}=\rho_j^n-\frac{\Delta t}{\Delta x}\left(\mathcal{\tilde{F}}_{j+1/2}^n-\mathcal{\tilde{F}}_{j-1/2}^n\right),\quad\mbox{for any } j\in{\mathbb{Z}},\,n\in{\mathbb{N}},$$ where $\mathcal{\tilde{F}}_{j+1/2}^n$ is defined by \begin{align*}
&\mathcal{\tilde{F}}_{j+1/2}^n=\left\{\begin{array}{l@{\quad\text{if }}l}
F\left (\rho_j^n,\rho_{j+1}^n\right )&j\ne j_c,\\[6pt]
\min\left \{F\left(\rho_j^n,\rho_{j+1}^n\right),\hat{q}^n\right \}&j=j_c.
\end{array}\right. \end{align*}
Then using the definitions of $(\rho_j^n)_{j\in{\mathbb{Z}},\,n\in{\mathbb{N}}}$ and $(\tilde{\rho}_j^n)_{j\in{\mathbb{Z}},\,n\in{\mathbb{N}}}$, we have for any $n=1,\dots,N$, $$\rho_j^n=\tilde{\rho}_j^n\quad\mbox{if}\quad j\notin\{j_c,j_c+1\}$$ and \begin{align*} &\rho_{j_c}^n-\tilde{\rho}_{j_c}^n= -\frac{\Delta t}{\Delta x}\left(\min\left \{F\left (\rho_{j_c}^{n-1},\rho_{j_c+1}^{n-1}\right ),q^{n-1}\right\} +\min\left \{F\left (\rho_{j_c}^{n-1},\rho_{j_c+1}^{n-1}\right ),\hat{q}^{n-1}\right \}\right),&\\ &\rho_{j_c+1}^n-\tilde{\rho}_{j_c+1}^n= \frac{\Delta t}{\Delta x}\left(\min\left \{F\left(\rho_{j_c}^{n-1},\rho_{j_c+1}^{n-1}\right),q^{n-1}\right \} - \min\left \{F\left(\rho_{j_c}^{n-1},\rho_{j_c+1}^{n-1}\right),\hat{q}^{n-1}\right \} \right), \end{align*} which implies that \begin{align*} &\modulo{\rho_{j_c}^n-\tilde{\rho}_{j_c}^n}\le\frac{\Delta t}{\Delta x} ~ \modulo{q^{n-1}-\hat{q}^{n-1}},& &\modulo{\rho_{j_c+1}^n-\tilde{\rho}_{j_c+1}^n}\le\frac{\Delta t}{\Delta x} ~ \modulo{q^{n-1}-\hat{q}^{n-1}}. \end{align*} Therefore we deduce that, for any $n=1,\dots,N$, \begin{equation} \label{pro.diff.solutions.eq1} \sum_{j\in{\mathbb{Z}}}\modulo{\rho_{j}^n-\tilde{\rho}_{j}^n}\le2\frac{\Delta t}{\Delta x} ~ \modulo{q^{n-1}-\hat{q}^{n-1}}. \end{equation} Besides, observe that the modification of the numerical flux at the interface $x_{j_c+1/2}$ introduced in~\eqref{def.flux.num.contrainte} does not affect the monotonicity of the scheme~\eqref{schema}-\eqref{def.flux.num.contrainte} (see~\cite[Proposition~4.2]{scontrainte}). Therefore, for any $n=1,\dots,N$, we have \begin{equation} \label{pro.diff.solutions.eq2} \sum_{j\in{\mathbb{Z}}}\modulo{\tilde{\rho}_j^n-\hat{\rho}_j^n}\le \sum_{j\in{\mathbb{Z}}}\modulo{\rho_j^{n-1}-\hat{\rho}_j^{n-1}}. \end{equation} Hence thanks to~\eqref{pro.diff.solutions.eq1} and~\eqref{pro.diff.solutions.eq2}, we can write \begin{align*}
\sum_{j\in{\mathbb{Z}}}|\rho_j^1-\hat{\rho}_j^1|&\le\sum_{j\in{\mathbb{Z}}}|\rho_j^1-\tilde{\rho}_j^1|+\sum_{j\in{\mathbb{Z}}}|\tilde{\rho}_j^1-\hat{\rho}_j^1|\le2\frac{\Delta t}{\Delta x} ~ \modulo{q^0-\hat{q}^0}+\sum_{j\in{\mathbb{Z}}}|\rho_j^0-\hat{\rho}_j^0| =2\frac{\Delta t}{\Delta x} ~ \modulo{q^0-\hat{q}^0}. \end{align*} Then an induction argument shows that for any $n=1,\dots,N$,
$$\sum_{j\in{\mathbb{Z}}}\modulo{\rho_j^n-\hat{\rho}_j^n}\le2\frac{\Delta t}{\Delta x}\sum_{k=0}^{n-1}|q^k-\hat{q}^k|\le\frac{2}{\Delta x}~\|q_\Delta-\hat{q}_\Delta\|_{L^1([0,t^n])}.$$ In conclusion, we find that \begin{align*}
\|\rho_\Delta-\hat{\rho}_\Delta\|_{L^1(Q)}&=\Delta t\,\Delta x\sum_{n=1}^N\sum_{j\in{\mathbb{Z}}}|\rho_j^n-\hat{\rho}_j^n|
\le 2 \|q_\Delta-\hat{q}_\Delta\|_{L^1([0,T])}~\sum_{n=1}^N \Delta t
\le 2 T \|q_\Delta-\hat{q}_\Delta\|_{L^1([0,T])} \end{align*} and this ends the proof. \end{proof}
\begin{comment} \begin{proof} Assume first $n=1$. Then we easily see from the scheme~\eqref{schema}-\eqref{def.flux.num.contrainte} that we have $\rho_j^1=\hat{\rho}_j^1$ for any $j\not\in\left \{ j_c, j_c+1\right \}$. Moreover we can write \begin{align*} &\rho_{j_c}^1-\hat{\rho}_{j_c}^1= -\frac{\Delta t}{\Delta x}\left(\min\left \{F\left (\rho_{j_c}^0,\rho_{j_c+1}^0\right ),q^0\right\} -\min\left \{F\left (\rho_{j_c}^0,\rho_{j_c+1}^0\right ),\hat{q}^0\right \}\right),&\\ &\rho_{j_c+1}^1-\hat{\rho}_{j_c+1}^1= \frac{\Delta t}{\Delta x}\left(\min\left \{F\left(\rho_{j_c}^0,\rho_{j_c+1}^0\right),q^0\right \} - \min\left \{F\left(\rho_{j_c}^0,\rho_{j_c+1}^0\right),\hat{q}^0\right \} \right), \end{align*} and therefore \begin{align*} &\modulo{\rho_{j_c}^1-\hat{\rho}_{j_c}^1}\le\frac{\Delta t}{\Delta x} ~ \modulo{q^0-\hat{q}^0},& &\modulo{\rho_{j_c+1}^1-\hat{\rho}_{j_c+1}^1}\le\frac{\Delta t}{\Delta x} ~ \modulo{q^0-\hat{q}^0}. \end{align*} Then an induction argument allows to prove that for any $n\in{\mathbb{N}}$, $n\ge1$, we have \begin{align*} &\rho_j^n=\hat{\rho}_j^n\quad\mbox{if }j\not\in \left \{j_c,\,j_c+1\right \},& &\modulo{\rho_{j_c}^n-\hat{\rho}_{j_c}^n} \le \frac{\Delta t}{\Delta x} ~ \modulo{q^{n-1}-\hat{q}^{n-1}},& &\modulo{\rho_{j_c+1}^n-\hat{\rho}_{j_c+1}^{n}} \le \frac{\Delta t}{\Delta x} ~ \modulo{q^{n-1}-\hat{q}^{n-1}}. \end{align*} Therefore we have \begin{align*} \norma{\rho_{\Delta}-\hat{\rho}_{\Delta}}_{\L1(Q)}&=\Delta t ~ \Delta x\sum_{n=0}^{\lfloor T/\Delta t\rfloor}\sum_{j\in{\mathbb{Z}}}\modulo{\rho_j^n-\hat{\rho}_j^n} =\Delta t ~ \Delta x\sum_{n=0}^{\lfloor T/\Delta t\rfloor}\left(\modulo{\rho_{j_c}^n-\hat{\rho}_{j_c}^n} + \modulo{\rho_{j_c+1}^n-\hat{\rho}_{j_c+1}^n}\right)\\ &\le2 ~ \Delta t^2\sum_{n=1}^{\lfloor T/\Delta t\rfloor}\modulo{q^{n-1}-\hat{q}^{n-1}} \le2 ~ \Delta t ~ \norma{q_\Delta-\hat{q}_\Delta}_{\L1([0,T])}, \end{align*} which ends the proof. \end{proof} \end{comment}
\begin{comment} \begin{proof} For notational simplicity, let $N = \lfloor T/\Delta t\rfloor$. By definition we have that \begin{align*} &\norma{\rho_{\Delta}-\hat{\rho}_{\Delta}}_{\L1(Q)}= \Delta t ~ \Delta x \left [\sum_{n=1}^N \sum_{i=1-n}^{n} \modulo{\rho^n_{j_c+i} - \hat{\rho}^n_{j_c+i}}\right ] \end{align*} where by~\eqref{CFL} \begin{align*} &\sum_{n=1}^N \sum_{i=1-n}^{n} \modulo{\rho^n_{j_c+i} - \hat{\rho}^n_{j_c+i}} \leq \sum_{n=2}^N \sum_{i=1-n}^{n} \modulo{\rho^{n-1}_{j_c+i} - \hat{\rho}^{n-1}_{j_c+i}} +2 ~ \frac{\Delta t}{\Delta x} \sum_{n=1}^N \modulo{q^{n-1}-\hat{q}^{n-1}} \\ &+\frac{\Delta t}{\Delta x} ~ \mathrm{Lip}(F) \sum_{n=2}^N \sum_{i=1-n}^{n} \left [ \modulo{\rho^{n-1}_{j_c+i-1} - \hat{\rho}^{n-1}_{j_c+i-1}} +2\modulo{\rho^{n-1}_{j_c+i} - \hat{\rho}^{n-1}_{j_c+i}} +\modulo{\rho^{n-1}_{j_c+i+1} - \hat{\rho}^{n-1}_{j_c+i+1}} \right ] \\ &\leq 2 ~ \frac{\Delta t}{\Delta x} \sum_{n=1}^N \modulo{q^{n-1}-\hat{q}^{n-1}} +\left [1+4\frac{\Delta t}{\Delta x} ~ \mathrm{Lip}(F)\right ] \left [\sum_{n=2}^N \sum_{i=1-n}^{n} \modulo{\rho^{n-1}_{j_c+i} - \hat{\rho}^{n-1}_{j_c+i}}\right ] \\ &\leq 2 ~ \frac{\Delta t}{\Delta x} \sum_{n=1}^N \modulo{q^{n-1}-\hat{q}^{n-1}} +3 \sum_{n=1}^{N-1} \sum_{i=1-n}^{n} \modulo{\rho^{n}_{j_c+i} - \hat{\rho}^{n}_{j_c+i}} \end{align*} \end{proof} \end{comment}
\noindent Let us now notice that as in~\cite[Proposition~4.2]{scontrainte}, under the CFL condition \begin{equation} \label{CFL} \mathrm{Lip}(F) ~ \frac{\Delta t}{\Delta x}\le\frac{1}{2}, \end{equation} we have the $\L\infty$ stability of the scheme~\eqref{schema}-\eqref{def.flux.num.contrainte}-\eqref{contrainte.approx} that is \begin{align} \label{stability.Linfty} &0\le\rho_\Delta(t,x)\le R&&\mbox{for a.e.~}(t,x)\in Q. \end{align} \noindent This stability result allows to prove the statement below.
\begin{pro} \label{prop.estimation.bv.q} Let $q_\Delta$ be defined by~\eqref{contrainte.approx}-\eqref{def.q.delta}. Then under the CFL condition~\eqref{CFL}, for any $T>0$, there exists $C>0$ only depending on $T$, $f$, $F$, $p$, $w$ and $R$ such that: \begin{equation} \label{estimation.bv.q} \modulo{q_\Delta}_{BV([0,T])}\le C. \end{equation} \end{pro}
\begin{proof} Let $N=\lfloor{T/\Delta t}\rfloor$ and $j_w$ be an integer such that supp$(w)\subset\underset{j_w\le j\le j_c}\cup K_j$. Then for any $n=0,\dots,N-1$, we have \begin{align*} \modulo{q^{n+1}-q^n}& = \modulo{p\left(\Delta x\sum_{j_w\le j\le j_c}w(x_j)\rho_j^{n+1}\right)-p\left(\Delta x\sum_{j_w\le j\le j_c}w(x_j)\rho_j^n\right)}\\
&\le\Delta x\,\mathrm{Lip}(p)\left|\sum_{j_{w}\le j\le j_c}w(x_j)(\rho_j^{n+1}-\rho_j^n)\right|
=\Delta t\,\mathrm{Lip}(p)\left|\sum_{j_{w}\le j\le j_c}w(x_j)\left(\mathcal{F}_{j+1/2}^n-\mathcal{F}_{j-1/2}^n\right)\right|. \end{align*} Now, using a summation by part, we have \begin{align*} \sum_{j_{w}\le j\le j_c}w(x_j)\left(\mathcal{F}_{j+1/2}^n-\mathcal{F}_{j-1/2}^n\right) =w(x_{j_c})\mathcal{F}_{j_c+1/2}^n-w(x_{j_w})\mathcal{F}_{j_w-1/2} - \!\!\!\sum_{j_w\le j\le j_c-1}\left(w(x_{j+1})-w(x_j)\right)\mathcal{F}_{j+1/2}^n. \end{align*} Then, it follows that
$$|q^{n+1}-q^n|\le\Delta t\,\mathrm{Lip}(p)\,\|w\|_{\L\infty({\mathbb{R}}_-;{\mathbb{R}})}\sum_{j_w-1\le j\le j_c}|\mathcal{F}_{j+1/2}^n|.$$ Now, from~\eqref{def.flux.num.contrainte}, for any $j\in{\mathbb{Z}}$ we have the estimate \begin{align*} \modulo{\mathcal{F}_{j+1/2}^n} & \le \modulo{F(\rho_{j}^n,\rho_{j+1}^n)} \le \modulo{F(\rho_{j}^n,\rho_{j+1}^n)-F(\rho_{j}^n,\rho_{j}^n)} + \modulo{f(\rho_{j}^n)} \le \mathrm{Lip}(F) \, \modulo{\rho_{j+1}^n-\rho_{j}^n}+\mathrm{Lip}(f)\, \modulo{\rho_{j}^n} \le R\left (\mathrm{Lip}(F)+\mathrm{Lip}(f)\right ). \end{align*} Hence we deduce that $$\modulo{q_\Delta}_{BV([0,T])}=\sum_{n=0}^{N-1}\modulo{q^{n+1}-q^n}\le C,$$ where $C=(j_c-j_w+2)\,T\,R\,\mathrm{Lip}(p)\,\norma{w}_{\L\infty({\mathbb{R}}_-;{\mathbb{R}})}\,\left (\mathrm{Lip}(F)+\mathrm{Lip}(f)\right )$. \end{proof}
We are now in a position to prove a convergence result for the scheme ~\eqref{schema}-\eqref{def.flux.num.contrainte}-\eqref{contrainte.approx}.
\begin{thm} \label{th.convergence} Under the CFL condition~\eqref{CFL}, the constrainted finite volume scheme~\eqref{schema}-\eqref{def.flux.num.contrainte}-\eqref{contrainte.approx} converges in $\L1(Q)$ to the unique entropy solution to~\eqref{eq:constrianed}. \end{thm}
\begin{proof} Let $(\rho_\Delta,q_\Delta)$ be constructed by the scheme~\eqref{schema}-\eqref{def.flux.num.contrainte}-\eqref{contrainte.approx}. Proposition~\ref{prop.estimation.bv.q} and Helly's lemma give the existence of a subsequence, still denoted $q_\Delta$ and a constraint function $q\in \L\infty([0,T])$ such that $q_\Delta$ converges to $q$ strongly in $\L1([0,T])$ as $\Delta t\to0$. Let $\rho\in \L\infty({\mathbb{R}}_+\times{\mathbb{R}};[0,R])$ be the unique entropy solution to~\eqref{eq:constrianed1}-\eqref{eq:constrianed3}-\eqref{eq:constrianed2bis} associated to $q$. It remains to prove that the subsequence $\rho_\Delta$ converges to $\rho$ strongly in $\L1(Q)$ as $\Delta t,\,\Delta x\to0$. The uniqueness of the entropy solution to~\eqref{eq:constrianed1}-\eqref{eq:constrianed3}-\eqref{eq:constrianed2bis} will then imply that the full sequence $\rho_\Delta$ converges to $\rho$ and, as a consequence, the full sequence $q_\Delta$ converges to $q=p\left (\int_{{\mathbb{R}}_-}w(x)\,\rho(t,x)\,{\rm{d}} x\right )$.\\ Let $\hat{q}_\Delta$ be a piecewise constant approximation of $q$ such that $\hat{q}_\Delta$ converges to $q$ strongly in $\L1([0,T])$. Furthermore, we also introduce $\hat{\rho}_\Delta$ constructed by the scheme~\eqref{schema}-\eqref{def.flux.num.contrainte} and associated to $\hat{q}_\Delta$. Now we have $$\norma{\rho-\rho_\Delta}_{\L1(Q)}\le\norma{\rho-\hat{\rho}_\Delta}_{\L1(Q)}+\norma{\rho_\Delta-\hat{\rho}_\Delta}_{\L1(Q)}. $$ But, thanks to~\cite[Theorem~4.9]{scontrainte}, under the CFL condition~\eqref{CFL}, $\norma{\rho-\hat{\rho}_\Delta}_{\L1(Q)}$ tends to $0$ as $\Delta t$, $\Delta x\to0$. Furthermore, thanks to Proposition~\ref{pro.diff.solutions}, we have $$\norma{\rho_\Delta-\hat{\rho}_\Delta}_{\L1(Q)}\le2 ~ T~ \norma{q_\Delta-\hat{q}_\Delta}_{\L1([0,T])}$$ which also shows that $\norma{\rho_\Delta-\hat{\rho}_\Delta}_{\L1(Q)}$ tends to $0$ as $\Delta t$, $\Delta x\to0$. \end{proof}
\subsection{Validation of the numerical scheme}\label{sec:validation}
\begin{figure}
\caption{The functions ${[\rho \mapsto f(\rho)]}$ and ${[\xi \mapsto p(\xi)]}$ as in Section~\ref{sec:validation}.}
\label{fig:validation1}
\end{figure}
\begin{figure}
\caption{Representation of the solution constructed in~\cite[Section~6]{BorisCarlottaMax-M3AS} and described in Subsection \ref{sec:validation}.}
\label{fig:gull1}
\label{fig:gull2}
\label{fig:gull3}
\label{fig:gull4}
\label{fig:validation2}
\end{figure}
We propose here to validate the numerical scheme~\eqref{schema}-\eqref{def.flux.num.contrainte}-\eqref{contrainte.approx} using the Godounov numerical flux (see e.g.~\cite{GodlewskiRaviartBook, LevequeBook}) which will be used in the remaining of this paper: \begin{eqnarray*}
F(a,b) &= \left\{
\begin{array}{l@{\quad\text{ if }}l}
\underset{[a,b]}\min f & a\le b,\\
\underset{[b,a]}\max f & a>b.
\end{array}
\right. \end{eqnarray*}
\noindent We consider the explicit solution to~\eqref{eq:constrianed} constructed in~\cite[Section~6]{BorisCarlottaMax-M3AS} by applying the wave front tracking algorithm. The set up for the simulation is as follows. Consider the domain of computation $[-6,1]$, take a normalized flux $f(\rho) = \rho(1-\rho)$ (namely the maximal velocity and the maximal density are assumed to be equal to one) and a linear weight function $w(x) = 2 (1 + x)\,\chi_{[-1, 0]}(x)$. Assume a uniform distribution of maximal density in $[x_A, x_B]$ at time $t=0$, namely $\bar\rho = \chi_{[x_A, x_B]}$. The efficiency of the exit, $p$, see Figure~\ref{fig:validation1}, is of the form \begin{eqnarray*}
p(\xi) &= \left\{
\begin{array}{l@{\quad\text{ if }}l}
p_0 & 0\le\xi<\xi_1,\\
p_1 & \xi_1\le\xi<\xi_2,\\
p_2 & \xi_2\le\xi\le 1.
\end{array}
\right. \end{eqnarray*} The explicit solution $\rho$ corresponding to the values \begin{align*}
&p_0 =0.21, && p_1 =0.168, && p_2 =0.021, && \xi_1 \sim 0.566,
&x_A = -5.75, && x_B =-2, && \xi_2 \sim 0.731, \end{align*} is represented in Figure~\ref{fig:validation2}. The above choices for the flux $f$ and the efficiency $p$ ensure that the solution to each Riemann problem is unique, see~\cite{AndreianovDonadelloRosiniRazafisonProc}. We defer to~\cite[Section~6]{BorisCarlottaMax-M3AS} for the details of the construction of the solution $\rho$ and its physical interpretation. \begin{figure}
\caption{With reference to Subsection \ref{sec:validation}: The numerically computed solution $x \mapsto \rho_\Delta(t,x)$ and the explicitly computed solution $x \mapsto \rho(t,x)$ at different fixed times $t$.}
\label{fig:tiger1}
\label{fig:tiger2}
\label{fig:tiger3}
\label{fig:tiger4}
\label{fig:zebra1}
\label{fig:zebra2}
\label{fig:zebra3}
\label{fig:zebra4}
\label{fig:tiger5}
\label{fig:tiger6}
\label{fig:tiger7}
\label{fig:tiger8}
\label{fig:zebra5}
\label{fig:zebra6}
\label{fig:zebra7}
\label{fig:zebra8}
\label{fig:validation3}
\end{figure}
\noindent A qualitative comparison between the numerically computed solution $x \mapsto \rho_\Delta(t,x)$ and the explicitly computed solution $x \mapsto \rho(t,x)$ at different fixed times $t$ is in Figure~\ref{fig:validation3}. We observe good agreements between $x \mapsto \rho(t,x)$ and $x \mapsto\rho_\Delta(t,x)$. The parameters for the numerically computed solution are $\Delta x=3.5\times10^{-4}$ and $\Delta t=7\times10^{-5}$.\\ A convergence analysis is also performed for this test. We introduce the relative $\L1$-error for the density $\rho$, at a given time $t^n$, defined by
$$E_{\L1}^n=\left [\sum_{j}\left|\rho(t^n,x_j)-\rho_j^n\right|\right ]\,\Big/\left [\sum_{j}\left|\rho(t^n,x_j)\right|\right ].$$ In Table~\ref{erreurs}, we computed the relative $\L1$-errors for different numbers of space cells at the fixed time $t=10$. We deduce that the order of convergence is approximatively $0.906$. As in~\cite{scontrainte}, we observe that the modification~\eqref{def.flux.num.contrainte} of the numerical flux does not affect the accuracy of the scheme.
\begin{table}[ht] \begin{center}
\begin{tabular}{|c|c|} \hline Number of cells & $\L1$-error\\ \hline $625$ & $9.6843\times10^{-3}$\\ \hline $1250$ & $6.2514\times10^{-3}$\\ \hline $2500$ & $3.4143\times10^{-3}$\\ \hline $5000$ & $1.3172\times10^{-3}$\\ \hline $10000$ & $1.03\times10^{-3}$\\ \hline $20000$ & $4.2544\times10^{-4}$\\ \hline Order & $0.906$\\ \hline \end{tabular} \tiny\caption{Relative $\L1$-error at time $t=10$.} \label{erreurs} \end{center} \end{table}
\section{Numerical simulations}\label{sec:simulations}
This section is devoted to the phenomenological description of some collective effects in crowd dynamics related to capacity drop, namely the Braess' paradox and the Faster Is Slower (FIS) effect.
\subsection{Faster is Slower effect}\label{sec:FIS}
The FIS effect was first described in~\cite{Helbing2000Simulating, Parisi2005606} in the context of the room evacuation problem. The authors studied the evolution of the evacuation time as a function of the maximal velocity reached by the pedestrians, and they shown that there exists an optimal velocity for which the evacuation time attains a minimum. Therefore, any acceleration beyond the optimal velocity worses the evacuation time. Following the studies above, the curve representing the evacuation time as a function of the average velocity takes a characteristic shape \cite[Figure 1c]{Parisi2005606}.
The first numerical tests we performed aim to verify if such shape is obtained starting from the ADR model. To this end, we consider the corridor modeled by the segment [-6,1], with an exit at $x=0$. We consider the flux $f(\rho)=\rho \, v_{\max} \, (1-\rho)$ where $v_{\max}$ is the maximal velocity of the pedestrians and the maximal density is equal to one. We use the same weight function as for the validation of the scheme, $w(x)=2(1+x)\chi_{[-1,0]}(x)$ and, the same initial density, $\bar{\rho}=\chi_{[-5.75,-2]}$. The efficiency of the exit $p$ is now given by the following continuous function \begin{eqnarray} \label{P_continuous}
p(\xi) &= \left\{
\begin{array}{l@{\quad\text{ if }}l}
p_0 & 0\le\xi<\xi_1,\\[6pt]
\displaystyle\frac{(p_0-p_1)\xi+p_1\xi_1-p_0\xi_2}{\xi_1-\xi_2} & \xi_1\le\xi<\xi_2,\\[10pt]
p_1 & \xi_2\le\xi\le1,
\end{array}
\right. \end{eqnarray} where \begin{align*}
&p_0 =0.24, && p_1 =0.05, && \xi_1=0.5, && \xi_2=0.9. \end{align*}
The space and time steps are fixed to $\Delta x=5\times10^{-3}$ and $\Delta t=5\times10^{-4}$. In Figure~\ref{fis_f_p} are plotted the flux $f$ corresponding to the maximal velocity $v_{\max}=1$ and the above efficiency of the exit.
\begin{figure}\label{fis_f_p}
\end{figure}
\begin{figure}\label{fis_exit_times_velocities}
\end{figure}
\begin{figure}
\caption{With reference to Subsection \ref{sec:FIS}: Densities at the exit as a function of time for different velocities.}
\label{fis_densities_doors}
\end{figure}
\begin{figure}\label{p_beta}
\end{figure}
\begin{figure}\label{fis_stabilities}
\end{figure}
Figure~\ref{fis_exit_times_velocities} represents the evacuation time as a function of the maximal velocity $v_{\max}$, as $v_{\max}$ varies in the interval $[0.1 , 5]$. As we can observe, the general shape described above is recovered. The numerical minimal evacuation time is $19.007$ and is obtained for $v_{\max}=1$.
In addition, we reported in Figure~\ref{fis_densities_doors} the density at the exit as a function of time for different values of the maximal velocity $v_{\max}$ around the optimal one. We notice that the maximal density at the exit and the time length where the density is maximal increase with the velocity. This expresses the jamming at the exit that leads to the FIS effect.
\begin{figure}
\caption{With reference to Subsection \ref{sec:Braess}: Evacuation time as a function of the position of the obstacle.}
\label{braess_exit_times}
\end{figure}
\begin{figure}
\caption{With reference to Subsection \ref{sec:Braess}: Braess paradox simulations: density profiles at times $t=1$ (first line), $t=7$ (second line), $t=15$ (third line), $t=19$ (fourth line) and $t=24.246$ (last line).}
\label{Braess_snapshots}
\end{figure}
Then we performed some series of tests to see how the general shape obtained in Figure~\ref{fis_exit_times_velocities} changes with respect to variations of the parameters of the model.
In Figure~\ref{fis_stabilities}~(a), we show this variation when we consider different initial densities, namely, $\bar\rho$, $\bar{\rho}_1$ and $\bar{\rho}_2$ with $\bar\rho_1(x)=0.8\chi_{[-5.75,-2]}$ and $\bar\rho_2(x)=0.6\chi_{[-5.75,-2]}$. The general shape of the curves is conserved. We observe that the evacuation time increases with the initial amount of pedestrians while the optimal velocity decreases as the initial amount of pedestrians increases. The minimal evacuation time and the corresponding optimal maximal velocity are $12.259$ and $1.07$ for $\bar{\rho}_2$ and $15.691$ and $1.03$ for $\bar{\rho}_1$.
Next we explore the case where the efficiency of the exit varies. We consider the function $p$ defined in~\eqref{P_continuous} and the modification $p_\beta$ such that $p_\beta(\xi)=p(\beta\xi)$. In Figure~\ref{p_beta}, we plotted the functions $p$, $p_\beta$ for $\beta=0.8$ and $\beta=0.9$. Then, in Figure~\ref{fis_stabilities}~(b) are plotted the evacuation time curves corresponding to these three efficiencies of the exit. As minimum evacuation times, we obtain $18.586$ and $18.827$ for $\beta=0.8$, $0.9$ respectively. As expected, the minimal evacuation time increases with lower efficiency of the exit. The corresponding velocities are approximatively $1.06$ and $1.02$ respectively.
Finally, we change the location of the initial density. In addition to the corridor $[-6,1]$, we consider two other corridors modeled by the segments $[-12,1]$ and $[-20,1]$. In these two corridors we take as initial densities $\bar{\rho}_3(x)=\chi_{[-11.75,-8]}$ and $\bar{\rho}_4(x)=\chi_{[-19.75,-16]}$ respectively. We have reported the obtained evacuation time curves in Figure~\ref{fis_stabilities}~(c). As expected, the minimal evacuation time increases with the distance between the exit and the initial density location.
\subsection{Braess' paradox}\label{sec:Braess}
The presence of obstacles, such as columns upstream from the exit, may prevent the crowd density from reaching dangerous values and may actually help to minimize the evacuation time, since in a moderate density regime the full capacity of the exit can be exploited. From a microscopic point of view, the decrease of the evacuation time may seem unexpected, as some of the pedestrians are forced to chose a longer path to reach the exit.
The ADR model is able to reproduce the Braess' paradox for pedestrians, as we show in the following simulations. We consider, as in the previous subsection, the corridor modeled by the segment $[-6,1]$ with an exit at $x=0$. We compute the solution corresponding to the flux $f(\rho)=\rho(1-\rho)$, the initial density $\bar\rho(x)=\chi_{[-5.75,-2]}(x)$, the efficiency of the exit $p$ of the form~\eqref{P_continuous} with the parameters \begin{align*}
&p_0 =0.21, && p_1 =0.1, && \xi_1=0.566, && \xi_2=0.731 \end{align*} and the same weight function $w(x)=2(1+x)\chi_{[-1,0]}(x)$. The space and time steps are fixed to $\Delta x=5\times10^{-3}$ and $\Delta t=5\times10^{-4}$. Without any obstacle, the numerical evacuation time is $29.496$. In these following simulations we place an obstacle at $x=d$, with $-2<d<0$. The obstacle reduces the capacity of the corridor and can be seen as a door, which we assume larger than the one at $x=0$. Following these ideas we define an efficiency function $p_d(\xi)=1.15 p(\xi)$ and a weight function $w_d(x)=2(x-d+1)\chi_{[d-1,d]}(x)$ associated to the obstacle.
In Figure~\ref{braess_exit_times} we have reported the evolution of the evacuation time when the position of the obstacle varies in the interval $[-1.9,-0.01]$ with a step of 0.01. We observe that for $-1.8\le d\le-1.72$, the evacuation time is lower than in the absence of the obstacle. The optimal position of the obstacle is obtained for $d=-1.72$ and the corresponding evacuation time is $24.246$. We compare in Figure~\ref{Braess_snapshots} five snapshots of the solution without obstacle and the solutions with an obstacle placed at $d=-1.72$ and $d=-1.85$. This latter location corresponds to a case where the evacuation time is greater than the one without an obstacle. In these snapshots, we see that the obstacle placed at $d=-1.85$ becomes congested very soon. This is due to the fact that the obstacle is too close to the location of the initial density. When the obstacle is placed at $d=-1.72$, it delays the congestion at the exit.
\subsection{Zone of low velocity}\label{sec:Braess+FIS}
\begin{figure}\label{slowzone}
\end{figure}
\noindent In this section, we perform a series of simulations where the obstacle introduced in Subsection~\ref{sec:Braess} is now replaced by a zone where the velocity of pedestrians is lower than elsewhere in the domain. The effect we want to observe here is similar to the one we see in Braess' Paradox. Namely we prevent an high concentration of pedestrians in front of the exit by constraining their flow in an upstream portion of the corridor. In this case however the constraint is local, as the maximal value allowed for the flow only depends on the position in the corridor.
\begin{figure}
\caption{With reference to Subsection \ref{sec:Braess+FIS}: Braess' paradox and zone of low velocity simulations: density profiles at times $t=1$ (first line), $t=7$ (second line), $t=15$ (third line), $t=19$ (fourth line) and $t=20.945$ (last line).}
\label{column_snapshots}
\end{figure} We consider again the corridor modeled by the segment $[-6,1]$ with an exit at $x=0$. The efficiency of the exit and the initial density are the same as in the previous subsection. Assume that the slow zone is of size one and is centred at $x=d$, where $-1.9\le d\le0$. Define the following function
\begin{eqnarray} \label{def_k}
k(x) &= \left\{
\begin{array}{l@{\quad\text{ if }}l}
1 & x\le d-0.5,\\[6pt]
-2(x-d) & d-0.5\le x\le d,\\[10pt]
2(x-d) & d\le x\le d+0.5,\\[10pt]
1 & x\ge d+0.5,
\end{array}
\right. \end{eqnarray} and the following velocity $v(x,\rho)= \left[\lambda+(1-\lambda) \, k(x)\right] v_{\max} \, (1-\rho)$, where $\lambda\in[0,1]$ and $v_{\max}\ge1$ is the maximal velocity. With such velocity, the maximal velocity of pedestrians decreases in the interval $[d-0.5,d]$, reaching its minimal value $\lambda \, v_{\max}$ at $x=d$. Then the velocity increases in the interval $[d,d+0.5]$ reaching the maximum value $v_{\max}$, that corresponds to the maximal velocity away from the slow zone. Finally we consider the flux $f(x,\rho)=\rho\,v(x,\rho)$ and the space and time steps are fixed to $\Delta x=5\times10^{-3}$ and $\Delta t=5\times10^{-4}$.
\noindent Figure~\ref{slowzone}~(a) shows the evolution of the evacuation time as a function of the parameter $\lambda$ varying in the interval $[0.1,1]$ when the center of the slow zone is fixed at $d=-1.5$. We observe that the optimal minimal velocity in the slow zone is for $\lambda=0.88$ and the corresponding evacuation time is $20.945$. Recalling that without the slow zone the evacuation time is $29.496$, we see that the introduction of the slow zone allows to reduce the evacuation time. In Figure~\ref{slowzone}~(b), we show the evolution of the evacuation time when varying the center of the slow zone $d$ in the interval $[-1.9,0]$ and when the minimal and the maximal velocities are fixed and correspond to $\lambda=0.88$ and $v_{\max}=1$. We observe here that, unlike in the Braess paradox tests case, the evacuation time does not depend on the location of the slow zone, except when this latter is close enough to the exit. Indeed, when the slow zone gets too close to the exit, the evacuation time grows. This is due to the fact that pedestrians do not have time to speed up before reaching the exit.
Fix now $d=-1.5$ and $\lambda=0.88$ and assume that $v_{\max}$ varies in the interval $[0.1,5]$. The evolution of the evacuation time as a function of $v_{\max}$ is reported in Figure~\ref{slowzone}~(c). We observe that we get the characteristic shape already obtained in the FIS effect.
Finally we present in Figure~\ref{column_snapshots} five snapshots for three different solutions. The first two solutions are the ones computed in Subsection~\ref{sec:Braess}, without obstacle and with an obstacle located at $d=~-1.72$ respectively. The third solution is computed with a zone of low velocity centered at $d=-1.72$, $\lambda=0.88$ and $v_{\max}=1$. In order to have a good resolution of this third solution, the space and time steps where fixed to $\Delta x=3.5\times10^{-4}$ and $\Delta t=7\times10^{-5}$.We note that in the case where a zone of low velocity is placed in the domain, we do not see the capacity drop, as the density of pedestrians never attains very high values in the region next to the exit.
\section{Conclusions}\label{sec:conclusions}
Qualitative features that are characteristic of pedestrians' macroscopic behaviour at bottlenecks (Faster is Slower, Braess' paradox) are reproduced in the setting of the simple scalar model with non-local point constraint introduced in~\cite{BorisCarlottaMax-M3AS}. These effects are shown to be persistent for large intervals of values of parameters. The validation is done by means of a simple and robust time-explicit splitting finite volume scheme which is proved to be convergent, with experimental rate close to one.
The results presented in this paper allow to consider more complex models. Indeed, as ADR is a first order model, it is not able to capture more complicated effects related to crowd dynamics. Typically, ADR fails to reproduce the amplification of small perturbations. This leads to consider second order model such as the model proposed by Aw, Rascle and Zhang~\cite{Aw_SIAP_2000, Zhang_Method_2002} in the framework of vehicular traffic.
Another extension of this work is to consider the ADR model with constraints that are non-local in time. Such constraints allow to tackle optimal management problems in the spirit of \cite{ColomboFacchiMaterniniRosini, CGRESAIM}.
Finally, this work can also be extended to two-dimensional models where experimental validations may be possible.
\section*{Acknowledgment}
All the authors are supported by French ANR JCJC grant CoToCoLa and Polonium 2014 (French-Polish cooperation program) No.331460NC. The first author is grateful to IRMAR, Universit\'e de Rennes, for the hospitality during the preparation of this paper. The second author is also supported by the Universit\'e de Franche-Comt\'e, soutien aux EC 2014.
Projekt zosta{\l} sfinansowany ze \'srodk\'ow Narodowego Centrum Nauki przyznanych na podstawie decyzji nr: DEC-2011/01/B/ST1/03965.
\section*{References}
\end{document}
|
arXiv
|
{
"id": "1503.02826.tex",
"language_detection_score": 0.6587545275688171,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\title{Projection sparse principal component analysis: \\ An efficient least squares method} \if00{ \author[A1]{Giovanni Maria Merola\corref{cor1}} \author[A1,A2]{Gemai Chen} \address[A1]{Department of Mathematical Sciences, Xi'an Jiaotong-Liverpool University, 111 Ren€™ai Road,\\Suzhou Industrial Park, Suzhou, Jiangsu Province, P.R. China 215123} \address[A2]{Department of Mathematics and Statistics, University of Calgary, Calgary, Alberta, Canada T2N IN4 } \cortext[cor1]{Corresponding author. Email address: \url{[email protected]}} }\fi
\begin{abstract} We propose a new sparse principal component analysis (SPCA) method in which the solutions are obtained by projecting the full cardinality principal components onto subsets of variables. The resulting components are guaranteed to explain a given proportion of variance. The computation of these solutions is very efficient. The proposed method compares well with the optimal least squares sparse components. We show that other SPCA methods fail to identify the best sparse approximations of the principal components and explain less variance than our solutions. We illustrate and compare our method with others with extensive simulations and with the analysis of the computational results for nine datasets of increasing dimensions up to 16,000 variables. \end{abstract}
\begin{keyword} Dimension reduction \sep Power method \sep SPCA \sep Variable selection. \end{keyword} \end{frontmatter}
\section{Introduction}
Principal components analysis (PCA) is the oldest and most popular data dimensionality reduction method used to approximate a set of variables in a lower dimensional space \cite{pea}. Effective use of the method can approximate a large number of variables by a few linear combinations of them, called principal components (PCs). PCA has been extensively used over the past century and in recent times the interest in this method has surged, due to the availability of very large datasets. Applications of PCA include the analysis of gene expression analysis, market segmentation, handwriting classification, image recognition, and other types of data.
PCs are usually difficult to interpret and not informative on important features of the dataset because they are combinations of all the observed variables, as already pointed out by \citet{jef}. A common approach used to increase their interpretability is to threshold the coefficients of the combinations defining the PCs, which are called loadings. That is, variables corresponding to loadings that are lower than a given threshold are ignored. However, this practice can give misleading results \citep{jol} and the retained variables can be highly correlated among themselves. This means that the variables included in the interpretation actually carry similar information.
In recent years a large number of methods for sparse principal components analysis (SPCA) have been proposed; see, e.g., \citep{jol00, mog, zou, sri, she, wan}. These methods compute solutions in which some of the coefficients to be estimated are equal to zero. In addition to increased interpretability of the results, sparse methods are recommended under the sparsity principle \citep{has}.
Conventional SPCA methods replace the ordinary PCs with PCs of subsets (blocks) of variables. The resulting sparse PCs (SPCs) are combinations of only a few of the observed variables. That is, the SPCs are linear combinations of all the variables with only few loadings not equal to zero, the number of which is called cardinality. The difference among conventional SPCA methods is in the optimization approach used to select the variables to be included in the blocks. In this context, the variable selection problem is non-convex NP-hard \citep{mog}, hence computationally intractable. Some methods use a genuine cardinality penalty (improperly called $\ell_0$ norm), others an $\ell_1$ penalty. The most popular of these methods seems to be the Lasso based SPCA \citep{zou}.
SPCA methods are expressly recommended for large fat datasets \citep{has2}, i.e., samples with fewer observations than variables. By the nature of the objective function maximized, the components computed maximize the variance explained of each block, instead of that of the whole data matrix. As a consequence, the selected blocks contain highly correlated variables \citep{mer}. Furthermore, with this approach the exact sparse reduction of the PCs of fat matrices cannot be identified, as we will show later.
A least squares SPCA method (LS SPCA) in which the sparse components are obtained by minimizing the $\ell_2$ norm of the approximation error was proposed in \cite{mer}. This approach produces sparse components that explain the largest possible proportion of variance for a given cardinality. LS SPCA can identify the equivalent sparse representation of the PCs of fat matrices. However, the variable selection approaches suggested are not scalable to large matrices because they are top-down and require the computation of (generalized) eigenvectors of large matrices.
In this paper we suggest an efficient variable selection strategy for LS SPCA, based on projecting the full cardinality PCs on blocks of variables. This approach is based on a property we prove which says that if the regression of a PC on a block of variables yields an $R^2$ statistics equal to $\alpha \in (0, 1)$, then the LS SPCA components computed on that block of variables will explain a proportion of the variance not smaller than $\alpha$. With this approach, the NP-hard SPCA variable selection problem is reduced to a more manageable univariate regression variable selection problem. This procedure, to which we refer to as Projection SPCA (PSPCA), also gives as by-products the projections of the PCs, which are sparse components in themselves.
We show that algorithms using PSPCA variable selection are very efficient for computing LS SPCA components, having a growth order of about the number of variables squared. We also show that the performance of the projected PCs is comparable to that of the LS SPCA components. This is relevant because these projections are easier to understand by researchers and also easier and more economical to compute.
In the next section we review PCA and LS SPCA, and give a novel interpretation of the latter. The methodological details of PSPCA are discussed in Section~\ref{sec:pspca}. In Section~\ref{sec:selectblocks} we discuss the use of PSPCA for variable selection. We compare its performance on fat matrices with that of conventional SPCA methods and explain the details of the PSPCA algorithm. The proposed method is illustrated by using simulated and real datasets in Section~\ref{sec:examples}. We give some final comments in Section~\ref{sec:conclusions}. The Appendix contains some of the proofs.
\section{Full cardinality and sparse principal components}\label{sec:sparsecomp}
We assume that ${\mathbf{X}}$ is an $n\times p$ matrix containing $n$ observations on $p$ mean centred variables which have been scaled so that ${\mathbf{S}} = {\mathbf{X}}\trasp{\mathbf{X}}$ is the sample covariance or correlation matrix. Because of this, we will use the terms uncorrelated (correlated) and nonorthogonal (orthogonal) interchangeably. The information contained in the dataset is summarized by its total variance, defined by the squared (Frobenius) norm $||{\mathbf{X}}||^2 = \mathrm{tr}({\mathbf{S}})$, where $\mathrm{tr}$ is the trace operator. In the following the term norm refers to this norm, unless otherwise specified.
A component is any linear combination of the columns of ${\mathbf{X}}$, generically denoted by ${\mathbf{t}} = {\mathbf{X}}\mathbf{a}$, where the vector $\mathbf{a}$ is the vector of loadings (or just the loadings, for short). A set of ordered components $({\mathbf{t}}_1, \ldots, {\mathbf{t}}_d) = {\mathbf{X}}(\mathbf{a}_1, \ldots, \mathbf{a}_d)$ is denoted as ${\mathbf{T}_{[d]}} = {\mathbf{X}}{\mathbf{A}_{[d]}}$, where the subscript ${[j]}$ denotes the first $j$ columns of a matrix and ${\mathbf{A}_{[d]}} = (\mathbf{a}_1, \ldots, \mathbf{a}_d)$ is called the matrix of loadings.
The least squares estimates of $d$ components, with $d \le p$, are obtained by minimizing the squared norm of the difference of the data matrix from its projection onto the components, $\Pi_{{\mathbf{T}_{[d]}}}{\mathbf{X}}$, where $\Pi_{{\mathbf{M}}}={\mathbf{M}}({\mathbf{M}}\trasp{\mathbf{M}})\matinv{\mathbf{M}}\trasp$ denotes the projector onto the column space of the matrix ${\mathbf{M}}$. By Pythagoras' theorem, the solutions must satisfy \begin{equation}\label{eq:vexp}
{\mathbf{T}_{[d]}} = \argmin\limits_{{\mathbf{M}}\in \mathbb{R}^{n\times d}} ||{\mathbf{X}} - \Pi_{{\mathbf{M}}}{\mathbf{X}}||^2 = \argmax\limits_{{\mathbf{M}}\in \mathbb{R}^{n\times d}} ||\Pi_{{\mathbf{M}}}{\mathbf{X}}||^2. \end{equation} The term on the right-hand side of Eq.~(\ref{eq:vexp}), called the variance explained by the components ${\mathbf{T}_{[d]}}$ and denoted as $\mathrm{vexp}({\mathbf{T}_{[d]}})$, is used to measure the performance of the components in approximating the data and is equal to \begin{align} \mathrm{vexp}({\mathbf{T}_{[d]}}) = \mathrm{tr} \Big \{ {\mathbf{X}}\trasp{\mathbf{T}_{[d]}} ({\mathbf{T}\traspd{[d]}}{\mathbf{T}_{[d]}} )^{-1}{\mathbf{T}\traspd{[d]}}{\mathbf{X}} \Big \} = \mathrm{tr} \Big \{ {\mathbf{S}}{\mathbf{A}_{[d]}} ({\mathbf{A}\traspd{[d]}}{\mathbf{S}}{\mathbf{A}_{[d]}} )^{-1}{\mathbf{A}\traspd{[d]}}{\mathbf{S}} \Big \}. \label{eq:vexp1} \end{align}
\subsection{Principal components}
The principal components, denoted as ${\mathbf{u}}_j = {\mathbf{X}}{\mathbf{v}}_j$ with $j \in \{ 1,\ldots, d\}$ and $d \leq \mathrm{rank}(S)$, are obtained by maximizing $\mathrm{vexp}({\mathbf{T}_{[d]}})$ under the orthogonality requirements ${\mathbf{u}}\trasps{i}{\mathbf{u}}_j = 0$ if $i\neq j$. That is, the PCs' loadings are found from \begin{align}\label{eq:pcaprob} \forall_{j \in \{ 1, \ldots, d\}} \quad {\mathbf{v}_j} = &\argmax\limits_{{\mathbf{a}_j}\in\mathbb{R}^p} {{\mathbf{a}\trasps{j}}{\mathbf{X}}\trasp{\mathbf{X}}{\mathbf{X}}\trasp{\mathbf{X}}{\mathbf{a}_j}}/{{\mathbf{a}\trasps{j}}{\mathbf{X}\trasp}{\mathbf{X}}{\mathbf{a}_j}}, \\
&\text{subject to}\, {\mathbf{v}}\trasps{i}{\mathbf{S}}{\mathbf{a}_j}=\mathbf{0},\, i < j, \,\text{if } j >1.\notag \end{align} The solution loadings ${\mathbf{v}}_j$ are the eigenvectors of ${\mathbf{S}}$, such that ${\mathbf{S}}{\mathbf{v}}_j = {\mathbf{v}}_j\lambda_j$, corresponding to the eigenvalues in nondecreasing order, $\lambda_1 \geq \cdots \geq \lambda_d$. By the orthogonality of the components, the total variance explained can be broken down as the sum of the individual variances explained as $$ \mathrm{vexp}\left({\mathbf{U}}_{[d]}\right) = \sum_{j=1}^d\mathrm{vexp}({\mathbf{u}_j}) = \sum_{j=1}^d {{\mathbf{u}\trasps{j}}{\mathbf{X}}{\mathbf{X}\trasp}{\mathbf{u}_j}}/{{\mathbf{u}\trasps{j}}{\mathbf{u}_j}} = \sum_{j=1}^d \lambda_j. $$
The PCs can be regarded as solutions to a number of different problems, such as the singular value decomposition \citep{eck}. Most notably, \citet{hot} showed that when the loadings are scaled to unit norm, we have $\mathrm{vexp}({\mathbf{u}}_j) ={\mathbf{u}}\trasps{j}{\mathbf{u}}_j = \lambda_j$. Noting also that, by the properties of the spectral decomposition of a symmetric matrix, the loadings are orthogonal, then the PCA problem can be formulated as finding \begin{align}\label{eq:pcahot}
\forall_{j \in \{ 1, \ldots, d\}} \quad {\mathbf{v}}_j =& \argmax_{\mathbf{a}_j\in\mathbb{R}^p} \mathbf{a}\trasps{j}{\mathbf{S}}\mathbf{a}_j, \\
&\text{subject to } \mathbf{a}\trasps{j}\mathbf{a}_j = 1, \mathbf{a}\trasps{j}{\mathbf{v}}_i = 0,\, \text{if}\ j > i \geq 1.\nonumber \end{align}
\subsection{Least squares sparse principal components}
A sparse component is a linear combination of a subset, $\dX_j$, of columns of the ${\mathbf{X}}$ matrix, or block of variables, defined by the sparse loadings $\da_j$ as ${\mathbf{t}}_j = \dX_j\da_j$. The number of variables in the block is the cardinality of the loadings. The least squares sparse PCA (LS SPCA) problem is defined by adding sparsity constraints directly into the PCA objective \eqref{eq:pcaprob}, which gives \begin{align}\label{eq:spcaprob}
\forall_{j \in \{ 1, \ldots, d\}} \quad \dbj =& \argmax\limits_{\daj\in\mathbb{R}^{c_j}} {\dajT\dXjT{\mathbf{X}}{\mathbf{X}}\trasp\dXj\daj}/{\dajT\dXjT\dXj\daj} \\
& \text{subject to}\, {\mathbf{b}}\traspd{i}{\mathbf{S}}{\mathbf{a}_j} = 0,\, i < j, \,\text{if } j >1,\nonumber \end{align} where $\dXj$ is a block of $c_j$ variables, $\mathbf{a}_j = {\mathbf{J}}_j\daj$ and ${\mathbf{b}_j} = {\mathbf{J}}_j\dbj$ are the full cardinality representations of the sparse loadings, defined by means of the matrix ${\mathbf{J}}_j$, which is formed by the columns of the order-$p$ identity matrix corresponding to the variables in $\dXj$. Note that the SPCA objective must be maximized sequentially. We will refer to the components in the order with which they are computed. The solutions are given in the following proposition \citep{mer}.
\begin{proposition}[Uncorrelated LS SPCA]\label{prop:uspca} Given a block of $c_j$ linearly independent variables $\dX_j$, the solutions to objective \eqref{eq:spcaprob} are the generalized eigenvectors satisfying \begin{equation}\label{eq:uspca} {\mathbf{C}}_j\dXjT{\mathbf{X}}{\mathbf{X}\trasp}\dXj\dbj = {\mathbf{C}}_j{\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j\dbj = \dSj\dbj\xi_{j}, \end{equation} where $$ {\mathbf{C}}_j = \{ {\mathbf{I}}_{c_j} - {\mathbf{H}\trasps{j}}({\mathbf{H}_j}\bdS\matinvj{j}{\mathbf{H}\trasps{j}})\matinv{\mathbf{H}_j}\bdS\matinvj{j} \} , \quad {\mathbf{C}}_1 = {\mathbf{I}}_{c_1}, $$ the sparse loadings $\dbj$, the orthogonality constraints ${\mathbf{H}_j} = {\mathbf{Y}}\trasps{[j-1]}\dXj$, with $\bdS_j = \dXjT\dXj$ and $\xi_{j} = \mathrm{vexp}({\mathbf{y}}_j)$ the largest eigenvalue. The SPCs ${\mathbf{y}}_j = \dXj\dbj$ are mutually orthogonal and maximize the variance explained. \end{proposition}
The optimal orthogonal LS SPCA components are highly constrained; for example their cardinality cannot be smaller than their order. Due to the greedy nature of the optimization carried out, these locally optimal orthogonal components often stride away from the global optimum, while globally better solutions can be found by removing the orthogonality constraints \cite{mer}.
If the orthogonality constraints are dropped, the net increment in total variance explained due to a new component is the variance explained by the residuals orthogonal to the components already in the model. This extra variance explained, which we denote as $\mathrm{evexp}$, is equal to \begin{align}
\mathrm{evexp}({\mathbf{t}_j}) &= ||\Pi_{[\bQTj{T}{\mathbf{a}_j}]}{\mathbf{X}}||^2 =
{{\mathbf{a}\trasps{j}}\bQTjT{T}{\mathbf{X}}{\mathbf{X}}\trasp\bQTj{T}{\mathbf{a}_j}}/{{\mathbf{a}\trasps{j}}\bQTjT{T}\bQTj{T}{\mathbf{a}_j}}\nonumber\\
&= {{\mathbf{a}\trasps{j}}{\mathbf{X}}\trasp\bQTj{T}\bQTjT{T}{\mathbf{X}}{\mathbf{a}_j}}/{{\mathbf{a}\trasps{j}}\bQTjT{T}\bQTj{T}{\mathbf{a}_j}}, \label{eq:evexp} \end{align} where $\bQTj{T} = ({\mathbf{I}}_n - \Pi_{{\mathbf{T}}_{[j-1]}}){\mathbf{X}}$ is the orthogonal complement of the ${\mathbf{X}}$ matrix (${\mathbf{Q}}_{T_{[0]}} = {\mathbf{X}}$), to which we refer as deflated ${\mathbf{X}}$ matrix (with respect to ${\mathbf{T}}_{[j-1]}$), and $\bQTj{T}\mathbf{a}_j = {\mathbf{t}}_j - \Pi_{{\mathbf{T}}_{[j-1]}} {\mathbf{t}}_j$ is the orthogonal residual of ${\mathbf{t}}_j$. For the first component $\mathrm{evexp}({\mathbf{t}}_1) = \mathrm{vexp}({\mathbf{t}}_1)$. The total variance explained by a set of correlated components is equal to the sum of the extra variances explained, viz. $$ \mathrm{vexp}({\mathbf{T}_{[d]}}) = \sum_{j = 1}^d \mathrm{evexp}({\mathbf{t}}_j). $$
The sparse solutions cannot be determined from the maximization of objective \eqref{eq:evexp} because this is defined in terms of the deflated components $\bQTj{T}\mathbf{a}_j$, while the cardinality constraints must be imposed on the $x$-variables.
For this reason, \citet{mer} derives nonorthogonal SPCs ${\mathbf{z}}_j = {\mathbf{X}}{\mathbf{d}}_j = \dXj\ddj$ which maximize the variance of $\bQTj{Z}$ explained by a component ${\mathbf{z}}_j$. This is defined in terms of the variables $\dXj$ as \begin{equation}\label{eq:vexpq}
\mathrm{vexp}_Q({\mathbf{z}_j}) = ||\Pi_{{\mathbf{z}}_j}\bQTj{Z}||^2 = {{\mathbf{z}}\trasps{j}\bQTj{Z}\bQTjT{Z}{\mathbf{z}_j}}/{{\mathbf{z}}\trasps{j}{\mathbf{z}_j}} = {\ddjT\dXjT\bQTj{Z}\bQTjT{Z}\dXj\ddj}/{\ddjT\dXjT\dXj\ddj}. \end{equation}
\begin{proposition}[Correlated LS SPCA] Given a block of linearly independent variables $\dX_j$, the correlated SPCs, ${\mathbf{z}}_j = \dXj\ddj = {\mathbf{X}}{\mathbf{d}}_j$, that successively maximize $\mathrm{vexp}_Q$ are the generalized eigenvectors satisfying \begin{equation}\label{eq:lsspcasol}
\dXjT\bQTj{Z}\bQTjT{Z}\dXj\ddj = \dXjT\dXj\ddj\gamma_j = \dSj\ddj\gamma_j, \end{equation} where $\gamma_j$ is the largest generalized eigenvalue, which is equal to $\mathrm{vexp}_Q({\mathbf{z}_j})$. The full cardinality loadings are equal to ${\mathbf{d}}_j = {\mathbf{J}}_j\ddj$. The first component is identical to the first orthogonal component. \end{proposition}
We will refer to the SPCs derived from the minimization of the least squares criterion generically as LS SPCA components and use USPCA and CSPCA to refer to the uncorrelated and correlated solutions, respectively.
The variance of ${\mathbf{Q}}_{{\mathbf{T}}_{[j]}}$ that a component explains is a lower bound for the variance of ${\mathbf{X}}$ that this component can explain, as stated in the following proposition, whose proof is given in the Appendix.
\begin{proposition}\label{prop:vqltev} Given an ordered set of $d$ components, ${\mathbf{t}}_j = {\mathbf{X}}\mathbf{a}_j, j = 1, \ldots, d$, the different types of variance explained as defined in Eqs.~\eqref{eq:vexp1}, \eqref{eq:evexp} and \eqref{eq:vexpq} satisfy \begin{equation}\label{eq:ineqvexp} \mathrm{vexp}_Q({\mathbf{t}_j}) \leq \mathrm{evexp}({\mathbf{t}_j}) \leq \mathrm{vexp}({\mathbf{t}_j}), \end{equation} where $\mathrm{vexp}({\mathbf{t}_j}) = {{\mathbf{t}\trasps{j}}{\mathbf{X}}{\mathbf{X}\trasp}{\mathbf{t}_j}}/{{\mathbf{t}\trasps{j}}{\mathbf{t}_j}}$. Equality is achieved for the first component or if a component is orthogonal to the preceding ones. \end{proposition}
The difference between $\mathrm{evexp}$ and $\mathrm{vexp}_Q$ lies in the different spaces onto which the matrix $\bQTj{T}$ is projected. The extra variance explained measures the norm of the projection of $\bQTj{T}$ onto a component in the span of $$ \mathscr{C}(\dQTj{T})\subseteq \mathscr{C}(\bQTj{T}). $$ Instead, $\mathrm{vexp}_Q$ measures the norm of the projection onto a component in the span of $$ \mathscr{C}(\dXj) = \mathscr{C}(\dQTj{T} + \Pi_{T_{[j-1]}}\dXj)\nsubseteq \mathscr{C}(\dQTj{T}), $$ where $\mathscr{C}({\mathbf{A}})$ denotes the column space of the matrix ${\mathbf{A}}$. This leads to the simple interpretation of the LS SPCA solutions as the first PCs of two different projections of the ${\mathbf{X}}$ matrix, as shown in the next theorem, the proof of which is in the Appendix.
\begin{theorem}\label{th:spcaisproj} Let $\dXj$ be a block of linearly independent variables. Then, \begin{enumerate} \item[(i)] The orthogonal LS SPCA components, ${\mathbf{y}}_j = \dXj\dbj$, are the first PCs of the matrices $ ( \Pi_{\dXj} - \Pi_{\hat{Y}_\dXj} ) {\mathbf{X}}$, where $$ \hat{Y}_\dXj = \Pi_{\dXj}{\mathbf{Y}}_{[j-1]}. $$ \item[(ii)] The nonorthogonal LS SPCA components, ${\mathbf{z}_j} = \dXj\ddj$, are the first PCs of the matrices $$ \hQTj{Z} = \Pi_{\dXj}\bQTj{Z} = \Pi_{\dXj} ({\mathbf{I}} - \Pi_{{\mathbf{Z}}_{[j-1]}} ){\mathbf{X}}. $$ \end{enumerate} \end{theorem}
\subsection{Conventional sparse principal components}
Conventional SPCA methods compute the sparse components as the first PCs of blocks of variables deflated in different ways~\citep{mog}. These solutions are derived with different motivations, either from a constrained LS approach (see, among others, \citep{zou, she}) or by directly adding sparsifying penalties to the Hotelling's formulation of PCA in Eq.~(\ref{eq:pcahot}); see \citep{mer} for a discussion. Hence, the SPCs are the PCs of the (possibly deflated) blocks of variables and the loadings are the solution to \begin{align*} \max\limits_{{\mathbf{a}_j}} {\mathbf{a}\trasps{j}}\mathbf{\cal{Q}}\trasps{j}\mathbf{\cal{Q}}_j{\mathbf{a}_j} \quad \Leftrightarrow \quad \max\limits_{\daj} \dajT\dQcjT\dQcj\daj, \quad \mathrm{card}({\mathbf{a}_j}) = c_j, \quad {\mathbf{a}\trasps{j}}{\mathbf{a}_j}=1 \quad \dajT\daj=1, \end{align*} where $\mathbf{\cal{Q}}_j$ denotes the ${\mathbf{X}}$ matrix deflated using one of the different existing methods; for a review, see \citep{mac}. In conventional SPCA the norm of the components is considered equivalent to the variance of ${\mathbf{X}}$ that they explain. It is easy to show that this latter assumption is not true \citep{mer}, thus these components do not maximize the variance explained. Furthermore, the blocks selected will contain highly correlated variables because the more correlated the variables in the block are, the larger the first eigenvalue of their covariance (or correlation) matrix. Hence, conventional SPCs will have larger cardinality and explain less variance than LS SPCs.
\section{Projection sparse principal components}\label{sec:pspca}
The idea underpinning PSPCA is to iteratively project the (full cardinality) first principal components of the deflated matrices onto blocks of variables $\dXj$. These projections, to which we refer to as projection sparse components, are SPCs in themselves and the variance of the PCs that they explain is a lower bound for the extra variance of ${\mathbf{X}}$ explained by an LS SPCA component, as we prove next.
Let $\bQTj{{\widehat{\mathbf{R}}}}$ denote the ${\mathbf{X}}$ matrix deflated of the first $j-1$ projection SPCs, $\widehat{\mathbf{r}}_{i_1}$ for all $i \in \{ 1,\ldots, j-1\}$, and $$ \bQTjT{{\widehat{\mathbf{R}}}}\bQTj{{\widehat{\mathbf{R}}}} = {\mathbf{W}}_j{\mathbf{M}_{j}}{\mathbf{W}}\trasps{j}, $$ with ${\mathbf{M}_{j}} = \mathrm{diag} (\mu_{j_1} \geq\cdots\geq \mu_{j_p})$ the eigendecomposition of its covariance matrix, the PCs of $\bQTj{{\widehat{\mathbf{R}}}}$ are $$ {\mathbf{R}}_j = \bQTj{{\widehat{\mathbf{R}}}}{\mathbf{W}}_j. $$
Since the PCs, ${\mathbf{r}}_{j_i}$, are orthogonal to the previously computed components, then $\mathrm{vexp}({\mathbf{r}_{j_1}}) = \mathrm{evexp}({\mathbf{r}_{j_1}}) = {\mathbf{r}\trasps{j_1}}{\mathbf{r}_{j_1}}= \mu_{j_1}$. Hence, $\mathrm{vexp}({\mathbf{r}_{j_1}})$ is an upper bound for the extra variance explained by any component, ${\mathbf{t}}_j = {\mathbf{X}}\mathbf{a}_j$, added to the model, i.e., $\mathrm{evexp}({\mathbf{t}}_j)\leq \mu_{j_1}$.
Assume that the variables in a block $\dXj$ are linearly independent and explain a proportion of the variance of ${\mathbf{r}_{j_1}}$ not less than $\alpha \in (0,1)$, i.e., ${\widehat{\mathbf{r}}_{j_1}} = \Pi_{\dXj}{\mathbf{r}_{j_1}}$ is such that \begin{equation}\label{eq:rsqu} {{\widehat{\mathbf{r}}\trasps{j}}{\widehat{\mathbf{r}}_{j_1}}}/{{\mathbf{r}\trasps{j_1}}{\mathbf{r}_{j_1}}} \geq \alpha\; \mathrm{or, equivalently,} \; {\widehat{\mathbf{r}}\trasps{{j_1}}}{\widehat{\mathbf{r}}_{j_1}} \geq \alpha\mu_{j}. \end{equation} Since $\mathscr{C}(\bQTj{{\widehat{\mathbf{R}}}})\subseteq \mathscr{C}({\mathbf{X}})$, a subset of variables $\dXj$ satisfying (\ref{eq:rsqu}) can be found for any $\alpha\in[0,1]$. The projection ${\widehat{\mathbf{r}}_{j_1}}$ is an SPC defined by ${\widehat{\mathbf{r}}_{j_1}} = \dXj{\widehat{\mathbf{w}}_{j_1}}$, with loadings \begin{equation}\label{eq:pspcaloads}
{\widehat{\mathbf{w}}_{j_1}} = (\dXjT\dXj)^{-1}\dXjT{\mathbf{r}_{j_1}}. \end{equation} A lower bound for the extra variance explained by ${\widehat{\mathbf{r}}_{j_1}}$ is given in the following theorem.
\begin{theorem}\label{th:prspca} Let ${\widehat{\mathbf{r}}_{j_1}}$ be the projection of the first PC of $\bQTj{{\widehat{\mathbf{R}}}}$ on a block of variables $\dXj$ such that ${\widehat{\mathbf{r}}\trasps{{j_1}}}{\widehat{\mathbf{r}}_{j_1}}\geq \alpha\mu_{j_1}$. Then, $\mathrm{evexp}({\widehat{\mathbf{r}}_{j_1}}) \geq \alpha\mu_{j_1}$. \end{theorem}
\noindent \textbf{Proof}. By substituting the eigendecomposition $\bQTj{{\widehat{\mathbf{R}}}}\bQTjT{{\widehat{\mathbf{R}}}} = {\mathbf{R}_j}{\mathbf{R}}\trasps{j}$ into Eq.~(\ref{eq:vexpq}), it is easy to verify that \begin{equation*}\label{eq:evexprhat} \mathrm{evexp}({\widehat{\mathbf{r}}_{j_1}}) \geq \mathrm{vexp}_Q({\widehat{\mathbf{r}}_{j_1}}) = {\widehat{\mathbf{r}}\trasps{{j_1}}}{\widehat{\mathbf{r}}_{j_1}} + \sum_{i>1} {({\widehat{\mathbf{r}}\trasps{{j_1}}}{\mathbf{r}_{j_i}})^2}/{{\widehat{\mathbf{r}}\trasps{{j_1}}}{\widehat{\mathbf{r}}_{j_1}}} \geq {\widehat{\mathbf{r}}\trasps{{j_1}}}{\widehat{\mathbf{r}}_{j_1}} \geq \alpha \mu_{j_1}, \end{equation*} because of Eqs.~\eqref{eq:ineqvexp} and \eqref{eq:rsqu}.
$\Box$
The intricacy of the iterated projections renders the comparison between SPCs computed with different methods difficult. In the following theorem we show that ${\widehat{\mathbf{r}}\trasps{{j_1}}}{\widehat{\mathbf{r}}_{j_1}}$ is a lower bound for the extra variances explained by the LS SPCs.
\begin{theorem}\label{th:ineqvexp} Let $\dXj$ be a block of linearly independent variables and ${\widehat{\mathbf{r}}\trasps{{j_1}}}{\widehat{\mathbf{r}}_{j_1}}\geq \alpha\mu_{j_1}$. Also let ${\mathbf{z}}_j$ and ${\mathbf{y}}_j$ be the correlated and uncorrelated SPCA components, respectively, and assume that, when $j>1$, all components have been computed with respect to the same set of previous components ${\mathbf{T}}_{[j-1]}$. Then, the following properties hold:
\begin{subequations}\label{eq:th3}
\begin{equation}\label{th3:i}
\alpha \mu_{j}\leq \mathrm{vexp}_Q({\widehat{\mathbf{r}}_{j_1}}) \leq \mathrm{evexp}({\widehat{\mathbf{r}}_{j_1}})\leq \mathrm{evexp}({\mathbf{y}_j})\leq \mu_j,
\end{equation}
\begin{equation}\label{th3:ii}
\alpha \mu_{j}\leq \mathrm{vexp}_Q({\widehat{\mathbf{r}}_{j_1}}) \leq \mathrm{vexp}_Q({\mathbf{z}_j}) \leq \mathrm{evexp}({\mathbf{z}_j})\leq \mathrm{evexp}({\mathbf{y}_j})\leq \mu_j.
\end{equation}
\end{subequations} \end{theorem}
\noindent \textbf{Proof}. By definition, ${\mathbf{y}_j}$ is the linear combination of the variables in $\dXj$ that explains the most possible extra variance of ${\mathbf{X}}$ and inequality (\ref{th3:i}) follows from Theorem~\ref{th:prspca}. By substituting the PCA decomposition into Eq.~\eqref{eq:vexpq}, it can be verified that \begin{equation}\label{eq:vlspr} \mathrm{vexp}_Q({\mathbf{z}_j}) = \max\limits_{{\mathbf{t}}_j=\dXj\daj}\sum_{i = 1}^p {({\mathbf{t}}\trasps{j}{\mathbf{r}}_{j_i})^2}/{{\mathbf{t}}\trasps{j}{\mathbf{t}}_j} \geq \sum_{i = 1}^p {({\widehat{\mathbf{r}}\trasps{{j_1}}}{\mathbf{r}}_{j_i})^2}/{{\widehat{\mathbf{r}}\trasps{{j_1}}}{\widehat{\mathbf{r}}_{j_1}}} = \mathrm{vexp}_Q({\widehat{\mathbf{r}}_{j_1}}) \geq \alpha\mu_j. \end{equation} This, together with the optimality of $\mathrm{evexp}({\mathbf{y}_j})$, proves inequality (\ref{th3:ii}). When considering the first components, the inequalities reduce to $\alpha \mu_{1}\leq \mathrm{evexp}(\widehat{\mathbf{r}}_{1_1}) \leq \mathrm{evexp}({\mathbf{z}}_1)= \mathrm{evexp}({\mathbf{y}}_1)\leq \mu_1$, because for the first components $\mathrm{vexp}({\mathbf{t}}_1) = \mathrm{vexp}_Q({\mathbf{t}}_1) = \mathrm{evexp}({\mathbf{t}}_1)$.
$\Box$
In principle it cannot be excluded that $\mathrm{evexp}({\widehat{\mathbf{r}}_{j_1}}) > \mathrm{evexp}({\mathbf{z}_j})$. The question of how different are ${\widehat{\mathbf{r}}_{j_1}}$ and ${\mathbf{z}_j}$ when computed with respect to the same deflated matrix ${\mathbf{Q}_j}$ does not have a straightforward answer. We have that ${\mathbf{z}_j}$ is the linear combination of the variables in $\dXj$ which maximizes $\mathrm{vexp}_Q$, while ${\widehat{\mathbf{r}}_{j_1}}$ is the component most correlated with the first PC ${\mathbf{r}_{j_1}}$. Given that ${\mathbf{t}}_j= \dXj\da_j$, $ {({\mathbf{t}}\trasps{j}{\mathbf{r}}_{j_i})^2}/{{\mathbf{t}}\trasps{j}{\mathbf{t}}_j} = \mathrm{corr}^2({\mathbf{t}}_j, {\mathbf{r}}_{j_i})\mu_{j_i}$, from Eq.~(\ref{eq:vlspr}) it follows that \begin{align*}
\mathrm{vexp}_Q({\mathbf{z}_j})-\mathrm{vexp}_Q({\widehat{\mathbf{r}}_{j_1}}) &= \sum_{i=1}^p (\beta_{j_i} - \alpha_{j_i})\mu_{j_i}
= \sum_{i>1} (\beta_{j_i} - \alpha_{j_i})\mu_{j_i} - (\alpha_{j_1} - \beta_{j_1})\mu_{j_1}\geq 0, \end{align*} where $\beta_{j_i} = \mathrm{corr}^2({\mathbf{z}}_j, {\mathbf{r}}_{j_i})$ and $\alpha_{j_i} = \mathrm{corr}^2({\widehat{\mathbf{r}}_{j_1}}, {\mathbf{r}}_{j_i})$, and necessarily $\alpha_{j_1} \geq \beta_{j_1}$.
The following lemma, which is proved in the Appendix, is useful to characterize the difference in variance of ${\mathbf{Q}}_j$ explained by the two components.
\setcounter{lemma}{0} \begin{lemma}\label{le:corr} Let $t$ and $x$ be two random variables and ${\mathbf{y}} = ( y_1, \ldots, y_d)\trasp$ a set of $d$ random variables uncorrelated with $x$. If $\mathrm{corr}^2(t, x) =\alpha$, then for all $i \in \{ 1, \ldots, d\}$, $\mathrm{corr}^2(t, y_i) \leq 1 - \alpha$. If the $y_i$ variables are mutually uncorrelated, it follows that
\begin{equation*}\label{eq:lemma1ii}
\sum_{i=1}^d \mathrm{corr}^2(t, y_i) \leq 1 - \alpha.
\end{equation*} \end{lemma}
Since the PCs ${\mathbf{r}_{j_i}}$ are mutually uncorrelated, assuming that $\beta_{j_1} > \beta_{j_2}$, by the inequalities in Lemma \ref{le:corr}, $$ 0\leq \max \{\mathrm{vexp}_Q({\mathbf{z}_j}) - \mathrm{vexp}_Q({\widehat{\mathbf{r}}_{j_1}})\} \leq \beta_{j_1}\mu_{{j_1}} + (1 - \beta_{{j_1}})\mu_\jind{2} - \alpha_{{j_1}}\mu_{{j_1}} = (1-\beta_{{j_1}})\mu_\jind{2} -(\alpha_{{j_1}} - \beta_{{j_1}})\mu_{{j_1}}. $$ Therefore, the squared correlations $\alpha_{{j_1}}$ and $\beta_{{j_1}}$ are such that $$ \alpha_{{j_1}} - {\mu_\jind{2}(1 - \alpha_{{j_1}})}/{(\mu_{{j_1}} - \mu_\jind{2})} \leq \beta_{{j_1}} \leq \alpha_{{j_1}}. $$ Hence, when $\alpha_{{j_1}}$ is large and the eigenvalues $\mu_{{j_1}}$ and $\mu_\jind{2}$ are well separated, ${\widehat{\mathbf{r}}_{j_1}}$ and ${\mathbf{z}_j}$ will be very close because they have similar correlation with ${\mathbf{r}_{j_1}}$. In our studies we found that the extra variance explained by the PSPCA components and the LS SPCA components is very similar. This is to be expected because the PSPCA components have the largest possible correlation with the first PCs ${\mathbf{r}_{j_1}}$, which are the components that explain the most extra variance.
It should be noted that the inequalities in Theorem~\ref{th:ineqvexp} apply when the different components are computed after the same set of previous components. This is hardly possible in practice because the solutions computed with the various methods are (slightly) different, and different optimality paths are determined by greedy algorithms.
\section{Using projection SPCA to select the variables for the sparse components}\label{sec:selectblocks}
Finding an efficient algorithm for the selection of the variables forming the SPCs is fundamental for the scalability of an SPCA algorithm. Greedy approaches are required because searching the $2^{(p-1)d}$ possible subsets of indices for $d$ SPCs of unknown cardinality is a non-convex NP-hard, hence computationally intractable, problem. A first simplification adopted by most SPCA methods is to select the blocks of variables sequentially for each component. Also in this case, each problem is NP-hard \citep{mog}. The several greedy solutions proposed for conventional SPCA cannot be used for the LS SPCA problem because they seek subsets of variables that are highly correlated. \citet{mer} suggested a branch-and-bound and a backward selection algorithms for LS SPCA. Neither of these is efficient because they are top-down and require the computation of SPCs of large cardinality to evaluate the variance explained.
PSPCA provides a simple yet effective supervisor for the selection of subsets of variables for the computation of LS SPCs. In fact, by Theorem~\ref{th:ineqvexp}, it is enough to select a block of variables that explains a given percentage of the current first PC to be guaranteed that an LS SPCA component computed on that block will explain more than that percentage of the variance of the whole data matrix. Hence, by using PSPCA the LS SPCA variable selection problem is transformed into a more economical regression variable selection problem, thus eliminating the need of computing costly SPCs to evaluate the objective function (the variance explained by the SPCs).
Regression model selection has been extensively researched and several approaches for this task have been proposed. Any of these approaches can be used to select the blocks of variables, including Lasso and regularized lasso, if preferred. The regression approach has also the advantage of being familiar to most data analysts and provides the projection SPCs as a by-product.
\subsection{Comparison with conventional SPCA methods on ``fat'' matrices}\label{sec:convspca}
A particular concern with conventional SPCA methods is that their objective function increases even when perfectly correlated variables are added to the model. Therefore, another important advantage of using a regression model selection method is that the blocks of variables can be chosen to have full column rank and not contain highly correlated variables. This property is important because parsimonious approximations of the PCs should not contain redundant variables. Hence, this property is a further reason for preferring the LS SPCA to conventional SPCA methods which, conversely, generate solution from blocks of highly correlated variables.
Another drawback of conventional SPCA methods, connected with the concern mentioned above, is that they cannot identify the most sparse representation of the PCs when applied to column rank deficient matrices. ``Fat'' matrices, i.e., datasets made up of more features than objects, are very common in the analysis of gene expression microarrays or near-infrared spectroscopy data, for example. In this case, the features are linearly dependent and the PCs can be expressed as linear combinations of as many variables as the rank of the matrix, as stated in the following lemma; see the Appendix for a proof.
\begin{lemma} When $\mathrm{rank}({\mathbf{X}}) = r < p$ the principal components can be expressed as sparse components of cardinality $r$ and loadings that have norm larger than $1$. \end{lemma}
When applied to column rank deficient matrices, conventional SPCA methods compute components of cardinality larger than the rank of the matrix because the only components with unit norm loadings that are equal to the PCs are the full cardinality PCs themselves. This fact is well documented by several examples available in the SPCA literature; see, e.g., \citep {wan,zou}. This means that the model is overfitted by the inclusion of redundant perfectly correlated variables.
When $\mathrm{rank}({\mathbf{X}}) = r$, the LS SPCA components computed on a block of $r$ independent variables will be equal to the full cardinality PC, because of Theorem~\ref{th:spcaisproj}. The same is true for PSPCA, when $r$ variables are enough to explain 100\% of the variance of the PCs. In fact, LS SPCA and PSPCA components of cardinality larger than the rank of the data matrix cannot be computed because in that case a matrix $\dXjT\dXj$ would be singular.
The following example shows the overfitting resulting from applying conventional SPCA to a matrix with perfectly correlated variables. Consider a matrix with 100 observations on five perfectly collinear variables defined, for all $i \in \{ 1, \ldots, 100\}$ and $j \in \{ 1, \ldots, d\}$, by \begin{equation*} x_{ij} = (-1)^{i}\sqrt{j}. \end{equation*} The covariance matrix of these variables, ${\mathbf{S}} = {\mathbf{X}}\trasp{\mathbf{X}}$, has rank $1$, and the only nonzero eigenvalue is equal to 1500. The first PC explains all the variance and can be written in terms of any of the variables as ${\mathbf{x}}_j\sqrt{1500/s_{jj}}$ with $j \in \{ 1, \ldots, 5\}$, i.e., as cardinality $1$ components with loading larger than $1$.
The loadings, norms and relative norms (norm of a component/total variance) of the conventional SPCA optimal solutions for the complete set of conventional SPCs of increasing cardinality are shown in Table~\ref{tab:expl}. We see that $x_5$ ``explains'' just 33\% of the norm of the PC, $x_4$ and $x_5$ together only 60\% (with a net increase equal to 27\%) and so on. These results suggest that the variables with larger variances explain more variance than other collinear variables and that only the full cardinality PC explains the maximum variance. This is true because the norm of a component with unit norm loadings is bounded by $\mathrm{tr}(\dXjT\dXj)$.
\begin{table}[t!]
\centering
\caption{Loadings and norms of the conventional SPCA solutions for the covariance matrix.}
\begin{tabular}{lrrrrr}
& \multicolumn{5}{c}{Cardinality} \\ \cmidrule{2-6} Variable & 1 & 2 & 3 & 4 & 5 \\ \midrule $x_1$ & 0 & 0 & 0 & 0 & 0.58 \\ $x_2$ & 0 & 0 & 0 & 0.60 & 0.52 \\ $x_3$ & 0 & 0 & 0.65 & 0.53 & 0.45 \\ $x_4$ & 0 & 0.75 & 0.58 & 0.46 & 0.37 \\ $x_5$ & 1 & 0.67 & 0.50 & 0.38 & 0.26 \\ \midrule Norm & 500 & 900 & 1200 & 1400 & 1500 \\ Rel norm & 0.33 & 0.60 & 0.80 & 0.93 & 1.0 \\ \bottomrule \end{tabular} \label{tab:expl} \end{table}
The results of applying conventional SPCA to the correlation matrix of the variables in the above example is even more revealing. The correlation matrix is a matrix of $1$s with only one nonzero eigenvalue equal to 5. The conventional SPCA optimal solutions are shown in Table~\ref{tab:expl2}. The results are given irrespectively of which variables are included, because the standardized variables are identical. The cardinality five component is the first PC, which explains all the variability. These results lead to the absurd conclusion that a linear combination of identical variables explains more variance than just one of them.
\begin{table}[t!]
\centering
\caption{Loadings and norms of the conventional SPCA solutions for the correlation matrix.}
\begin{tabular}{lrrrrr}
& \multicolumn{5}{c}{Cardinality} \\ \cmidrule{2-6} & 1 & 2 & 3 & 4 & 5 \\ \midrule Norm & 1 & 2 & 3 & 4 & 5 \\ Rel norm & 0.2 & 0.4 & 0.6 & 0.8 & 1.0 \\ \bottomrule \end{tabular} \label{tab:expl2} \end{table}
Applying LS SPCA to this dataset would yield a single cardinality one component both for the covariance and the correlation matrices, as expected. This can be seen by considering that the variance explained by any variable $x_j$ is equal to $$ \mathrm{vexp}({\mathbf{x}}_j) = {{\mathbf{x}}\trasps{j}{\mathbf{X}}{\mathbf{X}}\trasp{\mathbf{x}}_j}/{{\mathbf{x}}\trasps{j}{\mathbf{x}}_j} =
{\sum_{i=1}^5 s_{ij}^2}/{s_{jj}} = \sum_{i=1}^5 s_{ii} = \mathrm{tr}(S), $$ because $\mathrm{corr}(x_i, x_j)^2 = s_{ij}^2/(s_{ii}s_{jj}) = 1$. Which variable is chosen depends on the algorithm used. The same is true for PSPCA because any variable explains 100\% of the variance of the first PC. An analogous example for blocks of collinear variables can be derived using the artificial data example introduced in \cite{zou}, as illustrated in \cite{mer}, where other unconvincing aspects of conventional SPCA are discussed.
\subsection{Computational considerations}\label{sec:compdet}
The basic algorithm for computing LS SPCA or PSPCA SPCs is outlined in Algorithm~\ref{algo:fspca}. The algorithm is straightforward and simple to implement.
The computation of the PSPCA loadings (line~\ref{algo:projection}) can be simplified as follows. Let us assume, without loss of generality, that the first $c_j$ variables in ${\mathbf{X}}$ form the block $\dXj$ and that the block $\cX_j$ contains the remaining variables, so that ${\mathbf{X}} = (\dXj, \cX_j)$. We write, correspondingly, ${\mathbf{w}}_{j_1} = (\dwjuT, \cwjuT)\trasp$, then we have that $$ {\mathbf{X}\trasp}{\mathbf{r}_{j_1}} = \begin{pmatrix}
\dXjT{\mathbf{r}_{j_1}}\\
\cXjT{\mathbf{r}_{j_1}} \end{pmatrix} = \begin{pmatrix}
\dwju\mu_{j_1}\\
\cwju\mu_{j_1} \end{pmatrix}, $$ because ${\mathbf{X}\trasp}{\mathbf{r}_{j_1}} = \bQTjT{{\widehat{\mathbf{R}}}}\bQTj{{\widehat{\mathbf{R}}}}{\mathbf{w}}_{j_1} = {\mathbf{w}}_{j_1}\mu_{j_1}$.
Substituting this expression into Eq.~(\ref{eq:pspcaloads}), we can write the PSPCA loadings ${\widehat{\mathbf{w}}_j}$ as $$ {\widehat{\mathbf{w}}_{j_1}} = (\dXjT\dXj)^{-1}\dXjT{\mathbf{r}_{j_1}} = (\dXjT\dXj)^{-1}\dwju\mu_{j_1}. $$
The CSPCA loadings in Eq.~(\ref{eq:uspca}) can be computed as the generalized eigenvectors satisfying $$ {\mathbf{C}}_j{\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j {\mathbf{C}}\trasps{j}\dbj = \dSj\dbj\xi_{j}, $$ as shown as Statement 1 in the Appendix. The algorithm requires careful implementation because in its simplest form it is highly computationally demanding. The most demanding operations are the computation of the PCs of the ${\mathbf{Q}}_j$ matrices, the extraction of submatrices (which is a costly operation when the number of variables is large), the deflation of the ${\mathbf{X}}$ matrix, the multiplication $\bQTj{Z}\bQTjT{Z}$ (for CSPCA) and the computation of generalized eigenvalues, when the cardinality is large.
The algorithm can be sped up by computing the first eigenvector of the deflated covariance matrix ${\mathbf{Q}}\trasps{j}{\mathbf{Q}}_j$ and, if necessary the generalized eigenvectors for the sparse loadings, with the iterative power method. The simple power method algorithm is not necessarily very efficient, especially when the first two eigenvalues are not well separated, and other more efficient but complex algorithms could be used, e.g., Lanczos iterations \citep{bjo} or LOBPCG \citep{kny}.
\begin{algorithm}[H]\caption{Projection LS SPCA}
\begin{algorithmic}[1]
\Procedure {plsspca}{${\mathbf{X}},\, \alpha,\, computePSPCA,\, computeCSPCA,
stopRuleCompute$}
\State \textbf{initialize}
\State {\hspace{1em}${\mathbf{Q}}_1 \gets {\mathbf{X}}$}
\State{\hspace{1em}$j \gets 0$}
\State{\hspace{1em}stopCompute $\gets$ FALSE}
\State \textbf{end initialize}
\While{(stopCompute = FALSE)}\Comment{\textbf{start components computation}}
\State{$j \gets j + 1$}
\State{${\mathbf{r}_{j_1}} = {\mathbf{Q}_j}{\mathbf{w}_{j_1}} :\, {\mathbf{Q}}_j{\mathbf{Q}}\trasps{j}{\mathbf{r}_{j_1}} = {\mathbf{r}_{j_1}}\mu_j$} \Comment{compute first
PC of ${\mathbf{Q}_j}$}\label{algo:pwm}
\State{$ind_j \gets \{i_1,\ldots, i_{c_j}\}:\, ||{\widehat{\mathbf{r}}_j}||^2 \geq
\alpha\mu_j$} \Comment{variable selection output}\label{algo:fwdselect}
\State{$\dXj\leftarrow {\mathbf{X}}[,\, ind_j]$}\Comment{$\dX_j$ are columns of ${\mathbf{X}}$ in
$ind_j$}
\If{(computePSPCA)}\Comment{PSPCA}
\State{$\daj \gets (\dX\traspd{j}\dX_j)\matinv\dwju$}\label{algo:projection}
\Comment{$\dwju$ are the elements of ${\mathbf{w}_{j_1}}$ in $ind_j$}
\ElsIf{(computeCSPCA)}\Comment{Correlated LS SPCA}
\State{$\daj: \dXjT{\mathbf{Q}}_j{\mathbf{Q}\trasps{j}}\dXj\da_j = \dXjT\dXj\daj\gamma_j$}
\Else\Comment{Uncorrelated LS SPCA}
\State{$\daj: {\mathbf{C}}_j\dXjT{\mathbf{X}_j}{\mathbf{X}\trasps{j}}\dXj\daj = \dXjT\dXj\daj\xi_j$}
\EndIf
\State{$ {\mathbf{t}_j} \gets \dX_j\daj$} \Comment{$j$-th sparse component}
\If{(stopRuleCompute = FALSE)}\quad\Comment{stop rule on total variance explained or
number of components}
\State{$ {\mathbf{Q}}_{j+1} \gets {\mathbf{Q}_j} - \frac{{\mathbf{t}_j}{\mathbf{t}\trasps{j}}}{{\mathbf{t}\trasps{j}}{\mathbf{t}_j}}{\mathbf{Q}_j}$}
\Comment{deflate ${\mathbf{X}}$ of current component}\label{algo:deflX}
\State{$\mathrm{cvexp}(j) \gets \mathrm{tr}({\mathbf{X}}\trasp{\mathbf{X}}) - \mathrm{tr}({\mathbf{Q}}\trasps{j+1}{\mathbf{Q}}_{j+1})$}
\Comment{cumulative \rm{vexp}}\label{algo:cvexp}
\Else
\State stopCompute $\gets$ TRUE \Comment{\textbf{terminate components
computation}}
\EndIf
\EndWhile
\EndProcedure
\end{algorithmic}\label{algo:fspca} \end{algorithm}
In our implementation we used the simple version of the power method, which has complexity growth rate of about $O(p^2)$, while direct algorithms that compute the whole set of eigenvectors accurately are about $O(p^3)$. The power method is used in extremely high dimensional problems (for example, Google page ranking \citep{bry}) and in various algorithms for conventional SPCA, including \citep{jou, wan}.
The computational complexity of the algorithm depends also on which variable selection algorithm is used (line \ref{algo:fwdselect}). For our implementation we chose the fast greedy forward selection in which the variables that explain the most extra variance conditionally on the variables already in the model are selected until a given percentage of variance is explained. This method can be seen as a QR decomposition with supervised pivoting and can be implemented efficiently using updating formulas; see, e.g., Section~2.4.7 in~\citep {bjo}. Since the QR decomposition can be stopped after $c_j$ iterations, identifying a block will be an operation of order about $O(2c_jnp)$.
When applied to fat matrices, the solutions can be computed more economically by using a ``reverse svd'' approach (see Section~12.1.4 in~\citep{bish}), which means computing the eigenvectors of ${\mathbf{X}}\trasp{\mathbf{X}}$ starting from the ${\mathbf{X}}{\mathbf{X}}\trasp$ matrix.
The theoretical time complexity of the PLSSPCA algorithms cannot be computed exactly because the time taken to compute the eigenvectors with power iterations (if used) and to select the variables and extract submatrices depends on the implementation and the structure of the data. However, we expect the order of growth of the whole algorithm to be not higher than the complexity of computing the first PC of ${\mathbf{Q}_j}$ as it does not contain operations of higher complexity. Therefore, the computation of each vector of loadings should be about $O(p^3)$ when direct eigendecomposition of ${\mathbf{Q}}_j$ is used and roughly $O(p^2)$ when the power method is used. In the next section we will analyze the run time empirically.
\section{Numerical results}\label{sec:examples}
In this section we report the results of running repetitions of USPCA, CSPCA and PSPCA on simulated and real datasets, each with different number of variables, from 100 to over 16,000. To compute the components we applied forward selection requiring that each sparse component explained at least 95\% of the corresponding PC's variance. Since the number of factors considered is large and displaying the results would require very large tables, we present the results mostly graphically, highlighting the main features.
The computational times reported were measured using an implementation of the algorithm in \textsf{C++} embedded in \textsf{R} using the packages \texttt{Rcpp} and \texttt{RcppEigen}. The execution times were measured on an eptacore Intel$^\circledR$\ Core(TM) i7-4770S CPU @ 3.10GHz using Windows 7 operating system. It is well known that \textsf{R} is an inefficient language \citep{mor}, so the run times are slower than what they would be if the programs had been written in a lower level language.
\subsection{Simulations}
\begin{figure}
\caption{Median computational times (in milliseconds) for computing sparse components with underlying latent dimension of 100.}
\label{fig:doetime}
\end{figure}
In order to assess the behavior of the different methods, USPCA, CSPCA and PSPCA, we simulated different datasets according to an experimental design. We considered three different levels of latent dimension (number of latent variables), $d \in \{5, 50, 100\}$; different number of variables, $p \in \{100, 300, 500\}$ and different signal to noise ratio (snr), $s \in \{0.2, 1, 3\}$. The model we considered is ${\mathbf{X}}(d, p, s) = {\mathbf{T}}{\mathbf{P}}\trasp + \sqrt{s}\,{\mathbf{E}}$, where ${\mathbf{T}}$ are $d$ independent $\mathcal{N}(0, 1)$ latent variables, ${\mathbf{P}}$ is a $p \times d$ matrix with unit-norm rows and ${\mathbf{E}}$ is a matrix of $p$ independent $\mathcal{N}(0, 1)$ errors. Therefore, the theoretical correlation matrix of ${\mathbf{X}}$ is equal to $$ {\mathbf{S}} = \mathrm{corr}({\mathbf{X}}) = ({\mathbf{P}}{\mathbf{P}}\trasp + s{\mathbf{I}} )/(1+s). $$
When the snr, $s$, is small this correlation matrix is almost of rank $d$, while as snr increases the correlation matrix becomes closer to the identity matrix. The ${\mathbf{P}}$ matrices were created by generating the entries as independent $\mathcal{U}(-1, 1)$ variables and rescaling the rows. The error correlation matrix was generated as the correlation matrix of a sample of $4p$ pseudo-random realizations of a $\mathcal{N}(0,1)$ variable, without removing possible random correlations. For each combination of levels we ran 1000 repetitions, computing five components when the latent dimension was equal to five and 25 when it was larger, for each method. Variables were selected using a forward selection with stopping criterion $\alpha = 0.95$. The PCs of the ${\mathbf{Q}}_j$ matrices were computed with the power method but USPCA and CSPCA sparse loadings were computed with a direct generalized eigendecomposition algorithm.
In the following we highlight the main findings from these simulations. More details can be found in the Online Supplement to the paper.
The time taken to compute USPCA and CSPCA components are very similar and are indistinguishable on the plot. The higher efficiency of PSPCA shows when the number of variables or the snr grows, as can be appreciated by observing the median computational (CPU) times, shown in Figure~\ref{fig:doetime}.
\begin{table}[b!]
\centering
\caption{Log-log regression of computational times on experimental factors.}
\begin{tabular}{Clllll}
& & & & \multicolumn{2}{c}{95\% Confidence Interval} \\
\cmidrule{5-6} Term & \multicolumn{1}{l}{Estimate} & \multicolumn{1}{l}{Standard Error} & \multicolumn{1}{l}{$p$-value} & \multicolumn{1}{l}{Low} & \multicolumn{1}{l}{High} \\ \midrule
Intercept & \multicolumn{1}{r}{1.58} & \multicolumn{1}{r}{0.0050} & \multicolumn{1}{r}{0.0000} & \multicolumn{1}{r}{1.57} & \multicolumn{1}{r}{1.59} \\
nvar (p) & \multicolumn{1}{r}{2.21} & \multicolumn{1}{r}{0.0008} & \multicolumn{1}{r}{0.0000} & \multicolumn{1}{r}{2.21} & \multicolumn{1}{r}{2.21} \\
latDim (d) & \multicolumn{1}{r}{0.28} & \multicolumn{1}{r}{0.0018} & \multicolumn{1}{r}{0.0000} & \multicolumn{1}{r}{0.27} & \multicolumn{1}{r}{0.28} \\
snr (s) & \multicolumn{1}{r}{0.28} & \multicolumn{1}{r}{0.0005} & \multicolumn{1}{r}{0.0000} & \multicolumn{1}{r}{0.28} & \multicolumn{1}{r}{0.28} \\
Comp. No. (j) & \multicolumn{1}{r}{1.17} & \multicolumn{1}{r}{0.0030} & \multicolumn{1}{r}{0.0000} & \multicolumn{1}{r}{1.16} & \multicolumn{1}{r}{1.17} \\
CSPCA & \multicolumn{1}{r}{$-0.04$} & \multicolumn{1}{r}{0.0012} & \multicolumn{1}{r}{0.0000} & \multicolumn{1}{r}{$-0.05$} & \multicolumn{1}{r}{-0.04} \\
PSPCA & \multicolumn{1}{r}{$-0.17$} & \multicolumn{1}{r}{0.0012} & \multicolumn{1}{r}{0.0000} & \multicolumn{1}{r}{$-0.17$} & \multicolumn{1}{r}{-0.17} \\ \bottomrule
& & & & & \\ \multicolumn{6}{l}{Residual standard error: 0.1446 on 80,993 degrees of freedom} \\
\multicolumn{6}{l}{Multiple R$^2$: 0.9946, Adjusted R$^2$: 0.9946 } \\
\multicolumn{6}{l}{F-statistic: 2471982 on 6 and 80993 DF, $p$-value: 0} \\ \end{tabular} \label{tab:doeregtime} \end{table}
We assumed a polynomial dependence of time on the parameters $d, p, s$ and the order of the component computed, $j$. Hence, we estimated the polynomial terms by regressing the logarithm of time on the logarithm of these parameters and adding indicator variables for the method used. The results of the regression, shown in Table~\ref{tab:doeregtime}, confirm the conclusions given above. The fit is excellent, as indicated by the coefficient of determination $R^2 >0.99$, and the final time equation is $$ t(d, p, s, j, M) = e^{1.58} p^{2.21} d^{0.28} s^{0.28} j^{1.17} (0.96)^{I_{CSPCA}} ({0.84})^{I_{PSPCA}} \epsilon, $$ where $I_M$ denotes the indicator variable equal to $1$ when method $M$ is used and $0$ otherwise. The coefficients of these indicator variables measure the ratio of time with the corresponding times taken by USPCA. This result confirms that using the power method to compute the PCs the complexity growth rate is about $O(p^{2.2})$. The time increases almost linearly with the number of components computed. PSPCA is slightly faster than the other methods.
The components computed to explain 95\% of the variance explained by the PCs have relatively low cardinality. As expected, the cardinality of the components increases with the number of variables in the set, the latent dimension and the snr. The variability is low and it increases when the number of variables and the snr increase.
The results of the log-log regression of cardinality on the experimental factors gave an excellent fit, with coefficient of determination $R^2 \approx 0.97$. The final cardinality equation is $$ c(d, p, s, j, M) = e^{-0.73} p^{0.49} d^{0.48} s^{0.39} j^{1.01} \epsilon. $$ The cardinality increases less than linearly with the number of variables, latent dimension and snr, while it grows almost linearly with the components' order. The method used was not found to significantly change the cardinality of the solutions.
The proportions of variance explained by the three methods is very similar in value and in ratio, and differences are only observable at the third or fourth decimal figure. The variance explained by the CSPCA components is always very close to that explained by the USPCA components or is slightly higher. In most cases, the PSPCA components explain the least proportion of variance. The USPCA components of higher order tend to explain less variance than the CSPCA components. This phenomenon has already been observed and it is due to the greediness of the approach, when the local optimality of the USPCA components leads to globally inferior paths.
To compare the variance explained by different methods we use the cumulative variances explained by the sparse components relative to the variances explained by the same number of PCs, $$ \mathrm{rCvexp} = {\sum_{i = 1}^j \mathrm{evexp}({\mathbf{t}}_i)}{\Big /} {\sum_{i = 1}^j\mathrm{vexp}({\mathbf{p}}_i)}. $$ Figure~\ref{fig:doecvexp} shows the median $\mathrm{rCvexp}$ for various number of variables and latent dimensions at a constant snr, $s = 0.2$. This shows how USPCA performs noticeably worse than the other methods when the latent dimension is small ($d = 5$) and the snr ratio is low. This is because, under this setting, the rank of the $\bQTj{Y}$ matrices is almost equal to $5 - j + 1$ and orthogonality determines a more severe departure from the optimal path. However, also in this case, the differences are very small.
Another aspect that we investigated is the correlation among the components computed with CSPCA and PSPCA. Since each component is highly correlated with the corresponding first PC of ${\mathbf{Q}}_j$, which are mutually orthogonal, by Lemma \ref{le:corr}, we expect their correlation to be small. This is confirmed by the distribution of the $nc (nc-1)/2$ correlations between each pair of components computed for each experimental set up, of which the summary statistics are shown in Table~\ref{tab:corcomp}. The correlations are extremely small and do not show a particular pattern with respect to any of the experimental factors, except that, in most cases, the variability is slightly larger for the PSPCA components.
\begin{table}[b!]
\centering
\caption{Summary statistics of the correlations among sparse components computed with the same method on each simulated data set.}
\begin{tabular}{cccccc} \toprule
\multicolumn{1}{l}{Minimum} & \multicolumn{1}{l}{1st Quartile} & \multicolumn{1}{l}{Median} & \multicolumn{1}{l}{Mean} & \multicolumn{1}{l}{3rd Quartile} & \multicolumn{1}{l}{Maximum} \\
0.004 & 0.005 & 0.006 & 0.008 & 0.009 & 0.016 \\ \bottomrule \end{tabular}
\label{tab:corcomp} \end{table}
\begin{figure}
\caption{Median $\mathrm{rCvexp}$ for the sparse components computed with different methods for different simulated datasets, with constant snr = 0.2. The median values are computed over 1000 runs.}
\label{fig:doecvexp}
\end{figure}
\subsection{Real datasets}
The datasets that we consider in this section, listed in Table~\ref{tab:descdata}, have been taken from various sources, mostly from the data distributed with the book ``Elements of Statistical Learning'' (ESL) \citep{has}. Other sets were taken from the UCI Machine Learning Repository \citep{uci}. The remaining sets were taken from different sources; see Table~\ref{tab:descdata} for details. Most of these are fat datasets as they have a large number of features and fewer objects. The largest dataset, \cite{rama}, has been used to test other SPCA methods, including \citep{zou, sri, wan}.
\begin{table}[b!]
\caption{Description of the datasets used for numerical comparison.}
\centering
\begin{threeparttable}
\begin{tabular}{lrrlll} Name & Samples &Features & Type &Description & Source \\ \midrule Crime& 1994& 99 ®ular& social data &UCI Repository\tnote{a} \\ Isolet & 6238& 716 ®ular & character recognition & UCI Repository\tnote{b} \\ Ross (NCI60) &60 & 1375 &fat & gene expression & \textsf{R} package \texttt{made4}\tnote{c} \\ Khanh &88& 2308 &fat & gene expression & ESL\tnote{d} \\ Phoneme &257 & 4509 &fat & speech recognition & ESL \\ NCI60 &60 & 6830 &fat & gene expression & ESL \\ Protein &11 & 7466 &fat & protein cytometry& ESL \\ Radiation &58 & 12625 &fat & gene expression& ESL \\ Ramaswamy &144 & 16063 &fat & gene expression & Broadinstitute repository\tnote{e} \\ \bottomrule \end{tabular} \begin{tablenotes} {\footnotesize \item[a] \url{https://archive.ics.uci.edu/ml/datasets/Communities+and+Crime} \item[b] \url{https://archive.ics.uci.edu/ml/datasets/ISOLET} \item[c] \url{http://bioconductor.org/packages/release/bioc/html/made4.html} \item[d] \url{https://statweb.stanford.edu/~tibs/ElemStatLearn/} \item[e] \url{http://software.broadinstitute.org/cancer/software/genepattern/datasets}} \end{tablenotes} \end{threeparttable} \label{tab:descdata} \end{table}
First we compared the performance of USPCA, CSPCA and PSPCA on the fat dataset described in Table~\ref{tab:descdata}. We computed 10 components for each dataset using the reverse svd approach and requiring that each of them explained at least a proportion $\alpha = 0.95$ of the variance explained by the corresponding PC. Both the PCs and the sparse loadings were computed using direct eigendecomposition. Figure~\ref{fig:fatty_times} shows the median computational times over 25 repetitions. The plots are shown in increasing order of number of observations in the datasets.
The computational times of the three methods are very close for all datasets with the exception of Protein (Prot.). For this dataset the computation of the USPCA components takes longer because the orthogonality constraints require the cardinality to be larger than that of the other methods, as shown in Figure~\ref{fig:fatty_cards}.
Even though the computation of the eigenvectors of the ${\mathbf{Q}_j}{\mathbf{Q}}\trasps{j}$ matrices is $O(n^3)$, in some cases the computational time on some sets (for example Radiation) is greater than that on dataset with fewer variables and more observations (note the different scales on the vertical axes). This is because the computation of the PC loadings is $O(n^3 + n^2p)$, hence there is a cross-over effect, due to the number of variables.
\begin{figure}
\caption{Median computational times taken to compute the first 10 sparse components with $\alpha = 0.95$ on
seven fat datasets. Time is expressed in milliseconds.}
\label{fig:fatty_times}
\end{figure}
\begin{figure}
\caption{Comparison of the performance of LS SPCA methods on two fat datasets, Protein (top) and Phonema (bottom).}
\label{fig:fatty_cards}
\end{figure}
\subsection{Comparison with conventional SPCA}\label{sec:compspca}
In this section we compare the performances of the first sparse components computed with a conventional SPCA method and with PSPCA. As our conventional SPCA method we used SPCA-IE \citep{wan} with the amvl criterion. This method was shown to perform similarly to other SPCA methods. It does not require to choose arbitrary sparsity parameters and is simple to implement.
Since the results for fat matrices are quite similar, we present the results only for four datasets: Crime, Isolets, Khanh and Ramaswamy. For the last dataset, we computed the conventional sparse components using simple thresholding, stopping the computation at cardinality 200; details of the performance of components with larger cardinality computed with different conventional SPCA methods for this dataset can be found in the papers cited above.
Figure~\ref{fig:compspca} compares the relative norm ($||{\mathbf{t}}||^2/||{\mathbf{X}}||^2$), $\mathrm{rCvexp}$ and their correlation with the full PCs for increasing cardinality of the first components computed with PSPCA and SPCA-IE on different datasets. The PSPCA values for the rank deficient Khanh and Ramaswamy datasets are available only until the solutions reach full rank cardinality (87 and 143, respectively) at which the components explain the maximum possible variance. Clearly SPCA-IE outperforms PSPCA in the norm of the components. However, the latter method guarantees higher variance explained and closer convergence to the PC with much lower cardinality.
\begin{figure}
\caption{Norm, $\mathrm{rCvexp}$ and correlation with the PCs versus cardinality of the first sparse components computed with SPCA-IE and PSPCA.}
\label{fig:compspca}
\end{figure} The differences in performance of the two approaches are more evident for large rank deficient datasets, when conventional sparse components with cardinality in the hundreds explain less variance than PSPCA components of much lower cardinality. The plots also show clearly that the components' norms are not related to the variance that they explain or to their correlation with the PC. This confirms the theoretical conclusions given in Section \ref{sec:convspca}.
\begin{table}[b!]
\centering
\caption{Cardinality needed to reach 99.9\% $\mathrm{rCvexp}$ by the components computed with PSPCA.} \begin{threeparttable}
\begin{tabular}{lrrr}
Dataset& Rank & \multicolumn{2}{c}{Cardinality Needed for $99.9\%\, \mathrm{rCvexp}$}\\ \cmidrule{3-4} && PSPCA & SPCA\tnote{a}\\ \midrule Crime & 99& 38 & 74 \\ Isolet & 716& 123 & 439 \\ Khanh & 87& 28 & 1338 \\ Ramaswamy &143& 23 & $>$ 200\\ \bottomrule \end{tabular} \begin{tablenotes} \item[a] The values for the first three datasets were obtained using SPCA-IE and the values for the Ramaswamy dataset using simple thresholding \end{tablenotes} \end{threeparttable} \label{tab:compspca} \end{table}
Table~\ref{tab:compspca} shows the cardinality with which the components computed with the two methods reached 99.9\% $\mathrm{rCvexp}$. In all cases the cardinality of the PSPCA components is much lower than that of the SPCA-IE components.
\section{Discussion}\label{sec:conclusions}
Projection SPCA is a very efficient method for selecting variables for computing a sparse approximation to the PCs. The methodology is intuitive and can be understood by users who do not have a deep knowledge of numerical optimization. The only parameter to be set for computing the solutions is the proportion of variance explained, the meaning of which is also easily understandable. The algorithm is simple to implement and scalable to large datasets. Users can choose their preferred regression variable selection algorithm to select the variables. Most conventional SPCA methods, instead, are based on special numerical optimizations methods and require setting values for parameters with a difficult to understand effect. Future research could explore the results of using $\ell_1$ norm selection methods, such as least angle or Lasso regression, for example, on the computation of LS SPCA components.
In this work we have developed a framework for computing LS SPCA components that closely approximate the full PCs with low cardinality. We showed that sparse USPCA, CSPCA and PSPCA components can be efficiently computed for very large datasets. We also show that conventional SPCA methods suffer from a number of drawbacks which yield less attractive solutions than the corresponding LS SPCA solutions.
Conventional SPCA methods have been shown to give results similar to simple thresholding if not worse; see, e.g., \citep{zou, wan}. Thresholding has been proven to give misleading results; see, e.g., \citep{cad}. Since the loadings are proportional to the covariances of the variables with the PCs, the largest loadings correspond to variables that are highly correlated with the current PC and among themselves. These sets of variables are not very informative because they contain different measures of the same features.
\citet{zou} proposed three properties of a good SPCA method: \begin{itemize} \item [(i)] Without any sparsity constraint, the method should reduce to PCA. \item [(ii)] It should be computationally efficient for both small \emph{p} and big \emph{p} data. \item [(iii)] It should avoid misidentifying the important variables. \end{itemize} The first property is not enough for a good method. The second is not necessary for the most commonly analyzed datasets and the third is vague because importance is not defined and variables known to be unimportant could be directly eliminated from the analysis.
We suggest the following properties for a good SPCA method: \begin{itemize} \item [(i)] Without any sparsity constraint, the method should reduce to PCA. \item [(ii)] It should identify the sparsest expression of the principal components. \item [(iii)] The addition of a variable perfectly correlated with one or more variables already in the solution should not improve the objective function. \end{itemize} The last property eliminates redundant variables from the solution and should deter the inclusion of highly correlated ones. Conventional SPCA methods do not have the last two properties while methods based on LS SPCA do. It is possible that other methods could have these properties.
LS SPCA is implemented in the \textsf{R} package \texttt{spca} available on GitHub. PSPCA will be added to this package in the future.
\section*{Acknowledgments}
We thank the Editor-in-Chief, Christian Genest, an Associate Editor and the anonymous referees for their useful comments and suggestions that improved the paper. Gemai Chen's research is partially supported by a grant from the Natural Sciences and Engineering Research Council of Canada.
\section*{Appendix}
\setcounter{equation}{0} \setcounter{proposition}{0} \renewcommand{A.\arabic{equation}}{A.\arabic{equation}} \renewcommand{\Alph{proposition}}{\Alph{proposition}} \renewcommand{\Alph{theorem}}{\Alph{theorem}} \renewcommand{\Alph{lemma}}{\Alph{lemma}}
\begin{proposition} Given an ordered set of $d$ components, ${\mathbf{t}}_j = {\mathbf{X}}\mathbf{a}_j$ with $ j \in \{ 1, \ldots, d\}$, the different types of variance explained as defined in Eqs.~\eqref{eq:vexp1}, \eqref{eq:evexp} and \eqref{eq:vexpq} satisfy \begin{equation*} \mathrm{vexp}_Q({\mathbf{t}_j}) \leq \mathrm{evexp}({\mathbf{t}_j}) \leq \mathrm{vexp}({\mathbf{t}_j}), \end{equation*} where $\mathrm{vexp}({\mathbf{t}_j}) = {{\mathbf{t}\trasps{j}}{\mathbf{X}}{\mathbf{X}\trasp}{\mathbf{t}_j}}/{{\mathbf{t}\trasps{j}}{\mathbf{t}_j}}$. Equality is achieved for the first component or if a component is orthogonal to the preceding ones. \end{proposition}
\noindent \textbf{Proof}. The extra variance explained cannot be smaller than the variance of ${\mathbf{Q}}$ explained by the same components because \begin{equation*} \mathrm{vexp}_Q({\mathbf{t}_j}) = \mathrm{evexp}({\mathbf{t}}_j) \, \frac{{\mathbf{a}\trasps{j}}\bQTjT{T}\bQTj{T}{\mathbf{a}_j}}{{\mathbf{a}\trasps{j}}{\mathbf{X}\trasp}{\mathbf{X}}{\mathbf{a}_j}}
= \mathrm{evexp}({\mathbf{t}}_j) \left(\frac{{\mathbf{a}\trasps{j}}{\mathbf{X}\trasp}{\mathbf{X}}{\mathbf{a}_j} - {\mathbf{a}\trasps{j}}{\mathbf{X}\trasp}\Pi_{{\mathbf{T}_{[j-1]}}}{\mathbf{X}}{\mathbf{a}_j}}{{\mathbf{a}\trasps{j}}{\mathbf{X}\trasp}{\mathbf{X}}{\mathbf{a}_j}} \right) \leq \mathrm{evexp}({\mathbf{t}}_j). \end{equation*} It is well known that extra variance explained is not larger than the variance explained by a regressor. In fact, \begin{equation*}
\mathrm{vexp}({\mathbf{t}_j}) = ||\Pi_{{\mathbf{t}_j}}{\mathbf{X}}||^2 = ||\Pi_{[\Pi_{{\mathbf{T}}_{[j-1]}}{\mathbf{t}_j}]}{\mathbf{X}} + \Pi_{[\bQTj{T}{\mathbf{a}_j}]}{\mathbf{X}}||^2 \geq
||\Pi_{[\bQTj{T}{\mathbf{a}_j}]}{\mathbf{X}}||^2 = \mathrm{evexp}({\mathbf{t}_j}) \end{equation*} because ${\mathbf{t}}_j = \Pi_{{\mathbf{T}}_{[j-1]}}{\mathbf{t}_j} + \Pi_{[({\mathbf{I}} - {\mathbf{T}}_{[j-1]})]}{\mathbf{t}_j}$ and $\Pi_{[({\mathbf{I}} - {\mathbf{T}}_{[j-1]}){\mathbf{t}_j}]} = \Pi_{[\bQTj{T}{\mathbf{a}_j}]}$. Therefore, $\Pi_{{\mathbf{t}_j}} = \Pi_{\Pi_{{\mathbf{T}}_{[j-1]}}{\mathbf{t}_j}} + \Pi_{\Pi_{({\mathbf{I}} - {\mathbf{T}}_{[j-1]}){\mathbf{t}_j}}}$; see, e.g., Theorem~8.8 in~\citep{pun}.
The statement about the equality is true because if a component ${\mathbf{t}_j}$ is orthogonal to all preceding variables, then ${\mathbf{t}_j} = {\mathbf{X}}\mathbf{a}_j = \bQTj{T}{\mathbf{a}_j}$, and ${\mathbf{Q}}_{{\mathbf{T}}_{[0]}}= {\mathbf{X}}$.
$\Box$
\setcounter{theorem}{0} \begin{theorem} Let $\dXj$ be a block of linearly independent variables. Then, \begin{enumerate} \item[(i)] The orthogonal LS SPCA components, ${\mathbf{y}}_j = \dXj\dbj$, are the first PCs of the matrices $(\Pi_{\dXj} - \Pi_{\hat{Y}_\dXj} ){\mathbf{X}}$, where $\hat{Y}_\dXj = \Pi_{\dXj}{\mathbf{Y}}_{[j-1]}$. \item[(ii)] The nonorthogonal LS SPCA components, ${\mathbf{z}_j} = \dXj\ddj$, are the first PCs of the matrices $\hQTj{Z} = \Pi_{\dXj}\bQTj{Z} = \Pi_{\dXj} ({\mathbf{I}} - \Pi_{{\mathbf{Z}}_{[j-1]}} ){\mathbf{X}}$. \end{enumerate} \end{theorem}
\noindent \textbf{Proof}. Premultiplying Eq.~(\ref{eq:uspca}) by $\dXj\dSjinv$ gives \begin{equation*}
\Big(\Pi_{\dXj} - \Pi_{\hat{Y}_\dXj}\Big){\mathbf{X}}{\mathbf{X}}\trasp {\mathbf{y}}_j = {\mathbf{y}}_j\xi_{j_\text{max}}, \end{equation*} where $\Pi_{\hat{Y}_\dXj} = \Pi_{\dXj} {\mathbf{Y}_{[j-1]}} ({\mathbf{Y}\trasps{[j-1]}}\Pi_{\dXj}{\mathbf{Y}_{[j-1]}} )\matinv{\mathbf{Y}\trasps{[j-1]}}\Pi_{\dXj}$. Since, $ (\Pi_{\dXj} - \Pi_{\hat{Y}_\dXj} ){\mathbf{y}}_j = {\mathbf{y}}_j$, we can write $$ \Big(\Pi_{\dXj} - \Pi_{\hat{Y}_\dXj}\Big) {\mathbf{X}}{\mathbf{X}}\trasp \Big(\Pi_{\dXj} - \Pi_{\hat{Y}_\dXj}\Big){\mathbf{y}}_j = {\mathbf{y}}_j\xi_{j_\text{max}}, $$ which proves part (i). Given that $ \mathscr{C}(\hat{Y}_\dXj)\subset \mathscr{C}(\dXj)$, $\Pi_{\dXj} - \Pi_{\hat{Y}_\dXj} $ is a projector onto $\mathscr{C}(\dXj)\cap \mathscr{C}(\hat{Y}_\dXj)^\bot$, where $\mathscr{C}({\mathbf{A}})^\bot$ denotes the orthocomplement of $\mathscr{C}({\mathbf{A}})$ with respect to ${\mathbf{I}}$; see Chapter~7 in~\citep{pun}.
In a similar fashion, premultiplying Eq.~\eqref{eq:lsspcasol} by $\dXj\bdS\matinvj{j}$ we obtain $$ \Pi_{\dXj}\bQTj{Z}\bQTjT{Z}{\mathbf{z}_j} = \hQTj{Z}\hQTjT{Z}{\mathbf{z}_j} = {\mathbf{z}_j}\gamma_j, $$ which proves part~(ii).
$\Box$
\setcounter{lemma}{0} \begin{lemma} Let $t$ and $x$ be two random variables and ${\mathbf{y}} = (y_1, \ldots, y_d) \trasp$ a set of $d$ random variables uncorrelated with $x$. If $\mathrm{corr}^2(t, x) =\alpha$, then for all $i \in \{ 1, \ldots, d\}$, $ \mathrm{corr}^2(t, y_i) \leq 1 - \alpha$.
If the $y_i$ variables are mutually uncorrelated, it follows that
\begin{equation*}
\sum_{i=1}^d \mathrm{corr}^2(t, y_i) \leq 1 - \alpha.
\end{equation*} \end{lemma}
\noindent \textbf{Proof.} Let $\boldsymbol{\rho}\trasps{t{\mathbf{y}}} = ( \rho_{ty_1}, \ldots,\rho_{ty_d})$, where $\rho_{ty_i}= \mathrm{corr}(t,y_i)$. The squared multiple correlation coefficient of the regression of $t$ on $[x,\mathbf{y}\trasp]\trasp$ is such that \[ \rho_{t.x{\mathbf{y}}}^2 = [\sqrt{\alpha}, \boldsymbol{\rho}\trasps{t{\mathbf{y}}}]
\begin{bmatrix}
1& \mathbf{0}\trasp\\
\mathbf{0} & {\mathbf{R}}\matinv
\end{bmatrix}
\begin{bmatrix}
\sqrt{\alpha}\\ \boldsymbol{\rho}_{t{\mathbf{y}}}
\end{bmatrix}
= \alpha + \rho_{t.{\mathbf{y}}}^2\leq 1, \] where ${\mathbf{R}}$ is the correlation matrix of ${\mathbf{y}}$ and $\rho_{t.{\mathbf{y}}}^2 = \boldsymbol{\rho}\trasps{t{\mathbf{y}}} {\mathbf{R}}\matinv\boldsymbol{\rho}_{t{\mathbf{y}}}$ is the squared coefficient of multiple correlation between $t$ and ${\mathbf{y}}$.
Since $\rho_{t.{\mathbf{y}}}^2$ can be written as the sum of the squared correlation of the response variable with one of the regressors, $y_i$, say, and the multiple correlation of the response variable with the orthogonal complement of the remaining variables, $\{y_j,\ j\neq i\}$, namely, $\rho_{t.{\mathbf{y}}}^2 = \mathrm{corr}^2(t,y_i) + \rho_{t.{\mathbf{y}}_{/i}.y_i}^2\geq \mathrm{corr}^2(t,y_i)$, it follows that \[
0\leq \mathrm{corr}^2(t,y_i) \leq \boldsymbol{\rho}\trasps{t{\mathbf{y}}} {\mathbf{R}}\matinv\mathbf{\rho}_{t{\mathbf{y}}} \leq 1 - \alpha. \] When $\mathrm{corr}(y_i,y_j)=0, i\neq j$, $$ \rho_{t.{\mathbf{y}}}^2 = \boldsymbol{\rho}\trasps{t{\mathbf{y}}} {\mathbf{R}}\matinv\boldsymbol{\rho}_{t{\mathbf{y}}} = \sum_{i=1}^p \mathrm{corr}^2({\mathbf{t}}, {\mathbf{y}}_i), $$ from which follows the second statement to be proved. For a more general proof, see \cite{pun05}.
$\Box$
\begin{lemma} When $\mathrm{rank}({\mathbf{X}}) = r < p$ the principal components can be expressed as sparse components of cardinality $r$ and loadings that have norm larger than $1$. \end{lemma}
\noindent \textbf{Proof}. Given that $\mathrm{rank}({\mathbf{X}}) = r$, there are $p - r$ columns of ${\mathbf{X}}$ which are linearly dependent. Assume without loss of generality that the first $r$ columns of ${\mathbf{X}}$ are linearly independent and denote them as $\dX$.
Also let $\cX$ be the remaining columns. Then, we can write \begin{equation*} {\mathbf{X}} = [\dX, \cX] = \dX \Big\{ {\mathbf{I}}_r, \dX\trasp(\dX\dX\trasp)\matginv\cX \Big\} = \dX \Big \{{\mathbf{I}}_r, (\dX\trasp\dX)\matinv\dX\trasp\cX\Big\} \nonumber =\Pi_{\dX}{\mathbf{X}} = \dX{\mathbf{G}}, \end{equation*} where the superscript `$^+$' denotes the Moore--Penrose generalized inverse and ${\mathbf{G}} = {\mathbf{I}}_r, (\dX\trasp\dX)\matinv \dX\trasp\cX$.
Hence, the PCs can be written as ${\mathbf{u}}_j = {\mathbf{X}}{\mathbf{v}}_j = \dX({\mathbf{G}}{\mathbf{v}}_j) = \dX\dot{\mathbf{v}}_j$, where $\dot{\mathbf{v}}_j = {\mathbf{G}}{\mathbf{v}}_j$ has length $r$. Since the largest singular value of ${\mathbf{X}}_j$ ($\bQTj{U}$) is not smaller than the largest singular value of $\dXj$ ($\dQTj{U}$), it must follow that $\dot{\mathbf{v}}\trasps{j}\dot{\mathbf{v}}_j \geq 1$. Therefore, the PCs can be defined as sparse components of cardinality $r$ with loadings of norm larger than $1$.
$\Box$
\begin{state} The CSPCA loadings can be computed as the generalized eigenvectors satisfying $${\mathbf{C}}_j {\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j {\mathbf{C}}\trasps{j}\dbj = \dSj\dbj\xi_{j}, $$ where ${\mathbf{C}}_j = {\mathbf{I}}_{c_j} - {\mathbf{H}\trasps{j}}({\mathbf{H}_j}\bdS\matinvj{j}{\mathbf{H}\trasps{j}})\matinv{\mathbf{H}_j}\bdS\matinvj{j}$. \end{state}
\noindent \textbf{Proof}. From Proposition~\ref{prop:uspca}, we have that the USPCA loadings satisfy \begin{equation*} {\mathbf{C}}_j\dXjT{\mathbf{X}}{\mathbf{X}\trasp}\dXj\dbj = {\mathbf{C}}_j{\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j\dbj = \dSj\dbj\xi_{j}. \end{equation*} Then $$
{\mathbf{C}}_j{\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j\dbj = \dSj\dbj\xi_{j} \quad \Leftrightarrow \quad
{\mathbf{C}}_j{\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j \dSjinv (\dSj\dbj ) = \dSj\dbj\xi_{j}. $$ Since ${\mathbf{C}}_j$ is idempotent and $\mathscr{C}(\dSj\dbj)\subseteq\mathscr{C}({\mathbf{C}}_j)$, because ${\mathbf{C}}_j{\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j
\dSjinv (\dSj\dbj ) \propto \dSj\dbj$, $\dbj$ must satisfy $$ {\mathbf{C}}_j{\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j
\dSjinv{\mathbf{C}}_j (\dSj\dbj ) = {\mathbf{C}}_j{\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j (\dSjinv{\mathbf{C}}_j\dSj )\dbj =
\dSj\dbj\xi_{j}. $$ Now, \begin{equation*} \dSjinv{\mathbf{C}}_j\dSj = \dSjinv \Big \{ {\mathbf{I}} - {\mathbf{H}\trasps{j}} ({\mathbf{H}_j}\dSjinv{\mathbf{H}\trasps{j}} )\matinv{\mathbf{H}_j}\dSjinv \Big \} \dSj = {\mathbf{I}} - \dSjinv{\mathbf{H}\trasps{j}} ({\mathbf{H}_j}\dSjinv{\mathbf{H}\trasps{j}} )\matinv{\mathbf{H}_j} = {\mathbf{C}}\trasps{j}. \end{equation*} Therefore, $\dbj$ is the generalized eigenvector satisfying $ {\mathbf{C}}_j{\mathbf{J}}\trasps{j}{\mathbf{S}}{\mathbf{S}}{\mathbf{J}}_j
{\mathbf{C}}\trasps{j}\dbj = \dSj\dbj\xi_{j}$. This completes the argument.
$\Box$
\section*{References}
\end{document}
|
arXiv
|
{
"id": "1612.00939.tex",
"language_detection_score": 0.7490407824516296,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Cauchy-Davenport Theorem for linear maps: Simplification and Extension} \author{ John Kim\thanks{Department of Mathematics, Rutgers University. Research supported in part by NSF Grant Number DGE-1433187. {\tt [email protected]}.} \and Aditya Potukuchi\thanks{Department of Computer Science, Rutgers University. {\tt [email protected]}.}} \maketitle
\begin{abstract} We give a new proof of the Cauchy-Davenport Theorem for linear maps given by Herdade et al., (2015) in~\cite{HKK}. This theorem gives a lower bound on the size of the image of a linear map on a grid. Our proof is purely combinatorial and offers a partial insight into the range of parameters not handled in~\cite{HKK}. \end{abstract}
\section{Introduction}
Let $\mathbb{F}_p$ be the field containing $p$ elements, where $p$ is a prime, and let $A,B \subseteq \mathbb{F}_p$. The Cauchy-Davenport Theorem gives a lower bound on the size of the sumset $A + B \defeq \{a + b \;\ifnum\currentgrouptype=16 \middle\fi|\; a \in A,b \in B\}$ (for more on sumsets, see, for example,~\cite{TV}). The size of the sumset can be thought of as the size of the image of the linear map $(x,y) \rightarrow x + y$, where $x \in A$, and $y \in B$. Thus the theorem can be restated as follows:
\begin{theorem}[Cauchy-Davenport Theorem] Let $p$ be a prime, and $L:\mathbb{F}_p \times \mathbb{F}_p \rightarrow \mathbb{F}_p$ be a linear map that takes $(a,b)$ to $a+b$. For $A,B \subseteq \mathbb{F}_p$, Let $L(A,B)$ be the image of $L$ on $A \times B$. Then,
$$
|L(A,B)| \geq \min(|A|+|B|-1,p) $$ \end{theorem}
In~\cite{HKK}, this notion was extended to study the sizes of images of general linear maps on product sets. A lower bound was proved using the polynomial method (via a nonstandard application of the Combinatorial Nullstellensatz~\cite{Alon}). In this paper, we give a simpler, and combinatorial proof of the same using just the Cauchy-Davenport Theorem. \\
Notation: For a linear map $L : \mathbb{F}_p^{n} \rightarrow \mathbb{F}_p^m$, and for $S_1, S_2, \ldots S_n \subseteq \mathbb{F}_p$, we use $L(S_1, S_2 \ldots S_n)$ to denote the image of $L$ on $S_1 \times S_2 \times \cdots S_n$. The \emph{support} of a vector is the set of nonzero entries in the vector. A \emph{min-support vector} in a set $V$ of vectors is a nonzero vector of minimum support size in $V$.
\begin{theorem}[Main Theorem] \label{the:main}
Let $p$ be a prime, and $L:\mathbb{F}_p^{m+1} \rightarrow \mathbb{F}_p^m$ be a linear map of rank $m$. Let $A_1, A_2, \ldots A_{m+1} \subseteq \mathbb{F}_p$ with $|A_i| = k_i$. Further, suppose that $\min_i(k_i) + \max_i(k_i) < p$. Let $S$ be the support of $\ker(L)$, and $S' = [n] \setminus S$. Then
$$
|L(A_1,A_2, \ldots, A_n)| \geq \left( \prod_{j \in S'}k_j \right) \cdot \left( \prod_{i\in S}k_i - \prod_{i \in S}(k_i - 1) \right) $$ \end{theorem}
As noted in~\cite{HKK}, this bound is tight for every $m$ and $p$. We restrict our theorem to study only maps from $\mathbb{F}_p^{m+1}$ to $\mathbb{F}_p^m$ of rank $m$ for two reasons mainly:(1) It is simpler to state, and contains the tight case and (2) We are unable prove any better bounds if the rank is not $m$. It is not clear to us what the correct bound for the general case is. \\
We also show the following result for the size of the image for certain full rank linear maps from $\mathbb{F}_p^n \rightarrow \mathbb{F}_p^{n-1}$ when the size of the sets it is evaluated on are all large enough.
\begin{theorem} \label{cover}
Let $L:\mathbb{F}_p^n \rightarrow \mathbb{F}_p^{n-1}$ be a linear map given by $L(x_1,\ldots x_n) = (x_1 + x_n, x_2 + x_n \ldots x_{n-1} + x_n)$. Let $S_1, \ldots S_n \subseteq \mathbb{F}_p$ with $|S_i| = k$ for $i \in [n]$ such that $k > \frac{(n-1)p}{n}$, then $|L(S_1, \ldots S_n)| = p^{n-1}$ (i.e., $L(S_1, \ldots S_n) = \mathbb{F}_p^{n-1}$). \end{theorem}
The theorems do not, however, give tight bounds for all set sizes, for example if $\min_i |A_i| > p/2$. It would be interesting to obtain a tight bound even for the simple linear map $(x,y,z) \rightarrow (x+z, y+z)$ on the product set $A_1 \times A_2 \times A_3 \subseteq \mathbb{F}_p^3$ which holds for all sizes of the $A_i$'s.
\section{The Theorem}
\subsection{The Main Lemma}
The idea is that since the size of the image is invariant under row operations of $L$, we perform row operations to isolate a `hard' part, which gives the main part of the required lower bound \\
Our proof proceeds by induction on the dimension of the linear map. The base case is given by the Cauchy Davenport Theorem.
\begin{lemma} \label{lem:main}
Let $L:\mathbb{F}_p^n \rightarrow \mathbb{F}_p^{n-1}$ be a linear map such that $L(x_1,\ldots, x_n) = (x_1 + x_n, x_2 + x_n, \ldots, x_{n-1} + x_n)$. Let $S_1, \ldots S_n \subseteq \mathbb{F}_p$ with $|S_i| = s_i$ such that $\min_i(s_i) + \max_i(s_i) \leq p+1$. Then $|L(S_1, \ldots S_n)| \geq \prod_{i=1}^n s_i-\prod_{i=1}^n (s_i-1)$ \end{lemma}
\begin{proof}
We use the shorthand notation $|L| \defeq |L(S_1,S_2 \ldots S_n)|$. W.L.O.G, let $S_1$ be such that $|S_1| = \min_{i \in [n-1]}(|S_i|)$. \\
A preliminary observation is that $|S_1| + |S_n| \leq p+1$, and therefore, by the Cauchy-Davenport Theorem, \begin{equation} \label{CD}
|S_1 + S_n| \geq s_1 + s_n - 1 \end{equation}
The proof proceeds by induction on $n$. If $n = 2$, the result $|L| \geq s_1 \cdot s_2 - (s_1-1) \cdot (s_2-1) = s_1 + s_2 - 1$ is given by the Cauchy-Davenport Theorem. \\
For every $a \in \mathbb{F}_p$, we have $T_a \defeq \{x_n \in S_n \;\ifnum\currentgrouptype=16 \middle\fi|\; \exists x_1 \in S_1, x_1 + x_n = a\}$, and $t_a \defeq |T_a|$. We now look at the restricted linear map $L|_{x_1 + x_n = a}$. In this case, the induction is on sets $S_2, \ldots S_{n-1} \times T_a$. This is equivalent to restricting $S_n$ to the set $T_a$, and dropping $S_1$, since for every $x_n \in S_n$, there is a unique $x_1 \in S_1$ such that $x_1 + x_n = a$. \\
We first observe that the conditions are satisfied, i.e., $\min_i(|S_i|) + \max_i(|S_i|) \leq p + 1$, since $t_a \leq \min(|S_1|,|S_n|)$. Also the resulting linear map is of the same form, i.e., $L|_{x_1 + x_n = a}(x_2,\ldots x_n) = (x_2 + x_n \ldots x_{n-1} + x_n)$. (In reality, $L|_{x_1 + x_n = a}$ is a map from $\mathbb{F}_p^n$ to $\mathbb{F}_p^{n-1}$, given by $L|_{x_1 + x_n = a}(x_1,x_2,\ldots x_n) = (a, x_2 + x_n \ldots x_{n-1} + x_n)$ but we drop the first coordinate because it is fixed, i.e., $a$)\\
By induction hypothesis, the number of points in the image of $L_{x_1 + x_n = a}$ is at least:
$$ \left( \prod_{i=2}^{n-1}s_i \right) t_a - \left( \prod_{i=2}^{n-1}(s_i - 1) \right)(t_a - 1) $$
Summing over all $a \in \mathbb{F}_p$, we get a bound on the number of points in the image:
\begin{eqnarray*}
|L| &\geq& \sum_{a \in \mathbb{F}_p, t_a \neq 0} \left( \left( \prod_{i=2}^{n-1}s_i \right) t_a - \left( \prod_{i=2}^{n-1}(s_i - 1) \right)(t_a - 1) \right) \\ &=& \left( \prod_{i=2}^{n-1}s_i \right)\sum_{a \in \mathbb{F}_p}t_a - \left( \prod_{i=2}^{n-1}(s_i - 1) \right)\sum_{a\in \mathbb{F}_p, t_a \neq 0}(t_a - 1) \\ &\geq& \prod_{i=1}^{n}s_i - \prod_{i=1}^n(s_i - 1) \end{eqnarray*}
The last inequality comes from observing that $\sum_{a \in \mathbb{F}_p}t_a = s_1s_n$, and an upper bound on $\sum_{a \in \mathbb{F}_p, t_a \neq 0}(t_a - 1)$, by using~\ref{CD}. We have $\sum_{a \in \mathbb{F}_p, t_a \neq 0}(t_a - 1) = \sum_{a \in \mathbb{F}}t_a - \sum_{a \in \mathbb{F}_p}\mathbbm{1}_{t_a \neq 0} = \sum_{a \in \mathbb{F}}t_a + |S_1 + S_n| \leq s_1s_n - (s_1 + s_n - 1)$.
\end{proof}
\subsection{Arriving at the Main Theorem}
The first step in arriving at the main theorem is exactly as in~\cite{HKK}. For completeness, we describe it here. The idea is to transform a general linear map into a specific form, without reducing the size of the image (in fact, here it remains the same). This step is very intuitive, but describing it requires some setup. \\
Let $L:\mathbb{F}_p^{m+1} \rightarrow \mathbb{F}_p^m$ be an $\mathbb{F}_p$-linear map of rank $m$. Let $v$ be a non-zero min-support vector of $\ker(L)$. So, we have $Lv = 0$. The main observation is that under row operations, two quantities remain unchanged: the size of the image of $L$, and the size of the support of the min-support vector in the kernel. \\
Let $r_1, \ldots r_m$ be the rows, and $c_1, c_2, \ldots c_{m+1}$ be the columns of associated to $L$ with respect to the standard basis. We show that one can perform elementary row operations, and some column operations on $L$ while preserving the size of the image.
\begin{lemma} \label{lem:linop} The size of the image of $L$ does not change under
\begin{enumerate} \item Elementary row operations. \item Scaling any column $c_i$ by some $d \in \mathbb{F}_p \setminus \{0\}$ and scaling every element of $A_i$ by $d$. \item Swapping any two columns $c_i$ and $c_j$, and swapping sets $A_i$ and $A_j$. \end{enumerate} \end{lemma}
\begin{proof}
We prove this by considering each given operation separately.
\begin{enumerate} \item Suppose $L'$ was obtained from $L$ by elementary row operations. There is an invertible linear map $M$ such that $M \cdot L = L'$. This gives the bijection from every vector $v$ in the image of $L$, to the vector $M \cdot v$ in the image of $L'$.
\item Suppose $L'$ was obtained from $L$ by scaling column $c_i$ by $d \in \mathbb{F}_p \setminus \{0\}$, and scaling the set $A_i$ by $d^{-1}$. We map every vector $(u_1, \ldots, u_m) \in$ $L(A_1,\ldots ,A_i, \ldots, A_{m+1})$, to the vector $(u_1, \ldots, u_m) \in L'(A_1,\ldots , d^{-1} \cdot A_i, \ldots, A_{m+1})$. Here $d^{-1} \cdot A_i \defeq \{d^{-1}a_i \;\ifnum\currentgrouptype=16 \middle\fi|\; a_i \in A_i\}$ This map is invertible.
\item Suppose $L'$ was obtained from $L$ by switching columns $c_i$ and $c_j$, and swapping the sets $A_i$ and $A_j$. We map every vector $(u_1,\ldots, u_m) \in L(A_1,\ldots ,A_i, \ldots, A_j, \ldots, A_{m+1})$ to the identical vector $(u_1, \ldots, u_m) \in$
$ L'(A_1,\ldots ,A_j, \ldots, A_i, \ldots, A_{m+1})$. This map is invertible. \end{enumerate}
For every given operation, we have a bijection between the images of $L$ before and after the operation.
\end{proof}
\begin{observation} After the operations stated in Lemma~\ref{lem:linop}, the size of the support of the min-support vector in $\ker(L)$ does not change. \end{observation} To see this, we first observe that the kernel has rank $1$, and is orthogonal to the row span of $L$. Therefore, all nonzero vectors in $\ker(L)$ have the same support. Since, row operations do not change the row span of $L$, the resulting kernel spans the same subspace of $\mathbb{F}^{m+1}$, and therefore, the size of the support of the vectors in $\ker(L)$ does not change. \\
Next, we do the following operations, each of which preserves the size of the image.
\begin{enumerate} \item Perform row operations so that the last $m$ columns form an identity matrix.
\item Scale the rows so that the first column of every row is $1$.
\item Scale the last $m$ columns so that every nonzero entry in $L$ is $1$. \end{enumerate}
After we perform these operations, we have a linear map where the first column consists of $1$'s and $0$'s and the remaining $m$ columns form an identity matrix. Let the $S'$ be the set of indices of rows containing $1$'s in the first column. Consider the vector $v = -e_1 + \sum_{i \in S'}e_{i+1}$. This vector has support $|S'| + 1$, and lies in the kernel of $L$. Therefore, $|S| = |S'| + 1$.
\begin{proof}[Proof of Theorem~\ref{the:main}] Apply the transformation from Lemma~\ref{lem:linop} to $L$ to reduce it to the simple form. Let $S'$ be the set of rows where the first column is nonzero. Consider the restriction of $L$ on the the coordinates given by $S$. By Lemma~\ref{lem:main}, the size of this image is at least $\left( \prod_{i\in S}k_i - \prod_{i \in S}(k_i - 1) \right)$.
The linear map restricted to the coordinates $[m] \setminus S$ is nothing but the identity map, so the size of the image is $\prod_{i \not \in S} |A_i|$, and is independent of the linear map restricted to $S$. Putting them together, we have the desired result. \end{proof}
\section{The case when $2k > p+1$}
The proof of Lemma~\ref{lem:main} breaks down when $s_1+s_n > p + 1$ and, unfortunately, we do not know how to fix this issue. Consider, for example, the simplest nontrivial case where $m = 2$, i.e., $L(x,y,z) = (x+z, y+z)$, and we are interested in the size of the image of $L$ on $X \times Y \times Z$, further suppose, for simplicity, that $|X|=|Y|=|Z| = k$. If $k < \frac{p + 1}{2}$, then the above bound holds, and is tight. If $k>\frac{2p}{3}$, then $L$ covers $\mathbb{F}_p^2$, i.e., $|L(A,B,C)| = p^2$. This makes the case in between the interesting one. We conjecture that the correct lower bound is the size of the image of $L$ when $X = Y = Z = \{1,2,\ldots k\}$. Towards this, we are able to prove a partial result (Lemma~\ref{2d}) using the above method.
We will need the following Lemma:
\begin{lemma} \label{boundtx}
Let $X,Y\subseteq \mathbb{F}_p$ and $t_a = |\{(x,y)\in X\times Y: x+y = a\}|$. Then for every $a \in \mathbb{F}_p$:
$$|X|+|Y|-p\leq t_a \leq \min(|X|,|Y|).$$ \end{lemma}
\begin{proof} The bounds follow from the fact that $t_a$ can be written as the size of the
intersection of two sets of sizes $|X|$ and $|Y|$:
$$t_a = |X \cap (a-Y)|.$$ \end{proof}
Now we state the partial result:
\begin{theorem} \label{2d} Let $L:\mathbb{F}_p^3\rightarrow \mathbb{F}_p^2$ be the linear map defined by $L(x,y,z) = (x+z,y+z)$. Let $X,Y,Z\subset F_p$ be sets of size $k$, where $k \geq \frac{p+1}{2}$. Then we have the following lower bound:
$$|L(X,Y,Z)| \geq \min(p^2 + 3k^2 - (2p+1)k,p^2).$$ \end{theorem}
\begin{proof}
Let $T_a \defeq \{z \in Z \;\ifnum\currentgrouptype=16 \middle\fi|\; \exists x \in X, x+z = a\}$, with $t_a\defeq |T_a|$. Looking at this restriction, $L|_{x+z = a}$, by Cauchy-Davenport Theorem, there are at least $\min(t_a + k-1,p)$ points of $L(X,Y,Z)$ on $L_{x+z=a}(Y,T_a)$. By summing over all $a\in \mathbb{F}_p$, we get a lower bound on the size of $L(X,Y,Z)$:
\begin{eqnarray*}
|L(X,Y,Z)| &\geq& \displaystyle\sum_{a\in\mathbb{F}_p}{\min(t_a + k-1,p)} \\ &=& \displaystyle\sum_{a\in\mathbb{F}_p}{\min(t_a,p-k+1)} + p(k-1) \\ &=& \displaystyle\sum_{a:t_a \leq p-k+1}{t_a} + \displaystyle\sum_{a:t_a > p-k+1}{(p-k+1)} + p(k-1). \end{eqnarray*}
We now want to remove the dependence of the lower bound on the $t_a$ by considering the worst case scenario, where the $t_a$ take values that minimize the lower bound.
First, we observe $\sum_{a\in\mathbb{F}_p}{t_a} = k^2$, a fixed quantity. So to minimize the above lower bound for $|L(X,Y,Z)|$, we need $t_a$ to be maximal for as many $a\in\mathbb{F}_p$ as possible.
By Lemma~\ref{boundtx}, we know that $2k-p\leq t_a \leq k$. We set $t_a = k$ for as many $a\in\mathbb{F}_p$ as possible, and the remainder of the $t_a = 2k-p$. This gives:
\begin{eqnarray*}
|L(X,Y,Z)| &\geq& \displaystyle\sum_{a:t_a \leq p-k+1}{t_a} + \displaystyle\sum_{a:t_a > p-k+1}{(p-k+1)} + p(k-1) \\ &\geq& k(2k-p) + (p-k)(p-k+1) + p(k-1) \\ & = & 3k^2 + p^2 - (2p-1)k. \end{eqnarray*}
\end{proof}
As a corollary, we get, independent of theorem~\ref{cover}, the following corrolary:
\begin{corollary}
If the linear map $L$, and the sets $A$, $B$, $C$ were as above, with $|A|=|B|=|C|=k$, and $k > \frac{2p}{3}$, then $L(A,B,C) = p^2$. \end{corollary}
We would like to point out that at the two extremes, i.e., when $k = \frac{p+1}{2}$, and when $k =\lceil\frac{2p}{3} \rceil$, the above bound matches the `correct' lower bound.
\subsection{Proof of Theorem~\ref{cover}}
We prove theorem~\ref{cover} via a slightly stronger claim \begin{claim} \label{lem:cover}
Let $L:\mathbb{F}_p^n \rightarrow \mathbb{F}_p^{n-1}$ be a linear map given by $L(x_1,\ldots x_n) = (x_1 + x_n, x_2 + x_n \ldots x_{n-1} + x_n)$. Let $S_1, \ldots S_n \subseteq \mathbb{F}_p$ with $|S_i| = k$ for $i \in [n-1]$, and $|S_n| = k'$. Further, suppose that $(n-1)k + k' \geq (n-1)p + 1$, then $|L(S_1, \ldots S_n)| = p^{n-1}$. \end{claim}
\begin{proof} We prove this by induction on $n$, analogous to Lemma~\ref{lem:main}. The case where $n = 2$ is, again, given by the Cauchy-Davenport Theorem. \\
For $a \in \mathbb{F}_p$, $T_a \defeq \{x_n \in S_n \;\ifnum\currentgrouptype=16 \middle\fi|\; \exists x_1 \in S_1, x_1 + x_n = a\}$ with $t_a \defeq |T_a|$. Looking at this restriction of $L$ (i.e., $x_1 + x_n = a$), we have a linear map, $L_{x_1 + x_n = a}$ on the sets $S_2 \times S_3 \times \cdots T_a$, given by the $L_{x_1 + x_n = a}(x_2 ,\ldots x_n) = (x_2 + x_n, \ldots x_{n-1} + x_n)$. (similar to Lemma~\ref{lem:main}, we drop the first coordinate). \\
Here, $|S_i| = k$ for $i = 2,\ldots n-1$, and $|T_a| \geq k+k'-p$, by Lemma~\ref{boundtx}. Further, the required condition holds, i.e.,:
$$(n-2)k + t_a \geq (n-2)k + k + k' - p = (n-1)k + k' - p \geq (n-2)p+1.$$
Therefore, by induction hypothesis $|L|_{x_1 + x_n = a}(S_2, \ldots S_{n-1}, T_a)| = p^{n-2}$. Since this holds for every $a \in \mathbb{F}_p$, we have $|L(S_1, \ldots S_n)| = p^{n-1}$. \end{proof}
In particular, Lemma~\ref{lem:cover} tells that for the linear map $L$ given by $L(x_1,\ldots x_n) = (x_1 + x_n, x_2 + x_n, \ldots, x_{n-1} + x_n)$ on $S_1 \times S_2 \times \cdots S_n$, if $|S_i| \geq \frac{(n-1)p}{n}$, then $L(S_1, \ldots, S_n) = \mathbb{F}_p^{n-1}$.
\end{document}
|
arXiv
|
{
"id": "1612.09217.tex",
"language_detection_score": 0.7149213552474976,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{A $(k + 3)/2$-approximation algorithm for monotone submodular $k$-set packing and general $k$-exchange systems}
\begin{abstract} We consider the monotone submodular $k$-set packing problem in the context of the more general problem of maximizing a monotone submodular function in a $k$-exchange system. These systems, introduced by Feldman et al.\ \cite{Feldman-2011}, generalize the matroid k-parity problem in a wide class of matroids and capture many other combinatorial optimization problems. We give a deterministic, non-oblivious local search algorithm that attains an approximation ratio of $(k + 3)/2 + \epsilon$ for the problem of maximizing a monotone submodular function in a $k$-exchange system, improving on the best known result of $k + \epsilon$, and answering an open question posed in Feldman et al. \end{abstract}
\section{Introduction}
In the general $k$-set packing problem, we are given a collection $\mathcal{G}$ of sets, each with at most $k$ elements, and an objective function $f : 2^\mathcal{G} \to \ensuremath{\mathbb{R}}_+$ assigning each subset of $\mathcal{G}$ a value, and seek a collection of pairwise-disjoint sets $S \subseteq \mathcal{G}$ that maximizes $f$. In the special case that $f(A) = |A|$, we obtain the \emph{unweighted $k$-set packing problem}. Similarly, if $f$ is linear function, so that $f(A) = \sum_{e \in A}w(e)$ for some weight function $w : \mathcal{G} \to \ensuremath{\mathbb{R}}_+$ we obtain the \emph{weighted $k$-set packing problem}. In this paper we consider the case in which $f$ may be any monotone submodular function.
For unweighted $k$-set packing, Hurkens and Schrijver \cite{Hurkens-1989} and Halld\'orsson \cite{Halldorsson-1995} independently obtained a $k/2 + \epsilon$ approximation via a simple local search algorithm. Using similar techniques, Arkin and Hassin \cite{Arkin-1997} obtained a $k - 1 + \epsilon$ approximation for weighted $k$-set packing, and showed that this result is tight for their simple local search algorithm. Chandra and Halld\'orsson \cite{Chandra-1999} showed that a more sophisticated local search algorithm, which starts with a greedy solution and always chooses the best possible local improvement at each stage, attains an approximation ratio of $2(k + 1)/3 + \epsilon$. This was improved further by Berman \cite{Berman-2000}, who gave a \emph{non-oblivious} local search algorithm yielding a $(k + 1)/2 + \epsilon$ approximation for weighted $k$-set packing. Non-oblivious local search \cite{Khanna-1994} is a variant of local search in which an auxiliary objective function to evaluate solutions, rather than the problem's given objective. In the case of Berman, the local search procedure repeatedly seeks to improve the sum of \emph{square} of the weights in the current solution, rather than the sum of the weights.
Many of the above local search algorithms for $k$-set packing yield the same approximations for the more general problem of finding maximum independent sets in $(k + 1)$-claw free graphs. Additionally, local search techniques have proved promising for other generalizations of $k$-set packing, including variants of the matroid $k$-parity problem \cite{Lee-2010,Soto-2011a}. Motivated by the similarities between these problems, Feldman et al.\ \cite{Feldman-2011} introduced the class of \emph{$k$-exchange} systems, which captures problems amenable to approximation by local search algorithms. These systems are formulated in the general language of independence systems, which we now briefly review.
An independence system is specified by a ground set $\mathcal{G}$, and a hereditary (i.e.\ non-empty and downward-closed) family $\mathcal{I}$ of subsets of $\mathcal{G}$. These subsets of $\mathcal{G}$ contained in $\mathcal{I}$ are called \emph{independent sets}, and the inclusion-wise maximal sets of $\mathcal{I}$ are called \emph{bases} of the independence system $(\mathcal{I}, \mathcal{G})$. Given an independence system $(\mathcal{G}, \mathcal{I})$ and a function $f : 2^\mathcal{G} \to \ensuremath{\mathbb{R}}_+$, we consider the problem of finding an independent set $S \in \mathcal{I}$ that maximizes $f$.
The class of $k$-exchange systems satisfy the following additional property: \begin{definition}[$k$-exchange system \cite{Feldman-2011}] \label{def:k-exchange}
A hereditary system $\mathcal{I}$ is a \emph{$k$-exchange system} if, for all $A$ and $B$ in $\mathcal{I}$, there exists a multiset $Y = \{ Y_e \subseteq B \setminus A\ |\ e \in A \setminus B \}$, containing a subset $Y_e$ of $B \setminus A$ for each element $e \in A \setminus B$, that satisfies: \begin{enumerate}[itemsep=0em,topsep=0.5em,leftmargin=0.65in,label={\rm (K\arabic{*})}]
\item $|Y_e| \le k$ for each $x \in A$. \label{N1} \item Every $x \in B \setminus A$ appears in at most $k$ sets of $Y$. \label{N2} \item For all $C \subseteq A \setminus B$, $(B \setminus \left(\bigcup_{e \in C}Y_e\right)) \cup C
\in \mathcal{I}$. \label{N3} \end{enumerate} \end{definition} We call the set $Y_e$ in Definition \ref{def:k-exchange} the \emph{neighborhood} of $e$ in $B$. For convenience, we extend the collection $Y$ in Definition \ref{def:k-exchange} by including the set $Y_x = \{x\}$ for each element $x \in A \cap B$. It is easy to verify that the resulting collection still satisfies conditions \ref{N1}--\ref{N3}.
The 1-exchange systems are precisely the class of strongly base orderable matroids described by Brualdi \cite{Brualdi-1971}. This class is quite large and includes all gammoids, and hence all transversal and partition matroids. For $k > 1$, the class of $k$-exchange systems may be viewed as a common generalization of the matroid $k$-parity problem in strongly base orderable matroids and the independent set problem in $(k + 1)$-claw free graphs. Feldman et al. showed that $k$-exchange systems encompass a wide variety of combinatorial optimization problems, including as $k$-set packing, intersection of $k$ strongly base orderable matroids, hypergraph $b$-matching (here $k = 2$), as well as problems such as asymmetric traveling salesperson (here $k = 3$).
Our results hold for any $k$-exchange system, and so we present them in the general language of Definition \ref{def:k-exchange}. However, the reader may find it helpful to think in terms of a concrete problem, such as the $k$-set packing problem. In that case, the ground set $\mathcal{G}$ is the given collection of sets, and a sub-collection of sets $S \subseteq \mathcal{G}$ is independent if and only if all of the sets in $S$ are disjoint. Given $A$ and $B$ as in Definition \ref{def:k-exchange}, $Y_e$ is the set of all sets in $B$ that contain any element contained by the set $e \in A$ (i.e. the set of all sets in $B$ that are not disjoint from $e$). Then, property \ref{N3} is immediate, and \ref{N1} and \ref{N2} follow directly from the fact that each set in $\mathcal{G}$ contains at most $k$ elements.
\subsection{Related Work}
Recently, the problem of maximizing submodular functions subject to various constraints has attracted much attention. We focus here primarily on results pertaining to matroid constraints and related independence systems.
In the case of an arbitrary single matroid constraint, Calinescu et al.\ have attained a $e/(e - 1)$ approximation for monotone submodular maximization, via the \emph{continuous greedy algorithm}. This result is tight, provided that $P \neq NP$ \cite{Feige-1998}. In the case of $k \ge 2$ simultaneous matroid constraints, an early result of Fisher, Nemhauser, and Wolsey \cite{Fisher-1978} shows that the standard greedy algorithm attains a $k + 1$ approximation for monotone submodular maximization. Fischer et al.\ state further that the result can be generalized to $k$-systems (a full proof appears in Calinescu et al.\ \cite{Calinescu-2007}). More recently, Lee, Sviridenko, and Vondr\`ak\ \cite{Lee-2010a} have improved this result to give a $k + \epsilon$ approximation for monotone submodular maximization over $k \ge 2$ arbitrary matroid constraints, via a simple, oblivious local search algorithm. A similar analysis was used by Feldman et al.\ \cite{Feldman-2011} to show that oblivious local search attains a $k + \epsilon$ approximation for the class of $k$-exchange systems (here, again, $k \ge 2$). For the more general class of $k$-systems, Gupta et al.\ \cite{Gupta-2010} give a $(1 + \beta)(k + 2 + 1/k)$ approximation, where $\beta$ is the best known approximation ratio for unconstrained non-monotone submodular maximization.
In the case of unconstrained non-monotone submodular maximization, Feige, Mirrokni, and Vondr\'ak \cite{Feige-2007} gave a randomized $2.5$ approximation, which was iteratively improved by Gharan and Vondr\'{a}k \cite{Gharan-2011} and then Feldman, Naor, and Shwartz \cite{Feldman-2011a} to $\approx 2.38$. For non-monotone maximization subject to $k$ matroid constraints, Lee, Sviridenko, and Vondr\'ak \cite{Lee-2009} gave a $k + 2 + 1/k + \epsilon$ approximation, and later improved \cite{Lee-2010a} this to a $k + 1 + 1/(k - 1) + \epsilon$ approximation. Again, the latter result is obtained by a standard local search algorithm. Feldman et al.\ \cite{Feldman-2011} apply similar techniques to yield a $k + 1 + 1/(k - 1) + \epsilon$ approximation for non-monotone submodular maximization the general class of $k$-exchange systems.
\subsection{Our Contribution}
In the restricted case of a linear objective function, Feldman et al.\ \cite{Feldman-2011} gave a non-oblivious local search algorithm inspired by Berman's algorithm \cite{Berman-2000} for $(k + 1)$-claw free graphs. They showed that the resulting algorithm is a $(k + 1)/2 + \epsilon$ approximation for linear maximization in any $k$-exchange system. Here we consider a question posed in \cite{Feldman-2011}: namely, whether a similar technique can be applied to the case of monotone submodular maximization in $k$-exchange systems. In this paper, we give a successful application of the non-oblivious local search techniques to the case of monotone submodular maximization in a $k$-exchange system. As in \cite{Feldman-2011}, the $k$-exchange property is used only in the analysis of our algorithm. The resulting non-oblivious local search algorithm attains an approximation factor of $\frac{k + 3}{2} + \epsilon$. For $k > 3$, this improves upon the $k + \epsilon$ approximation obtained by the oblivious local search algorithm presented in \cite{Feldman-2011}. Additionally, we note that our algorithm runs in time polynomial in $\epsilon^{-1}$, while the $k + \epsilon$ approximation algorithm of \cite{Feldman-2011} requires time exponential in $\epsilon^{-1}$.
As a consequence of our general result, we obtain an improved approximation guarantee of $\frac{k + 3}{2}$ for a variety of monotone submodular maximization problems (some of which are generalizations of one another) including: $k$-set packing, independent sets in $(k + 1)$-claw free graphs, $k$-dimensional matching, intersection of $k$ strongly base orderable matroids, and matroid $k$-parity in a strongly base orderable matroid. In all cases, the previously best known result was $k + \epsilon$.
\section{A First Attempt at the Submodular Case}
Before presenting our algorithm, we describe some of the difficulties that arise when attempting to adapt the non-oblivious local search algorithm of \cite{Feldman-2011} and \cite{Berman-2000} to the submodular case. Our hope is that this will provide some intuition for our algorithm, which we present in the next section.
We recall that a function $f : 2^\mathcal{G} \to \ensuremath{\mathbb{R}}_+$ is submodular if $f(A) + f(B) \ge f(A \cup B) + f(A \cap B)$ for all $A, B \subseteq \mathcal{G}$. Equivalently, $f$ is submodular if for all $S \subseteq T$ and all $x \not\in T$, $f(S + x) - f(S) \ge f(T + x) - f(T)$. In other words, submodular functions are characterized by decreasing marginal gains. We say that a submodular function $f$ is monotone if it additionally satisfies $f(S) \le f(T)$ for all $S \subseteq T$.
The non-oblivious algorithm of \cite{Feldman-2011} for the linear case is shown in Algorithm \ref{alg:nols}. It proceeds from one solution to another by applying a \emph{$k$-replacement} $(A,B)$. Formally, we call the pair of sets $(A,B)$, where $A \subseteq \mathcal{G} \setminus (S \setminus B)$ and $B \subseteq S$ a $k$-replacement if $|A| \le k$, $|B| \le k^2 - k + 1$ and $(S \setminus B) \cup A \in \mathcal{I}$. Algorithm \ref{alg:nols} repeatedly searches for a $k$-replacement that improves an auxiliary potential function $w^2$. Because $f$ is linear, it can be represented as a sum of weights, one for each element in $S$. If $w(e) = f(\{e\})$ is the weight assigned to an element $e$ in this representation, then the non-oblivious potential function is given by $w^2(S) = \sum_{e \in S}w(e)^2$. That is, our non oblivious potential function $w^2(S)$ is simply the sum of the \emph{squared} weights of the elements of $S$.\footnote{To ensure polynomial-time convergence, Algorithm \ref{alg:nols} first round the weights down to integer multiples of a suitable small value $\alpha$, related to the approximation parameter $\epsilon$. The algorithm then converges in time polynomial in $\epsilon^{-1}$ and $n$, at a loss of only $(1 - \epsilon)^{-1}$ in the approximation factor.}
\begin{algorithm} \KwIn{\parbox[t]{7in}{\begin{itemize}[parsep=0em,itemsep=0em,topsep=0em,leftmargin=1em]
\item Ground set $\mathcal{G}$
\item Membership oracle for $\mathcal{I} \subseteq 2^\mathcal{G}$
\item Value oracle for monotone submodular function $f : 2^\mathcal{G} \to \ensuremath{\mathbb{R}}_+$
\item Approximation parameter $\epsilon \in (0,1)$
\end{itemize}}}
Let $S_\mathit{init} = \{ \arg\max_{e \in \mathcal{G}} w(e) \}$\;
Let $\alpha = w(S_\mathit{init})\epsilon/n$\;
Round all weights $w(e)$ down to integer multiples of $\alpha$\;
$S \gets S_\mathit{init}$\;
$\old{S} \gets S$\;
\Repeat{$\old{S} = S$}{
\ForEach{$k$-replacement $(A,B)$}{
\If{$w^2(A) > w^2(B)$}{
$\old{S} \gets S$\;
$S \gets (S \setminus B) \cup A$\;
\Break\;
}
}
}
\Return $S$\; \caption{Non-Oblivious Local Search for Linear Objective Functions} \label{alg:nols} \end{algorithm}
In the monotone submodular case, we can no longer necessarily represent $f$ as a sum of weights. However, borrowing some intuition from the greedy algorithm, we might decide to replace each weight $w(e)$ in the potential function $w$ with the marginal gain/loss associated with $e$. That is, at the start of each iteration of the local search algorithm, we assign each element $e \in \mathcal{G}$ weight $w(e) = f(S + e) - f(S - e)$, where $S$ is the algorithm's current solution, then we proceed as before. Note that $w(e)$ is simply the marginal gain attained by adding $e$ to $S$, in the case that $e \not\in S$ or the marginal loss suffered by removing $e$ from $S$, in the case that $e \in S$. We define the non-oblivious potential function $w^2$ in terms of the (current) weight function $w$, as before.
Unfortunately, the resulting algorithm may fail to terminate, as the following small example shows. We consider a simple, unweighted coverage function on the universe $U = \{a,b,c,x,y,z\}$, defined as follows. Let: \begin{align*} S_1 &= \{a, b \} & S_3 &= \{x, y \} \\ S_2 &= \{a, c \} & S_4 &= \{x, z \} \end{align*}
We define $\mathcal{G} = \{1,2,3,4\}$ and $f(A) = \left| \bigcup_{i \in A} S_i \right|$ for all $A \subseteq \mathcal{G}$. Finally, we consider the 2-exchange system with only 2 bases: $P = \{1,2\}$ and $Q = \{3,4\}$. For current solution $S = P$ we have $w(1) = w(2) = 1$ and $w(3) = w(4) = 2$. Since $w^2(\{1,2\}) = 2 < 8 = w^2(\{3,4\})$, the 2-replacement $(\{3,4\}, \{1,2\})$ is applied, and the current solution becomes $Q$. In the next iteration, we have $S = Q$, and $w(1) = w(2) = 2$ and $w(3) = w(4) = 1$, so the 2-replacement $(\{1,2\}, \{3,4\})$ is applied by the algorithm. This returns us to the solution to $P$, where the process repeats indefinitely.
\section{The New Algorithm} Intuitively, the problem with this initial approach stems from the fact that the weight function used at each step of the algorithm depends on the current solution $S$ (since all marginals are taken with respect to $S$). Each time the algorithm makes an improvement, it changes the current solution, thereby changing the weights assigned to all elements in the next iteration. Hence, we are effectively making use of an entire family of non-oblivious potential functions, indexed by the current solution $S$. It may be the case that a $k$-replacement $(A,B)$ results in an improvement with respect to the current solution's potential function, but in fact results in a \emph{decreased} potential value in the next iteration, after the weights have been updated.
Surprisingly, we can solve the problem by introducing \emph{even more} variation in the potential function. Specifically, we allow the algorithm to use a different weight function not only for each current solution $S$, but also for \emph{each $k$-replacement $(A,B)$ that is considered}. We give the full algorithm at the end of this section and a detailed analysis in the next.
Our general approach is to consider the elements of a set $X$ in some order $\prec$, and assign to each $e \in X$ in the set the marginal gain in $f$ obtained when it is added to the set containing all the preceding elements of $X$. By carefully updating $\prec$ together with the current solution $S$ at each step, we ensure that the algorithm converges to a local optimum. We now give the details for how are weights are calculated, given the current ordering $\prec$.
At each iteration of the algorithm, before searching an improving $k$-replacement, we assign weights $w$ to all of the elements of the current solution $S$. The weights will necessarily depend on $S$, but remain fixed for all $k$-replacements considered in the current phase. Let $s_i$ be the $i$th element of $S$ in the ordering $\prec$ and let $S_i = \{s_j \in S : j \le i\}$ be the set containing the first $i$ elements of $S$ in the ordering $\prec$. Then, the weight function $w : S \to \ensuremath{\mathbb{R}}_+$ is given by \[ w(s_i) = f(S_{i - 1} + s_i) - f(S_{i - 1}) \] for all $s_i \in S$. Note that our weight function satisfies \begin{equation}
\label{eq:15}
\sum_{x \in S}w(x) = \sum_{i = 1}^{|S|} f(S_{i - 1} + s_i) - f(S_{i - 1}) = f(S) \end{equation}
In order to evaluate each $k$-replacement $(A,B)$, we need to assign weights to the elements in $A \subseteq \mathcal{G} \setminus (S \setminus B)$. We use a different weight function for each $k$-replacement. Suppose that we are considering the $k$-replacement $(A,B)$. Let $a_i$ be the $i$th element of $A$ in the ordering $\prec$ and let $A_i = \{a_j \in A : j \le i\}$ be the set containing the first $i$ elements of $A$ in the ordering $\prec$. Then, we define the weight function $w_{(A,B)} : A \to \ensuremath{\mathbb{R}}_+$ by \[ w_{(A,B)}(a_i) = f((S \setminus B) \cup A_{i - 1} + a_i) - f((S \setminus B) \cup A_{i - 1}) \] for all $a_i \in A$. Note that for every $k$-replacement $(A,B)$, \begin{multline}
\label{eq:16}p
\sum_{x \in A}w_{(A,B)}(x) \ge \sum_{i = 1}^{|A|}\left(f((S \setminus B) \cup A_{i - 1} + a_i) - f((S \setminus B) \cup A_{i - 1})\right) \\ = f((S \setminus B) \cup A) - f(S \setminus B) \ge f(S \cup A) - f(S) \enspace, \end{multline} where the last inequality follows from the decreasing marginals characterization of submodularity. Note that since the function $f$ is \emph{monotone} submodular, all of the weights $w$ and $w_{(A,B)}$ that we consider will be nonnegative. This fact plays a crucial role in our analysis.
Our final algorithm appears in Algorithm \ref{alg:1}. We start from an initial solution $S_\mathit{init} = \{\arg\max_{e \in \mathcal{G}} f(\{e\})\}$, consisting of the singleton set of largest value. Note that after applying a $k$-replacement $(A,B)$, the algorithm updates $\prec$ to ensure that all of the elements of $S \setminus B$ precede those of $A$. As we shall see in the next section, this ensures that the algorithm will converge to a local optimum. As in the linear case, we use the non-oblivious potentials $w^2(B) = \sum_{b \in B}w(b)^2$ and $w_{(A,B)}(A) = \sum_{a \in A}w_{(A,B)}(a)^2$. Again, we note that while \emph{all} of our weights implicitly depend on the current solution, the weights $w_{(A,B)}$ additionally depend on the $k$-replacement $(A,B)$ considered.
Additionally, to ensure polynomial-time convergence, we round all of our weights down to the nearest integer multiple of $\alpha$, depending on the parameter $\epsilon$. This will ensure that every improvement improves the current solution by an additive factor of at least $\alpha^2$. Because of this rounding factor, we must actually work with the following analogs of (\ref{eq:15}) and (\ref{eq:16}): \begin{equation}
\label{eq:15r}
\sum_{x \in S}w(x) \le \sum_{i = 1}^{|S|} \left(f(S_{i - 1} + s_i) - f(S_{i - 1})\right) = \sum_{i = 1}^{|S|} \left(f(S_{i}) - f(S_{i - 1})\right) = f(S) - f(\emptyset) \le f(S) \end{equation} \begin{multline}
\label{eq:16r}
\sum_{x \in A}w_{(A,B)}(x) \ge \sum_{i = 1}^{|A|}\left(f((S \setminus B) \cup A_{i - 1} + a_i) - f((S \setminus B) \cup A_{i - 1}) - \alpha\right) \\ = f((S \setminus B) \cup A) - f(S \setminus B) - |A|\alpha \ge f(S \cup A) - f(S) - |A|\alpha \end{multline}
\begin{algorithm}[t] \caption{Non-Oblivious Local Search} \label{alg:1} \KwIn{\parbox[t]{7in}{\begin{itemize}[parsep=0em,itemsep=0em,topsep=0em,leftmargin=1em] \item Ground set $\mathcal{G}$ \item Membership oracle for $\mathcal{I} \subseteq 2^\mathcal{G}$ \item Value oracle for monotone submodular function $f : 2^\mathcal{G} \to \ensuremath{\mathbb{R}}_+$ \item Approximation parameter $\epsilon \in (0,1)$ \end{itemize}}} \DontPrintSemicolon Let $S_\mathit{init} = \{ \arg\max_{e \in \mathcal{G}} f(\{e\} )\}$\; Let $\delta = \left(1 + \frac{k + 3}{2\epsilon}\right)^{-1}$ and $\alpha = f(S_\mathit{init})\delta/n$\; $S \gets S_\mathit{init}$, $\prec\ \gets $ an arbitrary total ordering on $\mathcal{G}$, and $\old{S} \gets S$\; \Repeat{$\old{S} = S$}{
Sort $S$ according to $\prec$ and let $s_i$ be the $i$th element in $S$\;
$X \gets \emptyset$\;
\For{$i = 1$ \KwTo $|S|$}
{
$w(s_i) \gets \displaystyle\left\lfloor (f(X + s_i) - f(X))/\alpha\right\rfloor\alpha$\;
$X \gets X + s_i$\;
}
\ForEach{$k$ replacement $(A,B)$}{
Sort $A$ according to $\prec$ and let $a_i$ be the $i$th element in $A$\;
$X \gets S \setminus B$\;
\For{$i = 1$ \KwTo $|A|$}{
$w_{(A,B)}(a_i) \gets \displaystyle\left\lfloor (f(X + a_i) - f(X))/\alpha\right\rfloor\alpha$\;
$X \gets X + a_i$\;
}
\If{$w_{(A,B)}^2(A) > w^2(S)$}{
$\prec\ \gets $ the ordering $\prec'$ defined by
$\begin{cases}
x \prec' y & \text{for all $x \in S \setminus B$, $y \in A$}\\
x \prec' y & \text{if $x \prec y$ for all other $x,y$}
\end{cases}$\;
$\old{S} \gets S$\;
$S \gets (S \setminus B) \cup A$\;
\Break\;
}
} } \Return $S$\; \end{algorithm}
\section{Analysis of Algorithm \ref{alg:1}}
We now analyze the approximation and runtime performance of Algorithm \ref{alg:1}. We now consider the worst-case ratio, or \emph{locality gap}, $f(O)/f(S)$ where $S$ is any locally optimal solution (with respect to Algorithm \ref{alg:1}'s potential function) and $O$ is a globally optimal solution. We shall need the following technical lemma, which is a direct consequence of Lemma 1.1 in \cite{Lee-2010a}. We give a proof here for the sake of completeness.
\begin{lemma} \label{lem:submod} Let $f$ be a submodular function on $\mathcal{G}$, Let $T,S \subseteq \mathcal{G}$, and $\{T_i\}_{i=1}^t$ be a partition of $T \setminus S$. Then, \[ \sum_{i=1}^t\left(f(S \cup T_i) - f(S)\right) \ge f(S \cup T) - f(S) \] \end{lemma} \begin{proof}
Define $A_0 = S$ and $A_i = T_i \cup A_{i - 1}$ for all $1 \le i \le t$. Suppose that $T_i = \{ t_j \}_{j = 1}^{|T_i|}$. Then, note that $S \subseteq A_{i - 1}$ and $T_i \cap A_{i - 1} = \emptyset$. Submodularity of $f$ implies that: \begin{align*} f(T_i \cup S) - f(S) &= \sum_{j = 1}^{T_i}\left(f(\{t_l\}_{l = 1}^{j + 1} \cup S) - f(\{t_{l}\}_{l = 1}^j \cup S)\right) \\ &\ge \sum_{j = 1}^{T_i}\left(f(\{t_l\}_{l = 1}^{j + 1} \cup A_{i - 1}) - f(\{t_{l}\}_{l = 1}^j \cup A_{i - 1})\right) \\ &= f(T_i \cup A_{i - 1}) - f(A_{i - 1}) = f(A_i) - f(A_{i - 1}) \end{align*} Now, we have \begin{equation*}
\sum_{i = 1}^t\left(f(S \cup T_i) - f(S)\right) \ge \sum_{i = 1}^t\left(f(A_i) - f(A_{i - 1})\right) = f(A_t) - f(A_0) = f(S \cup T) - f(S) \qedhere \end{equation*} \end{proof}
We begin by considering the approximation ratio of Algorithm \ref{alg:1}. Suppose that $S$ is the locally optimal solution returned by the algorithm on some instance, while $O$ is a global optimum for this instance. Then, for every $k$-replacement $(A,B)$, we must have $w^2_{(A,B)}(A) \le w^2(B)$, where $w$ and each $w_{(A,B)}$ are weight functions determined by the solution $S$. We consider only a particular subset of $k$-replacements in our analysis.
We have $S, O \in \mathcal{I}$ for the $k$-exchange system $\mathcal{I}$. Thus, there must be a collection $Y$ assigning each $e$ of $O$ a neighborhood $Y_e \subseteq S$, satisfying the conditions of Definition \ref{def:k-exchange}. For each $x \in S$, let $P_x$ be the set of all elements in $e \in O$ for which: (1) $x \in Y_e$ and (2) for all $z \in Y_e$, $w(z) \le w(x)$. That is, $P_x$ is the set of all elements of $O$ in which $x$ is the heaviest element. Note that the construction of $P_x$ depends on the fact that the weights $w$ assigned to elements in $S$ are \emph{fixed} throughout each iteration, and do \emph{not} depend on the particular improvement under consideration.
We define $N_x = \bigcup_{e \in P_x}Y_e$, and consider $(P_x, N_x)$. Property \ref{N2} of $Y$ ensures that $|P_x| \le k$. Similarly, property \ref{N1}, together with the fact that all elements $e \in P_x$ have as a common neighbor $x \in Y_e$, ensures that $|N_x| \le 1 + k(k - 1) = k^2 - k + 1$. Finally, property \ref{N3} ensures that $(S \setminus N_x) \cup P_x \in \mathcal{I}$. Thus, $(P_x, N_x)$ is a valid $k$-replacement for all sets $P_x \subseteq O$, $x \in S$. Observe that $\{P_x\}_{x \in S}$ is a partition of $O$. Furthermore, by the definition of $P_x$, we have $w(x) \ge w(z)$ for all $z \in N_x$. Again, this depends on the fact that the weights of elements in $S$ are the same for all $k$-replacements considered by the algorithm during a given phase.
The following extension of a theorem from \cite{Berman-2000} allows us to relate the non-oblivious potentials $w^2$ and $w^2_{(P_x,N_x)}$ to the weight functions $w$ and $w_{(P_x,N_x)}$ for each of our $k$-replacements $(P_x, N_x)$.
\begin{lemma} \label{thm:axy-lemma} For all $x \in S$, and $e \in P_x$, $$w_{(P_x,N_x)}^2(e) - w^2(Y_e - x) \ge w(x) \cdot \left(2 w_{(P_x,N_x)}(e) - w(Y_e)\right) \enspace.$$ \end{lemma} \begin{proof} Let $a = \frac{1}{2}w(Y_e)$, and $b, c$ be such that $w(x) = a + b$ and $w_{(P_x,N_x)}(e) = a + c$ (note that $b$ and $c$ are not necessarily positive). Then, since $e \in P_x$, every element $z$ in $Y_e$ has weight at most $w(x) = a + b$. Furthermore, $w(Y_e - x) = w(Y_e) - w(x) = a - b$. Thus, \begin{equation} w^2(Y_e - x) = \sum_{z \in Y_e - x} w(z)^2 \le \sum_{z \in Y_e - x}(a + b) w(z) = (a + b)(a - b) \label{eq:6} \end{equation} Using \eqref{eq:6} and our definition of $a$, $b$, and $c$, we have \begin{multline*}
w_{(P_x,N_x)}^2(e) - w^2(Y_e - x) - w(x) \cdot (2 w_{(P_x,N_x)}(e) - w(Y_e)) \\ \ge (a + c)^2 - (a + b)(a - b) - (a + b)(2a + 2c - 2a) = (b - c)^2 \ge 0 \enspace. \qedhere \end{multline*} \end{proof}
Using Lemma \ref{thm:axy-lemma} we can prove the following lemma, which uses the local optimality of $S$ to obtain a lower bound on the weight $w(x)$ of each element $x \in S$.
\begin{lemma} For each $x \in S$, $w(x) \ge\sum\limits_{e \in P_x} \left(2w_{(P_x,N_x)}(e) - w(Y_e)\right)$. \label{thm:charge} \end{lemma} \begin{proof} Because $S$ is locally optimal with respect to $k$-replacements, including in particular $(P_x,N_x)$, we must have: \begin{equation}
\label{eq:7} w^2_{(P_x,N_x)}(P_x) \le w^2(N_x) \end{equation} First, we consider the case $w(x) = 0$. Recall that all the weights produced by the algorithm are non-negative. Because $w(x)$ is the largest weight in $N_x$, we must have $w(e) = 0$ for all $e \in N_x$ and so $w^2(N_x) = 0$. Moreover, (\ref{eq:7}) implies that $w^2_{(P_x,N_x)}(P_x) = 0$ as well, and so, in particular, $w_{(P_x,N_x)}(e) = 0$. The claim then follows.
Now, suppose that $w(x) \neq 0$. From (\ref{eq:7}), together with the fact that $x \in Y_e$ for all $e \in P_x$, and the non-negativity of all the weights $w$, we have: \begin{equation} w_{(P_x,N_x)}^2(P_x) \le w^2(N_x) \le w^2(x) + \sum\limits_{e \in P_x} w^2(Y_e - x) \enspace.\label{eq:8} \end{equation} Rearranging \eqref{eq:8} using $w^2_{(P_x,N_x)}(P_x) = \sum_{e \in P_x} w^2_{(P_x,N_x)}(e)$ we obtain: \begin{equation} \sum\limits_{e \in P_x} w^2_{(P_x,N_x)}(e) - w^2(Y_e - x) \le w^2(x) \enspace. \label{eq:9} \end{equation} Applying Lemma \ref{thm:axy-lemma} to each term on the left of \eqref{eq:9} we have: \begin{equation} \sum\limits_{e \in P_x} w(x)\cdot(2w_{(P_x,N_x)}(e) - w(Y_e)) \le \sum\limits_{e \in P_x} w^2_{(P_x,N_x)}(e) - w^2(Y_e - x) \le w^2(x) = w(x)^2 \enspace. \label{eq:19} \end{equation} Dividing by $w(x)$ (recall that $w(x) \neq 0$) then yields \begin{equation*} \sum\limits_{e \in P_x} \left(2w_{(P_x,N_x)}(e) - w(Y_e)\right) \le w(x) \enspace.\qedhere \end{equation*} \end{proof}
We now prove our main result, which gives an upper bound on the locality gap of Algorithm \ref{alg:1}.
\begin{theorem} \label{thm:locality-gap} $\left( \frac{k + 3}{2} + \epsilon \right) f(S) \ge f(O)$ \end{theorem} \begin{proof} Lemma \ref{thm:charge} gives us one inequality for each $x \in S$.
We now add all $|S|$ inequalities to obtain \begin{equation}
\sum_{x \in S}\sum_{e \in P_x}\left(2w_{(P_x,N_x)}(e) -
w(Y_e)\right) \le \sum_{x \in S} w(x) \enspace. \label{eq:10} \end{equation}
We have $\sum_{x \in S}w(x) \le f(S)$ by (\ref{eq:15r}). Additionally, from (\ref{eq:16r}), $ f(S \cup P_x) - f(S) - |P_x|\alpha \le \sum_{e \in P_x} w_{(P_x,N_x)}(e)$ for every $P_x$. Thus, (\ref{eq:10}) implies \begin{equation}
2\sum_{x \in S}\left(f(S \cup P_x) - f(S) - |P_x|\alpha \right) -
\sum_{x \in S}\sum_{e \in P_x}w(Y_e) \le f(S) \enspace. \label{eq:11} \end{equation} Since $\{P_x\}_{x \in S}$ is a partition of $O$, (\ref{eq:11}) is equivalent to \begin{equation}
2\sum_{x \in S}\left(f(S \cup P_x) - f(S)\right) - 2|O|\alpha
- \sum_{e \in O}w(Y_e) \le f(S) \enspace. \label{eq:12} \end{equation} We have $w(x) \ge 0$ for all $x \in S$, and there are at most $k$ distinct $e$ for which $x \in Y_e$, by property \ref{N2} of $Y$. Thus, we have \[\sum_{e \in O}w(Y_e) \le k\sum_{x \in S}w(x) \le kf(S)\enspace,\] by (\ref{eq:15}). Combining this with (\ref{eq:12}), we obtain \begin{equation}
2\sum_{x \in S}\left(f(S \cup P_x) - f(S)\right) - 2|O|\alpha - kf(S) \le f(S) \label{eq:2} \end{equation} Using again the fact that $P$ is a partition of $O$, we can apply Lemma \ref{lem:submod} to the remaining sum on the left of \ref{eq:2}, yielding \begin{equation*}
2\left(f(S \cup O) - f(S)\right) - 2|O|\alpha - kf(S) \le f(S)\enspace \end{equation*} which simplifies to \begin{equation} \label{eq:17}
f(S \cup O) - |O|\alpha \le \frac{k + 3}{2}f(S)\enspace. \end{equation} From the definition of $\alpha$ and the optimality of $O$, we have \begin{equation*} \label{eq:1}
|O|\alpha \le n\alpha = \delta f(S_\mathit{init}) \le \delta f(O)\enspace. \end{equation*} Finally, since $f$ is monotone, we have $f(S \cup O) \ge f(O)$. Thus, (\ref{eq:17}) implies: \begin{equation*} (1 - \delta)f(O) \le \frac{k + 3}{2}f(S)\enspace, \end{equation*} which, after expanding the definition of $\delta$ and simplifying, is equivalent to $f(O) \le \left(\frac{k + 3}{2} + \epsilon\right)f(S)$. \end{proof}
Next, we consider the runtime of Algorithm \ref{alg:1}. Each iteration requires time $O(n)$ to compute the weights for $S$, plus time to evaluate all potential $k$-replacements. There are $O(n^{k + k(k - 1) + 1}) = O(n^{k^2 + 1})$ such $k$-replacements $(A,B)$, and each one can be evaluated in time $O(k^2)$, including the computation of the weights $w_{(A,B)}$. Thus, the total runtime of Algorithm \ref{alg:1} is $O(Ik^2n^{k^2 + 1})$, where $I$ is the number of improvements it makes. The main difficulty remaining in our analysis is showing that Algorithm \ref{alg:1} constantly improves some global quantity, and so $I$ is bounded. Here, we show that although the weights $w$ assigned to elements of $S$ change at each iteration, the non-oblivious potential $w^2(S)$, is monotonically increasing. While the preceding analysis of the locality gap is valid regardless of the particular ordering $\prec$ used to generate the weights, our analysis of the convergence of Algorithm \ref{alg:1} requires that $\prec$ be updated at each phase to maintain the relative ordering of all elements in the current solution.
Finally, we consider what happens to the total squared weight $w^2(S)$ of the current solution after applying an $k$-replacement $(A,B)$. In order to show that our algorithm terminates, we would like to show that this value is strictly increasing. To show this, it is sufficient to show that each weight $w(x)$ for $x \in (S \setminus B) \cup A$ is strictly greater after applying the $k$-replacement than the corresponding weight before. Unfortunately, the weight assigned to an element is highly sensitive to the ordering $\prec$ in which elements are considered. Let $w$ be the weight function for solution $S$, and $w'$ be the weight function for solution $(S \setminus B) \cup A$. If $x \prec y$ for some $x \in A$, $y \in S\setminus B$, then in the updated weight function, we could have $w'(y) < w(y)$, since $w'$ considers the marginal gain of $y$ with respect to a set containing $x$, while $w$ does not (since $x \not\in S$). We avoid this phenomenon by updating the ordering $\prec$ each time an improvement is made. In particular, we ensure that all the elements of $S$ and $A$ are considered in the same relative order, but that all of $A$ comes after all of $S$. As we shall show, this ensures that the weights assigned each individual element in the $S \setminus B$ and $A$ solution do not decrease after applying the $k$-replacement $(A,B)$.
\begin{lemma} Suppose that for some $k$-replacement $(A,B)$ and $\alpha > 0$, we have $w^2_{(A,B)}(A) \ge w^2(B) + \alpha$, and Algorithm \ref{alg:1} applies the replacement $(A,B)$ to $S$ to obtain solution $T = (S \setminus B) \cup A$. Let $w_S$ be the weight function for solution $S$ and $w_T$ be the weight function for solution $T$. Then, $w_T^2(T) \ge w^2_S(S) + \alpha$. \label{lem:monotonic} \end{lemma} \begin{proof} After applying the $k$-replacement $(A,B)$ to $S$, we obtain a new current solution $T = (S \setminus B) \cup A$ and a new ordering $\prec_T$. We now show that for any element $x \in S \setminus B$, we must have $w(x) = w_S(x) \le w_T(x)$ and for any element $y \in A$, we must have $w_{(A,B)}(y) \le w_T(y)$.
In the first case, let $S_x$ be the set of all elements in $S$ that come before $x$ in the ordering $\prec$, and similarly let $T_x$ be the set of all elements in $T$ that come before $x$ in $\prec_T$. Suppose that for some element $z \in T$ we have $z \prec_T x$. Then, since $x \in S \setminus B$, we must have $z \prec x$. Thus, $T_x \subseteq S_x$. It follows directly from the submodularity of $f$ that \begin{equation*} w(x) = w_S(x) = \left\lfloor\frac{f(S_x + x) - f(S_x)}{\alpha}\right\rfloor\alpha \le \left\lfloor\frac{f(T_x + x) - f(T_x)}{\alpha}\right\rfloor\alpha = w_T(x)\enspace. \label{eq:13} \end{equation*}
In the second case, let $A_y$ be the set of all elements of $A$ that come before $y$ in the ordering $\prec$, and let $T_y$ be the set of all elements of $T$ that come before $y$ in the ordering $\prec_T$. Suppose that for some element $z \in T$ we have $z \prec_T y$. Then, since $y \in A$, we must have either $z \in S \setminus B$ or $z \in A$ and $z \prec y$. Thus, $T_y \subseteq (S \setminus B) \cup A_y$, and so \begin{equation*} w_{(A,B)}(y)\! =\! \left\lfloor\frac{f((S\! \setminus\! B) \cup A_y + y) - f((S\!\setminus\! B) \cup A_y)}{\alpha}\right\rfloor\!\alpha \le \left\lfloor\frac{f(T_y + y) - f(T_y)}{\alpha}\right\rfloor\!\alpha = w_T(y) \enspace . \label{eq:14} \end{equation*}
From the above bounds on $w$ and $w_{(A,B)}$, together with the assumption of the lemma, we now have \begin{multline*} w^2_S(S) = \sum_{x \in S \setminus B}w_S(x)^2 + \sum_{x \in B}w_S(x)^2 \le \sum_{x \in S \setminus B}w_S(x)^2 + \sum_{y \in A}w_{(A,B)}(y)^2 + \alpha \\\le \sum_{x \in S \setminus B}w_T(x)^2 + \sum_{y \in A}w_{T}(y)^2 + \alpha = w_T^2(T) + \alpha \enspace .\qedhere \end{multline*} \end{proof}
\begin{theorem} \label{thm:runtime} For any value $\epsilon \in (0,1)$, Algorithm \ref{alg:1} makes at most $O(n^3\epsilon^{-2})$ improvements. \end{theorem} \begin{proof} Note submodularity implies that for any element $e$ and any set $T \subseteq \mathcal{G}$, we must have $f(T + e) - f(T) \le f(\{e\}) \le f(S_\mathit{init})$. In particular, for any solution $S \subseteq \mathcal{G}$ with associated weight function $w$, we have
\[w^2(S) = \sum_{e \in S}w(e)^2 \le |S|f(S_\mathit{init})^2 \le nf(S_\mathit{init})^2\enspace.\]
Consider a given improvement $(A,B)$ applied by the algorithm. Because every weight used in the algorithm is a multiple of $\alpha$, we have $w_{(A,B)}^2(A) > w^2(B)$ only if $w_{(A,B)}^2(A) \ge w^2(B) + \alpha^2$. Let $T = (S \setminus B) \cup A$ be the solution resulting from the improvement, and, as in the proof of Lemma \ref{lem:monotonic}, let $w_S$ be the weight function associated with $S$ and $w_T$ be the weight function associated with $T$. For any $\epsilon > 0$, we have $\alpha > 0$, and hence $\alpha^2 > 0$. Thus, from Lemma \ref{lem:monotonic}, after applying the improvement must we must have $w^2_T(T) \ge w^2_S(S) + \alpha^2$.
Thus, the number of improvements we can make is at most \[\frac{nf(S_\mathit{init})^2 - f(S_\mathit{init})^2}{\alpha^2} = (n - 1)\left(\frac{f(S_\mathit{init})}{\alpha}\right)^2 = (n - 1)\frac{n^2}{\delta^2} = O(n^3\epsilon^{-2}) \enspace.\qedhere \] \end{proof}
\begin{corollary} For any $\epsilon > 0$, Algorithm \ref{alg:1} is a $\frac{k + 3}{2} + \epsilon$ approximation algorithm, running in time $O(\epsilon^{-2}k^2n^{k^2 + 4})$.
\end{corollary}
\section{Open Questions} \label{sec:open-questions} We do not currently have an example for which the locality gap of Algorithm \ref{alg:1} can be as bad as stated, even for specific $k$-exchange systems such as $k$-set packing. In the particular case of weighted independent set in $(k + 1)$-claw free graphs Berman \cite{Berman-2000} gives a tight example that shows his algorithm can return a set $S$ with $\frac{k + 1}{2}w(S) = w(O)$. His example uses only unit weights, and so the non-oblivious potential function is identical to the oblivious one. However, the algorithm of Feldman et al.\ (given here as Algorithm \ref{alg:nols}) considers a larger class of improvements than those considered by Berman, and so Berman's tight example no longer applies, even in the linear case. For the unweighted variant, Hurkens and Schrijver give a lower bound of $k/2 + \epsilon$, where $\epsilon$ depends on the size of the improvements considered. Because the non-oblivious local search routine performs the same as oblivious local search on instances with unit weights (since $1 = 1^2$), this lower bound applies to Algorithm \ref{alg:nols} in the linear case. From a hardness perspective, the best known bound is the $\Omega(k/\ln k)$ NP-hardness result of Hazan, Safra, and Schwartz \cite{Hazan-2006}, for the special case of unweighted $k$-set packing.
In addition to providing a tight example for our analysis, it would be interesting to see if similar techniques could be adapted to apply to more general problems such as matroid $k$ parity in arbitrary matroids (here, even an improvement over $k$ for the general linear case would be interesting) or to non-monotone submodular functions. A major difficulty with the latter generalization is our proof's dependence on the weights' non-negativity, as this assumption no longer holds if our approach is applied directly to non-monotone submodular functions.
\subparagraph*{Acknowledgment}
The author thanks Allan Borodin for providing comments on a preliminary version of this paper.
\end{document}
|
arXiv
|
{
"id": "1108.4983.tex",
"language_detection_score": 0.7296319007873535,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Propagation by Selective Initialization\
and Its Application to\
Numerical Constraint Satisfaction Problems}
\begin{abstract} Numerical analysis has no satisfactory method for the more realistic optimization models. However, with constraint programming one can compute a cover for the solution set to arbitrarily close approximation. Because the use of constraint propagation for composite arithmetic expressions is computationally expensive, consistency is computed with interval arithmetic. In this paper we present theorems that support, selective initialization, a simple modification of constraint propagation that allows composite arithmetic expressions to be handled efficiently.
\end{abstract}
\section{Introduction} The following attributes all make an optimization problem more difficult: having an objective function with an unknown and possibly large number of local minima, being constrained, having nonlinear constraints, having inequality constraints, having both discrete and continuous variables. Unfortunately, faithfully modeling an application tends to introduce many of these attributes. As a result, optimization problems are usually linearized, discretized, relaxed, or otherwise modified to make them feasible according to conventional methods.
One of the most exciting prospects of constraint programming is that such difficult optimization problems can be solved without these possibly invalidating modifications. Moreover, constraint programming solutions are of known quality: they yield intervals guaranteed to contain all solutions. Equally important, constraint programming can prove the absence of solutions.
In this paper we only consider the core of the constraint programming approach to optimization, which is to solve a system of nonlinear inequalities:
\begin{equation} \label{nonLinSys} \begin{array}{ccccccccc} g_1(x_1&,& x_2 &,& \ldots &,& x_m) & \leq & 0 \\ g_2(x_1 &,& x_2 &,& \ldots &,& x_m) & \leq & 0 \\ \multicolumn{9}{c}{\dotfill} \\ g_k(x_1 &,& x_2 &,& \ldots &,& x_m) & \leq & 0 \\ \end{array} \end{equation} It is understood that it may happen that $g_i= -g_j$ for some pairs $i$ and $j$, so that equalities are a special case. If this occurs, then certain obvious optimizations are possible in the methods described here.
The ability to solve systems such as (\ref{nonLinSys}) supports optimization in more ways than one. In the first place, these systems occur as conditions in some constrained optimized problems. Moreover, one of \myVec{g}{k} could be defined as $f(\myVec{x}{m}) - c$, where $f$ is the objective function and where $c$ is a constant. By repeatedly solving such a system for suitably chosen $c$, one can find the greatest value of $c$ for which (\ref{nonLinSys}) is found to have no solution. That value is a lower bound for the global minimum \cite{vnmdn03}.
This approach handles nonlinear inequalities with real variables. It also allows some or all variables to be integer by regarding integrality as a constraint on a real variable \cite{bnldr97}.
All constraint programming work in this direction has been based on interval arithmetic. The earliest work \cite{BNR88} used a generic propagation algorithm based directly on domain reduction operators for primitive arithmetic constraints. These constraints included $sum(x,y,z)$ defined as $x+y=z$ for all reals $x$, $y$, and $z$. Also included was $prod(x,y,z)$ defined as $xy=z$ for all reals $x$, $y$, and $z$.
This was criticized in \cite{bmcv94} which advocated the use of composite arithmetic expression directly rather than reducing them to primitive arithmetic constraints. In \cite{bggp99,vnmdn01b} it was acknowledged that the generic propagation algorithm is not satisfactory for CSPs that derive from composite arithmetic expressions. These papers describe propagation algorithms that exploit the structure of such expressions and thereby improve on what is attainable by evaluating such expressions in interval arithmetic. Selective Initialization was first described in \cite{vnmdn03a}. This was done under the tacit assumption that all default domains are $[-\infty,+\infty]$. As a result some of the theorems in that paper are not as widely applicable as claimed.
All of these researches are motivated by the severe difficulties experienced by conventional numerical analysis to solve practical optimization problems. They can be regarded as attempts to fully exploit the potential of interval arithmetic. In this paper we also take this point of view. We show that, though Equation (\ref{nonLinSys}) can contain arbitrarily large expressions, only a small modification of the generic propagation algorithm is needed to optimally exploit the structure of these expressions. This is made possible by a new canonical form for Equation (\ref{nonLinSys}) that we introduce in this paper. In addition to supporting our application of constraint processing to solving systems similar to Equation (\ref{nonLinSys}), this canonical form exploits the potential for parallelism in such systems.
\section{A software architecture for optimization problems} \begin{figure}
\caption{ A software architecture for optimization problems. }
\label{softArch}
\end{figure}
In Figure~\ref{softArch} we propose a hierarchical software architecture for optimization problems. Each layer is implemented in terms of the layer below.
In the introduction we briefly remarked on how layer 4 can be reduced to layer 3. More detail is given in \cite{vnmdn03}. For the transition between layers 0 and 1 there is much material in the interval arithmetic literature. The part that is relevant to constraint processing can be found in \cite{hckvnmdn01}. In the present paper we present a new method for implementing layer 3 in terms of layer 2. But first we review the transition between layers 1 and 2.
\section{Preliminaries} In this section we provide background by reviewing some basic concepts. These reviews also serve to establish the terminology and notation used in this paper. The first few sections apply to all constraint satisfaction problems, not only to numerical ones.
\subsection{Constraint satisfaction problems} A {\em constraint satisfaction problem (CSP)} consists of a set of {\em constraints}. Each of the variables in the constraint is associated with a {\em domain}, which is the set of values that are possible for the variable concerned. Typically, not all sets of values can be domains. For example, sets of real values are restricted to intervals, as described later.
A {\em valuation} is a tuple indexed by variables where the component indexed by $v$ is an element of the domain of $v$. A {\em solution} is a valuation such that each constraint is true if every variable in the constraint is substituted by the component of the valuation indexed by the variable. The set of solutions is a set of valuations; hence a set of tuples; hence a \emph{relation}. We regard this relation as the relation defined by the CSP. In this way the relation that is the meaning of a constraint in one CSP can be defined by another. This gives CSPs an hierarchical structure.
With each constraint, there is an associated {\em domain reduction operator}\/; DRO for short. This operator may remove from the domains of each of the variables in the constraint certain values that do not satisfy the constraint, given that the other variables of the constraint are restricted to their associated domains. Any DRO is contracting, monotonic, and idempotent.
When the DROs of the constraints are applied in a ``fair'' order, the domains converge to a limit or one of the domains becomes empty. A sequence of DROs activations is \emph{fair} if every one of them occurs an infinite number of times. The resulting Cartesian product of the domains becomes the greatest common fixpoint of the DROs \cite{vnmd97,aptEssence}. If one of the domains becomes empty, it follows that no solutions exist within the initial domains.
In practice, we are restricted to the domains that are representable in a computer. As there are only a finite number of these, any fair sequence of DRO applications yields domains that remain constant from a certain point onwards.
\subsection{Constraints} According to the usual terminology in constraint programming, a constraint states that a certain relation holds between its arguments. But in first-order predicate logic the same role is played by an \emph{atomic formula}. In this paper we adopt the terminology of first-order predicate logic for the meaning of ``atomic formula'' and we reserve ``constraint'' for a special case.
Thus an atomic formula consists of a predicate symbol with terms as arguments. A \emph{term} is a function symbol with terms as arguments. What makes an atomic formula first-order is that a predicate symbol can only occur as the outermost symbol in the formula.
At first sight, the inequalities in Equation (\ref{nonLinSys}) are atomic formulas. In fact, they follow the related, but different, usage that prevails in informal mathematics. The inequality
\begin{equation} \label{inEq} \begin{array}{ccccccccc} g_i(x_1&,& x_2 &,& \ldots &,& x_m) & \leq & 0 \end{array} \end{equation} is an atomic formula where $\leq$ is the predicate symbol with two arguments, which are the terms $g_i(x_1, x_2 , \ldots , x_m)$ and $0$. A possible source of confusion is that in mathematics $g_i(x_1, x_2 , \ldots , x_m)$ is not necessarily interpreted as a syntactically concrete term, but as an abstractly conceived function $g_i$ applied to the arguments \myVc{x}{1}{m}. The function could be defined by means of a term that looks quite different; such a term could be nested and contain several function symbols. For example, Equation (\ref{inEq}) could be $g_i(x_1,x_2) \leq 0$ with $g_i$ defined as $g_i(x,y) = x^2 + xy - y^2 $ for all $x$ and $y$. Accordingly, the atomic formula corresponding to Equation (\ref{inEq}) is \begin{equation} \label{inEqEx} \begin{array}{ccccccccc} \leq(+(sq(x),-(\times(x,y),sq(y))),0). \end{array} \end{equation} Taking advantage of infix and postfix notation this becomes $x^2 + xy - y^2 \leq 0$.
A \emph{constraint} is an atomic formula without function symbols. An example of such an atomic formula is $sum(x,y,z)$, which is a ternary constraint whose relation is defined by $x+y=z$ for all reals $x$, $y$, and $z$.
In this paper we translate Equation~(\ref{inEqEx}) to a CSP with the set of constraints $$ \{t_1 \leq 0,sum(t_2,t_3,t_1),sq(x,t_2),sum(t_5,t_3,t_4),
prod(x,y,t_5),sq(y,t_4) \} $$
Consider a constraint $c(\myVec{x}{n})$. The meaning of predicate symbol $c$ is a relation, say $r$. For all $i \in \{1,\ldots,n\}$, a value $a_i$ for variable $x_i$ is \emph{inconsistent} with respect to $r$ and domains $X_1,\ldots,X_{i-1},X_{i+1},\ldots,X_{n}$ iff it is not the case that \begin{eqnarray*}
\exists a_1 \in X_1,\ldots,
\exists a_{i-1} \in X_{i-1},\exists a_{i+1}\in X_{i+1},\ldots,
\exists a_{n} \in X_{n} \hbox{ such that } \\
\langle \myVec{a}{n} \rangle \in r. \end{eqnarray*} A DRO for $c$ may replace one or more of the domains \myVec{X}{n} by a subset of it if the set difference between the old and the new domains contains inconsistent values only. A DRO is \emph{optimal} if every domain is replaced by the smallest domain containing all its consistent values. We call a constraint \emph{primitive} if an optimal DRO is available for it that is sufficiently efficiently computed. What is sufficient depends on the context.
\subsection{Constraint propagation} To gain information about the solution set, inconsistent values are removed as much as possible with modest computation effort. For example, DROs can be applied as long as they remove inconsistent values. It is the task of a constraint propagation algorithm to reach as quickly as possible a set of domains that cannot be reduced by any DRO. Many versions of this algorithm exist \cite{aptEssence,rthpra01}. They can be regarded as refinements of the algorithm in Figure~\ref{lGPA}, which we refer to as the \emph{generic propagation algorithm}\/; GPA for short.
GPA maintains a pool of DROs, called {\em active set}. No order is specified for applying these operators. Notice that the active set $A$ is initialized to contain all constraints.
\begin{figure}
\caption{ A pseudo-code for GPA. }
\label{lGPA}
\end{figure}
\subsection{Intervals} A {\em floating-point number} is any element of $F \cup \{-\infty, +\infty\}$, where $F$ is a finite set of reals. A {\em floating-point interval} is a closed connected set of reals, where the bounds, in so far as they exist, are floating-point numbers. When we write ``interval'' without qualification, we mean a floating-point interval. For every bounded non-empty interval $X$, $lb(X)$ and $rb(X)$ denote the least and the greatest element of $X$ respectively. They are referred to as the left and the right bound of $X$. If $X$ is not bounded from below, then $lb(X) = -\infty$. Similarly, if $X$ is not bounded from above, then $rb(X) = +\infty$. Thus, $ X = [lb(X), rb(X)]$ is a convenient notation for all non-empty intervals, bounded or not.
A {\em box} is a Cartesian product of intervals.
\subsection{Solving inequalities in interval arithmetic}
Moore's idea of solving inequalities such as those in Equation (\ref{nonLinSys}) by means of interval arithmetic is at least as important as the subsequent applications of interval constraints to this problem.
Suppose we wish to investigate the presence of solutions of a single inequality in Equation (\ref{nonLinSys}) in a box $X_1 \times \cdots \times X_m$. Then one evaluates in interval arithmetic the expression in the left-hand side. As values for the variables $x_1 , \ldots , x_m$ one uses the intervals $X_1 , \cdots , X_m$. Suppose the result is the interval $[a_i,b_i]$. We have exactly one of the following three cases. If $0 < a_i$ for at least one $i$, then there are no solutions. If $b_i \leq 0$ for all $i$, then all tuples in $X_1 \times \cdots \times X_m$ are solutions. If $a_i \leq 0 < b_i$ for at least one $i$, then either of the above may be true. In this same case of $a_i \leq 0 < b_i$, it may also be that some of the tuples in $X_1 \times \cdots \times X_m$ are solutions, while others are not.
Again, in the case where $a_i \leq 0 < b_i$, it may be possible to split $X_1 \times \cdots \times X_m$. In this way, a more informative outcome may be obtained for one or both of the results of splitting. Such splits can be repeated as long as possible and necessary.
\subsection{Interval CSPs}
Problems in a wide variety of application areas can be expressed as CSPs. Domains can be as different as booleans, integers, finite symbolic domains, and reals. In this paper we consider {\em Interval CSPs} (ICSPs), which are CSPs where all domains are intervals and all constraints are primitive.
ICSPs are important because they encapsulate what can be efficiently computed; they represent Layer 2 in the software architecture of Figure~\ref{softArch}. The layer is distinct from Layer 3 because in Equation (\ref{nonLinSys}) there typically occur atomic formulas that contain function symbols.
To emphasize the role of ICSPs as a layer of software architecture, we view them as a \emph{virtual machine}, with a function that is similar to those for Prolog or Java. Just as a program in Prolog or Java is translated to virtual machine instructions, a system such as Equation (\ref{nonLinSys}) can be translated to an ICSP, as described in a later section.
The instructions of the ICSP level are DROs, one for each constraint. As an example of such an ICSP virtual machine instruction, consider the DRO for product constraint. It reduces the box $[a,b]\times [c,d]\times [e,f]$ to the box that has the projections \begin{eqnarray} \varphi([a,b] & \cap & ([e,f]/[c,d])) \nonumber\\ \varphi([c,d] & \cap & ([e,f]/[a,b])) \nonumber\\ \varphi([e,f] & \cap & ([a,b]*[c,d])) \nonumber \end{eqnarray} Here $\varphi$ is the function that yields the smallest interval containing its argument.
Of particular interest is the effect of the DRO when all variables have $[-\infty,+\infty]$ as domain. For each argument, the domain after application of the DRO is defined as the \emph{default domain} of that argument. Typically, default domains are $[-\infty,+\infty]$. Notable exceptions include the constraint $sin(x,y)$ (defined as $y = \sin(x)$), where the default domain of $y$ is $[-1,1]$. Another is $\hbox{sq}(x,y)$ (defined as $y = x^2$), where the default domain of $y$ is $[0,\infty]$.
A difference with other virtual machines is that a program for the ICSP virtual machine is an unordered collection of DROs. Programs for other virtual machines are ordered sequences of instructions. In those other virtual machines, the typical instruction does not specify the successor instruction. By default this is taken to be the next one in textual order. Execution of the successor is implemented by incrementing the instruction counter by one.
The simplicity of the instruction sequencing in conventional virtual (and actual) machines is misleading. Many instruction executions concern untypical instructions, where the next instruction is specified to be another than the default next instruction. Examples of such untypical instructions are branches (conditional or unconditional) and subroutine jumps.
In the ICSP virtual machine, the DROs are the instructions, and they form an unordered set. Instead of an instruction counter specifying the next instruction, there is the active set of GPA containing the set of possible next instructions. Instead of an instruction or a default rule determining the next instruction to be executed, GPA selects in an unspecified way which of the DROs in the active set to execute. In this way, programs can be declarative: instructions have only meaning in terms of \emph{what} is to be computed. \emph{How} it is done (instruction sequencing), is the exclusive task of the virtual machine.
\subsection{A canonical form for nonlinear numerical inequalities} \label{arch}
Equation (\ref{nonLinSys}) may have multiple occurrences of variables in the same formula. As there are certain advantages in avoiding such occurrences, we rewrite without loss of generality the system in Equation~(\ref{nonLinSys}) to the canonical form shown in Figure~\ref{singleSys}.
\begin{figure}\label{singleSys}
\end{figure}
In Figure~\ref{singleSys}, the expressions for the functions $g_1, \ldots, g_k$ have no multiple occurrences of variables. As a result, they have variables \myVec{y}{n} instead of \myVec{x}{m}, with $m \leq n$ as in Equation~(\ref{nonLinSys}). This canonical form is obtained by associating with each of the variables $x_j$ in Equation~(\ref{nonLinSys}) an equivalence class of the variables in Figure~\ref{singleSys}. This is done by replacing in Equation~(\ref{nonLinSys}) each occurrence of a variable by a different element of the corresponding equivalence class. This is possible by ensuring that each equivalence class is as large as the largest number of multiple occurrences. The predicate $\hbox{{\it allEq }}$ is true if and only if all its real-valued arguments are equal.
An advantage of this translation is that evaluation in interval arithmetic of each expression gives the best possible result, namely the range of the function values. At the same time, the $\hbox{{\it allEq }}$ constraint is easy to enforce by making all intervals of the variables in the constraint equal to their common intersection. This takes information into account from all $k$ inequalities. If the system in its original form as in Equation~\ref{nonLinSys}, with multiple occurrences, would be translated to a CSP, then only multiple occurrences in a single expression would be exploited at one time.
In the coming sections and without loss of generality, we will only consider expressions without multiple occurrences of variables.
\subsection{Translating nonlinear inequalities to ICSPs} \label{translation} ICSPs represent what we \emph{can} solve. They consist of atomic formulas without function symbols that, moreover, have efficient DROs. Equation (\ref{nonLinSys}) exemplifies what we \emph{want} to solve: it consists of atomic formulas typically containing deeply nested terms.
\paragraph{The tree form of a formula} We regard a first-order atomic formula as a tree. The unique predicate symbol is the root. The terms that are the arguments of the formula are also trees and they are the subtrees of the root. If the term is a variable, then the tree only has a root, which is that variable. A term may also be a function symbol with one or more arguments, which are terms. In that case, the function symbol is the root with the argument terms as subtrees.
In the tree form of a formula the leaves are variables. In addition, we label every node that is a function symbol with a unique variable.
Any constants that may occur in the formula are replaced by unique variables. We ensure that the associated domains contain the constants and are as small as possible.
\paragraph{Translating a formula to an ICSP} The tree form of a formula thus labeled is readily translated to an ICSP. The translation has a set of constraints in which each element is obtained by translating an internal node of the tree.
The root translates to $p(\myVc{x}{0}{n-1})$ where $p$ is the predicate symbol that is the root and \myVc{x}{0}{n-1} are the variables labeling the children of the root.
A non-root internal node of the form $f(\myVc{t}{0}{n-1})$ translates to $F(\myVc{x}{0}{n-1}, y)$, where \begin{itemize} \item $y$ is the variable labeling the node \item \myVc{x}{0}{n-1} are the variables labeling the child nodes \item $F$ is the relation defined by $F(\myVc{a}{0}{n-1}, v)$ iff $v= f(\myVc{a}{0}{n-1})$ for all $\myVc{a}{0}{n-1},v$. \end{itemize}
\subsection{Search}
Propagation may terminate with small intervals for all variables of interest. This is rare. More likely, propagation leaves a large box as containing all solutions, if any. To obtain more information about possibly existing solutions, it is necessary to split an ICSP into two ICSPs and to apply propagation to both. An ICSP $S'$ is a result of splitting an ICSP $S$ if $S'$ has the same constraints as $S$ and differs only in the domain for one variable, say, $x$. The domain for $x$ in $S'$ is the left or right half of the domain for $x$ in $S$.
A \emph{search strategy} for an ICSP is a binary tree representing the result of successive splits. Search strategies can differ greatly in the effort required to carry them to completion. The most obvious search strategy is the \emph{greedy strategy}\/: the one that ensures that all intervals become small enough by choosing a widest domain as the one to be split.
This is a plausible strategy in the case where the ICSP has a few point solutions. In general, the set of solutions is a continuum: a line segment, a piece of a surface, or a variety in a higher dimensional space that has positive volume. In such cases we prefer the search to result in a single box containing all solutions. Of course we also prefer such a box to be as small as possible. The greedy search strategy splits the continuum of solutions into an unmanageably large number of small boxes. It is not clear that the greedy strategy is preferable even in the case of a few well-isolated point solutions. In general we need a search strategy other than the greedy one.
A more promising search strategy was first described in the \verb+absolve+ predicate of the BNR Prolog system \cite{BNR88} and by \cite{bmcv94}, where it is called \emph{box consistency}.
The box consistency search strategy selects a variable and a domain bound. Box consistency uses binary search to determine a boundary interval that can be shown to contain no solutions. This boundary interval can then be removed from the domain, thus shrinking the domain. This is repeated until a boundary interval with width less than a certain tolerance is found that cannot be shown to contain no solutions. When this is the case for both boundaries of all variables, the domains are said to be \emph{box consistent} with respect to the tolerance used and with respect to the method for showing inconsistency. When this method is interval arithmetic, we obtain \emph{functional box consistency}. When it is propagation, then it is called \emph{relational box consistency} \cite{vnmdn01b}.
All we need to know about search in this paper is that greedy search and box consistency are both search strategies and that both can be based on propagation. Box consistency is the more promising search strategy. Thus we need to compare interval arithmetic and propagation as ways of showing that a nonlinear inequality has no solutions in a given box. This we do in section~\ref{PSI}.
\section{Propagation with selective initialization}
\label{PSI}
Suppose we have a term that can be evaluated in interval arithmetic. Let us compare the interval that is the result of such an evaluation with the effect of GPA on the ICSP obtained by translating the term as described in section~\ref{translation}.
To make the comparison possible we define evaluation of a term in interval arithmetic. The definition follows the recursive structure of the term: a term is either a variable or it is a function symbol with terms as arguments. If the term is an interval, then the result is that interval. If the argument is function $f$ applied to arguments, then the result is $f$ evaluated in interval arithmetic applied to the results of evaluating the arguments in interval arithmetic. This assumes that every function symbol denotes a function that is defined on reals as well as on intervals. The latter is called the \emph{interval extension} of the former. For a full treatment of interval extensions, see \cite{moore66,nmr90,hnsn92}. The following lemma appears substantially as Theorem 2.6 in \cite{chn98}.
\begin{lemma} \label{basic} Let $t$ be a term that can be evaluated in interval arithmetic. Let the variables of $t$ be $x_1,\ldots,x_n$. Let $y$ be the variable associated with the root of the tree form of $t$. Let $S$ be the ICSP that results from translating $t$, where the domains of $x_1,\ldots,x_n$ are $X_1,\ldots,X_n$ and where the domains of the internal variables are $[-\infty,+\infty]$. After termination of GPA applied to $S$, the domain for $y$ is the interval that results from interval arithmetic evaluation of $t$. \end{lemma}
\begin{proof} Suppose that a variable of a constraint has domain $[-\infty,+\infty]$. After applying the DRO for that constraint, this domain has become the result of the interval arithmetic operation that obtains the domain for this variable from the domains of the other variables of the constraint.\\ According to \cite{vnmd97,aptEssence}, every fair sequence of DROs converges to the same domains for the variables. These are also the domains on termination of GPA. Let us consider a fair sequence that begins with a sequence $s$ of DROs that mimics the evaluation of $t$ in interval arithmetic. At the end of this, $y$ has the value computed by interval arithmetic. This shows that GPA gives a result that is a subinterval of the result obtained by interval arithmetic.\\ GPA terminates after activating the DROs in $s$. This is because in the interval arithmetic evaluation of $t$ an operation is only performed when its arguments have been evaluated. This means that the corresponding DRO only changes one domain. This domain is the domain of a unique variable that occurs in only one constraint that is already in the active set. Therefore none of the DRO activations adds a constraint to the active set, which is empty after $s$. \end{proof}
GPA yields the same result whatever the way constraints are selected in the active set. Therefore GPA always gives the result of interval arithmetic evaluation. However, GPA may obtain this result in an inefficient way by selecting constraints that have no effect. This suggests that the active set be structured in a way that reflects the structure of $t$. This approach has been taken in \cite{bggp99,vnmdn01b}.
The proof shows that, if the active set had not contained any of the constraints only involving internal variables, these constraints would have been added to the active set by GPA. This is the main idea of selective initialization. By initializing and ordering the active set in a suitable way and leaving GPA otherwise unchanged, it will obtain the interval arithmetic result with no more operations than interval arithmetic. This assumes the optimization implied by the Totality Theorem in \cite{hckmdw98}.
\begin{definition} A constraint is a \emph{seed constraint} iff at least one of its variables has a domain that differs from the default domain assigned to that variable. \end{definition}
For example, the term $\sin(x_1)+\sin(x_2)$ translates to an ICSP with constraints $sum(u,v,y)$, $sin(x_1,u)$, and $sin(x_2,v)$. When the domains are $[-\infty,+\infty]$ for all variables, then the seed constraints are $sin(x_1,u)$ and $sin(x_2,v)$. When the domains are $[-\infty,+\infty]$ for $x_1$, $x_2$, and $y$; $[-1,1]$ for $u$ and $v$, then $sum(u,v,y)$ is the one seed constraint. When the domains are $[-\infty,+\infty]$ for $x_2$ and $y$; $[-1,1]$ for $x_1$, $u$ and $v$, then the seed constraints are $sum(u,v,y)$ and $sin(x_1,u)$.
\begin{definition} Let PSI (Propagation with Selective Initialization) be GPA except for the following modifications.\\ (a) PSI only applies to ICSPs generated by translation from an atomic formula. \\ (b) The active set is a priority queue that is ordered according to the distance from the root of the node that generated the constraint. The greater that distance, the earlier the item is removed from the queue. \\ (c) The active set contains all seed constraints and no other ones. \end{definition}
Lemma~\ref{basic} says that GPA simulates interval arithmetic as far as the result is concerned. It does not say anything about the efficiency with which the result is obtained. Theorem~\ref{simulate} says that PSI obtains the result as efficiently as it is done in interval arithmetic. This assumes the functionality optimization in the DROs \cite{hckmdw98}.
\begin{theorem} \label{simulate} Let $S$ be the ICSP obtained by translating a term $t$ in variables \myVc{x}{1}{n}, where these variables have domains \myVc{X}{1}{n}. Applying PSI to S terminates after selecting no constraint more than once. Moreover, the root variable ends up with $Z$ as domain where $Z$ is the interval resulting from evaluating $t$ with \myVc{x}{1}{n} substituted by \myVc{X}{1}{n}. \end{theorem}
\begin{proof} Suppose GPA is applied to $S$ in such a way that all non-seed constraints are selected first. The execution of the DRO corresponding to these non-seed constraints does not change any domains. Therefore these DRO executions do not add any constraints. As a result, the effect of applying GPA is the same as when the active set would have been initialized with only the seed constraints.\\ Suppose the seed constraints are selected according to priority order. This ensures that no future constraint selection re-introduces a constraint previously selected. Thus GPA terminates after activating every seed constraint exactly once.\\ Such an execution of GPA coincides step by step with that of PSI. As GPA terminates with the correct result, so does PSI. \end{proof}
\section{Using ICSPs to solve inequalities}
We briefly reviewed how interval arithmetic can solve systems of nonlinear inequalities. The fundamental capability turned out to be that of evaluating a term in interval arithmetic. We saw that this can also be done by applying propagation to ICSPs generated from arithmetic terms. We now investigate how to extend this to the use of ICSPs to solve nonlinear inequalities.
\subsection{Using ICSPs to solve a single inequality}
Suppose $S$ is the ICSP resulting from translating $$ g_i(\myVc{x}{1}{m}) \leq 0 $$ Let $y$ be the variable labeling the left child of the root; that is, the variable representing the value of the left-hand side. Let \myVc{X}{1}{m} be the domains in $S$ of \myVc{x}{1}{m}, respectively.
Now suppose that GPA is applied to $S$. One possible initial sequence of DRO activations is the equivalent of interval arithmetic evaluation of the left-hand side, leaving $y \leq 0$ as the only constraint in the active set with the domain for $y$ equal to $[a_i,b_i]$, the value in interval arithmetic of $g_i(\myVc{X}{1}{m})$.
At this stage the DRO for $y \leq 0$ is executed. If $0 < a_i$, then failure occurs. If $b_i \leq 0$, then the domain for $y$ is unchanged. Therefore, no constraint is added to the active set. Termination occurs with nonfailure. There is no change in the domain of any of \myVc{x}{1}{m}. The third possibility is that $a_i \leq 0 < b_i$. In this case, the domain for $y$ shrinks: the upper bound decreases from $b_i$ to $0$. This causes the constraints to be brought into the active set that correspond to nodes at the next lower level in the tree. This propagation may continue all the way down to the lowest level in the tree, resulting in shrinking of the domain of one or more of \myVc{x}{1}{m}.
Let us compare this behaviour with the use of interval arithmetic to solve the same inequality. In all three cases, GPA gives the same outcome as interval arithmetic: failure or nonfailure. In the first two cases, GPA gives no more information than interval arithmetic. It also does no more work.
In the third case, GPA may give more information than interval arithmetic: in addition to the nonfailure outcome, it may shrink the domain of one or more of \myVc{x}{1}{m}. This is beyond the capabilities of interval arithmetic, which is restricted to transmit information about arguments of a function to information about the value of the function. It cannot transmit information in the reverse direction. To achieve this extra capability, GPA needs to do more work than the equivalent of interval arithmetic evaluation.
In the above, we have assumed that GPA magically avoids selecting constraints in a way that is not optimal. In such an execution of GPA we can recognize two phases: an initial phase that corresponds to evaluating the left-hand side in interval arithmetic, followed by a second phase that starts with the active set containing only the constraint $y \leq 0$. When we consider the nodes in the tree that correspond to the constraints that are selected, then it is natural to call the first phase bottom-up (it starts at the leaves and ends at the root) and the second phase top-down (it starts at the root and may go down as far as to touch some of the leaves). The bottom-up phase can be performed automatically by the PSI algorithm.
The start of the top-down phase is similar to the situation that occurs in search. In both search and in the top-down phase a different form of selective initialization can be used, shown in the next section.
The bottom-up phase and the top-down phase are separated by a state in which the active set only contains $y \leq 0$. For reasons that become apparent in the next section, we prefer a separate treatment of this constraint: not to add it to the active set and to execute the shrinking of the domain for $y$ as an extraneous event. This is then a special case of termination of GPA, or its equivalent PSI, followed by the extraneous event of shrinking one domain.
The Pseudo-code for PSI algorithm is given in Figure~\ref{PSIAlg}.
\begin{figure}
\caption{ Pseudo-code for Propagation with Selective Initialization (PSI). }
\label{PSIAlg}
\end{figure} The correctness of PSI algorithm can be easily deduced from the following theorem.
\begin{theorem} \label{modEval} Consider the ICSP $\mathcal{S}$ obtained from the tree $T$ of the atomic formula $g_i(x_1,\ldots,x_n) \leq 0$. Suppose we modify GPA so that the active set is initialized to contain instead of all constraints only seed constraints. Suppose also that the active set is a priority queue in which the constraints are ordered according to the level they occupy in the tree $T$, with those that are further away from the root placed nearer to the front of the queue. Then GPA terminates with the same result as when the active set would have been initialized to contain all constraints. \end{theorem}
\begin{proof} As we did before, suppose that in GPA the active set $A$ is initialized with all constraints such that the seed constraints are at the end of the active set. Applying any DRO of a constraint that is not a seed constraint will not affect any domain. Thus, the constraints that are not seed constraints can be removed from the active set without changing the result of GPA. Since the GPA does not specify any order, $A$ can be ordered as desired. Here we choose to order it in such a way we get an efficient GPA when used to evaluate an expression (see previous section). \end{proof}
\section{Selective Initialization for search}
Often we find that after applying GPA to an ICSP $S$, the domain $X$ for one of the variables, say $x$, is too wide. Search is then necessary. This can take the form of splitting $S$ on the domain for $x$. The results of such a split are two separate ICSPs $S_1$ and $S_2$ that are the same as $S$ except for the domain of $x$. In $S_1$, $x$ has as domain the left half of $X$; in $S_2$, it is the right half of $X$.
However, applying GPA to $S_1$ and $S_2$ entails duplicating work already done when GPA was applied to $S$. When splitting on $x$ after termination of the application of GPA to $S$, we have the same situation as at the beginning of the downward phase of applying GPA to an inequality: the active set is empty and an extraneous event changes the domain of one variable to a proper subset.
The following theorem justifies a form of the PSI algorithm where the active set is initialized with what is typically a small subset of all constraints.
\begin{theorem} \label{modSolveGeneral} Let $T$ be the tree obtained from the atomic formula $g_i(x_1,\ldots,x_n) \leq 0$. Let $\mathcal{S}$ be the ICSP obtained from $T$. Let $x$ be a variable in $\mathcal{S}$. Suppose we apply GPA to $\mathcal{S}$. After the termination of GPA, suppose the domain of $x$ is changed to an interval that is a proper subset of it. If we apply GPA to $\mathcal{S}$ with an active set initialized with the constraints only involving $x$, then GPA terminates with the same result as when the active set would have been initialized to contain all constraints. \end{theorem}
\begin{proof} To prove Theorem~\ref{modSolveGeneral}, we should show that initializing GPA with all constraints gives the same results as when it is initialized with only the constraints involving $x$.
Since no ordering is specified for the active set of GPA, we choose an order in which the constraints involving $x$ are at the end of the active set. Because DROs are idempotent, all constraints at the front of the active set, different from those involving $x$, do not affect any domain. Thus removing them from the active set in the initialization process does not change the fixpoint of the GPA. Thus, Theorem~\ref{modSolveGeneral} is proved. \end{proof}
\section{Further work} We have only considered the application of selective initialization to solve a single inequality. A conjunction of inequalities such as Equation~(\ref{nonLinSys}) can be solved by solving each in turn. This has to be iterated because the solving of another inequality affects the domain of an inequality already solved. This suggests performing the solving of all inequalities in parallel. Doing so avoids the waste of completing an iteration on the basis of unnecessarily wide intervals. It also promises speed-up because many of the DRO activations only involve variables that are unique to the inequality. In the current version of the design of our algorithm, we combine this parallelization with a method of minimizing the complexity usually caused by multiple occurrences of variables.
\section{Conclusions} Before interval methods it was not clear how to tackle numerically realistic optimization models. Only with the advent of interval arithmetic in the 1960s \cite{moore66} one could for the first time at least say: ``If only we had so much memory and so much time, then we could solve this problem.''
Interval arithmetic has been slow in developing. Since the 1980s constraint programming has added fresh impetus to interval methods. Conjunctions of nonlinear inequalities, the basis for optimization, can be solved both with interval arithmetic and with constraint programming. In this paper we relate these two approaches.
It was known that constraint propagation subsumes interval arithmetic. It was also clear that using propagation for the special case of interval arithmetic evaluation is wasteful. In this paper we present an algorithm for propagation by Selective Initialization that ensures that propagation is as efficient in the special case of interval arithmetic evaluation. We also apply Selective Initialization for search and for solving inequalities. Preliminary results on a parallel version of the methods presented here suggest that realistic optimization models will soon be within reach of modest computing resources.
\section*{Acknowledgments} We acknowledge generous support by the University of Victoria, the Natural Science and Engineering Research Council NSERC, the Centrum voor Wiskunde en Informatica CWI, and the Nederlandse Organisatie voor Wetenschappelijk Onderzoek NWO.
\end{document}
|
arXiv
|
{
"id": "0404034.tex",
"language_detection_score": 0.8973519206047058,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\title{ On some new difference sequence spaces of fractional order}
\author[label1]{Serkan Demiriz\corref{cor1}} \ead{[email protected]}\cortext[cor1]{Corresponding Author (Tel: +90 356 252 16 16, Fax: +90 356 252 15 85)} \author[label3]{Osman Duyar} \ead{[email protected]}
\address[label1]{Department of Mathematics, Faculty of Arts and Science, Gaziosmanpa\c{s}a University,\\
60250 Tokat, Turkey } \address[label3]{Anatolian High School, 60200 Tokat, Turkey }
\begin{abstract} Let $\Delta^{(\alpha)}$ denote the fractional difference operator. In this paper, we define new difference sequence spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$. Also, the $\beta-$ dual of the spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ are determined and calculated their Schauder basis. Furthermore, the classes $(\mu(\Gamma,\Delta^{(\alpha)},u):\lambda)$ where $\mu\in \{c_{0},c\}$ and $\lambda \in \{c_{0},c,\ell_{\infty},\ell_1\}$ . \end{abstract}
\begin{keyword}
Difference operator $\Delta^{(\alpha)}$, Sequence spaces, $\beta-$dual, Matrix transformations.
\end{keyword}
\end{frontmatter}
\noindent
\section{Preliminaries,background and notation}
By a \textit{sequence space}, we mean any vector subspace of $\omega$, the space of all real or complex valued sequences $x=(x_k)$. The well-known sequence spaces that we shall use throughout this paper are as following:
$\ell_{\infty}$: the space of all bounded sequences,
$c$: the space of all convergent sequences,
$c_{0}$: the space of all null sequences,
$cs$: the space of all sequences which form convergent series,
$\ell_1$: the space of all sequences which form absolutely convergent series,
$\ell_p$: the space of all sequences which form $p$-absolutely convergent series,\\ where $1< p<\infty$.
Let $X,Y$ be two sequence spaces and $A=(a_{nk})$ be an infinite matrix of real or complex numbers $a_{nk}$, where $n,k\in \mathbb{N}$. Then, we say that $A$ defines a matrix mapping from $X$ into $Y$, and we denote it by writing $A:X\rightarrow Y$, if for every sequence $x=(x_{k})\in \lambda$ the sequence $Ax=\{(Ax)_{n}\}$, the $A$-transform of $x$, is in $Y$; where \begin{equation}\label{1} (Ax)_{n}=\sum_{k} a_{nk}x_{k}, \quad (n\in \mathbb{N}). \end{equation} For simplicity in notation, here and in what follows, the summation without limits runs from $0$ to $\infty$. By $(X:Y)$, we denote the class of all matrices $A$ such that $A:X\rightarrow Y$. Thus, $A\in (X:Y)$ if and only if the series on the right side of (\ref{1}) converges for each $n\in \mathbb{N}$ and every $x\in X$, and we have $Ax=\{(Ax)_{n}\}_{n\in \mathbb{N}}\in Y$ for all $x\in X$. A sequence $x$ is said to be $A$-summable to $\alpha$ if $Ax$ converges to $\alpha$ which is called as the $A$-limit of $x$.
If a normed sequence space $X$ contains a sequence $(b_{n})$ with the property that for every $x\in X$ there is a unique sequence of scalars $(\alpha_{n})$ such that $$ \lim_{n\rightarrow \infty}
\|x-(\alpha_{0}b_{0}+\alpha_{1}b_{1}+...+\alpha_{n}b_{n})\|=0, $$ then $(b_{n})$ is called a \emph{Schauder basis} (or briefly \emph{basis}) for $X$. The series $\sum \alpha_{k}b_{k}$ which has the sum $x$ is then called the expansion of $x$ with respect to $(b_{n})$, and written as $x=\sum \alpha_{k}b_{k}$.
A matrix $A=(a_{nk})$ is called a \emph{triangle} if $a_{nk}=0$ for $k>n$ and $a_{nn}\neq 0$ for all $n\in \mathbb{N}$. It is trivial that $A(Bx)=(AB)x$ holds for triangle matrices $A,B$ and a sequence $x$. Further, a triangle matrix $U$ uniquely has an inverse $U^{-1}=V$ that is also a triangle matrix. Then, $x=U(Vx)=V(Ux)$ holds for all $x\in \omega$. We write additionally $\mathcal{U}$ for the set of all sequences $u$ such that $u_k\neq0$ for all $k \in \mathbb{N}$.
For a sequence space $X$, the \emph{matrix domain} $X_{A}$ of an infinite matrix $A$ is defined by \begin{equation}\label{2} X_{A}=\left\{x=(x_{k})\in \omega: Ax\in \lambda\right\}, \end{equation} which is a sequence space. The approach constructing a new sequence space by means of the matrix domain of a particular limitation method has been recently employed by Wang \cite{w}, Ng and Lee \cite{nglee}, Ayd{\i}n and Ba\c{s}ar \cite{cafb} and Altay and Ba\c{s}ar \cite{bafb5}.
The gamma function may be regarded as a generalization of $n!$ ($n-$factorial), where $n$ is any positive integer. The gamma function $\Gamma$ is defined for all $p$ real numbers except the negative integers and zero. It can be expressed as an improper integral as follows: \begin{equation}\label{gamma} \Gamma(p)=\int_0^\infty e^{-t}t^{p-1}dt. \end{equation}
From the equality (\ref{gamma}) we deduce following properties:\\
(i) If $n\in\mathbb{N}$ then we have $\Gamma(n+1)=n!$
(ii) If $n\in\mathbb{R}-\{0,-1,-2,-3...\}$ then we have $\Gamma(n+1)=n\Gamma(n)$.\\
For a proper fraction $\alpha$, Baliarsingh and Dutta have defined a fractional difference operators $\Delta^{\alpha}:w\rightarrow w$, $\Delta^{(\alpha)}:w\rightarrow w$ and their inverse in \cite{bail3} as follows: \begin{equation}\label{op} \Delta^{\alpha}(x_{k})=\sum_{i=0}^\infty(-1)^i\frac{\Gamma(\alpha+1)}{i!\Gamma(\alpha+1-i)}x_{k+i} \end{equation}
\begin{equation}\label{op2} \Delta^{(\alpha)}(x_{k})=\sum_{i=0}^\infty(-1)^i\frac{\Gamma(\alpha+1)}{i!\Gamma(\alpha+1-i)}x_{k-i} \end{equation}
\begin{equation}\label{op3} \Delta^{-\alpha}(x_{k})=\sum_{i=0}^\infty(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)}x_{k+i} \end{equation} and \begin{equation}\label{op4} \Delta^{(-\alpha)}(x_{k})=\sum_{i=0}^\infty(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)}x_{k-i} \end{equation} where we assume throughout the series defined in (\ref{op})-(\ref{op4}) are convergent. In particular, for $\alpha=\frac{1}{2}$,
\begin{eqnarray}\label{space1} \nonumber \Delta^{1/2}x_k&=&x_{k}-\frac{1}{2}x_{k+1}-\frac{1}{8}x_{k+2}-\frac{1}{16}x_{k+3}-\frac{5}{128}x_{k+4} -\frac{7}{256}x_{k+5}-...\\ \nonumber \Delta^{-1/2}x_k&=&x_{k}+\frac{1}{2}x_{k+1}+\frac{3}{8}x_{k+2}+\frac{5}{16}x_{k+3}+\frac{35}{128}x_{k+4} +\frac{63}{256}x_{k+5}+...\\ \nonumber \Delta^{(1/2)}x_k&=&x_{k}-\frac{1}{2}x_{k-1}-\frac{1}{8}x_{k-2}-\frac{1}{16}x_{k-3}-\frac{5}{128}x_{k-4} -\frac{7}{256}x_{k-5}-...\\ \nonumber \Delta^{(-1/2)}x_k&=&x_{k}+\frac{1}{2}x_{k-1}+\frac{3}{8}x_{k-2}+\frac{5}{16}x_{k-3}+\frac{35}{128}x_{k-4} +\frac{63}{256}x_{k-5}+.... \end{eqnarray}
Baliarsingh have been defined the spaces $X(\Gamma,\Delta^{\alpha},u)$ for $X\in\{\ell_\infty,c_0,c\}$ by introducing the fractional difference operator $\Delta^{\alpha}$ and a positive fraction $\alpha$ in \cite{bail1}. In this article, Baliarsingh have been studied some topological properties of the spaces $X(\Gamma,\Delta^{\alpha},u)$ and established their $\alpha-$, $\beta-$ and $\gamma-$duals.
Following \cite{bail1}, we introduce the sequence spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ and obtain some results related to these sequence spaces. Furthermore, we compute the $\beta-$dual of the spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ . Finally, we characterize some matrix transformations on new sequence spaces.
\section{The sequence spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ }
In this section, we define the sequence spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ and examine the some topological properties of this sequence spaces.
The notion of difference sequence spaces was introduced by Kýzmaz \cite{kzmaz}. It was generalized by Et and Çolak \cite{metolak} as follows:
Let $m$ be a non-negative integer. Then $$ \Delta^{m}(X)=\{x=(x_k): \Delta^{m}x_k\in X\} $$ where $\Delta^{0}x=(x_k)$, $\Delta^{m}x=(\Delta^{m-1}x_k-\Delta^{m-1}x_{k+1})$ for all $k\in\mathbb{N}$ and
\begin{equation}\label{delta} \Delta^{m}x_k=\sum_{i=0}^{m}(-1)^i\left(m \atop i\right)x_{k+i}. \end{equation}
Furthermore, Malkowsky E., et al.\cite{emmmss} have been introduced the spaces \begin{equation}\label{delta} \Delta_u^{(m)}X=\{x\in \omega: \Delta_u^{(m)}x\in X\} \end{equation} where $\Delta_u^{(m)}x=u\Delta^{(m)}x$ for all $x\in\omega$ . In this study, the operator $\Delta_u^{(m)}:\omega\rightarrow\omega$ was defined as follows: \begin{equation}\label{delta2} \Delta^{(m)}x_k=\sum_{i=0}^{m}(-1)^i\left(m \atop i\right)x_{k-i}. \end{equation}
Let $\alpha$ be a proper fraction and $u\in\mathcal{U}$. We define the sequence spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ as follows:
\begin{equation}\label{space1} c_0(\Gamma,\Delta^{(\alpha)},u)=\{x\in\omega: (\sum_{j=0}^ku_j\Delta^{(\alpha)}x_j)\in c_0\} \end{equation} and \begin{equation}\label{space2} c(\Gamma,\Delta^{(\alpha)},u)=\{x\in\omega: (\sum_{j=0}^ku_j\Delta^{(\alpha)}x_j)\in c\}. \end{equation}
Now, we define the triangle matrix $\Delta_u^{(\alpha)}(\Gamma)=(\tau_{nk})$, \begin{equation}\label{matrix} \tau_{nk}=\left\{\begin{array}{ll}
\displaystyle
\sum_{i=0}^{n-k}(-1)^i\frac{\Gamma(\alpha+1)}{i!\Gamma(\alpha+1-i)}u_{i+k},
& (0\leq k\leq n)\\ \displaystyle 0, & k>n \end{array}\right. \end{equation} for all $k,n\in \mathbb{N}$. Further, for any sequence $x=(x_k)$ we define the sequence $y=(y_k)$ which will be used, as the $\Delta_u^{(\alpha)}(\Gamma)-$transform of $x$, that is \begin{eqnarray}\label{space1} \nonumber y_k&=&\sum_{j=0}^ku_j\Delta^{(\alpha)}x_j~=~\sum_{j=0}^ku_j(x_{j}-\alpha x_{j-1}+\frac{\alpha(\alpha-1)}{2!}x_{j-3}+...)\\ &=&\sum_{j=0}^{k}\bigg(\sum_{i=0}^{k-j}(-1)^i\frac{\Gamma(\alpha+1)}{i!\Gamma(\alpha+1-i)}u_{i+j}\bigg)x_{j} \end{eqnarray} for all $k\in\mathbb{N}$.
It is natural that the spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ may also be defined with the notation of (\ref{2}) that \begin{equation}\label{fac1} c_0(\Gamma,\Delta^{(\alpha)},u)=(c_0)_{\Delta_u^{(\alpha)}(\Gamma)}~~ \textrm{and}~~~~c(\Gamma,\Delta^{(\alpha)},u)=c_{\Delta_u^{(\alpha)}(\Gamma)}. \end{equation} Before the main result let us give some lemmas, which we use frequently throughout this study, with respect to $\Delta^{(\alpha)}$ operator.
\begin{lm}\label{lem1}\cite[Theorem 2.2]{bail3} $$\Delta^{(\alpha)}o\Delta^{(\beta)}=\Delta^{(\beta)}o\Delta^{(\alpha)}=\Delta^{(\alpha+\beta)}.$$ \end{lm}
\begin{lm}\label{lem2}\cite[Theorem 2.3]{bail3}
$$\Delta^{(\alpha)}o\Delta^{(-\alpha)}=\Delta^{(-\alpha)}o\Delta^{(\alpha)}=Id$$ where $Id$ the identity operator on $\omega$. \end{lm}
\begin{thm} The sequence spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ are $BK-$spaces with the norm \begin{equation}\label{norm}
\|x\|_{c_0(\Gamma,\Delta^{(\alpha)},u)}=\|x\|_{c(\Gamma,\Delta^{(\alpha)},u)}
=\sup_k\bigg|\sum_{j=0}^ku_j\Delta^{(\alpha)}x_j\bigg|. \end{equation}
\end{thm} \begin{pf} Since (\ref{fac1}) holds and $c_{0},c$ are $BK-$ spaces with respect to their natural norms (see \cite[pp. 16-17]{fb}) and the matrix $\Delta_u^{(\alpha)}(\Gamma)=(\tau_{nk})$ is a triangle, Theorem 4.3.12 Wilansky \cite[pp. 63]{aw} gives the fact that $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ are $BK-$ spaces with the given norms. This completes the proof . \end{pf}
Now, we may give the following theorem concerning the isomorphism between the spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$, $c(\Gamma,\Delta^{(\alpha)},u)$ and $c_0,c$, respectively:
\begin{thm} The sequence spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ are linearly isomorphic to the spaces $c_0$ and $c$, respectively, i.e, $c_0(\Gamma,\Delta^{(\alpha)},u)\cong c_0$ and $c(\Gamma,\Delta^{(\alpha)},u)\cong c$. \end{thm} \begin{pf} We prove the theorem for the space $c(\Gamma,\Delta^{(\alpha)},u)$. To prove this, we should show the existence of a linear bijection between the spaces $c(\Gamma,\Delta^{(\alpha)},u)$ and $c$. Consider the transformation $T$ defined, with the notation of (\ref{space1}), from $c(\Gamma,\Delta^{(\alpha)},u)$ to $c$ by $x\mapsto y=Tx=\Delta_u^{(\alpha)}(\Gamma)x$. The linearity of $T$ is clear. Further, it is trivial that $x=\theta$ whenever $Tx=\theta$ and hence $T$ is injective. Let be $y=(y_k)\in c$ and we define a sequence $x=(x_k)\in c(\Gamma,\Delta^{(\alpha)},u)$ by
\begin{equation}\label{ters} x_k=\sum_{i=0}^{\infty}(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)}\frac{y_{k-i}-y_{k-i-1}}{u_{k-i}}. \end{equation} Then by Lemma \ref{lem2}, we deduce that \begin{eqnarray}\label{space5} \nonumber \sum_{j=0}^ku_j\Delta^{(\alpha)}x_j&=&\sum_{j=0}^ku_j\Delta^{(\alpha)}\bigg(\sum_{i=0}^{\infty}(-1)^i \frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)}\frac{y_{j-i}-y_{j-i-1}}{u_{j-i}}\bigg)\\ \nonumber &=&\sum_{j=0}^ku_j\Delta^{(\alpha)}\bigg(\Delta^{(-\alpha)}\bigg(\frac{y_{j}-y_{j-1}}{u_{j}}\bigg)\bigg)\\ &=&\sum_{j=0}^k(y_{j}-y_{j-1})~=~y_k \end{eqnarray} Hence, $x\in c(\Gamma,\Delta^{(\alpha)},u)$ so $T$ is surjective. Furthermore one can easily show that $T$ is norm preserving. This complete the proof. \end{pf}
\section{The $\beta$-Dual of The Spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$ }
In this section, we determine the $\beta$-dual of the spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$. For the sequence spaces $X$ and $Y$, define the set $S(X,Y)$ by \begin{equation}\label{3.1} S(X,Y)=\{z=(z_{k})\in \omega: xz=(x_{k}z_{k})\in Y \textrm{ for all } x\in X\}. \end{equation} With the notation of (\ref{3.1}), $\beta-$ dual of a sequence space $X$ is defined by $$ X^{\beta}=S(X,cs). $$
\begin{lm}\label{e4.10} $A\in (c_0:c)$ if and only if \begin{equation}\label{e5.2} \lim_{n\rightarrow\infty} a_{nk}=\alpha_k \ \textrm{for each fixed}\ k\in \mathbb{N}, \end{equation}
\begin{equation}\label{e5.3}
\sup_{n\in \mathbb{N}}\sum_k|a_{nk}|<\infty. \end{equation} \end{lm}
\begin{lm}\label{e4..2} $A\in (c:c)$ if and only if (\ref{e5.2}) and (\ref{e5.3}) hold, and \begin{equation} \lim_{n\rightarrow\infty}\sum_ka_{nk}\quad\ \textrm{exists}. \end{equation} \end{lm}
\begin{lm}\label{lem3} $A=(a_{nk})\in (\ell_\infty:\ell_\infty)$ if and only if \begin{equation}\label{7c}
\sup_n\sum_{k} |a_{nk}|<\infty. \end{equation} \end{lm}
\begin{thm}\label{t3} Define the sets $\Gamma_{1}$, $\Gamma_{2}$ and a matrix $T=(t_{nk})$ by $$ t_{nk}=\left\{\begin{array}{ll}
\displaystyle t_{k}-t_{k+1},
& (k<n)\\ \displaystyle t_{n}, & (k=n) \\\displaystyle\ 0, & (k>n) \end{array}\right. $$ for all $k,n\in\mathbb{N}$ where $t_{k}=a_{k}\sum_{i=0}^{k}(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)} \frac{1}{u_{k-i}}$ \begin{eqnarray*} \Gamma_{1}&=&\bigg\{ a=(a_n)\in w:\lim_{n\rightarrow\infty}t_{nk}=\alpha_k ~\textrm{exists for each}\ k\in \mathbb{N}\bigg\} \\
\Gamma_{2}&=&\bigg\{ a=(a_n)\in w:\sup_{n\in\mathbb{N}}\sum_{k}^{}|t_{nk}|<\infty\bigg\}\\ \Gamma_{3}&=&\bigg\{ a=(a_n)\in w:\sup_{n\in\mathbb{N}}\lim_{n\rightarrow\infty}\sum_k t_{nk}\quad\ \textrm{exists}\bigg\}. \end{eqnarray*}
Then, $\{c_0(\Gamma,\Delta^{(\alpha)},u)\}^{\beta}=\Gamma_1\cap \Gamma_2$ and $\{c(\Gamma,\Delta^{(\alpha)},u)\}^{\beta}=\Gamma_1\cap \Gamma_2\cap \Gamma_3$.
\end{thm} \begin{pf} We prove the theorem for the space $c_0(\Gamma,\Delta^{(\alpha)},u)$. Let $a=(a_n)\in w$ and $x=(x_k)\in c_0(\Gamma,\Delta^{(\alpha)},u) $. Then, we obtain the equality \begin{eqnarray}\label{beta} \nonumber \sum_{k=0}^{n}a_kx_k&=&\sum_{k=0}^{n}\bigg[a_{k}\sum_{i=0}^{\infty}(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)} \frac{1}{u_{k-i}}\bigg](y_{k-i}-y_{k-i-1})\\ \nonumber &=&\sum_{k=0}^{n}\bigg[a_{k}\sum_{i=0}^{k}(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)} \frac{1}{u_{k-i}}\bigg](y_{k-i}-y_{k-i-1})\\ \nonumber &=&\sum_{k=0}^{n-1}\bigg[a_{k}\sum_{i=0}^{k}(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)} \frac{1}{u_{k-i}}-a_{k+1}\sum_{i=0}^{k+1}(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)} \frac{1}{u_{k+1-i}}\bigg]y_k\\ \nonumber & &+~\bigg[a_{n}\sum_{i=0}^{n}(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)} \frac{1}{u_{n-i}}\bigg]y_n\\ &=&T_ny \end{eqnarray}
Then we deduce by (\ref{beta}) that $ax=(a_kx_k)\in cs$ whenever $x=(x_k)\in c_0(\Gamma,\Delta^{(\alpha)},u)$ if and only if $T y\in c$ whenever $y=(y_k)\in c_0$. This means that $a=(a_k)\in \{c_0(\Gamma,\Delta^{(\alpha)},u)\}^\beta$ if and only if $T\in (c_0:c)$. Therefore, by using Lemma \ref{e4.10}, we obtain : \begin{eqnarray} &&\lim_{n\rightarrow\infty}t_{nk}=\alpha_k\qquad \textrm{exists for each}\ k\in\mathbb{N},\\
&&\sup_{n\in\mathbb{N}}\sum_{k}^{}|t_{nk}|<\infty. \end{eqnarray} Hence, we conclude that
$\{c_0(\Gamma,\Delta^{(\alpha)},u)\}^{\beta}=\Gamma_1\cap \Gamma_2$.
\end{pf}
\section{Some matrix transformations related to the sequence spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$}
In this final section, we state some results which characterize various matrix mappings on the spaces $c_0(\Gamma,\Delta^{(\alpha)},u)$ and $c(\Gamma,\Delta^{(\alpha)},u)$.
We shall write throughout for brevity that \begin{equation}\label{z1} \widetilde{a}_{nk}=z_{nk}-z_{n,k+1}~~\textrm{and}~~b_{nk}=\sum_{j=0}^{n}\bigg(\sum_{i=0}^{n-j}(-1)^i \frac{\Gamma(\alpha+1)} {i!\Gamma(\alpha+1-i)}u_{i+j}\bigg) a_{jk} \end{equation} for all $k,n\in \mathbb{N}$ , where $$z_{nk}=a_{nk}\sum_{i=0}^{k}(-1)^i\frac{\Gamma(1-\alpha)}{i!\Gamma(1-\alpha-i)} \frac{1}{u_{k-i}}.$$
Now, we may give the following theorem.
\begin{thm}\label{teo4} Let $\lambda$ be any given sequence space and $\mu\in \{c_{0},c\}$. Then, $A =(a_{nk})\in (\mu(\Gamma,\Delta^{(\alpha)},u):\lambda) $ if and only if ~$C\in (\mu: \lambda)$ and \begin{eqnarray}\label{Bn} C^{(n)}\in (\mu:c) \end{eqnarray}
for every fixed $n\in\mathbb{N}$, where $c_{nk}=\widetilde{a}_{nk}$ and $C^{(n)}=(c_{mk}^{(n)})$ with $$ c^{(n)}_{mk}=\left\{\begin{array}{ll}
\displaystyle z_{nk}-z_{n,k+1},
& (k<m)\\ \displaystyle z_{nm}, & (k=m) \\\displaystyle\ 0, & (k>m) \end{array}\right. $$ for all $k,m\in\mathbb{N}.$ \end{thm}
\begin{pf} Let $\lambda$ be any given sequence space. Suppose that (\ref{z1}) holds between the entries of $A=(a_{nk})$ and $C=(c_{nk})$, and take into account that the spaces $\mu(\Gamma,\Delta^{(\alpha)},u)$ and $\mu$ are linearly isomorphic.
Let $A =(a_{nk})\in (\mu(\Gamma,\Delta^{(\alpha)},u):\lambda) $ and take any $y=(y_k)\in \mu$. Then, $C\Delta_u^{(\alpha)}(\Gamma)$ exists and $\{a_{nk}\}_{k\in \mathbb{N}}\in \mu(\Gamma,\Delta^{(\alpha)},u)^{\beta}$ which yields that (\ref{Bn}) is necessary and $\{c_{nk}\}_{k\in \mathbb{N}}\in \mu^{\beta}$ for each $n\in \mathbb{N}$. Hence, $Cy$ exists for each $y\in\mu$ and thus by letting $m\rightarrow\infty$ in the equality \begin{equation} \sum_{k=0}^{m}a_{nk}x_k=\sum^{m-1}_{k=0} (z_{nk}-z_{n,k+1})y_{k}+z_{nm}y_m; \quad (m,n\in \mathbb{N}) \end{equation} we have that $Cy=Ax$ and so we have that $C\in (\mu:\lambda)$.
Conversely, suppose that $C\in (\mu:\lambda)$ and (\ref{Bn}) hold, and take any $x=(x_{k})\in \mu(\Gamma,\Delta^{(\alpha)},u)$. Then, we have $\{c_{nk}\}_{k\in \mathbb{N}}\in \mu^{\beta}$ which gives together with (\ref{Bn}) that $\{a_{nk}\}_{k\in \mathbb{N}}\in \mu(\Gamma,\Delta^{(\alpha)},u)^{\beta}$ for each $n\in \mathbb{N}$. So, $Ax$ exists. Therefore, we obtain from the equality \begin{equation} \sum_{k=0}^{m} c_{nk}y_{k}=\sum_{k=0}^{m}\bigg[\sum_{j=k}^{m}\bigg(\sum_{i=0}^{j-k}(-1)^i\frac{\Gamma(\alpha+1)} {i!\Gamma(\alpha+1-i)}u_{i+k}\bigg) c_{nj}\bigg]x_{k} \quad \textrm{for all}\ n\in \mathbb{N}, \end{equation} as $m\rightarrow \infty$ that $Ax=Cy$ and this shows that $A\in (\mu(\Gamma,\Delta^{(\alpha)},u):\lambda)$. This completes the proof. \end{pf}
\begin{thm}\label{teo5} Suppose that the entries of the infinite matrices $A=(a_{nk})$ and $B=(b_{nk})$ are connected with the relation (\ref{z1}) and $\lambda$ be given sequence space and $\mu\in \{c_0,c\}$. Then $A=(a_{nk})\in(\lambda:\mu(\Gamma,\Delta^{(\alpha)},u))$ if and only if $B=(b_{nk})\in(\lambda:\mu)$ . \end{thm}
\begin{pf} Let $z=(z_k)\in\lambda$ and consider the following equality with (\ref{z1}) $$ \sum_{k=0}^{m} b_{nk}z_{k}=\sum_{j=0}^{n}\bigg[\sum_{k=0}^{m}\bigg(\sum_{i=0}^{n-j}(-1)^i\frac{\Gamma(\alpha+1)} {i!\Gamma(\alpha+1-i)}u_{i+j}\bigg) a_{jk}\bigg]z_{k} \quad \ (m,n\in \mathbb{N}), $$ which yields as $m\rightarrow\infty$ that $(Bz)_n=[\Delta_u^{(\alpha)}(\Gamma)(Az)]_n$. Hence, we obtain that $Az\in\mu(\Gamma,\Delta^{(\alpha)},u)$ whenever $z\in\lambda$ if and only if $Bz\in\mu$ whenever $z\in\lambda$. \end{pf}
We will have several consequences by using Theorem \ref{teo4} and Theorem
\ref{teo5}. But we must give firstly some
relations which are important for consequences:
\begin{eqnarray}
\label{co1}&& \sup_{n\in\mathbb{N}}\sum_k|a_{nk}|<\infty \\ \label{co2}&& \lim_{n\rightarrow\infty}a_{nk}=\alpha_k\quad \textrm{exists for each fixed}\ k\in \mathbb{N} \\ \label{co3}&& \lim_{k\rightarrow\infty}a_{nk}=0 \quad \textrm{for each fixed}\ n\in \mathbb{N}\\ \label{co4}&& \lim_{n\rightarrow\infty}\sum_ka_{nk}\quad\ \textrm{exists}\\ \label{co5}&&\lim_{n\rightarrow\infty}\sum_ka_{nk}=0\\
\label{co6}&& \sup_{K\in \mathcal{F}}\sum_n\bigg|\sum_{k\in K}a_{nk}\bigg|<\infty\\
\label{co7}&& \lim_{n\rightarrow\infty}\sum_k |a_{nk}|=0\\
\label{co8}&&\sup_{n,k}|a_{nk}|<\infty\\
\label{co9}&&\lim_{m\rightarrow\infty}\sum_k|a_{nk}|=\sum_k|\alpha_{k}| \end{eqnarray}
Now, we can give the corollaries:
\begin{cor} The following statements hold:
(i) $A=(a_{nk})\in (c_0(\Gamma,\Delta^{(\alpha)},u):\ell_\infty)= (c(\Gamma,\Delta^{(\alpha)},u):\ell_\infty)$ if and only if (\ref{co1}) holds with $\tilde{a}_{nk}$ instead of $a_{nk}$ and (\ref{Bn}) also holds.
(ii) $A=(a_{nk})\in (c_0(\Gamma,\Delta^{(\alpha)},u):c)$ if and only if (\ref{co1})and (\ref{co2}) hold with $\tilde{a}_{nk}$ instead of $a_{nk}$ and (\ref{Bn}) also holds.
(iii) $A=(a_{nk})\in (c_0(\Gamma,\Delta^{(\alpha)},u):c_0)$ if and only if (\ref{co1}) and (\ref{co3}) hold with $\tilde{a}_{nk}$ instead of $a_{nk}$ and (\ref{Bn}) also holds.
(iv) $A=(a_{nk})\in (c(\Gamma,\Delta^{(\alpha)},u):c)$ if and only if (\ref{co1}), (\ref{co2}) and (\ref{co4}) hold with $\tilde{a}_{nk}$ instead of $a_{nk}$ and (\ref{Bn}) also holds.
(v) $A=(a_{nk})\in (c(\Gamma,\Delta^{(\alpha)},u):c_0)$ if and only if (\ref{co1}), (\ref{co3}) and (\ref{co5}) hold with $\tilde{a}_{nk}$ instead of $a_{nk}$ and (\ref{Bn}) also holds.
(vi) $A=(a_{nk})\in (c_0(\Gamma,\Delta^{(\alpha)},u):\ell)=(c(\Gamma,\Delta^{(\alpha)},u):\ell)$ if and only if (\ref{co6}) holds with $\tilde{a}_{nk}$ instead of $a_{nk}$ and (\ref{Bn}) also holds. \end{cor}
\begin{cor} The following statements hold:
(i)$A=(a_{nk})\in (\ell_\infty:c_0(\Gamma,\Delta^{(\alpha)},u))$ if and only if (\ref{co7}) hold with $b_{nk}$ instead of $a_{nk}$.
(ii)$A=(a_{nk})\in (c:c_0(\Gamma,\Delta^{(\alpha)},u))$ if and only if (\ref{co1}), (\ref{co3}) and (\ref{co5}) hold with $b_{nk}$ instead of $a_{nk}$.
(iii) $A=(a_{nk})\in (c_0:c_0(\Gamma,\Delta^{(\alpha)},u))$ if and only if (\ref{co1}) and (\ref{co3}) hold with $b_{nk}$ instead of $a_{nk}$.
(iv)$A=(a_{nk})\in (c:c_0(\Gamma,\Delta^{(\alpha)},u))$ if and only if (\ref{co3}) and (\ref{co8}) hold with $b_{nk}$ instead of $a_{nk}$.
(v) $A=(a_{nk})\in (\ell_\infty:c(\Gamma,\Delta^{(\alpha)},u))$ if and only if (\ref{co2}) and (\ref{co9}) hold with $b_{nk}$ instead of $a_{nk}$.
(vi) $A=(a_{nk})\in (c:c(\Gamma,\Delta^{(\alpha)},u))$ if and only if (\ref{co1}), (\ref{co2}) and (\ref{co4}) hold with $b_{nk}$ instead of $a_{nk}$.
(vii) $A=(a_{nk})\in (c_0:c(\Gamma,\Delta^{(\alpha)},u))$ if and only if (\ref{co1}) and (\ref{co2}) hold with $b_{nk}$ instead of $a_{nk}$.
(viii) $A=(a_{nk})\in (\ell:c(\Gamma,\Delta^{(\alpha)},u))$ if and only if (\ref{co2}) and (\ref{co8}) hold with $b_{nk}$ instead of $a_{nk}$.
\end{cor}
\end{document}
|
arXiv
|
{
"id": "1408.1269.tex",
"language_detection_score": 0.5634278655052185,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\title{Extensions and applications of ACF mappings}
\author{Jean-Philippe Chancelier}
\ead{[email protected]}
\address{Université Paris-Est, CERMICS (ENPC), 6-8 Avenue Blaise Pascal, Cité Descartes , F-77455 Marne-la-Vallée}
\begin{abstract} Using a definition of ASF sequences derived from the definition of asymptotic contractions of the final type of ACF, we give some new fixed points theorem for cyclic mappings and alternating mapping which extend results from \cite[Theorem 2]{suzuki2007} and \cite[Theorem 1]{zhang2007}. \end{abstract}
\begin{keyword}
Nonexpansive mappings \sep Fixed points \sep Meir-Keeler contraction \sep ACF mappings
\end{keyword} \end{frontmatter}
\def\mathop{\normalfont Fix}{\mathop{\normalfont Fix}} \def\stackrel{\mbox{\tiny def}}{=}{\stackrel{\mbox{\tiny def}}{=}} \def\mathop{\mbox{\rm Argmax}}{\mathop{\mbox{\rm Argmax}}} \def\mathop{\mbox{\rm Argmin}}{\mathop{\mbox{\rm Argmin}}} \def{\mathbb P}{{\mathbb P}} \def{\mathbb C}{{\mathbb C}} \def{\mathbb E}{{\mathbb E}} \def{\cal E}{{\cal E}} \def{\cal F}{{\cal F}} \def{\cal H}{{\cal H}} \def{\cal V}{{\cal V}} \def{\cal W}{{\cal W}} \def{\cal U}{{\cal U}} \def{\mathbb R}{{\mathbb R}} \def{\mathbb R}{{\mathbb R}} \def{\mathbb N}{{\mathbb N}} \def{\mathbb M}{{\mathbb M}}
\def{\mathbb R}{{\mathbb R}} \def{\mathbb I}{{\mathbb I}} \def{\mathbb U}{{\mathbb U}} \def{{\cal T}_{\mbox{\tiny ad}}}{{{\cal T}_{\mbox{\tiny ad}}}} \def\texte#1{\quad\mbox{#1}\quad} \def\Proba#1{{\mathbb P}\left\{ #1 \right\}} \def\Probax#1#2{{{\mathbb P}}_{#1}\left\{ #2 \right\}} \def\ProbaU#1#2{{{\mathbb P}}^{#1} \left\{ #2 \right\}} \def\ProbaxU#1#2#3{{{\mathbb P}}^{#1}_{#2} \left\{ #3 \right\}} \def\valmoy#1{{\mathbb E}\left[ #1 \right]} \def\valmoyDebut#1{{\mathbb E} [ #1 } \def\valmoyFin#1{ #1 ]} \def\valmoyp#1#2{{\mathbb E}_{#1}\left[ #2 \right]} \def\valmoypDebut#1#2{{\mathbb E}_{#1} \left[ #2 \right.} \def\valmoypFin#1{ \left. #1 \right]} \def\valmoypU#1#2#3{{\mathbb E}_{#1}^{#2}\left[ #3 \right]} \def\norminf#1{ {\Vert #1 \Vert}_{\infty}} \def\norm#1{ {\Vert #1 \Vert}} \def${\text{\bf H}}_1${${\text{\bf H}}_1$} \def${\text{\bf H}}_2${${\text{\bf H}}_2$} \def${\text{\bf H}}_3${${\text{\bf H}}_3$} \def\psca#1{\left< #1 \right>} \def\sigma\mbox{-}\lim{\sigma\mbox{-}\lim} \def\seq#1{\{{#1}_n\}_{n\in{\mathbb N}}} \def{\mathbb X}{{\mathbb X}}
\newenvironment{myproof}{{
\small{\it Proof:}}}{
$\Box$\normalsize \\
}
\setenumerate{labelindent=\parindent,label=\emph{($\mbox{C}_{\arabic*}$)},ref=($\mbox{C}_{\arabic*}$)}
\section{Introduction}
Many extensions of the well known Banach contraction principle \cite{banach} have been proposed in nonlinear analysis literature. Among them fixed point theorems for Meir-Keeler contraction have been extensively studied \cite{meir-keeler,kirk-2,suzuki-2006} and a final (in some sense) generalization defined as \emph{asymptotic contraction of the final type} (ACF, for short) has been stated by T.Suzuki \cite[Theorem 5]{suzuki2007}. Our aim in this paper is to extend the results of T.Suzuki to more general cases with regards to the mappings. More precisely, we want to be able to use the same framework for proving fixed point theorems for alternating mappings \S\ref{alternate} or for cyclic mappings \S\ref{cycling}. For that purpose we propose the definition of $p$-ASF-$1$ and $p$-ASF-$2$ sequences which are defined without references to a mapping and prove some Cauchy properties of such sequences in Theorem~\ref{thmpasf}. In \S\ref{defacf}, we recall the definition of ACF mapping and relate ACF mapping to $p$-ASF mappings. When the $p$-ASF sequences are generated using $\{T^nx\}$ we show that the two definitions coincide (Theorem~\ref{asf-acf}). We give an application to cyclic mappings in \S\ref{cycling} by providing a fixed point theorem which extends \cite[Theorem 2]{suzuki2007} to continuous $p$-ASF mappings. In \S\ref{alternate} we give an application to alternating mapping through Theorem~\ref{thm:ptfixe} which extends the results of~\cite{zhang2007}.
\section{ACF sequences}
In \cite{suzuki2007} T.Suzuki introduces the definition of an {\em asymptotic
contraction of the final type} (ACF, for short) and proves that if a mapping $T$ is ACF then the sequence $\seq{x}$ defined by $x_n\stackrel{\mbox{\tiny def}}{=} T^n x$ is a Cauchy sequence for all $x\in {\mathbb X}$. Since our aim is to extend T.Suzuki results when sequences $\seq{x}$ are generated by more general processes, we introduce a new definition that we call ASF, which stands for {\em asymptotic sequences of the
final type}. The definition characterizes two sequences and not a mapping. The link between the two definitions is the following. Suppose that the mapping $T$ is ACF and for $x$, $y \in {\mathbb X}$ define two sequences $\seq{x}$, $\seq{y}$ by $x_n\stackrel{\mbox{\tiny def}}{=} T^n x$ and $y_n\stackrel{\mbox{\tiny def}}{=} T^n y$. If for all $n\in {\mathbb N}$ we have $x_n \not=y_n$ then the two sequences are ASF. Properties of ASF sequences are given in Lemma~\ref{lem:ASF} and a proof is given but note that the proof is mostly a simple rephrase of \cite[Lemma 1 and 2]{suzuki2007}. We first start by the ASF definition.
In the sequel $({\mathbb X},d)$ is a complete metric space and $p$ is a given function from ${\mathbb X}\times {\mathbb X}$ into $[0,\infty)$.
\begin{defn} \label{def:acf} We say that two sequences $\seq{x}$, $\seq{y}$ with
$x_n,y_n \in {\mathbb X}$ are $p$-ASF-$1$ if the following are satisfied:
\begin{enumerate}
\item \label{it:acfun} For each $\epsilon >0$ there exists $\delta >0$ such that
if for $i \in {\mathbb N}$ we have $p(x_i,y_i) < \delta$ then
$\limsup_{n \to \infty} p(x_{n},y_{n}) \le \epsilon$\,;
\item \label{it:acfdeux} For each $\epsilon > 0$, there exists $\delta > 0$
such that for $i$,$j \in {\mathbb N}$ with $\epsilon < p(x_i, y_i) < \epsilon + \delta$,
there exists $\nu \in {\mathbb N}$ such that $p(x_{\nu+i},y_{\nu+i}) \le \epsilon$\,;
\item \label{it:acftrois} For each given $(x_i,y_i)$ such that $p(x_i,y_i)\ne 0$
there exists $\nu \in {\mathbb N}$ such that $$p(x_{\nu+i}, y_{\nu+i}) < p(x_i,y_i)\,.$$
\end{enumerate} \end{defn}
\begin{lem}\label{lem:ASF} Let $\seq{x}$, $\seq{y}$ be two $p$-ASF-1 sequences then $\lim_{n\to \infty} p(x_n,y_n)=0$. \end{lem}
\begin{myproof}
We follow \cite[Lemma 2]{suzuki2007}.
If we suppose that there exists $i\in{\mathbb N}$ such that $p(x_i,y_i)=0$,
we conclude directly using \ref{it:acfun} that $\lim_{n\to \infty} p(x_n,y_n)=0$.
Thus we assume now that $p(x_n,y_n)\not=0$ for all $n\in {\mathbb N}$.
We first prove that if $\seq{x}$, $\seq{y}$ satisfy \ref{it:acfdeux} and
\ref{it:acftrois} then $\liminf_{n \to \infty} p(x_n,y_n)=0$.
Using the fact that $p$ is nonnegative and repeatedly using Property~\ref{it:acftrois}
it is possible to build an extracted decreasing sub-sequence
$p(x_{\sigma(n)},y_{\sigma(n)})$ such that
$0 \le p(x_{\sigma(n)},y_{\sigma(n)}) \le p(x_0,y_0)$ which implies that
$\liminf_{n \to \infty} p(x_n,y_n)=\alpha$ exists and is finite.
Suppose that $\alpha >0$. We first show that we must have $\alpha <
p(x_n,y_n)$ for all $n \in {\mathbb N}$. Indeed suppose that there exists $n_0$ such
that $p(x_{n_0},y_{n_0}) \le \alpha$ then repeatedly using
\ref{it:acftrois} we
can build an extracted decreasing sequence $p(x_{\sigma(n)},y_{\sigma(n)})$ such
that $p(x_{\sigma(n)},y_{\sigma(n)}) < p(x_{n_0},y_{n_0}) \le \alpha$. This
decreasing sequence will converge to a cluster point of $p(x_{n},y_{n})$
strictly smaller than $\alpha$ which is contradictory with the definition of
$\alpha$. Thus we have $\alpha < p(x_n,y_n)$ for all $n \in {\mathbb N}$ and $\alpha >0$.
We then consider $\delta(\alpha)$ given by
\ref{it:acfdeux} for
$\epsilon=\alpha$. By definition of $\alpha$ we can find $(x_i,y_i)$ such that
$\alpha < p(x_i, y_i) < \alpha + \delta(\alpha)$ and by
\ref{it:acfdeux} we will obtain $\nu \in {\mathbb N}$
such that $p(x_{\nu+i},y_{\nu+i}) \le \alpha$ which contradicts $\alpha < p(x_n,y_n)$
for all $n \in {\mathbb N}$. Thus we conclude that $\alpha=0$.
We prove now that $\liminf_{n \to \infty} p(x_n,y_n)=0$ and
\ref{it:acfun} imply that $\limsup_{n \to \infty} p(x_n,y_n)=0$.
For $\epsilon > 0$ given, we consider $\delta$ given by
\ref{it:acfun}. Since $\liminf_{n \to \infty} p(x_n,y_n)=0$ then we can
find $i\in {\mathbb N}$ such that $ p(x_i, y_i) < \delta$. Thus by \ref{it:acfun} we have
$\limsup_{n \to \infty} p(x_{n+i},y_{n+i}) \le \epsilon$ and thus
successively $\limsup_{n \to \infty} p(x_{n},y_{n}) \le
\epsilon$ and $\limsup_{n \to \infty} p(x_{n},y_{n}) =0$ and the result follows. \end{myproof}
\begin{defn} \label{def:acfd} We say that a sequences $\seq{x}$, with
$x_n \in {\mathbb X}$ is $p$-ASF-$2$ if we have the following property:
\begin{enumerate}[start=4]
\item \label{it:acfddeux} For each $\epsilon > 0$, there exist $\delta > 0$ and
$\nu \in {\mathbb N}$ such that
if for $i$,$j \in {\mathbb N}$ we have $\epsilon < p(x_i, x_j) < \epsilon + \delta$, then
$p(x_{\nu+i},x_{\nu+j}) \le \epsilon$\,.
\end{enumerate} \end{defn}
Let $q$ be a given function from ${\mathbb X}\times {\mathbb X}$ into $[0,\infty)$ and $p=G\circ q$ where the mapping $G$ is a nondecreasing right continuous function such that $G(t)>0$ for $t>0$. We first show here that when a sequence is $(G\circ q)$-ASF-$2$ then it is also a $q$-ASF-$2$ sequence if \ref{it:acftrois-g} is satisfied by $p$. Note that Property~\ref{it:acfddeux} (resp. \ref{it:acftrois-g}) is a kind of uniform extension of \ref{it:acfdeux} (resp. \ref{it:acftrois}) when only one sequence is involved.
\begin{lem}(\cite[In Theorem 6]{suzuki-2007})\label{lem:ASF-D} Let $\seq{x}$ be a $p$-ASF-2 sequence and suppose that $p=G\circ q$ where $G$ is a nondecreasing right continuous function such that $G(t)>0$ for $t>0$. Suppose that we have \begin{enumerate}[start=5] \item \label{it:acftrois-g} for each given $(x_i,x_j)$ such that $p(x_i,x_j)\ne 0$
there exists $\nu \in {\mathbb N}$ such that $$p(x_{\nu+i}, x_{\nu+j}) < p(x_i,x_j)\,,$$ \end{enumerate} then $\seq{x}$ is a $q$-ASF-2 sequence. \end{lem}
\begin{myproof} The proof is contained in \cite[Theorem 6]{suzuki-2007}.
Fix $\eta >0$ and consider $\epsilon = G(\eta)$.
Since $G(t) > 0$ for $t >0$ we have $\epsilon >0$. Then we can
use \ref{it:acfddeux} to obtain $\delta >0$ and $\nu \in {\mathbb N}$ such that
$\epsilon < p(x_i, x_j) < \epsilon + \delta$, for some $i$, $j\in {\mathbb N}$
implies $p(x_{\nu+i},x_{\nu+j}) \le \epsilon$\,.
Since $G$ is nondecreasing right continuous we can find $\beta$ such that
$G([\eta,\eta+\beta]) \subset [\epsilon,\epsilon+\delta)$.
Thus suppose that $\eta < q(x_i,x_j) < \eta+\beta$, we then have
$\epsilon \le G(q(x_i,x_j)) < \epsilon+\delta$. Since $G$ is
nondecreasing it can be constant and equal to $\epsilon$ on a non
empty interval $[\eta,\eta+\overline{\beta}) \subset \eta+\beta$
in the contrary we will have $\epsilon < G(\eta+\gamma)$ for
$\gamma \in(0,\beta)$. If we are in the second case then
$\epsilon < G(q(x_i,x_j)) < \epsilon+\delta$ and using \ref{it:acfddeux} we obtain $G(q(x_{i+\nu},x_{j+\nu})) \le \epsilon < G(\eta+\gamma)$ we thus have $q(x_{i+\nu},x_{j+\nu}) < \eta+\gamma$ for all $\gamma \in(0,\beta)$ and consequently $q(x_{i+\nu},x_{j+\nu})\le \eta$. In the first case we have $G(q(x_i,x_j))=\epsilon$ for $\eta < q(x_i,x_j) < \eta+\overline{\beta}$. Using \ref{it:acftrois-g} we can find $\nu\in {\mathbb N}$ such that \begin{equation}
G(q(x_{i+\nu},x_{j+\nu})) < G(q(x_i,x_j))=\epsilon = G(\eta) \end{equation} and thus $q(x_{i+\nu},x_{j+\nu}) \le \eta$. We thus have proved that Property~\ref{it:acfddeux} is satisfied by $q$. \end{myproof}
We prove now that $p$-ASF-$2$ sequences mixed with convergence properties of the sequence $p(x_n,x_{n+1})$ gives $p$-Cauchy properties. More precisely we have the following lemma.
\defr{r}
\begin{lem}\label{lem:ASFD} Let $\seq{x}$, be a $p$-ASF-2 sequence
and suppose that $p$ is such that $p(x,y) \le p(x,z)+r(z,y)$ and
$p(x,y) \le r(x,z)+ p(z,y)$ for all
$x$, $y$, $z\in {\mathbb X}$ where the mapping $r:{\mathbb X}\times {\mathbb X} \to [0,\infty)$
satisfies the triangle inequality $r(x,y) \le r(x,z)+r(z,y)$ for all
$x$, $y$, $z\in {\mathbb X}$. If the sequence $\seq{x}$ is such that
$\lim_{n \to \infty} r(x_{n},x_{n+1})=0$ and $\lim_{n \to \infty} p(x_{n},x_{n+1})=0$
then we have $\lim_{n\to \infty} \sup_{m>n} p(x_n,x_m) = 0$. \end{lem}
\begin{myproof}
We follow \cite[Lemma 2]{suzuki2007} where a similar proof is given when $r=p$.
Let $\epsilon >0 $ be fixed
and consider $\delta$ and $\nu$ given by \ref{it:acfddeux}.
There exists $N\in {\mathbb N}$ such that
$r(x_n,x_{n+1}) < \delta/\nu$ and $p(x_n,x_{n+1}) < \delta/\nu$ for all $n \ge N$. We first have for $k\le \nu$:
\begin{align}
r(x_{n},x_{n+k}) & \le \sum_{i=0}^{k-1} r(x_{n+i},x_{n+i+1}) < k \frac{\delta}{\nu} \le \delta
\label{recfornu} \\ \intertext{and}
p(x_{n},x_{n+k}) & \le p(x_{n},x_{n+1}) + \sum_{i=1}^{k-1} r(x_{n+i},x_{n+i+1}) < k \frac{\delta}{\nu} \le \delta \label{recfornd}
\end{align}
We suppose that $p(x_{n},x_{n+\alpha}) < \epsilon+\delta$ is satisfied for $\alpha \in[1,k]$ and
we want to prove that the same inequalities are satisfied for $\alpha \in[1,k+1]$.
Using \eqref{recfornd} we may assume that $k\ge \nu$. Using the mixed
triangle inequality satisfied by $p$ we have the two separate
inequalities:
\begin{align}
p(x_{n},x_{n+k+1}) & \le p(x_{n},x_{n+k+1-\nu}) + \sum_{i=1-\nu}^{0} r(x_{n+k+i},x_{n+k+i+1})\nonumber \\
& < p(x_{n},x_{n+k+1-\nu}) + \delta \label{inequn} \\ \intertext{and} p(x_{n},x_{n+k+1}) & \le r(x_{n},x_{n+\nu}) + p(x_{n+\nu},x_{n+k+1-\nu+\nu}) \label{ineqdeux} \end{align}
By hypothesis we have $p(x_n,x_{n+k+1-\nu}) < \epsilon+\delta$. If $p(x_n,x_{n+k+1-\nu})
\le \epsilon$ then using \eqref{inequn} we obtain $p(x_{n},x_{n+k+1}) < \epsilon+\delta$
else we can use \ref{it:acfddeux} to first get
$p(x_{n+\nu},x_{n+k+1-\nu+\nu}) \le \epsilon$ and using \eqref{recfornu} and \eqref{ineqdeux}
we obtain $p(x_{n},x_{n+k+1}) < \epsilon+\delta$. \end{myproof}
In \cite{suzuki-2001}, T. Suzuki introduces the definition of a $\tau$-distance. We just recall here two properties which are satisfied by $\tau$-distance: if a function $p$ from ${\mathbb X}\times {\mathbb X}$ into ${\mathbb R}^+$ is a $\tau$-distance it satisfies $p(x,y)\le p(x,z)+p(z,y)$ for all $x$, $y$, $z \in {\mathbb X}$ and if a sequence $\seq{x}$ in ${\mathbb X}$ satisfies $\lim_{n \to \infty} \sup_{m >n} p(x_{n},x_{m})=0$ then $\seq{x}$ is a Cauchy sequence. We thus have the following theorem.
\begin{thm}\label{thmpasf} Let $\seq{x}$ be a $p$-ASF-2 sequence in ${\mathbb X}$ such that $\seq{x}$ and $\seq{y}$ are $p$-ASF-1 for $y_{n}=x_{n+1}$ for all $n\in {\mathbb N}$. If one of the following assumptions holds true \begin{enumerate}[labelindent=\parindent,label=(\roman*),ref=\emph{(\roman*)}] \item \label{thmpasf:un} $p=q$ and $q$ is a $\tau$-distance\,; \item \label{thmpasf:deux} $p=G(q)$ where $q$ is a $\tau$-distance and where $G$ is a nondecreasing right continuous function such that $G(t)>0$ for $t >0$ and \emph{\ref{it:acftrois-g}} is satisfied by the sequence $\seq{x}$ (for the mapping $p=G(q)$)\,; \item \label{thmpasf:trois} $p$ is a $\tau$-distance such that $p(x,y) \le p(x,z)+q(z,y)$
and $p(x,y) \le q(x,z)+p(z,y)$ for all
$x$, $y$, $z\in {\mathbb X}$ where the mapping $q:{\mathbb X}\times {\mathbb X} \to [0,\infty)$
satisfies the triangle inequality $q(x,y) \le q(x,z)+q(z,y)$ for all
$x$, $y$, $z\in {\mathbb X}$ and $\lim_{n\to \infty} q(x_n,x_{n+1})=0$\,; \end{enumerate} then, $\seq{x}$ is a Cauchy sequence. \end{thm}
\begin{myproof} First note that, in the three cases, using Lemma~\ref{lem:ASF} we have that $$\lim_{n\to \infty} p(x_n,x_{n+1})=0\,.$$
\ref{thmpasf:un} We consider the case $p=q$. Since $\lim_{n\to \infty} p(x_n,x_{n+1})=0$ We can use Lemma~\ref{lem:ASFD} (with $r=p$) to obtain $\lim_{n\to \infty} \sup_{m>n} p(x_n, x_m) = 0$ and since $p$ is a $\tau$-distance we obtain the fact that $\seq{x}$ is a Cauchy sequence \cite[Lemma 1]{suzuki-2001}.
\ref{thmpasf:deux} Suppose now that $p=G(q)$, we have $\lim_{n\to \infty} G(q(x_n,x_{n+1}))=0$. This is only possible if $G(0)=0$ and we thus also obtain that $\lim_{n\to \infty} q(x_n,x_{n+1})=0$. Using Lemma~\ref{lem:ASF-D} we obtain that $\seq{x}$ is $q$-ASF-2 and we conclude as in the part \ref{thmpasf:un} using now the $\tau$-distance $q$.
\ref{thmpasf:trois} Here we can use Lemma~\ref{lem:ASFD} to obtain $\lim_{n\to \infty} \sup_{m>n} p(x_n, x_m) = 0$ and using the fact that $p$ is a $\tau$-distance the conclusion follows the lines of the case \ref{thmpasf:un}. \end{myproof}
\begin{rem} Note that we have proved during the proof of Theorem~\ref{thmpasf} that if we have two sequences $\seq{x}$ and $\seq{y}$ which are $p$-ASF-1 with $p=G\circ q$ then $G(0)=0$. \end{rem}
\section{Links with ACF sequences} \label{defacf} We first recall here the definition of an ACF mapping. Then we give a definition of a $T$-ASF mapping by defining properties which are to be satisfied by the sequences $\{T^nx\}$ for $x \in {\mathbb X}$. We prove in Theorem~\ref{asf-acf} that the two definitions are equivalent.
\begin{defn}\cite[Definition 1]{suzuki2007} Let $({\mathbb X}, d)$ be a metric space.
Then a mapping $T$ on ${\mathbb X}$ is said to be an asymptotic
contraction of the final type (ACF, for short) if the following hold: \begin{enumerate}[label=\emph{($\mbox{D}_{\arabic*}$)}] \item $\lim_{\delta \to 0^+} \sup \left\{ \limsup_{n\to \infty} d(T^n x,T^n y): d(x, y) <\delta \right\} = 0$. \item For each $\epsilon > 0$, there exists $\delta > 0$ such that for $x$, $y \in {\mathbb X}$
with $\epsilon < d(x, y) < \epsilon + \delta$, there exists $\nu \in N$ such that
$d(T^\nu x,T^\nu y) \le \epsilon$. \item For $x$, $y \in {\mathbb X}$ with $x\not=y$, there exists $\nu \in N$ such that
$d(T^\nu x,T^\nu y)<d(x,y)$. \item For $x \in {\mathbb X}$ and $\epsilon >0$, there exist $\delta >0$ and $\nu \in N$ such that \begin{equation}
\epsilon <d(T^i x,T^j x) < \epsilon +\delta \;\mbox{implies}\;
d(T^\nu \circ T^i x,T^\nu\circ T^j x )\le \epsilon\, \end{equation} for all $i$, $j \in N$. \end{enumerate} \end{defn}
\begin{thm}\label{asf-acf} Let $({\mathbb X}, d)$ be a metric space. A mapping $T$ on ${\mathbb X}$ is
said to be $p$-ASF if for all $x$, $y \in {\mathbb X}$ the sequences
$\{T^n x\}$ and $\{T^n y\}$ are $p$-ASF-1 and $\{T^n x\}$ is $p$-ASF-2.
Then, $T$ is an $ACF$ mapping is equivalent to $T$ is a $d$-ASF mapping. \end{thm}
\begin{myproof} Suppose that the mapping $T$ is $ACF$. For each $x$, $y\in {\mathbb X}$ it is very easy to check and left to the reader that $\{T^n x\}$ and $\{T^n y\}$ are $d$-ASF-1 and $\{T^n x\}$ is $d$-ASF-2. Thus, $T$ is $d$-ASF.
If $T$ is $d$-ASF, using Lemma~\ref{lem:ASF} we obtain $\lim_{n\to \infty} d(T^n x,T^n y) = 0$. If we consider the special case $y=Tx$ and the sequence $x_n = T^nx$ we obtain using Theorem~\ref{thmpasf} that $\{T^nx\}$ is a Cauchy sequence. Then using \cite[Theorem 6]{suzuki2007}\footnote{
We first recall from \cite[Theorem 6]{suzuki2007} that for a mapping $T$ on a metric space $({\mathbb X}, d)$ the following are equivalent: \begin{enumerate}[labelindent=\parindent,label=(\roman*),ref=\emph{(\roman*)}] \item $T$ is an ACF. \item $\lim_{n\to \infty} d(T^n x,T^n y) = 0$ holds true and $\{T^nx\}$ is a Cauchy sequence
for all $x$, $y \in {\mathbb X}$. \end{enumerate} } we obtain that the mapping $T$ is ACF. \end{myproof}
Existence and uniqueness of fixed points of $p$-ASF mappings is now obtained. Note that, in the special case where the mapping $p$ is equal to $d$ (i.e when we use the $\tau$-distance $p=d$ in $(i)$) the next theorem gives same results as \cite[Theorem 5]{suzuki2007}.
\begin{thm}\label{fixedpt} Let $({\mathbb X},d)$ be a complete metric space, $T$ be a $p$-ASF mapping which is such that $T^l$ is continuous for some $l\in {\mathbb N}$ ($l>0$). We suppose that the function $q$ is a $\tau$-distance and one of the following holds true for the mapping $p$: \begin{enumerate}[labelindent=\parindent,label=(\roman*),ref=\emph{(\roman*)}] \item $p=q$\,. \item $p=G(q)$ where $G$ is a nondecreasing right continuous function such that $G(t)>0$ for $t >0$ and \emph{\ref{it:acftrois-g}} is satisfied by the sequence $\seq{x}$ (for the mapping $p=G(q)$). \end{enumerate} then, there exists a fixed point $z\in{\mathbb X}$ of $T$. Moreover, if for every sequences $\seq{x}$ and $\seq{y}$ $\lim_{n\to \infty} p(x_n,y_n)=0$ implies that $\lim_{n\to \infty} d(x_n,y_n)=0$ then the fixed point is unique and $\lim_{n\to \infty} T^nx=z$ holds true for every $x\in {\mathbb X}$ \end{thm}
\begin{myproof} For every $x\in {\mathbb X}$ using Theorem~\ref{thmpasf} we know that $\{T^nx\}$ is a Cauchy sequence. By Lemma~\ref{lem:ASF} we know that $\lim_{n\to \infty} p(T^nx,T^ny)=0$. We then have all the ingredients of \cite[Theorem 4 and Lemma 3]{suzuki2007} to conclude the proof. \end{myproof}
\section{An application to $ACF$ cyclic mappings} \label{cycling}
We suppose here that ${\mathbb X}$ is a uniformly convex Banach space and thus $d(x,y)\stackrel{\mbox{\tiny def}}{=} \norm{x-y}$. We consider $A$ and $B$ two nonempty subsets of ${\mathbb X}$, $A$ being convex, and a cyclic mapping $T: A \cup B \to A \cup B$. We recall that $T$ is a cyclic mapping if $T(A) \subset B$ and $T(B) \subset A$. We define a mapping $p: {\mathbb X}\times {\mathbb X} \to {\mathbb R}^+$
by $p(x,y) \stackrel{\mbox{\tiny def}}{=} d(x,y) -d(A,B)$ where $d(A,B) \stackrel{\mbox{\tiny def}}{=} \inf \{ d(x,y) \,|\, x\in A, y\in B\}$.
Then, using previous results we can give a short proof of a theorem which extends~\cite[Theorem 1]{suzuki-2008}.
\begin{thm}\label{thm:bcauchy} Suppose that the mapping $T$ is $p$-ASF, then the sequence $\{T^{2n}x\}$ for $x\in A$ is a $d$-Cauchy sequence. \end{thm}
\begin{myproof} For a given $x\in {\mathbb X}$, we consider the sequence $x_n=T^nx$.
Since $T$ is $p$-ASF we have by Lemma~\ref{lem:ASF} that $\lim_{n\to \infty}
p(x_n,x_{n+1})=0$. Using the definition of $p$ we immediately also have
$\lim_{n\to \infty} d(x_n,x_{n+1})=d(A,B)$. Using Lemma~\cite[Lemma
4]{suzuki-2008} we obtain that $\lim_{n \to \infty} d(x_{2n},x_{2n+2})=0$
(convexity of $A$ and uniformly convexity of ${\mathbb X}$ are used here). We now
consider the sequence $\{T^{2n} x\}$ taking values in $A$. We have
$\lim_{n\to \infty} p(x_{2n},x_{2n+2})=0$ and as it was already shown $\lim_{n
\to \infty} d(x_{2n},x_{2n+2})=0$. If the sequence $x_n=T^nx$ is $p$-ASF-2
then it is the same for the sequence $\{T^{2n} x\}$. The distance $d$ satisfy
the triangle inequality and it is straightforward to see that we have the two
mixed triangle inequality $p(x,y) \le p(x,z)+d(z,y)$ and $p(x,y) \le
d(x,y)+p(z,y)$ for all $x$, $y$, $z\in {\mathbb X}$. We can thus apply
Lemma~\ref{lem:ASFD} to the sequence $\{T^{2n} x\}$ with $x\in A$ to obtain
that it is a $p$-Cauchy sequence. It is now easy to see by contradiction that
a $p$-Cauchy sequence is a $d$-Cauchy sequence \cite[Proof of Theorem
2]{suzuki-2008}. The key argument being again the use of \cite[Lemma
4]{suzuki-2008} \end{myproof}
We extend now \cite[Theorem 2]{suzuki-2008} which was stated for continuous cyclic Meir-Keeler contractions to continuous $p$-ASF mappings.
\begin{thm}\label{thm:cyclic}Suppose in addition that $A$ is closed,
$T$ is $p$-ASF and $T^l$ is continuous for
some $l\in{\mathbb N}$ ($l>0$) then there exists a unique best
proximity point $z\in A$ (i.e
$d(z,Tz)=d(A,B)$). Moreover $\lim_{n\to \infty} T^{2n} x=z$ for each $x\in A$. \end{thm}
\begin{myproof} Using Theorem~\ref{thm:bcauchy}, the sequence $\{T^{2n}x\}$ for each $x\in A$ is a $d$-Cauchy sequence. Using Lemma~\ref{lem:ASF}, we have $\lim_{n \to \infty} p(T^{n}x,T^{n}y)=0$ for each $x$, $y \in A$, hence for $(x,Tx)$ it gives $\lim_{n \to \infty} p(T^{2n}x,T^{2n+1}x)=0$ and for $(Tx,y)$ it gives $\lim_{n \to \infty} p(T^{2n+1}x,T^{2n}y)=0$. Using again \cite[Lemma 4]{suzuki-2008} we obtain $\lim d(T^{2n}x,T^{2n}y)=0$ and we can use \cite[Theorem 4 and Lemma 3]{suzuki2007} to conclude the proof. \end{myproof}
\section{ASMK Sequences}
We introduce in this section the definition of ASMK sequences. It is an adaptation to sequences of the ACMK (Asymptotic contraction of Meir-Keeler type) definition used for mappings \cite{suzuki-2006}. It is proved in \cite[Theorem 3]{suzuki2007} that an ACMK mapping on a metric space is an ACF mapping. We will prove in this section similar results which relate ASMK sequences to ASF sequences. These results will be used in next section for studying sequences of alternating mappings.
\begin{defn} \label{Famk} We say that two sequences $\seq{x}$, $\seq{y}$ with
$x_n,y_n \in {\mathbb X}$ are $p$-ASMK-$1$ if there exists a sequence $\{\psi_n\}$
of functions from $[0,\infty)$ into itself satisfying $\psi_n(0)=0$\footnote{Note that this assumption can be removed when $F(0)>0$.}
for all $n \in {\mathbb N}$ and the following:
\begin{enumerate}[start=6]
\item \label{it:famkun} $\limsup_n \psi_n(\epsilon) < \epsilon \texte{for all} \epsilon > 0$.
\item \label{it:famkdeux} For each $\epsilon >0$, there exists $\delta >0$
such that for each $t \in [\epsilon, \epsilon + \delta]$ there exists
$\nu \in {\mathbb N}$ such that $\psi_\nu (t) < \epsilon$.
\item \label{it:famktrois} $F\bp{p(x_{n+i},y_{n+i})} \le \psi_n\Bp{F\bp{p(x_i,y_i)}}$ for
all $n$, $i \in {\mathbb N}$. $F$ is a given right continuous nondecreasing mapping
such that $F(t) >0$ for $t\not=0$.
\end{enumerate} \end{defn}
\begin{lem}\label{lem:asmk-asf} Suppose that the two sequences $\seq{x}$, $\seq{y}$ are $p$-ASMK-$1$ then they are $p$-ASF-1. \end{lem}
\begin{myproof}
\ref{it:acfun}: For all $n$, $i \in {\mathbb N}$ we have by~\ref{it:famktrois}
and~\ref{it:famkun} when $F\bp{p(x_i,y_i}\ne 0$ that
\begin{align}
F\bp{p(x_{n+i},y_{n+i})} & \le \psi_n\Bp{F\bp{p(x_i,y_i)}} \nonumber \\
& \le \limsup_{n\to \infty} \psi_n\Bp{F\bp{p(x_i,y_i)}} \nonumber \\
& < F\bp{p(x_i,y_i)}\,. \nonumber
\end{align}
Since $F$ is nondecreasing and the inequality is strict we obtain for all
$n \in {\mathbb N}$:
\begin{align*}
p(x_{n+i},y_{n+i}) < p(x_i,y_i) \,,
\end{align*}
and thus
\begin{align*}
\limsup_{n\to \infty} p(x_{n+i},y_{n+i}) \le p(x_i,y_i) \,.
\end{align*}
Then \ref{it:acfun} follows easily when $F\bp{p(x_i,y_i}\ne 0$. When $F\bp{p(x_i,y_i}= 0$,
we have by~\ref{it:famktrois} $F\bp{p(x_{n+i},y_{n+i})}\le 0$ for all $n\in {\mathbb N}$.
Since $F$ is a right continuous mapping such that $F(t) >0$ we must have $F(0)\ge 0$.
Thus we have that $F\bp{p(x_{n+i},y_{n+i})}= 0$ for all $n\in {\mathbb N}$ and thus
$p(x_{n+i},y_{n+i})=0$ for all $n\in {\mathbb N}$ and the same conclusion holds.
\ref{it:acfdeux}:
for $\epsilon > 0$ we know that $F(\epsilon) >0$ and we can use
\ref{it:famkdeux} to find $\delta >0$ such that for each
$t \in [F(\epsilon), F(\epsilon) + \delta]$ we can find $\nu \in {\mathbb N}$ such that $\psi_\nu
(t) < F(\epsilon)$. Since
$F$ is right continuous and nondecreasing we can find $\delta'$ such that
$F([\epsilon,\epsilon+\delta']) \subset [F(\epsilon), F(\epsilon) + \delta]$.
Thus, taking $i\in {\mathbb N}$ such that $\epsilon < p(x_i,y_i) < \epsilon + \delta'$,
we can find $\nu$ such that $\psi_{\nu}(F(p(x_i,y_i))) < F(\epsilon)$. And we
conclude using \ref{it:famktrois} that:
\begin{equation}
F\bp{p(x_{\nu+i},y_{\nu+i})} \le \psi_{\nu} \Bp{F\bp{p(x_i,y_i)}} < F(\epsilon) \le F\bp{p(x_i,y_i)}\,.
\end{equation}
Thus we have $F\bp{p(x_{\nu+i},y_{\nu+i})} < F\bp{p(x_i,y_i)}$ and since $F$ is nondecreasing
and the inequality is strict we obtain \ref{it:acfdeux}.
\ref{it:acftrois}: Let $i$ be given such that $p(x_i,y_i)\not=0$ and start as in the previous
paragraph using $\epsilon =p(x_i,y_i)$. We can find $\nu \in {\mathbb N}$ such that
$\psi_{\nu}\Bp{F\bp{p(x_i,y_i)}} < F(\epsilon)$ which combined with \ref{it:famktrois} gives:
\begin{equation}
F\bp{p(x_{\nu+i},y_{\nu+i})} \le \psi_{\nu} \Bp{F\bp{p(x_i,y_i)}} < F(\epsilon) = F\bp{p(x_i,y_i)}\,.
\end{equation}
Since $F$ is nondecreasing and the inequality is strict the result follows. \end{myproof}
\begin{defn} We say that two sequences $\seq{x}$, $\seq{y}$ with
$x_n,y_n \in {\mathbb X}$ are $p$-ASMK-$2$ when \emph{\ref{it:famktrois}} is replaced by
\begin{enumerate}[start=9]
\item \label{it:famktrois-p} $F\bp{p(x_{n+i},y_{n+j})} \le \psi_n \Bp{F\bp{p(x_i,y_j)}}$
for all $n$, $p$, $i$,$j \in {\mathbb N}$.
\end{enumerate} \end{defn}
\begin{cor} If two sequences $\seq{x}$, $\seq{y}$ with $y_{n}=x_{n+1}$ are $p$-ASMK-$2$ then they are $p$-ASF-1 and the sequence $\seq{x}$ is $p$-ASF-2. Moreover, assumption \emph{\ref{it:acftrois-g}} holds true for $p$. \end{cor}
\begin{myproof} It is obvious to see that if two sequences $\seq{x}$, $\seq{y}$ are
$p$-ASMK-$2$ then they are $p$-ASMK-$1$. Thus by Lemma~\ref{lem:asmk-asf} they are
$p$-ASF-1. Proving that \ref{it:acfddeux} holds true is similar to the proof that \ref{it:acfdeux}
holds true in Lemma~\ref{lem:asmk-asf} and proving that \ref{it:acftrois-g} holds true
follows the same steps as the proof that \ref{it:acftrois} holds true in Lemma~\ref{lem:asmk-asf}. \end{myproof}
\section{A sequence of alternating mappings} \label{alternate} In this section $p$ is a given function from ${\mathbb X}\times {\mathbb X}$ into $[0,\infty)$ such that such $p(x,y) \le p(x,z)+p(z,y)$ for all $x$, $y$, $z\in {\mathbb X}$ and $p(x,y)=p(y,x)$ for all $x$, $y \in {\mathbb X}$.
\begin{defn} We will say that the pair $(T,S)$ satisfy the $(F,\psi)$-contraction property if we can find two functions $F$ and $\psi$ such that: \begin{equation}
F \bp{ p(Tx,Sy)} \le \psi \Bp{F\bp{ M(x,y)}} \end{equation} where \begin{equation}
M(x,y) \stackrel{\mbox{\tiny def}}{=} \max \left\{p(x,y),p(Tx,x),p(Sy,y),
\frac{1}{2}\left\{p(Tx,y)+p(Sy,x) \right\} \right\} \end{equation} The function $F:{\mathbb R}^+ \to {\mathbb R}^+$ is a given right continuous nondecreasing mapping such that $F(t) >0$ for $t\not=0$. The function $\psi:{\mathbb R}^+ \to {\mathbb R}^+$ is a given nondecreasing upper semicontinuous function satisfying $\psi(t)< t$ for each $t>0$ and $\psi(0)= 0$. \end{defn}
We first start by a technical lemma.
\begin{lem}\label{lemme_dix} Let the pair of mappings $(T,S)$ be a $(F,\psi)$-contraction. Suppose that $x=S\alpha$ and $p(x,Tx)\not=0$ then we have: \begin{equation}
\label{ineqT}
F \bp{ p(x,Tx)} \le \psi\Bp{ F \bp{p(S\alpha,\alpha)}} \,. \end{equation} Suppose that $y=T\alpha$ and $p(y,Sy)\not=0$ then we have: \begin{equation}
\label{ineqS}
F \bp{p(Sy,y)} \le \psi \Bp{ F\bp{p(\alpha,T\alpha)}} \,. \end{equation} \end{lem}
\begin{myproof} We prove the first inequality \eqref{ineqT}. We suppose that $x=S\alpha$ then we have \begin{align}
F \bp{p(x,Tx)} &= F \bp{p(Tx,x)} = F \bp{p(Tx,S\alpha)} \le \psi \Bp{ F \bp{M(x,\alpha)}} \nonumber \end{align} and we have: \begin{align}
M(x,\alpha) &= \max \left\{p(x,\alpha),p(Tx,x),p(S\alpha,\alpha),
\frac{1}{2}\left\{p(Tx,\alpha)+p(S\alpha,x) \right\} \right\} \nonumber \\
&= \max \left\{p(x,\alpha),p(Tx,x),\frac{1}{2} p(Tx,\alpha) \right\} \nonumber \\
&= \max \left\{ p(x,\alpha),p(Tx,x) \right\}\,. \end{align} We show now that the maximum cannot be achieved by $p(Tx,x)$. Indeed, suppose that $M(x,\alpha)=p(Tx,x)$ then we would have \begin{align}
F \bp{p(x,Tx)} & \le \psi \Bp{ F \bp{p(x,Tx)}} \nonumber \end{align} which is not possible since $p(x,Tx)\ne 0$ and there does not exist $x > 0$ such that $F(x) \le \psi(F(x))$ (since for $x>0$ we have that $F(x) \le \psi(F(x)) < F(x)$). The proof for the second inequality is very similar and thus omited. \end{myproof}
We now introduce the alternating sequence of mappings $\seq{\Gamma}$ defined by: \begin{equation}
\Gamma_n \stackrel{\mbox{\tiny def}}{=} \begin{cases} T, & \mbox{if } n\mbox{ is even} \\ S, & \mbox{if } n\mbox{ is odd}\,. \end{cases} \end{equation} Then, we consider the two sequences $\seq{x}$ and $\seq{y}$ defined by \begin{equation} x_{n+1} =\Gamma_{n} x_n \quad\mbox{and} \quad y_{n+1}= \Gamma_{n+1} y_n\,. \label{defxy} \end{equation} It is very easy to check that when the two sequences are initiated with $(x_0,y_0)=(Sx,x)$ for a given $x\in {\mathbb X}$ they are related by $y_{n+1}=x_n$ and that only the two following cases can occur: \begin{equation}
(x_{n+1},y_{n+1})= \begin{cases} (Sx_n,x_n)& \text{with }\, x_n=Tx_{n-1}
\text{ and }\, y_n = x_{n-1} \\
(T x_n,x_n) & \text{with }\, x_n= S x_{n-1} \text{ and }\, y_n = x_{n-1} \end{cases} \end{equation} If we are in the first (resp. the second) case we use \eqref{ineqS} (resp. \eqref{ineqT}) to obtain the inequality \begin{equation}
F (p(x_{n+1},y_{n+1})) \le \psi (F(p(x_n,y_n)))\label{ineqfpu} \end{equation} We thus have the following easy lemma:
\begin{lem}\label{lem:tscontr}Let the pair of mappings $(T,S)$ be a $(F,\psi)$-contraction and $\seq{x}$ and $\seq{y}$ be two sequences defined by \eqref{defxy}. If the two sequences are initiated by $(x_0,y_0)=(Sx,x)$ we have $y_{n+1}=x_n$ and \begin{equation}
F \bp{p(x_{n+1},y_{n+1})} \le \psi \Bp{ F \bp{ p(x_n,y_n)}} \label{ineqfp}\,. \end{equation} If $x_{n+1}=Tx_n$ (resp. $x_{n+1}=Sx_n$) and $y_{k+1}= Sy_k$ (resp. $y_{k+1}= Ty_k$) we have: \begin{equation}
F \bp{p(x_{n+1},y_{k+1})} \le \psi \Bp{ F \bp{p(x_n,y_k)+\max(p(x_{n},x_{n+1}),p(x_{k},x_{k+1}))}}\,.
\label{ineqmelange} \end{equation} \end{lem}
\begin{myproof} Since inequation \eqref{ineqfp} was proved by \eqref{ineqfpu}, it just remains to prove inequality \eqref{ineqmelange}. Suppose that $x_{n+1}=Tx_n$ and $y_{k+1}= Sy_k$ then we have \begin{align}
M(x_n,y_k) &=\max \left\{p(x_n,y_k),p(Tx_n,x_n),p(Sy_k,y_k),
\frac{1}{2}\left\{p(Tx_n,y_k)+p(Sy_k,x_n) \right\} \right\} \nonumber \\
&= \max \left\{p(x_n,y_k),p(x_{n+1},x_n),p(y_{k+1},y_k),
\frac{1}{2}\left\{p(x_{n+1},y_k)+p(y_{k+1},x_n) \right\} \right\} \nonumber \\
&\le \max \left\{p(x_n,y_k),p(x_{n+1},x_n),p(y_{k+1},y_k),\right. \nonumber \\
&\hspace{1.5cm} \left. p(x_{n},y_k) + \max(p(x_{n+1},x_n),p(y_{k+1},y_k)) \right\} \nonumber \\
&\le p(x_{n},y_k)+ \max(p(x_{n+1},x_n),p(y_{k+1},y_k)) \end{align} We thus have \begin{align}
F \bp{p(x_{n+1},y_{k+1})} &\le \psi \Bp{F \bp{M(x_n,y_k)}} \nonumber \\
& \le \psi \Bp{F \bp{ p(x_{n},y_k)+\max(p(x_{n+1},x_n),p(y_{k+1},y_k))}} \,. \label{ineqcauchy} \end{align} If the opposite situation is $x_{n+1}=Sx_n$ and $y_{k+1}=Ty_k$ we obtain the same result by the same arguments. \end{myproof}
We make here a direct proof of the fact that the sequence $\seq{x}$ is a Cauchy sequence when $\lim_{n\to \infty} p(x_{n+1},x_n)=0$ is assumed. This last property will be derived from $p$-ASMK-$1$ properties as proved in Theorem~\ref{thm:ptfixe}.
\begin{lem} Let the pair of mappings $(T,S)$ be a $(F,\psi)$-contraction. Suppose that $$\lim_{n\to \infty} p(x_{n+1},x_n)=0\,,$$ then the sequence $\seq{x}$ given by~\eqref{defxy} is a Cauchy sequence. \label{cauchylem} \end{lem}
\begin{myproof} We follow here \cite{zhang2007} to prove the result by contradiction.
If the sequence is not a Cauchy sequence we can find two subsequences $\sigma(n)$ and $\rho(n)$ such that for all $n \in {\mathbb N}$ $p(x_{\sigma(n)}, x_{\rho(n)}) \ge 2\epsilon$ and $\sigma(n) < \rho(n)$. Since the sequence $\{ p(x_n,x_{n+1})\}_{n\in{\mathbb N}}$ converges to zero we can choose $N$ such that $p(x_n,x_{n+1}) < \epsilon$ for all $n \ge N$. Using the triangle inequality $$ p(x_{\sigma(n)},x_{\rho(n)+1}) \ge p(x_{\sigma(n)},x_{\rho(n)}) - p(x_{\rho(n)+1},x_{\rho(n)})\, $$ we obtain that $p(x_{\sigma(n)},x_{\rho(n)+1}) > \epsilon$ for large $n$. Thus, we can always change the subsequence ${\rho(n)}$ in such a way that the parity between $\sigma(n)$ and $\rho(n)$ is conform to the one we need for applying inequality \eqref{ineqcauchy} and such that for all $n \in {\mathbb N}$ $p(x_{\sigma(n)}, x_{\rho(n)}) > \epsilon$.
We now define $k(n)$ as follows: \begin{equation}
k(n) \stackrel{\mbox{\tiny def}}{=} \min \left\{
k > \sigma(n) \, \vert \, p(x_{\sigma(n)},x_{k}) > \epsilon \quad
\mbox{ with same parity as $\rho(n)$}
\right\} \end{equation} $k(n)$ is well defined and by construction $\sigma(n) < k(n)\le \rho(n)$. We now have that: \begin{align}
\epsilon < p(x_{\sigma(n)},x_{k(n)}) \le p(x_{\sigma(n)},x_{k(n)-2}) +
p(x_{k(n)-2},x_{k(n)}) \le \epsilon + p(x_{k(n)-2},x_{k(n)})\,. \end{align} the sequence $\{p(x_{k(n)-2},x_{k(n)})\}_{n\in{\mathbb N}}$ converges to zero since we have $$p(x_{k(n)-2},x_{k(n)})\le p(x_{k(n)-2},x_{k(n)-1}) + p(x_{k(n)-1},x_{k(n)}) $$ and thus $p(x_{\sigma(n)},x_{k(n)}) \to \epsilon^+$ when $n$ goes to infinity. We also obtain that $p(x_{\sigma(n)-1},x_{k(n)-1}) \to \epsilon$ when $n$ goes to infinity since: \begin{align}
|p(x_{\sigma(n)},x_{k(n)}) - p(x_{\sigma(n)-1},x_{k(n)-1})|
\le p(x_{k(n)},x_{k(n)-1}) + p(x_{\sigma(n)},x_{\sigma(n)-1})\,. \end{align} We now use inequality~\eqref{ineqcauchy} to obtain \begin{align}
F \bp{p(x_{\sigma(n)},x_{k(n)})} \le \psi \Bp{F\bp{p(x_{\sigma(n)-1},x_{k(n)-1})+\delta_n}} \end{align} where $\delta_n \stackrel{\mbox{\tiny def}}{=} \max \bp{p(x_{\sigma(n)-1},x_{\sigma(n)}),p(x_{k(n)-1},x_{k(n)})}$. When $n$ goes to infinity, using the facts that $F$ is right continuous and nondecreasing and $\psi \circ F$ is upper semicontinuous we obtain that $F(\epsilon) \le \psi \bp{F(\epsilon)}$ which is a contradiction. \end{myproof}
\begin{rem}\label{rem:zhang} The proof remains valid is we assume as in \cite{zhang2007} that the function $F$ is nondecreasing and continuous with $F(0)=0$ and $F(t)> 0$ for $t>0$ and that the function $\psi:{\mathbb R}^+ \to {\mathbb R}^+$ is assumed to be nondecreasing and right upper semicontinuous and satisfy $\psi(t)< t$ for each $t>0 $. The idea is to build the sequences choosing the parity so as to use~\eqref{ineqcauchy} in the reverse situation where \[F \bp{p(x_{\sigma(n)+1},x_{k(n)+1})} \le \psi \Bp{F\bp{p(x_{\sigma(n)},x_{k(n)})+\delta'_n}}, \] and where $\delta'_n \stackrel{\mbox{\tiny def}}{=} \max \bp{p(x_{\sigma(n)+1},x_{\sigma(n)}),p(x_{k(n)+1},x_{k(n)})}$. \end{rem}
\begin{thm}\label{thm:ptfixe} Consider two mappings $T: {\mathbb X} \to {\mathbb X} $ and $S: {\mathbb X} \to {\mathbb X}$ and suppose that the pair $(T,S)$ has the $(F,\psi)$-contraction property. Let the sequence of function $\seq{\psi}$ be defined by $\psi_n \stackrel{\mbox{\tiny def}}{=} \overset{n}{\overbrace{\psi\circ\psi\circ\cdots\psi}}$ and assume that \emph{\ref{it:famkun}} and \emph{\ref{it:famkdeux}} are satisfied, then the sequence $\seq{x}$ defined by \eqref{defxy} and initialized by $x_0=Sx$ is a Cauchy sequence. \end{thm}
\begin{myproof} The only point to prove is that assumption~\ref{it:famktrois} is satisfied. We consider the sequence $\seq{x}$ and the sequence $\seq{y}$ defined by \eqref{defxy} and initialized by $y_0=x$. Using the fact that $\psi$ is non-decreasing, we repeatedly use Equation~\eqref{ineqfp} in Lemma~\ref{lem:tscontr} to obtain assumption \ref{it:famktrois} and conclude that the two sequences $\seq{x}$ and $\seq{y}$ are $p$-ASMK-$1$ and then by Lemma~\ref{lem:asmk-asf} and \ref{lem:ASF} we obtain that $\limsup_{n \to \infty} p(x_n,x_{n+1})=0$. Using Lemma~\ref{cauchylem} we conclude that $\seq{x}$ is a Cauchy sequence. \end{myproof}
We make a link here with the result of \cite{zhang2007} where it is assumed that $F(0)=0$ and $F(t)> 0$ for $t>0$ and $F$ is supposed to be nondecreasing and continuous. The function $\psi:{\mathbb R}^+ \to {\mathbb R}^+$ is assumed to be nondecreasing and right upper semicontinuous and satisfy $\psi(t)< t$ for each $t>0 $ and $\lim_{n\to \infty} \psi_n(t)=0$. It is proved in \cite{zhang2007} that $F(x) \le \psi(F(x))$ implies $x=0$. We prove in the next lemma that these properties of functions $F$ and $\psi$ imply Properties~\ref{it:famkun} and~\ref{it:famkdeux}.
\begin{lem} Let $\psi:{\mathbb R}^+ \to {\mathbb R}^+$ be a nondecreasing, right upper semicontinuous
function satisfying $\psi(t)< t$ for each $t>0$. Then the sequence of functions
$\seq{\psi}$ defined by
$\psi_n \stackrel{\mbox{\tiny def}}{=} \overset{n}{\overbrace{\psi\circ\psi\circ\cdots\psi}}$
satisfy \emph{\ref{it:famkun}} and \emph{\ref{it:famkdeux}}. \end{lem}
\begin{myproof} \ref{it:famkun}: For $t>0$, since $\psi$ is nondecreasing and $\psi(t) < t$ we have $\psi_n(t) \le \psi(t) < t$ and thus \ref{it:famkun} follows. \ref{it:famkdeux}: Using \cite[Theorem 2]{jacek-1997} we can find a right continuous function $\overline{\psi}: {\mathbb R}^+ \to {\mathbb R}^+$ such that $\psi(t)\le \overline{\psi}(t) <t$ for $t>0$. Thus we easily have \ref{it:famkdeux}, since proving \ref{it:famkdeux} (using $\nu=1$) for a right continuous function is easy. \end{myproof}
In \cite{zhang2007} it is proved that $T$ and $S$ have a common fixed point when ${\mathbb X}$ is a complete metric space and $p=d$. The proof follows the following steps: Since $\seq{x}$ is a Cauchy sequence it converges to $\overline{x} \in {\mathbb X}$. Using the definition of $M$ one easily checks that $M(x_{2n},\overline{x})\to d(S\overline{x},\overline{x})$ and $M(x_{2n},\overline{x})\ge d(S\overline{x},\overline{x})$. Moreover $Tx_{2n}=x_{2n+1}$ also converges to $\overline{x}$. We therefore have \begin{equation}
F \bp{d(Tx_{2n},S\overline{x})} \le \psi \Bp{F \bp{M(x_{2n},\overline{x})}} \,. \end{equation} Using next Lemma~\ref{lem:ptfixe} we obtain that $\overline{x}=S\overline{x}$. Then proving that $\overline{x}$ is also a fixed point of $T$ is given in \cite[Theorem 1]{zhang2007}. We therefore conclude that in order to obtain convergence of the sequence $\seq{x}$ to the unique fixed point of $T$ and $S$ requires to add continuity of $F$ in the hypothesis of Theorem~\ref{thm:ptfixe}.
\begin{lem}\label{lem:ptfixe} Suppose that $F$ is a continuous nondecreasing function, $\psi$ is a right upper semicontinuous function satisfying one of the following property: \begin{enumerate}[label=\emph{($\mbox{E}_{\arabic*}$)}] \item $\psi(t) < t$ for all $t >0$; \item $\psi$ is nondecreasing and for each $t>0$, there exists
$\nu \in {\mathbb N}$, $\nu \ge 1$ such that $\psi_{\nu}(t) < t$. \end{enumerate} Suppose that we have two sequences $\seq{\alpha}$ and $\seq{\beta}$ such that: \begin{equation}
F(\alpha_n) \le \psi \bp{F(\beta_n)} \,. \end{equation} If $\lim_{n\to \infty} \alpha_n = \lim_{n\to \infty} \beta_n =\gamma$ and $\beta_n\ge \gamma$ for all $n \in {\mathbb N}$ then we must have $\gamma=0$. \end{lem} \begin{myproof} We have
\begin{align}
F(\gamma)=\lim_{n\to \infty} F(\alpha_n) \le \limsup_{n\to \infty} \psi \bp{F(\beta_n)}
\le \psi \bp{ \limsup_{n\to \infty} F(\beta_n)} \le \psi \bp{F(\gamma)};
\end{align}
If $\gamma \not=0$ and $\psi(\gamma) < \gamma$ we conclude that $F(\gamma) < F(\gamma)$ which is a
contradiction.
If $\psi$ is nondecreasing we consider the value of $\nu$ associated to $\gamma$ to
obtain:
$F(\gamma) \le \psi_{\nu} \bp{F(\gamma)} < F(\gamma)$ and conclude again by
contradiction. \end{myproof}
\begin{rem} Note that using \cite[Theorem 2]{jacek-1997} we obtain that
a right upper semicontinuous function $\psi$ satisfying
$\psi(t)<t$ for all $t >0$ satisfies Property~\ref{it:famkdeux}
with $\nu=1$. \end{rem}
\end{document}
|
arXiv
|
{
"id": "1111.3464.tex",
"language_detection_score": 0.6521062850952148,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Robustness of quantum key distribution with discrete and continuous variables to channel noise}
\author{Miko{\l}aj Lasota} \author{Radim Filip} \author{Vladyslav C. Usenko} \email{Corresponding author. E-mail: [email protected]} \affiliation{Department of Optics, Palack\'{y} University, 17.\,listopadu 1192/12, 77146 Olomouc, Czech Republic} \pacs{03.67.Dd, 03.67.Hk, 42.50.Ex} \keywords{quantum key distribution; quantum cryptography; discrete variables; continuous variables; squeezed states; single-photon states}
\begin{abstract} We study the robustness of quantum key distribution protocols using discrete or continuous variables to the channel noise. We introduce the model of such noise based on coupling of the signal to a thermal reservoir, typical for continuous-variable quantum key distribution, to the discrete-variable case. Then we perform a comparison of the bounds on the tolerable channel noise between these two kinds of protocols using the same noise parametrization, in the case of implementation which is perfect otherwise. Obtained results show that continuous-variable protocols can exhibit similar robustness to the channel noise when the transmittance of the channel is relatively high. However, for strong loss discrete-variable protocols are superior and can overcome even the infinite-squeezing continuous-variable protocol while using limited nonclassical resources. The requirement on the probability of a single-photon production which would have to be fulfilled by a practical source of photons in order to demonstrate such superiority is feasible thanks to the recent rapid development in this field. \end{abstract}
\maketitle
\section{Introduction}
Quantum key distribution (QKD) is the method of sharing a secret key between two trusted parties using nonclassical properties of quantum states. This enables the security of the key based on physical principles contrary to the mathematical complexity in the classical cryptographic protocols. The first QKD protocols were suggested on the basis of single photons \cite{Bennett84} or entangled photon pairs \cite{Ekert1991} and, respectively, photon-counting measurements. The key bits were encoded to and obtained from the measurement of the states with the discrete spectrum and so the protocols were later referred to as discrete-variable (DV). Alternatively, schemes utilizing multiphoton quantum states of light and encoding the key using observables with the continuous spectrum \cite{Ralph99} were suggested on the basis of Gaussian modulation \cite{Weedbrook12} of squeezed \cite{Cerf2001} or coherent \cite{Grosshans2002,Weedbrook04} states and homodyne detection, and are referred to as continuous-variable (CV) QKD protocols. Both these families of protocols were successfully implemented \cite{Bennett92,*Muller95,*Jennewein00,*Naik00,*Tittel00,Grosshans2003,Lodewyck2007,*Huang2016,Madsen2012,Jouguet2013} and their security was analyzed with respect to individual \cite{Lutkenhaus96,*Slutsky98,*Bechmann06,Grosshans2003a}, collective \cite{Biham97,*Biham02,Navascues2006,*Garcia2006} or the most effective coherent attacks \cite{Kraus05,*Renner05,Leverrier2013}, also taking into account the effects of finite data ensemble size \cite{Hasegawa07,*Hayashi07,Scarani08,Leverrier2010,*Ruppert2014}.
The applicability of all QKD protocols is limited by the imperfections of the devices used to prepare and measure quantum states and also by the properties of quantum channels, which are inclined to losses and noise \cite{Lutkenhaus99,*Brassard00,*Gottesman04,Filip2008,*Usenko2010a,*Jouguet2012,*Usenko2016,Garcia2009}. While it is important to understand which kind of protocols may be advantageous in specific conditions, at present there are no simple criteria for choosing either of their families for a particular task. The main reason for this is that making a fair comparison between DV and CV QKD protocols is hard due to the relativity of practical conditions and even different physical mechanisms leading to imperfections in the devices typically used in the protocol implementations. The only attempt to compare the performance of DV and CV systems done so far concerned the measurement-device independent systems and discussed practical conditions which can strongly vary depending on the wavelength, types of sources, channels and detectors being used, and set of optimistic or pessimistic assumptions being made about the possibility of an eavesdropper to attack the devices \cite{Xu15,*Pirandola15}. In our work we limit the discussion of realistic implementations of DV and CV schemes to a minimum. We mainly focus our attention on comparing the robustness of different types of protocols to the channel noise in the otherwise perfect set-ups. Later, we consider only finite nonclassical resources, \emph{i.e.} quality of single-photon states and finite amount of quadrature squeezing.
Including the problem of channel noise in the QKD security analysis is more typical for the CV case. While it is well known that CV protocols can tolerate ideally any level of channel losses, the excess channel noise can be very harmful and even break the security of these protocols making QKD impossible. It can be considered as a main threat for their security. Indeed, the Gaussian excess noise, which is typically assumed in the CV QKD following the optimality of Gaussian collective attacks \cite{Navascues2006,*Garcia2006}, can break the security at the values below a shot-noise unit for a lossless channel and is further enforced by the channel losses \cite{Lodewyck2007,*Huang2016,Madsen2012,Jouguet2013}.
On the other hand, the analyses of DV QKD protocols performed so far usually focused on different setup imperfections, specifically multiphoton pulses and detection noise, which seem to be the main threats for security in this field. Even if various types of channel noise, originating \emph{e.g.}\,from birefringence effect present in optical fibers, inhomogeneity of the atmosphere, changes of temperature or background light were sometimes included in these investigations \cite{Castelletto03,*Dong11}, they were usually described in a very simplified way, typically by using a single constant parameter, estimation of which could be made experimentally for a given, specific setup. This is especially true for the analyses of free-space DV QKD considering background light, arriving at Bob's detectors from other sources than the one used by Alice \cite{Miao05,*Bonato09,*Bourgoin13}. Disturbances of the states of photons traveling through a given quantum channel in the case of fiber-based QKD schemes were typically treated in a similar way, basing on the assumption that although these kind of effects can generally vary in time, the variations can be considered to be very slow comparing to the time needed for a single photon to propagate from Alice to Bob \cite{Dong11}. In this case, it is reasonable to assume that in short periods of time channel noise affects all of the traveling photons in the same way and can be described by a single, constant parameter. Such noise, called collective, was analyzed in many articles and a lot of possible countermeasures against it have been proposed, utilizing \emph{e.g.} Faraday mirrors \cite{Muller97,*Stucki02}, decoherence-free subspaces \cite{Zanardi97,*Kempe01,*Walton03,*Boileau04,*Li08}, quantum error-rejection codes \cite{Wang04a,*Wang04b,*Kalamidas05,*Chen06}, dense coding \cite{Cai04,*Wang05b,*Li09} or entanglement swapping \cite{Yang13}. However, no detailed analysis of the relationships between the transmittance of the channel connecting Alice and Bob, and the amount of tolerable channel noise has been presented so far and the influence of this relationship on the security of DV QKD protocols has never been analyzed, at least to our knowledge. At the same time due to the continuous improvement of realistic single-photon sources and detectors taking place nowadays, this issue gradually gains importance, especially since the links connecting Alice and Bob in commercial QKD applications may be more noisy than in the typical quantum-optical laboratories \cite{Eraerds10,Qi10}.
In this paper we use in the DV protocols the model for excess channel noise basing on a typical model for CV QKD configuration. We analyze the security bound on such noise for both of these two cases under the assumption that Alice's sources and Bob's detection systems are perfect. Furthermore, we check the stability of the obtained results to the decreasing number of quantum signals exchanged by the trusted parties during the protocol in the finite-key regime. We also compare lower bounds on the secure key rate for the two schemes and find requirements for the nonclassicality of resources needed for their realistic implementations. For the case of ideal sources and detectors our study shows that while CV protocols can successfully compete with DV schemes when the channel transmittance is relatively high, the latter are superior than the former ones for long-distance channels. In this situation it turns out to be possible for DV protocols to beat infinite-squeezing CV schemes even when using realistic single-photon sources. The requirements on the quality of pulses produced by such sources, which would be needed in order to demonstrate this superiority in practice, turn out to be high but reachable by the current technology. For the thermal sources of noise with mean number of photons produced per pulse higher than $10^{-4}$ overcoming CV protocols with DV schemes can be possible only by using single-photon sources with at least $50\%$ probability of producing a non-empty pulse and negligible probability of multiphoton emission.
The paper is organized as follows. In Sec.\,\ref{Sec:Models} we describe the models for excess channel noise used in our analysis: first a standard model for CV QKD case and subsequently the analogous model for DV QKD case. We also derive there all the necessary formulae needed for assessing the security of Gaussian squeezed-state, BB84 and six-state protocols. Next, in Sec.\,\ref{Sec:NumericalResults} we numerically compare the maximal secure values of the channel noise that these protocols can tolerate on the transmittance of the channel connecting Alice and Bob in the case of perfect source and detection system. The comparison is done both in the asymptotic and the finite-key regimes. We also present analytical expressions approximating the maximal tolerable channel noise for DV and CV protocols in the limit of very low transmittance. More realistic situation is analyzed in Sec.\,\ref{Sec:RealisticCase}, where we investigate how the quality of Alice's source may influence the security of our models. Finally, Sec.\,\ref{Sec:Conclusions} concludes our work.
\section{Models for the channel noise in QKD} \label{Sec:Models}
To assess the security of DV and CV QKD protocols we estimate the lower bound on the secure key rate per one pulse emitted by Alice's source. In the DV case this quantity can be expressed as \cite{Scarani09} \begin{equation} K^{(DV)}=p_{exp}\Delta I, \label{eq:keyDV} \end{equation} where $p_{exp}$ denotes the probability for Bob to get a click in his detection system per pulse produced by the source and $\Delta I$ is the so-called secret fraction. Following the quantum generalization of the Csisz\'ar-K\"orner theorem \cite{Csiszar1978} performed by Devetak and Winter \cite{Devetak2005}, this quantity reads \begin{equation} \Delta I=\max[0,I_{AB}-\min\left\{I_{EA},I_{EB}\right\}], \label{eq:DeltaImain} \end{equation} where $I_{AB}$ is the mutual information between Alice and Bob and $I_{EA}$ ($I_{EB}$) represents the amount of information Eve can gain on Alice's (Bob's) data upon an eavesdropping attack. On the other hand in the case of CV QKD protocols the lower bound on the secure key rate can be written simply as \begin{equation} K^{(CV)}=\Delta I, \label{eq:keyCV} \end{equation} since, contrary to the DV case, all of the pulses emitted by Alice's source are registered by Bob's detection system in this situation. Generally speaking both the formulae (\ref{eq:keyDV}) and (\ref{eq:keyCV}) should also contain the so-called sifting probability, representing the chance for the chosen settings of Bob's measurement setup to be compatible with a given signal sent by Alice. However, in the theoretical, asymptotic case, in which we assume that the key produced by Alice and Bob is infinitely long, its generation rate can be increased without compromising its security by performing highly asymmetric version of a given DV or CV protocol, making the sifting probability arbitrary close to one \cite{Scarani09,Lo05b}.
While the methods of calculating the lower bound on the secure key rate (\ref{eq:DeltaImain}) are substantially different in the DV and CV QKD, as we discuss in the following subsections, we develop the model of noise which can be applied to both these families of protocols using the same parametrization. The model is based on coupling every signal mode to an independent thermal reservoir with the coupling ratio $T$, which corresponds to the channel transmittance, and with the reservoirs being characterized by the mean number of thermal photons $\mu$ emitted per pulse. We study robustness of the DV and CV protocols to such thermal noise and derive and compare the security bounds in terms of the maximum tolerable mean numbers of noise photons.
\subsection{Channel noise in CV QKD} \label{Sec:CVmodel}
CV QKD protocols typically use Gaussian states of light and respectively Gaussian modulation, which are compatible with the extremality of Gaussian states \cite{Wolf2006} and enable the security proofs against optimal Gaussian collective attacks \cite{Navascues2006,*Garcia2006}. In our study we consider the Gaussian squeezed-state protocol based on the quadrature modulation and homodyne detection \cite{Cerf2001}. The reason why we choose this scheme instead of the more popular GG02 protocol \cite{Grosshans2002} is that the squeezed-state protocol is more resistant to the channel noise than GG02. This conclusion can be confirmed by comparing the results of our analysis performed for the squeezed-state protocol, presented in Sec.\,\ref{Sec:NumericalResults}, with the analogous results obtained for the GG02 scheme, shown in the Appendix \ref{Sec:GG02protocol}. Moreover, the squeezed-state protocol is the best known Gaussian CV QKD protocol in terms of the resistance to the channel noise \cite{Madsen2012}. Hence, demonstration of its inferiority to the DV protocols in that regard, shown in Sec.\,\ref{Sec:NumericalResults}, automatically implies that also other existing Gaussian CV QKD protocols cannot compete with the DV schemes.
The squeezed-state protocol was shown to be secure against collective \cite{Navascues2006,*Garcia2006} and subsequently against general attacks \cite{Renner2009} in the asymptotic limit and against the collective attacks in the finite-size regime \cite{Leverrier2010}. In our analysis we assume that i) Alice uses a perfect source of quadrature-squeezed states with a quadrature variance $1/V \ll 1$ and that ii) Bob's homodyne detection is perfect with a unity efficiency and no uncontrollable noise. The scheme of the protocol, illustrated in Fig. \ref{fig:CVscheme}, is based on the squeezed signal state preparation by Alice using an optical parametric oscillator (OPO), phase/amplitude quadrature modulation based on the random Gaussian displacements applied in the modulator (M), and transmission along with the local oscillator (LO), being a phase reference for the homodyne measurement, through the Gaussian lossy and noisy channel. The remote party (Bob) splits the signal from the LO and performs homodyne measurement on the squeezed and modulated quadrature. The parties should swap between the bases (i.e. squeezing and modulating either of the two complementary quadratures) in order to perform the channel estimation, but in the following we assume that the channel estimation is perfect.
\begin{figure}
\caption{(color online) CV QKD scheme with lossy and noisy quantum channel connecting Alice and Bob. The following abbreviations were used in this picture: OPO -- optical parametric oscillator, M -- amplitude/phase quadrature modulator, PBS -- polarization beam-splitter, LO - local oscillator.}
\label{fig:CVscheme}
\end{figure}
We now use Gaussian asymptotic security analysis to estimate the security bounds on the CV QKD protocols. To do so, following the Gaussian security proofs, we calculate the lower bound on the secure key rate in the reverse reconciliation scenario, which is known to be more robust against channel loss \cite{Grosshans2003} and being no less sensitive to the channel noise: \begin{equation} \label{LBCV} K^{(CV)}=\max[0,I_{AB}-\chi_{BE}], \end{equation} where $I_{AB}$ is the mutual information shared between the trusted parties, and $\chi_{BE}$ is the Holevo bound \cite{Holevo2001}, upper limiting the information available to an eavesdropper from a collective attack in a given channel. To analyze the security of CV QKD we switch to the equivalent entanglement-based representation \cite{Grosshans2003a} so that Alice and Bob measure a two-mode entangled state shared between them through a quantum channel. The covariance matrix of the state is then given by \begin{equation} \label{gammaAB} \gamma_{AB} = \left( \begin{array}{cc} V\mathbb{I} & \sqrt{T} \sqrt{V^2-1}\sigma_z \\ \sqrt{T} \sqrt{V^2-1}\sigma_z & [V T + (1 - T)W]\mathbb{I} \end{array} \right), \end{equation} where the diagonal matrix $\mathbb{I}=diag(1,1)$, $\sigma_z=diag(1,-1)$ is the Pauli matrix, $V$ is the variance of the modulated squeezed signal states, and $W=2\mu+1$ is the quadrature variance of the thermal noise state. The mutual information between the trusted parties then reads \begin{equation} \label{CVmutinf} I_{AB}=\frac{1}{2}\log_2{\frac{V+W'}{1/V+W'}}, \end{equation}
where $W'=W(1-T)/T$. Following the pessimistic assumption that Eve is able to purify all the noise added to the signal, we estimate the Holevo bound as $\chi_{BE}=S(AB)-S(A|B)$ through the quantum (von Neumann) entropies $S(AB)$ derived from the symplectic eigenvalues \cite{Weedbrook12} $\lambda_{1,2}$ of the state described by the covariance matrix (\ref{gammaAB}), and $S(A|B)$ derived from the symplectic eigenvalue $\lambda_3$ of the state conditioned on Bob's measurement and described by the covariance matrix \begin{equation}
\gamma_{A|B}=\gamma_A-\sigma_{AB}(X \gamma_B X)^{MP}\sigma_{AB}^T, \end{equation} where $\gamma_A=diag(V,V)$, $\gamma_B=diag([V T + (1 - T)W],[V T + (1 - T)W])$ are the matrices, describing the modes A and B individually; $\sigma_{AB}=\sqrt{T(V^2-1)}\sigma_z$ is the matrix, which characterizes correlations between the modes A and B, all being submatrices of (\ref{gammaAB}). MP stands for Moore-Penrose inverse of a matrix (also known as pseudoinverse applicable to the singular matrices), and $X=diag(1,0)$. Here with no loss of generality we assume that the x-quadrature is measured by Bob. Now the Holevo bound can be directly calculated as \begin{equation}\label{holevo1} \chi_{BE}=G\bigg(\frac{\lambda_1-1}{2}\bigg)+G\bigg(\frac{\lambda_2-1}{2}\bigg)-G\bigg(\frac{\lambda_3-1}{2}\bigg), \end{equation} which together with the mutual information (\ref{CVmutinf}) gives the lower bound on the secure key rate (\ref{LBCV}). Here $G(x)=(x+1)\log_2{(x+1)}-x\log_2x$ is the bosonic entropic function \cite{Serafini2005}. The bounds on the channel noise, characterized by $\mu$, are then derived by turning the secure key rate (\ref{LBCV}) to zero.
We also consider the extension of the protocol, when trusted noise is added on the detection stage to improve the robustness of the protocol to the channel noise \cite{Garcia2009} (note that the heterodyne detection can be seen as the particular case of such noise addition and therefore was not considered in our study). This provides the maximum tolerable channel noise for a perfect CV QKD protocol with a given squeezing $1/V$ and upon given channel transmittance $T$.
\subsection{Channel noise in DV QKD} \label{Sec:DVmodel}
Alternatively to the above-described CV scheme we consider the use of polarization-based BB84 \cite{Bennett84} and six-state \cite{Bruss98} protocols, both belonging to the family of DV protocols, to generate secure key by Alice and Bob. The scheme which we analyze is illustrated in Fig.\,\ref{fig:DVscheme}. We assume that i) Alice's source is a perfect single-photon source and ii) Bob uses perfect single-photon detectors with no dark counts and unity efficiency. Our basic assumption on these detectors is that they do not have the ability to resolve the number of incoming photons. However, in the Appendix \ref{Sec:DifferentDetection} we analyze also the opposite possibility for the comparison. Since Alice's source is perfect, it never emits multiphoton pulses and Eve cannot perform photon-number-splitting attacks on the signal pulses. If so, Alice and Bob cannot gain anything by using decoy-pulse method \cite{Hwang03,*Wang05a,*Lo05} and we do not consider it in our analysis.
\begin{figure}
\caption{(color online) Our model for DV QKD scheme with lossy and noisy quantum channel connecting Alice and Bob. The following abbreviations were used in this picture: SPS -- single-photon source, PBS -- polarization beam-splitter.}
\label{fig:DVscheme}
\end{figure}
In the model presented in Fig.\,\ref{fig:DVscheme} the channel noise, coupled to the signal during its propagation between Alice and Bob, is generated in two orthogonal polarizations by two independent sources of thermal noise. In fact this model is completely analogous to the one analyzed in Sec.\,\ref{Sec:CVmodel}, where two polarization modes are used to transmit the signal and the local oscillator. Since the effect of this noise on the bright local oscillator is negligible, it is not considered in the CV case. We assume here that the photons emitted by a given source of noise have the same polarization as signal photons transmitted through the channel to which it is coupled. We denote the probability of emitting $n$ noise photons by a given source by $p_n(\mu)$, where $\mu$ is the mean number of noise photons produced per pulse. For thermal noise this probability is given by \begin{equation} p_n(\mu)=\frac{\mu^n}{(\mu+1)^{n+1}}. \end{equation}
Similarly to the CV case we assume that Eve fully controls the noise coupled to the signal in the quantum channel. Therefore, she can perform any attack which produces the same QBER as would be obtained by the trusted parties if there was no eavesdropper. We assume here that Eve executes the general collective attack, which is optimal for the DV QKD protocols under given QBER \cite{Kraus05,*Renner05}.
We also consider the possibility for Alice and Bob to perform so-called preprocessing \cite{Renner05}, allowing them to improve the security of the generated key by deliberately adding some noise to it before going to the stages of error correction and privacy amplification. This technique can be seen as the DV counterpart to the noise addition on the Bob's side considered in the CV case above in order to reduce the information which is available to Eve.
In the case without preprocessing, the most general collective attacks performed by Eve on BB84 protocol can give her $I_{EA}^{BB84}=I_{EB}^{BB84}=H(Q)$ \cite{Renner05}, where $H(Q)$ is Shannon entropy and $Q$ represents the level of QBER measured by Alice and Bob in their raw key. Since for the asymptotic case of infinitely long key, which we assume here, the mutual information between Alice and Bob when they are not performing preprocessing stage can be written as $I_{AB}=1-H(Q)$ \cite{Scarani09}, using equations (\ref{eq:keyDV}) and (\ref{eq:DeltaImain}) we can get the following expression for the lower bound on the secure key rate: \begin{equation} K^{(BB84)}=p_{exp}\max[0,1-2H(Q)]. \label{eq:DeltaIBB84} \end{equation}
On the other hand, the upper bound on the information Eve can get by making the most general collective attacks when Alice and Bob use six-state protocol can be written as \cite{Renner05} \begin{equation} I_{EA}^{6state}=I_{EB}^{6state}=F(Q)-H(Q), \end{equation} where \begin{equation} F(Q)=-\left(1-\frac{3Q}{2}\right)\log_2\left(1-\frac{3Q}{2}\right)-\frac{3Q}{2}\log_2\frac{Q}{2} \end{equation} If so, then from (\ref{eq:keyDV}) and (\ref{eq:DeltaImain}) we get \begin{equation}
K^{(6state)}=p_{exp}\max\left[0,1-F(Q)\right] \label{eq:DeltaI6state}. \end{equation}
The above formulae for $K^{(BB84)}$ and $K^{(6state)}$ get more complicated, when Alice and Bob perform preprocessing, which can be done \emph{e.g.} by randomly flipping some bits of the raw key by Alice \cite{Kraus05,*Renner05}. In this case the mutual information about the key shared by Alice and Bob transforms into \begin{equation} I_{AB}(Q,x)=1-H[(1-x)Q+x(1-Q)], \end{equation} where $x$ is the probability for Alice to flip a given bit of the raw key. In turn $I_{EA}^{BB84}$ (which is still equal to $I_{EB}^{BB84}$) can be written as \begin{widetext} \begin{eqnarray} I_{EA}^{BB84}(Q,x)&=&\max_{x\in[0,1/2]}\min_{\lambda\in[0,Q]}\left[\sum_{i=1}^4A_i\log_2A_i-(1+\lambda-2Q)\log_2(1+\lambda-2Q)-\nonumber\right.\\&-&\left.2(Q-\lambda)\log_2(Q-\lambda)-\lambda\log_2\lambda\right], \label{eq:IAEBB84preprocessing} \end{eqnarray} where \begin{equation} A_{1,2}=\frac{1-Q\pm\sqrt{(1-Q)^2+16x(1-x)(\lambda-2Q+1)(\lambda-Q)}}{2} \end{equation} \end{widetext} and \begin{equation} A_{3,4}=\frac{Q\pm\sqrt{Q^2+16x(1-x)\lambda(\lambda-Q)}}{2}, \end{equation} while for six-state protocol we have \begin{equation} I_{EA}^{6state}(Q,x)=\max_{x\in[0,1/2]}\left[\sum_{i=1}^{4}B_i\log_2B_i+F(Q)\right], \label{eq:IAE6statepreprocessing} \end{equation} where \begin{equation} B_{1,2}=\frac{1-Q\pm\sqrt{(1-Q)^2-4x(1-x)Q(2-3Q)}}{2} \end{equation} and \begin{equation} B_{3,4}=\frac{Q\left[1\pm(1-2x)\right]}{2}. \end{equation}
From the above analysis follows that the only parameter which Alice and Bob have to estimate in order to be able to assess the security of their DV QKD protocol is $Q$. We will further express this quantity in terms of the parameters of a given setup, taking into consideration the assumptions that were made at the beginning of this section. To do so let us first observe that since the scheme shown in Fig.\,\ref{fig:DVscheme} is perfectly symmetric in respect to polarizations, we don't have to consider separately the cases when Alice generates differently polarized photons. Instead of this, we can just consider one single case, in which Alice emits single photon in a randomly chosen polarization state, which we simply call \emph{right}. The orthogonal polarization state is called \emph{wrong} in this situation. Similarly, we call the detector to which signal photon emitted by Alice would go, if it is not lost during the propagation and if Bob chose the right basis for his measurement, the \emph{right} detector, and the other one -- the \emph{wrong} detector. Now by $p_+(k,l)$ [$p_-(k,l)$] let's denote the probability that signal photon would [would not] arrive at the right detector in a given attempt to generate a single bit of the key and at the same time $k$ noise photons would arrive at the right detector, while $l$ noise photons would arrive at the wrong detector. These two quantities are equal to \begin{equation} p_+(k,l)=T\pi_k(T)\pi_l(T) \label{eq:pplus} \end{equation} and \begin{equation} p_-(k,l)=(1-T)\pi_k(T)\pi_l(T), \label{eq:pminus} \end{equation} where \begin{equation} \pi_k(T)=\sum_{n=k}^\infty p_n(\mu){n \choose k} (1-T)^kT^{n-k}. \end{equation} Since we assume here that Bob's detectors do not have photon-number resolution, Alice and Bob automatically have to accept every situation in which all of the photons leaving the channel enter the same detector. Nevertheless, they can discard from the generated key all of the cases when both Bob's detectors click at the same time (we call this kind of event a \emph{double click} here). If they do so, the expected probability for accepting a given event by users of six-state protocol can be written as \begin{equation} p_{exp}=\sum_{k=0}^\infty p_+(k,0)+\sum_{k=1}^\infty p_-(k,0)+\sum_{l=1}^\infty p_-(0,l). \label{eq:pexpII} \end{equation} It is clear that only the last term in the above formula contributes to the error rate, so the expression for QBER in our model takes the following form: \begin{equation} Q=\frac{\sum_{l=1}^\infty p_-(0,l)}{p_{exp}}. \label{eq:QII} \end{equation}
\section{Numerical results and analytical expressions} \label{Sec:NumericalResults}
We now compare the security of the CV and DV QKD protocols in the presence of channel noise. To do so we perform numerical calculations in order to find the dependency of the maximal values of $\mu$, for which it is possible to generate secure key, on the transmittance $T$ of the channel connecting Alice and Bob in the cases when they use different QKD schemes. The relationships between such $\mu_\mathrm{max}^\mathrm{DV}(T)$ functions computed for BB84 and six-state protocols and the analogous function $\mu_\mathrm{max}^\mathrm{CV}(T)$ calculated for the squeezed-state scheme, both for the basic scenario and for the case when Alice and Bob try to improve the security of all these protocols by deliberately adding some noise to their raw keys (as was described in Sec.\,\ref{Sec:Models}), are presented in Fig.\,\ref{fig:relativeresults}.
Let us begin the analysis of Fig.\,\ref{fig:relativeresults} by focusing on the comparison between the six-state and squeezed-state protocols. As we can see in this picture for relatively high values of $T$ the former of these two cryptographic schemes allows for significantly higher values of $\mu$ than the latter one. However, this advantage quickly vanishes when $T$ decreases and for some intermediate values of the transmittance of the channel connecting Alice and Bob squeezed-state protocol appears to be slightly better suited for noisy quantum cryptography than the six-state scheme. Nevertheless, when $T$ decreases even further, at some point six-state protocol again starts to outperform the squeezed-state scheme and its advantage grows while $T\rightarrow 0$.
In fact, the relationship between BB84 and squeezed-state protocols is also very similar to the one described above. However, since for every value of $T$ BB84 protocol happens to be less resistant to the channel noise than the six-state scheme, the region of channel transmittance for which squeezed-state protocol allows for stronger channel noise than BB84 turns out to be significantly larger than in the case of the comparison between the six-state and squeezed-state protocols discussed before. Also the relative advantage of the CV protocol in this region is higher. In Fig.\,\ref{fig:relativeresults} we can also see that for the values of $T$ between roughly $10^{-0.5}$ and $10^{-2}$ adding noise to the raw key by the legitimate participants of a given QKD protocol can be more profitable for squeezed-state protocol than for DV protocols, while for $T<10^{-2}$ the situation is opposite.
\begin{figure}
\caption{(color online) Ratios between maximal values of $\mu$ for which it is possible to generate secure key using CV squeezed-state protocol ($\mu_\mathrm{max}^\mathrm{CV}$) and both DV protocols ($\mu_\mathrm{max}^\mathrm{DV}$) considered in our analysis, plotted as a function of channel transmission $T$ for the situation when Alice and Bob perform the randomization stage of their raw key in order to increase its security (dashed lines) or do not perform it (solid lines).}
\label{fig:relativeresults}
\end{figure}
Although it is not possible to find any simple, analytical expressions for the functions $\mu_{max}(T)$ in the general case, the analytical boundaries approximating it in the limit of $T\rightarrow 0$ can be derived for every protocol of our interest.
\textbf{Expression for DV QKD:} Derivation of the boundary for the case of six-state and BB84 protocols is relatively easy. To do it, we observe that when $T\rightarrow 0$ and $\mu\rightarrow 0$, the formula for QBER can be easily transformed into \begin{equation} Q\approx\frac{\mu}{2\mu+T}. \end{equation} If so, then for $T \ll 1$ the maximal secure value of $\mu$ depends on $T$ as follows: \begin{equation} \mu_{\mathrm max}(T)=\frac{TQ_\mathrm{th}}{1-2Q_\mathrm{th}}, \label{eq:approxmuDV} \end{equation} where $Q_\mathrm{th}$ is the threshold value of QBER, which for the cases of six-state and BB84 protocols are approximately equal to $12.6\%$ and $11\%$ respectively \cite{Renner05}.
\textbf{Expression for CV QKD:} In the case of the squeezed-state CV QKD protocol, when no noise is deliberately added on the receiver side, the analytical lower bound on the secure key rate can be simplified to \begin{equation} K^{(CV)} \approx (T-\mu)\log_2{e}+\mu\log_2{\mu}, \end{equation} by using series expansion around $T=0$, taking the limit of infinite modulation $V \to \infty$ and performing series expansion around $\mu=0$. The value of $\mu$ which turns this simplified expression to zero can be calculated analytically and expressed using Lambert W function as
\begin{equation} \mu_{\mathrm max}(T)=\exp[1+W_{-1}(-T/e)]. \label{eq:approxmuCV} \end{equation}
The comparison between the boundaries given by formulae (\ref{eq:approxmuDV}) and (\ref{eq:approxmuCV}) and the results of our numerical calculations of the functions $\mu_{\mathrm max}(T)$ performed for the cases of six-state and squeezed-state protocols, which is illustrated in Fig.\,\ref{fig:zoomresults}, shows good agreement between our analytical and numerical results in the limit of $T\rightarrow0$.
\begin{figure}
\caption{(color online) Maximal values of $\mu$ for which it is possible to generate secure key as a function of channel transmittance $T$ calculated numerically (solid lines) for the cases of Alice and Bob using six-state protocol (red lines) and squeezed-state protocol (black lines), plotted along with the analytical approximations (dot-dashed lines) of the functions $\mu_{\mathrm max}(T)$ valid for the case of $T\rightarrow0$, given by formulae (\ref{eq:approxmuDV}) and (\ref{eq:approxmuCV}) respectively.}
\label{fig:zoomresults}
\end{figure}
While our main goal in performing the analysis presented above was to identify the conditions in which only one of the two main families of QKD protocols can be used to provide security for the process of key generation, its results cannot help us with answering the question which protocol one should choose in a particular case when both CV and DV QKD schemes can be secure at the same time. Facing such a decision it is always good to compare the lower bounds for the secure key rate for different protocols. This kind of comparison, performed for the six-state and squeezed-state schemes, is presented in Fig.\,\ref{fig:KeyComparison} where the function of $K(T)$ was plotted for a few different values of $\mu$, ranging from $10^{-5}$ to $0.5$. Although typical values of $\mu$ in a dark fiber, dedicated solely for the generation of secret key, can be estimated to be on the level of $10^{-4}$--$10^{-5}$ (basing on the experimental results obtained in \cite{Jouguet2013}), in commercial QKD applications utilizing telecom fibers populated by strong classical signals the channel noise can be considerably stronger. In this situation the actual level of $\mu$ would primarily depend on the number of classical channels multiplexed in a given fiber and the power of classical signals transmitted through them \cite{Qi10}. For this reason in the analysis presented in this paper we decided not to focus on a particular level of $\mu$, but consider a broad range of its values, encompassing several orders of magnitude.
From the Fig.\,\ref{fig:KeyComparison} one can conclude that if only $T$ is considerably larger than the minimal secure transmittance of the channel connecting Alice and Bob for CV squeezed-state protocol, this scheme can always provide comparable but slightly higher lower bound on the secure key rate than the six-state DV QKD protocol. Similar conclusion can be drawn from the comparison of BB84 and squeezed-state schemes. The main reason for this advantage stems from the capability of encoding more than one bit of information in a single pulse by using CV QKD protocols, which in turn is impossible for the considered DV schemes based on qubits. The results presented in Fig.\,\ref{fig:KeyComparison} can be also used to predict the outcome of a possible comparison of the robustness of the six-state and squeezed-state protocols to the channel noise for a given non-zero value of $K$. In this case one should just compare the minimal secure values of $T$ for these protocols, which can be reached for different levels of $\mu$ for a desired $K$. It is important to note, however, that the lower bound on the secure key rate in our work is calculated per use of the channel, so it contains only partial information on the achievable rate of a particular implementation of a given QKD protocol. In order to calculate the lower bound on the amount of bits of the final key per unit of time, one would have to multiply the expression for $K$ (formula (\ref{eq:keyDV}) or (\ref{eq:keyCV}) for the DV or CV protocols respectively) by the repetition rate of the system, which depends on the setup. Therefore comparing the key rates in the general case can be misleading.
\begin{figure}
\caption{(color online) Lower bound for the secure key rate as a function of transmittance of the channel connecting Alice and Bob, plotted for $\mu=0.5$ (red lines), $\mu=10^{-1}$ (orange lines), $\mu=10^{-2}$ (yellow lines), $\mu=10^{-3}$ (green lines), $\mu=10^{-4}$ (blue lines) and $\mu=10^{-5}$ (black lines) for six-state protocol (solid lines) and squeezed-state protocol (dashed lines) with the assumption that Alice's sources and Bob's detection systems are perfect.}
\label{fig:KeyComparison}
\end{figure}
The analysis presented above was performed for the asymptotic case of infinite number of quantum signals exchanged by Alice and Bob during the key generation process. However, in realistic situation this number, denoted here by $N$, is always finite. Therefore, it is instructive to check the stability of the discussed results in the finite-key regime. In order to do that we utilize the calculation method introduced for the DV QKD case in \cite{Scarani08} and adopted for the CV protocols in \cite{Leverrier2010}. For definiteness we set the values of all of the failure probabilities present in the mathematical formulas introduced there to the level of $10^{-10}$. The results of this calculation are illustrated in Fig.\,\ref{fig:MuComparisonFinite}, where the ratio of $\mu^\mathrm{CV}_\mathrm{max}/\mu^\mathrm{DV}_\mathrm{max}$ for squeezed-state and six-state protocols is plotted for different numbers of $N$.
As it turns out, if only the transmittance of the quantum channel connecting Alice and Bob is not particularly high, the finite-size effects have more negative influence on the squeezed-state protocol than on the DV schemes. In particular, for any finite $N$ there exists a corresponding threshold value for $T$ below which generation of secure key by utilizing squeezed-state protocol becomes impossible even for $\mu\rightarrow 0$. On the other hand, as long as the lossy and noisy channel connecting Alice and Bob is the only imperfect setup element, no such threshold appears for the DV protocols. Therefore, for limited $N$ the ratio of $\mu^\mathrm{CV}_\mathrm{max}/\mu^\mathrm{DV}_\mathrm{max}$ decreases much faster and eventually reaches zero, contrary to the asymptotic case. Furthermore, Fig.\,\ref{fig:MuComparisonFinite} shows that the value of $T$, below which six-state protocol becomes more resistant to the channel noise than a given CV QKD scheme, grows with the decreasing number of quantum signals exchanged by Alice and Bob.
\begin{figure}
\caption{(color online) Ratio between maximal values of $\mu$ for which it is possible to generate secure key using squeezed-state protocol ($\mu_\mathrm{max}^\mathrm{CV}$) and six-state protocol ($\mu_\mathrm{max}^\mathrm{DV}$), plotted as a function of channel transmittance $T$ for the asymptotic case in which the number of quantum signals exchanged by Alice and Bob during the protocol is infinite (solid line) and for the situations when it equals to $N=10^{10}$ (dashed line), $N=10^8$ (dot-dashed line) and $N=10^6$ (dotted line). The calculations were made with the assumption that the trusted parties do not increase the security of their raw key by performing the randomization stage.}
\label{fig:MuComparisonFinite}
\end{figure}
\section{Requirements for nonclassical resources} \label{Sec:RealisticCase}
Knowing that for the case when Alice's source and Bob's detection system are perfect DV QKD protocols can provide one with the security of key generation process for slightly higher values of $\mu$ than the squeezed-state CV protocol if only the transmittance of the channel connecting Alice and Bob is low enough, we can now consider the possibility for realizing this kind of scenario in the situation when the sources of signal owned by Alice are not ideal.
In order to assess the quality of a single-photon source needed for secure realization of QKD protocols for the combinations of parameters $T$ and $\mu$ for which squeezed-state protocol is insecure, we will assume in this section that Alice's source produces genuine single-photon pulses with probability $p$ and empty pulses with probability $1-p$, \emph{i.e.}\,it never emits multiphoton pulses. The reason for adopting this particular model of Alice's source for our considerations is that while decreasing the probability for multiphoton emission to a very low level is possible these days for many different kinds of realistic single-photon sources \cite{Fasel04, Keller04,*Brokmann04,*Laurat06,*Pisanello10,*Mucke13,Claudon10}, constructing a high-quality source which would produce non-empty pulses with probability close to one remains a serious challenge for experimental physicists. This task is especially hard to be accomplished for the case of deterministic single-photon sources, which are usually affected by poor collection efficiency of generated photons \cite{Eisaman11}. However, very promising sources based on quantum dots embedded in photonic nanowires or micropillar cavities have been developed recently, with probability of producing a single-photon pulse exceeding $70\%$ and potentially reaching even $95\%$ \cite{Claudon10,Bulgarini12,*Gazzano12,Somaschi16}. Furthermore, relatively efficient probabilistic single-photon sources, based especially on the spontaneous parametric down-conversion (SPDC) process, with very low probabilities of emitting a multiphoton pulse and with the heralding efficiency exceeding $60\%$ were already developed more than a decade ago \cite{Fasel04}. Nowadays, reports on SPDC-based sources with $p>80\%$ can be find in the literature \cite{Pomarico12,*Pereira13,*Ramelow13}.
\begin{figure}
\caption{(color online) Requirements on the value of squeezing parameter (dashed lines) and the probability $p$ of producing non-empty signal pulse by a single-photon source (solid lines) needed to be reached for the security of, respectively, the squeezed-state protocol and a) six-state, b) BB84 protocol, plotted as functions of the transmittance of the channel connecting Alice and Bob for six different values of $\mu$: $\mu=0.5$ (red lines), $\mu=10^{-1}$ (orange lines), $\mu=10^{-2}$ (yellow lines), $\mu=10^{-3}$ (green lines), $\mu=10^{-4}$ (blue lines) and $\mu=10^{-5}$ (black lines). Vertical dotted lines denote the values of $T$ for which squeezed-state protocol becomes insecure for particular values of $\mu$.}
\label{fig:CVDVrealistic}
\end{figure}
\begin{figure}
\caption{(color online) Requirements on the probability $p$ of producing non-empty signal pulse by a single-photon source needed to be fulfilled for the generation of secure key by using six-state (red, dashed line) or BB84 (blue, solid line) protocol for the same value of $\mu$, for which squeezed-state protocol with infinite squeezing stops being secure at a given transmittance $T$.}
\label{fig:DVReqComparison}
\end{figure}
Adopting the model for realistic single-photon source described above, we investigated the dependence of the minimal probability $p$ of producing non-empty pulse by Alice's source, required for the six-state and BB84 protocols to be secure, on the transmittance of the channel connecting Alice and Bob for a few different values of the power of the source of noise in the DV QKD model, illustrated in Fig.\,\ref{fig:DVscheme}. The results of this investigation are plotted in Fig.\,\ref{fig:CVDVrealistic}. In the same figure we also plotted the dependency of the value of squeezing parameter, required for the security of the squeezed-state protocol in the model for CV QKD pictured in Fig.\,\ref{fig:CVscheme}, on $T$. In order to make necessary calculations for squeezed-state protocol in realistic case we used the generalized state preparation model for CV QKD in which modulation and squeezing of the states emitted by Alice's source can be parametrized separately \cite{Usenko2011}. While the plots given in Fig.\,\ref{fig:CVDVrealistic} were obtained for the very strong modulation variance ($10^3$ shot-noise units), varying this quantity does not significantly affect the results.
From Fig.\,\ref{fig:CVDVrealistic} one can deduce that the requirements for the probability $p$ of emitting non-empty pulse by Alice's single-photon source, which would have to be fulfilled in order to ensure security of the DV QKD protocols for the values of $T$ for which squeezed-state protocol is no longer secure, are generally quite demanding, especially if the power of the source of noise is relatively high. For different levels of $\mu$ the minimal values of $p$ which would be needed to realize this task are given by the crossing points of the solid and dotted lines of the same colors displayed in Fig.\,\ref{fig:CVDVrealistic}. While in practice overcoming the squeezed-state protocol by the DV QKD schemes may be very hard or even impossible to demonstrate for relatively high values of $\mu$, it is certainly achievable for realistic sources in the case of $\mu\ll 1$, as the requirements for $p$ shown in Fig.\,\ref{fig:CVDVrealistic} become more and more relaxed when $\mu\rightarrow 0$. This conclusion can be confirmed in Fig.\,\ref{fig:DVReqComparison}, where the minimal required values of $p$ are plotted as the functions of $\mu$ both for the BB84 and six-state protocol. The results of our analysis shown in Fig.\,\ref{fig:CVDVrealistic} and Fig.\,\ref{fig:DVReqComparison} indicate that even DV QKD schemes with inefficienct sources of photons can be capable to overcome the CV protocols for long-distance quantum cryptography with ultra low channel noise.
Not surprisingly, in Fig.\,\ref{fig:DVReqComparison} one can also see that for every level of $\mu$ the value of $p$ needed to overcome squeezed-state protocol is larger for the BB84 than for the six-state protocol. This means that a demonstration of the superiority of the six-state protocol over the squeezed-state scheme in realistic situation would be easier to perform than an analogous demonstration for BB84 protocol. This conclusion justifies our choice to focus more on the six-state protocol in this work, despite much larger popularity of the BB84 scheme.
\begin{figure}
\caption{(color online) Minimal values of the probability $p$ of producing non-empty signal pulse by a single-photon source, needed for the six-state protocol to be secure for a given pair of values of the channel transmittance $T$ and noise mean photon number $\mu$ for which squeezed-state protocol is already insecure. White color indicates the regions of the plot where either the squeezed-state protocol is still secure or the six-state protocol is insecure even for $p=1$.}
\label{fig:p1ContourPlot}
\end{figure}
While Fig.\,\ref{fig:DVReqComparison} shows only the minimal values of $p$ for which DV QKD protocols can still be secure for given $\mu$ and $T$ that already breaks the security of the CV QKD schemes, for higher $p$ demonstrating the superiority of BB84 or six-state protocol over the squeezed-state scheme may be realized also for the lower transmittance of the quantum channel connecting Alice and Bob. Therefore, it is reasonable to ask about the whole region of parameters $\mu$ and $T$ for which overcoming the performance of the squeezed-state protocol by a given DV QKD scheme is possible. Such a region, found for the case of six-state protocol, is illustrated in Fig.\,\ref{fig:p1ContourPlot}. One can see there that it is relatively narrow. This is because the closer $T$ is to the minimal secure transmittance of the quantum channel connecting Alice and Bob for a given $\mu$, the faster the minimal required value of $p$ goes to one. This tendency could actually be observed even before, in Fig.\,\ref{fig:CVDVrealistic}. Fig.\,\ref{fig:p1ContourPlot} also confirms that the requirement for $p$ relaxes when $\mu\rightarrow0$.
Beside sources of photons, another fundamental part of the setup needed for the implementation of DV QKD protocols are single-photon detectors. In some situations imperfection of these devices can also affect the security of such schemes in significant way. In particular, every realistic single-photon detector is characterized by a non-zero dark count rate. The influence of these unwanted clicks on the results presented in this work is negligible as long as the value of $T$ is more than two orders of magnitude higher than the probability $d$ to register a dark count per single detection window. However, for lower transmittance of the quantum channel dark counts considerably affect the security of DV QKD protocols and can become the major issue. They result in threshold values of channel transmittance $T_{th}$, below which overcoming squeezed-state protocol with DV schemes becomes impossible even if the single-photon source used by Alice is perfect. These thresholds strongly depend on the relationship between $d$ and the detection efficiency $\eta$ of the measurement devices utilized by Bob. Typical values of $d/\eta$ that can be found in the literature describing recent DV QKD experiments range from $10^{-4}$ to $10^{-7}$ \cite{Zhang08,*Walenta14,*Valivarthi15,*Takemoto15,*Wang15,*Tang16}. During our work we found out that in this region $T_{th}$ can be upper-bounded by \begin{equation} T_{th}^{\,6 state}\leq10^{1.07\log_{10}\left(d/\eta\right)+1.45} \end{equation} for the case when the trusted parties implement six-state protocol or \begin{equation} T_{th}^{BB84}\leq10^{1.15\log_{10}\left(d/\eta\right)+2.12}. \end{equation} when they choose BB84 scheme.
On the other hand, if Bob's measurement system does not register any dark counts, the limited detection efficiency does not affect the results of our calculations as long as $T<10^{-2}$. This is because for low values of $T$ almost all of the non-empty pulses arriving at Bob's measurement system contain either a single signal photon or a single noise photon. Therefore, since the limited detection efficiency reduces the fractions of registered signal and noise photons in exactly the same way its value does not matter for the security threshold. Only when the transmittance of the quantum channel connecting Alice and Bob is relatively high and the probability for more photons to arrive at Bob's detectors at the same time becomes significant, the situation can be different. In this case limited detection efficiency makes the requirement for the quality of Alice's source slightly more demanding.
\section{Conclusions} \label{Sec:Conclusions}
In the analysis presented above we compared the security of two DV protocols, namely BB84 and six-state, and CV squeezed-state protocol in the situation when the only imperfect element of the setup used by Alice and Bob is the quantum channel connecting them. We assumed here that this channel is lossy and that the noise coupled to the signal during its propagation through it is of the type of thermal reservoir, which can be seen as a typical scenario for CV QKD case. The results of our analysis, depicted in Fig.\,\ref{fig:relativeresults}, clearly show that while for some intermediate values of the channel transmittance continuous-variable squeezed-state protocol is comparably resilient to the channel noise as BB84 and six-state schemes, for the cases of $T \to 1$ and $T \ll 1$ both the DV protocols perform better. It suggests that in the scenario when Alice and Bob have high-quality sources and detectors, but the quantum channel connecting them is lossy and noisy, DV QKD technique can be seen as having more potential for generating a secure cryptographic key than CV QKD. Although exploiting this potential in practice may be challenging, it is within our reach. With the recent engineering progress in the field of single-photon sources it can be even possible to demonstrate the superiority of realistic DV protocols over the infinite-squeezing ideal CV schemes in the regime of $T \ll 1$, as can be seen in Fig.\,\ref{fig:CVDVrealistic}. This conclusion may provide some additional motivation for the experimental physicists to focus even more of their efforts on developing novel high-quality sources with high probability of producing non-empty pulse and very low probability for multiphoton emission or improving the performance of the existing ones.
\noindent {\bf Acknowledgments. --} The research leading to these results has received funding from the EU FP7 under Grant Agreement No. 308803 (project BRISQ2), co-financed by M\v SMT \v CR (7E13032). M.L. and V.C.U. acknowledge the project 13-27533J of the Czech Science Foundation. M.L. acknowledges support by the Development Project of Faculty of Science, Palacky University.
\appendix
\section{} \label{Sec:GG02protocol}
\begin{figure}
\caption{(color online) Ratios between maximal values of $\mu$ for which it is possible to generate secure key using CV GG02 protocol ($\mu_\mathrm{max}^\mathrm{CV}$) and both DV protocols ($\mu_\mathrm{max}^\mathrm{DV}$) considered in our analysis, plotted as a function of channel transmission $T$ for the situation when Alice and Bob perform the randomization stage of their raw key in order to increase its security (dashed lines) or do not perform it (solid lines).}
\label{fig:relativeresults2}
\end{figure}
In the main body of our paper we considered Gaussian squeezed-state CV QKD protocol, using it for the comparison with DV QKD protocols in terms of their robustness to the channel noise. However, due to the popularity of the GG02 scheme based on coherent states \cite{Grosshans2002}, it is meaningful to perform a similar analysis also for this protocol. In order to do all the necessary calculations in this situation, one can once again utilize the formulae introduced in Sec.\,\ref{Sec:CVmodel}, only assuming that this time the variance of the signal states is $V=1$. In Fig.\,\ref{fig:relativeresults2} we present the results of our comparison between the maximal values of the parameter $\mu$ ensuring the security of GG02 and the DV QKD protocols. This comparison is similar to the one made for the squeezed-state scheme in Sec.\,\ref{Sec:NumericalResults}, which results are depicted in Fig.\,\ref{fig:relativeresults}. By comparing the two aforementioned figures with each other one can confirm that the squeezed-state protocol is indeed more resistant to the channel noise than the GG02 scheme, as was already stated in the first paragraph of Sec.\,\ref{Sec:CVmodel}. As can be seen in Fig.\ref{fig:relativeresults2}, contrary to the case of the squeezed-state protocol, for every possible value of $T$, GG02 scheme allows for significantly lower values of $\mu_\mathrm{max}$ than the BB84 and six-state protocols.
\section{} \label{Sec:DifferentDetection}
In the analysis of DV QKD protocols presented in the main body of this article we assumed that Bob's detectors have perfect detection efficiency, but do not have the ability to resolve the number of photons entering them. At first sight it would seem that replacing them with photon-number-resolving detectors should improve the setup, making it more resilient to the channel noise. However this intuition does not necessarily has to be correct. Here we are going to show that in our model, when the source of channel noise has thermal statistics, equipping Bob's detectors with photon-number resolution does not change the function of $\mu_\mathrm{max}^\mathrm{DV}(T)$ in any way, while for Poisson statistics it can even have negative effect on QKD security.
In order to accomplish this task, we will start with adapting the expressions for $p_{exp}$ and $Q$, given previously by formulae (\ref{eq:pexpII}) and (\ref{eq:QII}) respectively, to the case of photon-number-resolving detetcors used by Bob. We get: \begin{equation} p_{exp}^{(II)}= p_+(0,0)+p_-(1,0)+p_-(0,1) \end{equation} and \begin{equation} Q^{(II)}=\frac{ p_-(0,1)}{p_{exp}^{(II)}}. \label{eq:QIII} \end{equation} Since for both DV protocols considered here the formulae for $\Delta I$ depend only on the parameter $Q$ (and optinally the probability $x$ to flip a bit by Alice, if the preprocessing stage is being performed), it is obvious that the condition for photon-number-resolving detectors to offer better security of our DV QKD schemes than simple on/off binary detectors can be written in the form of the following inequality: \begin{equation} Q^{(II)}<Q. \end{equation} Using equations (\ref{eq:QII}) and (\ref{eq:QIII}), and taking advantage of the facts that \begin{equation} p_+(k,l)=p_+(l,k) \end{equation} and \begin{equation} p_-(k,l)=\frac{1-T}{T}p_+(k,l), \end{equation} we can transfrom this condition into \begin{equation} p_+(1,0)\cdot\sum_{k=1}^\infty p_+(k,0)<p_+(0,0)\cdot\sum_{k=2}^\infty p_+(k,0). \end{equation} After inserting (\ref{eq:pplus}) and performing some algebraic calculations, we can get the following final version of this condition: \begin{eqnarray} &&\sum_{k=1}^\infty\sum_{n=k}^\infty\sum_{m=1}^\infty\left[p_n(\mu)p_m(\mu)m{n \choose k}-\right.\\&-&\left.p_{n+1}(\mu)p_{m-1}(\mu){n+1 \choose k+1}\right](1-T)^kT^{n+m-k-1}<0\nonumber. \label{eq:condition2} \end{eqnarray}
The above inequality cannot be solved analytically in the general case. However, it can be further simplified in two extreme cases of $T\rightarrow 0$ and $T\rightarrow 1$. If $T\rightarrow 0$, we can leave only the expression for $m=1$ and $n=k$ on the left-hand side of the condition (\ref{eq:condition2}). If we do it, we get: \begin{equation} \sum_{k=1}^\infty\left[p_k(\mu)p_1(\mu)-p_{k+1}(\mu)p_{0}(\mu)\right]<0, \label{eq:condition3} \end{equation} But for the thermal statistics we have \begin{equation} p_k(\mu)p_1(\mu)-p_{k+1}(\mu)p_{0}(\mu)=0 \end{equation} for every $k$. This means that equipping Bob's detectors with the ability to resolve the number of incoming photons does not have any effect on the function $\mu_\mathrm{max}(T)$ when $T\rightarrow 0$.
\begin{figure}
\caption{(color online) Maximal values of $\mu$ for which it is possible to generate secure key as a function of channel transmission $T$ plotted for the case of Alice and Bob using six-state protocol when the channel noise has thermal statistics and the detectors used by Bob have photon-number-resolving ability (dashed blue line) or do not have it (solid red line). Analogous results for the Poissonian type of noise are plotted with dashed green line (for detectors with photon number resolution) and solid orange line (for detectors without photon number resolution).}
\label{fig:detectionmodels}
\end{figure}
The situation for $T\rightarrow 1$ is more complicated. In this case we can leave on the left-hand side of inequality (\ref{eq:condition2}) only the expression with lowest possible power of $(1-T)$, that is for $k=1$. Then we have \begin{equation} \sum_{n,m=1}^\infty\left[nmp_n(\mu)p_m(\mu)-{n+1 \choose 2}p_{n+1}(\mu)p_{m-1}(\mu)\right]<0 \label{eq:condition4}. \end{equation} For thermal statistics of the source of noise this condition becomes \begin{equation} \sum_{n,m=1}^\infty \frac{\mu^{n+m}}{(\mu+1)^{n+m+2}}n\left[m-\frac{n+1}{2}\right]<0. \label{eq:inequality5} \end{equation} A good method to prove that the left hand side of this inequality is equal to zero is to show that for any $c$ the term standing beside $\mu^c/(\mu+1)^{(c+2)}$, which can be actually written as \begin{equation} \sum_{n=1}^{c-1}n\left[c-n-\frac{n+1}{2}\right], \end{equation} is equal to zero. This can be done by induction.
On the other hand for Poisson statistics of the sources of noise, inequality (\ref{eq:condition2}) would transform into \begin{equation} \sum_{k=1}^\infty e^{-2\mu}\mu^{k+1}\left[\frac{1}{k!}-\frac{1}{(k+1)!}\right]<0 \end{equation} for the case of $T\rightarrow 0$ or into \begin{equation} \sum_{n,m=1}^\infty e^{-2\mu}\frac{\mu^{n+m}}{2(n-1)!(m-1)!}<0 \end{equation} for the case of $T\rightarrow 1$. It is not difficult to see, that the left-hand sides of both these inequalities are actually larger than zero, which means that if the sources of noise in our DV QKD scheme had Poisson statistics, from the point of its resilience to noise it would be better for Bob to use simple on/off detectors instead of photon-number-resolving ones.
The conclusions which can be drawn from the above analysis can be confirmed in Fig.\,\ref{fig:detectionmodels}, where we present the results of the numerical calculations of the function $\mu_\mathrm{max}^\mathrm{DV}(T)$ for the cases of thermal and Poisson statistics of the source of noise both in the situation when Bob uses detectors with and without the ability to resolve the number of photons entering them.
In fact, it is quite easy to intuitively explain why using photon-number-resolving detectors by Bob does not seem to improve the security of our DV QKD scheme over the case of on/off detectors. The basic reason for this is that detectors with photon-number resolution exclude from the key not only all the situations in which more than one photon comes to the wrong detector (which is obviously good for the security), but also all the cases when more than one photon arrives in the right detector (which is obviously bad). So although using photon-number-resolving detectors reduces the number of errors in the key, QBER given by the formula (\ref{eq:QII}) can actually increase due to even greater reduction of $p_{exp}$ at the same time.
\end{document}
|
arXiv
|
{
"id": "1602.03122.tex",
"language_detection_score": 0.8583811521530151,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{On the exterior Dirichlet problem for Hessian quotient equations\footnotemark[1]}
\author{Dongsheng Li \footnotemark[3] \and Zhisu Li (\Letter)\footnotemark[2]~\footnotemark[3]}
\renewcommand{\fnsymbol{footnote}}{\fnsymbol{footnote}}
\footnotetext[1]{This research is supported by NSFC.11671316.} \footnotetext[2]{Corresponding author.} \footnotetext[3]{School of Mathematics and Statistics, Xi'an Jiaotong University, Xi'an 710049, China;\\ Dongsheng Li: \href{mailto:[email protected]}{[email protected]}; Zhisu Li: \href{mailto:[email protected]}{[email protected]}.}
\date{}
\maketitle
\begin{abstract} In this paper, we establish the existence and uniqueness theorem for solutions of the exterior Dirichlet problem for Hessian quotient equations with prescribed asymptotic behavior at infinity. This extends the previous related results on the Monge-Amp\`{e}re equations and on the Hessian equations, and rearranges them in a systematic way. Based on the Perron's method, the main ingredient of this paper is to construct some appropriate subsolutions of the Hessian quotient equation, which is realized by introducing some new quantities about the elementary symmetric functions and using them to analyze the corresponding ordinary differential equation related to the generalized radially symmetric subsolutions of the original equation. \end{abstract}
\noindent\emph{Keywords:} {Dirichlet problem, existence and uniqueness, exterior domain, Hessian quotient equation, Perron's method, prescribed asymptotic behavior, viscosity solution}
\noindent\emph{2010 MSC:} {35D40, 35J15, 35J25, 35J60, 35J96}
\section{Introduction}
In this paper, we consider the Dirichlet problem for the Hessian quotient equation \begin{equation}\label{eqn.hqe} \frac{\sigma_k(\lambda(D^2u))}{\sigma_l(\lambda(D^2u))}=1 \end{equation} in the exterior domain $\mathds R^n\setminus\overline{D}$, where $D$ is a bounded domain in $\mathds R^n$, $n\geq3$, $0\leq l<k\leq n$, $\lambda(D^2u)$ denotes the eigenvalue vector $\lambda:=(\lambda_1,\lambda_2,...,\lambda_n)$ of the Hessian matrix $D^2u$ of the function $u$, and \[\sigma_0(\lambda)\equiv1 \quad\text{and}\quad \sigma_j(\lambda):=\sum_{1\leq s_1<s_2<...<s_j\leq n} \lambda_{s_1}\lambda_{s_2}...\lambda_{s_j} ~(\forall 1\leq j\leq n)\] are the elementary symmetric functions of the $n$-vector $\lambda$.
Note that when $l=0$, \eref{eqn.hqe} is the Hessian equation $\sigma_k(\lambda(D^2u))=1$; when $l=0$, $k=1$, it is the Poisson equation $\Delta u=1$, a linear elliptic equation; when $l=0$, $k=n$, it is the famous Monge-Amp\`{e}re equation $\det(D^2u)=1$; and when $l=1$, $k=3$, $n=3$ or $4$, it is the special Lagrangian equation $\sigma_1(\lambda(D^2u))=\sigma_3(\lambda(D^2u))$ in three or four dimension (in three dimension, this is $\det(D^2u)=\Delta u$ indeed) which arises from the special Lagrangian geometry \cite{HL82}.
For linear elliptic equations of second order, there have been much extensive studies on the exterior Dirichlet problem, see \cite{MS60} and the references therein.
For the Monge-Amp\`{e}re equation, a classical theorem of J\"{o}rgens \cite{Jor54}, Calabi \cite{Cal58} and Pogorelov \cite{Pog72} states that any convex classical solution of $\det(D^2u)=1$ in $\mathds R^n$ must be a quadratic polynomial. Related results was also given by \cite{CY86}, \cite{Caf95}, \cite{TW00} and \cite{JX01}.
Caffarelli and Li \cite{CL03} extended the J\"{o}rgens-Calabi-Pogorelov theorem to exterior domains. They proved that if $u$ is a convex viscosity solution of $\det(D^2u)=1$ in the exterior domain $\mathds R^n\setminus\overline{D}$, where $D$ is a bounded domain in $\mathds R^n$, $n\geq3$, then there exist $A\in\mathds R^{n\times n},b\in\mathds R^n$ and $c\in\mathds R$ such that \begin{equation}\label{eqn.abc-n}
\limsup_{|x|\rightarrow+\infty}|x|^{n-2}
\left|u(x)-\left(\frac{1}{2}x^{T}Ax+b^Tx+c\right)\right|<\infty. \end{equation} With such prescribed asymptotic behavior at infinity, they also established an existence and uniqueness theorem for solutions of the Dirichlet problem of the Monge-Amp\`{e}re equation in the exterior domain of $\mathds R^n$, $n\geq3$. See \cite{FMM99}, \cite{FMM00} or \cite{Del92} for similar problems in two dimension.
Recently, J.-G. Bao, H.-G. Li and Y.-Y. Li \cite{BLL14} extended the above existence and uniqueness theorem of the exterior Dirichlet problem in \cite{CL03} for the Monge-Amp\`{e}re equation to the Hessian equation $\sigma_k(\lambda(D^2u))=1$ with $2\leq k\leq n$ and with some appropriate prescribed asymptotic behavior at infinity which is modified from \eref{eqn.abc-n}.
Before them, for the special case that $A=c_0 I$ with $c_0:=({C_n^k})^{-1/k}$ and $C_n^k:={n!}/(k!(n-k)!)$, the exterior Dirichlet problem for the Hessian equation has been investigated by Dai and Bao in \cite{DB11}.
At the same time, Dai \cite{Dai11} proved the existence theorem of the exterior Dirichlet problem for the Hessian quotient equation \eref{eqn.hqe} with $k-l\geq3$, and with the prescribed asymptotic behavior at infinity of the special case that $A=c_\ast I$, that is, \begin{equation}\label{eqn.cc-kl}
\limsup_{|x|\rightarrow+\infty}|x|^{k-l-2}
\left|u(x)-\left(\frac{c_\ast}{2}|x|^2+c\right)\right|<\infty, \end{equation} where \begin{equation}\label{eqn.cst} c_\ast:=\left(\frac{C_n^l}{C_n^k}\right)^{\frac{1}{k-l}} ~\text{with}~C_n^i:=\frac{n!}{i!(n-i)!} ~\text{for}~i=k,l. \end{equation} As they pointed out in \cite{LB14} that the restriction $k-l\geq3$ rules out an important example, the special Lagrangian equation $\det(D^2u)=\Delta u$ in three dimension. Later, \cite{LD12} improve the result in \cite{Dai11} for \eref{eqn.hqe} with $k-l\geq3$ to that for \eref{eqn.hqe} with $0\leq l<k\leq(n+1)/2$.
More recently, Li and Bao \cite{LB14} established the existence theorem of the exterior Dirichlet problem for a class of fully nonlinear elliptic equations related to the eigenvalues of the Hessian which include the Monge-Amp\`{e}re equations, Hessian equations, Hessian quotient equations and the special Lagrangian equations in dimension equal and lager than three, but with the prescribed asymptotic behavior at infinity only in the special case of \eref{eqn.abc-n} that $A=c^\ast I$ with $c^\ast$ some appropriate constant, like \eref{eqn.cc-kl} and \eref{eqn.cst}.
In this paper, we focus our attention on the Hessian quotient equation \eref{eqn.hqe} and establish the existence and uniqueness theorem for the exterior Dirichlet problem of it with prescribed asymptotic behavior at infinity of the type similar to \eref{eqn.abc-n}.
This extends the previous corresponding results on the Monge-Amp\`{e}re equations \cite{CL03} and on the Hessian equations \cite{BLL14} to Hessian quotient equations, and also extends those results on the Hessian quotient equations in \cite{Dai11}, \cite{LD12} and \cite{LB14} to be valid for general prescribed asymptotic behavior condition at infinity.
Since we do not restrict ourselves to the case $k-l\geq3$ or $0\leq l<k\leq(n+1)/2$ only, our theorems also apply to the special Lagrangian equations $\det(D^2u)=\Delta u$ in three dimension and $\sigma_1(\lambda(D^2u))=\sigma_3(\lambda(D^2u))$ in four dimension. Indeed, we will show in our forthcoming paper \cite{LL16} that our method still works very well for the special Lagrangian equations with higher dimension and with general phase.
We would like to remark that, for the interior Dirichlet problems there have been much extensive studies, see for example \cite{CIL92}, \cite{CNS85}, \cite{Ivo85}, \cite{Kry83}, \cite{Urb90}, \cite{Tru90} and \cite{Tru95}; see \cite{BCGJ03} and the references given there for more on the Hessian quotient equations; and for more on the special Lagrangian equations, we refer the reader to \cite{HL82}, \cite{Fu98}, \cite{Yuan02}, \cite{CWY09} and the references therein.
For the reader's convenience, we give the following definitions related to Hessian quotient equation (see also \cite{CIL92}, \cite{CC95}, \cite{CNS85}, \cite{Tru90}, \cite{Tru95} and the references therein).
We say that a function $u\in C^2\left(\mathds R^n\setminus\overline{D}\right)$ is \emph{$k$-convex}, if $\lambda(D^2u)\in\overline{\Gamma_k}$ in $\mathds R^n\setminus\overline{D}$, where $\Gamma_k$ is the connected component of
$\set{\lambda\in\mathds R^n|\sigma_k(\lambda)>0}$ containing the \emph{positive cone}
\[\Gamma^+:=\set{\lambda\in\mathds R^n|\lambda_i>0,\forall i=1,2,...,n}.\] It is well known that $\Gamma_k$ is an open convex symmetric cone with its vertex at the origin and that
\[\Gamma_k=\set{\lambda\in\mathds R^n|\sigma_j(\lambda)>0, \forall j=1,2,...,k},\] which implies
\[\set{\lambda\in\mathds R^n|\lambda_1+\lambda_2+...+\lambda_n>0} =\Gamma_1\supset...\supset\Gamma_k\supset\Gamma_{k+1} \supset...\supset\Gamma_n=\Gamma^+\] with the first term $\Gamma_1$ the half space and with the last term $\Gamma_n$ the positive cone $\Gamma^+$. Furthermore, we also know that \begin{equation}\label{eqn.pisjp} \partial_{\lambda_i}\sigma_{j}(\lambda)>0, ~\forall 1\leq i\leq n, ~\forall 1\leq j\leq k, ~\forall\lambda\in\Gamma_k, ~\forall 1\leq k\leq n \end{equation} (see \cite{CNS85} or \cite{Urb90} for more details).
Let $\Omega$ be an open domain in $\mathds R^n$ and let $f\in C^0(\Omega)$ be nonnegative. Suppose $0\leq l<k\leq n$. A function $u\in C^0(\Omega)$ is said to be a \emph{viscosity subsolution} of \begin{equation}\label{eqn.hqe-f} \frac{\sigma_k(\lambda(D^2u))}{\sigma_l(\lambda(D^2u))}=f \quad\text{in}~\Omega \end{equation} \big{(}or say that $u$ satisfies \[\frac{\sigma_k(\lambda(D^2u))}{\sigma_l(\lambda(D^2u))}\geq f \quad\text{in}~\Omega\] in the viscosity sense, similarly hereinafter\big{)}, if for any function $v\in C^2(\Omega)$ and any point $x^\ast\in\Omega$ satisfying \[v(x)\geq u(x),~\forall x\in\Omega \quad\text{and}\quad v(x^\ast)=u(x^\ast),\] we have \[\frac{\sigma_k(\lambda(D^2v))}{\sigma_l(\lambda(D^2v))}\geq f \quad\text{in}~\Omega.\] A function $u\in C^0(\Omega)$ is said to be a \emph{viscosity supersolution} of \eref{eqn.hqe-f} if for any \textit{$k$-convex function} $v\in C^2(\Omega)$ and any point $x^\ast\in\Omega$ satisfying \[v(x)\leq u(x),~\forall x\in\Omega \quad\text{and}\quad v(x^\ast)=u(x^\ast),\] we have \[\frac{\sigma_k(\lambda(D^2v))}{\sigma_l(\lambda(D^2v))}\leq f \quad\text{in}~\Omega.\] A function $u\in C^0(\Omega)$ is said to be a \emph{viscosity solution} of \eref{eqn.hqe-f}, if it is both a viscosity subsolution and a viscosity supersolution of \eref{eqn.hqe-f}.
A function $u\in C^0(\Omega)$ is said to be a \emph{viscosity subsolution} (respectively, \emph{supersolution, solution}) of \eref{eqn.hqe-f} and $u=\varphi$ on $\partial\Omega$ with some $\varphi\in C^0(\partial\Omega)$, if $u$ is a viscosity subsolution (respectively, supersolution, solution) of \eref{eqn.hqe-f} and $u\leq$(respectively, $\geq,=$)$\varphi$ on $\partial\Omega$.
Note that in the definitions of viscosity solution above, we have used the ellipticity of the Hessian quotient equations indeed. For completeness and convenience, this will be proved in the end of \ssref{subsec.prelem}. See also \cite{CC95}, \cite{CIL92}, \cite{CNS85}, \cite{Urb90} and the references therein.
Define
\[\mathscr{A}_{k,l}:=\left\{A\in S(n)\big{|}\lambda(A)\in\Gamma^+, \sigma_k(\lambda(A))=\sigma_l(\lambda(A))\right\}.\] Note that there are plenty of elements in $\mathscr{A}_{k,l}$. In fact, for any $A\in S(n)$ with $\lambda(A)\in\Gamma^+$, if we set \[\varrho:=\left(\frac{\sigma_k(\lambda(A))}{\sigma_l(\lambda(A))}\right) ^{-\frac{1}{k-l}},\] we then have $\varrho A\in\mathscr{A}_{k,l}$. Let \[\mathscr{\widetilde A}_{k,l}
:=\set{A\in\mathscr{A}_{k,l}\big{|}m_{k,l}(\lambda(A))>2},\] where $m_{k,l}(\lambda)$ is a quantity which plays an important role in this paper. We will give the specific definition of $m_{k,l}(\lambda)$ in \eref{eqn.mkla} in \ssref{subsec.xi}, and verify there that $\mathscr{\widetilde A}_{k,l}$ possesses the following fine properties. \begin{proposition}\label{prop.wtakl} Suppose $0\leq l<k\leq n$ and $n\geq3$. \begin{enumerate}[\quad(1)] \item If $k-l\geq2$, then $\mathscr{\widetilde A}_{k,l}=\mathscr{A}_{k,l}$.
\item $\mathscr{\widetilde A}_{n,0}=\mathscr{A}_{n,0}$ and $m_{n,0}\equiv n$.
\item $c_\ast I\in\mathscr{\widetilde A}_{k,l}$ and $m_{k,l}(c_\ast(1,1,...,1))=n$, where $c_\ast$ is the one defined in \eref{eqn.cst}. \end{enumerate} \end{proposition}
The main result of this paper now can be stated as below. \begin{theorem}\label{thm.hqe} Let $D$ be a bounded strictly convex domain in $\mathds R^n$, $n\geq 3$, $\partial D\in C^2$ and let $\varphi\in C^2(\partial D)$. Then for any given $A\in\mathscr{\widetilde A}_{k,l}$ with $0\leq l<k\leq n$, and any given $b\in \mathds R^n$, there exists a constant $\tilde{c}$ depending only on $n,D,k,l,A,b$ and $\norm{\varphi}_{C^2(\partial D)}$, such that for every $c\geq\tilde{c}$, there exists a unique viscosity solution $u\in C^0(\mathds R^n\setminus D)$ of \begin{equation}\label{eqn.hqe-abc} \left\{ \begin{aligned} \displaystyle \frac{\sigma_k(\lambda(D^2u))}{\sigma_l(\lambda(D^2u))}&=1 \quad\text{in}~\mathds R^n\setminus\overline{D},\\%[0.05cm] \displaystyle u&=\varphi\quad\text{on}~\partial D,\\%[0.05cm]
\displaystyle \limsup_{|x|\rightarrow+\infty}|x|^{m-2}
&\left|u(x)-\left(\frac{1}{2}x^{T}Ax+b^Tx+c\right)\right|<\infty, \end{aligned} \right. \end{equation} where $m\in(2,n]$ is a constant depending only on $n,k,l$ and $\lambda(A)$, which actually can be taken as $m_{k,l}(\lambda(A))$. \end{theorem}
\begin{remark} \begin{enumerate}[(1)] \item One can easily see that \tref{thm.hqe} still holds with $A\in\mathscr{\widetilde A}_{k,l}$ replaced by $A\in\mathscr{\widetilde A}^\ast_{k,l}$ and $\lambda(A)$ replaced by $\lambda(A^\ast)$, where \[\mathscr{\widetilde A}^\ast_{k,l}
:=\left\{A\in\mathds R^{n\times n}\big{|} \lambda(A^\ast)\in\Gamma^+, \sigma_k(\lambda(A^\ast)) =\sigma_l(\lambda(A^\ast)), m_{k,l}(\lambda(A^\ast))>2\right\}\] and $A^\ast:=(A+A^T)/2$. This is to say that the above theorem can be adapted to a slightly more general form by modifying the meaning of $\mathscr{\widetilde A}_{k,l}$.
\item For the special cases that $l=0$ (i.e., the Hessian equation $\sigma_k(\lambda(D^2u))=1$) and that $l=0$ and $k=n$ (i.e., the Monge-Amp\`{e}re equation $\det(D^2u)=1$), in view of \pref{prop.wtakl}-\textsl{(2)}, our \tref{thm.hqe} recovers the corresponding results \cite[Theorem 1.1]{BLL14} and \cite[Theorem 1.5]{CL03}, respectively.
\item For $A=c_\ast I$ with $c_\ast$ defined in \eref{eqn.cst}, by \pref{prop.wtakl}-\textsl{(1),(3)}, our results improve those in \cite{Dai11} and \cite{LD12}. Indeed, by \pref{prop.wtakl}-\textsl{(3)}, the main results in \cite{Dai11}, \cite{LD12} and those parts concerning the Hessian quotient equations in \cite{LB14} can all be recovered by \tref{thm.hqe} as special cases. Furthermore, our results also apply to the special Lagrangian equation $\det(D^2u)=\Delta u$ in three dimension (respectively, $\sigma_1(\lambda(D^2u))=\sigma_3(\lambda(D^2u))$ in four dimension), not only for $A=\sqrt{3}I$ (respectively, $A=I$), but also for any $A\in\mathscr{A}_{3,1}$. \qed \end{enumerate} \end{remark}
The paper is organized as follows. In \sref{sec.prel}, after giving some basic notations in \ssref{subsec.nott}, we introduce the definitions of $\Xi_k,\underline\xi_k,\overline\xi_k$ and $m_{k,l}$, and investigate their properties in \ssref{subsec.xi}. Then we collect in \ssref{subsec.prelem} some preliminary lemmas which will be used in this paper.
\sref{sec.pmaint} is devoted to the proof of the main theorem (\tref{thm.hqe}). To do this, we start in \ssref{subsec.csubsol} to construct some appropriate subsolutions of the Hessian quotient equation \eref{eqn.hqe}, by taking advantages of the properties of $\Xi_k,\underline\xi_k,\overline\xi_k$ and $m_{k,l}$ explored in \ssref{subsec.xi}. Then in \ssref{subsec.pmaint}, after reducing \tref{thm.hqe} to \lref{lem.hqe} by simplification and normalization, we prove \lref{lem.hqe} by applying the Perron's method to the subsolutions we constructed in \ssref{subsec.csubsol}.
\section{Preliminary}\label{sec.prel}
\subsection{Notation}\label{subsec.nott} \quad
In this paper, $S(n)$ denotes the linear space of symmetric $n\times n$ real matrices, and $I$ denotes the identity matrix.
For any $M\in S(n)$, if $m_1,m_2,...,m_n$ are the eigenvalues of $M$ (usually, the assumption $m_1\leq m_2\leq...\leq m_n$ is added for convenience), we will denote this fact briefly by $\lambda(M)=(m_1,m_2,...,m_n)$ and call $\lambda(M)$ the eigenvalue vector of $M$.
For $A\in S(n)$ and $\rho>0$, we denote by \[
E_\rho:=\left\{x\in\mathds R^n\big{|}x^TAx<\rho^2\right\}
=\left\{x\in\mathds R^n\big{|}r_A(x)<\rho\right\} \] the ellipsoid of size $\rho$ with respect to $A$, where we set $r_A(x):=\sqrt{x^TAx}$.
For any $p\in\mathds R^n$, we write \[\sigma_k(p):=\sum_{1\leq s_1<s_2<...<s_k\leq n} p_{s_1}p_{s_2}...p_{s_k}\quad(\forall 1\leq k\leq n)\] as the $k$-th elementary symmetric function of $p$. Meanwhile, we will adopt the conventions that $\sigma_{-1}(p)\equiv0$, $\sigma_0(p)\equiv1$ and $\sigma_k(p)\equiv0$, $\forall k\geq n+1$; and we will also define \[\sigma_{k;i}(p):=\left(\sigma_{k}(\lambda)
\big{|}_{\lambda_i=0}\right)\Big{|}_{\lambda=p} =\sigma_{k}\left(p_1,p_2,...,\widehat{p_i},...,p_n\right)\] for any $-1\leq k\leq n$ and any $1\leq i\leq n$, and similarly \[\sigma_{k;i,j}(p):=\left(\sigma_{k}(\lambda)
\big{|}_{\lambda_i=\lambda_j=0}\right)\Big{|}_{\lambda=p} =\sigma_{k}\left(p_1,p_2,...,\widehat{p_i}, ...,\widehat{p_j},...,p_n\right)\] for any $-1\leq k\leq n$ and any $1\leq i,j\leq n$, $i\neq j$, for convenience.
\subsection{Definitions and properties of $\Xi_k,\underline\xi_k,\overline\xi_k$ and $m_{k,l}$}\label{subsec.xi} \quad
To establish the existence of the solution of \eref{eqn.hqe}, by the Perron's method, the key point is to find some appropriate subsolutions of the equation. Since the Hessian quotient equation \eref{eqn.hqe} is a highly fully nonlinear equation which including polynomials of the eigenvalues of the the Hessian matrix $D^2u$, $\sigma_k(\lambda)$ and $\sigma_l(\lambda)$, of different order of homogeneities, to solve it we need to strike a balance between them. It will turn out to be clear that the quantities $\Xi_k,\underline\xi_k,\overline\xi_k$ and $m_{k,l}$, which we shall introduce below, are very natural and perfectly fit for this purpose.
\begin{definition} For any $0\leq k\leq n$ and any $a\in\mathds R^n\setminus\{0\}$, let \[\Xi_k:=\Xi_k(a,x) :=\frac{\sum_{i=1}^n\sigma_{k-1;i}(a)a_i^2x_i^2} {\sigma_k(a)\sum_{i=1}^{n}a_ix_i^2}, ~\forall x\in\mathds R^n\setminus\{0\},\] and define \[\overline{\xi}_k:=\overline{\xi}_k(a) :=\sup_{x\in\mathds R^n\setminus\{0\}}\Xi_k(a,x)\] and \[\underline{\xi}_k:=\underline{\xi}_k(a) :=\inf_{x\in\mathds R^n\setminus\{0\}}\Xi_k(a,x).\] \end{definition}
\begin{definition} For any $0\leq l<k\leq n$ and any $a\in\mathds R^n\setminus\{0\}$, let \begin{equation}\label{eqn.mkla} \displaystyle m_{k,l}:=m_{k,l}(a):=\frac{k-l}{\overline{\xi}_k(a)-\underline{\xi}_l(a)}. \end{equation} \end{definition}
We remark, for the reader's convenience, that $\Xi_k$ originates from the computation of $\sigma_k(D^2\Phi(x))$ where $\Phi(x)$ is a generalized radially symmetric function (see \lref{lem.skm} and the proof of \lref{lem.Phi-subsol}),
that $\underline\xi_k$ and $\overline\xi_k$ result from the comparison between $\sigma_k(\lambda)$ and $\sigma_l(\lambda)$ in the attempt to derive an ordinary differential equation from the original equation (see the last part of the proof of \lref{lem.Phi-subsol}),
and that $m_{k,l}$ arises in the process of solving this ordinary differential equation (see \eref{eqn.mdlnr} in the proof of \lref{lem.psi}).
By $\Xi_k,\underline\xi_k$ and $\overline\xi_k$, we get a good balance between $\sigma_k(\lambda)$ and $\sigma_l(\lambda)$, which can be measured by $m_{k,l}$. Furthermore, we will find that $m_{k,l}$ has also some special meaning related to the decay and asymptotic behavior of the solution (see \lref{lem.psi}\textsl{-(iii)}, \cref{cor.mub} and \tref{thm.hqe}).
It is easy to see that \[\overline{\xi}_k(\varrho a)=\overline{\xi}_k(a), ~\underline{\xi}_k(\varrho a)=\underline{\xi}_k(a), ~\forall\varrho\neq 0, ~\forall a\in\mathds R^n\setminus\{0\}, ~\forall 0\leq k\leq n,\] and \[\underline\xi_k(C(1,1,...,1))=\frac{k}{n} =\overline\xi_k(C(1,1,...,1)), ~\forall C>0,~\forall 0\leq k\leq n.\]
Furthermore, we have the following lemma. \begin{lemma}\label{lem.xik} Suppose $a=(a_1,a_2,...,a_n)$ with $0<a_1\leq a_2\leq...\leq a_n$. Then \begin{equation}\label{eqn.uxknox} 0<\frac{a_1\sigma_{k-1;1}(a)}{\sigma_k(a)}=\underline{\xi}_k(a) \leq\frac{k}{n}\leq\overline{\xi}_k(a) =\frac{a_n\sigma_{k-1;n}(a)}{\sigma_k(a)}\leq 1, ~\forall 1\leq k\leq n; \end{equation} \begin{equation}\label{eqn.olxiin} 0=\overline{\xi}_0(a)<\frac{1}{n}\leq\frac{a_n}{\sigma_1(a)} =\overline{\xi}_1(a)\leq\overline{\xi}_2(a) \leq...\leq\overline{\xi}_{n-1}(a)<\overline{\xi}_n(a)=1; \end{equation} and \begin{equation}\label{eqn.ulxiin} 0=\underline{\xi}_0(a)<\frac{a_1}{\sigma_1(a)} =\underline{\xi}_1(a)\leq\underline{\xi}_2(a) \leq...\leq\underline{\xi}_{n-1}(a)<\underline{\xi}_n(a)=1. \end{equation}
Moreover, \begin{equation}\label{eqn.xkkn} \underline\xi_k(a)=\frac{k}{n}=\overline\xi_k(a) \end{equation} for some $1\leq k\leq n-1$, if and only if $a=C(1,1,...,1)$ for some $C>0$. \end{lemma}
\begin{proof} ($1^\circ$) By the definitions of $\sigma_k(a)$ and $\sigma_{k;i}(a)$, we see that \begin{equation}\label{eqn.sk} \sigma_k(a)=\sigma_{k;i}(a)+a_i\sigma_{k-1;i}(a), ~\forall 1\leq i\leq n; \end{equation} and \[\sum_{i=1}^{n}\sigma_{k;i}(a) =\frac{nC_{n-1}^k}{C_n^k}\sigma_k(a) =(n-k)\sigma_k(a).\] Hence we obtain \begin{equation}\label{eqn.ksk} \sum_{i=1}^{n}a_i\sigma_{k-1;i}(a)=k\sigma_k(a). \end{equation}
Now we show that \begin{equation}\label{eqn.aiski} a_1\sigma_{k-1;1}(a) \leq a_2\sigma_{k-1;2}(a)\leq... \leq a_n\sigma_{k-1;n}(a). \end{equation} In fact, for any $i\neq j$, similar to \eref{eqn.sk}, we have \[a_i\sigma_{k-1;i}(a) =a_i\left(\sigma_{k-1;i,j}(a)+a_j\sigma_{k-2;i,j}(a)\right)\] and \[a_j\sigma_{k-1;j}(a) =a_j\left(\sigma_{k-1;i,j}(a)+a_i\sigma_{k-2;i,j}(a)\right),\] thus \[a_i\sigma_{k-1;i}(a)-a_j\sigma_{k-1;j}(a) =(a_i-a_j)\sigma_{k-1;i,j}(a).\] Hence if $a_i\lessgtr a_j$, then \begin{equation}\label{eqn.aslgeq} a_i\sigma_{k-1;i}(a)\lessgtr a_j\sigma_{k-1;j}(a). \end{equation}
By the definition of $\overline\xi_k$, we have \begin{eqnarray*} \displaystyle \overline{\xi}_k(a) &=&\sup_{x\neq 0}\frac{\sum_{i=1}^n\sigma_{k-1;i}(a)a_i^2x_i^2} {\sigma_k(a)\sum_{i=1}^{n}a_ix_i^2}\\ \displaystyle &\geq& \sup_{\substack{x_1=...=x_{n-1}=0,\\x_n\neq 0}} \frac{\sum_{i=1}^n\sigma_{k-1;i}(a)a_i^2x_i^2} {\sigma_k(a)\sum_{i=1}^{n}a_ix_i^2}\\ \displaystyle &=&\sup_{x_n\neq 0}\frac{\sigma_{k-1;n}(a)a_n^2x_n^2} {\sigma_k(a)a_n x_n^2}\\ \displaystyle &=&\frac{a_n\sigma_{k-1;n}(a)}{\sigma_k(a)} \end{eqnarray*} and \begin{eqnarray*} \displaystyle \overline{\xi}_k(a) &=&\sup_{x\neq 0}\frac{\sum_{i=1}^n\sigma_{k-1;i}(a)a_i^2x_i^2} {\sigma_k(a)\sum_{i=1}^{n}a_ix_i^2}\\ \displaystyle &\leq& \sup_{x\neq 0}\frac{a_n\sigma_{k-1;n}(a)\sum_{i=1}^na_ix_i^2} {\sigma_k(a)\sum_{i=1}^{n}a_ix_i^2}\qquad\text{by \eref{eqn.aiski}}\\ \displaystyle &=&\frac{a_n\sigma_{k-1;n}(a)}{\sigma_k(a)}. \end{eqnarray*} Hence we obtain \begin{equation}\label{eqn.olxik} \overline\xi_k(a)=\frac{a_n\sigma_{k-1;n}(a)}{\sigma_k(a)}. \end{equation} Similarly \begin{equation}\label{eqn.ulxik} \underline\xi_k(a)=\frac{a_1\sigma_{k-1;1}(a)}{\sigma_k(a)}. \end{equation}
From \eref{eqn.ksk}, we have \[\sum_{i=1}^{n}\frac{a_i\sigma_{k-1;i}(a)}{\sigma_k(a)}=k.\] Combining this with \eref{eqn.aiski}, \eref{eqn.olxik} and \eref{eqn.ulxik}, we deduce that \[\underline\xi_k(a)\leq\frac{k}{n}\leq\overline\xi_k(a).\] Thus the proof of \eref{eqn.uxknox} is complete, and \eref{eqn.xkkn} is also clear in view of \eref{eqn.aslgeq}.
($2^\circ$) Since it follows from \eref{eqn.sk} that \[a_i\sigma_{k-1;i}(a)<\sigma_k(a), ~\forall 1\leq i\leq n, ~\forall 1\leq k\leq n-1,\] we obtain \[\underline\xi_k(a)\leq\overline\xi_k(a)<1,~\forall 0\leq k\leq n-1.\]
On the other hand, we have $\overline{\xi}_n(a)=\underline{\xi}_n(a)=1$ which follows from \[a_i\sigma_{n-1;i}(a)=\sigma_n(a),~\forall 1\leq i\leq n.\]
Combining \eref{eqn.olxik} and \eref{eqn.sk}, we discover that \begin{eqnarray*} \overline\xi_k(a)&=&\frac{a_n\sigma_{k-1;n}(a)}{\sigma_k(a)} =\frac{a_n\sigma_{k-1;n}(a)}{\sigma_{k;n}(a)+a_n\sigma_{k-1;n}(a)}\\ &\leq&\frac{a_n\sigma_{k;n}(a)}{\sigma_{k+1;n}(a)+a_n\sigma_{k;n}(a)} =\frac{a_n\sigma_{k;n}(a)}{\sigma_{k+1}(a)} =\overline\xi_{k+1}(a), \end{eqnarray*} where we used the inequality \[\frac{\sigma_{k-1;n}(a)}{\sigma_{k;n}(a)} \leq\frac{\sigma_{k;n}(a)}{\sigma_{k+1;n}(a)}\] which is a variation of the famous Newton inequality(see \cite{HLP34}) \[\sigma_{k-1}(\lambda)\sigma_{k+1}(\lambda) \leq\left(\sigma_{k}(\lambda)\right)^2, ~\forall\lambda\in\mathds R^n.\] Thus the proof of \eref{eqn.olxiin}, and similarly of \eref{eqn.ulxiin}, is complete. \end{proof}
Since it follows from \eref{eqn.uxknox} that \[\frac{k-l}{n}\leq\overline{\xi}_k(a)-\underline{\xi}_l(a) <\overline{\xi}_k(a)\leq 1,\] we obtain \begin{corollary}\label{cor.m} If $0\leq l<k\leq n$ and $a\in\Gamma^+$, then \[1\leq k-l<m_{k,l}(a)\overline\xi_k(a)\leq m_{k,l}(a)\leq n.\] \end{corollary}
As an application of \cref{cor.m} and \lref{lem.xik}, we now verify \pref{prop.wtakl}. \begin{proof}[\textbf{Proof of \pref{prop.wtakl}}] \textsl{(1)} and \textsl{(2)} are clear. For \textsl{(3)}, we only need to note that $c_\ast I\in\mathscr{A}_{k,l}$ and $m_{k,l}(c_\ast(1,1,...,1))=n>2$. \end{proof}
To help the reader to become familiar with these new quantities, it is worth to give the following examples which are also the applications of the above lemma. \begin{example} Note that, for $a=(a_1,a_2,a_3)\in\mathds R^3$ with $0<a_1\leq a_2\leq a_3$, by \lref{lem.xik}, we have \[\overline\xi_3(a)\equiv1\equiv\underline\xi_3(a),\] \[\overline\xi_2(a)=\frac{a_3(a_1+a_2)}{a_1a_2+a_1a_3+a_2a_3}, \quad \underline\xi_2(a)=\frac{a_1(a_2+a_3)}{a_1a_2+a_1a_3+a_2a_3},\] \[\overline\xi_1(a)=\frac{a_3}{a_1+a_2+a_3}, \quad \underline\xi_1(a)=\frac{a_1}{a_1+a_2+a_3},\] and \[\overline\xi_0(a)\equiv0\equiv\underline\xi_0(a).\]
Thus we can compute, for $a=(1,2,3)$, that \[\overline\xi_2=\frac{9}{11},~\underline\xi_2=\frac{5}{11}, ~\overline\xi_1=\frac{1}{2},~\underline\xi_1=\frac{1}{6},\] \[m_{3,2}=\frac{11}{6}<2, ~m_{3,1}=\frac{12}{5}>2, ~m_{3,0}\equiv3>2,\] \[m_{2,1}=\frac{66}{43}<2, ~m_{2,0}=\frac{22}{9}>2 ~\text{and}~m_{1,0}=2,\]
and, for $a=(11,12,13)$, that \[\overline\xi_2=\frac{299}{431},~\underline\xi_2=\frac{275}{431}, ~\overline\xi_1=\frac{13}{36},~\underline\xi_1=\frac{11}{36},\] \[m_{3,2}=\frac{431}{156}>2, ~m_{3,1}=\frac{72}{25}>2, ~m_{3,0}\equiv3>2,\] \[m_{2,1}=\frac{15516}{6023}>2, ~m_{2,0}=\frac{862}{299}>2 ~\text{and}~m_{1,0}=\frac{36}{13}>2.\] \end{example}
\begin{remark} \begin{enumerate}[(1)] \item By definition of $m_{k,l}$, we can easily check that for any $1<k\leq n$, $m_{k,k-1}(a)>2$ if and only if $\underline\xi_{k-1}(a)\leq\overline\xi_{k}(a)\leq\underline\xi_{k-1}(a)+1/2$. This will show us how $m_{k,l}$ plays a role in the making of a balance between different order of homogeneities as we stated in the beginning of this subsection.
\item \pref{prop.wtakl}-\textsl{(1)} states that $\mathscr{\widetilde A}_{k,l}=\mathscr{A}_{k,l}$ provided $k-l\geq2$. Note that this is the best case we can expect, since in general $\mathscr{\widetilde A}_{k,k-1}\subsetneqq\mathscr{A}_{k,k-1}$, which is evident by the fact stated in the first item of this remark (and also by the above examples). For example, in $\mathds R^3$ we have \[m_{3,2}(a)>2 \Leftrightarrow \overline\xi_{3}(a)\leq\underline\xi_{2}(a)+1/2 \Leftrightarrow a_1>\frac{a_2a_3}{a_2+a_3},\] where the last inequality is not always true.\qed \end{enumerate} \end{remark}
\subsection{Some preliminary lemmas}\label{subsec.prelem} \quad
In this subsection, we collect some preliminary lemmas which will be mainly used in \sref{sec.pmaint}.
We first give a lemma to compute $\sigma_k(\lambda(M))$ with $M$ of certain type. ~If $\Phi(x):=\phi(r)$ with $\phi\in C^2$, $r=\sqrt{x^TAx}$, $A\in S(n)\cap\Gamma^+$ and $a=\lambda(A)$ (we may call $\Phi$ a \emph{generalized radially symmetric function} with respect to $A$, according to \cite{BLL14}), one can conclude that \[\partial_{ij}\Phi(x) =\frac{\phi'(r)}{r}a_i\delta_{ij}+ \frac{\phi''(r)-\frac{\phi'(r)}{r}}{r^2}(a_ix_i)(a_jx_j), ~\forall 1\leq i,j\leq n,\] provided $A$ is normalized to a diagonal matrix (see the first part of \ssref{subsec.pmaint} and the proof of \lref{lem.Phi-subsol} for details). As far as we know, generally there is no explicit formula for $\lambda(D^2\Phi(x))$ of this type, but luckily we have a method to calculate $\sigma_k\left(\lambda(D^2\Phi(x))\right)$ for each $1\leq k\leq n$, which can be presented as follows. \begin{lemma}\label{lem.skm} If $M=\left(p_i\delta_{ij}+s q_iq_j\right)_{n\times n}$ with $p,q\in\mathds R^n$ and $s\in\mathds R$, then \[\sigma_k\left(\lambda(M)\right) =\sigma_k(p)+s\sum_{i=1}^n\sigma_{k-1;i}(p)q_i^2, ~\forall 1\leq k\leq n.\] \end{lemma} \begin{proof} See \cite{BLL14}. \end{proof}
To process information on the boundary we need the following lemma. \begin{lemma}\label{lem.Qxi} Let $D$ be a bounded strictly convex domain of $\mathds R^n$, $n\geq 2$, $\partial D\in C^2$, $\varphi\in C^0(D)\cap C^2(\partial{D})$ and let $A\in S(n)$, $\det{A}\neq0$. Then there exists a constant $K>0$ depending only on $n$, $\mbox{\emph{diam}}\,D$, the convexity of $D$, $\norm{\varphi}_{C^2(\overline{D})}$, the $C^2$ norm of $\partial D$ and the upper bound of $A$, such that for any $\xi\in\partial D$, there exists $\bar{x}(\xi)\in\mathds R^n$ satisfying
\[\left|\bar{x}(\xi)\right|\leq K \quad \mbox{and} \quad Q_\xi(x)<\varphi(x), ~\forall x\in \overline{D}\setminus\{\xi\},\] where \[Q_\xi(x):=\frac{1}{2}\left(x-\bar{x}(\xi)\right)^TA\left(x-\bar{x}(\xi)\right) -\frac{1}{2}\left(\xi-\bar{x}(\xi)\right)^TA\left(\xi-\bar{x}(\xi)\right) +\varphi(\xi),~\forall x\in\mathds R^n.\] \end{lemma} \begin{proof} See \cite{CL03} or \cite{BLL14}. \end{proof}
\begin{remark}\label{rmk.Qxi} It is easy to check that $Q_\xi$ satisfy the following properties. \begin{enumerate}[\quad(1)] \item $Q_\xi\leq\varphi$ on $\overline{D}$ and $Q_\xi(\xi)=\varphi(\xi)$.
\item If $A\in\mathscr{A}_{k,l}$, then \[\frac{\sigma_k(\lambda(D^2Q_\xi))}{\sigma_l(\lambda(D^2Q_\xi))}=1 \quad\mbox{in}~\mathds R^n.\]
\item There exists $\bar{c}=\bar{c}(D,A,K)>0$ such that \[Q_\xi(x)\leq\frac{1}{2}x^TAx+\bar{c}, \quad \forall x\in\partial D,~\forall\xi\in\partial D.\] \end{enumerate} \end{remark}
Now we introduce the following well known lemmas about the comparison principle and Perron's method which will be applied to the Hessian quotient equations but stated in a slightly more general setting. These lemmas are adaptions of those appeared in \cite{CNS85} \cite{Jen88} \cite{Ish89} \cite{Urb90} and \cite{CIL92}. For specific proof of them one may also consult \cite{BLL14} and \cite{LB14}.
\begin{lemma}[Comparison principle]\label{lem.cp} Assume $\Gamma^+\subset\Gamma\subset\mathds R^n$ is an open convex symmetric cone with its vertex at the origin, and suppose $f\in C^1(\Gamma)$ and $f_{\lambda_i}(\lambda)>0$, $\forall \lambda\in\Gamma$, $\forall i=1,2,...,n$. Let $\Omega\subset\mathds R^n$ be a domain and let $\underline{u},\overline{u}\in C^0(\overline\Omega)$ satisfying \[f\left(\lambda\left(D^2\underline{u}\right)\right)\geq 1 \geq f\left(\lambda\left(D^2\overline{u}\right)\right)\] in $\Omega$ in the viscosity sense. Suppose $\underline{u}\leq \overline{u}$ on $\partial\Omega$ (and additionally
\[\lim_{|x|\rightarrow+\infty}\left(\underline{u}-\overline{u}\right)(x)=0\] provided $\Omega$ is unbounded). Then $\underline{u}\leq \overline{u}$ in $\Omega$. \end{lemma}
\begin{lemma}[Perron's method]\label{lem.pm} Assume that $\Gamma^+\subset\Gamma\subset\mathds R^n$ is an open convex symmetric cone with its vertex at the origin, and suppose $f\in C^1(\Gamma)$ and $f_{\lambda_i}(\lambda)>0$, $\forall \lambda\in\Gamma$, $\forall i=1,2,...,n$. Let $\Omega\subset\mathds R^n$ be a domain, $\varphi\in C^0(\partial\Omega)$ and let $\underline{u},\overline{u}\in C^0(\overline\Omega)$ satisfying \[f\left(\lambda\left(D^2\underline{u}\right)\right)\geq 1 \geq f\left(\lambda\left(D^2\overline{u}\right)\right)\] in $\Omega$ in the viscosity sense. Suppose $\underline{u}\leq \overline{u}$ in $\Omega$, $\underline{u}=\varphi$ on $\partial\Omega$ (and additionally
\[\lim_{|x|\rightarrow+\infty}\left(\underline{u}-\overline{u}\right)(x)=0\] provided $\Omega$ is unbounded). Then \begin{eqnarray*}
u(x)&:=&\sup\Big\{v(x)\big{|}v\in C^0(\Omega),~ \underline{u}\leq v\leq \overline{u}~\mbox{in}~\Omega,~ f\left(\lambda\left(D^2v\right)\right)\geq 1~\mbox{in}~\Omega\\ &~&\qquad\mbox{in the viscosity sense},~ v=\varphi~\mbox{on}~\partial\Omega\Big\} \end{eqnarray*} is the unique viscosity solution of the Dirichlet problem \[\left\{ \begin{aligned} f\left(\lambda\left(D^2u\right)\right)=1 \qquad &\mbox{in}& \Omega,\\ u=\varphi \qquad &\mbox{on}& \partial\Omega. \end{aligned}\right. \] \end{lemma}
\begin{remark} In order to apply the above lemmas to the Hessian quotient operator \[f(\lambda):=\frac{\sigma_k(\lambda)}{\sigma_l(\lambda)}\] in the cone $\Gamma:=\Gamma_k$, we need to show that \begin{equation}\label{eqn.pskl} \partial_{\lambda_i}\left(\frac{\sigma_k(\lambda)}{\sigma_l(\lambda)}\right) >0,~\forall 1\leq i\leq n, ~\forall 0\leq l<k\leq n, ~\forall\lambda\in\Gamma_k, \end{equation} which indeed indicates that the Hessian quotient equations \eref{eqn.hqe} are elliptic equations with respect to its $k$-convex solution $u$.
Indeed, for $l=0$, \eref{eqn.pskl} is clear in light of \eref{eqn.pisjp}. For $1\leq l<k\leq n$, since \[\partial_{\lambda_i}\sigma_k(\lambda) =\frac{\sigma_k(\lambda)-\sigma_{k;i}(\lambda)}{\lambda_i} =\sigma_{k-1;i}(\lambda)\] according to \eref{eqn.sk}, we have \[\partial_{\lambda_i}\left(\frac{\sigma_k(\lambda)} {\sigma_l(\lambda)}\right) =\frac{\sigma_{k-1;i}(\lambda)\sigma_l(\lambda) -\sigma_k(\lambda)\sigma_{l-1;i}(\lambda)} {(\sigma_l(\lambda))^2}.\] Thus to prove \eref{eqn.pskl}, it remains to verify \[\sigma_{k-1;i}(\lambda)\sigma_l(\lambda) \geq\sigma_k(\lambda)\sigma_{l-1;i}(\lambda).\] In view of \eref{eqn.sk}, this is equivalent to \[\sigma_{k-1;i}(\lambda)\sigma_{l;i}(\lambda) \geq\sigma_{k;i}(\lambda)\sigma_{l-1;i}(\lambda),\] which in turn is equivalent to \[\frac{\sigma_{l;i}(\lambda)}{\sigma_{l-1;i}(\lambda)} \geq\frac{\sigma_{k;i}(\lambda)}{\sigma_{k-1;i}(\lambda)},\] since $\sigma_{j;i}(\lambda)=\partial_{\lambda_i}\sigma_{j+1}(\lambda)>0$, $\forall 1\leq i\leq n$, $\forall 0\leq j\leq k-1$, $\forall\lambda\in\Gamma_k$, according to \eref{eqn.pisjp}. For the proof of the latter, we only need to note that \[\frac{\sigma_{j;i}(\lambda)}{\sigma_{j-1;i}(\lambda)} \geq\frac{\sigma_{j+1;i}(\lambda)}{\sigma_{j;i}(\lambda)},\] which is the variation of the Newton inequality(see \cite{HLP34}) \[\sigma_{j-1}(\lambda)\sigma_{j+1}(\lambda) \leq\left(\sigma_{j}(\lambda)\right)^2, ~\forall\lambda\in\mathds R^n,\] as we met in the proof of \lref{lem.xik}.\qed \end{remark}
\section{Proof of the main theorem}\label{sec.pmaint}
\subsection{Construction of the subsolutions}\label{subsec.csubsol} \quad
The purpose of this subsection is to prove the following key lemma and then use it to construct subsolutions of \eref{eqn.hqe}. We remark that for the generalized radially symmetric subsolution $\Phi(x)=\phi(r)$ that we intend to construct, the solution $\psi(r)$ discussed in the the following lemma actually is equivalent to $\phi'(r)/r$ (see the proof of the \lref{lem.Phi-subsol}). \begin{lemma}\label{lem.psi} Let $0\leq l<k\leq n$, $n\geq3$, $A\in\mathscr{\widetilde A}_{k,l}$, $a:=(a_1,a_2,...,a_n):=\lambda(A)$, $0<a_1\leq a_2\leq...\leq a_n$ and $\beta\geq 1$. Then the problem \begin{equation}\label{eqn.psi} \left\{ \begin{aligned} \psi(r)^k+\overline{\xi}_k(a)r\psi(r)^{k-1}\psi'(r)\quad&\\ -\psi(r)^l-\underline{\xi}_l(a)r\psi(r)^{l-1}\psi'(r)&=0,~r>1,\\[0.1cm] \psi(1)&=\beta, \end{aligned} \right. \end{equation} has a unique smooth solution $\psi(r)=\psi(r,\beta)$ on $[1,+\infty)$, which satisfies \begin{enumerate}[\quad(i)] \item[(i)] $1\leq\psi(r,\beta)\leq\beta$, $\partial_r\psi(r,\beta)\leq0$, $\forall r\geq1$, $\forall\beta\geq1$. More specifically, $\psi(r,1)\equiv 1$, $\psi(1,\beta)\equiv\beta$; and $1<\psi(r,\beta)<\beta$, $\forall r>1$, $\forall\beta>1$.
\item[(ii)] $\psi(r,\beta)$ is continuous and strictly increasing with respect to $\beta$ and \[\lim_{\beta\rightarrow+\infty}\psi(r,\beta)=+\infty,~\forall r\geq 1.\]
\item[(iii)] $\psi(r,\beta)=1+O(r^{-m})~(r\rightarrow+\infty)$, where $m=m_{k,l}(a)\in(2,n]$ and the $O(\cdot)$ depends only on $k$, $l$, $\lambda(A)$ and $\beta$. \end{enumerate} \end{lemma}
\begin{proof} For brevity, we will often write $\psi(r)$ or $\psi(r,\beta)$ (respectively, $\underline\xi(a),\overline\xi(a)$) simply as $\psi$ (respectively, $\underline\xi,\overline\xi$), when there is no confusion. The proof of this lemma now will be divided into three steps.
\emph{Step 1.}\quad We deduce from \eref{eqn.psi} that \begin{equation}\label{eqn.psi-kl} \psi^k-\psi^l=-\frac{r}{dr}\left(\overline{\xi}_k\psi^{k-1} -\underline{\xi}_l\psi^{l-1}\right)d\psi \end{equation} and \begin{equation}\label{eqn.psid} \frac{d\psi}{dr}=-\frac{1}{r}\cdot\frac{\psi^k-\psi^l} {\overline{\xi}_k\psi^{k-1}-\underline{\xi}_l\psi^{l-1}} =-\frac{1}{r}\cdot\frac{\psi}{\overline{\xi}_k}\cdot \frac{\psi^{k-l}-1}{\psi^{k-l}-\frac{\underline{\xi}_l}{\overline{\xi}_k}} =:\frac{g(\psi)}{r}, \end{equation} where we set \[g(\nu):=-\frac{\nu}{\overline{\xi}_k}\cdot \frac{\nu^{k-l}-1}{\nu^{k-l}-\frac{\underline{\xi}_l}{\overline{\xi}_k}}.\] Hence the problem \eref{eqn.psi} is equivalent to the following problem \begin{equation}\label{eqn.psi-g} \left\{ \begin{aligned} \psi'(r)&=\frac{g(\psi(r))}{r},~r>1,\\ \psi(1)&=\beta. \end{aligned} \right. \end{equation}
If $\beta=1$, then $\psi(r)\equiv 1$ is a solution of the problem \eref{eqn.psi-g} since $g(1)=0$. Thus, by the uniqueness theorem for the solution of the ordinary differential equation, we know that $\psi(r,1)\equiv 1$ is the unique solution satisfies the problem \eref{eqn.psi-g}.
Now if $\beta>1$, since \[h(r,\nu):=\frac{g(\nu)}{r} \in C^{\infty}((1,+\infty)\times(\nu_0,+\infty)),\] where \[\frac{\underline{\xi}_l}{\overline{\xi}_k}<\nu_0<1\] (note that $\nu_0$ exists, since we have $\underline{\xi}_l\leq l/n<k/n\leq\overline{\xi}_k$ by \lref{lem.xik}), by the existence theorem (the Picard-Lindel\"{o}f theorem) and the theorem of the maximal interval of existence for the solution of the initial value problem of the ordinary differential equation, we know that the problem \eref{eqn.psi-g} has a unique smooth solution $\psi(r)=\psi(r,\beta)$ locally around the initial point and can be extended to a maximal interval $[1,\zeta)$ in which $\zeta$ can only be one of the following cases: \begin{enumerate}[\qquad($1^\circ$)] \item $\zeta=+\infty$;
\item $\zeta<+\infty$, $\psi(r)$ is unbounded on $[1,\zeta)$;
\item $\zeta<+\infty$, $\psi(r)$ converges to some point on $\{\nu=\nu_0\}$ as $r\rightarrow\zeta-$. \end{enumerate} Since \[\frac{g(\psi(r))}{r}<0,~\forall \psi(r)>1,\] we see that $\psi(r)=\psi(r,\beta)$ is strictly decreasing with respect to $r$ which exclude the case ($2^\circ$) above. We claim now that the case ($3^\circ$) can also be excluded. Otherwise, the solution curve must intersect with $\{\nu=1\}$ at some point $(r_0,\psi(r_0))$ on it and then tends to $\{\nu=\nu_0\}$ after crossing it. But $\psi(r)\equiv 1$ is also a solution through $(r_0,\psi(r_0))$ which contradicts the uniqueness theorem for the solution of the initial value problem of the ordinary differential equation. Thus we complete the proof of the existence and uniqueness of the solution $\psi(r)=\psi(r,\beta)$ of the problem \eref{eqn.psi} on $[1,+\infty)$.
Due to the same reason, i.e., $\psi(r,\beta)$ is strictly decreasing with respect to $r$ and the solution curve can not cross $\{\nu=1\}$ provided $\beta>1$, assertion $\emph{(i)}$ of the lemma is also clear now, that is, $1<\psi(r,\beta)<\beta$, $\forall r>1$, $\forall\beta>1$.
\emph{Step 2.}\quad By the theorem of the differentiability of the solution with respect to the initial value, we can differentiate $\psi(r,\beta)$ with respect to $\beta$ as blew: \[\left\{ \begin{aligned} &\frac{\partial\psi(r,\beta)}{\partial r}=\frac{g(\psi(r,\beta))}{r},&\\ &\psi(1,\beta)=\beta;& \end{aligned}\right.\] \[\Rightarrow\left\{ \begin{aligned} &\frac{\partial^2 \psi(r,\beta)}{\partial \beta\partial r} =\frac{g'(\psi(r,\beta))}{r}\cdot\frac{\partial\psi(r,\beta)}{\partial\beta},&\\ &\frac{\partial\psi(1,\beta)}{\partial\beta}=1.& \end{aligned}\right.\] Let \[v(r):=\frac{\partial\psi(r,\beta)}{\partial\beta}.\] We have \[\left\{ \begin{aligned} &\frac{dv}{dr} =\frac{g'(\psi(r,\beta))}{r}\cdot v,&\\ &v(1)=1.& \end{aligned}\right.\] Therefore we can deduce that \[\frac{dv}{v} =\frac{g'(\psi(r,\beta))}{r}dr,\] and hence \[\frac{\partial\psi(r,\beta)}{\partial\beta}=v(r) =\exp\int_1^r{\frac{g'(\psi(\tau,\beta))}{\tau}d\tau}.\]
Since \[g'(\nu)=-\frac{\nu^{k-l}-1}{\overline{\xi}_k\nu^{k-l}-\underline{\xi}_l} -\frac{\nu}{\overline{\xi}_k}\cdot \frac{-\left(\frac{\underline{\xi}_l}{\overline{\xi}_k}-1\right)(k-l)\nu^{k-l-1}} {\left(\nu^{k-l}-\frac{\underline{\xi}_l}{\overline{\xi}_k}\right)^2}\] and \begin{equation}\label{eqn.onepsibeta} 0<\frac{\underline{\xi}_l}{\overline{\xi}_k} <1\leq\psi(r,\beta)\leq\beta=\psi(1),~\forall r\geq 1, \end{equation} we have \[g'(\psi(r,\beta))\leq -\frac{(k-l)\left(1-\frac{\underline{\xi}_l}{\overline{\xi}_k}\right)} {\overline{\xi}_k \left(\beta^{k-l}-\frac{\underline{\xi}_l}{\overline{\xi}_k}\right)^2} =-C\left(k,l,\lambda(A),\beta\right)<0,\] and hence \[0<\frac{\partial\psi(r,\beta)}{\partial\beta} \leq r^{-C}\leq 1,~\forall r\geq 1.\] Thus $\psi(r,\beta)$ is strictly increasing with respect to $\beta$.
\emph{Step 3.}\quad By \eref{eqn.psi-kl}, we have \begin{eqnarray*} -d\ln r=-\frac{dr}{r}&=&\frac{\overline{\xi}_k\psi^{k-1} -\underline{\xi}_l\psi^{l-1}}{\psi^k-\psi^l}d\psi =\frac{\overline{\xi}_k}{\psi}\cdot \frac{\psi^{k-l}-\frac{\underline{\xi}_l}{\overline{\xi}_k}}{\psi^{k-l}-1}d\psi\\ &=&\frac{\overline{\xi}_k}{\psi}\left(1 +\frac{1-\frac{\underline{\xi}_l}{\overline{\xi}_k}}{\psi^{k-l}-1}\right)d\psi =\left(\frac{\overline{\xi}_k}{\psi} +\frac{\overline{\xi}_k-\underline{\xi}_l}{\psi(\psi^{k-l}-1)}\right)d\psi\\ &=&\overline{\xi}_kd\ln\psi-\frac{\overline{\xi}_k-\underline{\xi}_l}{k-l} d\ln\frac{\psi^{k-l}}{\psi^{k-l}-1}\\ &=&d\ln\left(\psi^{\overline{\xi}_k}\left(1-\psi^{-k+l}\right) ^{\frac{\overline{\xi}_k-\underline{\xi}_l}{k-l}}\right). \end{eqnarray*}
Hence \begin{equation}\label{eqn.mdlnr} -md\ln r=d\ln\left(\psi(r)^{m\overline{\xi}_k} \left(1-\psi(r)^{-k+l}\right)\right), \end{equation} where \[m:=m_{k,l}(a):=\frac{k-l}{\overline{\xi}_k-\underline{\xi}_l},\] which has been already defined in \eref{eqn.mkla} in \ssref{subsec.xi}. Note that, by the assumptions on $A$ and \cref{cor.m}, we have $2<m\leq n$ and $m\overline{\xi}_k>k-l$.
Integrating \eref{eqn.mdlnr} from $1$ to $r$ and recalling $\psi(1)=\beta\geq1$, we get \[\ln\left(\psi(r)^{m\overline{\xi}_k} \left(1-\psi(r)^{-k+l}\right)\right) =\ln\left(\beta^{m\overline{\xi}_k} \left(1-\beta^{-k+l}\right)\right)+\ln r^{-m},\] and hence \[\psi(r)^{m\overline{\xi}_k} \left(1-\psi(r)^{-k+l}\right) =\beta^{m\overline{\xi}_k} \left(1-\beta^{-k+l}\right)r^{-m}:=B(\beta)r^{-m},\] where we set \[B(\beta):=\beta^{m\overline{\xi}_k}\left(1-\beta^{-k+l}\right) =\beta^{m\overline{\xi}_k-k+l}\left(\beta^{k-l}-1\right).\] Since \begin{eqnarray*} &~&\psi(r)^{m\overline{\xi}_k} \left(1-\psi(r)^{-k+l}\right) =\psi(r)^{m\overline{\xi}_k-k+l} \left(\psi(r)^{k-l}-1\right)\\ &=&\psi(r)^{m\overline{\xi}_k-k+l}\left(\psi(r)-1\right) \left(\psi(r)^{k-l-1}+\psi(r)^{k-l-2}+...+\psi(r)+1\right), \end{eqnarray*} we thus conclude that \begin{equation}\label{eqn.psimoar} \frac{\psi(r)-1}{r^{-m}} =\left(\psi(r)^{m\overline{\xi}_k-k+l} \left(\psi(r)^{k-l-1}+\psi(r)^{k-l-2}+ ...+\psi(r)+1\right)\right)^{-1}B(\beta). \end{equation}
Note that $m\overline{\xi}_k-k+l>0$ and \[\beta-1=\left(\beta^{m\overline{\xi}_k-k+l} \left(\beta^{k-l-1}+\beta^{k-l-2}+ ...+b+1\right)\right)^{-1}B(\beta).\] Recalling \eref{eqn.onepsibeta}, we obtain \begin{equation}\label{eqn.psi-b} \beta-1\leq\frac{\psi(r,\beta)-1}{r^{-m}} \leq\frac{B(\beta)}{k-l}, ~\forall r\geq 1. \end{equation} Thus we have \[\lim_{\beta\rightarrow+\infty}\psi(r,\beta)=+\infty, ~\forall r\geq 1,\] and \[\psi(r,\beta)\rightarrow 1~(r\rightarrow+\infty), ~\forall \beta\geq 1.\] Substituting the latter to \eref{eqn.psimoar}, we get \[\frac{\psi(r,\beta)-1}{r^{-m}} \rightarrow\frac{B(\beta)}{k-l}~(r\rightarrow+\infty), ~\forall \beta\geq 1.\]
Therefore \[\psi(r,\beta)=1+\frac{B(\beta)}{k-l}r^{-m}+o(r^{-m}) =1+O(r^{-m})~(r\rightarrow+\infty),\] where $o(\cdot)$ and $O(\cdot)$ depend only on $k$, $l$, $\lambda(A)$ and $\beta$.
This completes the proof of the lemma. \end{proof}
\begin{remark} For $l=0$, i.e., the Hessian equation $\sigma_k(\lambda)=1$, we have an easy proof. Consider the problem \begin{equation}\label{eqn.psi-he} \left\{ \begin{aligned} \psi(r)^k+\overline{\xi}_k(a)r\psi(r)^{k-1}\psi'(r)&=1,~r>1,\\[0.1cm] \psi(1)&=\beta. \end{aligned} \right. \end{equation}
Set $m:=m_{k,0}(a)=k/\overline\xi_k$. We have \[\psi^k-1=-r\overline\xi_k\psi^{k-1}\frac{d\psi}{dr} =-\frac{1}{m}\cdot r\cdot\frac{d\left(\psi^k-1\right)}{dr},\] \[\frac{d\left(\psi^k-1\right)}{\psi^k-1} =-m\frac{dr}{r}\] and \[d\ln\left(\psi(r)^k-1\right)=-md\ln r=d\ln r^{-m}.\] Integrating it from $1$ to $r$ and recalling $\psi(1)=\beta\geq1$, we get \[\psi(r)^k-1=\left(\psi(1)^k-1\right)r^{-m} =(\beta^k-1)r^{-m}\] and \begin{eqnarray} \nonumber\psi(r)&=&\left(1+(\beta^k-1)r^{-m}\right)^{\frac{1}{k}}\\ &=&\left(1+(\beta^k-1)r^{-\frac{k}{\overline\xi_k}}\right)^{\frac{1}{k}} \label{eqn.compare}\\ \nonumber&=&1+\frac{\beta^k-1}{k}r^{-m}+o(r^{-m})=1+O(r^{-m})~(r\rightarrow+\infty). \end{eqnarray} It is obvious that the $\psi(r)$ that we here solved from \eref{eqn.psi-he} for $l=0$ satisfies all the conclusions of \lref{lem.psi}. Moreover, comparing \eref{eqn.compare} with the corresponding ones in \cite{BLL14} and in \cite{CL03}, we observe that our method actually provides a systematic way for construction of the subsolutions, which gives results containing the previous ones as special cases. \qed \end{remark}
Set \[\mu_R(\beta):=\int_R^{+\infty}\tau\big(\psi(\tau,\beta)-1\big)d\tau, \quad\forall R\geq 1,~\forall\beta\geq 1.\] Note that the integral on the right hand side is convergent in view of \lref{lem.psi}-\textsl{(iii)}. Moreover, as an application of \lref{lem.psi}, we have the following. \begin{corollary}\label{cor.mub} $\mu_R(\beta)$ is nonnegative, continuous and strictly increasing with respect to $\beta$. Furthermore, \[\mu_R(\beta)\geq\int_R^{+\infty}(\beta-1)r^{-m+1}d\tau \rightarrow+\infty~(\beta\rightarrow+\infty),~\forall R\geq 1;\] and \[\mu_R(\beta)=O(R^{-m+2})~(R\rightarrow+\infty),~\forall\beta\geq 1.\] \end{corollary}
\begin{proof} By \lref{lem.psi}-\textsl{(ii),(iii)} and the above property \eref{eqn.psi-b} of $\psi(r,\beta)$. \end{proof}
For any $\alpha,\beta,\gamma\in\mathds R$, $\beta,\gamma\geq1$ and for any diagonal matrix $A\in\mathscr{\widetilde A}_{k,l}$, let \[\phi(r):=\phi_{\alpha,\beta,\gamma}(r) :=\alpha+\int_\gamma^r\tau\psi(\tau,\beta)d\tau, ~\forall r\geq\gamma,\] and \[\Phi(x):=\Phi_{\alpha,\beta,\gamma,A}(x):=\phi(r) :=\phi_{\alpha,\beta,\gamma}(r_A(x)), ~\forall x\in\mathds R^n\setminus{E_\gamma},\] where $r=r_A(x)=\sqrt{x^TAx}$. Then we have \begin{eqnarray} \phi_{\alpha,\beta,\gamma}(r) &=&\nonumber\int_\gamma^r\tau\big(\psi(\tau,\beta)-1\big)d\tau +\frac{1}{2}r^2-\frac{1}{2}\gamma^2+\alpha\\ &=&\frac{1}{2}r^2+\left(\mu_\gamma(\beta)+\alpha -\frac{1}{2}\gamma^2\right)-\mu_r(\beta)\label{eqn.phi-mu}\\ &=&\frac{1}{2}r^2+\left(\mu_\gamma(\beta)+\alpha -\frac{1}{2}\gamma^2\right)+O(r^{-m+2}) ~(r\rightarrow+\infty),\quad\label{eqn.phi-O} \end{eqnarray} according to \cref{cor.mub}, and now we can assert that \begin{lemma}\label{lem.Phi-subsol} $\Phi$ is a smooth $k$-convex subsolution of \eref{eqn.hqe} in $\mathds R^n\setminus\overline{E_\gamma}$, that is, \[\sigma_j\left(\lambda\left(D^2{\Phi(x)}\right)\right)\geq0, ~\forall 1\leq j\leq k,~\forall x\in\mathds R^n\setminus\overline{E_\gamma},\] and \[\frac{\sigma_k\left(\lambda\left(D^2{\Phi(x)}\right)\right)} {\sigma_l\left(\lambda\left(D^2{\Phi(x)}\right)\right)}\geq 1, ~\forall x\in\mathds R^n\setminus\overline{E_\gamma}.\] \end{lemma}
\begin{proof} By definition we have $\phi'(r)=r\psi(r)$ and $\phi''(r)=\psi(r)+r\psi'(r)$. Since \[r^2=x^TAx=\sum_{i=1}^{n}a_ix_i^2,\] we deduce that \[2r\partial_{x_i}r=\partial_{x_i}\left(r^2\right)=2a_ix_i \quad\text{and}\quad \partial_{x_i}r=\frac{a_ix_i}{r}.\] Consequently \[\partial_{x_i}\Phi(x)=\phi'(r)\partial_{x_i}r =\frac{\phi'(r)}{r}a_ix_i,\] \begin{eqnarray*} \partial_{x_ix_j}\Phi(x) &=&\frac{\phi'(r)}{r}a_i\delta_{ij}+ \frac{\phi''(r)-\frac{\phi'(r)}{r}}{r^2}(a_ix_i)(a_jx_j)\\ &=&\psi(r)a_i\delta_{ij}+\frac{\psi'(r)}{r}(a_ix_i)(a_jx_j), \end{eqnarray*} and therefore \[D^2\Phi=\left(\psi(r)a_i\delta_{ij} +\frac{\psi'(r)}{r}(a_ix_i)(a_jx_j)\right)_{n\times n}.\]
So we can conclude from \lref{lem.skm} that \begin{eqnarray*} \sigma_j\left(\lambda\left(D^2\Phi\right)\right) &=&\sigma_j(a)\psi(r)^j+\frac{\psi'(r)}{r}\psi(r)^{j-1} \sum_{i=1}^n\sigma_{j-1;i}(a)a_i^2x_i^2\\ &=&\sigma_j(a)\psi^j+\Xi_j(a,x)\sigma_j(a)r\psi^{j-1}\psi'\\ &\geq&\sigma_j(a)\psi^j+\overline{\xi}_j(a)\sigma_j(a)r\psi^{j-1}\psi'\\ &=&\sigma_j(a)\psi^{j-1}\left(\psi+\overline{\xi}_j(a)r\psi'\right), ~\forall 1\leq j\leq n, \end{eqnarray*} where we have used the facts that $\psi(r)\geq1>0$ and $\psi'(r)\leq0$ for all $r\geq1$, according to \lref{lem.psi}-\textsl{(i)}.
For any fixed $1\leq j\leq k$, in view of \lref{lem.xik} and \lref{lem.psi}-\textsl{(i)}, we have \[0\leq\frac{\psi^{k-l}-1}{\psi^{k-l}-\frac{\underline{\xi}_l(a)}{\overline{\xi}_k(a)}} <1\leq\frac{\overline{\xi}_k(a)}{\overline{\xi}_j(a)}.\] Hence it follows from \eref{eqn.psid} that \[\psi'=-\frac{1}{r}\cdot\frac{\psi}{\overline{\xi}_k(a)}\cdot \frac{\psi^{k-l}-1}{\psi^{k-l}-\frac{\underline{\xi}_l(a)}{\overline{\xi}_k(a)}} >-\frac{1}{r}\cdot\frac{\psi}{\overline{\xi}_j(a)},\] which yields $\psi+\overline{\xi}_j(a)r\psi'>0$.
Since $A\in\mathscr{\widetilde A}_{k,l}$ implies $a\in\Gamma^+$, that is, $\sigma_i(a)>0$ for all $1\leq i\leq n$, we thus conclude that \[\sigma_j\left(\lambda\left(D^2\Phi\right)\right)>0, ~\forall 1\leq j\leq k.\] In particular, we have \[\sigma_k\left(\lambda\left(D^2\Phi\right)\right)>0 \quad\text{and}\quad \sigma_l\left(\lambda\left(D^2\Phi\right)\right)>0.\]
On the other hand, \begin{eqnarray*} &~&\sigma_k\left(\lambda\left(D^2\Phi\right)\right) -\sigma_l\left(\lambda\left(D^2\Phi\right)\right)\\ &=&\sigma_k(a)\psi^k+\Xi_k(a,x)\sigma_k(a)r\psi^{k-1}\psi' -\sigma_l(a)\psi^l-\Xi_l(a,x)\sigma_l(a)r\psi^{l-1}\psi'\\ &\geq&\sigma_k(a)\psi^k+\overline{\xi}_k(a)\sigma_k(a)r\psi^{k-1}\psi' -\sigma_l(a)\psi^l-\underline{\xi}_l(a)\sigma_l(a)r\psi^{l-1}\psi'\\ &=&\sigma_k(a)\left(\psi^k+\overline{\xi}_k(a)r\psi^{k-1}\psi' -\psi^l-\underline{\xi}_l(a)r\psi^{l-1}\psi'\right)\\ &=&0. \end{eqnarray*} Therefore \[\frac{\sigma_k\left(\lambda\left(D^2{\Phi}\right)\right)} {\sigma_l\left(\lambda\left(D^2{\Phi}\right)\right)}\geq 1.\] This completes the proof of \lref{lem.Phi-subsol}. \end{proof}
\subsection{Proof of \tref{thm.hqe}}\label{subsec.pmaint} \quad
We first introduce the following lemma which is a special and simple case of \tref{thm.hqe} with the additional condition that the matrix $A$ is diagonal and the vector $b$ vanishes. \begin{lemma}\label{lem.hqe} Let $D$ be a bounded strictly convex domain in $\mathds R^n$, $n\geq 3$, $\partial D\in C^2$ and let $\varphi\in C^2(\partial D)$. Then for any given \textsl{diagonal matrix} $A\in\mathscr{\widetilde A}_{k,l}$ with $0\leq l<k\leq n$, there exists a constant $\tilde{c}$ depending only on $n,D,k,l,A$ and $\norm{\varphi}_{C^2(\partial D)}$, such that for every $c\geq\tilde{c}$, there exists a unique viscosity solution $u\in C^0(\mathds R^n\setminus D)$ of \begin{equation}\label{eqn.hqe-ac} \left\{ \begin{aligned} \displaystyle \frac{\sigma_k(\lambda(D^2u))}{\sigma_l(\lambda(D^2u))}&=1 \quad\text{in}~\mathds R^n\setminus\overline{D},\\%[0.05cm] \displaystyle u&=\varphi\quad\text{on}~\partial D,\\%[0.05cm]
\displaystyle \limsup_{|x|\rightarrow+\infty}|x|^{m-2}
&\left|u(x)-\left(\frac{1}{2}x^{T}Ax+c\right)\right|<\infty, \end{aligned} \right. \end{equation} where $m=m_{k,l}(\lambda(A))\in(2,n]$. \end{lemma}
To prove \tref{thm.hqe}, it suffices to prove \lref{lem.hqe}. Indeed, suppose that $D,\varphi,A$ and $b$ satisfy the hypothesis of \tref{thm.hqe}. Consider the decomposition $A=Q^TNQ$, where $Q$ is an orthogonal matrix and $N$ is a diagonal matrix which satisfies $\lambda(N)=\lambda(A)$. Let
\[\tilde{x}:=Qx,\quad\widetilde{D}:=\left\{Qx|x\in D\right\}\] and \[\tilde{\varphi}(\tilde{x}):=\varphi(x)-b^Tx =\varphi(Q^T\tilde{x})-b^TQ^T\tilde{x}.\] By \lref{lem.hqe}, we conclude that there exists a constant $\tilde{c}$ depending only on $n,\widetilde{D},k,l,N$ and $\norm{\tilde{\varphi}}_{C^2(\partial\widetilde{D})}$, such that for every $c\geq\tilde{c}$, there exists a unique viscosity solution $\tilde{u}\in C^0(\mathds R^n\setminus\widetilde{D})$ of \begin{equation}\label{eqn.hqe-nc} \left\{ \begin{aligned} \displaystyle \frac{\sigma_k(\lambda(D^2\tilde{u}))} {\sigma_l(\lambda(D^2\tilde{u}))}&=1 \quad\text{in}~\mathds R^n\setminus\overline{\widetilde{D}},\\%[0.05cm] \displaystyle \tilde{u}&=\tilde{\varphi} \quad\text{on}~\partial\widetilde{D},\\%[0.05cm]
\displaystyle \limsup_{|\tilde{x}|\rightarrow+\infty}|\tilde{x}|^{m-2}
&\left|\tilde{u}(\tilde{x})-\left(\frac{1}{2}\tilde{x}^{T}N\tilde{x}
+c\right)\right|<\infty, \end{aligned} \right. \end{equation} where $m=m_{k,l}(\lambda(N))=m_{k,l}(\lambda(A))\in(2,n]$. Let \[u(x):=\tilde{u}(\tilde{x})+b^Tx=\tilde{u}(Qx)+b^Tx =\tilde{u}(\tilde{x})+b^TQ^T\tilde{x}.\] We claim that $u$ is the solution of \eref{eqn.hqe-abc} in \tref{thm.hqe}. To show this, we only need to note that \[D^2u(x)=Q^TD^2\tilde{u}(\tilde{x})Q,\quad \lambda\left(D^2u(x)\right)=\lambda\left(D^2\tilde{u}(\tilde{x})\right);\] \[u=\varphi\quad\text{on}~\partial D;\] and \begin{eqnarray*}
&&\displaystyle |\tilde{x}|^{m-2}\left|\tilde{u}(\tilde{x})
-\left(\frac{1}{2}\tilde{x}^{T}N\tilde{x}+c\right)\right|\\
&=&\displaystyle \left(x^TQ^TQx\right)^{(m-2)/2}\left|u(x)-b^Tx
-\left(\frac{1}{2}x^TQ^TNQx+c\right)\right|\\
&=&\displaystyle |x|^{m-2}
\left|u(x)-\left(\frac{1}{2}x^{T}Ax+b^Tx+c\right)\right|. \end{eqnarray*} Thus we have proved that \tref{thm.hqe} can be established by \lref{lem.hqe}.
\begin{remark} \begin{enumerate}[(1)] \item We may see from the above demonstration that the lower bound $\tilde{c}$ of $c$ in \tref{thm.hqe} can not be discarded generally. Indeed, for the radial solutions of the Hessian equation $\sigma_k(\lambda(D^2u))=1$ in $\mathds R^n\setminus\overline{B_1}$, \cite[Theorem 2]{WB13} states that there is no solution when $c$ is too small.
\item Unlike the Poisson equation and the Monge-Amp\`{e}re equation, generally, for the Hessian quotient equation, the matrix $A$ in \tref{thm.hqe} can only be normalized to a diagonal matrix, and can not be normalized to $I$ multiplied by some constant. This is the reason why we study the generalized radially symmetric solutions, rather than the radial solutions, of the original equation \eref{eqn.hqe}. See also \cite{BLL14}.\qed \end{enumerate} \end{remark}
Now we use the Perron's method to prove \lref{lem.hqe}.
\begin{proof}[\textbf{Proof of \lref{lem.hqe}}] We may assume without loss of generality that $E_1\subset\subset D\subset\subset E_{\bar{r}} \subset\subset E_{\hat{r}}$ and $a:=(a_1,a_2,...,a_n):=\lambda(A)$ with $0<a_1\leq a_2\leq...\leq a_n$. The proof now will be divided into three steps.
\emph{Step 1.}\quad Let \[\eta:=\inf_{\substack{x\in \overline{E_{\bar{r}}}\setminus D\\ \xi\in\partial D}}Q_\xi(x), \quad Q(x):=\sup_{\xi\in\partial D}Q_\xi(x)\] and \[\Phi_\beta(x):=\eta+\int_{\bar{r}}^{r_A(x)}\tau\psi(\tau,\beta)d\tau, \quad\forall r_A(x)\geq 1,~\forall \beta\geq 1,\] where $Q_{\xi}(x)$ and $\psi(r,\beta)$ are given by \lref{lem.Qxi} and \lref{lem.psi}, respectively. Then we have \begin{enumerate}[\quad(1)] \item Since $Q$ is the supremum of a collection of smooth solutions $\{Q_\xi\}$ of \eref{eqn.hqe}, it is a continuous subsolution of \eref{eqn.hqe}, i.e., \[\frac{\sigma_k(\lambda(D^2Q))}{\sigma_l(\lambda(D^2{Q}))}\geq 1\] in $\mathds R^n\setminus \overline{D}$ in the viscosity sense (see \cite[Proposition 2.2]{Ish89}).
\item $Q=\varphi$ on $\partial D$. To prove this we only need to show that for any $\xi\in\partial D$, $Q(\xi)=\varphi(\xi)$. This is obvious since $Q_\xi\leq\varphi$ on $\overline{D}$ and $Q_\xi(\xi)=\varphi(\xi)$, according to \rref{rmk.Qxi}-\textsl{(1)}.
\item By \lref{lem.Phi-subsol}, $\Phi_\beta$ is a smooth subsolution of \eref{eqn.hqe} in $\mathds R^n\setminus\overline{D}$.
\item $\Phi_\beta\leq\varphi$ on $\partial D$ and $\Phi_\beta\leq Q$ on $\overline{E_{\bar{r}}}\setminus D$. To show them we first note that $\Phi_\beta(x)$ is strictly increasing with respect to $r_A(x)$ since $\psi(r,\beta)\geq 1>0$ by \lref{lem.psi}-\textsl{(i)}. Invoking $\Phi_\beta=\eta$ on $\partial E_{\bar{r}}$ and $\eta\leq Q$ on $\overline{E_{\bar{r}}}\setminus D$ by their definitions, we have $\Phi_\beta\leq\eta\leq Q$ on $\overline{E_{\bar{r}}}\setminus D$. On the other hand, according to \rref{rmk.Qxi}-\textsl{(1)}, we have $Q_\xi\leq\varphi$ on $\overline{D}$ which implies that $\eta\leq\varphi$ on $\overline{D}$. Combining these two aspects we deduce that $\Phi_\beta\leq\eta\leq\varphi$ on $\partial D$.
\item $\Phi_\beta(x)$ is strictly increasing with respect to $\beta$ and \begin{equation}\label{eqn.Phib} \lim_{\beta\rightarrow+\infty}\Phi_\beta(x)=+\infty, ~\forall r_A(x)\geq 1, \end{equation} by the definition of $\Phi_\beta(x)$ and \lref{lem.psi}-\textsl{(ii)}.
\item As we showed in \eref{eqn.phi-mu} and \eref{eqn.phi-O}, for any $\beta\geq 1$, we have \begin{eqnarray*} \Phi_\beta(x) &=&\eta+\int_{\bar{r}}^{r_A(x)}\tau \psi(\tau,\beta)d\tau\\ &=&\eta+\frac{1}{2}(r_A(x)^2-\bar{r}^2) +\int_{\bar{r}}^{r_A(x)}\tau\big(\psi(\tau,\beta)-1\big)d\tau\\ &=&\frac{1}{2}r_A(x)^2+\left(\eta-\frac{1}{2}\bar{r}^2 +\mu_{\bar{r}}(\beta)\right)-\mu_{r_A(x)}(\beta)\\ &=&\frac{1}{2}r_A(x)^2+\mu(\beta)-\mu_{r_A(x)}(\beta)\\
&=&\frac{1}{2}x^TAx+\mu(\beta)+O\left(|x|^{-m+2}\right)
~(|x|\rightarrow+\infty), \end{eqnarray*} where we set \[\mu(\beta):=\eta-\frac{1}{2}\bar{r}^2+\mu_{\bar{r}}(\beta),\]
and used the fact that $x^TAx=O(|x|^2)~(|x|\rightarrow+\infty)$ since $\lambda(A)\in\Gamma^+$. \end{enumerate}
\emph{Step 2.}\quad For fixed $\hat{r}>\bar{r}$, there exists $\hat{\beta}>1$ such that \[\min_{\partial E_{\hat{r}}}\Phi_{\hat{\beta}} >\max_{\partial E_{\hat{r}}}Q,\] in light of \eref{eqn.Phib}. Thus we obtain \begin{equation}\label{eqn.PhiQ} \Phi_{\hat{\beta}}>Q\quad\mbox{on}~\partial E_{\hat{r}}. \end{equation}
Let \[\tilde{c}:=\max\left\{\eta,\mu(\hat{\beta}),\bar{c}\right\},\] where the $\bar{c}$ comes from \rref{rmk.Qxi}-\textsl{(3)}, and hereafter fix $c\geq\tilde{c}$.
By \lref{lem.psi} and \cref{cor.mub} we deduce that \[\psi(r,1)\equiv 1\Rightarrow\mu_{\bar{r}}(1)=0 \Rightarrow\mu(1)=\eta-\frac{1}{2}\bar{r}^2<\eta\leq\tilde{c}\leq c,\] and \[\lim_{\beta\rightarrow+\infty}\mu_{\bar{r}}(\beta)=+\infty \Rightarrow\lim_{\beta\rightarrow+\infty}\mu(\beta)=+\infty.\] On the other hand, it follows from \cref{cor.mub} that $\mu(\beta)$ is continuous and strictly increasing with respect to $\beta$ \big{(}which indicates that the inverse of $\mu(\beta)$ exists and $\mu^{-1}$ is strictly increasing\big{)}. Thus there exists a unique $\beta(c)$ such that $\mu(\beta(c))=c$. Then we have \[\Phi_{\beta(c)}(x)=\frac{1}{2}r_A(x)^2+c-\mu_{r_A(x)}(\beta(c))
=\frac{1}{2}x^TAx+c+O\left(|x|^{-m+2}\right)~(|x|\rightarrow+\infty),\] and \[\beta(c)=\mu^{-1}(c)\geq\mu^{-1}(\tilde{c})\geq\hat{\beta}.\] Invoking the monotonicity of $\Phi_\beta$ with respect to $\beta$ and \eref{eqn.PhiQ}, we obtain \begin{equation}\label{eqn.PhigQ} \Phi_{\beta(c)}\geq\Phi_{\hat{\beta}}>Q\quad\mbox{on} ~\partial E_{\hat{r}}. \end{equation} Note that we already know \[\Phi_{\beta(c)}\leq Q\quad\mbox{on} ~\overline{E_{\bar{r}}}\setminus D,\] from (4) of \emph{Step 1}.
Let \begin{equation*} \underline{u}(x):= \begin{cases} \max\left\{\Phi_{\beta(c)}(x),Q(x)\right\},& x\in E_{\hat{r}}\setminus D,\\ \Phi_{\beta(c)}(x),& x\in \mathds R^n\setminus E_{\hat{r}}. \end{cases} \end{equation*} Then we have \begin{enumerate}[\quad(1)] \item $\underline{u}$ is continuous and satisfies \[\frac{\sigma_k(\lambda(D^2{\underline{u}}))} {\sigma_l(\lambda(D^2{\underline{u}}))}\geq1\] in $\mathds R^n\setminus\overline{D}$ in the viscosity sense, by (1) and (3) of \emph{Step 1}.
\item $\underline{u}=Q=\varphi$ on $\partial D$, by (2) of \emph{Step 1}.
\item If $r_A(x)$ is large enough, then \[\underline{u}(x)=\Phi_{\beta(c)}(x)
=\frac{1}{2}x^TAx+c+O\left(|x|^{-m+2}\right)~(|x|\rightarrow+\infty).\] \end{enumerate}
\emph{Step 3.}\quad Let \[\overline{u}(x):=\frac{1}{2}x^TAx+c, ~\forall x\in\mathds R^n.\] Then $\overline{u}$ is obviously a supersolution and
\[\lim_{|x|\rightarrow+\infty}\left(\underline{u}-\overline{u}\right)(x)=0.\]
To use the Perron's method to establish \lref{lem.hqe}, we now only need to prove that \[\underline{u}\leq \overline{u}\quad\mbox{in}~\mathds R^n\setminus D.\]
In fact, since \[\mu_{r_A(x)}(\beta)\geq 0, \quad\forall x\in\mathds R^n\setminus E_1, ~\forall\beta\geq 1,\] according to \cref{cor.mub}, we have \begin{equation}\label{eqn.Phiolu} \Phi_{\beta(c)}(x)=\frac{1}{2}x^TAx+c-\mu_{r_A(x)}(\beta(c)) \leq\frac{1}{2}x^TAx+c=\overline{u}(x),~\forall x\in\mathds R^n\setminus D. \end{equation} \Big{(}We remark that this \eref{eqn.Phiolu} can also be proved by using the comparison principle, in view of \[\Phi_{\beta(c)}\leq\eta\leq\tilde{c}\leq c\leq\overline{u} \quad\mbox{on}~\partial D,\] and
\[\lim_{|x|\rightarrow+\infty}\left(\Phi_{\beta(c)}-\overline{u}\right)(x)=0.\Big{)}\]
On the other hand, for every $\xi\in\partial D$, since \[Q_\xi(x)\leq\frac{1}{2}x^TAx+\bar{c} \leq\frac{1}{2}x^TAx+\tilde{c} \leq\frac{1}{2}x^TAx+c=\overline{u}(x), ~\forall x\in\partial D,\] and \[Q_\xi\leq Q<\Phi_{\beta(c)}\leq\overline{u} \quad\mbox{on}~\partial E_{\hat{r}}\] follows from \eref{eqn.PhigQ} and \eref{eqn.Phiolu}, we obtain \[Q_\xi\leq\overline{u}\quad\mbox{on} ~\partial\left(E_{\hat{r}}\setminus D\right).\] In view of \[\frac{\sigma_k(\lambda(D^2{Q_\xi}))}{\sigma_l(\lambda(D^2{Q_\xi}))} =1=\frac{\sigma_k(\lambda(D^2{\overline{u}}))} {\sigma_l(\lambda(D^2{\overline{u}}))} \quad\mbox{in}~E_{\hat{r}}\setminus D,\] we deduce from the comparison principle that \[Q_\xi\leq\overline{u}\quad\mbox{in}~E_{\hat{r}}\setminus D.\] Hence \begin{equation}\label{eqn.Qu} Q\leq\overline{u}\quad\mbox{in}~E_{\hat{r}}\setminus D. \end{equation} Combining \eref{eqn.Phiolu} and \eref{eqn.Qu}, by the definition of $\underline{u}$, we get \[\underline{u}\leq \overline{u}\quad\mbox{in}~\mathds R^n\setminus D.\] This finishes the proof of \lref{lem.hqe}. \end{proof}
\begin{remark} To prove \lref{lem.hqe} we have used above \lref{lem.cp} and \lref{lem.pm} presented in \ssref{subsec.prelem}. In fact, one can follow the techniques in \cite{CL03} (see also \cite{DB11} \cite{Dai11} and \cite{LD12}) instead of \lref{lem.pm} to rewrite the whole proof. These two kinds of presentation look a little different but are essentially the same. \end{remark}
\end{document}
|
arXiv
|
{
"id": "1709.04712.tex",
"language_detection_score": 0.5810593366622925,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{{\bf\Large{On the K\"ahler-Ricci flow with small initial $E_1$ energy (I)}
\tableofcontents \section{Introduction and main results } \subsection{The motivation} In \cite{[chen-tian2]} \cite{[chen-tian1]}, a family of functionals $E_k (k = 1,2,\cdots n)$ was introduced by the first named author and G. Tian to prove the convergence of the K\"ahler-Ricci flow under appropriate curvature assumptions. The aim of this program (cf. \cite{[chen1]}) is to study how the lower bound of $E_1$ is used to derive the convergence of the K\"ahler-Ricci flow, i.e., the existence of K\"ahler-Einstein metrics. We will address this question in Subsection 1.2. The corresponding problem of the relation between the lower bound of $E_0$, which is the $K$-energy introduced by T. Mabuchi, and the existence of K\"ahler-Einstein metrics has been extensively studied (cf. \cite{[BaMa]}, \cite{[chen-tian2]}, \cite{[chen-tian1]}, \cite{[Do]}). One interesting question in this program is how the lower bound of $E_1$ compares to the lower bound of $E_0$. We will give a satisfactory answer to this question in Subsection 1.3. \\
\subsection{The lower bound of $E_1$ and K\"ahler-Einstein metrics} Let $(M, [\omega])$ be a polarized compact K\"ahler manifold with $[\omega]=2\pi c_1(M)>0 $ (the first Chern class) in this paper. In
\cite{[chen1]}, the first named author proved a stability theorem of the K\"ahler-Ricci flow near the infimum of $E_1$ under the assumption that the initial metric has $Ric>-1$ and $|Rm|$ bounded. Unfortunately, this stability theorem needs a topological assumption \begin{equation}(-1)^n([c_1(M)]^{[2]}[\omega]^{[n-2]}-\frac {2(n+1)}{n}[c_2(M)][\omega]^{[n-2]})\geq 0.\label{c1}\end{equation} The only known compact manifold which satisfies this condition is ${\mathbb C} P^n$, which restricts potential applications of this result. The main purpose of this paper is to remove this assumption.
\begin{theo}\label{main2} Suppose that $M$ is pre-stable, and $E_1$ is bounded from below in $ [\omega]$. For any $\delta, \Lambda>0,$ there exists a small positive constant $\epsilon(\delta, \Lambda)>0$ such that for any metric $\omega_0$ in the subspace ${\mathcal A}(\delta, \Lambda, \omega, \epsilon)$ of K\"ahler metrics
$$\{\omega_{\phi}=\omega+\sqrt{-1}\partial\bar\partial \phi\;|\; Ric(\omega_{\phi})>-1+\delta, \; |Rm|(\omega_{\phi})\leq \Lambda,\; E_1(\omega_{\phi})\leq \inf E_1+\epsilon \},$$where $E_1(\omega')=E_{1, \omega}(\omega')$, the K\"ahler-Ricci flow will deform it exponentially fast to a K\"ahler-Einstein metric in the limit. \end{theo}
\begin{rem} The condition that $M$ is pre-stable (cf. Defi \ref {prestable}), roughly means that the complex structure doesn't jump in the limit (cf. \cite{[chen1]}, \cite{[PhSt]}). In G. Tian's definition of $K$-stability, this condition appears to be one of three necessary conditions for a complex structure to be $K$-stable (cf. \cite{[Do]}, \cite{[Tian2]}). \end{rem}
\begin{rem}This gives a sufficient condition for the existence of K\"ahler-Einstein metrics. More interestingly, by a theorem of G. Tian \cite{[Tian2]}, this also gives a sufficient condition for an algebraic manifold being weakly K-stable. One tempting question is: does this condition imply weak K-stability directly? \end{rem}
\begin{rem} If we call the result in \cite{[chen1]} a ``pre-baby" step in this ambitious program, then Theorem 1.1 and 1.5 should be viewed as a ``baby step" in this program. We wish to remove the assumption on the bound of the bisectional curvature. More importantly (cf. Theorem 1.8 below), we wish to replace the condition on the Ricci curvature in both Theorem 1.1 and 1.5 by a condition on the scalar curvature. Then our theorem really becomes a ``small energy" lemma. \end{rem}
If we remove the condition of ``pre-stable", then
\begin{theo}\label{main} Suppose that $(M, [\omega])$ has no nonzero holomorphic vector fields and $E_1$ is bounded from below in $[\omega].$ For any $\delta, B, \Lambda>0,$ there exists a small positive constant $\epsilon(\delta, B, \Lambda, \omega)>0$ such that for any metric $\omega_0$ in the subspace ${\mathcal A}(\delta, B, \Lambda, \epsilon)$ of K\"ahler metrics
$$\{\omega_{\phi}=\omega+\sqrt{-1}\partial\bar\partial \phi\;|\; Ric(\omega_{\phi})>-1+\delta,\; |\phi|
\leq B, \;|Rm|(\omega_{\phi})\leq \Lambda, \; E_1(\omega_{\phi})\leq \inf E_1+\epsilon \}$$ the K\"ahler-Ricci flow will deform it exponentially fast to a K\"ahler-Einstein metric in the limit. \end{theo}
\begin{rem} In light of Theorem \ref{main4} below, we can replace the condition on $E_1$ by a corresponding condition on $E_0.\;$ \end{rem}
\subsection{The relations between energy functionals $E_k$ }
Song-Weinkove \cite{[SoWe]} recently proved that $E_k$ have a lower bound
on the space of K\"ahler metrics with nonnegative Ricci curvature for K\"ahler-Einstein manifolds. Moreover, they also showed that modulo holomorphic vector fields, $E_1$
is proper if and only if there exists a K\"ahler-Einstein metric. Shortly afterwards, N. Pali
\cite{[Pali]} gave a formula between $E_1$ and the $K$-energy $E_0$, which says that the $E_1$ energy
is always bigger than the $K$-energy.
Tosatti \cite{[Tosatti]} proved that under some curvature assumptions,
the critical point of $E_k$ is a K\"ahler-Einstein metric. Pali's theorem means that $E_1$ has a lower bound if the $K$-energy has a lower bound.
A natural question is if the converse holds. To our own surprise, we proved the following result.
\begin{theo}\label{main4}$E_1$ is bounded from below if and only if the $K$-energy is bounded from below in the class $[\omega].$
Moreover, we have\footnote{For simplicity of notation, we will often drop the subscript $\phi$ and write $|\nabla f|^2$ for $|\nabla f|_{\phi}^2 $. But in an integral, $|\nabla f|^2$ is with respect to the metric of the volume form. } $$\inf_{\omega'\in [\omega]} E_{1, \omega}(\omega')=2\inf_{\omega'\in [\omega]} E_{0,
\omega}(\omega')-\frac 1{nV}\int_M\;|\nabla h_{\omega}|^2 \omega^n,$$ where $h_{\omega}$ is the Ricci potential function with respect to $\omega$. \end{theo}
A crucial observation leads to this theorem is
\begin{theo} Along the K\"ahler-Ricci flow, $E_1$ will decrease after finite time. \label{main3} \end{theo}
Theorem \ref{main4} and \ref{main3} of course lead more questions than they answer to. For instance, is the properness of $E_k$ equivalent to the properness of $E_l$ for $k\neq l$? More subtlely, are the corresponding notions of semi-stability ultimately equivalent to each other? Is there a preferred functional in this family, or a better linear combination of these $E_k$ functionals? The first named author genuinely believes that this observation opens the door for more interesting questions. \\
Another interesting question is the relation of $E_k$ with various notions of stability defined by algebraic conditions. Theorem 1.1 and 1.5 suggest an indirect link of these functionals $E_k$ and stability. According to A. Futaki \cite{[futaki04]}, these functionals may directly link to the asymptotic Chow semi-stability (note that the right hand side of (1.2) in \cite{[futaki04]} is precisely equal to $\frac{d E_k}{dt}$ if one takes $p=k+1$ and $\phi = c_1^k$, cf. Theorem 2.4 below). It is highly interesting to explore further in this direction.
\subsection{Main ideas of proofs of Theorem 1.1 and 1.5}
In \cite{[chen1]}, a topological condition is used to control the $L^2$ norm of the bisectional curvature once the Ricci curvature is controlled. Using the parabolic Moser iteration arguments, this gives a uniform bound on the full bisectional curvature. In the present paper, we need to find a new way to control the full bisectional curvature under the flow. The whole scheme of obtaining this uniform estimate on curvatures depends on two crucial steps and their dynamical interplay.\\
{\bf STEP 1:} The first step is to follow the approach of the celebrated work of Yau on the Calabi conjecture (cf. {\cite{[Cao]}}, \cite{[Yau]}). The key point here is to control the $C^0$ norm of the evolving K\"ahler potential $\phi(t)$, in particular, the growth of $ u = {{\partial\phi}\over {\partial t}}$ along the K\"ahler-Ricci flow. Note that $u$ satisfies \[ { {\partial u}\over {\partial t}} = \triangle_{\phi} u + u. \] Therefore, the crucial step is to control the first eigenvalue of the Laplacian operator (assuming the traceless Ricci form is controlled via an iteration process which we will describe as {\bf STEP} 2 below).
For our purpose, we need to show that the first eigenvalue of the evolving Laplacian operator is bigger than $1 + \gamma$ for some fixed $\gamma > 0.\;$ Such a problem already appeared in \cite{[chen0]} and \cite{[chen-tian2]} since the first eigenvalue of the Laplacian operator of K\"ahler-Einstein metrics is exactly $1.\;$ If $Aut_r(M, J)\neq 0,\;$ the uniqueness of K\"ahler-Einstein metrics implies that the dimension of the first eigenspace is fixed; while the vanishing of the Futaki invariant implies that $u(t)$ is essentially perpendicular to the first eigenspace of the evolving metrics. These are two crucial ingredients which allow one to squeeze out a small gap $\gamma$ on the first eigenvalue estimate. Following the approach in \cite {[chen0]} and \cite{[chen-tian2]}, we can show that $u$ decays exponentially. This in turn implies the $C^0$ bound on the evolving K\"ahler potential. Consequently, this leads to control of all derivatives of the evolving potential, in particular, the bisectional curvature. In summary, as long as we have control of the first eigenvalue, one controls the full bisectional curvature of the evolving K\"ahler metrics.
For Theorem 1.5, a crucial technique step is to use an estimate obtained in \cite{[chenhe]} on the Ricci curvature tensor.\\
{\bf STEP 2:} Here we follow the Moser iteration techniques which appeared in \cite{[chen1]}. Assuming that the full bisectional curvature is bounded by some large but fixed number, the norm of the traceless bisectional curvature and the traceless Ricci tensor both satisfy the following inequality: \[
{{\partial u}\over {\partial t}} \leq \triangle_{\phi} u + |Rm| u \leq
\triangle_{\phi} u + C\cdot u. \] If the curvature of the evolving metric is controlled in $L^p(p> n)$, then the smallness of the energy $E_1$ allows you to control the norm of the traceless Ricci tensor of the evolving K\"ahler metrics (cf. formula (\ref{2})). According to Theorem \ref{lem2.8} and \ref{theo4.18}, this will in turn give an improved estimate on the first eigenvalue in a slightly longer period, but perhaps without full uniform control of the bisectional curvature in the ``extra" time. However, this gives uniform control on the K\"ahler potential which in turn gives the bisectional curvature in the extended ``extra" time. We use the Moser iteration again to obtain sharper control on the norm of the traceless bisectional curvature.\\
Hence, the combination of the parabolic Moser iteration together with Yau's estimate, yields the desired global estimate. In comparison, the iteration process in \cite{[chen1]} is instructive
and more direct. The first named author believes that this approach is perhaps more important than the mild results we obtained there.
\subsection{The organization} This paper is roughly organized as follows: In Section 2, we review some basic facts in K\"ahler geometry and necessary information on the K\"ahler-Ricci flow. We also include some basic facts on the energy functionals $E_k$. In Section 3, we prove Theorem \ref{main4} and \ref{main3}. In Section 4, we prove several technical theorems on the K\"ahler-Ricci flow. The key results are the estimates of the first eigenvalue of the evolving Laplacian, which are proved in Section 4.3. In Section 5, 6 we prove Theorem \ref{main2} and Theorem \ref{main}.\\
\noindent {\bf Acknowledgements}: Part of this work was done while the second named author was visiting University of Wisconsin-Madison and he would like to express thanks for the hospitality. The second named author would also like to thank
Professor W. Y. Ding and X. H. Zhu for their help and encouragement. The first named author would like to thank Professor P. Li of his interest and encouragement in this project. The authors would like to thank the referees for numerous suggestions which helped to improve the presentation. \vskip 1cm \section{Setup and known results} \subsection{Setup of notations} Let $M$ be an $n$-dimensional compact K\"ahler manifold. A K\"ahler metric can be given by its K\"ahler form $\omega$ on $M$. In local coordinates $z_1, \cdots, z_n $, this $\omega$ is of the form \[ \omega = \sqrt{-1} \displaystyle \sum_{i,j=1}^n\;g_{i \overline{j}} d\,z^i\wedge d\,z^{\overline{j}} > 0, \] where $\{g_{i\overline {j}}\}$ is a positive definite Hermitian matrix function. The K\"ahler condition requires that $\omega$ is a closed positive (1,1)-form. In other words, the following holds \[ {{\partial g_{i \overline{k}}} \over {\partial z^{j}}} = {{\partial g_{j \overline{k}}} \over {\partial z^{i}}}\qquad {\rm and}\qquad {{\partial g_{k \overline{i}}} \over {\partial z^{\overline{j}}}} = {{\partial g_{k \overline{j}}} \over {\partial z^{\overline{i}}}}\qquad\forall\;i,j,k=1,2,\cdots, n. \] The K\"ahler metric corresponding to $\omega$ is given by \[ \sqrt{-1} \;\displaystyle \sum_1^n \; {g}_{\alpha \overline{\beta}} \; d\,z^{\alpha}\;\otimes d\, z^{ \overline{\beta}}. \] For simplicity, in the following, we will often denote by $\omega$ the corresponding K\"ahler metric. The K\"ahler class of $\omega$ is its cohomology class $[\omega]$ in $H^2(M,{\mathbb R}).\;$ By the Hodge theorem, any other K\"ahler metric in the same K\"ahler class is of the form \[ \omega_{\phi} = \omega + \sqrt{-1} \displaystyle \sum_{i,j=1}^n\; {{\partial^2 \phi}\over {\partial z^i \partial z^{\overline{j}}}} \;dz^i\wedge dz^{\bar j}
> 0 \] for some real valued function $\phi$ on $M.\;$ The functional space in which we are interested (often referred to as the space of K\"ahler potentials) is \[ {\cal P}(M,\omega) = \{ \phi\in C^{\infty}(M, {\mathbb R}) \;\mid\; \omega_{\phi} = \omega + \sqrt{-1} {\partial} \overline{\partial} \phi > 0\;\;{\rm on}\; M\}. \] Given a K\"ahler metric $\omega$, its volume form is \[ \omega^n = n!\;\left(\sqrt{-1} \right)^n \det\left(g_{i \overline{j}}\right) d\,z^1 \wedge d\,z^{\overline{1}}\wedge \cdots \wedge d\,z^n \wedge d \,z^{\overline{n}}. \] Its Christoffel symbols are given by \[ \Gamma^k_{i\,j} = \displaystyle \sum_{l=1}^n\;g^{k\overline{l}} {{\partial g_{i \overline{l}}} \over {\partial z^{j}}} ~~~{\rm and}~~~ \Gamma^{\overline{k}}_{\overline{i} \,\overline{j}} = \displaystyle \sum_{l=1}^n\;g^{\overline{k}l} {{\partial g_{l \overline{i}}} \over {\partial z^{\overline{j}}}}, \qquad\forall\;i,j,k=1,2,\cdots n. \] The curvature tensor is \[ R_{i \overline{j} k \overline{l}} = - {{\partial^2 g_{i \overline {j}}} \over {\partial z^{k} \partial z^{\overline{l}}}} + \displaystyle \sum_ {p,q=1}^n g^{p\overline{q}} {{\partial g_{i \overline{q}}} \over {\partial z^{k}}} {{\partial g_{p \overline{j}}} \over {\partial z^{\overline{l}}}}, \qquad\forall\;i,j,k,l=1,2,\cdots n. \] We say that $\omega$ is of nonnegative bisectional curvature if \[ R_{i \overline{j} k \overline{l}} v^i v^{\overline{j}} w^k w^ {\overline{l}}\geq 0 \] for all non-zero vectors $v$ and $w$ in the holomorphic tangent bundle of $M$. The bisectional curvature and the curvature tensor can be mutually determined. The Ricci curvature of $\omega$ is locally given by \[ R_{i \overline{j}} = - {{\partial}^2 \log \det (g_{k \overline{l}}) \over {\partial z_i \partial \bar z_j }} . \] So its Ricci curvature form is \[ {\rm Ric}({\omega}) = \sqrt{-1} \displaystyle \sum_{i,j=1}^n \;R_{i \overline{j}}\; d\,z^i\wedge d\,z^{\overline{j}} = -\sqrt{-1} \partial \overline {\partial} \log \;\det (g_{k \overline{l}}). \] It is a real, closed (1,1)-form. Recall that $[\omega]$ is called a canonical K\"ahler class if this Ricci form is cohomologous to $\lambda \,\omega$ for some constant $\lambda$\,. In our setting, we require $\lambda = 1.\;$
\subsection{The K\"ahler-Ricci flow}
Let us assume that the first Chern class $c_1(M)$ is positive. Choose an initial K\"ahler metric $\omega$ in $2\pi c_1(M).\;$ The normalized K\"ahler-Ricci flow (cf. \cite{[Ha82]}) on a K\"ahler manifold $M$ is of the form \begin{equation} {{\partial g_{i \overline{j}}} \over {\partial t }} = g_{i \overline{j}} - R_{i \overline{j}}, \qquad\forall\; i,\; j= 1,2,\cdots ,n. \label{eq:kahlerricciflow} \end{equation} It follows that on the level of K\"ahler potentials, the Ricci flow becomes \begin{equation} {{\partial \phi} \over {\partial t }} = \log {{\omega_{\phi}}^n \over {\omega}^n } + \phi - h_{\omega} , \label{eq:flowpotential} \end{equation} where $h_{\omega}$ is defined by \[ {\rm Ric}({\omega})- \omega = \sqrt{-1} \partial \overline{\partial} h_ {\omega}, \; {\rm and}\;\displaystyle \int_M\; (e^{h_{\omega}} - 1) {\omega}^n = 0. \] Then the evolution equation for bisectional curvature is
\begin{eqnarray}{{\partial }\over {\partial t}} R_{i \overline{j} k \overline{l}} & = & \bigtriangleup R_{i \overline{j} k \overline{l}} + R_{i \overline{j} p \overline{q}} R_{q \overline{p} k \overline{l}} - R_{i \overline{p} k \overline{q}} R_{p \overline{j} q \overline{l}} + R_{i \overline{l} p \overline{q}} R_{q \overline{p} k \overline{j}} + R_{i \overline{j} k \overline{l}} \nonumber\\ & & \;\;\; -{1\over 2} \left( R_{i \overline{p}}R_{p \overline{j} k \overline{l}} + R_{p \overline{j}}R_{i \overline{p} k \overline{l}} + R_{k \overline{p}}R_{i \overline{j} p \overline{l}} + R_{p \overline{l}}R_{i \overline{j} k \overline{p}} \right). \label{eq:evolutio of curvature1} \end{eqnarray} Here $\Delta$ is the Laplacian of the metric $g(t).$ The evolution equation for Ricci curvature and scalar curvature are \begin{eqnarray} {{\partial R_{i \bar j}}\over {\partial t}} & = & \triangle R_{i\bar j} + R_{i\bar j p \bar q} R_{q \bar p} -R_{i\bar p} R_{p \bar j},\label {eq:evolutio of curvature2}\\ {{\partial R}\over {\partial t}} & = & \triangle R + R_{i\bar j} R_{j\bar i}- R. \label{eq:evolutio of curvature3} \end{eqnarray}
For direct computations and using the evolving frames, we can obtain the following evolution equations for the bisectional curvature: \begin{equation} \pd {R_{i\bar jk\bar l}}{t} =\Delta R_{i\bar jk\bar l}- R_{i\bar jk\bar l}+R_{i \bar j m\bar n}R_{n\bar m k\bar l}-R_{i\bar m k\bar n}R_{m\bar j n\bar l}+R_{i\bar l m\bar n}R_{n\bar m k\bar l} \label{eq:evolution of curvature4} \end{equation}
As usual, the flow equation (\ref{eq:kahlerricciflow}) or (\ref{eq:flowpotential}) is referred to as the K\"ahler-Ricci flow on $M$. It was proved by Cao \cite{[Cao]}, who followed Yau's celebrated work \cite{[Yau]}, that the K\"ahler-Ricci flow exists globally for any smooth initial K\"ahler metric. It was proved by S. Bando \cite{[Bando]} in dimension 3 and N. Mok \cite{[Mok]} in all dimensions that the positivity of the bisectional curvature is preserved under the flow. In \cite{[chen-tian2]} and \cite{[chen-tian1]}, the first named author and G. Tian proved that the K\"ahler-Ricci flow, in a K\"ahler-Einstein manifold, initiated from a metric with positive bisectional curvature converges to a K\"ahler-Einstein metric with constant bisectional curvature. In unpublished work on the K\"ahler-Ricci flow, G. Perelman proved, along with other results, that the scalar curvature is always uniformly bounded.
\subsection{Energy functionals $E_k$} In \cite{[chen-tian2]}, a family of energy functionals $E_k (k=0, 1, 2,\cdots, n)$ was introduced and these functionals played an important role there. First, we recall the definitions of these functionals. \begin{defi} For any $k=0, 1, \cdots, n,$ we define a functional $E_k^0$ on ${\mathcal P}(M, \omega)$ by $$E_{k, \omega}^0(\phi)=\frac 1V\int_M\; \Big(\log \frac {\omega^n_{\phi}} {\omega^n}-h_{\omega}\Big)\Big( \sum_{i=0}^k\; Ric(\omega_{\phi})^i\wedge \omega^{k-i}\Big)\wedge \omega_{\phi}^{n-k}+\frac 1V\int_M\; h_{\omega}\Big( \sum_{i=0}^k\; Ric(\omega)^i\wedge \omega^{k-i}\Big)\wedge \omega^{n-k}.$$ \end{defi}
\begin{defi}For any $k=0, 1, \cdots, n$, we define $J_{k, \omega}$ as follows $$J_{k, \omega}(\phi)=-\frac {n-k}{V}\int_0^1\;\int_M\;\pd {\phi(t)}{t}(\omega_{\phi(t)} ^{k+1}-\omega^{k+1})\wedge \omega_{\phi(t)}^{n-k-1} \wedge dt,$$ where $\phi(t)(t\in [0, 1])$ is a path from $0$ to $\phi$ in ${\mathcal P}(M, \omega)$. \end{defi}
\begin{defi}For any $k=0, 1, \cdots, n,$ the functional $E_{k, \omega}$ is defined as follows $$E_{k, \omega}(\phi)=E_{k, \omega}^0(\phi)-J_{k, \omega}(\phi).$$ For simplicity, we will often drop the subscript $\omega$. \end{defi}
By direct computation, we have \begin{theo} For any $k=0, 1, 2,\cdots, n,$ we have \begin{eqnarray*} \frac {dE_k}{dt}=&&\frac {k+1}{V}\int_M\;\Delta_{\phi}\dot\phi Ric(\omega_{\phi})^k\wedge\omega_{\phi}^{n-k}\\&&-\frac {n-k}{V}\int_M\;\dot\phi (Ric(\omega_{\phi})^{k+1}-\omega_{\phi}^{k+1})\wedge \omega_{\phi}^{n-k-1}. \end{eqnarray*} Here $\phi(t)$ is any path in ${\mathcal P}(M, \omega)$. \end{theo}
\begin{rem}Note that $$ \frac {dE_0}{dt}=-\frac nV\int_M\; \dot\phi (Ric(\omega_{\phi})-\omega_{\phi})\wedge \omega_{\phi}^{n-1}.$$ Thus, $E_0$ is the well-known $K$-energy. \end{rem} \begin{theo} Along the K\"ahler-Ricci flow where $Ric(\omega_{\phi})>-\omega_{\phi}$ is preserved, we have $$\frac {dE_k}{dt}\leq-\frac {k+1}{V}\int_M\;(R(\omega_{\phi})-r) Ric(\omega_{\phi})^k\wedge\omega_{\phi}^{n-k}.$$
When $k=0, 1,$ we have \begin{eqnarray*}
\frac {dE_0}{dt}&=&-\frac 1V\int_M\;|\nabla \dot\phi|^2\omega_{\phi}^n \leq 0,\\ \frac {dE_1}{dt}&\leq&-\frac 2V\int_M\;(R(\omega_{\phi})-r)^2\omega_{\phi}^n\leq 0. \end{eqnarray*} \end{theo}
Recently, Pali in \cite{[Pali]} found the following formula, which will be used in this paper. \begin{theo}\label{pali}For any $\phi\in {\mathcal P}(M, \omega)$, we have
$$E_{1, \omega}(\phi)=2E_{0, \omega}(\phi)+\frac 1{nV}\int_M\; |\nabla u|^2
\omega_{\phi}^n -\frac 1{nV}\int_M\; |\nabla h_{\omega}|^2 \omega^n,$$ where $$u=\log\frac {\omega^n_{\phi}}{\omega^n}+\phi-h_{\omega}.$$ \end{theo}
\begin{rem}This formula directly implies that if $E_0$ is bounded from below, then $E_1$ is bounded from below on ${\mathcal P}(M, \omega)$. \end{rem}
\begin{rem}In a forthcoming paper \cite{[chenli]}, the second named author will generalize Theorem \ref{pali} to all the functionals $E_k(k\geq 1)$, and discuss some interesting relations between $E_k.$ \end{rem} \section{Energy functionals $E_0$ and $E_1$} In this section, we want to prove Theorem \ref{main4} and \ref{main3}.
\subsection{Energy functional $E_1$ along the K\"ahler-Ricci flow} The following theorem is well-known in literature (cf. \cite{[chen2]}).
\begin{lem}\label{sec2}The minimum of the scalar curvature along the K\"ahler-Ricci flow, if negative, will increase to zero exponentially. \end{lem} \begin{proof} Let $\mu(t)=-\min_M R(x, 0)e^{-t},$ then
\begin{eqnarray*}\pd {}t(R+\mu(t))&=&\Delta (R+\mu(t))+|Ric|^2-(R+\mu(t))\\ &\geq &\Delta (R+\mu(t))-(R+\mu(t)). \end{eqnarray*} Since $R(x, 0)+\mu(0)\geq 0$, by the maximum principle we have $$R(x, t)\geq -\mu(t)=\min_M R(x, 0)e^{-t}.$$ \end{proof}
Using the above lemma, the following theorem is an easy corollary of Pali's formula. \begin{theo}\label{thm3.2}Along the K\"ahler-Ricci flow, $E_1$ will decrease after finite time. In particular, if the initial scalar curvature $R(0)>-n+1$, then there is a small constant $\delta>0$ depending only on
$n$ and $\min_{x\in M} R(0)$ such that for all time $t>0$, we have
\begin{equation} \frac {d}{dt}E_1\leq -\frac {\delta}V \int_M\;|\nabla \dot \phi|^2 \omega_{\phi}^n\leq
0.\label{aaa}\end{equation}
\end{theo}
\begin{proof} Along the K\"ahler-Ricci flow, the evolution equation for $|\nabla \dot \phi|^2$ is
$$\pd {}t|\nabla \dot \phi|^2=\Delta_{\phi} |\nabla \dot \phi|^2-|\nabla \nabla \dot
\phi|^2-|\nabla \bar \nabla\dot \phi|^2+|\nabla \dot \phi|^2.$$ By Theorem
\ref{pali}, we have \begin{eqnarray*}\frac {d}{dt}E_1&=&-\frac 2V\int_M\;|\nabla
\dot \phi|^2\omega_{\phi}
^n +\frac 1{nV}\frac {d}{dt}\int_M\; |\nabla \dot \phi|^2\omega_{\phi}^n\\
&= &-\frac 2V\int_M\;|\nabla \dot \phi|^2\omega_{\phi}^n +\frac 1{nV}\int_M\;(-|\nabla \nabla \dot \phi|^2-|\nabla \bar \nabla\dot
\phi|^2+|\nabla \dot \phi|^2+|\nabla \dot \phi|^2(n-R))\omega_{\phi}^n \end{eqnarray*} If the scalar curvature at the initial time $R(x, 0)\geq -n+1+n\delta (n\geq 2)$ for some small $\delta>0$, by Lemma \ref{sec2} for all time $t>0$ we have $R(x, t)\geq -n+1+n\delta.$ Then we have
\begin{eqnarray}\frac {d}{dt}E_1 &\leq&-\frac 2V\int_M\;|\nabla \dot
\phi|^2\omega_{\phi}^n +\frac 1{nV}\int_M\;(-|\nabla \nabla \dot \phi|^2-
|\nabla \bar \nabla\dot \phi|^2+|\nabla \dot \phi|^2+|\nabla \dot
\phi|^2(2n-1-n \delta))\omega_{\phi}^n\nonumber\\
&\leq&-\frac {\delta}V \int_M\;|\nabla \dot \phi|^2\omega_{\phi}^n. \label{aeq1}\end{eqnarray} Otherwise, by Lemma \ref{sec2} after finite time $T_0=\log\frac {-\min_M R(x, 0)}{n-1-n\delta},$ we still have $R(x, t)>-n+1+n\delta$ for small $\delta
>0.$ Thus, the inequality (\ref{aeq1}) holds.
If $n=1$, by direct calculation we have $$\frac {dE_1}{dt}=-\frac 2V\int_M\; (R(\omega_{\phi})-1)^2\omega_{\phi}=-\frac 2V\int_M\; (\Delta_{\phi}\dot\phi)^2\omega_{\phi}.$$ If the initial scalar curvature $R(0)>0$, by R. Hamilton's results in \cite{[Ha88]} the scalar curvature has a uniformly positive lower bound. Thus, $R(t)\geq c>0$ for some constant $c>0$ and all $t>0.$ Therefore, by the proof of Lemma \ref{lem2.13} in section \ref{section4.4.1} the first eigenvalue of $\Delta_{\phi}$ satisfies $\lambda_1(t)\geq c.$ Then $$\frac {dE_1}{dt}=-\frac 2V\int_M\; (\Delta_{\phi}\dot\phi)^2\omega_
{\phi}\leq -\frac {2c}V\int_M\; |\nabla\dot\phi|^2\omega_{\phi}.$$ The theorem is proved.
\end{proof}
\subsection{On the lower bound of $E_0$ and $E_1$} In this section, we will prove Theorem \ref{main4}. Recall the generalized energy: \begin{eqnarray*} I_{\omega}(\phi)&=&\frac 1V\sum_{i=0}^{n-1}\int_M\; \sqrt{-1} \partial \phi\wedge\bar \partial \phi\wedge \omega^i\wedge\omega_{\phi}^ {n-1-i},\\ J_{\omega}(\phi)&=&\frac 1V\sum_{i=0}^{n-1}\frac {i+1}{n+1}\int_M\; \sqrt{-1}\partial \phi\wedge\bar \partial \phi\wedge \omega^i\wedge\omega_{\phi}^{n-1-i}. \end{eqnarray*} By direct calculation, we can prove $$0\leq I-J\leq I\leq (n+1)(I-J)$$ and for any K\"ahler potential $\phi(t)$ $$\frac {d}{dt}(I_{\omega}-J_{\omega})(\phi(t))=-\frac 1V\int_M\;\phi\Delta_{\phi} \dot\phi \;\omega_{\phi}^n.$$
The behaviour of $E_1$ for the family of K\"ahler potentials $\phi(t)$ satisfying the equation (\ref{eq1}) below has been studied by Song and Weinkove in \cite{[SoWe]}. Following their ideas, we have the following lemma.
\begin{lem}\label{lema3.3}For any K\"ahler metric $\omega_0\in[\omega]$, there exists a K\"ahler metric $\omega_0'\in [\omega]$ such that $Ric(\omega_0')>0$ and $$E_0(\omega_0)\geq E_0(\omega_0').$$ \end{lem}
\begin{proof}We consider the complex Monge-Amp\`{e}re equation \begin{equation}(\omega_0+\sqrt{-1}\partial\bar\partial \varphi)^n=e^{th_{0}+c_t}\omega_0^n,\label{eq1}\end{equation} where $h_{0}$ satisfies the following equation $$Ric(\omega_0)-\omega_0=\sqrt{-1}\partial\bar\partial h_{0},\qquad \frac 1V\int_M\;e^{h_0} \omega_0^n=1$$ and $c_t$ is the constant chosen so that $$\int_M\; e^{th_0+c_t}\omega_0^n=V.$$ By Yau's results in \cite{[Yau]}, there exists a unique $\varphi(t)(t\in [0, 1])$ to the equation (\ref{eq1}) with $\int_M\; \varphi \omega_0^n=0.$ Then $\varphi(0)=0.$ Note that the equation (\ref{eq1}) implies \begin{equation} Ric(\omega_{\varphi})=\omega_{\varphi}+(1-t)\sqrt{-1}\partial\bar\partial h_0-\sqrt{-1}\partial\bar\partial\varphi,\label{eq2}\end{equation} and $$\Delta_{\varphi}\dot\varphi=h_0+c_t'.$$ By the definition of $E_0$ we have \begin{eqnarray*} \frac {d}{dt}E_0(\varphi(t))&=&-\frac 1V\int_M\; \dot\varphi(R(\omega_{\varphi})-n)\omega_{\varphi}^n\\ &=&-\frac 1V\int_M\; \dot\varphi((1-t)\Delta_{\varphi}h_0-\Delta_{\varphi}\varphi)\omega_ {\varphi}^n\\ &=&-\frac {1-t}{V}\int_M\;\Delta_{\varphi} \dot\varphi h_0\; \omega_{\varphi}^n+\frac 1V\int_M\;\varphi\Delta_{\varphi}\dot\varphi\omega_{\varphi}^n\\ &=&-\frac {1-t}{V}\int_M\;(\Delta_{\varphi} \dot\varphi)^2\;\omega_ {\varphi}^n-\frac {d}{dt}(I-J)_{\omega_0}(\varphi). \end{eqnarray*} Integrating the above formula from $0$ to $1$, we have $$E_0(\varphi(1))-E_0(\omega_0)=-\frac 1V\int_0^1(1-s)\int_M\;(\Delta_ {\varphi} \dot\varphi)^2\; \omega_{\varphi}^n\wedge ds-(I-J)_{\omega_0}(\varphi(1))\leq 0.$$ By the equation (\ref{eq2}), we know $Ric(\omega_{\varphi(1)})> 0$. This proves the lemma. \end{proof}
Now we can prove Theorem \ref{main4}. \begin{theo}$E_1$ is bounded from below if and only if the $K$-energy is bounded from below in the class $[\omega].$ Moreover, we have $$\inf_{\omega'\in [\omega]} E_{1}(\omega')=2\inf_{\omega'\in [\omega]} E_{0}(\omega')-
\frac 1{nV}\int_M\; |\nabla h_{\omega}|^2\omega^n.$$ \end{theo}
\begin{proof}It is sufficient to show that if $E_1$ is bounded from below, then $E_0$ is bounded from below. For any K\"ahler metric $\omega_0$, by Lemma \ref{lema3.3} there exists a K\"ahler metric $\omega_0'=\omega+\sqrt{-1}\partial\bar\partial \varphi_0$ such that $$Ric(\omega_0')\geq c>0,\qquad E_0(\omega_0)\geq E_0(\omega_0'),$$ where $c$ is a constant depending only on $\omega_0.$ Let $\varphi(t)$ be the solution to the K\"ahler-Ricci flow with the initial metric $\omega_0',$ $$\pd {\varphi}{t}=\log\frac {\omega_{\varphi}^n}{\omega^n}+\varphi-h_{\omega}, \qquad \varphi(0)=\varphi_0.$$ Then for any $t>s\geq0$, by Theorem \ref{thm3.2} we have \begin{equation} E_1(t)-E_1(s)\leq 2\delta(E_0(t)-E_0(s)),\label{qeq3}\end{equation} where $E_1(t)=E_1(\omega, \omega_{\varphi(t)})$ and $\delta=\frac {n-1}{2n}$ if $n\geq 2,$ or $\delta=c>0 $ if $n=1$. Here $c$ is a constant obtained in the proof of Theorem \ref{thm3.2}. By Theorem \ref{pali} we have
$$E_1(t)-2E_0(s)-\frac 1{nV}\int_M\;|\nabla\dot\varphi|^2\omega_{\varphi}^n
(s)+C_{\omega}\leq \delta(E_1(t) -\frac 1{nV}\int_M\;|\nabla\dot\varphi|^2\omega_{\varphi}^n(t)+C_{\omega})-2\delta E_0(s).$$ i.e. \begin{equation} E_1(t)-\frac 1{n(1-\delta)V}\int_M\;|\nabla\dot\varphi|^2\omega_{\varphi}^n(s)+ \frac
{\delta}{n(1-\delta)V}\int_M\;|\nabla\dot\varphi|^2\omega_{\varphi}^n(t)+C_{\omega}\leq 2E_0(s),\label{qeq4}\end{equation} where $C_{\omega}=\frac 1{nV}\int_M\; |\nabla h_{\omega}|^2\omega^n.$ By (\ref{qeq3}) we know $E_0$ is bounded from below along the K\"ahler-Ricci flow. Thus there exists a sequence of times $t_m$ such that
$$\int_M\;|\nabla\dot\varphi|^2\omega_{\varphi}^n(t_m)\rightarrow 0,\qquad m\rightarrow \infty.$$ We choose $t=t_m$ and let $m\rightarrow \infty$ in (\ref{qeq4}),
$$\inf E_1-\frac 1{n(1-\delta)V}\int_M\;|\nabla\dot\varphi|^2\omega_{\varphi}^n(s)+C_{\omega}\leq 2E_0(s)\leq 2E_0(\omega_0'), $$ where the last inequality is because $E_0$ is decreasing along the K\"ahler-Ricci flow. Thus, we choose $s=t_m$ again and let $m\rightarrow \infty$, $$\inf E_1+C_{\omega}\leq 2E_0(\omega_0')\leq 2E_0(\omega_0).$$ Thus, $E_0$ is bounded from below in $[\omega]$, and $$\inf E_1\leq 2\inf E_0-C_{\omega}.$$ On the other hand, for any $\omega'\in [\omega]$ we have $$E_1(\omega')\geq 2E_0(\omega')-C_{\omega}.$$ Combining the last two inequalities, we have $\inf E_1=2\inf E_0-C_{\omega}$. Thus, the theorem is proved. \end{proof}
\section{Some technical lemmas} In this section, we will prove some technical lemmas, which will be used in the proof of Theorem \ref{main2} and \ref{main}. These lemmas are based on the K\"ahler-Ricci flow $$\pd {\phi}{t}=\log\frac {\omega_{\phi}^n}{\omega^n}+\phi-h_{\omega}.$$ Most of these results are taken from \cite{[chen1]}-\cite{[chen-tian1]}. The readers are referred to these papers for the details. Here we will prove some of them for completeness.
\subsection{Estimates of the Ricci curvature} The following result shows that we can control the curvature tensor in a short time.
\begin{lem}\label{lem2.1}(cf. \cite{[chen1]}) Suppose that for some $\delta>0$, the curvature of $\omega_0=\omega+\sqrt{-1}\partial\bar\partial \phi(0)$ satisfies the following conditions
$$\left \{\begin{array}{lll}|Rm|(0)&\leq& \Lambda,\\ R_{i\bar j}(0)&\geq& -1+\delta. \end{array}\right. $$ Then there exists a constant $T(\delta, \Lambda)>0$, such that for the evolving K\"ahler metric $\omega_t(0\leq t\leq 6T)$, we have the following
\begin{equation} \left \{\begin{array}{lll}|Rm|(t)&\leq& 2\Lambda,\\ R_{i\bar j}(t)&\geq& -1+\frac {\delta}2. \end{array}\right. \label{1}\end{equation} \end{lem}
\begin{lem}
\label{lem2.2}(cf. \cite{[chen1]})If $E_1(0)\leq \inf_{\omega'\in [\omega] }E_1(\omega')+\epsilon,$ and $$Ric(t)+\omega(t)\geq \frac {\delta}2>0,\qquad \forall t\in
[0, T],$$ then along the K\"ahler-Ricci flow we have \begin{equation}\frac 1V\int_0^{T}\int_M\;\;|Ric-\omega|^2(t)\omega_ {\phi}^n\wedge dt\leq \frac {\epsilon}2.\label{2}\end{equation} \end{lem}
Since we have the estimate of the Ricci curvature, the following theorem shows that the Sobolev constant is uniformly bounded if $E_1$ is small.
\begin{prop}\label{lem2.3}(cf. \cite{[chen1]}) Along the K\"ahler-Ricci flow, if $E_1(0)\leq \inf_{\tilde\omega\in [\omega] }E_1(\tilde\omega)+\epsilon,$ and for any $t\in [0, T],$ $$Ric(t)+\omega(t)\geq 0, $$ the diameter of the evolving metric $\omega_{\phi}$ is uniformly bounded for $t\in [0, T].$ As $\epsilon\rightarrow 0, $ we have $D\rightarrow \pi.$ Let $\sigma(\epsilon)$ be the maximum of the Sobolev and Poincar\'e constants with respect to the metric $\omega_{\phi}$. As $\epsilon\rightarrow 0,$ we have $\sigma(\epsilon)\leq \sigma<+\infty.$ Here $\sigma$ is a constant independent of $\epsilon.$ \end{prop}
Next we state a parabolic version of Moser iteration argument (cf. \cite{[chen-tian1]}).
\begin{prop}\label{lem2.17} Suppose the Sobolev and Poincar\'e constants of the evolving K\"ahler metrics $g(t)$ are both uniformly bounded by $\sigma$. If a nonnegative function $u$ satisfies the following inequality $$\pd {}tu\leq \Delta_{\phi} u+f(t, x)u, \;\; \forall \,t\in (a, b),$$
where $|f|_{L^p(M, g(t))}$ is uniformly bounded by some constant $c$ for some $p>\frac m2 $, where $m=2n=\dim_{{\mathbb R}}M$, then for any $\tau\in (0, b-a)$ and any $t\in (a+\tau, b)$, we have\footnote{The constant $C$ may differ from line to line. The notation $C(A, B, ...)$ means that the constant $C$ depends only on $A, B, ...$.} $$u(t)\leq \frac {C(n, \sigma, c)}{\tau^{\frac {m+2}{4}}}\Big(\int_{t- \tau}^t\int_M \;u^2 \,\omega_{\phi}^n\wedge ds\Big)^{\frac 12}.$$ \end{prop}
By the above Moser iteration, we can show the following lemma. \begin{lem}\label{lem2.5} For any $\delta, \Lambda>0,$ there exists a small positive constant $\epsilon(\delta, \Lambda)>0$ such that if the initial metric $\omega_0$ satisfies the following condition: \begin{equation} Ric(0)>-1+\delta,\;
|Rm(0)|\leq \Lambda, \;E_1(0)\leq\inf E_1+\epsilon,\label{z1}\end{equation} then after time $2T$ along the K\"ahler-Ricci flow, we have
\begin{equation}|Ric-\omega|(t)\leq C_1(T, \Lambda)\epsilon, \qquad\forall t\in [2T, 6T]\label{2.5}\end{equation} and \begin{equation}|\dot\phi-c(t)|_{C^0}\leq C(\sigma)C_1(T, \Lambda)\epsilon, \qquad\forall t \in [2T, 6T],\label{z2}\end{equation} where $c(t)$ is the average of $\dot\phi$ with respect to the metric $g(t)$, and $\sigma$ is the uniformly upper bound of the Sobolev and Poincar\'e constants in Proposition \ref{lem2.3}. \end{lem}
\begin{proof} Let $Ric^0=Ric-\omega.$ Then $u=|Ric^0|^2(t)$ satisfies the parabolic inequality
$$\pd ut\leq \Delta_{\phi} u+c(n)|Rm|_{g(t)}u,$$
Note that by Lemma \ref{lem2.1}, $|Rm|(t)\leq 2\Lambda,$ for $0\leq t\leq 6T.$ Then applying Lemma \ref{lem2.1} again and Lemma \ref{lem2.2} for $t\in [2T, 6T]$, we have \footnote{Since the volume $V$ of the K\"ahler manifold $M$ is fixed for the metrics in the same K\"ahler class, the constant $C(T, \Lambda)$ below should depend on $V$, but we don't specify this for simplicity.} \begin{eqnarray*}
|Ric^0|^2(t)&\leq &C(\Lambda, T)\Big(\int_{0}^{6T}\int_M\;\;|Ric-\omega|^4(t)\omega_{\phi}^n\wedge dt\Big)^{\frac 12}\\
&\leq &C(\Lambda, T)(1+\Lambda)\Big(\int_{0}^{6T}\int_M\;\;|Ric-\omega
|^2(t)\omega_{\phi}^n\wedge dt\Big)^{\frac 12}\\
&\leq &C(\Lambda, T)\sqrt{\epsilon}. \end{eqnarray*} Thus, \begin{equation}|Ric-\omega|(t)\leq C(\Lambda, T)\epsilon^{\frac 14}. \label{z3}\end{equation} Recall that $\Delta_{\phi} \dot\phi=n-R(\omega_{\phi})$, by the above estimate and Proposition
\ref{lem2.3} we have \begin{equation}|\dot\phi-c(t)|_{C^0}\leq C(\sigma)C(T, \Lambda)\epsilon^{\frac 14}, \forall t\in [2T, 6T].\label{z4}\end{equation} For simplicity, we can write $\epsilon^{\frac 14}$ in the inequalities (\ref{z3}) and (\ref{z4}) as $\epsilon$, since we can assume $E_1(0)\leq \inf E_1+\epsilon^4$ in the assumption. The lemma is proved. \end{proof}
\subsection{Estimate of the average of $\pd {\phi}t$} In this section, we want to control $c(t)=\frac 1V\int_M\;\dot \phi \omega^n_{\phi}$. Here we follow the argument in \cite{[chen-tian2]}. Notice that the argument essentially needs the lower bound of the $K$-energy, which can be obtained by Theorem \ref{main4} in our case. Observe that for any solution $\phi(t)$ of the K\"ahler-Ricci flow, $$\pd {\phi}t=\log\frac {\omega_{\phi}^n}{\omega^n}+\phi-h_{\omega},$$
the function $\tilde\phi(t)=\phi(t)+Ce^t$ also satisfies the above equation for any constant $C$. Since $$\pd {\tilde \phi}{t}(0)=\pd {\phi}{t}(0)+C,$$ we have $\tilde c(0)=c(0)+C.$ Thus we can normalize the solution $\phi(t)$ such that the average of $\dot\phi(0)$ is any given constant.
The proof of the following lemma will be used in section 5 and 6, so we include a proof here.
\begin{lem}\label{lem2.6}(cf. \cite{[chen-tian2]})Suppose that the $K $-energy is bounded from below along the K\"ahler-Ricci flow. Then we can normalize the solution $\phi(t)$ so that
$$c(0)=\frac 1V\int_0^{\infty}\;e^{-t}\int_M\;|\nabla \dot\phi|^2\omega_ {\phi}^n\wedge dt<\infty. $$ Then for all time $t>0$, we have $$0<c(t),\;\;\int_0^{\infty}\;c(t)dt<E_0(0)-E_0(\infty),$$ where $E_0(\infty)=\lim_{t\rightarrow \infty}E_0(t)$. \end{lem} \begin{proof} A simple calculation yields
$$c'(t)=c(t)-\frac 1V\int_M\;|\nabla \dot\phi|^2\omega_{\phi}^n.$$ Define
$$\epsilon(t)=\frac 1V\int_M\;|\nabla \dot\phi|^2\omega_{\phi}^n.$$ Since the $K$-energy has a lower bound along the K\"ahler-Ricci flow, we have
$$\int_0^{\infty}\;\epsilon(t)dt=\frac 1V\int_0^{\infty}\int_M\;|\nabla \dot
\phi|^2\omega_{\phi}^n\wedge dt= E_0(0)-E_0(\infty).$$ Now we normalize our initial value of $c(t)$ as \begin{eqnarray*} c(0)&=&\int_0^{\infty}\;\epsilon(t)e^{-t}dt\\
&=&\frac 1V\int_0^{\infty}\;e^{-t}\int_M\;|\nabla
\dot\phi|^2\omega_{\phi}^n \wedge dt\\
&\leq &\frac 1V\int_0^{\infty}\int_M\;|\nabla \dot\phi|^2\omega_ {\phi}^n\wedge dt\\ &= &E_0(0)-E_0(\infty). \end{eqnarray*} From the equation for $c(t)$, we have $$(e^{-t}c(t))'=-\epsilon(t)e^{-t}.$$ Thus, we have \beqs0<c(t)=\int^{\infty}_t \;\epsilon(\tau)e^{-(\tau-t)}d\tau \leq E_0(0)-E_0(\infty) \end{eqnarray*} and $$\lim_{t\rightarrow \infty}c(t)=\lim_{t\rightarrow \infty}\int^{\infty}_t \;\epsilon(\tau) e^{-(\tau-t)}d\tau=0.$$ Since the $K$-energy is bounded from below, we have
$$\int_0^{\infty}\;c(t)dt=\frac 1V\int_0^{\infty}\int_M\;|\nabla \dot
\phi|^2\omega_{\phi}^n\wedge dt-c(0)\leq E_0(0)-E_0(\infty).$$ \end{proof}
\begin{lem}\label{lem5.9}Suppose that $E_1$ is bounded from below on ${\mathcal P}(M, \omega)$. For any solution $\phi(t)$ of the K\"ahler-Ricci flow with the initial metric $\omega_0$ satisfying $$E_1(0)\leq \inf E_1+\epsilon, $$ after normalization for the K\"ahler potential $\phi(t)$ of the solution, we have $$0<c(t),\;\;\int_0^{\infty}c(t)\omega^n_{\phi}\leq \frac {\epsilon}2.$$ \end{lem}
\begin{proof}By Theorem \ref{main4}, the $K$-energy is bounded from below, then one can find a sequence of times $t_m\rightarrow \infty$ such that $$\int_M\; |\nabla \dot\phi|^2\omega^n_{\phi}\Big|_{t=t_m}\rightarrow 0.$$ By Theorem \ref{pali}, we have
$$E_1(t)=2E_0(t)+\frac 1V\int_M\; |\nabla\dot\phi|^2\omega_{\phi}^n-C_{\omega}.$$
Then \begin{eqnarray*} 2(E_0(0)-E_0(t_m))&=&E_1(0)-E_{1}(t_m)-\frac 1V\int_M\;|\nabla\dot\phi|^2\omega_ {\phi}^n\Big|_{t=0}+ \frac 1V\int_M\;|\nabla\dot\phi|^2\omega_{\phi}^n\Big|_{t=t_m}\\&\leq&\epsilon+\frac 1V\int_M\;| \nabla\dot\phi|^2\omega_{\phi}^n\Big|_{t=t_m}\\&\rightarrow&\epsilon.\end{eqnarray*} Since the $K$-energy is decreasing along the K\"ahler-Ricci flow, we have $$E_0(0)-E_0(\infty)\leq \frac {\epsilon}2.$$ By the proof of Lemma \ref{lem2.6}, for any solution of the K\"ahler-Ricci flow we can normalize $\phi(t)$ such that $$0<c(t),\;\;\int_0^{\infty}c(t)\omega^n_{\phi}\leq E_0(0)-E_0(\infty) \leq \frac {\epsilon}2. $$ The lemma is proved. \end{proof}
\subsection{Estimate of the first eigenvalue of the Laplacian operator} \subsubsection{Case 1: $M$ has no nonzero holomorphic vector fields}\label{section4.4.1} In this subsection, we will estimate the first eigenvalue of the Laplacian when $M$ has no nonzero holomorphic vector fields. In order to show that the norms of $\phi$ decay exponentially in section 4.5, we need to prove that the first eigenvalue is strictly greater than $1$.
\begin{theo}\label{lem2.8}Assume that $M$ has no nonzero holomorphic vector fields. For any $A, B>0$, there exist $\eta(A, B, \omega)>0$ such that for any metric $\omega_{\phi}=\omega+\sqrt{-1}\partial\bar\partial \phi, $ if $$-\eta \omega_{\phi}\leq Ric(\omega_{\phi})-\omega_{\phi}\leq A\omega_{\phi}\; \;\;{ and}\;\;\;
|\phi|\leq B, $$ then the first eigenvalue of the Laplacian $\Delta_{\phi}$ satisfies $$\lambda_1>1+\gamma(\eta, B, A, \omega),$$ where $\gamma>0$ depends only on $\eta, B, A$ and the background metric $ \omega$. \end{theo}
The following lemma is taken from \cite{[chen-tian2]}. \begin{lem}\label{lem2.9}(cf. \cite{[chen-tian2]}) If the K\"ahler metric $\omega_{\phi}$ satisfies
$$Ric(\omega_{\phi})\geq \alpha\omega_{\phi},\; \;\;{ and}\;\;\; |\phi|\leq B$$ for two constants $\alpha$ and $B,$ then there exists a uniform constant $C$ depending only on $\alpha, B$ and $\omega$ such that $$\inf_M \log \frac {\omega_{\phi}^n}{\omega^n}(x)\geq -4C(\alpha, B, \Lambda) e^{2 (1+\int_M\; \log \frac {\omega_{\phi}^n}{\omega^n}\omega^n_{\phi})}.$$ \end{lem}
The following crucial lemma is taken from Chen-He \cite{[chenhe]}. Here we include a proof. \begin{lem}\label{lem2.10}For any constant $A, B>0$, if
$|Ric(\omega_{\phi})|\leq A$ and $|\phi|\leq B,$ then there is a constant $C$ depending only on $A, B$ and the background metric
$\omega$ such that $|\phi|_{C^{3, \beta}(M, \omega)}\leq C(A, B, \omega, \beta)$ for any $\beta\in (0, 1)$. In particular, one can find two constants $C_2(A, B, \omega)$ and $C_3(A, B, \omega)$ such that $$C_2(A, B, \omega)\omega\leq \omega_{\phi}\leq C_3(A, B, \omega)\omega.$$
\end{lem}
\begin{proof} We use Yau's estimate on complex Monge-Amp\`ere equation to obtain the $C^{3, \beta}$ norm of $|\phi|.$ Let $F=\log \frac {\omega^n_{\phi}}{\omega^n}.$ Then we have \begin{eqnarray*} \Delta_{\omega} F=g^{i\bar j}\partial_i\partial_{\bar j}\log \frac {\omega^n_{\phi}}{\omega^n} =-g^{i\bar j}R_{i\bar j}(\phi)+R(\omega), \end{eqnarray*} where $\Delta_{\omega}$ denotes the Laplacian of $\omega$. On the other hand, we choose normal coordinates at a point such that $ g_{i\bar j}=\delta_{ij}$ and $g_{i\bar j}(\phi)=\lambda_i\delta_{ij},$ then \begin{eqnarray*} g^{i\bar j}R_{i\bar j}({\phi})=\sum_i R_{i\bar i}(\phi) \leq A\sum_i g_{i\bar i}(\phi) =A(n+\Delta_{\omega} \phi) \end{eqnarray*} and $$g^{i\bar j}R_{i\bar j}(\phi)\geq -A(n+\Delta_{\omega}\phi).$$ Hence, we have \begin{eqnarray} \Delta_{\omega} (F-A\phi)&\leq &R(\omega)+An\label{f1}\\ \Delta_{\omega} (F+A\phi)&\geq &R(\omega)-An. \label{f2}\end{eqnarray} Appling the Green formula, we can bound $F$ from above. In fact, \begin{eqnarray*} F+A\phi&\leq &\frac 1V\int_M\; -G(x, y)\Delta_{\omega} (F+A\phi)(y)\omega^n (y)+\frac 1V\int_M\;(F+A\phi) \omega^n\\ &\leq &\frac 1V\int_M\; -G(x, y)(R(\omega)-An)\omega^n(y)+\frac 1V\int_M\; (F+A\phi) \omega^n\\ &\leq &C(\Lambda, A, B), \end{eqnarray*} where $\Lambda$ is an upper bound of
$|Rm|_{\omega}$. Notice that in the last inequality we used $$\frac 1V\int_M\; F\omega^n\leq \log \Big(\frac 1V\int_M\; e^F \omega^n\Big)=0.$$ Hence, $F\leq C(\Lambda, A, B).$ Consider complex Monge-Amp\`ere equation \begin{equation}(\omega+\sqrt{-1}\partial\bar\partial\phi)^n=e^F\omega^n,\label{f3}\end{equation} by Yau's estimate we have \begin{eqnarray*} \Delta_{\phi}(e^{-k\phi}(n+\Delta_{\omega}\phi))&\geq & e^{-k\phi} (\Delta_{\omega} F-n^2\inf_{i\neq j}R_{i\bar ij\bar j}(\omega))\\ &-&ke^{-k\phi}n(n+\Delta_{\omega}\phi)+(k+\inf_{i\neq j}R_{i\bar ij\bar j}(\omega))e^{-k\phi+ \frac {-F}{n-1}}(n+\Delta_{\omega}\phi)^{1+\frac 1{n-1}}\\ &\geq &e^{-k\phi}(R(\omega)-An-\Delta_{\omega}\phi-n^2\inf_{i\neq j}R_{i\bar ij \bar j}(\omega))\\ &-&ke^{-k\phi}n(n+\Delta_{\omega}\phi)+(k+\inf_{i\neq j}R_{i\bar ij\bar j} (\omega))e^{-k\phi+ \frac {-F}{n-1}}(n+\Delta_{\omega}\phi)^{1+\frac 1{n-1}}. \end{eqnarray*} The function $e^{-k\phi}(n+\Delta_{\omega}\phi)$ must achieve its maximum at some point $p.$ At this point, $$0\geq -An -\Delta_{\omega}\phi(p)-kn(n+\Delta_{\omega}\phi)+(k-\Lambda)e^{\frac {-F(p)} {n-1}}(n+\Delta_{\omega}\phi)^{1+\frac 1{n-1}}(p).$$ Notice that we can bound $\sup F$ by $C(\Lambda, A, B).$ Thus, the above inequality implies $$n+\Delta_{\omega}\phi \leq C_4(\Lambda, A, B).$$ Since we have an upper bound on $F$, the lower bound of $F$ can be obtained by Lemma \ref{lem2.9} \begin{eqnarray*}\inf F\geq -4C(\Lambda, A, B) \exp (2+2\int_M\; F \omega_{\phi}^n) =C(\Lambda, A, B).\end{eqnarray*} On the other hand, \begin{eqnarray*} \inf F\leq\log \frac {\omega^n_{\phi}}{\omega^n}=\log\prod_i \;(1+\phi_{i\bar i})\leq \log (\prod_i(n+\Delta_{\omega} \phi)^{n-1}(1+\phi_{i\bar i})). \end{eqnarray*} Hence, $1+\phi_{i\bar i}\geq C_5(\Lambda, A, B)>0.$ Thus, $$C_5(\Lambda, A, B)\leq n+\Delta_{\omega} \phi \leq C_4(\Lambda, A, B).$$ By (\ref{f1}) and (\ref{f2}), we have
$$|\Delta_{\omega} F|\leq C(A, B, \Lambda).$$ By the elliptic estimate, $F\in W^{2, p}(M, \omega)$ for any $p>1.$ Recall that $F$ satisfies the equation (\ref{f3}), we have the H\"older estimate $\phi\in C^{2, \alpha}(M, \omega)$ for some $\alpha\in (0, 1)$ (cf. \cite{[Siu]},\cite{[Tru]}). Let $\psi$ be a local potential of $\omega$ such that $\omega=\sqrt{-1}\partial\bar\partial \psi$. Differential the equation (\ref{f3}), we have $$\Delta_{\phi}\pd {}{z^i}(\phi+\psi)-\pd {}{z^i}\log \omega^n=\pd {F}{z^i}\in W^{1, p}(M, \omega).$$ Note that the coefficients of $\Delta_{\phi}$ is in $C^{\alpha}(M, \omega)$, by the elliptic estimate $\phi\in W^{4, p}(M, \omega)$. Then by the Sobolev embedding theorem for any $\beta\in (0, 1),$
$$|\phi|_{C^{3, \beta}(M, \omega)}\leq C(A, B, \omega, \beta).$$ The lemma is proved.
\end{proof}
For convenience, we introduce the following definition. \begin{defi} For any K\"ahler metric $\omega,$ we define
$$W(\omega)=\inf_f \Big\{\int_M \;|f_{\alpha\beta}|^2\omega^n\;\;\Big|\;f\in W^ {2,2}(M, \omega), \int_M\;f^2\omega^n=1, \int_M\;f\omega^n=0\Big\}.$$ \end{defi} Assume that $M$ has no nonzero holomorphic vector fields, then the following lemma gives a positive lower bound of $W(\omega).$
\begin{lem}\label{lem2.12}Assume that $M$ has no nonzero holomorphic vector fields. For any constant $A, B>0$, there exists a positive constant $C_6$ depending on $A, B$ and the background metric $\omega$, such that for any K\"ahler metric $\omega_{\phi}=\omega+\sqrt{-1}\partial\bar\partial\phi,$ if
$$|Ric(\omega_{\phi})|\leq A, \; \;\;{ and}\;\;\; |\phi|\leq B,$$ then $$W(\omega_{\phi})\geq C_6>0.$$ \end{lem} \begin{proof}Suppose not, we can find a sequence of metrics $\omega_m=\omega +\sqrt{-1}\partial\bar\partial\phi_m$ and functions $f_m$ satisfying
$$|Ric(\omega_m)|\leq A, \qquad |\phi_m|\leq B,$$ and
$$\int_M\;f_m^2\omega_{m}^n=1, \int_M\;f_m\omega_{m}^n=0, \int_M \;|f_{m,
\alpha\beta}|_{g_m}^2\omega_{m}^n\ri0.$$ Note that the Sobolev constants with respect to the metrics $\omega_m$ are uniformly bounded. By Lemma \ref{lem2.10}, we can assume that $\omega_m$ converges to a K\"ahler metric $\omega_{\infty}$ in $C^{1, \beta}(M, \omega)$ norm for some $\beta\in (0, 1).$ Now define a sequence of vector fields \begin{equation} X_m^i=g_m^{i\bar k}\pd {f_{m}}{z^{\bar k}},\qquad X_m=X_m^i\pd {}{z^i}.\label{f4}\end{equation} By direct calculation, we have
$$|X_m|^2_{g_m}=|\nabla f_m|_{g_m}^2,$$
and \begin{eqnarray*} \Big|\pd {X_m}{\bar z}\Big|_{g_m}^2=\sum_{i, j}\Big|\pd
{X_m^i}{z^{\bar j}}\Big|_{g_m}^2=|f_{m, \alpha\beta}|_{g_m}^2. \end{eqnarray*} Then
\begin{equation}\int_M \;\Big|\pd {X_m}{\bar z}\Big|_{g_m}^2\omega_{g_m}^n\ri0.\label{la1}\end{equation} Next we claim that there exist two positive constants $C_7$ and $C_8$ which depend only on $A$ and the Poincar\'e constant $\sigma$ such that
\beq0<C_7(\sigma)\leq \int_M \; |X_m|_{g_m}^2 \omega_{g_m}^n\leq C_8(A).\label{la2}\end{equation} In fact, since the Poincar\'e constant is uniformly bounded in our case,
$$\int_M \; |X_m|_{g_m}^2 \omega_{g_m}^n=\int_M \; |\nabla f_m|_{g_m}^2 \omega_ {g_m}^n\geq C(\sigma)\int_M\;f_m^2\omega_{g_m}^n=C(\sigma).$$ On the other hand, since the Ricci curvature has a upper bound, we have
\begin{eqnarray*}\int_M \; |\Delta_m f_m|_{g_m}^2 \omega_{g_m}^n&=&\int_M \;|f_{m,
\alpha \beta}|_{g_m}^2\omega_{g_m}^n+\int_M \;R_{i\bar j}f_{m, \bar i}f_{m, j}\omega_ {g_m}^n\\
&\leq &\int_M \;|f_{m, \alpha\beta}|_{g_m}^2\omega_{g_m}^n+A\int_M \;|\nabla f_m| _{g_m}^2\omega_{g_m}^n\\
&\leq &\int_M \;|f_{m, \alpha\beta}|_{g_m}^2\omega_{g_m}^n+\frac 12\int_M
\; |
\Delta_m f_m|_{g_m}^2 \omega_{g_m}^n+\frac {A^2}2 \int_M \;f_m^2\omega_{g_m}^n\\ \end{eqnarray*} Then
$$\int_M \; |\Delta_m f_m|_{g_m}^2 \leq 1+A^2.$$
Therefore, \begin{eqnarray*}\int_M \; |X_m|_{g_m}^2 \omega_{g_m}^n&=&\int_M \;
|\nabla f_m|_{g_m} ^2 \omega_{g_m}^n\\
&\leq&\frac 12\int_M \; |\Delta_m f_m|_{g_m}^2 + \frac 12\int_M\;f_m^2 \omega_{g_m}^n\\ &\leq &C(A). \end{eqnarray*} This proves the claim.
Now we have
$$\int_M\;f_m^2\omega_{m}^n=1, \qquad \int_M\;|\nabla\bar \nabla f_m|_{g_m}^2
\omega_{m}^n\leq C(A),\; \int_M\; |f_{m, \alpha\beta}|_{g_m}^2 \omega_m^n\rightarrow 0,$$ then $f_m\in W^{2,2}(M, \omega_m).$ Note that the metrics $\omega_m$ are $C^{1, \beta}$ equivalent to $\omega_{\infty}$, then $f_m\in W^{2, 2}(M, \omega_{\infty}),$ thus we can assume $f_m$ strongly converges to $f_{\infty}$ in $W^{1, 2}(M, \omega_{\infty})$. By (\ref{f4}) $X_{m}$
strongly converges to $ X_{\infty}$ in $L^2(M, \omega_{\infty}).$ Thus, by (\ref{la2}), \begin{equation} 0<C_7\leq \int_M \; |X_{\infty}|^2 \omega_{\infty}^n\leq C_8.\label{f5}\end{equation} Next we show that $X_{\infty}$ is holomorphic. In fact, for any vector valued smooth function
$\xi=(\xi^1, \xi^2, \cdots, \xi^n),$ \begin{eqnarray*} \Big|\int_M\; \xi\cdot\bar
\partial X_m\omega_{\infty}^n \Big|^2&=&\Big|\int_M\; \xi^k \frac {\partial X_m}{\partial \bar z^{ k}}\;\omega_{\infty}^n
\Big|^2\\&\leq &\int_M\; |\xi|^2\omega_{\infty}^n\int_M\; \Big|\pd
{X_m}{\bar z}\Big|^2\omega_{\infty}^n\\&\leq &C \int_M\;
|\xi|^2\omega_{\infty}^n\int_M\; \Big|\pd {X_m}{\bar z}\Big|_{g_m}^2\omega_{g_m}^n \rightarrow 0. \end{eqnarray*} On the other hand, $$\int_M\; \xi\cdot\bar \partial X_m\omega_{\infty}^n =-\int_M\; \bar \partial \xi\cdot X_m \; \omega_{\infty}^n\rightarrow -\int_M\; \bar \partial \xi\cdot X_{\infty} \; \omega_{\infty}^n.$$ Then $X_{\infty}$ is a weak holomorphic vector field, thus it must be holomorphic. By (\ref{f5}) $X_{\infty}$ is a {{nonzero}} holomorphic vector field, which contradicts the assumption that $M$ has no nonzero holomorphic vector fields. The lemma is proved. \end{proof}
\begin{lem}\label{lem2.13} If the K\"ahler metric $\omega_g$ satisfies $Ric(\omega_g)\geq (1-\eta)\omega_g$ where $0<\eta< \frac {\sqrt{C_6}}2$. Here $C_6$ is the constant obtained in Lemma \ref{lem2.12}. Then the first eigenvalue of $\Delta_g$ satisfies $\lambda_1\geq 1+\gamma$, where $\gamma=\frac {\sqrt{C_6}}{2}.$ \end{lem} \begin{proof} Let $ u$ is any eigenfunction of $\omega_g$ with eigenvalue $\lambda_1,$ so $\Delta_g u=-\lambda_1u.$ Then by direct calculation, we have \begin{eqnarray*} \int_M\; u_{ij}u_{\bar i\bar j}\;\omega_g^n&=&-\int_M\;u_{ij\bar j} u_{\bar i}\;\omega_g^n\\ &=& -\int_M\;(u_{j\bar ji}+R_{i\bar k}u_{k})u_{\bar i}\;\omega_g^n\\ &=&\int_M ((\Delta_g u)^2-R_{i\bar j}u_{j}u_{\bar i})\;\omega_g^n. \end{eqnarray*} This implies \begin{eqnarray*} C_6\int_M\; u^2 \omega_g^n&\leq & \int_M\; ((\Delta_g u)^2-R_{i\bar j}u_{j}u_{\bar i})\;\omega_g^n\\
&\leq &\lambda_1^2 \int_M\; u^2 \omega^n-(1-\eta)\int_M\;|\nabla u|^2 \omega_g^n\\ &=&(\lambda_1^2-(1-\eta)\lambda_1)\int_M\; u^2 \omega_g^n. \end{eqnarray*} Thus, we have $\lambda_1^2-(1-\eta)\lambda_1-C_6\geq 0.$ Then, $$\lambda_1\geq 1+\frac {\sqrt{C_6}}{2}.$$ \end{proof}
\begin{flushleft} \begin{proof}[Proof of Theorem \ref{lem2.8}]
The theorem follows directly from the above Lemma \ref{lem2.12} and \ref{lem2.13}.
\end{proof} \end{flushleft}
\subsubsection{Case 2: $M$ has nonzero holomorphic vector fields}\label{section4.4.2} In this subsection, we will consider the case when $M$ has nonzero holomorphic vector fields. Denote by $Aut(M)^{\circ}$ the connected component containing the identity of the holomorphic transformation group of $M$. Let $K$ be a maximal compact subgroup of $Aut(M)^{\circ}$. Then there is a semidirect decomposition of $Aut(M)^{\circ}$(cf. \cite{[FM]}), $$Aut(M)^{\circ}=Aut_r(M)\propto R_u,$$ where $Aut_r(M)\subset Aut(M)^{\circ}$ is a reductive algebraic subgroup and the complexification of $K$, and $R_u$ is the unipotent radical of $Aut(M)^{\circ}$. Let $\eta_r(M, J)$ be the Lie algebra of $Aut_r(M, J).$
Now we introduce the following definition which is a mild modification from \cite{[chen1]} and \cite{[PhSt]}. \begin{defi}\label{prestable} The complex structure $J$ of $M$ is called pre-stable, if no complex structure of the orbit of diffeomorphism group contains larger (reduced) holomorphic automorphism group (i.e., $Aut_r(M)$).
\end{defi}
Now we recall the following $C^{k, \alpha}$ convergence theorem of a sequence of K\"ahler metrics, which is well-known in literature (cf. \cite{[PhSt]}, \cite{[Tian4]}).
\begin{theo}\label{conv1} Let $M$ be a compact K\"ahler manifold. Let $(g(t), J(t))$ be any sequence of metrics $g(t)$ and complex structures $J(t)$ such that $g(t)$ is K\"ahler with respect to $J(t)$. Suppose the following is true:
\begin{enumerate}\item For some integer $k\geq 1$, $|\nabla^lRm|_{g(t)}$ is uniformly bounded for any integer $l (0\leq l< k)$;
\item The injectivity radii $i(M, g(t))$ are all bounded from below;
\item There exist two uniform constant $c_1$ and $c_2$ such that $0<c_1\leq {\rm Vol}(M, g(t))\leq c_2$. \end{enumerate} Then there exists a subsequence of $t_j$, and a sequence of diffeomorphism $F_j: M\rightarrow M$ such that the pull-back metrics $\tilde g(t_j)=F_j^*g(t_j)$ converge in $C^{k, \alpha}(\forall \,\alpha\in (0, 1))$ to a $C^{k, \alpha}$ metric $g_{\infty}$. The pull-back complex structure tensors $\tilde J(t_j)=F_j^*J(t_j)$ converge in $C^{k, \alpha}$ to an integral complex structure tensor $\tilde J_{\infty}$. Furthermore, the metric $ g_{\infty}$ is K\"ahler with respect to the complex structure $\tilde J_{\infty}$.
\end{theo} \begin{theo}\label{theo4.18}Suppose $M$ is pre-stable. For any $\Lambda_0, \Lambda_1>0$, there exists $\eta>0$ depending only on $\Lambda_0$ and $\Lambda_1$ such that for any metric $\omega\in 2\pi c_1(M),$
if \begin{equation}|Ric(\omega)-\omega|\leq \eta,\;\; |Rm(\omega)|\leq \Lambda_0,\;\;|\nabla Rm(\omega)|\leq \Lambda_1, \label{r1}\end{equation} then for any smooth function $f$ satisfying $$ \int_M\; f\omega^n=0 {\;\;{ and}\;\;} Re\left(\int_M\; X(f)\omega^n\right)=0, \qquad \forall X\in \eta(M, J),$$ we have
$$\int_M\; |\nabla f|^2\omega^n>(1+\gamma(\eta, \Lambda_0, \Lambda_1))\int_M\; |f|^2\omega^n,$$ where $\gamma>0$ depends only on $\eta, \Lambda_0$ and $\Lambda_1.$
\end{theo} \begin{proof}Suppose not, for any positive numbers $\eta_m\rightarrow 0$, there exists a sequence of K\"ahler metrics $\omega_m\in 2\pi c_1(M)$
such that \begin{equation} |Ric(\omega_m)-\omega_{m}|\leq \eta_m,\;\; |Rm(\omega_m)|\leq
\Lambda_0,\;\;\;|\nabla_m Rm(\omega_m)|\leq \Lambda_1,\label{r2}\end{equation} where $Rm_m$ is with respect to the metric $\omega_m$, and smooth functions $f_m$ satisfying $$\int_M\; f_m\omega_m^n=0, \qquad Re\left(\int_M\; X(f_m)\omega_m^n\right)=0, \qquad \forall X\in \eta(M, J),$$
\begin{equation}\int_M\; |\nabla_m f_m|^2\omega_m^n<(1+\gamma_m)\int_M\;
|f_m|^2\omega_m^n,\label{eq5.25}\end{equation} where $0<\gamma_m\rightarrow 0.$ Without loss of generality, we may assume that \[ \int_M\; f_m^2 \omega_m^n = 1, \qquad \forall m \in {\mathbb N}, \] which means \[ \int_M\;
|\nabla_m f_m|^2\omega_m^n\leq 1 + \gamma_m < 2. \] Then, $f_m $ will converge in $W^{1,2}$ if $(M, \omega_m)$ converges. However, according to our stated condition, $(M, \omega_m, J)$ will converge in $C^{2, \alpha}(\alpha\in (0, 1))$ to $(M, \omega_\infty, J_\infty).\;$ In fact, by (\ref{r2}) the diameters of $\omega_m$ are uniformly bounded. Note that all the metrics $\omega_m$ are in the same K\"ahler class, the volume is fixed. Then by (\ref{r2}) again, the injectivity radii are uniformly bounded from below. Therefore, all the conditions of Theorem \ref{conv1} are satisfied.
Note that the complex structure $J_\infty$ lies in the closure of the orbit of diffeomorphisms, while $\omega_\infty$ is a K\"ahler-Einstein metric in $(M, J_\infty)$. By the standard deformation theorem in complex structures, we have \[
\dim Aut_r (M, J) \leq \dim Aut_r(M, J_\infty). \] By abusing notation, we can write \[ Aut_r(M, J) \subset Aut_r(M, J_\infty). \] By our assumption of pre-stable of $(M, J)$, we have the inequality the other way around. Thus, we have \[ \dim Aut_r(M, J) = \dim Aut_r(M, J_\infty),\;\;\;{\rm or}\;\;\; Aut_r(M, J) = Aut_r(M, J_\infty). \] Now, let $f_\infty $ be the $W^{1,2}$ limit of $f_m$, then we have \[
1 \leq |f_\infty|_{W^{1,2}(M,\, \omega_\infty)} \leq 3
\]
and
\[
\int_M f_\infty \omega_\infty^n = 0, \qquad Re\left(\int_M\; X(f_\infty) \omega_\infty^n\right) =
0,\qquad \forall X\in \eta(M, J).
\]
Thus, $f_\infty$ is a non-trivial function. Since $\omega_\infty$ is a K\"ahler-Einstein metric, we have
\[
\int_M\; \theta_X f_\infty \omega_\infty^n = 0,
\]
where \[
{\cal L}_X \omega_\infty =\sqrt{-1}\partial\bar\partial\theta_X.
\]
This implies that $f_\infty $ is perpendicular to the first eigenspace\footnote{Note that
$\triangle \theta_X = -\theta_X $ is totally real for
$X \in Aut_r(M, J_\infty).\;$ Moreover, the first eigenspace consists of
all such $\theta_X.\;$} of $\triangle_{\omega_\infty}.\;$
In other words, there is a $\delta > 0$ such that
\[
\int_M |\nabla f_\infty|^2 \omega_\infty^n > (1+ \delta) \int_M f_\infty^2 \omega_\infty^n > 1+ \delta.
\]
However, this contradicts the following fact:
\begin{eqnarray*} \int_M \;|\nabla f_\infty|^2 \omega_\infty^n & \leq & \displaystyle \lim_{m\rightarrow \infty}
\int_M |\nabla_m f_m|^2 \omega_m^n \\
& \leq & \displaystyle \lim_{m\rightarrow \infty} (1+ \gamma_m) \int_M f_\infty^2 \omega_\infty^n = 1.
\end{eqnarray*} The lemma is then proved. \end{proof}
\subsection{Exponential decay in a short time}\label{section4.5} In this subsection, we will show that the $W^{1,2}$ norm of $\dot \phi$ decays exponentially in a short time. Here we follow the argument in \cite{[chen-tian2]} and use the estimate of the first eigenvalue obtained in the previous subsection. \begin{lem}\label{lem2.14} Suppose for any time $t\in [T_1, T_2]$, we have
$$|Ric-\omega|(t)\leq C_1\epsilon\; \;\;{ and}\;\;\; \lambda_1(t)\geq 1+\gamma>1.$$ Let $$\mu_0(t)=\frac 1V\int_M\;(\dot\phi-c(t))^2\omega_{\phi}^n.$$ If $\epsilon$ is small enough, then there exists a constant $\alpha_0>0$ depending only on $\gamma, \sigma$ and $C_1\epsilon$ such that
$$\mu_0(t)\leq e^{-\alpha_0 (t-T_1)}\mu_0(T_1), \qquad\forall t\in [T_1, T_2].$$
\end{lem} \begin{proof} By direct calculation, we have \begin{eqnarray*} \frac {d}{dt}\mu_0(t)&=&\frac 2V\int_M\;(\dot\phi-c(t))(\ddot \phi-c(t)')\omega_{\phi}^n+\frac 1V\int_M\;(\dot \phi-c(t))^2\Delta_{\phi}\dot
\phi\omega_{\phi}^n\\&=&-\frac 2V\int_M\;(1+\dot \phi-c(t))|\nabla (\dot
\phi-c(t))|^2\omega_{\phi}^n+\frac 2V\int_M\;(\dot \phi-c(t))^2\omega_{\phi}^n. \end{eqnarray*} By the assumption, we have for
$t\in [T_1, T_2]$ \begin{eqnarray*} \frac {d}{dt}\mu_0(t)&=&-\frac 2V\int_M\;(1+\dot\phi-c(t))|\nabla
\dot\phi|^2\omega_{\phi}^n+\frac 2V\int_M\;(\dot\phi-c(t))^2\omega_{\phi}^n\\
&\leq&-\frac 2V\int_M\;(1-C(\sigma)C_1\epsilon)|\nabla
\dot\phi|^2\omega_{\phi}^n +\frac 2V\int_M\;(\dot \phi-c(t))^2\omega_{\phi}^n\\ &\leq&-\frac 2V\int_M\;(1-C(\sigma)C_1\epsilon)(1+\gamma)(\dot \phi-c(t))^2\omega_{\phi}^n+\frac 2V\int_M\;(\dot \phi-c(t))^2\omega_{\phi}^n\\&=&-\alpha_0\mu_0(t). \end{eqnarray*} Here $$\alpha_0=2(1-C(\sigma)C_1\epsilon)(1+\gamma)-2>0,$$ if we choose $\epsilon$ small enough. Thus, we have $$\mu_0(t)\leq e^{-\alpha_0 (t-T_1)}\mu_0(T_1).$$ \end{proof}
\begin{lem}\label{lem2.15}Suppose for any time $t\in [T_1, T_2]$, we have
$$|Ric-\omega|(t)\leq C_1\epsilon\; \;\;{ and}\;\;\; \lambda_1(t)\geq 1+\gamma>1.$$
Let $$\mu_1(t)=\frac 1V\int_M\;|\nabla \dot \phi|^2\omega_{\phi}^n.$$ If $ \epsilon$ is small enough, then there exists a constant $\alpha_1>0$ depending only on $\gamma$ and $C_1\epsilon$ such that $$\mu_1(t)\leq e^{-\alpha_1(t-T_1)}\mu_1(T_1), \qquad\forall t\in [T_1, T_2].$$ \end{lem}
\begin{proof} Recall that the evolution equation for $|\nabla \dot\phi|^2$ is
$$\pd {}t|\nabla \dot \phi|^2=\Delta_{\phi} |\nabla \dot \phi|^2-|\nabla \nabla \dot
\phi|^2-|\nabla \bar \nabla\dot \phi|^2+|\nabla \dot \phi|^2.$$ Then for any time $t\in [T_1, T_2],$ \begin{eqnarray*} \frac d{dt}\mu_1(t)&=&\frac 1V\int_M\;(-|\nabla \nabla \dot \phi|^2- |\nabla \bar \nabla\dot \phi|^2+|\nabla
\dot \phi|^2+|\nabla \dot \phi|^2\Delta_{\phi} \dot\phi)\;\omega_{\phi}^n\\
&\leq &\frac 1V\int_M\;(-\gamma|\nabla \dot
\phi|^2+(n-R(\omega_{\phi}))|\nabla \dot \phi|^2) \omega_{\phi}^n\\ &\leq& -(\gamma-C_1\epsilon)\mu_1(t). \end{eqnarray*} Thus, we have $$\mu_1(t)\leq e^{-\alpha_1(t-T_1)}\mu_1(T_1)$$ where $\alpha_1=\gamma-C_1\epsilon>0$ if we choose $\epsilon$ small. \end{proof}
\subsection {Estimate of the $C^0$ norm of $\phi(t)$ }\label{section4.6} In this subsection, we derive some estimates on the $C^0$ norm of
$|\phi|$. Recall that in the previous subsection we proved that the
$W^{1, 2}$ norm of $|\dot\phi-c(t)|$ decays exponentially. Based on this result we will use the parabolic Moser iteration to show that the $C^0$ norm of $|\dot\phi-c(t)|$ also decays exponentially.
\begin{lem}\label{lem2.18} Suppose that $\mu_0(t), \mu_1(t)$ decay exponentially for $t\in [T_1, T_2]$ as in Lemma \ref{lem2.14} and \ref{lem2.15}, then we have
$$\Big|\pd {\phi}t-c(t)\Big|_{C^{0}}\leq \frac {C_9(n, \sigma)} {\tau^{\frac {m}{4}}}\Big(\mu_0(t-\tau)+\frac 1{\alpha_1^2}\mu_1^2(t-\tau)\Big)^{\frac 12},\;\;\;\forall \;t\in [T_1+\tau, T_2]$$ where $m=\dim_{{\mathbb R}}M$ and $\tau<T_2-T_1$.
\end{lem} \begin{proof} Let $u=\pd {\phi}t-c(t),$ the evolution equation for $u$ is $$\pd ut=\Delta_{\phi} u+u+\mu_1(t),$$
where $\mu_1(t)=\frac 1V\int_M\;|\nabla \dot \phi|^2\omega_{\phi}^n.$ Note that in the proof of Lemma \ref{lem2.15}, we derived $$\pd {}t\mu_1(t)\leq -\alpha_1 \mu_1(t).$$ Thus, we have $$\pd {}t(u_++\frac 1{\alpha_1}\mu_1)\leq \Delta_{\phi} (u_++\frac 1{\alpha_1} \mu_1)+(u_++\frac 1{\alpha_1}\mu_1).$$ where $u_+=\max\{u, 0\}$. Since $u_++\frac 1{\alpha_1}\mu_1$ is a nonnegative function, we can use the parabolic Moser iteration, $$(u_++\frac 1{\alpha_1}\mu_1)(t)\leq \frac {C(n, \sigma)}{\tau^{\frac {m +2}{4}}} \Big(\int_{t-\tau}^t\;\int_M (u_++\frac 1{\alpha_1}\mu_1)^2(s) \;\omega_{\phi}^n\wedge ds\Big)^{\frac 12}.$$ Since $\mu_0$ and $\mu_1$ are decreasing, \begin{eqnarray}& &(u_++\frac 1{\alpha_1}\mu_1)(t)\nonumber\\ &\leq&\frac {C(n, \sigma)}{\tau^{\frac {m+2}{4}}}\Big(\int_{t-\tau}^t \;(\mu_0(s)+\frac 1{\alpha_1^2} \mu_1^2(s))ds\Big)^{\frac 12}\nonumber\\ &\leq&\frac {C(n, \sigma)}{\tau^{\frac {m}{4}}}\Big(\mu_0(t-\tau)+\frac 1{\alpha_1^2}\mu_1^2(t-\tau)\Big)^{\frac 12}. \label{x1}\end{eqnarray} On the other hand, the evolution equation for $-u$ is $$\pd {}t(-u)=\Delta_{\phi} (-u)+(-u)-\mu_1(t)\leq \Delta_{\phi} (-u)+(-u).$$ Thus, $$\pd {}t(-u)_+\leq \Delta_{\phi} (-u)_++(-u)_+.$$ By the parabolic Moser iteration, we have \begin{eqnarray} (-u)_+&\leq &\frac {C(n, \sigma)}{\tau^{\frac {m+2}{4}}}\Big(\int_{t-\tau}^t\; \int_M (-u)_+^2\omega_{\phi}^n\wedge ds\Big)^{\frac 12}\nonumber\\ &\leq&\frac {C(n, \sigma)}{\tau^{\frac {m}{4}}}\mu_0(t-\tau)^{\frac 12}. \label{x2}\end{eqnarray} Combining the two inequalities (\ref{x1})(\ref{x2}), we obtain the estimate
$$\Big|\pd {\phi}t-c(t)\Big|_{C^{0}}\leq \frac {C(n, \sigma)}{\tau^ {\frac {m}{4}}}\Big (\mu_0(t-\tau)+\frac 1{\alpha_1^2}\mu_1^2(t-\tau)\Big)^{\frac 12}.$$ This proved the lemma.\end{proof}
\begin{lem}\label{lem2.19}Under the same assumptions as in Lemma \ref{lem2.18}, we have
$$|\phi(t)|\leq |\phi(T_1+\tau)|+\frac {C_{10}(n, \sigma)}{\alpha \tau^{\frac {m}{4}}}(\sqrt{\mu_0(T_1)}+\frac 1{\alpha_1}\mu_1(T_1)) + \tilde C,\qquad \forall\, t\in [T_1+\tau, T_2].$$ Here $\tilde C=E_0(0)-E_0(\infty)$ is a constant in Lemma \ref{lem2.6}.
\end{lem} \begin{proof}
\begin{eqnarray*} |\phi(t)|&\leq &|\phi(T_1+\tau)|+\int_{T_1+\tau}^{t}\;
\Big|\pd {\phi(s)}
s-c(s)\Big| ds+\int_{T_1+\tau}^{t}\;c(s) ds\\
&\leq &|\phi(T_1+\tau)|+\frac {C(n, \sigma)}{\tau^{\frac {m}{4}}}\int_{T_1+\tau} ^{t}\;\Big(\mu_0(s-\tau)+\frac 1{\alpha_1^2}\mu_1^2(s-\tau)\Big)^{\frac 12}ds + \tilde C\\
&\leq &|\phi(T_1+\tau)|+\frac {C(n, \sigma)}{\tau^{\frac {m}{4}}}(\sqrt {\mu_0(T_1)}+\frac 1{\alpha_1}\mu_1(T_1))\int_{T_1+\tau}^{t}\;e^{- \alpha (s- \tau-T_1)}ds +\tilde C\\
&\leq&|\phi(T_1)|+\frac {C(n, \sigma)}{\alpha \tau^{\frac {m}{4}}}(\sqrt{\mu_0(T_1)}+\frac 1{\alpha_1}\mu_1(T_1)) +\tilde C \end{eqnarray*} where $\alpha=\min\{\frac {\alpha_0}2, \alpha_1\}$ and $\tilde C=E_0(0)-E_0(\infty)$ is a constant in Lemma \ref{lem2.6}. \end{proof}
\subsection{Estimate of the $C^k$ norm of $\phi(t)$ } In this subsection, we shall obtain uniform $C^k$ bounds for the solution $\phi(t)$ of the K\"ahler-Ricci flow $$\pd {\phi}t=\log\frac {\omega^n_{\phi}}{\omega^n}+\phi-h_{\omega}$$ with respect to any background metric $\omega$. For simplicity, we normalize $h_{\omega}$ to satisfy $$\int_M\; h_{\omega}\;\omega^n=0.$$ The following is the main result in this subsection. \begin{theo}\label{theoRm}For any positive constants $\Lambda, B>0$ and small $\eta>0$, there exists a constant $C_{11}$ depending only on $B, \eta, \Lambda$ and the Sobolev constant $\sigma$ such that if the background metric $\omega$ satisfies
$$|Rm(\omega)|\leq \Lambda, \qquad |Ric(\omega)-\omega|\leq \eta, $$
and $|\phi(t)|, |\dot\phi(t)|\leq B,$ then
$$|Rm|(t)\leq C_{11}(B, \Lambda, \eta, \sigma).$$ \end{theo} \begin{proof}Note that $R(\omega)-n=\Delta_{\omega} h_{\omega}$, by the assumption we have
$$|\Delta_{\omega} h_{\omega}|\leq \eta.$$ Since the Sobolev constant with respect to the metric $\omega$ is uniformly bounded by a constant $\sigma$, we have
$$|h_{\omega}|_{C^0}\leq C(\sigma)\eta.$$ Now we use Yau's estimate to obtain higher order estimate of $\phi.$ Define $$F=\dot\phi-\phi+h_{\omega},$$ then the K\"ahler-Ricci flow can be written as $$(\omega+\sqrt{-1}\partial\bar\partial \phi)^n=e^F \omega^n.$$ By Yau's estimate we have \begin{eqnarray*} \Delta_{\phi}(e^{-k\phi}(n+\Delta_{\omega}\phi))&\geq & e^{-k\phi}(\Delta_{\omega} F-n^2 \inf_{i\neq j}R_{i\bar ij\bar j}(\omega))\\ &-&ke^{-k\phi}n(n+\Delta_{\omega}\phi)+(k+\inf_{i\neq j}R_{i\bar ij\bar j}(\omega))e^{-k\phi+ \frac {-F}{n-1}}(n+\Delta_{\omega}\phi)^{1+\frac 1{n-1}}. \end{eqnarray*} Note that \begin{eqnarray*} \pd {}t(e^{-k\phi}(n+\Delta_{\omega} \phi))&=&-k\dot\phi e^{-k \phi}(n+\Delta_{\omega} \phi)+e^{-k\phi}\Delta_{\omega} \dot\phi\\ &=&-k\dot\phi e^{-k\phi}(n+\Delta_{\omega} \phi)+e^{-k\phi}\Delta_{\omega} (F+\phi-h_{\omega}).\end{eqnarray*} Combing the above two inequalities, we have \begin{eqnarray*} (\Delta_{\phi}-\pd {}t)(e^{-k\phi}(n+\Delta_{\omega}\phi))&\geq & e^{-k\phi}(\Delta_{\omega} h_{\omega}+n-n^2\inf_{i\neq j}R_{i\bar ij\bar j}(\omega))\\&+&(k\dot\phi-kn-1) e^{-k\phi}(n+\Delta_{\omega} \phi)\\&+&(k+\inf_{i\neq j}R_{i\bar ij\bar j}(\omega))e^{-k\phi+ \frac {-F}{n-1}}(n+\Delta_{\omega}\phi)^{1+\frac 1{n-1}}.\end{eqnarray*} Since $\phi,
\Delta_{\omega} h_{\omega}, |h_{\omega}|, |Rm(\omega)|, \dot\phi$ are bounded, by the maximum principle we can obtain the following estimate $$n+\Delta_{\omega} \phi\leq C_{12}(B, \eta, \Lambda, \sigma).$$ By the definition of $F$, $$\log\frac {\omega^n_{\phi}}{\omega^n}=F\geq -C_{13}(B, \eta, \sigma).$$ On the other hand, we have \begin{eqnarray*}\log\frac {\omega^n_{\phi}}{\omega^n}&=&\log\prod_{i=1}^n(1+\phi_{i\bar i})\leq\log ((n+\Delta_{\omega}\phi)^{n-1}(1+\phi_{i\bar i})). \end{eqnarray*} Thus, $1+\phi_{i\bar i}\geq e^{-C_{13}}C_{12}^{-\frac 1{n-1}},$i.e. $C_{14}\omega\leq \omega_{\phi}\leq C_{12}\omega.$ Following Calabi's computation (cf. \cite{[chen-tian1]},\cite{[Yau]}), we can obtain the following $C^3$ estimate:
$$|\phi|_{C^3(M, \,\omega)}\leq C_{14}(B, \eta, \Lambda, \sigma).$$ Since the metrics $\omega_{\phi}$ are uniformly equivalent, the flow is uniform parabolic with $C^1$ coefficients. By the standard parabolic estimates, the $C^4$ norm of $\phi$ is bounded, and then all the curvature tensors are also bounded. The theorem is proved. \end{proof}
\section{Proof of Theorem \ref{main} } In this section, we shall prove Theorem \ref{main}. This theorem needs the technical condition that $M$ has no nonzero holomorphic vector fields, which will be removed in Section \ref{section6}. The idea is to use the estimate of the first eigenvalue proved in Section \ref{section4.4.1}.
\begin{theo}Suppose that $M$ has no nonzero holomorphic vector fields and $E_1$ is bounded from below in $[\omega].$ For any $\delta, B, \Lambda>0,$ there exists a small positive constant $\epsilon(\delta, B, \Lambda, \omega)>0$ such that for any metric $\omega_0$ in the subspace ${\mathcal A}(\delta, B, \Lambda, \epsilon)$ of K\"ahler metrics
$$\{\omega_{\phi}=\omega+\sqrt{-1}\partial\bar\partial \phi\;|\; Ric(\omega_{\phi})>-1+\delta,\; |\phi|
\leq B, \;|Rm|(\omega_{\phi})\leq \Lambda, \; E_1(\omega_{\phi})\leq \inf E_1+\epsilon \}$$ the K\"ahler-Ricci flow will deform it exponentially fast to a K\"ahler-Einstein metric in the limit. \end{theo}
\begin{flushleft} \begin{proof} Let $\omega_0=\omega+\sqrt{-1}\partial\bar\partial \phi(0)\in {\mathcal A}(\delta, B, \Lambda, \epsilon)$, where $\epsilon$ will be determined later. Note that $E_1(0)\leq \inf E_1 +\epsilon, $ by Lemma \ref{lem5.9} we have $$E_0(0)-E_0(\infty)\leq \frac {\epsilon}2<1.$$ Here we choose $\epsilon<2.$ Therefore, we can normalize the K\"ahler-Ricci flow such that for the normalized solution $\psi(t)$,\begin{equation} 0< c(t), \;\int_0^{\infty} c(t)dt<1,\label{a1}\end{equation} where $c(t)=\frac 1V\int_M\;\dot\psi\omega^n_{\psi}.$ Now we give the details on how to normalize the solution to satisfy (\ref{a1}). Since $\omega_0=\omega+\sqrt{-1}\partial\bar\partial \phi(0)\in {\mathcal A}(\delta, B, \Lambda, \epsilon)$, by Lemma {\ref{lem2.10}} we have $$C_2(\Lambda, B, \omega)\omega\leq \omega_0\leq C_3(\Lambda, B, \omega)\omega.$$ By the equation of K\"ahler-Ricci flow, we have
$$|\dot\phi|(0)= \Big|\log \frac {\omega_{\phi}^n}{\omega^n}+\phi-h_{\omega}
\Big|_{t=0}\leq C_{16}(\omega, \Lambda, B).$$ Set $\psi(t)=\phi(t)+C_0e^t,$ where
$$C_0=\frac 1V\int_0^{\infty}\;e^{-t}\int_M\; |\nabla \dot\phi|^2\;\omega^n_ {\phi}\wedge dt -\frac 1V\int_M\; \dot\phi
\omega_{\phi}^n\Big|_{t=0}.$$ Then (\ref{a1}) holds and
$$|C_0|\leq 1+C_{16},$$
and $$|\psi|(0), \;|\dot\psi|(0)\leq B+1+C_{16}:=B_0.$$ \vskip 10pt
\textbf{STEP 1.}(Estimates for $t\in [2T_1, 6T_1]$). By Lemma \ref{lem2.1} there exists a constant $T_1(\delta, \Lambda)$ such that
\begin{equation} Ric(t)>-1+\frac {\delta}2,\; \;\;{ and}\;\;\; |Rm|(t)\leq 2\Lambda, \qquad \forall t\in [0, 6T_1].\label{5.20}\end{equation} By Lemma {\ref{lem2.5}} and the equation (\ref{5.20}), we can choose $\epsilon$ small enough so that
\begin{equation}|Ric-\omega|(t)\leq C_1(T_1, \Lambda)\epsilon<\frac 12,\qquad \forall t\in
[2T_1, 6T_1],\label{5.21}\end{equation} and \begin{equation}|\dot \psi-c(t)|\leq C(\sigma)C_1(T_1, \Lambda)\epsilon<1, \qquad\forall t\in [2T_1, 6T_1].\label{5.22}\end{equation} Then by the inequality
(\ref{a1})\begin{equation}|\dot\psi|(t)\leq 1+|c(t)|\leq 2,\qquad \forall t\in [2T_1, 6T_1]. \label{5.25}\end{equation} Note that the equation for $\dot\psi$ is $$\pd {}t\dot\psi=\Delta_{\psi} \dot\psi+\dot\psi,$$
we have \begin{equation} |\dot\psi|(t)\leq |\dot\psi|(0)e^{2T_1}\leq B_0e^{2T_1},\qquad \forall t\in [0, 2T_1]. \label{5.24} \end{equation} Thus, for any $t\in [2T_1, 6T_1]$ we have \begin{eqnarray*}|\psi|(t)&\leq
&|\psi|(0)+\int_0^{2T_1}\;
|\dot\psi|ds+\int_{2T_1}^t|\dot\psi|ds\\ &\leq &B_0+2T_1B_0e^{2T_1}+8T_1,\end{eqnarray*} where the last inequality used (\ref{5.25}) and (\ref{5.24}). For simplicity, we define \begin{eqnarray*} B_1:&=&B_0+2T_1B_0e^{2T_1}+8T_1+2,\\ B_k:&=&B_{k-1}+2,\qquad 2\leq k\leq 4. \end{eqnarray*} Then
$$|\dot\psi|(t),\;|\psi|(t)\leq B_1,\qquad \forall t\in [2T_1, 6T_1].$$ By Theorem \ref{theoRm} we have
$$|Rm|(t)\leq C_{11}(B_1, \Lambda_{\omega}, 1),\qquad \forall t\in [2T_1, 6T_1],$$ where $\Lambda_{\omega}$ is an upper bound of curvature tensor with respect to the metric $\omega,$ and $C_{11}$ is a constant obtained in Theorem \ref{theoRm}. Set $\Lambda_0=C_{11}(B_4, \Lambda_{\omega}, 1)$, we have
$$|Rm|(t)\leq \Lambda_0, \qquad \forall t\in [2T_1, 6T_1].$$
\vskip 10pt \textbf{STEP 2.}(Estimate for $t\in [2T_1+2T_2, 2T_1+6T_2]$). By STEP 1, we have
$$|Ric-\omega|(2T_1)\leq C_1\epsilon<\frac 12 \; \;\;{ and}\;\;\; |Rm|(2T_1)\leq\Lambda_0.$$ By Lemma {\ref{lem2.1}}, there exists a constant $T_2(\frac 12, \Lambda_0)\in (0, T_1]$ such that
$$|Rm|(t)\leq 2\Lambda_0,\; \;\;{ and}\;\;\; Ric(t)\geq 0,\qquad \forall t\in [2T_1, 2T_1 +6T_2].$$ Recall that $E_1\leq \inf E_1+\epsilon ,$ by Lemma \ref{lem2.2} and Lemma \ref{lem2.5} there exists a constant $C_1'(T_2, \Lambda_0)$ such that
$$|Ric-\omega|(t)\leq C_1'(T_2, \Lambda_0)\epsilon, \qquad\forall t\in [2T_1 +2T_2, 2T_1+6T_2].$$ Choose $\epsilon$ small enough so that $C_1'(T_2, \Lambda_0)\epsilon<\frac 12.$ Then by Lemma \ref{lem2.5},
$$|\dot\psi-c(t)|_{C^0}\leq C(\sigma)C_1'(T_2, \Lambda_0)\epsilon,\qquad \forall t\in [2T_1+2T_2, 2T_1+6T_2].$$ Choose $\epsilon$ small enough so that $ C(\sigma)C_1'(T_2, \Lambda_0)\epsilon<1.$ Thus, we can estimate the $C^0$ norm of $\psi$ for any $t\in [2T_1+2T_2, 2T_1+6T_2]$
\begin{eqnarray*}|\psi(t)|&\leq &|\psi|(2T_1+2T_2)+\Big|\int_{2T_1+2T_2}^t\;
\Big(\pd {\psi}{s}-c(s)\Big)ds\Big|+\Big|\int_0^t\;c(s)ds\Big|\\ &\leq &B_1+ 4T_2C(\sigma)C_1'(T_2, \Lambda_0)\epsilon+1. \end{eqnarray*} Choose $\epsilon$ small enough such that $4T_2C(\sigma)C_1'(T_2, \Lambda_0)\epsilon<1,$ then
$$|\psi(t)|\leq B_2, \qquad\forall t\in [2T_1+2T_2, 2T_1+6T_2].$$ Since $M$ has no nonzero holomorphic vector fields, applying Theorem
\ref{lem2.8} for the parameters $\eta=C_1'\epsilon,\;A=1\;, |\phi|\leq B_4,$ if we choose $\epsilon$ small enough, there exists a constant $\gamma(C_1'\epsilon, B_4, 1, \omega)$ such that the first eigenvalue of the Laplacian $\Delta_{\psi}$ satisfies $$\lambda_1(t)\geq 1+\gamma>1, \qquad \forall t\in [2T_1+2T_2, 2T_1+6T_2].$$
\textbf{STEP 3.} In this step, we want to prove the following claim: \begin{claim}For any positive number $S\geq 2T_1+6T_2$, if
$$|Ric-\omega|(t)\leq C_1'(T_2, \Lambda_0)\epsilon<\frac 12 \; \;\;{ and}\;\;\; |\psi(t)| \leq B_3, \qquad\forall t\in [2T_1+2T_2, S],$$ then we can extend the solution $g(t)$ to $[2T_1+2T_2, S+4T_2]$ such that the above estimates still hold for $t\in [2T_1+2T_2, S+4T_2]$. \end{claim} \begin{proof}By the assumption and Lemma {\ref{lem2.5}}, we have
$$|\dot \psi(t)-c(t)|_{C^0}\leq C(\sigma)C_1'(T_2, \Lambda_0)\epsilon, \qquad \forall t\in [2T_1+2T_2, S].$$ Note that in step 2, we know that $C(\sigma)C_1'(T_2, \Lambda_0)\epsilon<1.$ Then
$$|\dot\psi|(t)\leq 2,\qquad \forall t\in [2T_1+2T_2, S].$$
Therefore, $|\psi|, |\dot\psi|\leq B_3$. By Theorem \ref{theoRm} and the definition of $\Lambda_0,$ we have
$$|Rm|(t)\leq \Lambda_0,\qquad \forall t\in [2T_1+2T_2, S].$$ By Lemma {\ref{lem2.1}} and the definition of $T_2$,
$$|Rm|(t)\leq 2\Lambda_0, \;\;Ric(t)\geq 0, \qquad\forall t\in [S-2T_2, S +4T_2].$$ Thus, by Lemma \ref{lem2.2} and Lemma \ref{lem2.5} we have
$$|Ric-\omega|(t)\leq C_1'(T_2, \Lambda_0)\epsilon,\qquad \forall t\in [S, S +4T_2],$$ and
$$|\dot\psi-c(t)|_{C^0}\leq C(\sigma)C_1'(T_2, \Lambda_0)\epsilon, \qquad \forall t\in [S, S+4T_2].$$ Then we can estimate the $C^0$ norm of
$\psi$ for $t\in [S, S+4T_2],$ \begin{eqnarray*}|\psi(t)|&\leq
&|\psi|(S)+\Big|\int_S^{S+4T_2}\; \Big(\pd
{\psi}{s}-c(s)\Big)ds\Big|
+\Big|\int_0^{\infty}\;c(s)ds\Big|\\ &\leq &B_3+4T_2C(\sigma)C_1'(T_2, \Lambda_0)\epsilon+1 \\ &\leq &B_4. \end{eqnarray*} Then by Theorem \ref{lem2.8} and the definition of $\gamma$, the first eigenvalue of the Laplacian $\Delta_{\psi}$ $$\lambda_1(t)\geq 1+\gamma>1, \qquad\forall t\in [2T_1+2T_2, S+4T_2].$$ Note that $$\mu_0(2T_1+2T_2)=\frac 1V\int_M\;(\dot\psi-c(t))^2\omega_{\psi}^n
\leq(C(\sigma)C_1'\epsilon)^2 $$ and \begin{eqnarray*}\mu_1(2T_1+2T_2)&=&\frac 1V\int_M\;|\nabla\dot\psi|^2\omega_ {\psi}^n\\ &=&\frac 1V\int_M\;(\dot\psi-c(t))(R(\omega_{\psi})-n)\omega_{\psi} ^n\\
&\leq&C(\sigma)(C_1'\epsilon)^2. \end{eqnarray*} By Lemma \ref{lem2.19}, we can choose $\epsilon$ small enough such that \begin{eqnarray*}|\psi(t)|&\leq&
|\psi(2T_1+3T_2)|+\frac {C(n, \sigma)}{\alpha T_2^{\frac {m}{4}}} (\sqrt{\mu_0(2T_1+2T_2)}+\frac 1{\alpha_1}\mu_1(2T_1+2T_2)) +1\\ &\leq&B_2+\frac {C(n, \sigma)}{\alpha T_2^{\frac {m}{4}}}(1+\frac 1{\alpha_1}C_1'\epsilon) C(\sigma)C_1'\epsilon+1\\ &\leq &B_3 \end{eqnarray*} for $t\in [S, S+4T_2].$ Note that $\epsilon$ doesn't depend on $S$ here, so it won't become smaller as $S\rightarrow \infty.$ \end{proof}
\textbf{STEP 4.} By step 3, we know the bisectional curvature is uniformly bounded and the first eigenvalue $\lambda_1(t)\geq 1+\eta>1$ uniformly for some positive constant $\eta>0.$ Thus, following the argument in \cite{[chen-tian2]}, the K\"ahler-Ricci flow converges to a K\"ahler-Einstein metric exponentially fast. This theorem is proved.
\end{proof}
\end{flushleft}
\section{Proof of Theorem \ref{main2}}\label{section6} In this section, we shall use the pre-stable condition to drop the assumptions that $M$ has no nonzero holomorphic vector fields, and the dependence of the initial K\"ahler potential. The proof here is roughly the same as in the previous section, but there are some differences.
In the STEP 1 of the proof below, we will choose a new background metric at time $t=2T_1$, so the new K\"ahler potential with respect to the new background metric at $t=2T_1$ is $0$, and has nice estimates afterwards. Notice that all the estimates, particularly in Theorem \ref{theo4.18} and \ref{theoRm}, are essentially independent of the choice of the background metric. Therefore the choice of $\epsilon$ will not depend on the initial K\"ahler potential $\phi(0)$. This is why we can remove the assumption on the initial K\"ahler potential.
As in Theorem \ref{main}, the key point of the proof is to use the improved estimate on the first eigenvalue in Section \ref{section4.4.2} (see Claim \ref{last} below). Since the curvature tensors are bounded in some time interval, by Shi's estimates the gradient of curvature tensors are also bounded. Then the assumptions of Theorem \ref{theoRm} are satisfied, and we can use the estimate of the first eigenvalue.
Now we state the main result of this section.
\begin{theo}Suppose $M$ is pre-stable, and $E_1$ is bounded from below in $[\omega]$. For any $\delta, \Lambda>0,$ there exists a small positive constant $\epsilon(\delta, \Lambda)>0$ such that for any metric $\omega_0$ in the subspace ${\mathcal A}(\delta, \Lambda, \omega, \epsilon)$ of K\"ahler metrics
$$\{\omega_{\phi}=\omega+\sqrt{-1}\partial\bar\partial \phi\;|\; Ric(\omega_{\phi})>-1+\delta, \; |Rm|(\omega_{\phi})\leq \Lambda,\; E_1(\omega_{\phi})\leq \inf E_1+\epsilon \},$$ the K\"ahler-Ricci flow will deform it exponentially fast to a K\"ahler-Einstein metric in the limit. \end{theo}
\begin{flushleft} \begin{proof} Let $\omega_0\in {\mathcal A}(\delta, \Lambda, \omega, \epsilon)$, where $\epsilon$ will be determined later. By Lemma \ref{lem2.1} there exists a constant $T_1(\delta, \Lambda)$ such that
$$Ric(t)>-1+\frac {\delta}2 \; \;\;{ and}\;\;\; |Rm|(t)\leq 2\Lambda, \qquad\forall t\in
[0, 6T_1].$$ By Lemma {\ref{lem2.5}}, we can choose $\epsilon$ small enough so that \begin{equation}|Ric-\omega|(t)\leq C_1(T_1, \Lambda)\epsilon<\frac 12,
\qquad\forall t\in [2T_1, 6T_1],\label{8.29}\end{equation} and \begin{equation}|\dot
\phi-c(t)|\leq C(\sigma)C_1(T_1, \Lambda)\epsilon<1, \qquad\forall t\in [2T_1, 6T_1].\label{8.30}\end{equation}
\textbf{STEP 1.}(Choose a new background metric). Let $\underline \omega=\omega+\sqrt{-1}\partial\bar\partial \phi(2T_1)$ and let $\underline \phi(t)$ be the solution to the following K\"ahler-Ricci flow $$\left\{\begin{array}{l}
\pd {\underline \phi(t)}t=\log \frac {(\underline \omega+\sqrt{-1}\partial\bar\partial \underline\phi)^n}{\underline \omega^n} +\underline \phi-h_{\underline\omega},\qquad t\geq 2T_1,\\ \underline \phi(2T_1)=0.\\ \end{array} \right. $$ Here $h_{\underline \omega}$ satisfies the following conditions $$Ric(\underline \omega)-\underline \omega=\sqrt{-1}\partial\bar\partial h_{\underline \omega}\; \;\;{ and}\;\;\; \int_M\; h_{\underline \omega} \underline \omega^n=0.$$ Then the metric $\underline \omega(t)=\underline \omega+\sqrt{-1}\partial\bar\partial \underline\phi(t)$ satisfies $$\pd {}t\underline\omega(t)=-Ric (\underline \omega(t))+\underline \omega(t)\; \;\;{ and}\;\;\; \underline \omega(2T_1)=\omega+ \sqrt{-1}\partial\bar\partial \phi(2T_1).$$ By the uniqueness of K\"ahler-Ricci flow, we have $$\underline \omega(t)=\omega+\sqrt{-1}\partial\bar\partial \phi(t),\qquad \forall t\geq 2T_1.$$ Since the Sobolev constant is bounded and
$$|\Delta_{\underline \omega}h_{\underline \omega}|=|R(\underline \omega)-n|\leq C_1(T_1, \Lambda)\epsilon,$$
we have $$\Big|\pd {\underline \phi}{t}\Big|(2T_1)=|h_{\underline \omega}|\leq C(\sigma)C_1(T_1, \Lambda)\epsilon.$$ Since $E_1$ is decreasing in our case, we have $$E_1(\underline \omega)\leq E_1(\omega+\sqrt{-1}\partial\bar\partial \phi(0))\leq \inf E_1+\epsilon.$$ By Lemma \ref{lem5.9}, we have $$E_0(\underline \omega)\leq \inf E_0+\frac \ee2.$$ Thus, by Lemma \ref{lem2.6} we have
$$\frac 1V\int_{2T_1}^{\infty}\; e^{-t}\int_M\; \Big|\nabla \pd {\underline\phi}
{t}\Big|^2\underline\omega(t)^n\wedge dt <\frac {\epsilon}2<1.$$ Here we choose $\epsilon<2.$ Set $\psi(t)=\underline \phi(t)+C_0e^{t-2T_1},$ where
$$C_0=\frac 1V\int_{2T_1}^{\infty}\; e^{-t}\int_M\; \Big|\nabla \pd {\underline
\phi}{t}\Big|^2\underline\omega(t)^n\wedge dt-\frac 1V \int_M\; \pd {\underline
\phi}t\underline\omega(t)^n\Big|_{t=2T_1}. $$
Then $$|\psi(2T_1)|,\;\; \Big|\pd {\psi}{t}\Big|(2T_1)\leq 2,$$ and $$0< \underline c(t),\;\;\int_{2T_1}^{\infty}\; \underline c(t)dt<1,$$ where $\underline c(t)=\frac 1V\int_M\; \pd {\psi}{t} \underline\omega_{\psi}^n.$ Since
$$|\dot\psi-\underline c(t)|=|\dot\phi-c(t)|\leq C(\sigma)C_1(T_1, \Lambda) \epsilon,\qquad \forall t\in [2T_1, 6T_1],$$ we have
$$|\dot\psi|(t)\leq 2, \qquad \forall t\in [2T_1, 6T_1],$$
and \begin{eqnarray*} |\psi|(t)&\leq& |\psi|(2T_1)+\Big|\int_{2T_1}^{t}\;
(\dot\psi-\underline c(t))\Big|+\Big|\int_{2T}^{\infty}\; \underline c(s)ds\Big|\\&\leq&3+4T_1C(\sigma)C_1(T_1, \Lambda)\epsilon, \qquad \forall t\in [2T_1, 6T_1].\end{eqnarray*} Choose $\epsilon$ small enough such that $4T_1 C(\sigma)C_1(T_1, \Lambda)\epsilon<1$, and define $B_k=2k+2$. Then
$$|\psi|, |\dot\psi|\leq B_1, \qquad \forall t\in [2T_1, 6T_1].$$ By Theorem \ref{theoRm}, we have
$$|Rm|(t)\leq C_{11}(B_1, 2\Lambda, 1),\qquad \forall t\in [2T_1, 6T_1].$$ Here $C_{11}$ is a constant obtained in Theorem \ref{theoRm}. Let $\Lambda_0:= C_{11}(B_3, 2\Lambda, 1)$, then
$$|Rm|(t)\leq \Lambda_0,\qquad \forall t\in [2T_1, 6T_1].$$
\vskip 10pt \textbf{STEP 2.}(Estimates for $t\in [2T_1+2T_2, 2T_1+6T_2]$). By step 1, we have
$$|Ric-\omega|(2T_1)\leq C_1(T_1, \Lambda)\epsilon<\frac 12,\; \;\;{ and}\;\;\; |Rm|(2T_1)\leq
\Lambda_0.$$ By Lemma {\ref{lem2.1}}, there exists a constant $T_2(\frac 12, \Lambda_0)\in (0, T_1]$ such that \begin{equation}|Rm|(t)\leq 2\Lambda_0,\; \;\;{ and}\;\;\; Ric(t)\geq 0,\qquad \forall t\in [2T_1, 2T_1+6T_2].\label{7.28}\end{equation} Recall that $E_1\leq \inf E_1+\epsilon ,$ by Lemma \ref{lem2.2} and Lemma \ref{lem2.5} there exists a constant $C_1'(T_2, \Lambda_0)$ such that
$$|Ric-\omega|(t)\leq C_1'(T_2, \Lambda_0)\epsilon, \qquad \forall t\in [2T_1+2T_2, 2T_1+6T_2].$$ Choose $\epsilon$ small enough so that $C_1'(T_2, \Lambda_0)\epsilon<\frac 12$. Then by Lemma \ref{lem2.5},
$$|\dot\psi(t)-\underline c(t)|_{C^0}\leq C(\sigma)C_1'(T_2, \Lambda_0)\epsilon, \qquad \forall t\in [2T_1+2T_2, 2T_1+6T_2].$$ {{Choose $\epsilon$ small such that $C(\sigma)C_1'(T_2, \Lambda_0)\epsilon<1.$}} Thus, we can estimate the $C^0$ norm of $\psi$ for any $t\in [2T_1+2T_2, 2T_1+6T_2]$
\begin{eqnarray*}| {\psi}(t)|&\leq&|\psi|(2T_1+2T_2)+ \Big|\int_{2T_1+2T_2}^t\;
\Big(\pd { \psi}s-\underline c(s)\Big)ds\Big|+\Big|\int_{2T_1+2T_2}^{\infty}\; \underline c(s)ds\Big|\\ &\leq &B_1+4T_2 C(\sigma)C_1'(T_2, \Lambda_0)\epsilon +1\\&\leq &B_2. \end{eqnarray*} Here we choose $\epsilon$ small enough such that $4T_2 C(\sigma)C_1'(T_2, \Lambda_0)\epsilon<1$. Thus, by the definition of $\Lambda_0,$ we have
$$|Rm|(t)\leq \Lambda_0, \qquad\forall t\in [2T_1+2T_2, 2T_1+6T_2].$$
\textbf{STEP 3.} In this step, we want to prove the following claim: \begin{claim}\label{last}For any positive number $S\geq 2T_1+6T_2$, if
$$|Ric-\omega|(t)\leq C_1'(T_2, \Lambda_0)\epsilon<\frac 12, \; \;\;{ and}\;\;\; |Rm|(t)\leq \Lambda_0,\qquad \forall t\in [2T_1+2T_2, S],$$ then we can extend the solution $g(t)$ to $[2T_1+2T_2, S+4T_2]$ such that the above estimates still hold for $t\in [2T_1+2T_2, S+4T_2]$. \end{claim} \begin{proof} By Lemma {\ref{lem2.1}} and the definition of $T_2$,
$$|Rm|(t)\leq 2\Lambda_0,\; Ric(t)\geq 0, \qquad\forall t\in [2T_1+2T_2, S +4T_2].$$ Thus, by Lemma \ref{lem2.2} and Lemma \ref{lem2.5} we have
$$|Ric-\omega|(t)\leq C_1'(T_2, \Lambda_0)\epsilon,\qquad \forall t\in [S-2T_2, S +4T_2].$$ Therefore, we have
$$|\dot\psi(t)-\underline c(t)|_{C^0}\leq C(\sigma)C_1'(T_2, \Lambda_0)\epsilon, \qquad \forall t\in [2T_1+2T_2, S+4T_2].$$ By Theorem \ref{main4} the $K$-energy is bounded from below, then the Futaki invariant vanishes. Therefore, we have $$\int_M\; X(\dot\psi)\underline\omega_{\psi}^n=0,\qquad \forall X\in \eta_r(M, J).$$ By the assumption that $M$ is pre-stable and Theorem \ref{theo4.18}, if $\epsilon$ is small enough, there exists a constant $\gamma(C_1'\epsilon, 2\Lambda_0)$ such that
$$\int_M\; |\nabla \dot\psi|\underline\omega^n_{\psi}\geq (1+\gamma)\int_M\; |
\dot\psi-\underline c(t)|^2\underline\omega^n_{\psi}.$$ Therefore, Lemma \ref{lem2.14} still holds, i.e. there exists a constant $\alpha(\gamma, C_1'\epsilon, \sigma)>0$ such that for any $t\in [2T_1+2T_2, S+4T_2]$ $$\mu_1(t)\leq \mu_1(2T_1+2T_2)e^{-\alpha(t-2T_1-2T_2)},$$ and $$\mu_0(t)\leq \frac {1}{1-C_1'\epsilon}\mu_1(t)\leq 2\mu_1(2T_1+2T_2)e^{- \alpha(t-2T_1-2T_2)}. $$
Then by Lemma \ref{lem2.19}, we can choose $\epsilon$ small enough such that \begin{eqnarray*}|\psi(t)|&\leq& |\psi(2T_1+3T_2)|+\frac {C_{10}(n, \sigma)}{\alpha T_2^{\frac {m}{4}}} (\sqrt{\mu_0(2T_1+2T_2)}+\frac 1{\alpha_1}\mu_1(2T_1+2T_2)) +1\\ &\leq&B_2+\frac {C_{10}(n, \sigma)}{\alpha T_2^{\frac {m}{4}}}(1+\frac 1{\alpha_1}C_1'\epsilon) C(\sigma)C_1'\epsilon+1\\ &\leq &B_3 \end{eqnarray*} for $t\in [S, S+4T_2].$ By the definition of $\Lambda_0,$ we have
$$|Rm|(t)\leq \Lambda_0,\qquad \forall t\in [S, S+4T_2].$$ \end{proof}
\textbf{STEP 4.} By step 3, we know the bisectional curvature is uniformly bounded and the $W^{1, 2}$ norm of $\underline {\dot \phi}-\underline c(t)$ decays exponentially. Thus, following the argument in \cite{[chen-tian2]}, the K\"ahler-Ricci flow converges to a K\"ahler-Einstein metric exponentially fast. This theorem is proved.
\end{proof}
\end{flushleft}
\vskip3mm
Xiuxiong Chen, Department of Mathematics, University of Wisconsin-Madison, Madison WI 53706, USA; [email protected]\\
Haozhao Li, School of Mathematical Sciences, Peking University, Beijing, 100871, P.R. China; [email protected]\\
Bing Wang, Department of Mathematics, University of Wisconsin-Madison, Madison WI 53706, USA; [email protected]\\
\end{document}
|
arXiv
|
{
"id": "0609694.tex",
"language_detection_score": 0.5970340371131897,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\title{An introduction to Lie group integrators -- basics, new developments and applications}
\author[ntnu]{Elena Celledoni}
\ead{[email protected]}
\author[ntnu]{H{\aa}kon Marthinsen} \ead{[email protected]}
\author[ntnu]{Brynjulf Owren}
\ead{[email protected]}
\address[ntnu]{Department of Mathematical Sciences, NTNU, N--7491 Trondheim, Norway}
\begin{abstract} We give a short and elementary introduction to Lie group methods. A selection of applications of Lie group integrators are discussed. Finally, a family of symplectic integrators on cotangent bundles of Lie groups is presented and the notion of discrete gradient methods is generalised to Lie groups. \end{abstract}
\begin{keyword} Lie group integrators, symplectic methods, integral preserving methods \end{keyword} \end{frontmatter}
\section{Introduction} The significance of the geometry of differential equations was well understood already in the nineteenth century, and in the last few decades such aspects have played an increasing role in numerical methods for differential equations. Nowadays, there is a rich selection of integrators which preserve properties like symplecticity, reversibility, phase volume and first integrals, either exactly or approximately over long times~\cite{hairer10gni}. Differential equations are inherently connected to Lie groups, and in fact one often sees applications in which the phase space is a Lie group or a manifold with a Lie group action. In the early nineties, two important papers appeared which used the Lie group structure directly as a building block in the numerical methods. Crouch and Grossman \cite{crouch93nio} suggested to advance the numerical solution by computing flows of vector fields in some Lie algebra. Lewis and Simo \cite{lewis94caf} wrote an influential paper on Lie group based integrators for Hamiltonian problems, considering the preservation of symplecticity, momentum and energy. These ideas were developed in a systematic way throughout the nineties by several authors. In a series of three papers, Munthe-Kaas \cite{munthe-kaas95lbt, munthe-kaas98rkm,munthe-kaas99hor} presented what are now known as the Runge--Kutta--Munthe-Kaas methods. By the turn of the millennium, a survey paper \cite{iserles00lgm} summarised most of what was known by then about Lie group integrators. More recently a survey paper on structure preservation appeared with part of it dedicated to Lie group integrators \cite{christiansen11tis}.
The purpose of the present paper is three-fold. First, in section~\ref{sec:basics} we give an elementary, geometric introduction to the ideas behind Lie group integrators. Secondly, we present some examples of applications of Lie group integrators in sections \ref{sec:appcm} and \ref{sec:dataanalysis}. There are many such examples to choose from, and we give here only a few teasers. These first four sections should be read as a survey. But in the last two section, new material is presented. Symplectic Lie group integrators have been known for some time, derived by Marsden and coauthors \cite{marsden01dma} by means of variational principles. In section~\ref{sec:symplie} we consider a group structure on the cotangent bundle of a Lie group and derive symplectic Lie group integrators using the model for vector fields on manifolds defined by Munthe-Kaas in~\cite{munthe-kaas99hor}. In section~\ref{sec:discdiff} we extend the notion of discrete gradient methods as proposed by Gonzalez \cite{gonzalez96tia} to Lie groups, and thereby we obtain a general method for preserving first integrals in differential equations on Lie groups.
We would also like to briefly mention some of the issues we are \emph{not} pursuing in this article. One is the important family of Lie group integrators for problems of linear type, including methods based on the Magnus and Fer expansions. An excellent review of the history, theory and applications of such integrators can be found in \cite{blanes09tme}. We will also skip all discussions of order analysis of Lie group integrators. This is a large area by itself which involves technical tools and mathematical theory which we do not wish to include in this relatively elementary exposition. There have been several new developments in this area recently, in particular by Lundervold and Munthe-Kaas, see e.g.\ \cite{lundervold11hao}.
\section{Lie group integrators}
\label{sec:basics} The simplest consistent method for solving ordinary differential equations is the Euler method. For an initial value problem of the form \begin{equation*}
\dot{y} = F(y),\quad y(0)=y_0, \end{equation*}
one takes a small time increment $h$, and approximates $y(h)$ by the simple formula
\begin{equation*}
y_{1} = y_0 + hF(y_0),
\end{equation*} advancing along the straight line coinciding with the tangent at $y_0$. Another way of thinking about the Euler method is to consider the constant vector field $F_{y_0}(y) \coloneqq F(y_0)$ obtained by parallel translating the vector~$F(y_0)$ to all points of phase space. A step of the Euler method is nothing else than computing the exact $h$-flow of this simple vector field starting at $y_0$. In Lie group integrators, the same principle is used, but allowing for more advanced vector fields than the constant ones. A Lie group generalisation of the Euler method is called the Lie--Euler method, and we shall illustrate its use through an example \cite{crouch93nio}. \paragraph{Example, the Duffing equation} Consider the system in $\mathbf{R}^2$ \begin{equation} \label{eq:duffing} \begin{aligned}
\dot{x} &= y, \\
\dot{y} &= -a x - b x^3, \end{aligned} \qquad a\geq 0, b\geq 0, \end{equation} a model used to describe the buckling of an elastic beam. Locally, near a point $(x_0,y_0)$ we could use the approximate system \begin{equation} \begin{alignedat}{2} \label{eq:duff:sl2:frozen}
\dot{x} &= y, &\qquad x(0) &= x_0,\\
\dot{y} &= -(a+b x_0^2) x, & y(0) &= y_0, \end{alignedat} \end{equation} which has the exact solution \begin{equation}\label{eq:duffing:sl2:flow}
\bar{x}(t) = x_0\cos\omega t+ \frac{y_0}{\omega}\sin\omega t,\quad
\bar{y}(t) = y_0\cos\omega t- \omega x_0\sin\omega t,\quad \omega=\sqrt{a+bx_0^2}. \end{equation} Alternatively, we may consider
the local problem \[ \begin{aligned}
\dot{x} &= y, \\
\dot{y} &= -ax - bx_0^3, \end{aligned} \] having exact solution \[ \begin{aligned} \bar{x}(t) &= x_0\cos\alpha t + \frac{y_0}{\alpha}\sin\alpha t + b\,x_0^3\,\frac{\cos\alpha t-1}{\alpha^2},\\ \bar{y}(t) &= y_0\cos\alpha t - \alpha x_0\sin\alpha t - b\,x_0^3\, \frac{\sin\alpha t}{\alpha}, \end{aligned} \qquad \alpha=\sqrt{a}. \] In each of the two cases, one may take $x_1=\bar{x}(h)$, $y_1=\bar{y}(h)$ as the numerical approximation at time $t=h$. The same procedure is repeated in subsequent steps. \begin{figure}
\caption{$(\mathbf{R}^d,+)$-frozen vector field (left) and $\mathfrak{sl}(2)$-frozen vector field (right) for the Duffing equation. Both are frozen at $(x_0,y_0)=(0.75,0.75)$. The thin black curve in each plot shows the flows of the frozen vector fields $0\leq t\leq 20$. The thicker curve in each plot is the exact flow of the Duffing equation.
}
\label{fig:duffing}
\end{figure}
A common framework for discussing these two cases is provided by the use of frames, i.e.\ a set of
of vector fields which at each point is spanning the tangent space.
In the first case, the numerical method applies the frame \begin{equation} \label{eq:XYsl2}
X = \begin{bmatrix} y\\0 \end{bmatrix} \eqqcolon y\, \partial x,\quad
Y = \begin{bmatrix} 0\\x \end{bmatrix} \eqqcolon x\, \partial y. \end{equation} Taking the negative Jacobi--Lie bracket (also called the commutator) between $X$ and $Y$ yields the third element of the standard basis for the Lie algebra $\mathfrak{sl}(2)$, i.e.\ \begin{equation} \label{eq:Hsl2}
H = -[X,Y] = x\,\partial x - y\,\partial y, \end{equation} so that the frame may be augmented to consist of $\{X, Y, H\}$.
In the second case, the vector fields $E_1=y\,\partial x - ax\,\partial y$ and $E_2=\partial y$ can be used as a frame, but again we choose to augment these two fields with the commutator $E_3=-[E_1,E_2]=\partial x$ to obtain the Lie algebra of the special Euclidean group $\SE(2)$ consisting of translations and rotations in the plane. The situation is illustrated in Figure~\ref{fig:duffing}. In the left part, we have considered the constant vector field corresponding to the Duffing vector field evaluated at $(x_0,y_0)=(0.75,0.75)$, and the exact flow of this constant field is just the usual Euler method, a straight line. In the right part, we have plotted the vector field defined in \eqref{eq:duff:sl2:frozen} with the same $(x_0,y_0)$ along with its flow~\eqref{eq:duffing:sl2:flow}. The exact flow of \eqref{eq:duffing} is shown in both plots (thick curve).
In general, a way to think about Lie group integrators is that we have a manifold $M$ where there is such a frame available; $\{E_1,\ldots,E_d\}$ such that at any point $p\in M$ one has \[
\operatorname{span}\{E_1(p),\ldots,E_d(p)\} = T_pM. \] Frames with this property are said to be locally transitive. The frame may be a linear space or in many cases even a Lie algebra $\ensuremath{\mathfrak{g}}$ of vector fields. In the example with Duffing's equation, the set $\{X, Y, H\}$ is locally transitive on $\mathbf{R}^2\setminus\{0\}$ and $\{E_1,E_2,E_3\}$ is locally transitive on $\mathbf{R}^2$.
Given an arbitrary vector field $F$ on $M$, then at any point $p\in M$ there exists a vector field $F_p$ in the span of the frame vector fields such that $F_p(p) = F(p)$. An explicit way of writing this is by using a set of basis vector fields $E_1,\ldots,E_d$ for $\ensuremath{\mathfrak{g}}$, such that any smooth vector field $F$ has a representation \begin{equation} \label{VFframerep}
F(y) = \sum_{k=1}^d f_k(y) E_k(y), \end{equation} for some functions $f_k \colon M\rightarrow\mathbf{R}$. The vector fields $F_p\in\ensuremath{\mathfrak{g}}$, called \emph{vector fields with frozen coefficients} by Crouch and Grossman \cite{crouch93nio}, are then obtained as \[
F_p(y) = \sum_{k=1}^d f_k(p) E_k(y). \] In the example with the Duffing equation we took $E_1=X, E_2=Y$,
$f_1(x,y)=1$ and $f_2(x,y)=-(a+bx^2)$. The Lie--Euler method reads in general \begin{equation} \label{eq:Lie--Euler}
y_{n+1} = \exp(hF_{y_n})y_n, \end{equation} where $\exp$ denotes the flow of a vector field.
A more interesting example, also found in \cite{crouch93nio} is obtained by choosing $M=S^2$, the 2-sphere. A suitable way to induce movements of the sphere is that of rotations, that is, by introducing the Lie group $\SO(3)$ consisting of orthogonal matrices with unit determinant. The corresponding Lie algebra $\mathfrak{so}(3)$ of vector fields are spanned by \[
E_1(x,y,z)=-z\,\partial y + y\,\partial z,\quad
E_2(x,y,z)=z\,\partial x - x\,\partial z,\quad
E_3(x,y,z)=-y\,\partial x + x\,\partial y. \] We note that $xE_1(x,y,z)+yE_2(x,y,z)+zE_3(x,y,z)=0$, showing that the functions $f_k$ in \eqref{VFframerep} are not unique. A famous example of a system whose solution evolves on $S^2$ is the free rigid body Euler equations \begin{equation} \label{eq:euler_frb}
\dot{x} = \left(\frac1{I_3}-\frac1{I_2}\right)\,yz, \quad
\dot{y} = \left(\frac1{I_1}-\frac1{I_3}\right)\,xz, \quad
\dot{z} = \left(\frac1{I_2}-\frac1{I_1}\right)\,xy, \end{equation} where $x, y, z$ are the coordinates of the angular momentum relative to the body, and $I_1, I_2, I_3$ are the principal moments of inertia. A choice of representation \eqref{VFframerep} is obtained with \[
f_1(x,y,z) = -\frac{x}{I_1},\quad f_2(x,y,z)=-\frac{y}{I_2},\quad f_3(x,y,z) = -\frac{z}{I_3}, \] so that the ODE vector field can be expressed in the form \[ F(x,y,z)= -\frac{x}{I_1} \begin{bmatrix*}[r] 0\\ -z\\ y \end{bmatrix*} -\frac{y}{I_2} \begin{bmatrix*}[r] z\\ 0\\ -x \end{bmatrix*} -\frac{z}{I_3} \begin{bmatrix*}[r] -y\\ x\\ 0 \end{bmatrix*}. \] We compute the vector field with coefficients frozen at $p_0=(x_0,y_0,z_0)$, \[ F_{p_0}(x,y,z)= \mathbf{F}_{p_0} \begin{bmatrix} x\\ y \\ z \end{bmatrix} \coloneqq \begin{bmatrix*}[r] 0 & \frac{z_0}{I_3} & -\frac{y_0}{I_2} \\ -\frac{z_0}{I_3} & 0 & \frac{x_0}{I_1} \\ \frac{y_0}{I_2} & -\frac{x_0}{I_1} & 0 \end{bmatrix*} \begin{bmatrix} x\\ y \\ z \end{bmatrix}. \] The $h$-flow of this vector field is the solution of a linear system of ODEs and can be expressed in terms of the matrix exponential $\mathtt{expm}(h\mathbf{F}_{p_0})$. The Lie--Euler method can be expressed as follows:
\begin{algorithmic}
\State $p_0 \gets (x_0,y_0,z_0)$
\For{$n \gets 0, 1, \dotsc$}
\State $p_{n+1} \gets \mathtt{expm}(h\mathbf{F}_{p_n}) p_n$
\EndFor \end{algorithmic}
Notice that the matrix to be exponentiated belongs to the matrix group $\mathfrak{so}(3)$ of real skew-symmetric matrices. The celebrated Rodrigues' formula \[
\mathtt{expm}(A) = I + \frac{\sin\alpha}{\alpha} A + \frac{1-\cos\alpha}{\alpha^2}A^2,\qquad
\alpha^2=\lVert A\rVert_2^2=\tfrac12\lVert A\rVert_F^2,\quad A\in\mathfrak{so}(3), \] provides an inexpensive way to compute this.
Whereas the notion of frames was used by Crouch and Grossman in their pioneering work \cite{crouch93nio}, a different type of notation was used in a series of papers by Munthe-Kaas \cite{munthe-kaas95lbt,munthe-kaas98rkm,munthe-kaas99hor}, see also \cite{lundervold11hao} for a more modern treatment. Let $G$ be a finite dimensional Lie group acting transitively on a manifold $M$. A Lie group action is generally a map from $G\times M$ into $M$, having the properties that \begin{equation*} \label{eq:LGaction}
e\cdot m=m,\;\forall m\in M,\qquad g\cdot(h\cdot m)=(g\cdot h)\cdot m,\ \forall g,h\in G,\; m\in M, \end{equation*} where $e$ is the group identity element, and the first $\cdot$ in the right hand side of the second identity is the group product. Transitivity means that for any two points $m_1, m_2\in M$ there exists a group element $g\in G$ such that $m_2=g\cdot m_1$. We denote the Lie algebra of $G$ by $\ensuremath{\mathfrak{g}}$. For any element $\xi\in\ensuremath{\mathfrak{g}}$ there exists a vector field on~$M$ \begin{equation} \label{eq:groupaction}
X_{\xi}(m) = \left.\frac{\ensuremath{\mathrm{d}}}{\ensuremath{\mathrm{d}} t}\right\rvert_{t=0} \exp(t\xi)\cdot m \eqqcolon \lambda_*(\xi)(m). \end{equation} Munthe-Kaas introduced a generic representation of a vector field $F\in\mathcal{X}(M)$ by a map $f \colon M\rightarrow\ensuremath{\mathfrak{g}}$ such that \begin{equation} \label{eq:genpres}
F(m) =\lambda_*(f(m))(m). \end{equation} The corresponding frame is obtained as $E_i=\lambda_*(e_i)$ where $\{e_1,\ldots,e_d\}$ is some basis for $\ensuremath{\mathfrak{g}}$ and one chooses the functions $f_i \colon M\rightarrow\mathbf{R}$ such that $f(m) = \sum_{i=1}^d f_i(m) e_i$. The map $\lambda_*$ is an anti-homomorphism of the Lie algebra $\ensuremath{\mathfrak{g}}$ into the Lie algebra of vector fields $\mathcal{X}(M)$ under the Jacobi--Lie bracket, meaning that \[
\lambda_*([X_m, Y_m]_{\ensuremath{\mathfrak{g}}}) =- [\lambda_*(X_m), \lambda_*(Y_m)]_{\mathrm{JL}}. \] This separation of the Lie algebra $\ensuremath{\mathfrak{g}}$ from the manifold $M$ allows for more flexibility in the way we represent the frame vector fields. For instance, in the example with Duffing's equation and the use of $\mathfrak{sl}(2)$, we could have used the matrix Lie algebra with basis elements \[
X_m =
\begin{bmatrix}
0 & 1 \\ 0 & 0
\end{bmatrix},\quad
Y_m =
\begin{bmatrix}
0 & 0 \\ 1 & 0
\end{bmatrix},\quad
H_m =
\begin{bmatrix*}[r]
1 & 0 \\ 0 & -1
\end{bmatrix*}, \] rather than the basis of vector fields \eqref{eq:XYsl2}, \eqref{eq:Hsl2}. The group action by $g\in \SL(2)$ on a point $m\in\mathbf{R}^2$ would then be simply $g\cdot m$, matrix-vector multiplication, and the $\exp$ in \eqref{eq:groupaction} would be the matrix exponential. The map~$f(x,y)$ would in this case be \[
f \colon (x,y) \mapsto \begin{bmatrix}
0 & y \\ -(a+bx^2) & 0
\end{bmatrix}, \] but note that since the dimension of the manifold is just two whereas the dimension of $\mathfrak{sl}(2)$ is three, there is freedom in the choice of $f$. In the example we chose not to use the third basis element $H$.
\subsection{Generalising Runge--Kutta methods} \label{GeneralizingRK} In order to construct general schemes, as for instance a Lie group version of the Runge--Kutta methods, one needs to introduce intermediate stage values. This can be achieved in a number of different ways. They all have in common that when the methods are applied in Euclidean space where the Lie group is $(\mathbf{R}^m,+)$, they reduce to conventional Runge--Kutta schemes. Let us begin by studying the simple second order Heun method, sometimes called the improved Euler method. \[
k_1 = F(y_n),\quad k_2=F(y_n+hk_1),\qquad y_{n+1} = y_n + \tfrac12 h(k_1+k_2). \]
Geometrically, we may think of $k_1$ and $k_2$ as constant vector fields, coinciding with the exact ODE $F(y)$ at the points $y_n$ and $y_n+hk_1$ respectively. The update $y_{n+1}$ can be interpreted in at least three different ways, \begin{equation} \label{eq:flowalts}
\exp\left(\frac{h}{2}(k_1+k_2)\right) \cdot y_n, \quad
\exp\left(\frac{h}{2}k_1\right) \cdot \exp\left(\frac{h}{2}k_2\right) \cdot y_n, \quad
\exp\left(\frac{h}{2}k_2\right) \cdot \exp\left(\frac{h}{2}k_1\right) \cdot y_n. \end{equation}
The first is an example of a Runge--Kutta--Munthe-Kaas method and the second is an example of a Crouch--Grossman method. All three fit into the framework of \emph{commutator-free} Lie group methods. All three suggestions above are generalisations that will reduce to Heun's method in $(\mathbf{R}^m,+)$. In principle we could extend the idea to Runge--Kutta methods with several stages
\[
y_{n+1} = y_n + h\sum_{i=1}^sb_iF(Y_i),\quad Y_i = y_n + h\sum_{j=1}^s a_{ij} F(Y_j),\
i=1,\ldots,s,
\]
by for instance interpreting the summed expressions as vector fields with frozen coefficients whose flows we apply to the point $y_n\in M$.
But it is unfortunately not true that one in this way will retain the order of the Runge--Kutta method when applied to cases where the acting group is non-abelian.
Let us first describe methods as proposed by Munthe-Kaas \cite{munthe-kaas99hor}, where one may think of the method simply as a change of variable. As before, we assume that the action of $G$ on $M$ is locally transitive. Since the exponential mapping is a local diffeomorphism in some open set containing $0\in\ensuremath{\mathfrak{g}}$, it is possible to represent \emph{any} smooth curve $y(t)$ on $M$ in some neighbourhood of a point $p\in M$ by means of a curve
$\sigma(t)$ through the origin of $\ensuremath{\mathfrak{g}}$ as follows
\begin{equation} \label{eq:solrep}
y(t) = \exp(\sigma(t))\cdot p,\qquad \sigma(0)=0,
\end{equation}
though $\sigma(t)$ is not necessarily unique. We may differentiate this curve with respect to $t$ to obtain \begin{equation} \label{eq:doty}
\dot{y}(t) = \lambda_*\bigl(\dexp_{\sigma(t)}\dot{\sigma}(t)\bigr)(y(t)) = F(y(t))=
\lambda_*\bigl(f(\exp(\sigma(t))\cdot p)\bigr)(y(t)). \end{equation} The details are given in \cite{munthe-kaas99hor} and the map $\dexp_\sigma \colon \ensuremath{\mathfrak{g}}\rightarrow\ensuremath{\mathfrak{g}}$ was derived by Hausdorff in \cite{hausdorff06dse} as an infinite series of commutators \begin{equation} \label{eq:dexp}
\dexp_{\sigma}(v) = v + \frac{1}{2}[\sigma,v] + \frac{1}{6}[\sigma,[\sigma,v]]+\dotsb
= \sum_{k=0}^\infty \frac{1}{(k+1)!} \ad_\sigma^k v = \left.\frac{\exp(z)-1}{z}\right\rvert_{z=\ad_\sigma} v, \end{equation} with the usual definition of $\ad_u(v)$ as the commutator $[u,v]$. The map $\lambda_*$ does not have to be injective, but a sufficient condition for \eqref{eq:doty} to hold is that \[
\dot{\sigma} = \dexp_{\sigma}^{-1} (f(\exp(\sigma)\cdot p)). \] This is a differential equation for $\sigma(t)$ on a linear space, and one may choose any conventional integrator for solving it. The map $\dexp_\sigma^{-1} \colon \ensuremath{\mathfrak{g}}\rightarrow\ensuremath{\mathfrak{g}}$ is the inverse of $\dexp_\sigma$ and can also be obtained by differentiating the logarithm, i.e.\ the inverse of the exponential map.
From \eqref{eq:dexp} we find that one can write $\dexp_{\sigma}^{-1}(v)$ as \begin{equation} \label{eq:dexpinv} \dexp_{\sigma}^{-1}(v) = \left.\frac{z}{\exp(z)-1}\right\rvert_{z=\ad_\sigma}v = v - \frac{1}{2}[\sigma,v] + \frac{1}{12}[\sigma, [\sigma, v]]+\cdots. \end{equation} The coefficients appearing in this expansion are scaled Bernoulli numbers $\frac{B_k}{k!}$, and $B_{2k+1}=0$ for all $k\geq 1$. One step of the resulting Runge--Kutta--Munthe-Kaas method is then expressed in terms of evaluations of the map $f$ as follows \begin{align*} y_{1} &= \exp\Bigl(h\sum_{i=1}^s b_i k_i\Bigr) \cdot y_0, \\ k_i &= \dexp_{h\sum_j a_{ij}k_j}^{-1} f\biggl(\exp\Bigl(h\sum_j a_{ij}k_j\Bigr) \cdot y_0\biggr),\quad i=1,\ldots,s. \end{align*} This is not so surprising seen from the perspective of the first alternative in \eqref{eq:flowalts}, the main difference is that the stages $k_i$ corresponding to the frozen vector fields $\lambda_*(k_i)$ need to be ``corrected'' by the $\dexp^{-1}$ map. Including this map in computational algorithms may seem awkward, however, fortunately truncated versions of \eqref{eq:dexpinv} may frequently be used. In fact, by applying some clever tricks involving graded free Lie algebras, one can in many cases replace $\dexp^{-1}$ with a low order Lie polynomial while retaining the convergence order of the original Runge--Kutta method. Details of this can be found in \cite{munthe-kaas99cia, casas03cel}.
There are also some important cases of Lie algebras for which $\dexp_\sigma^{-1}$ can be computed exactly in terms of elementary functions, among those is $\mathfrak{so}(3)$ reported in \cite{celledoni03lgm}. Notice that the representation \eqref{eq:solrep} does not depend on the use of the exponential map from $\ensuremath{\mathfrak{g}}$ to $G$. In principle, one can replace this map with any local diffeomorphism $\varphi$, where one usually scales $\varphi$ such that $\varphi(0)=e$ and $T_0\varphi = \mathrm{Id}_\ensuremath{\mathfrak{g}}$. An example of such map is the Cayley transformation \cite{diele98tct} which can be used for matrix Lie groups of the type $G_P = \{X\in\mathbf{R}^{n\times n} \mid X^TPX = P\}$ for a nonsingular $n\times n$-matrix $P$.
These include the orthogonal group~$\LieO(n)=G_I$ and the linear symplectic group~$\Sp(n)=G_J$ where $J$ the skew-symmetric matrix of the standard symplectic form. Another possibility is to replace the exponential map by canonical coordinates of the second kind \cite{owren01imb}.
We present here the well-known Runge--Kutta--Munthe-Kaas method based on the popular fourth order method of Kutta \cite{kutta01bzn}, having Butcher tableau \begin{equation} \label{eq:kutta} \morespacearray
\begin{array}{r|rrrr} 0 & \\ \tfrac{1}{2} & \tfrac{1}{2} \\ \tfrac{1}{2} & 0 & \tfrac{1}{2} \\ 1 & 0 & 0 & 1 \\ \hline
& \tfrac{1}{6} & \tfrac{1}{3} & \tfrac{1}{3} & \tfrac{1}{6} \end{array} \end{equation} In the Lie group method, the $\dexp^{-1}$ map has been replaced by the optimal Lie polynomials. \begin{align*}
k_1 &= h f(y_0),\\
k_2 &= h f(\exp(\tfrac{1}{2}k_1) \cdot y_0), \\
k_3 &= h f(\exp(\tfrac{1}{2}k_2-\tfrac{1}{8}[k_1,k_2])\cdot y_0), \\
k_4 &= h f(\exp(k_3)\cdot y_0), & \\
y_1 &= \exp(\tfrac{1}{6}(k_1+2k_2+2k_3+k_4-\tfrac12[k_1,k_4]))\cdot y_0. \end{align*} An important advantage of the Runge--Kutta--Munthe-Kaas schemes is that it is easy to preserve the convergence order when extending them to Lie group integrators. This is not the case with for instance the schemes of Crouch and Grossman \cite{crouch93nio, owren99rkm}, where it is necessary to develop order conditions for the non-abelian case. This is also true for the commutator-free methods developed by Celledoni et al.\ \cite{celledoni03cfl}. In fact, these methods include those of Crouch and Grossman. The idea here is to allow compositions of exponentials or flows instead of commutator corrections. With stages $k_1,\ldots, k_s$ in the Lie algebra, one includes expressions of the form \[
\exp\Bigl(\sum_{i}\beta_J^i k_i\Bigr) \dotsm \exp\Bigl(\sum_{i}\beta_2^i k_i\Bigr) \cdot \exp\Bigl(\sum_{i}\beta_1^i k_i\Bigr) \cdot y_0, \] both in the definition of the stages and the update itself. In some cases it is also possible to reuse flow calculations from one stage to another, and thereby lower the computational cost of the scheme. An extension of \eqref{eq:kutta} can be obtained as follows, setting $k_i=hf(Y_i)$ for all $i$, \begin{align*}
Y_1&= y_0, \\
Y_2&=\exp(\tfrac{1}{2}k_1)\cdot y_0, \\
Y_3&=\exp(\tfrac{1}{2}k_2)\cdot y_0 \\
Y_4&= \exp(k_3-\tfrac{1}{2}k_1)\cdot Y_2, \\
y_{\frac{1}{2}} &=
\exp(\tfrac{1}{12}(3k_1+2k_2+2k_3-k_4))\cdot y_0, \\
y_{1} &=
\exp(\tfrac{1}{12}(-k_1+2k_2+2k_3+3k_4))\cdot y_{\frac{1}{2}}.
\end{align*} Note in particular in this example how the expression for $Y_4$ involves $Y_2$ and thereby one exponential calculation has been saved.
\subsection{A plenitude of group actions} We saw in the first examples with Duffing's equation that the manifold $M$, the group $G$ and even the way $G$ acts on $M$ can be chosen in different ways. It is not obvious which action is the best or suits the purpose in the problem at hand. Most examples we know from the literature are using matrix Lie groups~$G\subseteq \GL(n)$, but the choice of group action depends on the problem and the objectives of the simulation. We give here several examples of situations where Lie group integrators can be used.
\paragraph{$G$ acting on $G$} In the case $M=G$, it is natural to use either left or right multiplication as the action \[
L_g(m) = g\cdot m\quad\text{or}\quad R_g(m) = m\cdot g,\qquad g,m\in G. \] The correspondence between the vector field $F\in\mathcal{X}(M)$ and the map \eqref{eq:genpres} is then just the tangent map of left or right multiplication \[
F(g) = T_eL_g(f(g))\quad\text{or}\quad F(g)=T_eR_{g}(\tilde{f}(g)),\quad g\in G. \] When working with matrices, this simply amounts to setting $F(g) = g\cdot f(g)$ or $F(g) = \tilde{f}(g) \cdot g$. Note that $\tilde{f}(g)$ is related to $f(g)$ through the adjoint representation of $G$, $\Ad \colon G\rightarrow \operatorname{Aut}(\ensuremath{\mathfrak{g}})$,
\[
\tilde{f}(g) = \Ad_g f(g),\qquad \Ad_g = T_eL_g\circ T_eR_g^{-1}. \]
\paragraph{The affine group and its use in semilinear PDE methods} Lie group integrators can also be used for approximating the solution to partial differential equations, the most obvious choice of PDE model being the semilinear problem \begin{equation} \label{eq:semilinear}
u_t = Lu + N(u), \end{equation} where $L$ is a linear differential operator and $N(u)$ is some nonlinear map, typically containing derivatives of lower order than $L$. After discretising in space, \eqref{eq:semilinear} is turned into a system of $n_d$ ODEs, for some large $n_d$, $L$ becomes an $n_d\times n_d$-matrix, and $N \colon \mathbf{R}^{n_d}\rightarrow \mathbf{R}^{n_d}$ a nonlinear function. We may now as in \cite{munthe-kaas99hor} introduce the action on $\mathbf{R}^{n_d}$ by some subgroup of the affine group represented as the semidirect product $G=\GL(n_d)\ltimes\mathbf{R}^{n_d}$. The group product, identity, and inverse are given as \[
(A_1,b_1)\cdot (A_2,b_2) = (A_1 A_2, A_1b_2+b_1),\quad
e=(I,0),\quad (A,b)^{-1}=(A^{-1}, -A^{-1}b). \] The action on $\mathbf{R}^{n_d}$ is \[
(A,b)\cdot x = Ax + b,\qquad (A,b)\in G,\ x\in\mathbf{R}^{n_d}, \] and the Lie algebra and commutator are given as \[
\ensuremath{\mathfrak{g}} = (\xi, c),\ \xi\in\mathfrak{gl}(n_d),\ c\in\mathbf{R}^{n_d},\quad
[(\xi_1,c_1),(\xi_2,c_2)] = ([\xi_1,\xi_2], \xi_1c_2-\xi_2c_1+c_1). \] In many interesting PDEs, the operator $L$ is constant, so it makes sense to consider the $n_d+1$-dimensional subalgebra $\ensuremath{\mathfrak{g}}_L$ of $\ensuremath{\mathfrak{g}}$ consisting of elements $(\alpha L, c)$ where $\alpha\in\mathbf{R}$, $c\in\mathbf{R}^d$, so that the map $f \colon \mathbf{R}^{n_d}\rightarrow \ensuremath{\mathfrak{g}}_L$ is given as \[
f(u) = (L,N(u)). \] One parameter subgroups are obtained through the exponential map as follows \[
\exp(t(L,b)) = (\exp(tL), \phi(tL) tb). \] Here the entire function $\phi(z)=(\exp(z)-1)/z$ familiar from the theory of exponential integrators appears. As an example, one could now consider the Lie--Euler method \eqref{eq:Lie--Euler} in this setting, which coincides with the exponential Euler method \[
u_{1} = \exp(h(L,N(u_0))\cdot u_0 = \exp(hL)u_0 + h\phi(hL)N(u_0). \] There is a large body of literature on exponential integrators, going almost half a century back in time, see~\cite{hochbruck10ei} and the references therein for an extensive account.
\paragraph{The coadjoint action and Lie--Poisson systems} Lie group integrators for this interesting case were studied by Eng{\o} and Faltinsen \cite{engo01nio}. Suppose $G$ is a Lie group and the manifold under consideration is the dual space~$\ensuremath{\mathfrak{g}}^*$ of its Lie algebra~$\ensuremath{\mathfrak{g}}$. The coadjoint action by $G$ on $\ensuremath{\mathfrak{g}}^*$ is denoted $\Ad_g^*$ defined for any $g\in G$ as \begin{equation} \label{eq:coadjoint-action} \langle \Ad_g^*\mu, \xi\rangle = \langle \mu, \Ad_g\xi\rangle,\quad\forall\xi\in\ensuremath{\mathfrak{g}}, \end{equation} for a duality pairing $\langle {\cdot}, {\cdot} \rangle$ between $\ensuremath{\mathfrak{g}}^*$ and $\ensuremath{\mathfrak{g}}$. It is well known (see e.g.\ section~13.4 in \cite{marsden99itm}) that mechanical systems formulated on the cotangent bundle $T^*G$ with a left or right invariant Hamiltonian can be reduced to a system on $\ensuremath{\mathfrak{g}}^*$ given as \[
\dot{\mu} = \pm\ad^*_{\frac{\partial H}{\partial\mu}} \mu, \] where the negative sign is used in case of right invariance. The solution to this system preserves coadjoint orbits, which makes it natural to suggest the group action \[
g\cdot\mu = \Ad_{g^{-1}}^*\mu, \] so that the resulting Lie group integrator also respects this invariant. For Euler's equations for the free rigid body, the Hamiltonian is left invariant and the coadjoint orbits are spheres in $\ensuremath{\mathfrak{g}}^*\cong\mathbf{R}^3$.
\paragraph{Homogeneous spaces and the Stiefel and Grassmann manifolds} The situation when $G$ acts on itself by left of right multiplication is a special case of a homogeneous space \cite{munthe-kaas97nio}, where the assumption is only that $G$ acts transitively and continuously on some manifold $M$. Homogeneous spaces are isomorphic to the quotient~$G/G_x$ where $G_x$ is the \emph{isotropy group} for the action at an arbitrarily chosen point $x\in M$ \[
G_x = \{ h\in G \mid h\cdot x = x\}. \] Note that if $x$ and $z$ are two points on $M$, then by transitivity of the action, $z=g\cdot x$ for some $g\in G$. Therefore, whenever $h\in G_z$ it follows that $g^{-1} \cdot h \cdot g\in G_x$ so isotropy groups are isomorphic by conjugation~\cite{bryant95ait}. Therefore the choice of $x\in M$ is not important for the construction of the quotient. For a readable introduction to this type of construction, see \cite{bryant95ait}, in particular Lecture 3.
A much encountered example is the hypersphere $M=S^{d-1}$ corresponding to the left action by $G=\SO(d)$,
the Lie group of orthogonal $d\times d$ matrices with unit determinant.
One has $S^{d-1} = \SO(d)/\SO(d-1)$. We have in fact already discussed the example of the free rigid body \eqref{eq:euler_frb} where $M=S^2$.
The Stiefel manifold $\St(d,k)$ can be represented by the set of $d\times k$-matrices with orthonormal columns. An action on this set is obtained by left multiplication by $G=\SO(d)$. Lie group integrators for Stiefel manifolds are extensively studied in the literature, see e.g.\ \cite{celledoni03oti,krogstad03alc} and some applications involving Stiefel manifolds are discussed in Section~\ref{sec:dataanalysis}. An important subclass of the homogeneous spaces is the symmetric spaces, also obtained through a transitive action by a Lie group $G$, where $M=G/G_x$, but here one requires in addition that the isotropy subgroup is an open subgroup of the fixed point set of an involution of $G$~\cite{munthe-kaas01aos}. A prominent example of a symmetric space in applications is the Grassmann manifold, obtained as $\SO(d)/(\SO(k)\times \SO(d-k))$.
\paragraph{Isospectral flows} In isospectral integration one considers dynamical systems evolving on the manifold of $d\times d$-matrices sharing the same Jordan form. Considering the case of symmetric matrices, one can use the transitive group action by $\SO(d)$ given as \[
g\cdot m = g m g^T. \] This action is transitive, since any symmetric matrix can be diagonalised by an appropriately chosen orthogonal matrix. If the eigenvalues are distinct, then the isotropy group is discrete and consists of all matrices in $\SO(d)$ which are diagonal.
Lie group integrators for isospectral flows have been extensively studied, see for example \cite{calvo96rkm1,calvo97nso}. See also~\cite{celledoni01ano} for an application to the KdV equation.
\paragraph{Tangent and cotangent bundles} For mechanical systems the natural phase space will often be the tangent bundle $TM$ as in the Lagrangian framework or the cotangent bundle $T^*M$ in the Hamiltonian framework. The seminal paper by Lewis and Simo \cite{lewis94caf} discusses several Lie group integrators for mechanical systems on cotangent bundles, deriving methods which are symplectic, energy and momentum preserving. Eng{\o}~\cite{engo03prk} suggested a way to generalise the Runge--Kutta--Munthe-Kaas methods into a partitioned version when $M$ is a Lie group. Marsden and collaborators have developed the theory of Lie group integrators from the variational viewpoint over the last two decades. See \cite{marsden01dma} for an overview. For more recent work pertaining to Lie groups in particular, see \cite{lee07lgv,bou-rabee09hpi,saccon09mrf}. In Section~\ref{sec:symplie} we present what we believe to be the first symplectic partitioned Lie group integrators on $T^*G$ phrased in the framework we have discussed here. Considering trivialised cotangent bundles over Lie groups is particularly attractive since there is a natural way to extend action by left multiplication from $G$ to $G\times\ensuremath{\mathfrak{g}}^*$ via \eqref{eq:prodGxgs}.
\subsection{Isotropy -- challenges and opportunities} An issue which we have already mentioned a few times is that the map $\lambda_* \colon \ensuremath{\mathfrak{g}} \rightarrow \mathcal{X}(M)$ defined in \eqref{eq:groupaction} is not necessarily injective. This means that the choice of $f \colon M\rightarrow\ensuremath{\mathfrak{g}}$ is not unique. In fact, if $g \colon M\rightarrow\ensuremath{\mathfrak{g}}$ is any map satisfying $\lambda_*(g(m))(m)=0$ for all $m\in M$, then we could replace the map $f$ by $f+g$ in \eqref{eq:genpres} without altering the vector field $F$. But such a modification of $f$ \emph{will have} an impact on the numerical schemes that we consider. This freedom in the setup of the methods makes it challenging to prove general results for Lie group methods, it might seem that some restrictions should apply to the isotropy choice for a more well defined class of schemes. However, the freedom can of course also be taken advantage of to obtain approximations of improved quality.
An illustrative example is the two-sphere $S^2$ acted upon linearly by the special orthogonal group $\SO(3)$. Representing elements of the Lie algebra $\mathfrak{so}(3)$ by vectors in $\mathbf{R}^3$, and points on the sphere as unit length vectors in $\mathbf{R}^3$, we may facilitate \eqref{eq:genpres} as \[
F(m)= f(m) \times m = (f(m) + \alpha(m)m)\times m, \] for any scalar function $\alpha \colon S^2 \rightarrow \mathbf{R}$. Using for instance the Lie--Euler method one would get \begin{equation} \label{eq:LieEulerS2}
m_1 = \exp(f(m_0)+\alpha(m_0)m_0) m_0, \end{equation} where the $\exp$ is the matrix exponential of the $3\times 3$ skew-symmetric matrix associated to a vector in $\mathbf{R}^3$ via the hat-map~\eqref{eq:hatmap}. Clearly the approximation depends on the choice of $\alpha(m)$. The approach of Lewis and Olver \cite{lewis08gia} was to use the isotropy to improve certain qualitative features of the solution. In particular, they studied how the orbital error could be reduced by choosing the isotropy in a clever way. In Figure~\ref{fig:isotropy} we illustrate the issue of isotropy for the Euler free rigid body equations. The curve drawn from the initial point~$z_0$ to $z_1$ is the exact solution, i.e.\ the momenta in body coordinates. The broken line shows the terminal points using the Lie--Euler method for $\alpha$ varying between $0$ and $25$. \begin{figure}
\caption{The effect of isotropy on $S^2$ for Euler's free rigid body equations.
The curve drawn from the initial point $z_0$ to $z_1$ is the exact solution, i.e.\ the momenta in body coordinates. The broken line shows the terminal points using the Lie--Euler method for $\alpha(z_0)$ (as in \eqref{eq:LieEulerS2}) varying between $0$ and $25$. }
\label{fig:isotropy}
\end{figure}
Another potential difficulty with isotropy is the increased computational complexity when the group~$G$ has much higher dimension than the manifold $M$. This could for instance be the case with the Stiefel manifold~$\St(d,k)$ if $d\gg k$. Linear algebra operations used in integrating differential equations on the Stiefel manifold should preferably be of complexity $\mathcal{O}(dk^2)$. But solving a corresponding problem in the Lie algebra $\mathfrak{so}(d)$ would typically require linear algebra operations of complexity $\mathcal{O}(d^3)$, see for example \cite{celledoni03oti} and references therein. By taking advantage of the many degrees of freedom provided by the isotropy, it is actually possible to reduce the cost down to the required $\mathcal{O}(dk^2)$ operations as explained in for instance \cite{celledoni03lgm} and \cite{krogstad03alc}.
\section{Applications to nonlinear problems of evolution in classical mechanics} \label{sec:appcm}
The emphasis on the use of Lie groups in modelling and simulation of engineering problems in classical mechanics started in the eighties with the pioneering and fundamental work of J.C.~Simo and his collaborators. In the case of rod dynamics, for example, models based on partial differential equations were considered where the configuration of the centreline of the rod is parametrised via arc-length, and the movement of a rigid frame attached to each of the cross sections of the rod is considered (see Figure~\ref{fig1}). This was first presented in a geometric context in \cite{simo85afs}.
\begin{figure}
\caption{Geometric rod model. Here $\phi$ is the line of centroids and a cross section is identified by the frame $\Lambda=[\mathbf{t}_1,\mathbf{t}_2,\mathbf{t}_3]$, $\phi_r$ is the initial position of the line of centroids. }
\label{fig1}
\end{figure}
In robot technology, especially robot locomotion and robot grasping, the occurrence of non-holonomically constrained models is very common. The motion of robots equipped with wheels is not always locally controllable, but is often globally controllable. A classical example is the parking of a car that cannot be moved in the direction orthogonal to its wheels. The introduction of Lie groups and Lie brackets to describe the dynamics of such systems, has been considered by various authors, see for example \cite{murray93nmp}. The design of numerical integration methods in this context has been addressed in the paper of Crouch and Grossman,~\cite{crouch93nio}. These methods have had a fundamental impact to the successive developments in the field of Lie group methods.
The need for improved understanding of non-holonomic numerical integration has been for example advocated in \cite{mclachlan96aso}. Recent work in this field has led to the construction of low order non-holonomic integrators based on a discrete Lagrange--d'Alembert principle, \cite{cortes01nhi,mclachlan06ifn}. The use of Lie group integrators in this context has been considered in \cite{leok05alg,mclachlan06ifn}.
We have already mentioned the relevance of rigid body dynamics to the numerical discretisation of rod models. There are many other research areas in which the accurate and efficient simulation of rigid body dynamics is crucial: molecular dynamics, satellite dynamics, and celestial mechanics just to name a few, \cite{leimkuhler04shd}. In some of these applications, it is desirable to produce numerical approximations which are accurate possibly to the size of roundoff. The simulations of interest occur over very long times and/or a large number of bodies and this inevitably causes propagation of errors even when the integrator is designed to be very accurate. For this reason accurate symplectic rigid body integrators are of interest because they can guarantee that the roundoff error produced by the accurate computations can stay bounded also in long time integration. This fact seems to be of crucial importance in celestial mechanics simulations, \cite{laskar04aln}. A symplectic and energy preserving Lie group integrator for the free rigid body motion was proposed in \cite{lewis94caf}. The method computes a time re-parametrisation of the exact solution. Some recent and promising work in this field has been presented in \cite{mclachlan05tdm,celledoni06ets,celledoni07eco,hairer06pdm}. The control of rigid bodies with variational Lie group integrators was considered in \cite{leok05alg}.
In the next section we illustrate the use of Lie group methods in applications on a particular case study, the pipe-laying process from ships to the bottom of the sea.
\subsection{Rigid body and rod dynamics}
\paragraph{Pipe-laying problem}
The simulation of deep-water risers, pipelines and drill rigs requires the use of models of long and thin beams subject to large external forces. These are complicated nonlinear systems with highly oscillatory components.
We are particularly interested in the correct and accurate simulation of the pipe-laying process
from ships on the bottom of the sea, see Figure~\ref{fig2}. The problem comprises the modelling of two interacting structures: a long and thin pipe (modelled as a rod) and a vessel (modelled as a rigid body). The system is subject to environmental forces (such as sea and wind effects). The control parameters for this problem are the vessel position and velocity, the pay-out speed and the pipe tension while the control objectives consist in determining the touchdown position of the pipe as well as ensuring the integrity of the pipe and to avoid critical deformations, \cite{jensen10anp,safstrom09mas}.
\begin{figure}
\caption{The pipe-laying process. }
\label{fig2}
\end{figure}
The vessel rigid body equations determine the boundary conditions of the rod. They are expressed in six degrees of freedom as \begin{equation*}\label{eq:vesseleq} M\dot{\boldsymbol\nu} + C(\boldsymbol\nu)\boldsymbol\nu
+ D(\boldsymbol\nu)\boldsymbol\nu + g(\boldsymbol\eta) = \boldsymbol\tau, \end{equation*} where $M$ is the system inertia matrix, $C(\boldsymbol\nu)$ the Coriolis-centripetal matrix, $D(\boldsymbol\nu)$ the damping matrix, $g(\boldsymbol\eta)$ the vector of gravitational and buoyancy forces and moments, and $\boldsymbol\tau$ the vector of control inputs and environmental disturbances such as wind, waves and currents (see \cite{perez07kmf} for details). The vector $\boldsymbol\nu$ contains linear and angular velocity and $\boldsymbol\eta$ is the position vector. It has been shown in \cite{jensen10anp} that the rigid body vessel equations are input-output passive.
The equations can be integrated numerically with a splitting and composition technique where the vessel equations are split into a free rigid body part and a damping and control part. The free rigid body equations can be solved with a method proposed in \cite{celledoni06ets} where the angular momentum is accurately and efficiently computed by using Jacobi elliptic functions, the attitude rotation is obtained using a Runge--Kutta--Munthe-Kaas Lie group method, and the control and damping part is solved exactly.
Simulations of the whole pipe-lay problem with local parametrisations of the pipe and the vessel based on Euler angles have been obtained in \cite{jensen10anp}.
\paragraph{Rod dynamics} At fixed time each cross section of the pipe is the result of a rigid rotation in space of a reference cross section, and analogously, for each fixed value of the space variable the corresponding cross section evolves in time as a forced rigid body, see Figure~\ref{fig1}. In absence of external forces the equations are \begin{align*} \rho_A \partial_{tt} \phi &= \partial_S\mathbf{n},\qquad S\in[0,L],\quad t\ge0,\\ \partial_t \pi+ (I_{\rho}^{-1} \pi)\times \pi &= \partial_S\mathbf{m}+(\partial_S \phi)\times \mathbf{n}, \end{align*} here $\phi=\phi(S,t)$ is the line of centroids of the rod, $\mathbf{m}$ and $\mathbf{n}$ are the stress couple and stress resultant, $\mathbf{\pi}$ is the angular momentum density, $I_{\rho}$ is the inertia in spatial coordinates, and $\rho_A=\rho_A(S)$ is the mass per unit length of the rod (see \cite{simo87otd} and \cite{celledoni10aha}). The kinematic equations for the attitude rotation matrix are \[
\partial_t\Lambda=\hat{\mathbf{w}} \Lambda, \qquad \partial_S\Lambda=\hat{\mathbf{M}} \Lambda, \]
where $\Lambda(S,t)=[\mathbf{t}_1,\mathbf{t}_2,\mathbf{t}_3]$, $I_{\rho}^{-1}\pi=\mathbf{w}$, $\mathbf{M}=C_2 \Lambda^T\mathbf{m}$ and $C_2$ is a constant diagonal matrix.
We denote by ``$\hat{\quad}$" the hat-map identifying $\mathbf{R}^3$ with $\mathfrak{so}(3)$: \begin{equation} \label{eq:hatmap}
\mathbf{v}=\begin{bmatrix}
v_1\\ v_2\\ v_3
\end{bmatrix} \mapsto
\hat{\mathbf{v}}= \begin{bmatrix*}[r]
0 &-v_3&v_2\\v_3&0&-v_1\\ -v_2&v_1&0
\end{bmatrix*}. \end{equation} With no external forces one assumes pure displacement boundary conditions providing $\phi$ and $\Lambda$ on the boundaries $S=0$ and $S=L$.
In \cite{simo87otd}, partitioned Newmark integrators, of Lie group type, and of moderate order were considered for this problem. While classical Newmark methods are variational integrators and as such are symplectic when applied to Hamiltonian systems \cite{marsden01dma}, the particular approach of \cite{simo87otd} involves the use of exponentials for the parametrisation of the Lie group $\SO(3)$, and the geometric properties of this approach are not trivial to analyse.
Moreover, since the model is a partial differential equation, space and time discretisations should be designed so that the overall discrete equations admit solutions and are stable. It turns out that conventional methods perform poorly on such problem in long time simulations. To obtain stable methods reliable in long-time simulation, an energy-momentum method was proposed for the rod problem in \cite{simo95ndo}.
Later, this line of thought has been further developed in \cite{romero01aof}. The Hamiltonian formulation of this model allows one to derive natural structure preserving discretisations into systems of coupled rigid bodies \cite{leyendecker06oem}.
Following the geometric space-time integration procedure proposed in \cite{frank04gst}, a multi-Hamiltonian formulation\footnote{For a definition of the multi-symplectic structure of Hamiltonian partial differential equations, see \cite{bridges01msi}.} of these equations has been proposed in \cite{celledoni10aha}, using the Lie group of Euler parameters. The design of corresponding multi-symplectic Lie group discretisations is still under investigation.
\section{Applications to problems of data analysis and statistical signal processing} \label{sec:dataanalysis}
The solution of the many-body Schr{\"o}dinger eigenvalue problem \begin{equation} \label{eq:schr} \hat{\mathbf{H}}\Psi =E\Psi, \end{equation} where the so called electronic ground state (the smallest eigenstate) is sought, is an important problem of computational chemistry. The main difficulty is the curse of dimensionality. Since $\hat{\mathbf{H}}$ is a differential operator in several space dimensions, a realistic simulation of (\ref{eq:schr}) would require the numerical discretisation and solution of a partial differential equation in several space dimensions. The number of space dimensions grows with the number of electrons included in the simulation.
The eigenvalue problem admits an alternative variational formulation. Instead of looking for the smallest eigenvalue and eigenfunction of the Schr{\"o}dinger equation, one minimises directly the ground state energy.
After appropriate spatial discretisation, the problem becomes a minimisation problem on a Riemannian manifold $\mathcal{M}$, \begin{equation}\label{genopt} {\min_{x\in\mathcal{M}}}\, \phi(x), \end{equation} where $\phi \colon \mathcal{M}\rightarrow \mathbf{R}$ is a smooth discrete energy function to be minimised on $\mathcal{M}$. The discrete energy $\phi$ considered here is the so called Kohn--Sham energy, \cite{baarman12dma}. For a related application of Lie group techniques in quantum control, see \cite{degani09qcw}.
The general optimisation problem giving rise to (\ref{genopt}) appears in several applied fields, ranging from engineering to applied physics and medicine. Some specific examples are principal component/subspace analysis, eigenvalue and generalised eigenvalue problems, optimal linear compression, noise reduction, signal representation and blind source separation.
\subsection{Gradient-based optimisation on Riemannian manifolds}
Let $\mathcal{M}$ be a Riemannian manifold with metric $\langle\cdot, \cdot \rangle_{\mathcal{M}}$ and $\phi \colon \mathcal{M}\rightarrow \mathbf{R}$ be a smooth cost function to be minimised on $\mathcal{M}$. We want to solve (\ref{genopt}).
The optimisation method based on gradient flow -- written for the minimisation problem only, for the sake of easy reading -- consists in setting up the differential equation on the manifold,
\begin{equation}\label{eqdiff} \dot{x}(t)=-\grad \phi\big(x(t)\big), \end{equation}
with appropriate initial condition $x(0)=x_0\in\mathcal{M}$. The equilibria of equation (\ref{eqdiff}) are the critical points of the function $\phi$.
In the above equation, the symbol $\grad \phi$ denotes the Riemannian gradient of the function $\phi$ with respect to the chosen metric. Namely, $\grad \phi(x)\in T_x\mathcal{M}$ and $T_x \phi(v)=\langle \grad \phi (x),v\rangle_{\mathcal{M}}$ for all $v\in T_x\mathcal{M}$.
The solution of (\ref{eqdiff}) on $\mathcal{M}$ may be locally expressed in terms of a curve on the tangent space $T_{x_0}\mathcal{M}$ using a retraction map $\mathcal{R}$. Retractions are tangent space parametrisations of $\mathcal{M}$, and allow us to write
\[x(t)=\mathcal{R}_{x_0}(\sigma(t)), \quad \sigma(t)\in T_{x_0}\mathcal{M}, \quad t\in [0,t_f],\] for small enough $t_f$, see \cite{shub86sro} for a precise definition.
In most applications of interest, see for example \cite{brockett91dst,helmke94oad}, $\mathcal{M}$ is a matrix manifold endowed with a Lie group action and there is a natural way to define a metric and a retraction.
In fact, let $\mathcal{M}$ be a manifold acted upon by a Lie group $G$, with a locally transitive group action $\Lambda(g,x)=\Lambda_x(g)$. Let us also consider a coordinate map $\psi$,
\[ \psi \colon \ensuremath{\mathfrak{g}} \rightarrow G, \quad \text{and} \quad \rho_x \coloneqq T_0(\Lambda_x\circ \psi). \]
One can prove that if there exists a linear map $a_x \colon T_x\mathcal{M}\rightarrow \ensuremath{\mathfrak{g}}$ such that $\rho_x\circ a_x =\mathrm{Id}_{T_x\mathcal{M}}$, then $\mathcal{R}_x$, given by
\[Â \mathcal{R}_x(v) \coloneqq (\Lambda_x \circ \psi \circ a_x )(v), \]
is a retraction, see \cite{celledoni03oti}. The existence of $a_x$ is guaranteed, at least locally, by the transitivity of the action and the fact that $\psi$ is a local diffeomorphism. The approach is analogous to the one described for differential equations in section~\ref{GeneralizingRK}. Therefore, we can construct retractions using any coordinate map from the Lie algebra $\ensuremath{\mathfrak{g}}$ to the group.
Any metric on $\mathfrak{g}$, $\langle\cdot ,\cdot\rangle_{\mathfrak{g}}$ induces a metric on $\mathcal{M}$ by \[\langle v_x,w_x\rangle_{\mathcal{M}}=\langle a_x(v_x) ,a_x(w_x)\rangle_{\mathfrak{g}}.\]
Also, we may define the image of the tangent space under the map $a_x$:
\[ \mathfrak{m}_x \coloneqq a_x(T_x\mathcal{M})\subset \ensuremath{\mathfrak{g}}. \]
The set $\ensuremath{\mathfrak{m}}_x$ is a linear subspace of the Lie algebra $\ensuremath{\mathfrak{g}}$, often of lower dimension. Parametrisations of the solution of (\ref{eqdiff}) involving the whole Lie algebra are in general more computationally intensive than those restricted to $\ensuremath{\mathfrak{m}}_x$, but, if the isotropy is chosen suitably, they might lead to methods which converge faster to the optimum.
For the sake of illustration, we consider the minimisation on a two-dimensional torus $T^2=S^1\times S^1$. Here we denote by $S^1$ the circle, i.e.\ \begin{gather*} S^1=\{g(\alpha) \mathbf{e}_1 \in \mathbf{R}^2 \mid g(\alpha)\in \SO(2)\}, \\ g(\alpha)=\exp(\alpha E), \quad E=\begin{bmatrix*}[r]
0 & -1\\
1 & 0\\ \end{bmatrix*}, \quad 0 \le \alpha < 2\pi, \end{gather*} where $\mathbf{e}_1$ is the first canonical vector and $\SO(2)$ is the commutative Lie group of planar rotations. Any element in $T^2$ is of the form \[x_0\in T^2,\quad x_0=(g(\theta)\mathbf{e}_1,g(\varphi)\mathbf{e}_1), \quad g(\theta), g(\varphi)\in \SO(2).\] The Lie group acting on $T^2$ is $\SO(2)\times \SO(2)$, its corresponding Lie algebra is $\mathfrak{so}(2)\times \mathfrak{so}(2)$, which has dimension $d=2$ and basis $\{ (E,O), (O,E)\}$, where $O$ is the zero element in $\mathfrak{so}(2)$.
The Lie group action is \[\Lambda_{x_0}(h_1,h_2)=(h_1 g(\theta)\mathbf{e}_1,h_2 g(\varphi)\mathbf{e}_1),\qquad (h_1,h_2)\in \SO(2)\times \SO(2),\] and $\psi=\exp$. Any $v_{x_0}\in T_{x_0}T^2$ can be written as \[v_{x_0}=(\alpha E\mathbf{e}_1,\beta E\mathbf{e}_1),\] for some $\alpha, \beta \in\mathbf{R},$ so \[a_{x_0}(v_{x_0})=(\alpha E,\beta E).\]
Assume the cost function we want to minimise is simply the distance from a fixed plane in $\mathbf{R}^3$, say the plane with equation $y=8$. This gives \begin{equation*} \label{cost1} \phi(g(\theta)\mathbf{e}_1,g(\varphi)\mathbf{e}_1)=\lvert(1+\cos (\theta)) \sin (\varphi)-8\rvert, \end{equation*} and the minimum is attained in $\theta =0$ and $\varphi=\pi/2$.\footnote{ We have used a parameterisation of $T^2$ in $\mathbf{R}^3$ in angular coordinates, obtained applying the following mapping
\[ (g(\theta)\mathbf{e}_1,g(\varphi)\mathbf{e}_1) \mapsto \left\{ \begin{alignedat}{2}
x &= (1+\mathbf{e}_1^Tg(\theta)\mathbf{e}_1) \mathbf{e}_1^Tg(\varphi)\mathbf{e}_1 & &= (1+\cos (\theta)) \cos (\varphi),\\
y &= (1+\mathbf{e}_1^Tg(\theta)\mathbf{e}_1) \mathbf{e}_2^Tg(\varphi)\mathbf{e}_1 & &= (1+\cos (\theta)) \sin (\varphi),\\
z &= \mathbf{e}_2^Tg(\theta)\mathbf{e}_1 & &= \sin (\theta), \end{alignedat} \right. \] with $0 \le \theta , \varphi < 2\pi$. This is equivalent to the composition of two planar rotations and one translation in $\mathbf{R}^3.$} In Figure~\ref{fig0} we plot $-\grad \phi$, the negative gradient vector field for the given cost function. The Riemannian metric we used is \[ \langle(\alpha_1E\mathbf{e}_1,\beta_1E\mathbf{e}_1),(\alpha_2E\mathbf{e}_1,\beta_2E\mathbf{e}_1)\rangle_{T^2}=\alpha_1\alpha_2+\beta_1\beta_2, \] and $(\alpha_1E\mathbf{e}_1,\beta_1E\mathbf{e}_1)\in T_{(\mathbf{e}_1,\mathbf{e}_1)}T^2$. This metric can be easily interpreted as a metric on the Lie algebra~$\mathfrak{g}=\mathfrak{so}(2)\times \mathfrak{so}(2)$: \[ \langle(\alpha_1E,\beta_1E),(\alpha_2E,\beta_2E)\rangle_{\mathfrak{g}}=\alpha_1\alpha_2+\beta_1\beta_2.\] At the point $p_0=( g(\theta_0)\mathbf{e}_1, g(\varphi_0)\mathbf{e}_1)\in T^2$, the gradient vector field can be represented by \[ (\gamma E g(\theta_0)\mathbf{e}_1,\delta E g(\varphi_0)\mathbf{e}_1), \] where $\gamma$ and $\delta$ are real values given by \[ \gamma=-C\sin(\theta_0)\sin(\varphi_0),\qquad \delta=C(1+\cos(\theta_0))\cos(\varphi_0),\] and \[C=2((1+\cos(\theta_0)\sin(\varphi_0)-8). \]
Gradient flows are not the only type of differential equations which can be used to solve optimisation problems on manifolds. Alternative equations have been proposed in the context of neural networks \cite{celledoni04nlb, celledoni08dmf}. Often they arise naturally as the Euler--Lagrange equations of a variational problem.
\begin{figure}
\caption{The gradient vector field of the cost function $\phi (g(\theta)\mathbf{e}_1,g(\varphi)\mathbf{e}_1)= ((1+\mathbf{e}_1^Tg(\theta)\mathbf{e}_1) \mathbf{e}_2^Tg(\varphi)\mathbf{e}_1-8)^2$ on the torus. The vector field points towards the two minima, the global minimum is marked with a black spot in the middle of the picture. }
\label{fig0}
\end{figure}
\subsection{Principal component analysis}
Data reduction techniques are statistical signal processing methods that aim at providing efficient representations of data. A well-known data compression technique consists of mapping a high-dimensional data space into a lower dimensional representation space by means of a linear transformation. It requires the computation of the data covariance matrix and then the application of a numerical procedure to extract its eigenvalues and the corresponding eigenvectors. Compression is then obtained by representing the signal in a basis consisting only of those eigenvectors associated with the most significant eigenvalues.
In particular, principal component analysis (PCA) is a second-order adaptive statistical data processing technique that helps removing the second-order correlation among given random signals. Let us consider a stationary multivariate random process $x(t)\in\mathbf{R}^{n}$ and suppose its covariance matrix $A=E[(x-E[x])(x-E[x])]^{T}]$ exists and is bounded. Here the symbol $E[\cdot]$ denotes statistical expectation. If $A\in\mathbf{R}^{n\times n}$ is not diagonal, then the components of $x(t)$ are statistically correlated. One can remove this redundancy by partially diagonalising $A$, i.e.\ computing the operator $F$ formed by the eigenvectors of the matrix $A$ corresponding to its largest eigenvalues. This is possible since the covariance matrix $A$ is symmetric (semi) positive-definite, and $F\in\St(n,p)$.
To compute $F$ and the corresponding $p$ eigenvalues of the $n\times n$ symmetric and positive-definite matrix~$A$, we consider the maximisation of the function \[
\phi(X)=\frac{1}{2}\operatorname{trace}(X^TAX), \] on the Stiefel manifold, and solve numerically the corresponding gradient flow with a Lie group integrator.
As a consequence the new random signal defined by $y(t) \coloneqq F^{T}(x(t)-E[x(t)])\in\mathbf{R}^{p}$ has uncorrelated components, with $p\leq n$ properly selected. The component signals of $y(t)$ are the so called \emph{principal components of the signal} $x(t)$, and their relevance is proportional to the corresponding eigenvalues $\sigma_{i}^2=E[y_{i}^2]$ which here are arranged in descending order ($\sigma_{i}^{2}\geq\sigma_{i+1}^{2}$).
Thus, the data stream $y(t)$ is a compressed version of the data stream $x(t)$. After the reduced-size data has been processed (i.e.\ stored, transmitted), it needs to be recovered, that is, it needs to be brought back to the original structure. However, the principal-component-based data reduction technique is not lossless, thus only an approximation $\hat{x}(t)\in\mathbf{R}^{n}$ of the original data stream may be recovered. An approximation of $x(t)$ is given by $\hat{x}(t)=Fy(t)+E[x]$. Such approximate data stream minimises the reconstruction error~$E[\lVert x-\hat{x}\rVert_2^{2}]=\sum_{i=n+1}^{p}\sigma_{i}^2$.
For a scalar or a vector-valued random variable $x\in\mathbf{R}^n$ endowed with a probability density function~$p_x \colon x\in\mathbf{R}^n\rightarrow p_x(x)\in\mathbf{R}$, the expectation of a function $\beta \colon \mathbf{R}^n\rightarrow \mathbf{R}$ is defined as
\[ E[\beta] \coloneqq \int_{\mathbf{R}^n}\beta(x)p_x(x) \, \ensuremath{\mathrm{d}}^nx. \]
Under the hypothesis that the signals whose expectation is to be computed are ergodic, the actual expectation (ensemble average) may be replaced by temporal-average on the basis of the available signal samples, namely
\[ E[\beta]\approx \frac{1}{T}\sum_{t=1}^T\beta(x(t)). \]
\subsection{Independent component analysis}
An interesting example of a problem that can be tackled via statistical signal processing is the \emph{cocktail-party problem}. Let us suppose $n$ signals $x_1(t),\dots ,x_n(t)$ were recorded from $n$ different positions in a room where there are $p$ sources or speakers. Each recorded signal is a linear mixture of the voices of the sources~$s_1(t),\dots , s_p(t)$, namely
\begin{align*} x_1(t) &= a_{1,1} s_1(t) +\dotsb +a_{1,p} s_p(t),\\ &\vdotswithin{=} \\ x_n(t) &= a_{n,1} s_1(t) +\dotsb +a_{n,p} s_p(t), \end{align*}
where the $n p$ coefficients $a_{i,j}\in\mathbf{R}$ denote the mixing proportions. The mixing matrix $A=(a_{i,j})$ is unknown. The cocktail party problem consists in estimating signals $s_1(t),\dots , s_p(t)$ from only the knowledge of their mixtures $x_1(t), \dots ,x_n(t)$. The main assumption on the source signals is that $s_1(t), \dots , s_p(t)$ are statistically independent. This problem can be solved using independent component analysis (ICA).
Typically, one has $n>p$, namely, the number of observations exceeds the number of actual sources. Also, a typical assumption is that the source signals are spatially white, which means $E[ss^T]=I_p$, the $p \times p$~identity matrix. The aim of independent component analysis is to find estimates $y(t)$ of signals in $s(t)$ by constructing a de-mixing matrix $W\in\mathbf{R}^{n\times p}$ and by computing $y(t) \coloneqq W^Tx(t)$. Using statistical signal processing methods, the problem is reformulated into an optimisation problem on homogeneous manifolds for finding the de-mixing matrix $W$.
The geometrical structure of the parameter space in ICA comes from a signal pre-processing step named \emph{signal whitening}, which is operated on the observable signal $x(t)\rightarrow \tilde{x}(t)\in\mathbf{R}^p$ in such a way that the components of the signal $\tilde{x}(t)$ are uncorrelated and have variances equal to $1$, namely $E[\tilde{x}\tilde{x}^T]=I_p$.
This also means that redundant observations are eliminated and the ICA problem is brought back to the smallest dimension $p$. This can be done by computing $E[xx^T]=VDV^T$, with $V\in\St(n,p)$ and $D\in\mathbf{R}^{p\times p}$ diagonal invertible. Then
\[ \tilde{x}(t) \coloneqq D^{-\frac{1}{2}}V^Tx(t), \]
and with $\tilde{A} \coloneqq D^{-\frac{1}{2}}V^TA$ we have $E[\tilde{x}\tilde{x}^T]=\tilde{A}E[ss^T]\tilde{A}^T=\tilde{A}\tilde{A}^T=I_p$.
After observable signal whitening, the de-mixing matrix may be searched for such that it solves the optimisation problem
\begin{equation*}\label{pri1} {\max_{W\in \LieO(p)}}\,\phi(W). \end{equation*}
As explained, after whitening, the number of projected observations in the signal $\tilde{x}(t)$ equals the number of sources. However, in some applications it is known that not all the source signals are useful, so it is sensible to analyse only a few of them. In these cases, if we denote by $\overline{p}\ll p$ the actual number of independent components that are sought after, the appropriate way to cast the optimisation problem for ICA is
\begin{equation*}\label{pri2} \max_{W\in \St(n,\overline{p})} \phi(W), \quad \text{with} \quad \overline{p}\ll p. \end{equation*} The corresponding gradient flows obtained in this case are differential equations on the orthogonal group or on the Stiefel manifold, and can be solved numerically by Lie group integrators.
As a possible principle for reconstruction, the maximisation or minimisation of non-Gaussianity is viable. It is based on the notion that the sum of independent random variables has distribution closer to Gaussian than the distributions of the original random variables. A measure of non-Gaussianity is the kurtosis, defined for a scalar signal $z\in\mathbf{R}$ as
\[ \operatorname{kurt}(z) \coloneqq E[z^4]-3E^2[z^2]. \]
If the random signal $z$ has unitary variance, then the kurtosis computes as $\operatorname{kurt}(z)=E[z^4]-3$. Maximising or minimising kurtosis is thus a possible way of estimating independent components from their linear mixtures, see \cite{celledoni08dmf} and references therein for more details.
\subsection{Computation of Lyapunov exponents}
The Lyapunov exponents of a continuous dynamical system $\dot x=F(x)$, $x(t)\in \mathbf{R}^n$, provide a qualitative measure of its complexity. They are numbers related to the linearisation $A(t)$ of $\dot x=F(x)$ along a trajectory~$x(t)$.
Consider the solution $U$ of the matrix problem \[ \dot{U}=A(t)U, \qquad U(0)=U_0, \qquad U(t) \in \mathbf{R}^{n\times n}. \] The logarithms of the eigenvalues of the matrix \[ \Lambda=\lim_{t\rightarrow \infty}\left(U(t)^{T}U(t)\right)^{\frac{1}{2t}}, \] are the Lyapunov exponents for the given dynamical system. In \cite{dieci95coa} a procedure for computing just $k$ of the $n$ Lyapunov exponents of a dynamical system is presented. The exponents are computed by solving an
initial value problem on $\St (n,k)$ and computing a quadrature of the diagonal entries of a $k\times k$ matrix-valued function. The initial value problem is defined as follows: \[ \dot{Q}=(A-QQ^{T}A+QSQ^{T})Q, \] with random initial value in $\St(n,k)$ and \[ S_{k,j}= \begin{cases}
(Q^TAQ)_{k,j}, & k>j,\\
0, & k=j,\\
-(Q^TAQ)_{j,k}, & k<j,
\end{cases}
\qquad k,j=1,\dots ,p. \]
It can be shown that the $i$-th Lyapunov exponent $\lambda_i$ can be obtained as \begin{equation} \label{int} \lambda_i=\lim_{t\rightarrow
\infty}\frac{1}{t}\int_0^tB_{i,i}(s) \, \ensuremath{\mathrm{d}} s, \quad i=1,\dots ,k, \end{equation} and \[ B=Q^TAQ-S. \]
One could use for example the trapezoidal rule to approximate the integral (\ref{int}) and compute $\lambda_i$ ($i=1,\dots ,k$). We refer to \cite{dieci95coa} for further details on the method, and to \cite{celledoni02aco} for the use of Lie group integrators on this problem. Lie group methods for ODEs on Stiefel manifolds have also been considered in \cite{celledoni03oti,krogstad03alc,celledoni03cfl}.
We have here presented a selection of applications that can be dealt with by solving differential equations on Lie groups and homogeneous manifolds. For these problems, Lie group integrators are a natural choice. We gave particular emphasis to problems of evolution in classical mechanics and problems of signal processing. This is by no means an exhaustive survey; other interesting areas of application are for example problems in vision and medical imaging, see for instance \cite{xu12aol,lee07gds}.
\section{Symplectic integrators on the cotangent bundle of a Lie group} \label{sec:symplie} In this section we shall assume that the manifold is the cotangent bundle $T^*G$ of a Lie group $G$. Let $R_g\colon G\rightarrow G$ be the right multiplication operator such that $R_g(h)=h \cdot g$ for any $h\in G$. The tangent map of $R_g$ is denoted $R_{g*} \coloneqq TR_g$. Any cotangent vector $p_g\in T_g^*G$ can be associated to $\mu\in\ensuremath{\mathfrak{g}}^*$ by right trivialisation as follows: Write $v_g\in T_gG$ in the form $v_g = R_{g*}\xi$ where $\xi\in\ensuremath{\mathfrak{g}}$, so that $\langle p_g, v_g\rangle=\langle p_g, R_{g*}\xi\rangle = \langle R_g^* p_g, \xi\rangle$, where we have used $R_g^*$ for the dual map of $R_{g*}$, and $\langle{\cdot},{\cdot}\rangle$ is a duality pairing. We therefore represent $p_g\in T_g^*G$ by $\mu=R_g^* p_g\in\ensuremath{\mathfrak{g}}^*$. Thus, we may use as phase space $G\times\ensuremath{\mathfrak{g}}^*$ rather than $T^*G$. For applying Lie group integrators we need a transitive group action on $G\times\ensuremath{\mathfrak{g}}^*$ and this can be achieved by lifting the group structure of $G$ and using left multiplication in the extended group. The semidirect product structure on $\mathbf{G} \coloneqq G\ltimes\ensuremath{\mathfrak{g}}^*$ is defined as \begin{equation} \label{eq:prodGxgs}
(g_1, \mu_1)\cdot (g_2, \mu_2) = (g_1\cdot g_2, \mu_1 + \Ad_{g_1^{-1}}^*\mu_2), \end{equation} where the coadjoint action $\Ad^{*}$ is defined in \eqref{eq:coadjoint-action}. Similarly, the tangent map of right multiplication extends as \[ TR_{(g,\mu)}(R_{h*}\,\zeta, \nu) = (R_{hg*}\;\zeta, \nu-\ad_\zeta^*\,\Ad_{h^{-1}}^*\mu),\quad g,h \in G,\ \zeta\in\ensuremath{\mathfrak{g}},\ \mu,\nu\in\ensuremath{\mathfrak{g}}^*. \] Of particular interest is the restriction of $TR_{(g,\mu)}$ to $T_e\mathbf{G}\cong \ensuremath{\mathfrak{g}} \times \ensuremath{\mathfrak{g}}^*$, \[
T_eR_{(g,\mu)}(\zeta,\nu) = (R_{g*}\zeta, \nu - \ad_\zeta^*\mu). \] The natural symplectic form on $T^*G$ (which is a differential two-form) is defined as \[ \Omega_{(g,p_g)}((\delta v_1, \delta \pi_1),(\delta v_2,\delta\pi_2)) =\langle \delta \pi_2, \delta v_1\rangle - \langle \delta \pi_1, \delta v_2\rangle, \] and by right trivialisation it may be pulled back to $\mathbf{G}$ and then takes the form \begin{equation} \label{eq:sympform}
\omega_{(g,\mu)}( (R_{g*} \xi_1, \delta\nu_1), (R_{g*}\xi_2, \delta\nu_2))
= \langle\delta\nu_2,\xi_1 \rangle - \langle\delta\nu_1, \xi_2 \rangle - \langle\mu, [\xi_1,\xi_2]\rangle, \qquad \xi_1,\xi_2 \in \ensuremath{\mathfrak{g}}. \end{equation} The presentation of differential equations on $T^*G$ as in \eqref{eq:genpres} is now achieved via the action by left multiplication, meaning that any vector field $F\in\mathcal{X}(\mathbf{G})$ is expressed by means of a map $f\colon \mathbf{G} \rightarrow T_e \mathbf{G}$, \begin{equation} \label{eq:Fpres}
F(g,\mu) = T_e R_{(g,\mu)} f(g,\mu) = (R_{g*} f_1, f_2-\ad_{f_1}^*\mu), \end{equation} where $f_1=f_1(g,\mu)\in\ensuremath{\mathfrak{g}}$, $f_2=f_2(g,\mu)\in\ensuremath{\mathfrak{g}}^*$ are the two components of $f$. We are particularly interested in the case that $F$ is a Hamiltonian vector field which means that $F$ satisfies the relation \begin{equation} \label{eq:iF}
\mathrm{i}_{F}\omega = \ensuremath{\mathrm{d}} H, \end{equation} for some Hamiltonian function $H\colon T^*G\rightarrow\mathbf{R}$ and $\mathrm{i}_F$ is the interior product defined as $\mathrm{i}_F \omega (X) \coloneqq \omega(F, X)$ for any vector field $X$. From now on we let $H \colon \mathbf{G} \to \mathbf{R}$ denote the trivialised Hamiltonian. A simple calculation using \eqref{eq:sympform}, \eqref{eq:Fpres} and \eqref{eq:iF} shows that the corresponding map $f$ for such a Hamiltonian vector field is \[
f(g,\mu) = \left(\frac{\partial H}{\partial\mu}(g,\mu), -R_g^*\frac{\partial H}{\partial g}(g,\mu)\right). \] We have come up with the following family of symplectic Lie group integrators on $G\times\ensuremath{\mathfrak{g}}^*$ \begin{align*} (\xi_i, \bar{n}_i) &= h f(G_i, M_i),\qquad n_i = \Ad_{\exp(X_i)}^* \bar{n}_i,\quad i=1,\ldots s, \\ (g_1, \mu_1) &= \exp\Bigl(Y, (\dexp_{Y}^{-1})^{*}\sum_{i=1}^s b_i n_i\Bigr)\cdot (g_0, \mu_0), \\ Y&= \sum_{i=1}^s b_i \xi_i,\qquad X_i = \sum_{j=1}^s a_{ij} \xi_j,\quad i=1,\ldots,s, \\ G_i &= \exp(X_i)\cdot g_0,\quad i=1,\ldots, s, \\ M_i &= \dexp_{-Y}^* \mu_0 + \sum_{j=1}^s \left(b_j\dexp_{-Y}^* - \frac{b_ja_{ji}}{b_i}\dexp_{-X_j}^*\right) n_j,\quad i=1,\ldots,s. \end{align*} Here, $a_{ij}$ and $b_i$ are coefficients where it is assumed that $\sum_{i=1}^s b_i = 1$ and that $b_i\neq 0,\ 1\leq i\leq s$. The symplecticity of these schemes is a consequence of their derivation from a variational principle, following ideas similar to that of \cite{bou-rabee09hpi} and \cite{saccon09mrf}. One should be aware that order barriers for this type of schemes may apply, and that further stage corrections may be necessary to obtain high order methods.
\paragraph{Example, the $\theta$-method for a heavy top} Let us choose $s=1$ with coefficients $b_1=1$ and $a_{11}=\theta$, i.e.\ the RK coefficients of the well known $\theta$-method. Inserting this into our method and simplifying gives us the method \begin{align*}
(\xi, \bar n) &= h f\bigl(\exp(\theta \xi) \cdot g_0, \dexp_{-\xi}^{*} \mu_0 + (1 - \theta) \dexp_{-(1-\theta)\xi}^{*} \bar n\bigr), \\
(g_1, \mu_1) &= (\exp(\xi), \Ad_{\exp(-(1 -\theta) \xi)}^{*} \bar n) \cdot (g_0, \mu_0). \end{align*} \begin{figure}
\caption{Heavy top simulations with the symplectic (SLGI) $\theta$-methods and RKMK $\theta$-methods with $\theta=0,\tfrac{1}{2}$. The curves show the time evolution of the centre of mass of the body. The simulations were run over $10^5$ steps with step size $h=0.05$. See the text, page \pageref{vale:pgref}, for all other parameter values. }
\label{fig:heavytop}
\end{figure} In Figure~\ref{fig:heavytop} we show numerical experiments for the heavy top, where the Hamiltonian is given as \[
H \colon \mathbf{G} \to \mathbf{R}, \qquad H(g,\mu) =\frac{1}{2} \langle \mu, \mathbb{I}^{-1}\mu\rangle + \mathbf{e}_3^Tgu_0, \] where $G = \SO(3)$, $\mathbb{I}\colon \ensuremath{\mathfrak{g}}\rightarrow \ensuremath{\mathfrak{g}}^*$ is the inertia tensor, here represented as a diagonal \label{vale:pgref} $3\times 3$ matrix, $u_0$ is the initial position of the top's centre of mass, and $\mathbf{e}_3$ is the canonical unit vector in the vertical direction. We have chosen $\mathbb{I}=10^3\,\operatorname{diag}(1,5,6)$ and $u_0=\mathbf{e}_3$. The initial values used were $g_0=I$ (the identity matrix), and $\mu_0=10\,\mathbb{I}\,(1,1,1)^T$. We compare the behaviour of the symplectic schemes presented here to the Runge--Kutta--Munthe-Kaas (RKMK) method with the same coefficients. In Figure~\ref{fig:heavytop} we have drawn the time evolution of the centre of mass, $u_n=g_n u_0$. The characteristic band structure observed for the symplectic methods was reported in \cite{celledoni06ets}. The RKMK method with $\theta=\frac{1}{2}$ exhibits a similar behaviour, but the bands are expanding faster than for the symplectic ones. We have also found in these experiments that none of the symplectic schemes, $\theta=0$ and $\theta=\frac{1}{2}$ have energy drift, but this is also the case for the RKMK method with $\theta=\frac{1}{2}$. This may be related to the fact that both methods are symmetric for $\theta=\frac{1}{2}$. For $\theta=0$, however, the RKMK method shows energy drift as expected. These tests were done with step size $h=0.05$ over $10^5$ steps. See Table~\ref{tab:experiment} for a summary of the properties of the tested methods. \begin{table} \centering \morespacearray
\begin{tabular}{r|c|c|c|c|}
\cline{2-5}
& \multicolumn{2}{c|}{RKMK} & \multicolumn{2}{c|}{SLGI} \\
\cline{2-5}
& $\theta = 0$ & $\theta = 1/2$ & $\theta = 0$ & $\theta = 1/2$ \\
\hline
\multicolumn{1}{|r|}{Symplectic} & no & no & yes & yes \\
\multicolumn{1}{|r|}{Symmetric} & no & yes & no & yes \\
\multicolumn{1}{|r|}{No energy drift} & no & yes & yes & yes \\
\hline \end{tabular} \caption{Properties of the tested methods. The energy drift was observed numerically.} \label{tab:experiment} \end{table}
\section{Discrete gradients and integral preserving methods on Lie groups} \label{sec:discdiff} The discrete gradient method for preserving first integrals has to a large extent been made popular through the works of Gonzalez \cite{gonzalez96tia} and McLachlan et al.\ \cite{mclachlan99giu}. The latter proved the result that under relatively general circumstances, a differential equation which has a first integral $I(x)$ can be written in the form \[
\dot{x} = S(x) \nabla I(x), \] for some non-unique solution-dependent skew-symmetric matrix $S(x)$. The idea is to introduce a mapping which resembles the true gradient; a \emph{discrete gradient} $\overline{\nabla}I \colon \mathbf{R}^d\times\mathbf{R}^d\rightarrow \mathbf{R}^d$ which is a continuous map that satisfies the following two conditions: \begin{align*} \overline{\nabla}I(x,x) &= \nabla I(x),\quad \forall x, \\ I(y)-I(x) &= \overline{\nabla}I(x,y)^T(y-x),\quad \forall x\neq y. \end{align*} An integrator which preserves $I$, that is, $I(x_n)=I(x_0)$ for all $n$ is now easily devised as \[
\frac{x_{n+1}-x_n}{h} = \tilde{S}(x_n,x_{n+1})\overline{\nabla}I(x_n,x_{n+1}), \] where $\tilde{S}(x,y)$ is some consistent approximation to $S(x)$, i.e.\ $\tilde{S}(x,x)=S(x)$. There exist several discrete gradients, two of the most popular are \begin{equation} \label{eq:cavf}
\overline{\nabla}I(x,y) = \int_0^1 \nabla I(\zeta y + (1-\zeta) x) \, \ensuremath{\mathrm{d}} \zeta, \end{equation} and \begin{equation} \label{eq:gonz} \overline{\nabla} I(x,y) = \nabla I\left(\frac{x+y}{2}\right) + \frac{I(y)-I(x)-\nabla I\left(\frac{x+y}{2}\right)^T(y-x)}{\lVert y-x\rVert^2} (y-x). \end{equation} The matrix $\tilde{S}(x,y)$ can be constructed with the purpose of increasing the accuracy of the resulting approximation, see e.g.\ \cite{quispel08anc}.
We now generalise the concept of the discrete gradient to a Lie group $G$. We consider differential equations which can, for a given dual two-form\footnote{By dual two-form, we here mean a differential two-form on $G$ such that on each fibre of the cotangent bundle we have $\omega_x \colon T_x^*G\times T_x^*G\rightarrow\mathbf{R}$, a bilinear, skew-symmetric form. Such forms are sometimes called bivectors or 2-vectors.}
$\omega\in\Omega_2(G)$ and a function $H \colon G\rightarrow\mathbf{R}$ be written in the form \begin{equation*} \label{eq:intprod}
\dot{x} = \mathrm{i}_{\ensuremath{\mathrm{d}} H}\omega, \end{equation*} where $\mathrm{i}_{\alpha}$ is the interior product $\mathrm{i}_\alpha\omega(\beta)=\omega(\alpha,\beta)$ for any two one-forms $\alpha, \beta\in\Omega^1(G)$. The function $H$ is a first integral since \[
\frac{\ensuremath{\mathrm{d}}}{\ensuremath{\mathrm{d}} t} H(x(t)) = \ensuremath{\mathrm{d}} H_{x(t)}(\dot{x}(t))=\omega(\ensuremath{\mathrm{d}} H, \ensuremath{\mathrm{d}} H)=0. \] We define the \emph{trivialised discrete differential} (TDD) of the function $H$ to be a continuous map $\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H \colon G\times G \rightarrow \ensuremath{\mathfrak{g}}^*$ such that \begin{align*}
H(x')-H(x) &= \langle \ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x'), \log(x' \cdot x^{-1}) \rangle, \\
\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x) &= R_x^*\,\ensuremath{\mathrm{d}} H_x. \end{align*} A numerical method can now be defined in terms of the discrete differential as \[
x' = \exp(h \, \mathrm{i}_{\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x')}\bar{\omega}(x,x')) \cdot x. \] where $\bar{\omega}$ is a continuous map from $G\times G$ into the space of exterior two-forms on $\ensuremath{\mathfrak{g}}^*$, $\Lambda^2(\ensuremath{\mathfrak{g}}^*)$.
This exterior form is some local trivialised approximation to $\omega$, meaning that we impose the following consistency condition \begin{equation*} \label{eq:consomb}
\bar{\omega}(x,x)(R_x^*\alpha, R_x^*\beta) = \omega_x(\alpha,\beta), \qquad \text{for all }
\alpha, \beta\in T_x^*G. \end{equation*} We easily see that this method preserves $H$ exactly, since \begin{align*} H(x')-H(x) &= \langle\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x'), \log(x' \cdot x^{-1})\rangle \\
&= \langle\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x'), h\,\mathrm{i}_{\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x')}\bar\omega(x, x')\rangle \\
&= h\,\bar{\omega}(x,x') (\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x'), \ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x')) = 0. \end{align*} Extending \eqref{eq:cavf} to the Lie group setting, we define the following TDD: \begin{equation*} \label{eq:tdgavf}
\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x') = \int_0^1 R_{\ell(\xi)}^* \, \ensuremath{\mathrm{d}} H_{\ell(\xi)} \, \ensuremath{\mathrm{d}}\xi,\qquad
\ell(\xi) = \exp(\xi\log(x'\cdot x^{-1})) \cdot x. \end{equation*} Similarly, for any given inner product on $\ensuremath{\mathfrak{g}}$, we may extend the discrete gradient \eqref{eq:gonz} to \begin{equation*} \label{eq:gonz1}
\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x') = R_{\bar{x}}^* \, \ensuremath{\mathrm{d}} H_{\bar{x}} +
\frac{H(x')-H(x)-\langle R_{\bar{x}}^* \, \ensuremath{\mathrm{d}} H_{\bar{x}},\eta\rangle
}{\lVert\eta\rVert^2}\eta^\flat,\quad\eta=\log(x' \cdot x^{-1}),
\end{equation*} where $\bar{x}\in G$ for instance could be $\bar{x}=\exp(\eta/2) \cdot x$, a choice which would cause $\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x,x')=\ensuremath{\overline{\ensuremath{\mathrm{d}}}} H(x',x)$. The standard notation $\eta^\flat$ is used for index-lowering, the inner product $({\cdot},{\cdot})$ associates to any element $\eta\in\ensuremath{\mathfrak{g}}$ the dual element $\eta^\flat\in\ensuremath{\mathfrak{g}}^*$ through $\langle\eta^\flat, \zeta\rangle = (\eta,\zeta)$, $\forall\zeta\in\ensuremath{\mathfrak{g}}$.
Suppose that the ODE vector field $F$ is known as well as the invariant $H$. A dual two-form $\omega$ can now be defined in terms of a Riemannian metric on $G$. By index raising applied to $\ensuremath{\mathrm{d}} H$, we obtain the Riemannian gradient vector field $\grad H$, and we define \[
\omega = \frac{\grad H\wedge F}{\lVert\grad H\rVert^2}\qquad
\Rightarrow\qquad \mathrm{i}_{\ensuremath{\mathrm{d}} H}\omega = F. \] \paragraph{Example} We consider the equations for the attitude rotation of a free rigid body expressed using Euler parameters. The set $S^3=\{ \mathbbm{q}\in \mathbf{R}^4 \mid \lVert\mathbbm{q}\rVert_2=1\}$ with $\mathbbm{q}=(q_0,\mathbf{q})$, ($q_0\in\mathbf{R}$ and $\mathbf{q}\in\mathbf{R}^3$), is a Lie group with the quaternion product \[
\mathbbm{p}\cdot \mathbbm{q}=(p_0q_0-\mathbf{p}^T\mathbf{q}, p_0\mathbf{q}+q_0\mathbf{p}+\mathbf{p}\times \mathbf{q}), \] with unit $\mathbbm{e}=(1, 0, 0, 0)$ and inverse $\mathbbm{q}_c=(q_0,\,-\mathbf{q})$. We denote by ``$\hat{\quad}$" the hat-map defined in \eqref{eq:hatmap}.
The Lie group $S^3$ can be mapped into $\SO(3)$ by the Euler--Rodrigues map: \[
\mathcal{E}(\mathbbm{q})=I_3+2q_0\hat{\mathbf{q}}+2\hat{\mathbf{q}}^2, \] where $I_3$ denotes the $3\times 3$ identity matrix. The Lie algebra $\mathfrak{s}^3$ of $S^3$ is the set of so called pure quaternions, the elements of $\mathbf{R}^4$ with first component equal to zero, identifiable with $\mathbf{R}^3$ and with $\mathfrak{so}(3)$ via the hat-map.
The equations for the attitude rotation of a free rigid body on $S^3$ read \[
\dot{\mathbbm{q}}=f(\mathbbm{q})\cdot \mathbbm{q},\qquad f(\mathbbm{q})=\mathbbm{q}\cdot \mathbbm{v} \cdot \mathbbm{q}_c, \] and
\[
\mathbbm{v}=(0, \mathbf{v}),\qquad \mathbf{v} = \frac{1}{2} \mathbb{I}^{-1} \mathcal{E} (\mathbbm{q}_c)\mathbf{m}_0, \] where $\mathbf{m}_0$ is the initial body angular momentum and $\mathbb{I}$ is the diagonal inertia tensor, and according to the notation previously used in this section $F(\mathbbm{q})=f(\mathbbm{q})\cdot \mathbbm{q}$. The energy function is \[H(\mathbbm{q})=\frac{1}{2}\,\mathbf{m}_0^T\mathcal{E}(\mathbbm{q})\mathbb{I}^{-1}\mathcal{E}(\mathbbm{q}_c)\mathbf{m}_0.\]
We consider the $\mathbf{R}^3$ Euclidean inner product as metric in the Lie algebra $\mathfrak{s}^3$, and obtain by right translation a Riemannian metric on $S^3$. The Riemannian gradient of $H$ with respect to this metric is then \[\grad H=(I_4-\mathbbm{q}\mathbbm{q}^T)\nabla H,\] where $I_4$ is the identity in $\mathbf{R}^{4\times4}$ and $\nabla H$ is the usual gradient of $H$ as a function from $\mathbf{R}^4$ to $\mathbf{R}$. We identify $\mathfrak{s}^3$ with its dual, and using $\grad H$ in \eqref{eq:gonz} we obtain the (dual) discrete differential $\overline{\grad H}(\mathbbm{q},\mathbbm{q}')\in \mathfrak{s}^3$.
The
two-form $\omega = \frac{\grad H\wedge F}{\lVert\grad H\rVert^2}$ with respect to the right trivialisation can be identified with the $4\times 4$ skew-symmetric matrix
\[\omega_R(\mathbbm{q})=\frac{\xi\,\gamma^T-\gamma\,\xi^T}{\lVert\gamma\rVert^2},\qquad \xi ,\, \gamma\in\mathfrak{s}^3,\qquad \xi\cdot \mathbbm{q}=F(\mathbbm{q}), \qquad \gamma \cdot \mathbbm{q}=\grad H(\mathbbm{q}),\]
where $\omega_R(\mathbbm{q})$ has first row and first column equal to zero. We choose $\bar{\omega}$ to be \[\bar{\omega}(\mathbbm{q},\mathbbm{q}')=\omega_R(\bar{\mathbbm{q}}),\qquad \bar{\mathbbm{q}}=\exp(\eta/2) \cdot \mathbbm{q},\qquad \eta=\log (\mathbbm{q}'\cdot \mathbbm{q}_c),\] i.e.\ $\omega_R$ frozen at the mid-point $\bar{\mathbbm{q}}$.
The energy-preserving Lie group method of second order is
\[\mathbbm{q}'=\exp( h\,\bar{\omega}(\mathbbm{q},\mathbbm{q}') \overline{\grad H}(\mathbbm{q},\mathbbm{q}'))\cdot \mathbbm{q},\]
and $\exp$ is the exponential map from $\mathfrak{s}^3$ to $S^3$ with $\log \colon S^3\rightarrow \mathfrak{s}^3$ as its inverse, defined locally around the identity.
In Figure~\ref{fig:EP} we plot the body angular momentum vector $\mathbf{m}=\mathcal{E}(\mathbbm{q}_c)\mathbf{m}_0$ on a time interval $[0,T]$, $T=1000$, for four different methods: the Lie group energy-preserving integrator just described (top left), the built-in \textsc{Matlab} routine \texttt{ode45} with absolute and relative tolerance $10^{-6}$ (top right); the \texttt{ode45} routine with tolerances $10^{-14}$ (bottom left); and the explicit Heun RKMK Lie group method (bottom right). \begin{figure}
\caption{Free rigid body angular momentum, time interval $[0,1000]$, moments of inertia $I_1=1$, $I_2=5$, $I_3=60$, initial angular velocity $\mathbb{I}\,\mathbf{m}_0=(1, 1/2, -1)^T$. (Top left) energy-preserving Lie group method, $h=1/64$; (top right) \texttt{ode45} with tolerances $10^{-6}$; (bottom left) \texttt{ode45} with tolerances $10^{-14}$; (bottom right) Heun RKMK, $h=1/64$.}
\label{fig:EP}
\end{figure} The two Lie group methods both have order $2$. The energy preserving method is both symmetric, energy preserving and it preserves the constraint $\lVert\mathbbm{q}\rVert_2=1$. The Lie group integrators use a step-size $h=1/64.$ The solution of the built-in \textsc{Matlab} routine at high precision is qualitatively similar to the highly accurate solution produced by \textsc{Matlab} with tolerances $10^{-14}$. The energy error is also comparable for these two experiments. The performance of other \textsc{Matlab} built-in routines we tried was worse than for \texttt{ode45}. We remark that the equations are formulated as differential equations on $S^3$, a formulation of the problem in form of a differential algebraic equation would possibly have improved the performance of the \textsc{Matlab} built-in routines. However it seems that the preservation of the constraint alone can not guarantee the good performance of the method. In fact the explicit (non-symmetric) Lie group integrator preserves the constraint $\lVert\mathbbm{q}\rVert_2=1$, but performs poorly on this problem (see Figure~\ref{fig:EP} bottom right). The cost per step of the explicit Lie group integrator is much lower than for the energy-preserving symmetric Lie group integrator.
We have given an introduction to Lie group integrators for differential equations on manifolds using the notions of frames and Lie group actions. A few application areas have been discussed. An interesting observation is that when the Lie group featuring in the method can be chosen to be the Euclidean group, the resulting integrator always reduce to some well-known numerical scheme like for instance a Runge-Kutta method. In this way, one may think of the Lie group integrators as a superset of the traditional integrators and a natural question to ask is whether the Euclidean choice will always be superior to any other Lie group and group action.
Lie group integrators that are symplectic for Hamiltonian problems in the general setting presented here are, as far as we know, not known. However, we have shown that such methods exist in the important case of Hamiltonian problems on the cotangent bundle of a Lie group. There are however still many open questions regarding this type of schemes, like for instance how to obtain high order methods.
The preservation of first integrals in Lie group integrators has been achieved in the literature by imposing a group action in which the orbit, i.e. the reachable set of points, is contained in the level set of the invariant. But it is not always convenient to impose such group actions, and we have here suggested a type of Lie group integrator which can preserve any prescribed invariant for the case where the manifold is a Lie group acting on itself by left or right multiplication. An interesting idea to pursue is the generalisation of this approach to arbitrary smooth manifolds with a group action.
\end{document}
|
arXiv
|
{
"id": "1207.0069.tex",
"language_detection_score": 0.7588862776756287,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Analytic Continuation of Holomorphic Correspondences and
Equivalence of Domains in ${\mathbb C}^n$.} \author{Rasul Shafikov \\{\small \tt [email protected]}} \date{\today} \maketitle
\begin{abstract}
The following result is proved: Let $D$ and $D'$ be bounded domains in $\mathbb C^n$, $\partial D$ is smooth, real-analytic, simply connected, and $\partial D'$ is connected, smooth, real-algebraic. Then there exists a proper holomorphic correspondence $f:D\to D'$ if and only if $\partial D$ and $\partial D'$ a locally CR-equivalent. This leads to a characterization of the equivalence relationship between bounded domains in $\mathbb C^n$ modulo proper holomorphic correspondences in terms of CR-equivalence of their boundaries.
\end{abstract}
\section{Definitions and Main Results.}
Following Stein \cite{st1} we say that a holomorphic correspondence between two domains $D$ and $D'$ in ${\mathbb C}^n$ is a complex-analytic set $A \subset D\times D'$ which satisfies: (i) $A$ is of pure complex dimension $n$, and (ii) the natural projection $\pi:A\to D$ is proper. A correspondence $A$ is called proper, if in addition the natural projection $\pi': A\to D'$ is proper. We may also think of $A$ as the graph of a multiple valued mapping defined by $F:=\pi'\circ\pi^{-1}$. Holomorphic correspondences were studied for instance, in \cite{bb}, \cite{bb2}, \cite{be2}, \cite{bsu}, \cite{dp1}, \cite{dfy}, \cite{p3}, and \cite{v}.
In this paper we address the following question: Given two domains $D$ and $D'$, when does there exist a proper holomorphic correspondence $F:D\to D'$? Note (see \cite{st2}) that the existence of a correspondence $F$ defines an equivalence relation $D \sim D'$. This equivalence relation is a natural generalization of biholomorphic equivalence of domains in ${\mathbb C}^n$.
To illustrate the concept of equivalence modulo holomorphic correspondences, consider domains of the form
$\Omega_{p,q}= \{|z_1|^{2p}+|z_2|^{2q}<1 \}, \ \ p,q\in{\mathbb Z}^+$. Then $f(z)=({z_1}^{p/s},{z_2}^{q/t})$ is a proper holomorphic correspondence between $\Omega_{p,q}$ and $\Omega_{s,t}$, while a
proper holomorphic map from $\Omega_{p,q}$ to $\Omega_{s,t}$ exists only if $s|p$ and $t|q$, or $s|q$ and $t|p$. For details see \cite{bd} or \cite{l}.
The main result of this paper is the following theorem.
\begin{theorem}\label{t1} Let $D$ and $D'$ be bounded domains in ${\mathbb C}^n$, $n>1$. Let $\partial D$ be smooth, real-analytic, connected and simply connected, and let $\partial D'$ be connected, smooth, real-algebraic. Then there exists a proper holomorphic correspondence $F:D\to D'$ if and only if there exist points $p\in\partial D$, $p'\in\partial D'$, neighborhoods $U\ni p$ and $U'\ni p'$, and a biholomorphic map $f:U\to U'$, such that $f(U\cap\partial D)= U'\cap\partial D'$. \end{theorem}
In other words, we show that a germ of a biholomorphic mapping between the boundaries extends analytically to a holomorphic correspondence of the domains. By the example above, the extension will not be in general single-valued. Note that we do not require pseudoconvexity of either $D$ or $D'$. Also $\partial D'$ is not assumed to be simply connected. If both $D$ and $D'$ have real-algebraic boundary, i.e. each is globally defined by an irreducible real polynomial, then we can drop the requirement of simple connectivity of $\partial D$.
\begin{theorem}\label{t2} Let $D$ and $D'$ be bounded domains in ${\mathbb C}^n$, $n>1$, with connected smooth real-algebraic boundaries. Then there exists a proper holomorphic correspondence between $D$ and $D'$ if and only if there exist points $p\in\partial D$, $p'\in\partial D'$, neighborhoods $U\ni p$ and $U'\ni p'$, and a biholomorphic map $f:U\to U'$, such that $f(U\cap\partial D)= U'\cap\partial D'$. \end{theorem}
The proof of Theorem \ref{t2} is simpler than that of Theorem \ref{t1} due to the fact that when both domains are algebraic, Webster's theorem \cite{w} can be applied. We give a separate complete proof of Theorem \ref{t2} in Section 3 to emphasize the ideas of the proof of Theorem \ref{t1} and the difficulties that arise in the case when one of the domains is not algebraic.
Local CR-equivalence of the boundaries, which is used in the above theorems to characterize correspondence equivalence, is a well-studied subject. Chern and Moser \cite{cm} gave the solution to local equivalence problem for real-analytic hypersurfaces with non-degenerate Levi form both in terms of normalization of Taylor series of the defining functions and in terms of intrinsic differential-geometric invariants of the hypersurfaces. See also \cite{se}, \cite{ca} and \cite{t}. Note that for a bounded domain in ${\mathbb C}^n$ with smooth real-analytic boundary, the set of points where the Levi form is degenerate is a closed nowhere dense set. Thus we may reformulate Theorem \ref{t1} in the following way.
\begin{theorem} Let $D$ and $D'$ be as in Theorem \ref{t1}. Then $D$ and $D'$ are correspondence equivalent if and only if there are points $p\in\partial D$ and $p'\in\partial D'$ such that the Levi form of the defining functions are non-degenerate at $p$ and $p'$, and the corresponding Chern-Moser normal forms are equivalent. \end{theorem}
Theorem \ref{t1} generalizes the result in \cite{p1}, which states that a bounded, strictly pseudoconvex domain $D\subset{\mathbb
C}^n$ with connected, simply-connected, real-analytic boundary is biholomorphically equivalent to ${\mathbb B}^n={\mathbb B}(0,1)$, the unit ball centered at the origin, if and only if there exists a local biholomorphism of the boundaries. It was later shown in \cite{cj} (see also \cite{ber}) that simple connectivity of $\partial D$ can be replaced by simple connectivity of the domain $D$ and connectedness of $\partial D$. In \cite{p2} Pinchuk also established sufficient and necessary conditions of equivalence for bounded, strictly pseudoconvex domains with real-analytic boundaries. Two such domains $D$ and $D'$ are biholomorphically equivalent if and only if there exist points $p\in\partial D$ and $p'\in\partial D'$ such that $\partial D$ and $\partial D'$ are locally equivalent at $p$ and $p'$.
If in Theorem \ref{t1} $\partial D$ is not assumed to be simply connected, then the result is no longer true. Indeed, in a famous example Burns and Shnider \cite{bs} constructed a domain $\O$ with the boundary given by \begin{equation}\label{bad}
\partial \O=\{z\in{\mathbb C}^2 : \sin(\ln|z_2|) + |z_1|^2=0, \
e^{-\pi}\le|z_1|\le 1\}, \end{equation} which is real-analytic and strictly pseudoconvex but not simply connected. There exists a mapping $f:{\mathbb B}^2\to \O$ such that $f$ does not extend even continuously to a point on $\partial {\mathbb B}^2$. The inverse to $f$ gives a local biholomorphic map from $\partial \O$ to $\partial {\mathbb B}^2$, but nevertheless $f^{-1}$ does not extend to a global proper holomorphic correspondence between $\O$ and ${\mathbb B}^2$. Furthermore, suppose that there exists a proper correspondence $g:\O\to {\mathbb B}^2$. Since ${\mathbb B}^2$ is simply connected, $g^{-1}$ is a proper holomorphic mapping, which extends holomorphically to a neighborhood of $\overline{ {\mathbb B}^2}$. Let $p\in \partial \O$, $q^1\in f^{-1}(p)$, and $q^2\in g(p)$. By the result in \cite{bb} $g$ splits at every point in ${\overline \O}$. Therefore, with a suitable choice of the branch of $g$ near $p$, $\phi:=g\circ f$ defines a local biholomorphic mapping from a neighborhood $U_1\ni q^1$ to some neighborhood $U_2\ni q^2$. Moreover, $\phi(U_1\cap \partial {\mathbb B}^2)\subset (U_2\cap \partial {\mathbb B}^2)$. By the theorem of Poincar\'e-Alexander (see e.g. \cite{a}), $\phi$ extends to a global automorphism of ${\mathbb B}^2$. Thus after a biholomorphic change of coordinates, by the uniqueness theorem $f$ and $g^{-1}$ must agree in ${\mathbb B}^2$. But $f:{\mathbb B}^2\to \Omega$ is not a proper map. This contradiction shows that there are no proper holomorphic correspondences between $\O$ and ${\mathbb
B}^2$. Thus the condition of simple connectivity of $\partial D$ in Theorem \ref{t1} cannot be in general weakened.
One direction in the proof of Theorems \ref{t1} and \ref{t2} is essentially contained in the work of Berteloot and Sukhov \cite{bsu}. The proof in the other direction is based on the idea of extending the mapping $f$ along $\partial D$ as a holomorphic correspondence.
It is not known whether Theorem \ref{t1} holds if $\partial D'$ is real-analytic. The main difficulty is to prove local analytic continuation of holomorphic mappings along real-analytic hypersurfaces in the case when hypersurfaces are not strictly pseudoconvex. In particular, Lemma \ref{l:alongSV} cannot be directly established for real-analytic hypersurfaces.
In Section 2 we present background material on Segre varieties and holomorphic correspondences. In Section 3 we prove a technical result, important for the proof of the main theorems. Section 4 contains the proof of Theorem~\ref{t2}. In Section 5 we prove local extendability of holomorphic correspondences along hypersurfaces. Theorem \ref{t1} is proved in Section 6.
\section{Background Material.}
Let $\G$ be an arbitrary smooth real-analytic hypersurface with a defining function $\rho(z,\overline z)$ and let $z^0\in\G$. In a suitable neighborhood $U\ni z^0$ to every point $w\in U$ we can associate its so-called Segre variety defined as \begin{equation} Q_w=\left\{ z\in U: \rho(z,\overline w)=0 \right\}, \end{equation} where $\rho(z,\overline w)$ is the complexification of the defining function of $\G$. After a local biholomorphic change of coordinates near $z^0$, we can find neighborhoods $U_1 \Subset U_2$ of $z^0$, where \begin{equation}\label{e:1.00} U_2={'U_2}\times {{''U_{2}}}\subset {\mathbb C}^{n-1}_{{\ 'z}}\times {\mathbb C}_{z_n}, \end{equation} such that for any $w\in U_1$, $Q_w$ is a closed smooth complex-analytic hypersurface in $U_2$. Here $z=({'z},z_n)$. Furthermore, a Segre variety can be written as a graph of a holomorphic function, \begin{equation}\label{e:implicit} Q_w=\left\{({'z},z_n)\in {'U_2}\times {''U_{2}} : \ z_n=h({'z}, \overline w)\right\}, \end{equation} where $h(\cdot,\overline w)$ is holomorphic in ${'U_{2}}$. Following \cite{dp1} we call $U_1$ and $U_2$ a \textsl{standard} pair of neighborhoods of $z^0$. A detailed discussion of elementary properties of Segre varieties can be found in \cite{dw}, \cite{df2}, \cite{dp1} or \cite{ber1}.
The map $\lambda:z\to Q_z$ is called the Segre map. We define $I_w=\l^{-1}\circ \l(w)$. This is equivalent to \begin{equation} I_w=\left\{z\in U_1: Q_z=Q_w\right\}. \end{equation} If $\G$ is of finite type in the sense of D'Angelo, or more generally, if $\G$ is essentially finite, then there exists a neighborhood $U$ of $\G$ such that for any $w\in U$ the set $I_w$ is finite. Due to the result in \cite{df1}, this is the case for compact smooth real-analytic hypersurfaces, in particular for the boundaries of $D$ and $D'$ in Theorem \ref{t1}. The last observation is crucial for the construction of proper holomorphic correspondences used throughout this paper. We remark that our choice of a standard pair of neighborhoods of any point $z\in\G$ is always such that for any $w\in U_2$, the set $I_w$ is finite.
For the proof of Theorem \ref{t1} we will need the following lemma.
\begin{lemma}\label{l-finite} Let $\G$ be a compact, smooth, real-algebraic hypersurface. Then there exist a neighborhood $U$ of $\G$ and an integer $m\ge 1$ such that for almost any $z\in U$, $\#I_z = m$. \end{lemma}
\noindent{\it Proof.} Let $P(z,\overline z)$ be the defining polynomial of $\G$ of degree $d$ in $z$. The complexified defining function can be written in the form \begin{equation}
P(z,\overline w) = \sum_{|K|=0}^d{a_K(\overline w) z^K},\ \ K=(k_1,\dots, k_n). \end{equation} We may consider the projectivized version of the polynomial $P:\mathbb P^n \times\mathbb C^n \to \mathbb C$: \begin{equation}\label{proj-poly} \tilde P(\tilde\zeta,\overline w)=\zeta_0^d
\sum_{|K|=0}^d{a_K(\overline w) \left(\frac{\zeta}{\zeta_0}\right)^K}=
\sum_{|K|\le d} a_K(\overline w)\tilde \zeta^K, \end{equation} where $z_j=\frac{\zeta_j}{\zeta_0}$, $\zeta=(\zeta_1,\dots,\zeta_n)$, and $\tilde\zeta = (\zeta_0,\zeta)$. Let $\tilde Q_w = \{\tilde\zeta\in \mathbb P^n : \tilde P(\tilde\zeta,\overline w)=0\}$. Then $Q_w=\tilde Q_w\cap \{\zeta_0=1\}$.
Define the map $\hat \l$ between the set of points in $\mathbb C^n$ and the space of complex hypersurfaces in $\mathbb P^n$, by letting $\hat \l (w) = \{\tilde\zeta\in\mathbb P^n : \tilde P(\tilde\zeta, \overline w)=0\}$. Hypersurfaces in $\mathbb P^n$ can be parametrized by points in $\mathbb P^N$, where $N$ is some large integer, thus $\hat \l: \mathbb C^n\to \mathbb P^{N}$. Note that each component of $\hat\l$ is defined by a (antiholomorphic) polynomial.
It follows that $\hat\l^{-1}\circ\hat\l (w)=I_w$ for every $w\in\mathbb C^n$, for which $Q_w$ is defined. Indeed, suppose $\xi\in I_w$. Then $\{P(z,\overline\xi)=0\}=\{P(z,\overline w)=0\}$, and therefore, \begin{equation}\label{p=p} \{\tilde P(\tilde\zeta,\overline\xi)=0\}= \{\tilde P(\tilde\zeta,\overline w)=0\}. \end{equation} Thus $\hat\l(\xi)=\hat\l(w)$. The converse clearly also holds.
From the reality condition on $P(z,\overline z)$ (see \cite{dw}), \begin{equation} I_w\subset \G, {\rm\ \ for\ } w\in\G. \end{equation} Since $\G$ is compact, $I_w=\hat\l^{-1}\circ\hat\l (w)$ is finite. Let $Y=\hat\l(\mathbb C^n)$. Then $\dim Y = n$, and $\hat\l$ is a dominating regular map. It follows (see e.g. \cite{m}), that there exists an algebraic variety $Z\subset Y$ such that for any $q\in Y\setminus Z$, \begin{equation}\label{deg} \#\hat\lambda^{-1} (q)=\deg(\hat\lambda)=m, \end{equation} where $m$ is some positive integer. From (\ref{deg}) the assertion follows. $\square$
The following statement describes the invariance property of Segre varieties under holomorphic mappings. It is analogous to the classical Schwarz reflection principle in dimension one.
Suppose that $\G$ and $\G'$ are real-analytic hypersurfaces in ${\mathbb C}^n$, $(U_1, U_2)$ and $(U'_1, U'_2)$ are standard pairs of neighborhoods for $z_0\in\G$ and $z'_0\in\G'$ respectively. Let $f: U_2 \to U'_2$ be a holomorphic map, $f(U_1)\subset U'_1$ and $f(\G\cap U_2)\subset(\G'\cap U'_2)$. Then \begin{equation}\label{e-invar} f(Q_w\cap U_2)\subset Q'_{f(w)}\cap U'_2,\ \ {\rm for\ all\ } w\in U_1. \end{equation} Moreover, a similar invariance property also holds for a proper holomorphic correspondence $f:U_2 \to U'_2$, $f(\G\cap U_2)\subset(\G'\cap U'_2)$. In this case (\ref{e-invar}) indicates that any branch of $f$ maps any point from $Q_w$ to $Q_{w'}$ for any $w'\in f(w)$. For details, see \cite{dp1} or \cite{v}.
Let $f:D\to D'$ be a holomorphic correspondence. We say that $f$ is {\it irreducible}, if the corresponding analytic set $A\subset D\times D'$ is irreducible. The condition that $f$ is proper is equivalent to the condition that \begin{equation} \sup \{ {\rm dist}(f(z),\partial D') \}\to 0,\ \ {\rm as}\ \ {\rm dist}(z,\partial D)\to 0. \end{equation}
Recall that if $A\subset D\times D'$ is a proper holomorphic correspondence, then $\pi: A\to D$ and $\pi':A\to D'$ are proper. There exists a complex subvariety $S\subset D$ and a number $m$ such that \begin{equation} f:=\pi'\circ \pi^{-1}=(f^1(z),\dots,f^m(z)), \ \ z\in D, \end{equation} where $f^j$ are distinct holomorphic functions in a neighborhood of $z\in D\setminus S$. The set $S$ is called the {\it branch locus} of $f$. We say that the correspondence $f$ {\it splits} at $z\in {\overline D}$ if there is an open subset $U\ni z$ and holomorphic maps $f^j:D\cap U\to D'$, $i=1,2,\dots,m$ that represent $f$.
Given a proper holomorphic correspondence $A$, one can find the system of canonical defining functions \begin{equation}\label{e-canon}
\Phi_I(z,z') = \sum_{|J|\le m}\phi_{IJ}(z){z'}^J,\ \ |I|=m, \ \ (z,z')\in{\mathbb C}^n\times{\mathbb C}^n, \end{equation} where $\phi_{IJ}(z)$ are holomorphic on $D$, and $A$ is precisely the set of common zeros of the functions $\Phi_I(z,z')$. For details see e.g. \cite{c}.
We define analytic continuation of analytic sets as follows. Let $A$ be a locally complex analytic set in ${\mathbb C}^n$ of pure dimension $p$. We say that $A$ {\it extends analytically} to an open set $U\subset{\mathbb C}^n$ if there exists a (closed) complex-analytic set $A^*$ in $U$ such that (i) $\dim A^*=p$, (ii) $A\cap U\subset A^*$ and (iii) every irreducible component of $A^*$ has a nonempty intersection with $A$ of dimension $p$. Note that if conditions (i) and (ii) are satisfied, then the last condition can always be met by removing certain components of $A^*$. It follows from the uniqueness theorem that such analytic continuation of $A$ is uniquely defined. From this we define analytic continuation of holomorphic correspondences:
\begin{definition} Let $U$ and $ U'$ be open sets in ${\mathbb C}^n$. Let $f:U\to U'$ be a holomorphic correspondence, and let $A\subset U\times U'$ be its graph. We say that $f$ extends as a holomorphic correspondence to an open set $V$, $U\cap V\not=\varnothing$, if there exists an open set $V'\subset {\mathbb C}^n$ such that $A$ extends analytically to a set $A^*\subset V\times V'$ and $\pi:A^*\to V$ is proper. \end{definition}
Note that we can always choose $V'={\mathbb C}^n$ in the definition above. In general, correspondence $g=\pi'\circ\pi^{-1}:V\to V'$, where $\pi':A^*\to V'$ is the natural projection, may have more branches in $U\cap V$ than $f$. The following lemma gives a simple criterion for the extension to have the same number of branches.
\begin{lemma}\label{l-intersection} Let $A^*\subset V\times{\mathbb C}^n$ be a holomorphic correspondence which is an analytic extension of the correspondence $A\subset U\times {\mathbb C}^n$. Suppose that for any $z\in (V\cap U)$, \begin{equation}\label{pre} \#\{\pi^{-1}(z)\}=\#\{\pi^{*-1}(z)\}, \end{equation} where $\pi:A\to U$ and $\pi^*:A^*\to V$. Then $A\cup A^*$ is a holomorphic correspondence in $(U\cup V)\times {\mathbb C}^n$. \end{lemma}
\noindent{\it Proof.} We only need to check that $A\cup A^*$ is closed in $(U\cup V)\times {\mathbb C}^n$. If not, then there exists a sequence $\{q^j\}\subset A^*$ such that $q^j\to q^0$ as $j\to\infty$, $q^0\in U\times {\mathbb C}^n$, and $q^0\not\in A$. Then $q^j\not\in A$ for $j$ sufficiently large. Since by the definition of analytic continuation of correspondences $A\cap (U\cap V)\times {\mathbb C}^n\subset A^*$, we have $$ \#\{\pi^{-1}(\pi^*(q^j))\} < \#\{\pi^{*-1}(\pi^*(q^j))\}. $$ But this contradicts (\ref{pre}). $\square$
\section{Extension along Segre Varieties.}
Before we prove the main results of the paper we need to establish a technical result of local extension of holomorphic correspondence along Segre varieties. This will be used on several occasions in the proof of the main theorems.
\begin{lemma}\label{l:alongSV} Let $\G\subset{\mathbb C}^n$ be a smooth, real-analytic, essentially finite hypersurface, and let $\G'\subset{\mathbb C}^n$ be a smooth, real-algebraic, essentially finite hypersurface. Let $0\in\G$, and let $U_1$, $U_2$ be a sufficiently small standard pair of neighborhoods of the origin. Let $f:U\to{\mathbb C}^n$ be a germ of a holomorphic correspondence such that $f(U\cap\G)\subset\G'$, where $U$ is some neighborhood of the origin. Then there exists a neighborhood $V$ of $Q_0\cap U_1$ and an analytic set $\L\subset V$, $\dim_{\mathbb C} \L\le n-1$, such that $f$ extends to $V\setminus \L$ as a holomorphic correspondence. \end{lemma}
\noindent{\it Proof.} In the case when $\G'$ is strictly pseudoconvex and $f$ is a germ of a biholomorphic mapping, the result was established in \cite{s1}. Here we prove the lemma in a more general situation.
Lemma \ref{l:alongSV} only makes sense if $U\subset U_1$. We shrink $U$ and choose $V$ in such a way that for any $w\in V$, the set $Q_w\cap U$ is non-empty and connected. Note that if $w\in Q_0$, then $0\in Q_w$ and $Q_w\cap U\not=\varnothing$. Let $S\subset U$ be the branch locus of $f$, and let \begin{equation} \S=\{z\in V: (Q_z\cap U)\subset S\} \end{equation} Since $\dim_{\mathbb C} S=n-1$ and $\G$ is essentially finite, $\S$ is a finite set. Define \begin{equation}\label{e:A1} A=\left\{ (w,w')\in (V\setminus\S)\times {\mathbb C}^n : f\left(Q_w\cap U\right)\subset Q'_{w'} \right\} \end{equation} We establish the following facts about the set $A$: \begin{list}{}{} \item[(i)] $A$ is not empty \item[(ii)] $A$ is locally complex analytic \item[(iii)] $A$ is closed \item[(iv)] $\S\times{\mathbb C}^n$ is a removable singularity for $A$. \end{list}
(i) $A\not=\varnothing$ since by the invariance property of Segre varieties, $A$ contains the graph of $f$.
(ii) Let $(w,w')\in A$. Consider an open simply connected set $\O\in(U\setminus S)$ such that $Q_w\cap\Omega\not=\varnothing$. Then the branches of $f$ are correctly defined in $\O$. Since $Q_w\cap U$ is connected, the inclusion $f(Q_w\cap U)\subset Q'_{w'}$ is equivalent to \begin{equation}\label{e:omega1} f^j(Q_w\cap\Omega)\subset Q'_{w'}, \ j=1,\dots,m, \end{equation} where $f^j$ denote the branches of $f$ in $\Omega$. Note that such neighborhood $\Omega$ exists for any $w\in V\setminus\S$. Inclusion (\ref{e:omega1}) can be written as a system of holomorphic equations as follows. Let $\rho'(z,\overline z)$ be the defining function of $\G'$. Then \begin{equation} \rho'(f^j(z), \overline{w'})=0, \ {\rm\ for\ any}\ z\in (Q_w\cap \Omega), \ \ j=1,2,\dots,m. \end{equation} We can choose $\Omega$ in the form \begin{equation} \O={'\O}\times{\O_n}\subset{\mathbb C}^{n-1}_{'z}\times{\mathbb C}_{z_n} \end{equation} Combining this with (\ref{e:implicit}) we obtain \begin{equation}\label{e-system1} \rho'(f^j({'z},h('z,\overline w)), \overline{w'})=0 \end{equation} for any $'z\in {'\Omega}$. Then (\ref{e-system1}) is an infinite system of holomorphic equations in $(w,w')$ thus defining $A$ as a locally complex analytic variety in $(V\setminus\S)\times{\mathbb C}^n$.
(iii) Let us show now that $A$ is a closed set. Suppose that $(w^j, {w'}^j)\to(w^0, {w'}^0)$, as $j\to\infty$, where $(w^j, {w'}^j)\in A$ and $(w^0,{w'}^0)\in (V\setminus\S)\times{\mathbb C}^n$. Then by the definition of $A$, $f(Q_{w^j}\cap U)\subset Q'_{{w'}^j}$. Since $Q_{w^j}\to Q_{w^0}$, and $Q'_{{w'}^j}\to Q'_{w^0}$ as $j\to\infty$, by analyticity $f(Q_{w^0}\cap U)\subset Q'_{{w'}^0}$, which implies that $(w^0,{w'}^0)\in A$ and thus $A$ is a closed set. Since $A$ is locally complex-analytic and closed, it is a complex variety in $(V\setminus\S)\times{\mathbb C}^n$. We now may restrict considerations only to the irreducible component of $A$ which coincides with the graph of $f$ at the origin. Then $\dim A = n$.
(iv) Let us show now that $\overline A$ is a complex variety in $V\times{\mathbb C}^n$. Let $q\in \S$, then \begin{equation} \overline {A}\cap(\{q\}\times{\mathbb C}^n) \subset \{q\}\times \{z': f(Q_q)\subset Q'_{z'}\}. \end{equation} Notice that if $w'\in f(Q_q)\subset Q'_{z'}$, then $z'\in Q'_{w'}$. Hence the set $\{z': f(Q_q)\subset Q'_{z'}\}$ has dimension at most $2n-2$, and $\overline{A}\cap(\S\times{\mathbb C}^n)$ has Hausdorff $2n$-measure zero. It follows from Bishop's theorem on removable singularities of analytic sets (see e.g. \cite{c}) that $\overline{A}$ is an analytic set in $V\times {\mathbb C}^n$.
Thus from (i) - (iv) we conclude that (\ref{e:A1}) defines a complex-analytic set in $V\times {\mathbb C}^n$ which we denote again by $A$. Also we observe that since $\G'$ is algebraic, the system of holomorphic equations in (\ref{e-system1}) is algebraic in $w'$ and thus we can define the closure of $A$ in $V\times {\mathbb P}^n$. Let $\pi:A\to V$ and $\pi': A\to {\mathbb P}^n$ be the natural projections. Since ${\mathbb P}^n$ is compact, $\pi^{-1}(K)$ is compact for any compact set $K\subset V$, and thus $\pi$ is proper.
This, in particular, implies that $\pi(A)=V$. We let $\L_1=\pi({\pi'}^{-1}(H_0))$, where $H_0\subset{\mathbb P}^n$ is the hypersurface at infinity. It is easy to see that $\L_1$ is a complex analytic set of dimension at most $n-1$. We also consider the set $\L_2:=\pi\{ (w,w')\in A : \dim_{\mathbb C} \pi^{-1}(w) \ge 1\}$. It was shown in \cite{s1} Prop.~3.3, that $\L_2$ is a complex-analytic set of dimension at most $n-2$. Let $\L=\L_1\cup\L_2$. Then $\pi'\circ
\pi^{-1}|_{V\setminus\L}$ is the desired extension of $f$ as a holomorphic correspondence. $\square$
\section{Proof of Theorem \ref{t2}.}
For completeness let us repeat the argument of \cite{bsu} to prove the ``only if'' direction of Theorems \ref{t1} and \ref{t2}. Suppose that $f:D\to D'$ is a proper holomorphic correspondence. Let us show that $\partial D$ and $\partial D'$ are locally CR-equivalent.
If $D$ is not pseudoconvex, then for $p\in\widehat D$, there exists a neighborhood $U\ni p$ such that all the functions in the representation (\ref{e-canon}) of $f$ extend holomorphically to $U$. Here $\widehat D$ refers to the envelope of holomorphy of $D$. Moreover, we can replace $p$ by a nearby point $q\in U\cap\partial D$ so that $f$ splits at $q$ and at least one of the holomorphic mappings of the splitting is biholomorphic at $q$.
If $D$ is pseudoconvex, then $D'$ is also pseudoconvex. By \cite{bsu} $f$ extends continuously to $\partial D$ and we can choose $p\in \partial D$ such that $f$ splits in some neighborhood $U\ni p$ to holomorphic mappings $f^{j}: D\cap U\to D'$, $j=1,\dots,m$. Since $f^{-1}:D'\to D$ also extends continuously to $\partial D'$, the set $\{f^{-1}\circ f(p)\}$ is finite. Therefore, by \cite{be3}, $f^j$ extend smoothly to $\partial D\cap U$. It follows that $f^j$ extend holomorphically to a neighborhood of $p$ by \cite{be} and \cite{df2}. Finally, choose $q\in U\cap\partial D$ such that $f^j$ is biholomorphic at $q$ for some $j$.
To prove Theorem \ref{t2} in the other direction, consider a neighborhood $U$ of $p\in\partial D$ and a biholomorphic map $f:U\to{\mathbb C}^n$ such that $f(U\cap\partial D)\subset \partial D'$. Let us show that $f$ extends to a proper holomorphic correspondence $F:D\to D'$.
Let $\G=\partial D$ and $\G'=\partial D'$. Since the set of Levi non-degenerate points is dense in $\G$, by Webster's theorem \cite{w}, $f$ extends to an algebraic mapping, i.e. the graph of $f$ is contained in an algebraic variety $X\subset{\mathbb C}^n \times {\mathbb C}^n$ of dimension $n$. Without loss of generality assume that $X$ is irreducible, as otherwise consider only the irreducible component of $X$ containing $\G_f$, the graph of the mapping $f$.
Let $E=\{ z\in \mathbb C^n : \dim \pi^{-1}(z)>0\}$, where $\pi:X\to \mathbb C^n$ is the coordinate projection to the first component. Then $E$ is an algebraic variety in $\mathbb C^n$. Let $f:{\mathbb C}^n \setminus E \to {\mathbb C}^n$ now denote the multiple valued map corresponding to $X$. Let $S\subset{\mathbb
C}^n\setminus E$ be the branch locus of $f$, in other words, for any $z\in S$ the coordinate projection onto the first component is not locally biholomorphic near $\pi^{-1}(z)$. To prove Theorem \ref{t2} it is enough to show that $E\cap \G=\varnothing$.
\begin{lemma}\label{l2.1} Let $p\in\G$. If $Q_p\not\subset E$, then $p\not\in E$. \end{lemma}
\noindent{\it Proof.} Suppose, on the contrary, that $p\in E$. Since $Q_p\not\subset E$, there exist a point $b\in Q_p$ and a small neighborhood $U_b$ of $b$ such that $U_b\cap E = \varnothing$. Choose neighborhoods $U_b$ and $U_p$ such that for any $z\in U_p$, the set $Q_z\cap U_b$ is non-empty and connected. Let \begin{equation} \S=\{z\subset U_p: Q_z\cap U_b\subset S\}. \end{equation} Similar to (\ref{e:A1}), consider the set \begin{equation} A=\left\{ (w,w')\in (U_p\setminus\S)\times {\mathbb C}^n : f\left(Q_w\cap U_b\right)\subset Q'_{w'} \right\}. \end{equation}
Then $A\not=\varnothing$. Indeed, since $\dim_{\mathbb C} E\le n-1$, there exists a sequence of points $\{p^j\}\subset (U_p\cap\G)\setminus (E\cup\S)$ such that $p^j\to p$ as $j\to\infty$. By the invariance property of holomorphic correspondences, for every $p^j$ there exists a neighborhood $U_j\ni p^j$ such that $f(Q_{p^j}\cap U_j)\subset Q'_{{p'}^j}$, where ${p'}^j\in f(p^j)$. But this implies that $f(Q_{p^j}\cap U_b)\subset Q'_{{p'}^j}$, and therefore $(p^j,{p'}^j)\in A$. Moreover, it follows that \begin{equation}\label{e:A=X}
A|_{U_{j}\times{\mathbb C}^n}=X|_{U_{j}\times{\mathbb C}^n}, \ \ j=1,2,\dots,m. \end{equation}
Similar to the proof of Lemma \ref{l:alongSV}, one can show that $A$ is a complex analytic variety in $(U_p\setminus\S)\times{\mathbb C}^n$, and that $\S\times{\mathbb C}^n$ is a removable singularity for $A\subset U_p\times{\mathbb C}^n$. Denote the closure of $A$ in $U_p\times{\mathbb C}^n$ again by $A$.
Without loss of generality we assume that $A$ is irreducible, therefore in view of (\ref{e:A=X}) we conclude that
$A|_{U_{p}\times{\mathbb C}^n}=X|_{U_{p}\times{\mathbb C}^n}$. Let $\hat f$ be a multiple valued mapping corresponding to $A$. Then by analyticity, there exists $p'\in\G'\cap \hat f(p)$. Moreover, by construction, $\hat f(p) = I'_{p'}$. By \cite{dw}, \begin{equation}\label{e-inG} I'_{z'}\subset\G', \ \ {\rm for\ any\ } z'\in\G'. \end{equation} Now choose $U_p$ so small that $\overline A \cap (U_p \times \partial U') =\varnothing$, where $U'$ is a neighborhood of $\G'$. This is always possible, since otherwise there exists a sequence of points $\{(z^j, {z'}^j), j=1,2,\dots\}\subset A$, such that $z^j\to p$ and ${z'}^j\to {z'}^0 \in\partial U'$ as $j\to\infty$. Then $(p,{z'}^0)\in A$ and ${z'}^0\not\in \G'$. But this contradicts (\ref{e-inG}).
This shows that $\hat f:U_p\to U'$ is a holomorphic correspondence extending $f$, which contradicts the assumption $p\in E$. $\square$
\begin{lemma}\label{l2.2} Let $p\in\G$. Then there exists a change of variables, which is biholomorphic near $\overline {D'}$, such that in the new coordinate system $Q_p\not\subset E$. \end{lemma}
{\it Proof.} Suppose that $Q_p\subset E$. Then we find a point $a\in(\G\setminus E)$ such that $Q_a\cap Q_p\not=\varnothing$. The existence of such $a$ follows, for example, from \cite{s1} Prop~4.1. (Note, that $\dim E\cap \G \le 2n-3$). By Lemma \ref{l:alongSV} the germ of the correspondence $f$ defined at $a$, extends holomorphically to a neighborhood $V$ of $Q_a$. Let $\L_1$ and $\L_2$ be as in Lemma~\ref{l:alongSV}. Since $\dim \L_2<n-1$, we may assume that $(Q_p\cap V)\not\subset \L_2$. If $(Q_p\cap V)\subset\L_1$, we can perform a linear-fractional transformation such that $H_0$ is mapped onto another complex hyperplane $H\subset{\mathbb P}^n$ and such that $H\cap\G'=\varnothing$. Note that after such transformation $\G'$ remains compact in ${\mathbb C}^n$. Then we may also assume that $(Q_p\cap V)\not\subset \L_1$. Thus holomorphic extension along $Q_a$ defines $f$ on a non-empty set in $Q_p$. $\square$
Theorem \ref{t2} now follows. Indeed, from Lemmas \ref{l2.1} and \ref{l2.2} we conclude that $E\cap\G=\varnothing$. Since $D$ is bounded, $D\cap E=\varnothing$, and $X\cap(D\times D')$ defines a proper holomorphic correspondence from $D$ to $D'$.
\section{Local Extension.}
To prove Theorem \ref{t1} we first establish local extension of holomorphic correspondences.
\begin{definition} Let $\G$ and $\G'$ be smooth, real-analytic hypersurfaces in ${\mathbb
C}^n$. Let $f: U\to \mathbb C^n$ be a holomorphic correspondence such that $f(U\cap\G)\subset \G'$. Then $f$ is called {\it complete} if for any
$z\in U\cap\G$, $f(Q_z\cap U)\subset Q'_{z'}$ and $f(z)=I_{z'}$. \end{definition}
By the invariance property of Segre varieties, $f(_{z}Q_z)\subset Q_{f(z)}$, where $_{z}Q_z$ is the germ of $Q_z$ at $z$, for any $z\in U\cap \G$. The condition $f(Q_z\cap U)\subset Q'_{z'}$ in the definition is somewhat stronger: it states that every connected component of $Q_z\cap U$ is mapped by $f$ into the same Segre variety. Note that in general Segre varieties are defined only locally, while the set $U$ can be relatively large. In this case the inclusion $f(Q_z\cap U)\subset Q'_{z'}$ should be considered only in a suitable standard pair of neighborhoods of $z$.
The condition $f(z)=I_{z'}$ in the definition above indicates that $f$ has the maximal possible number of branches. It is convenient to establish analytic continuation of complete correspondences, as such continuation does not introduce additional branches.
\begin{lemma}\label{l3.2}\label{l:sublocal} Let $f:U\to {\mathbb C}^n$ be a complete holomorphic correspondence, $f(\G\cap U)\subset \G'$, where $\G=\partial D$ and $\G'=\partial D'$, $D$ and $D'$ are as in Theorem \ref{t1}. Suppose $p\in \partial U\cap\G$ is such that $Q_p\cap U\ne\varnothing$. Then there exists a neighborhood $U_p$ of $p$ such that $f$ extends to a holomorphic correspondence $\hat f: U_p\to {\mathbb C}^n$. \end{lemma}
\noindent{\it Proof.} The proof of this lemma repeats that of Lemma \ref{l2.1}. Let $b\in Q_p\cap U$. Consider a small neighborhood $U_b$ of $b$, $U_b\subset U$, and a neighborhood $U_p$ of $p$ such that for any $z\in U_b$, the set $Q_z\cap U_p$ is non-empty and connected. As before, let $S\subset U$ be the branch locus of $f$, and $\S=\{z\subset U_p: Q_z\cap U_b\subset S\}$. Define \begin{equation}\label{e:A} A=\left\{ (w,w')\in (U_p\setminus\S)\times {\mathbb C}^n : f\left(Q_w\cap U_b\right)\subset Q'_{w'} \right\}. \end{equation} Observe that since $f$ is complete, for any $w\in U \cap\G$, the inclusion $f(Q_w\cap U_b)\subset Q'_{w'}$ implies that $f(Q_w\cap U)\subset Q'_{w'}$. In particular, this holds for any $w$ arbitrary close to $p$. Therefore $A$ is well-defined if the neighborhood $U_p$ is chosen sufficiently small. Analogously to Lemma \ref{l2.1}, one can show that $A$ is a non-empty closed complex analytic set in $(U_p\setminus\S)\times{\mathbb C}^n$. Similar argument also shows that $\S\times{\mathbb C}^n$ is a removable singularity for $A$, and thus $\overline A$ defines a closed-complex analytic set in $U_p\times{\mathbb C}^n$.
Let us show now that $A$ defines a holomorphic correspondence $\hat f:U_p\to U'$, where $U'$ is a suitable neighborhood of $\G'$.
Consider the closure of $A$ in $U_p\times {\mathbb P}^n$. Recall, that since ${\mathbb P}^n$ is compact, the projection $\pi:\overline A \to U_p$ is proper. In particular, $\pi (\overline A)= U_p$. Let $U'$ be a neighborhood of $\G'$ as in Lemma \ref{l-finite}. To simplify the notation, denote the restriction of $\overline A$ to $U_p \times U'$ again by $A$. Let $\pi: A\to U_p$ and $\pi': A\to U'$ be the natural projections, and let $\hat f=\pi'\circ \pi$.
Let $Z={\pi'}^{-1}(\G')$. Then since $f(\G\cap U)\subset \G'$, $\pi^{-1}(\G\cap U\cap U_p)\subset Z$. Therefore there exists at least one irreducible component of $\pi^{-1}(\G\cap U_p)$ which is contained in $Z$. Thus for any $z\in\G\cap U_p$, there exists $z'\in\G'$ such that $z'\in \hat f(z)$. By construction, if $z\in U_p$ and $z'\in \hat f(z)$, then $\hat f(z) = I'_{z'}$. In view of (\ref{e-inG}) we conclude that $\hat f(\G\cap U_p)\subset \G'$. Now the same argument as in Lemma \ref{l2.1} shows that $U_p$ can be chosen so small that $\hat f$ is a holomorphic correspondence. $\square$
\begin{theorem}[Local extension] \label{t:local} Let $D$ and $D'$ be as in Theorem \ref{t1}, $\G=\partial D$ and $\G'=\partial D'$. Let $f:U\to {\mathbb C}^n$ be a complete holomorphic correspondence, such that $f(\G\cap U)\subset \G'$, where $\G\cap U$ is connected, and $\G\cap \partial U$ is a smooth submanifold. Let $p\in\partial U\cap \G$. Then there exists a neighborhood $U_p$ of the point $p$ such that $f$ extends to a holomorphic correspondence $\hat f:U_p\to {\mathbb
C}^n$. Moreover, $\hat f|_{U\cap U_p} = f|_{U\cap U_p}$, and the resulting correspondence $F:U\cup U_p \to {\mathbb C}^n$ is complete. \end{theorem}
\noindent{\it Proof.} We call a point $p\in\partial U\cap \G$ {\it regular}, if $\partial U\cap \G$ is a generic submanifold of $\G$, i.e. $T^c_p(\partial U\cap\G)=n-2$. We prove the theorem in three steps. First we prove the result under the assumption $Q_p\cap U\not=\varnothing$, then for regular points in $\partial U\cap\G$, and finally for arbitrary $p\in \partial U\cap\G$.
{\it Step 1.} Suppose that $Q_p\cap U\not=\varnothing$. Then by Lemma \ref{l:sublocal} $f$ extends as a holomorphic correspondence $\hat f$ to some neighborhood $U_p$ of $p$. It follows from the construction that for any $z\in U\cap U_p$ the number of preimages of $f(z)$ and $\hat f(z)$ is the same. Thus by Lemma~\ref{l-intersection}, $f$ and $\hat f$ define a holomorphic correspondence in $U\cup U_p$. Denote this correspondence by $F$.
We now show that $F$ is also complete in $U\cup U_p$. Since $f$ is complete, for any $z\in U\cap U_p\cap\G$, arbitrarily close to $p$, $f(Q_z\cap U)\subset Q'_{f(z)}$. Thus if $U_p$ is chosen sufficiently small, then for any $z\in U_p\cap\G$, \begin{equation} F(Q_z\cap(U\cup U_p))\subset Q'_{F(z)}. \end{equation} Suppose now that there exists some point $z$ in $(U\cup U_p)\cap\G$ such that not all components of $Q_z\cap (U\cup U_p)$ are mapped by $F$ into the same Segre variety. From the argument above, $z\notin U_p$. Since $U\cap\G$ is connected, there exists a simple smooth curve $\gamma\subset \G\cap U$ connecting $z$ and $p$. By Lemma \ref{l:alongSV} for every point $\zeta\in\g$, the germ of a correspondence $F$ at $\zeta$ extends as a holomorphic correspondence along the Segre variety $Q_\zeta$. Moreover, for $\zeta\in\g$ which are close to $p$, the extension of $F$ along $Q_\zeta$ coincides with the correspondence $f$ in $U$ (even if $Q_\zeta\cap U$ is disconnected). Since $\cup_{\zeta\in\g}Q_\zeta$ is connected, this property holds for all $\zeta\in\g$. The extension of $F$ along $Q_\zeta$ clearly maps $Q_\zeta$ into $Q'_{F(\zeta)}$, and therefore $Q_\zeta\cap U$ is mapped by $f$ into the same Segre variety. But this contradicts the assumption that the components of $Q_z\cap U$ are mapped into different Segre varieties. This shows that $F$ is also a complete correspondence.
{\it Step 2.} Suppose now that $Q_p\cap U=\varnothing$, but $p$ is a regular point. Then by \cite{s1} Prop. 4.1, there exists a point $a\in U$ such that $Q_a\cap Q_p\not=\varnothing$. We now apply Lemma \ref{l:alongSV} to extend the germ of the correspondence at $a$ along $Q_a$. We note that such extension along $Q_a$ may not in general define a complete correspondence, since apriori $Q_a\cap \G$ may be disconnected from $U\cap\G$. Let $\L$ be as in Lemma \ref{l:alongSV}. Then after performing, if necessary, a linear-fractional transformation in the target space, we can find a point $b\in Q_p\cap Q_a$, such that $b\notin \L$. Let $U_b$ be a small neighborhood of $b$ such that $U_b\cap\L=\varnothing$ and $f$ extends to $U_b$ as a holomorphic correspondence $f_b$. Then for any $z\in U\cap\G$ such that $Q_z\cap U_b\not=\varnothing$, the sets $f(Q_z\cap U)$ and $f_b(Q_z\cap U_b)$ are contained in the same Segre variety. Indeed, if not, then we can connect $a$ and $z$ by a smooth path $\g\subset\G\cap U$ and apply the argument that we used to prove completeness of $F$ in Step 1.
Now the same proof as in Step 1 shows that $f$ extends as a holomorphic correspondence to some neighborhood of $p$, and that the resulting extension is also complete.
{\it Step 3.} Suppose now that $p\in \partial U\cap \G$ is not a regular point. Let \begin{equation} M=\left\{z\in\partial U\cap\G: T_z(\partial U\cap\G)=T^c_z(\G)\right\}. \end{equation} It is easy to see that $M$ is a locally real-analytic subset of $\G$. Moreover, since $\G$ is essentially finite, $\dim M < 2n-2$. Choose the coordinate system such that $p=0$ and the defining function of $\G$ is given in the so-called normal form (see \cite{cm}): \begin{equation}
\rho(z,\overline z)= 2x_n+\sum_{|k|,|l|\ge
1}\rho_{k,l}(y_n)('z)^k(\overline{'z})^l, \end{equation} where $'z=(z_1,\dots,z_{n-1})$. Since the extendability of $f$ through regular points is already established, after possibly an additional change of variables, we may assume that $f$ extends as a holomorphic correspondence to the points $\{z\in\partial U\cap\G: x_1>0\}$. Let $L_c$ denote the family of real hyperplanes in the form $\{z\in {\mathbb C}^n : x_1=c\}$. Then there exists $\epsilon>0$ such for any $c\in[-\epsilon,\epsilon]$, \begin{equation}\label{good-c} T^c_z(\G) \not= T_z(L_c\cap\G), {\rm \ \ for\ any\ } z\in L_c\cap\G\cap {\mathbb B}(0,\epsilon). \end{equation}
Let $\Omega_{c,\delta}$ be the intersection of $\G$, the $\delta$-neighborhood of $x_1$-axis and the set bounded by $L_c$ and $L_{c+\delta}$, that is \begin{equation} \Omega_{c,\delta}=\left\{ z\in\G\cap {\mathbb B}(0,\epsilon) :
c<x_1<c+\delta, \ y_1^2+\sum_{j=2}^n|z_j|^2<\delta \right\}. \end{equation}
Then there exist $\delta>0$ and $c>0$, such that $f$ extends as a holomorphic correspondence to a neighborhood of the set $\Omega_{c,\delta}$. Since $L_c\cap\G$ consists only of regular points, from Steps 1 and 2 we conclude that $f$ extends to a neighborhood of any point in $L_c\cap\G$ that belongs to the boundary of $\Omega_{c,\delta}$. Let $c_0$ be the smallest number such that $f$ extends past $L_c\cap\G$. Then from (\ref{good-c}) and previous steps, $c_0<0$, and therefore, $f$ extends to a neighborhood of the origin. $\square$
\section{Proof of Theorem \ref{t1}.}
The proof of the local equivalence of boundaries $\partial D$ and $\partial D'$ is equivalent to that of Theorem \ref{t2}.
To prove the theorem in the other direction let us first show that a germ of a biholomorphic map $f:U\to {\mathbb C}^n$, $f(U\cap\G)\subset\G'$ can be replaced by a complete correspondence. Without loss of generality we assume that $0\in U\cap \G$. We choose a neighborhood $U_0$ of the origin and shrink $U$ in such a way, that $Q_w\cap U$ is non-empty and connected for any $w\in U_0$. Define \begin{equation}\label{comp} A=\left\{ (w,w')\in U_0\times{\mathbb C}^n: f(Q_w\cap U)\subset Q'_{w'} \right\}. \end{equation} Then (\ref{comp}) defines a holomorphic correspondence, which in particular contains the germ of the graph of $f$ at the origin. Let $\hat f$ be the multiple valued mapping corresponding to $A$. Then by construction and from (\ref{e-inG}), $w'\in\hat f(w)$ implies $\hat f(w)=I'_{w'}$. Thus $\hat f$ is a complete correspondence.
If $\partial U_0$ is smooth, then By Theorem \ref{t:local} we can locally extend $\hat f$ along $\G$ past the boundary of $U_0\cap\G$ to a larger open set $\Omega$. However, local extension in general does not imply that $A$ is a closed set in $\Omega\times{\mathbb C}^n$. Indeed, there may exist a point $p\in \G\cap\partial \Omega$ such that for any sufficiently small neighborhood $V$ of $p$, $\Omega\cap V\cap\G$ consists of two connected components, say $\G_1$ and $\G_2$. Then local extension from $\G_1$ to a neighborhood of $p$ may not coincide with the correspondence $\hat f$ defined in $\G_2$. Therefore, local extension past the boundary of $\Omega\cap\G$ does not lead to a correspondence defined globally in a neighborhood of $\overline\Omega\times \mathbb C^n$. Note that this cannot happen if $\Omega$ is sufficiently small.
We now show that $\hat f$ extends analytically along any path on $\G$.
\begin{lemma}\label{along-paths} Let $\g:[0,1]\to\G$ be a simple curve without self-intersections, and $\g(0)=0$. Then there exist a neighborhood $V$ of $\g$ and a holomorphic correspondence $F:V\to\mathbb C^n$ which extends $\hat f$. \end{lemma}
\noindent{\it Proof.} Suppose that $\hat f$ does not extend along $\g$. Then let $\zeta\in\gamma$ be the first point to which $\hat f$ does not extend. Let $\epsilon_0>0$ be so small that $\mathbb B(\zeta,\epsilon)\cap\G$ is connected and simply connected for any $\epsilon \le \epsilon_0$. Choose a point $z\in B(\zeta,\epsilon_0/2)\cap \g$ to which $\hat f$ extends. Let $\delta$ be the largest positive number such that $\hat f$ extends holomorphically to $\mathbb B(z,\delta)\cap\G$. By Theorem \ref{t:local} $\hat f$ extends to a neighborhood of every point in $\partial\mathbb B(z,\delta)\cap\G$. Moreover, if $\mathbb B(z,\delta)\subset \mathbb B(\zeta,\epsilon_0)$, then the extension of $\hat f$ is a closed complex analytic set. Thus $\delta>\epsilon/2$. This shows that $\hat f$ also extends to $\zeta$, and therefore extends along $\g$. $\square$
Note that analytic continuation of $\hat f$ along $\g$ in Lemma \ref{along-paths} always yields a complete correspondence.
The Monodromy theorem cannot be directly applied for multiple valued mappings, and we need to show that analytic continuation is independent of the choice of a curve connecting two points on $\G$.
\begin{lemma} Suppose that $\g\subset\G$ is a Jordan curve $\g(0)=\g(1)=0$. Let $F$ be the holomorphic correspondence defined near the origin and obtained by analytic continuation of $\hat f$ along $\g$. Then $F=\hat f$ in some neighborhood of the origin. \end{lemma}
\noindent{\it Proof.} Since $\G$ is simply connected and compact, there exists $\epsilon_0 >0$ such that for any $z\in\G$, $\mathbb B(z,\epsilon)\cap\G$ is connected and simply connected for any $\epsilon\le\epsilon_0$.
Let $\phi$ be the homotopy map, that is $\phi(t,\tau):I\times I\to\G$, $\phi(t,0)=\g(t)$, $\phi(t,1)\equiv 0$, $I=[0,1]$. Let $\{(t_j,\tau_k)\in I\times I$, $j,k=0,1,2,\dots,m\}$ be the set of points satisfying:
\begin{enumerate} \item[(i)] $t_0=\tau_0=0$, $t_m=\tau_m=1$, \item[(ii)] $\{\phi(t,\tau_k): t_j\le t \le t_{j+1}\} \subset \mathbb B(\phi(t_{j},\tau_{k}),\epsilon_0/2)$, \\ $\{\phi(t_{j},\tau): \tau_k \le \tau \le \tau_{k+1}\} \subset \mathbb B(\phi(t_{j},\tau_{k}),\epsilon_0/2)$, for any $j,k<m$. \end{enumerate}
Suppose that $f$ is a complete holomorphic correspondence defined in a ball $B$ of small radius centered at $\phi(tj,\tau_k)\in\G$. By Theorem \ref{t:local}, $f$ extends holomorphically past every boundary point of $\partial B\cap\G$. Since $B(\phi(t_j,\tau_k),\epsilon_0)$ is connected and simply connected, $f$ extends at least to a ball of radius $\epsilon_0/2$. Consider the closed path $\g_{j,k}=\{\phi(t,\tau_k): t_j\le t \le t_{j+1}\}\cup \{\phi(t_{j+1},\tau): \tau_k\le \tau \le \tau_{k+1}\}\cup \{\phi(t,\tau_{k+1}): t_j\le t \le t_{j+1}\}\cup \{\phi(t_{j},\tau): \tau_k\le \tau\le \tau_{k+1}\},$ where the second and fourth pieces are traversed in the opposite direction. Then $\g_{j,k}$ is entirely contained in $\mathbb (B(\phi(tj,\tau_k),\epsilon_0/2)$. Therefore, analytic continuation of $f$ along $\g_{j,k}$ defines the same correspondence at $\phi(t_j,\tau_k)$.
Analytic continuation of $\hat f$ along $\g$ can be reduced to continuation along paths $\g_{j,k}$. Since continuation along each path $\g_{j,k}$ does not introduce new branches of $\hat f$, $F=\hat f$. $\square$
Thus simple connectivity of $\G$ implies that the process of local extension of $\hat f$ leads to a global extension of $\hat f$ to some neighborhood of $\G$. Since $\hat f(\G)\subset \G'$, there exist neighborhoods $U$ of $\G$ and $U'$ of $\G'$ such that $\hat f: U\to U'$ is a proper holomorphic correspondence. Let $A$ be the analytic set corresponding to $\hat f$. By (\ref{e-canon}) there exist functions
$\phi_{IJ}$ holomorphic in $U$ such that $A$ is determined from the system $\sum_{|J|\le m}\phi_{IJ}(z){z'}^J=0$. By Hartog's theorem all $\phi_{IJ}$ extend holomorphically to a neighborhood of $\overline D$, (recall that $\G=\partial D$). This defines a proper holomorphic correspondence $f:D\to D'$. $\square$
\begin{small}
\end{small}
\end{document}
|
arXiv
|
{
"id": "0201247.tex",
"language_detection_score": 0.76568204164505,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Implementation: The conjugacy problem in right-angled Artin groups} \author{\small{\uppercase{Gemma Crowe, Michael Jones}}} \date{} \maketitle \begin{abstract}
In 2009, Crisp, Godelle and Wiest constructed a linear-time algorithm to solve the conjugacy problem in right-angled Artin groups. This algorithm has now been implemented in Python, and the code is freely available on \href{https://github.com/gmc369/Conjugacy-problem-RAAGs}{GitHub}. This document provides a summary of how the code works. As well as determining whether two elements $w_{1}, w_{2}$ are conjugate in a RAAG $A_{\Gamma}$, our code also returns a conjugating element $x \in A_{\Gamma}$ such that $w_{1} = x^{-1}w_{2}x$, if $w_{1}$ and $w_{2}$ are conjugate. \end{abstract}
\section{Introduction} In their paper `\textit{The conjugacy problem in subgroups of right-angled Artin groups}', Crisp, Godelle and Wiest created a linear-time algorithm to solve the conjugacy problem in right-angled Artin groups (RAAGs) \cite{CGW}. This algorithm has now been implemented as a Python program, and is freely available on \href{https://github.com/gmc369/Conjugacy-problem-RAAGs}{GitHub}. \par This document provides an overview of the code. We recommend the reader first takes a look at the original paper \cite{CGW}, to understand the key tools and steps of this algorithm. \begin{rmk} Our Python code requires the \texttt{networkx} module. You may also need to install Pillow and nose. Details on how to install these modules can be found here: \url{https://networkx.org/documentation/stable/install.html}. \end{rmk}
\subsection{Notation} Throughout we let $A_{\Gamma}$ denote a RAAG with defining graph $\Gamma$. We let $N$ denote the number of vertices in the generating set, i.e. the size of the standard generating set for $A_{\Gamma}$. We assume generators commute if and only if there does not exist an edge between corresponding vertices in $\Gamma$, to match with the convention used in \cite{CGW}. \par Examples are provided both in the code as well as this document to assist the user. Throughout this summary, we will often use the RAAG defined in Example 2.4 from \cite{CGW}, which has the following presentation: \begin{equation}\label{CGW example}
A_{\Gamma} = \langle a_{1}, a_{2}, a_{3}, a_{4} \; | \; [a_{1}, a_{4}] = 1, [a_{2}, a_{3}] = 1, [a_{2}, a_{4}] = 1 \rangle. \end{equation} This RAAG is defined by the following graph: \begin{figure}
\caption{Defining graph $\Gamma$}
\label{fig:my_label}
\end{figure}
\section{Summary of algorithm} We summarise the key steps of the algorithm from \cite{CGW}. For further details on piling representations, see Chapter 14 of \cite{OfficeHour}. \par \underline{ALGORITHM}: Conjugacy problem in RAAGs. \par \textbf{Input}: \begin{enumerate}
\item RAAG $A_{\Gamma}$: number of vertices of the defining graph, and list of commuting generators.
\item Words $v,w$ representing group elements in $A_{\Gamma}$. \end{enumerate} \textbf{Step 1: Cyclic reduction} \begin{adjustwidth}{1.5cm}{}
Produce the piling representation $\pi(v)$ of the word $v$, and apply cyclic reduction to $\pi(v)$ to produce a cyclically reduced piling $p$. Repeat this step for the word $w$ to get a cyclically reduced piling $q$. \end{adjustwidth} \textbf{Step 2: Factorisation} \begin{adjustwidth}{1.5cm}{}
Factorise each of the pilings $p$ and $q$ into non-split factors. If the collection of subgraphs do not coincide, \textbf{Output} = \texttt{\color{blue}False}.\\
Otherwise, continue to Step 3. \end{adjustwidth} \textbf{Step 3: Compare non-split factors} \begin{adjustwidth}{1.5cm}{}
If $p = p^{(1)}\dots p^{(k)}$ and $q = q^{(1)}\dots q^{(k)}$ are the factorisations found in Step 2, then for each $i = 1, \dots, k$, do the following:
\begin{adjustwidth}{1.5cm}{}
\begin{enumerate}[label=(\roman*)]
\item Transform the non-split cyclically reduced pilings $p^{(i)}$ and $q^{(i)}$ into pyramidal pilings $\Tilde{p}^{(i)}$ and $\Tilde{q}^{(i)}$.
\item Produce the words representing these pilings in cyclic normal form $\sigma^{\ast}\left(\Tilde{p}^{(i)}\right)$ and $\sigma^{\ast}\left(\Tilde{q}^{(i)}\right)$.
\item Decide whether $\sigma^{\ast}\left(\Tilde{p}^{(i)}\right)$ and $\sigma^{\ast}\left(\Tilde{q}^{(i)}\right)$ are equal up to a cyclic permutation. If not, \textbf{Output} = \texttt{\color{blue}False}.
\end{enumerate}
\end{adjustwidth}
\textbf{Output} = \texttt{\color{blue}True}. \end{adjustwidth}
\begin{exmp} We provide an example of how to implement this algorithm with our Python code, using the RAAG $A_{\Gamma}$ defined in Equation \ref{CGW example}. Consider the following two words: \[ w_{1} = a^{-2}_{2}a^{-1}_{4}a_{3}a_{2}a_{4}a_{1}a_{2}a^{-1}_{1}a^{2}_{2}a^{-1}_{4}, \quad w_{2} = a_{4}a_{3}a^{-1}_{4}a_{2}a_{1}a_{2}a^{-1}_{1}a^{-1}_{4}. \] These represent conjugate elements in $A_{\Gamma}$, since \[ w_{1} = \left(a^{2}_{4}a^{2}_{2}\right)^{-1}\cdot w_{2} \cdot \left(a^{2}_{4}a^{2}_{2}\right). \] To check this, we input the following code. See \cref{sec:inputs} on how to input words and commutation relations from a RAAG. \begin{align*}
\texttt{w\_{1}} &= \texttt{[-2,-2,-4,3,2,4,1,2,-1,2,2,-4]}\\
\texttt{w\_{2}} &= \texttt{[4,3,-4,2,1,2,-1,-4]} \\
\texttt{CGW\_Edges} &= \texttt{[(1,4), (2,3), (2,4)]}\\
&\texttt{is\_conjugate(w\_1, w\_2, 4, CGW\_Edges)} \end{align*} The output from this is \[ \texttt{\color{blue}True}, \texttt{[-2, -2, -4, -4]} \] The first argument is a \texttt{\color{blue}True} or \texttt{\color{blue}False} statement of whether $w_{1}$ and $w_{2}$ are conjugate. The second argument returns a conjugating element $x$ such that $w_{1} = x^{-1}w_{2}x$. \end{exmp} It is important to note that the order of the input words determines the conjugating element. In particular, when we input $w_{1}$ following by $w_{2}$, we obtain a conjugator $x$ such that $w_{1} = x^{-1}w_{2}x$. If we reverse the order of $w_{1}$ and $w_{2}$, the program will return the inverse of this conjugator, since $w_{2} = xw_{1}x^{-1}$.
\section{Inputs}\label{sec:inputs} \subsection{Words} The approach of using symbols $a,b,c...$ to represent letters in a word is limiting, since there are only finitely many letters in the alphabet. Instead, we use the positive integers to represent the generators, and the negative integers to represent their inverses: \begin{center}
\begin{tabular}{|c|c||c|c|}
\hline
Generator & Representation & Generator & Representation\\
\hline\hline
$a$ & \texttt{1} & $a^{-1}$ & \texttt{-1}\\
\hline
$b$ & \texttt{2} & $b^{-1}$ & \texttt{-2}\\
\hline
$c$ & \texttt{3} & $c^{-1}$ & \texttt{-3}\\
\hline
... & ... & ... & ... \\
\hline \end{tabular} \end{center} This also matches with our convention of ordering which will be shortlex, i.e. \[ 1 < -1 < 2 < -2 < \dots \] We note this convention is the opposite of the normal form convention in \cite{CGW}. \par Words in $A_{\Gamma}$ are represented by lists of integers. The following table gives some examples. Here $\varepsilon$ denotes the empty word. \begin{center}
\begin{tabular}{|c|c|}
\hline
Word & Representation\\
\hline\hline
$abc$ & \texttt{[1,2,3]}\\
\hline
$\varepsilon$ & \texttt{[]}\\
\hline
$a^{-2}c^3b^{-1}$ & \texttt{[-1,-1,3,3,3,-2]}\\
\hline
$a^{100}b^{-1}$ & \texttt{([1]*100)+[-2]}\\
\hline
$abc...xyz$ & \texttt{\color{pu}list\color{black}(\color{pu}range\color{black}(1,27))}\\
\hline \end{tabular} \end{center}
\subsection{Commuting generators} In many of the functions in our code, we need to define which generators commute in $A_{\Gamma}$. This is represented by a list of tuples, each of which represents a pair of commuting generators. Here are some examples: \begin{center}
\begin{tabular}{|c|c|}
\hline
Group & Commuting generators list\\
\hline\hline
$\mathbb{F}_n$ & \texttt{[]}\\
\hline
$\mathbb{Z}^2$ & \texttt{[(1,2)]}\\
\hline
$\mathbb{F}_2\times\mathbb{Z}$ & \texttt{[(1,3),(2,3)]}\\
\hline
$\mathbb{Z}^n$ & \texttt{[(i,j) \color{orange}for \color{black}i \color{orange}in \color{pu}range\color{black}(1,n) \color{orange}for \color{black}j \color{orange}in \color{pu}range\color{black}(i+1,n+1)]}\\
\hline \end{tabular} \end{center} By convention, we do not include edges $(n,n)$, and we order our list by shortlex ordering. For example, if $(3,4)$ is in our list, we do not need to also include $(4,3)$. Also, the code does not require inverses of generators to be taken into account. In particular, for every tuple $(n,m)$ in the list, we do not need to also include $(-n, m), (n, -m)$ or $(-n, -m)$. \par This list should contain precisely one tuple for each non-edge in the defining graph. If no list is entered, the program takes \texttt{[]} as default (i.e. the free group).
\begin{rmk} In the Python code, we use the command \texttt{commuting\_elements} to denote commuting generators. \end{rmk}
\section{Pilings} Pilings are represented by lists of lists. Each list within the main list corresponds to a column of the piling, reading from bottom to top. A `$+$' bead is represented by \texttt{1}, a `$-$' bead is represented by \texttt{-1} and a `$0$' bead is represented by \texttt{0}. Here are some examples: \begin{center}
\begin{tabular}{|c|c|}
\hline
Piling & Representation\\
\hline\hline
\includegraphics[scale=0.1]{piling2.png} & \texttt{[[0],[1,0,-1],[0,1,0]]}\\
\hline
\includegraphics[scale=0.1]{piling3.png} & \texttt{[[1,0,-1,0],[0,1,0,1],[0,0],[]]}\\
\hline
\includegraphics[scale=0.1]{commutative_group_piling.png} & \texttt{[[-1],[1],[1],[],[1]]}\\
\hline \end{tabular} \end{center} We note the empty piling is represented by \texttt{[[]$\ast N$]} where $N$ denotes the number of generators. \par Reading pilings as a list of lists is not as user friendly as a pictorial representation. We refer the reader to Appendix \ref{appen:draw}, which explains how to use the \texttt{draw\_piling()} function. This allows the user to create a pictorial representation of pilings.
\subsection{Programs} One of the most useful steps of the algorithm is converting words, which represent group elements in a RAAG, into piling representations. One key fact from these constructions is that if two words $u,v \in A_{\Gamma}$ represent the same group element, then the piling representations for $u$ and $v$ will be the same. In particular, every group element in $A_{\Gamma}$ is uniquely determined by its piling. We describe this function here. \par \textbf{Function:} \texttt{piling(w, N, commuting\_elements=[])} \par \textbf{Input}: \begin{enumerate}
\item Word $w$ representing a group element in $A_{\Gamma}$.
\item $N=$ number of vertices in defining graph.
\item List of commuting generators. \end{enumerate} \textbf{Output}: Piling representation of $w$, as a list of lists. \par By construction, the piling will reduce any trivial cancellations in $w$ after shuffling, so we do not require our input word $w$ to be reduced. \begin{exmp} Suppose you wish to compute the piling for the word $ab^{2}a^{-1}b\in\mathbb{F}_2$. Then you would input \texttt{piling([1,2,2,-1,2], 2)}, which outputs the following: \[ \texttt{[[1, 0, 0, -1, 0], [0, 1, 1, 0, 1]]} \] Similarly we could compute the piling for the word $a^{-2}_{2}a^{-1}_{4}a_{3}a_{2}a_{4}a_{1}a_{2}a^{-1}_{1}a^{2}_{2}a^{-1}_{4} \in A_{\Gamma}$ from Equation \ref{CGW example}. This outputs the following piling: \[ \texttt{[[0, 0, 1, 0, -1, 0, 0], [-1, 0, 1, 0, 1, 1], [0, 1, 0, 0], [-1, 0]]} \] \begin{figure}
\caption{$ab^{2}a^{-1}b\in\mathbb{F}_2$ and $a^{-2}_{2}a^{-1}_{4}a_{3}a_{2}a_{4}a_{1}a_{2}a^{-1}_{1}a^{2}_{2}a^{-1}_{4} \in A_{\Gamma}$ }
\label{fig:ex pilings}
\end{figure} Figure \ref{fig:ex pilings} gives a pictorial representation of these pilings. Each list represents a column in the piling. \end{exmp} We now present a function which computes the normal form representative from a piling. For us, this is the shortlex shortest word which represents the piling. We remind the reader that this convention is the opposite ordering for normal forms in \cite{CGW}.
\textbf{Function:} {\texttt{word(p, commuting\_elements=[])}} \par \textbf{Input}: \begin{enumerate}
\item $p$ = piling.
\item List of commuting generators. \end{enumerate} \textbf{Output}: normal form word which represents the piling $p$.
\begin{exmp}
The following is an example of how to use this function for the following piling in $A_{\Gamma} = \langle a,b,c \; | \; ac=ca\rangle$: \begin{align*}
&\texttt{p=[[1,0],[0,0,-1],[-1,0]]}\\
&\texttt{w=word(p,[(1,3)])}\\
&\texttt{\color{pu}print\color{black}(w)} \end{align*} The output is: \begin{center}
\texttt{[1,-3,-2]} \end{center} \end{exmp} \section{Cyclic reduction} After constructing pilings from our input words, the next step of the conjugacy algorithm is to cyclically reduce each piling. \par \textbf{Function:} {\texttt{cyclically\_reduce(p, commuting\_elements=[])}} \par \textbf{Input}: \begin{enumerate}
\item $p =$ piling.
\item List of commuting generators. \end{enumerate} \textbf{Output}: \begin{enumerate}
\item Cyclically reduced piling.
\item Conjugating element. \end{enumerate} \begin{exmp}
Let $w = abca^{-1}\in\langle a,b,c\; | \; bc=cb\rangle$. We can compute all possible cyclic reductions on $w$ as follows: \begin{align*}
&\texttt{p=piling([1,2,3,-1], 3,[(2,3)])}\\
&\texttt{p\_cr=cyclically\_reduce(p,[(2,3)])}\\
&\texttt{w=word(p\_cr[0],[(2,3)])}\\
&\texttt{\color{pu}print\color{black}(w, p\_cr[1])} \end{align*} The output from this is: \begin{center}
\texttt{[2,3], [1]} \end{center} This function is the first example where we output a conjugating element. In this example, when we cyclically reduce $w = a\cdot bc \cdot a^{-1}$ to the word $bc$, we have taken out the conjugating element $a$. \end{exmp}
\section{Graphs} We remind the reader that for the functions in this section, we need to import the \texttt{networkx} module into Python. This makes working with the defining graph much easier, in particular computing induced subgraphs and connected components. \par The next step in the conjugacy algorithm is to compute the non-split factors from a word based on the defining graph. First, we need a method to input our defining graph. \par \textbf{Function:} {\texttt{graph\_from\_edges(N, commuting\_edges=[])}} \par \textbf{Input}: \begin{enumerate}
\item $N =$ number of vertices in defining graph.
\item List of commuting generators. \end{enumerate} \textbf{Output}: \texttt{networkx} graph, which represents the defining graph. \par We recall that in \cite{CGW}, the convention is to add edges for non-commuting vertices. This is precisely what this function does - the output is a graph on $N$ vertices, with edges between vertices $(x,y)$ if and only if $(x,y)$ is not in the list of commuting generators. \begin{exmp}\label{exmp: graph} Let's compute the defining graph from Equation \ref{CGW example}. We input the following: \begin{align*}
&\texttt{CGW\_Edges = [(1,4), (2,3), (2,4)]} \\
&\texttt{g = graph\_from\_edges(4, CGW\_Edges)} \\
&\texttt{nx.draw(g)} \end{align*} This outputs the following graph: \begin{figure}
\caption{\texttt{networkx} graph defined in Example 6.1.}
\label{fig:my_label2}
\end{figure} \end{exmp} \subsection{Factorising} We now want to build up a function which takes our defining graph $g$ and our input word $w$, and computes the non-split factors related to $w$. \par \textbf{Function:} {\texttt{factorise(g, p)}} \par \textbf{Input}: \begin{enumerate}
\item $g =$ \texttt{networkx} graph representing the defining graph of $A_{\Gamma}$.
\item $p =$ piling. \end{enumerate} \textbf{Output}: List of subgraphs of $g$, one for each non-split factor of $p$. \par This function first checks which columns in the piling contain a \texttt{1} or \texttt{-1} bead (this is the \texttt{supp\_piling()} function). From this subset of vertices, we build the induced subgraph, and then return a list of the connected components of this subgraph. Each of these connected components represents the non-split factors. \par At this stage, we have extracted each of the subgraphs which correspond to the factors. We now want to extract pilings which represent each factor. \comm{ \par \textbf{Function:} {\texttt{graphs\_to\_nsfactor(g, w)}} \par \textbf{Input}: \begin{enumerate}
\item $g =$ connected component of defining graph.
\item $w =$ word representing a group element of $A_{\Gamma}$. \end{enumerate} \textbf{Output}: non-split factor of $w$ based on $g$.} \par \textbf{Function:} {\texttt{graphs\_to\_nsfactors(l, w, N, commuting\_elements)}} \par \textbf{Input}: \begin{enumerate}
\item $l =$ list of subgraphs representing connected components.
\item $w =$ word representing a group element of $A_{\Gamma}$.
\item $N =$ number of vertices in defining graph.
\item List of commuting generators. \end{enumerate} \textbf{Output}: list of the non-split factors from $w$ represented by pilings. \begin{exmp} We use the example from Figure 4 of \cite{CGW}. We input the following: \begin{align*}
&\texttt{w = [2,3,-4]}\\
&\texttt{CGW\_Edges = [(1,4), (2,3), (2,4)]} \\
&\texttt{g = graph\_from\_edges(4, CGW\_Edges)}\\
&\texttt{graphs = factorise(g, piling(w, 4, CGW\_Edges))} \\
&\texttt{\color{pu}print\color{black}(graphs\_to\_nsfactors(graphs, w, 4, CGW\_Edges))} \end{align*} The output is \[ \texttt{[[[0], [1], [], []], [[0], [], [1, 0], [0, -1]]]} \] where the first piling represents the factor $a_{2}$, and the second piling represents the factor $a_{3}a^{-1}_{4}$. \end{exmp}
\section{Pyramidal Pilings} For the final step in the algorithm, we need a function which converts pilings into pyramidal pilings. After this, we need to check when two pyramidal pilings are equal up to cyclic permutation. \par \textbf{Function:} {\texttt{pyramidal\_decomp(p, commuting\_elements=[])}} \par \textbf{Input}: \begin{enumerate}
\item $p = $ non-split piling.
\item List of commuting generators. \end{enumerate} \textbf{Output}: factors $p_{0}, p_{1}$ as pilings. \par It is important to note that the input piling must be non-split, otherwise we cannot achieve a pyramidal piling and the function will run forever. \begin{exmp} Let's compute the decomposition in Figure 5 of \cite{CGW}. We input the following: \begin{align*}
&\texttt{CGW\_Edges = [(1,4), (2,3), (2,4)]} \\
&\texttt{p = [[0,1,0,-1,0], [0,1,0,1], [0,1,0,0], [-1,0]]} \\
&\texttt{decomp = pyramidal\_decomp(p, CGW\_Edges)}\\
&\texttt{\color{pu}print\color{black}(decomp)} \end{align*} The output is then \[ \texttt{([[0], [], [0, 1], [-1, 0]], [[1, 0, -1, 0], [0, 1, 0, 1], [0, 0], []])} \] \end{exmp}
\textbf{Function:} {\texttt{pyramidal(p, N, commuting\_elements=[])}} \par \textbf{Input}: \begin{enumerate}
\item $p = $ non-split cyclically reduced piling.
\item $N = $ number of vertices in defining graph.
\item List of commuting generators. \end{enumerate} \textbf{Output}: \begin{enumerate}
\item Pyramidal piling of $p$.
\item Conjugating element. \end{enumerate} The \texttt{pyramidal()} function iteratively applies \texttt{pyramidal\_decomp()} until the factor $p_{0}$ is empty. Since the only operation we apply on the piling is cyclic permutations, we can add these steps to our conjugating element. \begin{exmp} Again we take the same example above. Our input is: \begin{align*}
&\texttt{CGW\_Edges = [(1,4), (2,3), (2,4)]} \\
&\texttt{p = [[0,1,0,-1,0], [0,1,0,1], [0,1,0,0], [-1,0]]} \\
&\texttt{pyr = pyramidal(p, 4, CGW\_Edges)} \end{align*} Figure \ref{fig:pyr} gives the pictorial representation of this pyramidal piling. The conjugating element output in this example is \texttt{\color{blue}[-4,3,-4]}. \begin{figure}
\caption{Pyramidal piling.}
\label{fig:pyr}
\end{figure} \end{exmp}
\textbf{Function:} {\texttt{is\_cyclic\_permutation(w,v)}} \par \textbf{Input}: Two words $w,v$ representing group elements in $A_{\Gamma}$.\\ \textbf{Output}: If $w$ is a cyclic permutation of $v$, the function outputs two arguments: \begin{enumerate}
\item \texttt{\color{blue}True}
\item Conjugating element. \end{enumerate} Otherwise, return \texttt{\color{blue}False}.
\section{Implementation of the Conjugacy Problem} \subsection{Simple Cases} If we want to solve the conjugacy problem in either $\mathbb{F}_n$ or $\mathbb{Z}^n$, the algorithm for implementing the conjugacy problem is more straightforward, and using the \texttt{is\_conjugate()} function detailed below is wasteful. Certainly in the latter case, the conjugacy problem is trivial; it is equivalent to checking if two words are equal, which is just a case of comparing the number of occurrences of each letter in the words. For free groups, two cyclically reduced words are conjugate if and only if they are cyclic permutations of each other. Hence one can make a shorter program for free groups by simply using the \texttt{cyclically\_reduce()} and \texttt{is\_cyclic\_permutation()} functions. \par For general RAAGs, we now have the necessary functions to implement the conjugacy problem. \par \textbf{Function:} {\texttt{is\_conjugate(w1, w2, N, commuting\_elements=[])}} \par \textbf{Input}: \begin{enumerate}
\item Two words $w_{1}, w_{2}$ representing group elements in $A_{\Gamma}$.
\item $N = $ number of vertices in defining graph.
\item List of commuting generators. \end{enumerate} \textbf{Output}: If $w_{1}$ is conjugate to $w_{2}$ in $A_{\Gamma}$ then return: \begin{enumerate}
\item \texttt{\color{blue}True}
\item Conjugating element $x$ such that $w_{1} = x^{-1}w_{2}x$. \end{enumerate} Otherwise, return \texttt{\color{blue}False}. \par We note that when computing the conjugating element, we find a reduced word $x$ which represents an element which conjugates $w_{1}$ to $w_{2}$.
\appendix \section{The `draw piling' Function}\label{appen:draw} The representation of pilings as a list of lists of \texttt{1}s, \texttt{-1}s and \texttt{0}s is not very readable in the Python interpreter. For small pilings it is tolerable, but for larger ones the \texttt{draw\_piling()} function is useful. \par \noindent\texttt{draw\_piling()} can take a total of nine arguments, however only the first one, which is the piling to be drawn, is needed. The function returns nothing, but by default it will show a picture of the piling in a new window, as well as save a \texttt{PNG} file of it. \par The following table lists all the arguments of the \texttt{draw\_piling()} function and what they do: \begin{center}
\begin{tabular}{|c|c|c|}
\hline
Argument & Type & Description\\
\hline\hline
\texttt{piling} & Piling & \tiny The (mandatory) piling to draw.\\
\hline
\texttt{scale} & Float & \tiny The scale to draw the piling at. Default is \texttt{100.0}.\\
\hline
\texttt{plus\_colour} & Colour & \tiny The colour to draw the `$+$' beads. Default is red.\\
\hline
\texttt{zero\_colour} & Colour & \tiny The colour to draw the `$0$' beads. Default is grey.\\
\hline
\texttt{minus\_colour} & Colour & \tiny The colour to draw the `$-$' beads. Default is blue.\\
\hline
\texttt{anti\_aliasing} & Integer & \tiny The super-sampling resolution for anti-aliasing.\\
& & \tiny Only allows positive integers. Default is 4.\\
\hline
\texttt{filename} & String & \tiny The filename to save the piling as. Default is \texttt{\color{gr}"piling.png"}.\\
\hline
\texttt{show} & Boolean & \tiny Whether to show the piling in a window. Default is \texttt{\color{orange}True}.\\
\hline
\texttt{save} & Boolean & \tiny Whether to save the piling. Default is \texttt{\color{orange}True}.\\
\hline \end{tabular} \end{center}
\section{Solution to the Word Problem} With the tools that \texttt{pilings.py} provides, it is not hard to solve the Word Problem in any given RAAG. The following code defines a function that decides if a given word is equal to the identity: \begin{align*}
\texttt{\color{orange}def} \; &\texttt{\color{blue}identity\color{black}(w, N, commuting\_elements=[]):}\\
&\texttt{\indent p=piling(w, N, commuting\_elements)\color{red}\#generate reduced piling}\\
&\texttt{\indent reduced\_w=word(p, commuting\_elements)\color{red}\#read off reduced word}\\
&\texttt{\indent \color{orange}return\color{black}(reduced\_w==[])\color{red}\#return whether it equals the identity} \end{align*} The following code defines a function that decides if two given words are equal in a given RAAG: \begin{align*}
\texttt{\color{orange}def} \; &\texttt{\color{blue}equal\color{black}(w1, w2, N, commuting\_elements=[]):}\\
&\texttt{\indent \color{red}\#generate reduced pilings}\\
&\texttt{\indent p1=piling(w1, N, commuting\_elements)}\\
&\texttt{\indent p2=piling(w2, N, commuting\_elements)}\\
&\texttt{\indent \color{red}\#read off reduced words}\\
&\texttt{\indent reduced\_w1=word(p1, commuting\_elements)}\\
&\texttt{\indent reduced\_w2=word(p2, commuting\_elements)}\\
&\texttt{\indent \color{red}\#return whether they are equal}\\
&\texttt{\indent \color{orange}return\color{black}(reduced\_w1==reduced\_w2)} \end{align*}
\section*{Acknowledgments} The second author would like to thank the Department of Mathematics at Heriot-Watt University for supporting this summer project. Both authors would like to thank Laura Ciobanu for supervising this project. \par Whilst we have made our best efforts to debug and correctly implement code for this algorithm, there may still be mistakes! If you spot any errors or issues with our code, please let us know.
\uppercase{School of Mathematical and Computer Sciences, Heriot-Watt University, Edinburgh, Scotland, EH14 4AS} \par Email address: \texttt{[email protected], [email protected]}
\end{document}
|
arXiv
|
{
"id": "2305.06636.tex",
"language_detection_score": 0.6958787441253662,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[Vanishing cohomology]{Vanishing cohomology and Betti bounds for complex projective hypersurfaces}
\author[L. Maxim ]{Lauren\c{t}iu Maxim} \address{L. Maxim : Department of Mathematics, University of Wisconsin-Madison, 480 Lincoln Drive, Madison WI 53706-1388, USA} \email {[email protected]} \thanks{L. Maxim is partially supported by the Simons Foundation Collaboration (Grant \#567077) and by the Romanian Ministry of National Education (CNCS-UEFISCDI grant PN-III-P4-ID-PCE-2020-0029).}
\author[L. P\u{a}unescu ]{Lauren\c{t}iu P\u{a}unescu} \address{L. P\u{a}unescu: Department of Mathematics, University of Sydney, Sydney, NSW, 2006, Australia} \email {[email protected]}
\author[M. Tib\u{a}r]{Mihai Tib\u{a}r} \address{M. Tib\u{a}r : Universit\' e de Lille, CNRS, UMR 8524 -- Laboratoire Paul Painlev\'e, F-59000 Lille, France} \email {[email protected]} \thanks{M. Tib\u{a}r acknowledges the support of the Labex CEMPI (ANR-11-LABX-0007). }
\keywords{singular projective hypersurface, vanishing cycles, vanishing cohomology, Betti numbers, Milnor fiber, Lefschetz hyperplane theorem}
\subjclass[2010]{32S30, 32S50, 55R55, 58K60}
\date{\today}
\begin{abstract} We employ the formalism of vanishing cycles and perverse sheaves to introduce and study the vanishing cohomology of complex projective hypersurfaces. As a consequence, we give upper bounds for the Betti numbers of projective hypersurfaces, generalizing those obtained by different methods by Dimca in the isolated singularities case, and by Siersma-Tib\u{a}r in the case of hypersurfaces with a $1$-dimensional singular locus. We also prove a supplement to the Lefschetz hyperplane theorem for hypersurfaces, which takes the dimension of the singular locus into account, and we use it to give a new proof of a result of Kato. \end{abstract}
\maketitle
\section{Introduction. Results}\label{intro} Let $V=\{f=0\} \subset {\mathbb C} P^{n+1}$ be a reduced complex projective hypersurface of degree $d$, with $n \geq 1$. By the classical Lefschetz Theorem, the inclusion map $j:V \hookrightarrow {\mathbb C} P^{n+1}$ induces cohomology isomorphisms \begin{equation}\label{one} j^k:H^k( {\mathbb C} P^{n+1};{\mathbb Z}) \overset{\cong}\longrightarrow H^k(V;{\mathbb Z}) \ \ \text{for all} \ \ k<n, \end{equation}
and a primitive monomorphism for $k=n$ (e.g., see \cite[Theorem 5.2.6]{Di}). Moreover, if $s=\dim V_{\rm sing}<n$ is the complex dimension of the singular locus of $V$ (with $\dim \emptyset=-1$), then Kato \cite{Ka} showed that (see also \cite[Theorem 5.2.11]{Di}) \begin{equation}\label{two} H^k(V;{\mathbb Z}) \cong H^k( {\mathbb C} P^{n+1};{\mathbb Z}) \ \ \text{for all} \ \ n+s+2\leq k\leq 2n, \end{equation} and the homomorphism $j^k$ induced by inclusion is given in this range (and for $k$ even) by multiplication by $d=\deg(V)$. It therefore remains to study the cohomology groups $H^k(V;{\mathbb Z})$ for $n \leq k \leq n+s+1$.
In the case when $V\subset {\mathbb C} P^{n+1}$ is a {\it smooth} degree $d$ hypersurface, the above discussion yields that $H^k(V;{\mathbb Z}) \cong H^k( {\mathbb C} P^{n};{\mathbb Z})$ for all $k \neq n$. This is in fact the only information we take as an input in this note (it also suffices to work with \eqref{one}, its homology counterpart, and Poincar\'e duality). The Universal Coefficient Theorem also yields in this case that $H^n(V;{\mathbb Z})$ is free abelian, and its rank $b_n(V)$ can be easily deduced from the formula for the Euler characteristic of $V$ (e.g., see \cite[Proposition 10.4.1]{M}): \begin{equation}\label{chi} \chi(V)=(n+2)-\frac{1}{d} \big[1+(-1)^{n+1}(d-1)^{n+2}\big]. \end{equation} Specifically, if $V\subset {\mathbb C} P^{n+1}$ is a smooth degree $d$ projective hypersurface, one has: \begin{equation}\label{bsm} b_n(V)=\frac{(d-1)^{n+2}+(-1)^{n+1}}{d}+\frac{3(-1)^n+1}{2}. \end{equation}
The case when $V$ has only isolated singularities was studied by Dimca \cite{Di0,Di}, (see also \cite{Mi} and \cite{ST}) while projective hypersurfaces with a one-dimensional singular locus have been more recently considered by Siersma-Tib\u{a}r \cite{ST}.
In the singular case, let us fix a Whitney stratification ${\mathcal V}$ of $V$ and consider a one-parameter smoothing of degree $d$, namely $$V_t:=\{f_t=f-tg=0\}\subset {\mathbb C} P^{n+1} \ \ (t \in {\mathbb C}),$$ for $g$ a general polynomial of degree $d$. Here, the meaning of ``general'' is that the hypersurface $W:=\{g=0\}$ is smooth and transverse to all strata in the stratification ${\mathcal V}$ of $V$.
Then, for $t \neq 0$ small enough, all the $V_t$ are smooth and transverse to the stratification ${\mathcal V}$. Let $$B=\{f=g=0\}$$ be the base locus (axis) of the pencil. Consider the incidence variety $$V_D:=\{(x,t)\in {\mathbb C} P^{n+1} \times D \mid x \in V_t \}$$ with $D$ a small disc centered at $0 \in {\mathbb C}$ so that $V_t$ is smooth for all $t \in D^*:=D\setminus \{0\}$. Denote by $\pi:V_D \to D$ the proper projection map, and note that $V=V_0=\pi^{-1}(0)$ and $V_t=\pi^{-1}(t)$ for all $t \in D^*$. In what follows we write $V$ for $V_0$ and use $V_t$ for a smoothing of $V$ (i.e., with $t \in D^*$). In this setup, one can define the Deligne vanishing cycle complex of the family $\pi$, see \cite[Section 10.3]{M} for a quick introduction. More precisely, one has a bounded constructible complex $$\varphi_\pi \underline{{\mathbb Z}}_{V_D} \in D^b_c(V)$$ on the hypersurface $V$, whose hypercohomology groups fit into a long exact sequence (called the {\it specialization sequence}): \begin{equation}\label{spec} \cdots \longrightarrow H^k(V;{\mathbb Z}) \overset{sp^k}{\longrightarrow} H^k(V_t;{\mathbb Z}) \overset{\alpha^k}{\longrightarrow} {\mathbb H}^k(V; \varphi_\pi \underline{{\mathbb Z}}_{V_D}) \longrightarrow H^{k+1}(V;{\mathbb Z}) \overset{sp^{k+1}}{\longrightarrow} \cdots \end{equation} The maps $sp^k$ are called the {\it specialization} morphisms, while the $\alpha^k$'s are usually referred to as the {\it canonical} maps. For any integer $k$, we define $$H^k_\varphi(V):={\mathbb H}^k(V; \varphi_\pi \underline{{\mathbb Z}}_{V_D})$$ and call it the {\it $k$-th vanishing cohomology group of $V$}. This is an invariant of $V$, i.e., it does not depend on the choice of a particular smoothing of degree $d$ (since all smooth hypersurfaces of a fixed degree are diffeomorphic). By its very definition, the vanishing cohomology measures the difference between the topology of a given projective hypersurface $V$ and that of a smooth hypersurface of the same degree.
\begin{rem}\label{r1} Since the incidence variety $V_D=\pi^{-1}(D)$ deformation retracts to $V=\pi^{-1}(0)$, and the {\it specialization map} $sp^k:H^k(V;{\mathbb Z}) \to H^k(V_t;{\mathbb Z})$ of \eqref{spec} factorizes as $$H^k(V;{\mathbb Z}) \overset{\cong}{\longrightarrow} H^k(V_D;{\mathbb Z}) \longrightarrow H^k(V_t;{\mathbb Z})$$ with $H^k(V_D;{\mathbb Z}) \to H^k(V_t;{\mathbb Z})$ induced by the inclusion map, it follows readily that the vanishing cohomology of $V$ can be identified with the relative cohomology of the pair $(V_D,V)$, i.e., \begin{equation} H^k_\varphi(V) \cong H^{k+1}(V_D,V_t;{\mathbb Z}).\end{equation} In particular, the groups $H^k_\varphi(V)$ are the cohomological version of the {\it vanishing homology groups} $$H_k^{\curlyvee}(V):=H_{k}(V_D,V_t;{\mathbb Z})$$ introduced and studied in \cite{ST} in special situations. For the purpose of computing Betti numbers of projective hypersurfaces, the two ``vanishing'' theories yield the same answer, but additional care is needed to handle torsion when computing the actual integral cohomology groups. \end{rem}
Our first result gives the concentration degrees of the vanishing cohomology of a projective hypersurface in terms of the dimension of the singular locus. \begin{thm}\label{th1} Let $V \subset {\mathbb C} P^{n+1}$ be a reduced complex projective hypersurface with $s=\dim V_{\rm sing}$ the complex dimension of its singular locus. Then \begin{equation} H^k_\varphi(V) \cong 0 \ \ \text{ for all integers} \ \ k \notin [n, n+s]. \end{equation} Moreover, $H^n_\varphi(V)$ is a free abelian group. \end{thm}
In view of Remark \ref{r1}, one gets by Theorem \ref{th1} and the Universal Coefficient Theorem the concentration degrees of the vanishing homology groups $H_k^{\curlyvee}(V)$ of a projective hypersurface in terms of the dimension of its singular locus: \begin{cor}\label{corh} With the above notations and assumptions, we have that \begin{equation} H_k^{\curlyvee}(V) \cong 0 \ \ \text{ for all integers} \ \ k \notin [n+1, n+s+1]. \end{equation} Moreover, $H_{n+s+1}^{\curlyvee}(V)$ is free. \end{cor}
\begin{rem} In the case when the projective hypersurface $V \subset {\mathbb C} P^{n+1}$ has a $1$-dimensional singular locus, it was shown in \cite[Theorem 4.1]{ST} that $H_k^{\curlyvee}(V) \cong 0$ for all $k \neq n+1, n+2$. Moreover, Theorem 6.1 of \cite{ST} shows that in this case one also has that $H_{n+2}^{\curlyvee}(V)$ is free. So, Corollary \ref{corh} provides a generalization of the results of \cite{ST} to projective hypersurfaces with arbitrary singularities. Nevertheless, the methods used in its proof are fundamentally different from those in \cite{ST}. \end{rem}
As a consequence of Theorem \ref{th1}, the specialization sequence \eqref{spec} together with the fact that the integral cohomology of a smooth projective hypersurface is free, yield the following result on the integral cohomology of a complex projective hypersurface (where the estimate on the $n$-th Betti number uses formula \eqref{bsm}): \begin{cor}\label{corgen} Let $V \subset {\mathbb C} P^{n+1}$ be a degree $d$ reduced projective hypersurface with a singular locus $V_{\rm sing}$ of complex dimension $s$. Then: \begin{itemize} \item[(i)] $H^k(V;{\mathbb Z}) \cong H^k(V_t;{\mathbb Z}) \cong H^k({\mathbb C} P^n;{\mathbb Z})$ \ for all integers $k \notin [n, n+s+1]$. \item[(ii)] $H^n (V;{\mathbb Z}) \cong \ker(\alpha^n)$ is free. \item[(iii)] $H^{n+s+1} (V;{\mathbb Z}) \cong H^{n+s+1} ({\mathbb C} P^n;{\mathbb Z}) \oplus \mathop{{\mathrm{coker}}}\nolimits(\alpha^{n+s})$. \item[(iv)] $H^k(V;{\mathbb Z}) \cong \ker(\alpha^k) \oplus \mathop{{\mathrm{coker}}}\nolimits(\alpha^{k-1})$ for all integers $k \in [n+1,n+s]$, $s\ge 1$.
\end{itemize} In particular, $$b_n(V) \leq b_n(V_t)=\frac{(d-1)^{n+2}+(-1)^{n+1}}{d}+\frac{3(-1)^n+1}{2},$$ and $$b_{k}(V) \leq \mathop{{\mathrm{rank}}}\nolimits \ H^{k-1}_{\varphi}(V) + b_{k}({\mathbb C} P^{n}) \ \ \text{ for all integers} \ k \in [n+1,n+s+1], \ s\ge 0.$$ \end{cor}
The homological version of the specialisation sequence \eqref{spec} identifies to the long exact sequence of the pair $(V_{D}, V_{t})$, namely: \[ \cdots \to H_{k+1}(V_{t};{\mathbb Z}) \to H_{k+1}(V_{D};{\mathbb Z}) \to H_{k+1}^{\curlyvee}(V;{\mathbb Z}) \ \overset{\alpha_{k}}{\longrightarrow} \ H_{k}(V_{t};{\mathbb Z}) \to \cdots \] The inclusions $V_{t}\hookrightarrow V_{D}\hookrightarrow {\mathbb C} P^{n+1}\times D$ induce in homology a commutative triangle, where $H_{k}(V_{t};{\mathbb Z}) \to H_{k}({\mathbb C} P^{n+1}\times D;{\mathbb Z})$ is injective for $k\not= n$ (by the Lefschetz Theorem for $k<n$, and it is multiplication by $d$ for $k>n$, see e.g. Remark \ref{Katoh} for the homological version of the proof of Theorem \ref{Kato}). This shows that the morphisms $H_{k}(V_{t};{\mathbb Z}) \to H_{k}(V_{D};{\mathbb Z})$ is also injective for all $k\not= n$, and therefore $\alpha_{k} =0$ for $k\not= n$. Consequently, the above long exact sequence splits into a $5$-term exact sequence, and short exact sequences: \begin{equation}\label{sp1} \begin{split}
0 \to H_{n+1}(V_{t};{\mathbb Z}) \to H_{n+1}(V;{\mathbb Z}) \to H_{n+1}^{\curlyvee}(V;{\mathbb Z}) \stackrel{\alpha_{n}}{\to}
H_{n}(V_{t};{\mathbb Z}) \rightarrow H_{n}(V;{\mathbb Z}) \to 0. \\
0 \to H_{k}(V_{t};{\mathbb Z}) \to H_{k}(V_{D};{\mathbb Z}) \to H_{k}^{\curlyvee}(V;{\mathbb Z}) \to 0 \ \ \ \ \ \ \ \ \mbox{ for } \ k\ge n+1. \end{split} \end{equation} We then get the following homological version of Corollary \ref{corgen}(i-iv), with the same upper bounds for Betti numbers, but with an interesting improvement for (iii) and (iv) showing more explicitly the dependence of the homology of $V$ on the vanishing homology groups:
\begin{cor}\label{corgenhom} Let $V \subset {\mathbb C} P^{n+1}$ be a degree $d$ reduced projective hypersurface with a singular locus $V_{\rm sing}$ of complex dimension $s$. Then:
\begin{itemize} \item[(i')] $H_{k}(V;{\mathbb Z}) \cong H_{k}(V_t;{\mathbb Z}) \cong H_{k}({\mathbb C} P^n;{\mathbb Z})$ \ for all $k\le n-1$ and all $k\ge n+s+2$. \item[(ii')] $H_{n}(V;{\mathbb Z}) \cong \mathop{{\mathrm{coker}}}\nolimits(\alpha_{n})$. \item[(iii')] $H_{n+1}(V;{\mathbb Z}) \cong \ker(\alpha_{n}) \oplus H_{n+1}({\mathbb C} P^n;{\mathbb Z})$. \item[(iv')] $H_{k}(V;{\mathbb Z}) \cong H_k^{\curlyvee}(V;{\mathbb Z}) \oplus H_{k}({\mathbb C} P^n;{\mathbb Z})$, for all $ n+2 \le k \le n+s+1$, whenever $s\ge 1$, \\ and $H_{n+s+1}(V;{\mathbb Z})$ is free. \end{itemize} \end{cor}
The ranks of the (possibly non-trivial) vanishing (co)homology groups can be estimated in terms of the local topology of singular strata and of their generic transversal types by making use of the hypercohomology spectral sequence. Such estimates can be made precise for hypersurfaces with low-dimensional singular loci.
Concretely, as special cases of Corollaries \ref{corgen} and \ref{corgenhom}, in Section \ref{bounds} we recast Siersma-Tib\u{a}r's \cite{ST} result for $s\le 1$, and in particular Dimca's \cite{Di0,Di} computation for $s=0$. Concerning the estimation of the rank of the highest interesting (co)homology group, we prove the following general result:
\begin{thm}\label{th2} Let $V \subset {\mathbb C} P^{n+1}$ be a degree $d$ reduced projective hypersurface with a singular locus $V_{\rm sing}$ of complex dimension $s$. For each connected stratum $S_i \subseteq V_{\rm sing}$ of top dimension $s$ in a Whitney stratification of $V$, let $F_i^\pitchfork$ denote its transversal Milnor fiber with corresponding Milnor number $\mu_i^\pitchfork$. Then: \begin{equation}\label{bt} b_{n+s+1}(V) \leq 1+ \sum_i \mu_i^\pitchfork, \end{equation}
and the inequality is strict for $n+s$ even. \end{thm}
In fact, the inequality in \eqref{bt} is deduced from
\begin{equation}\label{btb} b_{n+s+1}(V) \leq 1+\mathop{{\mathrm{rank}}}\nolimits \ H^{n+s}_{\varphi}(V),\end{equation} together with \begin{equation} \mathop{{\mathrm{rank}}}\nolimits \ H^{n+s}_{\varphi}(V) \leq \sum_i \mu_i^\pitchfork, \end{equation} and the inequality \eqref{btb} is strict for $n+s$ even. For further refinements of Theorem \ref{th2}, see Remark \ref{rem31}. Note also that if $s=0$, i.e., $V$ has only isolated singularities, then $\mu_i^\pitchfork$ is just the usual Milnor number of such a singularity of $V$.
Let us remark that if the projective hypersurface $V \subset {\mathbb C} P^{n+1}$ has singularities in codimension $1$, i.e., $s=n-1$, then $b_{n+s+1}(V)=b_{2n}(V)=r$, where $r$ denotes the number of irreducible components of $V$. Indeed, in this case, one has (e.g., see \cite[(5.2.9)]{Di}): \begin{equation}\label{top} H^{2n}(V;{\mathbb Z}) \cong {\mathbb Z}^{r}.\end{equation} In particular, Theorem \ref{th2} yields the following generalization of \cite[Corollary 7.6]{ST}: \begin{cor} If the reduced projective hypersurface $V \subset {\mathbb C} P^{n+1}$ has singularities in codimension $1$, then the number $r$ of irreducible components of $V$ satisfies the inequality: \begin{equation} r \leq 1+\sum_i \mu_i^\pitchfork. \end{equation} \end{cor}
\begin{rem}\label{fr} Note that if the projective hypersurface $V \subset {\mathbb C} P^{n+1}$ is a rational homology manifold, then the Lefschetz isomorphism \eqref{one} and Poincar\'e duality over the rationals yield that $b_i(V)=b_i({\mathbb C} P^n)$ for all $i \neq n$. Moreover, $b_n(V)$ can be deduced by computing the Euler characteristic of $V$, e.g., as in \cite[Section 10.4]{M}.\end{rem}
The computation of Betti numbers of a projective hypersurface which is a rational homology manifold can be deduced without appealing to Poincar\'e duality by using the vanishing cohomology instead, as the next result shows: \begin{prop}\label{p1} If the projective hypersurface $V \subset {\mathbb C} P^{n+1}$ is a ${\mathbb Q}$-homology manifold, then $H^k_{\varphi}(V) \otimes {\mathbb Q} \cong 0$ for all $k \neq n$. In particular, in this case one gets: $b_i(V)=b_i(V_t)=b_i({\mathbb C} P^n)$ for all $i \neq n$, and $b_n(V)=b_n(V_t)+\mathop{{\mathrm{rank}}}\nolimits H^n_{\varphi}(V)$. \end{prop}
At this end, we note that Corollary \ref{corgen}(i) reproves Kato's isomorphism \eqref{two} about the integral cohomology of $V$, by using only the integral cohomology of a smooth hypersurface (for this it suffices to rely only on the Lefschetz isomorphism \eqref{one}, its homological version, and Poincar\'e duality). In Section \ref{supLHT}, we give a new proof of Kato's result (see Theorem \ref{Kato}), which relies on the following supplement to the Lefschetz hyperplane section theorem for hypersurfaces, which may be of independent interest: \begin{thm}\label{thapi} Let $V \subset {\mathbb C} P^{n+1}$ be a reduced complex projective hypersurface with $s=\dim V_{\rm sing}$ the complex dimension of its singular locus. (By convention, we set $s=-1$ if $V$ is nonsingular.) Let $H \subset {\mathbb C} P^{n+1}$ be a generic hyperplane. Then \begin{equation}\label{34api} H^k(V,V\cap H; {\mathbb Z})=0 \ \ \text{for} \ \ k < n \ \ \text{and} \ \ n+s+1 < k < 2n. \end{equation} Moreover, $H^{2n}(V,V \cap H; {\mathbb Z})\cong{\mathbb Z}^r$, where $r$ is the number of irreducible components of $V$, and $H^{n}(V,V \cap H; {\mathbb Z})$ is (torsion-)free. \end{thm}
Note that the vanishing \eqref{34api} for $k<n$ is equivalent to the classical Lefschetz hyperplane section theorem. The proof of \eqref{34api} for $n+s+1 < k < 2n$ reduces to understanding the homotopy type of the complement of a smooth affine hypersurface transversal to the hyperplane at infinity; see \cite[Corollary 1.2]{Lib} for such a description. Homological counterparts of Theorem \ref{thapi} and of Kato's result are also explained in Section \ref{supLHT}, see Corollary \ref{corap} and Remark \ref{Katoh}.
Finally, let us note that similar techniques apply to the study of Milnor fiber cohomology of complex hypersurface singularity germs. This is addressed by the authors in the follow-up paper \cite{MPT} (see also \cite{ST0} for the case of $1$-dimensional singularities).
\noindent{\bf Acknowledgements.} L. Maxim thanks the Sydney Mathematical Research Institute (SMRI) for support and hospitality, and J\"org Sch\"urmann for useful discussions.
\section{Concentration degrees of vanishing cohomology}
The proof of Theorem \ref{th1} makes use of the formalism of perverse sheaves and their relation to vanishing cycles, see \cite{Di1,M} for a brief introduction.
\subsection{Proof of Theorem \ref{th1}} By definition, the incidence variety $V_D$ is a complete intersection of pure complex dimension $n+1$. It is non-singular if $V=V_0$ has only isolated singularities, but otherwise it has singularities where the base locus $B=V\cap W$ of the pencil $\{f_t\}_{t\in D}$ intersects the singular locus $\Sigma:=V_{\rm sing}$ of $V$.
If $\underline{{\mathbb Z}}_{V_D}$ denotes the constant sheaf with stalk ${\mathbb Z}$ on the complete intersection $V_D$, a result of L\^e \cite{Le} implies that the complex $\underline{{\mathbb Z}}_{V_D}[n+1]$ is a perverse sheaf on $V_D$. It then follows that $\varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]$ is a ${\mathbb Z}$-perverse sheaf on $\pi^{-1}(0)=V$ (see, e.g., \cite[Theorem 10.3.13]{M} and the references therein).
Recall that the stalk of the cohomology sheaves of $\varphi_\pi \underline{{\mathbb Z}}_{V_D}$ at a point $x \in V$ are computed by (e.g., see \cite[(10.20)]{M}): \begin{equation} {\mathcal H}^j(\varphi_\pi \underline{{\mathbb Z}}_{V_D})_x \cong H^{j+1}(B_{x}, B_{x}\cap V_t;{\mathbb Z}), \end{equation} where $B_{x}$ denotes the intersection of $V_D$ with a sufficiently small ball in some chosen affine chart ${\mathbb C}^{n+1} \times D$ of the ambient space ${\mathbb C} P^{n+1} \times D$ (hence $B_x$ is contractible). Here $B_{x}\cap V_t=F_{\pi,x}$ is the Milnor fiber of $\pi$ at $x$. Let us now consider the function $$h=f/g:{\mathbb C} P^{n+1} \setminus W \to {\mathbb C}$$ where $W:=\{g=0\}$, and note that $h^{-1}(0)=V\setminus B$ with $B=V\cap W$ the base locus of the pencil. If $x \in V \setminus B$, then in a neighborhood of $x$ one can describe $V_t$ ($t \in D^*$) as $$\{x \mid f_t(x)=0\}=\{x \mid h(x)=t\},$$ i.e., as the Milnor fiber of $h$ at $x$. Note also that $h$ defines $V$ in a neighborhood of $x \notin B$. Since the Milnor fiber of a complex hypersurface singularity germ does not depend on the choice of a local equation (e.g., see \cite[Remark 3.1.8]{Di}), we can therefore use $h$ or a local representative of $f$ when considering Milnor fibers (of $\pi$) at points in $V \setminus B$. From here on we will use the notation $F_x$ for the Milnor fiber of the hypersurface singularity germ $(V,x)$, and we note for future reference that the above discussion also yields that $F_x$ is a manifold, which moreover is contractible if $x \in V \setminus B$ is a smooth point.
It was shown in \cite[Proposition 5.1]{PP} (see also \cite[Proposition 4.1]{MSS} or \cite[Lemma 4.2]{ST}) that there are no vanishing cycles along the base locus $B$, i.e., \begin{equation} \varphi_\pi \underline{{\mathbb Z}}_{V_D} \vert_B \simeq 0.\end{equation} Therefore, if $u:V\setminus B \hookrightarrow V$ is the open inclusion, we get that \begin{equation}\label{s6} \varphi_\pi \underline{{\mathbb Z}}_{V_D} \simeq u_! u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}. \end{equation} Since pullback to open subvarieties preserves perverse sheaves, we note that $u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]$ is a perverse sheaf on the {\it affine} variety $V \setminus B$. Artin's vanishing theorem for perverse sheaves (e.g., \cite[Corollary 6.0.4]{Sc}) then implies that:
\begingroup \allowdisplaybreaks \begin{equation}\label{s1} \begin{split}
H^k_\varphi(V)
& := {\mathbb H}^{k}(V; \varphi_\pi \underline{{\mathbb Z}}_{V_D}) \\
& \cong {\mathbb H}^{k-n}(V; \varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]) \\
& \cong {\mathbb H}^{k-n}(V; u_! u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]) \\
& \cong {\mathbb H}_c^{k-n}(V \setminus B; u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]) \\ & \cong 0 \end{split} \end{equation} \endgroup
for all $k-n<0$, or equivalently, for all $k<n$.
Contractibility of Milnor fibers at smooth points of $V \setminus B$ implies that the support of $\varphi_\pi \underline{{\mathbb Z}}_{V_D}$ is in fact contained in $\Sigma \setminus B$, with $\Sigma$ denoting as before the singular locus of $V$. In particular, if $v:\Sigma \setminus B \hookrightarrow V\setminus B$ is the closed inclusion, then \begin{equation}\label{s5} u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D} \simeq v_!v^*u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}.\end{equation} Next, consider the composition of inclusion maps $$\Sigma \setminus B \overset{q}{\hookrightarrow}\Sigma \overset{p}{\hookrightarrow} V$$ with $p\circ q=u \circ v$. By using \eqref{s6} and \eqref{s5}, we get: \begingroup \allowdisplaybreaks \begin{equation}\label{s3} \begin{split} \varphi_\pi \underline{{\mathbb Z}}_{V_D} & \simeq u_!v_!v^* u^*\varphi_\pi \underline{{\mathbb Z}}_{V_D} \\ & \simeq (u\circ v)_! (u\circ v)^* \varphi_\pi \underline{{\mathbb Z}}_{V_D} \\ & \simeq (p\circ q)_! (p\circ q)^* \varphi_\pi \underline{{\mathbb Z}}_{V_D} \\ & \simeq p_!q_!q^*p^* \varphi_\pi \underline{{\mathbb Z}}_{V_D} \\ &\simeq p_*p^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}, \end{split} \end{equation} \endgroup where the last isomorphism uses the fact that $p^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}$ is supported on $\Sigma \setminus B$, hence $p^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}\simeq q_!q^*p^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}$. Since the support of the perverse sheaf $\varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]$ on $V$ is contained in the closed subset $\Sigma$, we get that $p^*\varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]$ is a perverse sheaf on $\Sigma$ (e.g., see \cite[Corollary 8.2.10]{M}). Since the complex dimension of $\Sigma$ is $s$, the support condition for perverse sheaves together with the hypercohomology spectral sequence yield that $${\mathbb H}^{\ell}(\Sigma; p^*\varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]) \cong 0$$ for all $\ell \notin [-s,s]$. This implies by \eqref{s3} that \begin{equation}\label{s2} H^k_\varphi(V)={\mathbb H}^{k-n}(V; \varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]) \cong {\mathbb H}^{k-n}(\Sigma;p^*\varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]) \cong 0\end{equation} for all $k \notin [n-s,n+s]$.
The desired concentration degrees for the vanishing cohomology is now obtained by combining \eqref{s1} and \eqref{s2}.
Let us finally show that $H^n_\varphi(V)$ is free. Fix a Whitney stratification ${\mathcal V}$ of $V$, so that $V \setminus \Sigma$ is the top stratum. (Note that together with $\pi^{-1}(D^*)$, this also yields a Whitney stratification of $V_D$.) Since $W$ intersects $V$ transversally (i.e., $W$ intersects each stratum $S$ in ${\mathcal V}$ transversally in ${\mathbb C} P^{n+1}$), we can assume without any loss of generality that the base locus $B=V \cap W$ is a closed union of strata of ${\mathcal V}$. Next, we have by \eqref{s1}
that $$ H^n_\varphi(V) \cong {\mathbb H}_c^{0}(V \setminus B; u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]),$$
with $${\mathcal P}:=u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}[n]$$ a ${\mathbb Z}$-perverse sheaf on the affine variety $V \setminus B$ and $u:V \setminus B \hookrightarrow V$ the open inclusion. In particular, this implies that if $S \in {\mathcal V}$ is any stratum in $V \setminus B$ with inclusion $i_S:S \hookrightarrow V \setminus B$ then ${\mathcal H}^k(i_S^!{\mathcal P}) \simeq 0$ for all integers $k<-\dim_{{\mathbb C}} S$. By the Artin-Grothendieck type result of \cite[Corollary 6.0.4]{Sc}, in order to show that ${\mathbb H}^0_c(V\setminus B;{\mathcal P})$ is free it suffices to check that the perverse sheaf ${\mathcal P}$ satisfies the following costalk condition (see \cite[Example 6.0.2(3)]{Sc}):\footnote{We thank J\"org Sch\"urmann for indicating the relevant references to us.} \begin{equation}\label{co1} {\mathcal H}^{-\dim_{{\mathbb C}} S}(i_S^!{\mathcal P})_x \ \text{ is free } \end{equation} for any point $x$ in any stratum $S$ in $V \setminus B$ with inclusion $i_S:S \hookrightarrow V \setminus B$. Let us now fix a stratum $S \in {\mathcal V}$ contained in $V \setminus B$ and let $x \in S$ be a point with inclusion map $k_x:\{x\} \hookrightarrow S$. Consider the composition $i_x:=i_S \circ k_x: \{x\} \hookrightarrow V \setminus B$. Using the fact that $$k_x^*i_S^! \simeq k_x^!i_S^! [2 \dim_{{\mathbb C}} S] \simeq i_x^! [2 \dim_{{\mathbb C}} S]$$ (e.g., see \cite[Remark 6.0.2(1)]{Sc}), the condition \eqref{co1} for $x \in S$ is equivalent to the following: \begin{equation}\label{co2} {\mathcal H}^{\dim_{{\mathbb C}} S}(i_x^!{\mathcal P}) \ \text{ is free}. \end{equation} In fact, the above discussion applies to any algebraically constructible complex ${\mathcal F}^\centerdot \in {^pD}^{\geq 0}$, with $({^pD}^{\leq 0}, {^pD}^{\geq 0})$ denoting the perverse t-structure on $D^b_c(V \setminus B)$. Furthermore, in our setup (i.e., working with PID coefficients and having finitely generated stalk cohomology) ${\mathcal F}^\centerdot \in {^pD}^{\geq 0}$ satisfies the additional costalk condition \eqref{co1} (or, equivalently, \eqref{co2}) if and only if the Verdier dual ${\mathcal D}{\mathcal F}^\centerdot $ satisfies ${\mathcal D}{\mathcal F}^\centerdot \in {^pD}^{\leq 0}$.
Let $i:V=V_0 \hookrightarrow V_D$ denote the closed inclusion, and consider the following {\it variation triangle} for the projection map $\pi:V_D \to D$: \begin{equation}\label{var} i^![1] \longrightarrow \varphi_\pi \overset{var}{\longrightarrow} \psi_\pi \overset{[1]}{\longrightarrow} \end{equation} with $\psi_\pi $ denoting the corresponding nearby cycle functor for $\pi$ (e.g., see \cite[(5.90)]{Sc}). Apply the functor $u^!=u^*$ to the triangle \eqref{var}, and the apply the resulting triangle of functors to the complex $\underline{{\mathbb Z}}_{V_D}[n]$ to get the following triangle of constructible complexes on $V\setminus B$: \begin{equation}\label{var2} {\mathcal Z}:=u^!i^!\underline{{\mathbb Z}}_{V_D}[n+1] \longrightarrow {\mathcal P}:=u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}[n] \longrightarrow {\mathcal R}:=u^* \psi_\pi \underline{{\mathbb Z}}_{V_D}[n] \overset{[1]}{\longrightarrow} \end{equation} Let $x \in S$ be a point in a stratum of $V \setminus B$ with inclusion map $i_x:\{x\} \hookrightarrow V \setminus B$ as before, and apply the functor $i_x^!$ to the triangle \eqref{var2} to get the triangle: \begin{equation}\label{var3} i_x^!{\mathcal Z} \longrightarrow i_x^!{\mathcal P} \longrightarrow i_x^!{\mathcal R} \overset{[1]}{\longrightarrow} \end{equation} The cohomology long exact sequence associated to \eqref{var3} contains the terms $$\cdots \longrightarrow {\mathcal H}^{\dim_{{\mathbb C}} S}(i_x^!{\mathcal Z}) \longrightarrow {\mathcal H}^{\dim_{{\mathbb C}} S}(i_x^!{\mathcal P}) \longrightarrow {\mathcal H}^{\dim_{{\mathbb C}} S}(i_x^!{\mathcal R}) \longrightarrow \cdots$$ Since the category of (torsion-)free abelian groups is closed under extensions, in order to prove \eqref{co2} it suffices to check that ${\mathcal H}^{\dim_{{\mathbb C}} S}(i_x^!{\mathcal Z})$ and ${\mathcal H}^{\dim_{{\mathbb C}} S}(i_x^!{\mathcal R})$ are (torsion-)free. (Note that, in fact, all costalks in question are finitely generated.)
Let us first show that ${\mathcal H}^{\dim_{{\mathbb C}} S}(i_x^!{\mathcal Z})$ is free. Regard the stratum $S$ containing $x$ as a stratum in $V_D$, and let $r_x:\{x\} \to V_D$ be the point inclusion, i.e., $r_x=i\circ u \circ i_x$. So $i_x^!{\mathcal Z}=r_x^!\underline{{\mathbb Z}}_{V_D}[n+1]$. Recall that $\underline{{\mathbb Z}}_{V_D}[n+1]$ is a ${\mathbb Z}$-perverse sheaf on $V_D$, i.e., $\underline{{\mathbb Z}}_{V_D}[n+1] \in {^pD}^{\leq 0}(V_D) \cap{^pD}^{\geq 0}(V_D)$. As already indicated above, in order to show that ${\mathcal H}^{\dim_{{\mathbb C}} S}(r_x^!\underline{{\mathbb Z}}_{V_D}[n+1])$ is free it suffices to verify that ${\mathcal D}(\underline{{\mathbb Z}}_{V_D}[n+1]) \in {^pD}^{\leq 0}(V_D)$, or equivalently, ${\mathcal D}\underline{{\mathbb Z}}_{V_D} \in {^pD}^{\leq -n-1}(V_D)$. This fact is a consequence of \cite[Definition 6.0.4, Example 6.0.11]{Sc}, where it is shown that the complete interesection $V_D$ has a {\it rectified homological depth} equal to its complex dimension $n+1$.
Next note that, due to the local product structure, the Milnor fiber $F_x$ of the hypersurface singularity germ $(V,x)$ with $x \in S$ has the homotopy type of a finite CW complex of real dimension $n-\dim_{{\mathbb C}} S$. In particular, $H_{n-\dim_{{\mathbb C}}S}(F_x;{\mathbb Z})$ is free. Since by the costalk calculation (cf. \cite[(5.92)]{Sc}) and Poincar\'e duality we have for $x\in S$ that \begin{equation} {\mathcal H}^{\dim_{{\mathbb C}} S}(i_x^!{\mathcal R}) \cong H_c^{n+\dim_{{\mathbb C}}S}(F_x;{\mathbb Z}) \cong H_{n-\dim_{{\mathbb C}}S}(F_x;{\mathbb Z}), \end{equation} it follows that ${\mathcal H}^{\dim_{{\mathbb C}} S}(i_x^!{\mathcal R})$ is free. This completes the proof of Theorem \ref{th1}.
\subsection{Proof of Proposition \ref{p1}} Since $V$ is a ${\mathbb Q}$-homology manifold, it follows by standard arguments involving the Hamm fibration (e.g., see \cite[Theorem 3.2.12]{Di}) that $V_D$ is also a ${\mathbb Q}$-homology manifold (with boundary). Thus $\underline{{\mathbb Q}}_{V_D}[n+1]$ is a self-dual ${\mathbb Q}$-perverse sheaf on $V_D$. Moreover, since $\varphi_\pi[-1]$ commutes with the Verdier dualizing functor (see \cite[Theorem 3.1]{Ma} and the references therein), we get that ${\mathcal Q}:=\varphi_\pi\underline{{\mathbb Q}}_{V_D}[n]$ is a Verdier self-dual perverse sheaf on $V$. Using the Universal Coefficients Theorem, we obtain: $$H^k_{\varphi}(V) \otimes {\mathbb Q}={\mathbb H}^{k-n}(V;{\mathcal Q})\cong {\mathbb H}^{k-n}(V;{\mathcal D}{\mathcal Q}) \cong {\mathbb H}^{n-k}(V;{\mathcal Q})^\vee = (H^{2n-k}_{\varphi}(V)\otimes {\mathbb Q})^\vee.$$ The desired vanishing follows now from Theorem \ref{th1}.
\section{Bounds on Betti numbers of projective hypersurfaces}\label{bounds} In this section, we prove Theorem \ref{th2} and specialize it, along with Corollary \ref{corgen}, in the case when the complex dimension $s$ of the singular locus is $\leq 1$.
\subsection{Proof of Theorem \ref{th2}} Let $\Sigma:=V_{\rm sing}$ be the singular locus of $V$, of complex dimension $s$, and fix a Whitney stratification ${\mathcal V}$ of $V$ so that $V \setminus \Sigma$ is the top open stratum. We have by Corollary \ref{corgen} (or by the specialization sequence \eqref{spec}) that $$b_{n+s+1}(V) \leq 1+\mathop{{\mathrm{rank}}}\nolimits \ H^{n+s}_{\varphi}(V).$$ So it suffices to show that \begin{equation}\label{rvg} \mathop{{\mathrm{rank}}}\nolimits \ H^{n+s}_{\varphi}(V) \leq \sum_i \mu_i^\pitchfork, \end{equation} where the summation on the right-hand side runs over the top $s$-dimensional connected strata $S_i$ of $\Sigma$, and $\mu_i^\pitchfork$ denotes the corresponding transversal Milnor number for such a stratum $S_i$.
If $s=0$, an easy computation shows that \eqref{rvg} is in fact an equality, see \eqref{nee} below. Let us next investigate the case when $s\geq 1$.
For any $\ell \leq s$, denote by $\Sigma_{\ell}$ the union of strata in $\Sigma$ of complex dimension $\leq \ell$. In particular, we can filter $\Sigma$ by closed (possibly empty) subsets $$\Sigma=\Sigma_s \supset \Sigma_{s-1} \supset \cdots \supset \Sigma_0 \supset \Sigma_{-1}=\emptyset.$$ Let $$U_\ell:=\Sigma_{\ell} \setminus \Sigma_{\ell-1}$$ be the union of $\ell$-dimensional strata, so $\Sigma_\ell=\sqcup_{k\leq \ell} U_k$. (Here, $\sqcup$ denotes disjoint union.) Recall that the smooth hypersurface $W=\{g=0\}$ was chosen so that it intersects each stratum in $\Sigma$ transversally.
In the notations of the proof of Theorem \ref {th1} , it follows from equations \eqref{s1} and \eqref{s5} that: $$ H^{n+s}_\varphi(V) \cong {\mathbb H}_c^{n+s}(V \setminus B; u^* \varphi_\pi \underline{{\mathbb Z}}_{V_D}) \cong {\mathbb H}_c^{n+s}(\Sigma \setminus B; v^* u^*\varphi_\pi \underline{{\mathbb Z}}_{V_D}),$$ with $B=V \cap W$ the axis of the pencil, and with $v:\Sigma\setminus B \hookrightarrow V \setminus B$ and $u:V \setminus B \hookrightarrow V$ the inclusion maps. We also noted that either $h$ or a local representative of $f$ can be used when considering Milnor fibers of $\pi$ at points in $V \setminus B$. For simplicity, let us use the notation $${\mathcal R} := v^* u^*\varphi_\pi \underline{{\mathbb Z}}_{V_D} \in D^b_c(\Sigma \setminus B),$$ and consider the part of the long exact sequence for the compactly supported hypercohomology of ${\mathcal R}$ associated to the disjoint union $$\Sigma \setminus B= (U_s \setminus B) \sqcup (\Sigma_{s-1} \setminus B)$$ involving $H^{n+s}_{\varphi}(V)$, namely: $$\cdots \to {\mathbb H}^{n+s}_c(U_s \setminus B; {\mathcal R}) \to H^{n+s}_{\varphi}(V) \to {\mathbb H}^{n+s}_c(\Sigma_{s-1} \setminus B; {\mathcal R}) \to \cdots$$
We claim that
\begin{equation}\label{cl1}
{\mathbb H}^{n+s}_c(\Sigma_{s-1} \setminus B; {\mathcal R}) \cong 0,
\end{equation}
so, in particular, there is an epimorphism:
\begin{equation}\label{nee2}{\mathbb H}^{n+s}_c(U_s \setminus B; {\mathcal R}) \twoheadrightarrow H^{n+s}_{\varphi}(V).\end{equation}
In order to prove \eqref{cl1}, consider
the part of the long exact sequence for the compactly supported hypercohomology of ${\mathcal R}$ associated to the disjoint union $$\Sigma_{s-1} \setminus B= (U_{s-1} \setminus B) \sqcup (\Sigma_{s-2} \setminus B)$$ involving $ {\mathbb H}^{n+s}_c(\Sigma_{s-1} \setminus B; {\mathcal R})$, namely: $$\cdots \to {\mathbb H}^{n+s}_c(U_{s-1} \setminus B; {\mathcal R}) \to {\mathbb H}^{n+s}_c(\Sigma_{s-1} \setminus B; {\mathcal R}) \to {\mathbb H}^{n+s}_c(\Sigma_{s-2} \setminus B; {\mathcal R}) \to \cdots$$ We first show that \begin{equation}\label{cl2}{\mathbb H}^{n+s}_c(U_{s-1} \setminus B; {\mathcal R}) \cong 0.\end{equation} Indeed, the $(p,q)$-entry in the $E_2$-term of the hypercohomology spectral sequence computing ${\mathbb H}^{n+s}_c(U_{s-1} \setminus B; {\mathcal R})$ is given by $$E^{p,q}_2=H^p_c(U_{s-1} \setminus B; {\mathcal H}^q( {\mathcal R})),$$ and we are interested in those pairs of integers $(p,q)$ with $p+q=n+s$. Since a point in a $(s-1)$-dimensional stratum of $V$ has a Milnor fiber which has the homotopy type of a finite CW complex of real dimension $n-s+1$, it follows that $${\mathcal H}^q( {\mathcal R}) \vert_{U_{s-1} \setminus B}\simeq 0 \ \ \text{ for any } \ q>n-s+1.$$ Also, by reasons of dimension, we have that $E_2^{p,q}=0$ if $p> 2s-2$. In particular, the only possibly non-trivial entries on the $E_2$-page of the above spectral sequence are those corresponding to pairs $(p,q)$ with $p\leq 2s-2$ and $q\leq n-s+1$, none of which add up to $n+s$. This proves \eqref{cl2}. If $s=1$, this completes the proof of \eqref{cl1} since $\Sigma_{-1}=\emptyset$. If $s>1$, the long exact sequences for the compactly supported hypercohomology of ${\mathcal R}$ associated to the disjoint union $$\Sigma_{\ell} \setminus B= (U_{\ell} \setminus B) \sqcup (\Sigma_{\ell-1} \setminus B),$$ $0 \leq \ell \leq s-1$, can be employed to reduce the proof of \eqref{cl1} to showing that \begin{equation}\label{cl3}{\mathbb H}^{n+s}_c(U_{\ell} \setminus B; {\mathcal R}) \cong 0\end{equation} for all $0 \leq \ell \leq s-1$. To prove \eqref{cl3}, we make use of the hypercohomology spectral sequence whose $E_2$-term is computed by $$E^{p,q}_2=H^p_c(U_{\ell} \setminus B; {\mathcal H}^q( {\mathcal R})),$$ and we are interested again in those pairs of integers $(p,q)$ with $p+q=n+s$. Since a point in an $\ell$-dimensional stratum of $V$ has a Milnor fiber which has the homotopy type of a finite CW complex of real dimension $n-\ell$, it follows that $${\mathcal H}^q( {\mathcal R}) \vert_{U_{\ell} \setminus B}\simeq 0 \ \ \text{ for any } \ q>n-\ell.$$ Moreover, by reasons of dimension, $E_2^{p,q}=0$ if $p> 2\ell$. So the only possibly non-trivial entries on the $E_2$-page are those corresponding to pairs $(p,q)$ with $p\leq 2\ell$ and $q\leq n-\ell$, none of which add up to $n+s$. This proves \eqref{cl3}, and completes the proof of \eqref{cl1} in the general case.
In order to prove \eqref{rvg}, we make use of the epimorphism \eqref{nee2} as follows. Recall that, in our notations, $U_s \setminus B$ is a disjoint union of connected strata $S_i \setminus B$ of complex dimension $s$. Each $S_i \setminus B$ has a generic transversal Milnor fiber $F_i^\pitchfork$, which has the homotopy type of a bouquet of $\mu_i^\pitchfork$ $(n-s)$-dimensional spheres. So the integral cohomology of $F_i^\pitchfork$ in concentrated in degree $n-s$. Moreover, for each $i$, there is a local system ${\mathcal L}_i^\pitchfork$ on $S_i \setminus B$ with stalk $\widetilde{H}^{n-s}(F_i^\pitchfork;{\mathbb Z})$, whose monodromy is usually refered to as the {\it vertical monodromy}. This is exactly the restriction of the constructible sheaf ${\mathcal H}^{n-s}( {\mathcal R})$ to $S_i \setminus B$. It then follows from the hypercohomology spectral sequence computing ${\mathbb H}^{n+s}_c(U_s \setminus B; {\mathcal R}) $ and by Poincar\'e duality that \begin{equation}\label{last}{\mathbb H}^{n+s}_c(U_s \setminus B; {\mathcal R}) \cong \bigoplus_i \ H^{2s}_c(S_i \setminus B;{\mathcal L}_i^\pitchfork) \cong \bigoplus_i \ H_0(S_i \setminus B;{\mathcal L}_i^\pitchfork)\end{equation} which readily gives \eqref{rvg}. $
$ $\square$
\begin{rem}\label{rem31} Note that the upper bound on $b_{n+s+1}(V)$ can be formulated entirely in terms of coinvariants of vertical monodromies along the top dimensional singular strata of $V$. Indeed, if in the notations of the above proof we further let $h_i^v$ denote the vertical monodromy along $S_i \setminus B$, then each term on the right-hand side of \eqref{last} is computed by the coinvariants of $h_i^v$, i.e., $H_0(S_i \setminus B;{\mathcal L}_i^\pitchfork) \cong \widetilde{H}^{n-s}(F_i^\pitchfork;{\mathbb Z})_{h^v_i}.$ Note that the latter statement, when combined with \eqref{nee2}, yields an epimorphism \begin{equation}\label{nee2b} \bigoplus_i \widetilde{H}^{n-s}(F_i^\pitchfork;{\mathbb Z})_{h^v_i} \twoheadrightarrow H^{n+s}_{\varphi}(V),\end{equation} the summation on the left hand side being over the top dimensional singular strata of $V$. One can, moreover, proceed like in \cite{MPT} and give a more precise dependence of all (possibly non-trivial) vanishing cohomology groups $H^{k}_{\varphi}(V)$, $n \leq k \leq n+s$, in terms of the singular strata of $V$. We leave the details to the interested reader. \end{rem}
\subsection{Isolated singularities} Assume that the projective hypersurface $V \subset {\mathbb C} P^{n+1}$ has only isolated singularities (i.e., $s=0$). Then the incidence variety $V_D$ is smooth since the pencil has an empty base locus, and the projection $\pi:V_D \to D$ has isolated singularities exactly at the singular points of $V$.
The only non-trivial vanishing homology group, $H_{n+1}^{\curlyvee}(V)$, is free, and is computed as: \begin{equation}\label{nee} H_{n+1}^{\curlyvee}(V) \cong \bigoplus_{x \in V_{\rm sing}} \widetilde{H}_{n}(F_x;{\mathbb Z}) \cong \bigoplus_{x \in V_{\rm sing}} {\mathbb Z}^{\mu_x}, \end{equation} where $F_x$ denotes the Milnor fiber of the isolated hypersurface singularity germ $(V,x)$, with corresponding Milnor number $\mu_x$. The second isomorphism follows from the fact that $F_x$ has the homotopy type of a bouquet of $\mu_x$ $n$-spheres.
The $5$-term exact sequence \eqref{sp1} then reads as: \begin{equation}\label{speciso} 0 \to H_{n+1}(V_t;{\mathbb Z}) \to H_{n+1}(V;{\mathbb Z}) {\to} \bigoplus_{x \in V_{\rm sing}} \widetilde{H}_{n}(F_x;{\mathbb Z}) \overset{\alpha_{n}}{\to} H_{n}(V_t;{\mathbb Z}) \to H_{n}(V;{\mathbb Z}) \to 0. \end{equation}
Therefore Corollary \ref{corgen}(i)--(iii), together with the following bound via Theorem \ref{th2}:
$$b_{n+1}(V) \leq 1+\sum_{x \in V_{\rm sing}} \mu_x.$$
recover \cite[Proposition 2.2]{ST}, which in turn is a homology counterpart of Dimca's result \cite[Theorem 5.4.3]{Di}.
In fact, Dimca's result was formulated in cohomology, and it is a direct
consequence of the specialization sequence \eqref{spec} via Theorem \ref{th1}, together with the observation that the only non-trivial vanishing cohomology group, $H^n_{\varphi}(V)$, is computed as: \begin{equation}\label{nee} H^n_{\varphi}(V) \cong \bigoplus_{x \in V_{\rm sing}} \widetilde{H}^n(F_x;{\mathbb Z}).\end{equation}
\begin{rem}\label{rem3.3} Let us recall here that if $V\subset {\mathbb C} P^{n+1}$ is a degree $d$ reduced projective hypersurface with only isolated singularities, then its Euler characteristic is computed by the formula (e.g., see \cite[Exercise 5.3.7(i) and Corollary 5.4.4]{Di} or \cite[Proposition 10.4.2]{M}): \begin{equation}\label{chii} \chi(V)=(n+2)-\frac{1}{d} \big[1+(-1)^{n+1}(d-1)^{n+2}\big] +(-1)^{n+1}\sum_{x \in V_{\rm sing}} \mu_x, \end{equation} with $\mu_x$ denoting as before the Milnor number of the isolated hypersurface singularity germ $(V,x)$. In particular, if $V$ is a projective {\it curve} (i.e., $n=1$), then $H_{0}(V;{\mathbb Z}) \cong {\mathbb Z}$, $H_{2}(V;{\mathbb Z})\cong {\mathbb Z}^r$, with $r$ denoting the number of irreducible components of $V$, and $H_{1}(V;{\mathbb Z})$ is a free group whose rank is computed from \eqref{chii} by the formula: \begin{equation}\label{b1}b_1(V)=r+1+d^2-3d-\sum_{x \in V_{\rm sing}} \mu_x.\end{equation} \end{rem}
\subsection{$1$-dimensional singular locus}
This particular case was treated in homology in \cite[Proposition 7.7]{ST}. Let us recall the preliminaries, in order to point out once more that in this paper we have transposed them to a fully general setting.
One starts with $V \subset {\mathbb C} P^{n+1}$, a degree $d$ projective hypersurface with a singular locus $\Sigma:=V_{\rm sing}$ of complex dimension $1$. The singular locus $\Sigma$ consists of a union of irreducible projective curves $\Sigma_i$ and a finite set $I$ of isolated singular points. Each curve $\Sigma_i$ has a generic transversal type of transversal Milnor fiber $F_i^\pitchfork \simeq \bigvee_{\mu_i^\pitchfork} S^{n-1}$ with corresponding transversal Milnor number $\mu_i^\pitchfork$. Each $\Sigma_i$ also contains a finite set $S_i$ of special points of non-generic transversal type. One endows $V$ with the Whitney stratification whose strata are: \begin{itemize} \item the isolated singular points in $I$, \item the special points in $S=\bigcup_i S_i$, \item the (top) one-dimensional components of $\Sigma \setminus S$, \item the open stratum $V \setminus \Sigma$. \end{itemize} The genericity of the pencil $\{V_{t}\}_{t\in D}$ implies that the base locus $B$ intersects each $\Sigma_i$ in a finite set $B_i$ of general points, which are not contained in $I \cup S_i$. The total space $V_D$ of the pencil has in this case only isolated singularities (corresponding to the points where $B$ intersects $\Sigma$), and the projection $\pi:V_D \to D$ has a $1$-dimensional singular locus $\Sigma \times \{0\}$.
With the above specified landscape, the Siersma-Tib\u ar result \cite[Proposition 7.7]{ST} reads now as the specialisation for $s=1$ of Corollary \ref{corgenhom}, together with the bound provided by Theorem \ref{th2}.
\section{Examples} In this section we work out a few specific examples. In particular, in \S\ref{quad} we show that the upper bound given by Theorem \ref{th2} is sharp, \S\ref{rathom} deals with a hypersurface which is a rational homology manifold, while \S\ref{projc} discusses the case of a projective cone on a singular curve. However, as pointed out in \cite{Di0} already in the case of isolated singularities, it is difficult in general to compute the integral cohomology of a hypersurface by means of Corollary \ref{corgen}. It is therefore important to also develop alternative methods for exact calculations of cohomology and/or Betti numbers, e.g., see \cite{Di} for special situations.
\subsection{Singular quadrics}\label{quad} Let $n$ and $q$ be integers satisfying $4 \leq q \leq n+1$, and let $$f_q(x_0,\ldots x_{n+1})=\sum_{0\leq i,j \leq n+1} q_{ij} x_i x_j$$ be a quadric of rank $q:=\mathop{{\mathrm{rank}}}\nolimits (Q)$ with $Q=(q_{ij})$. The singular locus $\Sigma$ of the quadric hypersurface $V_q=\{f_q=0\} \subset {\mathbb C} P^{n+1}$ is a linear space of complex dimension $s=n+1-q$ satisfying $0 \leq s \leq n-3$. The generic transversal type for $\Sigma={\mathbb C} P^s$ is an $A_1$-singularity, so $\mu^\pitchfork=1$. Theorem \ref{th2} yields that \begin{equation}\label{ub2} b_{n+s+1}(V_q) \leq 2.\end{equation} In what follows, we show that if the rank $q$ is even (i.e., $n+s+1$ is even), the upper bound on $ b_{n+s+1}(V_q)$ given in \eqref{ub2} is sharp. Indeed, in our notation, the quadric $V_q$ is a projective cone with vertex $\Sigma$ over a smooth quadric $W_q \subset {\mathbb C} P^{n-s}$. Moreover, since $n-s\geq 3$, the homotopy version of the Lefschetz hyperplane theorem yields that $W_q$ is simply-connected (see, e.g., \cite[Theorem 1.6.5]{Di}). Let $U=V_q \setminus \Sigma$ and consider the long exact sequence $$ \cdots \to H^k_c(U;{\mathbb Z}) \to H^k(V_q;{\mathbb Z}) \to H^k(\Sigma;{\mathbb Z}) \to H^{k+1}_c(U;{\mathbb Z}) \to \cdots $$ Note that projecting from $\Sigma$ gives $U$ the structure of a vector bundle of rank $s+1$ over $W_q$. Let $p:U \to W_q$ denote the bundle map. Then $$H^k_c(U;{\mathbb Z})\cong H^k(W_q;Rp_!\underline{{\mathbb Z}}_U)$$ can be computed by the corresponding hypercohomology spectral sequence (i.e., the compactly supported Leray-Serre spectral sequence of the map $p$), with $E^{a,b}_2=H^a(W_q;R^bp_!\underline{{\mathbb Z}}_U)$. Since $\pi_1(W_q)=0$, the local system $R^bp_!\underline{{\mathbb Z}}_U$ is constant on $W_q$ with stalk $H^b_c({\mathbb C}^{s+1};{\mathbb Z})$. Since the latter is ${\mathbb Z}$ if $b=2s+2$ and $0$ otherwise, the above spectral sequence yields isomorphisms $H^k_c(U;{\mathbb Z}) \cong H^{k-2-2s}(W_q;{\mathbb Z})$ if $k \geq 2s+2$ and $H^k_c(U;{\mathbb Z}) \cong 0$ if $k < 2s+2$. On the other hand, $H^k(\Sigma;{\mathbb Z})=0$ if $k>2s$, so the above long exact sequence yields: \begin{equation} H^k(V_q;{\mathbb Z})\cong \begin{cases} H^k(\Sigma;{\mathbb Z}) & 0 \leq k \leq 2s \\ 0 & k=2s+1 \\ H^{k-2-2s}(W_q;{\mathbb Z}) & 2s+2 \leq k \leq 2n. \end{cases} \end{equation} Since $W_q$ is a smooth quadric, its integral cohomology is known from \eqref{one}, \eqref{two} and \eqref{bsm}. Altogether, this gives: \begin{equation} H^k(V_q;{\mathbb Z})\cong \begin{cases} 0 & k \text{ odd} \\ {\mathbb Z} & k \text{ even}, \ k \neq n+s+1 \\ {\mathbb Z}^2 & k= n+s+1 \text{ even}. \end{cases} \end{equation}
\subsection{One-dimensional singular locus with a two-step filtration}\label{rathom} Let $V=\{f=0\}\subset {\mathbb C} P^4$ be the $3$-fold in homogeneous coordinates $[x:y:z:t:v]$, defined by $$f=y^2z+x^3+tx^2+v^3.$$ The singular locus of $V$ is the projective line $\Sigma=\{[0:0:z:t:0] \mid z,t \in {\mathbb C}\}$. By \eqref{one}, we get: $b_0(V)=1$, $b_1(V)=0$, $b_2(V)=1$. Since $V$ is irreducible, \eqref{top} yields: $b_6(V)=1$. We are therefore interested to understand the Betti numbers $b_3(V)$, $b_4(V)$ and $b_5(V)$.
It was shown in \cite[Example 6.1]{M0} that $V$ has a Whitney stratification with strata: $$S_3:=V \setminus \Sigma, \ \ S_1:=\Sigma \setminus [0:0:0:1:0], \ \ S_0:=[0:0:0:1:0],$$ giving $V$ a two-step filtration $V \supset \Sigma \supset [0:0:0:1:0].$
The transversal singularity for the top singular stratum $S_1$ is the Brieskorn type singularity $y^2+x^3+v^3=0$ at the origin of ${\mathbb C}^3$ (in a normal slice to $S_1$), with corresponding transversal Milnor number $\mu_1^\pitchfork =4$. So Theorem \ref{th2} yields that $b_5(V) \leq 5$, while Corollary \ref{corgen} gives $b_3(V) \leq 10$. As we will indicate below, the actual values of $b_3(V)$ and $b_5(V)$ are zero.
It was shown in \cite[Example 6.1]{M0} that the hypersurface $V$ is in fact a ${\mathbb Q}$-homology manifold, so it satisfies Poincar\'e duality over the rationals. In particular, $b_5(V)=b_1(V)=0$ and $b_4(V)=b_2(V)=1$. To determine $b_3(V)$, it suffices to compute the Euler characteristic of $V$, since $\chi(V)=4-b_3(V)$. Let us denote by $Y\subset {\mathbb C} P^4$ a smooth $3$-fold which intersects the Whitney stratification of $V$ transversally. Then \eqref{chi} yields that $\chi(Y)=-6$ and we have by \cite[(10.40)]{M} that \begin{equation}\label{plug} \chi(V)=\chi(Y)-\chi(S_1 \setminus Y) \cdot \mu_1^\pitchfork -\chi(S_0) \cdot (\chi(F_0)-1), \end{equation} where $F_0$ denotes the Milnor fiber of $V$ at the singular point $S_0$. As shown in \cite[Example 6.1]{M0}, $F_0 \simeq S^3 \vee S^3$. So, using the fact that the general $3$-fold $Y$ intersects $S_1$ at $3$ points, we get from \eqref{plug} that $\chi(V)=4$. Therefore, $b_3(V)=0$, as claimed. Moreover, since $H^3(V;{\mathbb Z})$ is free, this also shows that in fact $H^3(V;{\mathbb Z})\cong 0$.
\begin{rem} Note that the hypersurface of the previous example has the same Betti numbers as ${\mathbb C} P^3$. This fact can also be checked directly, by noting that the monodromy operator acting on the reduced homology of the Milnor fiber of $f$ at the origin in ${\mathbb C}^5$ has no eigenvalue equal to $1$ (see \cite[Corollary 5.2.22]{Di}).
More generally, consider a degree $d$ homogeneous polynomial $g(x_0,\ldots, x_n)$ with associated Milnor $F_g$ such that the monodromy operator $h_*$ acting on $\widetilde{H}_*(F_g;{\mathbb Q})$ is the identity. Then the hypersurface $V=\{g(x_0,\ldots, x_n)+x_{n+1}^{d}=0\}\subset {\mathbb C} P^{n+1}$ has the same ${\mathbb Q}$-(co)homology as ${\mathbb C} P^n$. For example, the hypersurface $V_n=\{x_0x_1\ldots x_n+x_{n+1}^{n+1}=0\}$ has singularities in codimension $2$, but the same ${\mathbb Q}$-(co)homology as ${\mathbb C} P^n$. However, $V_n$ does not have in general the ${\mathbb Z}$-(co)homology of ${\mathbb C} P^n$; indeed, $H^3(V_2;{\mathbb Z})$ contains $3$-torsion (cf. \cite[Proposition 5.4.8]{Di}). \end{rem}
\subsection{Projective cone on a curve}\label{projc} The projective curve $C=\{xyz=0\}\subset {\mathbb C} P^2$ has three irreducible components and three singularities of type $A_1$ (each having a corresponding Milnor number equal to $1$). Therefore, by Remark \ref{rem3.3} and formula \eqref{b1}, the integral cohomology of $C$ is given by: $$H^0(C;{\mathbb Z})\cong {\mathbb Z}, \ H^1(C;{\mathbb Z}) \cong {\mathbb Z}, \ H^2(C;{\mathbb Z})\cong {\mathbb Z}^3.$$ The projective cone on $C$ is the surface $V=\{xyz=0\}\subset {\mathbb C} P^3$. The singular locus of $V$ consists of three projective lines intersecting at the point $[0:0:0:1]$, each having a (generic) transversal singularity of type $A_1$, i.e., with corresponding transversal Milnor number equal to $1$. By \cite[(5.4.18)]{Di}, we have that $$H^k(V;{\mathbb Z}) \cong H^{k-2}(C;{\mathbb Z}), \ \ \text{for all} \ k \geq 2.$$ Together with \eqref{one}, this yields: \begin{equation}\label{comp1} H^0(V;{\mathbb Z})\cong {\mathbb Z}, \ H^1(V;{\mathbb Z})\cong 0, \ H^2(V;{\mathbb Z})\cong {\mathbb Z}, \ H^3(V;{\mathbb Z}) \cong {\mathbb Z}, \ H^4(V;{\mathbb Z})\cong {\mathbb Z}^3. \end{equation} By Theorem \ref{th1}, the only non-trivial vanishing cohomology groups of $V$ are $H^2_{\varphi}(V)$, which is free, and $H^3_{\varphi}(V)$. These can be explicitly computed by using \eqref{bsm}, \eqref{sp1} and \eqref{comp1}, to get: $$H^2_{\varphi}(V)\cong {\mathbb Z}^7, \ H^3_{\varphi}(V)\cong {\mathbb Z}^2$$ (compare with \cite[Example 7.5]{ST}).
\section{Supplement to the Lefschetz hyperplane theorem and applications}\label{supLHT} In this section, we give a new proof of Kato's result mentioned in the Introduction. Our proof is different from that of \cite[Theorem 5.2.11]{Di}, and it relies on a supplement to the Lefschetz hyperplane section theorem (Theorem \ref{thapi}), which is proved in Theorem \ref{thap} below.
\subsection{A supplement to the Lefschetz hyperplane theorem} In this section, we prove the following result of Lefschetz type: \begin{thm}\label{thap} Let $V \subset {\mathbb C} P^{n+1}$ be a reduced complex projective hypersurface with $s=\dim V_{\rm sing}$ the complex dimension of its singular locus. (By convention, we set $s=-1$ if $V$ is nonsingular.) Let $H \subset {\mathbb C} P^{n+1}$ be a generic hyperplane (i.e., transversal to a Whitney stratification of $V$), and denote by $V_H:=V\cap H$ the corresponding hyperplane section of $V$. Then \begin{equation}\label{34ap} H^k(V,V_H; {\mathbb Z})=0 \ \ \text{for} \ \ k < n \ \ \text{and} \ \ n+s+1 < k < 2n. \end{equation} Moreover, $H^{2n}(V,V_H; {\mathbb Z})\cong{\mathbb Z}^r$, where $r$ is the number of irreducible components of $V$, and $H^{n}(V,V_H; {\mathbb Z})$ is (torsion-)free. \end{thm} \begin{proof} Let us first note that the long exact sequence for the cohomology of the pair $(V,V_H)$ together with \eqref{top} yield that: $$H^{2n}(V,V_H; {\mathbb Z})\cong H^{2n}(V;{\mathbb Z})\cong{\mathbb Z}^r.$$ Moreover, we have isomorphisms: $$H^k(V,V_H; {\mathbb Z}) \cong H^k_c(V^a;{\mathbb Z}),$$ where $V^a:=V\setminus V_H$. Therefore, the vanishing in \eqref{34ap} for $k<n$ is a consequence of the Artin vanishing theorem (e.g., see \cite[Corollary 6.0.4]{Sc}) for the perverse sheaf $\underline{{\mathbb Z}}_{V^a}[n]$ (cf. \cite{Le}) on the affine hypersurface $V^a$ obtained from $V$ by removing the hyperplane section $V_H$. Indeed, $$H^k_c(V^a;{\mathbb Z})={\mathbb H}^{k-n}_c(V^a;\underline{{\mathbb Z}}_{V^a}[n]) \cong 0$$ for all $k-n<0$. (Note that vanishing in this range is equivalent to the classical Lefschetz hyperplane section theorem.)
Since $V$ is reduced, we have that $s<n$. If $n=s+1$ then $n+s+1=2n$ and there is nothing else to prove in \eqref{34ap}. So let us now assume that $n>s+1$. For $n+s+1<k<2n$, we have the following sequence of isomorphisms: \begin{equation}\label{35} \begin{split} H^k(V, V_H; {\mathbb Z}) &\cong H^k(V \cup H, H; {\mathbb Z}) \\ &\cong H_{2n+2-k} ({\mathbb C} P^{n+1}\setminus H, {\mathbb C} P^{n+1} \setminus (V \cup H); {\mathbb Z}) \\ &\cong H_{2n+1-k}({\mathbb C} P^{n+1}\setminus (V \cup H); {\mathbb Z}), \end{split} \end{equation} where the first isomorphism follows by excision, the second is an application of the Poincar\'e-Alexander-Lefschetz duality, and the third follows from the cohomology long exact sequence of a pair. Set $$U={\mathbb C} P^{n+1}\setminus (V \cup H),$$ and let $L = {\mathbb C} P^{n-s}$ be a generic linear subspace (i.e., transversal to both $V$ and $H$). Then, by transversality, $L \cap V$ is a nonsingular hypersurface in $L$, transversal to the hyperplane at infinity $L \cap H$ in $L$. Therefore, $U \cap L=L\setminus (V \cup H) \cap L$ has the homotopy type of a wedge $$U \cap L \simeq S^1 \vee S^{n-s} \vee \ldots \vee S^{n-s},$$ e.g., see \cite[Corollary 1.2]{Lib}. Thus, by the Lefschetz hyperplane section theorem (applied $s+1$ times), we obtain: $$H_i(U;{\mathbb Z}) \cong H_i(U \cap L;{\mathbb Z}) \cong 0$$ for all integers $i$ in the range $1 < i < n-s$. Substituting $i = 2n+1-k$ in \eqref{35}, we get that $H^k(V, V_H; {\mathbb Z})\cong 0$ for all integers $k$ in the range $n+s+1 < k < 2n$.
It remains to show that $H^{n}(V,V_H; {\mathbb Z})\cong H^n_c(V^a;{\mathbb Z})\cong {\mathbb H}^{0}_c(V^a;\underline{{\mathbb Z}}_{V^a}[n])$ is (torsion-)free. This follows as in the proof of Theorem \ref{th1} since the affine hypersurface $V^a$ has rectified homological depth equal to its complex dimension $n$. This completes the proof of the theorem. \end{proof}
Theorem \ref{thap} and the Universal Coefficient Theorem now yield the following consequence: \begin{cor}\label{corap} In the notations of Theorem \ref{thap} we have that: \begin{equation}\label{36ap} H_k(V,V_H; {\mathbb Z})=0 \ \ \text{for} \ \ k < n \ \ \text{and} \ \ n+s+1 < k < 2n. \end{equation} Moreover, $H_{2n}(V,V_H; {\mathbb Z})\cong{\mathbb Z}^r$, where $r$ is the number of irreducible components of $V$. \end{cor}
\subsection{Kato's theorem for hypersurfaces} The isomorphism \eqref{two} from the introduction was originally proved by Kato \cite{Ka}, and it holds more generally for complete intersections. We derive it here as a consequence of Theorem \ref{thap}.
\begin{thm}[Kato]\label{Kato} Let $V\subset {\mathbb C} P^{n+1}$ be a reduced degree $d$ complex projective hypersurface with $s=\dim V_{\rm sing}$ the complex dimension of its singular locus. (By convention, we set $s=-1$ if $V$ is nonsingular.) Then \begin{equation}\label{twoap} H^k(V;{\mathbb Z}) \cong H^k( {\mathbb C} P^{n+1};{\mathbb Z}) \ \ \text{for all} \ \ n+s+2\leq k\leq 2n. \end{equation} Moreover, if $j:V \hookrightarrow {\mathbb C} P^{n+1}$ denotes the inclusion, the induced cohomology homomorphisms \begin{equation}\label{threeap} j^k:H^k( {\mathbb C} P^{n+1};{\mathbb Z}) \longrightarrow H^k(V;{\mathbb Z}), \ \ n+s+2\leq k\leq 2n, \end{equation} are given by multiplication by $d$ if $k$ is even. \end{thm}
\begin{proof} The statement of the theorem is valid only if $n\geq s+2$, so in particular we can assume that $V$ is irreducible and hence $H^{2n}(V;{\mathbb Z})\cong {\mathbb Z}$. Moreover, the fact that $j^{2n}$ is multiplication by $d=\deg(V)$ is true regardless of the dimension of singular locus, see \cite[(5.2.10)]{Di}. If $n=s+2$ there is nothing else to prove, so we may assume (without any loss of generality) that $n \geq s+3$.
We next proceed by induction on $s$.
If $V$ is nonsingular (i.e., $s=-1$), the assertions are well-known for any $n \geq 1$. We include here a proof for completeness. The isomorphism \eqref{twoap} can be obtained in this case from the Lefschetz isomorphism \eqref{one}, its homology analogue, and Poincar\'e duality. The statement about $j^k$ can also be deduced from \eqref{one} and Poincar\'e duality, but we include here a different argument inspired by \cite{Di}. Consider the isolated singularity at the origin for the affine cone $CV \subset {\mathbb C}^{n+2}$ on $V$, and the corresponding link $L_V:=S^{2n+3} \cap CV$, for $S^{2n+3}$ a small enough sphere at the origin in ${\mathbb C}^{n+2}$. Then $L_V$ is a $(n-1)$-connected closed oriented manifold of real dimension $2n+1$, so its only possibly nontrivial integral (co)homology appears in degrees $0$, $n$, $n+1$ and $2n+1$. The Hopf fibration $S^1 \hookrightarrow S^{2n+3} \longrightarrow {\mathbb C} P^{n+1}$ induces by restriction to $CV$ a corresponding Hopf fibration for $V$, namely $S^1 \hookrightarrow L_V \longrightarrow V$. Then for any $n+1 \leq k \leq 2n-2$, the cohomology Gysin sequences for the diagram of fibrations $$\xymatrix{ S^{2n+3} \ar[r] & {\mathbb C} P^{n+1}\\ L_V \ar[u] \ar[r] & V \ar[u]. }$$ yield commutative diagrams (with ${\mathbb Z}$-coefficients): \begin{equation}\label{Gys} \CD 0=H^{k+1}(S^{2n+3}) @>>> H^k({\mathbb C} P^{n+1}) @>{\psi}>{\cong}> H^{k+2}({\mathbb C} P^{n+1}) @>>> H^{k+2}(S^{2n+3})=0\\ @VVV @V {j^k}VV @V {j^{k+2}}VV @VVV \\ 0=H^{k+1}(L_V) @>>> H^k(V) @>{\psi_V}>{\cong}> H^{k+2}(V) @>>> H^{k+2}(L_V)=0\\ \endCD \end{equation} Here, if $k=2\ell$ is even, the isomorphism $\psi$ is the cup product with the cohomology generator $a\in H^2({\mathbb C} P^{n+1};{\mathbb Z})$, and similarly, $\psi_V$ is the cup product with $j^2(a)$. The assertion about $j^k$ follows now from \eqref{Gys} by decreasing induction on $\ell$, using the fact mentioned at the beginning of the proof that $j^{2n}$ is given by multiplication by $d$.
Let us next choose a generic hyperplane $H \subset {\mathbb C} P^{n+1}$ (i.e., $H$ is transversal to a Whitney stratification of $V$), and set as before $V_H=V \cap H$. It then follows from Theorem \ref{thap} and the cohomology long exact sequence of the pair $(V,V_H)$ that $H^{2n-1}(V;{\mathbb Z}) \cong 0$. It therefore remains to prove \eqref{twoap} and the corresponding assertion about $j^k$ for $k$ in the range for $n+s+2\leq k\leq 2n-2$. Let us consider the commuting square $$\CD V_H @> {\delta} >> H={\mathbb C} P^n\\ @V {\gamma}VV @VVV\\\ V @>> {j} > {\mathbb C} P^{n+1} \endCD$$ and the induced commutative diagram in cohomology: \begin{equation}\label{di} \CD H^k({\mathbb C} P^{n+1};{\mathbb Z}) @> {j^k} >> H^k(V;{\mathbb Z}) \\ @V {\cong}VV @VV{\gamma^k}V\\\ H^k({\mathbb C} P^{n};{\mathbb Z}) @>> {\delta^k} > H^k(V_H;{\mathbb Z}) \endCD \end{equation} By Theorem \ref{thap} and the cohomology long exact sequence of the pair $(V,V_H)$ we get that $\gamma^k$ is an isomorphism for all integers $k$ in the range $n+s+2\leq k\leq 2n-2$. Moreover, since $V_H \subset {\mathbb C} P^n$ is a degree $d$ reduced projective hypersurface with a $(s-1)$-dimensional singular locus (by transversality), the induction hypothesis yields that $H^k(V_H;{\mathbb Z}) \cong H^k({\mathbb C} P^n;{\mathbb Z})$ for $n+s \leq k \leq 2n-2$ and that, in the same range and for $k$ even, the homomorphism $\delta^k$ is given by multiplication by $d$. The commutativity of the above diagram \eqref{di} then yields \eqref{twoap} for all integers $k$ satisfying $n+s+2\leq k\leq 2n-2$, and the corresponding assertion about the induced homomorphism $j^k$ for $k$ even in the same range. This completes the proof of the theorem. \end{proof}
\begin{rem} Let us remark here that the proof of Kato's theorem in \cite[Theorem 5.2.11]{Di} relies on the Kato-Matsumoto result \cite{KM} on the connectivity of the Milnor fiber of the singularity at the origin of the affine cone $CV \subset {\mathbb C}^{n+2}$. \end{rem}
\begin{rem}\label{Katoh}
One can prove the homological version of Theorem \ref{Kato} in the similar manner, namely by using Corollary \ref{corap} instead of Theorem \ref{thap}. This yields the isomorphisms: \begin{equation} H_k(V;{\mathbb Z}) \cong H_k( {\mathbb C} P^{n+1};{\mathbb Z}) \ \ \text{for all} \ \ n+s+2\leq k\leq 2n, \end{equation} and the homomorphisms induced by the inclusion $j:V\hookrightarrow {\mathbb C} P^{n+1}$ in homology are given in this range (and for $k$ even) by multiplication by $d=\deg(V)$. \end{rem}
\begin{rem} We already noted that Theorem \ref{th1} yields the isomorphism \eqref{two} of Kato's theorem (see Corollary \ref{corgen}(i)). On the other hand, Kato's Theorem \ref{Kato} may be used to obtain a weaker version of Theorem \ref{th1} by more elementary means. Indeed, in the notations from the Introduction consider the diagram: $$ \CD H^k({\mathbb C} P^{n+1};{\mathbb Z}) @> {\cong}>> H^k({\mathbb C} P^{n+1} \times D;{\mathbb Z}) @> {b^k} >> H^k(V_D;{\mathbb Z}) @> {c^k} >> H^k(V_t;{\mathbb Z})\\
@. @. @V {\cong}VV @. \\ @. @. H^k(V;{\mathbb Z}) \endCD $$ and let $a^k:=c^k \circ b^k$. By Theorem \ref{Kato}, we have that: \begin{itemize} \item[(i)] $a^k$ is the multiplication by $d$ if $k>n$ even and an isomorphism for $k<n$; \item[(ii)] $b^k$ is the multiplication by $d$ if $n+s+2\leq k \leq 2n$ ($k$ even) and an isomorphism for $k<n$. \end{itemize} Therefore, $c^k$ is an isomorphism if $n+s+2 \leq k \leq 2n$ or $k<n$. The cohomology long exact sequence of the pair $(V_D,V_t)$ then yields that $H^k_{\varphi}(V)\cong H^{k+1}(V_D,V_t;{\mathbb Z})\cong 0$ for all integers $k \notin [n-1,n+s+1]$. \end{rem}
\end{document}
|
arXiv
|
{
"id": "2004.07686.tex",
"language_detection_score": 0.6749274134635925,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Measurement of the short-range attractive force between Ge plates using a torsion balance}
\author{W. J. Kim} \affiliation{Yale University, Department of Physics, P.O. Box 208120, New Haven, CT 06520, USA}
\author{A. O. Sushkov} \affiliation{Yale University, Department of Physics, P.O. Box 208120, New Haven, CT 06520, USA}
\author{D. A. R. Dalvit} \affiliation{Theoretical Division, MS B213, Los Alamos National Laboratory, Los Alamos, NM 87545, USA}
\author{S. K. Lamoreaux} \affiliation{Yale University, Department of Physics, P.O. Box 208120, New Haven, CT 06520, USA}
\date{\today}
\begin{abstract} We have measured the short-range attractive force between crystalline Ge plates, and found contributions from both the Casimir force and an electrical force possibly generated by surface patch potentials. Using a model of surface patch effects that generates an additional force due to a distance dependence of the apparent contact potential, the electrical force was parameterized using data at distances where the Casimir force is relatively small. Extrapolating this model, to provide a correction to the measured force at distances less than 5 $\mu$m, shows a residual force that is in agreement, within experimental uncertainty, with five models that have been used to calculate the Casimir force.
\end{abstract}
\pacs{12.20.Fv, 11.10.Wx, 73.40.Cg, 04.80.Cc}
\maketitle
{\it Introduction.-} The Casimir force has been a subject of great interest, both theoretically and experimentally, because it is a macroscopic manifestation of quantum vacuum effects \cite{Casimir,Milonni,Bordag}, and it can have significant effects in nanomechanical systems. Despite a number of successful measurements celebrated over the last decade \cite{Steve,Casexp}, early investigations of short-range forces \cite{Nancy,Stipe} report the possible systematic effects due to residual electrostatic forces. In particular, the observed variation in the effective contact potential, recently reported in \cite{KimPRARC} and later confirmed in \cite{Sven}, presents a problem of fundamental importance when setting limits to predicted submicron corrections to Newtonian gravity in a Casimir force measurement \cite{Roberto}. The optical response of a particular sample under study must also be carefully considered \cite{Svetovoy}, as the accuracy of data on the optical properties of materials typically limits calculational accuracy to no better than 5\%. In principle, both electric and optical studies of a given sample are subject to a combination of various surface effects of electric origin, and it is important to understand these issues in order to accurately characterize fundamental interactions, such as the Casimir force and non-Newtonian gravity.
{\it Our torsion balance set-up.-} In this Letter, we present results of force measurements between crystalline Ge plates \cite{ispoptics} in a sphere-plane geometry. Our apparatus, shown schematically in Fig. \ref{fig1}, is based on the design presented in \cite{Lamoreaux2} and improves on the apparatus described in \cite{Steve,fan}. On one side of a torsion pendulum a flat Ge plate is mounted, and approached by a Ge plate with a spherical surface, with radius of curvature $R=(15.10\pm0.05)$ cm, mounted on a Thorlabs T25 XYZ motion stage (8 nm resolution). When a force exists between these plates, the torsion body rotates and thereby generates an imbalance in capacitance on the other side of the pendulum, which carries a flat plate, situated in between two fixed ``compensator plates", that are attached to the support frame. An AC voltage is applied to the compensator plates, and the capacitance imbalance creates an AC voltage that is amplified and sent to a phase sensitive detector (PSD), providing an error signal to a proportional-integral-differential (PID) feedback circuit. A small correction voltage ($S_{\rm{PID}}$) is applied to the compensator plates keeping the system in equilibrium. The correction voltage is added to a large constant voltage $V_0 (\approx 9$V) to linearize the restoring force, $F\propto (S_{\rm{PID}}+V_0)^2\approx V_0^2+2V_0S_{\rm PID}$. This correction voltage provides a measure of the force between the Casimir plates and is recorded during the measurement.
\begin{figure}
\caption{Experimental setup of torsion balance (top-view). A pendulum body of length 15 cm hangs from a tungsten wire connected to a motorized rotation stage via the pivot point that is mounted on a support frame. The wire diameter is 25 $\mu$m, with length 2.5 cm, shorter than the previous experiment (66 cm) \cite{Steve} in order to minimize effects of tilt of the apparatus. At the bottom of the pendulum body (not shown in the figure) is a NdFeB magnet to damp the swinging modes of the pendulum at a natural frequency of 3 Hz. The mechanical assembly is covered by a glass bell jar (vacuum $5\times 10^{-7}$ torr) and is supported on a vibration isolation slab that has its foundation separate from the laboratory building.}
\label{fig1}
\end{figure}
The measured signal $S_{\rm{PID}}$ has contributions from several sources: \begin{equation} \label{PID} S_{\rm{PID}}(d,V_{\rm a})=S_{\rm{DC}}(d\rightarrow\infty)+S_{r}(d)+S_{a}(d,V_{\rm{a}}), \end{equation} where $S_{\rm{DC}}$ is the force-free component of the signal at large distances, $S_{r}$ is the residual signal due to distance-dependent forces, such as Casimir-Lifshitz force, and $S_{\rm{p}}$ is the signal due to the electrostatic force in response to an applied external voltage $V_{\rm{a}}$. For the sphere-plane geometry, this latter signal can be written in the proximity force approximation (PFA), valid when $d \ll R$, as $S_{a}(d,V_{\rm{a}})=\pi\epsilon_0 R(V_{\rm{a}}-V_{\rm{m}})^2/\beta d$, where $\beta$ is a calibration factor that converts $S_{\rm{PID}}$ in units of voltage to the actual units of force. The electrostatic signal is minimized ($S_{a}=0$) when $V_{\rm{a}}=V_{\rm{m}}$, and the electrostatic minimizing potential $V_{\rm{m}}$ is then defined to be the contact potential between the plates.
A range of plate voltages $V_{\rm{a}}$ is applied, and at a given separation the response $S_{\rm{PID}}$ is fitted to a parabola \begin{equation} S_{\rm{PID}}(d,V_{\rm{a}})=S_0+k(V_{\rm{a}}-V_{\rm{m}})^2. \label{para} \end{equation} The first two terms in Eq. (\ref{PID}) are absorbed in $S_0$ and represent the minimized signal when $V_{\rm{a}}=V_{\rm{m}}$. Repeating the parabola measurements shown in Fig. 2a, sequentially moving from the farthest to closest plate separations, enables us to inspect the $d$ dependence of the fitting parameters $k(d)$, $V_{\rm{m}}(d)$, and $S_0(d)$. The procedure outlined here was first implemented as a calibration routine in \cite{Iannuzzi} and more recently in \cite{KimPRARC} in an effort to detect a distance dependence of $V_{\rm{m}}$.
As the gap between the plates is reduced, the parabola curvature $k$ rapidly increases as shown in Fig. \ref{fig2}b. These curvature values are fitted to $k(d)=\alpha/d$, the expected dependence for the plane-sphere geometry, where the absolute distance $d\equiv d_0-d_{\rm{r}}$ is defined in terms of the asymptotic limit $d_0$ and the relative distance $d_r$ recorded during a parabola measurement. The conversion factor $\beta$ is then obtained through $\alpha\equiv\pi\epsilon_0R/\beta$. Obviously, $\alpha$ can be also used to determine the absolute distance through $d=\alpha/k$, implying a significant correlation of $\alpha$ with $d_0$. Consistency between these two methods of distance determination reflects validity of the use of the $1/d$ power law as implied by a value of $\chi^2_0$ close to unity for our data set. Fig. \ref{fig2}c shows the electric potentials $V_{\rm{m}}$ at minima of the parabola curvatures plotted versus $d$, indicating the distance-dependent minimizing potential $V_{\rm{m}}(d)$, a behavior that has been observed in other experiments \cite{KimPRARC,Sven}.
\begin{figure}\label{fig2}
\end{figure}
To see the trend in $V_{\rm m}(d)$ more clearly and to determine short-range forces with higher statistical accuracy, we have repeated 200 times the experimental sequence described in Fig. \ref{fig2}, yielding a total of 5800 data points. Each group of five data points taken at a given fixed distance with varying applied potential are used to determine the three parabola parameters discussed above, in addition to the force and distance. The mean value of the calibration factor after analyzing all data is $\beta=(1.35\pm0.04)\times 10^{-7}$~N/V. Both the asymptotic limit $d_0$, shown in Fig \ref{fig2}b, and the DC offset of the PID signals $S_{\rm{DC}}$ drift slightly during a run. The uncertainty in position is roughly 10 \% at a given distance and about 50 nm at the typical closest gap separation, consistent with the actuator minimum displacement of 40 nm. The DC offset drift has been corrected by monitoring $S_{\rm{PID}}$ before and after each consecutive run and applying a linear correction.
{\it Varying minimizing potential.-} An outstanding feature of our data is the distance variation of the applied voltage $V_{\rm{m}}$ that minimizes the force, as clearly shown in Fig. \ref{fig3}. It must be recognized that this variation can lead to an extra force of electrical origin, as demonstrated in \cite{Lamoreauxcont}. However, the model used in \cite{Lamoreauxcont} assumes that the variation in the minimizing potential is due to a varying contact potential, specifically modeled as a voltage source in series with the plates. The varying minimizing potential observed in our data is more likely due to large-scale gradients in the contact potential across the surface of the plates, due to, for example, polishing stresses or the curvature of the spherical surface plate changing the crystal plane orientation at the surface. Such variations have been observed for many materials \cite{Robertson,INFN}, with typical large scale fluctuations on the order of a few mV. The variation in the apparent contact potential is due to the effective averaging area changing as the curved and flat surfaces of the plates are brought together. A numerical analysis \cite{Note} of a wide range of surface potential variations shows that the variation of $V_{\rm{m}}(d)$ leads to an electrostatic force of the form $F^{\rm el}_{r_1}(d)= \pi R\epsilon_0[V_{\rm{m}}(d)+V_1]^2 / d$, where $V_1$ is a constant offset parameter at large distances that can be determined from the experimental data. The origin of this effect is due to the plate curvature, together with large scale variations in the surface contact potential. Note that although the parabola measurement {\it minimizes} the electrostatic force across the plates, it does not necessarily {\it nullify} all the electric forces that possibly exist. \begin{figure}
\caption{The value of the force minimizing potential as a function of plate separation. The red points are the average of the data. Our measurement reveals a slow rise of the minimizing potential as the plates approach each other, of order of 6 mV over 100 $\mu$m. This variation in our data set shows a similar trend observed in a recent measurement \cite{Sven} where the variation of 6 mV over 1 $\mu$m is reported between Au coated samples, significantly larger than the value found in our study.}
\label{fig3}
\end{figure}
{\it Random small-scale patches.-} In addition to large scale gradients in the surface potential, there can be small-scale (i.e., much smaller than the plate diameter) random fluctuations in the surface potential associated with strains, irregularities, and impurities. It is straightforward to show that the electrostatic energy per unit area between two flat plates with random patch voltages is \cite{Speake,Note} \begin{equation} E_{\rm patch}(d)=\frac{\epsilon_0 \pi V_{\rm rms}^2}{4 d} \int_0^\infty du S(u/d) \frac{e^{-2u}}{\sinh^2(u)}, \label{patch_residual} \end{equation} where $V_{\rm rms}$ is the rms value of the random patch voltages. For simplicity, we have assumed isotropic patches with surface correlation functions $\langle V_{k,i} V_{k',j} \rangle = V^2_{\rm rms} S(k) \delta_{i,j} \delta(k-k')$, where $i,j=1,2$ denote the plates, and $S(k)$ is the unity-normalized spectral density. The residual electrostatic force between the sphere and the plane due to these patches can be obtained from PFA as $F^{\rm el}_{r_2}(d)=2 \pi R E_{\rm patch}(d)$. For example, for random-voltage patches of radius $\lambda$ uniformly distributed on the surfaces, the spectral density is $S(k) \approx \sin k \lambda / \pi^2 k^2$. It is easy to see that in the limit $\lambda \gg d$ the residual patch force in the sphere-plane geometry scales as $F^{\rm el}_{r_2}(d)=\pi R\epsilon_0 V^2_{\rm{rms}}/d$ \cite{Note}.
{\it Electrostatic residual force.-} We fit the data of the residual force at the minimizing potential (Fig 2.d) with a force of electric origin $F^{\rm el}_r = F_0+F^{\rm el}_{r_1} + F^{\rm el}_{r_2} =F_0+\pi R \epsilon_0 \{[V_{\rm{m}}(d)+V_1]^2 + V^2_{\rm rms} \}/d$, where $F_0$ is an offset parameter at large distances. A least-squares fit of the observed force using data for $d>5$ $\mu$m (a regime where the Casimir force should be vanishingly small) and the measured $V_{\rm{m}}(d)$, while leaving $F_0$, $V_1$, and $V_{\rm{rms}}$ as adjustable fit parameters, yields an excellent description of the observed large distance force \cite{powerlaw}, as shown plotted with the data in Fig. 4. Including data at shorter distances ($d<5$ $\mu$m) causes a significant fit deviation, indicating an interference with the actual Casimir force which is highly nonlinear at short distances. A similar long-range force has been previously observed in the measurement of van der Waals interaction and the corresponding correction is applied to the data based on work function anisotropies and their related patch charges \cite{Nancy}.
{\it Casimir residual force.-} We have tested our data for the presence of a residual Casimir force $F_r^{\rm Cas}(d)$ between the Ge plates, which we have computed in PFA from the plane-plane Casimir-Lifshitz energy, $F^{\rm Cas}_r(d) = 2 \pi R E^{\rm Cas}_{pp}(d)$. We have calculated the corresponding reflection coefficients using five different theoretical models for the Ge plates \cite{Diego, Galina}: ideal dielectric; ideal dielectric + Drude conductivity corrections; ideal dielectric + plasma conductivity corrections; quasi-static Debye-H\"uckel screening model; and charge drift model. Fig. 4 (bottom) shows the experimental data for the residual force after subtraction of the joint (contact potential and patch potential) electrostatic forces, and the theory curves for the Casimir-Lifshitz force between Ge plates at $T=300$ K for these five theoretical models. The error bars take into account all statistical uncertainties (2-3 \%) as well as fitting uncertainties from the electrostatic force analysis (10\%). Within experimental uncertainty, our data agrees well with the theoretical predictions for the Casimir force, but it is not of sufficient accuracy to distinguish among the different models for the Ge plates. High precision measurements of the Casimir force require a careful evaluation of the electrostatic effects considered in detail for the first time in our present study. \begin{figure}\label{fig4}
\end{figure}
{\it Conclusions.-} We have performed measurements of the short-range force between Ge plates in the sphere-plane geometry, and have observed that the potential $V_{\rm{m}}$ that minimizes the electrostatic force depends on the gap between the plates. We have considered two contributions of electric origin present in the residual data for the force. The first contribution is due to large-scale variations in the contact potential along the surface of the plates, that leads to the gap-dependent minimizing potential and, as a result, to an electrostatic force proportional to $(V_{\rm{m}}(d)+V_1)^2/d$. The second contribution can be modeled as arising from potential patches on the surfaces that, in the case when they have typical sizes much smaller that the plate diameters and much larger than the plate separation, leads to a further electrostatic force proportional to $V_{\rm rms}^2/d$. We have fitted our experimental data at large distances ($d>5$ $\mu$m, where the Casimir force is expected to be negligible) with these two electrostatic force effects, and found we could establish good agreement between our model and the experimental data. Furthermore, we have subtracted these forces from the data at short separations ($d<5$ $\mu$m) and found a residual force that is in agreement with the theoretical predictions for the Casimir-Lifshitz force between Ge plates. Our measurements do not have enough accuracy to distinguish between the different theoretical models used to characterize the optical properties of the Ge plates. Future measurements are deemed necessary in light of our discussion, in particular to better understand the physical origins of the observed electrostatic forces. We are currently exploring similar surface effects in a pair of Au samples. We acknowledge support from Yale University for the construction of the experimental apparatus and data acquisition, and from Los Alamos LDRD program. We thank G. Klimchitskaya for useful discussions.
\end{document}
|
arXiv
|
{
"id": "0812.2798.tex",
"language_detection_score": 0.8563132286071777,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Colorful Subhypergraphs in Uniform Hypergraphs}
\begin{abstract} There are several topological results ensuring the existence of a large complete bipartite subgraph in any properly colored graph satisfying some special topological regularity conditions. In view of $\mathbb{Z}_p$-Tucker lemma, Alishahi and Hajiabolhassan [{\it On the chromatic number of general Kneser hypergraphs, Journal of Combinatorial Theory, Series B, 2015}] introduced a lower bound for the chromatic number of Kneser hypergraphs ${\rm KG}^r({\mathcal H})$. Next, Meunier [{\it Colorful subhypergraphs in Kneser hypergraphs, The Electronic Journal of Combinatorics, 2014}] improved their result by proving that any properly colored general Kneser hypergraph ${\rm KG}^r({\mathcal H})$ contains a large colorful $r$-partite subhypergraph provided that $r$ is prime. In this paper, we give some new generalizations of $\mathbb{Z}_p$-Tucker lemma. Hence, improving Meunier's result in some aspects. Some new lower bounds for the chromatic number and local chromatic number of uniform hypergraphs are presented as well. \\
\noindent{\it Keyword:} chromatic number of hypergraphs, $\mathbb{Z}_p$-Tucker-Ky~Fan lemma, colorful complete hypergraph, $\mathbb{Z}_p$-box-complex, $\mathbb{Z}_p$-hom-complex \end{abstract}
\section{\bf Introduction} \subsection{{\bf Background and Motivations}} In 1955, Kneser~\cite{MR0068536} posed a conjecture about the chromatic number of Kneser graphs. In 1978, Lov{\'a}sz~\cite{MR514625} proved this conjecture by using algebraic topology. The Lov{\'a}sz's proof marked the beginning of the history of topological combinatorics. Nowadays, it is an active stream of research to study the coloring properties of graphs by using algebraic topology. There are several lower bounds for the chromatic number of graphs related to the indices of some topological spaces defined based on the structure of graphs. However, for hypergraphs, there are a few such lower bounds, see~\cite{2013arXiv1302.5394A,MR953021,Iriye20131333,MR1081939,MR2279672}.
A {\it hypergraph} ${\mathcal H}$ is a pair $(V({\mathcal H}),E({\mathcal H}))$, where $V({\mathcal H})$ is a finite set, called the vertex set of ${\mathcal H}$, and $E({\mathcal H})$ is a family of nonempty subsets of $V({\mathcal H})$, called the edge set of ${\mathcal H}$. Throughout the paper, by a nonempty hypergraph, we mean a hypergraph with at least one edge. If any edge $e\in E({\mathcal H})$ has the cardinality $r$, then the hypergraph ${\mathcal H}$ is called {\it $r$-uniform.} For a set $U\subseteq V({\mathcal H})$, the {\it induced subhypergraph on $U$,} denoted ${\mathcal H}[U]$, is a hypergraph with the vertex set $U$ and the edge set $\{e\in E({\mathcal H}):\; e\subseteq U\}$. Throughout the paper, by a {\it graph,} we mean a $2$-uniform hypergraph. Let $r\geq 2$ be a positive integer and $q\geq r$ be an integer. An $r$-uniform hypergraph ${\mathcal H}$ is called {\it $q$-partite} with parts $V_1,\ldots,V_q$ if \begin{itemize} \item $V({\mathcal H})=\displaystyle\bigcup_{i=1}^q V_i$ and \item each edge of ${\mathcal H}$ intersects each part $V_i$ in at most one vertex. \end{itemize} If ${\mathcal H}$ contains all possible edges, then we call it a {\it complete $r$-uniform $q$-partite hypergraph.}
Also, we say the hypergraph ${\mathcal H}$ is {\it balanced} if the values of $|V_j|$ for
$j =1,\ldots,q$ differ by at most one, i.e., $|V_i|-|V_j|\leq 1$ for each $i,j\in[q]$.
Let ${\mathcal H}$ be an $r$-uniform hypergraph and $U_1,\ldots, U_q$ be $q$ pairwise disjoint
subsets of $V({\mathcal H})$. The hypergraph ${\mathcal H}[U_1,\ldots, U_q]$ is a subhypergraph of ${\mathcal H}$ with the vertex set $\displaystyle\bigcup_{i=1}^q U_i$ and the edge set $$E({\mathcal H}[U_1,\ldots, U_q])=\left\{e\in E({\mathcal H}):\; e\subseteq \displaystyle\bigcup_{i=1}^q U_i\mbox{ and } |e\cap U_i|\leq 1\mbox{ for each } i\in[q]\right\}.$$ Note that ${\mathcal H}[U_1,\ldots, U_q]$ is an $r$-uniform $q$-partite hypergraph with parts $U_1,\ldots,U_q$. By the symbol ${[n]\choose r}$, we mean the family of all $r$-subsets of the set $[n]$. The hypergraph $K^r_n=\left([n],{[n]\choose r}\right)$ is celled the complete $r$-uniform hypergraph with $n$ vertices. For $r=2$, we would rather use $K_n$ instead of $K_n^2$. The largest possible integer $n$ such that ${\mathcal H}$ contains $K^r_n$ as a subhypergraph is called the {\it clique number of ${\mathcal H}$}, denoted $\omega({\mathcal H})$.
A {\it proper $t$-coloring} of a hypergraph ${\mathcal H}$ is a map $c:V({\mathcal H})\longrightarrow [t]$ such that there is no monochromatic edge. The minimum possible such a $t$ is called {\it the chromatic number of ${\mathcal H}$}, denoted $\chi({\mathcal H})$. If there is no such a $t$, we define the chromatic number to be infinite. Let $c$ be a proper coloring of ${\mathcal H}$ and $U_1,\ldots, U_q$ be $q$ pairwise disjoint subsets of $V({\mathcal H})$. The hypergraph ${\mathcal H}[U_1,\ldots, U_q]$ is said to be {\it colorful } if for each $j\in[q]$, the vertices of $U_j$ get pairwise distinct colors. For a properly colored graph $G$, a subgraph is called {\it multicolored} if its vertices get pairwise distinct colors.
For a hypergraph ${\mathcal H}$, {\it the Kneser hypergraph} ${\rm KG}^r({\mathcal H})$ is an $r$-uniform hypergraph with the vertex set $E({\mathcal H})$ and whose edges are formed by $r$ pairwise vertex-disjoint edges of ${\mathcal H}$, i.e., $$E({\rm KG}^r({\mathcal H}))=\left\{ \{e_1,\ldots,e_r\}:\; e_i\cap e_j=\varnothing\mbox{ for each } i\neq j\in[r] \right\}.$$ For any graph $G$, it is known that there are several hypergraphs ${\mathcal H}$ such that ${\rm KG}^2({\mathcal H})$ and $G$ are isomorphic.
The Kneser hypergraph ${\rm KG}^r\left(K_n^k\right)$ is called the ``usual" Kneser hypergraph which is denoted by ${\rm KG}^r(n,k)$. Coloring properties of Kneser hypergraphs have been studied extensively in the literature. Lov\'asz~\cite{MR514625} (for $r=2$) and Alon, Frankl and Lov\'asz~\cite{MR857448} determined the chromatic number of ${\rm KG}^r(n,k)$. For an integer $r\geq 2$, they proved that $$\chi\left({\rm KG}^r(n,k)\right)= \left\lceil{n-r(k-1)\over r-1}\right\rceil.$$ For a hypergraph ${\mathcal H}$, the $r$-colorability defect of ${\mathcal H}$, denoted ${\rm cd}_r({\mathcal H})$, is the minimum number of vertices which should be removed such that the induced hypergraph on the remaining vertices is $r$-colorable, i.e.,
$${\rm cd}_r({\mathcal H})=\min\left\{|U|:\; {\mathcal H}[V({\mathcal H})\setminus U]\mbox{ is $r$-colorable}\right\}.$$ For a hypergraph ${\mathcal H}$, Dol'nikov~\cite{MR953021}~(for $r=2$) and K{\v{r}}{\'{\i}}{\v{z}}~{\rm \cite{MR1081939} proved taht $$\chi({\rm KG}^r({\mathcal H}))\geq \left\lceil{{\rm cd}_r({\mathcal H})\over r-1}\right\rceil,$$ which is a generalization of the results by Lov\'asz~\cite{MR514625} and Alon, Frankl and Lov\'asz~\cite{MR857448}.
For a positive integer $r$, let $\mathbb{Z}_r=\{\omega,\omega^2\ldots,\omega^r\}$ be a cyclic group of order $r$ with generator $\omega$. Consider a vector $X=(x_1,x_2,\ldots,x_n)\in(\mathbb{Z}_r\cup\{0\})^n$. An alternating subsequence of $X$ is a sequence $x_{i_1},x_{i_2},\ldots,x_{i_m}$ of nonzero terms of $X$ such that $i_1<\cdots<i_m$ and $x_{i_j}\neq x_{i_{j+1}}$ for each $j\in [m-1]$. We denote by ${\rm alt}(x)$ the maximum possible length of an alternating subsequence of $X$. For a vector $X=(x_1,x_2,\ldots,x_n)\in (\mathbb{Z}_r\cup\{0\})^n$ and for an $\epsilon\in\mathbb{Z}_p$, set $X^\epsilon=\{i\in[n]:\; x_i=\epsilon\}$. Note that, by abuse of notation, we can write $X=(X^\epsilon)_{\epsilon\in \mathbb{Z}_r}$. For two vectors $X,Y\in (\mathbb{Z}_r\cup\{0\})^n$, by $X\subseteq Y$, we mean $X^\epsilon\subseteq Y^\epsilon$ for each $\epsilon\in\mathbb{Z}_r$.
For a hypergraph ${\mathcal H}$ and a bijection $\sigma:[n]\longrightarrow V({\mathcal H})$, define $${\rm alt}_r({\mathcal H},\sigma)=\displaystyle\max\left\{{\rm alt}(X):\; X\in (\mathbb{Z}_r\cup\{0\})^n\mbox{ such that } E({\mathcal H}[\sigma(X^\epsilon)])=\varnothing\mbox{ for each } \epsilon\in\mathbb{Z}_r \right\}.$$ Also, let $${\rm alt}_r({\mathcal H})=\displaystyle\min_{\sigma} {\rm alt}_r({\mathcal H},\sigma),$$
where the minimum is taken over all bijection $\sigma:[n]\longrightarrow V({\mathcal H})$. One can readily check that for any hypergraph ${\mathcal H}$, $|V({\mathcal H})|-{\rm alt}_r({\mathcal H})\geq {\rm cd}_r({\mathcal H})$ and the inequality is often strict, see~\cite{2013arXiv1302.5394A}. Alishahi and Hajiabolhassan~\cite{2013arXiv1302.5394A} improved Dol'nikov-K{\v{r}}{\'{\i}}{\v{z}} result by proving that
for any hypergraph ${\mathcal H}$ and for any integer $r\geq 2$, the quantity $\left\lceil{ |V({\mathcal H})|-{\rm alt}_r({\mathcal H})\over r-1}\right\rceil$ is a lower bound for the chromatic number of ${\rm KG}^r({\mathcal H})$.
Using this lower bound, the chromatic number of some families of graphs and hypergraphs are computed, see~\cite{2014arXiv1401.0138A,2014arXiv1403.4404A,2014arXiv1407.8035A,2015arXiv150708456A,2013arXiv1302.5394A,HaMe16}. There are some other lower bounds for the chromatic number of graphs which are better than the former discussed lower bounds. They are based on some topological indices of some topological spaces connected to the structure of graphs. In spite of these lower bounds being better, they are~not combinatorial and most of the times they are difficult to compute.
The existence of large colorful bipartite subgraphs in a properly colored graph has been extensively studied in the literature~\cite{2013arXiv1302.5394A,MR2971704,MR2763055,SiTaZs13,MR2279672,MR2351519}. To be more specific, there are several theorems ensuring the existence of a colorful bipartite subgraph in any properly colored graph such that the bipartite subgraph has a specific number of vertices related to some topological parameters connected to the graph. Simonyi and Tardos~\cite{MR2351519} improved Dol'nikov's lower bound and proved that in any proper coloring of a Kneser graph ${\rm KG}^2({\mathcal H})$, there is a multicolored complete bipartite graph $K_{\left\lceil{{\rm cd}_2({\mathcal H})\over 2}\right\rceil,\left\lfloor{{\rm cd}_2({\mathcal H})\over 2}\right\rfloor}$ such that the ${\rm cd}^2({\mathcal H})$ different colors occur alternating on the two parts of the bipartite graph with respect to their natural order. By a combinatorial proof, Alishahi and Hajiabolhassan~\cite{2013arXiv1302.5394A} improved this result. They proved that the the result remains true if we replace ${\rm cd}^2({\mathcal H})$ by $n-{\rm alt}_2({\mathcal H})$. Also, a stronger result is proved by Simonyi, Tardif, and Zsb{\'{a}}n~\cite{SiTaZs13}. \begin{alphtheorem}{\rm (Zig-zag Theorem~\cite{SiTaZs13}).}\label{zigzag} Let $G$ be a nonempty graph which is properly colored with arbitrary number of colors. Then $G$ contains a multicolored complete bipartite subgraph $K_{\lceil{t\over2}\rceil,\lfloor{t\over2}\rfloor}$, where ${\rm Xind}({\rm Hom}(K_2,G))+2= t$. Moreover, colors appear alternating on the two sides of the bipartite subgraph with respect to their natural ordering. \end{alphtheorem}
The quantity ${\rm Xind}({\rm Hom}(K_2,G))$ is the cross-index of hom-complex ${\rm Hom}(K^2,G)$ which will be defined in Subsection~\ref{Boxdefin}. We should mention that there are some other weaker similar results in terms of some other topological parameters, see~\cite{MR2279672,MR2351519}.
Note that prior mentioned results concern the existence of colorful bipartite subgraphs in properly colored graphs ($2$-uniform hypergraphs). In 2014, Meunier~\cite{Meunier14} found the first colorful type result for the uniform hypergraphs. He proved that for any prime number $p$, any properly colored Kneser hypergraph $\operatorname{KG}^p({\mathcal H})$ must contain a colorful balanced complete $p$-uniform $p$-partite subhypergraph with a specific number of vertices, see Theorem~\ref{colorfulhyper}.
\subsection{\bf Main Results} For a given graph $G$, there are several complexes defined based on the structure of $G$. For instance, the box-complex of $G$, denoted ${\rm B}_0(G)$, and the hom-complex of $G$, denoted ${\rm Hom}(K_2,G)$, see~\cite{MR1988723,SiTaZs13,MR2279672}. Also, there are some lower bounds for the chromatic number of graphs related to some indices of these complexes~\cite{SiTaZs13,MR2279672}. In this paper, we naturally generalize the definitions of box-complex and hom-complex of graphs to uniform hypergraphs. Also, the definition of $\mathbb{Z}_p$-cross-index of $\mathbb{Z}_p$-posets will be introduced. Using these complexes, as a first main result of this paper, we generalize Meunier's result~\cite{Meunier14} (Theorem~\ref{colorfulhyper}) to the following theorem. \begin{theorem}\label{maincolorfulindex} Let $r\geq 2$ be a positive integer and $p\geq r$ be a prime number. Assume that ${\mathcal H}$ is an $r$-uniform hypergraph and $c:V({\mathcal H})\longrightarrow[C]$ is a proper coloring of ${\mathcal H}$ {\rm (}$C$ arbitrary{\rm )}. Then we have the following assertions. \begin{itemize} \item[{\rm (i)}] There is some colorful balanced complete $r$-uniform $p$-partite subhypergraph in ${\mathcal H}$ with ${\rm ind}_{\mathbb{Z}_p}(B_0({\mathcal H},\mathbb{Z}_p))+1$ vertices. In particular, $$\chi({\mathcal H})\geq {{\rm ind}_{\mathbb{Z}_p}(B_0({\mathcal H},\mathbb{Z}_p))+1\over r-1}.$$ \item[{\rm (ii)}] If $p\leq \omega({\mathcal H})$, then there is some colorful balanced complete $r$-uniform $p$-partite subhypergraph in ${\mathcal H}$ with ${\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))+p$ vertices. In particular, $$\chi({\mathcal H})\geq {{\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))+p\over r-1}.$$ \end{itemize} \end{theorem}
Quantities ${\rm ind}_{\mathbb{Z}_p}(B_0({\mathcal H},\mathbb{Z}_p))$ and ${\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))$ appearing in the statement of Theorem~\ref{maincolorfulindex} are respectively the $\mathbb{Z}_p$-index and $\mathbb{Z}_p$-cross-index of the $\mathbb{Z}_p$-box-complex $B_0({\mathcal H},\mathbb{Z}_p)$ and $\mathbb{Z}_p$-hom-complex ${\rm Hom}(K^r_p,{\mathcal H})$ which will be defined in Subsection~\ref{Boxdefin}. Using these complexes, we introduce some new lower bounds for the chromatic number of uniform hypergraphs. In view of Theorem~\ref{maincolorfulindex}, next theorem provides a hierarchy of lower bounds for the chromatic number of $r$-uniform hypergraphs. \begin{theorem}\label{inequalities} Let $r\geq 2$ be a positive integer and $p\geq r$ be a prime number. For any $r$-uniform hypergraph ${\mathcal H}$, we have the following inequalities.\\ {\rm (i)} If $p\leq \omega({\mathcal H})$, then $${\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))+p \geq {\rm ind}_{\mathbb{Z}_p}(B_0({\mathcal H},\mathbb{Z}_p))+1.$$ {\rm (ii)} If ${\mathcal H}={\rm KG}^r({\mathcal F})$ for some hypergraph ${\mathcal F}$, then $$
{\rm ind}_{\mathbb{Z}_p}(B_0({\mathcal H},\mathbb{Z}_p))+1
\geq |V({\mathcal F})|-{\rm alt}_p({\mathcal F}) \geq {\rm cd}_p({\mathcal F}). $$ \end{theorem}
In view of Theorem~\ref{inequalities}, Theorem~\ref{maincolorfulindex} is a common extension of Theorem~\ref{zigzag} and Theorem~\ref{colorfulhyper}. Furthermore, for $r=2$, Theorem~\ref{maincolorfulindex} implies the next corollary which also is a generalization of Theorem~\ref{zigzag}. \begin{corollary}\label{cor1} Let $p$ be a prime number and let $G$ be a nonempty graph which is properly colored with arbitrary number of colors. Then there is a multicolored complete $p$-partite subgraph $K_{n_1,n_2,\ldots,n_p}$ of $G$ such that \begin{itemize} \item $\displaystyle\sum_{i=1}^pn_i={\rm ind}_{\mathbb{Z}_p}(B_0(G,\mathbb{Z}_p))+1$,
\item $|n_i-n_j|\leq 1$ for each $i,j\in[p]$. \end{itemize} Moreover, if $p\leq \omega(G)$, then ${\rm ind}_{\mathbb{Z}_p}(B_0(G,\mathbb{Z}_p))+1$ can be replaced with ${\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K_p,G))+p$. \end{corollary}
In view of the prior mentioned results, the following question naturally arises. \begin{question} Do Theorem~\ref{maincolorfulindex} and Theorem~\ref{inequalities} remain true for non-prime $p$? \end{question}
\subsection{\bf Applications to Local Chromatic Number of Uniform Hypergraphs} For a graph $G$ and a vertex $v\in V(G)$, the {\it closed neighborhood of $v$,} denoted $N[v]$, is the set $\{v\}\cup\{u:\; uv\in E(G)\}$. The {\it local chromatic number} of $G$, denoted $\chi_l(G)$, is defined in~\cite{ERDOS198621} as follows:
$$\chi_l(G)=\displaystyle\min_c\max\{|c(N[v])|:\; v\in V(G)\}$$ where the minimum in taken over all proper coloring $c$ of $G$. Note that Theorem~\ref{zigzag} gives the following lower bound for the local chromatic number of a nonempty graph $G$: \begin{equation}\label{localloerzigzag} \chi_l(G)\geq \left\lceil{{\rm Xind}({\rm Hom}(K_2,G))+2\over 2}\right\rceil+1. \end{equation} Note that for a Kneser hypergraph $\operatorname{KG}^2({\mathcal H})$, by using Simonyi and Tardos colorful result~\cite{MR2351519} or the extension given by Alishahi and Hajiabolhassan~\cite{2013arXiv1302.5394A}, there are two similar lower bounds for $\chi_l(\operatorname{KG}^2({\mathcal H}))$ which respectively used $\operatorname{cd}_2({\mathcal H})$ and
$|V({\mathcal H})|-\operatorname{alt}_2({\mathcal H})$ instead of ${\rm Xind}({\rm Hom}(K_2,G))+2$. However, as it is stated in Theorem~\ref{inequalities}, the lower bound in terms of ${\rm Xind}({\rm Hom}(K_2,G))+2$ is better than these two last mentioned lower bounds. Using Corollary~\ref{cor1}, we have the following lower bound for the local chromatic number of graphs. \begin{corollary}\label{locallowerp} Let $G$ be a nonempty graph and $p$ be a prime number. Then $$\chi_l(G)\geq t-\left\lfloor{t\over p}\right\rfloor+1,$$ where $t={\rm ind}_{\mathbb{Z}_p}(B_0(G,\mathbb{Z}_p))+1$. Moreover, if $p\leq \omega(G)$, then ${\rm ind}_{\mathbb{Z}_p}(B_0(G,\mathbb{Z}_p))+1$ can be replaced with ${\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K_p,G))+p$. \end{corollary} Note that if we set $p=2$, then previous theorem implies Simonyi and Tardos lower bound for the local chromatic number. Note that, in general, this lower bound might be better than Simonyi and Tardos lower bound. To see this, let $k\geq 2$ be a fixed integer. Consider the Kneser graph $\operatorname{KG}^2(n,k)$ and let $p=p(n)$ be a prime number such that $p=O(\ln n)$. By Theorem~\ref{inequalities}, for $n\geq pk$, we have $${\rm ind}_{\mathbb{Z}_p}(B_0(\operatorname{KG}^2(n,k),\mathbb{Z}_p))+1\geq \operatorname{cd}_p(K_n^k)=n-p(k-1).$$ Note that the lower bound for $\chi_l(\operatorname{KG}^2(n,k))$ coming form Inequality~\ref{localloerzigzag} is \begin{equation}\label{equation2} 1+\left\lceil{n-2k+2\over 2}\right\rceil={n\over 2}-o(1), \end{equation} while, in view of Corollary~\ref{locallowerp}, we have $$\chi_l(\operatorname{KG}^2(n,k))\geq n-p(k-1)-\left\lfloor{n-p(k-1)\over p}\right\rfloor+1=n-o(n),$$ which is better than the quantity in Equation~\ref{equation2} if $n$ is sufficiently large. However, since the induced subgraph on the neighbors of any vertex of $\operatorname{KG}(n,k)$ is isomorphic to $\operatorname{KG}(n-k,k)$, we have $$\chi_l(\operatorname{KG}(n,k))\geq n-3(k-1).$$ \begin{corollary} Let $\mathcal{F}$ be a hypergraph and $\alpha(\mathcal{F})$ be its independence number. Then for any prime number $p$, we have
$$\chi_l(\operatorname{KG}^2(\mathcal{F}))\geq \left\lceil{(p-1)|V(\mathcal{F})|\over p}\right\rceil-(p-1)\cdot\alpha(\mathcal{F})+1.$$ \end{corollary} \begin{proof} In view of Theorem~\ref{inequalities}, we have
$${\rm ind}_{\mathbb{Z}_p}(B_0(\operatorname{KG}^2(\mathcal{F}),\mathbb{Z}_p))+1\geq \operatorname{cd}_p(\mathcal{F})\geq |V(\mathcal{F})|-p\cdot\alpha(\mathcal{F}).$$ Now, Corollary~\ref{locallowerp} implies the assertion. \end{proof} Meunier~\cite{Meunier14} naturally generalized the definition of local chromatic number of graphs to uniform hypergraphs as follows. Let ${\mathcal H}$ be a uniform hypergraph. For a set $X\subseteq V({\mathcal H})$, the closed neighborhood of $X$ is the set $X\cup {\mathcal N}(X),$ where $${\mathcal N}(X)=\{v\in V({\mathcal H}):\; \exists\; e\in E({\mathcal H})\mbox{ such that } e\setminus X=\{v\}\}.$$ For a uniform hypergraph ${\mathcal H}$, the local chromatic number of ${\mathcal H}$ is defined as follows:
$$\chi_l({\mathcal H})=\displaystyle\min_c\max\{|c({\mathcal N}[e\setminus \{v\}])|:\; e\in E({\mathcal H})\mbox{ and } v\in e\},$$ where the minimum is taken over all proper coloring $c$ of ${\mathcal H}$.
Meunier~\cite{Meunier14}, by using his colorful theorem (Theorem~\ref{colorfulhyper}), generalized Simonyi and Tardos lower bound~\cite{MR2351519} for the local chromatic number of Kneser graphs to the local chromatic number of Kneser hypergraphs. He proved: $$
\chi_l(\operatorname{KG}^p({\mathcal H}))\geq \min\left(\left\lceil{|V({\mathcal H})|-\operatorname{alt}_p({\mathcal H})\over p}\right\rceil+1, \left\lceil{|V({\mathcal H})|-\operatorname{alt}_p({\mathcal H})\over p-1}\right\rceil \right)$$ for any hypergraph ${\mathcal H}$ and any prime number $p$. In what follows, we generalize this result. \begin{theorem}\label{localhyper} Let ${\mathcal H}$ be an $r$-uniform hypergraph with at least one edge and $p$ be a prime number, where $r\leq p\leq \omega({\mathcal H})$. Let $t={\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))+p$. If $t=ap+b$, where $a$ and $b$ are nonnegative integers and $0\leq b\leq p-1$, then $$\chi_l({\mathcal H})\geq \min\left(\left\lceil{(p-r+1)a+\min\{p-r+1,b\}\over r-1}\right\rceil+1, \left\lceil{t\over r-1}\right\rceil \right).$$ \end{theorem} \begin{proof} Let $c$ be an arbitrary proper coloring of ${\mathcal H}$ and let ${\mathcal H}[U_1,\ldots,U_p]$ be the colorful balanced complete $r$-uniform $p$-partite subhypergraph of ${\mathcal H}$ whose existence is ensured by Theorem~\ref{maincolorfulindex}. Note that $b$ numbers of $U_i$'s, say $U_1,\ldots,U_b$, have the cardinality $\lceil{t\over r}\rceil$ while the others have the cardinality $\lfloor{t\over r}\rfloor\geq 1$. Consider $U_1,\ldots,U_{p-r+1}$. Two different cases will be distinguished. \begin{itemize}
\item[{\bf Case 1.}] If $\left|\displaystyle\bigcup_{i=1}^{p-r+1} c(U_i)\right|<\left\lceil{t\over r-1}\right\rceil$, then there is a vertex $v\in \displaystyle\bigcup_{i=p-r+2}^p U_i$ whose color is~not in $\displaystyle\bigcup_{i=1}^{p-r+1} c(U_i)$. Consider an edge of ${\mathcal H}[U_1,\ldots,U_p]$ containing $v$
and such that $|e\cap U_{p-r+1}|=1$ and $e\cap U_i=\varnothing$ for $i=1,\ldots, p-r$. Let $e\cap U_{p-r+1}=\{u\}$. One can check that $$\{c(v)\}\cup\displaystyle\left(\bigcup_{i=1}^{p-r+1} c(U_i)\right)\subseteq c({\mathcal N}(e\setminus \{u\})).$$ Therefore, since any color is appeared in at most $r-1$ number of $U_i$'s, we have
$$\left|\bigcup_{i=1}^{p-r+1} c(U_i)\right|\geq \displaystyle\left\lceil{\sum_{i=1}^{p-r+1}|U_i|\over r-1}\right\rceil,$$ and consequently,
$$|c({\mathcal N}(e\setminus \{u\}))|\geq 1+\displaystyle\left\lceil{\sum_{i=1}^{p-r+1}|U_i|\over r-1}\right\rceil=1+\left\lceil{(p-r+1)a+\min\{p-r+1,b\}\over r-1}\right\rceil,$$ which completes the proof in Case~1.
\item[{\bf Case 2.}] If $\left|\displaystyle\bigcup_{i=1}^{p-r+1} c(U_i)\right|\geq\left\lceil{t\over r-1}\right\rceil$, then consider an edge of ${\mathcal H}[U_1,\ldots,U_p]$ such that
$|e\cap U_{p-r+1}|=1$ and $e\cap U_i=\varnothing$ for $i=1,\ldots, p-r$. Let $e\cap U_{p-r+1}=\{u\}$. One can see that $$\displaystyle\bigcup_{i=1}^{p-r+1} c(U_i)\subseteq c({\mathcal N}(e\setminus \{u\})),$$ which completes the proof in Case~2. \end{itemize} \end{proof}
\begin{corollary} Let ${\mathcal H}$ be a $p$-uniform hypergraph with at least one edge, where $p$ is a prime number. Then $$\chi_l({\mathcal H})\geq \min\left(\left\lceil{{\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^p_p,{\mathcal H}))+p\over p}\right\rceil+1, \left\lceil{{\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^p_p,{\mathcal H}))+p\over p-1}\right\rceil \right).$$ \end{corollary} \begin{proof} Since ${\mathcal H}$ has at least one edge, we have $\omega({\mathcal H})\geq p$. Therefore, in view of Theorem~\ref{localhyper}, we have the assertion. \end{proof} Note that if ${\mathcal H}=\operatorname{KG}^p(\mathcal{F})$, then, in view of Theorem~\ref{inequalities}, we have
$${\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^p_p,{\mathcal H}))+p\geq |V(\mathcal{F})|-\operatorname{alt}_p(\mathcal{F}).$$ This implies that the previous corollary is a generalization of Meunier's lower bound for the local chromatic number of $\operatorname{KG}^p(\mathcal{F})$
\subsection{\bf Plan} Section~\ref{intro} contains some backgrounds and essential definitions used elsewhere in the paper. In Section~\ref{definition}, we present some new topological tools which help us for the proofs of main results. Section~\ref{sec:proofs} is devoted to the proofs of Theorem~\ref{maincolorfulindex} and Theorem~\ref{inequalities}.
\section{\bf Preliminaries}\label{intro}
\subsection{{\bf Topological Indices and Lower Bound for Chromatic Number}}
We assume basic knowledge in combinatorial algebraic topology. Here, we are going to bring a brief review of some essential notations and definitions which will be needed throughout the paper. For more, one can see the book written by Matou{\v{s}}ek~\cite{MR1988723}. Also, the definitions of box-complex, hom-complex, and cross-index will be generalized to $\mathbb{Z}_p$-box-complex, $\mathbb{Z}_p$-hom-complex, and $\mathbb{Z}_p$-cross-index, respectively.
Let $\mathbb{G}$ be a finite nontrivial group which acts on a topological space $X$. We call $X$ a {\it topological $\mathbb{G}$-space} if for each $g\in \mathbb{G}$, the map $g:X\longrightarrow X$ which $x\mapsto g\cdot x$ is continuous. A {\it free topological $\mathbb{G}$-space $X$} is a topological $\mathbb{G}$-space such that $\mathbb{G}$ acts on it freely, i.e., for each $g\in \mathbb{G}\setminus\{e\}$, the map $g:X\longrightarrow X$ has no fixed point. For two topological $\mathbb{G}$-spaces $X$ and $Y$, a continuous map $f:X\longrightarrow Y$ is called a $\mathbb{G}$-map if $f(g\cdot x)=g\cdot f(x)$ for each $g\in \mathbb{G}$ and $x\in X$. We write $X\stackrel{\mathbb{G}}{\longrightarrow} Y$ to mention that there is a $\mathbb{G}$-map from $X$ to $Y$. A map $f:X\longrightarrow Y$ is called a $\mathbb{G}$-equivariant map, if $f(g\cdot x)=g\cdot f(x)$ for each $g\in \mathbb{G}$ and $x\in X$.
Simplicial complexes provide a bridge between combinatorics and topology. A simplicial complex can be viewed as a combinatorial object, called abstract simplicial complex, or as a topological space, called geometric simplicial complex. Here, we just remind the definition of an abstract simplicial complex. However, we assume that the reader is familiar with the concept of how an abstract simplicial complex and its geometric realization are connected to each other. A {\it simplicial complex} is a pair $(V,K)$, where $V$ is a finite set and $K$ is a family of subsets of $V$ such that if $F\in K$ and $F'\subseteq F$, then $F'\in K$. Any set in $K$ is called a simplex. Since we may assume that $V=\bigcup_{F\in K}F$, we can write $K$ instead of $(V, K)$. The {\it dimension of $K$} is defined as follows:
$${\rm dim}(K)=\max\{|F|-1:\; F\in K\}.$$
The geometric realization of $K$ is denoted by $||K||$. For two simplicial complexes $C$ and $K$, by a {\it simplicial map $f:C\longrightarrow K$,} we mean a map from $V(C)$ to $V(K)$ such that the image of any simplex of $C$ is a simplex of $K$. For a nontrivial finite group $\mathbb{G}$, a {\it simplicial $\mathbb{G}$-complex} $K$ is a simplicial complex with a $\mathbb{G}$-action on its vertices such that each $g\in \mathbb{G}$ induces a simplicial map from $K$ to $K$, that is the map which maps $v$ to $g\cdot v$ for each $v\in V(K)$. If for each $g\in \mathbb{G}\setminus\{e\}$, there is no fixed simplex under the simplicial map made by $g$, then $K$ is called a
{\it free simplicial $\mathbb{G}$-complex.} For a simplicial $\mathbb{G}$-complex $K$, if we take the affine extension, then $K$ is free if and only if $||K||$ is free. For two simplicial $\mathbb{G}$-complexes $C$ and $K$, a simplicial map $f:C\longrightarrow K$ is called a simplicial $\mathbb{G}$-map if $f(g\cdot v)=g\cdot f(v)$ for each $g\in \mathbb{G}$ and $v\in V(C)$. We write $C\stackrel{\mathbb{G}} {\longrightarrow} K$, if there is a simplicial $\mathbb{G}$-map from $C$ to $K$. Note that if $C\stackrel{\mathbb{G}}{\longrightarrow} K$,
then $||C||\stackrel{\mathbb{G}}{\longrightarrow} ||K||$. A map $f:C\longrightarrow K$ is called a $\mathbb{G}$-equivariant map, if $f(g\cdot v)=g\cdot f(v)$ for each $g\in \mathbb{G}$ and $v\in V(C)$.
For an integer $n\geq 0$ and a nontrivial finite group $\mathbb{G}$, {\it $E_n \mathbb{G}$ space} is a free $(n-1)$-connected $n$-dimensional simplicial $\mathbb{G}$-complexes. A concrete example of an $E_n \mathbb{G}$ space is the $(n+1)$-fold join $\mathbb{G}^{*(n+1)}$. As a topological space $\mathbb{G}^{*(n+1)}$ is a $(n+1)$-fold join of an $(n+1)$-point discrete space. This is known that for any two $E_n \mathbb{G}$ space $X$ and $Y$, there is a $\mathbb{G}$-map from $X$ to $Y$.
For a $\mathbb{G}$-space $X$, define $${\rm ind}_{\mathbb{G}}(X)=\min\{n:\; X\stackrel{\mathbb{G}}{\longrightarrow} E_n\mathbb{G}\}.$$ Note that here $E_n \mathbb{G}$ can be any $E_n \mathbb{G}$, since there is a $\mathbb{G}$-map between any two $E_n \mathbb{G}$ spaces, see~\cite{MR1988723}. Also, for a simplicial complex $K$, by ${\rm ind}_{\mathbb{G}}(K)$, we mean
${\rm ind}_{\mathbb{G}}(||K||)$. Throughout the paper, for $\mathbb{G}=\mathbb{Z}_2$, we would rather use ${\rm ind}(-)$ instead of ${\rm ind}_{\mathbb{Z}_2}(-)$.
\noindent{\bf Properties of the $\mathbb{G}$-index.} \cite{MR1988723} Let $\mathbb{G}$ be a finite nontrivial group. \begin{itemize} \item[{\rm (i)}] ${\rm ind}_{\mathbb{G}}(X)>{\rm ind}_{\mathbb{G}}(Y)$ implies $X\stackrel{\mathbb{G}}{\centernot\longrightarrow} Y$. \item[{\rm (ii)}] ${\rm ind}_{\mathbb{G}}(E_n \mathbb{G})=n$ for any $E_n \mathbb{G}$ space. \item[{\rm (iii)}] ${\rm ind}_{\mathbb{G}}(X*Y)\leq {\rm ind}_{\mathbb{G}}(X)+{\rm ind}_{\mathbb{G}}(Y)+1$. \item[{\rm (iv)}] If $X$ is $(n-1)$-connected, then ${\rm ind}_{\mathbb{G}}(X)\geq n$. \item[{\rm (v)}] If $K$ is a free simplicial $\mathbb{G}$-complex of dimension $n$, then ${\rm ind}_{\mathbb{G}}(K)\leq n$. \end{itemize}
\subsection{{\bf $\mathbb{Z}_p$-Box-Complex, $\mathbb{Z}_p$-Poset, and $\mathbb{Z}_p$-Hom-Complex}}\label{Boxdefin} In this subsection, for any $r$-uniform hypergraph ${\mathcal H}$, we are going to define two objects; $\mathbb{Z}_p$-box-complex of ${\mathcal H}$ and $\mathbb{Z}_p$-hom-complex of ${\mathcal H}$ which the first one is a simplicial $\mathbb{Z}_p$-complex and the second one is a $\mathbb{Z}_p$-poset. Moreover, for any $\mathbb{Z}_p$-poset $P$, we assign a combinatorial index to $P$ called the cross-index of $P$. \\
\noindent{\bf $\mathbb{Z}_p$-Box-Complex.} Let $r\geq 2$ be a positive integer and $p\geq r$ be a prime number. For an $r$-uniform hypergraph ${\mathcal H}$, define the {\it $\mathbb{Z}_p$-box-complex of ${\mathcal H}$,} denoted ${\rm B}_0({\mathcal H},{\mathbb{Z}_p})$, to be a simplicial complex with the vertex set $\displaystyle\biguplus_{i=1}^pV({\mathcal H})=\mathbb{Z}_p\times V({\mathcal H})$ and the simplex set consisting of all $\{\omega^1\}\times U_1\cup\cdots\cup \{\omega^p\}\times U_p$, where
\begin{itemize} \item $U_1,\ldots,U_p$ are pairwise disjoint subsets of $V({\mathcal H})$, \item $\displaystyle\bigcup_{i=1}^p U_i\neq\varnothing$, and \item the hypergraph ${\mathcal H}[U_1,U_2,\ldots,U_p]$ is a complete $r$-uniform $p$-partite hypergraph.
\end{itemize} Note that some of $U_i$'s might be empty. In fact, if $U_1,\ldots,U_p$ are pairwise disjoint subsets of $V({\mathcal H})$ and the number of nonempty $U_i$'s is less than $r$, then ${\mathcal H}[U_1,U_2,\ldots,U_p]$ is a complete $r$-uniform $p$-partite hypergraph and thus $\{\omega^1\}\times U_1\cup\cdots\cup \{\omega^p\}\times U_p\in {\rm B}_0({\mathcal H},{\mathbb{Z}_p})$. For each $\epsilon\in{\mathbb{Z}_p}$ and each $(\epsilon',v)\in V({\rm B}_0({\mathcal H},{\mathbb{Z}_p}))$, define $\epsilon\cdot(\epsilon',v)=(\epsilon\cdot\epsilon',v)$. One can see that this action makes ${\rm B}_0({\mathcal H},{\mathbb{Z}_p})$ a free simplicial $\mathbb{Z}_p$-complex. It should be mentioned that the $\mathbb{Z}_2$-box-complex ${\rm B}_0({\mathcal H},\mathbb{Z}_2)$ is extensively studied in the literature, see~\cite{MR2279672,MR2351519}. In the literature, for a graph $G$, the simplicial complex ${\rm B}_0(G,{\mathbb{Z}_2})$ is shown by $B_0(G)$. This simplicial complex is used to introduce some lower bounds for the chromatic number of a given graph $G$, see~\cite{MR2279672}. In particular, we have the following inequalities $$\chi(G)\geq {\rm ind}(B_0(G))+1\geq {\rm coind}(B_0(G))+1\geq n-{\rm alt}({\mathcal F})\geq{\rm cd}_2({\mathcal F}),$$ where ${\mathcal F}$ is any hypergraph such that ${\rm KG}^2({\mathcal F})$ and $G$ are isomorphic, see~\cite{2014arXiv1403.4404A,2013arXiv1302.5394A,MR2279672}.\\
\noindent{\bf $\mathbb{Z}_p$-Poset.} A partially ordered set, or simply a {\it poset}, is defined as an ordered pair $P=(V(P),\preceq)$, where $V(P)$ is a set called the ground set of $P$ and $\preceq$ is a partial order on $V(P)$. For two posets $P$ and $Q$, by an order-preserving map $\phi:P\longrightarrow Q$, we mean a map $\phi$ from $V(P)$ to $V(Q)$ such that for each $u,v\in V(P)$, if $u\preceq v$, then $\phi(u)\preceq \phi(v)$.
A poset $P$ is called a {\it $\mathbb{Z}_p$-poset}, if $\mathbb{Z}_p$ acts on $V(P)$
and furthermore, for each $\epsilon\in \mathbb{Z}_p$, the map $\epsilon:V(P)\longrightarrow V(P)$ which $v\mapsto \epsilon\cdot v$ is an automorphism of $P$ (order preserving bijective map). If for each $\epsilon\in \mathbb{Z}_p\setminus\{e\}$, this map has no fixed point, then $P$ is called a {\it free $\mathbb{Z}_p$-poset}. For two $\mathbb{Z}_p$-poset $P$ and $Q$, by an order-preserving $\mathbb{Z}_p$-map $\phi:P\longrightarrow Q$, we mean
an order-preserving map from $V(P)$ to $V(Q)$ such that for each $v\in V(P)$ and $\epsilon\in \mathbb{Z}_p$, we have $\phi(\epsilon\cdot v)=\epsilon\cdot\phi(v)$.
If there exists such a map, we write $P\stackrel{\mathbb{Z}_p}{\longrightarrow} Q$.
For a nonnegative integer $n$ and a prime number $p$, let $Q_{n,p}$ be a free $\mathbb{Z}_p$-poset with ground set $\mathbb{Z}_p\times[n+1]$ such that for any two members $(\epsilon,i),(\epsilon',j)\in Q_{n,p}$, $(\epsilon,i)<_{Q_{n,p}}(\epsilon',j)$ if $i<j$. Clearly, $Q_{n,p}$ is a free $\mathbb{Z}_p$-poset with the action $\epsilon\cdot(\epsilon',j)=(\epsilon\cdot\epsilon',j)$ for each $\epsilon\in\mathbb{Z}_p$ and $(\epsilon',j)\in Q_{n,p}$. For a $\mathbb{Z}_p$-poset $P$, the {\it $\mathbb{Z}_p$-cross-index} of $P$, denoted ${\rm Xind}_{\mathbb{Z}_p}(P)$, is the least integer $n$ such that there is a $\mathbb{Z}_p$-map from $P$ to $Q_{n,p}$. Throughout the paper, for $p=2$, we speak about ${\rm Xind}(-)$ rather than ${\rm Xind}_{\mathbb{Z}_2}(-)$. It should be mentioned that ${\rm Xind}(-)$ is first defined in~\cite{SiTaZs13}.
Let $P$ be a poset. We can define an {\it order complex} $\Delta P$ with the vertex set same as the ground set of $P$ and simplex set consisting of all chains in $P$. One can see that if $P$ is a free $\mathbb{Z}_p$-poset, then $\Delta P$ is a free simplicial $\mathbb{Z}_p$-complex. Moreover, any order-preserving $\mathbb{Z}_p$-map $\phi:P\longrightarrow Q$ can be lifted to a simplicial $\mathbb{Z}_p$-map from $\Delta P$ to $\Delta Q$. Clearly, there is a simplicial $\mathbb{Z}_p$-map from $\Delta Q_{n,p}$ to $\mathbb{Z}_p^{*(n+1)}$ (identity map). Therefore, if ${\rm Xind}_{\mathbb{Z}_p}(P)=n$, then we have a simplicial $\mathbb{Z}_p$-map from $\Delta P$ to $\mathbb{Z}_p^{*(n+1)}$. This implies that ${\rm Xind}_{\mathbb{Z}_p}(P)\geq {\rm ind}_{\mathbb{Z}_p}(\Delta P)$. Throughout the paper, for each $(\epsilon, j)\in Q_{n,p}$, when we speak about the sign of $(\epsilon, j)$ and the absolute value of $(\epsilon, j)$, we mean $\epsilon$ and $j$, respectively.
\begin{alphtheorem}{\rm \cite{AliHajiMeu2016}}\label{altercrossindex} Let $P$ be a free ${\mathbb Z}_2$-poset and $\phi:P\longrightarrow Q_{s,2}$ be an order preserving ${\mathbb Z}_2$-map. Then $P$ contains a chain $p_1\prec_P\cdots\prec_Pp_{ k}$ such that $k= {\rm Xind}(P)+1$ and the signs of $\phi(p_i)$ and $\phi(p_{i+1})$ differ for each $i\in[k-1]$. Moreover, if $s= {\rm Xind}(P)$, then for any $(s+1)$-tuple $(\epsilon_1,\ldots,\epsilon_{s+1})\in\mathbb{Z}_2^{s+1}$, there is at least one chain $p_1\prec_P\cdots\prec_Pp_{ s+1}$ such that $\phi(p_i)=(\epsilon_i,i)$ for each $i\in[s+1]$. \end{alphtheorem}
\noindent{\bf $\mathbb{Z}_p$-Hom-Complex.} Let ${\mathcal H}$ be an $r$-uniform hypergraph. Also, let $p\geq r$ be a prime number. The {\it $\mathbb{Z}_p$-hom-complex} ${\rm Hom}(K^r_p,{\mathcal H})$ is a free $\mathbb{Z}_p$-poset with the ground set consisting of all ordered $p$-tuples $(U_1,\cdots,U_p)$, where $U_i$'s are nonempty pairwise disjoint subsets of $V$ and ${\mathcal H}[U_1,\ldots,U_p]$ is a complete $r$-uniform $p$-partite hypergraph. For two $p$-tuples $(U_1,\cdots,U_p)$ and $(U'_1,\cdots,U'_p)$ in ${\rm Hom}(K^r_p,{\mathcal H})$, we define $(U_1,\cdots,U_p)\preceq(U'_1,\cdots,U'_p)$ if $U_i\subseteq U'_i$ for each $i\in[p]$. Also, for each $\omega^i\in \mathbb{Z}_p=\{\omega^1,\ldots,\omega^p\}$, let $\omega^i\cdot (U_1,\cdots,U_p)=(U_{1+i},\cdots,U_{p+i})$, where $U_j=U_{j-p}$ for $j>p$. Clearly, this action is a free $\mathbb{Z}_p$-action on ${\rm Hom}(K^r_p,{\mathcal H})$. Consequently, ${\rm Hom}(K^r_p,{\mathcal H})$ is a free $\mathbb{Z}_p$-poset with this $\mathbb{Z}_p$-action.
For a nonempty graph $G$ and for $p=2$, it is proved~\cite{2014arXiv1403.4404A,2013arXiv1302.5394A,SiTaZs13,MR2279672} that \begin{equation}\label{equation} \begin{array}{lll} \chi(G) &\geq & {\rm Xind}({\rm Hom}(K_2,G))+2 \geq {\rm ind}(\Delta {\rm Hom}(K_2,G))+2 \geq {\rm ind}(B_0(G))+1\\
&\geq & {\rm coind}(B_0(G))+1\geq |V({\mathcal F})|-{\rm alt}_2({\mathcal F}) \geq {\rm cd}_2({\mathcal F}), \end{array} \end{equation} where ${\mathcal F}$ is any hypergraph such that ${\rm KG}^2({\mathcal F})$ and $G$ are isomorphic.\\
\section{\bf Notations and Tools}\label{definition}
For a simplicial complex $K$, by $\operatorname{sd} K$, we mean the first barycentric subdivision of $K$. It is the simplicial complex whose vertex set is the set of nonempty simplices of $K$ and whose simplices are the collections of simplices of $K$ which are pairwise comparable by inclusion. Throughout the paper, by $\sigma_{t-1}^{r-1}$, we mean the $(t-1)$-dimensional simplicial complex with vertex set $\mathbb{Z}_r$ containing all $t$-subsets of $\mathbb{Z}_r$ as its maximal simplices. The join of two simplicial complexes $C$ and $K$, denoted $C*K$, is a simplicial complex with the vertex set $V(C)\biguplus V(K)$ and such that the set of its simplices is $\{F_1\biguplus F_2:\; F_1\in C\mbox{ and } F_2\in K\}$. Clearly, we can see $\mathbb{Z}_r$ as a $0$-dimensional simplicial complex. Note that the vertex set of simplicial complex $\operatorname{sd}\mathbb{Z}_r^{*\alpha}$ can be identified with $(\mathbb{Z}_r\cup\{0\})^\alpha\setminus\{\boldsymbol{0}\}$ and the vertex set of $(\sigma^{r-1}_{t-1})^{*n}$ is the set of all pairs $(\epsilon,i)$, where $\epsilon\in \mathbb{Z}_r$ and $i\in [n]$.
\subsection{{\bf $\mathbb{Z}_p$-Tucker-Ky Fan lemma}} The famous Borsuk-Ulam theorem has many generalizations which have been extensively used in investigating graph coloring properties. Some of these interesting generalizations are Tucker lemma~\cite{MR0020254}, $Z_p$-Tucker Lemma~\cite{MR1893009}, and Tucker-Ky Fan~\cite{MR0051506}. For more details about the Borsuk-Ulam theorem and its generalizations, we refer the reader to \cite{MR1988723}.
Actually, Tucker lemma is a combinatorial counterpart of Borsuk-Ulam theorem. There are several interesting and surprising applications of Tucker Lemma in combinatorics, including a combinatorial proof of Lov{\'a}sz-Kneser theorem by Matou{\v{s}}ek \cite{MR2057690}. \begin{alphlemma}\label{tuckeroriginal} {\rm(Tucker lemma \cite{MR0020254}).} Let $m$ and $n$ be positive integers and $\lambda:\{-1,0,+1\}^n\setminus \{(0,\ldots,0)\} \longrightarrow \{\pm 1, \pm 2,\ldots ,\pm m\}$ be a map satisfying the following properties: \begin{itemize} \item for any $X\in \{-1,0,+1\}^n\setminus \{\boldsymbol{0}\}$, we have $\lambda(-X)=-\lambda(X)$ {\rm (}a
$Z_2$-equivariant map{\rm ),} \item no two signed vectors $X$ and $Y$ are such that $X\subseteq Y$ and $\lambda(X) =-\lambda(Y)$. \end{itemize} Then, we have $m \geq n$. \end{alphlemma} Another interesting generalization of the Borsuk-Ulam theorem is Ky~Fan's lemma~\cite{MR0051506}. This generalization ensures that with the same assumptions as in Lemma~\ref{tuckeroriginal}, there is odd number of chains $X_1\subseteq X_2\subseteq \cdots \subseteq X_n$ such that $$\{\lambda(X_1),\ldots,\lambda(X_n)\}=\{+c_1,-c_2,\ldots , (- 1)^{n-1}c_n\},$$ where $1\leq c_1 < \cdots < c_n \leq m$. Ky~Fan's lemma has been used in several articles to study some coloring properties of graphs, see \cite{AliHajiMeu2016,MR2763055,MR2837625}. There are also some other generalizations of Tucker Lemma. A $\mathbb{Z}_p$ version of Tucker Lemma, called $\mathbb{Z}_p$-Tucker Lemma, is proved by Ziegler~\cite{MR1893009} and extended by Meunier~\cite{MR2793613}. In next subsection, we present a $\mathbb{Z}_p$ version of Ky~Fan's lemma which is called $\mathbb{Z}_p$-Tucker-Ky Fan lemma.
\subsection{{\bf New Generalizations of Tucker Lemma}} Before presenting our results, we need to introduce some functions having key roles in the paper. Throughout the paper, we are going to use these functions repeatedly. Let $m$ be a positive integer. We remind that $(\sigma^{p-1}_{p-2})^{*m}$ is a free simplicial $\mathbb{Z}_p$-complex with vertex set $\mathbb{Z}_p\times [m]$. \\
\noindent{\bf The value function $l(-)$.} Let $\tau\in (\sigma^{p-1}_{p-2})^{*m}$ be a simplex. For each $\epsilon\in \mathbb{Z}_p$, define $\tau^\epsilon=\left\{(\epsilon,j):\; (\epsilon,j)\in \tau\right\}.$ Moreover, define
$$l(\tau)=\max\left\{\displaystyle|\bigcup_{\epsilon\in\mathbb{Z}_p} B^\epsilon|:\; \forall\epsilon\in\mathbb{Z}_p ,\; B^\epsilon\subseteq \tau^\epsilon\mbox{ and } \forall \epsilon_1,\epsilon_2\in\mathbb{Z}_p,\; |\;|B^{\epsilon_1}|-|B^{\epsilon_2}|\;|\leq 1 \right\}.$$
Note that if we set $h(\tau)=\displaystyle\min_{\epsilon\in \mathbb{Z}_p}|\tau^\epsilon|$, then
$$l(\tau)=p\cdot h(\tau)+|\{\epsilon\in\mathbb{Z}_p:\; |\tau^\epsilon|>h(\tau)\}|.$$\\
\noindent{\bf The sign functions $s(-)$ and $s_0(-)$.} For an $a\in[m]$,
let $W_a$ be the set of all simplices $\tau\in (\sigma_{p-2}^{p-1})^{*m}$ such that $|\tau^\epsilon|\in\{0,a\}$ for each $\epsilon\in\mathbb{Z}_p$. Let $W=\displaystyle\bigcup_{a=1}^{m}W_a$. Choose an arbitrary $\mathbb{Z}_p$-equivariant map $s:W\longrightarrow \mathbb{Z}_p$. Also, consider an $\mathbb{Z}_p$-equivariant map $s_0:\sigma_{p-2}^{p-1}\longrightarrow \mathbb{Z}_p$. Note that since $\mathbb{Z}_p$ acts freely on both $\sigma_{p-2}^{p-1}$ and $W$, these maps can be easily built by choosing one representative in each orbit. It should be mentioned that both functions $s(-)$ and $s_0(-)$ are first introduced in~\cite{Meunier14}.
Now, we are in a position to generalize Tucker-Ky Fan lemma to $\mathbb{Z}_p$-Tucker-Ky Fan lemma. \begin{lemma}{\rm ($\mathbb{Z}_p$-Tucker-Ky Fan lemma).}\label{Z_pfanlemma} Let $m,n,p$ and $\alpha$ be nonnegative integers, where $m,n\geq 1$, $m\geq \alpha\geq 1$, and $p$ is prime. Let $$ \begin{array}{crcl} \lambda: & (\mathbb{Z}_p\cup\{0\})^n\setminus\{\boldsymbol{0}\} &\longrightarrow & \mathbb{Z}_p\times[m]\\ & X &\longmapsto & (\lambda_1(X),\lambda_2(X)) \end{array}$$ be a $\mathbb{Z}_p$-equivariant map satisfying the following conditions. \begin{itemize} \item For $X_1\subseteq X_2\in \left(\mathbb{Z}_p\cup\{0\}\right)^n\setminus\{\boldsymbol{0}\}$, if $\lambda_2(X_1)=\lambda_2(X_2)\leq \alpha$, then $\lambda_1(X_1)=\lambda_1(X_2)$. \item For $X_1\subseteq X_2\subseteq\cdots \subseteq X_p\in \left(\mathbb{Z}_p\cup\{0\}\right)^n\setminus\{\boldsymbol{0}\}$, if $\lambda_2(X_1)=\lambda_2(X_2)=\cdots=\lambda_2(X_p)\geq\alpha+1$, then
$$\left|\left\{\lambda_1(X_1),\lambda_1(X_2),\ldots,\lambda_1(X_p)\right\}\right|<p.$$ \end{itemize} Then there is a chain $$Z_1\subset Z_2\subset\cdots\subset Z_{n-\alpha}\in \left(\mathbb{Z}_p\cup\{0\}\right)^n\setminus\{\boldsymbol{0}\}$$ such that \begin{enumerate} \item for each $i\in [n-\alpha]$, $\lambda_2(Z_i)\geq \alpha+1$, \item for each $i\neq j\in [n-\alpha]$, $\lambda(Z_i)\neq \lambda(Z_j)$, and \item\label{condition3} for each $\epsilon\in\mathbb{Z}_p$,
$$\left\lfloor{n-\alpha\over p}\right\rfloor\leq \left|\left\{j:\; \lambda_1(Z_j)=\epsilon\right\}\right|\leq \left\lceil{n-\alpha\over p}\right\rceil.$$ \end{enumerate} In particular, $n-\alpha\leq (p-1)(m-\alpha)$. \end{lemma} \begin{proof} Note that the map $\lambda$ can be considered as a simplicial $\mathbb{Z}_p$-map from $\operatorname{sd} \mathbb{Z}_p^{*n}$ to $(\mathbb{Z}_p^{*\alpha})*((\sigma_{p-2}^{p-1})^{*(m-\alpha)}).$ Let $K={\rm Im}(\lambda)$. Note that each simplex in $K$ can be represented in a unique form $\sigma\cup\tau$ such that $\sigma\in \mathbb{Z}_p^{*\alpha}$ and $\tau \in (\sigma_{p-2}^{p-1})^{*m-\alpha}.$
In view of definition of the function $l(-)$ and the properties which $\lambda$ satisfies in, to prove the assertion, it suffices to show that there is a simplex $\sigma\cup\tau\in K$ such that $l(\tau)\geq n-\alpha$. For a contradiction, suppose that for each $\sigma\cup\tau\in K$, we have $l(\tau)\leq n-\alpha-1$.
Define the map $$\Gamma: \operatorname{sd} K\longrightarrow \mathbb{Z}_p^{*(n-1)}$$ such that for each vertex $\sigma\cup\tau\in V(\operatorname{sd} K)$, \begin{itemize} \item if $\tau=\varnothing$, then $\Gamma(\sigma\cup\tau)=(\epsilon, j)$, where $j$ is the maximum possible value such that $(\epsilon, j)\in\sigma$. Note that since $\sigma\in \mathbb{Z}_p^{*\alpha}$, there is only one $\epsilon\in\mathbb{Z}_p$ for which
the maximum is attained. Therefore, in this case, the function $\Gamma$ is well-defined.
\item if $\tau\neq\varnothing$. Define $h(\tau)=\displaystyle\min_{\epsilon\in \mathbb{Z}_p}|\tau^\epsilon|.$ \begin{enumerate}[label={\rm (\roman*)}] \item If $h(\tau)=0$, then define $\bar{\tau}=\{\epsilon\in \mathbb{Z}_p:\;
\tau^\epsilon= \varnothing\}\in \sigma^{p-1}_{p-2}$ and
$$\Gamma(\sigma\cup\tau)=\left(s_0(\bar\tau), \alpha+l(\tau)\right).$$
\item If $h(\tau)> 0$, then define $\bar{\tau}=\displaystyle
\bigcup_{\{\epsilon\in\mathbb{Z}_p:\; |\tau^\epsilon|=h(\tau)\}} \tau^\epsilon\in W$
and $$\Gamma(\sigma\cup\tau)=\left(s(\bar\tau), \alpha+l(\tau)\right).$$
\end{enumerate} \end{itemize} Now, we claim that $\Gamma$ is a simplicial $\mathbb{Z}_p$-map from $\operatorname{sd} K$ to $\mathbb{Z}_p^{*(n-1)}$. It is clear that $\Gamma$ is a $\mathbb{Z}_p$-equivariant map. For a contradiction, suppose that there are $\sigma\cup\tau,\sigma'\cup\tau' \in \operatorname{sd} K$ such that $\sigma\subseteq \sigma'$, $\tau\subseteq\tau'$, $\Gamma(\sigma\cup\tau)=(\epsilon,\beta)$, and $\Gamma(\sigma'\cup\tau')=(\epsilon',\beta)$, where $\epsilon\neq \epsilon'$. First note that in view of the definition of $\Gamma$ and the assumption $\Gamma(\sigma\cup\tau)=(\epsilon,\beta)$ and $\Gamma(\sigma'\cup\tau')=(\epsilon',\beta)$, the case $\tau=\varnothing$ and $\tau'\neq\varnothing$ is not possible. If $\tau'=\varnothing$, then $\tau=\tau'=\varnothing$ and we should have $(\epsilon,\beta),(\epsilon',\beta)\in\sigma'\in \mathbb{Z}_p^{*\alpha}$ which implies that $\epsilon=\epsilon'$, a contradiction. If $\varnothing\neq \tau\subseteq \tau'$, then in view of definition of $\Gamma$, we should have $l(\tau)=l(\tau')$. We consider three different cases.\\ \begin{enumerate}[label={\rm (\roman*)}] \item If $h(\tau)=h(\tau')=0$, then $$\epsilon=s_0(\{\epsilon\in \mathbb{Z}_p:\;
\tau^\epsilon= \varnothing\})\neq s_0(\{\epsilon\in \mathbb{Z}_p:\;
{\tau'}^\epsilon= \varnothing\})=\epsilon'.$$
Therefore, $ \{\epsilon\in \mathbb{Z}_p:\;
{\tau'}^\epsilon= \varnothing\}\subsetneq\{\epsilon\in \mathbb{Z}_p:\;
\tau^\epsilon= \varnothing\}$. This implies that
$$l(\tau')=p-|\{\epsilon\in \mathbb{Z}_p:\;
{\tau'}^\epsilon= \varnothing\}|>p-|\{\epsilon\in \mathbb{Z}_p:\;
\tau^\epsilon= \varnothing\}|=l(\tau),$$
a contradiction.\\
\item If $h(\tau)=0$ and $h(\tau')>0.$ We should have
$l(\tau)\leq p-1$ and $l(\tau')\geq p$ which contradicts the fact that $l(\tau)=l(\tau')$.\\
\item If $h(\tau)>0$ and $h(\tau')>0.$
Note that
$$ l(\tau)=p\cdot h(\tau)+|\{\epsilon\in\mathbb{Z}_p:\; |\tau^\epsilon|>h(\tau)\}|
\mbox{ and } l(\tau')=p\cdot h(\tau')+|\{\epsilon\in\mathbb{Z}_p:\;
|{\tau'}^\epsilon|>h(\tau')\}|.$$ For this case, two different sub-cases will be distinguished.
\begin{itemize}
\item[(a)] If $h(\tau)=h(\tau')=h$, then
$$\epsilon=s(\displaystyle\bigcup_{\{\epsilon\in\mathbb{Z}_p:\; |\tau^\epsilon|=h\}} \tau^\epsilon)\neq s(\displaystyle\bigcup_{\{\epsilon\in\mathbb{Z}_p:\; |{\tau'}^\epsilon|=h\}} {\tau'}^\epsilon)=\epsilon'.$$
Clearly, it implies that $$\displaystyle\bigcup_{\{\epsilon\in\mathbb{Z}_p:\; |\tau^\epsilon|=h\}} \tau^\epsilon\neq \displaystyle\bigcup_{\{\epsilon\in\mathbb{Z}_p:\; |{\tau'}^\epsilon|=h\}} {\tau'}^\epsilon.$$
Note that $\tau\subseteq \tau'$ and $h=\displaystyle\min_{\epsilon\in \mathbb{Z}_p}|\tau^\epsilon|=\displaystyle\min_{\epsilon\in \mathbb{Z}_p}|{\tau'}^\epsilon|.$ Therefore, we should have
$$
\{\epsilon\in\mathbb{Z}_p:\; |{\tau'}^\epsilon|=h\} \subsetneq \{\epsilon\in\mathbb{Z}_p:\; |{\tau}^\epsilon|=h\}$$
and consequently $l(\tau)<l(\tau')$ which is a contradiction.
\item[(b)] If $h(\tau)<h(\tau')$, then
$$l(\tau)\leq p\cdot h(\tau)+p-1< p\cdot (h(\tau)+1)\leq l(\tau'),$$
a contradiction.
\end{itemize} \end{enumerate} Therefore, $\Gamma$ is a simplicial $\mathbb{Z}_p$-map from $\operatorname{sd} K$ to $\mathbb{Z}_p^{*(n-1)}$. Naturally, $\lambda$ can be lifted to a simplicial $\mathbb{Z}_p$-map $\bar\lambda:\operatorname{sd}^2 \mathbb{Z}_p^{*n}\longrightarrow \operatorname{sd} K$. Thus $\Gamma\circ\bar\lambda$ is a simplicial $\mathbb{Z}_p$-map from $\operatorname{sd}^2 \mathbb{Z}_p^{*n}$ to $\mathbb{Z}_p^{*(n-1)}$. In view of Dold's theorem~\cite{MR711043,MR1988723}, the dimension of $\mathbb{Z}_p^{*(n-1)}$ should be strictly larger than the connectivity of $\operatorname{sd}^2 \mathbb{Z}_p^{*n}$, that is $n-2>n-2$, which is not possible. \end{proof}
Lemma~\ref{Z_pfanlemma} provides a short simple proof of Meunier's colorful result for Kneser hypergraphs (next Theorem) as follows. \begin{alphtheorem}{\rm \cite{Meunier14}}\label{colorfulhyper} Let ${\mathcal H}$ be a hypergraph and let $p$ be a prime number. Then any proper coloring $c:V({\rm KG}^p({\mathcal H}))\longrightarrow [C]$ {\rm(}$C$ arbitrary{\rm)} must contain a colorful balanced complete $p$-uniform $p$-partite hypergraph with
$|V({\mathcal H})|-{\rm alt}_p({\mathcal H})$ vertices. \end{alphtheorem} \begin{proof} Consider a bijection $\pi:[n]\longrightarrow V({\mathcal H})$ such that ${\rm alt}_p({\mathcal H},\pi)={\rm alt}_p({\mathcal H}).$ We are going to define a map $$\begin{array}{cccc} \lambda: & (\mathbb{Z}_p\cup\{0\})^n\setminus\{\boldsymbol{0}\} &\longrightarrow & \mathbb{Z}_p\times[m]\\ & X &\longmapsto & (\lambda_1(X),\lambda_2(X)) \end{array}$$ satisfying the conditions of Lemma~\ref{Z_pfanlemma}
and with parameters $n= |V({\mathcal H})|$, $m={\rm alt}_p({\mathcal H})+C$, and $\alpha={\rm alt}_p({\mathcal H})$. Assume that $2^{[n]}$ is equipped with a total ordering $\preceq$. For each $X\in(\mathbb{Z}_p\cup\{0\})^n\setminus\{\boldsymbol{0}\}$, define $\lambda(X)$ as follows. \begin{itemize} \item If ${\rm alt}(X)\leq {\rm alt}_p({\mathcal H},\pi)$, then let $\lambda_1(X)$ be the first nonzero coordinate of $X$ and $\lambda_2(X)={\rm alt}(X)$. \item If ${\rm alt}(X)\geq {\rm alt}_p({\mathcal H},\pi)+1$, then in view of the definition of ${\rm alt}_p({\mathcal H},\pi)$, there is some $\epsilon\in\mathbb{Z}_p$ such that $E(\pi(X^\epsilon))\neq \varnothing$. Define $$c(X)=\max\left\{c(e):\; \exists\epsilon\in\mathbb{Z}_p\mbox { such that } e\subseteq \pi(X^\epsilon)\right\}$$ and $\lambda_2(X)={\rm alt}_p({\mathcal H},\pi)+c(X)$. Choose $\epsilon\in\mathbb{Z}_p$ such that there is at least one edge $e\in\pi (X^\epsilon)$ with $c(X)=c(e)$ and such that
$X^\epsilon$ is the maximum one having this property. By the maximum, we mean
the maximum according to the total ordering $\preceq$. It is clear that $\epsilon$ is defined uniquely. Now, let $\lambda_1(X)=\epsilon$. \end{itemize} One can check that $\lambda$ satisfies the conditions of Lemma~\ref{Z_pfanlemma}. Consider the chain $Z_1\subset Z_2\subset\cdots\subset Z_{n-{\rm alt}_p({\mathcal H},\pi)}$ whose existence is ensured by Lemma~\ref{Z_pfanlemma}. Note that for each $i\in[n-{\rm alt}_p({\mathcal H},\pi)]$, we have $\lambda_2(Z_i)>{\rm alt}_p({\mathcal H},\pi)$. Consequently, $\lambda_2(Z_i)={\rm alt}_p({\mathcal H},\pi)+c(Z_i)$. Let $\lambda(Z_i)=(\epsilon_i,j_i)$. Note that for each $i$, there is at least one edge $e_{i,\epsilon_i}\subseteq \pi(Z_i^{\epsilon_i})\subseteq \pi(Z_{n-{\rm alt}_p({\mathcal H},\pi)}^{\epsilon_i})$ such that $c(e_{i,\epsilon_i})=j_i-{\rm alt}_p({\mathcal H},\pi)$. For each $\epsilon\in\mathbb{Z}_p$, define $U_\epsilon=\{e_{i,\epsilon_i}:\; \epsilon_i=\epsilon\}.$ We have the following three properties for $U_\epsilon$'s. \begin{itemize} \item Since the chain $Z_1\subset Z_2\subset\cdots\subset
Z_{n-{\rm alt}_p({\mathcal H},\pi)}$ is satisfying Condition~\ref{condition3} of
Lemma~\ref{Z_pfanlemma}, we have
$\left\lfloor{n-{\rm alt}_p({\mathcal H},\pi)\over p}\right\rfloor\leq
|U_\epsilon|\leq \left\lceil{n-{\rm alt}_p({\mathcal H},\pi)\over p}\right\rceil.$ \item The edges in $U_\epsilon$ get distinct colors.
If there are two edges $e_{i,\epsilon}$ and $e_{i',\epsilon}$ in $U_\epsilon$ such that
$c(e_{i,\epsilon})=c(e_{i',\epsilon})$, then $\lambda(Z_i)=\lambda(Z_{i'})$
which is not possible. \item If $\epsilon\neq \epsilon'$, then for each $e\in U_\epsilon$ and $f\in U_{\epsilon'}$,
we have $e\cap f=\varnothing$. It is clear because
$e\subseteq\pi(Z_{n-{\rm alt}_p({\mathcal H},\pi)}^\epsilon)$,
$f\subseteq\pi(Z_{n-{\rm alt}_p({\mathcal H},\pi)}^{\epsilon'})$,
and
$$\pi(Z_{n-{\rm alt}_p({\mathcal H},\pi)}^\epsilon)\cap
\pi(Z_{n-{\rm alt}_p({\mathcal H},\pi)}^{\epsilon'})=\varnothing.$$ \end{itemize} Now, it is clear that the subhypergraph ${\rm KG}^p({\mathcal H})[U_{\omega^1},\ldots,U_{\omega^p}]$ is the desired subhypergraph. \end{proof}
The proof of next lemma is similar to the proof of Lemma~\ref{Z_pfanlemma}. \begin{lemma}\label{genfanlemma} Let $C$ be a free simplicial $\mathbb{Z}_p$-complex such that ${\rm ind}_{\mathbb{Z}_p}(C)\geq t$ and let $\lambda:C\longrightarrow (\sigma^{p-1}_{p-2})^{*m}$ be a simplicial $\mathbb{Z}_p$-map. Then there is at least one $t$-dimensional simplex $\sigma\in C$ such that $\tau=\lambda(\sigma)$ is a $t$-dimensional simplex and for each $\epsilon\in \mathbb{Z}_p$, we have
$\lfloor{t+1\over p}\rfloor\leq |\tau^\epsilon|\leq\lceil{t+1\over p}\rceil.$ \end{lemma} \begin{proof} For simplicity of notation, let $K={\rm Im}(\lambda)$. Clearly, to prove the assertion, it is enough to show that there is a $t$-dimensional simplex $\tau\in K$
such that $l(\tau)\geq t$. Suppose, contrary to the assertion, that there is no such a $t$-dimensional simplex. Therefore, for each simplex $\tau$ of $K$, we have $l(\tau)\leq t$. For each vertex $\tau\in V(\operatorname{sd} K)$, set $h(\tau)=\displaystyle\min_{\epsilon\in \mathbb{Z}_p}|\tau^\epsilon|$.
Let $\Gamma:\operatorname{sd} K\longrightarrow \mathbb{Z}_p^{*t}$ be a map such that for each vertex $\tau$ of $\operatorname{sd} K$, $\Gamma(\tau)$ is defined as follows. \begin{enumerate}[label={\rm (\roman*)}] \item If $h(\tau)=0$, then define $\bar{\tau}=\{\epsilon\in \mathbb{Z}_p:\;
\tau^\epsilon= \varnothing\}\in \sigma^{p-1}_{p-2}$ and
$$\Gamma(\sigma\cup\tau)=\left(s_0(\bar\tau), l(\tau)\right).$$
\item If $h(\tau)> 0$, then define $\bar{\tau}=\displaystyle
\bigcup_{\{\epsilon\in\mathbb{Z}_p:\; |\tau^\epsilon|=h(\tau)\}} \tau^\epsilon\in W$
and $$\Gamma(\sigma\cup\tau)=\left(s(\bar\tau), l(\tau)\right).$$ \end{enumerate} Similar to the proof of Lemma~\ref{Z_pfanlemma}, $\Gamma\circ\bar{\lambda}:\operatorname{sd} C\longrightarrow \mathbb{Z}_p^{*t}$ is a simplicial $\mathbb{Z}_p$-map. This implies that ${\rm ind}_{\mathbb{Z}_p}(C)\leq t-1$ which is~not possible. \end{proof}
Next proposition is an extension of Theorem~\ref{altercrossindex}. However, we lose some properties by this extension.
\begin{proposition}\label{Xindposet} Let $P$ be a free ${\mathbb Z}_p$-poset and $$\begin{array}{rll} \psi: P & \longrightarrow & Q_{s,p}\\
p &\longmapsto & (\psi_1(p),\psi_2(p)) \end{array}$$ be an order preserving ${\mathbb Z}_p$-map. Then $P$ contains a chain $p_1\prec_P\cdots\prec_Pp_{ k}$ such that \begin{itemize} \item $k= {\rm ind}_{\mathbb{Z}_p}(\Delta P)+1$, \item for each $i\in[k-1]$, $\psi_2(p_i)< \psi_2(p_{i+1})$, and \item for each $\epsilon\in\mathbb{Z}_p$,
$$\left\lfloor{k\over p}\right\rfloor\leq \left|\left\{j:\; \psi_1(p_j)=\epsilon\right\}\right|\leq \left\lceil{k\over p}\right\rceil.$$ \end{itemize} \end{proposition} \begin{proof} Note $\psi$ can be considered as a simplicial ${\mathbb Z}_p$-map from $\Delta P$ to $ \mathbb{Z}_p^{*n}\subseteq (\sigma_{p-2}^{p-1})^{*n}$. Now, in view of Lemma~\ref{genfanlemma}, we have the assertion. \end{proof} Note that, for $p=2$, since ${\rm Xind}(P)\geq {\rm ind}(\Delta P)$,
Theorem~\ref{altercrossindex} is better than proposition~\ref{Xindposet}. However, we cannot prove that proposition~\ref{Xindposet} is valid if we replace ${\rm ind}(\Delta P)$ by ${\rm Xind}(P)$.
In an unpublished paper, Meunier~\cite{unpublishedMeunier} introduced a generalization of Tuckey-Ky~Fan lemma. He presented a version of $\mathbb{Z}_q$-Fan lemma which is valid for each odd integer $q\geq 3$. To be more specific, he proved that if $q$ is an odd positive integer and $\lambda:V(T)\longrightarrow \mathbb{Z}_q\times[m]$ is a $\mathbb{Z}_q$-equivariant labeling of an $\mathbb{Z}_q$-equivariant triangulation of a $(d-1)$-connected free $\mathbb{Z}_q$-spaces $T$, then there is at least one simplex in $T$ whose vertices are labelled with labels $(\epsilon_0,j_0),(\epsilon_1,j_1),\ldots,(\epsilon_n,j_n)$, where $\epsilon_i\neq \epsilon_{i+1}$ and $j_i<j_{i+1}$ for all $i\in\{0,1,\ldots,n-1\}$. Also, he asked the question if the result is true for even value of $q$. This question received a positive answer owing to the work of B.~Hanke et~al.~\cite{Hanke2009404}. In both mentioned works, the proofs of $\mathbb{Z}_q$-Fan lemma are built in involved construction. Here, we take the opportunity of this paper to propose the following generalization of this result with a short simple proof because we are using similar techniques in the paper. \begin{lemma}{\rm($\mathbb{G}$-Fan lemma).}\label{Gtucker} Let $\mathbb{G}$ be a nontrivial finite group and let $T$ be a free $\mathbb{G}$-simplicial complex such that ${\rm ind}_{\mathbb{G}}(T)= n$. Assume that $\lambda:V(T)\longrightarrow \mathbb{G}\times[m]$ be a $\mathbb{G}$-equivariant labeling such that there is no edge in $T$ whose vertices are labelled with $(g,j)$ and $(g',j)$ with $g\neq g'$ and $j\in[m]$. Then there is at least one simplicial complex in $T$ whose vertices are labelled with labels $(g_0,j_0),(g_1,j_1),\ldots,(g_n,j_n)$, where $g_i\neq g_{i+1}$ and $j_i<j_{i+1}$ for all $i\in\{0,1,\ldots,n-1\}$. In particular, $m\geq n+1$. \end{lemma} \begin{proof} Clearly, the map $\lambda$ can be considered as a $\mathbb{G}$-simplicial map from $T$ to $\mathbb{G}^{*m}$. Naturally, each nonempty simplex $\sigma\in \mathbb{G}^{*m}$ can be identified with a vector $X=(x_1,x_2,\ldots,x_m)\in (\mathbb{G}\cup\{0\})^n\setminus\{\boldsymbol{0}\}$. To prove the assertion, it is enough to show that there is a simplex $\sigma\in T$ such that ${\rm alt}(\lambda(\sigma))\geq n+1$. For a contradiction, suppose that, for each simplex $\sigma\in T$, we have ${\rm alt}(\lambda(\sigma))\leq n$. Define $$\begin{array}{lrll} \Gamma:&V(\operatorname{sd} T) &\longrightarrow & \mathbb{G}\times[n]\\
&\sigma&\longmapsto & \left(g,{\rm alt}(\lambda(\sigma)\right)), \end{array}$$ where $g$ is the first nonzero coordinate of the vector $\lambda(\sigma)\in (\mathbb{G}\cup\{0\})^n\setminus\{\boldsymbol{0}\}.$ One can check that $\Gamma$ is a simplicial $\mathbb{G}$-map from $\operatorname{sd} T$ to $\mathbb{G}^{*n}$. Note $\mathbb{G}^{*n}$ is an $E_{n-1} \mathbb{G}$ space. Consequently, ${\rm ind}_{\mathbb{G}}(\mathbb{G}^{*n})= n-1$. This implies that ${\rm ind}_{\mathbb{G}}(T)\leq n-1$ which is a contradiction. \end{proof}
\subsection{\bf Hierarchy of Indices} The aim of this subsection is introducing some tools for the proof of Theorem~\ref{inequalities}.
Let $n,\alpha$, and $p$ be integers where $n\geq 1$, $n\geq\alpha\geq 0$, and $p$ is prime. Define $$\displaystyle\Sigma_p(n,\alpha)=\Delta\left\{X\in(\mathbb{Z}_p\cup\{0\})^n:\; {\rm alt}(X)\geq \alpha+1\right\}.$$ Note that $\displaystyle\Sigma_p(n,\alpha)$ is a free simplicial $\mathbb{Z}_p$-complex with the vertex set $$\left\{X\in(\mathbb{Z}_p\cup\{0\})^n:\; {\rm alt}(X)\geq \alpha+1\right\}.$$
\begin{lemma}\label{indsigma} Let $n,\alpha$, and $p$ be integers where $n\geq 1$, $n\geq\alpha\geq 0$, and $p$ is prime. Then $${\rm ind}_{\mathbb{Z}_p}(\displaystyle\Sigma_p(n,\alpha))\geq n-\alpha-1.$$ \end{lemma} \begin{proof} Define $$ \begin{array}{crcl} \lambda: & \operatorname{sd} \mathbb{Z}_p^{*n} & \longrightarrow &
(\mathbb{Z}_p^{*\alpha})*\left(\displaystyle\Sigma_p(n,\alpha)\right)\\
& X & \longmapsto &
\left
\{\begin{array}{cl}
(\epsilon,{\rm alt}(X)) & \mbox{ if ${\rm alt}(X)\leq \alpha$}\\
X & \mbox{ if ${\rm alt}(X)\geq \alpha+1$},
\end{array}
\right. \end{array}$$ where $\epsilon$ is the first nonzero term of $X$. Clearly, the map $\lambda$ is a simplicial $\mathbb{Z}_p$-map. Therefore, $$ \begin{array}{lll} n-1={\rm ind}_{\mathbb{Z}_p}( \operatorname{sd} \mathbb{Z}_p^{*n}) & \leq & {\rm ind}_{\mathbb{Z}_p}\left(\mathbb{Z}_p^{*\alpha}*\displaystyle\Sigma_p(n,\alpha)\right)\\ & \leq & {\rm ind}_{\mathbb{Z}_p}(\mathbb{Z}_p^{*\alpha})+{\rm ind}_{\mathbb{Z}_p}(\displaystyle\Sigma_p(n,\alpha))+1\\ &\leq &\alpha+{\rm ind}_{\mathbb{Z}_p}(\displaystyle\Sigma_p(n,\alpha)) \end{array} $$ which completes the proof. \end{proof} \begin{proposition}\label{inequalityI} Let ${\mathcal H}$ be a hypergraph. For any integer $r\geq 2$ and any prime number $p\geq r$, we have
$${\rm ind}_{\mathbb{Z}_p}({\rm B}_0({\rm KG}^r({\mathcal H}),\mathbb{Z}_p))+1\geq |V({\mathcal H})|-{\rm alt}_p({\mathcal H}).$$ \end{proposition} \begin{proof}
For convenience, let $|V({\mathcal H})|=n$ and $\alpha=n-{\rm alt}_p({\mathcal H})$. Let $\pi:[n]\longrightarrow V({\mathcal H})$ be the bijection such that ${\rm alt}_p({\mathcal H},\pi)={\rm alt}_p({\mathcal H})$. Define $$ \begin{array}{lrll} \lambda:& \Sigma_p(n,\alpha)& \longrightarrow & \operatorname{sd}{\rm B}_0({\rm KG}^r({\mathcal H}),\mathbb{Z}_p))\\
& X&\longmapsto & \{\omega^1\}\times U_1\cup\cdots\cup \{\omega^p\}\times U_p, \end{array} $$ where $U_i=\{e\in E({\mathcal H}):\; e\subseteq \pi(X^{\omega^i})\}.$ One can see that $\lambda$ is a simplicial $\mathbb{Z}_p$-map. Consequently, $${\rm ind}_{\mathbb{Z}_p}({\rm B}_0({\rm KG}^r({\mathcal H}),\mathbb{Z}_p))\geq {\rm ind}_{\mathbb{Z}_p}(\Sigma_p(n,\alpha))\geq n-{\rm alt}_p({\mathcal H})-1. $$ \end{proof}
\begin{proposition}\label{inequalityII} Let ${\mathcal H}$ be an $r$-uniform hypergraph and $p\geq r$ be a prime number. Then $$ {\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))+p\geq {\rm ind}_{\mathbb{Z}_p}(\Delta{\rm Hom}(K^r_p,{\mathcal H}))+p\geq {\rm ind}_{\mathbb{Z}_p}(B_0({\mathcal H},\mathbb{Z}_p))+1.$$ \end{proposition} \begin{proof} Since already we know ${\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))\geq {\rm ind}_{\mathbb{Z}_p}(\Delta{\rm Hom}(K^r_p,{\mathcal H}))$, to prove the assertion, it is enough to show that ${\rm ind}_{\mathbb{Z}_p}(\Delta{\rm Hom}(K^r_p,{\mathcal H}))+p\geq {\rm ind}_{\mathbb{Z}_p}(B_0({\mathcal H},\mathbb{Z}_p))+1.$ To this end, define $$\begin{array}{llll} \lambda: & \operatorname{sd} B_0({\mathcal H},\mathbb{Z}_p) & \longrightarrow & \left(\operatorname{sd}\sigma_{p-2}^{p-1}\right)*\displaystyle\left(\Delta{\rm Hom}(K^r_p,{\mathcal H})\right)\\ \end{array}$$ such that for each vertex $\tau=\displaystyle\bigcup_{i-1}^p\left(\{\omega^i\}\times U_i\right)$ of $\operatorname{sd} B_0({\mathcal H},\mathbb{Z}_p)$, $\lambda(\tau)$ is defined as follows. \begin{itemize} \item If $U_i\neq\varnothing$ for each $i\in[p]$, then $\lambda(\tau)=\tau.$ \item If $U_i=\varnothing$ for some $i\in[p]$, then $$\lambda(\tau)=\{\omega^i\in\mathbb{Z}_p:\; U_i=\varnothing\}.$$ \end{itemize} One can check that the map $\lambda$ is a simplicial $\mathbb{Z}_p$-map. Also, since $\sigma_{p-2}^{p-1}$ is a free simplicial $\mathbb{Z}_p$-complex of dimension $p-2$, we have ${\rm ind}_{\mathbb{Z}_p}(\sigma_{p-2}^{p-1})\leq p-2$ (see properties of the $\mathbb{G}$-index in Section~\ref{intro}). This implies that $$ \begin{array}{lll}
{\rm ind}_{\mathbb{Z}_p}(B_0({\mathcal H},\mathbb{Z}_p))& \leq & {\rm ind}_{\mathbb{Z}_p}\left(\left(\operatorname{sd}\sigma_{p-2}^{p-1}\right)* \left(\Delta{\rm Hom}(K^r_p,{\mathcal H})\right)\right)\\ & \leq &{\rm ind}_{\mathbb{Z}_p}(\sigma_{p-2}^{p-1})+ {\rm ind}_{\mathbb{Z}_p}(\Delta{\rm Hom}(K^r_p,{\mathcal H}))+1\\ &\leq & p-1+{\rm ind}_{\mathbb{Z}_p}(\Delta{\rm Hom}(K^r_p,{\mathcal H})) \end{array} $$ which completes the proof. \end{proof}
\section{\bf Proofs of Theorem~\ref{maincolorfulindex} and Theorem~\ref{inequalities}}\label{sec:proofs} Now, we are ready to prove Theorem~\ref{maincolorfulindex} and Theorem~\ref{inequalities}.\\
\noindent{\bf Proof of Theorem~\ref{maincolorfulindex}: Part (i).} For convenience, let ${\rm ind}_{\mathbb{Z}_p}({\rm B}_0({\mathcal H},{\mathbb{Z}_p}))=t$. Note that $$ \begin{array}{crcl} \Gamma: &\mathbb{Z}_p\times V({\mathcal H}) & \longrightarrow & \mathbb{Z}_p\times [C]\\
& (\epsilon,v) & \longmapsto & (\epsilon,c(v)) \end{array}$$ is a simplicial $\mathbb{Z}_p$-map from ${\rm B}_0({\mathcal H},{\mathbb{Z}_p})$ to $(\sigma^{p-1}_{r-2})^{*C}$. Therefore, in view of Lemma~\ref{genfanlemma}, there is a $t$-dimensional simplex $\tau\in{\rm im}(\Gamma)$ such that, for each $\epsilon\in \mathbb{Z}_p$, we have
$\lfloor{t+1\over p}\rfloor\leq |\tau^\epsilon|\leq\lceil{t+1\over p}\rceil.$ Let $\displaystyle\bigcup_{i=1}^p(\{\omega^i\} \times U_i)$ be the minimal simplex in $\Gamma^{-1}(\tau)$. One can see that ${\mathcal H}[U_1,\ldots,U_p]$ is the desired subhypergraph. Moreover, since every color can be appeared in at most $r-1$ number of $U_i$'s, we have $$C\geq {{\rm ind}_{p}({\rm B}_0({\mathcal H},{\mathbb{Z}_p}))+1\over r-1}.$$
\noindent{\bf Part (ii).} For convenience, let ${\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))=t$. Define the map $$\lambda:{\rm Hom}(K^r_p,{\mathcal H})\longrightarrow \operatorname{sd}(\sigma_{r-2}^{p-1})^{*C}$$ such that for each $(U_1,\cdots,U_p)\in {\rm Hom}(K^r_p,{\mathcal H})$, $$\lambda(U_1,\cdots,U_p)=\{\omega^1\}\times c(U_1) \cup\cdots\cup \{\omega^p\}\times c(U_p).$$ {\bf Claim.} There is a $p$-tuple $(U_1,\cdots,U_p)\in {\rm Hom}(K^r_p,{\mathcal H})$ such that for $\tau=\lambda(U_1,\cdots,U_p)$, we have $l(\tau)\geq {\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))+p$.
\noindent{\bf Proof of Claim.} Suppose, contrary to the claim, that for each $\tau\in {\rm Im}(\lambda)$, we have $l(\tau)\leq t+p-1$. Note that $\operatorname{sd}(\sigma_{r-2}^{p-1})^{*C}$ can be considered as a free $\mathbb{Z}_p$-poset ordered by inclusion. One can readily check that $\lambda$ is an order-preserving $\mathbb{Z}_p$-map. Clearly, for each $\tau\in {\rm Im}(\lambda)$, we have $h(\tau)=\displaystyle\min_{\epsilon\in\mathbb{Z}_p}|\tau^\epsilon|\geq 1$ and consequently, $l(\tau)\geq p$. Now, define $$\bar{\tau}=\displaystyle
\bigcup_{\{\epsilon\in\mathbb{Z}_p:\; |\tau^\epsilon|=h(\tau)\}} \tau^\epsilon\in W\quad {\rm and }\quad \Gamma(\tau)=\left(s(\bar\tau), l(\tau)-p+1\right).$$ One can see that the map $\Gamma:{\rm im}(\lambda)\longrightarrow Q_{t-1,p}$ is an order-preserving $\mathbb{Z}_p$-map. Therefore, $$\Gamma\circ\lambda:{\rm Hom}(K^r_p,{\mathcal H})\longrightarrow Q_{t-1,p}$$ is an order-preserving $\mathbb{Z}_p$-map, which contradicts the fact that ${\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))=t$.
$\square$
Now, let $(U_1,\cdots,U_p)$ be a minimal $p$-tuple in ${\rm Hom}(K^r_p,{\mathcal H})$ such that for $$\tau=\lambda(U_1,\cdots,U_p)=\{\omega^1\}\times c(U_1) \cup\cdots\cup \{\omega^p\}\times c(U_p),$$ we have $l(\tau)= t+p$. One can check that ${\mathcal H}[U_1,\cdots,U_p]$ is the desired complete $r$-uniform $p$-partite subhypergraph. Similar to the proof of Part (i), since every color can be appeared in at most $r-1$ number of $U_i$'s, we have $$C\geq {{\rm Xind}_{\mathbb{Z}_p}({\rm Hom}(K^r_p,{\mathcal H}))+p\over r-1}.$$
$\square$
\noindent{\bf Proof of Theorem~\ref{inequalities}.} It is simple to prove that
$|V({\mathcal F})|-{\rm alt}_p({\mathcal F}) \geq {\rm cd}_p({\mathcal F})$ for any hypergraph ${\mathcal F}$. Therefore, the proof follows by Proposition~\ref{inequalityI} and Proposition~\ref{inequalityII}.
$\square$\\
\noindent{\bf Acknowledgements.} I would like to acknowledge Professor Fr\'ed\'eric~Meunier for interesting discussions about the paper and his invaluable comments. Also, I would like to thank Professor Hossein~Hajiabolhasan and Mrs~Roya~Abyazi~Sani for their useful comments.
\def$'$} \def\cprime{$'${$'$} \def$'$} \def\cprime{$'${$'$}
\end{document}
|
arXiv
|
{
"id": "1605.06701.tex",
"language_detection_score": 0.6468298435211182,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{The Triple Point Spectrum of closed orientable $3$-manifolds}
\author{\'Alvaro Lozano Rojo}
\address{Centro Universitario de la Defensa Zaragoza, Academia General Militar Carretera de Huesca s/n. 50090 Zaragoza, Spain --- IUMA, Universidad de Zaragoza} \email{[email protected]}
\thanks{Partially supported by the European Social Fund and Diputaci\'on General de Arag\'on (Grant E15 Geometr{\'\i}a).}
\author{Rub\'en Vigara Benito} \address{Centro Universitario de la Defensa Zaragoza, Academia General Militar Carretera de Huesca s/n. 50090 Zaragoza, Spain --- IUMA, Universidad de Zaragoza} \email{[email protected]}
\subjclass[2000]{Primary 57N10, 57N35}
\keywords{$3$-manifold, homology $3$-sphere, immersed surface, filling Dehn surface, triple points, complexity of $3$-manifolds}
\begin{abstract}
The triple point numbers and the triple point spectrum of a closed
$3$-manifold were defined in~\cite{tesis}.
They are topological invariants that give a measure of the complexity of a
$3$-manifold
using the number of triple points of minimal filling Dehn surfaces.
Basic
properties of these invariants are presented, and the triple point spectra of
$\mathbb{S}^2\times \mathbb{S}^1$ and $\mathbb{S}^3$ are computed. \end{abstract}
\maketitle
\section{Introduction}\label{sec:intro}
Through the whole paper all 3-manifolds and surfaces are assumed to be orientable and closed, that is, compact connected and without boundary, $M$ denotes a 3-manifold and $S$ a surface. All objects are assumed to be in the smooth category: manifolds have a differentiable structure and all maps are assumed to be smooth. By simplicity a genus $g$ surface, $g=0,1,2,\ldots$, is called a \emph{$g$-torus}.
Let $M$ be a 3-manifold.
A subset $\Sigma\subset M$ is a \emph{Dehn surface} in $M$~\cite{Papa} if there exists a surface $S$ and a general position immersion $f:S\rightarrow M$ such that $\Sigma=f\left( S\right)$. In this situation we say that $S$ is the \emph{domain} of $\Sigma$ and that $f$ \emph{parametrizes} $\Sigma$. If $S$ is a $2$-sphere, a torus, or a $g$-torus then $\Sigma$ is a \emph{Dehn sphere}, a \emph{Dehn torus}, or a \emph{Dehn $g$-torus} respectively. For a Dehn surface $\Sigma\subset M$, its singularities are divided into \emph{double points}, where two sheets of $\Sigma$ intersect transversely, and \emph{triple points}, where three sheets of $\Sigma$ intersect transversely, and they are arranged along \emph{double curves} (see Section~\ref{sec:Dehn-surfaces-Johansson-diagrams} below for definitions and pictures).
A Dehn surface $\Sigma\subset M$ \emph{fills} $M$~\cite{Montesinos} if it defines a cell-decomposition of $M$ in which the 0-skeleton is the set of triple points of $\Sigma$, the 1-skeleton is the set of double and triple points of $\Sigma$, and the 2-skeleton is $\Sigma$ itself. Filling Dehn spheres are introduced (following ideas of W. Haken~\cite{Haken1}) in~\cite{Montesinos}, where it is proved that every closed, orientable 3-manifold has a filling Dehn sphere. In ~\cite{Montesinos} filling Dehn spheres and their Johansson diagrams (cf. Section~\ref{sec:Dehn-surfaces-Johansson-diagrams}) are proposed as a suitable way for representing all closed orientable 3-manifolds. A weaker version of filling Dehn surfaces are the \emph{quasi-filling Dehn surfaces} defined in~\cite{Amendola02}. A quasi-filling Dehn surface in $M$ is a Dehn surface whose complementary set in $M$ is a disjoint union of open $3$-balls. In~\cite{FennRourke} it is proved that every $3$-manifold has a quasi-filling Dehn sphere. As it is pointed out in~\cite{Amendola02} (see also~\cite{Funar}), filling Dehn surfaces are closely related with \emph{cubulations} of 3-manifolds: cell decompositions where the building blocks are cubes. The dual decomposition of that defined by a filling Dehn surface is a cubulation and vice versa, the dual cell decomposition of a cubulation is a filling Dehn surface (perhaps with nonconnected or nonorientable domain).
\begin{figure}
\caption{A triple point of $\Sigma$ and its triplet
in $S$}
\label{fig:triple-point-00}
\end{figure} The number of triple points of filling Dehn surfaces measures the complexity of $3$-manifolds. A filling Dehn surface $\Sigma\subset M$ with domain $S$ is \emph{minimal} if there is no filling Dehn surface in $M$ with domain $S$ with less triple points than $\Sigma$.
In~\cite{RHomotopies} it is defined the \emph{genus $g$ triple point number} $t_g(M)$ of $M$ as the number of triple points of a minimal filling $g$-torus of $M$, and the \emph{triple point spectrum} of $M$ as the sequence of numbers \[
\mathscr{T}(M)=(t_0(M),t_1(M),t_2(M),\ldots)\,. \] The genus $0$ triple point number of $M$ is the \emph{Montesinos complexity} $M$~\cite{racsam}. These invariants are related to the \emph{surface-complexity} of $M$, defined in~\cite{Amendola02} as the number of triple points of a quasi-filling Dehn surface of $M$ (perhaps with nonconnected or nonorientable domain).
We want to clarify how much information about $M$ encloses $\mathscr{T}(M)$. Since the original work~\cite{RHomotopies}, nothing is known about the triple point numbers and the triple point spectrum apart from their definitions. In this paper we introduce some properties of $\mathscr{T}(M)$ (Section~\ref{sec:spectrum-surgerysurfaces}), and we present the two unique known triple point spectra
\begin{align}
\mathscr{T}(\mathbb{S}^2\times \mathbb{S}^1)&=(2,1,3,5,\ldots)\,,\label{eq:triple-point-spectrum-S2-S1}\\
\mathscr{T}(\mathbb{S}^3)&=(2,4,6,8,\ldots)\,.\label{eq:triple-point-spectrum-S3}
\end{align}
The proof of \eqref{eq:triple-point-spectrum-S2-S1} relies on two previously known examples of minimal filling Dehn surfaces of $\mathbb{S}^2\times \mathbb{S}^1$, together with the inequality (Lemma~\ref{lem:triple-point-inequality})
\begin{equation}
t_{g+1}(M)\leq t_{g}(M)+2\,,\; g=0,1,2,\ldots ,\label{eq:consecutive-triple-point-numbers-2}
\end{equation}
that arises from a simple surgery operation (the \emph{handle piping}, see Figure~\ref{Fig:handle}). This will be detailed in Section~\ref{sec:spectrum-surgerysurfaces}.
The characterization of the triple point spectrum of $\mathbb{S}^3$ needs
more work.
The identity \eqref{eq:triple-point-spectrum-S3} is proposed in~\cite{racsam}
as an open question.
Using that $t_0(\mathbb{S}^3)=2$~\cite{racsam}, the handle piping provides $g$ filling
Dehn $g$-tori in $\mathbb{S}^3$ with $2+2g$ triple points for all $g=1,2,\ldots$.
The harder part is to prove that all these filling Dehn surfaces in $\mathbb{S}^3$ are minimal.
To do so, in Section~\ref{sec:fundamentalgroup} we introduce a presentation of the
fundamental group a Dehn $g$-torus. This is a generalization of a presentation of
the fundamental group of Dehn spheres due to W. Haken and detailed in~\cite{tesis}.
Using it, in Section~\ref{sec:checkers-homology-spheres} we prove:
\begin{thmA}
If $M$ is a $\mathbb{Z}/2$-homology $3$-sphere
\begin{equation*}
t_g(M) \geq 2 + 2g.
\end{equation*} \end{thmA}
This theorem completes the proof of \eqref{eq:triple-point-spectrum-S3}.
\begin{figure}
\caption{A filling Dehn sphere in $\mathbb{S}^2\times \mathbb{S}^1$}
\label{Fig:audi}
\end{figure}
\section{Dehn surfaces and their Johansson's diagrams}\label{sec:Dehn-surfaces-Johansson-diagrams}
Let $\Sigma$ be a Dehn surface in $M$ and consider a parametrization $f:S\to M$
of $\Sigma$.
The \emph{singularities} of $\Sigma$ are the points $x\in\Sigma$ such that
$\#f^{-1}(x)>1$. The \emph{singularity set} $S(\Sigma)$ of $\Sigma$ is the set of
singularities of $\Sigma$. As $f$ is in general position, the singularities of
$\Sigma$ can be
divided into double points ($\#f^{-1}(x)=2$), and triple points
($\#f^{-1}(x)=3$). Following~\cite{Shima}, we denote by $T(\Sigma)$ the set
of triple points of $\Sigma$. The preimage under $f$ in $S$ of the
singularity set of $\Sigma$, together with the information about how its
points become identified by $f$ in $\Sigma$ is the \emph{Johansson diagram} $\mathcal{D}$
of $\Sigma$ (see~\cite{Johansson1,Montesinos}). We say that two points of $S$
are \emph{related} if they project onto the same point of $\Sigma$.
In the following a \emph{curve} in $S$, $\Sigma$ or $M$ is the image of an immersion
(a \emph{parametrization} of the curve) from $\mathbb{S}^1$ or $\mathbb{R}$
into $S$, $\Sigma$ or $M$, respectively. A \emph{path} in $S$, $\Sigma$ or $M$ is a map $\eta$
from the interval $[0,1]$ into $S$, $\Sigma$ or $M$, respectively. We say that $\eta$ \emph{joins}
$\eta (0)$ \emph{with} $\eta (1)$, or that $\eta$ \emph{starts} at
$\eta (0)$ and \emph{ends} at $\eta (1)$, and if $\eta(0)=\eta(1)$ we say that $\eta$
is a \emph{loop}.
A double curve of $\Sigma$ is a curve in $M$ contained in $S(\Sigma)$.
Because $S$ is closed, double curves are closed and there is a finite number of them. The number of triple points is also finite. Because $S$ and $M$ are orientable, the preimage under $f$ of a double curve of $\Sigma$ is the union of two different closed curves in $S$, and we will say that these two curves are \emph{sister curves} of $\mathcal{D}$. Thus, the Johansson diagram of $\Sigma$ is composed by an even number of different closed curves in $S$, and we will identify $\mathcal{D}$ with the set of different curves that compose it. For any curve $\alpha\in\mathcal{D}$ we denote by $\tau \alpha$ the sister curve of $\alpha$ in $\mathcal{D}$. This defines a free involution $\tau:\mathcal{D}\rightarrow\mathcal{D}$, the \emph{sistering} of $\mathcal{D}$, that sends each curve of $\mathcal{D}$ into its sister curve of $\mathcal{D}$.
The curves of $\mathcal{D}$ intersect with others or with themselves transversely at the \emph{crossings} of $\mathcal{D}$. The crossings of $\mathcal{D}$ are the preimage under $f$ of the triple points of $\Sigma$. If $P$ is a triple point of $\Sigma$, the three crossings of $\mathcal{D}$ in $f^{-1}(D)$ compose \emph{the triplet of} $P$ (see Figure~\ref{fig:triple-point-00}). In Section~\ref{sec:fundamentalgroup} we will consider paths contained in curves of $\mathcal{D}$ or in double curves. In this special case, we will consider that paths are also immersed, and therefore they continue ``straight ahead'' when they arrive to a crossing.
If $\Sigma$ is a Dehn surface in $M$, a connected component of $M-\Sigma$ is a \emph{region} of $\Sigma$, a connected component of $\Sigma-S(\Sigma)$ is a \emph{face} of $\Sigma$, and a connected component of $S(\Sigma)-T(\Sigma)$ is an \emph{edge} of $\Sigma$. The Dehn surface $\Sigma$ fills $M$ if and only if all its edges, faces and regions are open $1$, $2$ or $3$-dimensional disks respectively.
In Figure~\ref{Fig:audi} we have depicted (left) one of the simplest Johansson diagrams of filling Dehn spheres. This is the diagram of a Dehn sphere $\Sigma$ (right) that fills $\mathbb{S}^2\times \mathbb{S}^1$. In this figure we consider $\mathbb{S}^2\times \mathbb{S}^1$ as $\mathbb{S}^2\times [-1,1]$ where the top and bottom covers $\mathbb{S}^2\times \{+1\}$ and $\mathbb{S}^2\times \{-1\}$ are identified by the vertical projection. This example appears completely detailed in~\cite[Example 7.5.1]{tesis}.
If we are given an \emph{abstract diagram}, i.e., an even collection of curves
in $S$ coherently identified in pairs, it is possible to know if this
abstract diagram is \emph{realizable}: if it is actually the Johansson diagram
of a Dehn surface in a $3$-manifold (see~\cite{Johansson1,Johansson2,tesis}).
It is also possible to know if the abstract diagram is \emph{filling}: if it is
the Johansson diagram of a filling Dehn surface of a 3-manifold (see
~\cite{tesis}). If $\Sigma$ fills $M$, it is possible to build $M$ out of the Johansson diagram of $\Sigma$. Thus, filling Johansson diagrams represent all closed, orientable $3$-manifolds.
It must be noted that when a diagram $(\mathcal{D},\tau)$ in $S$ is not realizable, the quotient space of $S$ under the equivalence relation defined by the diagram is something very close to a Dehn surface: it is a 2-dimensional complex with simple, double and triple points, but it cannot be embedded in any $3$-manifold. We reserve the name \emph{pseudo Dehn surface} for these objects. Many constructions about Dehn surfaces, as the presentation of their fundamental group given in Section~\ref{sec:fundamentalgroup}, for example, are also valid for pseudo Dehn surfaces.
\begin{figure}
\caption{Handle piping}
\label{Fig:handle}
\end{figure}
\section{The triple point spectrum}
\label{sec:spectrum-surgerysurfaces}
In order to understand the structure of the sequence $\mathscr{T}(M)$, we need to relate the different triple point numbers of $M$. For a shorter notation we will omit the word ``filling'' in the expression ``minimal filling Dehn surface''. If we are given a filling Dehn $g$-torus $\Sigma_g$ of $M$ with $q$ triple points, we can always obtain a filling Dehn $(g+1)$-torus $\Sigma_{g+1}$ of $M$ with $q+2$ triple points by applying to $\Sigma_g$ the \emph{handle piping} modification depicted in Figure~\ref{Fig:handle} in a small neighbourhood of a triple point. If the original $\Sigma_g$ is minimal, we have:
\begin{lemma}\label{lem:triple-point-inequality}
For any $g=0,1,2,\ldots$
\begin{equation}
\pushQED{\qed}
t_{g+1}(M)\leq t_{g}(M)+2\,,\label{eq:consecutive-triple-point-numbers-3}\qedhere
\popQED
\end{equation}
\end{lemma}
This lemma suggests the following definitions.
\begin{definition}\label{def:exceptional}
A minimal Dehn $g$-torus of $M$ is \emph{exceptional} if it has less than $t_{g-1}(M)+2$ triple points.
\end{definition}
\begin{definition}\label{def:height}
The \emph{height} $\mathscr{H}(M)$ of $M$ is the highest genus among all exceptional Dehn $g$-tori of $M$.
If $M$ has no exceptional Dehn $g$-torus, $\mathscr{H}(M)=0$.
\end{definition}
It is clear that if $\mathscr{H}(M)$ is finite, the equality in \eqref{eq:consecutive-triple-point-numbers-3} holds for all $g\geq \mathscr{H}(M)$.
An important result for filling Dehn surfaces which is of great interest here is the following.
\begin{proposition}
\label{prop:triplets-regions}
A Dehn $g$-torus in $M$ with $q$ triple points and $r$ regions fills $M$ if and only if
\begin{equation}
r=q+(2-2g)\,.\label{eq:triplets-regions}
\end{equation}
\end{proposition}
This proposition follows from~\cite[Theorem 3.7.1]{tesis} (see also~\cite[Lemma 43]{RHomotopies}). Its proof relies on Euler's characteristic techniques and it has strong consequences in this context:
\begin{remark}\label{rmk:at-least-1+2g-triple-points}
A filling Dehn $g$-torus has at least $2g-1$ triple points.
\end{remark}
\begin{remark}\label{rmk:handle-regions-unchanged}
After a handle piping on a filling Dehn $g$-torus the number of regions remains unchanged. More generally, if $\Sigma_g$ is a filling Dehn $g$-torus of $M$ with $q$ triple points and $\Sigma_{g+1}$ is a filling Dehn $(g+1)$-torus of $M$ with $q+2$ triple points, the number of regions of $\Sigma_g$ and $\Sigma_{g+1}$ is the same.
\end{remark}
\begin{remark}\label{rmk:exceptional-reduce-regions}
Exceptional Dehn $g$-tori reduce the number of regions, that is, an exceptional Dehn $g$-torus
in $M$ has less regions than any minimal Dehn $g'$-torus in $M$ with $g'<g$.
\end{remark}
\begin{figure}
\caption{Dual curves to the curves of $\mathcal{D}$}
\label{Fig:dual-curve-flat}
\label{Fig:dual-curve}
\label{Fig:dual-curve-flat-dual-curve}
\end{figure}
As the number of regions is bounded from below, by Remark~\ref{rmk:exceptional-reduce-regions} there cannot be exceptional Dehn $g$-tori in $M$ with arbitrarily high genus. Therefore,
\begin{theorem}\label{thm:height-finite}
The height of $M$ is finite.\qed
\end{theorem}
In~\cite[Example 7.5.2]{tesis} it is shown a filling Dehn $1$-torus of $\mathbb{S}^2\times \mathbb{S}^1$ with just $1$ triple point. It is clearly a minimal Dehn $1$-torus, and by Proposition~\ref{prop:triplets-regions} it has only one region. By Remark~\ref{rmk:exceptional-reduce-regions}, there cannot be an exceptional Dehn $g$-torus in $\mathbb{S}^2\times \mathbb{S}^1$ of genus $g>1$. On the other hand, a Dehn sphere in any $3$-manifold has an even number of triple points~\cite{Haken1}, and therefore the filling Dehn sphere of Figure~\ref{Fig:audi} is minimal. It turns out that $\mathscr{H}(\mathbb{S}^2\times \mathbb{S}^1)=1$ and
\begin{theorem}
$\mathscr{T}(\mathbb{S}^2\times \mathbb{S}^1)=(2,1,3,5,\ldots )\,.$\qed
\end{theorem}
\section{The fundamental group of a Dehn $g$-torus}
\label{sec:fundamentalgroup}
In this section the manifold $M$ containing $\Sigma$ is no longer needed. Although we still talk about ``Dehn surfaces'', the construction only makes use of the Johansson diagram $\mathcal{D}$, so it is valid for pseudo Dehn surfaces.
Let $f:S\to M$ be a Dehn surface in $M$ and denote by $\mathcal{D}$ its Johansson diagram.
Fix a simple point $x\notin\mathcal{D}$ as the base point of the fundamental group
$\pi_1(\Sigma)$ of $\Sigma$. We identify $x$ with its preimage
under $f$.
A path in $\Sigma$ is \emph{surfacewise} if it is a path on $S$ mapped to
$\Sigma$ through $f$. We denote by $\pi_S = f_*\pi_1(S)$ the subgroup of $\pi_1(\Sigma)$
generated by the sufacewise loops based at $x$.
In general we will use the same notation for a surfacewise path in $\Sigma$
and its preimage path in $S$.
Let $\alpha,\tau\alpha$ be two sister curves of $\mathcal{D}$, and take two paths
$\lambda_{\alpha},\lambda_{\tau\alpha}$ in $S$, starting from $x$ and
arriving to related points on $\alpha,\tau\alpha$, respectively
(see Figure~\ref{Fig:dual-curve-flat-dual-curve}).
\begin{definition}\label{def:dual-paths}
We say that the loop $\lambda_\alpha \lambda_{\tau\alpha}^{-1}$ in
$\Sigma$ based at $x$ is \emph{dual} to $\alpha$.
\end{definition}
\begin{proposition}\label{prop:surfacewise-and-duals-generate-pi1}
Surfacewise loops based at $x$ and loops dual to the curves
of $\mathcal{D}$ generate $\pi_1(\Sigma)$.
\end{proposition}
\begin{figure}
\caption{Crossing-type and corner-type intersections with $S(\Sigma)$}
\label{Fig:crossing-type}
\label{Fig:corner-type}
\label{Fig:crossing-corner-type}
\end{figure}
\begin{proof}
Consider a loop $\sigma$ in $\Sigma$ based at $x$. Up to homotopy, we can
assume that all the intersections of $\sigma$ with the singular set of $\Sigma$ are of
``crossing type'' or of ``corner type'' as in Figure~\ref{Fig:crossing-type} and~\ref{Fig:corner-type} respectively.
Therefore, $\sigma$ can be written as a product
$\sigma = b_1 b_2\cdots b_k$ of surfacewise paths
(Figure~\ref{Fig:holed_plane}). If $k=1$ or $2$, either $\sigma$ is
surfacewise or it is dual to a
curve of $\mathcal{D}$, and so there is nothing to prove.
In other case, for each $i=2,\ldots,k-1$ we choose a midpoint $x_i$ of $b_i$
and a path $c_i$ joining $x$ with $x_i$ (Figure~\ref{Fig:holed_plane}).
We write $b_i=b_i^-b_i^+$, with
$b_i^-$ ending and $b_i^+$ starting at $x_i$. The loop $\sigma$
is homotopic to
\[
b_1 b_2^- c_2^{-1} c_2 b_2^+ b_3^-
\cdots c_{k-2}b_{k-2}^+ b_{k-1}^- c_{k-1}^{-1}
c_{k-1} b_{k-1}^+ b_k\,,
\]
and this expression is a product of loops dual to curves of $\mathcal{D}$.
\end{proof}
\begin{figure}
\caption{Loops as product of dual loops}
\label{Fig:holed_plane}
\end{figure}
We know that $\pi_S$ is finitely generated because so is $\pi_1(S)$. The
following lemma allows us to find a finite set of generators for
$\pi_1(\Sigma)$.
\begin{lemma}
\label{lem:fundamental-dual-paths-surface-conjugate}
If $a,a'$ are loops on $\Sigma$ dual to $\alpha\in\mathcal{D}$, they are
\emph{surfacewise conjugate}, that is, there are two surfacewise loops
$s ,t$ based at $x$ such that $a=s a' t$.
\end{lemma}
\begin{figure}
\caption{Duals to $\alpha$ are surfacewise conjugate}
\label{Fig:holed_plane_duals-conjugate}
\end{figure}
\begin{proof}
Put $a=\lambda_\alpha \lambda_{\tau\alpha}^{-1}$, $a'=\mu_\alpha
\mu_{\tau\alpha}^{-1}$, where
$\lambda_\alpha,\lambda_{\tau\alpha},\mu_\alpha,\mu_{\tau\alpha}$ are
surfacewise paths as in Definition~\ref{def:dual-paths}. Let
$x_\alpha,x_{\tau\alpha},x'_\alpha,x'_{\tau\alpha}$ be the endpoints of
$\lambda_\alpha,\lambda_{\tau\alpha},\mu_\alpha,\mu_{\tau\alpha}$ in $\mathcal{D}$,
respectively (Figure~\ref{Fig:holed_plane_duals-conjugate}). Let $d$ be one
path in $S$ contained in $\alpha$ joining $x_\alpha$ with $x'_\alpha$, and let $\tau d$ be the
sister path of $d$ inside $\tau\alpha$ joining $x_{\tau\alpha}$ with
$x'_{\tau\alpha}$. In $\Sigma$ the paths $d$ and $\tau d$ coincide, and so we have:
$$a=\lambda_\alpha \lambda_{\tau\alpha}^{-1}=\lambda_\alpha d (\tau d)^{-1}
\lambda_{\tau\alpha}^{-1}=
\lambda_\alpha d \mu_{\alpha}^{-1}\mu_\alpha\mu_{\tau\alpha}^{-1}
\mu_{\tau\alpha} (\tau d)^{-1} \lambda_{\tau\alpha}^{-1}= s a' t\,,$$ where $s =\lambda_\alpha d \mu_{\alpha}^{-1}$ and $t=\mu_{\tau\alpha} (\tau d)^{-1} \lambda_{\tau\alpha}^{-1}$ are surfacewise loops in $\Sigma$ based at $x$. \end{proof}
From now on, we will fix a set of generators $s_1,s_2,\ldots,s_{2g}$ of $\pi_S$, and we will also fix a set of preferred dual loops to the curves of $\mathcal{D}$ as follows. For each diagram curve $\alpha\in\mathcal{D}$ we choose a \emph{basepoint} $x_\alpha$ and a joining arc $\lambda_\alpha$ from $x$ to $x_\alpha$, such that the basepoints of sister curves are related (see Figure~\ref{Fig:dual-curve-flat-dual-curve}). The preferred dual loop of $\alpha\in\mathcal{D}$ is
\[
a = \lambda_\alpha \lambda_{\tau\alpha}^{-1}\,.
\] If a curve of $\mathcal{D}$ is denoted with lowercase greek letters \[
\alpha, \beta, \gamma, \ldots,
\alpha_1, \alpha_2, \ldots, \alpha_i, \ldots, \tau\alpha\,, \] its preferred dual loop will be denoted with the corresponding lowercase roman letters \[
a, b, c,\ldots, a_1, a_2,\ldots, a_i,\ldots, \tau a\,. \]
\begin{figure}
\caption{Contractible loop near a triple point}
\label{Fig:contractible-loop}
\end{figure}
By Proposition~\ref{prop:surfacewise-and-duals-generate-pi1} and Lemma \ref{lem:fundamental-dual-paths-surface-conjugate}, we have: \begin{theorem}\label{thm:finite-set-of-generators-pi1} If $\alpha_1,\alpha_2,\ldots ,\alpha_{2k}$ are the curves of $\mathcal{D}$, then $\pi_1(\Sigma)$ is generated by \[
\pushQED{\qed}
s_1,s_2,\ldots,s_{2g},a_1,a_2,\ldots,a_{2k}\,.\qedhere
\popQED \]
\end{theorem}
\begin{figure}
\caption{Triple point relation (diagram view)}
\label{Fig:triple-point-diagram}
\end{figure}
In order to obtain a presentation of $\pi_1(\Sigma)$, we need to establish a set of relators associated to the generators of Theorem~\ref{thm:finite-set-of-generators-pi1}. These relators will be a natural extension of Haken's relators for Dehn spheres (see also~\cite{racsam,tesis}).
\begin{enumerate}[\em(R1)]
\item \emph{Dual loop relations}. By construction, if
$\alpha,\tau \alpha$ are two sister curves of $\mathcal{D}$, their dual loops
$a,\tau a$ verify $\tau a= a^{-1}$.
\item \emph{Triple point relations}. The idea behind this relation is that any small circuit around
a triple point as that of Figure~\ref{Fig:contractible-loop}
is homotopically trivial.
Assume that all the curves of $\mathcal{D}$ are oriented,
with sister curves coherently oriented.
Let $P$ be a triple point of $\Sigma$ and let $P_1,P_2,P_3$ be the
three crossings of $\mathcal{D}$ in the triplet of $P$. Label the curves of
$\mathcal{D}$ intersecting at these crossings as in
Figure~\ref{Fig:triple-point-diagram}.
Consider three paths
$\omega_1,\omega_2,\omega_3$ joining $x$ with $P_1,P_2,P_3$ respectively.
At $P_1$, consider the path $d_\alpha$ contained in $\alpha$ that travels
from $P_1$ along the branch labelled $\alpha$ in the positive
sense until it reaches the basepoint $x_\alpha$ of $\alpha$. Consider
also
the similarly defined paths $d_\beta,d_\gamma$ contained in
$\beta,\gamma$ and joining $P_2,P_3$ with $x_\beta,x_\gamma$
respectively, and the sister paths
$d_{\tau\alpha},d_{\tau\beta},d_{\tau\gamma}$ of
$d_{\alpha},d_{\beta},d_{\gamma}$ respectively, that join $P_2,P_3,P_1$
with $x_{\tau\alpha},x_{\tau\beta},x_{\tau\gamma}$ along
$\tau\alpha,\tau\beta,\tau\gamma$, respectively. If we introduce the
loops
\begin{align*}
t_\alpha&=\omega_1 d_\alpha \lambda_\alpha^{-1} &
t_\beta&=\omega_2 d_\beta \lambda_\beta^{-1} &
t_\gamma&=\omega_3 d_\gamma \lambda_\gamma^{-1} \\
t_{\tau\alpha}&=\omega_2 d_{\tau\alpha} \lambda_{\tau\alpha}^{-1} &
t_{\tau\beta}&=\omega_3 d_{\tau\beta} \lambda_{\tau\beta}^{-1} &
t_{\tau\gamma}&=\omega_1 d_{\tau\gamma} \lambda_{\tau\gamma}^{-1}
\end{align*}
we get the triple point relation around $P$:
\begin{equation}
\label{ec:3pto_relation}
t_\alpha a t_{\tau\alpha}^{-1} \,
t_\beta b t_{\tau\beta}^{-1} \,t_\gamma c t_{\tau\gamma}^{-1}=1\,.
\end{equation}
Note that the loops
$t_\alpha,t_{\tau\alpha}, t_\beta,t_{\tau\beta}, t_\gamma$ and
$t_{\tau\gamma}$ are surfacewise elements of $\Sigma$ and so they can be
expressed as words in $s_1,s_2,\ldots,s_{2g}$.
\begin{figure}
\caption{Triple point relation (Dehn surface
view)}
\label{Fig:triple-point-Sigma}
\end{figure}
\item \emph{Double curve relations}.
In~\cite{racsam,RHomotopies} this relators did not appear because in $\mathbb{S}^2$ all loops
are homotopically trivial, while in a $g$-torus this is not true.
Let $\alpha,\tau\alpha$ be a pair of sister curves
of $\mathcal{D}$. We orient both curves coherently starting from their basepoints
$x_\alpha,x_{\tau\alpha}$. The loops
$\lambda_\alpha \alpha \lambda_{\alpha}^{-1}$ and
$\lambda_{\tau\alpha} \tau\alpha \lambda_{\tau\alpha}^{-1}$ in $S$ are
related in $\Sigma$ because $\alpha$ and $\tau\alpha$ coincide in
$\Sigma$:
\[
\lambda_\alpha \alpha \lambda_{\alpha}^{-1}= \lambda_\alpha\lambda_{\tau\alpha}^{-1} \lambda_{\tau\alpha} \tau\alpha \lambda_{\tau\alpha}^{-1} \lambda_{\tau\alpha} \lambda_{\alpha}^{-1}\,.
\]
If we also denote by $\alpha$ and $\tau\alpha$ the surfacewise elements
of $\pi_1(\Sigma)$ represented by
$\lambda_\alpha \alpha \lambda_{\alpha}^{-1}$ and
$\lambda_{\tau\alpha} \tau\alpha \lambda_{\tau\alpha}^{-1}$,
respectively, this relation can be written as
\begin{equation}\label{eq:double-curve-relator}
\alpha=a \,\tau\alpha \, a^{-1}\,.
\end{equation}
\item \emph{Surface relations}. Those are the relations that the
surfacewise generators $s_1,s_2,\ldots,s_{2g}$ verify when considered as
elements of $\pi_1(S)$.
\end{enumerate}
Any relation among those specified above define an associated
\emph{dual loop relator}, \emph{triple point relator},
\emph{double curve relator} or \emph{surface relator} in the natural way.
\begin{theorem}
\label{thm:fundamental-group-presentation}
If
$\mathcal{D}=\{\alpha_1,\alpha_2,\ldots,\alpha_{2k}\}$, the fundamental group
$\pi_{1}(\Sigma)$ is isomorphic to
\[
\pi (\mathcal{D})=\langle\, s_1,s_2,\ldots,s_{2g},a_1,a_2,\ldots,a_{2k} \mid
\text{R1, R2, R3 and R4 relators}\,\rangle\,.
\]
\end{theorem}
We need to introduce some notation and a previous result before being able
to give the proof of this theorem.
If we consider the generators of $\pi (\mathcal{D})$ as elements of $\pi_1 (\Sigma)$,
we are defining implicitly a homomorphism
$\varepsilon:\pi (\mathcal{D})\to\pi_1(\Sigma)$ which is surjective by
Theorem~\ref{thm:finite-set-of-generators-pi1}. To prove that
$\varepsilon$ is also injective we will prove first that $\pi(\mathcal{D})$ behaves
exactly as $\pi_1(\Sigma)$ when dealing with covering spaces:
the covering spaces of $\Sigma$ are completely characterized by
representations of $\pi(\mathcal{D})$ on a permutation group. For $m\in\mathbb{N}$ we
denote by $\Omega_m$ the group of permutations of the finite set
$\{1,2,\ldots, m\}$, and we denote by $\Omega_\infty$ the group of
permutations of the countable set $\mathbb{N}$.
\begin{lemma}
\label{lem:representations-factorize}
Every representation $\rho:\pi (\mathcal{D})\to \Omega_m$, with $m\in\mathbb{N}\cup\{\infty\}$, factors
through $\varepsilon$, i.e., there exists a representation
$\hat\rho:\pi_1(\Sigma)\to\Omega_m$ such that the following diagram
commutes.
\[
\xymatrix{
\pi(\mathcal{D}) \ar[d]_{\varepsilon} \ar[r]^{\rho} & \Omega_m \\
\pi_1(\Sigma) \ar[ur]_{\hat\rho}
}
\]
\end{lemma}
\begin{proof}
Fix an $m\in\mathbb{N}\cup\{\infty\}$, and consider a representation $\rho:\pi(\mathcal{D})\to \Omega_m$.
As $\pi(\mathcal{D})$ verifies the surface relations we have a
natural homomorphism
\[
\zeta:\pi_1(S)\to \pi(\mathcal{D})\,.
\]
The map $\rho_S=\rho\circ \zeta$ is a representation of $pi_1(S)$, it is the
monodromy homomorphism of a $m$-sheeted covering map $h:\hat{S}\to S$.
Let $x^1,x^2,\ldots$ be the different points of $\hat{S}$ on the fiber of $h$
over $x$, labelled in such way that for any $s\in\pi_1(S)$ the lift of $s$
starting in $x^i$ has its end in $x^j$, where $j=\rho_S(s)(i)$.
If we lift the diagram $\mathcal{D}$ to $\hat{S}$, we obtain another diagram $\hat{\mathcal{D}}$ in $\hat{S}$. We will define a ``sistering'' $\hat{\tau}:\hat{\mathcal{D}}\to\hat{\mathcal{D}}$ between the curves of $\hat{\mathcal{D}}$, with the expected compatibility with $\tau$. For any curve $\alpha\in\mathcal{D}$, we call $\lambda_\alpha^i$ the lift of $\lambda_\alpha$ starting at $x^i$, and $x_\alpha^i$ the endpoint of $\lambda_\alpha^i$. The sistering $\hat{\tau}$ is defined as follows: $\hat{\tau}$ sends the curve of $\hat{\mathcal{D}}$ passing through $x_\alpha^i$ to the curve of $\hat{\mathcal{D}}$ passing through $x_{\tau\alpha}^j$, where $j=\rho(a)(i)$. The points of both curves are related by $\hat{\tau}$ starting from $x_\alpha^i,x_{\tau\alpha}^j$ lifting the relation among the the points of $\alpha$ and $\tau\alpha$ near $x_\alpha$ and $x_{\tau\alpha}$ in the natural way. If we think about the circle $\mathbb{S}^1$ as the real numbers modulo $2\pi$, we can take parametrizations $\alpha,\tau\alpha:\mathbb{R}\to S$ of the curves $\alpha,\tau\alpha$ respectively such that $\alpha(0)=x_\alpha$, $\tau\alpha(0)=x_{\tau\alpha}$ and such that $\alpha(t)$ is related by $\tau$ with $\tau\alpha(t)$ for all $t\in\mathbb{R}$. Taking lifts $\alpha^i,\tau\alpha^j:\mathbb{R}\to \hat{S}$ of these parametrizations with $\alpha^i(0)=x^i_\alpha$, $\tau\alpha^j(0)=x^j_{\tau\alpha}$ we state that $\alpha^i(t)$ is related by $\hat{\tau}$ with $\tau\alpha^j(t)$ for all $t\in\mathbb{R}$. We want to prove that $(\hat{\mathcal{D}},\hat{\tau})$ is an abstract Johansson diagram on $\hat{S}$. \begin{claim}\label{claim:2-hat-tau-well-defined} The sistering $\hat{\tau}$ is well defined: if $x_{\alpha}^i,x_{\alpha}^{i'}$ are different lifts of $x_\alpha$ lying on the same curve $\hat{\alpha}$ of $\hat{\mathcal{D}}$, then $x_{\tau\alpha}^{j}$ and $x_{\tau\alpha}^{j'}$ lie on the same curve of $\hat{\mathcal{D}}$, where $j=\rho(a)(i)$ and $j'=\rho(a)(i')$. \end{claim} \begin{proof}[Proof of Claim~\ref{claim:2-hat-tau-well-defined}] Assume first that $x_{\alpha}^i$ and $x_{\alpha}^{i'}$ are consecutive in $\hat{\alpha}$: we can travel along $\hat{\alpha}$ from $x_{\alpha}^i$ to $x_{\alpha}^{i'}$ without crossing any other point of $h^{-1}(x_{\alpha})$. Choosing appropiately the orientation of $\alpha$ or the indices $i,i'$, we can assume that the lift of $\alpha$ starting at $x_{\alpha}^i$ ends at $x_{\alpha}^{i'}$, and in terms of $\pi(\mathcal{D})$ this implies that $\rho(\alpha)(i)=i'$. Therefore, \begin{align*} \rho(\tau\alpha)(j)&=\rho(\tau\alpha)(\rho(a)(i))=\rho(a\tau\alpha)(i)\\ & =\rho(\alpha a)(i)= \rho(a)(\rho(\alpha)(i))=\rho(a)(i')=j'\,, \end{align*} because by the double curve relation \eqref{eq:double-curve-relator} it is $\alpha a=a \tau\alpha$. The points $x_{\tau\alpha}^j$ and $x_{\tau\alpha}^{j'}$ are consecutive in a lift of $\tau\alpha$ in $\hat{\mathcal{D}}$.
By repeating this argument, the statement is also true if $x_{\alpha}^i$ and $x_{\alpha}^{i'}$ are not consecutive in $\hat{\alpha}$ \end{proof} \begin{claim}\label{claim:1-hat-tau-involutive} The sistering $\hat{\tau}$ is an involution on $\hat{\mathcal{D}}$. \end{claim}
\begin{proof}[Proof of Claim~\ref{claim:1-hat-tau-involutive}]
We have $\tau a=a^{-1}$ in $\pi(\mathcal{D})$ by dual loop relations. Then $\rho(\tau a)=\rho(a)^{-1}$ because $\rho$ is a homomorphism. Therefore if we have $\rho(a)(i)=j$, it is also $\rho(\tau a)(j)=i$. In other words, the sister of the curve of $\hat{\mathcal{D}}$ passing through $x_\alpha^i$ is the curve of $\hat{\mathcal{D}}$ passing through $x_{\tau\alpha}^j$, and the sister of the curve of $\hat{\mathcal{D}}$ passing through $x_{\tau\alpha}^j$ is the curve of $\hat{\mathcal{D}}$ passing through $x_{\alpha}^i$. \end{proof}
The sistering $\hat{\tau}$ defines an equivalence relation into the points of $\hat{\mathcal{D}}$.
By Claims~\ref{claim:2-hat-tau-well-defined} and~\ref{claim:1-hat-tau-involutive}, each point of $\hat{\mathcal{D}}$ which is not a crossing has exactly another point related with it.
\begin{claim}
\label{claim:4-crossings-in-triples}
The crossings of $\hat{\mathcal{D}}$ are related in triplets by $\hat{\tau}$. \end{claim}
\begin{proof}[Proof of Claim~\ref{claim:4-crossings-in-triples}] Let $P$ be a triple point of $\Sigma$, and let $P_1,P_2,P_3$ be the three crossings of $\mathcal{D}$ in the triplet of $P$. We label the curves of $\mathcal{D}$ intersecting at $P_1,P_2,P_3$ as in Figure~\ref{Fig:triple-point-diagram}, and we consider the paths $$\omega_1,\omega_2,\omega_3,\quad d_{\alpha},d_{\tau\alpha},d_{\beta},d_{\tau\beta},d_{\gamma},d_{\tau\gamma},\quad t_{\alpha},t_{\tau\alpha},t_{\beta},t_{\tau\beta},t_{\gamma},t_{\tau\gamma}$$ as in the construction of the triple point relations (Figure~\ref{Fig:triple-point-diagram}).
For each $n=1,2,3$ and $i=1,2,\ldots$, let $\omega_n^i$ be the lift of $\omega_n$ starting at $x^{i}$, and let $P_n^i$ be the endpoint of $\omega_n^i$. If we choose one of the lifts $P_1^i$ of $P_1$, it is related by $\hat{\tau}$ with one lift $P_2^j$ of $P_2$ by means of the sister curve of the lift of $\alpha$ passing through $P_1^i$. Let label $\alpha^i$ the lift of $\alpha$ passing through $P_1^i$. In order to find $P_2^j$ we need to find the sister curve of $\alpha^i$, and to do this we follow $\alpha^i$ until we reach a lift of $x_\alpha$. This is made by taking the lift of $d_\alpha$ starting at $P_1^i$, which ends at a lift $x_\alpha^{i'}$ of $x_\alpha$. The lift of $\lambda_\alpha$ ending at $x_\alpha^{i'}$ starts at $x^{i'}$, where $i'=\rho(t_\alpha)(i)$. The sister curve of $\alpha^i$ is the lift of $\tau \alpha$ passing through $x_{\tau\alpha}^{j'}$, where $j'=\rho(a)(i')$. Finally, $P_2^j$ is located at the starting point of the lift of $d_{\tau \alpha}$ ending at $x_{\tau\alpha}^{j'}$, and it verifies that $j=\rho(t_{\tau\alpha}^{-1})(j')$. In other words, it is $$j=\rho(t_\alpha a t_{\tau\alpha}^{-1})(i)\,.$$ By the same argument, we have that $P_2^j$ is related with $P_3^\ell$, with $\ell=\rho(t_\beta b t_{\tau\beta}^{-1})(j)$ and that $P_3^\ell$ is related with $P_1^{i^*}$, where $i^*=\rho(t_\gamma c t_{\tau\gamma}^{-1})(\ell)$. Triple point relation~\eqref{ec:3pto_relation} at $P$ implies that $i^*=i$, and thus there is no more points related with $P_1^i$ different from $P_2^j$ and $P_3^\ell$. This construction is valid in general and so the crossings of $\mathcal{D}$ are related in triples by $\hat{\tau}$. \end{proof}
We have proved that $(\hat{\mathcal{D}},\hat{\tau})$ is an abstract Johansson diagram in $\hat{S}$: points in $\hat{\mathcal{D}}$ are identified in pairs with the exception of crossings, which are identified in triplets. Moreover, this equivalence relation is compatible with $f$, that is, if $\hat{Y},\hat{Z}\in\hat{S}$ are related by $\hat{\tau}$, their projections $h(\hat{Y}),h(\hat{Z})\in S$ are related by $\tau$. The quotient space $\hat{\Sigma}$ of $\hat{S}$ under the equivalence relation defined by $\hat{\tau}$ is a (pseudo) Dehn surface and we can define a map $\hat{h}:\hat{\Sigma}\to\Sigma$ making the following diagram commutative: \[
\xymatrix{
\hat{S} \ar[r]^{\hat{f}} \ar[d]_{h} & \hat\Sigma \ar[d]^{\hat{h}} \\
S \ar[r]_ {f} & \Sigma
} \] where $\hat{f}:\hat{S}\to\hat\Sigma$ is the quotient map. The map $\hat{h}$ is in fact an $m$-sheeted covering map whose monodromy homomorphism $\hat\rho$ verifies by construction $\rho=\hat\rho\circ\varepsilon$. \end{proof}
\begin{proof}[Proof of Theorem~\ref{thm:fundamental-group-presentation}]
Due to Cayley's Theorem ~\cite[p. 9]{Hall}, every group admits an injective
representation into the group of permutations of its own elements. Because
$\pi(\mathcal{D})$ is finite or countable, the group of permutations of its elements
is isomorphic to $\Omega_m$, for an $m\in\mathbb{N}\cup\{\infty\}$. Thus, we can construct a
faithful representation $\rho:\pi(\mathcal{D})\to\Omega_m$, and by
Lemma~\ref{lem:representations-factorize}, this implies that $\varepsilon$
must be injective. \end{proof}
Hurewicz's Theorem implies that the first homology group of $\Sigma$ is isomorphic to the abelianization of its fundamental group. Let us see how to adapt the presentation of Theorem~\ref{thm:fundamental-group-presentation} to the abelian case. Following the notation of above: \begin{enumerate}[\em({AR}1)]
\item \emph{Dual loop relations}.\label{AR1}
These relations remains the same, if
$\alpha$ and $\tau\alpha$ are sister curves, their preferred dual loops
satisfy $a = -\tau a$.
\item \emph{Triple point relations}. The triple points relations can be
simplified using commutativity to
\[
a + b + c = w,
\]
where $w$ is a word in surfacewise generators and $a$, $b$ and $c$ are
preferred dual loops of double curves.
\item \emph{Double curve relations}. Conjugacies on commuting elements are
equalities, so the double curve relation can be written as
$\alpha = \tau\alpha$.
\item \emph{Surface relations}. The generators of the fundamental group of
the $g$-torus $S$ are related by a product of commutators. In the abelian
context, this relation is trivial.
\end{enumerate} Relations~AR\ref{AR1} allow us to select $k$ generators $a_1,a_2,\ldots,a_k$ from the set of preferred dual loops, dropping $k$ of them. Finally,
\begin{theorem}
\label{thm:homology}
The abelian group $H_1(\Sigma)$ is isomorphic to
\[
H_1(\mathcal{D}) = \langle\, s_1,s_2,\ldots,s_{2g},a_1,a_2,\ldots,a_k
\mid \text{AR2 and AR3 relations} \,\rangle.
\] \end{theorem}
\begin{remark}\label{rmk:homology-M-Sigma}
If $\Sigma$ is filling, it is the $2$-skeleton of $M$ as a CW-complex, so
$\pi_1(\Sigma)=\pi_1(M)$. Hence $H_1(\Sigma)=H_1(M)$, which is isomorphic to
the second homology group due to Poincar\'e duality. In this way,
Theorem~\ref{thm:homology} characterizes completely the homology of the
closed orientable manifold $M$. \end{remark}
\section{Checkerboards and homology spheres}\label{sec:checkers-homology-spheres}
Consider a closed orientable $3$-manifold $M$ and a Dehn surface
$S\to\Sigma\subset M$. We say \emph{$M$ is checkered by $\Sigma$} if
we can assign a
color from $\{0,1\}$ to each region of $\Sigma$ such that if two of them
share a face they have different colors. In the same way, $S$ is checkered by
the Johansson diagram $\mathcal{D}$ of $\Sigma$ if we can assign a color from $\{0,1\}$ to
each face of $S-\mathcal{D}$ such that if two of them share an edge they have
different colors. Checkerboards are a usual tool in knot theory
(see \cite{Carter-Kamada-Saito}, for example).
\begin{lemma}
\label{lem:checkered_paths}
The manifold $M$ is checkered by $\Sigma$ if and only if any loop
on $M$ meeting $\Sigma$ transversely on the faces intersects $\Sigma$ in an even
number of points.
\end{lemma}
\begin{proof}
Given the decomposition of $M$ in regions of $\Sigma$, we can build a graph $G_\Sigma$
describing it combinatorially. Each vertex corresponds to a
region and there is an edge connecting two vertices if the corresponding
regions share a face. It is clear that $M$ is checkered by $\Sigma$ if and only if
$G_\Sigma$ is $2$-colourable.
Each path $\gamma$ in $M$ meeting $\Sigma$ transversely on the faces has an associated
path $c_\gamma$ in $G_\Sigma$ that encodes the sequence of ``crossings'' of $\gamma$ between
different regions of $\Sigma$ across the faces of $\Sigma$. The number of points of $\gamma\cap\Sigma$ is just the
number of edges of $c_\gamma$, and if $\gamma$ is closed, $c_\gamma$ is closed too.
As a graph is $2$-colorable if and only if it is bipartite, the statement
of the lemma is equivalent to: $G_\Sigma$ is bipartite if and only if all
loops have an even number of edges, which is trivially true.
\end{proof}
\begin{proposition}
\label{prop:checkered}
A $\mathbb{Z}/2$-homology $3$-sphere is checkered by any
Dehn surface.
\end{proposition}
\begin{proof}
Let $M$ be a $\mathbb{Z}/2$-homology $3$-sphere and let $\Sigma$ be a Dehn surface in
$M$. Let $\gamma$ be a loop in $M$ intersecting $\Sigma$ transversely at
the faces. Both $\gamma$ and $\Sigma$ are $\mathbb{Z}/2$-nullhomologous.
Therefore the set of intersection points of $\gamma$ and $\Sigma$ must be a
$\mathbb{Z}/2$-nulhomologous $0$-cycle, and then, the number of intersection points
should be even. Lemma~\ref{lem:checkered_paths} completes the proof.
\end{proof}
If the surface $S$ is orientable, we can fix a well defined normal field on
$\Sigma$ in $M$. We assign to each face of $\Sigma$ the color of
the region pointed by the normal vector field on the face. The faces of $S$
sharing an edge cannot share also the color, as the corresponding regions
share a face (see Figure~\ref{fig:checkers}). Therefore:
\begin{figure}
\caption{Coloring faces with region colors.}
\label{fig:checkers}
\end{figure}
\begin{lemma}\label{lem:orientable-checkered-surface-checkered-diagram}
If $M$ and $S$ are orientable and $M$ is checkered by $\Sigma$, the
surface $S$ is also checkered by $\mathcal{D}$.
\end{lemma}
The converse of this lemma does not hold. The 2-sphere $\mathbb{S}^2$ is
checkered by the diagram of Figure~\ref{Fig:audi}, but
$\mathbb{S}^2\times \mathbb{S}^1$ is not checkered by the corresponding
filling Dehn sphere. The existence of checkerboards has strong
implications on the number of triple points.
\begin{proposition}
\label{prop:checkered-even}
If $M$ is checkered by $\Sigma$, the number of triple points of $\Sigma$ is
even.
\end{proposition}
\begin{proof}
For each curve of $\mathcal{D}$, consider a \emph{neighbouring curve} running parallel to it and intersecting $\mathcal{D}$ transversely near the crossings of $\mathcal{D}$
(see~\cite[Fig. 26]{RHomotopies}). As $S$ is
checkered by $\mathcal{D}$, by a similar argument to that of the proof of
Lemma~\ref{lem:checkered_paths} each neighbouring curve must intersect $\mathcal{D}$
in an even number of points.
Because the number of curves in $\mathcal{D}$ is even, the total amount of intersection
points between neighbouring curves and diagram curves is a multiple of $4$.
Each crossing of the diagram
corresponds to two of those intersections, so the number of crossings is
even. Finally, the number of triple points is also even, as it is the
number of crossings over three.
\end{proof}
From Remark~\ref{rmk:at-least-1+2g-triple-points} and Proposition~\ref{prop:checkered-even} we get:
\begin{theorem}
\label{thm:spectrum_bound}
Let $\Sigma$ be a filling Dehn $g$-torus
of $M$. If $M$ is checkered by $\Sigma$, the number of
triple points of $\Sigma$ is greater than or equal to $2g$. \qed
\end{theorem}
All these results allows us to give a sharper lower bound for the
triple point numbers of $\mathbb{Z}/2$-homology $3$-spheres.
\begin{theorem}\label{thm:spectrum-homology-sphere_bound}
If $M$ is a $\mathbb{Z}/2$-homology $3$-sphere
\begin{equation*}
t_g(M) \geq 2 + 2g.
\end{equation*}
\end{theorem}
\begin{proof}
By Theorem~\ref{thm:spectrum_bound} it is enough to show that
the number of triple points cannot be $2g$.
Let $\Sigma$ be a Dehn $g$-torus in $M$ with exactly $2g$ triple
points. Let $S$ be the domain of $\Sigma$.
Let us compute the first homology group of $\Sigma$ with coefficients in
$\mathbb{Z}/2$. By Theorem~\ref{thm:homology} this abelian group is
generated by $s_1,s_2,\ldots,s_{2g}$, the curves generating $H_1(S)$, and the
preferred dual loops $a_1,a_2,\ldots,a_{k}$ extracted from the $k$ double curves of $\Sigma$.
The relations come from the $2g$ AR2 relations and the $k$ AR3 relations by
just taking coefficients in $\mathbb{Z}/2$. With this presentation $H_1(\Sigma;\mathbb{Z}/2)$
has the same number of generators as of relators.
Consider only the $k$ AR3 relators. If $\alpha$ and $\tau\alpha$
are sister curves, as they are surfacewise curves they can be written
(up to homology, see \eqref{eq:double-curve-relator}) as
\begin{align*}
\alpha &= \sum_{i=1}^{2g} n^i_\alpha s_i, &
\tau\alpha &= \sum_{i=1}^{2g} m^i_\alpha s_i,
\end{align*}
with $n_\alpha^i,m_\alpha^i\in\mathbb{Z}/2$. Therefore, the relation AR3 relation
given by the pair $\alpha,\tau\alpha$ is
\[
\sum_{i=1}^{2g} q^i_\alpha s_i=0\,,
\]
with $q^i_\alpha=n^i_\alpha + m^i_\alpha$.
All this information can be written as a tableau
\[
\begin{array}{c|cccc}
& s_1 & s_2 & \cdots & s_{2g} \\[.5mm]
\hline
\alpha_1\vphantom{\raisebox{2mm}1}
& q_{\alpha_1}^1 & q_{\alpha_1}^2 & \cdots & q_{\alpha_1}^{2g}
\\[1mm]
\alpha_2 & q_{\alpha_2}^1 & q_{\alpha_2}^2 & \cdots & q_{\alpha_2}^{2g}
\\
\vdots & \vdots & \vdots & \ddots & \vdots \\[1mm]
\alpha_k & q_{\alpha_k}^1 & q_{\alpha_k}^2 & \cdots & q_{\alpha_k}^{2g}
\\[1mm]
\end{array}
\]
Rows in this tableau represents double curves relators.
Using intersection theory, it can be seen that columns count
intersection between the surfacewise generators and diagram curves.
By Proposition~\ref{prop:checkered} and Lemma~\ref{lem:orientable-checkered-surface-checkered-diagram}
the surface $S$ is checkered by $\mathcal{D}$, so the intersection number of each curve $s_i$
with
the diagram should be even. This implies that the sum of each column is $0\pmod2$.
In other words, one of the AR3 relators can be written as a linear combination
of the rest. Therefore, the group $H_1(\Sigma;\mathbb{Z}/2)$ has
$2g+k$ generators and at most $2g+k-1$ nontrivial relations, and it cannot be trivial.
By Remark~\ref{rmk:homology-M-Sigma} this implies that $\Sigma$ cannot fill the $\mathbb{Z}/2$-homology $3$-sphere $M$.
\end{proof}
It is well known that $t_0(\mathbb{S}^3)=2$ (see~\cite{racsam,tesis}).
From Lemma~\ref{lem:triple-point-inequality} and
Theorem~\ref{thm:spectrum-homology-sphere_bound}, $\mathbb{S}^3$ cannot have exceptional Dehn $g$-tori, therefore $\mathscr{H}(\mathbb{S}^3)=0$ and finally
\begin{corollary}
$\mathscr{T}(\mathbb{S}^3)=(2,4,6,\ldots)$.\qed
\end{corollary}
\end{document}
|
arXiv
|
{
"id": "1412.1637.tex",
"language_detection_score": 0.7489063739776611,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\sloppy
\title{Path Deviations Outperform Approximate Stability in Heterogeneous Congestion Games}
\begin{abstract} We consider non-atomic network congestion games with heterogeneous players where the latencies of the paths are subject to some bounded deviations. This model encompasses several well-studied extensions of the classical Wardrop model which incorporate, for example, risk-aversion, altruism or travel time delays. Our main goal is to analyze the worst-case deterioration in social cost of a \emph{deviated\xspace Nash flow} (i.e., for the perturbed latencies) with respect to an original Nash flow.
We show that for homogeneous players deviated\xspace Nash flows coincide with approximate Nash flows and derive tight bounds on their inefficiency. In contrast, we show that for heterogeneous populations this equivalence does not hold. We derive tight bounds on the inefficiency of both deviated\xspace and approximate Nash flows for \emph{arbitrary} player sensitivity distributions.
Intuitively, our results suggest that the negative impact of path deviations (e.g., caused by risk-averse behavior or latency perturbations) is less severe than approximate stability\xspace (e.g., caused by limited responsiveness or bounded rationality).
We also obtain a tight bound on the inefficiency of deviated\xspace Nash flows for matroid congestion games and homogeneous populations if the path deviations can be decomposed into edge deviations.
In particular, this provides a tight bound on the Price of Risk-Aversion for matroid congestion games. \end{abstract}
\section{Introduction}
In 1952, Wardrop~\cite{Wardrop1952} introduced a simple model, also known as the \emph{Wardrop model}, to study outcomes of selfish route choices in traffic networks which are affected by congestion. In this model, there is a continuum of non-atomic players, each controlling an infinitesimally small amount of flow, whose goal is to choose paths in a given network to minimize their own travel times. The latency (or delay) of each edge is prescribed by a non-negative, non-decreasing latency function which depends on the total flow on that edge.
Ever since its introduction, the Wardrop model has been used extensively, both in operations research and traffic engineering studies, to investigate various aspects of selfish routing in networks.
More recently, the classical Wardrop model has been extended in various ways to capture more complex player behaviors. Examples include the incorporation of uncertainty attitudes (e.g., risk-aversion, risk-seeking), cost alterations (e.g., latency perturbations, road pricing), other-regarding dispositions (e.g., altruism, spite) and player biases (e.g., responsiveness, bounded rationality).
Several of these extensions can be viewed as defining some modified cost for each path which combines the original latency with some `deviation' (or perturbation) along that path. Such deviations are said to be \emph{$\beta$-bounded} if the total deviation along each path is at most $\beta$ times the latency of that path. The player objective then becomes to minimize the combined cost of latency and deviation along a path (possibly using different norms). An equilibrium outcome corresponds to a \emph{$\beta$-deviated\xspace Nash flow}, i.e., a Nash flow with respect to the combined cost. The deviations might be given explicitly (e.g., as in the altruism model of Chen et al.~\cite{Chen2014}) or be defined implicitly (e.g., as in the risk-aversion model of Nikolova and Stier-Moses~\cite{Nikolova2015}). Further, different fractions of players might perceive these deviations differently, i.e., players might be heterogeneous with respect to the deviations.
Another extension, which is closely related to the one above, is to incorporate different degrees of `responsiveness' of the players. For example, each player might be willing to deviate to an alternative route only if her latency decreases by at least a certain fraction. In this context, an equilibrium outcome corresponds to an \emph{$\epsilon$-approximate Nash flow} for some $\epsilon \ge 0$, i.e., for each player the latency is at most $(1+\epsilon)$ times the latency of any other path. Here, $\epsilon$ is a parameter which reflects the responsiveness of the players. An analogue definition can be given for populations with heterogeneous responsiveness parameters.
To illustrate the relation between deviated\xspace and approximate Nash flows, suppose we are given a $\beta$-deviated\xspace Nash flow $f$ for some $\beta \geq 0$, where the latency $\ell_P(f)$ of each path $P$ is perturbed by an arbitrary $\beta$-bounded deviation $\delta_P(f)$ satisfying $0 \leq \delta_P(f) \leq \beta l_P(f)$. Intuitively, the deviations inflate the latency on each path by at most a factor of $(1+\beta)$. Further, assume that the population is homogeneous. From the Nash flow conditions (see Section~\ref{sec:pre} for formal definitions), it follows trivially that $f$ is also an $\epsilon$-approximate Nash flow with $\epsilon = \beta$. But does the converse also hold? That is, can every $\epsilon$-approximate Nash flow be induced by a set of bounded path deviations? More generally, what about the relation between deviated\xspace and approximate Nash flows for heterogenous populations? Can we bound the inefficiency of these flows?
In this paper, we answer these questions by investigating the relation between the two equilibrium notions. Our main goal is to quantify the inefficiency of deviated\xspace and approximate Nash flows, both for homogeneous and heterogeneous populations. To this aim, we study the (relative) worst-case deterioration in social cost of a $\beta$-deviated\xspace Nash flow with respect to an original (unaltered) Nash flow; we use the term \emph{$\beta$-deviation\xspace ratio} to refer to this ratio. This ratio has recently been studied in the context of risk aversion \cite{Lianeas2016,Nikolova2015} and in the more general context of bounded path deviations \cite{Kleer2016}. Similarly, for approximate Nash flows we are interested in bounding the \emph{$\epsilon$-stability\xspace ratio}, i.e., the worst-case deterioration in social cost of an $\epsilon$-approximate Nash flow with respect to an original Nash flow.
Note that these notions differ from the classical \emph{price of anarchy} notion~\cite{Koutsoupias1999}, which refers to the worst-case deterioration in social cost of a $\beta$-deviated\xspace (respectively, $\varepsilon$-approximate) Nash flow with respect to an \emph{optimal} flow. While the price of anarchy typically depends on the class of latency functions (see, e.g., \cite{Chen2014,Christodoulou2011,Kleer2016,Nikolova2015} for results in this context), the deviation\xspace ratio is independent of the latency functions but depends on the topology of the network (see \cite{Kleer2016,Nikolova2015}).
\paragraph{Our contributions.}
The main contributions of this paper are as follows:
\begin{enumerate}\itemsep7pt
\item We show that for homogeneous populations the set of $\beta$-deviated\xspace Nash flows coincides with the set of $\epsilon$-approximate Nash flows for $\beta = \epsilon$. Further, we derive an upper bound on the $\epsilon$-stability\xspace ratio (and thus also on the $\epsilon$-deviation\xspace ratio) which is at most $(1+\epsilon)/(1-\epsilon n)$, where $n$ is the number of nodes, for single-commodity networks. We also prove that the upper bound we obtain is tight for \emph{generalized Braess graphs}. These results are presented in Section~\ref{sec:approx}.
\item We prove that for heterogenous populations the above equivalence does not hold. We derive tight bounds for both the $\beta$-deviation\xspace ratio and the $\epsilon$-stability\xspace ratio for single-commodity instances on series-parallel graphs and arbitrary sensitivity distributions of the players. To the best of our knowledge, these are the first inefficiency results in the context of heterogenous populations which are tight for \emph{arbitrary} sensitivity distributions. Our bounds show that both ratios depend on the demands and sensitivity distribution $\gamma$ of the heterogenous players (besides the respective parameters $\beta$ and $\epsilon$). Further, it turns out that the $\beta$-deviation\xspace ratio is always at most the $\epsilon$-stability\xspace ratio for $\epsilon = \beta \gamma$. These results are given in Section~\ref{sec:het}.
\item We also derive a tight bound on the $\beta$-deviation\xspace ratio for single-commodity matroid congestion games and homogeneous populations if the path deviations can be decomposed into edge deviations. To the best of our knowledge, this is the first result in this context which goes beyond network congestion games. In particular, this gives a tight bound on the Price of Risk-Aversion \cite{Nikolova2015} for matroid congestion games. This result is of independent interest and presented in Section~\ref{sec:approx}.
\end{enumerate}
In a nutshell, our results reveal that for homogeneous populations there is no quantitative difference between the inefficiency of deviated\xspace and approximate Nash flows in the worst case. In contrast, for heterogenous populations the $\beta$-deviation\xspace ratio is always at least as good as the $\epsilon$-stability\xspace ratio with $\epsilon = \beta \gamma$. Intuitively, our results suggest that the negative impact of path deviations (e.g., caused by risk-averse behavior or latency perturbations) is less severe than approximate stability\xspace (e.g., caused by limited responsiveness or bounded rationality).
\paragraph{Related work.}
We give a brief overview of the works which are most related to our results. Christodoulou et al. \cite{Christodoulou2011} study the inefficiency of approximate equilibria in terms of the price of anarchy and price of stability (for homogeneous populations). Generalized Braess graphs were introduced by Roughgarden \cite{Roughgarden2006} and are used in many other lower bound constructions (see, e.g., \cite{Englert2008,Kleer2016,Roughgarden2006}). Chen et al. \cite{Chen2014} study an altruistic extension of the Wardrop model and, in particular, also consider heterogeneous altruistic populations. They obtain an upper bound on the ratio between an altruistic Nash flow and a social optimum for parallel graphs, which is tight for two sensitivity classes. It is mentioned that this bound is most likely not tight in general. Meir and Parkes \cite{Meir2014Arxiv} study player-specific cost functions in a smoothness framework \cite{Roughgarden2015}. Some of their inefficiency results are tight, although none of their bounds seems to be tight for arbitrary sensitivity distributions. Matroids have also received some attention in the Wardrop model. In particular, Fujishige et al. \cite{Fujishige2015} show that matroid congestion games are immune against the Braess paradox (and their analysis is tight in a certain sense). We refer the reader to \cite{Kleer2016} for additional references and relations of other models to the bounded path deviation model considered here.
\section{Preliminaries}\label{sec:pre}
Let $\mathcal{I} = (E,(l_e)_{e \in E},(\mathcal{S}_i)_{i \in [k]},(r_i)_{i \in [k]})$ be an instance of a non-atomic congestion game. Here, $E$ is the set of resources (or edges, or arcs) that are equipped with a non-negative, non-decreasing, continuous latency function $l_e: \mathbb{R}_{\ge 0} \rightarrow \mathbb{R}_{\ge 0}$. Each commodity $i \in [k]$ has a strategy set $\mathcal{S}_i \subseteq 2^E$ and demand $r_i \in \mathbb{R}_{>0}$. Note that in general the strategy set $\mathcal{S}_i$ of player $i$ is defined by arbitrary resource subsets. If each strategy $P \in \mathcal{S}_i$ corresponds to an $s_i, t_i$-path in a given directed graph, then the corresponding game is called a \emph{network} congestion game.\footnote{If a network congestion game with a single commodity is considered (i.e., $k = 1$), we omit the commodity index for ease of notation.} We slightly abuse terminology and use the term \emph{path} also to refer to a strategy $P \in \mathcal{S}_i$ of player $i$ (which does not necessarily correspond to a path in a graph); no confusion shall arise. We denote by $\mathcal{S} = \cup_i \mathcal{S}_i$ the set of all paths.
An outcome of the game is a (feasible) flow $f^i: \mathcal{S}_i \rightarrow \mathbb{R}_{\geq 0}$ satisfying $\sum_{P \in \mathcal{S}_i} f_P^i = r_i$ for every $i \in [k]$. We use $\mathcal{F}(\mathcal{S})$ to denote the set of all feasible flows $f = (f^1,\dots,f^k)$.
Given a flow $f = (f^i)_{i \in [k]} \in \mathcal{F}(\mathcal{S})$, we use $f^i_e$ to denote the total flow on resource $e \in E$ of commodity $i \in [k]$, i.e., $f_e^i = \sum_{P \in \mathcal{S}_i : e \in P} f_P^i$. The total flow on edge $e \in E$ is defined as $f_e = \sum_{i \in [k]} f_e^i$.
The latency of a path $P \in \mathcal{S}$ with respect to $f$ is defined as $l_P(f) := \sum_{e \in P} l_e(f_e)$. The cost of commodity $i$ with respect to $f$ is $C_i(f) = \sum_{P \in \mathcal{S}_i} f_P l_P(f)$. The \textit{social cost} $C(f)$ of a flow $f$ is given by its total average latency, i.e., $C(f) = \sum_{i \in [k]} C_i(f) = \sum_{e \in E} f_e l_e(f_e)$. A flow that minimizes $C(\cdot)$ is called \textit{(socially) optimal}.
If the population is heterogenous, then each commodity $i \in [k]$ is further partitioned in $h_i$ \emph{sensitivity classes}, where class $j \in [h_i]$ has demand $r_{ij}$ such that $r_i = \sum_{j \in [h_i]} r_{ij}$. Given a path $P \in \mathcal{S}_i$, we use $f_{P,j}$ to refer to the amount of flow on path $P$ of sensitivity class $j$ (so that $\sum_{j \in [h_i]} f_{P,j} = f_P$).
\begin{comment} We say that $f$ is a \emph{Nash flow} if \begin{equation} \forall i \in [k], \forall P \in \mathcal{S}_i, f_P > 0: \qquad l_{P}(f) \leq l_{P'}(f) \ \ \forall P' \in \mathcal{S}_i. \label{def:nash} \end{equation} Note that under a Nash flow $f$ all flow-carrying paths $P \in \mathcal{S}_i$ of commodity $i \in [k]$ have the same latency. \end{comment}
\paragraph{Deviated Nash flows.}
We consider a \emph{bounded deviation model} similar to the one introduced in \cite{Kleer2016}.\footnote{In fact, in \cite{Kleer2016} more general path deviations are introduced; the path deviations considered here correspond to \emph{$(0, \beta)$-path deviations} in \cite{Kleer2016}.} We use $\delta= (\delta_P)_{P \in \mathcal{S}}$ to denote some arbitrary path deviations, where $\delta_P : \mathcal{F}(\mathcal{S}) \rightarrow \mathbb{R}_{\geq 0}$ for all $P \in \mathcal{S}$. Let $\beta \geq 0$ be fixed. Define the set of \emph{$\beta$-bounded path deviations} as $
\Delta(\beta) = \{ (\delta_P)_{P \in \mathcal{S}} \ | \ 0 \leq \delta_P(f) \leq \beta l_P(f) \text{ for all } f \in \mathcal{F}(\mathcal{S}) \}. $
Every commodity $i \in [k]$ and sensitivity class $j \in [h_i]$ has a non-negative sensitivity $\gamma_{ij}$ with respect to the path deviations. The population is \emph{homogeneous} if $\gamma_{ij} = \gamma$ for all $i \in [k]$, $j \in [h_i]$ and some $\gamma \ge 0$; otherwise, it is \emph{heterogeneous}. Define the \emph{deviated latency} of a path $P \in \mathcal{S}_i$ for sensitivity class $j \in [h_i]$ as $q_{P}^j(f) = l_P(f) + \gamma_{ij} \delta_P(f)$.
We say that a flow $f$ is a \emph{$\beta$-deviated\xspace Nash flow} if there exist some $\beta$-bounded path deviations $\delta \in \Delta(\beta)$ such that \begin{equation} \forall i \in [k], \forall j \in [h_i], \forall P \in \mathcal{S}_i, f_{P,j} > 0: \qquad q_{P}^j(f) \leq q_{P'}^j(f) \ \ \forall P' \in \mathcal{S}_i. \label{eq:nash} \end{equation}
We define the \emph{$\beta$-deviation\xspace ratio} $\text{$\beta$-DR}(\mathcal{I})$ as the maximum ratio $C(f^\beta)/C(f^0)$ of an $\beta$-deviated\xspace Nash flow $f^\beta$ and an original Nash flow $f^0$. Intuitively, the deviation ratio measures the worst-case deterioration in social cost as a result of (bounded) deviations in the path latencies. Note that here the comparison is done with respect to an \emph{unaltered} Nash flow to measure the impact of these deviations.
The set $\Delta(\beta)$ can also be restricted to path deviations which are defined as a function of edge deviations along that path. Suppose every edge $e \in E$ has a deviation $\delta_e : \mathbb{R}_{\geq 0} \rightarrow \mathbb{R}_{\geq 0}$ satisfying $0 \leq \delta_e(x) \leq \beta l_e(x)$ for all $x \geq 0$. For example, feasible path deviations can then be defined by the $L_1$-norm objective $\delta_P(f) = \sum_{e \in P} \delta_e(x)$ (as in \cite{Kleer2016,Nikolova2015}) or the $L_2$-norm objective $\delta_P(f) = \sqrt{\sum_{e \in P} \delta_e(x)^2)}$ (as in \cite{Nikolova2015,Lianeas2016}).
The \emph{Price of Risk-Aversion} introduced by Nikolova and Stier-Moses \cite{Nikolova2015} is technically the same ratio as the deviation\xspace ratio for the $L_1$- and $L_2$-norm (see \cite{Kleer2016} for details).
\paragraph{Approximate Nash flows.}
We introduce the notion of an approximate Nash flow. Also here, each commodity $i \in [k]$ and sensitivity class $j \in [h_i]$ has a non-negative sensitivity $\epsilon_{ij}$. We say that the population is \emph{homogeneous} if $\epsilon_{ij} = \epsilon$ for all $i \in [k]$, $j \in [h_i]$ and some $\epsilon \ge 0$; otherwise, it is \emph{heterogeneous}.
A flow $f$ is an \emph{$\epsilon$-approximate Nash flow} with respect to sensitivities $\epsilon = (\epsilon_{ij})_{i \in [k], j \in [h_i]}$ if \begin{equation} \forall i \in [k], \ \forall j \in [h_i], \ \forall P \in \mathcal{S}_i, f_{P,j} > 0: \qquad l_{P}(f) \leq (1+\epsilon_{ij}) l_{P'}(f) \ \ \forall P' \in \mathcal{S}_i \label{def:epsnash} \end{equation} Note that a $0$-approximate Nash flow is simply a Nash flow. We define the \textit{$\epsilon$-stability\xspace ratio} $\text{$\epsilon$-SR}(\mathcal{I})$ as the maximum ratio $C(f^\epsilon)/C(f^0)$ of an $\epsilon$-approximate Nash flow $f^\epsilon$ and an original Nash flow $f^0$.
Some of the proofs are missing in the main text below and can be found in the appendix.
\section{Heterogeneous populations}\label{sec:het}
We first elaborate on the relation between deviated\xspace and approximate Nash flows for general congestion games with heterogeneous populations.
\begin{proposition}\label{prop:inclusion} Let $\mathcal{I}$ be a congestion game with heterogeneous players. If $f$ is a $\beta$-deviated\xspace Nash flow for $\mathcal{I}$, then $f$ is an $\epsilon$-approximate Nash flow for $\mathcal{I}$ with $\epsilon_{ij} = \beta \gamma_{ij}$ for all $i \in [k]$ and $j \in [h_i]$ (for the same demand distribution $r$). \end{proposition}
\paragraph{Discrete sensitivity distributions.}
Subsequently, we show that the reverse of Proposition \ref{prop:inclusion} does not hold.
We do this by providing tight bounds on the $\beta$-deviation\xspace ratio and the $\epsilon$-stability ratio for instances on (single-commodity) series-parallel graphs and arbitrary discrete sensitivity distributions.
\begin{theorem}\label{thm:hetero} Let $\mathcal{I}$ be a single-commodity network congestion game on a series-parallel graph with heterogeneous players, demand distribution $r = (r_i)_{i \in [h]}$ normalized to $1$, i.e., $\sum_{j \in [h]} r_i = 1$, and sensitivity distribution $\gamma = (\gamma_i)_{i \in [h]}$, with $\gamma_1 < \gamma_2 < \dots < \gamma_h$.
Let $\beta \ge 0$ be fixed and define $\epsilon = (\beta \gamma_i)_{i \in [h]}$. Then the $\epsilon$-stability ratio and the $\beta$-deviation\xspace ratio are bounded by: \begin{align} \text{$\epsilon$-SR}(\mathcal{I}) \le 1 + \beta \sum_{j = 1}^h r_j\gamma_j \quad\text{and}\quad \text{$\beta$-DR}(\mathcal{I}) \le 1 + \beta \cdot \max_{j \in [h]} \bigg\{ \gamma_j \bigg( \sum_{p = j}^h r_p \bigg) \bigg\}. \label{eq:comparison_epsilon}
\end{align} Further, both bounds are tight for all distributions $r$ and $\gamma$. \end{theorem}
It is not hard to see that the bound on the $\beta$-deviation\xspace ratio is always smaller than the bound on the $\epsilon$-stability ratio.\footnote{This follows from Markov's inequality: for a random variable $Y$, $P(Y \geq t) \leq E(Y)/t$.} Our bound on the $\beta$-deviation\xspace ratio also yields tight bounds on the \emph{Price of Risk-Aversion} \cite{Nikolova2015} for series-parallel graphs and arbitrary heterogeneous risk-averse populations, both for the $L_1$-norm and $L_2$-norm objective.\footnote{Observe that we show tightness of the bound on parallel arcs, in which case these objectives coincide.}
We need the following technical lemma for the proof of the $\beta$-deviation ratio.
\begin{lemma}\label{lem:max_seq} Let $0 \leq \tau_{k-1} \leq \dots \leq \tau_1 \leq \tau_0$ and $c_i \geq 0$ for $i = 1,\dots,k$ be given. We have $ c_1 \tau_0 + \sum_{i = 1}^{k-1} (c_{i+1} - c_i)\tau_i \leq \tau_0 \cdot \max_{i=1,\dots,k} \{c_i\}. $ \end{lemma}
\begin{proof}[Theorem~\ref{thm:hetero}, $\beta$-deviation\xspace ratio] Let $x = f^\beta$ be a $\beta$-deviated\xspace Nash flow with path deviations $(\delta_P)_{P \in \mathcal{S}} \in \Delta(\beta)$ and let $z = f^0$ be an original Nash flow. Let $X = \{a \in A : x_a > z_a\}$ and $Z = \{a \in A : z_a \geq x_a \text{ and } z_a > 0\}$ (arcs with $x_a = z_a = 0$ may be removed without loss of generality).
In order to analyze the ratio $C(x)/C(z)$ we first argue that we can assume without loss of generality that the latency function $l_a(y)$ is constant for values $y \geq x_a$ for all arcs $a \in Z$. To see this, note that we can replace the function $l_a(\cdot)$ with the function $\hat{l}_a$ defined by $\hat{l}_a(y) = l_a(x_a)$ for all $y \geq x_a$ and $\hat{l}_a(y) = l_a(y)$ for $y \leq x_a$. In particular, this implies that the flow $x$ is still a $\beta$-deviated\xspace Nash flow for the same path deviations as before. This holds since for any path $P$ the latency $l_P(x)$ remains unchanged if we replace the function $l_a$ by $\hat{l}_a$.
By definition of arcs in $Z$, we have $x_a \leq z_a$ and therefore $\hat{l}_a(z_a) = l_a(x_a) \leq l_a(z_a)$. Let $z'$ be an original Nash flow for the instance with $l_a$ replaced by $\hat{l}_a$. Then we have $C(z') \leq C(z)$ using the fact that series-parallel graphs are immune to the Braess paradox, see Milchtaich \cite[Lemma 4]{Milchtaich2006}. Note that, in particular, we find $C(x)/C(z) \leq C(x)/C(z')$. By repeating this argument, we may without loss of generality assume that all latency functions $l_a$ are constant between $x_a$ and $z_a$ for $a \in Z$. Afterwards, we can even replace the function $\hat{l}_a$ by a function that has the constant value of $l_a(x_a)$ everywhere.
In the remainder of the proof, we will denote $P_j$ as a flow-carrying arc for sensitivity class $j \in [h]$ that maximizes the path latency amongst all flow-carrying path for sensitivity class $j \in [h]$, i.e., $ P_j = \text{argmax}_{P \in \mathcal{P} : x_{P,j} > 0} \{ l_P(x) \}. $ Moreover, there also exists a path $P_0$ with the property that $z_a \geq x_a$ and $z_a > 0$ for all arcs $a \in P_0$ (see, e.g., Lemma 2 \cite{Milchtaich2006}).
For fixed $a < b \in \{1,\dots,h\}$, the Nash conditions imply that (these steps are of a similar nature as Lemma 1 \cite{Fleischer2005}) \begin{align*} l_{P_a}(x) + \gamma_{a}\cdot \delta_{P_a}(x) &\leq l_{P_b}(x) + \gamma_{a} \cdot \delta_{P_b}(x) \\ l_{P_b}(x) + \gamma_{b}\cdot \delta_{P_b}(x) &\leq l_{P_a}(x) + \gamma_{b} \cdot \delta_{P_a}(x). \end{align*} Adding up these inequalities implies that $(\gamma_b - \gamma_a) \delta_{P_b}(x) \leq (\gamma_b - \gamma_a) \delta_{P_a}(x)$, which in turn yields that $\delta_{P_b}(x) \leq \delta_{P_a}(x)$ (using that $\gamma_a < \gamma_b$ if $a < b$). Furthermore, we also have \begin{equation}\label{eq:nash1} l_{P_1}(x) + \gamma_1 \delta_{P_1}(x) \leq l_{P_0}(x) + \gamma_1 \delta_{P_0}(x), \end{equation} and $l_{P_0}(x) = l_{P_0}(z) \leq l_{P_1}(z) \leq l_{P_1}(x)$, which can be seen as follows. The equality follows from the fact that $l_a$ is constant for all $a \in Z$ and, by choice, $P_0$ only consists of arcs in $Z$. The first inequality follows from the Nash conditions of the original Nash flow $z$, since there exists a flow-decomposition in which the path $P_0$ is used (since the flow on all arcs of $P_0$ is strictly positive in $z$). The second inequality follows from the fact that $$ \sum_{e \in P_1} l_e(z_e) = \sum_{e \in P_1 \cap X} l_e(z_e) + \sum_{e \in P_1 \cap Z} l_e(z_e) \leq \sum_{e \in P_1 \cap X} l_e(x_e) + \sum_{e \in P_1 \cap Z} l_e(x_e) $$ using that $z_e \leq x_e$ for $e \in X$ and the fact that latency functions for $e \in Z$ are constant. In particular, we find that $l_{P_0}(x) \leq l_{P_1}(x)$. Adding this inequality to (\ref{eq:nash1}), we obtain $\gamma_1\delta_{P_1}(x) \leq \gamma _1 \delta_{P_0}(x)$ and therefore $\delta_{P_1}(x) \leq \delta_{P_0}(x)$. Thus $ \delta_{P_h}(x) \leq \delta_{P_{h-1}}(x) \leq \dots \leq \delta_{P_1}(x) \leq \delta_{P_0}(x). $ Moreover, by using induction (see appendix) it can be shown that \begin{equation}\label{eq:induction} l_{P_j}(x) \leq l_{P_0}(x) + \gamma_1 \delta_{P_0}(x) + \bigg[\sum_{g = 1}^{j-1} (\gamma_{g+1} - \gamma_g)\delta_{P_g}(x)\bigg] - \gamma_j \delta_{P_j}(x). \end{equation}
Using (\ref{eq:induction}), we then have \begin{eqnarray} C(x) &\leq & \sum_{j=1}^h r_j l_{P_j}(x) \ \ \ \ \text{(by choice of the paths $P_j$)} \nonumber \\ &\leq& \sum_{j=1}^h r_j \left(l_{P_0}(x) + \gamma_1 \delta_{P_0}(x) + \bigg[\sum_{g = 1}^{j-1} (\gamma_{g+1} - \gamma_g)\delta_{P_g}(x)\bigg] - \gamma_j \delta_{P_j}(x) \right) \nonumber \\ &=& l_{P_0}(x) + \gamma_1 \delta_{P_0}(x) + \sum_{j=1}^{h} (r_{j+1}+\dots+ r_h)(\gamma_{j+1} - \gamma_j)\delta_{P_j}(x) - r_j\gamma_j \delta_{P_j}(x) \nonumber \\ &\leq & l_{P_0}(x) + \gamma_1 \delta_{P_0}(x) \nonumber \\ & & +\sum_{j=1}^{h-1} \bigg[(r_{j+1}+\dots+ r_h)\gamma_{j+1} - (r_j+r_{j+1}+\dots+ r_h)\gamma_j\bigg]\delta_{P_j}(x) \nonumber \end{eqnarray} In the last inequality, we leave out the last negative term $-r_h\gamma_h\delta_{P_h}(x)$. Note that $\gamma_1 = (r_1 + \dots + r_h)\gamma_1$ since we have normalized the demand to $1$.
We can then apply Lemma \ref{lem:max_seq} with $\tau_i = \delta_{P_i}(x)$ for $i = 0,\dots, h-1$ and $ c_i = \gamma_i \cdot \sum_{p = i}^h r_p $ for $i = 1,\dots, k$. Continuing the estimate, we get $$ C(x) \leq l_{P_0}(x) + \max_{j \in [h]} \bigg\{ \gamma_j \cdot \sum_{p = j}^h r_p \bigg\} \cdot \delta_{P_0}(x) \leq \bigg[1 + \beta \cdot \max_{j \in [h]} \bigg\{ \gamma_j \bigg( \sum_{p = j}^h r_p \bigg) \bigg\}\bigg] C(z) $$ where for the second inequality we use that $\delta_{P_0}(x) \leq \beta l_{P_0}(x)$, which holds by definition, and $l_{P_0}(x) = l_{P_0}(z) = C(z)$, which holds because $z$ is an original Nash flow and all arcs in $P_0$ have strictly positive flow in $z$ (and because of the fact that that all arcs in $P_0$ have a constant latency functions).
To prove tightness, fix $j \in [h]$ and consider the following instance on two arcs. We take $(l_1(y),\delta_1(y)) = (1,\beta)$ and $(l_2(y),\delta_2(y))$ with $\delta_2(y) = 0$ and $l_2(y)$ a strictly increasing function satisfying $l_2(0) = 1 + \epsilon$ and $l_2(r_j+r_{j+1}+\dots+r_h) = 1 + \gamma_j \beta$, where $\epsilon < \gamma_j \beta$. The (unique) original Nash flow is given by $z = (z_1,z_2) = (1,0)$ with $C(z) = 1$. The (unique) $\beta$-deviated\xspace Nash flow $x$ is given by $x = (x_1,x_2) = (r_1 + r_2 + \dots + r_{j-1}, r_j+r_{j+1}+\dots+r_h)$ with $C(x) = 1 + \beta \cdot \gamma_j (r_j + \dots + r_h)$. Since this construction holds for all $j \in [h]$, we find the desired lower bound.\qed
\end{proof}
\paragraph{Continuous sensitivity distributions.}
We obtain a similar result for more general (not necessarily discrete) sensitivity distributions. That is, we are given a Lebesgue integrable \emph{sensitivity density function} $\psi : \mathbb{R}_{\geq 0} \rightarrow \mathbb{R}_{\geq 0}$ over the total demand. Since we can normalize the demand to $1$, we have the condition that $\int_0^\infty \psi(y) dy = 1$. We then find the following natural generalizations of our upper bounds: \begin{enumerate} \item $\text{$\epsilon$-SR}(\mathcal{I}) \le 1 + \beta \int_{0}^\infty y\cdot \psi(y)dy$, and \item $\text{$\beta$-DR}(\mathcal{I}) \le 1 + \beta \cdot \sup_{t \in \mathbb{R}_{\geq 0}} \big\{ t \cdot \int_t^\infty \psi(y) dy \big\}$. \end{enumerate} These bounds are both asymptotically tight for all distributions. Details are given in Corollary \ref{cor:het} in the appendix.
\section{Homogeneous population}\label{sec:approx}
The reverse of Proposition~\ref{prop:inclusion} also holds for homogeneous players in single-commodity instances. As a consequence, the set of $\beta$-deviated\xspace Nash flows and the set of $\epsilon$-approximate Nash flows with $\epsilon = \beta \gamma$ coincide in this case.
Recall that for homogeneous players we have $\gamma_{ij} = \gamma$ for all $i \in [k]$, $j \in [h_i]$ and some $\gamma \ge 0$.
\begin{proposition}\label{prop:equiv} Let $\mathcal{I}$ be a single-commodity congestion game with homogeneous players. $f$ is an $\epsilon$-approximate Nash flow for $\mathcal{I}$ if and only if $f$ is a $\beta$-deviated\xspace Nash flow for $\mathcal{I}$ with $\epsilon = \beta\gamma$. \end{proposition}
\noindent \emph{Upper bound on the stability ratio.} Our main result in this section is an upper bound on the $\epsilon$-stability ratio. Given the above equivalence, this bound also applies to the $\beta$-deviation\xspace ratio with $\epsilon = \beta\gamma$.
The following concept of alternating paths is crucial. For single-commodity instances an alternating path always exists (see, e.g., \cite{Nikolova2015}).
\begin{definition}[Alternating path \cite{Lin2011,Nikolova2015}] Let $\mathcal{I}$ be a single-commodity network congestion game and let $x$ and $z$ be feasible flows. We partition the edges $E = X \cup Z$ such that $Z = \{a \in E : z_a \geq x_a \text{ and } z_a > 0\}$ and $X = \{a \in E : z_a < x_a \text{ or } z_a = x_a = 0\}$. We say that $\pi$
is an alternating $s,t$-path if the arcs in $\pi \cap Z$ are oriented in the direction of $t$, and the arcs in $\pi \cap X$ are oriented in the direction of $s$. We call the number of backward arcs on $\pi$ the \emph{backward length} of $\pi$ and refer to it by $q(\pi) = |\pi \cap X|$.
\label{def:alt_path} \end{definition}
\begin{theorem}\label{thm:approx_single} Let $\mathcal{I}$ be a single-commodity network congestion game. Let $\epsilon \ge 0$ be fixed and consider an arbitrary alternating path $\pi$ with backward length $q = q(\pi)$. If $\epsilon < 1/q$, then the $\epsilon$-stability ratio is bounded by $$ \text{$\epsilon$-SR}(\mathcal{I}) \le \frac{1 + \epsilon}{1 - \epsilon \cdot q} \leq \frac{1 + \epsilon}{1 - \epsilon \cdot n}. $$
\end{theorem}
Note that the restriction on $\epsilon$ stated in the theorem always holds if $\epsilon < 1/n$. In particular, for $\epsilon \ll 1/n$ we roughly get $\text{$\epsilon$-SR}(\mathcal{I}) \leq 1 + \epsilon n$. The proof of Theorem~\ref{thm:approx_single} is inspired by a technique of Nikolova and Stier-Moses \cite{Nikolova2015}, but technically more involved.
\begin{proof}
Let $x = f^\epsilon$ be an $\epsilon$-approximate Nash flow and let $z = f^0$ an original Nash flow. Let $\pi = Z_1X_1Z_2X_2\dots Z_{\eta-1}X_{\eta-1}Z_{\eta}$ be an alternating path for $x$ and $z$, where $Z_i$ and $X_i$ are maximal sections consisting of consecutive arcs, respectively, in $Z$ and $X$ (i.e., $Z_i \subseteq Z$ and $X_i \subseteq X$ for all $i$). Furthermore, we let $q_i = |X_i|$ and write $X_i = (X_{iq_i},\dots,X_{i2},X_{i1})$, where $X_{ij}$ are the arcs in the section $X_i$. By definition, for every arc $X_{ij}$ there exists a path $C_{ij}X_{ij}D_{ij}$ that is flow-carrying for $x$.\footnote{Note that for a Nash flow one can assume that there is a flow-carrying path traversing all arcs $X_{iq_i},\dots,X_{i1}$; but this cannot be done for an approximate Nash flow.}
For convenience, we define $C_{01} = D_{\eta,0} = \emptyset$. Furthermore, we denote $P^{\max}$ as a path maximizing $l_P(x)$ over all paths $P \in \mathcal{S}$. For convenience, we will abuse notation, and write $Q = Q(x) = \sum_{a \in Q} l_{a}(x)$ for $Q \subseteq E$.
Note that for all $i,j$: \begin{equation}\label{eq:pmax_upperbound} C_{ij}(x) + X_{ij}(x) + D_{ij}(x) \leq P^{\max}(x). \end{equation}
\begin{figure}
\caption{Sketch of the situation in the proof of Theorem \ref{thm:approx_single} with $q_1 = 1$ and $q_2 = 2$.}
\label{fig:braess}
\end{figure} Fix some $i \in \{1,\dots,\eta-1\}$. Then we have $C_{i1} + X_{i1} + D_{i1} \leq (1+\epsilon)(C_{i-1,q_{i-1}} + Z_i + D_{i1})$ by definition of an $\epsilon$-approximate Nash flow. This implies that (leaving out $D_{i1}$ on both sides) \begin{eqnarray} C_{i1} + X_{i1} &\leq &(1+\epsilon)Z_i + C_{i-1,q_{i-1}} + \epsilon(C_{i-1,q_{i-1}} + D_{i1}). \nonumber
\end{eqnarray}
Furthermore, for all $j \in \{2,\dots,q_i\}$, we have $C_{ij} + X_{ij} + D_{ij} \leq (1+\epsilon)(C_{i,j-1} + D_{ij})$ which implies (again leaving out $D_{ij}$ on both sides) $$ C_{ij} + X_{ij} \leq C_{i,j-1} + \epsilon(C_{i,j-1} + D_{ij}). $$
Adding up these inequalities for $j \in\{1,\dots,q_i\}$ and subtracting $\sum_{j=1}^{q_i-1} C_{ij}$ from both sides, we obtain for all $i \in \{1,\dots,\eta-1\}$
\begin{equation}\label{eq:alt_general_i} C_{i,q_i} + \sum_{j = 1}^{q_i} X_{ij} \leq C_{i-1,q_{i-1}} + (1+\epsilon)Z_i + \epsilon \bigg( \sum_{j = 1}^{q_i} D_{ij} + C_{i-1,q_{i-1}} + \sum_{j=1}^{q_i-1}C_{ij}\bigg). \end{equation}
Moreover, we also have \begin{equation}\label{eq:alt_pmax} P^{\max} \leq (1+\epsilon)(C_{\eta-1,\eta-1} + Z_{\eta}) = C_{\eta-1,\eta-1} + (1+\epsilon)Z_{\eta} + \epsilon C_{\eta-1,\eta-1}. \end{equation} Adding up the inequalities in (\ref{eq:alt_general_i}) for all $i \in \{1,\dots,\eta-1\}$, and the inequality in (\ref{eq:alt_pmax}), we obtain $$ P^{\max} + \sum_{i=1}^{\eta-1} C_{i,q_i} + \sum_{i = 1}^{\eta-1} \sum_{j=1}^{q_i} X_{ij} \leq \sum_{i=1}^{\eta-1} C_{i,q_i} + (1+\epsilon) \sum_{i = 1}^{\eta} Z_i + \epsilon \bigg(\sum_{i = 1}^{\eta-1} \sum_{j=1}^{q_i} C_{ij} + D_{ij} \bigg) $$ which simplifies to \begin{equation}\label{eq:cum} P^{\max} + \sum_{i = 1}^{\eta-1} \sum_{j=1}^{q_i} X_{ij} \leq (1+\epsilon) \sum_{i = 1}^{\eta} Z_i + \epsilon \bigg(\sum_{i = 1}^{\eta-1} \sum_{j=1}^{q_i} C_{ij} + D_{ij} \bigg). \end{equation} Using (\ref{eq:pmax_upperbound}), we obtain $$ \sum_{i = 1}^{\eta-1} \sum_{j=1}^{q_i} C_{ij} + D_{ij} \leq \sum_{i = 1}^{\eta-1} \sum_{j=1}^{q_i} P^{\max} - X_{ij} = \bigg(\sum_{i=1}^{\eta-1} q_i\bigg)P^{\max} - \sum_{i = 1}^{\eta-1} \sum_{j=1}^{q_i} X_{ij}. $$ Combining this with (\ref{eq:cum}), and rearranging some terms, we get \begin{align*}\label{eq:cum_final} (1 - \epsilon \cdot q)P^{\max} &\leq (1+\epsilon) \bigg[ \sum_{i = 1}^{\eta} Z_i - \sum_{i = 1}^{\eta-1} \sum_{j=1}^{q_i} X_{ij}\bigg] = (1+\epsilon) \bigg[ \sum_{e \in Z \cap \pi} l_e(x_e) - \sum_{e \in X \cap \pi} l_e(x_e)\bigg] \end{align*} where $q = q(\pi) = \sum_{i = 1}^{\eta-1} q_i$ is the backward length of $\pi$.
Similarly (see also \cite[Lemma 4.5]{Nikolova2015}), it can be shown that \begin{equation}\label{eq:z} l_Q(z) \geq \sum_{e \in Z \cap \pi} l_e(z_e) - \sum_{e \in X \cap \pi} l_e(z_e) \end{equation} for any path $Q$ with $z_Q > 0$ (these all have the same latency, since $z$ is an original Nash flow). Using a similar argument as in \cite[Theorem 4.6]{Nikolova2015}, we obtain \begin{eqnarray} (1 - \epsilon \cdot q)l_{P^{\max}}(x) &\leq & (1+\epsilon) \bigg[ \sum_{e \in Z \cap \pi} l_e(x_e) - \sum_{e \in X \cap \pi} l_e(x_e)\bigg] \nonumber \\ & \leq & (1+\epsilon) \bigg[ \sum_{e \in Z \cap \pi} l_e(z_e) - \sum_{e \in X \cap \pi} l_e(z_e)\bigg]
\nonumber
\le (1+\epsilon)l_Q(z).
\end{eqnarray} By multiplying both sides with the demand $r$, we obtain $ (1 - \epsilon \cdot q) C(x) \leq (1 - \epsilon \cdot q) r \cdot l_{P^{\max}}(x) \leq (1+\epsilon)r \cdot l_Q(z) = (1+\epsilon) C(z) $ for $\epsilon < 1/q$, which proves the claim.
\qed \end{proof}
\paragraph{Tight bound on the stability ratio.}
In this section, we consider instances for which all backward sections of the alternating path $\pi$ consist of a single arc., i.e., $q_i = 1$ for all $i = 1,\dots,\eta-1$. We then have $q = \sum_{i=1}^{\eta-1} q_i \leq \lfloor n/2 \rfloor - 1$ since every arc in $X$ must be followed directly by an arc in $Z$ (and we can assume w.l.o.g. that the first and last arc are contained in $Z$). By Theorem \ref{thm:approx_single}, we obtain
$ \text{$\epsilon$-SR}(\mathcal{I}) \leq (1 + \epsilon)/(1 - \epsilon \cdot (\lfloor n/2 \rfloor - 1)) $ for all $\epsilon < 1/(\lfloor n/2 \rfloor - 1)$. We show that this bound is tight. Further, we show that there exist instances for which $\text{$\epsilon$-SR}(\mathcal{I})$ is unbounded for $\epsilon \geq 1/(\lfloor n/2 \rfloor - 1)$. This completely settles the case of $q_i = 1$ for all $i$.
Our construction is based on the \emph{generalized Braess graph} \cite{Roughgarden2006}.
By construction, alternating paths for these graphs satisfy $q_i = 1$ for all $i$ (see Figure \ref{fig:braess_5} in the appendix for an example and a formal definition of these graphs).
\begin{theorem}\label{thm:braess_tight} Let $n = 2m$ be fixed and let $\mathcal{B}^m$ be the set of all instances on the generalized Braess graph with $n$ nodes. Then $$ \sup_{\mathcal{I} \in \mathcal{B}^m} \text{$\epsilon$-SR}(\mathcal{I})
= \begin{cases} \frac{1 + \epsilon}{1 - \epsilon \cdot (\lfloor n/2 \rfloor - 1)} & \text{if $\epsilon < \frac{1}{\lfloor n/2 \rfloor - 1}$,} \\ \infty & \text{otherwise.} \end{cases} $$ \end{theorem}
\paragraph{Non-symmetric matroid congestion games.}\label{sec:matroid}
In the previous sections, we considered (symmetric) network congestion games only. It is interesting to consider other combinatorial strategy sets as well. In this section we make a first step in this direction by focusing on the bases of matroids as strategies.
A matroid congestion game is given by $\mathcal{J} = (E,(l_e)_{e \in E},(\mathcal{S}_i)_{i \in [k]},(r_i)_{i \in [k]})$, and matroids $\mathcal{M}_i = (E,\mathcal{I}_i)$ over the ground set $E$ for every $i \in [k]$.\footnote{A matroid over $E$ is given by a collection $\mathcal{I} \subseteq 2^E$ of subsets of $E$ (called \emph{independent sets}). The pair $\mathcal{M} = (E,\mathcal{I})$ is a \emph{matroid} if the following three properties hold: i) $\emptyset \in \mathcal{I}$; ii) If $A \in \mathcal{I}$ and $B \subseteq A$, then $B \in \mathcal{I}$. iii) If $A,B \in \mathcal{I}$ and $|A| > |B|$, then there exists an $a \in A \setminus B$ such that $B + a \in \mathcal{I}$.} The strategy set $\mathcal{S}_i$ consists of the \emph{bases} of the matroid $\mathcal{M}_i$, which are the independent sets of maximum size, e.g., spanning trees in an undirected graph. We refer the reader to Schrijver \cite{schrijver2003} for an extensive overview of matroid theory.
As for network congestion games, it can be shown that in general the $\epsilon$-stability ratio can be unbounded (see Theorem \ref{thm:matroid_unbounded} in the appendix); this also holds for general path deviations because the proof of Proposition~\ref{prop:equiv} in the appendix holds for arbitrary strategy sets. However, if we consider path deviations induced by the sum of edge deviations (as in \cite{Nikolova2015,Kleer2016}), then we can obtain a more positive result for general matroids.
Recall that for every resource $e \in E$ we have a deviation function $\delta_e : \mathbb{R}_{\geq 0} \rightarrow \mathbb{R}_{\geq 0}$ satisfying $0 \leq \delta_e(x) \leq \beta l_e(x)$ for all $x \geq 0$. The deviation of a basis $B$
is then given by $\delta_B(f) = \sum_{e \in B} \delta_e(f_e)$.
\begin{theorem}\label{thm:matroid} Let $\mathcal{J} = (E,(l_e)_{e \in E},(\mathcal{S}_i)_{i \in [k]},(r_i)_{i \in [k]})$ be a matroid congestion game with homogeneous players. Let $\beta \ge 0$ be fixed and consider $\beta$-bounded basis deviations as defined above. Then the $\beta$-deviation\xspace ratio is upper bounded by $\text{$\beta$-DR}(\mathcal{J}) \le 1 + \beta$. Further, this bound is tight already for $1$-uniform matroid congestion games. \end{theorem}
\noindent \textbf{Acknowledgements.} We thank the anonymous referees for their very useful comments, and one reviewer for pointing us to Lemma 1 \cite{Fleischer2005}.
\appendix
\section{Omitted material of Section \ref{sec:het}}
\begin{rtheorem}{Proposition}{\ref{prop:inclusion}} Let $\mathcal{I}$ be a congestion game with heterogeneous players. If $f$ is a $\beta$-deviated\xspace Nash flow for $\mathcal{I}$, then $f$ is an $\epsilon$-approximate Nash flow for $\mathcal{I}$ with $\epsilon_{ij} = \beta \gamma_{ij}$ for all $i \in [k]$ and $j \in [h_i]$ (for the same demand distribution $r$). \end{rtheorem} \begin{proof} Let $f$ be a $\beta$-deviated Nash flow for some set of path deviations $\delta \in \Delta(0,\beta)$. For $P \in \mathcal{S}$, we have $$ l_P(f) \leq l_P + \gamma_{ij}\delta_P(f) \leq l_{P'}(f) + \gamma_{ij} \delta_{P'}(f)\leq (1+ \gamma_{ij} \beta) l_{P'}(f) $$ where we use the non-negativity of $\delta_P(f)$ in the first inequality, the Nash condition for $f$ in the second inequality, and the fact that $\delta_{P'}(f) \leq \beta l_{P'}(f)$ in the third inequality. By definition, it now holds that $f$ is also a $\epsilon$-approximate equilibrium (for $\epsilon = \beta \gamma$). \qed \end{proof}
\begin{rtheorem}{Theorem}{\ref{thm:hetero}} Let $\mathcal{I}$ be a network congestion game on a series-parallel graph with heterogeneous players, demand distribution $r = (r_i)_{i \in [h]}$ normalized to $1$, i.e., $\sum_{j \in [h]} r_i = 1$, and sensitivity distribution $\gamma = (\gamma_i)_{i \in [h]}$, with $\gamma_1 < \gamma_2 < \dots < \gamma_h$.
Let $\beta \ge 0$ be fixed and define $\epsilon = (\beta \gamma_i)_{i \in [h]}$. Then the $\epsilon$-stability ratio and the $\beta$-deviation\xspace ratio are bounded by: \begin{align} \text{$\epsilon$-SR}(\mathcal{I}) \le 1 + \beta \sum_{j = 1}^h r_j\gamma_j \quad\text{and}\quad \text{$\beta$-DR}(\mathcal{I}) \le 1 + \beta \cdot \max_{j \in [h]} \bigg\{ \gamma_j \bigg( \sum_{p = j}^h r_p \bigg) \bigg\}. \label{eq:comparison_epsilon}
\end{align} Further, both bounds are tight for all distributions $r$ and $\gamma$. \end{rtheorem}
\subsection{Proof of statement for $\epsilon$-$SR(\mathcal{I})$.} \begin{proof}[Theorem \ref{thm:hetero}, $\epsilon$-stability ratio]
The statement for $\epsilon$-approximate equilibria can be proven almost similar as in Kleer and Sch\"afer \cite{Kleer2016} (the bound there is used for path deviations, but the proof extends directly to approximate equilibria). For completeness, we give the argument here.
For $j \in [k]$, let $\bar{P}_j$ be a path maximizing $l_P(x)$ over all flow-carrying paths $P \in \mathcal{P}$ of type $j$. Moreover, there exists a path $\pi$ such that $x_a \leq z_a$ and $z_a > 0$ for all $a \in \pi$ (see, e.g., Milchtaich \cite{Milchtaich2006}). We then have (this is also reminiscent of an argument by Lianeas et al. \cite{Lianeas2016}): $$ l_{\bar{P}_j}(x) \leq (1+\beta \gamma_j) l_{\pi}(x) = (1 + \beta \gamma_j)\sum_{a \in \pi} l_a(x_a). $$ Note that, by definition of the alternating path $\pi$, we have $x_a \leq z_a$ for all $a \in \pi$. Continuing with the estimate, we find $l_{\bar{P}_j}(x) \leq (1 + \beta \gamma_j)\sum_{a \in \pi} l_a(z_a)$ and thus
$$ C(x) \leq \sum_{j \in [h]} r_j l_{\bar{P}_j}(x) \leq \sum_{j \in [h]} r_j (1 + \beta \gamma_j)\sum_{a \in \pi} l_a(z_a) = C(z) \bigg(\sum_{j \in [h]} r_j (1 + \beta \gamma_j)\bigg) $$ Since $\sum_{j \in [h]} r_j = 1$, we get the desired result. Note that we use $C(z) = \sum_{a \in \pi} l_a(z_a)$, which is true because there exists a flow-decomposition of $z$ in which $\pi$ is flow-carrying (here we use $z_a > 0$ for all $a \in \pi$).
Tightness follows by considering an instance with arc set $\{0,1,\dots,h\}$ where the zeroth arc has latency $l_0(y) = 1$ and the arcs $j \in \{1,\dots,h\}$ have latency $l_j(y) = 1 + \beta \gamma_j$. An original Nash flow is given by $f^0 = (z_0,z_1,\dots,z_h) = (1,0,\dots,0)$, and an $\epsilon$-approximate Nash flow is given by $f^\epsilon = (x_0,x_1,\dots,x_h) = (0, r_1, r_2, \dots, r_h)$.\qed \end{proof}
\subsection{Missing arguments for proof of $\beta$-deviation ratio.}
\begin{rtheorem}{Lemma}{\ref{lem:max_seq}} Let $0 \leq \tau_{k-1} \leq \dots \leq \tau_1 \leq \tau_0$ and $c_i \geq 0$ for $i = 1,\dots,k$ be given. Then $$ c_1 \tau_0 + \sum_{i = 1}^{k-1} (c_{i+1} - c_i)\tau_i \leq \tau_0 \cdot \max_{i=1,\dots,k} \{c_i\}. $$ \end{rtheorem} \begin{proof} The statement is clearly true for $k = 1$. Now suppose the statement is true for some $k \in \mathbb{N}$. We will prove the statement for $k + 1$. \\ \textbf{Case 1: $c_{k+1} - c_{k} \leq 0$.} Then we have \begin{eqnarray} c_1 \tau_0 + \sum_{i = 1}^{k} (c_{i+1} - c_i)\tau_i &\leq & c_1 \tau_0 + \sum_{i = 1}^{k-1} (c_{i+1} - c_i)\tau_i \text{ \ \ \ \ (using $\tau_k \geq 0$) } \nonumber \\ &\leq & \tau_0 \cdot \max_{i=1,\dots,k} \{c_i\} \text{ \ \ \ \ (using induction hypothesis) } \nonumber \\ &\leq & \tau_0 \cdot \max_{i=1,\dots,k+1} \{c_i\}\text{ \ \ \ \ (using non-negativity of $\tau_0$) } \nonumber \end{eqnarray}\\ \textbf{Case 2: $c_{k+1} - c_{k} > 0$.} \begin{eqnarray} c_1 \tau_0 + \sum_{i = 1}^{k} (c_{i+1} - c_i)\tau_i &= & c_1 \tau_0 + \bigg[\sum_{i = 1}^{k-1} (c_{i+1} - c_i)\tau_i \bigg] + (c_{k+1}-c_k)\tau_{k} \nonumber \\ &\leq & c_1 \tau_0 + \bigg[\sum_{i = 1}^{k-1} (c_{i+1} - c_i)\tau_i \bigg] + (c_{k+1}-c_k)\tau_{k-1} \text{ \ \ ($\tau_k \leq \tau_{k-1}$) } \nonumber \\ &=& c_1 \tau_0 + \bigg[\sum_{i = 1}^{k-2} (c_{i+1} - c_i)\tau_i \bigg] + (c_{k+1}-c_{k-1})\tau_{k-1} \nonumber \\ &\leq & \tau_0 \cdot \max_{i=1,\dots,k-2,k-1,k+1} \{c_i\} \text{ \ \ \ \ (using induction hypothesis) } \nonumber \\ &\leq & \tau_0 \cdot \max_{i=1,\dots,k+1} \{c_i\}\text{ \ \ \ \ (using non-negativity of $\tau_0$) } \nonumber \end{eqnarray} Note that we apply the induction hypothesis with the set $\{c_1,\dots,c_{k-1},c_{k+1}\}$ of size $k$. \qed \end{proof}
\subsubsection{Induction step to prove inequality (\ref{eq:induction}) in proof of Theorem \ref{thm:hetero}:}
The case $j = 1$ is precisely (\ref{eq:nash1}). Now suppose it holds for some $j$, then we have \begin{eqnarray} l_{P_{j+1}}(x) &\leq& l_{P_j}(x) + \gamma_{j+1}\delta_{P_j}(x) - \gamma_{j+1}\delta_{P_{j+1}}(x) \ \ \ \ \text{(Nash condition for path $P_{j+1}$)} \nonumber \\ &\leq& l_{P_0}(x) + \gamma_1 \delta_{P_0}(x) + \bigg[\sum_{g = 1}^{j-1} (\gamma_{g+1} - \gamma_g)\delta_{P_g}(x)\bigg] - \gamma_j \delta_{P_j}(x) \nonumber \\ & & + \ \gamma_{j+1}\delta_{P_j}(x) - \gamma_{j+1}\delta_{P_{j+1}}(x) \nonumber \ \ \ \text{(induction hypothesis)} \nonumber \\ & = &l_{P_0}(x) + \gamma_1 \delta_{P_0}(x) + \bigg[\sum_{g = 1}^{j} (\gamma_{g+1} - \gamma_g)\delta_{P_g}(x)\bigg] - \gamma_{j+1} \delta_{P_{j+1}}(x), \nonumber \end{eqnarray} which shows the result for $j + 1$.\qed
\subsection{Bounded deviation model for $(0,\beta)$-path deviations and general sensitivity density distributions (single-commodity case).} We use the section to explain the generalization of the bounds in Theorem \ref{thm:hetero} to more general sensitivity density distributions (see Corollary \ref{cor:het} further on). For sake of completeness, we adjust the definitions given in the description of the bounded deviation model (see Section \ref{sec:pre}).
We are given a single-commodity instance $\mathcal{I} = (E,(l_e)_{e \in E},\mathcal{S})$ and $\beta \geq 0$ fixed. We consider the set $$
\Delta(0,\beta) = \{ \delta = (\delta_P)_{P \in \mathcal{S}} \ \big| \ 0 \leq \delta_P(f) \leq \beta l_P(f) \text{ for all } f \in \mathcal{F}(\mathcal{S}) \} $$ of $(0,\beta)$-path deviation vectors $\delta= (\delta_P)_{P \in \mathcal{S}}$ where $\delta_P : \mathcal{F}(\mathcal{S}) \rightarrow \mathbb{R}_{\geq0}$ for all $P \in \mathcal{S}$. Moreover we have a Lesbesgue integrable (or just continuous) density function $\psi : \mathbb{R}_{\geq 0} \rightarrow \mathbb{R}_{\geq 0}$ over the the demand, which is normalized to $1$, i.e., $\int_0^\infty \phi(y)dy = 1$. For feasible flow $f$, we have an indicator function $$ \mathbb{I} : \mathcal{S} \times [0,\infty) \rightarrow \{0,1\} $$ which is $1$ if sensitivity $\gamma \in [0,\infty)$ is present on path $P$ in flow $f$, and $0$ otherwise.
Define the \emph{deviated latency} of a path $P \in \mathcal{S}_i$ for sensitivity class $j \in [h_i]$ as $q_{P}^j(f) = l_P(f) + \gamma_{ij} \delta_P(f)$. We say that a flow $f$ is a \emph{$\beta$-deviated\xspace Nash flow} if there exist some $\beta$-bounded path deviations $\delta \in \Delta(\beta)$ such that \begin{equation} \forall \gamma [0,\infty), \forall s \in \mathcal{S} \text{ with } \mathbb{I}(P,\gamma) = 1: \qquad q_{P}^j(f) \leq q_{P'}^j(f) \ \ \forall P' \in \mathcal{S}. \label{eq:nash} \end{equation}
We define the \emph{$\beta$-deviation\xspace ratio} $\text{$\beta$-DR}(\mathcal{I})$ as the maximum ratio $C(f^\beta)/C(f^0)$ of an $\beta$-deviated\xspace Nash flow $f^\beta$ and an original Nash flow $f^0$.
\noindent The result of Theorem \ref{thm:hetero} for general sensitivity demand functions (as mentioned at the end of Section \ref{sec:het}) can now be stated as follows.
\begin{corollary}\label{cor:het}
Let $\mathcal{I}$ be a network congestion game on a series-parallel graph with heterogeneous players, given by a Lebesgue integrable demand density function $\psi : \mathbb{R}_{\geq 0} \rightarrow \mathbb{R}_{\geq 0}$. Let $\beta \geq 0$ be fixed and let $\epsilon = (\beta\gamma_i)_{i \in [h]}$. Then the $\epsilon$-stability ratio and the $\beta$-deviation\xspace ratio are bounded by: \begin{align} \text{$\epsilon$-SR}(\mathcal{I}) \le 1 + \beta \int_{0}^\infty y\cdot \psi(y)dy \quad\text{and}\quad \text{$\beta$-DR}(\mathcal{I}) \le 1 + \beta \sup_{t \in \mathbb{R}_{\geq 0}} \bigg\{ t \cdot \int_t^\infty \psi(y) dy \bigg\}. \label{eq:comparison_epsilon_app}
\end{align} Further, both bounds are asymptotically tight for all distributions $r$ and $\gamma$.
\end{corollary} \begin{proof} We first show that we can reduce to a discrete instance as considered in Theorem \ref{thm:hetero}. For every path $P \in \mathcal{S}$ we can set the sensitivity of all the flow on that path to $$ \gamma_P^* = \inf\{ \gamma : \mathbb{I}(P,\gamma) = 1 \}. $$ In particular, this induces a demand distribution over a discrete (finite) valued sensitivity population (here we implicitly use the continuity of the latency functions). The upper bounds in Theorem \ref{thm:hetero} for the resulting sensitivity distributions will never be worse than the quantities in (\ref{eq:comparison_epsilon_app}). This concludes the proofs of the upper bounds (since the result now follows from the proof of Theorem \ref{thm:hetero}).
For tightness of the $\beta$-deviation ratio, fix $t \in [0,\infty)$, and let $T = \int_t^\infty \phi(y)dy$. Consider the following instance on two arcs. We take $(l_1(y),\delta_1(y)) = (1,\beta)$ and take $(l_2(y),\delta_2(y))$ with $\delta_2(y) = 0$ and $l_2(y)$ a strictly increasing function satisfying $l_2(0) = 1 + \epsilon'$ and $l_2(T) = 1 + t\cdot \beta$, where $\epsilon' < t \cdot \beta$. The (unique) original Nash flow is given by $z = (z_1,z_2) = (1,0)$ with $C(z) = 1$, and the (unique) deviated Nash flow $x$ is given by $x = (x_1,x_2) = (1-T,T)$ with $C(x) = 1 + \beta \cdot \gamma \cdot T$. Since this construction holds for all $t \in [0,\infty)$, the bound in (\ref{eq:comparison_epsilon_app}) is asymptotically tight. This holds since we can get arbitrarily close to the supremum if it is finite, and otherwise we can create a sequence of instances of which its ratio goes approaches infinity in case the supremum is not finite.
For tightness of the $\epsilon$-stability ratio, we create a discretized version of the construction in the proof of Theorem \ref{thm:hetero}. Fix some arbitrary $A> 0$. Choose $\alpha$ large enough so that $$ A = \int_\alpha^\infty \psi(y) dy $$ For the density under $\alpha$, we discretize the interval $[0,\alpha]$ into the intervals $[k\cdot \epsilon', (k+1)\epsilon']$ where $k = 0,\dots,q$ for some $q = \alpha/\epsilon'$ (without loss of generality $q$ can be assumed to be integral, and it can be chosen as large as desired). We then consider the discrete demand distribution $r = (r_1,\dots,r_q,A)$, where $$ r_i = \int_{k\epsilon'}^{(k+1)\epsilon'} \psi(y) dy. $$ We then create an instance with $q+1$ arcs, with latency functions $l_k(y) = (1 + k\epsilon')$ for $k = 1,\dots,q$ and $l_{q+1}(y) = 1 + A$. We now can use the same construction as in the proof of statement (\ref{eq:comparison_epsilon}), and by sending $A,\epsilon' \rightarrow 0$ we can get arbitrarily close to the value $1 + \beta \int_{0}^\infty y\cdot \psi(y)dy$. Note that the number of arcs in the instance goes to infinity as $A,\epsilon' \rightarrow 0$. \qed \end{proof}
\section{Omitted material of Section \ref{sec:approx}}
\begin{rtheorem}{Proposition}{\ref{prop:equiv}}
Let $\mathcal{I}$ be a single-commodity congestion game with homogeneous players. $f$ is an $\epsilon$-approximate Nash flow for $\mathcal{I}$ if and only if $f$ is a $\beta$-deviated\xspace Nash flow for $\mathcal{I}$ with $\epsilon = \beta\gamma$. \end{rtheorem} \begin{proof} The first part of the proof is a special case of the proof of Proposition \ref{prop:inclusion}. Let $f$ be a $\beta$-deviated\xspace Nash flow for some set of path deviations $\delta \in \Delta(0,\beta)$. For $P \in \mathcal{S}$, we have $$ l_P(f) \leq l_P + \gamma \delta_P(f) \leq l_{P'}(f) + \gamma \delta_{P'}(f)\leq (1+ \gamma \beta) l_{P'}(f) $$ where we use the non-negativity of $\delta_P(f)$ in the first inequality, the Nash condition for $f$ in the second inequality, and the fact that $\delta_{P'}(f) \leq \beta l_{P'}(f)$ in the third inequality. By definition, it now holds that $f$ is also a $\epsilon$-approximate equilibrium (for $\epsilon = \beta \gamma$).
Reversely, let $f$ be a $\epsilon$-approximate Nash flow (w.l.o.g., we can take $\gamma = 1$, so that $\beta = \epsilon$). We show that there exist $(0,\beta)$-path deviations $\delta_P$ such that $f$ is inducible with respect to these path deviations. Let $P_1,\dots,P_k$ be the set of flow-carrying paths under $f$, and assume without loss of generality that $l_{P_1}(f) \leq l_{P_2}(f) \leq \dots \leq l_{P_k}(f)$. We define $\delta_{P_i}(f) = l_{P_k}(f) - l_{P_i}(f)$ for $i = 1,\dots,k$. Using the Nash condition for the path $P_k$, we find $$ \delta_{P_i}(f) = l_{P_k}(f) - l_{P_i}(f) \leq (1+ \beta)l_{P_i}(f) - l_{P_i}(f) = \beta l_{P_i}(f) $$ which shows that these path deviations are feasible. Moreover, we take $\delta_{Q}(f) = \beta l_Q(f)$ for all the paths $Q \in \mathcal{P} \setminus \{P_1,\dots,P_k\}$ which are not flow-carrying under $f$. Now, let $i \in \{1,\dots,k\}$ be fixed. Then for any $j \in \{1,\dots,k\}$, we have $$ l_{P_i}(f) + \delta_{P_i}(f) = l_{P_k}(f) = l_{P_j}(f) + \delta_{P_j}(f) $$ and for any $Q \in \mathcal{P} \setminus \{P_1,\dots,P_k\}$, we have $$ l_{P_i}(f) + \delta_{P_i}(f) = l_{P_k}(f) \leq (1+\beta)l_Q(f) = l_Q(f) + \delta_Q(f), $$ using the Nash condition for the path $P_k$ and the definition of $\delta_Q(f)$. This shows that $f$ is indeed a $\beta$-deviated Nash flow. \qed \end{proof}
\subsubsection{Generalized Braess graph.} The \emph{$m$-th (generalized) Braess graph} $G^m = (V^m, A^m)$ is defined by $V^m = \{s,v_1,\dots,v_{m-1},w_1,\dots,w_{m-1},t\}$ and $A^m$ as the union of three sets: $E^m_1 = \{(s,v_j),(v_j,w_j),(w_j,t): 1 \leq j \leq m-1\}$, $E^m_2 = \{(v_j,w_{j-1}) : 2 \leq j \leq m\}$ and $E^m_3 = \{(v_1,t)\cup\{(s,w_{m-1}\}\}$ (see Figure \ref{fig:braess_5} for an example).
\begin{rtheorem}{Theorem}{\ref{thm:braess_tight}} Let $n = 2m$ be fixed and let $\mathcal{B}^m$ be the set of all instances on the generalized Braess graph with $n$ nodes. Then $$ \sup_{\mathcal{I} \in \mathcal{B}^m} \text{$\epsilon$-SR}(\mathcal{I})
= \begin{cases} \frac{1 + \epsilon}{1 - \epsilon \cdot (\lfloor n/2 \rfloor - 1)} & \text{if $\epsilon < \frac{1}{\lfloor n/2 \rfloor - 1}$,} \\ \infty & \text{otherwise.} \end{cases} $$
\end{rtheorem}
The proof of the upper bound for $\epsilon < 1/(\lfloor n/2 \rfloor - 1)$ follows from the discussion in the main text of Section \ref{sec:approx}. That is, we either have a path consisting solely of $Z$-edges (in which case we get a bound of $1 + \epsilon$ which follows from similar arguments as the proof of Theorem \ref{thm:hetero}) or we have an alternating path $q_i = 1$ for all sections of consecutive $X$-edges in the alternating path (in Figure \ref{fig:braess_5}, these are the edges with latency $T$). That is, every such section consists of one edge in $X$.
\begin{proof}[Lower bound for $\epsilon \geq 1/(\lfloor n/2 \rfloor - 1)$] Let $\tau \geq 0$ be a fixed constant. Let $y_m : \mathbb{R}_{\geq 0} \rightarrow \mathbb{R}_{\geq 0}$ be a non-decreasing, continuous function with $y_m(1/m) = 0$ and $y_m(1/(m-1)) = \tau$. We define $$ l^{m}_a(g) = \left\{ \begin{array}{ll} (m - j)\cdot y_m(g) & \text{ for } a \in \{(s,v_j) : 1 \leq j \leq m-1\}\\ j \cdot y_m(g) & \text{ for } a \in \{(w_j,t) : 1 \leq j \leq m-1\}\\ 1 & \text{ for } a \in \{(v_i,w_{i-1}): 2 \leq i \leq m-1\} \cup E_3\\ (1+\epsilon) + ((m-1)\epsilon - 1)\tau & \text{ for } a \in \{(v_i,w_i): 1 \leq i \leq m-1\} \end{array}\right. $$ Note that $l_a^m$ is non-negative for all $a \in A$. This is clear for the first three cases, and the last case follows from the assumption $\epsilon \geq 1/(m-1)$.
\begin{figure}
\caption{The fifth Braess graph with $l_a^5$ on the arcs as defined in the proof of Theorem \ref{thm:braess_tight}, with $T = (1+\epsilon) + ((m-1)\epsilon - 1)\tau$. The bold arcs indicate the alternating path $\pi_1$.}
\label{fig:braess_5}
\end{figure}
A Nash flow $z = f^0$ is given by routing $1/m$ units of flow over the paths $(s,w_{m-1},t),(s,v_1,t)$ and the paths in $\{(s,v_j,w_{j-1},t) : 2 \leq j \leq m-1\}$. Note that all these paths have latency one, and the path $(s,v_j,w_j,t)$, for some $1 \leq m \leq j$, has latency $(1+\epsilon) + ((m-1)\epsilon - 1)\tau \geq 1$ for all $\tau \geq 0$ and $ \epsilon \geq 1/(m-1)$. We conclude that $C(z) = 1$.
An $\epsilon$-approximate Nash flow $x = f^\epsilon$ is given by routing $1/(m-1)$ units of flow over the paths in $\{(s,v_j,w_j,t) : 1 \leq j \leq m - 1\}$. Each such path $P$ then has a latency of $$ l_P(x) = (1+\epsilon) + ((m-1)\epsilon - 1)\tau + (m-j+j)\tau = (1+\epsilon)(1 + (m-1)\tau). $$ Furthermore, we have $l_P(x) = (1+\epsilon)l_{P'}(x)$ for all $P' = (s,v_j,w_{j-1},t)$, where $2 \leq j \leq m-1$. The same argument holds for the paths $(s,w_{m-1},t)$ and $(s,v_1,t)$, so $x$ is indeed an $\epsilon$-approximate Nash flow. We have $C(x) = (1+\epsilon)(1 + (m-1)\tau)$. Sending $\tau \rightarrow \infty$ then gives the desired result. \qed \end{proof}
\begin{proof}[Lower bound for $\epsilon < 1/(\lfloor n/2 \rfloor - 1)$] Let $y_m : \mathbb{R}_{\geq 0} \rightarrow \mathbb{R}_{\geq 0}$ be a non-decreasing, continuous function with $y_m(1/m) = 0$ and $y_m(1/(m-1)) = \epsilon/(1 - \epsilon\cdot(m-1))$. We define $$ l^{m}_a(g) = \left\{ \begin{array}{ll} (m - j)\cdot y_m(g) & \text{ for } a \in \{(s,v_j) : 1 \leq j \leq m-1\}\\ j \cdot y_m(g) & \text{ for } a \in \{(w_j,t) : 1 \leq j \leq m-1\}\\ 1 & \text{ for } a \in \{(v_i,w_{i-1}): 2 \leq i \leq m-1\} \cup E_3\\ 1 & \text{ for } a \in \{(v_i,w_i): 1 \leq i \leq m-1\} \end{array}\right. $$ Note that $l_a^m$ is non-negative for all $a \in A$.
A Nash flow $z = f^0$ is given by routing $1/m$ units of flow over the paths $(s,w_{m-1},t),(s,v_1,t)$ and the paths in $\{(s,v_j,w_{j-1},t) : 2 \leq j \leq m-1\}$. Note that all these paths have latency one, and the path $(s,v_j,w_j,t)$, for $1 \leq m \leq j$, has latency one as well. We conclude that $C(z) = 1$.
An $\epsilon$-approximate Nash flow $x = f^\epsilon$ is given by routing $1/(m-1)$ units of flow over the paths in $\{(s,v_j,w_j,t) : 1 \leq j \leq m - 1\}$. Each such path $P$ then has a latency of $$ l_P(x) = 1 + (m - j + j) \cdot \frac{\epsilon}{1 - \epsilon\cdot (m-1)} = \frac{1+\epsilon}{1 - \epsilon\cdot (m-1)}. $$ Furthermore, for all $P' = (s,v_j,w_{j-1},t)$, where $2 \leq j \leq m-1$, we have $$ (1+\epsilon)l_{P'}(x) = (1+\epsilon)\left(1 + (m - 1) \cdot \frac{\epsilon}{1 - \epsilon\cdot (m-1)}\right) = l_P(x) $$ The same argument holds for the paths $(s,w_{m-1},t)$ and $(s,v_1,t)$, so $x$ is indeed an $\epsilon$-approximate Nash flow. The result now follows since $C(x)/C(z) = (1+\epsilon)/(1 - \epsilon\cdot (m-1))$ as desired. \qed \end{proof}
\subsection{Unboundedness of the $\epsilon$-stability ratio for matroid congestion games.} We first show that the ratio between a $\beta$-approximate Nash flow and an original Nash flow can, in general, be unbounded.
\begin{theorem}\label{thm:matroid_unbounded} Let $\epsilon \geq 0$ fixed and let $\mathcal{J} = (E,(l_e)_{e \in E},\mathcal{S})$ be a single-commodity matroid congestion game in which $\mathcal{S}$ contains all subsets of $E$ of precisely size $k$. Then $$ \sup_{\mathcal{J}} \text{$\epsilon$-SR}(\mathcal{J}) \ \geq \ \left\{ \begin{array}{ll} \frac{1 + \epsilon}{1 - \epsilon (k-1)} & \text{ \ \ \ if } \epsilon < 1/(k - 1) \\ \infty & \text{ \ \ \ if } \epsilon \geq 1/(k - 1), \end{array}\right. $$ For $k = 2$, the resulting lower bound $(1+\epsilon)/(1 - \epsilon)$ for $\epsilon < 1$ is tight (i.e., it is also an upper bound). This settles the case of $k=2$ completely. \end{theorem} \begin{proof}[Unboundedness] Fix some $k \in \mathbb{N}$ and consider an instance with demand $r = 1$ and $E = \{e_0,\dots,e_k\}$, and, as stated above, let $\mathcal{S}$ be the family of all subsets of size $k$. Let $l_0(y) = 1$ be the constant latency function of resource $e_0$ and for $e \in \{e_1,\dots,e_k\}$, let $l_e(y)$ be a non-negative, continuous, increasing function satisfying $$ l_e((k-1)/k) = 1 \ \ \text{ and } \ \ l_e(1) = M, $$ where $M$ is a constant to be determined later.
A classical Nash flow $z = f^0$ is given by assigning $1/k$ units of flow to the strategies $E \setminus \{e_j\}$ for $j = 1,\dots,k$, so that the load on $e_0$ is $z_0 = 1$, and the load on resources $e_j$ is $x_{j} = (k-1)/k$. Every strategy has latency $k$ (since $l_e((k-1)/k) = 1$ for all resources $j \in \{1,\dots,k\}$), and therefore also $C(z) = k$.
We construct an $\epsilon$-approximate Nash flow $x = f^\epsilon$ as follows.The flow $x$ is defined by assigning all flow to the strategy $s = \{e_1,\dots,e_k\}$. For suitable choices of $M$, we show that $x$ is then indeed an $\epsilon$-approximate Nash flow. The set of other strategies is given by $s^j = E \setminus \{e_j\}$ for $j \in \{1,\dots,k\}$. The Nash conditions for $x$ are then equivalent to \begin{equation}\label{eq:nash_matroid} l_s(x) = k\cdot M \leq (1 + \epsilon)[1 + (k-1)M] = l_{s_j}(x) \end{equation} for all $j \in \{1,\dots,k\}$. If $\epsilon \geq 1/(k-1)$, then $1 \leq \epsilon (k-1)$, and thus $$ k \cdot M \leq 1 + \epsilon + (k-1)M + M \leq 1 + \epsilon + (k-1)M + \epsilon (k-1)M = (1+\epsilon)(1 + (k-1)M) $$ for any non-negative $M$, i.e., the Nash conditions are always satisfied. Note that $C(x) = kM$ so that $C(x)/C(z) = M \rightarrow \infty$ as $M \rightarrow \infty$. Therefore the $\epsilon$-ratio is unbounded for $\epsilon \geq 1/(k-1)$.
If $\epsilon < 1/(k-1)$, then the Nash condition in (\ref{eq:nash_matroid}) is equivalent to $$ M \leq \frac{1 + \epsilon}{1 - \epsilon(k-1)} $$ which is strictly positive since $\epsilon < 1/(k-1)$. In particular, by choosing $M = (1 + \epsilon)/(1 - \epsilon(k-1))$, we get $C(x)/C(z) = (k\cdot M)/k = (1 + \epsilon)/(1 - \epsilon(k-1))$ for $\epsilon < 1/(k-1)$. \qed
\end{proof}
For the case $k = 2$, the construction in the previous proof yields a lower bound of $(1+\epsilon)/(1 - \epsilon)$ if $\epsilon< 1$, and $\infty$ otherwise. For $\epsilon < 1$, we show that the bound $(1+\epsilon)/(1-\epsilon)$ is also an upper bound.
\begin{proof}[Upper bound for $k = 2$] Let $z = f^0$ be a classical Nash flow and let $x = f^\epsilon$ be a worst-case $\epsilon$-approximate Nash flow (both with normalized demand $r = 1$). We partition $E = X \cup Z$, where $Z = \{a \in E : z_a \geq x_a \text{ and } z_a > 0\}$ and $X = \{a \in E : z_a < x_a \text{ or } z_a = x_a = 0\}$.
Let us first elaborate on the \emph{structure} of the flow $z$. One of the following cases holds. \begin{enumerate}[i)] \item There are two resources $e$ and $e'$ which are used by every player, that is, all flow is assigned to the strategy $\{e,e'\}$. By definition, both these resources will then be part of $Z$ (since $x_a \leq 1 = z_a$ for $a = e,e'$). \item There is precisely one resource $e_0$ used by every player. Again, this resource will then be part of $Z$ by definition. Moreover, we then also have $$ l_{e_0}(1) = l_{e_0}(z_e) \leq l_{e'}(z_{e'}) $$ for every other resource $e'$. Also, all other resources $e'$ with $z_{e'} > 0$ have equal latency (which is true because of base exchange arguments), which in turn is greater or equal than that of $e_0$. \item There are no resources which are used by every player. Then all flow-carrying resources in $z$ have equal latency (again because of base exchange arguments). \end{enumerate}
\noindent \textbf{Claim 1:} Without loss of generality, there is at most one resource in $Z$. \begin{proof} If there are at least two resources in $Z$, then there exists a pair of resources in $Z$, forming a strategy with a strictly positive amount of flow assigned to it in $z$ (by what was said above regarding the structure of $z$).\footnote{We emphasize that this is not necessarily true for all combinations of resources in $Z$.} Using the fact that matroids are immune against the Braess paradox, as shown by Fujishige et al. \cite{Fujishige2015}, we can then carry out similar arguments as in the proof of Theorem \ref{thm:hetero} for the $\beta$-deviation ratio, but then for the more simple homogeneous case.\footnote{Phrased differently, the proof of Theorem \ref{thm:hetero} for the $\beta$-deviation ratio essentially relies on the fact that for the flow $z$ we can without loss of generality assume that all latency functions of resources in $Z$ are constant (by using immunity against the Braess paradox), and that there exists a strategy, with positive amount of flow assigned to it in $z$, only using resources in $Z$.} This would result in a bound of $1 + \epsilon \leq (1+\epsilon)/(1 - \epsilon)$. \qed \end{proof}
Let $P_{\max}$ be a strategy maximizing $l_P(x)$ over all flow-carrying strategies $P$ in $x$.
\noindent \textbf{Claim 2:} Without loss of generality, $P_{\max}$ uses only resources in $X$ (and not the resource $e_0$ in $Z$).
\begin{proof} Suppose that every strategy with maximum latency is of the form $\{f,e_0\}$ for some resource $f \in E \setminus \{e_0\}$. Let $\{f_1,\dots,f_q\}$ be the set of all such resources $f$. It follows that, for any fixed $j = 1,\dots,q$, the strategy $\{e,f_j\}$, for some $e \in E \setminus \{e_0\}$, cannot be flow-carrying, otherwise, \begin{eqnarray} l_{e_0}(x_{e_0}) &\leq& l_{e_0}(z_{e_0}) \ \ \ \ \text{(definition of $Z$)} \nonumber \\ &\leq& l_{e}(z_{e}) \ \ \ \ \ \ \text{(structure of Nash flow $z$ discussed above)} \nonumber \\ &\leq& l_e(x_e) \ \ \ \ \ \ \text{(definition of $X$)} \nonumber \end{eqnarray} which implies that $l_{e_0}(x_{e_0}) + l_{f_j}(x_{f_j}) \leq l_e(x_e) + l_{f_j}(x_{f_j})$. Since $\{e_0,f_j\}$ is a strategy of maximum latency, then also $\{e,f_j\}$ is a strategy of maximum latency, which contradicts our assumption.
Moreover, not all players in $x$ can use resource $e_0$, otherwise $1 = x_{e_0} \leq z_{e_0}$, meaning that also all players use resource $e_0$ in $z$. But since $x_e > z_e$ for all other resources, which are in $X$ since we have only one resource in $Z$, this leads to a contradiction, since the total flow on all edges in $x$ is then higher than the total flow on all edges in $z$.
Now, let $\{a,b\}$ be a flow-carrying strategy for some $a,b \in E\setminus \{e_0\}$, which exists by what was said above. For any flow-carrying strategy of the form $\{a,e\}$ for some $e \in E$, we have $$ l_a(x_a) + l_e(x_e) \leq l_{P_{\max}}(x) \leq (1+\beta)l_P(x) $$ for any other strategy $P$. We can then raise the value of $l_a(x_a)$ until the latency of some strategy of the form $\{a,e\}$ becomes tight with respect to $l_{P_{\max}}(x)$. Note that, since $a \in X$, we have $x_a > z_a$ and therefore this does not contradict the fact that $l_a$ is non-decreasing. More precisely, we replace $l_a$ by some function $\hat{l}_a(y)$ which is non-negative, continuous and non-decreasing, and that satisfies $\hat{l}_a(z_a) = l_a(z_a)$ and $\hat{l}_a(x_a) = l_a(x_a) + \alpha$, where $\alpha$ is the smallest value such that $l_a(x_a) + \alpha + l_e(x_e) = l_{P_{\max}}(x)$ for some flow-carrying resource of the form $\{e,a\}$. If $a \neq e_0$, we have found the desired result, and otherwise, we can use a similar argument as in the beginning of the claim to show that $\{a,b\}$ is also a flow-carrying strategy with maximum latency. Note that the social cost $C(x)$ of $x$ can only get worse if we replace the function $l_a$ by $\hat{l}_a$.\qed \end{proof}
We now use Claims 1 and 2 to establish the bound $(1+\epsilon)/(1-\epsilon)$ for $\epsilon < 1$. Let $P_{\max} = \{a,b\}$ for $a,b \in E \setminus \{e_0\}$ (which is justified because of Claim 2). The Nash conditions then give (by comparing $\{a,b\}$ with $\{a,e_0\}$ and $\{b,e_0\}$) $$ l_a(x_a) + l_b(x_b) \leq (1+\epsilon)[l_a(x_a) + l_{e_0}(x_{e_0})] $$ $$ l_a(x_a) + l_b(x_b) \leq (1+\epsilon)[l_b(x_b) + l_{e_0}(x_{e_0})] $$ Adding up these inequalities, and rewriting, gives \begin{equation}\label{eq:nash_matroid2} (1 -\epsilon)[l_a(x_a) + l_b(x_b)] \leq (1+\epsilon)[l_{e_0}(x_{e_0})+ l_{e_0}(x_{e_0})] \end{equation} By definition of $Z$, we have $l_{e_0}(x_{e_0}) \leq l_{e_0}(z_{e_0})$, and, moreover, by the structure of the flow $z$ (as discussed in the beginning of the proof), we have $l_{e_0}(z_{e_0}) \leq l_{e}(z_{e})$ for any other resource with $z_e > 0$. On top of that, it also holds that $l_{e_0}(z_{e_0}) + l_{e}(z_{e}) = C(z)$, for any $e$ with $z_e > 0$, because of the fact that the demand is normalized to $r = 1$. Combining all this implies that $l_{e_0}(x_{e_0})+ l_{e_0}(x_{e_0}) \leq C(z)$. Using this observation, combined with (\ref{eq:nash_matroid2}), and the fact that $\epsilon < 1$, it follows that $$ C(x) \leq l_a(x_a) + l_b(x_b) \leq \frac{1 + \epsilon}{1 - \epsilon} \cdot [l_{e_0}(x_{e_0})+ l_{e_0}(x_{e_0})] \leq \frac{1 + \epsilon}{1 - \epsilon} \cdot C(z) $$ which is the desired result. \qed \end{proof}
\subsection{Proof of Theorem \ref{thm:matroid}} \begin{rtheorem}{Theorem}{\ref{thm:matroid}} Let $\mathcal{J} = (E,(l_e)_{e \in E},(\mathcal{S}_i)_{i \in [k]},(r_i)_{i \in [k]})$ be a matroid congestion game with homogeneous players. Let $\beta \ge 0$ be fixed and consider $\beta$-bounded basis deviations as defined above. Then the $\beta$-deviation\xspace ratio is upper bounded by $\text{$\beta$-DR}(\mathcal{J}) \le 1 + \beta$. Further, this bound is tight already for $1$-uniform matroid congestion games.
\end{rtheorem} \begin{proof} The proof consists of establishing the following two claims. \begin{enumerate}[i)] \item We have $l_e(x_e) \leq (1+\beta)l_e(z_e)$ for all $e \in E$ with $x_e > z_e$. \item It holds that $$ \sum_{e : x_e > z_e} (x_e - z_e)l_e(x_e) \leq (1 + \beta) \sum_{e : z_e \geq x_e} (z_e - x_e)l_e(x_e). $$ \end{enumerate} From these two claims, the result can be derived as follows. Rewriting the inequality in $ii)$, we find $$ \sum_{e : x_e > z_e} x_el_e(x_e) \leq \sum_{e : x_e > z_e} z_el_e(x_e) + (1 + \beta) \sum_{e : z_e \geq x_e} (z_e - x_e)l_e(x_e). $$ Adding $\sum_{e : z_e \geq x_e} x_el_e(x_e)$ to both sides of the inequality, and using the definition of $C(x)$, implies $$ C(x) \leq \sum_{e : x_e > z_e} z_el_e(x_e) + (1 + \beta)\sum_{e : z_e \geq x_e} z_e l_e(x_e) - \beta \sum_{e : z_e \geq x_e} x_el_e(x_e) $$ In the first term, we now use that the fact that $l_e(x_e) \leq (1+\beta) l_e(z_e)$ for every $e$ in this summation, because of $i)$; in the second term, we use the fact that $l_e(x_e) \leq l_e(z_e)$ for all resources $e$ in the summation (which follows from $z_e \geq x_e$ and the fact that the latency functions are non-decreasing); the third term is left out using the fact that $\beta \sum_{e : z_e \geq x_e} x_el_e(x_e) \geq 0$ because of the non-negativity of the latency functions and $\beta$. This then implies that $$ C(x) \leq (1+\beta) \sum_{e : x_e > z_e} z_el_e(z_e) + (1 + \beta)\sum_{e : z_e \geq x_e} z_e l_e(z_e) - 0 = (1+\beta)C(z) $$ which gives the desired result.
It now remains to show that $i)$ and $ii)$ holds. We first prove $i)$, of which the proof is of a similar nature as a proof of Fujishige et al. \cite{Fujishige2015}. In particular, using similar arguments as Lemma 3.2 \cite{Fujishige2015}, it can be shown that for every $e \in E$ with $x_e > z_e$, there exists an $f \in E \setminus \{e\}$ with $z_f > x_f$ such that $$ l_e(x_e) + \delta_e(x_e) \leq l_f(x_f) + \delta(x_f) \text{ \ \ \ and \ \ \ } l_f(z_f) \leq l_e(z_e).\footnote{We refer the reader to the proof of Lemma 3.2 \cite{Fujishige2015} for details. We omit it here for notational reasons.} $$ The proof of this argument uses the fact that $x$ is a Nash flow w.r.t. the costs $l_e + \delta_e$, and $z$ a Nash flows w.r.t. to the latencies $l_e$. It follows that $$ l_e(x_e) \leq l_e(x_e) + \delta_e(x_e) \leq l_f(x_f) + \delta(x_f) \leq (1+\beta)l_f(x_f) \leq (1 + \beta)l_f(z_f) \leq (1+\beta)l_e(z_e) $$ using the fact that $0 \leq \delta_e(x_e) \leq \beta l_e(x_e)$ and the properties of the resources $e$ and $f$.
We now prove the second claim. We use a `variational inequality' argument similar to, e.g., the proof of Theorem 4 \cite{Lianeas2016}. Because of the fact that $x$ is $\beta$-deviated Nash flow, it follows that $$ \sum_{e \in E} x_e(l_e(x_e)+ \delta_e(x_e)) \leq \sum_{e \in E} z_e(l_e(x_e) + \delta_e(x_e)). $$ Rewriting gives $$ \sum_{e: x_e > z_e} (x_e - z_e)(l_e(x_e)+ \delta_e(x_e)) \leq \sum_{e: z_e \geq x_e} (z_e - x_e)(l_e(x_e)+ \delta_e(x_e)). $$ Using the fact that $0 \leq \delta_e(x_e) \leq \beta l_e(x_e)$ it follows that $$ \sum_{e: x_e > z_e} (x_e - z_e)l_e(x_e) \leq \sum_{e: z_e \geq x_e} (z_e - x_e)(l_e(x_e)+ \delta_e(x_e)) \leq (1+\beta)\sum_{e: z_e \geq x_e} (z_e - x_e)l_e(x_e) $$ which completes the proof.
Tightness follows, e.g., from a similar construction as in the last tightness construction of the proof of Theorem \ref{thm:hetero} (if applied to a homogeneous population). \qed \end{proof}
\end{document}
|
arXiv
|
{
"id": "1707.01278.tex",
"language_detection_score": 0.7399153113365173,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\title{About the spectra of a real nonnegative matrix and its signings}
\author[B]{Kawtar Attas} \ead{[email protected]}
\author[B]{Abderrahim Boussa\"{\i}ri\corref{cor1}} \ead{[email protected]}
\author[B]{Mohamed Zaidi} \ead{[email protected]}
\cortext[cor1]{Corresponding author}
\address[B]{Facult\'e des Sciences A\"{\i}n Chock, D\'epartement de Math\'ematiques et Informatique,
Km 8 route d'El Jadida, BP 5366 Maarif, Casablanca, Maroc}
\begin{abstract} For a real matrix $M$, we denote by $sp(M)$ the spectrum of $M$ and by $\left \vert M\right \vert $ its absolute value, that is the matrix obtained from $M$ by replacing each entry of $M$ by its absolute value. Let $A$ be a nonnegative real matrix, we call a \emph{signing} of $A$ every real matrix $B$ such that $\left \vert B\right \vert =A$. In this paper, we study the set of all signings of $A$ such that $sp(B)=\alpha sp(A)$ where $\alpha$ is a complex unit number. Our work generalizes some results obtained in \cite{anradha2013, cuihou13, shader}. \end{abstract}
\begin{keyword}
Spectra; digraph; nonnegative matrices; irreducible matrices.
\MSC 05C20, 05C50 \end{keyword}
\end{frontmatter}
\section{Introduction}
Throughout this paper, all matrices are real or complex. The identity matrix of order $n$ is denoted by $I_{n}$ and the transpose of a matrix $A$ by $A^{T}$. Let $\Sigma$ be a set of nonzero real or complex numbers such that if $x\in \Sigma$ then $x^{-1}\in \Sigma$.\ Two square matrices $A$ and $B$ are $\Sigma$-\emph{diagonally similar} if $B=\Lambda^{-1}A\Lambda$ for some nonsingular diagonal matrix $\Lambda$ with diagonal entries in $\Sigma$. A square matrix $A$ is \emph{reducible }if there exists a permutation matrix $P$, so that $A$ can be reduced to the form $PAP^{T}=\left( \begin{array} [c]{cc} X & Y\\ 0 & Z \end{array} \right) $ where $X$ and $Z$ are square matrices. A matrix which is not reducible is said to be \emph{irreducible}. A real matrix $A$ is nonnegative, $A\geq0$, if all its entries are nonnegative.
Let $A$ be a $n\times n$ real or complex matrix. The multiset $\left \{ \lambda_{1},\lambda_{2},\ldots,\lambda_{n}\right \} $ of eigenvalues of $A$ is called the \emph{spectrum} of $A$ and is denoted by $sp(A)$. We usually assume that $\left \vert \lambda_{1}\right \vert \geq \left \vert \lambda_{2}\right \vert \geq \ldots \geq \left \vert \lambda_{n}\right \vert $. The \emph{spectral radius} of $A$ is $\left \vert \lambda_{1}\right \vert $ that is denoted by $\rho(A)$. Clearly, two real matrices $\left \{ -1,1\right \} $-diagonally similar have the same absolute value and the same spectrum.
In this paper, we address the following problem.
\begin{prob} Let $A$ be a nonnegative real matrix and let $\alpha$ be a complex unit number. Characterize the set of all signings of $A$ such that $sp(B)=\alpha sp(A)$. \end{prob}
Our motivation comes from the works of Shader and So \cite{shader} which correspond to adjacency matrix of a\emph{ }graph. Let $G$ be a finite simple graph with vertex set $V(G)=\left \{ v_{1},\ldots,v_{n}\right \} $ and edge set $E(G)$. The \emph{adjacency matrix} of\emph{ }$G$ is the symmetric matrix $A(G)=(a_{ij})_{1\leq i,j\leq n}$ where $a_{ij}=a_{ji}=1$\ if $\left \{ v_{i},v_{j}\right \} $ is an edge of $G$, otherwise $a_{ij}=a_{ji}=0$. Since the matrix $A(G)$ is symmetric, its eigenvalues are real. The \emph{adjacency spectrum} $Sp\left( G\right) $ of $G$ is defined as the spectrum of $A(G)$. Let $G^{\sigma}$ be an orientation of $G$, which assigns to each edge a direction so that the resultant graph $G^{\sigma}$ becomes an oriented graph. The \emph{skew-adjacency} matrix of $G^{\sigma}$ is the real skew symmetric matrix $S(G^{\sigma})=(a_{ij}^{\sigma})_{1\leq i,j\leq n}$\ where $a_{ij}^{\sigma }=-a_{ji}^{\sigma}=1$ if $(i,j)$ is an arc of $G^{\sigma}$ and $a_{ij} ^{\sigma}=0$, otherwise. The \emph{skew-spectrum} $Sp\left( G^{\sigma }\right) $ of $G^{\sigma}$ is defined as the spectrum of $S(G^{\sigma})$. Note that $Sp(G^{\sigma})$ consists of only purely imaginary eigenvalues because $S(G^{\sigma})$ is a real skew symmetric matrix.
There are several recent works about the relationship between $Sp(G^{\sigma})$ and $Sp(G)$ (see for example \cite{anradha2013, anuradha2014, cavers, cuihou13, shader}). In the last paper, Shader and So have obtained a result which is closely related to our work. To state this result, let $G$ be a bipartite graph with bipartition $[ I,J]$. The orientation $\varepsilon_{I,J}$ that assigns to each edge of $G$ a direction from $I$ to $J$ is called the \emph{canonical orientation}. Shader and So showed \cite{shader} that $Sp\left( G^{\varepsilon_{I,J}}\right) =iSp\left( G\right) $. Moreover, they proved that a graph $G$ is bipartite if and only if $Sp\left( G^{\sigma}\right) =isp\left( G\right) $ for some orientation $G^{\sigma}$ of $G$. For connected graphs, this result can be viewed as a particular case of Proposition \ref{shader generalis} because a graph $G$ is bipartite if and only if there exists a permutation matrix $P$ such that $PA(G)P^{T}=\left( \begin{array} [c]{cc} 0 & X\\ X^{T} & 0 \end{array} \right) $ where the zero diagonal blocks are square. Consider now an orientation $G^{\sigma}$ of a graph $G$ and let $W$ be a subset of $V(G)$. The orientation $G^{\tau}$ of $G$ obtained from $G^{\sigma}$ by reversing the direction of all arcs between $W$ and $V(G)\backslash W$ is said to be obtained from $G^{\sigma}$ by \emph{switching} with respect to $W$. Two orientations are \emph{switching-equivalent} if one can be obtained from the other by switching. Anuradha, Balakrishnan, Chen, Li, Lian and So \cite{anradha2013} proved that $Sp\left( G^{\sigma}\right) =isp\left( G\right) $ if an only if $G^{\sigma}$ is switching-equivalent to the canonical orientation. For a bipartite connected graph, this is a direct consequence of Proposition \ref{solution shader} because two orientations of a graph are switching-equivalent if and only if their skew-adjacency matrices are $\{-1,1\}$-diagonally similar.
In this work, we consider only the case of irreducible matrices (not necessarily symmetric). To state our main result, we need some terminology. A \emph{digraph }$D$ is a pair consisting of a finite set $V(D)$ of \emph{vertices} and a subset $E(D)$ of ordered pairs of vertices called \emph{arcs}. Let $v,v^{\prime}$ two vertices of $D$, a \emph{path} $P$ from $v$ to $v^{\prime}$ is a finite sequence $v_{0}=v,\ldots,v_{k}=v^{\prime}$ such that $(v_{0},v_{1}),\ldots,(v_{k-1},v_{k})$ are arcs of $D$. The \emph{length} of $P$ is the number $k$ of its arcs.\ If $v_{0}=v_{k}$, we say that $P$ is a \emph{closed path. }A digraph is said to be\emph{ strongly connected} if for any two vertices $v$ and $v^{\prime}$, there is a path joining $v$ to $v^{\prime}$. It is easy to see that a strongly connected digraph has a closed path. The \emph{period} of a digraph is the greatest common divisor of the lengths of its closed paths. A digraph is \emph{aperiodic }if its period is one.
With each $n\times n$ matrix $A=(a_{ij})_{1\leq i,j\leq n}$, we associate a digraph $D_{A}$ on the vertex set $\left[ n\right] =\left \{ 1,\ldots ,n\right \} $ and with arc set $E(D_{A})=\left \{ (i,j):a_{ij}\neq0\right \} $. It is easy to show that $A$ is irreducible if and only if $D_{A}$ is strongly connected. The \emph{period} of a matrix is the period of its associate digraph.
\begin{nt} Let $A$ be an irreducible nonnegative real matrix with period $p$ and let $\alpha$ be a complex unit number, we denote by $\mathcal{M}(\alpha,A)$ the set of all signings of $A$ such that $sp(B)=\alpha sp(A)$. \end{nt}
In Corollary \ref{forme de alpha}, we gave a necessary condition for $\mathcal{M}(\alpha,A)$ to be nonempty. More precisely, we will prove that if $\mathcal{M}(\alpha,A)$ is nonempty, then $\alpha=e^{\frac{i\pi k}{p}}$ for some $k\in \left \{ 0,\ldots,2p-1\right \} $. If $p>1$, we will construct (see Corollary \ref{nonempty period p}) for each $k\in \left \{ 0,\ldots,2p-1\right \} $ a matrix $\widetilde{A}$ such that $sp(\widetilde{A})=e^{\frac{i\pi k}{p}}sp(A)$. Finally, by using Corollary \ref{forme de alpha} and Proposition \ref{structurede malpha}, we prove our main theorem.
\begin{thm} \label{main theorem}Let $A$ be an irreducible nonnegative matrix with period $p$ and let $\alpha$ be an unit complex number. Then the following statements hold:
\begin{description} \item[i)] $\mathcal{M}(\alpha,A)$ is nonempty iff $\alpha=e^{\frac{i\pi k}{p} }$ for some $k\in \left \{ 0,\ldots,2p-1\right \} $;
\item[ii)] For $k\in \left \{ 0,\ldots,2p-1\right \} $, we have
\item[a)] if $k$ is even then $\mathcal{M}(e^{\frac{i\pi k}{p}},A)$ is the set of all matrices $\left \{ -1,1\right \} -$diagonally similar to $A$;
\item[b)] if $k$ is odd then $\mathcal{M}(e^{\frac{i\pi k}{p}},A)$ is the set of all matrices $\left \{ -1,1\right \} -$diagonally similar to $B_{0}$ where $B_{0}$ is an arbitrary signing of $A$ such that $sp(B_{0})=e^{\frac{i\pi}{p}}sp(A)$. \end{description} \end{thm}
\begin{req} Let $A$ be an aperiodic matrix. It follows from the above Theorem that $\mathcal{M}(\alpha,A)$ is empty whenever $\alpha \neq \pm1$. Moreover, $\mathcal{M}(1,A)$ (resp. $\mathcal{M}(-1,A)$) is the set of all matrices $\left \{ -1,1\right \} -$diagonally similar to $A$ (resp. to $-A$). \end{req}
\section{Some properties of $\mathcal{M}(\alpha,A)$}
We will use the following theorem due to Frobenius (see \cite{godsil}).
\begin{thm}
\label{frobenius} Let $A$ be an irreducible nonnegative $n\times n$ matrix and let $B$ be a complex $n\times n$ matrix such that $|B|\leq A$. Then $\left \vert \lambda \right \vert \leq \rho(A)$ for each eigenvalue $\lambda$ of $B$ and $\lambda=\rho(A)e^{i\theta}$ iff $B=e^{i\theta}LAL^{-1}$, where $L$ is a complex diagonal matrix such that $\left \vert L\right \vert =I_{n}$. \end{thm}
In the next Proposition, we describe the structure of $\mathcal{M}(\alpha,A)$.
\begin{prop} \label{structurede malpha} Let $A$ be an irreducible nonnegative matrix with period $p$ and let $\alpha$ be an unit complex number. Then $\mathcal{M} (\alpha,A)$ is empty or $\mathcal{M}(\alpha,A)$ is the set of all matrices $\left \{ -1,1\right \} -$diagonally similar to $B_{0}$ where $B_{0}$ is an arbitrary matrix in $\mathcal{M}(\alpha,A)$. \end{prop}
\begin{pf} Assume that $\mathcal{M}(\alpha,A)$ is nonempty and let $B_{0}\in \mathcal{M}(\alpha,A)$. It is easy to see that $\Lambda^{-1}B_{0}\Lambda\in \mathcal{M} (\alpha,A)$ for every $\left \{ -1,1\right \} $-diagonal matrix $\Lambda$. Conversely, let $B\in \mathcal{M}(\alpha,A)$. Then $sp(B)=sp(B_{0})=\alpha sp(A)$. Let $\lambda=\rho(A)e^{i\theta}$ be a common eigenvalue of $B$ and $B_{0}$. By Theorem \ref{frobenius}, we have $B=e^{i\theta}LAL^{-1}$ and $B_{0}=e^{i\theta}L^{\prime}AL^{\prime-1}$ where $L$, $L^{\prime}$ are complex diagonal matrices such that $\left \vert L\right \vert =\left \vert L^{\prime }\right \vert =I_{n}$. It follows that $B=(L^{^{\prime}}L)^{-1}B_{0}L^{\prime }L^{-1}$. To conclude, it suffices to apply Lemma \ref{realsimilarit} below. \qed\end{pf}
\begin{lem} \label{realsimilarit} Let $A$ be an $n\times n$ irreducible nonnegative matrix and let $B$, $B^{\prime}$ be two signings of $A$. If there exists a complex diagonal matrix $\Gamma$ such that $B^{\prime}=\Gamma B\Gamma^{-1}$ and $\left \vert \Gamma \right \vert =I_{n}$ then $B$ and $B^{\prime}$ are $\left \{ -1,1\right \} -$diagonally similar. \end{lem}
\begin{pf} Let $A:=(a_{ij})_{1\leq i,j\leq n}$, $B:=(b_{ij})_{1\leq i,j\leq n}$ and $B^{\prime}: =(b_{ij}^{\prime})_{1\leq i,j\leq n}$. We denote by $\gamma_{1},\ldots,\gamma_{n}$ the diagonal entries of $\Gamma$. Let $\Delta$ be the diagonal matrix with diagonal entries $\delta_{j} =\gamma_{j}\gamma_{1}^{-1}$ for $j=1,\ldots, n$. As $A$ is irreducible, the digraph $D_{A}$ is strongly connected and then there is a path $j=i_{1} ,\ldots,i_{r}=1$ of $D_{A}$ from $j$ to $1$. By definition of $D_{A}$, we have $a_{i_{1}i_{2}}\neq0,\ldots,a_{i_{r-1}i_{r}}\neq0$. It follows that $b_{i_{1}i_{2}}\neq0$,\ldots, $b_{i_{r-1}i_{r}}\neq0$. and $b_{i_{1}i_{2} }^{\prime}\neq0$,\ldots, $b_{i_{r-1}i_{r}}^{\prime}\neq0$ because $\left \vert B\right \vert =\left \vert B^{\prime}\right \vert =A$. Moreover, from the equality $B^{\prime}=\Gamma B\Gamma^{-1}$ we have $b_{i_{1}i_{2}}^{\prime }=\gamma_{i_{1}}b_{i_{1}i_{2}}\gamma_{i_{2}}^{-1}$ ,$b_{i_{2}i_{3}}^{\prime }=\gamma_{i_{2}}b_{i_{2}i_{3}}\gamma_{i_{3}}^{-1}$,\ldots, $b_{i_{r-1}i_{r} }^{\prime}=\gamma_{i_{r-1}}b_{i_{r-1}i_{r}}\gamma_{i_{r}}^{-1}$. Then $b_{i_{1}i_{2}}^{\prime}\ldots b_{i_{r-1}i_{r}}^{\prime}=\gamma_{i_{1}} \gamma_{i_{r}}^{-1}b_{i_{1}i_{2}}\ldots b_{i_{r-1}i_{r}}$. But by hypothesis, $B,B^{\prime}$ are real matrices and $\left \vert B\right \vert =\left \vert B^{\prime}\right \vert $, then $b_{i_{1}i_{2}}^{\prime}\ldots b_{i_{r-1}i_{r} }^{\prime}=\pm b_{i_{1}i_{2}}\ldots b_{i_{r-1}i_{r}}$ and hence $\delta _{j}=\gamma_{j}\gamma_{1}^{-1}=\gamma_{i_{1}}\gamma_{i_{r}}^{-1}\in \left \{ -1,1\right \} $. To conclude, it suffices to see that $\Delta B\Delta ^{-1}=\Gamma B\Gamma^{-1}=B^{\prime}$. \qed\end{pf}
By using Theorem \ref{frobenius}, we obtain the following.
\begin{prop} \label{eigenvalue of B} Let $A$ be an irreducible nonnegative real matrix with period $p$ and let $B$ be a signing of $A$ such that $\rho(B)=\rho(A)$. If $\lambda$ is an eigenvalue of $B$ such that $\left \vert \lambda \right \vert =\rho(A)$, then $\lambda =\rho(A)e^{\frac{i\pi k}{p}}$ for some $k\in \left \{ 0,\ldots,2p-1\right \} $. \end{prop}
\begin{pf}
Let $A:=(a_{ij})_{1\leq i,j\leq n}$, $B:=(b_{ij})_{1\leq i,j\leq n}$ and
$\lambda=\rho(A)e^{i\theta}$. By Theorem \ref{frobenius}, we have $B=e^{i\theta}LAL^{-1}$ where $L$ is a complex diagonal matrix such that $\left \vert L\right \vert =I_{n}$. It follows that $b_{ij}=e^{i\theta}l_{i}a_{ij}l_{j} ^{-1}$ for $i,j\in \left \{ 1,\ldots,n\right \} $, where $l_{1},\ldots,l_{n}$ are the diagonal entries of $L$. Consider now a closed path $C=(i_{1} ,i_{2},\ldots,i_{r},i_{1})$ of $D_{A}$. By the previous equality, we have \begin{align*} \text{ }\frac{b_{i_{1}i_{2}}\ldots b_{i_{r-1}i_{r}}b_{i_{r}i_{1}}} {a_{i_{1}i_{2}}\ldots a_{i_{r-1}i_{r}}a_{i_{r}i_{1}}} & =(e^{i\theta }l_{i_{1}}l_{i_{2}}^{-1})\ldots(e^{i\theta}l_{i_{r-1}}l_{i_{r}}^{-1} )(e^{i\theta}l_{i_{r}}l_{i_{1}}^{-1})\\ & =(e^{i\theta})^{r} \end{align*}
Then $(e^{i\theta})^{r}\in \left \{ 1,-1\right \} $ because $\left \vert B\right \vert =A$.
Since $p$ is the greatest common divisor of the lengths of the closed paths in $D_{A}$, we have $(e^{i\theta})^{p}\in \left \{ 1,-1\right \} $ and then $\lambda=\rho(A)e^{\frac{i\pi k}{p}}$ for some $k\in \left \{ 0,\ldots ,2p-1\right \} $. \qed\end{pf}
\begin{req} \label{eigenvalues of A} Let $A=(a_{ij})_{1\leq i,j\leq n}$ be an irreducible nonnegative real matrix with period $p$ and let $\lambda$ be an eigenvalue of $A$ such that $\left \vert \lambda \right \vert =\rho(A)$. By applying Proposition \ref{eigenvalue of B} to $B=A$, we have $\lambda=\rho (A)e^{\frac{i\pi k}{p}}$ for some $k\in \left \{ 0,\ldots,2p-1\right \} $. \end{req}
The following result gives a necessary condition under which $\mathcal{M} (\alpha,A)$ is nonempty.
\begin{cor} \label{forme de alpha} Let $A$ be an irreducible nonnegative real matrix of period $p$ and let $\alpha$ be a complex unit number. If $\mathcal{M} (\alpha,A)$ is nonempty then $\alpha=e^{\frac{i\pi k}{p}}$ for some $k\in \left \{ 0,\ldots,2p-1\right \} $, in particular $\alpha^{p}=\pm1$. \end{cor}
\begin{pf} Let $\lambda$ be an eigenvalue of $A$ such that $\left \vert \lambda \right \vert =\rho(A)$. By applying Remark \ref{eigenvalues of A}, we have $\lambda =\rho(A)e^{\frac{i\pi k}{p}}$ for some $k\in \left \{ 0,\ldots,2p-1\right \} $. Let $B\in \mathcal{M}(\alpha,A)$. We have $\alpha \rho(A)e^{\frac{i\pi k}{p}}\in sp(B)$ because $sp(B)=\alpha sp(A)$. It follows from Proposition \ref{eigenvalue of B} that $\alpha \rho(A)e^{\frac{i\pi k}{p}}=\rho (A)e^{\frac{i\pi h}{p}}$ for some $h\in \left \{ 0,\ldots,2p-1\right \} $ and hence $\alpha=e^{\frac{i\pi(h-k)}{p}}$. \qed\end{pf}
\section{Proof of the Main Theorem}
Let $n$ be a positive integer and let $(r_{1},\ldots,r_{p})$ be a partition of $n$, that is $r_{1},\ldots,r_{p}$ are positive integers and $r_{1} +\cdots+r_{p}=n$. For $i=1,\ldots,p-1$, let $A_{i}$ be a $r_{i}\times r_{i+1} $ matrix and let $A_{p}$ be a $r_{p}\times r_{1}$ matrix. The matrix $\left( \begin{array} [c]{ccccc} 0 & A_{1} & 0 & \cdots & 0\\ 0 & 0 & A_{2} & \cdots & 0\\ \vdots & \vdots & \ddots & \ddots & \vdots \\ 0 & 0 & 0 & \ddots & A_{p-1}\\ A_{p} & 0 & \cdots & 0 & 0 \end{array} \right) $ is denoted by $Cyc(A_{1},A_{2},\ldots,A_{p})$.
Each matrix of this form is called $p$-\emph{cyclic. }
Recall\ the well-known result of Frobenius about irreducible matrices with period $p>1$.
\begin{prop} \label{cyclic form} Let $A$ be an irreducible nonnegative real matrix with period $p>1$, then there exists a permutation matrix $P$ such that $PAP^{T}$ is $p$-cyclic. \end{prop}
\begin{req} \label{coverce aperiodic} For aperiodic matrices, the converse of Corollary \ref{forme de alpha} is true because $A\in \mathcal{M}(1,A)$ and $-A\in \mathcal{M}(-1,A)$. \end{req}
For $p>1$, we have the following result.
\begin{prop} \label{converse for periodic} Let $A=Cyc(A_{1},A_{2},\ldots,A_{p})$ be a nonnegative $p$-cyclic matrix where $A_{i}$ is a $r_{i}\times r_{i+1}$ matrix for $i=1,\ldots,p-1$ and $A_{p}$ is a $r_{p}\times r_{1}$ matrix. Let $\widetilde{A}$ be the matrix obtained from $A$ by replacing the block $A_{p}$ by $-A_{p}$. Given $\alpha=e^{\frac{i\pi k}{p}}$ where $k\in \left \{ 0,\ldots,2p-1\right \} $, then
\begin{description} \item[i)] if $k$ is even, $e^{\frac{i\pi k}{p}}A$ is diagonally similar to $A$, in particular $sp(A)=e^{\frac{i\pi k}{p}}sp(A)$;
\item[ii)] if $k$ is odd, $e^{\frac{i\pi k}{p}}A$ is diagonally similar to $\widetilde{A}$, in particular $sp(\widetilde{A})=e^{\frac{i\pi k}{p}}sp(A)$. \end{description} \end{prop}
\begin{pf} Let $L:=\left( \begin{array} [c]{ccccc} I_{r_{1}} & 0 & 0 & \cdots & 0\\ 0 & e^{\frac{i\pi}{p}}I_{r_{2}} & 0 & \cdots & 0\\ \vdots & \vdots & \ddots & \ddots & \vdots \\ 0 & 0 & 0 & \ddots & 0\\ 0 & 0 & \cdots & 0 & e^{\frac{i\pi(p-1)}{p}}I_{r_{p}} \end{array} \right) $. It easy to check that if $k$ is even, $e^{\frac{i\pi k}{p} }LAL^{-1}=A$ and if $k$ is odd, $e^{\frac{i\pi k}{p}}LAL^{-1}=\widetilde{A}$. \qed\end{pf}
The next Corollary is a direct consequence of the above Proposition and Theorem \ref{cyclic form}.
\begin{cor} \label{nonempty period p} Let $A$ be an irreducible nonnegative matrix with period $p>1$. Then $\mathcal{M}(e^{\frac{i\pi k}{p}},A)$ is nonempty for $k\in \left \{ 0,\ldots,2p-1\right \} $. \end{cor}
\begin{pf} By Proposition \ref{cyclic form}, there exists a permutation matrix $P$ such that $PAP^{T}$ is $p-$cyclic. Let $A^{\prime}:=PAP^{T}:=Cyc(A_{1},A_{2} ,\ldots,A_{p})$ and let $\widetilde{A^{\prime}}$ the matrix obtained from $A^{\prime}$ by replacing the block $A_{p}$ by $-A_{p}$. It follows from Proposition \ref{converse for periodic} that $sp(A^{\prime})=e^{\frac{i\pi k}{p}}sp(A^{\prime})$ if $k$ is even and $sp(\widetilde{A^{\prime}} )=e^{\frac{i\pi k}{p}}sp(A^{\prime})$ if $k$ is odd. If $k$ is even then $sp(A)=e^{\frac{i\pi k}{p}}sp(A)$ because $P^{T}A^{\prime}P=A$ and hence $A\in \mathcal{M}(e^{\frac{i\pi k}{p}},A)$. If $k$ is odd then $sp(P^{T} \widetilde{A^{\prime}}P)=e^{\frac{i\pi k}{p}}sp(P^{T}A^{\prime}P)=e^{\frac {i\pi k}{p}}sp(A)$. Moreover, since$\left \vert \widetilde{A^{\prime} }\right \vert =A^{\prime}$, we have $\left \vert P^{T}\widetilde{A^{\prime} }P\right \vert =P^{T}A^{\prime}P=A$ and then $P^{T}\widetilde{A^{\prime}} P\in \mathcal{M}(e^{\frac{i\pi k}{p}},A)$. \qed\end{pf}
\begin{req} \label{egalite des malpha} It follows from Corollary \ref{forme de alpha}, Remark \ref{coverce aperiodic} and Corollary \ref{nonempty period p} that $\mathcal{M}(\alpha,A)$ is nonempty iff $\alpha=e^{\frac{i\pi k}{p}}$ for some $k\in \left \{ 0,\ldots,2p-1\right \} $. Moreover, as $e^{\frac{i\pi k}{p} }sp(A)=sp(A)$ if $k$ is even, we have $e^{\frac{i\pi k}{p}}sp(A)=e^{\frac {i\pi}{p}}sp(A)$ if $k$ is odd and then \begin{align*} \mathcal{M}(1,A) & =\mathcal{M}(e^{\frac{2i\pi}{p}},A)=\cdots=\mathcal{M} (e^{\frac{2(p-1)i\pi}{p}},A)\\ \mathcal{M}(e^{\frac{i\pi}{p}},A) & =\mathcal{M}(e^{\frac{3i\pi}{p} },A)=\cdots=\mathcal{M}(e^{\frac{(2p-1)i\pi}{p}},A) \end{align*}
\end{req}
\begin{pot}
i) (Necessity) This follows from Corollary \ref{forme de alpha}.
(Sufficiency) If $p=1$ then from Corollary \ref{forme de alpha} we have $\alpha=\pm1$ and hence by Remark \ref{coverce aperiodic}, $\mathcal{M} (\alpha,A)$ is nonempty. If $p>1$, it suffices to apply Corollary \ref{nonempty period p}.
ii) By Remark \ref{egalite des malpha}, $\mathcal{M}(e^{\frac{i\pi k}{p} },A)=\mathcal{M}(1,A)$ if $k$ is even and $\mathcal{M}(e^{\frac{i\pi k}{p} },A)=\mathcal{M}(e^{\frac{i\pi}{p}},A)$ if $k$ is odd. To prove a) and b) it suffices to apply Proposition \ref{structurede malpha} respectively to $\alpha=1$ and to $\alpha=e^{\frac{i\pi}{p}}$. \qed\end{pot}
\section{The special case of symmetric matrices}
In this section we give some consequences of Theorem \ref{main theorem} when the matrix $A$ is symmetric.
\begin{prop} \label{shader generalis} Let $A$ be an irreducible nonnegative symmetric real matrix. Then the following statements are equivalent :
\begin{description} \item[i)] there is a real skew-symmetric matrix $B$ such that $\left \vert B\right \vert =A$ and $sp(B)=isp(A)$;
\item[ii)] there exists a permutation matrix $P$ such that $PAP^{T}=\left( \begin{array} [c]{cc} 0 & X\\ X^{T} & 0 \end{array} \right) $ where the zero diagonal blocks are square. \end{description} \end{prop}
\begin{pf} Note that the period of $A$ is a most $2$ because its associate digraph contains a closed path of length $2$.
First, we prove the implication i)$\implies$ii). It follows from i) that $\mathcal{M}(i,A)$ is nonempty. Then, by Corollary \ref{forme de alpha}, the period of $A$ is necessarily $2$. It follows from Proposition \ref{cyclic form} that there exists a permutation matrix $P$, so that $PAP^{T}=\left( \begin{array} [c]{cc} 0 & X\\ Y & 0 \end{array} \right) $ where the zero diagonal blocks are square. But, the matrix $PAP^{T}$ is symmetric, then $Y=X^{T}$.
To prove ii) $\implies$ i), it suffices to apply Proposition \ref{converse for periodic} for $p=2$ and $k=1$. \qed\end{pf}
\begin{prop} \label{solution shader} Let $A=\left( \begin{array} [c]{cc} 0 & X\\ X^{T} & 0 \end{array} \right) $ be an irreducible nonnegative symmetric matrix and let $B$ be a skew symmetric matrix such that $\left \vert B\right \vert =A$. Then the following statements are equivalent :
\begin{description} \item[i)] $sp(B)=isp(A)$;
\item[ii)] $B$ is $\{-1,1\}$-diagonally similar to $\widetilde{A}=\left( \begin{array} [c]{cc} 0 & X\\ -X^{T} & 0 \end{array} \right) $. \end{description} \end{prop}
\begin{pf} It follows from assertion ii) of Proposition \ref{converse for periodic} that $sp(\widetilde{A})=isp(A)$ and then $\widetilde{A}\in \mathcal{M}(i,A)$. The equivalence i) $\Longleftrightarrow$ ii) result from assertion b) of Theorem \ref{main theorem}. \qed\end{pf}
\end{document}
|
arXiv
|
{
"id": "1507.07860.tex",
"language_detection_score": 0.6295179724693298,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[Moduli spaces of weighted pointed stable rational curves]{Moduli spaces of weighted pointed stable rational curves via GIT} \date{March 2010}
\author{Young-Hoon Kiem} \address{Department of Mathematics and Research Institute of Mathematics, Seoul National University, Seoul 151-747, Korea} \email{[email protected]}
\author{Han-Bom Moon} \address{Department of Mathematics, Seoul National University, Seoul 151-747, Korea} \email{[email protected]}
\thanks{Partially supported by an NRF grant}
\begin{abstract} We construct the Mumford-Knudsen space $\overline{M}_{0,n} $ of $n$-pointed stable rational curves by a sequence of explicit blow-ups from the GIT quotient $(\mathbb{P}^1)^n/\!/ SL(2)$ with respect to the symmetric linearization $\mathcal{O} (1,\cdots,1)$. The intermediate blown-up spaces turn out to be the moduli spaces of weighted pointed stable curves $\overline{M}_{0,n\cdot \epsilon} $ for suitable ranges of $\epsilon$. As an application, we provide a new unconditional proof of M. Simpson's Theorem about the log canonical models of $\overline{M}_{0,n} $. We also give a basis of the Picard group of $\overline{M}_{0,n\cdot \epsilon} $. \end{abstract}
\maketitle
\section{Introduction}\label{sec1} Recently there has been a tremendous amount of interest in the birational geometry of moduli spaces of stable curves. See for instance \cite{AlexSwin, FedoSmyt,Hassett,Kapranov,Keel,Li, Mustata, Simpson} for the genus 0 case only. Most prominently, it has been proved in \cite{AlexSwin, FedoSmyt,Simpson} that the log canonical models for $(\overline{M}_{0,n} , K_{\overline{M}_{0,n} }+\alpha D)$, where $D$ is the boundary divisor and $\alpha$ is a rational number, give us Hassett's moduli spaces $\overline{M}_{0,n\cdot \epsilon} $ of weighted pointed stable curves with \emph{symmetric} weights $n\cdot \epsilon =(\epsilon,\cdots, \epsilon)$. See \S\ref{sec2.1} for the definition of $\overline{M}_{0,n\cdot \epsilon} $ and Theorem \ref{thm1.2} below for a precise statement. The purpose of this paper is to prove that actually all the moduli spaces $\overline{M}_{0,n\cdot \epsilon} $ can be constructed by explicit blow-ups from the GIT quotient $(\mathbb{P}^1)^n/\!/ SL(2)$ with respect to the symmetric linearization $\mathcal{O} (1,\cdots,1)$ where $SL(2)$ acts on $(\mathbb{P}^1)^n$ diagonally. More precisely, we prove the following. \begin{theorem}\label{thm1.1} There is a sequence of blow-ups \begin{equation}\label{eq1} \overline{M}_{0,n} =\overline{M}_{0,n\cdot \epsilon_{m-2}} \to \overline{M}_{0,n\cdot \epsilon_{m-3}} \to \cdots \to \overline{M}_{0,n\cdot \epsilon_2} \to \overline{M}_{0,n\cdot \epsilon_1} \to (\mathbb{P}^1)^n/\!/ SL(2) \end{equation} where $m=\lfloor\frac{n}{2}\rfloor$ and $\frac1{m+1-k}<\epsilon_k\le \frac1{m-k}$. Except for the last arrow when $n$ is even, the center for each blow-up is a union of transversal smooth subvarieties of same dimension. When $n$ is even, the last arrow is the blow-up along the singular locus which consists of $\frac12\binom{n}{m}$ points in $(\mathbb{P}^1)^n/\!/ SL(2)$, i.e. $\overline{M}_{0,n\cdot \epsilon_1} $ is Kirwan's partial desingularization (see \cite{Kirwan}) of the GIT quotient $(\mathbb{P}^1)^{2m}/\!/ SL(2)$.\end{theorem} If the center of a blow-up is the transversal union of smooth subvarieties in a nonsingular variety, the result of the blow-up is isomorphic to that of the sequence of smooth blow-ups along the irreducible components of the center in any order (see \S\ref{sec2.3}). So each of the above arrows can be decomposed into the composition of smooth blow-ups along the irreducible components.
As an application of Theorem \ref{thm1.1}, we give a new proof of the following theorem of M. Simpson (\cite{Simpson}) without relying on Fulton's conjecture. \begin{theorem}\label{thm1.2} Let $\alpha$ be a rational number satisfying $\frac{2}{n-1} < \alpha \le 1$ and let $D=\overline{M}_{0,n} -M_{0,n}$ denote the boundary divisor. Then the log canonical model $$\overline{M}_{0,n} (\alpha) = \mathrm{Proj}\;\left(\bigoplus_{l \ge 0} H^0(\overline{M}_{0,n} , \mathcal{O} (\lfloor l (K_{\overline{M}_{0,n} } +\alpha D) \rfloor))\right) $$ satisfies the following: \begin{enumerate} \item If $\frac{2}{m-k+2} < \alpha \le \frac{2}{m-k+1}$ for $1\le k\le m-2$, then $\overline{M}_{0,n} (\alpha) \cong \overline{M}_{0,n\cdot \epsilon_k} $.\item If $\frac{2}{n-1} < \alpha \le \frac{2}{m+1}$, then $\overline{M}_{0,n} (\alpha) \cong (\mathbb{P}^1)^n /\!/ G$ where the quotient is taken with respect to the symmetric linearization $\mathcal{O} (1,\cdots,1)$. \end{enumerate} \end{theorem} There are already two different \emph{unconditional} proofs of Theorem \ref{thm1.2} by Alexeev-Swinarski \cite{AlexSwin} and by Fedorchuk-Smyth \cite{FedoSmyt}. See Remark \ref{rem-compproofs} for a brief outline of the two proofs. In this paper we obtain the ampleness of some crucial divisors directly from Theorem \ref{thm1.1}. As another application, we give an explicit basis of the Picard group of $\overline{M}_{0,n\cdot \epsilon_k} $ for each $k$.
It is often the case in moduli theory that adding an extra structure makes a problem easier. Let $0\le k< n$. A pointed nodal curve $(C,p_1,\cdots, p_n)$ of genus $0$ together with a morphism $f:C\to \mathbb{P}^1$ of degree $1$ is called \emph{$k$-stable} if \begin{enumerate} \item[i.] all marked points $p_i$ are smooth points of $C$; \item[ii.] no more than $n-k$ of the marked points $p_i$ can coincide; \item[iii.] any ending irreducible component $C'$ of $C$ which is contracted by $f$ contains more than $n-k$ marked points; \item[iv.] the group of automorphisms of $C$ preserving $f$ and $p_i$ is finite. \end{enumerate} A. Mustata and M. Mustata prove the following in \cite{Mustata}. \begin{theorem} \cite[\S1]{Mustata} There is a fine moduli space $F_k$ of $k$-stable pointed parameterized curves $(C,p_1,\cdots,p_n,f)$. Furthermore, the moduli spaces $F_k$ fit into a sequence of blow-ups \begin{equation}\label{2010eq1} \xymatrix{ \mathbb{P}^1[n]\ar@{=}[r] & F_{n-2}\ar[r]^{\psi_{n-2}} & F_{n-3}\ar[r]^{\psi_{n-3}} & \cdots \ar[r]^{\psi_2} & F_1\ar[r]^{\psi_1} & F_0\ar@{=}[r] & (\mathbb{P}^1)^n } \end{equation} whose centers are transversal unions of smooth subvarieties. \end{theorem} The first term $\mathbb{P}^1[n]$ is the Fulton-MacPherson compactification of the configuration space of $n$ points in $\mathbb{P}^1$ constructed in \cite{FM}. The blow-up centers are transversal unions of smooth subvarieties and hence we can further decompose each arrow into the composition of smooth blow-ups along the irreducible components in any order. This blow-up sequence is actually a special case of L. Li's inductive construction of a \emph{wonderful compactification} of the configuration space and transversality of various subvarieties is a corollary of Li's result \cite[Proposition 2.8]{Li}. (See \S\ref{sec2.3}.) The images of the blow-up centers are invariant under the diagonal action of $SL(2)$ on $(\mathbb{P}^1)^n$ and so this action lifts to $F_k$ for all $k$. The aim of this paper is to show that the GIT quotient of the sequence \eqref{2010eq1} by $SL(2)$ gives us \eqref{eq1}.
To make sense of GIT quotients, we need to specify a linearization of the action of $G=SL(2)$ on $F_k$. For $F_0=(\mathbb{P}^1)^n$, we choose the symmetric linearization $L_0=\mathcal{O} (1,\cdots, 1)$. Inductively, we choose $L_k=\psi^*_kL_{k-1}\otimes \mathcal{O} (-\delta_kE_{k})$ where $E_k$ is the exceptional divisor of $\psi_k$ and $0<\delta_k<\!<\delta_{k-1}<\!< \cdots <\!<\delta_1<\!<1$. Let $F_k^{ss}$ (resp. $F_k^s$) be the semistable (resp. stable) part of $F_k$ with respect to $L_k$. Then by \cite[\S3]{Kirwan} or \cite[Theorem 3.11]{Hu}, we have \begin{equation}\label{2010eq2} \psi_k^{-1}(F^s_{k-1})\subset F_k^s\subset F_k^{ss}\subset \psi_k^{-1}(F_{k-1}^{ss}). \end{equation} In particular, we obtain a sequence of morphisms $$\bar\psi_k:F_k/\!/ G\to F_{k-1}/\!/ G.$$ It is well known that a point $(x_1,\cdots,x_n)$ in $F_0=(\mathbb{P}^1)^n$ is stable (resp. semistable) if $\ge \lfloor \frac{n}2\rfloor$ points (resp. $> \lfloor \frac{n}2\rfloor$ points) do not coincide (\cite{MFK,K1}).
Let us first consider the case where $n$ is odd. In this case, $F_0^s=F_0^{ss}$ because $\frac{n}2$ is not an integer. Hence $F_k^s=F_k^{ss}$ for any $k$ by \eqref{2010eq2}. Since the blow-up centers of $\psi_k$ for $k\le m+1$ lie in the unstable part, we have $F_k^s=F_0^s$ for $k\le m+1$. Furthermore, the stabilizer group of every point in $F_k^s$ is $\{\pm 1\}$, i.e. $\bar G=PGL(2)$ acts freely on $F_k^s$ for $0\le k\le n-2$ and thus $F_k/\!/ G=F_k^s/G$ is nonsingular. By the stability conditions, forgetting the degree 1 morphism $f:C\to \mathbb{P}^1$ gives us an invariant morphism $F_{n-m+k}^s\to \overline{M}_{0,n\cdot \epsilon_k} $ which induces a morphism $$\phi_k:F_{n-m+k}/\!/ G\to \overline{M}_{0,n\cdot \epsilon_k} \quad \text{for } k=0,\cdots, m-2.$$ Since both varieties are nonsingular, we can conclude that $\phi_k$ is an isomorphism by showing that the Picard numbers are identical. Since $\bar G$ acts freely on $F_{n-m+k}^s$, the quotient of the blow-up center of $\psi_{n-m+k+1}$ is again a transversal union of $\binom{n}{m-k}$ smooth varieties $\Sigma^S_{n-m+k}/\!/ G$ for a subset $S$ of $\{1,\cdots, n\}$
with $|S|=m-k$, which are isomorphic to the moduli space $\overline{M}_{0,(1,\epsilon_k,\cdots,\epsilon_k)}$ of weighted pointed stable curves with $n-m+k+1$ marked points (Remark \ref{2010rem1}). Finally we conclude that $$\varphi_k:\overline{M}_{0,n\cdot \epsilon_k} \cong F_{n-m+k}/\!/ G\mapright{\bar\psi_{n-m+k}} F_{n-m+k-1}/\!/ G\cong \overline{M}_{0,n\cdot \epsilon_{k-1}} $$ is a blow-up by using a lemma in \cite{Kirwan} which tells us that quotient and blow-up commute. (See \S\ref{sec2.2}.) It is straightforward to check that this morphism $\varphi_k$ is identical to Hassett's natural morphisms (\S\ref{sec2.1}). Note that the isomorphism $$\phi_{m-2}:\mathbb{P}^1[n]/\!/ G\mapright{\cong} \overline{M}_{0,n} $$ was obtained by Hu and Keel (\cite{HuKeel}) when $n$ is odd because $L_0$ is a \emph{typical} linearization in the sense that $F_0^{ss}=F_0^s$. The above proof of the fact that $\phi_k$ is an isomorphism in the odd $n$ case is essentially the same as Hu-Keel's. However their method does not apply to the even degree case.
\defF_{n-m+k} {F_{n-m+k} } \def\tilde{F}_{n-m+k} {\tilde{F}_{n-m+k} } \defY_{n-m+k} {Y_{n-m+k} }
The case where $n$ is even is more complicated because $F_k^{ss}\ne F_k^s$ for all $k$. Indeed, $F_m/\!/ G=\cdots =F_0/\!/ G= (\mathbb{P}^1)^n/\!/ G$ is singular with exactly $\frac12\binom{n}{m}$ singular points. But for $k\ge 1$, the GIT quotient of $F_{n-m+k}$ by $G$ is nonsingular and we can use Kirwan's partial desingularization of the GIT quotient $F_{n-m+k}/\!/ G$ (\cite{Kirwan}). For $k\ge 1$, the locus $Y_{n-m+k}$ of closed orbits in $F_{n-m+k}^{ss}-F_{n-m+k}^s$ is the disjoint union of the transversal intersections of smooth divisors $\Sigma^S_{n-m+k}$ and $\Sigma^{S^c}_{n-m+k}$ where
$S\sqcup S^c=\{1,\cdots,n\}$ is a partition with $|S|=m$. In particular, $Y_{n-m+k}$ is of codimension $2$ and the stabilizers of points in $Y_{n-m+k}$ are all conjugates of $\mathbb{C}^*$. The weights of the action of the stabilizer $\mathbb{C}^*$ on the normal space to $Y_{n-m+k}$ are $2,-2$. By Luna's slice theorem (\cite[Appendix 1.D]{MFK}), it follows that $F_{n-m+k}/\!/ G$ is smooth along the divisor $Y_{n-m+k}/\!/ G$. If we let $\tilde{F}_{n-m+k} \to F_{n-m+k} ^{ss}$ be the blow-up of $F_{n-m+k} ^{ss}$ along $Y_{n-m+k} $, $\tilde{F}_{n-m+k} ^{ss}=\tilde{F}_{n-m+k} ^s$ and $\tilde{F}_{n-m+k} /\!/ G=\tilde{F}_{n-m+k} ^s/G$ is nonsingular. Since blow-up and quotient commute (\S\ref{sec2.2}), the induced map $$\tilde{F}_{n-m+k} /\!/ G\to F_{n-m+k} /\!/ G$$ is a blow-up along $Y_{n-m+k} /\!/ G$ which has to be an isomorphism because the blow-up center is already a smooth divisor. So we can use $\tilde{F}_{n-m+k} ^s$ instead of $F_{n-m+k} ^{ss}$ and apply the same line of arguments as in the odd degree case. In this way, we can establish Theorem \ref{thm1.1}.
To deduce Theorem \ref{thm1.2} from Theorem \ref{thm1.1}, we note that by \cite[Corollary 3.5]{Simpson}, it suffices to prove that $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ is ample for $\frac{2}{m-k+2}<\alpha\le \frac{2}{m-k+1}$ where $D_k=\overline{M}_{0,n\cdot \epsilon_k} -M_{0,n}$ is the boundary divisor of $\overline{M}_{0,n\cdot \epsilon_k} $ (Proposition \ref{prop-amplerange}). By the intersection number calculations of Alexeev and Swinarski (\cite[\S3]{AlexSwin}), we obtain the nefness of $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ for $\alpha= \frac{2}{m-k+1}+s$ for some (sufficiently small) positive number $s$. Because any positive linear combination of an ample divisor and a nef divisor is ample, it suffices to show that $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ is ample for $\alpha=\frac{2}{m-k+2}+t$ for \emph{any} sufficiently small $t>0$. We use induction on $k$. By calculating the canonical divisor explicitly, it is easy to show when $k=0$. Because $\varphi_k$ is a blow-up with exceptional divisor $D^{m-k+1}_k$, $\varphi_k^*(K_{\overline{M}_{0,n\cdot \epsilon_{k-1}} }+\alpha D_{k-1})-\delta D^{m-k+1}_k$ is ample for small $\delta>0$ if $K_{\overline{M}_{0,n\cdot \epsilon_{k-1}} }+\alpha D_{k-1}$ is ample. By a direct calculation, we find that these ample divisors give us $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ with $\alpha=\frac{2}{m-k+2}+t$ for any sufficiently small $t>0$. So we obtain a proof of Theorem \ref{thm1.2}.
For the moduli spaces of \emph{unordered} weighted pointed stable curves \[\widetilde{M}_{0,n\cdot\epsilon_k}=\overline{M}_{0,n\cdot \epsilon_k} /S_n\] we can simply take the $S_n$ quotient of our sequence \eqref{eq1} and thus $\widetilde{M}_{0,n\cdot\epsilon_k}$ can be constructed by a sequence of \emph{weighted blow-ups} from $\mathbb{P}^n/\!/ G=\left((\mathbb{P}^1)^n/\!/ G\right)/S_n$. In particular, $\widetilde{M}_{0,n\cdot\epsilon_1}$ is a weighted blow-up of $\mathbb{P}^n/\!/ G$ at its singular point when $n$ is even.
Here is an outline of this paper. In \S2, we recall necessary materials about the moduli spaces $\overline{M}_{0,n\cdot \epsilon_k} $ of weighted pointed stable curves, partial desingularization and blow-up along transversal center. In \S3, we recall the blow-up construction of the moduli space $F_k$ of weighted pointed parameterized stable curves. In \S4, we prove Theorem \ref{thm1.1}. In \S5, we prove Theorem \ref{thm1.2}. In \S6, we give a basis of the Picard group of $\overline{M}_{0,n\cdot \epsilon_k} $ as an application of Theorem \ref{thm1.1}.
\textbf{Acknowledgement.} This paper grew out of our effort to prove a conjecture of Brendan Hassett (passed to us by David Donghoon Hyeon): When $n$ is even, $\widetilde{M}_{0,n\cdot\epsilon_1}$ is the (weighted) blow-up of $\mathbb{P}^n/\!/ G$ at the singular point. It is our pleasure to thank Donghoon Hyeon for useful discussions. We are also grateful to David Smyth who kindly pointed out an error in a previous draft.
\section{Preliminaries}\label{sec2}
\subsection{Moduli of weighted pointed stable curves}\label{sec2.1} We recall the definitions and basic facts on Hassett's moduli spaces of weighted pointed stable curves from \cite{Hassett}.
A family of nodal curves of genus $g$ with $n$ marked points over base scheme $B$ consists of \begin{enumerate} \item a flat proper morphism $\pi : C \to B$ whose geometric fibers are nodal connected curves of arithmetic genus $g$ and \item sections $s_1, s_2, \cdots, s_n$ of $\pi$. \end{enumerate} An $n$-tuple $\mathcal{A} =(a_1, a_2, \cdots, a_n) \in \mathbb{Q}^n$ with $0 < a_i \le 1$ assigns a weight $a_i$ to the $i$-th marked point. Suppose that $2g-2+a_1+a_2+\cdots+a_n > 0$.
\begin{definition}\cite[\S 2]{Hassett} A family of nodal curves of genus $g$ with $n$ marked points $(C, s_1, \cdots, s_n) \stackrel{\pi}{\to} B$ is stable of type $(g, \mathcal{A})$ if \begin{enumerate} \item the sections $s_1, \cdots, s_n$ lie in the smooth locus of $\pi$; \item for any subset $\{s_{i_1}, \cdots, s_{i_r}\}$ of nonempty intersection, $a_{i_1} + \cdots + a_{i_r} \le 1$; \item $K_{\pi} + a_1s_1 + a_2s_2+\cdots+a_ns_n$ is $\pi$-relatively ample. \end{enumerate} \end{definition}
\begin{theorem}\cite[Theorem 2.1]{Hassett} There exists a connected Deligne-Mumford stack $\overline{\mathcal{M}}_{g, \mathcal{A}}$, smooth and proper over $\mathbb{Z}$, representing the moduli functor of weighted pointed stable curves of type $(g, \mathcal{A})$. The corresponding coarse moduli scheme $\overline{M}_{g, \mathcal{A}}$ is projective over $\mathbb{Z}$. \end{theorem}
When $g = 0$, there is no nontrivial automorphism for any weighted pointed stable curve and hence $\overline{M}_{0, \mathcal{A}}$ is a projective \emph{smooth variety} for any $\mathcal{A} $.
There are natural morphisms between moduli spaces with different weight data. Let $\mathcal{A}=\{a_1, \cdots, a_n\}$, $\mathcal{B}=\{b_1, \cdots, b_n\}$ be two weight data and suppose $a_i \ge b_i$ for all $1 \le i \le n$. Then there exists a birational \emph{reduction} morphism \[
\varphi_{\mathcal{A}, \mathcal{B}}
: \overline{\mathcal{M}}_{g, \mathcal{A}} \to
\overline{\mathcal{M}}_{g, \mathcal{B}}. \] For $(C, s_1, \cdots, s_n) \in \overline{\mathcal{M}}_{g, \mathcal{A}}$, $\varphi_{\mathcal{A}, \mathcal{B}}(C, s_1, \cdots, s_n)$ is obtained by collapsing components of $C$ on which $\omega_C + b_1s_1+ \cdots + b_ns_n$ fails to be ample. These morphisms between moduli stacks induce corresponding morphisms between coarse moduli schemes.
The exceptional locus of the reduction morphism $\varphi_{\mathcal{A}, \mathcal{B}}$ consists of boundary divisors $D_{I, I^c}$ where $I = \{i_1, \cdots, i_r\}$ and $I^c=\{j_1, \cdots, j_{n-r}\}$ form a partition of $\{1, \cdots, n\}$ satisfying $r > 2$, $$a_{i_1} + \cdots + a_{i_r} > 1 \quad \text{and}\quad b_{i_1} + \cdots + b_{i_r} \le 1.$$ Here $D_{I, I^c}$ denotes the closure of the locus of $(C, s_1, \cdots, s_n)$ where $C$ has two irreducible components $C_1, C_2$ with $p_a(C_1) = 0$, $p_a(C_2) = g$, $r$ sections $s_{i_1}, \cdots s_{i_r}$ lying on $C_1$, and the other $n-r$ sections lying on $C_2$.
\begin{proposition}\cite[Proposition 4.5]{Hassett}\label{reduction} The boundary divisor $D_{I,I^c}$ is isomorphic to $ \overline{M}_{0, \mathcal{A}'_I} \times \overline{M}_{g, \mathcal{A}'_{I^c}},$ with $ \mathcal{A}'_I = (a_{i_1}, \cdots, a_{i_r}, 1) $ and $\mathcal{A}'_{I^c}= (a_{j_1}, \cdots, a_{j_{n-r}}, 1).$ Furthermore, $
\varphi_{\mathcal{A}, \mathcal{B}}(D_{I, I^c})
\cong \overline{M}_{g, \mathcal{B}'_{I^c}}$ with $ \mathcal{B}'_{I^c} = (b_{j_1}, \cdots, b_{j_{n-r}}, \sum_{k=1}^r b_{i_k}).$ \end{proposition}
From now on, we focus on the $g=0$ case. Let $$m = \lfloor \frac{n}{2}\rfloor,\quad \frac{1}{m-k+1} < \epsilon_k \le \frac{1}{m-k}\quad \text{and}\quad n\cdot \epsilon_k = (\epsilon_k, \cdots, \epsilon_k).$$ Consider the reduction morphism \[
\varphi_{n\cdot \epsilon_{k}, n \cdot \epsilon_{k-1}}
: \overline{M}_{0, n \cdot \epsilon_{k}} \to
\overline{M}_{0, n \cdot \epsilon_{k-1}}. \]
Then $D_{I, I^c}$ is contracted by $\varphi_{n\cdot \epsilon_{k}, n \cdot \epsilon_{k-1}}$ if and only if $|I| = m - k + 1$. Certainly, there are ${n} \choose {m-k+1}$ such partitions $I\sqcup I^c$ of $\{1,\cdots,n\}$.
For two subsets $I, J \subset \{1, \cdots, n\}$ such that $|I| =
|J| = m - k + 1$, $D_{I, I^c} \cap D_{J, J^c}$ has codimension at least two in $\overline{M}_{0,n\cdot \epsilon_k} $. So if we denote the complement of the intersections of the divisors by \[
\overline{M}_{0,n\cdot \epsilon_k} ' = \overline{M}_{0,n\cdot \epsilon_k} - \bigcup_{|I| = |J| = n-k+1, I \ne J} D_{I, I^c} \cap D_{J,
J^c}, \] we have $\mathrm{Pic}(\overline{M}_{0,n\cdot \epsilon_k} ') = \mathrm{Pic}(\overline{M}_{0,n\cdot \epsilon_k} )$. The restriction of $\varphi_{n \cdot \epsilon_{k}, n \cdot \epsilon_{k-1}}$ to $\overline{M}_{0,n\cdot \epsilon_k} '$ is a contraction of ${n} \choose {m-k+1}$ \emph{disjoint} divisors and its image is an open subset whose complement has codimension at least two. Therefore we obtain the following equality of Picard numbers: \begin{equation}\label{eq-eqPicNumber}
\rho(\overline{M}_{0, n \cdot \epsilon_{k}}) =
\rho(\overline{M}_{0, n \cdot \epsilon_{k-1}}) + {n \choose {m-k+1}}. \end{equation}
It is well known that the Picard number of $\overline{M}_{0,n} $ is \begin{equation}\label{eq-eqPic2} \rho(\overline{M}_{0,n} )=\rho(\overline{M}_{0, n \cdot \epsilon_{m-2}})=2^{n-1}-\binom{n}{2}-1. \end{equation} Hence we obtain the following lemma from \eqref{eq-eqPicNumber} and \eqref{eq-eqPic2}. \begin{lemma}\label{lem-PicNumMze}\begin{enumerate}\item If $n$ is odd, $\rho(\overline{M}_{0,n\cdot \epsilon_k} )= n+\sum_{i=1}^k\binom{n}{m-i+1}$.\item If $n$ is even, $\rho(\overline{M}_{0,n\cdot \epsilon_k} )=n+\frac12\binom{n}{m}+\sum_{i=2}^k\binom{n}{m-i+1}$.\end{enumerate} \end{lemma}
\subsection{Partial desingularization}\label{sec2.2} We recall a few results from \cite{Kirwan, Hu} on change of stability in a blow-up.
Let $G$ be a complex reductive group acting on a projective nonsingular variety $X$. Let $L$ be a $G$-linearized ample line bundle on $X$. Let $Y$ be a $G$-invariant closed subvariety of $X$, and let $\pi : \widetilde{X} \to X$ be the blow-up of $X$ along $Y$, with exceptional divisor $E$. Then for sufficiently large $d$, $L_d = \pi^*L^d \otimes \mathcal{O} (-E)$ becomes very ample, and there is a natural lifting of the $G$-action to $L_d$ ( \cite[\S3]{Kirwan}).
Let $X^{ss}$(resp. $X^s$) denote the semistable (resp. stable) part of $X$. With respect to the polarizations $L$ and $L_d$, the following hold (\cite[\S3]{Kirwan} or \cite[Theorem 3.11]{Hu}) :\begin{equation}\label{eq-StabBlowup} \widetilde{X}^{ss} \subset \pi^{-1}(X^{ss}), \qquad \widetilde{X}^{s} \supset \pi^{-1}(X^{s}).\end{equation} In particular, if $X^{ss} = X^s$, then $\widetilde{X}^{ss} = \widetilde{X}^s = \pi^{-1}(X^s)$.
For the next lemma, let us suppose $Y^{ss}=Y\cap X^{ss}$ is nonsingular. We can compare the GIT quotient of $\widetilde{X}$ by $G$ with respect to $L_d$ with the quotient of $X$ by $G$ with respect to $L$. \begin{lemma}\cite[Lemma 3.11]{Kirwan} \label{blowupGIT} For sufficiently large $d$, $\widetilde{X}/\!/ G$ is the blow-up of $X/\!/ G$ along the image $Y /\!/ G$ of $Y^{ss}$. \end{lemma}
Let $\mathcal{I}$ be the ideal sheaf of $Y$. In the statement of Lemma \ref{blowupGIT}, the blow-up is defined by the ideal sheaf $(\mathcal{I}^m)_G$ which is the $G$-invariant part of $\mathcal{I}^m$, for some $m$. (See the proof of \cite[Lemma 3.11]{Kirwan}.) In the cases considered in this paper, the blow-ups always take place along \emph{reduced} ideals, i.e. $\widetilde{X}/\!/ G$ is the blow-up of $X/\!/ G$ along the subvariety $Y/\!/ G$ because of the following. \begin{lemma}\label{lem-specialcasebl} Let $G=SL(2)$ and $\mathbb{C}^*$ be the maximal torus of $G$. Suppose $Y^{ss}$ is smooth. The blow-up $\widetilde{X}/\!/ G\to X/\!/ G$ is the blow-up of the reduced ideal of $Y/\!/ G$ if any of the following holds: \begin{enumerate} \item The stabilizers of points in $X^{ss}$ are all equal to the center $\{\pm 1\}$, i.e. $\bar G=SL(2)/\{\pm 1\}$ acts on $X^{ss}$ freely. \item If we denote the $\mathbb{C}^*$-fixed locus in $X^{ss}$ by $Z^{ss}_{\mathbb{C}^*}$, $Y^{ss}=Y\cap X^{ss}=GZ^{ss}_{\mathbb{C}^*}$ and the stabilizers of points in $X^{ss}-Y^{ss}$ are all $\{\pm 1\}$. Furthermore suppose that the weights of the action of $\mathbb{C}^*$ on the normal space of $Y^{ss}$ at any $y\in Z^{ss}_{\mathbb{C}^*}$ are $\pm l$ for some $l\ge 1$. \item There exists a smooth divisor $W$ of $X^{ss}$ which intersects transversely with $Y^{ss}$ such that the stabilizers of points in $X^{ss}-W$ are all $\mathbb{Z}_2=\{\pm 1\}$ and the stabilizers of points in $W$ are all isomorphic to $\mathbb{Z}_4$. \end{enumerate} In the cases (1) and (3), $Y/\!/ G=Y^s/G$ and $X/\!/ G=X^s/G$ are nonsingular and the morphism $\widetilde{X}/\!/ G\to X/\!/ G$ is the smooth blow-up along the smooth subvariety $Y/\!/ G$. \end{lemma} \begin{proof} Let us consider the first case. Let $\bar G=PGL(2)$. By Luna's \'etale slice theorem \cite[Appendix 1.D]{MFK}, \'etale locally near a point in $Y^{ss}$, $X^{ss}$ is $\bar{G}\times S$ and $Y^{ss}$ is $\bar{G}\times S^Y$ for some nonsingular locally closed subvariety $S$ and $S^Y=S\cap Y$. Then \'etale locally $\widetilde{X}^{ss}$ is $\bar{G}\times \mathrm{bl}_{S^Y}S$ where $\mathrm{bl}_{S^Y}S$ denotes the blow-up of $S$ along the nonsingular variety $S^Y$. Thus the quotients $X/\!/ G$, $Y/\!/ G$ and $\widetilde{X}/\!/ G$ are \'etale locally $S$, $S^Y$ and $\mathrm{bl}_{S^Y}S$ respectively. This implies that the blow-up $\widetilde{X}/\!/ G\to X/\!/ G$ is the smooth blow-up along the reduced ideal of $Y/\!/ G$.
For the second case, note that the orbits in $Y^{ss}$ are closed in $X^{ss}$ because the stabilizers are maximal. So we can again use Luna's slice theorem to see that \'etale locally near a point $y$ in $Y^{ss}$, the varieties $X^{ss}$, $Y^{ss}$ and $\widetilde{X}$ are respectively $G\times_{\mathbb{C}^*}S$, $G\times_{\mathbb{C}^*}S^0$ and $G\times_{\mathbb{C}^*}\mathrm{bl}_{S^0}S$ for some nonsingular locally closed $\mathbb{C}^*$-equivariant subvariety $S$ and its $\mathbb{C}^*$-fixed locus $S^0$. Therefore the quotients $X/\!/ G$, $Y/\!/ G$ and $\widetilde{X}/\!/ G$ are \'etale locally $S/\!/ \mathbb{C}^*$, $S^0$ and $(\mathrm{bl}_{S^0}S)/\!/ \mathbb{C}^*$. Thus it suffices to show $$(\mathrm{bl}_{S^0}S)/\!/ \mathbb{C}^*\cong \mathrm{bl}_{S^0}(S/\!/ \mathbb{C}^*).$$ Since $X$ is smooth, \'etale locally we can choose our $S$ to be the normal space to the orbit of $y$ and $S$ is decomposed into the weight spaces $S^0\oplus S^+\oplus S^-$. As the action of $\mathbb{C}^*$ extends to $SL(2)$, the nonzero weights are $\pm l$ by assumption. If we choose coordinates $x_1,\cdots, x_r$ for $S^+$ and $y_1,\cdots, y_s$ for $S^-$, the invariants are polynomials of $x_iy_j$ and thus $(I^{2m})_{\mathbb{C}^*}=(I_{\mathbb{C}^*})^m$ for $m\ge 1$ where $I=\langle x_1,\cdots,x_r,y_1,\cdots,y_s \rangle$ is the ideal of $S^0$. By \cite[II Exe. 7.11]{Hartshorne}, we have $$\mathrm{bl}_{S^0}S=\mathrm{Proj}_S(\oplus_m I^m)\cong \mathrm{Proj}_S(\oplus_m I^{2m})$$ and thus $$(\mathrm{bl}_{S^0}S)/\!/ \mathbb{C}^* =\mathrm{Proj}_{S/\!/ \mathbb{C}^*}(\oplus_m I^{2m})_{\mathbb{C}^*} =\mathrm{Proj}_{S/\!/ \mathbb{C}^*}\left(\oplus_m (I_{\mathbb{C}^*})^{m}\right)=\mathrm{bl}_{I_{\mathbb{C}^*}}(S/\!/ \mathbb{C}^*).$$ Since $S$ is factorial and $I$ is reduced, $I_{\mathbb{C}^*}$ is reduced. (If $f^m\in I_{\mathbb{C}^*}$, then $f\in I$ and $(g\cdot f)^m=f^m$ for $g\in \mathbb{C}^*$. By factoriality, $g\cdot f$ may differ from $f$ only by a constant multiple, which must be an $m$-th root of unity. Because $\mathbb{C}^*$ is connected, the constant must be $1$ and hence $f\in I_{\mathbb{C}^*}$.) Therefore $I_{\mathbb{C}^*}$ is the reduced ideal of $S^0$ on $S/\!/ \mathbb{C}^*$ and hence $(\mathrm{bl}_{S^0}S)/\!/ \mathbb{C}^*\cong \mathrm{bl}_{S^0}(S/\!/ \mathbb{C}^*)$ as desired.
The last case is similar to the first case. Near a point in $W$, $X^{ss}$ is \'etale locally $\bar G\times_{\mathbb{Z}_2}S$ where $S=S_W\times\mathbb{C}$ for some smooth variety $S_W$. $\mathbb{Z}_2$ acts trivially on $S_W$ and by $\pm 1$ on $\mathbb{C}$. Etale locally $Y^{ss}$ is $\bar G\times_{\mathbb{Z}_2}S_Y$ where $S_Y=(S_W\cap Y)\times \mathbb{C}$. The quotients $X/\!/ G$, $Y/\!/ G$ and $\widetilde{X}/\!/ G$ are \'etale locally $S_W\times \mathbb{C}$, $(S_W\cap Y)\times \mathbb{C}$ and $\mathrm{bl}_{S_W\cap Y}S_W\times \mathbb{C}$. This proves our lemma. \end{proof} \begin{corollary}\label{cor-blcomquot} Suppose that (1) of Lemma \ref{lem-specialcasebl} holds. If $Y^{ss}=Y_1^{ss}\cup\cdots\cup Y_r^{ss}$ is a transversal union of smooth subvarieties of $X^{ss}$ and if $\widetilde{X}$ is the blow-up of $X^{ss}$ along $Y^{ss}$, then $\widetilde{X}/\!/ G$ is the blow-up of $X/\!/ G$ along the reduced ideal of $Y/\!/ G$ which is again a transversal union of smooth varieties $Y_i/\!/ G$. The same holds under the condition (3) of Lemma \ref{lem-specialcasebl} if furthermore $Y_i$ are transversal to $W$. \end{corollary} \begin{proof} Because of the assumption (1), $X^{ss}=X^s.$ If $Y^{ss}=Y_1^{ss}\cup\cdots\cup Y_r^{ss}$ is a transversal union of smooth subvarieties of $X^{ss}$ and if $\pi:\widetilde{X}\to X^{ss}$ is the blow-up along $Y^{ss}$, then $\widetilde{X}^s=\widetilde{X}^{ss}=\pi^{-1}(X^s)$ is the composition of smooth blow-ups along (the proper transforms of) the irreducible components $Y_i^{ss}$ by Proposition \ref{prop-bltrcen} below. For each of the smooth blow-ups, the quotient of the blown-up space is the blow-up of the quotient along the reduced ideal of the quotient of the center by Lemma \ref{lem-specialcasebl}. Hence $\widetilde{X}/\!/ G\to X/\!/ G$ is the composition of smooth blow-ups along irreducible smooth subvarieties which are proper transforms of $Y_i/\!/ G$. Hence $\widetilde{X}/\!/ G$ is the blow-up along the union $Y/\!/ G$ of $Y_i/\!/ G$ by Proposition \ref{prop-bltrcen} again.
The case (3) of Lemma \ref{lem-specialcasebl} is similar and we omit the detail. \end{proof}
Finally we recall Kirwan's partial desingularization construction of GIT quotients. Suppose $X^{ss} \ne X^s$ and $X^s$ is nonempty. Kirwan in \cite{Kirwan} introduced a systematic way of blowing up $X^{ss}$ along a sequence of nonsingular subvarieties to obtain a variety $\widetilde{X}$ with linearized $G$ action such that $\widetilde{X}^{ss} = \widetilde{X}^s$ and $\widetilde{X}/\!/ G$ has at worst finite quotient singularities only, as follows:\begin{enumerate} \item Find a maximal dimensional connected reductive subgroup $R$ such that the $R$-fixed locus $Z_R^{ss}$ in $X^{ss}$ is nonempty. Then $$GZ_R^{ss}\cong G\times_{N^R}Z_R^{ss}$$ is a nonsingular closed subvariety of $X^{ss}$ where $N^R$ denotes the normalizer of $R$ in $G$. \item Blow up $X^{ss}$ along $GZ_R^{ss}$ and find the semistable part $X_1^{ss}$. Go back to step 1 and repeat this precess until there are no more strictly semistable points. \end{enumerate} Kirwan proves that this process stops in finite steps and $\widetilde{X}/\!/ G$ is called the \emph{partial desingularization} of $X /\!/ G$. We will drop ``partial" if it is nonsingular.
\subsection{Blow-up along transversal center}\label{sec2.3} We show that the blow-up along a center whose irreducible components are transversal smooth varieties is isomorphic to the result of smooth blow-ups along the irreducible components in any order. This fact can be directly proved but instead we will see that it is an easy special case of beautiful results of L. Li in \cite{Li}.
\begin{definition} \cite[\S1]{Li} (1) For a nonsingular algebraic variety $X$, an \emph{arrangement} of subvarieties $S$ is a finite collection of nonsingular subvarieties such that all nonempty scheme-theoretic intersections of subvarieties in $S$ are again in $S$.
(2) For an arrangement $S$, a subset $B\subset S$ is called a \emph{building set} of $S$ if for any $s \in S- B$, the minimal elements in $\{b \in B : b \supset s\}$ intersect transversally and the intersection is $s$.
(3) A set of subvarieties $B$ is called a \emph{building set} if all the possible intersections of subvarieties in $B$ form an arrangement $S$ (called the induced arrangement of $B$) and $B$ is a building set of $S$. \end{definition}
The \emph{wonderful compactification} $X_B$ of $X^0=X-\cup_{b\in B} b$ is defined as the closure of $X^0$ in $\prod_{b\in B}\mathrm{bl}_bX$. Li then proves the following. \begin{theorem}\cite[Theorem 1.3]{Li} \label{thm-Li1} Let $X$ be a nonsingular variety and $B = \{b_1, \cdots,b_n\}$ be a nonempty building set of subvarieties of $X$. Let $I_i$ be the ideal sheaf of $b_i \in B$. \begin{enumerate}\item The wonderful compactification $X_B$ is isomorphic to the blow-up of $X$ along the ideal sheaf $I_1I_2\cdots I_n$. \item If we arrange $B = \{b_1, \cdots,b_n\}$ in such an order that the first $i$ terms $b_1,\cdots,b_i$ form a building set for any $1\le i\le n$, then $X_B = \mathrm{bl}_{\tilde{b}_n} \cdots \mathrm{bl}_{\tilde{b}_2} \mathrm{bl}_{b_1} X$, where each blow-up is along a nonsingular subvariety $\tilde{b}_i$.\end{enumerate} \end{theorem} Here $\tilde{b}_i$ is the \emph{dominant transform} of $b_i$ which is obtained by taking the proper transform when it doesn't lie in the blow-up center or the inverse image if it lies in the center, in each blow-up. (See \cite[Definition 2.7]{Li}.)
Let $X$ be a smooth variety and let $Y_1, \cdots, Y_n$ be transversally intersecting smooth closed subvarieties. Here, \emph{transversal intersection} means that for any nonempty $S \subset \{1, \cdots, n\}$ the intersection $Y_S:=\cap_{i \in S}Y_i$ is smooth and the normal bundle $N_{Y_S/X}$ in $X$ of $Y_S$ is the direct sum of the restrictions of the normal bundles $N_{Y_i/X}$ in $X$ of $Y_i$, i.e.
$$N_{Y_S/X} = \bigoplus_{i\in S} N_{Y_i/X}|_{Y_S}.$$ If we denote the ideal of $Y_i$ by $I_i$, the ideal of the union $\cup_{i=1}^n Y_i$ is the product $I_1I_2\cdots I_n$. Moreover for any permutation $\tau\in S_n$ and $1\le i\le n$, $B=\{Y_{\tau(1)},\cdots,Y_{\tau(i)}\}$ is clearly a building set. By Theorem \ref{thm-Li1} we obtain the following. \begin{proposition}\label{prop-bltrcen} Let $Y=Y_1\cup\cdots \cup Y_n$ be a union of transversally intersecting smooth subvarieties of a smooth variety $X$. Then the blow-up of $X$ along $Y$ is isomorphic to \[ \mathrm{bl}_{\tilde Y_{\tau(n)}}\cdots \mathrm{bl}_{\tilde Y_{\tau(2)}}\mathrm{bl}_{Y_{\tau(1)}} X \] for any permutation $\tau\in S_n$ where $\tilde{Y}_i$ denotes the proper transform of $Y_i$. \end{proposition}
\subsection{Log canonical model} Let $X$ be a normal projective variety and $D = \sum a_i D_i$ be a rational linear combination of prime divisors of $X$ with $0 < a_i \le 1$. A \emph{log resolution} of $(X, D)$ is a birational morphism $\pi : Y \to X$ from a smooth projective variety $Y$ to $X$ such that $\pi^{-1}(D_i)$ and the exceptional divisors $E_i$ of $\pi$ are simple normal crossing divisors on $Y$. Then the discrepancy formula \[
K_Y + \pi^{-1}_* (D) \equiv
\pi^*(K_X + D) + \sum_{E_i : \mbox{exceptional}} a(E_i, X, D)E_i, \] defines the \emph{discrepancy} of $(X, D)$ by \[
\mathrm{discrep}(X,D) := \inf \{ a(E, X, D) : E : \mbox{exceptional}\}. \]
Let $(X, D)$ be a pair where $X$ is a normal projective variety and $D = \sum a_i D_i$ be a rational linear combination of prime divisors with $0 < a_i \le 1$. Suppose that $K_X + D$ is $\mathbb{Q}$-Cartier. A pair $(X, D)$ is \emph{log canonical (abbrev. lc)} if $\mathrm{discrep}(X,D) \ge -1$ and \emph{Kawamata log terminal (abbrev. klt)} if $\mathrm{discrep}(X,D) > -1$ and $\lfloor D \rfloor \le 0$.
When $X$ is smooth and $D$ is a normal crossing effective divisor, $(X, D)$ is always lc and is klt if all $a_i < 1$.
\begin{definition} For lc pair $(X, D)$, the \emph{canonical ring} is \[
R(X, K_X + D) := \oplus_{l \ge 0} H^0(X, \mathcal{O} _X(\lfloor l (K_X + D) \rfloor)) \] and the \emph{log canonical model} is \[
\mathrm{Proj}\; R(X, K_X + D). \] \end{definition} In \cite{BCHM}, Birkar, Cascini, Hacon and McKernan proved that for any klt pair $(X, D)$, the canonical ring is finitely generated, so the log canonical model always exists.
\section{Moduli of weighted parameterized stable curves}\label{sec3}
Let $X$ be a smooth projective variety. In this section, we decompose the map $$X[n]\to X^n$$ defined by Fulton and MacPherson (\cite{FM}) into a \emph{symmetric} sequence of blow-ups along transversal centers. A. Mustata and M. Mustata already considered this problem in their search for intermediate moduli spaces for the stable map spaces in \cite[\S1]{Mustata}. Let us recall their construction.
\noindent \textbf{Stage 0}: Let $F_0=X^n$ and $\Gamma_0=X^n\times X$. For a subset $S$ of $\{1,2,\cdots,n\}$, we let \[
\Sigma^S_0=\{(x_1,\cdots,x_n)\in X^n\,|\, x_i=x_j \text{ if }
i,j\in S\}, \quad \Sigma^k_0=\cup_{|S|=k}\Sigma_0^S \] and let $\sigma^i_0\subset \Gamma_0$ be the graph of the $i$-th projection $X^n\to X$. Then $\Sigma_0^n\cong X$ is a smooth subvariety of $F_0$. For each $S$, fix any $i_S\in S$.
\noindent \textbf{Stage $1$}: Let $F_1$ be the blow-up of $F_0$ along $\Sigma_0^n$. Let $\Sigma_1^n$ be the exceptional divisor and $\Sigma_1^S$ be the proper transform of $\Sigma_0^S$ for
$|S|\ne n$. Let us define $\Gamma_1$ as the blow-up of $F_1\times_{F_0}\Gamma_0$ along $\Sigma^n_1\times_{F_0}\sigma^1_0$ so that we have a flat family \[ \Gamma_1\to F_1\times_{F_0}\Gamma_0 \to F_1 \] of varieties over $F_1$. Let $\sigma_1^i$ be the proper transform of $\sigma_0^i$ in $\Gamma_1$. Note that $\Sigma^S_{1}$ for
$|S|=n-1$ are all disjoint smooth varieties of same dimension.
\noindent \textbf{Stage $2$}: Let $F_2$ be the blow-up of $F_1$
along $\Sigma_1^{n-1}=\sum_{|S|=n-1}\Sigma_1^S$. Let $\Sigma_2^S$
be the exceptional divisor lying over $\Sigma_1^S$ if $|S|=n-1$ and $\Sigma_2^S$ be the proper transform of $\Sigma_1^S$ for
$|S|\ne n-1$. Let us define $\Gamma_2$ as the blow-up of $F_2\times_{F_1}\Gamma_1$ along the disjoint union of
$\Sigma^S_2\times_{F_1}\sigma^{i_S}_1$ for all $S$ with $|S|=n-1$ so that we have a flat family \[ \Gamma_2\to F_2\times_{F_1}\Gamma_1 \to F_2 \] of varieties over $F_2$. Let $\sigma_2^i$ be the proper transform of $\sigma_1^i$ in $\Gamma_2$. Note that $\Sigma^S_{2}$ for
$|S|=n-2$ in $F_2$ are all transversal smooth varieties of same dimension. Hence the blow-up of $F_2$ along their union is smooth by \S\ref{sec2.3}.
We can continue this way until we reach the last stage.
\noindent \textbf{Stage $n-1$}: Let $F_{n-1}$ be the blow-up of
$F_{n-2}$ along $\Sigma_{n-2}^2=\sum_{|S|=2}\Sigma_{n-2}^S$. Let $\Sigma_{n-1}^S$ be the exceptional divisor lying over
$\Sigma_{n-2}^S$ if $|S|=2$ and $\Sigma_{n-1}^S$ be the proper transform of $\Sigma_{n-2}^S$ for $|S|\ne 2$. Let us define $\Gamma_{n-1}$ as the blow-up of $F_{n-1}\times_{F_{n-2}}\Gamma_{n-2}$ along the disjoint union of $\Sigma^S_{n-1}\times_{F_{n-2}}\sigma^{i_S}_{n-2}$ for all $S$
with $|S|=2$ so that we have a flat family \[ \Gamma_{n-1}\to F_{n-1}\times_{F_{n-2}}\Gamma_{n-2} \to F_{n-1} \] of varieties over $F_{n-1}$. Let $\sigma_{n-1}^i$ be the proper transform of $\sigma_{n-2}^i$ in $\Gamma_{n-1}$.
Nonsingularity of the blown-up spaces $F_k$ are guaranteed by the following.
\begin{lemma}\label{lem3-1}
$\Sigma^S_{k}$ for $|S|\ge n-k$ are transversal in $F_{k}$ i.e. the normal bundle in $F_{k}$ of the intersection
$\cap_i\Sigma^{S_i}_k$ for distinct $S_i$ with $|S_i|\ge n-k$ is the direct sum of the restriction of the normal bundles in $F_k$ of $\Sigma^{S_i}_k$. \end{lemma} \begin{proof} This is a special case of the inductive construction of the wonderful compactification in \cite{Li}. (See \S \ref{sec2.3}.) In our situation, the building set is the set of all diagonals $B_0 =
\{\Sigma_0^S | S \subset \{1, 2, \cdots, n\}\}$. By \cite[Proposition 2.8]{Li}, $B_k=\{\Sigma_k^S\}$ is a building set of an arrangement in $F_k$ and hence the desired transversality follows. \end{proof}
By construction, $F_k$ are all smooth and $\Gamma_k\to F_k$ are equipped with $n$ sections $\sigma_k^i$. When $\dim X=1$, $\Sigma^2_{n-2}$ is a divisor and thus $F_{n-1}=F_{n-2}$. In \cite[Proposition 1.8]{Mustata}, Mustata and Mustata prove that the varieties $F_k$ are fine moduli spaces for some moduli functors as follows.
\begin{definition}\label{def3.1} \cite[Definition 1.7]{Mustata} A family of \emph{$k$-stable parameterized rational curves} over $S$ consists of a flat family of curves $\pi:C\to S$, a morphism $\phi:C\to S\times \mathbb{P}^1$ of degree 1 over each geometric fiber $C_s$ of $\pi$ and $n$ marked sections $\sigma^1,\cdots, \sigma^n$ of $\pi$ such that for all $s\in S$, \begin{enumerate} \item no more than $n-k$ of the marked points $\sigma^i(s)$ in $C_s$ coincide; \item any ending irreducible curve in $C_s$, except the parameterized one, contains more than $n-k$ marked points; \item all the marked points are smooth points of the curve $C_s$; \item $C_s$ has finitely many automorphisms preserving the marked points and the map to $\mathbb{P}^1$. \end{enumerate} \end{definition}
\begin{proposition}\label{prop3.2} \cite[Proposition 1.8]{Mustata} Let $X=\mathbb{P}^1$. The smooth variety $F_k$ finely represents the functor of isomorphism classes of families of $k$-stable parameterized rational curves. In particular, $F_{n-2}=F_{n-1}$ is the Fulton-MacPherson space $\mathbb{P}^1[n]$. \end{proposition}
\section{Blow-up construction of moduli of pointed stable curves}\label{sec4}
In the previous section, we decomposed the natural map $\mathbb{P}^1[n]\to (\mathbb{P}^1)^n$ of the Fulton-MacPherson space into a sequence \begin{equation}\label{eq4-1}\xymatrix{ \mathbb{P}^1[n]\ar@{=}[r] & F_{n-2}\ar[r]^{\psi_{n-2}} & F_{n-3}\ar[r]^{\psi_{n-3}} & \cdots \ar[r]^{\psi_2} & F_1\ar[r]^{\psi_1} & F_0\ar@{=}[r] & (\mathbb{P}^1)^n }\end{equation} of blow-ups along transversal centers. By construction the morphisms above are all equivariant with respect to the action of $G=SL(2)$. For GIT stability, we use the \emph{symmetric} linearization $L_0=\mathcal{O} (1,\cdots,1)$ for $F_0$. For $F_k$ we use the linearization $L_k$ inductively defined by $L_k=\psi_k^*L_{k-1}\otimes \mathcal{O} (-\delta_kE_k)$ where $E_k$ is the exceptional divisor of $\psi_k$ and $\{\delta_k\}$ is a decreasing sequence of sufficiently small positive numbers. Let $m=\lfloor\frac{n}{2}\rfloor$. In this section, we prove the following.
\begin{theorem}\label{thm4-1} (i) The GIT quotient $F_{n-m+k}/\!/ G$ for $1\le k\le m-2$ is isomorphic to Hassett's moduli space of weighted pointed stable rational curves $\overline{M}_{0,n\cdot \epsilon_k} $ with weights $n\cdot \epsilon_k=(\epsilon_k,\cdots,\epsilon_k)$ where $\frac1{m+1-k}<\epsilon_k\le \frac1{m-k}$. The induced maps on quotients \[ \overline{M}_{0,n\cdot \epsilon_k} = F_{n-m+k}/\!/ G \to F_{n-m+k-1}/\!/ G =\overline{M}_{0,n\cdot \epsilon_{k-1}} \] are blow-ups along transversal centers for $k=2,\cdots, m-2$.
(ii) If $n$ is odd, $$F_{m+1}/\!/ G=\cdots =F_0/\!/ G=(\mathbb{P}^1)^n/\!/ G=\overline{M}_{0,n\cdot \epsilon_0} $$ and we have a sequence of blow-ups \[ \overline{M}_{0,n} =\overline{M}_{0,n\cdot \epsilon_{m-2}} \to \overline{M}_{0,n\cdot \epsilon_{m-3}} \to \cdots \to \overline{M}_{0,n\cdot \epsilon_1} \to \overline{M}_{0,n\cdot \epsilon_0} = (\mathbb{P}^1)^n/\!/ G \] whose centers are transversal unions of equidimensional smooth varieties.
(iii) If $n$ is even, $\overline{M}_{0,n\cdot \epsilon_1} $ is a desingularization of $$(\mathbb{P}^1)^n/\!/ G=F_0/\!/ G=\cdots =F_m/\!/ G,$$ obtained by blowing up $\frac12\binom{n}{m}$ singular points so that we have a sequence of blow-ups \[ \overline{M}_{0,n} =\overline{M}_{0,n\cdot \epsilon_{m-2}} \to \overline{M}_{0,n\cdot \epsilon_{m-3}} \to \cdots \to \overline{M}_{0,n\cdot \epsilon_1} \to (\mathbb{P}^1)^n/\!/ G. \] \end{theorem} \begin{remark} (1) When $n$ is even, $\overline{M}_{0,n\cdot \epsilon_0} $ is not defined because the sum of weights does not exceed $2$.
(2) When $n$ is even, $\overline{M}_{0,n\cdot \epsilon_1} $ is Kirwan's (partial) desingularization of the GIT quotient $(\mathbb{P}^1)^n/\!/ G$ with respect to the symmetric linearization $L_0=\mathcal{O} (1,\cdots,1)$. \end{remark}
Let $F_k^{ss}$ (resp. $F_k^s$) denote the semistable (resp. stable) part of $F_k$. By \eqref{eq-StabBlowup}, we have \begin{equation}\label{eq4-2} \psi_k(F_k^{ss})\subset F_{k-1}^{ss},\qquad \psi_k^{-1}(F_{k-1}^s)\subset F_k^s. \end{equation} Also recall from \cite{K1} that $x=(x_1,\cdots,x_n)\in (\mathbb{P}^1)^n$ is semistable (resp. stable) if $> \frac{n}2$ (resp. $\ge \frac{n}2$) of $x_i$'s are not allowed to coincide. In particular, when $n$ is odd, $\psi_k^{-1}(F_{k-1}^s)=F_k^s=F_k^{ss}$ for all $k$ and \begin{equation}\label{eq4-3} F_{m+1}^s=F_{m}^s=\cdots =F_0^s, \end{equation} because the blow-up centers lie in the unstable part. Therefore we have \begin{equation}\label{eq4-4} F_{m+1}/\!/ G = \cdots =F_0/\!/ G= (\mathbb{P}^1)^n/\!/ G. \end{equation} When $n$ is even, $\psi_k$ induces a morphism $F_k^{ss}\to F_{k-1}^{ss}$ and we have \begin{equation}\label{eq4-5} F_m^{ss}=F_{m-1}^{ss}=\cdots =F_0^{ss} \quad \text{and}\quad F_m/\!/ G=\cdots =F_0/\!/ G=(\mathbb{P}^1)^n/\!/ G. \end{equation}
Let us consider the case where $n$ is odd first. By forgetting the parameterization of the parameterized component of each member of family $(\Gamma_{m+k+1}\to F_{m+k+1},\sigma_{m+k+1}^i)$, we get a rational map $F_{m+k+1}\dashrightarrow \overline{M}_{0,n\cdot \epsilon_k} $ for $k=0,1,\cdots, m-2$. By the definition of the stability in \S\ref{sec2.1}, a fiber over $\xi\in F_{m+k+1}$ is not stable with respect to $n\cdot \epsilon_k=(\epsilon_k,\cdots,\epsilon_k)$ if and only if, in each irreducible component of the curve, the number $a$ of nodes and the number $b$ of marked points satisfy $b\epsilon_k+a\le 2$. Obviously this cannot happen on the (GIT) stable part $F_{m+k+1}^s$. Therefore we obtain a morphism $F_{m+k+1}^s\to \overline{M}_{0,n\cdot \epsilon_k} $. By construction this morphism is $G$-invariant and thus induces a morphism $$\phi_k:F_{m+k+1}/\!/ G\to \overline{M}_{0,n\cdot \epsilon_k} .$$ Since the stabilizer groups in $G$ of points in $F_0^s$ are all $\{ \pm 1\}$, the quotient $$\bar{\psi}_{m+k+1}:F_{m+k+1}/\!/ G\to F_{m+k}/\!/ G$$ of $\psi_{m+k+1}$ is also a blow-up along a center which consists of transversal smooth varieties by Corollary \ref{cor-blcomquot}.
Since the blow-up center has codimension $\ge 2$, the Picard number increases by $\binom{n}{m-k+1}$ for $k=1, \cdots, m-2$. Since the character group of $SL(2)$ has no free part, by the descent result in \cite{DN}, the Picard number of $F_{m+1}/\!/ G=F_0^s/G$ is the same as the Picard number of $F_0^s$ which equals the Picard number of $F_0$. Therefore $\rho(F_{m+1}/\!/ G)=n$ and the Picard number of $F_{m+k+1}/\!/ G$ is \[ n+\sum_{i=1}^{k}\binom{n}{m-i+1} \] which equals the Picard number of $\overline{M}_{0,n\cdot \epsilon_k} $ by Lemma \ref{lem-PicNumMze}. Since $\overline{M}_{0,n\cdot \epsilon_k} $ and $F_{m+k+1}/\!/ G$ are smooth and their Picard numbers coincide, we conclude that $\phi_k$ is an isomorphism as we desired. So we proved Theorem \ref{thm4-1} for odd $n$.
Now let us suppose $n$ is even. For ease of understanding, we divide our proof into several steps.
\noindent \underline{Step 1:} For $k\ge 1$, $F_{m+k}/\!/ G$ are nonsingular and isomorphic to the partial desingularizations $\tilde{F}_{m+k}/\!/ G$.
The GIT quotients $F_{m+k}/\!/ G$ may be singular because there are $\mathbb{C}^*$-fixed points in the semistable part $F_{m+k}^{ss}$. So we use Kirwan's partial desingularization of the GIT quotients $F_{m+k}/\!/ G$ (\S\ref{sec2.2}). The following lemma says that the partial desingularization process has no effect on the quotient $F_{m+k}/\!/ G$ for $k\ge 1$.
\begin{lemma}\label{lem4-3} Let $F$ be a smooth projective variety with linearized $G=SL(2)$ action and let $F^{ss}$ be the semistable part. Fix a maximal torus $\mathbb{C}^*$ in $G$. Let $Z$ be the set of $\mathbb{C}^*$-fixed points in $F^{ss}$. Suppose the stabilizers of all points in the stable part $F^{s}$ are $\{\pm 1\}$ and $Y=GZ$ is the union of all closed orbits in $F^{ss}-F^s$. Suppose that the stabilizers of points in $Z$ are precisely $\mathbb{C}^*$. Suppose further that $Y=GZ$ is of codimension $2$. Let $\tilde{F}\to F^{ss}$ be the blow-up of $F^{ss}$ along $Y$ and let $\tilde{F}^s$ be the stable part in $\tilde{F}$ with respect to a linearization as in \S\ref{sec2.2}. Finally suppose that for each $y\in Z$, the weights of the $\mathbb{C}^*$ action on the normal space to $Y$ is $\pm l$ for some $l>0$. Then $\tilde{F}/\!/ G=\tilde{F}^s/G\cong F/\!/ G$ and $F/\!/ G$ is nonsingular. \end{lemma} \begin{proof} Since $\bar G=G/\{\pm 1\}$ acts freely on $F^s$, $F^s/G$ is smooth. By assumption, $Y$ is the union of all closed orbits in $F^{ss}-F^s$ and hence $F/\!/ G-F^s/G=Y/G$. By Lemma \ref{lem-specialcasebl} (2), $\tilde{F}^s/G$ is the blow-up of $F/\!/ G$ along the reduced ideal of $Y/G$. By our assumption, $Z$ is of codimension $4$ and $$Y/G=GZ/G\cong G\times _{N^{\mathbb{C}^*}}Z/G\cong Z/\mathbb{Z}_2$$ where $N^{\mathbb{C}^*}$ is the normalizer of $\mathbb{C}^*$ in $G$. Since the dimension of $F/\!/ G$ is $\dim F-3$, the blow-up center $Y/G$ is nonsingular of codimension $1$. By Luna's slice theorem (\cite[Appendix 1.D]{MFK}), the singularity of $F/\!/ G$ at any point $[Gy]\in Y/G$ is $\mathbb{C}^2/\!/ \mathbb{C}^*$ where the weights are $\pm l$. Obviously this is smooth and hence $F/\!/ G$ is smooth along $Y/G$. Since the blow-up center is a smooth divisor, the blow-up map $\tilde{F}^s/G\to F/\!/ G$ has to be an isomorphism. \end{proof}
Let $Z_{m+k}$ be the $\mathbb{C}^*$-fixed locus in $F_{m+k}^{ss}$ and let $Y_{m+k}=GZ_{m+k}$. Then $Y_{m+k}$ is the disjoint union of
$$\Sigma_{m+k}^{S,S^c}:=\Sigma_{m+k}^S\cap \Sigma_{m+k}^{S^c}\cap F_{m+k}^{ss} \quad \text{for }|S|=m, S^c=\{1,\cdots,n\}-S $$ which are nonsingular of codimension $2$ for $k\ge 1$ by Lemma \ref{lem3-1}. For a point $$(C,p_1,\cdots,p_n,f:C\to\mathbb{P}^1)\in \Sigma^{S,S^c}_{m+k},$$ the parameterized component of $C$ (i.e. the unique component which is not contracted by $f$) has two nodes and no marked points. The normal space $\mathbb{C}^2$ to $\Sigma^{S,S^c}_{m+k}$ is given by the smoothing deformations of the two nodes and hence the stabilizer $\mathbb{C}^*$ acts with weights $2$ and $-2$.
The blow-up $\tilde{F}_{m+k}$ of $F_{m+k}^{ss}$ along $Y_{m+k}$ has no strictly semistable points by \cite[\S6]{Kirwan}. In fact, the unstable locus in $\tilde{F}_{m+k}$ is the proper transform of $\Sigma^S_{m+k}\cup \Sigma^{S^c}_{m+k}$ and the stabilizers of points in $\tilde{F}^s_{m+k}$ are either $\mathbb{Z}_2=\{\pm 1\}$ (for points not in the exceptional divisor of $\tilde{F}^s_{m+k}\to F^{ss}_{m+k}$) or $\mathbb{Z}_4=\{\pm 1,\pm i\}$ (for points in the exceptional divisor). Therefore, by Lemma \ref{lem4-3} and Lemma \ref{lem-specialcasebl} (3), we have isomorphisms \begin{equation}\label{eq4-10} \tilde{F}_{m+k}^s/G\cong F_{m+k}/\!/ G \end{equation} and $F_{m+k}/\!/ G$ are nonsingular for $k\ge 1$.
\noindent \underline{Step 2:} The partial desingularization $\tilde{F}_m/\!/ G$ is a nonsingular variety obtained by blowing up the $\frac12\binom{n}{m}$ singular points of $F_m/\!/ G=(\mathbb{P}^1)^n/\!/ G$.
Note that $Y_m$ in $F_m^{ss}$ is the disjoint union of
$\frac12\binom{n}{m}$ orbits $\Sigma_m^{S,S^c}$ for $|S|=m$. By Lemma \ref{lem-specialcasebl} (2), the morphism $\tilde{F}_m^s/G\to F_m/\!/ G$ is the blow-up at the $\frac12\binom{n}{m}$ points given by the orbits of the blow-up center. A point in $\Sigma^{S,S^c}_{m}$ is represented by $(\mathbb{P}^1,p_1,\cdots,p_n,\mathrm{id})$ with $p_i=p_j$ if $i,j\in S$ or $i,j\in S^c$. Without loss of generality, we may let $S=\{1,\cdots, m\}$. The normal space to an orbit $\Sigma^{S,S^c}_{m}$ is given by \[ (T_{p_1}\mathbb{P}^1)^{m-1}\times (T_{p_{m+1}}\mathbb{P}^1)^{m-1}=\mathbb{C}^{m-1}\times\mathbb{C}^{m-1} \] and $\mathbb{C}^*$ acts with weights $2$ and $-2$ respectively on the two factors. By Luna's slice theorem, \'etale locally near $\Sigma^{S,S^c}_{m}$, $F_m^{ss}$ is $G\times_{\mathbb{C}^*}(\mathbb{C}^{m-1}\times\mathbb{C}^{m-1})$ and $\tilde{F}_m$ is $G\times_{\mathbb{C}^*}\mathrm{bl}_{0}(\mathbb{C}^{m-1}\times\mathbb{C}^{m-1})$ while $\tilde{F}_m^s$ is $G\times_{\mathbb{C}^*}\left[\mathrm{bl}_{0}(\mathbb{C}^{m-1}\times\mathbb{C}^{m-1}) -\mathrm{bl}_{0}\mathbb{C}^{m-1}\sqcup \mathrm{bl}_{0}\mathbb{C}^{m-1}\right]$. By an explicit local calculation, the stabilizers of points on the exceptional divisor of $\widetilde{F}_m$ are $\mathbb{Z}_4=\{\pm 1,\pm i\}$ and the stabilizers of points over $F_m^s$ are $\mathbb{Z}_2=\{\pm 1\}$. Since the locus of nontrivial stabilizers for the action of $\bar G$ on $\tilde{F}^s_m$ is a smooth divisor with stabilizer $\mathbb{Z}_2$, $\tilde{F}_m/\!/ G=\tilde{F}^s_m/G$ is smooth and hence $\tilde{F}_m^s/G$ is the desingularization of $F_m/\!/ G$ obtained by blowing up its $\frac12\binom{n}{m}$ singular points.
\noindent \underline{Step 3:} The morphism $\bar\psi_{m+k+1}:F_{m+k+1}/\!/ G\to F_{m+k}/\!/ G$ is the blow-up along the union of transversal smooth subvarieties for $k\ge 1$. For $k=0$, we have $\tilde{F}^s_{m+1}=\tilde{F}^s_m$ and thus $$F_{m+1}/\!/ G\cong \tilde{F}^s_{m+1}/G=\tilde{F}^s_m/G=\tilde{F}_m/\!/ G$$ is the blow-up along its $\frac12\binom{n}{m}$ singular points.
From Lemma \ref{lem3-1}, we know $\Sigma^S_{m+k}$ for $|S|\ge m-k$ are transversal in $F_{m+k}$. In particular,
$$\bigcup_{|S|=m}\Sigma_{m+k}^S\cap \Sigma_{m+k}^{S^c}$$ intersects transversely with the blow-up center $$\bigcup_{|S'|=m-k} \Sigma^{S'}_{m+k}$$ for $\psi_{m+k+1}:F_{m+k+1}\to F_{m+k}$. Hence, by Proposition \ref{prop-bltrcen} we have a commutative diagram \begin{equation}\label{eq4-12} \xymatrix{ \tilde{F}_{m+k+1}\ar[r]\ar[d] & \tilde{F}_{m+k}\ar[d]\\ F_{m+k+1}^{ss}\ar[r] & F_{m+k}^{ss} } \end{equation} for $k\ge 1$ where the top horizontal arrow is the blow-up along the proper transforms $\tilde{\Sigma}^{S'}_{m+k}$ of
$\Sigma^{S'}_{m+k}$, $|S'|=m-k$. By Corollary \ref{cor-blcomquot}, we deduce that for $k\ge 1$, $\bar\psi_{m+k+1}$ is the blow-up along the transversal union of smooth subvarieties $\tilde{\Sigma}^{S'}_{m+k}/\!/ G\cong \Sigma^{S'}_{m+k}/\!/ G$.
For $k=0$, the morphism $\tilde{F}_{m+1}\to \tilde{F}_m$ is the blow-up along the proper transforms of $\Sigma_m^S$ and
$\Sigma_m^{S^c}$ for $|S|=m$. But these are unstable in $\tilde{F}_m$ and hence the morphism $\tilde{F}_{m+1}^s\to \tilde{F}_m^s$ on the stable part is the identity map. So we obtain $\tilde{F}_{m+1}^s= \tilde{F}_m^s$ and $\tilde{F}_{m+1}^s/G\cong \tilde{F}_m^s/G$.
\noindent \underline{Step 4:} Calculation of Picard numbers.
The Picard number of $F_m^{ss}=F_0^{ss}\subset F_0=(\mathbb{P}^1)^n$ is $n$ and so the Picard number of $\tilde{F}_m$ is $n+\frac12\binom{n}{m}$. By the descent lemma of \cite{DN} as in the odd degree case, the Picard number of $$F_{m+1}/\!/ G\cong \tilde{F}_{m+1}^s/G=\tilde{F}_m^s/G$$ equals the Picard number $n+\frac12\binom{n}{m}$ of $\tilde{F}_m^s$. Since the blow-up center of $\tilde{F}_{m+k}/\!/ G\to \tilde{F}_{m+k-1}/\!/ G$ has $\binom{n}{m-k+1}$ irreducible components, the Picard number of $\tilde{F}_{m+k}/\!/ G\cong F_{m+k}/\!/ G$ is \begin{equation}\label{eq4-13} n+\frac12\binom{n}{m}+\sum_{i=2}^{k}\binom{n}{m-i+1} \end{equation} for $k\ge 2$.
\noindent \underline{Step 5:} Completion of the proof.
As in the odd degree case, for $k\ge 1$ the universal family $\pi_k:\Gamma_{m+k}\to F_{m+k}$ gives rise to a family of pointed curves by considering the linear system $K_{\pi_k}+\epsilon_k\sum_i\sigma^i_{m+k}$. Over the semistable part $F_{m+k}^{ss}$ it is straightforward to check that this gives us a family of $n\cdot \epsilon_k$-stable pointed curves. Therefore we obtain an invariant morphism $$F_{m+k}^{ss}\to \overline{M}_{0,n\cdot \epsilon_k} $$ which induces a morphism $$F_{m+k}/\!/ G\to \overline{M}_{0,n\cdot \epsilon_k} .$$ By Lemma \ref{lem-PicNumMze}, the Picard number of $\overline{M}_{0,n\cdot \epsilon_k} $ coincides with that of $F_{m+k}/\!/ G$ given in \eqref{eq4-13}. Hence the morphism $F_{m+k}/\!/ G\to \overline{M}_{0,n\cdot \epsilon_k} $ is an isomorphism as desired. This completes our proof of Theorem \ref{thm4-1}.
\begin{remark}\label{2010rem1}
Let $S \subset \{1, 2, \cdots, n\}$ with $|S| = m-k$. On $\overline{M}_{0,n\cdot \epsilon_{k}}$, the blow-up center for $\overline{M}_{0,n\cdot \epsilon_{k+1}}\to\overline{M}_{0,n\cdot \epsilon_k} $ is the union of $n \choose {m-k}$ smooth subvarieties $\Sigma_{n-m+k}^S/\!/ G$. Each $\Sigma_{n-m+k}^S /\!/ G$ parameterizes weighted pointed stable curves with $m-k$ colliding marked points $s_{i_1}, s_{i_2}, \cdots, s_{i_{m-k}}$ for $i_j \in S$. On the other hand, for any member of $\overline{M}_{0,n\cdot \epsilon_k} $, no $m-k+1$ marked points can collide. So we can replace $m-k$ marked points $s_{i_j}$ with $i_j \in S$ by a single marked point which cannot collide with any other marked points. Therefore, an irreducible component $\Sigma_{n-m+k}^S /\!/ G$ of the blow-up center is isomorphic to the moduli space of weighted pointed rational curves $\overline{M}_{0, (1, \epsilon_k, \cdots, \epsilon_k)}$ with $n-m+k+1$ marked points as discovered by Hassett. (See Proposition \ref{reduction}.) \end{remark}
\begin{remark} For the moduli space of \emph{unordered} weighted pointed stable curves $\overline{M}_{0,n\cdot \epsilon_k} /S_n$, we can simply take quotients by the $S_n$ action of the blow-up process in Theorem \ref{thm4-1}. In particular, $\overline{M}_{0,n} /S_n$ is obtained by a sequence of weighted blow-ups from $\left((\mathbb{P}^1)^n/\!/ G\right)/ S_n=\mathbb{P}^n/\!/ G.$ \end{remark}
\section{Log canonical models of $\overline{M}_{0,n} $}\label{sec6}
In this section, we give a relatively elementary and straightforward proof of the following theorem of M. Simpson by using Theorem \ref{thm4-1}. Let $M_{0,n}$ be the moduli space of $n$ \emph{distinct} points in $\mathbb{P}^1$ up to $\mathrm{Aut}(\mathbb{P}^1)$.
\begin{theorem} \label{thm6.1}\emph{(M. Simpson \cite{Simpson})} Let $\alpha$ be a rational number satisfying $\frac{2}{n-1} < \alpha \le 1$ and let $D=\overline{M}_{0,n} -M_{0,n}$ denote the boundary divisor. Then the log canonical model \[
\overline{M}_{0,n} (\alpha) =
\mathrm{Proj}\;\left(\bigoplus_{l \ge 0} H^0(\overline{M}_{0,n} , \mathcal{O} (\lfloor l (K_{\overline{M}_{0,n} }
+\alpha D) \rfloor))\right) \] satisfies the following: \begin{enumerate} \item If $\frac{2}{m-k+2} < \alpha \le \frac{2}{m-k+1}$ for $1\le k\le m-2$, then $\overline{M}_{0,n} (\alpha) \cong \overline{M}_{0,n\cdot \epsilon_k} $.\item If $\frac{2}{n-1} < \alpha \le \frac{2}{m+1}$, then $\overline{M}_{0,n} (\alpha) \cong (\mathbb{P}^1)^n /\!/ G$ where the quotient is taken with respect to the symmetric linearization $\mathcal{O} (1,\cdots,1)$. \end{enumerate} \end{theorem} \begin{remark} Keel and McKernan prove (\cite[Lemma 3.6]{KeelMcKer}) that $K_{\overline{M}_{0,n} } + D$ is ample. Because $$\overline{M}_{0,n\cdot \epsilon_{m-2}} \cong\overline{M}_{0,n\cdot \epsilon_{m-1}} = \overline{M}_{0,n} $$ by definition, we find that (1) above holds for $k=m-1$ as well.\end{remark}
For notational convenience, we denote $(\mathbb{P}^1)^n/\!/ G$ by $\overline{M}_{0,n\cdot \epsilon_0} $ for even $n$ as well. Let $\Sigma_k^S$ denote the subvarieties of $F_k$ defined in \S\ref{sec3} for $S\subset \{1,\cdots,n\}$,
$|S|\le m$. Let $$D_k^S=\Sigma_{n-m+k}^S/\!/ G\subset F_{n-m+k}/\!/ G\cong \overline{M}_{0,n\cdot \epsilon_k} .$$
Then $D_k^S$ is a divisor of $\overline{M}_{0,n\cdot \epsilon_k} $ for $|S|=2$ or $m-k<|S|\le m$. Let $D_k^j=(\cup_{|S|=j}\Sigma_{n-m+k}^S)/\!/ G$ and $D_k=D_k^2+\sum_{j>m-k}D^j_k$. Then $D_k$ is the boundary divisor of $\overline{M}_{0,n\cdot \epsilon_k} $, i.e. $\overline{M}_{0,n\cdot \epsilon_k} -M_{0,n}=D_k$. When $k = m-2$ so $\overline{M}_{0,n\cdot \epsilon_k} \cong \overline{M}_{0,n} $, sometimes we will drop the subscript $k$. Note that if $n$ is even and
$|S|=m$, $D_k^S=D_k^{S^c}=\Sigma^{S,S^c}_{n-m+k}/\!/ G$.
By Theorem \ref{thm4-1}, there is a sequence of blow-ups \begin{equation}\label{eq6.1} \overline{M}_{0,n} \cong \overline{M}_{0,n\cdot \epsilon_{m-2}} \mapright{\varphi_{m-2}} \overline{M}_{0,n\cdot \epsilon_{m-3}} \mapright{\varphi_{m-3}}\cdots \mapright{\varphi_{2}}\overline{M}_{0,n\cdot \epsilon_1} \mapright{\varphi_{1}} \overline{M}_{0,n\cdot \epsilon_0} \end{equation} whose centers are transversal unions of smooth subvarieties, except for $\varphi_1$ when $n$ is even. Note that the irreducible components of the blow-up center of $\varphi_k$ furthermore intersect transversely with $D^j_{k-1}$ for $j>m-k+1$ by Lemma \ref{lem3-1} and by taking quotients. \begin{lemma}\label{computepushandpull} Let $1\le k\le m-2$. \begin{enumerate} \item $\varphi_k^* (D^j_{k-1}) = D^j_k$ for $j> m-k+1$. \item $\varphi_k^* (D^2_{k-1}) = D^2_k + {m-k+1 \choose 2} D^{m-k+1}_k$. \item $\varphi_{k *} (D^j_k) = D^j_{k-1}$ for $j > m-k+1$ or $j=2$. \item $\varphi_{k *} (D^j_k) =0$ for $j=m-k+1$.\end{enumerate} \end{lemma}
\begin{proof} The push-forward formulas (3) and (4) are obvious. Recall from \S\ref{sec4} that $\varphi_k=\bar\psi_{n-m+k}$ is the quotient of $\psi_{n-m+k}:F_{n-m+k}^{ss}\to F_{n-m+k-1}^{ss}$. Suppose $n$ is not even or $k$ is not $1$. Since $D^S_k$ for
$|S|>2$ does not contain any component of the blow-up center,
$\varphi_k^*(D^S_{k-1}) = D^S_k$. If $|S|=2$, $D^S_{k-1}$ contains a component $D^{S'}_{k-1}$ of the blow-up center if and only if $S' \supset S$. Therefore we have \[
\varphi_k^*(D^S_{k-1}) = D^S_k +
\sum_{S' \supset S, |S'| = m-k+1} D^{S'}_k. \]
By adding them up for all $S$ such that $|S|=2$, we obtain (2).
When $n$ is even and $k=1$, we calculate the pull-back before quotient. Let $\pi:\tilde{F}_{m}^s\to F_m^{ss}$ be the map obtained by blowing up $\cup_{|S|=m}\Sigma^{S,S^c}_m$ and removing unstable points. Recall that $\tilde{F}_m^s/G\cong F_{m+1}/\!/ G\cong \overline{M}_{0,n\cdot \epsilon_1} $ and the quotient of $\pi$ is $\varphi_1$. Then a direct calculation similar to the above gives us $\pi^*\Sigma^2_m=\tilde{\Sigma}_m^2+2\binom{m}{2}\tilde{\Sigma}^m_m$
where $\Sigma^2_m=\cup_{|S|=2}\Sigma^S_m$ and $\tilde{\Sigma}_m^2$ is the proper transform of $\Sigma^2_m$ while $\tilde{\Sigma}_m^m$ denotes the exceptional divisor. Note that by the descent lemma (\cite{DN}), the divisor $\Sigma_m^2$ and $\tilde{\Sigma}^2_m$ descend to $D^2_0$ and $D_{1}^2$. However $\tilde{\Sigma}_m^m$ does not descend because the stabilizer group $\mathbb{Z}_2$ in $\bar G=PGL(2)$ of points in $\tilde{\Sigma}_m^m$ acts nontrivially on the normal spaces. But by the descent lemma again, $2\tilde{\Sigma}^m_m$ descends to $D^m_1$. Thus we obtain (2). \end{proof}
Next we calculate the canonical divisors of $\overline{M}_{0,n\cdot \epsilon_k} $. \begin{proposition}\label{canonicaldiv1} \cite[Proposition 1]{Pand} The canonical divisor of $\overline{M}_{0,n} $ is \[
K_{\overline{M}_{0,n} } \cong -\frac{2}{n-1} D^2 + \sum_{j=3}^m
\left(-\frac{2}{n-1}{j \choose 2} +(j-2)\right) D^j. \] \end{proposition}
\begin{lemma}\label{canonicaldiv2} (1) The canonical divisor of $(\mathbb{P}^1)^n /\!/ G$ is \[
K_{(\mathbb{P}^1)^n /\!/ G} \cong -\frac{2}{n-1}D_0^2. \] (2) For $1\le k\le m-2$, the canonical divisor of $\overline{M}_{0,n\cdot \epsilon_k} $ is \[
K_{\overline{M}_{0,n\cdot \epsilon_k} } \cong -\frac{2}{n-1}D_k^2 + \sum_{j\ge m-k+1}^m
\left(-\frac{2}{n-1}{j \choose 2} + (j-2) \right)D_k^j. \] \end{lemma}
\begin{proof} It is well known by the descent lemma (\cite{DN}) that $\mathrm{Pic}((\mathbb{P}^1)^n /\!/ G)$ is a free abelian group of rank $n$(See \S6). The symmetric group $S_n$ acts on $(\mathbb{P}^1)^n /\!/ G$ in the obvious manner, and there is an induced action on its Picard group. Certainly the canonical bundle $K_{(\mathbb{P}^1)^n /\!/ G}$ and $D^2_0$ are $S_n$-invariant. On the other hand, the $S_n$-invariant part of the rational Picard group is a one dimensional vector space generated by the quotient $D_0^2$ of $\mathcal{O} _{(\mathbb{P}^1)^n}(n-1,\cdots,n-1)$ and hence we have $K_{(\mathbb{P}^1)^n /\!/ G} \cong c D_0^2$ for some $c\in \mathbb{Q}$.
Suppose $n$ is odd. The contraction morphisms $\varphi_k$ are all compositions of smooth blow-ups for $k \ge 1$. From the blow-up formula of canonical divisors (\cite[II Exe. 8.5]{Hartshorne}) and Lemma \ref{computepushandpull}, we deduce that \[
K_{\overline{M}_{0,n\cdot \epsilon_k} } = cD^2_k + \sum_{j\ge m-k+1}^m
\left(c{j \choose 2} + (j-2)\right)D^j_k. \] Since $\overline{M}_{0,n} \cong \overline{M}_{0,n\cdot \epsilon_{m-2}} $, we get $c = -\frac{2}{n-1}$ from Proposition \ref{canonicaldiv1}.
When $n$ is even, $\varphi_1^*(K_{(\mathbb{P}^1)^n /\!/ G}) = cD_1^2 + c{m \choose 2} D^m_1$ by Lemma \ref{computepushandpull}. We write $K_{\overline{M}_{0,n\cdot \epsilon_1} } = cD_1^2 + (c{m \choose 2} + a)D_1^m$. By the blow-up formula of canonical divisors (\cite[II Exe. 8.5]{Hartshorne}) again, we deduce that \[
K_{\overline{M}_{0,n\cdot \epsilon_k} } = cD_k^2 + \sum_{j \ge m-k+1}^{m-1}
\left( c{j \choose 2} + (j-2)\right)D_k^j + (c{m \choose 2} + a)D_k^m. \] From Proposition \ref{canonicaldiv1} again, we get $c = -\frac{2}{n-1}$ and $a = m-2$. \end{proof}
We are now ready to prove Theorem \ref{thm6.1}. By \cite[Corollary 3.5]{Simpson}, the theorem is a direct consequence of the following proposition. \begin{proposition}\label{prop-amplerange} (1) $K_{\overline{M}_{0,n\cdot \epsilon_0} }+\alpha D_0$ is ample if $\frac{2}{n-1} < \alpha \le \frac{2}{m+1}$.\\ (2) For $1\le k\le m-2$, $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ is ample if $\frac{2}{m-k+2} < \alpha \le \frac{2}{m-k+1}$.
\end{proposition}
Since any positive linear combination of an ample divisor and a nef divisor is ample \cite[Corollary 1.4.10]{Larz1}, it suffices to show the following: \begin{enumerate}\item[(a)] Nefness of $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ for $\alpha =\frac{2}{m-k+1}+s$ where $s$ is some (small) positive number; \item[(b)] Ampleness of $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ for $\alpha=\frac{2}{m-k+2}+t$ where $t$ is \emph{any} sufficiently small positive number. \end{enumerate} We will use Alexeev and Swinarski's intersection number calculation in \cite{AlexSwin} to achieve (a) (See Lemma \ref{lem-otherextreme}.) and then (b) will immediately follow from our Theorem \ref{thm4-1}.
\begin{definition} (\cite{Simpson}) Let $\varphi=\varphi_{n\cdot\epsilon_{m-2}, n\cdot\epsilon_k} : \overline{M}_{0,n} \to \overline{M}_{0,n\cdot \epsilon_k} $ be the natural contraction map (\S\ref{sec2.1}). For $k = 0,1,\cdots,m-2$ and $\alpha > 0$, define $A(k, \alpha)$ by \begin{eqnarray*}
A(k, \alpha) &:=& \varphi^*(K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k)\\
&=& \sum_{j = 2}^{m-k} {j \choose 2}\left(\alpha - \frac{2}{n-1}\right)D^j
+ \sum_{j \ge m-k+1}^m \left(\alpha - \frac{2}{n-1}{j \choose 2}
+ j-2\right)D^j. \end{eqnarray*} \end{definition} Notice that the last equality is an easy consequence of Lemma \ref{computepushandpull}.
By \cite{Kapranov}, there is a birational morphism $\pi_{\vec{x}} : \overline{M}_{0,n} \to (\mathbb{P}^1)^n /\!/ _{\vec{x}} G$ for any linearization $\vec{x} = (x_1, \cdots, x_n) \in \mathbb{Q}_+^n$. Since the canonical ample line bundle $\mathcal{O} _{(\mathbb{P}^1)^n }(x_1, \cdots, x_n)/\!/ G$ over $(\mathbb{P}^1)^n /\!/ _{\vec{x}} G$ is ample, its pull-back $L_{\vec{x}}$ by $\pi_{\vec{x}}$ is certainly nef.
\begin{definition}\cite[Definition 2.3]{AlexSwin}\label{def-symnefdiv} Let $x$ be a rational number such that $\frac{1}{n-1} \le x \le \frac{2}{n}$. Set $\vec{x} = \mathcal{O} (x, \cdots, x, 2-(n-1)x)$. Define \[
V(x, n) := \frac{1}{(n-1)!}\bigotimes_{\tau \in S_n} L_{\tau
\vec{x}}. \] Obviously the symmetric group $S_n$ acts on $\vec{x}$ by permuting the components of $\vec{x}$. \end{definition} Notice that $V(x,n)$ is nef because it is a positive linear combination of nef line bundles.
\begin{definition}\cite[Definition 3.5]{AlexSwin} Let $C_{a,b,c,d}$ be \emph{any} vital curve class corresponding to a partition $S_a \sqcup S_b \sqcup S_c \sqcup S_d$ of $\{1, 2, \cdots, n\}$
such that $|S_a| = a, \cdots, |S_d|=d$.\\ (1) Suppose $n=2m+1$ is odd. Let $C_i = C_{1, 1, m-i, m+i-1}$, for $i = 1, 2, \cdots, m-1$.\\ (2) Suppose $n=2m$ is even. Let $C_i = C_{1,1,m-i, m+i-2}$ for $i = 1, 2, \cdots, m-1$. \end{definition}
By \cite[Corollary 4.4]{KeelMcKer}, the following computation is straightforward. \begin{lemma}\label{lem-intAkalpha} The intersection numbers $C_i \cdot A(k, \alpha)$ are \[
C_i \cdot A(k, \alpha) = \left\{\begin{array}{ll}
\alpha&\mbox{if } i < k\\
\left(2-{m-k \choose 2}\right)\alpha + m-k-2&\mbox{if } i = k\\
\left({m-k+1 \choose 2}-1\right)\alpha-m+k+1
&\mbox{if } i = k+1\\
0&\mbox{if } i > k+1.
\end{array}\right. \] \end{lemma}
This lemma is in fact a slight generalization of \cite[Lemma 3.7]{AlexSwin} where the intersection numbers for $\alpha=\frac{2}{m-k+1}$ only are calculated.
The $S_n$-invariant subspace of Neron-Severi vector space of $\overline{M}_{0,n} $ is generated by $D^j$ for $j=2,3, \cdots, m$ (\cite[Theorem 1.3]{KeelMcKer}). Therefore, in order to determine the linear dependency of $S_n$-invariant divisors, we find $m-1$ linearly independent curve classes, and calculate the intersection numbers of divisors with these curves classes. Let $U$ be an $(m-1) \times (m-1)$ matrix with entries $U_{ij} = (C_i \cdot V(\frac{1}{m+j}, n))$ for $1 \le i,j \le m-1$. Since $V(\frac{1}{m+j}, n)$'s are all nef, all entries of $U$ are nonnegative. \begin{lemma}\cite[\S3.2, \S3.3]{AlexSwin}\label{lem-nefbdl} (1) The intersection matrix $U$ is upper triangular and if $i \le j$, then $U_{ij} > 0$. In particular, $U$ is invertible.\\ (2) Let $\vec{a} = ((C_1 \cdot A(k, \frac{2}{m-k+1})), \cdots, (C_{m-1} \cdot A(k,\frac{2}{m-k+1})))^t$ be the column vector of intersection numbers. Let $\vec{c} = (c_1, c_2, \cdots, c_{m-1})^t$ be the unique solution of the system of linear equations $U \vec{c} = \vec{a}$. Then $c_i > 0$ for $i \le k+1$ and $c_i = 0$ for $i \ge k+2$. \end{lemma}
This lemma implies that $A(k,\frac{2}{m-k+1})$ is a positive linear combination of $V(\frac{1}{m+j}, n)$ for $j = 1, 2, \cdots, k+1$. Note that $A(k,\frac{2}{m-k+2}) = A(k-1,\frac{2}{m-(k-1)+1})$ and that for $\frac{2}{m-k+2} \le \alpha \le \frac{2}{m-k+1}$, $A(k,\alpha)$ is a nonnegative linear combination of $A(k,\frac{2}{m-k+2})$ and $A(k,\frac{2}{m-k+1})$. Hence by the numerical result in Lemma \ref{lem-nefbdl} and the convexity of the nef cone, $A(k,\alpha)$ is nef for $\frac{2}{m-k+2} \le \alpha \le \frac{2}{m-k+1}$. Actually we can slightly improve this result by using continuity.
\begin{lemma}\label{lem-otherextreme} For each $k = 0,1,\cdots,m-2$, there exists $s > 0$ such that $A(k,\alpha)$ is nef for $\frac{2}{m-k+2} \le \alpha \le \frac{2}{m-k+1}+s$. Therefore, $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ is nef for $\frac{2}{m-k+2} \le \alpha \le \frac{2}{m-k+1}+s$. \end{lemma} \begin{proof} Let $\vec{a}^\alpha = ((C_1 \cdot A(k,\alpha)), \cdots, (C_{m-1} \cdot A(k,\alpha)))^t$ and let $\vec{c}^\alpha = (c^\alpha_1, \cdots, c^\alpha_{m-1})^t$ be the unique solution of equation $U \vec{c}^\alpha = \vec{a}^\alpha$. Then by continuity, the components $c^\alpha_1, c^\alpha_2, \cdots, c^\alpha_{k+1}$ remain positive when $\alpha$ is slightly increased. By Lemma \ref{lem-intAkalpha} and the upper triangularity of $U$, $c^\alpha_i$ for $i > k+1$ are all zero. Hence $A(k, \alpha)$ is still nef for $\alpha = \frac{2}{m-k+1}+s$ with sufficiently small $s > 0$. \end{proof}
With this nefness result, the proof of Proposition \ref{prop-amplerange} is obtained as a quick application of Theorem \ref{thm4-1}.
\begin{proof}[Proof of Proposition \ref{prop-amplerange}] We prove that in fact $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ is ample for $\frac{2}{m-k+2} < \alpha < \frac{2}{m-k+1}+s$ where $s$ is the small positive rational number in Lemma \ref{lem-otherextreme}. Since a positive linear combination of an ample divisor and a nef divisor is ample by \cite[Corollary 1.4.10]{Larz1}, it suffices to show that $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ is ample when $\alpha= \frac{2}{m-k+2}+t$ for any sufficiently small $t>0$ by Lemma \ref{lem-otherextreme}.
We use induction on $k$. It is certainly true when $k=0$ by Lemma \ref{canonicaldiv2} because $D^2_0$ is ample as the quotient of $\mathcal{O} (n-1,\cdots,n-1)$. Suppose $K_{\overline{M}_{0,n\cdot \epsilon_{k-1}} }+\alpha D_{k-1}$ is ample for $\frac{2}{m-k+3} < \alpha < \frac{2}{m-k+2}+s'$ where $s'$ is the small positive number in Lemma \ref{lem-otherextreme} for $k-1$. Since $\varphi_k$ is a blow-up with exceptional divisor $D_k^{m-k+1}$, \[
\varphi_k^*(K_{\overline{M}_{0,n\cdot \epsilon_{k-1}} }+\alpha D_{k-1})-\delta D_k^{m-k+1} \] is ample for any sufficiently small $\delta>0$ by \cite[II 7.10]{Hartshorne}. A direct computation with Lemmas \ref{computepushandpull} and \ref{canonicaldiv2} provides us with \[ \varphi_k^*(K_{\overline{M}_{0,n\cdot \epsilon_{k-1}} }+\alpha D_{k-1})-\delta D_k^{m-k+1}\] \[ =K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_{k}+\left( \binom{m-k+1}{2}\alpha -\alpha-(m-k-1)-\delta \right)D_k^{m-k+1}. \] If $\alpha=\frac{2}{m-k+2}$, $\binom{m-k+1}{2}\alpha -\alpha-(m-k-1)=0$ and thus we can find $\alpha>\frac{2}{m-k+2}$ satisfying $\binom{m-k+1}{2}\alpha -\alpha-(m-k-1)-\delta =0$. If $\delta$ decreases to $0$, the solution $\alpha$ decreases to $\frac{2}{m-k+2}$. Hence $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ is ample when $\alpha= \frac{2}{m-k+2}+t$ for any sufficiently small $t>0$ as desired. \end{proof}
\begin{remark}\label{rem-compproofs} There are already two different proofs of M. Simpson's theorem (Theorem \ref{thm6.1}) given by Fedorchuk--Smyth \cite{FedoSmyt}, and by Alexeev--Swinarski \cite{AlexSwin} without relying on Fulton's conjecture. Here we give a brief outline of the two proofs.
In \cite[Corollary 3.5]{Simpson}, Simpson proves that Theorem \ref{thm6.1} is an immediate consequence of the ampleness of $K_{\overline{M}_{0,n\cdot \epsilon_k} } + \alpha D_k$ for $\frac{2}{m-k+2} < \alpha \le \frac{2}{m-k+1}$ (Proposition \ref{prop-amplerange}). The differences in the proofs of Theorem \ref{thm6.1} reside solely in different ways of proving Proposition \ref{prop-amplerange}.
The ampleness of $K_{\overline{M}_{0,n\cdot \epsilon_k} } + \alpha D_k$ follows if the divisor $A(k, \alpha) = \varphi^*(K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k)$ is nef and its linear system contracts only $\varphi$-exceptional curves. Here, $\varphi : \overline{M}_{0,n} \to \overline{M}_{0,n\cdot \epsilon_k} $ is the natural contraction map (\S\ref{sec2.1}). Alexeev and Swinarski prove Proposition \ref{prop-amplerange} in two stages: First the nefness of $A(k, \alpha)$ for suitable ranges is proved and next they show that the divisors are the pull-backs of ample line bundles on $\overline{M}_{0,n\cdot \epsilon_k} $. Lemma \ref{lem-otherextreme} above is only a negligible improvement of the nefness result in \cite[\S3]{AlexSwin}. In \cite[Theorem 4.1]{AlexSwin}, they give a criterion for a line bundle to be the pull-back of an ample line bundle on $\overline{M}_{0,n\cdot \epsilon_k} $. After some rather sophisticated combinatorial computations, they prove in \cite[Proposition 4.2]{AlexSwin} that $A(k, \alpha)$ satisfies the desired properties.
On the other hand, Fedorchuk and Smyth show that $K_{\overline{M}_{0,n\cdot \epsilon_k} } + \alpha D_k$ is ample as follows. Firstly, by applying the Grothendieck-Riemann-Roch theorem, they represent $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ as a linear combination of boundary divisors and tautological $\psi$-classes. Secondly, for such a linear combination of divisor classes and for a complete curve in $\overline{M}_{0,n\cdot \epsilon_k} $ parameterizing a family of curves with smooth general member, they perform brilliant computations and get several inequalities satisfied by their intersection numbers (\cite[Proposition 3.2]{FedoSmyt}). Combining these inequalities, they prove in particular that $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ has positive intersection with any complete curve on $\overline{M}_{0,n\cdot \epsilon_k} $ with smooth general member (\cite[Theorem 4.3]{FedoSmyt}). Thirdly, they prove that if the divisor class intersects positively with any curve with smooth general member, then it intersects positively with all curves by an induction argument on the dimension. Thus they establish the fact that $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ has positive intersection with all curves. Lastly, they prove that the same property holds even if $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ is perturbed by any small linear combination of boundary divisors. Since the boundary divisors generate the Neron-Severi vector space, $K_{\overline{M}_{0,n\cdot \epsilon_k} }+\alpha D_k$ lies in the interior of the nef cone and the desired ampleness follows. \end{remark}
\section{The Picard groups of $\overline{M}_{0,n\cdot \epsilon_k} $}\label{sec5}
As a byproduct of our GIT construction of the moduli spaces of weighted pointed curves, we give a basis of the \emph{integral} Picard group of $\overline{M}_{0,n\cdot \epsilon_k} $ for $0 \le k \le m-2$.
Let $e_i$ be the $i$-th standard basis vector of $\mathbb{Z}^n$. For notational convenience, set $e_{n+1} = e_1$. For $S\subset
\{1,2,\cdots,n\}$, let $D_k^S=\Sigma_{n-m+k}^S/\!/ G\subset F_{n-m+k}/\!/ G\cong \overline{M}_{0,n\cdot \epsilon_k} $. Note that if $m-k<|S|\le m$ or
$|S|=2$, $D_k^S$ is a divisor of $\overline{M}_{0,n\cdot \epsilon_k} $.
\begin{theorem}\label{picardgroup} (1) If $n$ is odd, then the Picard group of $\overline{M}_{0,n\cdot \epsilon_k} $ is \[
\mathrm{Pic} (\overline{M}_{0,n\cdot \epsilon_k} ) \cong \bigoplus_{m-k < |S| \le m} \mathbb{Z} D^S_k \oplus
\bigoplus_{i=1}^n \mathbb{Z} D_k^{\{i, i+1\}} \] for $0 \le k \le m-2$.\\ (2) If $n$ is even, then the Picard group of $\overline{M}_{0,n\cdot \epsilon_k} $ is \[
\mathrm{Pic} (\overline{M}_{0,n\cdot \epsilon_k} ) \cong \bigoplus_{m-k < |S| < m} \mathbb{Z} D^S_k \oplus
\bigoplus_{1 \in S, |S| = m} \mathbb{Z} D^S_k \oplus
\bigoplus_{i=1}^{n-1} \mathbb{Z} D_k^{\{i, i+1\}}
\oplus \mathbb{Z} D_k^{\{1, n-1\}}. \] for $1 \le k \le m-2$. \end{theorem}
\begin{proof} Since the codimensions of unstable strata in $(\mathbb{P}^1)^n$ are greater than 1, $$\mathrm{Pic} (((\mathbb{P}^1)^n)^{ss}) = \mathrm{Pic} ((\mathbb{P}^1)^n) \cong \oplus_{1 \le i \le n} \mathbb{Z} \mathcal{O} (e_i).$$ For all $x \in ((\mathbb{P}^1)^n)^s$, $G_x \cong \{\pm 1\}$. If $n$ is even and $x$ is strictly semistable point with closed orbit, then $G_x \cong \mathbb{C}^*$. Since $G$ is connected, $G$ acts on the discrete set $\mathrm{Pic}((\mathbb{P}^1)^n)$ trivially. By Kempf's descent lemma (\cite[theorem 2.3]{DN}) and by checking the actions of the stabilizers on the fibers of line bundles, we deduce that $\mathcal{O} (a_1, a_2, \cdots, a_n)$ descends to $((\mathbb{P}^1)^n)^{ss} /\!/ G$ if and only if $2$ divides $\sum a_i$.
Consider the case when $n$ is odd first. It is elementary to check that the subgroup $\{(a_1, \cdots, a_n) \in \mathbb{Z}^n | \sum a_i \in 2\mathbb{Z} \}$ is free abelian of rank $n$ and $\{e_i + e_{i+1}\}$ for $1 \le i \le n$ form a basis of this group. Furthermore, for $S=\{i,j\}$ with $i\ne j$, the big diagonal $(\Sigma_{m+1}^S)^s = (\Sigma_0^S)^s$ satisfies $\mathcal{O} ((\Sigma_0^S)^s) \cong \mathcal{O} _{F_0^s}(e_i+e_j)$. Hence in $F_{m+1}/\!/ G = F_0 /\!/ G$, $\mathcal{O} (\Sigma_0^S /\!/ G) \cong \mathcal{O} (e_i + e_j)$. Therefore we have \[
\mathrm{Pic} (\overline{M}_{0,n\cdot \epsilon_0} ) = \mathrm{Pic} (F_{m+1} /\!/ G)
= \bigoplus_{i = 1}^n \mathbb{Z} D_0^{\{i, i+1\}}. \]
By Theorem \ref{thm4-1}, the contraction morphism $\varphi_k : \overline{M}_{0,n\cdot \epsilon_k} \to \overline{M}_{0,n\cdot \epsilon_{k-1}} $ is the blow-up along the union of transversally intersecting smooth subvarieties. By \S2.3, this is a composition of smooth blow-ups. In $\overline{M}_{0,n\cdot \epsilon_k} $, the exceptional divisors are
$D^S_k$ for $|S| = {m-k+1}$. So the Picard group of $\overline{M}_{0,n\cdot \epsilon_k} $ is \[
\mathrm{Pic} (\overline{M}_{0,n\cdot \epsilon_k} ) \cong \varphi_k^* \mathrm{Pic} (\overline{M}_{0,n\cdot \epsilon_{k-1}} ) \oplus
\bigoplus_{|S| = {m-k+1}}\mathbb{Z} D^S_k. \]
by \cite[II Exe. 8.5]{Hartshorne}. For any $S$ with $|S| = 2$, $D_{k-1}^S$ contains the blow-up center $D_{k-1}^{S'}$ if $S \subset {S'}$. So $\varphi_k^* (D_{k-1}^S)$ is the sum of $D_k^S$
and a linear combination of $D_k^{S'}$ for $S' \supset S, |S'| =
m-k+1$. If $|S| > 2$, then $\varphi_k^* D^S_{k-1} = D^S_k$ since it does not contain any blow-up centers. After obvious basis change, we get the desired result by induction.
Now suppose that $n$ is even. Still the group $\{ (a_1, \cdots, a_n) \in \mathbb{Z}^n | \sum a_i \in 2\mathbb{Z}\}$ is free abelian of rank $n$ and $\{e_i + e_{i+1}\}_{1 \le i \le n-1} \cup \{e_1 + e_{n-1}\}$ form a basis. In $F_m/\!/ G = F_0 /\!/ G$, $\mathcal{O} (\Sigma_m^S/\!/ G) \cong \mathcal{O} (e_i + e_j)$ when $S = \{i, j\}$ with $i\ne j$. Hence \[
\mathrm{Pic} (F_m /\!/ G) = \bigoplus_{i=1}^{n-1}
\mathbb{Z} D_0^{\{i, i+1\}} \oplus \mathbb{Z} D_0^{\{1, n-1\}}. \]
In $\tilde{F}_m$, the unstable loci have codimension two. Therefore we have \[
\mathrm{Pic} (\tilde{F}_m^s) = \mathrm{Pic}(\tilde{F}_m)
= \pi_m^* \mathrm{Pic} (F_m^{ss}) \oplus \bigoplus_{1 \in S, |S| = m}
\mathbb{Z} \tilde{\Sigma}_m^S, \] where $\pi_m : \tilde{F}_m \to F_m^{ss}$ is the blow-up morphism, and $\tilde{\Sigma}_m^S = \pi_m^{-1}(\Sigma_m^S \cap \Sigma_m^{S^c})$
for $|S|=m$.
By Kempf's descent lemma, $\mathrm{Pic}(F_{m+k}/\!/ G)$ is a subgroup of $\mathrm{Pic}(F_{m+k}^{ss})$ and $\mathrm{Pic}(\tilde{F}_{m+k}^s)$ for $0 \le k \le m-2$. From our blow-up description, all arrows except possibly $\bar{\psi}_{m+1}^*$ in following commutative diagram \[
\xymatrix{\mathrm{Pic}(\tilde{F}_{m+1}^s)&
\mathrm{Pic}(\tilde{F}_m^s)\ar@{=}[l]_{\tilde{\psi}_{m+1}^*}\\
\mathrm{Pic}(F_{m+1}^{ss})\ar[u]_{\pi_{m+1}^*}&
\mathrm{Pic}(F_m^{ss})\ar[l]_{\psi_{m+1}^*}\ar[u]_{\pi_m^*}\\
\mathrm{Pic}(F_{m+1}/\!/ G)\ar[u]&
\mathrm{Pic}(F_m /\!/ G)\ar[l]_{\bar{\psi}_{m+1}^*}\ar[u]} \] are injective, and thus the bottom arrow $\bar{\psi}_{m+1}^*$ is also injective. Hence $\mathrm{Pic}(F_{m+1}/\!/ G)$ contains the pull-back of $\mathrm{Pic}(F_m /\!/ G)$ as a subgroup. Also, for the quotient map $p : \tilde{F}_{m+1}^s \to F_{m+1}/\!/ G$, $p^*
D_1^S = \tilde{\Sigma}_{m+1}^S$ for $|S| = m$. Let $H$ be the subgroup of $\mathrm{Pic}(\tilde{F}_{m+1}^s)$ generated by the images of $\bar{\psi}_{m+1}^* \mathrm{Pic}(F_m /\!/ G)$ and the divisors $D_1^S$ with $|S| = m$. By definition, the image of $\mathrm{Pic}(F_{m+1}/\!/ G)$ contains $H$. Now by checking the action of stabilizers on the fibers of line bundles, it is easy to see that no line bundles in $\mathrm{Pic}(\tilde{F}_{m+1}^s)- H$ descend to $F_{m+1}/\!/ G$. Hence we have \begin{equation}\label{mzeo1}
\mathrm{Pic}(\overline{M}_{0,n\cdot \epsilon_1} ) = \mathrm{Pic}(F_{m+1} /\!/ G)
= \bar{\psi}_{m+1}^* \mathrm{Pic}(F_m /\!/ G) \oplus
\bigoplus_{1 \in S, |S| = m}\mathbb{Z} D_1^S. \end{equation}
For an $S$ with $|S|=2$, $\Sigma_m^S/\!/ G$ contains the blow-up center $\Sigma_m^{S'} \cap \Sigma_m^{{S'}^c}/\!/ G$ if $S \subset S'$ or $S \subset {S'}^c$. So $\bar{\psi}_{m+1}^* (D_0^S)$ is the sum of $D_{1}^S$ and a linear combination of divisors $D_1^{S'}$
for $S' \supset S$ or ${S'}^c\supset S$ with $|S'| = m$. From this and (\ref{mzeo1}), we get the following by an obvious basis change: \[
\mathrm{Pic}(\overline{M}_{0,n\cdot \epsilon_1} ) = \bigoplus_{1 \in S, |S| = m}\mathbb{Z} D_1^S
\oplus \bigoplus_{i=1}^{n-1} \mathbb{Z} D_1^{\{i,i+1\}}
\oplus \mathbb{Z} D_1^{\{1,n-1\}}. \] The rest of the proof is identical to the odd $n$ case and so we omit it. \end{proof}
\end{document}
|
arXiv
|
{
"id": "1002.2461.tex",
"language_detection_score": 0.6628998517990112,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Cohesive dynamics and brittle fracture}
\begin{abstract} We formulate a nonlocal cohesive model for calculating the deformation state inside a cracking body. In this model a more complete set of physical properties including elastic and softening behavior are assigned to each point in the medium. We work within the small deformation setting and use the peridynamic formulation. Here strains are calculated as difference quotients. The constitutive relation is given by a nonlocal cohesive law relating force to strain. At each instant of the evolution we identify a process zone where strains lie above a threshold value. Perturbation analysis shows that jump discontinuities within the process zone can become unstable and grow. We derive an explicit inequality that shows that the size of the process zone is controlled by the ratio given by the length scale of nonlocal interaction divided by the characteristic dimension of the sample. The process zone is shown to concentrate on a set of zero volume in the limit where the length scale of nonlocal interaction vanishes with respect to the size of the domain. In this limit the dynamic evolution is seen to have bounded linear elastic energy and Griffith surface energy. The limit dynamics corresponds to the simultaneous evolution of linear elastic displacement and the fracture set across which the displacement is discontinuous. We conclude illustrating how the approach developed here can be applied to limits of dynamics associated with other energies that $\Gamma$- converge to the Griffith fracture energy. \end{abstract}
\begin{flushleft} {\bf Keywords:} \,\,peridynamics, dynamic brittle fracture, fracture toughness, process zone, $\Gamma$- convergence \end{flushleft}
\begin{flushleft} {\bf Mathematics Subject Classification}: 34A34, 74H55, 74R10 \end{flushleft}
\pagestyle{myheadings}
\markboth{R. LIPTON}{Cohesive Dynamics and Brittle Fracture}
\setcounter{equation}{0} \setcounter{theorem}{0} \setcounter{lemma}{0}\setcounter{proposition}{0}\setcounter{remark}{0}\setcounter{definition}{0}\setcounter{hypothesis}{0}
\section{Introduction} \label{introduction}
Dynamic brittle fracture is a multiscale phenomenon operating across a wide range of length and time scales. Contemporary approaches to brittle fracture modeling can be broadly characterized as bottom-up and top-down. Bottom-up approaches take into account the discreteness of fracture at the smallest length scales and are expressed through lattice models. This approach has provided insight into the dynamics of the fracture process \cite{6,16,17,21}. Complementary to the bottom-up approaches are top-down computational approaches using cohesive surface elements \cite{Cox}, \cite{14}, \cite{22}, \cite{Remmers}. In this formulation the details of the process zone are collapsed onto an interfacial element with a force traction law given by the cohesive zone model \cite{Barenblatt}, \cite{Dougdale}. Cohesive surfaces have been applied within the extended finite element method \cite{5}, \cite{Duarte}, \cite{18} to minimize the effects of mesh dependence on free crack paths. Higher order multi-scale cohesive surface models involving excess properties and differential momentum balance are developed in \cite{WaltonSendova}. Comparisons between different cohesive surface models are given in \cite{Falk}. More recently variational approaches to brittle fracture based on quasi-static evolutions of global minimizers of Grifffith's fracture energy have been developed \cite{FrancfortMarigo}, \cite{BourdinFrancfortMarigo}, and \cite{FrancfortLarsen}. Phase field approaches have also been developed to model brittle fracture evolution from a continuum perspective \cite{BourdinFrancfortMarigo}, \cite{BourdinLarsenRichardson}, \cite{Miehe}, \cite{Hughes}, \cite{Ortiz}, \cite{Wheeler}. In the phase field approach a second field is introduced to interpolate between cracked and undamaged elastic material. The evolution of the phase field is used to capture the trajectory of the crack. A concurrent development is the emergence of the peridynamic formulation introduced in \cite{Silling1} and \cite{States}. Peridynamics is a nonlocal formulation of continuum mechanics expressed in terms of displacement differences as opposed to spatial derivatives of the displacement field. These features provide the ability to simultaneously simulate kinematics involving both smooth displacements and defect evolution. Numerical simulations based on peridynamic modeling exhibit the formation and evolution of sharp interfaces associated with defects and fracture \cite{Bobaru1}, \cite{BhattacharyaDyal}, \cite{Foster}, \cite{Bobaru2}, \cite{SillingBobaru}, \cite{SillingAscari2}, \cite{WecknerAbeyaratne}. In an independent development nonlocal formulations have been introduced for modeling the passage from discrete to continuum limits of energies for quasistatic fracture models \cite{Alicandro}, \cite{Buttazzo}, \cite{Braides}, \cite{BraidesGelli}, for smeared crack models \cite{Negri} and for image processing \cite{Gobbino1} and \cite{Gobbino3}. A complete review of contemporary methods is beyond the scope of this paper however the reader is referred to \cite{Agwai}, \cite{Bazant}, \cite{BelitchReview}, \cite{Bouchbinder}, \cite{BourdinFrancfortMarigo}, \cite{Braides1}, \cite{Braides2} for a more complete guide to the literature.
In this paper we formulate a nonlocal, multi-scale, cohesive continuum model for assessing the deformation state inside a cracking body. This model is expressed using the peridynamic formulation introduced in \cite{Silling1}, \cite{States}. Here strains are calculated as difference quotients of displacements between two points $x$ and $y$. In this approach the force between two points $x$ and $y$ is related to the strain through a nonlinear cohesive law that depends upon the magnitude and direction of the strain. The forces are initially elastic for small strains and soften beyond a critical strain. We introduce the dimensionless length scale $\epsilon$ given by the ratio of the length scale of nonlocal interaction to the characteristic length of the material sample $D$. Working in the new rescaled coordinates the nonlocal interactions between $x$ and its neighbors $y$ occur within a horizon of radius $\epsilon$ about $x$ and the characteristic length of $D$ is taken to be unity. This neighborhood of $x$ is denoted by $\mathcal{H}_\epsilon(x)$.
To define the potential energy we first assume the deformation $z$ is given by $z(x)=u(x)+x$ where $u$ is the displacement field. The strain between two points $x$ and $y$ inside $D$ is given by \begin{eqnarray}
\mathcal{S}=\frac{|z(y)-z(x)|-|y-x|}{|y-x|}. \label{bigdefstrain} \end{eqnarray}
In this treatment we assume small deformation kinematics and the displacements are small (infinitesimal) relative to the size of the body $D$. Under this hypothesis \eqref{bigdefstrain} is linearized and the strain is given by $$\mathcal{S}=\frac{u(y)-u(x)}{|y-x|}\cdot e,$$ where $e=\frac{y-x}{|y-x|}$. Both two and three dimensional problems will be considered and the dimension is denoted by $d=2,3$. The cohesive model is characterized through a nonlocal potential $$W^\epsilon(\mathcal{S},y-x),$$ associated with points $x$ and $y$. The associated energy density is obtained on integrating over $y$ for $x$ fixed and is given by \begin{eqnarray} {\bf{W}}^\epsilon(\mathcal{S},x)=\frac{1}{V_d}\int_{\mathcal{H}_\epsilon(x)}\,W^\epsilon(\mathcal{S},y-x)\,dy \label{densityy} \end{eqnarray} where $V_d=\epsilon^d\omega_d$ and $\omega_d$ is the (area) volume of the unit ball in dimensions $d=(2)3$. The potential energy of the body is given by \begin{eqnarray} PD^\epsilon(u)=\int_{D}\,{\bf{W}}^\epsilon(\mathcal{S},x)\,dx \label{the peridynamicenergy} \end{eqnarray}
\begin{figure}
\caption{\bf Cohesive potential as a function of $\mathcal{S}$ for $x$ and $y$ fixed.}
\label{ConvexConcave}
\end{figure}
\begin{figure}
\caption{{\bf Cohesive relation between force and strain for $x$ and $y$ fixed.}}
\label{SofteningBond}
\end{figure}
We introduce the class of potentials associated with a cohesive force that is initially elastic and then softens after a critical strain. These potentials are of the generic form given by \begin{eqnarray}
W^\epsilon(S,y-x)=|y-x|\mathcal{W}^\epsilon(\mathcal{S},y-x), \label{potentialdensity1a} \end{eqnarray} where $\mathcal{W}^\epsilon(\mathcal{S},y-x)$ is the peridynamic potential per unit length associated with $x$ and $y$ given by \begin{eqnarray}
\mathcal{W}^\epsilon(\mathcal{S},y-x)=\frac{1}{\epsilon}J^\epsilon\left(|y-x|\right)\left(\frac{1}{|y-x|}f\left(|y-x|\mathcal{S}^2\right)\right). \label{potentialdensity2a} \end{eqnarray} These potentials are of a general form and are associated with potential functions $f:[0,\infty)\rightarrow\mathbb{R}$ that are positive, smooth and concave with the properties \begin{eqnarray} \lim_{r\rightarrow 0^+}\frac{f(r)}{r}=f'(0)>0,&&\lim_{r\rightarrow\infty}f(r)=f_\infty <\infty. \label{properties} \end{eqnarray}
The composition of $f$ with $|y-x|\mathcal{S}^2$ given by \eqref{potentialdensity2a} delivers the convex-concave dependence of $\mathcal{W}^\epsilon(\mathcal{S},y-x)$ on $\mathcal{S}$ for fixed values of $x$ and $y$, see Figure \ref{ConvexConcave}. Here $J^\epsilon(|y-x|)$ is used to prescribe the influence of separation length $|y-x |$ on the force between $x$ and $y$ with $0\leq J^\epsilon(|y-x|)<M $ for $0\leq |y-x|< \epsilon$ and $J^\epsilon(|y-x|)=0$ for $\epsilon\leq |y-x |$. For fixed $x$ and $y$ the inflection point for the potential energy \eqref{potentialdensity2a} with respect to the strain $\mathcal{S}$ is given by $\overline{r}/\sqrt{|y-x|}$, where $\overline{r}$ is the inflection point for the function $r:\rightarrow f(r^2)$, see Figure \ref{ConvexConcave}. This choice of potential delivers an initially elastic and then softening constitutive law for the force per unit length along the direction $e$ given by \begin{eqnarray}
\hbox{\rm force per unit length}=\partial_\mathcal{S} \mathcal{W}^\epsilon(\mathcal{S},y-x)=\frac{2}{\epsilon}\left(J^\epsilon(|y-x|)f'\left(|y-x|\mathcal{S}^2\right)\mathcal{S}\right). \label{forcestate} \end{eqnarray} The force between points $y$ and $x$ begins to drop when the strain $\mathcal{S}$ exceeds the critical strain \begin{eqnarray}
|\mathcal{S}|>\frac{\overline{r}}{\sqrt{|y-x|}}=\mathcal{S}_c, \label{sqrtsingularity} \end{eqnarray} see Figure \ref{SofteningBond}. This is the same singularity strength associated with a strain concentration in the vicinity of a crack tip as in the classic theory of brittle fracture \cite{Freund}. A future goal will be to inform the cohesive model introduced here by resolving atomistic or molecular dynamics across smaller length scales.
We apply the principle of least action to recover the cohesive equation of motion describing the state of displacement inside the body $D\subset\mathbb{R}^d$ given by \begin{eqnarray}
\rho\partial_{tt}^2 u(t,x)=2\frac{1}{V_d}\int_{\mathcal{H}_\epsilon(x)}\,\partial_\mathcal{S} \mathcal{W}^\epsilon(\mathcal{S},y-x)\frac{y-x}{|y-x|}\,dy+b(t,x) \label{eqofmotion} \end{eqnarray}
where $\rho$ is the density and $b(t,x)$ is the body force. This is a well posed formulation in that existence and uniqueness (within a suitable class of evolutions) can be shown, see Section \ref{sec2} and Theorem \ref{existenceuniqueness} of Section \ref{EE}. In this model a more complete set of physical properties including elastic and softening behavior are assigned to each point in the medium. Here each point in the domain is connected to its neighbors by a cohesive law see Figure \ref{SofteningBond}. We define the {\em process zone} to be the collection of points $x$ inside the body $D$ associated with peridynamic neighborhoods $\mathcal{H}_\epsilon(x)$ for which the strain $\mathcal{S}$ between $x$ and $y$ exceeds a threshold value for a sufficiently large proportion of points $y$ inside $\mathcal{H}_\epsilon(x)$. Here the force vs. strain law departs from linear behavior when the strain exceeds the threshold value. The mathematically precise definition of the process zone is given in Section \ref{sec4}, see Definition \ref{processZone}. In this model the {\em fracture set} is associated with peridynamic neighborhoods $\mathcal{H}_\epsilon(x)$ with strains $|\mathcal{S}|>\mathcal{S}_c$ for which the force vs. strain law begins to soften and is defined in Section \ref{sec4}, see Definition \ref{Fractureset}. The nonlinear elastic--softening behavior put forth in this paper is similar to the ones used in cohesive zone models \cite{Dougdale}, \cite{Barenblatt}. However for this model the dynamics selects whether a material point lies inside or outside the process zone. The principle feature of the cohesive dynamics model introduced here is that the evolution of the process zone together with the the fracture set is goverened by one equation consistent with Newton's second law given by \eqref{eqofmotion}. This is a characteristic feature of peridynamic models \cite{Silling1}, \cite{States} and lattice formulations for fracture evolution \cite{6,16,17,21}.
In this paper the goal is to characterize the size of the process zone for cohesive dynamics as a function of domain size and the length scale of the nonlocal forces. The second focus is to identify properties of the distinguished limit evolution for these models in the limit of vanishing non-locality as characterized by the $\epsilon\rightarrow 0$ limit. For the model introduced here the parameter that controls the size of the process zone is given by the radius of the horizon $\epsilon$. We derive an explicit inequality that shows that the size of the process zone is controlled by the horizon radius. Perturbation analysis shows that jump discontinuities within the process zone can become unstable and grow. This analysis shows that {\em the horizon size $\epsilon$ for cohesive dynamics is a modeling parameter} that can be calibrated according to the size of the process zone obtained from experimental measurements.
Further calculation shows that the volume of the process zone goes to zero with $\epsilon$ in the limit of vanishing non-locality, $\epsilon\rightarrow 0$. Distinguished $\epsilon\rightarrow 0$ limits of cohesive evolutions are identified and are found to have both bounded linear elastic energy and Griffith surface energy. Here the limit dynamics corresponds to the simultaneous evolution of linear elastic displacement and a fracture set across which the displacement is discontinuous. Under suitable hypotheses it is found that for points in space-time away from the fracture set that the displacement field evolves according to the linear elastic wave equation. Here the linear wave equation provides a dynamic coupling between elastic waves and the evolving fracture path inside the media. The elastic moduli, wave speed and energy release rate for the evolution are explicitly recovered from moments of the peridynamic potential energy. These features are consistent with the asymptotic behavior seen in the convergence of solutions of the Barenblatt model to the Griffith model when cohesive forces confined to a surface act over a sufficiently short range \cite{MarigoTruskinovsky}, \cite{Willis}.
Earlier work has shown that linear peridynamic formulations recover the classic linear elastic wave equation in the limit of vanishing non-locality see \cite{EmmrichWeckner}, \cite{SillingLehoucq}. The convergence of linear peridynamics to Navier's equation in the sense of solution operators is demonstrated in \cite{MengeshaDu}. Recent work shows that analogous results can be found for dynamic problems and fully nonlinear peridynamics \cite{LiptonJElast2014} in the context of antiplane shear. There distinguished $\epsilon\rightarrow 0$ limits of cohesive evolutions are identified and are found to have both bounded linear elastic energy and Griffith surface energy. It is shown that the limiting displacement evolves according to the linear elastic wave equation away from the crack set, see \cite{LiptonJElast2014}. For large deformations, the connection between hyperelastic energies and the small horizon limits of nonlinear peridynamic energies is recently established in \cite{Bellido}. In the current paper both two and three dimensional problems involving multi-mode fracture are addressed. For these problems new methods are required to identify the existence of a limit dynamics as the length scale of nonlocal interaction $\epsilon$ goes to zero. A crucial step is to establish a suitable notion of compactness for sequences of cohesive evolutions. The approach taken here employs nonlocal Korn inequalities introduced in \cite{DuGunzbergerlehoucqmengesha}. This method is presented in Section \ref{CC}. We conclude noting that the cohesive dynamics model introduced here does not have an irreversibility constraint and that the constitutive law \eqref{forcestate} applies at all times in the fracture evolution. However with this caveat in mind, the nonlocal cohesive model offers new computational and analytical opportunities for understanding the effect of the process zone on fracture patterns.
In the next section we write down the Lagrangian formulation for the cohesive dynamics and apply the principle of least action to recover the equation of motion. In that section it is shown that the nonlinear-nonlocal cohesive evolution is a well posed initial boundary value problem. It is also shown that energy balance is satisfied by the cohesive dynamics. A formal stability analysis is carried out in Section \ref{sec3} showing that jump discontinuities within the process zone can become unstable and grow, see Proposition \ref{Nuccriteria}. In Section \ref{sec4} we provide a mathematically rigorous inequality explicitly showing how the volume of the process zone for the cohesive evolutions is controlled by the length scale of nonlocal interaction $\epsilon$, see Theorem \ref{epsiloncontropprocesszone}. It is shown that the process zone concentrates on a set of zero volume in the limit, $\epsilon\rightarrow 0$, see Theorem \ref{bondunstable}. In Sections \ref{sec5} and \ref{sec6} we introduce suitable technical hypothesis and identify the distinguished limit of the cohesive evolutions as $\epsilon\rightarrow 0$, see Theorem \ref{LimitflowThm}. It is shown that the dynamics can be expressed in terms of displacements that satisfy the linear elastic wave equation away from the crack set, see Theorem \ref{waveequation}. These displacements are shown to have bounded bulk elastic and surface energy in the sense of Linear Elastic Fracture Mechanics (LEFM), see Theorem \ref{LEFMMThm}. In Section \ref{EightIntro} we provide the mathematical underpinnings and proofs of the theorems. In Section \ref{Ninth} we apply the approach developed here to examine limits of dynamics associated with other energies that $\Gamma$- converge to the Griffith fracture energy. As an illustrative example we examine the Ambrosio-Tortorelli \cite{AT} approximation as applied to the dynamic problem in \cite{BourdinLarsenRichardson} and \cite{LarsenOrtnerSuli}.
\setcounter{equation}{0} \setcounter{theorem}{0} \setcounter{lemma}{0}\setcounter{proposition}{0}\setcounter{remark}{0}\setcounter{remark}{0}\setcounter{definition}{0}\setcounter{hypothesis}{0}
\section{Cohesive dynamics} \label{sec2}
We formulate the initial boundary value problem for the cohesive evolution. Since the problem is nonlocal the domain $D$ is split into a boundary layer called the constraint set $D_c$, and the interior $D_s$. To fix ideas the thickness of the boundary layer is denoted by $\alpha$ and $2\epsilon<\alpha$ where $2\epsilon$ is the diameter of nonlocal interaction see Figure \ref{Domains}. The boundary condition for the displacement $u$ is given by $u(t,x)=0$ for $x$ in $D_c$. To incorporate nonlocal boundary conditions we introduce the space $L^2_0(D;\mathbb{R}^d)$, of displacements that are square integrable over $D$ and zero on $D_c$. The initial conditions for the cohesive dynamics belong to $L^2_0(D;\mathbb{R}^d)$ and are given by \begin{eqnarray} u(x,0)=u_0(x),\hbox{ and }u_t(x,0)=v_0(x). \label{initialconditions} \end{eqnarray}
We will investigate the evolution of the deforming domain for general initial conditions. These can include an initially un-cracked body or one with a preexisting system of cracks. For two dimensional problems the cracks are given by a system of curves of finite total length, while for three dimensional problems the crack set is given by a system of surfaces of finite total surface area. Depending on the dimension of the problem the displacement suffers a finite jump discontinuity across each curve or surface. The initial condition is specified by a crack set $K$ and displacement $u_0$. The strain $\mathcal{E}u_0=(\nabla u_0+\nabla u_0^T)/2$ is defined off the crack set and the displacement $u_0$ can suffer jumps across $K$. Griffith's theory of fracture asserts that the energy necessary to produce a crack $K$ is proportional to the crack length (or surface area). For Linear Elastic Fracture Mechanics (LEFM) the total energy associated with bulk elastic and surface energy is given by \begin{eqnarray}
LEFM(u_0)=\int_{D}\left(2\mu|\mathcal{E}u_0|^2+\lambda|{\rm div}\,u_0|^2\right)\,dx+\mathcal{G}_c |K|, \label{Gcrackenergy} \end{eqnarray}
where $\mu$, $\lambda$ are the the shear and Lam\'e moduli and $\mathcal{G}_c$ is the critical energy release rate for the material. Here $|K|$ denotes the length or surface area of the crack. In what follows we will assume that the bulk elastic energy and surface energy of the initial displacement are bounded as well as the the initial velocity and displacement. For future reference we describe initial data $u_0$ and $v_0$ that satisfy these conditions as {\em LEFM initial data} and we have the inequality between the peridynamic energy and the energy of Linear Elastic Fracture Mechanics given by \begin{eqnarray} PD^\epsilon(u_0)\leq LEFM(u_0), \label{basicinequality} \end{eqnarray} when $\mu$, $\lambda$, and $\mathcal{G}_c$ are related to the nonlocal potentials according to \eqref{calibrate1} and \eqref{calibrate2}. This inequality is established in Section \ref{CC}, see \eqref{upperboundperi}.
In what follows we write $u(t,x)$ as $u(t)$ to expedite the presentation. The cohesive dynamics is described by the Lagrangian \begin{eqnarray} L^\epsilon(u(t),\partial_t u(t),t)=K(\partial_t u(t))-PD^\epsilon(u(t))+U(u(t)), \label{Lagrangian} \end{eqnarray} with \begin{eqnarray}
K(\partial_t u(t))&=&\frac{1}{2}\int_{D}\rho|\partial_t u(t,x)|^2\,dx, \hbox{ and }\nonumber\\ U(u(t))&=&\int_{D}b(t,x) u(t,x)\,dx, \label{Components} \end{eqnarray} where $\rho$ is the mass density of the material and $b(t,x)$ is the body force density. The initial conditions $u^\epsilon(0,x)=u_0(x)$ and $u^\epsilon_t(0,x)=v_0(x)$ are prescribed and the action integral for the peridynamic evolution is \begin{eqnarray} I^\epsilon(u)=\int_0^TL^\epsilon(u(t),\partial_t u(t),t)\,dt. \label{Action} \end{eqnarray} The Euler Lagrange Equation for this system delivers the the cohesive dynamics described by \begin{eqnarray} \rho u^\epsilon_{tt}&=&-\nabla PD^\epsilon(u^\epsilon)+b \label{stationary} \end{eqnarray} where \begin{eqnarray}
\nabla PD^\epsilon(u^\epsilon)=-\frac{2}{V_d}\int_{\mathcal{H}_\epsilon(x)}\partial_\mathcal{S} \mathcal{W}^\epsilon(\mathcal{S},y-x)\frac{y-x}{|y-x|}\,dy, \label{GradPD} \end{eqnarray} and \begin{eqnarray}
\mathcal{S}=\frac{u^\epsilon(y)-u^\epsilon(x)}{|y-x|}\cdot e. \label{sepsilon} \end{eqnarray} The displacement $u^\epsilon(t,x)$ is twice differentiable in time taking values in $L^2_0(D;\mathbb{R}^d)$. The space of such functions is denoted by $C^2([0,T];L^2_0(D;\mathbb{R}^d))$. The initial value problem for the peridynamic evolution \eqref{stationary} is seen to have a unique solution in this space, see Theorem \ref{existenceuniqueness} of Section \ref{EE}. The cohesive evolution $u^\epsilon(x,t)$ is uniformly bounded in the mean square norm over bounded time intervals $0<t<T$, i.e., \begin{eqnarray} \max_{0<t<T}\left\{\Vert u^\epsilon(x,t)\Vert_{L^2(D;\mathbb{R}^d)}^2\right\}<C. \label{bounds} \end{eqnarray}
Here $\Vert u^\epsilon(x,t)\Vert_{L^2(D;\mathbb{R}^d)}=\left(\int_D|u^\epsilon(x,t)|^2\,dx\right)^{1/2}$ and the upper bound $C$ is independent of $\epsilon$ and depends only on the initial conditions and body force applied up to time $T$, see Section \ref{GKP}.
\begin{figure}
\caption{{\bf Domain $D=D_c\cup D_s$.}}
\label{Domains}
\end{figure}
The cohesive evolution has the following properties that are established in Section \ref{GKP}. The evolution has uniformly bounded kinetic and elastic potential energy \begin{theorem} \label{Gronwall} {\rm \bf Bounds on kinetic and potential energy for cohesive dynamics}\\ There exists a positive constant $C$ depending only on $T$ and independent of $\epsilon$ for which \begin{eqnarray} \sup_{0\leq t\leq T}\left\{PD^{\epsilon}(u^{\epsilon}(t))+\frac{\rho}{2}\Vert u_t^{\epsilon}(t)\Vert^2_{L^2(D;\mathbb{R}^d)}\right\}\leq C. \label{boundenergy} \end{eqnarray} \end{theorem}
The evolution is uniformly continuous in time as measured by the mean square norm. \begin{theorem}{\rm \bf Continuous cohesive evolution in mean square norm}\\ There is a positive constant $K$ independent of $t_2 < t_1$ in $[0,T]$ and index $\epsilon$ for which \begin{eqnarray}
\Vert u^{\epsilon}(t_1)-u^{\epsilon}(t_2)\Vert_{L^2(D;\mathbb{R}^d)}\leq K |t_1-t_2|. \label{holderest} \end{eqnarray} \label{holdercont} \end{theorem}
The evolution satisfies energy balance. The total energy of the cohesive evolution at time $t$ is given by \begin{eqnarray} \mathcal{EPD}^\epsilon(t,u^\epsilon(t))=\frac{\rho}{2}\Vert u^\epsilon_t(t)\Vert^2_{L^2(D;\mathbb{R}^d)}+PD^\epsilon(u^\epsilon(t))-\int_{D}b(t)\cdot u^\epsilon(t)\,dx \label{energyt} \end{eqnarray} and the total energy of the system at time $t=0$ is \begin{eqnarray} \mathcal{EPD}^\epsilon(0,u^\epsilon(0))=\frac{\rho}{2}\Vert v_0\Vert^2_{L^2(D;\mathbb{R}^d)}+PD^\epsilon(u_0)-\int_{D}b(0)\cdot u_0\,dx. \label{energy0} \end{eqnarray} The cohesive dynamics is seen to satisfy energy balance at every instant of the evolution. \begin{theorem}{\rm \bf Energy balance for cohesive dynamics}\\ \label{Ebalance} \begin{eqnarray} \mathcal{EPD}^\epsilon(t,u^\epsilon(t))=\mathcal{EPD}^\epsilon(0,u^\epsilon(0))-\int_0^t\int_{D} b_t(\tau)\cdot u^\epsilon(\tau)\,dx\,d\tau.\label{BalanceEnergy} \end{eqnarray} \end{theorem}
\setcounter{equation}{0} \setcounter{theorem}{0} \setcounter{lemma}{0}\setcounter{proposition}{0}\setcounter{remark}{0}\setcounter{remark}{0}\setcounter{definition}{0}\setcounter{hypothesis}{0}
\section{Dynamic instability and fracture nucleation} \label{sec3} In this section we present a fracture nucleation condition that arises from the unstable force law \eqref{forcestate}. This condition is manifested as a dynamic instability. In the following companion section we investigate the localization of dynamic instability as $\epsilon_k\rightarrow 0$ and define the notion of process zone for the cohesive evolution. Fracture nucleation conditions can be viewed as instabilities and have been identified for peridynamic evolutions in \cite{SillingWecknerBobaru}. Fracture nucleation criteria formulated as instabilities for one dimensional peridynamic bars are developed in \cite{WecknerAbeyaratne}. In this treatment we define a source for crack nucleation as jump discontinuity in the displacement field that can become unstable and grow in time. Here we establish a direct link between the growth of jump discontinuities and the appearance of strain concentrations inside the deforming body.
We proceed with a formal perturbation analysis and consider a time independent body force density $b$ and a smooth equilibrium solution $u$ of \eqref{stationary}. Now perturb $u$ in the neighborhood of a point $x$ by adding a piecewise smooth discontinuity denoted by the vector field $\delta$. The perturbation takes the value zero on one side of a plane with normal vector $\nu$ passing through $x$ and on the other side of the plane takes the value $\delta=\overline{u}s(t)$. Here $s(t)$ is a scalar function of time and $\overline{u}$ is a constant vector. Consider the neighborhood $\mathcal{H}_\epsilon(x)$, then $\delta(y)=0$ for $(y-x)\cdot\nu<0$ and $\delta(y)=\overline{u}s(t)$ for $(y-x)\cdot\nu\geq 0$, see Figure \ref{plane}. The half space on the side of the plane for which $(y-x)\cdot\nu<0$ is denoted by $E^-_{\nu}$.
Write $u^p=u+\delta$ and assume \begin{eqnarray} \rho u^p_{tt}&=&-\nabla PD^\epsilon(u^p)+b.\label{stationarydiff1} \end{eqnarray} We regard $s(t)$ as a small perturbation and expand the integrand of $\nabla PD^\epsilon(u^p)$ in a Taylor series to recover the linearized evolution equation for the jump $s=s(t)$. The evolution equation is given by \begin{eqnarray} \rho s_{tt}\overline{u}=\mathcal{A}_{\nu}(x)\overline{u}s, \label{pertevolution} \end{eqnarray} where the stability matrix $\mathcal{A}_{\nu}(x)$ is a $d\times d$ symmetric matrix with real eigenvalues and is defined by \begin{eqnarray}
\mathcal{A}_{\nu}(x)&=&-\frac{2}{\epsilon V_d}\left\{\int_{\mathcal{H}_\epsilon(x)\cap E^-_{\nu}}\frac{1}{|y-x|}\partial^2_{\mathcal{S}}\mathcal{W}^\epsilon(\mathcal{S},y-x)\frac{y-x}{|y-x|}\otimes\frac{y-x}{|y-x|}dy\right\}, \label{instabilitymatrix} \end{eqnarray}
and $$\mathcal{S}=\mathcal{S}(y,x)=\left(\frac{u(y)-u(x)}{|y-x|}\right)\cdot\frac{y-x}{|y-x|}.$$ Calculation shows that \begin{eqnarray}
\partial^2_{\mathcal{S}}\mathcal{W}^\epsilon(\mathcal{S},y-x)=\frac{2}{\epsilon}J^\epsilon(|y-x|)\left(f'\left(|y|\mathcal{S}^2\right)+2f''\left(|y-x|\mathcal{S}^2\right)|y-x|\mathcal{S}^2\right),\label{expand} \end{eqnarray}
where $f'(|y|\mathcal{S}^2)>0$ and $f''(|y|\mathcal{S}^2)<0$. On writing \begin{eqnarray}
\mathcal{S}_c=\frac{\overline{r}}{\sqrt{|y-x|}} \label{critstrain} \end{eqnarray} we have that \begin{eqnarray}
\partial_\mathcal{S}^2\mathcal{W}^\epsilon(\mathcal{S},y)>0\hbox{ for }|\mathcal{S}(y,x)|<\mathcal{S}_c, \label{loss1} \end{eqnarray} and \begin{eqnarray}
\partial_\mathcal{S}^2W^\epsilon(\mathcal{S},y)<0\hbox{ for }|\mathcal{S}(y,x)|>\mathcal{S}_c. \label{loss2} \end{eqnarray} Here $\overline{r}$ is the inflection point for the function $r:\rightarrow f(r^2)$ and is the root of the equation \begin{eqnarray} f'(r)+2{r}f''({r})=0. \label{rootroverline} \end{eqnarray} Note that the critical strain $\mathcal{S}_c$ for which the cohesive force between a pair of points $y$ and $x$ begins to soften is akin to the square root singularity seen at the crack tip in classical brittle fracture mechanics.
For eigenvectors $\overline{u}$ in the eigenspace associated with positive eigenvalues $\lambda$ of $\mathcal{A}_\nu(x)$ one has \begin{eqnarray} \rho\partial^2_{tt}s(t)=\lambda s(t) \label{stabeq} \end{eqnarray} and the perturbation $s(t)$ can grow exponentially. Observe from \eqref{loss2} that the quadratic form \begin{eqnarray}
\mathcal{A}_{\nu}(x)\overline{w}\cdot\overline{w}=-\frac{2}{\epsilon V_d}\left\{\int_{\mathcal{H}_\epsilon(x)\cap E^-_{\nu}}\frac{1}{|y-x|}\partial^2_{\mathcal{S}}\mathcal{W}^\epsilon(\mathcal{S},y-x)(\frac{y-x}{|y-x|}\cdot\overline{w})^2dy\right\} \label{quadformunstable} \end{eqnarray} will have at least one positive eigenvalue provided a sufficiently large proportion of bonds $y-x$ inside the horizon have strains satisfying \begin{eqnarray}
|\mathcal{S}(x,y)|>\mathcal{S}_c \label{exceed} \end{eqnarray} for which the cohesive force is in the unstable phase. For this case we see that the jump can grow exponentially. The key feature here is that dynamic instability is explicitly linked to strain concentrations in this cohesive model as is seen from \eqref{loss2} together with \eqref{quadformunstable}. Collecting results we have the following proposition.
\begin{proposition}{\em \bf Fracture nucleation condition for cohesive dynamics}\\ \label{Nuccriteria}
A condition for crack nucleation at a point $x$ is that there is at least one direction $\nu$ for which $\mathcal{A}_{\nu}(x)$ has at least one positive eigenvalue. This occurs if there is a square root strain concentration $|\mathcal{S}(y,x)|>\mathcal{S}_c$ over a sufficiently large proportion of cohesive bonds inside the peridynamic horizon. \end{proposition} \noindent Proposition \ref{Nuccriteria} together with (\ref{loss2}) provide the explicit link between dynamic instability and the critical strain where the cohesive law begins to soften.
\begin{figure}
\caption{{\bf Jump discontinuity.}}
\label{plane}
\end{figure}
More generally we may postulate a condition for the direction along which the opposite faces of a nucleating fissure are oriented and the direction of the displacement jump across it. Recall that two symmetric matrices $A$ and $B$ satisfy $A\geq B$ in the sense of quadratic forms if $A\overline{w}\cdot\overline{w}\geq B\overline{w}\cdot\overline{w}$ for all $\overline{w}$ in $\mathbb{R}^d$. We say that a matrix $A$ is the maximum of a collection of symmetric matrices if $A\geq B$ for all matrices $B$ in the collection.
We postulate that the faces of the nucleating fissure are perpendicular to the direction $\nu^*$ associated with the the matrix $\mathcal{A}_{\nu^*}(x)$ for which \begin{eqnarray} \mathcal{A}_{\nu^*}(x)=\max\left\{\mathcal{A}_{\nu}(x);\, \hbox{over all directions } \nu\hbox{ such that } \mathcal{A}_{\nu}(x) \hbox{ has a positive eigenvalue}\right\}, \label{bestdirctionforgrowth} \end{eqnarray} and that the orientation of the jump in displacement across opposite sides of the fissure lies in the eigenspace associated with the largest positive eigenvalue of $\mathcal{A}_{\nu^*}$, {\em i.e., the fissure is oriented along the most unstable orientation and the displacement jump across the nucleating fissure is along the most unstable direction.}
\setcounter{equation}{0} \setcounter{theorem}{0} \setcounter{lemma}{0}\setcounter{proposition}{0}\setcounter{remark}{0}\setcounter{remark}{0}\setcounter{definition}{0}\setcounter{hypothesis}{0}
\section{The process zone for cohesive dynamics and its localization in the small horizon limit } \label{sec4}
In this section it is shown that the collection of centroids of peridynamic neighborhoods with strain exceeding a prescribed threshold concentrate on sets with zero volume in the limit of vanishing non-locality. In what follows we probe the dynamics to obtain mathematically rigorous and explicit estimates on the size of the process zone in terms of the radius of the peridynamic horizon $0<\epsilon<1$.
The continuity of the displacement inside the neighborhood $\mathcal{H}_\epsilon(x)$ is measured quantitatively by \begin{eqnarray}
|u(y)-u(x)|\leq\underline{k}\,|y-x|^\alpha, \hbox{ for $y\in \mathcal{H}_\epsilon(x)$}, \label{moduluscont} \end{eqnarray} with $0<\underline{k}$ and $0\leq\alpha\leq 1$. In what follows we focus on the reduction of continuity measured quantitatively by \begin{eqnarray}
|u(y)-u(x)|>\underline{k}\,|y-x|^\alpha \hbox{ for $y\in \mathcal{H}_\epsilon(x)$}. \label{modulusdiscont} \end{eqnarray}
Observe when \eqref{modulusdiscont} holds and $\alpha=1/2$ and $\underline{k}=\overline{r}$ then $|\mathcal{S}(y,x)|>\mathcal{S}_c$ and there is softening in the cohesive force-strain behavior given by \eqref{forcestate} .
We now consider solutions $u^\epsilon$ of \eqref{stationary} and define a mathematical notion of process zone based the strain exceeding threshold values associated with \eqref{modulusdiscont}.
The process zone is best described in terms of the basic unit of peridynamic interaction: the peridynamic neighborhoods $\mathcal{H}_\epsilon(x)$ of radius $\epsilon>0$ with centers $x\in D$. We fix a choice of $\underline{k}$ and $\alpha$ belonging to the intervals $0< \underline{k}\leq \overline{r}$ and $1/2\leq \alpha<1$. The strain between $x$ and a point $y$ inside the neighborhood is denoted by $\mathcal{S}^\epsilon(y,x)$. The collection of points $y$ inside $\mathcal{H}_\epsilon(x)$ for which the strain $|\mathcal{S}^\epsilon(y,x)|$ exceeds the threshold function $\underline{k}\,|y-x|^{\alpha-1}$ is denoted by
$\{y\hbox{ $ \in$ }\mathcal{H}_\epsilon(x):\, |\mathcal{S}^\epsilon(x,y)|>\underline{k}\,|y-x|^{\alpha-1}\}$. Note for $0<\underline{k}<\overline{r}$ and $1/2<\alpha<1$ that \begin{eqnarray}
&&\left\{y\hbox{ $\in$ }\mathcal{H}_\epsilon(x):\, |\mathcal{S}^\epsilon(y,x)|>\mathcal{S}_c\right\}\subset\{y\hbox{ $\in$ }\mathcal{H}_\epsilon(x):\, |\mathcal{S}^\epsilon(y,x)|>\frac{\underline{k}}{|y-x|^{1-\alpha}}\}.\label{nonlipshitz} \end{eqnarray}
The fraction of points inside the neighborhood $\mathcal{H}_\epsilon(x)$ with strains exceed the threshold is written \begin{eqnarray}
P\left(\{y\hbox{ in }\mathcal{H}_\epsilon(x):\, |\mathcal{S}^\epsilon(y,x)|>\underline{k}\,|y-x|^{\alpha-1}\}\right),\label{weight} \end{eqnarray} where the weighted volume fraction for any subset $B$ of $\mathcal{H}_\epsilon(x)$ is defined as \begin{eqnarray}
P(B)=\frac{1}{\epsilon^d m}\int_{B}\,(|y-x|/\epsilon)J(\vert y-x\vert/\epsilon)\,dy, \label{weightdefined} \end{eqnarray} with normalization constant \begin{eqnarray}
m=\int_{\mathcal{H}_1(0)}|\xi||J(|\xi|)\,d\xi \label{normalize} \end{eqnarray} chosen so that $P(\mathcal{H}_\epsilon(x))=1$.
\begin{definition}{\bf Process Zone.} \label{processZone}
Fix a volume fraction $0<\overline{\theta}\leq 1$, $0<\underline{k}\leq\overline{r}$, and $1/2\leq\alpha<1$ and at each time $t$ in the interval $0\leq t\leq T$, define the process zone $PZ^\epsilon(\underline{k},\alpha,\overline{\theta},t)$ to be the collection of centers of peridynamic neighborhoods for which the portion of points $y$ with strain $\mathcal{S}^\epsilon(t,y,x)$ exceeding the threshold $\underline{k}\,|y-x|^{\alpha-1}$ is greater than $\overline{\theta}$, i.e., $P\left(\{y\hbox{ in }\mathcal{H}_\epsilon(x):\, |\mathcal{S}^\epsilon(t,y,x)|>\underline{k}\,|y-x|^{\alpha-1}\}\right)>\overline{\theta}$. \end{definition}
The fracture set is defined to be the process zone for which strains exceed the threshold $\mathcal{S}_c$ and the force vs. strain curve begins to soften. \begin{definition}{\bf Fracture Set.} \label{Fractureset}
The fracture set is defined to be the process zone associated with the values ${\theta}=1/2$, $\underline{k}=\overline{r}$, and $\alpha=1/2$ and at each time $t$ in the interval $0\leq t\leq T$, and is defined by $PZ^\epsilon(\overline{r},1/2,1/2,t)$ to be the collection of centers of peridynamic neighborhoods for which the portion of points $y$ with strain $\mathcal{S}^\epsilon(t,y,x)$ exceeding the threshold $\mathcal{S}_c$ is greater than $1/2$, i.e., $P\left(\{y\hbox{ in }\mathcal{H}_\epsilon(x):\, |\mathcal{S}^\epsilon(t,y,x)|>\mathcal{S}_c\}\right)>1/2$. \end{definition} \noindent It is clear from the definition that the fracture set defined for this model contains the set of jump discontinuities for the displacement. This definition of fracture set given here is different that the usual one which collapses material damage onto a surface across which the displacement jumps.
It follows from Proposition \ref{Nuccriteria} that the process zone contains peridynamic neighborhoods associated with softening cohesive forces. Within this zone pre-existing jump discontinuities in the displacement field can grow. \begin{remark} Here we have described a range of process zones depending upon the choice of $\alpha$, $\underline{k}$ and $\overline{\theta}$. In what follows we show that for any choice of $\alpha$ in $1/2\leq \alpha <1$ and $\underline{k}$ in $0<\underline{k}\leq\overline{r}$ and $0<\overline{\theta}\leq 1$ the volume of the process zone is explicitly controlled by the radius of the peridynamic horizon $\epsilon$. \end{remark}
We consider problem formulations in two and three dimensions and the volume or area of a set is given by the $d$ dimensional Lebesgue measure denoted by $\mathcal{L}^d$, for $d=2,3$. We let \begin{eqnarray} \label{upbdconst} C(t)= \left( (2LEFM(u_0)+{\rho}\Vert v_0\Vert_{L^2(D;\mathbb{R}^d)}+1)^{1/2}+\sqrt{\rho^{-1}}\int_0^t\Vert b(\tau)\Vert_{L^2(D;\mathbb{R}^d)}\,d\tau\right)^2-1, \end{eqnarray} and note that $C(t)\leq C(T)$ for $t<T$.
We now give the following bound on the size of the process zone. \begin{theorem} {\bf Dependence of the process zone on the radius of the peridynamic horizon} \label{epsiloncontropprocesszone} \begin{eqnarray} \mathcal{L}^d\left(PZ^\epsilon(\underline{k},\alpha,\overline{\theta},t) \right)\leq \frac{\epsilon^{1-\beta}}{\overline{\theta}\underline{k}^2(f'(0)+o(\epsilon^\beta))}\times \frac{C(t)}{2m}, \label{controlbyepsilon} \end{eqnarray} where $0\leq\beta<1$ and $\beta=2\alpha-1$ and $0\leq t\leq T$. \end{theorem} \noindent Theorem \ref{epsiloncontropprocesszone} explicitly shows that the size of the process zone is controlled by the radius $\epsilon$ of the peridynamic horizon, uniformly in time. This theorem is proved in Section \ref{proofbondunstable}.
\begin{remark} \label{ModelParameter} This analysis shows that {\em the horizon size $\epsilon$ for cohesive dynamics is a modeling parameter} that may be calibrated according to the size of the process zone obtained from experimental measurements. \end{remark}
Next we show how the process zone localizes and concentrates on sets with zero volume in the small horizon limit. To proceed choose $\delta>0$ and consider the sequence of solutions $u^{\epsilon_k}(t,x)$ to the cohesive dynamics for a family of radii $\epsilon_k=\frac{1}{2^k}$, $k=1,\ldots$. The set of centers $x$ of neighborhoods $\mathcal{H}_{\epsilon_k}(x)$ that belong to at least one of the process zones $PZ^{\epsilon_k}(\underline{k},\alpha,\overline{\theta},t)$ for some $\epsilon_k<\delta$ at time $t$ is denoted by $CPZ^\delta(\underline{k},\alpha,\overline{\theta},t)$. Let $CPZ^0(\underline{k},\alpha,\overline{\theta},t)=\cap_{0<\delta}CPZ^\delta(\underline{k},\alpha,\overline{\theta},t)$ be the collection of centers of neighborhoods such that for every $\delta>0$ they belong to a process zone $PZ^{\epsilon_k}(\underline{k},\alpha,\overline{\theta},t)$ for some $\epsilon_k<\delta$. The localization and concentration of the process zone is formulated in the following theorem. \begin{theorem}{\rm\bf Localization of process zone in the small horizon limit.}\\ \label{bondunstable} The collection of process zones $CPZ^\delta(\underline{k},\alpha,\overline{\theta},t)$ is decreasing with $\delta\rightarrow 0$ and there is a positive constant $K$ independent of $t$ and $\delta$ for which \begin{eqnarray} && \mathcal{L}^d\left(CPZ^\delta(\underline{k},\alpha,\overline{\theta},t)\right)\leq K{\delta^{1-\beta}}, \hbox{ for, } 0\leq t\leq T,\,\,0<\beta=2\alpha-1\leq 1, \hbox{ with }\nonumber\\ && \mathcal{L}^d\left(CPZ^0(\underline{k},\alpha,\overline{\theta},t)\right)=\lim_{\delta\rightarrow 0}\mathcal{L}^d\left(CPZ^\delta(\underline{k},\alpha,\overline{\theta},t)\right)=0. \label{limdelta} \end{eqnarray} For any choice of $0<\overline{\theta}\leq1$ the collection of centers of neighborhoods for which there exists a positive $\delta$ such that \begin{eqnarray}
P\left(\{y\hbox{ in }H_{\epsilon_k}(x):\, |\mathcal{S}^{\epsilon_k}(t,y,x)|\leq\underline{k}\,|y-x|^{\alpha-1}\}\right)\geq1-\overline{\theta}, \label{contolledstrain} \end{eqnarray} for all $\epsilon_k<\delta$ is a set of full measure for every choice of $0<\underline{k}\leq\overline{r}$ and $1/2\leq\alpha<1$, i.e., $\mathcal{L}^d(D)=\mathcal{L}^d(D\setminus CPZ^0(\underline{k},\alpha,\overline{\theta},t))$. \end{theorem}
\begin{remark} \label{nearlipschitz} The theorem shows that the process zone concentrates on a set of zero volume in the limit of vanishing peridynamic horizon. Note \eqref{contolledstrain} holds for any $0<\overline{\theta}\leq 1$. On choosing $\overline{\theta}\approx 0$, and $\alpha\approx 1$ it is evident that the modulus of continuity for displacement field is close to Lipschitz outside of the process zone in the limit of vanishing nonolcality, $\epsilon_k\rightarrow 0$. The concentration of the process zone is inevitable for the cohesive model and is directly linked to the constraint on the energy budget associated with the cohesive dynamics as described by Theorem \ref{Gronwall}. This bound forces the localization of the process zone as shown in Section \ref{proofbondunstable}. \end{remark} \setcounter{equation}{0} \setcounter{theorem}{0} \setcounter{lemma}{0}\setcounter{proposition}{0}\setcounter{remark}{0}\setcounter{remark}{0}\setcounter{definition}{0}\setcounter{hypothesis}{0}
\section{The small horizon limit of cohesive dynamics}
\label{sec5}
In this section we identify the distinguished small horizon $\epsilon\rightarrow 0$ limit for cohesive dynamics. It is shown here that the limit dynamics has bounded bulk linear elastic energy and Griffith surface energy characterized by the shear modului $\mu$, Lam\'e modulus $\lambda$, and energy release rate $\mathcal{G}_c$ respectively. In order to make the connection between the limit dynamics and cohesive dynamics we will identify the relationship between the potentials $W^\epsilon(\mathcal{S},y-x)$ and the triple $\mu$, $\lambda$, $\mathcal{G}_c$.
To reveal this connection consider a family of cohesive evolutions $u^{\epsilon_k}$, each associated with a fixed potential $W^{\epsilon_k}$ and horizon length $\epsilon_k$, with $k=1,2,\ldots$ and $\epsilon_k\rightarrow 0$. Each $u^{\epsilon_k}(x,t)$ can be thought of as being the result of a perfectly accurate numerical simulation of a cohesive evolution associated with the potential $W^{\epsilon_k}$. It is shown in this section that the cohesive dynamics $u^{\epsilon_k}(x,t)$ converges to a limit evolution $u^0(x,t)$ in the limit, $\epsilon_k\rightarrow 0$. The limit evolution describes the dynamics of the cracked body when the scale of nonlocality is infinitesimally small with respect to the material specimen. Here the limiting free crack evolution is mediated through the triple $\mu$, $\lambda$, and $\mathcal{G}_c$ that are described by explicit formulas associated with the sequence of potentials $W^{\epsilon_k}$, see \eqref{calibrate1}, \eqref{calibrate2} and Theorem \ref{LEFMMThm} below.
\noindent {\em It is of fundamental importance to emphasize that we do do not impose a-priori relations between $W^{\epsilon_k}$ and the triple $\mu$, $\lambda$, and $\mathcal{G}_c$; we show instead that the cohesive dynamics $u^{\epsilon_k}(x,t)$ approaches the limit dynamics $u^0(x,t)$ characterized by $\mu$, $\lambda$, and $\mathcal{G}_c$ given by the formulas \eqref{calibrate1} and \eqref{calibrate2} in the limit when $\epsilon_k\rightarrow 0$} .
\noindent In what follows the sequence of cohesive dynamics described by $u^{\epsilon_k}$ is seen to converge to the limiting free crack evolution $u^0(x,t)$ in mean square, uniformly in time, see Theorem \ref{LimitflowThm}. The limit evolution is shown to have the following properties: \begin{itemize} \item It has uniformly bounded energy in the sense of linear elastic fracture mechanics for $0\leq t \leq T$. \item It satisfies an energy inequality involving the kinetic energy of the motion together with the bulk elastic and surface energy associated with linear elastic fracture mechanics for $0\leq t\leq T$. \end{itemize}
\noindent We provide explicit conditions under which these properties are realized for the limit dynamics. \begin{hypothesis} \label{remarkone} We suppose that the magnitude of the displacements $u^{\epsilon_k}$ for cohesive dynamics are bounded for $0\leq t\leq T$ uniformly in $\epsilon_k$, i.e., $\sup_{\epsilon_k}\sup_{0\leq t\leq T}\Vert u^{\epsilon_k}(t)\Vert_{L^\infty(D;\mathbb{R}^d)}<\infty$. \end{hypothesis}
The convergence of cohesive dynamics is given by the following theorem, \begin{theorem} \label{LimitflowThm} {\rm\bf Convergence of cohesive dynamics}\\ For each $\epsilon_k$ we prescribe identical LEFM initial data $u_0(x)$ and $v_0(x)$ and the solution to the cohesive dynamics initial value problem is denoted by $u^{\epsilon_k}$. Now consider a sequence of solutions $u^{\epsilon_k}$ associated with a vanishing peridynamic horizon $\epsilon_k\rightarrow 0$ and suppose Hypothesis \ref{remarkone} holds true. Then, on passing to a subsequence if necessary, the cohesive evolutions $u^{\epsilon_k}$ converge in mean square uniformly in time to a limit evolution $u^0$ with the same LEFM initial data, i.e., \begin{eqnarray} \lim_{\epsilon_k\rightarrow 0}\max_{0\leq t\leq T}\left\{\Vert u^{\epsilon_k}(t)-u^0(t)\Vert_{L^2(D;\mathbb{R}^d)}\right\}=0 \label{unifconvg} \end{eqnarray} and $u^0(x,0)=u_0(x)$ and $\partial_t u^0(x,0)=v_0(x)$. \end{theorem}
To appropriately characterize the LEFM energy for the limit dynamics with freely propagating cracks one needs a generalization of the strain tensor. The appropriate notion of displacement and strain useful for problems involving discontinuities is provided by functions of bounded deformation BD introduced in \cite{Matthies}, \cite{Suquet}. The subspace of BD given by the special functions of bounded deformation $SBD$ introduced in \cite{Bellettini} is appropriate for describing discontinuities associated with linear elastic fracture. Functions in $u$ SBD belong to $L^1(D;\mathbb{R}^d)$ and are approximately continuous, i.e., have Lebesgue limits for almost every $x\in D$ given by \begin{eqnarray}
\lim_{r\searrow 0}\frac{1}{\pi r^d}\int_{B(x,r)}\,|u(y)-u(x)|\,dy=0, \hbox{ $d=2,3$} \label{approx} \end{eqnarray} where $B(x,r)$ is the ball of radius $r$ centered at $x$. The jump set $J_{u}$ for elements of $SBD$ is defined to be the set of points of discontinuity which have two different one sided Lebesgue limits. One sided Lebesgue limits of $u$ with respect to a direction $\nu_u(x)$ are denoted by $u^-(x)$, $u^+(x)$ and are given by \begin{eqnarray}
\lim_{r\searrow 0}\frac{1}{\pi r^d}\int_{B^-(x,r)}\,|u(y)-u^-(x)|\,dy=0,\,\,\, \lim_{r\searrow 0}\frac{1}{\pi r^d}\int_{B^+(x,r)}\,|u(y)-u^+(x)|\,dy=0, \hbox{$d=2,3$}, \label{approxjump} \end{eqnarray} where $B^-(x,r)$ and $B^+(x,r)$ are given by the intersection of $B(x,r)$ with the half spaces $(y-x)\cdot \nu_u(x)<0$ and $(y-x)\cdot \nu_u(x)>0$ respectively. SBD functions have jump sets $J_u$, described by a countable number of components $K_1,K_2,\ldots$, contained within smooth manifolds, with the exception of a set $K_0$ that has zero $d-1$ dimensional Hausdorff measure \cite{AmbrosioCosicaDalmaso}. Here the notion of arc length or (surface area) is the $d-1$ dimensional Hausdorff measure of $J_{u}$ and $\mathcal{H}^{d-1}(J_{u})=\sum_i\mathcal{H}^{d-1}(K_i)$. The strain \cite{AmbrosioCosicaDalmaso} of a displacement $u$ belonging to SBD, written as $\mathcal{E}u$, is a generalization of the classic strain tensor and satisfies the property \begin{eqnarray}
\lim_{r\searrow 0}\frac{1}{\pi r^d}\int_{B(x,r)}\,\frac{|(u(t,y)-u(t,x)-\mathcal{E}u(t,x)(y-x))\cdot(y-x)|}{|y-x|^2}\,dy=0, \hbox{ $d=2,3$} \label{appgrad} \end{eqnarray} for almost every $x\in D$, with respect to $d$-dimensional Lebesgue measure $\mathcal{L}^d$. The symmetric part of the distributional derivative of $u$, $E u=1/2(\nabla u+\nabla u^T)$ for $SBD$ functions is a $d\times d$ matrix valued Radon measure with absolutely continuous part described by the density $\mathcal{E}u$ and singular part described by the jump set \cite{AmbrosioCosicaDalmaso}, \cite{Bellettini} and \begin{eqnarray} \langle E u,\Phi\rangle=\int_D\,\sum_{i,j=1}^d\mathcal{E}u_{ij}\Phi_{ij}\,dx+\int_{J_{u}}\,\sum_{i,j=1}^d(u^+_i - u^-_i)\nu_j\Phi_{ij}\,d\mathcal{H}^{d-1}, \label{distderiv} \end{eqnarray} for every continuous, symmetric matrix valued test function $\Phi$. A description of $BD$ functions including their fine properties and structure, together with the characterization of $SBD$ functions on slices is developed in \cite{AmbrosioCosicaDalmaso} and \cite{Bellettini}.
The energy of linear elastic fracture mechanics extended to the class of $SBD$ functions is given by: \begin{eqnarray}
LFEM(u,D)=\int_{D}\left(2\mu |\mathcal{E} u|^2+\lambda |{\rm div}\,u|^2\right)\,dx+\mathcal{G}_c\mathcal{H}^{d-1}(J_{u}), \hbox{ $d=2,3$,} \label{LEFMSBVDefinition} \end{eqnarray} for $u$ belonging to $SBD$. We now describe the elastic energy for the limit dynamics. \begin{theorem} {\rm\bf The limit dynamics has bounded LEFM energy}\\ The limit evolution $u^0$ belongs to SBD for every $t\in[0,T]$. Furthermore there exists a constant $C$ depending only on $T$ bounding the LEFM energy, i.e., \begin{eqnarray}
\int_{D}\,2\mu |\mathcal{E} u^0(t)|^2+\lambda |{\rm div}\,u^0(t)|^2\,dx+\mathcal{G}_c\mathcal{H}^{d-1}(J_{u^0(t)})\leq C, \hbox{ $d=2,3$,} \label{LEFMbound} \end{eqnarray} for $0\leq t\leq T$. Here $\mu$, $\lambda$, and $\mathcal{G}_c$ are given by the explicit formulas \begin{eqnarray} \mu=\lambda=\frac{1}{4} f'(0)\int_{0}^1r^2J(r)dr, \hbox{ $d=2$}&\hbox{ and }& \mu=\lambda=\frac{1}{5} f'(0)\int_{0}^1r^3J(r)dr, \hbox{ $d=3$}\label{calibrate1} \end{eqnarray} and \begin{eqnarray} \mathcal{G}_c=\frac{2\omega_{d-1}}{\omega_d}\, f_\infty \int_{0}^1r^dJ(r)dr, \hbox{ for $d=2.3$} \label{calibrate2} \end{eqnarray} where $f_\infty$ is defined by \eqref{properties} and $\omega_{n}$ is the volume of the $n$ dimensional unit ball, $\omega_1=2,\omega_2=\pi,\omega_3=4\pi/3$. The potential $f$ and influence function $J$ can always be chosen to satisfy \eqref{calibrate1} and \eqref{calibrate2} for any $\mu=\lambda>0$ corresponding to the Poisson ratio $\nu=1/3$, for $d=2$ and $\nu=1/4$, for $d=3$, and $\mathcal{G}_c>0$. \label{LEFMMThm} \end{theorem}
\begin{remark} \label{boundedhausdorffmeasure} The absolutely continuous part of the strain $\mathcal{E}u^0$ is defined for points away from the jump set $J_{u^0}$ and in this sense the process zone for the limit evolution can be viewed as being confined to the jump set $J_{u^0}$ . Theorem \ref{LEFMMThm} shows that the jump set $J_{u^0}$ for the limit evolution $u^0(t,x)$ is confined to a set of finite $d-1$ dimensional Hausdorff measure. \end{remark}
We now present an energy inequality for the limit evolution. The sum of energy and work for the displacement $u^0$ at time $t$ is written \begin{eqnarray} \mathcal{GF}(u^0(t),D)=\frac{\rho}{2}\Vert u_t^0(t)\Vert^2_{L^2(D;\mathbb{R}^d)}+LEFM(u^0(t),D)-\int_{D}b(t)\cdot u^0(t)\,dx. \label{sumtt} \end{eqnarray} The sum of energy and work for the initial data $u_0,v_0$ is written \begin{eqnarray} \mathcal{GF}(u_0,D)=\frac{\rho}{2}\Vert v_0\Vert^2_{L^2(D;\mathbb{R}^d)}+LEFM(u_0,D)-\int_{D}b(0)\cdot u_0\,dx. \label{sumt0} \end{eqnarray} The energy inequality for the limit evolution $u^0$ is given by, \begin{theorem} {\rm \bf Energy Inequality}\\ \label{energyinequality} For almost every $t$ in $[0, T]$, \begin{eqnarray} \mathcal{GF}(u^0(t),D)\leq\mathcal{GF}(u_0,D)-\int_0^t\int_{D} b_t(\tau) \cdot u^0(\tau)\,dx\,d\tau. \label{enegineq} \end{eqnarray} \end{theorem}
\begin{remark} \label{remarkfinal} The equality $\lambda=\mu$ appearing in Theorem \ref {LEFMMThm} is a consequence of the central force nature of the local cohesive interaction mediated by \eqref{forcestate}. More general non-central interactions are proposed in Section 15 of \cite{Silling1} and in the state based peridynamic formulation \cite{States}. The non-central formulations deliver a larger class of energy-volume-shape change relations for homogeneous deformations. Future work will address state based formulations that deliver general anisotropic elastic response for the bulk energy associated with the limiting dynamics. \end{remark}
\setcounter{equation}{0} \setcounter{theorem}{0} \setcounter{lemma}{0}\setcounter{proposition}{0}\setcounter{remark}{0}\setcounter{remark}{0}\setcounter{definition}{0}\setcounter{hypothesis}{0}
\section{Free crack propagation in the small horizon limit} \label{sec6}
We recall that the process zone concentrates on a set of zero volume (Lebesgue measure) in the small horizon limit and identify conditions for which the limit dynamics $u^0$ solves the wave equation away from the evolving crack set. To begin we make a technical hypothesis on the regularity of the jump set of the limit dynamics $u^0(x,t)$. \begin{hypothesis} \label{remarkthree} We suppose that the crack set given by $J_{u^0(t)}$ is a closed set for $0\leq t\leq T$. \end{hypothesis} \noindent The next hypothesis applies to the concentration of the process zones as $\epsilon\rightarrow 0$ and their relation to the crack set for the limit dynamics. \begin{hypothesis} \label{remark555} \noindent Theorem \ref{bondunstable} shows that the fracture sets defined as process zones with strains above $\mathcal{S}_c$, see Definition \ref{Fractureset}, concentrate on the set $CPZ^0(\overline{r},1/2,1/2,t)$. Here we assume that $J_{u^0(t)}=CPZ^0(\overline{r},1/2,1/2,t)$ for $0\leq t\leq T$. \end{hypothesis}
\noindent The next hypothesis applies to neighborhoods $\mathcal{H}_{\epsilon_k}(x)$ for which the strain is subcritical, i.e., $|\mathcal{S}^\epsilon|<\overline{r}/\sqrt{|y-x|}$, for $y$ in $\mathcal{H}_{\epsilon_k}(x)$. These neighborhoods will be referred to as neutrally stable. \begin{hypothesis} \label{remark556} We suppose that $\epsilon_k=\frac{1}{2^k}<\delta$ and $0\leq t\leq T$. Consider the collection of centers of peridynamic neighborhoods in $CPZ^\delta(\overline{r},1/2,1/2,t)$. We fatten out $CPZ^\delta(\overline{r},1/2,1/2,t)$ and consider $C\widetilde{PZ}^\delta(\overline{r},1/2,1/2,t)=\{x\in D:\, dist(x,CPZ^\delta(\overline{r},1/2,1/2,t)<\delta\}$. We suppose that all neighborhoods $H_{\epsilon_k}(x)$ that do not intersect the set $C\widetilde{PZ}^\delta(\overline{r},1/2,1/2,t)$ are neutrally stable. \end{hypothesis}
\noindent With these conditions satisfied the limit evolution $u^0$ is identified as a solution of the linear elastic wave equation.
\begin{theorem} \label{waveequation} Suppose Hypotheses \ref{remarkthree}, \ref{remark555} and \ref{remark556} hold true then the limit evolution $u^0(t,x)$ is a solution of the following wave equation (the first law of motion of Cauchy) in the sense of distributions on the domain $[0,T]\times D$ given by \begin{eqnarray} \rho u^0_{tt}= {\rm div}\sigma+b, \hbox{on $[0,T]\times D$}, \label{waveequationn} \end{eqnarray} where the stress tensor $\sigma$ is given by, \begin{eqnarray} \sigma =\lambda I_d Tr(\mathcal{E}\,u^0)+2\mu \mathcal{E}u^0, \label{stress} \end{eqnarray} where $I_d$ is the identity on $\mathbb{R}^d$ and $Tr(\mathcal{E}\,u^0)$ is the trace of the strain. Here the second derivative $u_{tt}^0$ is the time derivative in the sense of distributions of $u^0_t$ and ${\rm div}\sigma$ is the divergence of the stress tensor $\sigma$ in the distributional sense. \end{theorem}
\begin{remark} \label{Disjointsets} For completeness we recall that the strain $\mathcal{E} u^0(x,t)$ and jump set $J_{u^0(t)}$ are defined over disjoint sets in $[0,T]\times D$. \end{remark}
\begin{remark} \label{displacementcrack} The limit of the cohesive dynamics model is given by the displacement - crack set pair $u^0(t,x)$, $J_{u^0(t)}$. The wave equation provides the dynamic coupling between elastic waves and the evolving fracture path inside the media. \end{remark}
\begin{remark} \label{remarknearlyfinal} Hypotheses \ref{remarkthree}, \ref{remark555}, and \ref{remark556} are applied exclusively to establish Lemma \ref{twolimitsB} which identifies the absolutely continuous part of the limit strain
$\mathcal{E}_{ij}u^0e_ie_j$, $e=\xi/|\xi|$, with the weak $L^2 (D\times\mathcal{H}_1 (0))$ limit of the strain $\mathcal{S}^\epsilon$restricted to pairs $(x,\xi)\in D\times\mathcal{H}_1 (0)$ for which $|\mathcal{S}^\epsilon|\leq\mathcal{S}_c$. \end{remark}
\begin{remark} \label{remarkfinal1} We point out that the cohesive model addressed in this work does not have an irreversibility constraint and the constitutive law \eqref{forcestate} applies at all times in the peridynamic evolution. Because of this the crack set at each time is given by $J_{u^0(t)}$. For rapid monotonic loading we anticipate that crack growth is increasing for this model, i.e., $J_{u^0(t')}\subset J_{u^0(t)}$ for $t'<t$. For cyclic loading this is clearly not the case and the effects of irreversibility (damage) must be incorporated into in the cohesive model. \end{remark}
\setcounter{equation}{0} \setcounter{theorem}{0} \setcounter{lemma}{0}\setcounter{proposition}{0}\setcounter{remark}{0}\setcounter{remark}{0}\setcounter{definition}{0}\setcounter{hypothesis}{0}
\section{Mathematical underpinnings and analysis} \label{EightIntro}
In this section we provide the proofs of theorems stated in Sections \ref{sec2}, \ref{sec4}, \ref{sec5} and \ref{sec6}. The first subsection asserts the Lipschitz continuity of $\nabla PD^{\epsilon_k}(u)$ for $u$ in $L^2_0(D;\mathbb{R}^d)$ and applies the theory of ODE to deduce existence of the cohesive dynamics, see Section \ref{EE}. A Gronwall inequality is used to bound the cohesive potential energy and kinetic energy uniformly in time, see Section \ref{GKP}. Uniformly bounded sequences $\{u^{\epsilon_k}\}_{k=1}^\infty$ of cohesive dynamics are shown to be compact in $C([0,T]; L^2_0(D;\mathbb{R}^d))$, see Section \ref{CC}. Any limit point $u^0$ for the sequence $u^{\epsilon_k}$ is shown to belong to SBD for every $0\leq t\leq T$, see Section \ref{CC}. The limit evolutions $u^0$ are shown to have uniformly bounded elastic energy in the sense of linear elastic fracture mechanics for $0\leq t\leq T$, see Section \ref{CC}. In Section \ref{EI} we pass to the limit in the energy balance equation for cohesive dynamics \eqref{BalanceEnergy} to recover an energy inequality for the limit flow. The wave equation satisfied by the limit flow is obtained on identifying the weak $L^2$ limit of the sequence $\{\nabla PD^{\epsilon_k}(u^{\epsilon_k})\}_{k=1}^\infty$ and passing to the limit in the weak formulation of \eqref{stationary}, see Section \ref{SC}. We conclude with the proof of Theorems \ref{epsiloncontropprocesszone} and \ref{bondunstable}.
\subsection{Existence of a cohesive evolution} \label{EE}
The peridynamic equation \eqref{eqofmotion} for cohesive dynamics is written as an equivalent first order system. We set $y^{\epsilon_k}=(y^{\epsilon_k}_1,y^{\epsilon_k}_2)^T$ where $y^{\epsilon_k}_1=u^{\epsilon_k}$ and $y_2^{\epsilon_k}=u_t^{\epsilon_k}$. Set $F^{\epsilon_k}(y^{\epsilon_k},t)=(F^{\epsilon_k}_1(y^{\epsilon_k},t),F^{\epsilon_k}_2(y^{\epsilon_k},t))^T$ where \begin{eqnarray} F^{\epsilon_k}_1(y^{\epsilon_k},t)&=&y_2^{\epsilon_k}\nonumber\\ F^{\epsilon_k}_2(y^{\epsilon_k},t)&=&\nabla PD^{\epsilon_k}(y_1^{\epsilon_k})+b(t).\nonumber \end{eqnarray} The initial value problem for $y^{\epsilon_k}$ given by the first order system is \begin{eqnarray} \frac{d}{dt} y^{\epsilon_k}=F^{\epsilon_k}(y^{\epsilon_k},t)\label{firstordersystem} \end{eqnarray} with initial conditions $y^{\epsilon_k}(0)=(u_0,v_0)^T$ satisfying LEFM initial conditions. In what follows we consider the more general class of initial data $(u_0,v_0)$ belonging to $L^2_0(D;\mathbb{R}^d)\times L^2_0(D;\mathbb{R}^d)$. \begin{theorem} For $0\leq t\leq T$ there exists unique solution in $C^1([0,T];L^2_0(D;\mathbb{R}^d)$ for the mesoscopic dynamics described by \eqref{firstordersystem} with initial data in $L^2_0(D;\mathbb{R}^d)\times L^2_0(D;\mathbb{R}^d)$ and body force $b(t,x)$ in $C^1([0,T];L^2_0(D;\mathbb{R}^d)$. \label{existenceuniqueness} \end{theorem} \noindent It now follows that for $LEFM$ initial data that one has a unique solution $u^{\epsilon_k}$ of \eqref{stationary} in Section \ref{sec2} belonging to $C^2([0,T];L^2_0(D;\mathbb{R}^d)$.
{\bf Proof of Theorem \ref{existenceuniqueness} }. A straight forward calculation shows that for a generic positive constant $C$ independent of $\mathcal{S}$, $y-x$, and $\epsilon_k$, that \begin{eqnarray}
\sup_{\mathcal{S}}|\partial_{\mathcal{S}}^2 W^{\epsilon_k}(\mathcal{S},y-x)|\leq \frac{C}{\epsilon_k|y-x|} \times J(|y-x|/\epsilon_k). \label{secondderv} \end{eqnarray} From this it easily follows from H\"older and Minkowski inequalities that $\nabla PD^{\epsilon_k}$ is a Lipschitz continuous map from $L^2_0(D;\mathbb{R}^d)$ into $L^2_0(D;\mathbb{R}^d)$ and there is a positive constant $C$ independent of $0\leq t\leq T$, such that for any pair of vectors $y=(y_1,y_2)^T$, $z=(z_1,z_2)^T$ in $L^2_0(D;\mathbb{R}^d)\times L^2_0(D;\mathbb{R}^d)$ \begin{eqnarray} \Vert F^{\epsilon_k}(y,t)-F^{\epsilon_k}(z,t)\Vert_{L^2(D;\mathbb{R}^d)^2}\leq \frac{C}{\epsilon_k}\Vert y-z\Vert_{L^2(D;\mathbb{R}^d)^2} \hbox{ for $0\leq t\leq T$}. \label{lipschitz} \end{eqnarray} Here for any element $w=(w_1,w_2)$ of $L^2_0(D;\mathbb{R}^d)\times L^2_0(D;\mathbb{R}^d)$, $\Vert w \Vert^2_{L^2(D;\mathbb{R}^d)^2}=\Vert w_1\Vert_{L^2(D;\mathbb{R}^d)}^2+\Vert w_2\Vert_{L^2(D;\mathbb{R}^d)}^2$. Since \eqref{lipschitz} holds the theory of ODE in Banach space \cite{Driver} shows that there exists a unique solution to the initial value problem \eqref{firstordersystem} with $y^{\epsilon_k}$ and $\partial_t y^{\epsilon_k}$ belonging to $C([0,T]; L^2_0(D;\mathbb{R}^d))$ and Theorem \ref{existenceuniqueness} is proved. In this context we point out the recent work of \cite{EmmrichPhulst} where an existence theory for peridynamic evolutions for general pairwise force functions that are Lipschitz continuous with respect to the peridynamic deformation state is presented.
\subsection{Bounds on kinetic and potential energy for solutions of PD} \label{GKP} In this section we apply Gronwall's inequality to obtain bounds on the kinetic and elastic energy for peridynamic flows described by Theorem \ref{Gronwall}. The bounds are used to show that the solutions of the PD initial value problem are Lipschitz continuous in time
We now prove Theorem \ref{Gronwall}. Multiplying both sides of \eqref{stationary} by $u_t^{\epsilon_k}(t)$ and integration together with a straight forward calculation gives \begin{eqnarray} &&\frac{1}{2}\frac{d}{dt}\left\{2PD^{\epsilon_k}(u^{\epsilon_k}(t))+{\rho}\Vert u_t^{\epsilon_k}(t)\Vert^2_{L^2(D;\mathbb{R}^d)}\right\}\nonumber\\ &&=\int_{D}(\nabla PD^{\epsilon_k}(u^{\epsilon_k}(t))+\rho u_{tt}^{\epsilon_k}(t))\cdot u_t^{\epsilon_k}(t)\,dx\nonumber\\ &&=\int_{D} u_t^{\epsilon_k}(t)\cdot b(t)\,dx\,\leq \, \Vert u_t^{\epsilon_k}\Vert_{L^2(D;\mathbb{R}^d)}\Vert b(t)\Vert_{L^2(D;\mathbb{R}^d)}.\label{esttime1} \end{eqnarray} Set \begin{eqnarray} && W(t)=2PD^{\epsilon_k}(u^{\epsilon_k}(t))+{\rho}\Vert u_t^{\epsilon_k}(t)\Vert^2_{L^2(D;\mathbb{R}^d)}+1, \label{wt} \end{eqnarray} and applying \eqref{esttime1} gives \begin{eqnarray} &&\frac{1}{2}W'(t) \leq \, \Vert u_t^{\epsilon_k}\Vert_{L^2(D;\mathbb{R}^d)}\Vert b(t)\Vert_{L^2(D;\mathbb{R}^d)}\leq\frac{1}{\sqrt{\rho}}\sqrt{W(t)}\Vert b(t)\Vert_{L^2(D;\mathbb{R}^d)}\label{esttime2} \end{eqnarray} and \begin{eqnarray} \frac{1}{2}\int_0^t\frac{W'(\tau)}{\sqrt{W(\tau)}}\,d\tau\leq\frac{1}{\sqrt{\rho}}\int_0^t\Vert b(\tau)\Vert_{L^2(D;\mathbb{R}^d)}\,d\tau. \label{estime3} \end{eqnarray} Hence \begin{eqnarray} \sqrt{W(t)}-\sqrt{W(0)}\leq\frac{1}{\sqrt{\rho}}\int_0^t\Vert b(\tau)\Vert_{L^2(D;\mathbb{R}^d)}\,d\tau \label{estime4} \end{eqnarray} and \begin{eqnarray} && 2PD^{\epsilon_k}(u^{\epsilon_k}(t))+{\rho}\Vert u_t^{\epsilon_k}(t)\Vert^2_{L^2(D;\mathbb{R}^d)} \leq \left(\frac{1}{\sqrt{\rho}}\int_0^t\Vert b(\tau)\Vert_{L^2(D;\mathbb{R}^d)}\,d\tau +\sqrt{W(0)}\right )^2-1. \label{gineq} \end{eqnarray}
For now we postpone the proof of \eqref{basicinequality} to Section \ref{CC} (see the discussion preceeding \eqref{upperboundperi}) and apply \eqref{basicinequality} to get the upper bound \begin{eqnarray} PD^{\epsilon_k}(u_0)\leq LEFM (u_0,D)\hbox{ for every $\epsilon_k$, \,\,$k=1,2,\ldots$}, \label{upperbound} \end{eqnarray} where $LEFM(u_0,D)$ is the elastic potential energy for linear elastic fracture mechanics given by \eqref{Gcrackenergy} or equivelently \eqref{LEFMSBVDefinition}. Theorem \ref{Gronwall} now follows from \eqref{gineq} and \eqref{upperbound}.
Theorem \ref{Gronwall} implies that PD solutions are Lipschitz continuous in time; this is stated explicitly in Theorem \ref{holdercont} of Section \ref{sec2}. To prove Theorem \ref{holdercont} we write \begin{eqnarray}
&&\Vert u^{\epsilon_k}(t_1)-u^{\epsilon_k}(t_2)\Vert_{L^2(D;\mathbb{R}^d)}=\left (\int_{D}|\int_{t_2}^{t_1} u^{\epsilon_k}_\tau(\tau)\,d\tau |^2\,dx\right )^{\frac{1}{2}}\nonumber\\
&&=\left (\int_{D}|t_1-t_2|^{2}\left|\frac{1}{|t_1-t_2|}\int_{t_2}^{t_1} u^{\epsilon_k}_\tau(\tau)\,d\tau \right|^2\,dx\right )^{\frac{1}{2}}\nonumber\\
&&\leq\left (\int_{D}|t_1-t_2|\int_{t_2}^{t_1} |u^{\epsilon_k}_\tau(\tau)|^2\,d\tau\,dx\right )^{\frac{1}{2}}\nonumber\\
&&\leq\left(|t_1-t_2|\int_{t_2}^{t_1}\Vert u_\tau^{\epsilon_k}(\tau)\Vert_{L^2(D;\mathbb{R}^d)}^2\,d\tau\right)^{1/2}\nonumber\\
&&\leq K|t_1-t_2|, \label{lip} \end{eqnarray} where the third to last line follows from Jensen's inequality, the second to last line from Fubini's theorem and the last inequality follows from the upper bound for $\Vert u_t^{\epsilon_k}(t)\Vert^2_{L^2(D;\mathbb{R}^d)}$ given by Theorem \ref{Gronwall}.
\subsection{Compactness and convergence} \label{CC} In this section we prove Theorems \ref{LimitflowThm} and \ref{LEFMMThm} . We start by establishing the inequality \eqref{basicinequality} between the elastic energies $PD^{\epsilon_k}(u)$ and for $LEFM(u,D)$. This is illustrated for any $u$ in $L^2_0(D;\mathbb{R}^d)\cap L^\infty(D;\mathbb{R}^d)$ and $LEFM(u,D)$ given by \eqref{LEFMSBVDefinition}. Here \eqref{LEFMSBVDefinition} reduces to \eqref{Gcrackenergy} when $u$ is piecewise smooth and the crack $K$ consists of a finite number of smooth components. To obtain the upper bound we can directly apply the slicing technique of \cite{Gobbino3} to reduce to the one dimensional case to obtain an upper bound on one dimensional sections and then apply integral-geometric arguments to conclude. Here the slicing theorem and integral-geometric measure appropriate for this approach in the context of SBD are given by Theorems 4.5 and 4.10 of \cite{AmbrosioCosicaDalmaso}. These arguments deliver the following inequality \begin{eqnarray} &&PD^{\epsilon_k}(u)\leq LEFM(u,D), \hbox{ for every $u$ in $L^2_0(D;\mathbb{R}^d)$, and $\epsilon_k>0$}. \label{upperboundperi} \end{eqnarray}
To proceed with the proof of Theorem \ref{LimitflowThm} we require the compactness theorem. \begin{theorem} {\rm \bf Compactness.}\\ Given a sequence of functions $u^{\epsilon_k}\in L^2_0(D;\mathbb{R}^d)$, $\epsilon_k=1/k,k=1,2,\ldots$ such that \begin{eqnarray} \sup_{\epsilon_k}\left(PD^{\epsilon_k}(u^{\epsilon_k})+\Vert u^{\epsilon_k}\Vert_{L^\infty(D;\mathbb{R}^d)}\right)<\infty, \label{unifbound} \end{eqnarray} then there exists a subsequence $u^{{\epsilon_{k}}'}$ and limit point $u$ in $L_0^2(D;\mathbb{R}^d)$ for which \begin{eqnarray} u^{{\epsilon_k}'}\rightarrow u \hbox{ in $L^2(D;\mathbb{R}^d)$ as } {\epsilon_k}'\rightarrow 0. \label{compactness} \end{eqnarray} \label{L2compact} \end{theorem}
In what follows its convenient to change variables $y=x+\delta\xi$ for $|\xi|<1$ and $0<\delta<\alpha/2<1$, here the peridynamic neighborhood $\mathcal{H}_\delta(x)$ transforms to $\mathcal{H}_1(0)=\{\xi\in\mathbb{R}^d;\,|\xi|<1\}$. The unit vector $\xi/|\xi|$ is denoted by $e$. To prove Theorem \ref{L2compact} we need the following upper bound given by the following theorem. \begin{theorem}{\rm \bf Upper bound}\\ \label{coercivityth} For any $0<\delta<\alpha/2$ there exists positive constants $\tilde{K}_1$ and $\tilde{K}_2$ independent of $u\in L^2_0(D;\mathbb{R}^d)\cap L^\infty(D;\mathbb{R}^d)$ such that \begin{eqnarray}
\int_{\mathcal{H}_1(0)}\int_D|u(x+\delta\xi)-u(x)|^2dx\,J(|\xi|)d\xi\leq \delta(\tilde{K}_1+\tilde{K}_2\Vert u\Vert_{L^\infty(D;\mathbb{R}^d)}^2) PD^\delta(u). \label{coercivity} \end{eqnarray} \end{theorem} We establish the upper bound in two steps. \begin{lemma}{\rm \bf Coercivity}\\ \label{coercivitya} There exists a positive constant $C$ independent of $u\in L^2_0(D;\mathbb{R}^d)$ for which \begin{eqnarray}
\int_{\mathcal{H}_1(0)}\int_D|u(x+\delta\xi)-u(x)|^2dx\,J(|\xi|)d\xi\leq C \int_{\mathcal{H}_1(0)}\int_D|(u(x+\delta\xi)-u(x))\cdot e|^2dx\,J(|\xi|)d\xi \label{coercivea} \end{eqnarray} \end{lemma}
{\bf Proof of Lemma \ref{coercivitya}} We prove by contradiction. Suppose for every positive integer $N>0$ there is an element $u^N\in L_0^2(D;\mathbb{R}^d)$ for which \begin{eqnarray}
&&N\int_{\mathcal{H}_1(0)}\int_D|(u^N(x+\delta\xi)-u^N(x))\cdot e |^2dx\,J(|\xi|)d\xi\nonumber\\
&&\leq\int_{\mathcal{H}_1(0)}\int_D|(u^N(x+\delta\xi)-u^N(x)) |^2dx\,J(|\xi|)d\xi. \label{contra} \end{eqnarray} The Cauchy Schwartz inequality together with the triangle inequality deliver a constant $\overline{K}>0$ for which \begin{eqnarray}
\int_{\mathcal{H}_1(0)}\int_D|(u(x+\delta\xi)-u(x))|^2dx\,J(|\xi|)d\xi\leq\overline{K}\Vert u\Vert^2_{L^2(D;\mathbb{R}^d)}. \label{upineq} \end{eqnarray} An application of the nonlocal Korn inequality, Lemma 5.5 of \cite{DuGunzbergerlehoucqmengesha} gives the existence of a constant $\underline{K}>0$ independent of $u$ in $L_0^2(D;\mathbb{R}^d)$ for which \begin{eqnarray}
\underline{K} \Vert u\Vert^2_{L^2(D;\mathbb{R}^d)}\leq\int_{\mathcal{H}_1(0)}\int_D|(u(x+\delta\xi)-u(x))\cdot e|^2dx\,J(|\xi|)d\xi. \label{korn} \end{eqnarray} Applying the inequalities \eqref{contra}, \eqref{upineq}, and \eqref{korn} we discover that $\overline{K}/N\geq\underline{K}$ for all integers $N>0$ to conclude $\underline{K}=0$ which is a contradiction and Lemma \ref{coercivitya} is proved. Theorem \ref{coercivityth} now follows from Lemma \ref{coercivitya} and the upper bound given by \begin{lemma} {\rm Upper bound}\\ \label{coercivityb} \begin{eqnarray}
\int_{\mathcal{H}_1(0)}\int_D|(u(x+\delta\xi)-u(x))\cdot e|^2dx\,J(|\xi|)d\xi\leq \delta(\tilde{K}_1+\tilde{K}_2\Vert u\Vert_{L^\infty(D;\mathbb{R}^d)}^2) PD^\delta(u). \label{coerciveb} \end{eqnarray} \end{lemma} {\rm \bf Proof of Lemma \ref{coercivityb}}. Consider the concave potential function $f$ described in the introduction, recall $f(0)=0$ and given $M>0$ set $H_M=f(M)/M$. For $0<r<M$ one has $r<H_M^{-1}f(r)$ and set \begin{eqnarray}
A_{\delta\xi}=\{x\in D; |(u(x+\delta\xi)-u(x))\cdot e|^2>\delta|\xi|M\}, \label{Aset} \end{eqnarray} so \begin{eqnarray}
&&\int_{D\setminus A_{\delta\xi}}\,|(u(x+\delta\xi)-u(x))\cdot e|^2\,dx=
\delta|\xi|\int_{D\setminus A_{\delta\xi}}\delta|\xi||\mathcal{S}|^2\,dx\nonumber\\
&&\leq\frac{\delta|\xi|}{H_M}\int_{D\setminus A_{\delta\xi}}\frac{1}{\delta|\xi|}f\left(\delta|\xi|\mathcal{S}|^2\right)\,dx. \label{starter} \end{eqnarray} Now $f(r)>f(M)$ for $r>M$ gives \begin{eqnarray}
\frac{1}{\delta|\xi|}f(M)\mathcal{L}^d(A_{\delta\xi})\leq
\int_{A_{\delta\xi}}\frac{1}{\delta|\xi|}f\left(\delta|\xi|\mathcal{S}|^2\right)\,dx \label{vol1} \end{eqnarray} and \begin{eqnarray} \mathcal{L}^d(A_{\delta\xi}\label{ineqonA})\leq
\frac{\delta|\xi|}{f(M)}\int_{D}\frac{1}{\delta|\xi|}f\left(\delta|\xi|\mathcal{S}|^2\right)\,dx. \label{volumebound} \end{eqnarray} Noting that \begin{eqnarray}
\int_{A_{\delta\xi}}\,|(u(x+\delta\xi)-u(x))\cdot e|^2\,dx\leq 2\Vert u\Vert_{L^\infty(D;\mathbb{R}^d}^2\mathcal{L}^d(A_{\delta\xi}) \label{basiccc} \end{eqnarray} and collecting results one has \begin{eqnarray} \label{almostestimate}
\int_{D}\,|(u(x+\delta\xi)-u(x))\cdot e|^2\,dx\leq\delta|\xi|\left(\frac{1}{H_M}+\frac{2\Vert u\Vert_{L^\infty(D;\mathbb{R}^d)}^2}{f(M)}\right)
\int_{D}\frac{1}{\delta|\xi|}f\left(\delta|\xi|\mathcal{S}|^2\right)\,dx. \end{eqnarray}
Lemma \ref{coercivityb} follows on multiplying both sides of \eqref{almostestimate} by $J(|\xi|)$ and integration over $\mathcal{H}_1(0)$. Theorem \ref{coercivityth} follows from Lemmas \ref{coercivitya} and \ref{coercivityb}.
Arguing as in \cite{Gobbino3} we have the monotonicity given by \begin{lemma}{\rm\bf Monotonicity}\\ For any integer $M$, $\eta>0$ and $u\in L^\infty(D;\mathbb{R}^d)$ one has \begin{eqnarray} PD^{M\eta}(u)\leq PD^\eta(u). \label{monotone} \end{eqnarray} \label{monotonicity} \end{lemma} Now choose the subsequence $\epsilon_k=1/2^k,\,i=1,2,\ldots$ and from Theorem \ref{coercivityth} and Lemma \ref{monotonicity} we have for any $0<K<k$ with $\delta=2^{-K}$, $\epsilon_k=2^{-k}$, \begin{eqnarray}
\int_{\mathcal{H}_1(0)}\int_D|u^{\epsilon_k}(x+\delta\xi)-u^{\epsilon_k}(x)|^2dx\,J(|\xi|)d\xi\leq \delta(\tilde{K}_1+\tilde{K}_2\Vert u^{\epsilon_k}\Vert_{L^\infty(D;\mathbb{R}^d)}^2) PD^{\epsilon_k}(u^{\epsilon_k}).\label{precompactness} \end{eqnarray} Applying the hypothesis \eqref{unifbound} to inequality \eqref{precompactness} gives a finite constant $B$ independent of $\epsilon_k$ and $\delta$ for which \begin{eqnarray}
\int_{\mathcal{H}_1(0)}\int_D|u^{\epsilon_k}(x+\delta\xi)-u^{\epsilon_k}(x)|^2dx\,J(|\xi|)d\xi\leq \delta B,\label{precompactnesses} \end{eqnarray} for all $\epsilon_k<\delta$. One can then apply \eqref{precompactnesses} as in \cite{Gobbino3}, (or alternatively apply \eqref{precompactnesses} and arguments similar to the proof of the Kolomogorov-Riesz compactness theorem \cite{OlsenHolden}) to show that the sequence $\{u^{\epsilon_k}\}_{k=1}^\infty$ is a totally bounded subset of $L^2_0(D;\mathbb{R}^d)$ and Theorem \ref{L2compact} is proved.
Now it is shown that the family of mesoscopic dynamics $\{u^{\epsilon_k}\}_{k=1}^\infty$ is relatively compact in \\ $C([0,T];L_0^2(D;\mathbb{R}^d))$. For each $t$ in $[0,T]$ we apply Theorem \ref{Gronwall} and Hypothesis \ref{remarkone} to obtain the bound \begin{eqnarray} PD^{\epsilon_k}(u^{\epsilon_k}(t))+\Vert u^{\epsilon_k}(t)\Vert_{L^\infty(D)}<C \label{cpactbnd} \end{eqnarray} where $C<\infty$ and is independent of $\epsilon_k$, $k=1,2,\ldots$, and $0\leq t\leq T$. With this bound we apply Theorem \ref{L2compact} to assert that for each $t$ the sequence $\{u^{\epsilon_k}(t)\}_{k=1}^\infty$ is relatively compact in $L^2(D;\mathbb{R}^d)$. From Theorem \ref{holdercont} the sequence $\{u^{\epsilon_k}\}_{k=1}^\infty$, is seen to be uniformly equi-continuous in $t$ with respect to the $L^2(D;\mathbb{R}^d)$ norm and we immediately conclude from the Ascoli theorem that $\{u^{\epsilon_k}\}_{k=1}^\infty$ is relatively compact in $C([0,T];L^2(D;\mathbb{R}^d))$. Therefore we can pass to a subsequence also denoted by $\{u^{\epsilon_{k}}(t)\}_{k=1}^\infty$ to assert the existence of a limit evolution $u^0(t)$ in $C([0,T];L^2(D;\mathbb{R}^d))$ for which \begin{eqnarray} \lim_{k\rightarrow\infty}\left\{\sup_{t\in[0,T]}\Vert u^{\epsilon_{k}}(t)-u^0(t)\Vert_{\scriptscriptstyle{{L^2(D;\mathbb{R}^d)}}}\right\}=0 \label{unfconvergence} \end{eqnarray} and Theorem \ref{LimitflowThm} is proved.
We now prove Theorem \ref{LEFMMThm}. One has that limit points of sequences satisfying \eqref{unifbound} enjoy higher regularity.
\begin{theorem}{\rm\bf Higher regularity}\\ \label{higherreg} Every limit point of a sequence $\{u^{\epsilon_k}\}_{k=1}^\infty$ in $L_0^2(D;\mathbb{R}^d)$ satisfying \eqref{unifbound} belongs to $SBD$. \end{theorem}
{\bf Proof.} To recover higher regularity one can directly apply the slicing technique of \cite{Gobbino3} to reduce to the one dimensional case and construct sequences of functions converging in SBV to the limit point along one dimensional sections. One then applies Theorem 4.7 of \cite{AmbrosioCosicaDalmaso} to conclude that the limit point belongs to $SBD$.
It now follows from Theorem \ref{higherreg} that the limit evolution $u^0(t)$ belongs to $SBD$ for $0\leq t\leq T$. Next we recall the properties of $\Gamma$-convergence and apply them to finish the proof of Theorem \ref{LEFMMThm}. Consider a sequence of functions $\{F_j\}$ defined on a metric space $\mathbb{M}$ with values in $\overline{\mathbb{R}}$ together with a function $F$ also defined on $\mathbb{M}$ with values in $\overline{\mathbb{R}}$.
\begin{definition} \label{Gammaconvergence} We say that $F$ is the $\Gamma$-limit of the sequence $\{F_j\}$ in $\mathbb{M}$ if the following two properties hold: \begin{enumerate} \item for every $x$ in $\mathbb{M}$ and every sequence $\{x_j\}$ converging to $x$, we have that \begin{eqnarray} F(x)\leq \liminf_{j\rightarrow\infty} F_j(x_j),\label{lowerbound} \end{eqnarray} \item for every $x$ in $\mathbb{M}$ there exists a recovery sequence $\{x_j\}$ converging to $x$, for which \begin{eqnarray} F(x)=\lim_{j\rightarrow\infty} F_j(x_j).\label{recovery} \end{eqnarray} \end{enumerate} \end{definition}
For $u$ in $L^2_0(D;\mathbb{R}^d)$ define $PD^0:\,L^2_0(D;\mathbb{R}^d)\rightarrow [0,+\infty]$ by \begin{equation} PD^0(u)=\left\{ \begin{array}{ll} LEFM(u,D)&\hbox{if $u$ belongs to $SBD$}\\ +\infty&\hbox{otherwise} \end{array} \right. \label{Gammalimit} \end{equation}
A straight forward argument following Theorem 4.3 $(ii)$ and $(iii)$ of \cite{Gobbino3} and invoking Theorems 4.5, 4.7, and 4.10 of \cite{AmbrosioCosicaDalmaso} as appropriate delivers
\begin{theorem}{\bf $\Gamma$- convergence and point wise convergence of peridynamic energies for cohesive dynamics.} \begin{eqnarray} &&PD^0 \hbox{ is the $\Gamma$-limit of $\{PD^{\epsilon_k}\}$ in $L^2_0(D;\mathbb{R}^d)$}, \hbox{ and } \label{gammaconvpd}\\ &&\lim_{k\rightarrow\infty}PD^{\epsilon_k}(u)=PD^0(u), \hbox{ for every $u$ in $L^2_0(D;\mathbb{R}^d)$}.\label{pointwise} \end{eqnarray} \label{Gammaandpointwise} \end{theorem}
Observe that since the sequence of peridynamic energies $\{PD^{\epsilon_k}\}$ $\Gamma$-converge to $PD^0$ in $L^2(D;\mathbb{R}^d)$ we can apply the lower bound property \eqref{lowerbound} of $\Gamma$-convergence to conclude that the limit has bounded elastic energy in the sense of fracture mechanics, i.e., \begin{eqnarray} LEFM(u^0(t),D)=PD^0(u^0(t))\leq\liminf_{k\rightarrow\infty}PD^{\epsilon_{k}}(u^{\epsilon_{k}}(t))<C. \label{GSBV} \end{eqnarray} This concludes the proof of Theorem \ref{LEFMMThm}.
\subsection{Energy inequality for the limit flow} \label{EI} In this section we prove Theorem \ref{energyinequality}. We begin by showing that the limit evolution $u^0(t,x)$ has a weak derivative $u_t^0(t,x)$ belonging to $L^2([0,T]\times D;\mathbb{R}^d)$. This is summarized in the following theorem. \begin{theorem} \label{weaktimederiviative} On passage to subsequences if necessary the sequence $u_t^{\epsilon_k}$ weakly converges in $L^2([0,T]\times D;\mathbb{R}^d)$ to $u^0_t$ where \begin{eqnarray} -\int_0^T\int_D\partial_t\psi \cdot u^0\, dxdt=\int_0^T\int_D\psi \cdot u^0_t\, dxdt, \label{weakl2time} \end{eqnarray} for all compactly supported smooth test functions $\psi$ on $[0,T]\times D$. \end{theorem}
{\bf Proof.} The bound on the kinetic energy given in Theorem \ref{Gronwall} implies \begin{eqnarray} \sup_{\epsilon_k>0}\left(\sup_{0\leq t\leq T}\Vert u^{\epsilon_k}_t\Vert_{L^2(D;\mathbb{R}^d)}\right)< \infty. \label{bddd} \end{eqnarray} Therefore the sequence $u^{\epsilon_k}_t$ is bounded in $L^2([0,T]\times D;\mathbb{R}^d)$ and passing to a subsequence if necessary we conclude that there is a limit function $\tilde{u}^0$ for which $u_t^{\epsilon_k}\rightharpoonup\tilde{u}^0$ weakly in $L^2([0,T]\times D;\mathbb{R}^d)$. Observe also that the uniform convergence \eqref{unfconvergence} implies that $u^{\epsilon_k}\rightarrow u^0$ in $L^2([0,T]\times D;\mathbb{R}^d)$. On writing the identity \begin{eqnarray} -\int_0^T\int_D\partial_t\psi\cdot u^{\epsilon_k}\, dxdt=\int_0^T\int_D\psi \cdot u^{\epsilon_k}_t\, dxdt. \label{weakidentity} \end{eqnarray} applying our observations and passing to the limit it is seen that $\tilde{u}^0=u_t^0$ and the theorem follows.
To establish Theorem \ref{energyinequality} we require the following inequality. \begin{lemma} For almost every $t$ in $[0,T]$ we have \label{weakinequality} \begin{eqnarray} \Vert u^0_t(t)\Vert_{L^2(D;\mathbb{R}^d)}\leq \liminf_{\epsilon_k\rightarrow 0}\Vert u^{\epsilon_k}_t(t)\Vert_{L^2(D;\mathbb{R}^d)}. \label{limitweakineq} \end{eqnarray} \end{lemma} {\bf Proof.} We start with the identity \begin{eqnarray} \Vert u^{\epsilon_k}_t\Vert_{L^2(D;\mathbb{R}^d)}^2-2\int_D u^{\epsilon_k}_t\cdot u^0_t \,dx+\Vert u^0_t\Vert_{L^2(D;\mathbb{R}^d)}^2 = \Vert u^{\epsilon_k}_t-u^0_t\Vert_{L^2(D;\mathbb{R}^d)}^2 \geq 0, \label{locpositive} \end{eqnarray} and for every non-negative bounded measurable function of time $\psi(t)$ defined on $[0,T]$ we have \begin{eqnarray} \int_0^T\psi \Vert u^{\epsilon_k}_t-u^0_t\Vert_{L^2(D;\mathbb{R}^d)}^2\,dt\geq 0. \label{positive} \end{eqnarray} Together with the weak convergence given in Theorem \ref{weaktimederiviative} one easily sees that \begin{eqnarray} \liminf_{\epsilon_k\rightarrow 0}\int_0^T\psi\Vert u^{\epsilon_k}_t\Vert_{L^2(D;\mathbb{R}^d)}^2\,dt-\int_0^T\psi\Vert u^0_t\Vert_{L^2(D;\mathbb{R}^d)}^2\,dt\geq 0. \label{diff} \end{eqnarray}
\noindent Applying \eqref{bddd} and invoking the Lebesgue dominated convergence theorem we conclude \begin{eqnarray} \liminf_{\epsilon_k\rightarrow 0}\int_0^T\psi\Vert u^{\epsilon_k}_t\Vert_{L^2(D;\mathbb{R}^d)}^2\,dt=\int_0^T\psi\liminf_{\epsilon_k\rightarrow 0}\Vert u^{\epsilon_k}_t\Vert_{L^2(D;\mathbb{R}^d)}^2\,dt \label{equalitybalance} \end{eqnarray} to recover the inequality given by \begin{eqnarray} \int_0^T\psi\left(\liminf_{\epsilon_k\rightarrow 0}\Vert u^{\epsilon_k}_t\Vert_{L^2(D;\mathbb{R}^d)}^2-\Vert u^0_t\Vert_{L^2(D;\mathbb{R}^d)}^2\right)\,dt\geq 0. \label{diffinal} \end{eqnarray} The lemma follows noting that \eqref{diffinal} holds for every non-negative test function $\psi$.
Theorem \ref{energyinequality} now follows immediately on taking the $\epsilon_k\rightarrow 0$ limit in the peridynamic energy balance equation \eqref{BalanceEnergy} of Theorem \ref{Ebalance} and applying \eqref{pointwise}, \eqref{GSBV}, and \eqref{limitweakineq} of Lemma \ref{weakinequality}.
\subsection{Stationarity conditions for the limit flow} \label{SC}
In this section we prove Theorem \ref{waveequation}. The first subsection establishes Theorem \ref{waveequation} using Theorem \ref{convgofelastic}. Theorem \ref{convgofelastic} is proved in the second subsection.
\subsubsection{Proof of Theorem \ref{waveequation}}
To proceed we make the change of variables $y=x+\epsilon\xi$ where $\xi$ belongs to the unit disk $\mathcal{H}_1(0)$ centered at the origin and the local strain $\mathcal{S}$ is of the form \begin{eqnarray}
\mathcal{S}=\left(\frac{u(x+\epsilon\xi)-u(x)}{\epsilon|\xi|}\right)\cdot e. \label{strainrescaled} \end{eqnarray}
It is convenient for calculation to express the strain through the directional difference operator $D_{e}^{\epsilon|\xi|}u$ defined by \begin{eqnarray}
D_{e}^{\epsilon|\xi|}u(x)=\frac{u(x+\epsilon\xi)-u(x)}{\epsilon|\xi|}\hbox{ and }\mathcal{S}=D_{e}^{\epsilon|\xi|} u\cdot e, \label{strainrescaledderiv} \end{eqnarray}
with $e=\xi/|\xi|$. One also has \begin{eqnarray}
D_{-e}^{\epsilon|\xi|}u(x)=\frac{u(x-\epsilon\xi)-u(x)}{\epsilon|\xi|}, \label{strainrescaledderivopposite} \end{eqnarray} and the integration by parts formula for functions $u$ in $L_0^2(D;\mathbb{R}^d)$, densities $\phi$ in $L_0^2(D;\mathbb{R})$ and $\psi$ continuous on $\mathcal{H}_1(0)$ given by \begin{eqnarray}
\int_D\int_{\mathcal{H}_1(0)}(D_e^{\epsilon|\xi|}u\cdot e)\phi(x)\psi(\xi)\,d\xi\,dx=\int_D\int_{\mathcal{H}_1(0)}(u\cdot e) (D_{-e}^{\epsilon|\xi|}\phi)\psi(\xi)\,d\xi\,dx. \label{intbypts} \end{eqnarray} Note further for $v$ in $C^\infty_0(D;\mathbb{R}^d)$ and $\phi$ in $C^\infty_0(D;\mathbb{R})$ one has \begin{eqnarray}
\lim_{\epsilon_k\rightarrow 0}D_e^{{\epsilon_k}|\xi|} v\cdot e=\mathcal{E} v \,e\cdot e &\hbox{ and }& \lim_{\epsilon_k\rightarrow 0}D_e^{{\epsilon_k}|\xi|} \phi= e\cdot \nabla\phi\label{grad} \end{eqnarray} where the convergence is uniform in $D$.
Taking the first variation of the action integral \eqref{Action} gives the Euler equation in weak form \begin{eqnarray} &&\rho\int_0^T\int_{D} u_t^{\epsilon_k}\cdot\delta_t\,dx \,dt\nonumber\\
&&-\frac{1}{\omega_d}\int_0^T\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)f'\left(\epsilon_k|\xi||D_e^{\epsilon_k|\xi|}u^{\epsilon_k}\cdot e|^2\right)2(D_e^{\epsilon_k|\xi|}u^{\epsilon_k}\cdot e) (D_{e}^{\epsilon_k|\xi|}\delta\cdot e)\,d\xi\,dx\,dt\nonumber\\ &&+\int_0^T\int_{D} b\cdot\delta\,dx\,dt=0,\label{stationtxweaker} \end{eqnarray}
where the test function $\delta=\delta(x,t)=\psi(t)\phi(x)$ is smooth and has compact support in $[0,T]\times D$. Next we make the change of function and write $F_s (\mathcal{S})=\frac{1}{s}f(s\mathcal{S}^2)$, $F'_s(\mathcal{S})=2\mathcal{S}f'(s\mathcal{S}^2)$, and $s={\epsilon_k}|\xi|$ we transform \eqref{stationtxweaker} into \begin{eqnarray} &&\rho\int_0^T\int_{D} u_t^{\epsilon_k}\cdot\delta_t\,dx \,dt\nonumber\\
&&-\frac{1}{\omega_d}\int_0^T\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)F_{\epsilon_k|\xi|}'(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)D_{e}^{\epsilon_k}\delta\cdot e\,d\xi\,dx\,dt\nonumber\\ &&+\int_0^T\int_{D} b\cdot\delta\,dx\,dt=0,\label{stationtxweakerlimitform} \end{eqnarray} where \begin{eqnarray}
F_{\epsilon_k|\xi|}'(D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)=f'\left(\epsilon_k |\xi||D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e|^2\right)2D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e. \label{deriv} \end{eqnarray}
For future reference observe that $F_s(r)$ is convex-concave in $r$ with inflection point $\overline{r}_s=\overline{r}/\sqrt{s}$ where $\overline{r}$ is the inflection point of $f(r^2)=F_1(r)$. One also has the estimates \begin{eqnarray} &&F_s(r)\geq\frac{1}{s}F_1(\overline{r})\hbox{ for $r\geq\overline{r}_s$, and }\label{lowerestforF}\\
&&\sup_{0\leq r<\infty}|F'_s(r)|\leq\frac{2f'(\overline{r}^2){\overline{r}}}{\sqrt{s}},\label{boundderiv} \label{estforFprime} \end{eqnarray} We send $\epsilon_k\rightarrow 0$ in \eqref{stationtxweakerlimitform} applying the weak convergence Theorem \ref{weaktimederiviative} to the first term to obtain \begin{eqnarray}
&&\rho\int_0^T\int_{D} u_t^{0}\cdot\delta_t\,dx \,dt-\lim_{\epsilon_k\rightarrow 0}\frac{1}{\omega_d}\left(\int_0^T\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)F_{\epsilon_k|\xi|}'(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)D_{e}^{\epsilon_k}\delta\cdot e\,d\xi\,dx\,dt\right)\nonumber\\ &&+\int_0^T\int_{D} b\cdot\delta\,dx\,dt=0.\label{stationtxweakerlimit} \end{eqnarray} Theorem \ref{waveequation} follows once we identify the limit of the second term in \eqref{stationtxweakerlimit} for smooth test functions $\phi(x)$ with support contained in $D$. We state the following convergence theorem. \begin{theorem} \label{convgofelastic} Given any infinitely differentiable test function $\phi$ with compact support in $D$ then \begin{eqnarray}
\lim_{\epsilon_k\rightarrow 0}\frac{1}{\omega_d}\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)F_{\epsilon_k|\xi|}'(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k})D_{e}^{\epsilon_k}\phi\,d\xi\,dx=\int_{D}\mathbb{C}\mathcal{E} u^0:\mathcal{E}\phi\,dx, \label{limitdelta} \end{eqnarray} where $\mathbb{C}\mathcal{E} u^0:\mathcal{E}\phi=\sum_{ijkl=1}^d\mathbb{C}_{ijkl}\mathcal{E}u^0_{ij}\mathcal{E}\phi_{kl}$, $\mathbb{C}\mathcal{E}u^0=\lambda I_d Tr(\mathcal{E}u^0)+2\mu \mathcal{E}u^0$, and $\lambda$ and $\mu$ are given by \eqref{calibrate1}. \end{theorem} \noindent Theorem \ref{convgofelastic} is proved in Section \ref{prtheorem44}. The sequence of integrals on the left hand side of \eqref{limitdelta} are uniformly bounded in time, i.e., \begin{eqnarray}
\sup_{\epsilon_k>0}\left\{\sup_{0\leq t\leq T}\left\vert\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)F_{\epsilon_k|\xi|}'(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)D_{e}^{\epsilon_k}\phi\cdot e\,d\xi\,dx\right\vert\right\}<\infty, \label{uniboundt} \end{eqnarray} this is demonstrated in \eqref{Fprimesecond} of Lemma \ref{estimates} in Section \ref{prtheorem44}. Applying the Lebesgue bounded convergence theorem together with Theorem \ref{convgofelastic} with $\delta(t,x)=\psi(t)\phi(x)$ delivers the desired result \begin{eqnarray}
&&\lim_{\epsilon_k\rightarrow 0}\frac{1}{\omega_d}\left(\int_0^T\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)F_{\epsilon_k|\xi|}'(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)\psi D_{e}^{\epsilon_k}\phi\cdot e\,d\xi\,dx\,dt\right)\nonumber\\ &&=\int_0^T\int_{D}\mathbb{C}\mathcal{E} u^0:\mathcal{E}\phi\,dx\,dt, \label{limitidentity} \end{eqnarray} and we recover the identity \begin{eqnarray} &&\rho\int_0^T\int_{D} u_t^{0}(t,x)\cdot\psi_t(t)\phi(x)\,dx \,dt-\int_0^T\int_{D}\psi(t)\mathbb{C}\mathcal{E} u^0(t,x):\mathcal{E}\phi(x)\,dx\,dt \nonumber\\ &&+\int_0^T\int_{D} b(t,x)\cdot\psi(t)\phi(x)\,dx\,dt=0 \label{finalweakidentity} \end{eqnarray} from which Theorem \ref{waveequation} follows.
\subsubsection{ Proof of Theorem \ref{convgofelastic}} \label{prtheorem44}
We decompose the difference $D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e$ as \begin{eqnarray}
D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e=(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^- +(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^+ \label{decompose} \end{eqnarray} where \begin{equation}
(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}(x)\cdot e)^-=\left\{\begin{array}{ll}(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}(x)\cdot e),&\hbox{if $|D_e^{\epsilon_k |\xi|}u^{\epsilon_k}(x)\cdot e|<\frac{\overline{r}}{\sqrt{\epsilon_k|\xi|}}$}\\ 0,& \hbox{otherwise} \end{array}\right. \label{decomposedetails} \end{equation}
where $\overline{r}$ is the inflection point for the function $F_1(r)=f(r^2)$. Here $(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^+$ is defined so that \eqref{decompose} holds. We prove Theorem \ref{convgofelastic} by using the following two identities described in the Lemmas below. \begin{lemma} \label{twolimitsA} For any $\phi$ in $C^\infty_0(D;\mathbb{R}^d)$ \begin{eqnarray}
&&\lim_{\epsilon_k\rightarrow 0} \frac{1}{\omega_d}\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)F_{\epsilon_k|\xi|}'(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)D_{e}^{\epsilon_k|\xi|}\phi\cdot e\,d\xi\,dx\nonumber\\
&&-2\lim_{\epsilon_k\rightarrow 0}\frac{1}{\omega_d}\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)f'(0)(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-D_e^{\epsilon_k |\xi|}\phi\cdot e\,d\xi\,dx=0. \label{functiotolimit} \end{eqnarray} \end{lemma} \begin{lemma} \label{twolimitsB}
Assume that Hypotheses \ref{remarkthree}, \ref{remark555} and \ref{remark556} hold true and define the weighted Lebesgue measure $\nu$ by $d\nu=|\xi|J(|\xi|)d\xi\,dx$ for any Lebesgue measurable set $S\subset D\times\mathcal{H}_1(0)$. Passing to subsequences if necessary $\{(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-\}_{k=1}^\infty$ converges weakly in $L^2(D\times\mathcal{H}_1(0);\nu)$ to $\mathcal{E} u^0 e\cdot e$, i.e., \begin{eqnarray}
&&\lim_{\epsilon_k\rightarrow 0}\frac{1}{\omega_d}\int_{D}\int_{\mathcal{H}_1(0)}(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^- \psi\,d\nu\nonumber\\ &&=\frac{1}{\omega_d}\int_{D}\int_{\mathcal{H}_1(0)}(\mathcal{E} u^0 e \cdot e)\psi\,d\nu, \label{functiotofunction} \end{eqnarray} for any test function $\psi(x,\xi)$ in $L^2(D\times\mathcal{H}_1(0);\nu)$. \end{lemma}
We now apply the Lemmas. Observing that $D_e^{\epsilon_k |\xi|}\phi\cdot e $ converges strongly in $L^2(D\times\mathcal{H}_1(0):\nu)$ to $\mathcal{E}\phi\, e\cdot e$ for test functions $\phi$ in $C^\infty_0(D;\mathbb{R}^d)$ and from the weak $L^2(D\times\mathcal{H}_1(0):\nu)$ convergence of $(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-$ we deduce that \begin{eqnarray}
&&\lim_{\epsilon_k\rightarrow 0}\frac{1}{\omega_d}\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)f'(0)(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-(D_e^{\epsilon_k |\xi|}\phi\cdot e)\,d\nu\nonumber\\
&&=\frac{1}{\omega_d}\int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)f'(0)\,(\mathcal{E} u^0 e\cdot e)(\mathcal{E}\phi e\cdot e)\,d\nu\nonumber\\
&&=\frac{f'(0)}{\omega_d}\sum_{ijkl=1}^d\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)\,e_i e_j e_k e_l,d\xi\int_{D}\mathcal{E}u^0_{ij}\mathcal{E}\phi_{kl}\,dx.\label{limitproduct} \end{eqnarray} Now we show that \begin{eqnarray}
\frac{f'(0)}{\omega_d}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)\,e_i e_j e_k e_l\,d\xi=\mathbb{C}_{ijkl}=2\mu\left(\frac{\delta_{ik}\delta_{jl}+\delta_{il}\delta_{jk}}{2}\right)+\lambda \delta_{ij}\delta_{kl}\label{shear} \end{eqnarray} where $\mu$ and $\lambda$ are given by \eqref{calibrate1}. To see this we write \begin{eqnarray} \Gamma_{ijkl}(e)=e_ie_je_ke_l, \label{tensorId} \end{eqnarray} to observe that $\Gamma(e)$ is a totally symmetric tensor valued function defined for $e\in S^{d-1}$ with the property \begin{eqnarray} \Gamma_{ijkl}(Qe)=Q_{im}e_mQ_{jn}e_nQ_{ko}e_oQ_{lp}e_p=Q_{im}Q_{jn}Q_{ko}Q_{lp}\Gamma_{mnop}(e) \label{rot} \end{eqnarray} for every rotation $Q$ in $SO^d$. Here repeated indices indicate summation. We write \begin{eqnarray}
\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)\,e_i e_j e_k e_l\,d\xi=\int_0^1|\xi|d|\xi|\int_{S^{d-1}}\Gamma_{ijkl}(e)\,de \label{groupavgpre} \end{eqnarray} to see that for every $Q$ in $SO^d$ \begin{eqnarray} Q_{im}Q_{jn}Q_{ko}Q_{lp}\int_{S^{d-1}}\Gamma_{ijkl}(e)\,de=\int_{S^{d-1}}\Gamma_{mnop}(Qe)\,de=\int_{S^{d-1}}\Gamma_{mnop}(e)\,de. \label{groupavg} \end{eqnarray} Therefore we conclude that $\int_{S^{d-1}}\Gamma_{ijkl}(e)\,de$ is an isotropic symmetric $4^{th}$ order tensor and of the form \begin{eqnarray} \int_{S^{d-1}}\Gamma_{ijkl}(e)\,de=a\left(\frac{\delta_{ik}\delta_{jl}+\delta_{il}\delta_{jk}}{2}\right)+b \delta_{ij}\delta_{kl}. \label{groupavgaft} \end{eqnarray} Here we evaluate $a$ by contracting both sides of \eqref{groupavgaft} with a trace free matrix and $b$ by contracting both sides with the $d\times d$ identity and calculation delivers \eqref{shear}. Theorem \ref{convgofelastic} now follows immediately from \eqref{limitproduct} and \eqref{functiotolimit}.
To establish Lemmas \ref{twolimitsA} and \ref{twolimitsB} we develop the following estimates for the sequences\\ $(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-$ and $(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^+$. We define the set $K^{+,\epsilon_k}$ by \begin{eqnarray}
K^{+,\epsilon_k}=\{(x,\xi)\in D\times\mathcal{H}_1(0)\,: (D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^+\not=0\}. \label{suppset} \end{eqnarray}
We have the following string of estimates. \begin{lemma} We introduce the generic positive constant $0<C<\infty$ independent of $0<\epsilon_k<1$ and $0\leq t\leq T$ and state the following inequalities that hold for all $0<\epsilon_k<1$ and $0\leq t\leq T$ and for $C^\infty(D)$ test functions $\phi$ with compact support on $D$. \label{estimates} \begin{eqnarray}
&&\int_{K^{+,\epsilon_k}} |\xi|J(|\xi|)\,d\xi\,dx<C\epsilon_k,\label{suppsetupper}\\
&&\left\vert\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)F_{\epsilon_k|\xi|}'((D_{e}^{{\epsilon_k}|\xi|,+}u^{\epsilon_k}\cdot e)^+)(D_{e}^{\epsilon_k}\phi\cdot e)\,d\xi\,dx\right\vert<C\sqrt{\epsilon_k}\Vert\mathcal{E}\phi\Vert_{\scriptscriptstyle{{L^\infty(D;\mathbb{R}^{d\times d})}}},\label{Fprimefirst}\\
&&\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)|(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)^-|^2\,d\xi\,dx<C,\label{L2bound}\\
&&\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)|D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e|\,d\xi\,dx<C,\hbox{ and}\label{L1bound}\\
&&\left\vert\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)F_{\epsilon_k|\xi|}'(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e) (D_{e}^{\epsilon_k}\phi\cdot e)d\,\xi\,dx\right\vert<C\Vert\mathcal{E} \phi\Vert_{\scriptscriptstyle{L^\infty(D;\mathbb{R}^{d\times d})}}.\label{Fprimesecond} \end{eqnarray} \end{lemma} {\bf Proof.} For $(x,\xi)\in K^{+,\epsilon_k}$ we apply \eqref{lowerestforF} to get \begin{eqnarray}
J(|\xi|)\frac{1}{\epsilon_k}F_{1}(\overline{r})=|\xi|J(|\xi|)\frac{1}{\epsilon_k|\xi|}F_{1}(\overline{r})
\leq|\xi|J(|\xi|)F_{\epsilon_k|\xi|}(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)\label{first} \end{eqnarray}
and in addition since $|\xi|\leq 1$ we have \begin{eqnarray}
&&\frac{1}{\epsilon_k}F_{1}(\overline{r}) \int_{K^{+,\epsilon_k}} |\xi|J(|\xi|)\,d\xi\,dx\leq\frac{1}{\epsilon_k}F_{1}(\overline{r}) \int_{K^{+,\epsilon_k}} J(|\xi|)\,d\xi\,dx\nonumber\\
&&\leq\int_{K^{+,\epsilon_k}} |\xi|J(|\xi|)F_{\epsilon_k|\xi|}(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)\,d\xi\,dx\leq\sup_{t\in [0,T]}\sup_{\epsilon_k}PD^{\epsilon_k}(u^{\epsilon_k}) \label{firstb} \end{eqnarray}
where Theorem \ref{Gronwall} implies that the right most element of the sequence of inequalities is bounded and \eqref{suppsetupper} follows noting that the inequality \eqref{firstb} is equivalent to \eqref{suppsetupper}. More generally since $|\xi|\leq 1$ we may argue as above to conclude that \begin{eqnarray}
\int_{K^{+,\epsilon_k}} |\xi|^pJ(|\xi|)\,d\xi\,dx<C\epsilon_k. \label{power} \end{eqnarray} for $0\leq p$. We apply \eqref{estforFprime} and \eqref{power} to find \begin{eqnarray}
&&\left\vert\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)F_{\epsilon_k|\xi|}'((D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)^+)D_{e}^{\epsilon_k}\phi\,d\xi\,dx\right\vert\nonumber\\
&&\leq C\frac{2f'(\overline{r}^2)\overline{r}}{\sqrt{\epsilon_k}}\int_{K^{+,\epsilon_k}}\sqrt{|\xi|}J(|\xi|)\,d\xi\,dx \Vert\mathcal{E}\phi\Vert_{\scriptscriptstyle{L^\infty(D;\mathbb{R}^{d\times d})}} \leq\sqrt{\epsilon_k}C\Vert\mathcal{E}\phi\Vert_{\scriptscriptstyle{L^\infty(D;\mathbb{R}^{d\times d})}}, \label{second} \end{eqnarray} and \eqref{Fprimefirst} follows.
A basic calculation shows there exists a positive constant independent of $r$ and $s$ for which \begin{eqnarray} r^2\leq C F_s(r), \hbox{ for $r<\frac{\overline{r}}{\sqrt{s}}$}, \label{rsquaredd} \end{eqnarray} so \begin{eqnarray}
|D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e|^2\leq C F_{\epsilon_k|\xi|}(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e), \hbox{ for $|D_{e}^{\epsilon_k|\xi|}u^{\epsilon_k}\cdot e|<\frac{\overline{r}}{\sqrt{\epsilon_k |\xi|}}$}, \label{rsquared} \end{eqnarray} and \begin{eqnarray}
&&\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)|(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)^-|^2\,d\xi\,dx
=\int_{D\times\mathcal{H}_1(0)\setminus K^{+,\epsilon_k}}|\xi| J(|\xi|)|D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e|^2\,d\xi\,dx\nonumber\\
&&\leq C\int_{D\times\mathcal{H}_1(0)\setminus K^{+,\epsilon_k}}|\xi| J(|\xi|)F_{\epsilon_k |\xi|}(D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)\,d\xi\,dx\leq C\sup_{t\in [0,T]}\sup_{\epsilon_k}PD^{\epsilon_k}(u^{\epsilon_k}) \label{third} \end{eqnarray} where Theorem \ref{Gronwall} implies that the right most element of the sequence of inequalities is bounded and \eqref{L2bound} follows.
To establish \eqref{L1bound} we apply H\"older's inequality to find that \begin{eqnarray}
&&\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)|D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e|\,d\xi\,dx\nonumber\\
&&=\int_{K^{+,\epsilon_k}}|\xi| J(|\xi|)|D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e|\,d\xi\,dx+\int_{D\times\mathcal{H}_1(0)\setminus K^{+,\epsilon_k}}|\xi| J(|\xi|)|D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e|\,d\xi\,dx\nonumber\\
&&\leq \frac{2\Vert u^{\epsilon_k}\Vert_{L^\infty(D;\mathbb{R}^d)}}{\epsilon_k}\int_{K^{+,\epsilon_k}}|\xi|J(|\xi|)\,d\xi\,dx+\nonumber\\
&&+\nu(D\times\mathcal{H}_1(0))^{\frac{1}{2}}\left (\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)|(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)^-|^2\,d\xi\,dx\right)^{\frac{1}{2}}, \label{twoterms} \end{eqnarray} and \eqref{L1bound} follows from \eqref{suppsetupper} and \eqref{L2bound}.
We establish \eqref{Fprimesecond}. This bound follows from the basic features of the potential function $f$. We will recall for subsequent use that $f$ is smooth positive, concave and $f'$ is a decreasing function with respect to its argument. So for $A$ fixed and $0\leq h\leq A^2\overline{r}^2$ we have \begin{eqnarray}
|f'(h)-f'(0)|\leq |f'(A^2\overline{r}^2)- f'(0)|<2|f'(0)|^2. \label{ffact} \end{eqnarray} The bound \eqref{Fprimesecond} is now shown to be a consequence of the following upper bound valid for the parameter $0<A<1$ given by \begin{eqnarray}
&&\int_{D\times\mathcal{H}_1(0)}|\xi|J(|\xi|)|f'(\epsilon_k |\xi||(D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-|^2)-f'(0)|^2\, d\xi\,dx\nonumber\\
&&\leq \nu(D\times\mathcal{H}_1(0))\times|f'(A^2\overline{r}^2)-f'(0)|^2+C\epsilon_k\frac{4|f'(0)|^2}{A^2}. \label{usefulbound} \end{eqnarray}
We postpone the proof of \eqref{usefulbound} until after it is used to establish \eqref{Fprimesecond}. Set $h_{\epsilon_k}=(D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-$ to note \begin{eqnarray}
F_{\epsilon_k |\xi|}'(h_{\epsilon_k})-2f'(0)h_{\epsilon_k}=(f'(\epsilon_k |\xi| h^2_{\epsilon_k})-f'(0))2h_{\epsilon_k}. \label{diffeq} \end{eqnarray} Applying H\"olders inequality, \eqref{Fprimefirst}, \eqref{L2bound}, \eqref{usefulbound}, and \eqref{diffeq} gives \begin{eqnarray}
&&\left\vert\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)F_{\epsilon_k|\xi|}'(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e) (D_{e}^{\epsilon_k}\phi\,d\xi\cdot e)\,dx \right\vert\nonumber\\
&&\leq\left\vert\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)F_{\epsilon_k|\xi|}'((D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)^+) (D_{e}^{\epsilon_k}\phi\cdot e)\,d\xi\,dx \right\vert\nonumber\\
&&+\left\vert\int_{D\times\mathcal{H}_1(0)}|\xi| J(|\xi|)F_{\epsilon_k|\xi|}'((D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)^-) (D_{e}^{\epsilon_k}\phi\cdot e)\,d\xi\,dx \right\vert\nonumber\\
&&\leq C\sqrt{\epsilon_k}\Vert\mathcal{E}\phi\Vert_{\scriptscriptstyle{L^\infty(D;\mathbb{R}^{d\times d})}}+2\int_{D\times \mathcal{H}_1(0)}|\xi|J(|\xi|)f'(0)(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-(D_e^{\epsilon_k |\xi|}\phi\cdot e)\,d\xi\,dx\nonumber\\
&&+\int_{D\times\mathcal{H}_1(0)}|\xi|J(|\xi|)\left (F_{\epsilon_k|\xi|}'((D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)^-) -2f'(0)(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-\right )(D_e^{\epsilon_k |\xi|}\phi\cdot e)\,d\xi\,dx\nonumber\\
&&\leq C\left(f'(0)+\sqrt{\epsilon_k}+\left(\nu(D\times\mathcal{H}_1(0))\times|f'(A^2\overline{r}^2)-f(0)|^2+\epsilon_k\frac{4|f'(0)|^2}{A^2}\right)^{1/2}\right )\Vert\mathcal{E} \phi\Vert_{\scriptscriptstyle{L^\infty(D;\mathbb{R}^{d\times d})}}.\nonumber\\ \label{five} \end{eqnarray} and \eqref{Fprimesecond} follows.
We establish the inequality \eqref{usefulbound}. Set $h_{\epsilon_k}=(D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-$ and for $0<A<1$ introduce the set \begin{eqnarray}
K^{+,\epsilon_k}_A=\{(x,\xi)\in D\times\mathcal{H}_1(0)\,: A^2\overline{r}^2\leq \epsilon_k|\xi||h_{\epsilon_k}|^2\}. \label{suppAset} \end{eqnarray}
To summarize $(x,\xi)\in K^{+,\epsilon_k}_A$ implies $A^2\overline{r}^2\leq\epsilon_k|\xi||h_{\epsilon_k}|^2\leq\overline{r}^2$ and $(x,\xi)\not\in K^{+,\epsilon_k}_A$ implies $\epsilon_k|\xi||h_{\epsilon_k}|^2<A^2\overline{r}^2$ and $|f'(\epsilon_k|\xi||h_{\epsilon_k}|^2)-f'(0)|\leq|f'(A^2\overline{r}^2)-f'(0)|$. Inequality \eqref{L2bound} implies \begin{eqnarray}
&&C>\int_{K^{+,\epsilon_k}_A} |\xi|J(|\xi|) h_{\epsilon_k}^2\,d\xi\,dx\geq\frac{A^2\overline{r}^2}{\epsilon_k}\int_{K^{+,\epsilon_k}_A} J(|\xi|) \,d\xi\,dx\nonumber\\
&&\geq\frac{A^2\overline{r}^2}{\epsilon_k}\int_{K^{+,\epsilon_k}_A} |\xi|J(|\xi|) \,d\xi\,dx, \label{chebyA} \end{eqnarray}
the last inequality follows since $1\geq|\xi|>0$. Hence \begin{eqnarray}
\int_{K^{+,\epsilon_k}_A} |\xi|J(|\xi|) \,d\xi\,dx\leq C\frac{\epsilon_k}{A^2\overline{r}^2}, \label{chebyAUpper} \end{eqnarray} and it follows that \begin{eqnarray}
&&\int_{K^{+,\epsilon_k}_A} |\xi|J(|\xi|)|f'(\epsilon_k|\xi|h_{\epsilon_k}|^2-f'(0)|^2 \,d\xi\,dx\nonumber\\
&&\leq 4|f'(0)|^2\int_{K^{+,\epsilon_k}_A} |\xi|J(|\xi|) \,d\xi\,dx\leq C\epsilon_k\frac{4|f'(0)|^2}{A^2\overline{r}^2}. \label{kepsplus} \end{eqnarray} Collecting observations gives \begin{eqnarray}
&&\int_{D\times\mathcal{H}_1(0)\setminus K^{+,\epsilon_k}_A}|\xi|J(|\xi|)|f'(\epsilon_k |\xi||(D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-|^2)-f'(0)|^2\, d\xi\,dx\nonumber\\
&&\leq \nu(D\times\mathcal{H}_1(0))\times |f'(A^2\overline{r}^2)-f'(0)|^2, \label{2ndbd} \end{eqnarray} and \eqref{usefulbound} follows.
We now prove Lemma \ref{twolimitsA}. Write \begin{eqnarray}
F_{\epsilon_k|\xi|}'(D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)=F_{\epsilon_k|\xi|}'((D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^+)
+F_{\epsilon_k|\xi|}'((D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-), \label{plusminus} \end{eqnarray} and from \eqref{Fprimefirst} it follows that \begin{eqnarray}
&&\lim_{\epsilon_k\rightarrow 0} \int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)F_{\epsilon_k|\xi|}'(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)D_{e}^{\epsilon_k|\xi|}\phi\cdot e\,d\xi\,dx\nonumber\\
&&=\lim_{\epsilon_k\rightarrow 0} \int_{D}\int_{\mathcal{H}_1(0)}|\xi|J(|\xi|)F_{\epsilon_k|\xi|}'((D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)^-)(D_{e}^{\epsilon_k|\xi|}\phi\cdot e)\,d\xi\,dx. \label{functiotolimitminus} \end{eqnarray}
To finish the proof we identify the limit of the right hand side of \eqref{functiotolimitminus}. Set $h_{\epsilon_k}=(D_{e}^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-$ and apply H\'older's inequality to find \begin{eqnarray}
&&\int_{D\times\mathcal{H}_1(0)}|\xi|J(|\xi|)\left(F_{\epsilon_k|\xi|}'(h_{\epsilon_k}) -2f'(0)h_{\epsilon_k}\right)(D_e^{\epsilon_k|\xi|}\phi\cdot e)\,d\xi\,dx\nonumber\\
&&\leq C\int_{D\times\mathcal{H}_1(0)}|\xi|J(|\xi|)\left|F_{\epsilon_k|\xi|}'(h_{\epsilon_k}) -2f'(0)h_{\epsilon_k}\right|\,d\xi\,dx\Vert\mathcal{E}\phi\Vert_{\scriptscriptstyle{L^\infty(D;\mathbb{R}^{d\times d})}} \label{firstestimate} \end{eqnarray} We estimate the first factor in \eqref{firstestimate} and apply \eqref{diffeq}, H\"older's inequality, \eqref{L2bound}, and \eqref{usefulbound} to obtain \begin{eqnarray}
&&\int_{D\times\mathcal{H}_1(0)}|\xi|J(|\xi|)\left |F_{\epsilon_k|\xi|}'(h_{\epsilon_k}) -2f'(0)h_{\epsilon_k}\right |\,d\xi\,dx\nonumber\\
&&\leq\int_{D\times\mathcal{H}_1(0)}|\xi|J(|\xi|)\left |f'(\epsilon_k |\xi||h_{\epsilon_k}|^2) -2f'(0)\right |\left | h_{\epsilon_k}\right |\,d\xi\,dx \nonumber\\
&&\leq C\left(\nu(D\times\mathcal{H}_1(0))\times |f'(A^2\overline{r}^2)-f'(0)|^2+\epsilon_k\frac{4|f'(0)|^2}{A^2\overline{r}^2}\right)^{1/2}. \label{usefulLemmaA} \end{eqnarray} Lemma \ref{twolimitsA} follows on applying the bound \eqref{usefulLemmaA} to \eqref {firstestimate} and passing to the $\epsilon_k$ zero limit and noting that the choice of $0<A<1$ is arbitrary.
We now prove Lemma \ref{twolimitsB}. For $\tau>0$ sufficiently small define $K^\tau\subset D$ by $K^{\tau}=\{x\in D:\,dist(x,J_{u^0(t)})<\tau\}$. From Hypothesis \ref{remark555} the collection of centroids in $CPZ^\delta(\overline{r},1/2,1/2,t)$ lie inside $K^\tau$ for $\delta$ sufficiently small. (Otherwise the components of the collection $CPZ^\delta(\overline{r},1/2,1/2,t)$ would concentrate about a component of $CPZ^0(\overline{r},1/2,1/2,t)$ outside $K^\tau$; contradicting the hypothesis that $J_{u^0(t)}=CPZ^0(\overline{r},1/2,1/2,t)$). The collection of all points belonging to unstable neighborhoods associated with centroids in is easily seen to be contained in the slightly larger set $K^{\tau,\delta}=\{x\in \,D; dist(x,K^\tau)<\delta\}$. From Hypothesis \ref{remark556} we may choose test functions $\phi\in C_0^1(D\setminus K^{\tau,2\delta})$ such that for $\epsilon_k$ sufficiently small \begin{eqnarray}
(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-\phi=(D_e^{\epsilon_k|\xi|}u^{\epsilon_k}\cdot e)\phi. \label{identical} \end{eqnarray}
We form the test functions $\phi(x)\psi(\xi)$, with $\phi\in C_0^1(D\setminus K^{\tau,2\delta})$ and
$\psi\in C(\mathcal{H}_1(0))$. From \eqref{L2bound} we may pass to a subsequence to find that $(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}\cdot e)^-$ weakly converges to the limit $g(x,\xi)$ in $L^2(D\times\mathcal{H}_1(0);\nu)$. With this in mind we write \begin{eqnarray}
&&\int_{D\times\mathcal{H}_1(0)} g(x,\xi)\phi(x)\psi(\xi)\,d\nu=\int_{D\times\mathcal{H}_1(0)} g(x,\xi)\phi(x)\psi(\xi)|\xi|J(|\xi|)\,d\xi\,dx\nonumber\\
&&=\lim_{\epsilon_k\rightarrow 0}\int_{D\times\mathcal{H}_1(0)}(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}(x)\cdot e)^- \phi(x)\psi(\xi)|\xi|J(|\xi|)\,d\xi\,dx\nonumber\\
&&=\lim_{\epsilon_k\rightarrow 0}\int_{D\times\mathcal{H}_1(0)}(D_e^{\epsilon_k |\xi|}u^{\epsilon_k}(x)\cdot e) \phi(x)\psi(\xi)|\xi|J(|\xi|)\,d\xi\,dx\nonumber\\
&&=\lim_{\epsilon_k\rightarrow 0}\int_{D\times\mathcal{H}_1(0)}(u^{\epsilon_k}(x)\cdot e)(D_{-e}^{\epsilon_k |\xi|}\phi(x))\psi(\xi)|\xi|J(|\xi|)\,d\xi\,dx,\label{middlepart} \end{eqnarray}
where we have integrated by parts using \eqref{intbypts} in the last line of \eqref{middlepart}. Noting that $D_{-e}^{\epsilon_k |\xi|}\phi(x)$ converges uniformly to $-e\cdot\nabla\phi(x)$ and from the strong convergence of $u^{\epsilon_k}$ to $u^0$ in $L^2$ we obtain \begin{eqnarray}
&&=\lim_{\epsilon_k\rightarrow 0}\int_{D\times\mathcal{H}_1(0)}(u^{\epsilon_k}(x)\cdot e)(D_{-e}^{\epsilon_k |\xi|}\phi(x))\psi(\xi)|\xi|J(|\xi|)\,d\xi\,dx\nonumber\\
&&=-\int_{D\times\mathcal{H}_1(0)}(u^{0}(x)\cdot e) (e\cdot\nabla\phi(x))\psi(\xi)|\xi|J(|\xi|)\,d\xi\,dx\nonumber\\
&&=-\sum_{j,k=1}^d\int_{D}u^0_j(x)\,\partial_{x_k}\left(\phi(x) \int_{\mathcal{H}_1(0)} e_j e_k\psi(\xi)|\xi|J(|\xi|)\,d\xi\right)\,dx\nonumber\\
&&=\int_{D}\mathcal{E} u^0_{jk}(x)\left(\phi(x) \int_{\mathcal{H}_1(0)} e_j e_k\psi(\xi)|\xi|J(|\xi|)\,d\xi\right)\,dx\nonumber\\
&&=\int_{D\times{\mathcal{H}_1(0)} }(\mathcal{E} u^0(x) e\cdot e)\phi(x)\psi(\xi)|\xi|J(|\xi|)\,d\xi\,dx, \label{identifyweakl2} \end{eqnarray} where we have made use of $E u^0\lfloor D\setminus K^{\tau,2\delta}=\mathcal{E} u^0\,dx$ on the third line of \eqref{identifyweakl2}. From the density of the span of the test functions we conclude that $g(x,\xi)=\mathcal{E} u^0(x)e\cdot e$ almost everywhere on $D\setminus K^{\tau,2\delta}\times\mathcal{H}_1(0)$. Since $K^{\tau,2\delta}$ can be chosen to have arbitrarily small measure with vanishing $\tau$ and $\delta$ we conclude that $g(x,\xi)=\mathcal{E} u^0(x) e\cdot e$ on $D\times\mathcal{H}_1(0)$ a.e. and Lemma \ref{twolimitsB} is proved.
\subsubsection{ Proof of Theorems \ref{epsiloncontropprocesszone} and \ref{bondunstable}} \label{proofbondunstable} We begin with the proof on the upper bound on the size of the process zone given by Theorem \ref{epsiloncontropprocesszone}. The set $K^{+,\epsilon_k}_\alpha$ is defined by \
\begin{eqnarray}
K^{+,\epsilon_k}_\alpha=\{(x,\xi)\in D\times\mathcal{H}_1(0);\,|D_e^{\epsilon_k|\xi|}u^{\epsilon_k}\cdot e|>\underline{k}|\epsilon_k\xi|^{\alpha-1}\} \label{equivdescralpha} \end{eqnarray} where $0<\underline{k}\leq\overline{r}$ and $1/2\leq \alpha<1$. We set $\beta=2\alpha-1$ to see for $(x,\xi)\in K^{+,\epsilon_k}$ that \begin{eqnarray}
\epsilon_k|\xi||D_e^{\epsilon_k|\xi|}u^{\epsilon_k}(x)\cdot e|^2>\underline{k}^2\epsilon_k^\beta|\xi|^\beta. \label{inequalbasic} \end{eqnarray} and recall that the potential function $f(r)=F_1(r)$ is increasing to get \begin{eqnarray}
J(|\xi|)\frac{1}{\epsilon_k}F_{1}(\underline{k}^2\epsilon_k^\beta|\xi|^\beta)=|\xi|J(|\xi|)\frac{1}{\epsilon_k|\xi|}F_{1}(\underline{k}^2\epsilon_k^\beta|\xi|^\beta)
\leq|\xi|J(|\xi|)F_{\epsilon_k|\xi|}(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)\label{firstalpha} \end{eqnarray}
and in addition since $|\xi|^{1-\beta}\leq 1$ we have \begin{eqnarray}
&& \int_{K^{+,\epsilon_k}_\alpha}\frac{1}{\epsilon_k}F_{1}(\underline{k}^2\epsilon_k^\beta|\xi|^\beta)|\xi|^{1-\beta}J(|\xi|)\,d\xi\,dx\leq \int_{K^{+,\epsilon_k}_\alpha}\frac{1}{\epsilon_k}F_{1}(\underline{k}^2\epsilon_k^\beta|\xi|^\beta)J(|\xi|)\,d\xi\,dx\nonumber\\
&&\leq\int_{K^{+,\epsilon_k}_\alpha} |\xi|J(|\xi|)F_{\epsilon_k|\xi|}(D_{e}^{{\epsilon_k}|\xi|}u^{\epsilon_k}\cdot e)\,d\xi\,dx\leq\sup_{t\in [0,T]}\sup_{\epsilon_k}PD^{\epsilon_k}(u^{\epsilon_k}) \label{firstbalpha} \end{eqnarray} Taylor approximation gives \begin{eqnarray}
\frac{1}{\epsilon_k}F_1(\underline{k}^2\epsilon_k^\beta|\xi|^\beta)=\epsilon_k^{\beta-1}\underline{k}^2(f'(0)+o(\epsilon_k^\beta))|\xi|^\beta \label{taylor} \end{eqnarray} Substitution of \eqref{taylor} into the left hand side of \eqref{firstbalpha} gives \begin{eqnarray}
\int_{K^{+,\epsilon_k}_\alpha} |\xi|J(|\xi|)\,d\xi\,dx\leq \frac{\epsilon_k^{1-\beta}}{\underline{k}^2(f'(0)+o(\epsilon_k^\beta))}\left(\sup_{t\in [0,T]}\sup_{\epsilon_k}PD^{\epsilon_k}(u^{\epsilon_k})\right). \label{semifinal} \end{eqnarray} Introduce the characteristic function $\chi_\alpha^{+,\epsilon_k}(x,\xi)$ defined on $D\times\mathcal{H}_1(0)$ taking the value $1$ for $(x,\xi)$ in $K_\alpha^{+,\epsilon_k}$ and zero otherwise. Observe that \begin{eqnarray}
\frac{1}{m}\int_{\mathcal{H}_1(0)}\chi_\alpha^{+,\epsilon_k}(x,\xi)|\xi|J(|\xi|)\,d\xi=P(\{y\in\mathcal{H}_{\epsilon_k}(x);|S^{\epsilon_k}(y,x)|>\underline{k}|y-x|^{\alpha-1}\}), \label{equivlence} \end{eqnarray} so \begin{eqnarray}
&&\int_DP(\{y\in\mathcal{H}_{\epsilon_k}(x);|S^{\epsilon_k}(y,x)|>\underline{k}|y-x|^{\alpha-1}\})dx\nonumber\\ &&\leq \frac{\epsilon_k^{1-\beta}}{m\times \underline{k}^2(f'(0)+o(\epsilon_k^\beta))}\left(\sup_{t\in [0,T]}\sup_{\epsilon_k}PD^{\epsilon_k}(u^{\epsilon_k})\right). \label{semifinaler} \end{eqnarray} For $0<\overline{\theta}\leq 1$, Tchebyshev's inequality delivers \begin{eqnarray}
&& \mathcal{L}^d\left(\{x\in D;\, P\left(\{y\in\mathcal{H}_{\epsilon_k}(x);\,|S^{\epsilon_k}(y,x)|>\underline{k}|y-x|^{\alpha-1}\}\right)>\overline{\theta}\}\right)\nonumber\\ &&\leq \frac{\epsilon_k^{1-\beta}}{m\times \overline{\theta}\underline{k}^2(f'(0)+o(\epsilon_k^\beta))}\left(\sup_{t\in [0,T]}\sup_{\epsilon_k}PD^{\epsilon_k}(u^{\epsilon_k})\right). \label{probnewvariableboundchebyshcv} \end{eqnarray} and Theorem \ref{epsiloncontropprocesszone} follows on applying \eqref{gineq} and \eqref{upperbound}.
Choose $\epsilon_k=\frac{1}{2^k}$ and Theorem \ref{epsiloncontropprocesszone} imples $\mathcal{L}^d(PZ^{\epsilon_k}(\underline{k},\alpha,\overline{\theta},t))<C(\frac{1}{{2}^k})^{1-\beta}$, where $C$ is independent of $t\in [0,T]$. The collection of process zones for $\epsilon_k<\delta$ is written as \begin{eqnarray} CPZ^{\delta}(\underline{k},\alpha,\overline{\theta},t)=\cup_{\epsilon_k<\delta}PZ^{\epsilon_k}(\underline{k},\alpha,\overline{\theta},t) \label{unstablecollection} \end{eqnarray} and from the geometric series we find \begin{eqnarray} \mathcal{L}^d\left(CPZ^{\delta}(\underline{k},\alpha,\overline{\theta},t)\right)<C{\delta}^{1-\beta}. \label{boundcdelta} \end{eqnarray} Theorem \ref{bondunstable} follows noting further that $CPZ^{0}(\underline{k},\alpha,\overline{\theta},t)\subset CPZ^{\delta}(\underline{k},\alpha,\overline{\theta},t)\subset CPZ^{\delta'}(\underline{k},\alpha,\overline{\theta},t)$, for $0<\delta<\delta'$.
\setcounter{equation}{0} \setcounter{theorem}{0} \setcounter{lemma}{0}\setcounter{proposition}{0}\setcounter{remark}{0}\setcounter{remark}{0} \setcounter{definition}{0}\setcounter{hypothesis}{0}
\section{Dynamics and limits of energies that $\Gamma$-converge to Griffith fracture energies} \label{Ninth} In this final section we collect ideas and illustrate how the approach presented in the earlier sections can be used to examine limits of dynamics associated with other energies that $\Gamma$- converge to the Griffith fracture energy. As an example we consider the phase field aproach based on the Ambrosio-Tortorelli approximation for dynamic brittle fracture calculations \cite{BourdinLarsenRichardson}. This model is seen to be a well posed formulation in the sense that existence of solutions can be shown \cite{LarsenOrtnerSuli}. To illustrate the ideas we focus on anti-plane shear and the model is described by an out of plane elastic displacement $u^\epsilon(x,t)$ and phase field $0\leq v^\epsilon(x,t)\leq 1$ defined for points $x$ belonging to the domain $D\subset \mathbb{R}^2$. The potential energy associated with the cracking body is given by the Ambrosio-Tortorelli potential \begin{eqnarray} P^\epsilon(u^\epsilon(t),v^\epsilon(t))=E^\epsilon(u^\epsilon(t),v^\epsilon(t))+H^\epsilon(v^\epsilon(t)), \label{ab} \end{eqnarray} with \begin{eqnarray}
E^\epsilon(u^\epsilon(t),v^\epsilon(t))=\frac{\mu}{2}\int_D a^\epsilon(t)|\nabla u^\epsilon(t)|^2\,dx \label{detailsforab} \end{eqnarray} and \begin{eqnarray}
H^\epsilon(v^\epsilon(t))=\frac{\mathcal{G}_c}{2}\int_D \frac{(1-v^\epsilon(t))^2}{2\epsilon}+2\epsilon|\nabla v^\epsilon(t)|^2\,dx. \label{moredetailsforab} \end{eqnarray} here $a^\epsilon(t)=a^\epsilon(x,t)=(v^\epsilon(x,t))^2+\eta^\epsilon$, with $0<\eta^\epsilon<<\epsilon$. In this model the phase field $v^\epsilon$, provides an approximate description of a freely propagating crack taking the value $1$ for points $(x,t)$ away from the crack and zero on the crack. To formulate the problem we introduce the space $H^1_0(D)$ defined to be displacements $u$ in $H^1(D)$ with zero Dirichlet data on $\partial D$ and the set of functions $H_1^1(D)$ defined to be functions $v$ in $H^1(D)$ for which $v=1$ on $\partial D$. The total energy is given by \begin{eqnarray}
\mathcal{F}(t;u^\epsilon,\partial_t u^\epsilon,v^\epsilon)=\frac{1}{2}\int_D|\partial_t u^\epsilon|^2\,dx+P^\epsilon(u^\epsilon,v^\epsilon)-\int_D f(t)u^\epsilon\,dx. \label{eergytotala} \end{eqnarray} The body force $f(x,t)$ is prescribed and the displacement - phase field pair $(u^\epsilon,v^\epsilon)$ is a solution of the initial boundary value problem \cite{LarsenOrtnerSuli} given by: \begin{eqnarray} \partial_{tt}^2 u^\epsilon-{\rm{div}}\left(a^\epsilon(t)\nabla(u^\epsilon-\partial_t u^\epsilon)\right)=f(t),&& \hbox{ in $D$},\nonumber\\ u^\epsilon=0 \hbox{ and } v^\epsilon=1, &&\hbox{ on $\partial D$},\label{ibvp} \end{eqnarray} for $t\in (0,T]$ with initial conditions $u^\epsilon(0)=u^\epsilon_0$, $\partial_t u^\epsilon(0)=u^\epsilon_1\in H^1_0(D)$, satisfying the crack stability condition \begin{eqnarray} P^\epsilon(u^\epsilon(t),v^\epsilon(t))\leq\inf\left\{P^\epsilon(u^\epsilon(t),v):\, v\in H^1_1(D),\,\,v\leq v^\epsilon(t)\right\} \label{stability} \end{eqnarray} and energy balance \begin{eqnarray}
\mathcal{F}(t;u^\epsilon,\partial_t u^\epsilon,v^\epsilon)=\mathcal{F}(0;u^\epsilon_0,u^\epsilon_1,v^\epsilon_0)-\int_0^t\int_D\,a^\epsilon|\nabla\partial_\tau u^\epsilon|^2\,d\tau-\int_0^t\int_D\,\partial_\tau f u^\epsilon\,dx\,d\tau, \label{eergytotalbalance} \end{eqnarray} for every time $0\leq t\leq T$. Finally the initial condition for the phase field is chosen such that $v^\epsilon(0)=v^\epsilon_0$ where $v^\epsilon_0\in H^1_1(D)$ satisfies the unilateral minimality condition \eqref{stability}. In this formulation the pair $u^\epsilon(t)$, $v^\epsilon(t)$ provides a regularized model for free crack propagation. Here the phase field tracks the evolution of the crack with $v^\epsilon=1$ away from the crack and $v^\epsilon=0$ in the crack set.
For a body force $f(x,t)$ in $C^1([0,T];L^2(D))$ it is shown in \cite{LarsenOrtnerSuli} that there exists at least one trajectory $(u^\epsilon,v^\epsilon)\in H^2((0,T);L^2(D))\cap W^{1,\infty}((0,T); H^1_0(D))\times W^{1,\infty}((0,T); H_1^1(D))$ satisfying \ref{ibvp} in the weak sense, i.e. \begin{eqnarray} \int_D\partial_{tt}^2 u^\epsilon\varphi\,dx+\int_D(a^\epsilon(t)\nabla(u^\epsilon-\partial_t u^\epsilon)\cdot\nabla\varphi\,dx=\int_D f(t)u^\epsilon\,dx, \label{ibvpweak} \end{eqnarray} for all $\varphi$ in $H^1_0(D)$ for almost every $t$ in $(0,T]$, with $u^\epsilon(0)=u^\epsilon_0$, $\partial_t u^\epsilon(0)=u^\epsilon_1$, $v^\epsilon(0)=v^\epsilon_0$, and such that \eqref{stability} and \eqref{eergytotalbalance} are satisfied for all times $0<t\leq T$. We have formulated the problem in a simplified setting to illustrate the ideas and note that this type of evolution is shown to exist for evolutions with more general boundary conditions and for displacements in two and three dimensions, see \cite{LarsenOrtnerSuli}. For future reference we call the pair $(u^\epsilon(t),v^\epsilon(t))$ a phase field fracture evolution.
In what follows we pass to the $\epsilon\rightarrow 0$ limit in the phase field evolutions to show existence of a limiting evolution with bounded linear elastic fracture energy. Here the limit evolution $u^0(t)$ is shown to take values in the space of special functions of bounded variation $SBV(D)$. This space is well known and can be thought of as a specialization of the space SBD introduced earlier in Section \ref{sec5} that is appropriate for the scalar problem treated here. For a treatment of SBV and its relation to fracture the reader is referred to \cite{AmbrosioBrades}.
Applying the techniques developed in previous sections it is possible to state and prove following theorem on the dynamics associated with the phase field evolutions $(u^\epsilon,v^\epsilon)$ in the limit as $\epsilon\rightarrow 0$. \begin{theorem} {\bf Sharp interface limit of phase field fracture evolutions.}\\ \label{epszerolimitofLOS} Suppose for every $\epsilon>0$ that: (a) the potential energy of the initial data $(u_0^\epsilon,v_0^\epsilon)$ is uniformly bounded, ie. $\sup_{\epsilon>0}\left\{P^\epsilon(u^\epsilon_0,v_0^\epsilon)\right\}<\infty$, and that (b) $\Vert u^\epsilon(t)\Vert_{L^\infty(D)}<C$ for all $0<\epsilon$ and $0\leq t\leq T$. Then on passing to a subsequence if necessary in the phase field fracture evolutions $(u^\epsilon,v^\epsilon)$ there exists an anti-plane displacement field $u^0(x,t)$ in $SBV(D)$ for all $t\in [0,T]$ such that $u^0\in C([0,T];L^2(D))$ and \begin{eqnarray} \label{limsbv} \lim_{\epsilon\rightarrow 0}\max_{0\leq t\leq T}\{\Vert u^\epsilon(t)-u^0(t)\Vert_{L^2(D)}\}=0 \end{eqnarray} with \begin{eqnarray} \label{lefm}
GF(u^0)=\frac{\mu}{2}\int_D|\nabla u^0(t)|^2+{G_c}\mathcal{H}^1(J_{u^0(t)})<C \end{eqnarray} for $0\leq t\leq T$. \end{theorem}
For anti-plane shear deformations the energy $GF$ is a special form of the energy $LEFM$ introduced in Section \ref{sec5}.
The strategy we will use for proving Theorem \ref{epszerolimitofLOS} is the same as the one developed in the proofs of Theorems \ref{LimitflowThm} and \ref{LEFMMThm}. This strategy can be made systematic and applied to evolutions associated with potential energies that $\Gamma$- converge to the Griffith fracture energy. It consists of four component parts: \begin{enumerate} \item Constructing upper bounds on the kinetic and potential energy of the evolutions that hold uniformly for $0\leq t\leq T$ and $0<\epsilon$. \item Showing compactness of the evolution $u^\epsilon(t)$ in $L^2(D)$ for each time $0\leq t\leq T$. \item Showing limit points of the sequence $\{u^\epsilon(t)\}$ belong to $SBD(D)$ (or $SBV(D)$ as appropriate) for each time $0\leq t\leq T$. \item $\Gamma$-convergence of the potential energies to the Griffith energy $LEFM$ (or $GF$ as appropriate). \end{enumerate}
Assume first that Parts 1 through 4 hold for the the phase field fracture evolution with potential energies $P^\epsilon$ given by \eqref{ab}. These are used as follows to prove Theorem \ref{epszerolimitofLOS}. Part 1 is applied as in \eqref{lip} to show that the sequence of evolutions $u^\epsilon(t)$ is uniformly Lipschitz continuous in time with respect to the $L^2(D)$ norm, i.e. \begin{eqnarray} \label{lipschitzat}
\Vert u^\epsilon(t_1)-u^\epsilon(t_2)\Vert_{L^2(D)}\leq K |t_1-t_2| \end{eqnarray} for $K$ independent of $\epsilon$ and for any $0\leq t_1<t_2\leq T$. Part 2 together with \eqref{lipschitzat} and the Ascoli theorem imply the existence of a subsequence and a limit $u^0(x,t)\in C([0,T];L^2(D))$ such that the convergence \eqref{limsbv} holds. Part 3 shows that $u^0(x,t)$ belongs to $SBV(D)$ for every time in $[0,T]$. Part 4 together with Part 1 and the lower bound property of $\Gamma$-convergence shows that \eqref{lefm} holds and Theorem \ref{epszerolimitofLOS} follows.
We now establish Parts 1 through 4 for the dynamic phase field fracture evolution introduced in \cite{LarsenOrtnerSuli}. To obtain a uniform bound on the kinetic and potential energy differentiate both sides of the energy balance \eqref{eergytotalbalance} with respect to time to get \begin{eqnarray} \label{gronagain}
&&\frac{d}{dt}\left(\frac{1}{2}\int_D|\partial_t u^\epsilon(t)|^2\,dx+P^\epsilon(u^\epsilon(t),v^\epsilon(t))\right)-\frac{d}{dt}\int_D\,f(t) u^\epsilon\,dx\\
&&=-\int_D a^\epsilon|\nabla \partial_t u^\epsilon|^2\,dx-\int_D\partial_t f u^\epsilon\,dx.\nonumber \end{eqnarray} Manipulation and application of the identity $f\partial_t u^\epsilon=\partial_t(f u^\epsilon)-\partial_t f u^\epsilon$ to \eqref{gronagain} delivers the inequality \begin{eqnarray} \label{gronagainagain}
&&\frac{d}{dt}\left(\frac{1}{2}\int_D|\partial_t u^\epsilon(t)|^2\,dx+P^\epsilon(u^\epsilon(t),v^\epsilon(t))\right)\\ &&\leq \int_D f \partial_t u^\epsilon\,dx.\nonumber \end{eqnarray} Now set \begin{eqnarray}
W^\epsilon(t)=\left(\frac{1}{2}\int_D|\partial_t u^\epsilon(t)|^2\,dx+P^\epsilon(u^\epsilon(t),v^\epsilon(t))\right)+1 \label{westat} \end{eqnarray} and proceed as in Section \ref{GKP} to get \begin{eqnarray}
\left(\frac{1}{2}\int_D|\partial_t u^\epsilon(t)|^2\,dx+P^\epsilon(u^\epsilon(t),v^\epsilon(t))\right) \leq \left(\int_0^t\Vert f(\tau)\Vert_{L^2(D)}\,d\tau +\sqrt{W^\epsilon(0)}\right )^2-1. \label{gineqat} \end{eqnarray} Part 1 easily follows from \eqref{gineqat} noting that $\sup_{\epsilon>0}\{W^{\epsilon}(0)\}<\infty$ is a consequence of hypothesis (a) of Theorem \ref{epszerolimitofLOS}. For this example Parts 2 and 3 follow from the uniform bound of Part 1, hypothesis (b) of Theorem \ref{epszerolimitofLOS} and the well known compactness result for the Ambrosio Tortorelli functional, see for example the remark following Theorem 2.3 of \cite{Giacomini}. Part 4 is given by the Ambrosio-Tortorelli result \cite{AT} as expressed in Theorem 2.3 of \cite{Giacomini}.
\section{Conclusions}
The cohesive model for dynamic brittle fracture evolution presented in this paper does not require extra constitutive laws such as a kinetic relation between crack driving force and crack velocity or a crack branching condition. Instead the evolution of the process zone together with the the fracture set is governed by one equation consistent with NewtonÕs second law given by \eqref{eqofmotion}. This is a characteristic feature of peridynamic models \cite{Silling1}, \cite{States}. This evolution is intrinsic to the formulation and encoded into the nonlocal cohesive constitutive law. Crack nucleation criteria although not necessary to the formulation follow from the dynamics and are recovered here by viewing nucleation as a dynamic instability, this is similar in spirit to \cite{SillingWecknerBobaru} and the work of \cite{BhattacharyaDyal} for phase transitions. Theorem \ref{epsiloncontropprocesszone} explicitly shows how the size of the process zone is controlled by the radius of the horizon. This analysis shows that {\em the horizon size $\epsilon$ for cohesive dynamics is a modeling parameter} that can be calibrated according to the size of the process zone obtained from experimental measurements. The process zone is seen to concentrate on a set of zero volume as the length scale of non-locality characterized by the radius of the horizon $\epsilon$ goes to zero, see Theorem \ref{bondunstable}. In this limit the dynamics is shown (under suitable hypotheses) to coincide with the simultaneous evolution of a displacement crack set pair. Here displacement satisfies the elastic wave equation for points in space-time away from the crack set. The shear and Lam\'e moduli together with the energy release rate are described in terms of moments of the nonlocal potentials.
\section{Acknowlegements} \label{Acknowlegements} The author would like to thank Stewart Silling, Richard Lehoucq and Florin Bobaru for stimulating and fruitful discussions. This research is supported by NSF grant DMS-1211066, AFOSR grant FA9550-05-0008, and NSF EPSCOR Cooperative Agreement No. EPS-1003897 with additional support from the Louisiana Board of Regents.
\end{document}
|
arXiv
|
{
"id": "1411.4609.tex",
"language_detection_score": 0.6058721542358398,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{The reverse Yang-Mills-Higgs flow in a neighbourhood of a critical point}
\author{Graeme Wilkin} \address{Department of Mathematics, National University of Singapore, Singapore 119076} \email{[email protected]}
\date{\today}
\begin{abstract} The main result of this paper is a construction of solutions to the reverse Yang-Mills-Higgs flow converging in the $C^\infty$ topology to a critical point. The construction uses only the complex gauge group action, which leads to an algebraic classification of the isomorphism classes of points in the unstable set of a critical point in terms of a filtration of the underlying Higgs bundle.
Analysing the compatibility of this filtration with the Harder-Narasimhan-Seshadri double filtration gives an algebraic criterion for two critical points to be connected by a flow line. As an application, we can use this to construct Hecke modifications of Higgs bundles via the Yang-Mills-Higgs flow. When the Higgs field is zero (corresponding to the Yang-Mills flow), this criterion has a geometric interpretation in terms of secant varieties of the projectivisation of the underlying bundle inside the unstable manifold of a critical point, which gives a precise description of broken and unbroken flow lines connecting two critical points. For non-zero Higgs field, at generic critical points the analogous interpretation involves the secant varieties of the spectral curve of the Higgs bundle. \end{abstract}
\maketitle
\thispagestyle{empty}
\baselineskip=16pt
\section{Introduction}
There is a well-known relationship between the Yang-Mills heat flow on a Riemann surface and the notion of stability from algebraic geometry. This began with work of Atiyah and Bott \cite{AtiyahBott83} and continued with Donaldson's proof \cite{Donaldson83} of the Narasimhan-Seshadri theorem \cite{NarasimhanSeshadri65} and subsequent work of Daskalopoulos \cite{Daskal92} and Rade \cite{Rade92}, which shows that the Yang-Mills flow converges to a unique critical point which is isomorphic to the graded object of the Harder-Narasimhan-Seshadri double filtration of the initial condition. In the setting of Higgs bundles, a theorem of Hitchin \cite{Hitchin87} and Simpson \cite{Simpson88} shows that a polystable Higgs bundle is gauge equivalent to the minimum of the Yang-Mills-Higgs functional and that this minimum is achieved by the heat flow on the space of metrics. The results of \cite{Wilkin08} show that the theorem of Daskalopoulos and Rade described above extends to the Yang-Mills-Higgs flow on the space of Higgs bundles over a compact Riemann surface. More generally, when the base manifold is compact and K\"ahler, then these results are due to \cite{Donaldson85}, \cite{Donaldson87-2}, \cite{UhlenbeckYau86}, \cite{Simpson88}, \cite{DaskalWentworth04}, \cite{Sibley15}, \cite{Jacob15} and \cite{LiZhang11}.
Continuing on from these results, it is natural to investigate flow lines between critical points. Naito, Kozono and Maeda \cite{NaitoKozonoMaeda90} proved the existence of an unstable manifold of a critical point for the Yang-Mills functional, however their method does not give information about the isomorphism classes in the unstable manifold, and their proof requires a manifold structure on the space of connections (which is not true for the space of Higgs bundles). Recent results of Swoboda \cite{Swoboda12} and Janner-Swoboda \cite{JannerSwoboda15} count flow lines for a perturbed Yang-Mills functional, however these perturbations destroy the algebraic structure of the Yang-Mills flow, and so there does not yet exist an algebro-geometric description of the flow lines in the spirit of the results described in the previous paragraph. Moreover, one would also like to study flow lines for the Yang-Mills-Higgs functional, in which case the perturbations do not necessarily preserve the space of Higgs bundles, which is singular.
The purpose of this paper is to show that in fact there is an algebro-geometric description of the flow lines connecting given critical points of the Yang-Mills-Higgs functional over a compact Riemann surface. As an application, we show that the Hecke correspondence for Higgs bundles studied by Witten in \cite{witten-hecke} has a natural interpretation in terms of gradient flow lines. Moreover, for the Yang-Mills flow, at a generic critical point there is a natural embedding of the projectivisation of the underlying bundle inside the unstable set of the critical point, and the results of this paper show that the isomorphism class of the limit of the downwards flow is determined if the initial condition lies in one of the secant varieties of this embedding, giving us a geometric criterion to distinguish between broken and unbroken flow lines. For the Yang-Mills-Higgs flow the analogous picture involves the secant varieties of the space of Hecke modifications compatible with the Higgs field. At generic critical points of the Yang-Mills-Higgs functional this space of Hecke modifications is the spectral curve of the Higgs bundle.
The basic setup for the paper is as follows. Let $E \rightarrow X$ be a smooth complex vector bundle over a compact Riemann surface with a fixed Hermitian metric and let $\mathcal{B}$ denote the space of Higgs pairs on $E$. The \emph{Yang-Mills-Higgs functional} is \begin{equation*}
\mathop{\rm YMH}\nolimits(\bar{\partial}_A, \phi) := \| F_A + [\phi, \phi^*] \|_{L^2}^2 \end{equation*} and the \emph{Yang-Mills-Higgs flow} is the downwards gradient flow of $\mathop{\rm YMH}\nolimits$ given by the equation \eqref{eqn:YMH-flow}. This flow is generated by the action of the complex gauge group $\mathcal{G}^\mathbb C$. Equivalently, one can fix a Higgs pair and allow the Hermitian metric on the bundle to vary in which case the flow becomes a nonlinear heat equation on the space of Hermitian metrics (cf. \cite{Donaldson85}, \cite{Simpson88}). At a critical point for this flow the Higgs bundle splits into Higgs subbundles and on each subbundle the Higgs structure minimises $\mathop{\rm YMH}\nolimits$. The \emph{unstable set} of a critical point $(\bar{\partial}_A, \phi)$ consists of all Higgs pairs for which a solution to the $\mathop{\rm YMH}\nolimits$ flow \eqref{eqn:YMH-flow} exists for all negative time and converges in the smooth topology to $(\bar{\partial}_A, \phi)$ as $t \rightarrow - \infty$. The first theorem of the paper gives an algebraic criterion for a complex gauge orbit to intersect the unstable set for the Yang-Mills-Higgs flow.
\begin{theorem}[Criterion for convergence of reverse heat flow]\label{thm:unstable-set-intro} Let $E$ be a complex vector bundle over a compact Riemann surface $X$, and let $(\bar{\partial}_A, \phi)$ be a Higgs bundle on $E$. Suppose that $E$ admits a filtration $(E^{(1)}, \phi^{(1)}) \subset \cdots \subset (E^{(n)}, \phi^{(n)}) = (E, \phi)$ by Higgs subbundles such that the quotients $(Q_k, \phi_k) := (E^{(k)}, \phi^{(k)}) / (E^{(k-1)}, \phi^{(k-1)})$ are Higgs polystable and $\slope(Q_k) < \slope(Q_j)$ for all $k < j$. Then there exists $g \in \mathcal{G}^\mathbb C$ and a solution to the reverse Yang-Mills-Higgs heat flow equation with initial condition $g \cdot (\bar{\partial}_A, \phi)$ which converges to a critical point isomorphic to $(Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$.
Conversely, if there exists a solution of the reverse heat flow from the initial condition $(\bar{\partial}_A, \phi)$ converging to a critical point $(Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$ where each $(Q_j, \phi_j)$ is polystable with $\slope(Q_k) < \slope(Q_j)$ for all $k < j$, then $(E, \phi)$ admits a filtration $(E^{(1)}, \phi^{(1)}) \subset \cdots \subset (E^{(n)}, \phi^{(n)}) = (E, \phi)$ whose graded object is isomorphic to $(Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$. \end{theorem}
A key difficulty in the construction is the fact that the space of Higgs bundles is singular, and so the existing techniques for constructing unstable sets (see for example \cite{NaitoKozonoMaeda90} for the Yang-Mills flow or \cite[Sec. 6]{Jost05} in finite dimensions) cannot be directly applied since they depend on the manifold structure of the ambient space. One possibility is to study the unstable set of the function $\| F_A + [\phi, \phi^*] \|_{L^2}^2 + \| \bar{\partial}_A \phi \|_{L^2}^2$ on the space of all pairs $(\bar{\partial}_A, \phi)$ without the Higgs bundle condition $\bar{\partial}_A \phi = 0$, however one would then need a criterion to determine when a point in this unstable set is a Higgs bundle and one would also need a method to determine the isomorphism classes of these points.
The construction in the proof of Theorem \ref{thm:unstable-set-intro} is intrinsic to the singular space since it uses the action of the complex gauge group to map the unstable set for the linearised $\mathop{\rm YMH}\nolimits$ flow (for which we can explicitly describe the isomorphism classes) to the unstable set for the Yang-Mills-Higgs flow. The method used here to compare the flow with its linearisation is called the ``scattering construction'' in \cite{Hubbard05} and \cite{Nelson69} since it originates in the study of wave operators in quantum mechanics (see \cite{ReedSimonVol3} for an overview). The method in this paper differs from \cite{Hubbard05} and \cite{Nelson69} in that (a) the construction here is done using the gauge group action in order to preserve the singular space and (b) the distance-decreasing formula for the flow on the space of metrics \cite{Donaldson85} is used here in order to avoid constructing explicit local coordinates as in \cite{Hubbard05} (the construction of \cite{Hubbard05} requires a manifold structure around the critical points).
As a consequence of Theorem \ref{thm:unstable-set-intro}, we have an algebraic criterion for critical points to be connected by flow lines.
\begin{corollary}[Algebraic classification of flow lines]\label{cor:algebraic-flow-line-intro} Let $x_u = (\bar{\partial}_{A_u}, \phi_u)$ and $x_\ell = (\bar{\partial}_{A_\ell}, \phi_\ell)$ be critical points with $\mathop{\rm YMH}\nolimits(x_u) > \mathop{\rm YMH}\nolimits(x_\ell)$. Then $x_u$ and $x_\ell$ are connected by a flow line if and only if there exists a Higgs pair $(E, \phi)$ which has Harder-Narasimhan-Seshadri double filtration whose graded object is isomorphic to $x_\ell$, and which also admits a filtration $(E^{(1)}, \phi^{(1)}) \subset \cdots \subset (E^{(n)}, \phi^{(n)}) = (E, \phi)$ by Higgs subbundles such that the quotients $(Q_k, \phi_k) := (E^{(k)}, \phi^{(k)}) / (E^{(k-1)} \phi^{(k-1)})$ are polystable and satisfy $\slope(Q_k) < \slope(Q_j)$ for all $k < j$ and the graded object $(Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$ is isomorphic to $x_u$. \end{corollary}
As an application of the previous theorem, we can construct Hecke modifications of Higgs bundles via Yang-Mills-Higgs flow lines. First consider the case of a Hecke modification at a single point (miniscule Hecke modifications in the terminology of \cite{witten-hecke}).
\begin{theorem}\label{thm:hecke-intro}
\begin{enumerate}
\item Let $0 \rightarrow (E', \phi') \rightarrow (E, \phi) \stackrel{v}{\rightarrow} \mathbb C_p \rightarrow 0$ be a Hecke modification such that $(E, \phi)$ is stable and $(E', \phi')$ is semistable, and let $L_u$ be a line bundle with $\deg L_u + 1 < \slope(E') < \slope(E)$. Then there exist sections $\phi_u, \phi_\ell \in H^0(K)$, a line bundle $L_\ell$ with $\deg L_\ell = \deg L_u + 1$ and a metric on $E \oplus L_u$ such that $x_u = (E, \phi) \oplus (L_u, \phi_u)$ and $x_\ell = (E_{gr}', \phi_{gr}') \oplus (L_\ell, \phi_\ell)$ are critical points connected by a $\mathop{\rm YMH}\nolimits$ flow line, where $(E_{gr}', \phi_{gr}')$ is isomorphic to the graded object of the Seshadri filtration of $(E', \phi')$.
\item Let $x_u = (E, \phi) \oplus (L_u, \phi_u)$ and $x_\ell = (E', \phi') \oplus (L_\ell, \phi_\ell)$ be critical points connected by a $\mathop{\rm YMH}\nolimits$ flow line such that $L_u, L_\ell$ are line bundles with $\deg L_\ell = \deg L_u + 1$, $(E, \phi)$ is stable and $(E', \phi')$ is polystable with $\deg L_u + 1 < \slope(E') < \slope(E)$. Then $(E', \phi')$ is the graded object of the Seshadri filtration of a Hecke modification of $(E, \phi)$. If $(E', \phi')$ is Higgs stable then it is a Hecke modification of $(E, \phi)$.
\end{enumerate}
\end{theorem}
For Hecke modifications defined at multiple points, we can inductively apply the above theorem to obtain a criterion for two critical points to be connected by a broken flow line. For non-negative integers $m, n$, the definition of $(m, n)$ stability is given in Definition \ref{def:m-n-stable}. The space $\mathcal{N}_{\phi, \phi_u}$ denotes the space of Hecke modifications compatible with the Higgs fields $\phi$ and $\phi_u$ (see Definition \ref{def:Hecke-compatible}).
\begin{corollary}\label{cor:broken-hecke-intro} Consider a Hecke modification $0 \rightarrow (E', \phi') \rightarrow (E, \phi) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$ defined by $n > 1$ distinct points $\{ v_1, \ldots, v_n \} \in \mathbb{P} E^*$, where $(E, \phi)$ is $(0,n)$ stable. If there exists $\phi_u \in H^0(K)$ such that $v_1, \ldots, v_n \in \mathcal{N}_{\phi, \phi_u}$, then there is a broken flow line connecting $x_u = (E, \phi) \oplus (L_u, \phi_u)$ and $x_\ell = (E_{gr}', \phi_{gr}') \oplus (L_\ell, \phi_\ell)$, where $(E_{gr}', \phi_{gr}')$ is the graded object of the Seshadri filtration of the semistable Higgs bundle $(E', \phi')$. \end{corollary}
For any gradient flow, given upper and lower critical sets $C_u$ and $C_\ell$, one can define the spaces $\mathcal{F}_{\ell, u}$ (resp. $\mathcal{BF}_{\ell,u}$) of unbroken flow lines (resp. broken or unbroken flow lines) connecting these sets, and the spaces $\mathcal{P}_{\ell, u}$ (resp. $\mathcal{BP}_{\ell, u}$) of pairs of critical points connected by an unbroken flow line (resp. broken or unbroken flow line). These spaces are correspondences with canonical projection maps to the critical sets given by the projection taking a flow line to its upper and lower endpoints. \begin{equation*} \xymatrix{
& \mathcal{F}_{\ell, u} \ar[d] \ar@/_/[ddl] \ar@/^/[ddr] & & & \mathcal{BF}_{\ell, u} \ar[d] \ar@/_/[ddl] \ar@/^/[ddr] \\
& \mathcal{P}_{\ell, u} \ar[dl] \ar[dr] & & & \mathcal{BP}_{\ell, u} \ar[dl] \ar[dr] & \\ C_\ell & & C_u & C_\ell & & C_u } \end{equation*}
In the setting of Theorem \ref{thm:hecke-intro}, let $d = \deg E$ and $r = \rank(E)$ and let $C_u$ and $C_\ell$ be the upper and lower critical sets. There are natural projection maps to the moduli space of semistable Higgs bundles $C_u \rightarrow \mathcal{M}_{ss}^{Higgs}(r, d)$ and $C_\ell \rightarrow \mathcal{M}_{ss}^{Higgs}(r, d-1)$. Suppose that $\gcd(r,d) = 1$ so that $\mathcal{M}_{ss}^{Higgs}(r, d)$ consists solely of stable Higgs pairs and hence any Hecke modification is semistable. Since the flow is $\mathcal{G}$-equivariant, then there is an induced correspondence variety, denoted $\mathcal{M}_{\ell, u}$ in the diagram below. \begin{equation*} \xymatrix{
& \mathcal{F}_{\ell, u} \ar[d] \ar@/_/[ddl] \ar@/^/[ddr] & \\
& \mathcal{P}_{\ell, u} \ar[dl] \ar[dr] \ar[d] & \\ C_\ell \ar[d] & \mathcal{M}_{\ell,u} \ar[dl] \ar[dr] & C_u \ar[d] \\ \mathcal{M}_{ss}^{Higgs}(r, d-1) & & \mathcal{M}_{ss}^{Higgs}(r,d) } \end{equation*}
As a consequence of Theorem \ref{thm:hecke-intro}, we have the following result.
\begin{corollary}\label{cor:hecke-correspondence-intro} $\mathcal{M}_{\ell,u}$ is the Hecke correspondence. \end{corollary}
A natural question from Floer theory is to ask whether a pair of critical points connected by a broken flow line can also be connected by an unbroken flow line, i.e whether $\mathcal{BP}_{\ell, u} = \mathcal{P}_{\ell, u}$. The methods used to prove the previous theorems can be used to investigate this question using the geometry of secant varieties of the space of Hecke modifications inside the unstable set of a critical point. For critical points of the type studied in Theorem \ref{thm:hecke-intro}, generically this space of Hecke modifications is the spectral curve of the Higgs field, and so the problem reduces to studying secant varieties of the spectral curve. This is explained in detail in Section \ref{sec:secant-criterion}. In particular, Corollary \ref{cor:rank-2-classification} gives a complete classification of the unbroken flow lines on the space of rank $2$ Higgs bundles.
The paper is organised as follows. In Section \ref{sec:preliminaries} we set the notation for the paper, prove a slice theorem around the critical points and derive some preliminary estimates for the $\mathop{\rm YMH}\nolimits$ flow near a critical point. Section \ref{sec:local-analysis} contains the main part of the analysis of the $\mathop{\rm YMH}\nolimits$ flow around a critical point, which leads to the proof of Theorem \ref{thm:unstable-set-intro} and Corollary \ref{cor:algebraic-flow-line-intro}. In Section \ref{sec:hecke} we interpret the analytic results on flow lines in terms of the Hecke correspondence, leading to the proof of Theorem \ref{thm:hecke-intro}, Corollary \ref{cor:broken-hecke-intro} and Corollary \ref{cor:hecke-correspondence-intro}. Appendix \ref{sec:uniqueness} contains a proof that a solution to the reverse $\mathop{\rm YMH}\nolimits$ flow with a given initial condition is necessarily unique.
{\bf Acknowledgements.} I would like to thank George Daskalopoulos, M.S. Narasimhan and Richard Wentworth for their interest in the project, as well as George Hitching for useful discussions about \cite{ChoeHitching10} and \cite{Hitching13}.
\section{Preliminaries}\label{sec:preliminaries}
\subsection{The Yang-Mills-Higgs flow on a compact Riemann surface}
Fix a compact Riemann surface $X$ and a smooth complex vector bundle $E \rightarrow X$. Choose a normalisation so that $\mathop{\rm vol}\nolimits(X) = 2\pi$. Fix $\bar{\partial}_{A_0} : \Omega^0(E) \rightarrow \Omega^{0,1}(E)$ such that $\bar{\partial}_{A_0}$ is $\mathbb C$-linear and satisfies the Leibniz rule $\bar{\partial}_{A_0}(fs) = (\bar{\partial} f) s + f (\bar{\partial}_{A_0} s)$ for all $f \in \Omega^0(X)$ and $s \in \Omega^0(E)$. Let $\mathcal{A}^{0,1}$ denote the affine space $\bar{\partial}_{A_0} + \Omega^{0,1}(\mathop{\rm End}\nolimits(E))$. A theorem of Newlander and Nirenberg identifies $\mathcal{A}^{0,1}$ with the space of holomorphic structures on $E$. The \emph{space of Higgs bundles on $E$} is \begin{equation} \mathcal{B} := \{ (\bar{\partial}_A, \phi) \in \mathcal{A}^{0,1} \times \Omega^{1,0}(\mathop{\rm End}\nolimits(E)) \, : \, \bar{\partial}_A \phi = 0 \} \end{equation}
The complex gauge group is denoted $\mathcal{G}^\mathbb C$ and acts on $\mathcal{B}$ by $g \cdot (\bar{\partial}_A, \phi) = (g \bar{\partial}_A g^{-1}, g \phi g^{-1})$. If $X$ is a complex manifold with $\dim_\mathbb C X > 1$ then we impose the extra integrability conditions $(\bar{\partial}_A)^2 = 0$ and $\phi \wedge \phi = 0$. Given a Hermitian metric on $E$, let $\mathcal{A}$ denote the space of connections on $E$ compatible with the metric, and let $\mathcal{G} \subset \mathcal{G}^\mathbb C$ denote the subgroup of unitary gauge transformations. The Chern connection construction defines an injective map $\mathcal{A}^{0,1} \hookrightarrow \mathcal{A}$ which is a diffeomorphism when $\dim_\mathbb C X = 1$. Given $\bar{\partial}_A \in \mathcal{A}^{0,1}$, let $F_A$ denote the curvature of the Chern connection associated to $\bar{\partial}_A$ via the Hermitian metric. The metric induces a pointwise norm $| \cdot | : \Omega^2(\mathop{\rm End}\nolimits(E)) \rightarrow \Omega^0(X, \mathbb R)$ and together with the Riemannian structure on $X$ an $L^2$ norm $\| \cdot \|_{L^2} : \Omega^2(\mathop{\rm End}\nolimits(E)) \rightarrow \mathbb R$. The \emph{Yang-Mills-Higgs functional} $\mathop{\rm YMH}\nolimits : \mathcal{B} \rightarrow \mathbb{R}$ is defined by \begin{equation}\label{eqn:YMH-def}
\mathop{\rm YMH}\nolimits(\bar{\partial}_A, \phi) = \| F_A + [\phi, \phi^*] \|_{L^2}^2 = \int_X | F_A + [ \phi, \phi^*] |^2 \, dvol \end{equation}
When $\dim_\mathbb C = 1$, the Hodge star defines an isometry $* : \Omega^2(\mathop{\rm End}\nolimits(E)) \rightarrow \Omega^0(\mathop{\rm End}\nolimits(E)) \cong \mathop{\rm Lie}\nolimits \mathcal{G}^\mathbb C$. For any initial condition $(A_0, \phi_0)$, the following equation for $g_t \in \mathcal{G}^\mathbb C$ has a unique solution on the interval $t \in [0, \infty)$ (cf. \cite{Donaldson85}, \cite{Simpson88}) \begin{equation}\label{eqn:gauge-flow} \frac{\partial g}{\partial t} g_t^{-1} = - i * ( F_{g_t \cdot A_0} + [g_t \cdot \phi_0, (g_t \cdot \phi_0)^*]) , \quad g_0 = \id . \end{equation} This defines a unique curve $(A_t, \phi_t) = g_t \cdot (A_0, \phi_0) \in \mathcal{B}$ which is a solution to the downwards Yang-Mills-Higgs gradient flow equations \begin{align}\label{eqn:YMH-flow} \begin{split} \frac{\partial A}{\partial t} & = i \bar{\partial}_A * (F_A + [\phi, \phi^*]) \\ \frac{\partial \phi}{\partial t} & = i \left[ \phi, *(F_A + [\phi, \phi^*]) \right] . \end{split} \end{align} for all $t \in [0, \infty)$. The result of \cite[Thm 3.1]{Wilkin08} shows that the solutions converge to a unique limit $(A_\infty, \phi_\infty)$ which is a critical point of $\mathop{\rm YMH}\nolimits$. Moreover \cite[Thm. 4.1]{Wilkin08} shows that the isomorphism class of this limit is determined by the graded object of the Harder-Narasimhan-Seshadri double filtration of the initial condition $(A_0, \phi_0)$.
\begin{remark}
Since the space $\mathcal{B}$ of Higgs bundles is singular, then we define the gradient of $\mathop{\rm YMH}\nolimits$ as the gradient of the function $\| F_A + [\phi, \phi^*] \|_{L^2}^2$ defined on the ambient smooth space $T^* \mathcal{A}^{0,1}$, which contains the space $\mathcal{B}$ as a singular subset. When the initial condition is a Higgs bundle, then a solution to \eqref{eqn:YMH-flow} is generated by the action of the complex gauge group $\mathcal{G}^\mathbb C$ which preserves $\mathcal{B}$. Therefore the solution to \eqref{eqn:YMH-flow} is contained in $\mathcal{B}$ and so from now on we can consider the flow \eqref{eqn:YMH-flow} as a well-defined gradient flow on the singular space $\mathcal{B}$. Throughout the paper we define a critical point to be a stationary point for the Yang-Mills-Higgs flow. \end{remark}
\begin{definition}\label{def:critical-point} A \emph{critical point} for $\mathop{\rm YMH}\nolimits$ is a pair $(A, \phi) \in \mathcal{B}$ such that \begin{equation}\label{eqn:critical-point} \bar{\partial}_A * (F_A + [\phi, \phi^*]) = 0, \quad \text{and} \quad \left[ \phi, *(F_A + [\phi, \phi^*]) \right] = 0 . \end{equation}
\end{definition}
The critical point equations \eqref{eqn:critical-point} imply that the bundle $E$ splits into holomorphic $\phi$-invariant sub-bundles $E_1 \oplus \cdots \oplus E_n$, such that the induced Higgs structure $(\bar{\partial}_{A_j}, \phi_j)$ on the bundle $E_j$ minimises the Yang-Mills-Higgs functional on the bundle $E_j$ (cf. \cite[Sec. 5]{AtiyahBott83} for holomorphic bundles and \cite[Sec. 4]{Wilkin08} for Higgs bundles). In particular, each Higgs pair $(\bar{\partial}_{A_j}, \phi_j)$ is polystable. The decomposition is not necessarily unique due to the possibility of polystable bundles with the same slope, however it is unique if we impose the condition that $(E_1, \phi_1) \oplus \cdots \oplus (E_n, \phi_n)$ is the graded object of the socle filtration of the Higgs bundle $(E, \phi)$ (see \cite{HuybrechtsLehn97} for holomorphic bundles and \cite[Sec. 4]{BiswasWilkin10} for Higgs bundles). With respect to this decomposition the curvature $*(F_A + [\phi, \phi^*]) \in \Omega^0(\mathop{\rm ad}\nolimits(E)) \cong \mathop{\rm Lie}\nolimits(\mathcal{G}$ has the following block-diagonal form \begin{equation}\label{eqn:critical-curvature} i * (F_A + [\phi, \phi^*]) = \left( \begin{matrix} \lambda_1 \cdot \id_{E_1} & 0 & \cdots & 0 \\ 0 & \lambda_2 \cdot \id_{E_2} & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & \lambda_n \cdot \id_{E_n} \end{matrix} \right) \end{equation} where $\lambda_j =\slope(E_j)$ and we order the eigenvalues by $\lambda_j < \lambda_k$ for all $j < k$.
\begin{definition} A \emph{Yang-Mills-Higgs flow line} connecting an upper critical point $x_u = (\bar{\partial}_{A_u}, \phi_u)$ and a lower critical point $x_\ell = (\bar{\partial}_{A_\ell}, \phi_\ell)$ is a continuous map $\gamma : \mathbb{R} \rightarrow \mathcal{B}$ such that \begin{enumerate}
\item $\frac{d\gamma}{dt}$ satisfies the Yang-Mills-Higgs flow equations \eqref{eqn:YMH-flow}, and
\item $\lim_{t \rightarrow - \infty} \gamma(t) = x_u$ and $\lim_{t \rightarrow \infty} \gamma(t) = x_\ell$, where the convergence is in the $C^\infty$ topology on $\mathcal{B}$.
\end{enumerate}
\end{definition}
\begin{definition}\label{def:unstable-set} The \emph{unstable set} $W_{x_u}^-$ of a non-minimal critical point $x_u = (\bar{\partial}_{A_u}, \phi_u)$ is defined as the set of all points $y_0 \in \mathcal{B}$ such that a solution $y_t$ to the Yang-Mills-Higgs flow equations \eqref{eqn:YMH-flow} exists for all $(-\infty, 0]$ and $y_t \rightarrow x$ in the $C^\infty$ topology on $\mathcal{B}$ as $t \rightarrow - \infty$. \end{definition}
\subsection{A local slice theorem}\label{subsec:local-slice}
In this section we define local slices around the critical points and describe the isomorphism classes in the negative slice.
\begin{definition}\label{def:slice} Let $x = (\bar{\partial}_A, \phi) \in \mathcal{B}$. The \emph{slice} through $x$ is the set of deformations orthogonal to the $\mathcal{G}^\mathbb C$ orbit at $x$. \begin{equation}\label{eqn:slice-def} S_x = \{ (a, \varphi) \in \Omega^{0,1}(\mathop{\rm End}\nolimits(E)) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)) \mid \bar{\partial}_A^* a -*[*\phi^*, \varphi] = 0, (\bar{\partial}_A + a, \phi + \varphi) \in \mathcal{B} \} . \end{equation} If $x$ is a critical point of $\mathop{\rm YMH}\nolimits$ with $\beta = *(F_A + [\phi, \phi^*])$, then the \emph{negative slice} $S_x^-$ is the subset \begin{equation}\label{eqn:neg-slice-def} S_x^- = \{ (a, \varphi) \in S_x \mid \lim_{t \rightarrow \infty} e^{i \beta t} \cdot (a, \varphi) = 0 \} . \end{equation} \end{definition}
To prove Lemma \ref{lem:slice-theorem} and Proposition \ref{prop:filtered-slice-theorem} below, one needs to first define the slice on the $L_1^2$ completion of the space of Higgs bundles with the action of the $L_2^2$ completion of the gauge group. The following lemma shows that if the critical point $x$ is $C^\infty$ then the elements in the slice $S_x$ are also $C^\infty$.
\begin{lemma}\label{lem:slice-smooth} Let $x = (\bar{\partial}_A, \phi)$ be a critical point of $\mathop{\rm YMH}\nolimits$ in the space of $C^\infty$ Higgs bundles, let $S_x$ be the set of solutions to the slice equations in the $L_1^2$ completion of the space of Higgs bundles and let $\delta x = (a, \varphi) \in S_x$. Then $\delta x$ is $C^\infty$. \end{lemma}
\begin{proof} The slice equations are \begin{align*} \bar{\partial}_A \varphi + [a, \phi] + [a, \varphi] & = 0 \\ \bar{\partial}_A^* a - *[\phi^*, *\varphi] & = 0 \end{align*} Since $(a, \varphi) \in L_1^2$ and $(\bar{\partial}_A, \phi)$ is $C^\infty$, then the second equation above implies that $\bar{\partial}_A^* a \in L_1^2$ and so $a \in L_2^2$ by elliptic regularity. After applying Sobolev multiplication $L_2^2 \times L_1^2 \rightarrow L^4$, then $[a, \varphi] \in L^4$ and so the first equation above implies that $\bar{\partial}_A \varphi \in L^4$, hence $\varphi \in L_1^4$. Repeating this again shows that $\varphi \in L_2^2$, and then one can repeat the process inductively to show that $\delta x = (a, \varphi)$ is $C^\infty$. \end{proof}
The following result gives a local description of the space of Higgs bundles in terms of the slice. The infinitesimal action of $\mathcal{G}^\mathbb C$ at $x \in \mathcal{B}$ is denoted by $\rho_x : \mathop{\rm Lie}\nolimits(\mathcal{G}^\mathbb C) \cong \Omega^0(\mathop{\rm End}\nolimits(E)) \rightarrow \Omega^{0,1}(\mathop{\rm End}\nolimits(E)) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E))$. Explicitly, for $x = (\bar{\partial}_A, \phi)$ and $u \in \Omega^0(\mathop{\rm End}\nolimits(E))$, we have $\rho_x(u) = -(\bar{\partial}_A u, [\phi, u])$. The $L^2$-orthogonal complement of $\ker \rho_x \subseteq \Omega^0(\mathop{\rm End}\nolimits(E))$ is denoted $(\ker \rho_x)^\perp$.
\begin{lemma}\label{lem:slice-theorem} Fix $x \in \mathcal{B}$. Then the map $\psi : (\ker \rho_x)^\perp \times S_x \rightarrow \mathcal{B}$ given by $\psi(u, \delta x) = \exp(u) \cdot (x + \delta x)$ is a local homeomorphism. \end{lemma}
\begin{proof} The result of \cite[Prop. 4.12]{Wilkin08} shows that the statement is true for the $L_1^2$ completion of the space of Higgs bundles and the $L_2^2$ completion of the gauge group, and so it only remains to show that it remains true on restricting to the space of $C^\infty$ Higgs bundles with the action of the group of $C^\infty$ gauge transformations. The proof of this statement follows from elliptic regularity using the same method as \cite[Cor. 4.17]{Wilkin08}. \end{proof}
Now let $x = (\bar{\partial}_A, \phi)$ be a critical point and let $\beta = \mu(x) := *(F_A + [\phi, \phi^*])$. The Lie algebra $\mathop{\rm Lie}\nolimits(\mathcal{G}^\mathbb C) \cong \Omega^0(\mathop{\rm End}\nolimits(E))$ decomposes into eigenbundles for the adjoint action of $e^{i \beta}$. We denote the positive, zero and negative eigenspaces respectively by $\Omega^0(\mathop{\rm End}\nolimits(E)_+)$, $\Omega^0(\mathop{\rm End}\nolimits(E)_0)$ and $\Omega^0(\mathop{\rm End}\nolimits(E)_-)$. The positive and negative eigenspaces are nilpotent Lie algebras with associated unipotent groups $\mathcal{G}_+^\mathbb C$ and $\mathcal{G}_-^\mathbb C$. The subgroups of $\mathcal{G}$ and $\mathcal{G}^\mathbb C$ consisting of elements commuting with $e^{i \beta}$ are denoted $\mathcal{G}_\beta$ and $\mathcal{G}_\beta^\mathbb C$ respectively. Since $\Omega^0(\mathop{\rm End}\nolimits(E)_0) \oplus \Omega^0(\mathop{\rm End}\nolimits(E)_+)$ is also a Lie algebra then there is a corresponding subgroup denoted $\mathcal{G}_*^\mathbb C$.
Let $\mathcal{G}_x$ and $\mathcal{G}_x^\mathbb C$ denote the respective isotropy groups of $x$ in $\mathcal{G}$ and $\mathcal{G}^\mathbb C$. There is an inclusion $(\mathcal{G}_x)^\mathbb C \subseteq \mathcal{G}_x^\mathbb C$, however at a non-minimal critical point the two groups may not be equal (in the context of reductive group actions on finite-dimensional affine spaces, this question has been studied by Sjamaar in \cite[Prop. 1.6]{Sjamaar95}). At a general critical point, the Higgs bundle $(E, \phi)$ splits into polystable Higgs sub-bundles $(E_1, \phi_1) \oplus \cdots \oplus (E_n, \phi_n)$, where we order by increasing slope. Then a homomorphism $u \in \mathop{\rm Hom}\nolimits(E_j, E_k)$ satisfying $u \phi_j = \phi_k u$ will be zero if $j > k$ since $(E_j, \phi_j)$ and $(E_k, \phi_k)$ are polystable and $\slope(E_j) > \slope(E_k)$, however if $j < k$ then the homomorphisms do not necessarily vanish in which case $(\mathcal{G}_x)^\mathbb C \subsetneq \mathcal{G}_x^\mathbb C$. Therefore $\ker \rho_x = \mathop{\rm Lie}\nolimits(\mathcal{G}_x^\mathbb C) \subset \Omega^0(\mathop{\rm End}\nolimits(E)_+) \oplus \Omega^0(\mathop{\rm End}\nolimits(E)_0)$, and so $\mathcal{G}_x^\mathbb C \subset \mathcal{G}_*^\mathbb C$.
The result of \cite[Thm. 2.16]{Daskal92} shows that the $L_2^2$ completion of the gauge group satisfies $\mathcal{G}^\mathbb C \cong \mathcal{G}_*^\mathbb C \times_{\mathcal{G}_\beta} \mathcal{G}$. We will use $(\ker \rho_x)_*^\perp$ to denote $(\ker \rho_x)^\perp \cap (\Omega^0(\mathop{\rm End}\nolimits(E))_+ \oplus \Omega^0(\mathop{\rm End}\nolimits(E)_0)$. At a critical point $x$, the above argument shows that isotropy group $\mathcal{G}_x^\mathbb C$ is contained in $\mathcal{G}_*^\mathbb C$, and so we have the following refinement of Lemma \ref{lem:slice-theorem}. \begin{proposition}\label{prop:filtered-slice-theorem} Let $x \in \mathcal{B}$ be a critical point of $\mathop{\rm YMH}\nolimits$. Then there exists a $\mathcal{G}$-invariant neighbourhood $U$ of $x$ and a neighbourhood $U'$ of $[\id, 0, 0]$ in $\mathcal{G} \times_{\mathcal{G}_\beta} \left( (\ker \rho_x)_*^\perp \times S_x \right)$ such that $\psi : U' \rightarrow U$ is a $\mathcal{G}$-equivariant homeomorphism. \end{proposition}
The results of Section \ref{sec:local-analysis} show that the negative slice $S_x^-$ is complex gauge-equivalent to the unstable set $W_x^-$ of a critical point. The next lemma gives a complete classification of the isomorphism classes in $S_x^-$. Together with the results of Section \ref{sec:local-analysis}, this is used in Section \ref{sec:hecke} to classify critical points connected by flow lines.
\begin{lemma}\label{lem:classify-neg-slice} Let $x = (E_1, \phi_1) \oplus \cdots \oplus (E_n, \phi_n)$ be a critical point of $\mathop{\rm YMH}\nolimits$ with curvature as in \eqref{eqn:critical-curvature} with the Higgs polystable subbundles ordered so that $\slope(E_j) < \slope(E_k)$ iff $j < k$. If $\delta x \in S_x^- \cap U$ then $x + \delta x$ has a filtration $(E^{(1)}, \phi^{(1)}) \subset \cdots \subset (E^{(n)}, \phi^{(n)})$ by Higgs subbundles such that the successive quotients are $(E^{(k)}, \phi^{(k)}) / (E^{(k-1)}, \phi^{(k-1)}) = (E_k, \phi_k)$. Conversely, there exists a neighbourhood $U$ of $x$ such that if a Higgs bundle $y = (E, \phi) \in U$ admits such a filtration then it is gauge equivalent to $x + \delta x$ for some $\delta x \in S_x^-$. \end{lemma}
\begin{proof} The first statement follows directly from the definition of the negative slice in \eqref{eqn:neg-slice-def}.
Let $\mathop{\rm End}\nolimits(E)_-$ be the subbundle of $\mathop{\rm End}\nolimits(E)$ corresponding to the negative eigenspaces of $i \beta$ and let $\rho_x^- : \Omega^0(\mathop{\rm End}\nolimits(E)_-) \rightarrow \Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-)$ be the restriction of the infinitesimal action to the negative eigenspaces. Then \begin{equation*} \im \rho_x^- = \im \rho_x \cap \Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-) \end{equation*} and \begin{equation}\label{eqn:negative-orthogonal} \ker (\rho_x^-)^* \supseteq \ker \rho_x^* \cap \Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-) \end{equation} Since $\im \rho_x \oplus \ker \rho_x^* \cong \Omega^{0,1}(\mathop{\rm End}\nolimits(E)) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E))$ by \cite[Lem. 4.9]{Wilkin08} then \begin{align*} \Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-) & = \left( \im \rho_x \oplus \ker \rho_x^* \right) \cap \left( \Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-) \right) \\
& \subseteq \left( \im \rho_x^- \oplus \ker (\rho_x^-)^* \right) \subseteq \Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-) \end{align*} and so \eqref{eqn:negative-orthogonal} must be an equality, therefore $\Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-) \cong \im \rho_x^- \oplus \ker (\rho_x^-)^*$. Therefore the function \begin{align*} \psi^- : (\ker \rho_x^-)^\perp \times \ker (\rho_x^-)^* & \rightarrow \Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-) \\
(u, \delta x) & \mapsto e^u \cdot (x + \delta x) \end{align*} is a local diffeomorphism at $0$. If $\delta x \in S_x^-$ then $x + \delta x \in \mathcal{B}$, and so $e^u \cdot (x + \delta x) \in \mathcal{B}$, since the complex gauge group preserves the space of Higgs bundles. Conversely, if $e^u \cdot (x + \delta x) \in \mathcal{B}$ then $x + \delta x \in \mathcal{B}$ and so $\delta x \in S_x^-$. Therefore $\psi$ restricts to a local homeomorphism $(\ker \rho_x^-)^\perp \times S_x^- \rightarrow \mathcal{B} \cap \left( \Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-) \right)$. \end{proof}
The next two results concern a sequence of points $g_t \cdot z$ in a $\mathcal{G}^\mathbb C$ orbit which approach a critical point $x$ in the $L_k^2$ norm and for which $\mathop{\rm YMH}\nolimits(z) < \mathop{\rm YMH}\nolimits(x)$. Since $x$ is critical and $\mathop{\rm YMH}\nolimits(z) < \mathop{\rm YMH}\nolimits(x)$ then $x \in \overline{\mathcal{G}^\mathbb C \cdot z} \setminus \mathcal{G}^\mathbb C \cdot z$, and therefore $\| g_t \|_{L_{k+1}^2} \rightarrow \infty$. The result below shows that the $C^0$ norm of the function $\sigma(h_t) = \mathop{\rm Tr}\nolimits(h_t) + \mathop{\rm Tr}\nolimits(h_t^{-1}) - 2 \rank (E)$ must also blow up.
\begin{lemma}\label{lem:GIT-C0-norm-blows-up} Let $x \in \mathcal{B}$ be a critical point of $\mathop{\rm YMH}\nolimits$ and let $z \in \mathcal{B}$ be any point such that $\mathop{\rm YMH}\nolimits(z) < \mathop{\rm YMH}\nolimits(x)$. Suppose that there exists a sequence of gauge transformations $g_t \in \mathcal{G}^\mathbb C$ such that $g_t \cdot z \rightarrow x$ in $L_k^2$. Then the change of metric $h_t = g_t^* g_t$ satisfies $\sup_X \sigma(h_t) \rightarrow \infty$. \end{lemma}
\begin{proof} Let $U$ be the neighbourhood of $x$ from Lemma \ref{lem:slice-theorem}. Since $g_t \cdot z \rightarrow x$, then there exists $T$ such that $g_t \cdot z \in U$ for all $t \geq T$. Therefore there exists $f_t$ in a neighbourhood of the identity in $\mathcal{G}^\mathbb C$ such that $f_t \cdot g_t \cdot z \in S_x$. The uniqueness of the decomposition from the slice theorem shows that if $t > T$, then $f_{t} \cdot g_{t} \cdot z = f_{t, T} \cdot f_{T} \cdot g_{T} \cdot z$ with $f_{t,T} \in \mathcal{G}_x^\mathbb C$. Therefore $t \rightarrow \infty$ implies that $f_{t,T}$ diverges in $\mathcal{G}_x^\mathbb C$. Fix a point $p$ on the surface $X$, and let $\mathcal{G}_0^\mathbb C$ be the normal subgroup of complex gauge transformations that are the identity at $p$, as in \cite[Sec. 13]{AtiyahBott83}. We have the following short exact sequence of groups \begin{equation*} 1 \rightarrow \mathcal{G}_0^\mathbb C \rightarrow \mathcal{G}^\mathbb C \rightarrow \mathsf{GL}(n, \mathbb C) \rightarrow 1 . \end{equation*} Since $\mathcal{G}_0^\mathbb C$ acts freely on the space of connections (and hence on $\mathcal{B}$), then restriction to the fibre over $p$ defines a bijective correspondence between $\mathcal{G}_x^\mathbb C \subset \mathcal{G}^\mathbb C$ and a subgroup of $\mathsf{GL}(n, \mathbb C)$ via the exact sequence above. Therefore $f_{t,T}$ diverges in $\mathcal{G}_x^\mathbb C$ implies that the restriction of $f_{t,T}$ to the fibre over $p$ diverges in $\mathsf{GL}(n, \mathbb C)$, and so the $C^0$ norm of $f_{t,T}$ diverges to $\infty$, and hence the same is true for $g_t = f_t^{-1} \cdot f_{t, T} \cdot f_T \cdot g_T \cdot z$ since $g_T$ is fixed and both $f_t$ and $f_T$ are contained in a fixed neighbourhood of the identity in $\mathcal{G}^\mathbb C$. Therefore $\sup_X \sigma(h_t) \rightarrow \infty$. \end{proof}
\begin{corollary}\label{cor:bounded-metric-away-from-critical} Let $x$ be a critical point of $\mathop{\rm YMH}\nolimits$. Then for each neighbourhood $V$ of $x$ in the $L_k^2$ topology on $\mathcal{B}$ and each constant $C > 0$, there exists a neighbourhood $U$ of $x$ such that if $z \notin V$ and $\mathop{\rm YMH}\nolimits(z) < \mathop{\rm YMH}\nolimits(x)$, then $y = g \cdot z$ with $h = g^* g$ satisfying $\sup_X \sigma(h) \leq C$ implies that $y \notin U$. \end{corollary}
\begin{proof} If no such neighbourhood $U$ exists, then we can construct a sequence $y_t = g_t \cdot z$ converging to $x$ in $L_k^2$ such that $h_t = g_t^* g_t$ satisfies $\sup_X \sigma(h_t) \leq C$ for all $t$, however this contradicts the previous lemma. \end{proof}
\subsection{Modifying the $\mathop{\rm YMH}\nolimits$ flow in a neighbourhood of a critical point}
Let $x$ be a critical point, let $\beta = \mu(x) = *(F_A + [\phi, \phi^*])$, and let $\mathcal{G}_*^\mathbb C$ be the subgroup defined in the previous section. In this section we explain how to modify the $\mathop{\rm YMH}\nolimits$ flow near $x$ so that the gauge transformation generating the flow is contained in $\mathcal{G}_*^\mathbb C$. The reason for modifying the flow is so that we can apply the distance-decreasing formula of Lemma \ref{lem:modified-distance-decreasing}, which is used for the convergence result of Section \ref{subsec:inverse-construction}.
Let $U$ be a $\mathcal{G}$-invariant neighbourhod of $x$ such that $U$ is homeomorphic to a neighbourhood of $[\id, 0, 0]$ in $\mathcal{G} \times_{\mathcal{G}_\beta} \left( (\ker \rho_x)_*^\perp \times S_x \right)$ by Proposition \ref{prop:filtered-slice-theorem}. Let $V \subset U$ be the image of $(\ker \rho_x)_*^\perp \times S_x$ under the homeomorphism from Proposition \ref{prop:filtered-slice-theorem}. For each $y \in V$, let $\gamma_-(y)$ be the component of $\mu(y)$ in $\Omega^0(\mathop{\rm End}\nolimits(E)_-)$. Since $\mu$ is $\mathcal{G}$-equivariant then we can extend $\gamma_-$ equivariantly from $V$ to all of $U$ using the action of $\mathcal{G}$. Define the map $\gamma : U \rightarrow \mathop{\rm Lie}\nolimits(\mathcal{G})$ by \begin{equation}\label{eqn:def-gamma} \gamma(y) = \gamma_-(y) - \gamma_-(y)^* \end{equation}
\begin{definition}\label{def:modified-flow} The \emph{modified flow} with initial condition $y_0 \in U$ is the solution to \begin{equation}\label{eqn:modified-flow} \frac{dy}{dt} = - I \rho_x(\mu(y)) + \rho_x(\gamma(y)) . \end{equation} More explicitly, on the space of Higgs bundles $y = (\bar{\partial}_A, \phi)$ satisfies \begin{align*} \frac{\partial A}{\partial t} & = i \bar{\partial}_A * (F_A + [\phi, \phi^*]) - \bar{\partial}_A \gamma(\bar{\partial}_A, \phi) \\ \frac{\partial \phi}{\partial t} & = i [\phi, *(F_A + [\phi, \phi^*])] - [\phi, \gamma(\bar{\partial}_A, \phi)] \end{align*} \end{definition}
In analogy with \eqref{eqn:gauge-flow}, the modified flow is generated by the action of the gauge group $y_t = g_t \cdot y_0$, where $g_t$ satisfies the equation \begin{equation}\label{eqn:modified-gauge-flow} \frac{\partial g_t}{\partial t} g_t^{-1} = -i \mu(g_t \cdot y_0) + \gamma(g_t \cdot y_0), \quad g_0 = \id . \end{equation} As before, let $V \subset U$ be the image of $(\ker \rho_x)_*^\perp \times S_x$ under the homeomorphism from the slice theorem (Proposition \ref{prop:filtered-slice-theorem}). Note that if $y_0 \in V$ then $\frac{\partial g_t}{\partial t} g_t^{-1} \in \mathop{\rm Lie}\nolimits(\mathcal{G}_*^\mathbb C)$, so $g_t \in \mathcal{G}_*^\mathbb C$ and the solution to the modified flow remains in $V$ for as long as it remains in the neighbourhood $U$.
\begin{lemma}\label{lem:relate-flows} Let $y_t = g_t \cdot y_0$ be the solution to the $\mathop{\rm YMH}\nolimits$ flow \eqref{eqn:gauge-flow} with initial condition $y_0$. Then there exists $s_t \in \mathcal{G}$ solving the equation \begin{equation}\label{eqn:unitary-modification} \frac{ds}{dt} s_t^{-1} = \gamma(s_t \cdot y_t), \quad s_0 = \id \end{equation} such that $\tilde{y}_t = s_t \cdot y_t$ is a solution to the modified flow equation \eqref{eqn:modified-flow} with initial condition $y_0$. \end{lemma}
\begin{proof} Since $\gamma$ is $\mathcal{G}$-equivariant then \eqref{eqn:unitary-modification} reduces to \begin{equation*} \frac{ds}{dt} s_t^{-1} = \mathop{\rm Ad}\nolimits_{s_t} \gamma(y_t) . \end{equation*} Since $\gamma(y_t) \in \mathop{\rm Lie}\nolimits(\mathcal{G})$ is already defined by the gradient flow $y_t$, then this equation reduces to solving an ODE on the fibres of the bundle, and therefore existence of solutions follows from ODE existence theory. Let $\tilde{g}_t = s_t \cdot g_t$. A calculation shows that \begin{align*} \frac{d \tilde{g}_t}{dt} \tilde{g}_t^{-1} & = \frac{ds}{dt} s_t^{-1} + \mathop{\rm Ad}\nolimits_{s_t} \left( \frac{dg}{dt} g_t^{-1} \right) \\
& = \gamma(s_t \cdot y_t) - i \mathop{\rm Ad}\nolimits_{s_t} \mu(y_t) \\
& = \gamma(\tilde{y}_t) - i \mu(\tilde{y}_t) \\
& = \gamma(\tilde{g}_t \cdot y_0) - i \mu(\tilde{g}_t \cdot y_0) , \end{align*} and so $\tilde{y}_t = \tilde{g}_t \cdot y_0 = s_t \cdot y_t$ is a solution to the modified flow \eqref{eqn:modified-flow} with initial condition $y_0$. \end{proof}
As a corollary, we see that the change of metric is the same for the YMH flow \eqref{eqn:gauge-flow} and the modified flow \eqref{eqn:modified-gauge-flow}.
\begin{corollary}\label{cor:metrics-same} Let $y_t = g_t \cdot y_0$ be a solution to the Yang-Mills-Higgs flow equation \eqref{eqn:gauge-flow} and $\tilde{y}_t = \tilde{g}_t \cdot y_0$ be a solution to the modified flow equation \eqref{eqn:modified-gauge-flow}. Then $h_t = g_t^* g_t = \tilde{g}_t^* \tilde{g}_t$. \end{corollary}
Finally, we prove that convergence for the upwards $\mathop{\rm YMH}\nolimits$ flow implies convergence for the modified flow. \begin{lemma}\label{lem:unstable-sets-same} Let $x$ be a critical point and let $y_0 \in W_x^-$. Then the modified flow with initial condition $y_0$ exists for all $t \in (-\infty, 0]$ and converges in the $C^\infty$ topology to a point in $\mathcal{G} \cdot x$. \end{lemma}
\begin{proof} Let $y_t$ be the $\mathop{\rm YMH}\nolimits$ flow with initial condition $y_0$ and $\tilde{y_t} = s_t \cdot y_t$ the modified flow. By the definition of $W_x^-$ the $\mathop{\rm YMH}\nolimits$ flow exists for all $t \in (-\infty, 0]$ and $y_t \rightarrow x$ in the $C^\infty$ topology. Existence of the modified flow then follows from Lemma \ref{lem:relate-flows}. Proposition \ref{prop:exponential-convergence} shows that $y_t \rightarrow x$ exponentially in $L_k^2$ for all $k$, and so the same is true for $\gamma(y_t)$. Therefore the length of the modified flow line satisfies \begin{align*}
\int_{-\infty}^0 \| I \rho_{\tilde{y}_t}(\mu(\tilde{y}_t)) - \rho_{\tilde{y}_t}(\gamma(\tilde{y}_t)) \|_{L_k^2} \, dt & = \int_{-\infty}^0 \| I \rho_{y_t}(\mu(y_t)) - \rho_{y_t}(\gamma(y_t)) \|_{L_k^2} \, dt \\
& \leq \int_{-\infty}^0 \| \rho_{y_t}(\mu(y_t)) \|_{L_k^2} \, dt + \int_{-\infty}^0 \| \rho_{y_t}(\gamma(y_t)) \|_{L_k^2} \, dt \end{align*}
which is finite since the length $\int_{-\infty}^0 \| \rho_{y_t}(\mu(y_t)) \|_{L_k^2} \, dt$ of the $\mathop{\rm YMH}\nolimits$ flow line is finite, $y_t$ is bounded and $\gamma(y_t) \rightarrow 0$ exponentially. This is true for all $k$, and so the modified flow converges in the $C^\infty$ topology. \end{proof}
\subsection{Preliminary estimates for the $\mathop{\rm YMH}\nolimits$ flow in a neighbourhood of a critical point}
Given eigenvalues for $i \beta$ labelled by $\lambda_1 \leq \cdots \leq \lambda_k < 0 \leq \lambda_{k+1} \leq \cdots$, for any $y \in S_x^-$ and any norm, we have the Lipschitz bounds \begin{equation}\label{eqn:lipschitz-slice}
e^{\lambda_1 t} \| y - x \| \leq \| e^{i \beta t} \cdot y - x \| \leq e^{\lambda_k t} \| y - x \| . \end{equation}
\begin{lemma}\label{lem:moment-map-quadratic}
For any critical point $x$ there exists $C>0$ such that for any $y \in S_{x}^-$, we have $\| \mu(y) - \beta \|_{C^0} \leq C \| y - x \|_{C^0}^2$. \end{lemma}
\begin{proof} Let $y \in S_x^-$ and define $\delta y := y-x \in V \cong T_x V$. Then the defining equation for the moment map shows that for all $v \in \mathfrak{k}$, we have \begin{equation*} d \mu_x (\delta y) \cdot v = \omega(\rho_x(v), \delta y) = \left< I \rho_x(v), \delta y \right> \end{equation*} By the definition of the slice, each $\delta y \in S_x^-$ is orthogonal to the infinitesimal action of $\mathcal{G}^\mathbb C$ at $x$, and so $\left< I \rho_x(v), \delta y \right>=0$ for all $v \in \mathfrak{k}$. Therefore $d \mu_x(\delta y) = 0$. Since the moment map $\mu(\bar{\partial}_A, \phi) = F_A + [\phi, \phi^*]$ is quadratic, then we have \begin{equation*}
\| \mu(y) - \mu(x) \|_{C^0} \leq \| d\mu_x(\delta y) \|_{C^0} + C \| \delta y \|_{C^0}^2 = C \| \delta y \|_{C^0}^2 . \end{equation*} Since the moment map is $\mathcal{G}$-equivariant and the norms above are all $\mathcal{G}$-invariant, then the constant $C$ is independent of the choice of critical point in the orbit $\mathcal{G} \cdot x$. \end{proof}
Given $g \in \mathcal{G}^\mathbb C$, let $g^*$ denote the adjoint with respect to the Hermitian metric on $E$ and let $\mathcal{G}$ act on $\mathcal{G}^\mathbb C$ by left multiplication. In every equivalence class of the space of metrics $\mathcal{G}^\mathbb C/ \mathcal{G}$ there is a unique positive definite self-adjoint section $h$, which we use from now on to represent elements of $\mathcal{G}^\mathbb C/ \mathcal{G}$. Given $h = g^* g \in \mathcal{G}^\mathbb C/ \mathcal{G}$, define $\mu_h : \mathcal{B} \rightarrow \Omega^0(\mathop{\rm End}\nolimits(E)) \cong \mathop{\rm Lie}\nolimits (\mathcal{G}^\mathbb C)$ by \begin{equation}\label{eqn:def-muh} \mu_h(y) = \mathop{\rm Ad}\nolimits_{g^{-1}} \left( \mu(g\cdot y) \right) . \end{equation} Since the moment map is $\mathcal{G}$-equivariant, then for any $k \in \mathcal{G}$ we have \begin{equation*} \mathop{\rm Ad}\nolimits_{g^{-1}} \mathop{\rm Ad}\nolimits_{k^{-1}} \left( \mu(k \cdot g \cdot y) \right) = \mathop{\rm Ad}\nolimits_{g^{-1}} \left( \mu(g\cdot y) \right) \end{equation*} and so $\mu_h$ is well-defined on $\mathcal{G}^\mathbb C/ \mathcal{G}$. The length of a geodesic in the space of positive definite Hermitian matrices is computed in \cite[Ch. VI.1]{Kobayashi87}. Following \cite[Prop. 13]{Donaldson85} (see also \cite[Prop. 6.3]{Simpson88}), it is more convenient to define the distance function $\sigma : \mathcal{G}^\mathbb C/ \mathcal{G}\rightarrow \mathbb R$ \begin{equation}\label{eqn:def-sigma} \sigma(h) = \mathop{\rm Tr}\nolimits h + \mathop{\rm Tr}\nolimits h^{-1} - 2 \rank (E) . \end{equation} As explained in \cite{Donaldson85}, the function $\sup_X \sigma$ is not a norm in the complete metric space $\mathcal{G}^\mathbb C/ \mathcal{G}$, however we do have $h_t \stackrel{C^0}{\longrightarrow} h_\infty$ in $\mathcal{G}^\mathbb C/ \mathcal{G}$ if and only if $\sup_X \sigma(h_t h_\infty^{-1}) \rightarrow 0$. Note that if $h_1 = g_1^* g_1$ and $h_2 = g_2^* g_2$, then \begin{equation}\label{eqn:metric-difference} \sigma(h_1 h_2^{-1}) = \sigma \left( g_1^* g_1 g_2^{-1} (g_2^*)^{-1} \right) = \sigma \left( (g_1 g_2^{-1})^* g_1 g_2^{-1} \right) . \end{equation}
Recall from \cite{Donaldson85}, \cite{Simpson88} that we have the following distance-decreasing formula for a solution to the downwards $\mathop{\rm YMH}\nolimits$ flow. Since the change of metric is the same for the modified flow by Corollary \ref{cor:metrics-same}, then \eqref{eqn:distance-decreasing} is also valid for the modified flow.
\begin{lemma}\label{lem:distance-decreasing} Let $y_1, y_2 \in \mathcal{B}$ and suppose that $y_1 = g_0 \cdot y_2$ for $g \in \mathcal{G}^\mathbb C$. For $j = 1,2$, define $y_j(t)$ to be the solution of the $\mathop{\rm YMH}\nolimits$ flow \eqref{eqn:YMH-flow} with initial condition $y_j$. Define $g_t$ by $y_1(t) = g_t \cdot y_2(t)$ and let $h_t = g_t^* g_t$ be the associated change of metric. Then \begin{equation}\label{eqn:distance-decreasing} \left( \frac{\partial}{\partial t} + \Delta \right) \sigma(h_t) \leq 0 . \end{equation} \end{lemma}
Since $\mathop{\rm Lie}\nolimits(\mathcal{G}_*^\mathbb C) = \Omega^0(\mathop{\rm End}\nolimits(E))_0 \oplus \Omega^0(\mathop{\rm End}\nolimits(E))_+$ and the adjoint action of $e^{-i \beta t}$ is the identity on $\Omega^0(\mathop{\rm End}\nolimits(E))_0$ and strictly contracting on $\Omega^0(\mathop{\rm End}\nolimits(E))_+$, then we have the following lemma which is used in Section \ref{subsec:inverse-construction}.
\begin{lemma}\label{lem:modified-distance-decreasing} Given any $g_0 \in \mathcal{G}_*^\mathbb C$, let $g_t = e^{-i \beta t} g_0 e^{i \beta t}$ and $h_t = g_t^* g_t$. Then $\frac{\partial}{\partial t} \sigma(h_t) \leq 0$. \end{lemma}
As part of the proof of the distance-decreasing formula in \cite{Donaldson85} we also have the following inequalities. This result is used in the proof of Lemma \ref{lem:uniform-bound-sigma}. \begin{lemma}\label{lem:metric-inequalities} For any metric $h \in \mathcal{G}^\mathbb C / \mathcal{G}$ and any $y \in \mathcal{B}$, we have \begin{align*} -2i \mathop{\rm Tr}\nolimits \left( (\mu_h(y) - \mu(y)) h \right) + \Delta \mathop{\rm Tr}\nolimits(h) & \leq 0 \\ 2i \mathop{\rm Tr}\nolimits \left( (\mu_h(y) - \mu(y)) h^{-1} \right) + \Delta \mathop{\rm Tr}\nolimits(h) & \leq 0 . \end{align*} \end{lemma}
\subsection{Exponential convergence of the backwards flow}
In this section we prove that if a solution to the backwards $\mathop{\rm YMH}\nolimits$ flow converges to a critical point, then it must do so exponentially in each Sobolev norm.
\begin{proposition}\label{prop:exponential-convergence}
Let $y_t$ be a solution to the $\mathop{\rm YMH}\nolimits$ flow \eqref{eqn:YMH-flow} such that $\lim_{t \rightarrow -\infty} y_t = x$. Then for each positive integer $k$ there exist positive constants $C_1$ and $\eta$ such that $\| y_t - x \|_{L_k^2} \leq C_1 e^{\eta t}$ for all $t \leq 0$. \end{proposition}
The proof of the proposition reduces to the following lemmas. First recall from the slice theorem that there is a unique decomposition \begin{equation*} y = e^u \cdot (x + z) \end{equation*} for $u \in (\ker \rho_x)^\perp$ and $z \in S_x$. We can further decompose $z = z_{\geq 0} + z_-$, where $z_- \in S_x^-$ is the component of $z$ in the negative slice and $z_{\geq 0} = z - z_-$. At the critical point $x$ we have the decomposition $\mathop{\rm End}\nolimits(E) \cong \mathop{\rm End}\nolimits(E)_+ \oplus \mathop{\rm End}\nolimits(E)_0 \oplus \mathop{\rm End}\nolimits(E)_-$ according to the eigenspaces of $i \beta$ (cf. Sec. \ref{subsec:local-slice}). Then with respect to this decomposition $z_{\geq 0}$ is the component of $z$ in $\Omega^{0,1}(\mathop{\rm End}\nolimits(E))_+ \oplus \mathop{\rm End}\nolimits(E)_0) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_+ \oplus \mathop{\rm End}\nolimits(E)_0)$ and $z_-$ is the component in $\Omega^{0,1}(\mathop{\rm End}\nolimits(E)_-) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_-)$. In terms of the action of $\beta = \mu(x)$ we have $\lim_{t \rightarrow \infty} e^{i \beta t} \cdot z_- = 0$ and $\lim_{t \rightarrow \infty} e^{- i \beta t} \cdot z_{\geq 0} = z_0$, where $z_0$ is the component of $z$ in $\Omega^{0,1}(\mathop{\rm End}\nolimits(E)_0) \oplus \Omega^{1,0}(\mathop{\rm End}\nolimits(E)_0)$. Note that if $y = e^u \cdot (x + z)$ is a Higgs bundle, then $x+z$ is a Higgs bundle since $e^u \in \mathcal{G}^\mathbb C$ preserves the space of Higgs bundles, however $x+z_{\geq 0}$ may not be a Higgs bundle as the pair $(\bar{\partial}_{A_{\geq 0}}, \phi_{\geq 0})$ representing $x + z_{\geq 0}$ may not satisfy $\bar{\partial}_{A_{\geq 0}} \phi_{\geq 0} = 0$. Even though $\phi_{\geq 0}$ may not be holomorphic, we can still apply the principle that curvature decreases in subbundles and increases in quotient bundles and follow the same idea as \cite[Sec. 8 \& 10]{AtiyahBott83} to prove the following lemma.
\begin{lemma}\label{lem:non-holomorphic-extensions} \begin{enumerate}
\item $\mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0})) \geq \mathop{\rm YMH}\nolimits(x)$.
\item $\grad \mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0}))$ is tangent to the set $\{z_- = 0\}$. \end{enumerate} \end{lemma}
The next lemma shows that the component in the negative slice is decreasing exponentially.
\begin{lemma}
Let $y_t = e^u \cdot (x + z_{\geq 0} + z_-)$ be a solution to the $\mathop{\rm YMH}\nolimits$ flow such that $\lim_{t \rightarrow -\infty} y_t = x$. Then there exist positive constants $K_1$ and $K_2$ such that $\| z_- \|_{L_1^2}^2 \leq K_1 e^{K_2 t}$ for all $t \leq 0$. \end{lemma}
\begin{proof} The proof follows the idea of \cite[Sec. 10]{Kirwan84}. The downwards gradient flow equation for $z_-$ is \begin{equation*} \frac{\partial z_-}{\partial t} = L z_- + N_-(u, z_{\geq 0}, z_-) \end{equation*}
where $L$ is a linear operator and the derivative of $N_-$ vanishes at the origin. Since $z_-$ is orthogonal to the $\mathcal{G}^\mathbb C$ orbit through $x$, then the Laplacian term in $\grad \mathop{\rm YMH}\nolimits$ vanishes on $z_-$ and so the linear part satisfies $e^{Lt} z_- = e^{-i \beta t} \cdot z_-$. Since $z_-$ is in the negative slice then there exists $\lambda_{min} > 0$ such that $\left< L z_- , z_- \right>_{L_1^2} \geq \lambda_{min} \| z_- \|_{L_1^2}$. Now Lemma \ref{lem:non-holomorphic-extensions} shows that the $\mathop{\rm YMH}\nolimits$ flow preserves the set $\{ z_- = 0 \}$, and so $N_-(u, z_{\geq 0}, 0) = 0$. Since $N_-$ is $C^1$ with vanishing derivative at the origin then for all $\varepsilon > 0$ there exists $\delta > 0$ such that if $\| y_t - x \|_{L_1^2} < \delta$ then \begin{equation*}
\| N_-(u, z_{\geq 0}, z_-) \|_{L_1^2} \leq \varepsilon \| z_- \|_{L_1^2} \end{equation*} Therefore \begin{equation*}
\frac{1}{2} \frac{\partial}{\partial t} \| z_- \|_{L_1^2}^2 = \left< L z_-, z_- \right>_{L_1^2} + \left< N_-(u, z_{\geq 0}, z_- ), z_- \right>_{L_1^2} \geq (\lambda_{min} - \varepsilon) \| z_- \|_{L_1^2}^2 , \end{equation*}
and so if $\varepsilon > 0$ is small enough (e.g. $\varepsilon < \frac{1}{2} \lambda_{min}$) then there exist positive constants $K_1$ and $K_2$ such that $\| z_- \|_{L_1^2}^2 \leq K_1 e^{K_2 t}$ for all $t \leq 0$. \end{proof}
The next lemma shows that the difference $\mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(y_t)$ is decreasing exponentially.
\begin{lemma}\label{lem:f-exponential} Let $y_t = e^u \cdot (x+z_{\geq 0} + z_-)$ be a solution to the $\mathop{\rm YMH}\nolimits$ flow such that $\lim_{t \rightarrow -\infty} y_t = x$. Then there exist positive constants $K_1'$ and $K_2'$ such that \begin{equation*}
\mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0} + z_-)) \leq K_1' e^{K_2' t} \end{equation*} for all $t \leq 0$. \end{lemma}
\begin{proof} Recall that the Morse-Kirwan condition from Lemma \ref{lem:non-holomorphic-extensions} implies \begin{equation*} \mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0})) - \mathop{\rm YMH}\nolimits(x) \geq 0 \end{equation*}
Since $x$ is a critical point of $\mathop{\rm YMH}\nolimits$, then for all $\varepsilon > 0$ there exists $\delta > 0$ such that if $\| y_t - x \|_{L_1^2} < \delta$ we have \begin{equation*}
\mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0} + z_-)) - \mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0})) \geq -\varepsilon \| z_- \|_{L_1^2} . \end{equation*} Therefore \begin{align*} \mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0} + z_-)) - \mathop{\rm YMH}\nolimits(x) & = \mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0} + z_-)) - \mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0})) \\
& \quad \quad + \mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0})) - \mathop{\rm YMH}\nolimits(x) \\
& \geq - \varepsilon \|z_- \|_{L_1^2} \geq - \varepsilon \sqrt{K_1} e^{\frac{1}{2} K_2 t} \end{align*} Since $\mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0} + z_-))$ is monotone decreasing with $t$ and $\lim_{t \rightarrow -\infty} \mathop{\rm YMH}\nolimits(e^u \cdot (x+z_{\geq 0} + z_-)) = \mathop{\rm YMH}\nolimits(x)$, then $\mathop{\rm YMH}\nolimits(e^u \cdot (x + z_{\geq 0} + z_-)) \leq \mathop{\rm YMH}\nolimits(x)$, and so the above equation implies that \begin{equation*}
\left| \mathop{\rm YMH}\nolimits(y_t) - \mathop{\rm YMH}\nolimits(x) \right| \leq K_1' e^{K_2' t} \end{equation*} for positive constants $K_1' = \varepsilon \sqrt{K_1}$ and $K_2' = \frac{1}{2} K_2$. \end{proof}
\begin{lemma}\label{lem:interior-bound} Let $y_t$ be a solution to the $\mathop{\rm YMH}\nolimits$ flow such that $y_t \rightarrow x$ as $t \rightarrow -\infty$. Then for each positive integer $k$ there exists a constant $C$ and a constant $\tau_0 \in \mathbb R$ such that \begin{equation*}
\| y_\tau - x \|_{L_k^2} \leq C \int_{-\infty}^\tau \| \grad \mathop{\rm YMH}\nolimits(y_s) \|_{L^2} \, ds \end{equation*} for all $\tau \leq \tau_0$. \end{lemma}
\begin{proof} Recall the interior estimate from \cite[Lem. 7.3]{Rade92}, \cite[Prop. 3.6]{Wilkin08} which says that for all positive integers $k$ there exists a neighbourhood $U$ of $x$ in the $L_k^2$ topology and a constant $C$ such that if $y_t \in U$ for all $t \in [0, T]$ then \begin{equation*}
\int_1^T \| \grad \mathop{\rm YMH}\nolimits(y_t) \|_{L_k^2} \, dt \leq C \int_0^T \| \grad \mathop{\rm YMH}\nolimits(y_t) \|_{L^2} \, dt . \end{equation*} The constant $C$ is uniform as long as the initial condition satisfies a uniform bound on the derivatives of the curvature of the underlying holomorphic bundle and the flow line $y_t$ remains in the fixed neighbourhood $U$ of the critical point $x$ (cf. \cite[Prop. A]{Rade92}). In particular, the estimates of \cite[Lem. 3.14, Cor 3.16]{Wilkin08} show that this bound on the curvature is satisfied for any initial condition along a given flow line $y_t$. \emph{A priori} the constant depends on $T$, however it can be made uniform in $T$ using the following argument. Let $C$ be the constant for $T = 2$. For any $T \geq 2$, let $N$ be an integer greater than $T$ such that $y_t \in U$ for all $t \in [0, N]$. We then have \begin{align*}
\int_1^T \| \grad \mathop{\rm YMH}\nolimits(y_t) \|_{L_k^2} \, dt & \leq \sum_{n=1}^{N-1} \int_n^{n+1} \| \grad \mathop{\rm YMH}\nolimits(y_t) \|_{L_k^2} \, dt \\
& \leq C \sum_{n=1}^{N-1} \int_{n-1}^{n+1} \| \grad \mathop{\rm YMH}\nolimits(y_t) \|_{L^2} \, dt \\
& \leq 2 C \int_0^N \| \grad \mathop{\rm YMH}\nolimits(y_t) \|_{L^2} \, dt \end{align*}
Since $\lim_{t \rightarrow - \infty} y_t = x$ in the $C^\infty$ topology, then for any $\varepsilon > 0$ there exists $\tau_0$ such that $\tau \leq \tau_0$ implies that $\| y_t - x \|_{L_k^2} < \varepsilon$ for all $t \leq \tau$ and therefore by choosing $\varepsilon$ small we can apply the above interior estimate on any interval $[t, \tau]$ for $\tau \leq \tau_0$. Therefore we have the bound \begin{equation*}
\int_t^\tau \| \grad \mathop{\rm YMH}\nolimits(y_s) \|_{L_k^2} \, ds \leq 2C \int_{-\infty}^\tau \| \grad \mathop{\rm YMH}\nolimits(y_s) \|_{L^2} \, ds \end{equation*} For fixed $\tau$ the right-hand side of the above inequality is constant, and so \begin{equation*}
\| y_\tau - x \|_{L_k^2} \leq \int_{-\infty}^\tau \| \grad \mathop{\rm YMH}\nolimits(y_s) \|_{L_k^2} \, ds \leq 2C \int_{-\infty}^\tau \| \grad \mathop{\rm YMH}\nolimits(y_s) \|_{L^2} \, ds \end{equation*} \end{proof}
\begin{proof}[Proof of Proposition \ref{prop:exponential-convergence}] After possibly shrinking the neighbourhood $U$ from the previous lemma, we can apply the Lojasiewicz inequality (cf. \cite[Prop. 3.5]{Wilkin08}) which implies that \begin{equation*}
\int_{-\infty}^\tau \| \grad \mathop{\rm YMH}\nolimits(y_s) \|_{L^2} \, ds \leq \frac{1}{C\theta} \left( f(x) - f(y_\tau) \right)^\theta \end{equation*} for constants $C > 0$ and $\theta \in (0, \frac{1}{2}]$. Lemma \ref{lem:f-exponential} shows that \begin{equation*} \left( f(x) - f(y_\tau) \right)^\theta \leq (K_1')^\theta e^{\theta K_2' t} \end{equation*} for all $t \leq 0$. These two estimates together with the result of Lemma \ref{lem:interior-bound} show that \begin{equation*}
\| y_t - x \|_{L_k^2} \leq C_1 e^{\eta t} \end{equation*} for some positive constants $C_1, \eta$ and all $t \leq 0$. \end{proof}
\section{The isomorphism classes in the unstable set}\label{sec:local-analysis}
Given a critical point $x \in \mathcal{B}$, in this section we show that for each $y \in S_x^-$ there exists a smooth gauge transformation $g \in \mathcal{G}^\mathbb C$ such that $g \cdot y \in W_x^-$ (Proposition \ref{prop:convergence-group-action}), and conversely for each $y \in W_x^-$ there exists $g \in \mathcal{G}^\mathbb C$ such that $g \cdot y \in S_x^-$ (Proposition \ref{prop:unstable-maps-to-slice}). As a consequence, the isomorphism classes in the unstable set are in bijective correspondence with the isomorphism classes in the negative slice, and so we have a complete description of these isomorphism classes by Lemma \ref{lem:classify-neg-slice}. This leads to Theorem \ref{thm:algebraic-flow-line} which gives an algebraic criterion for two points to be connected by a flow line.
\subsection{Convergence of the scattering construction}\label{sec:scattering-convergence}
The goal of this section is to prove Proposition \ref{prop:convergence-group-action}, which shows that every point in the negative slice $S_x^-$ is complex gauge equivalent to a point in the unstable set $W_x^-$.
The construction involves flowing up towards the critical point on the slice using the linearisation of the $\mathop{\rm YMH}\nolimits$ flow and then flowing down using the $\mathop{\rm YMH}\nolimits$ flow. A similar idea is used by Hubbard in \cite{Hubbard05} for analytic flows around a critical point in $\mathbb C^n$, where the flow on the slice is defined by projecting the flow from the ambient space. Hubbard's construction uses the fact that the ambient space is a manifold to (a) define this projection to the negative slice, and (b) define local coordinates in which the nonlinear part of the gradient flow satisfies certain estimates in terms of the eigenvalues for the linearised flow \cite[Prop. 4]{Hubbard05}, which is necessary to prove convergence. This idea originated in the study of the existence of scattering states in classical and quantum mechanics. In the context of this paper, one can think of the linearised flow and the YMH flow as two dynamical systems and the goal is to compare their behaviour as $t \rightarrow - \infty$ (see \cite[Ch. XI.1]{ReedSimonVol3} for an overview). As noted in \cite{Hubbard05}, \cite{Nelson69} and \cite{ReedSimonVol3}, the eigenvalues of the linearised flow play an important role in comparing the two flows.
The method of this section circumvents the need for a local manifold structure by defining the flow on the slice using the linearised flow and then using the distance-decreasing property of the flow on the space of metrics from \cite{Donaldson85}, \cite{Simpson88} (cf. Lemma \ref{lem:distance-decreasing}) in place of the estimate of \cite[Prop. 4]{Hubbard05} on the nonlinear part of the flow. The entire construction is done in terms of the complex gauge group, and so it is valid on any subset preserved by $\mathcal{G}^\mathbb C$, thus avoiding any problems associated with the singularities in the space of Higgs bundles. Moreover, using this method it follows naturally from the Lojasiewicz inequality and the smoothing properties of the heat equation that the backwards $\mathop{\rm YMH}\nolimits$ flow with initial condition in the unstable set converges in the $C^\infty$ topology.
\subsubsection{A $C^0$ bound on the metric}
First we derive an \emph{a priori} estimate on the change of metric along the flow. Fix an initial condition $y_0 \in S_x^-$ and let $\beta = \mu(x) = \Lambda(F_A + [\phi, \phi^*]) \in \Omega^0(\mathop{\rm ad}\nolimits(E)) \cong \mathop{\rm Lie}\nolimits(\mathcal{G})$. In this section we also use the function $\mu_h(y) = \mathop{\rm Ad}\nolimits_{g^{-1}} \left( \mu(g\cdot y) \right)$ from \eqref{eqn:def-muh}. The linearised flow with initial condition $y_0$ has the form $e^{- i \beta t} \cdot y_0$, and the $\mathop{\rm YMH}\nolimits$ flow \eqref{eqn:gauge-flow} has the form $g_t \cdot y_0$. Let $f_t = g_t \cdot e^{i \beta t}$ and define $h_t = f_t^* f_t \in \mathcal{G}^\mathbb C / \mathcal{G}$. This is summarised in the diagram below.
\begin{figure}
\caption{Comparison of the gradient flow and the linearised flow.}
\end{figure}
\begin{lemma}\label{lem:derivative-difference} For any initial condition $y_0 \in S_x^-$, the induced flow on $\mathcal{G}^\mathbb C / \mathcal{G}$ satisfies \begin{equation*} \frac{dh_t}{dt} = -2 i h_t \, \mu_h(e^{-i \beta t} \cdot y_0) + i \beta h_t + h_t (i\beta) \end{equation*} \end{lemma}
\begin{proof} First compute \begin{equation}\label{eqn:group-scattering} \frac{df}{dt} f_t^{-1} = \frac{dg}{dt} g_t^{-1} + g_t (i\beta) e^{i\beta t} f_t^{-1} = -i \mu(g_t \cdot y_0) + f_t (i\beta) f_t^{-1} \end{equation} Then \begin{align*} \frac{dh}{dt} & = \frac{df^*}{dt} f_t + f_t^* \frac{df}{dt} \\
& = f_t^* \left( \frac{df}{dt} f_t^{-1} \right)^* f_t + f_t^* \left( \frac{df}{dt} f_t^{-1} \right) f_t \\
& = - f_t^* i \mu(g_t \cdot y_0) f_t + i \beta h_t - f_t^* i \mu(g_t \cdot y_0) f_t + h_t (i\beta) \\
& = -2 f_t^* i \mu(g_t \cdot y_0) f_t + i \beta h_t + h_t (i\beta) \\
& = -2 i h_t \mathop{\rm Ad}\nolimits_{f_t^{-1}} \left( \mu(g_t \cdot y_0) \right) + i \beta h_t + h_t (i\beta) \\
& = -2 i h_t \, \mu_h(e^{-i \beta t} \cdot y_0) + i \beta h_t + h_t (i\beta) \end{align*} where the last step follows from the definition of $\mu_h$ in \eqref{eqn:def-muh} and the fact that $e^{-i \beta t} = f_t^{-1} \cdot g_t$. \end{proof}
The next estimate gives a bound for $\sup_X \sigma(h_t)$ in terms of $\| y_0 - x \|_{C^0}$.
\begin{lemma}\label{lem:uniform-bound-sigma}
For every $\varepsilon > 0$ there exists a constant $C > 0$ such that for any initial condition $y_0 \in S_{x}^-$ with $\| e^{-i \beta T} \cdot y_0 - x \|_{C^0} < \varepsilon$ we have the estimate $\sup_X \sigma(h_t) \leq C \| e^{-i \beta T} \cdot y_0 - x \|_{C^0}^2$ for all $0 \leq t \leq T$. \end{lemma}
\begin{proof} Taking the trace of the result of Lemma \ref{lem:derivative-difference} gives us \begin{align*} \frac{d}{dt} \mathop{\rm Tr}\nolimits h_t = \mathop{\rm Tr}\nolimits \left( \frac{dh}{dt} \right) & = - 2i \mathop{\rm Tr}\nolimits \left( (\mu_h(e^{-i \beta t} \cdot y_0) - \beta) h_t \right) \\ \frac{d}{dt} \mathop{\rm Tr}\nolimits h_t^{-1} = - \mathop{\rm Tr}\nolimits \left( h_t^{-1} \frac{dh}{dt} h_t^{-1} \right) & = 2i \mathop{\rm Tr}\nolimits \left( h_t^{-1} (\mu_h(e^{-i \beta t} \cdot y_0) - \beta) \right) \end{align*}
Therefore \begin{equation*} \frac{d}{dt} \mathop{\rm Tr}\nolimits ( h_t ) = -2i \mathop{\rm Tr}\nolimits \left( (\mu_h(e^{-i \beta t} \cdot y_0) - \mu(e^{-i\beta t} \cdot y_0) ) h_t \right) - 2i \mathop{\rm Tr}\nolimits \left( (\mu(e^{-i \beta t} \cdot y_0) - \beta) h_t \right) \end{equation*} Lemma \ref{lem:metric-inequalities} together with the fact that $h_t$ is positive definite then shows that \begin{align*} \left( \frac{\partial}{\partial t} + \Delta \right) \mathop{\rm Tr}\nolimits(h_t) & \leq -2i \mathop{\rm Tr}\nolimits \left( (\mu(e^{-i \beta t} \cdot y_0) - \beta) h_t \right) \\
& \leq C_1 \| \mu(e^{-i \beta t} \cdot y_0) - \beta \|_{C^0} \mathop{\rm Tr}\nolimits (h_t) \\
& \leq C_1 \| e^{-i \beta t} \cdot y_0 - x \|_{C^0}^2 \mathop{\rm Tr}\nolimits (h_t) \quad \text{(by Lemma \ref{lem:moment-map-quadratic})} \end{align*} A similar calculation shows that \begin{equation*}
\left( \frac{\partial}{\partial t} + \Delta \right) \mathop{\rm Tr}\nolimits(h_t^{-1}) \leq C_1 \| e^{-i \beta t} \cdot y_0 - x \|_{C^0}^2 \mathop{\rm Tr}\nolimits (h_t^{-1}) \end{equation*}
If we label the eigenvalues of $i \beta$ as $\lambda_1 \leq \cdots \leq \lambda_k < 0 \leq \lambda_{k+1} \leq \cdots \leq \lambda_n$, then the estimate $\| e^{i \beta s} \cdot (y_0 - x) \|_{C^0}^2 \leq e^{2\lambda_k s} \| y_0 - x \|_{C^0}^2$ from \eqref{eqn:lipschitz-slice} gives us \begin{align}\label{eqn:sigma-sub-estimate} \begin{split} \left( \frac{\partial}{\partial t} + \Delta \right) \sigma(h_t) & = \left( \frac{\partial}{\partial t} + \Delta \right) \left( \mathop{\rm Tr}\nolimits (h_t) + \mathop{\rm Tr}\nolimits (h_t^{-1}) \right) \\
& \leq C_1 \| e^{-i \beta t} \cdot (y_0 - x) \|_{C^0}^2 \left( \mathop{\rm Tr}\nolimits (h_t) + \mathop{\rm Tr}\nolimits (h_t^{-1}) \right) \\
& = C_1 \| e^{i \beta (T-t)} \cdot e^{-i \beta T} \cdot (y_0 - x) \|_{C^0}^2 \left( \mathop{\rm Tr}\nolimits (h_t) + \mathop{\rm Tr}\nolimits (h_t^{-1}) \right) \\
& \leq C_1 e^{2\lambda_k (T-t)} \| e^{-i \beta T} \cdot (y_0 - x) \|_{C^0}^2 \sigma(h_t) + C_1 e^{2 \lambda_k (T-t)} \| e^{-i \beta T} \cdot (y_0 - x) \|_{C^0}^2 \rank(E) \end{split} \end{align}
Let $K_1 = C_1 \| e^{-i \beta T} \cdot y_0 - x \|_{C^0}^2$ and $K_2 = C_1 \| e^{-i \beta T} \cdot y_0 - x \|_{C^0}^2 \rank(E)$. Define \begin{equation*} \nu_t = \sigma(h_t) \exp\left( \frac{K_1}{2\lambda_k} e^{2 \lambda_k (T-t)} \right) - \int_0^t K_2 e^{2 \lambda_k (T-s)} \exp \left( \frac{K_1}{2\lambda_k} e^{2 \lambda_k (T-s)} \right) \, ds \end{equation*} Note that $\nu_0 = 0$ since $h_0 = \id$. A calculation using \eqref{eqn:sigma-sub-estimate} then shows that \begin{equation*} \left( \frac{\partial}{\partial t} + \Delta \right) \nu_t \leq 0 \end{equation*} and so $\sup_X \nu_t \leq \sup_X \nu_0 = 0$ by the maximum principle. Therefore \begin{align*} \sup_X \sigma(h_t) & \leq \exp \left(-\frac{K_1}{2\lambda_k} e^{2 \lambda_k (T-t)} \right) \int_0^t K_2 e^{2 \lambda_k (T-s)} \exp \left(\frac{K_1}{2\lambda_k} e^{2 \lambda_k (T-s)} \right) \, ds \\
& \leq \exp \left( - \frac{K_1}{2\lambda_k} \right) \int_0^t K_2 e^{2\lambda_k (T-s)} \, ds \leq C \| e^{-i \beta T} \cdot y_0 - x \|_{C^0}^2 \end{align*}
for some constant $C$, since $\lambda_k < 0$, $0 \leq s \leq t < T$, $K_1$ is bounded since $\| e^{-i \beta T} \cdot y_0 - x \|_{C^0} < \varepsilon$ by assumption and $K_2$ is proportional to $\| e^{-i \beta T} \cdot y_0 - x \|_{C^0}^2$. \end{proof}
\subsubsection{$C^\infty$ convergence in the space of metrics}\label{subsec:metric-convergence}
Now consider the case of a fixed $y_0 \in S_x^-$ and define $y_t = e^{i \beta t} \cdot y_0$. Define $g_s(y_t) \in \mathcal{G}^\mathbb C$ to be the unique solution of \eqref{eqn:gauge-flow} such that $g_s(y_t) \cdot y_t$ is the solution to the $\mathop{\rm YMH}\nolimits$ flow at at time $s$ with initial condition $y_t$. Let $f_s(y_t) = g_s(y_t) \cdot e^{i \beta s} \in \mathcal{G}^\mathbb C$, and define $h_s(y_t) = f_s(y_t)^* f_s(y_t)$ to be the associated change of metric. The estimate from the previous lemma now becomes \begin{equation}\label{eqn:sigma-C0-estimate}
\sup_X \sigma(h_s(y_t)) \leq C \| y_t - x \|_{C^0}^2 = C \| e^{i \beta t} \cdot (y_0 - x) \|_{C^0}^2 \leq C e^{2 \lambda_k t} \| y_0 - x \|_{C^0}^2 \end{equation} This is summarised in the diagram below.
\begin{figure}
\caption{Comparison of $f_{t_1}(y_0) \cdot y_0$ and $f_{t_2}(y_0) \cdot y_0$.}
\label{fig:flow-up-flow-down}
\end{figure}
\begin{proposition}\label{prop:metrics-converge} $h_t(y_0) \stackrel{C^0}{\longrightarrow} h_\infty(y_0) \in \mathcal{G}^\mathbb C / \mathcal{G}$ as $t \rightarrow \infty$. The limit depends continuously on the initial condition $y_0$. The rate of convergence is given by \begin{equation}\label{eqn:metric-convergence-rate}
\sup_X \sigma(h_t(y_0) (h_\infty(y_0))^{-1}) \leq C_2 e^{2 \lambda_k t} \| y_0 - x \|_{C^0}^2 \end{equation} where $C_2 > 0$ is a constant depending only on the orbit $\mathcal{G} \cdot x$. \end{proposition}
\begin{proof} Let $t_1 > t_2 \geq T$. The estimate \eqref{eqn:sigma-C0-estimate} shows that \begin{equation*}
\sup_X \sigma(h_{t_1-t_2}(y_{t_2})) \leq C_2 \| y_{t_2} - x \|_{C^0}^2 \leq C e^{2 \lambda_k t_2} \| y_0 - x \|_{C^0}^2 \leq C e^{2 \lambda_k T} \| y_0 - x \|_{C^0}^2 . \end{equation*} Recall from \eqref{eqn:metric-difference} that \begin{equation*} \sigma(h_{t_1}(y_0) h_{t_2}(y_0)^{-1}) = \sigma\left( ( f_{t_1}(y_0) f_{t_2}(y_0)^{-1} )^* f_{t_1}(y_0) f_{t_2}(y_0)^{-1} \right) . \end{equation*} The distance-decreasing formula of Lemma \ref{lem:distance-decreasing} shows that \begin{equation*} \sup_X \sigma\left( ( f_{t_1}(y_0) f_{t_2}(y_0)^{-1} )^* f_{t_1}(y_0) f_{t_2}(y_0)^{-1} \right) \leq \sup_X \sigma( h_{t_1-t_2}(y_{t_2}) ) . \end{equation*} Therefore the distance (measured by $\sigma$) between the two metrics $h_{t_1}(y_0)$ and $h_{t_2}(y_0)$ satisfies the following bound \begin{align*} \sup_X \sigma(h_{t_1}(y_0) h_{t_2}(y_0)^{-1}) & = \sup_X \sigma\left( ( f_{t_1}(y_0) f_{t_2}(y_0)^{-1} )^* f_{t_1}(y_0) f_{t_2}(y_0)^{-1} \right) \\
& \leq \sup_X \sigma( h_{t_1-t_2}(y_{t_2}) ) \leq C_2 e^{2 \lambda_k T} \| y_0 - x \|_{C^0}^2 \end{align*} and so $h_t(y_0)$ is a Cauchy sequence in $C^0$ with a unique limit $h_\infty \in \mathcal{G}^\mathbb C / \mathcal{G}$, The above equation shows that the rate of convergence is given by \eqref{eqn:metric-convergence-rate}.
Since the finite-time Yang-Mills-Higgs flow and linearised flow both depend continuously on the initial condition, then $h_t(y_0)$ depends continuously on $y_0$ for each $t > 0$. Continuous dependence of the limit then follows from the estimate \eqref{eqn:metric-convergence-rate}. \end{proof}
Now we can improve on the previous estimates to show that $h_t(y_0)$ converges in the smooth topology along a subsequence, and therefore the limit $h_\infty$ is $C^\infty$. Define $z_t = f_t(y_0) \cdot y_0$, where $y_0 \in S_x^-$ and $f_t(y_0) \in \mathcal{G}^\mathbb C$ are as defined in the previous proposition. Given a Higgs bundle $z_t = (\bar{\partial}_A, \phi)$, let $\nabla_A$ denote the covariant derivative with respect to the metric connection associated to $\bar{\partial}_A$.
\begin{lemma}
For each initial condition $y_0 \in S_x^-$, there is a uniform bound on $\sup_X | \nabla_A^\ell \mu(z_t) |$ and $\sup_X |\nabla_A^\ell \phi|$ for each $\ell \geq 0$. \end{lemma}
\begin{proof}
Since $\{ e^{i \beta t} \cdot y_0 \, : \, t \in [0, \infty] \}$ is a compact curve in the space of $C^\infty$ Higgs bundles connecting two $C^\infty$ Higgs bundles $y_0$ and $x$, then $\sup_X \left| \mu(e^{i \beta t} \cdot y_0) \right|$ and $\sup_X \left| \nabla_A \phi \right|$ are both uniformly bounded along the sequence $e^{i \beta t} \cdot y_0$. By construction, $z_t$ is the time $t$ $\mathop{\rm YMH}\nolimits$ flow with initial condition $e^{i \beta t} \cdot y_0$. Along the $\mathop{\rm YMH}\nolimits$ flow, for each $\ell$ the quantities $\sup_X \left| \nabla_A^\ell \mu \right|$ and $\sup_X \left| \nabla_A^\ell \phi \right|$ are both uniformly bounded by a constant depending on the value of $\sup_X \left| \mu \right|$ and $\sup_X \left| \nabla_A \phi \right|$ at the initial condition (cf. \cite[Sec. 3.2]{Wilkin08}). Since these quantities are uniformly bounded for the initial conditions, then the result follows. \end{proof}
\begin{corollary} There is a subsequence $t_n$ such that $h_{t_n} \rightarrow h_\infty$ in the $C^\infty$ topology. Therefore $h_\infty$ is $C^\infty$. \end{corollary}
\begin{proof}
Since $z_t$ is contained in the complex gauge orbit of $y_0$ for all $t$, then \cite[Lem. 3.14]{Wilkin08} shows that the uniform bound on $\left| \nabla_A^\ell \mu(z_t) \right|$ from the previous lemma implies a uniform bound on $\left| \nabla_A^\ell F_A \right|$ for all $\ell$. Therefore, since Proposition \ref{prop:metrics-converge} shows that $h_t$ converges in $C^0$, then the estimates of \cite[Lem. 19 \& 20]{Donaldson85} show that $h_t$ is bounded in $C^\ell$ for all $\ell$, and so there is a subsequence $h_{t_n}$ converging in the $C^\infty$ topology. \end{proof}
\subsubsection{$C^\infty$ convergence in the space of Higgs bundles}\label{sec:convergence-in-B}
In this section we show that the scattering construction converges in the $C^\infty$ topology on the space of Higgs bundles. As a consequence of the methods, we obtain an estimate that shows the solution to the reverse heat flow constructed in Section \ref{subsec:construct-reverse-solution} converges to the critical point $x$ in the smooth topology.
This section uses a slightly modified version of the flow from the previous section, defined as follows. Given $y_0 \in S_x^-$ and $t > 0$, let $x_s = g_s \cdot e^{i \beta t} \cdot y_0$ be the time $s$ solution to the $\mathop{\rm YMH}\nolimits$ flow \eqref{eqn:gauge-flow} with initial condition $e^{i \beta t} \cdot y_0$, let $s(t)$ be the unique point in time such that $\mathop{\rm YMH}\nolimits(x_{s(t)}) = \mathop{\rm YMH}\nolimits(y_0)$ and define $t' = \min \{ t, s(t) \}$. Since the critical values of $\mathop{\rm YMH}\nolimits$ are discrete, then $t'$ is well-defined for small values of $\mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(y_0)$.
\begin{center} \begin{pspicture}(0,-0.5)(8,5) \psline(4,5)(4,0) \psline(4,4)(4,1) \pscurve[arrowsize=5pt]{->}(4,4)(4.4,2.3)(5,1.2)(5.2,1)
\psline[linestyle=dashed](0,1)(8,1) \psdots[dotsize=3pt](4,5)(4,4)(4,1)(5.2,1) \uput{4pt}[180](4,5){\small{$x$}} \uput{4pt}[180](4,4){\small{$e^{i \beta t} \cdot y_0$}} \uput{4pt}[180](4,0.7){\small{$y_0$}} \uput{4pt}[0](4.5,0.7){\small{$z_t = g_{t'} \cdot e^{i \beta t} \cdot y_0$}} \uput{2pt}[180](4,0){$S_x^-$}
\uput{3pt}[30](4.4,2.3){\small{$g_{t'}$}} \uput{3pt}[180](4,2.3){\small{$e^{i \beta t}$}} \uput{3pt}[90](1,1){\small{$\mathop{\rm YMH}\nolimits^{-1}(\mathop{\rm YMH}\nolimits(y_0))$}} \end{pspicture} \end{center}
Now define $z_t = g_{t'} \cdot e^{i \beta t} \cdot y_0$ and $y_t = e^{i \beta (t-t')} \cdot y_0$. Note that $z_t = g_{t'} \cdot e^{i \beta t'} \cdot y_t$ and so the results of the previous section show that the $C^0$ norm of the change of metric connecting $y_t$ and $z_t$ is bounded. Therefore Corollary \ref{cor:bounded-metric-away-from-critical} shows that $y_t$ and $z_t$ are both uniformly bounded away from $x$.
\begin{lemma}\label{lem:bounded-away-from-critical} There exists $T > 0$ such that $t-t' \leq T$ for all $t$. \end{lemma}
\begin{proof}
If $s(t) \geq t$ then $t' = t$ and the desired inequality holds. Therefore the only non-trivial case is $s(t) < t$. Since $\mathop{\rm YMH}\nolimits(z_t) = \mathop{\rm YMH}\nolimits(y_0)$ and $\mathop{\rm YMH}\nolimits$ is continuous in the $L_1^2$ norm on $\mathcal{B}$, then there exists a neighbourhood $V$ of $x$ such that $z_t \notin V$ for all $t$. We also have $z_t = f_{t'} \cdot y_t$ with $f_{t'} = g_{t'} e^{i \beta t'}$ such that $h_t = f_{t'}^* f_{t'}$ satisfies $\sup_X \sigma(h_t) \leq C \| y_t - x \|_{C^0}^2 \leq C \| y_0 - x \|_{C^0}^2$ by Lemma \ref{lem:uniform-bound-sigma}, and so Corollary \ref{cor:bounded-metric-away-from-critical} shows that there exists a neighbourhood $U$ of $x$ in the $L_1^2$ topology on $\mathcal{B}$ such that $y_t \notin U$. Therefore there exists $\eta > 0$ such that $\| y_t - x \|_{L_1^2} \geq \eta$ and $\| z_t - x \|_{L_1^2} \geq \eta$.
Since $y_t = e^{i \beta (t-t')} \cdot y_0$ and $e^{i \beta s} \cdot y_0$ converges to $x$ as $s \rightarrow \infty$, then there exists $T$ such that $t-t' \leq T$ for all $t$, since otherwise $\| y_t - x \|_{L_1^2} < \eta$ for some $t$ which contradicts the inequality from the previous paragraph. \end{proof}
Next we use the Lojasiewicz inequality to derive a uniform bound on $\| z_t - x \|_{L_1^2}$.
\begin{lemma}\label{lem:L12-bound}
Given $\varepsilon > 0$ there exists $\delta > 0$ such that for each $y_0 \in S_x^-$ with $\| y_0 - x \|_{L_1^2} < \delta$ there exists a neighbourhood $U$ of $x$ in the $L_1^2$ topology such that $\| z_t - x \|_{L_1^2} < \varepsilon$ for all $t$ such that $e^{i \beta t} \cdot y_0 \in U$. \end{lemma}
\begin{proof} Recall from \cite[Prop. 3.5]{Wilkin08} that there exists $\varepsilon_1 > 0$ and constants $C > 0$ and $\theta \in \left( 0, \frac{1}{2} \right)$ such that the Lojasiewicz inequality \begin{equation}\label{eqn:lojasiewicz}
\| \grad \mathop{\rm YMH}\nolimits(z) \|_{L^2} \geq C \left| \mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(z) \right|^{1-\theta} \end{equation}
holds for all $z$ such that $\| z - x \|_{L_1^2} < \varepsilon_1$. Recall the interior estimate \cite[Prop. 3.6]{Wilkin08} which says that for any positive integer $k$ there exists $\varepsilon_2 > 0$ and a constant $C_k'$ such that for any solution $x_s = g_s \cdot e^{i \beta t} \cdot y_0$ to the $\mathop{\rm YMH}\nolimits$ flow with initial condition $e^{i \beta t} \cdot y_0$ which satisfies $\| x_s - x \|_{L_k^2} < \varepsilon_2$ for all $0 \leq s \leq S$, then we have \begin{equation}\label{eqn:interior-estimate}
\int_1^S \| \grad \mathop{\rm YMH}\nolimits(x_s) \|_{L_k^2} \, dt \leq C_k' \int_0^S \| \grad \mathop{\rm YMH}\nolimits(x_s) \|_{L^2} \, dt . \end{equation}
where the constant $C_k'$ is uniform over all initial conditions in a given $\mathcal{G}^\mathbb C$ orbit and for all $S$ such that $\| x_s - x \|_{L_k^2} < \varepsilon_2$ for all $s \in [0, S]$ (cf. Lemma \ref{lem:interior-bound}). Define $\varepsilon' = \min \{ \varepsilon, \varepsilon_1, \varepsilon_2 \}$. A calculation using \eqref{eqn:lojasiewicz} (cf. \cite{Simon83}) shows that any flow line $x_s$ which satisfies $\| x_s - x \|_{L_1^2} < \varepsilon'$ for all $s \in [0,t']$ also satisfies the gradient estimate \begin{equation*}
C \theta \| \grad \mathop{\rm YMH}\nolimits(x_s) \|_{L^2} \leq \frac{\partial}{\partial s} \left| \mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(x_s) \right|^\theta \end{equation*}
and so if $\| x_s - x \|_{L_1^2} < \varepsilon'$ for all $s < t'$ then \begin{align}\label{eqn:flow-length-estimate} \begin{split}
\int_0^{t'} \| \grad \mathop{\rm YMH}\nolimits(x_s) \|_{L^2} \, ds & \leq \frac{1}{C\theta} \left( |\mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(x_{t'}) |^\theta - |\mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(x_0)|^\theta \right) \\
& \leq \frac{1}{C\theta} \left| \mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(x_{t'}) \right|^\theta \end{split} \end{align}
Let $k=1$ in \eqref{eqn:interior-estimate} and choose $\delta > 0$ so that $\| y_0 - x \|_{L_1^2} < \delta$ implies that that $\frac{1}{C \theta} |\mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(y_0)|^\theta \leq \frac{\varepsilon'}{3C_1'}$, where $C$ and $\theta$ are the constants from the Lojasiewicz inequality \eqref{eqn:lojasiewicz} and $C_1'$ is the constant from \eqref{eqn:interior-estimate} for $k=1$. Therefore, since $\mathop{\rm YMH}\nolimits(y_0) = \mathop{\rm YMH}\nolimits(x_{t'}) < \mathop{\rm YMH}\nolimits(x_\tau) \leq \mathop{\rm YMH}\nolimits(x)$ for all $\tau < t'$, then \begin{equation}\label{eqn:energy-bound}
\frac{1}{C\theta} \left| \mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(x_\tau) \right|^\theta \leq \frac{\varepsilon'}{3C_1'} \quad \text{for all $\tau < t'$}. \end{equation}
Since the finite-time $\mathop{\rm YMH}\nolimits$ flow depends continuously on the initial condition in the $L_1^2$ norm by \cite[Prop. 3.4]{Wilkin08}, then there exists a neighbourhood $U$ of $x$ such that $x_0 \in U$ implies that $\| x_1 - x \|_{L_1^2} < \frac{1}{3} \varepsilon'$. Choose $t$ large so that $e^{i \beta t} \cdot y_0 = e^{i \beta t'} \cdot y_t \in U$ and let $x_s = g_s \cdot e^{i \beta t} \cdot y_0$ be the solution to the $\mathop{\rm YMH}\nolimits$ flow at time $s$ with initial condition $x_0 = e^{i \beta t} \cdot y_0$. Note that $x_{t'} = z_t$. Define \begin{equation*}
\tau = \sup \{ s \mid \| x_r - x \|_{L_1^2} < \varepsilon' \, \, \text{for all $r \leq s$} \} \end{equation*} and note that $\tau > 0$. By definition of $\tau$, the Lojasiewicz inequality \eqref{eqn:lojasiewicz} and the interior estimate \eqref{eqn:flow-length-estimate} are valid for the flow line $x_s$ on the interval $[0,\tau]$. If $\tau < t'$, then \eqref{eqn:flow-length-estimate} and \eqref{eqn:energy-bound} imply that \begin{align*}
\| x_\tau - x \|_{L_1^2} & \leq \| x_1 - x \|_{L_1^2} + \| x_\tau - x_1 \|_{L_1^2} \\
& < \frac{1}{3} \varepsilon' + \int_1^\tau \| \grad \mathop{\rm YMH}\nolimits(x_s) \|_{L_1^2} \, ds \\
& \leq \frac{1}{3} \varepsilon' + C_1' \int_0^\tau \| \grad \mathop{\rm YMH}\nolimits(x_s) \|_{L^2} \, ds \\
& \leq \frac{1}{3} \varepsilon' + \frac{C_1'}{C \theta} \left| \mathop{\rm YMH}\nolimits(x) - \mathop{\rm YMH}\nolimits(x_\tau) \right|^\theta \leq \frac{1}{3} \varepsilon' + \frac{1}{3} \varepsilon' \end{align*}
contradicting the definition of $\tau$ as the supremum. Therefore $t' \leq \tau$ and the same argument as above shows that $\| x_{t'} - x_0 \|_{L_1^2} < \frac{2}{3} \varepsilon'$, so we conclude that $z_t = x_{t'}$ satisfies $\| z_t - x \|_{L_1^2} < \frac{2}{3} \varepsilon' < \varepsilon$ for all $t$ such that $e^{i \beta t} \cdot y_0 \in U$. \end{proof}
Now that we have a uniform $L_1^2$ bound on $z_t - x$, then we can apply the same idea using the interior estimate \eqref{eqn:interior-estimate} as well as continuous dependence on the initial condition in the $L_k^2$ norm from \cite[Prop. 3.4]{Wilkin08} to prove the following uniform $L_k^2$ bound on $z_t - x$.
\begin{lemma}\label{lem:Lk2-length}
Given $\varepsilon > 0$ and a positive integer $k$ there exists $\delta > 0$ such that if $\| y_0 - x \|_{L_1^2} < \delta$ then there exists a neighbourhood $U$ of $x$ in the $L_k^2$ topology such that $\| z_t - x \|_{L_k^2} < \varepsilon$ for all $t$ such that $e^{i \beta t} \cdot y_0 \in U$. \end{lemma}
Now we can prove that there is a limit $z_\infty$ in the space of $C^\infty$ Higgs bundles. In Section \ref{subsec:construct-reverse-solution} we will show that $z_\infty \in W_x^-$.
\begin{proposition}\label{prop:strong-convergence} For each $y_0 \in S_x^-$, let $z_t$ be the sequence defined above. Then there exists $z_\infty \in \mathcal{B}$ such that for each positive integer $k$ there exists a subsequence of $z_t$ converging to $z_\infty$ strongly in $L_k^2$. \end{proposition}
\begin{proof}
The previous estimate with $k=2$ shows that $\| z_t - x \|_{L_2^2}$ is bounded. Compactness of the embedding $L_{k+1}^2 \hookrightarrow L_k^2$ shows that there is a subsequence $\{ z_{t_n} \}$ converging strongly to a limit $z_\infty$ in $L_1^2$.
For any $k > 1$, the same argument applied to the subsequence $\{ z_{t_n} \}$ from the previous paragraph shows that there exists a further subsequence, which we denote by $\{ z_{t_{n_j}} \}$, which converges strongly in $L_k^2$. Since $z_{t_{n_j}} \stackrel{L_1^2}{\longrightarrow} z_\infty$ then the limit in $L_k^2$ of $z_{t_{n_j}}$ must be $z_\infty$ also. Therefore $z_\infty$ is a $C^\infty$ Higgs pair. \end{proof}
Finally, we can prove that $z_\infty$ is gauge-equivalent to $y_0$. Recall the constant $T$ from Lemma \ref{lem:bounded-away-from-critical} and let $\varphi(z_t,s)$ denote the time $s$ downwards $\mathop{\rm YMH}\nolimits$ flow \eqref{eqn:YMH-flow} with initial condition $z_t$. The gauge transformation $f_t(y_0) \in \mathcal{G}^\mathbb C$ from Proposition \ref{prop:metrics-converge} satisfies $f_t(y_0) \cdot y_0 = \phi(z_t, t-t')$.
For any $k$, let $z_{t_n}$ be a subsequence converging strongly to $z_\infty$ in $L_k^2$. Such a subsequence exists by Proposition \ref{prop:strong-convergence}. Since $0 \leq t_n - t_n' \leq T$ for all $n$ then there exists $s \in [0, T]$ and a subsequence $\{t_{n_\ell} \}$ such that $t_{n_\ell} - t_{n_\ell}' \rightarrow s$. Since the finite-time $\mathop{\rm YMH}\nolimits$ flow depends continuously on the initial condition in $L_k^2$, then $f_{t_{n_\ell}} (y_0) \cdot y_0 = \varphi(z_{t_{n_\ell}}, t_{n_\ell} - t_{n_\ell}')$ converges to $z_\infty^0 := \varphi(z_\infty, s)$ strongly in $L_k^2$. After taking a further subsequence if necessary, the method of Section \ref{subsec:metric-convergence} shows that the change of metric associated to $f_{t_{n_\ell}}(y_0)$ converges strongly in $L_{k+1}^2$. Therefore, since the action of the Sobolev completion $\mathcal{G}_{L_{k+1}^2}^\mathbb C$ on $\mathcal{B}_{L_k^2}$ is continuous, then $\varphi(z_\infty, s)$ (and hence $z_\infty$) is related to $y_0$ by a gauge transformation in $\mathcal{G}_{L_{k+1}^2}^\mathbb C$. Since $y_0$ and $z_\infty$ are both smooth Higgs pairs then an elliptic regularity argument shows that this gauge transformation is smooth. Therefore we have proved the following result.
\begin{proposition}\label{prop:limit-in-group-orbit} Given any $y_0 \in S_x^-$, let $z_\infty$ be the limit from Proposition \ref{prop:strong-convergence}. Then there exists a smooth gauge transformation $g \in \mathcal{G}^\mathbb C$ such that $z_\infty = g \cdot y_0$. \end{proposition}
Since $z_\infty^0 = \varphi(z_\infty, s)$ is related to $z_\infty$ by the finite-time flow and $s$ is bounded, then we have the following estimate for $\| z_\infty^0 - x \|_{L_k^2}$. Note that this requires a bound on $\| y_0 - x \|_{L_1^2}$ for the estimates of this section to work, and a bound on $\| y_0 - x \|_{C^0}$ for the estimates of Lemma \ref{lem:uniform-bound-sigma} to work.
\begin{corollary}\label{cor:flow-bound}
For all $\varepsilon > 0$ there exists $\delta > 0$ such that $\| y_0 - x \|_{L_1^2} + \| y_0 - x \|_{C^0} < \delta$ implies $\| z_\infty^0 - x \|_{L_k^2} < \varepsilon$. \end{corollary}
\begin{remark} The previous proof uses the fact that the \emph{finite-time} flow depends continuously on the initial condition. The limit of the downwards $\mathop{\rm YMH}\nolimits$ flow as $t \rightarrow \infty$ depends continuously on initial conditions within the same Morse stratum (cf. \cite[Thm. 3.1]{Wilkin08}). It is essential that the constant $T$ from Lemma \ref{lem:bounded-away-from-critical} is finite (which follows from Corollary \ref{cor:bounded-metric-away-from-critical}) in order to guarantee that $z_\infty$ and $\varphi(z_\infty, s)$ are gauge equivalent. Without a bound on $T$, it is possible that $z_\infty$ may be in a different Morse stratum to $\lim_{t \rightarrow \infty} \varphi(z_t, t-t')$. \end{remark}
\subsubsection{Constructing a convergent solution to the backwards $\mathop{\rm YMH}\nolimits$ flow}\label{subsec:construct-reverse-solution}
In this section we show that the limit $z_\infty$ is in the unstable set $W_x^-$.
\begin{proposition}\label{prop:convergence-group-action} For each $y_0 \in S_x^-$ there exists $g \in \mathcal{G}^\mathbb C$ such that $g \cdot y_0 \in W_x^-$. \end{proposition}
\begin{proof} In what follows, fix any positive integer $k$. Given $y_0 \in S_x^-$, let $z_t^0 = f_t(y_0) \cdot y_0$, where $f_t$ is the complex gauge transformation from Proposition \ref{prop:metrics-converge}. Then Proposition \ref{prop:limit-in-group-orbit} shows that there exists $z_\infty^0 := \phi(z_\infty, s)$ and a subsequence $\{ z_{t_n}^0 \}$ such that $z_{t_n}^0 \rightarrow z_\infty^0$ strongly in $L_k^2$.
For any $s > 0$, let $y_s = e^{i \beta s} \cdot y_0$ and define $z_t^{-s} = f_t(y_s) \cdot y_s$. By definition, $z_t^0$ is the downwards $\mathop{\rm YMH}\nolimits$ flow for time $s$ with initial condition $z_t^{-s}$. Applying Proposition \ref{prop:strong-convergence} to the subsequence $z_{t_n}^{-s}$ shows that there is a subsequence $z_{t_{n_j}}^{-s}$ converging in $L_k^2$ to some $z_\infty^{-s}$. Since the $\mathop{\rm YMH}\nolimits$ flow for finite time $s$ depends continuously on the initial condition (cf. \cite[Prop. 3.4]{Wilkin08}) then $z_{t_{n_j}}^{-s} \rightarrow z_\infty^{-s}$ and $z_{t_{n_j}}^0 \rightarrow z_\infty^0$ implies that $z_\infty^0$ is the time $s$ flow with initial condition $z_\infty^{-s}$. Therefore, for any $s > 0$ we have constructed a solution to the $\mathop{\rm YMH}\nolimits$ flow on $[-s, 0]$ connecting $z_\infty^0$ and $z_\infty^{-s}$. Proposition \ref{prop:backwards-uniqueness} shows that this solution must be unique for each $s$, and therefore there is a well-defined solution on the time interval $(-\infty, 0]$.
Moreover, we also have the uniform bound from Corollary \ref{cor:flow-bound} which shows that for all $\varepsilon > 0$ there exists $\delta > 0$ such that $\| z_\infty^{-s} - x \|_{L_k^2} \leq \varepsilon$ for all $y_0$ such that $\| y_0 - x \|_{L_1^2} < \delta$. Therefore as $s \rightarrow \infty$, the sequence $z_\infty^{-s}$ converges strongly to $x$ in the $L_k^2$ norm for any $k$, and so $z_\infty^0 = g \cdot y_0 \in W_x^-$. Proposition \ref{prop:exponential-convergence} then shows that the convergence is exponential in each Sobolev norm. \end{proof}
\subsection{Convergence of the inverse process}\label{subsec:inverse-construction}
In this section we consider the inverse procedure to that of the previous section and prove that each point in the unstable set $W_x^-$ is gauge equivalent to a point in the negative slice $S_x^-$. The idea is similar to that of the previous section, except here we use the modified flow.
\subsubsection{A $C^0$ bound in the space of metrics}
Given $y_0 \in W_x^-$, let $y_t = g_t \cdot y_0$ be the solution to the modified flow \eqref{eqn:modified-flow} with initial condition $y_0$. Define $f_t = g_t \cdot e^{i\beta t}$ and let $h_t = f_t^* f_t$. This is summarised in the diagram below.
Using a similar calculation as the previous section, we have the same expression for the change of metric as in Lemma \ref{lem:derivative-difference}. \begin{lemma}\label{lem:reverse-derivative-difference} For any initial condition $y_0 \in W_x^-$, the induced flow on $\mathcal{G}^\mathbb C / \mathcal{G}$ satisfies \begin{equation}\label{eqn:metric-derivative} \frac{dh_t}{dt} = -2i h_t \mu_h(e^{-i \beta t} \cdot y_0) + i \beta h_t + i h_t \beta . \end{equation} \end{lemma}
\begin{proof} A similar calculation as in the proof of Lemma \ref{lem:derivative-difference} (this time using the modified flow \eqref{eqn:modified-flow}) shows that \begin{equation*} \frac{df_t}{dt} f_t^{-1} = -i\mu(g_t \cdot y_0) + \gamma(g_t \cdot y_0) + f_t (i \beta) f_t^{-1} . \end{equation*} Then \begin{align*} \frac{dh_t}{dt} & = f_t^* \left( \frac{df}{dt} f_t^{-1} \right)^* f_t + f_t^* \left( \frac{df}{dt} f_t^{-1} \right) f_t \\
& = f_t^* \left( -i \mu(g_t \cdot y_0) - \gamma(g_t \cdot y_0) + (f_t^*)^{-1} (i \beta) f_t^* - i \mu(g_t \cdot y_0) + \gamma(g_t \cdot y_0) + f_t (i \beta) f_t^{-1} \right) f_t \\
& = -2i h_t f_t^{-1} \mu(g_t \cdot y_0) f_t + i \beta h_t + i h_t \beta \\
& = -2i h_t \mu_h(e^{-i \beta t} \cdot y_0) + i \beta h_t + i h_t \beta . \qedhere \end{align*} \end{proof}
\begin{lemma}\label{lem:reverse-uniform-bound-sigma}
For every $\varepsilon > 0$ there exists a constant $C > 0$ such that for any initial condition $y_0 \in W_{x}^-$ with $\| e^{-i \beta T} \cdot y_0 - x \|_{C^1} + \| g_T \cdot y_0 - x \|_{C^1} < \varepsilon$ we have the estimate \begin{equation*}
\sup_X \sigma(h_t) \leq C \left( \| e^{-i \beta T} \cdot y_0 - x \|_{C^1} + \| g_T \cdot y_0 - x \|_{C^1} \right) \end{equation*} for all $0 \leq t \leq T$. \end{lemma}
\begin{proof}
In contrast to the proof of Lemma \ref{lem:uniform-bound-sigma}, $e^{-i \beta t} \cdot y_0$ is not in the slice $S_x$ and so it satisfies the inequality $\| \mu(e^{-i \beta t} \cdot y_0) - \beta \|_{C^0} \leq C' \| e^{-i \beta t} \cdot y_0 - x \|_{C^1}$ instead of the quadratic bound of Lemma \ref{lem:moment-map-quadratic}. Using this inequality, the same idea as in the proof of Lemma \ref{lem:uniform-bound-sigma} leads to the bound \begin{align}\label{eqn:heat-operator-bounded} \begin{split}
\left( \frac{\partial}{\partial t} + \Delta \right) \sigma(h_t) & \leq C_1 \| e^{-i \beta t} \cdot (y_0-x) \|_{C^1} \left( \mathop{\rm Tr}\nolimits(h_t) + \mathop{\rm Tr}\nolimits(h_t^{-1}) \right) \\
& = C_1 \| e^{-i \beta t} \cdot (y_0 - x) \|_{C^1} \sigma(h_t) + 2 C_1 \| e^{-i \beta t} \cdot (y_0 - x) \|_{C^1} \rank(E) \end{split} \end{align} In general, if the heat operator is bounded for all $t \geq 0$ \begin{equation*} \left( \frac{\partial}{\partial t} + \Delta \right) f(p,t) \leq C(t) f(p,t) + D(t), \quad p \in X, t \in [0, \infty) \end{equation*} for some nonnegative functions $C(t)$ and $D(t)$ independent of $p \in X$, then $f(p,t)$ satisfies the bound \begin{equation}\label{eqn:general-heat-bound} f(p,t) \leq \exp \left( \int_0^t C(s) \, ds \right) \int_0^t D(s) \, ds + f(p,0) \end{equation}
Therefore \eqref{eqn:heat-operator-bounded} implies that the problem reduces to finding a bound for $\int_0^t \| e^{-i \beta s} \cdot (y_0 - x) \|_{C^1} \, ds$. Proposition \ref{prop:exponential-convergence} shows that the backwards flow with initial condition in $W_x^-$ converges exponentially to $x$ in every Sobolev norm. Therefore there exists a neighbourhood $U$ of $x$ such that if $g_T \cdot y_0 \in U$ then there exist positive constants $C_1$ and $\eta$ such that the following estimate holds \begin{equation*}
\| y_0 - x \|_{C^1} \leq C_1 e^{-\eta T} \| g_T \cdot y_0 - x \|_{C^1} . \end{equation*}
Recall the eigenbundles $\mathop{\rm End}\nolimits(E)_-$, $\mathop{\rm End}\nolimits(E)_0$ and $\mathop{\rm End}\nolimits(E)_+$ from Section \ref{sec:preliminaries}. The above estimate shows that each component of $y_0 - x$ in $\mathop{\rm End}\nolimits(E)_-$, $\mathop{\rm End}\nolimits(E)_0$ and $\mathop{\rm End}\nolimits(E)_+$ is bounded by $C_1 e^{-\eta T} \| g_T \cdot y_0 - x \|_{C^1}$. Since the component of $e^{-i \beta t} \cdot (y_0 - x)$ in $\mathop{\rm End}\nolimits(E)_+$ is exponentially decreasing with $t$ then \begin{equation*}
\int_0^T \| (e^{-i \beta t} \cdot y_0 - x)_{\mathop{\rm End}\nolimits(E)_+} \|_{C^1} dt \leq C_1' \| y_0 - x \|_{C^1} \leq C_1 e^{-\eta T} \| g_T \cdot y_0 - x \|_{C^1} . \end{equation*} The component of $e^{-i \beta t} \cdot (y_0 - x)$ in $\mathop{\rm End}\nolimits(E)_0$ is constant with respect to $t$, and so \begin{equation*}
\int_0^T \| (e^{-i \beta t} \cdot y_0 - x)_{\mathop{\rm End}\nolimits(E)_0} \|_{C^1} dt \leq C_2' T \| y_0 - x \|_{C^1} \leq C_2 T e^{-\eta T} \| g_T \cdot y_0 - x \|_{C^1} . \end{equation*} Finally, the component of $e^{-i \beta t} \cdot (y_0 - x)$ in $\mathop{\rm End}\nolimits(E)_-$ is exponentially increasing, and so we have the bound \begin{equation*}
\int_0^T \| (e^{-i \beta t} \cdot y_0 - x)_{\mathop{\rm End}\nolimits(E)_-} \|_{C^1} \leq C_3 \| e^{-i \beta T} \cdot (y_0 - x) \|_{C^1} . \end{equation*}
Combining the estimates for the three components shows that the integral \begin{equation*}
I(t) = \int_0^t \| e^{-i \beta s} \cdot (y_0-x) \|_{C^1} \, ds \end{equation*} is bounded by \begin{equation*}
I(t) \leq C_1 e^{-\eta T} \| g_T \cdot y_0 - x \|_{C^1} + C_2 T e^{-\eta T} \| g_T \cdot y_0 - x \|_{C^1} + C_3 \| e^{-i \beta T} \cdot y_0 - x \|_{C^1} \end{equation*}
The inequality \eqref{eqn:general-heat-bound} together with the assumption $\| g_T \cdot y_0 - x \|_{C^1} + \| e^{-i \beta T} \cdot y_0 - x \|_{C^1} < \varepsilon$ shows that there exists a constant $C$ such that \begin{equation*}
\sup_X \sigma(h_t) \leq C \left( \| e^{-i \beta T} \cdot y_0 - x \|_{C^1} + \| g_T \cdot y_0 - x \|_{C^1} \right) \qedhere \end{equation*} \end{proof}
\subsubsection{Convergence in the space of Higgs bundles}
In this section we use a method analogous to that of Section \ref{sec:convergence-in-B} to show that the sequence converges in the space of Higgs bundles and that the limit is gauge equivalent to $y_0$. Given $y_0 \in W_x^-$ and $t \in (-\infty, 0]$, define $s < 0$ by $\| e^{i \beta s} \cdot g_t(y_0) \cdot y_0 - x \|_{L_k^2} = \| y_0 - x \|_{L_k^2}$. Note that this is well-defined for small values of $\| y_0 - x \|_{L_k^2}$ since Lemma \ref{lem:unstable-sets-same} shows that $g_t(y_0) \cdot y_0 \rightarrow x$ in the $C^\infty$ topology as $t \rightarrow - \infty$ and for $s < 0$ the action of $e^{i \beta s}$ exponentially increases the $C^0$ norm of the component of $g_t(y_0) \cdot y_0$ in $\mathop{\rm End}\nolimits(E)_-$. Now define $t' := \max \{ t, s \} < 0$, let $f_t(y_0) = e^{i \beta t'} \cdot g_{t'}(g_{t-t'}(y_0) \cdot y_0) $ and $z_t := e^{i \beta t'} \cdot g_t(y_0) \cdot y_0 = f_t(y_0) \cdot g_{t-t'}(y_0) \cdot y_0$. Let $h_t = f_t^* f_t$ be the associated change of metric.
\begin{center} \begin{pspicture}(0,-0.5)(8,5.5) \psline[arrowsize=5pt]{->}(4,4)(4,1) \pscurve(3.9,5)(4,4)(4.4,2.3)(5,1.2)(5.2,1)(6,0.4)
\psline[linestyle=dashed](0,1)(8,1) \psdots[dotsize=3pt](3.9,5)(4,4)(4,1)(5.2,1) \uput{4pt}[180](3.9,5){\small{$x$}} \uput{4pt}[180](4,4){\small{$g_t(y_0) \cdot y_0$}} \uput{5pt}[270](3.8,1){\small{$z_t = f_t \cdot y_0$}} \uput{5pt}[270](5.2,1){\small{$y_0$}} \uput{2pt}[270](6,0.4){$W_x^-$}
\uput{3pt}[30](4.4,2.3){\small{$g_{t}$}} \uput{3pt}[180](4,2.3){\small{$e^{i \beta t'}$}}
\uput{3pt}[90](8,1){\small{$\| y_0 - x \|_{L_k^2} = \text{constant}$}} \end{pspicture} \end{center}
Lemma \ref{lem:reverse-uniform-bound-sigma} then shows that $\sup_X \sigma(h_t) \leq C \left( \| z_t - x \|_{C^1} + \| g_{t-t'}(y_0) \cdot y_0 - x \|_{C^1} \right)$. Since either $\| z_t - x \|_{L_k^2} = \| y_0 - x \|_{L_k^2}$ (when $t < t'$) or $g_{t-t'}(y_0) \cdot y_0 = y_0$ (when $t'=t$), then Corollary \ref{cor:bounded-metric-away-from-critical} shows that $g_{t-t'}(y_0) \cdot y_0$ and $z_t$ are both bounded away from $x$ in the $L_k^2$ norm. As a consequence, $|t-t'|$ is uniformly bounded in the same way as Lemma \ref{lem:bounded-away-from-critical}. Therefore \begin{equation}\label{eqn:bounded-linear-flow}
\| e^{i \beta t} \cdot g_t(y_0) \cdot y_0 - x \|_{L_k^2} = \| e^{i \beta (t-t')} \cdot z_t - x \|_{L_k^2} \leq C' \| z_t - x \|_{L_k^2} = C' \| y_0 - x \|_{L_k^2} \end{equation} for some constant $C'$, which implies that there is a subsequence of $e^{i \beta t} \cdot g_t(y_0) \cdot y_0$ converging strongly to a limit $z_\infty^0$ in $L_{k-1}^2$. Since this is true for all $k$, then $z_\infty^0$ is a $C^\infty$ Higgs pair.
A special case of \eqref{eqn:bounded-linear-flow} is \begin{equation}\label{eqn:bounded-linear-flow-C1}
\| e^{i \beta t} \cdot g_t(y_0) \cdot y_0 - x \|_{C^1} \leq C \| e^{i \beta t} \cdot g_t(y_0) \cdot y_0 - x \|_{L_k^2} \leq C' \| y_0 - x \|_{L_k^2} \end{equation} for any $k$ such that $L_k^2 \hookrightarrow C^1$ is an embedding.
By modifying the method of Proposition \ref{prop:metrics-converge} we can now show that the change of metric converges in $C^0$. For $t \in (-\infty, 0]$, define $f_t(y_0) = e^{i \beta t} \cdot g_{t}(y_0)$ and let $t_1 \leq t_2 \leq T < 0$. This is summarised in the diagram below.
\begin{proposition}\label{prop:reverse-metrics-converge} $h_t(y_0)$ converges in the $C^0$ norm to a unique limit $h_\infty(y_0) \in \mathcal{G}^\mathbb C / \mathcal{G}$ as $t \rightarrow -\infty$. The limit depends continuously on the initial condition $y_0 \in S_x^-$. The rate of convergence is given by \begin{equation}\label{eqn:reverse-metric-convergence-rate}
\sup_X \sigma(h_t(y_0) (h_\infty(y_0))^{-1}) \leq C_2 e^{2 \eta t} \| y_0 - x \|_{L_k^2} \end{equation} where $C_2 > 0$ is a constant depending only on the orbit $\mathcal{G} \cdot x$, the constant $\eta$ is from Proposition \ref{prop:exponential-convergence} and $k$ is a positive integer chosen so that $L_k^2 \hookrightarrow C^1$ is a continuous embedding. \end{proposition}
\begin{proof} The result follows from the same procedure as the proof of Proposition \ref{prop:metrics-converge}, except now we use the estimate from Lemma \ref{lem:reverse-uniform-bound-sigma} instead of the estimate from Lemma \ref{lem:uniform-bound-sigma} and the distance-decreasing formula for the modified flow from Lemma \ref{lem:modified-distance-decreasing}.
Let $h_{t_1-t_2}(y_{t_2})$ be the change of metric connecting $y_{t_2} = g_{t_2}(y_0) \cdot y_0$ and $e^{i \beta (t_1-t_2)} \cdot y_{t_1}$. Lemma \ref{lem:reverse-uniform-bound-sigma} and the estimate \eqref{eqn:bounded-linear-flow-C1} above show that $h_{t_1-t_2}(y_{t_2})$ satisfies \begin{align*}
\sup_X \sigma(h_{t_1-t_2}(y_{t_2})) & \leq C \left( \| e^{i \beta (t_1-t_2)} \cdot y_{t_1} - x \|_{C^1} + \| y_{t_2} - x \|_{C^1} \right) \\
& \leq C C' \| y_{t_2} - x \|_{L_k^2} + C \| y_{t_2} - x \|_{C^1} \\
& \leq C'' \| y_T - x \|_{L_k^2} \leq C_2 e^{2 \eta T} \| y_0 - x \|_{L_k^2} \end{align*} By the construction of the modified flow, the gauge transformation connecting $y_{t_2}$ and $e^{i \beta (t_1-t_2)} \cdot y_{t_1}$ is in $\mathcal{G}_*^\mathbb C$, The distance-decreasing formula for the action of $e^{i \beta (t_1 - t_2)}$ from Lemma \ref{lem:modified-distance-decreasing} then implies that \begin{equation*} \sigma(h_{t_1}(y_0) h_{t_2}(y_0)^{-1}) \leq \sigma(h_{t_1-t_2}(y_{t_2})) \end{equation*} and so the sequence $h_t(y_0)$ is Cauchy in the $C^0$ norm, by the same proof as Proposition \ref{prop:metrics-converge}. \end{proof}
Therefore $y_0$ is connected to $z_\infty^0$ by a $C^0$ gauge transformation. Elliptic regularity together with the fact that $z_\infty^0$ is a $C^\infty$ Higgs pair then shows that $y_0$ is gauge equivalent to $z_\infty^0$ by a $C^\infty$ gauge transformation.
The same method as the proof of Proposition \ref{prop:convergence-group-action} then allows us to explicitly construct a solution of the linearised flow $z_\infty^{-s} = e^{i\beta s} \cdot z_\infty^0$ converging to $x$ as $s \rightarrow +\infty$. Lemma \ref{lem:classify-neg-slice} then shows that $z_\infty^0$ is $\mathcal{G}^\mathbb C$ equivalent to a point in $S_x^-$, which is smooth by Lemma \ref{lem:slice-smooth}.
Therefore $y_0$ is $\mathcal{G}^\mathbb C$ equivalent to a point in $S_x^-$, and so we have proved the following converse to Proposition \ref{prop:convergence-group-action}.
\begin{proposition}\label{prop:unstable-maps-to-slice} For each $y_0 \in W_x^-$ there exists a $C^\infty$ gauge transformation $g \in \mathcal{G}^\mathbb C$ such that $g \cdot y_0 \in S_x^-$. \end{proposition}
\subsection{An algebraic criterion for the existence of flow lines}\label{sec:filtration-criterion}
The results of the previous two sections combine to give the following theorem. \begin{theorem}\label{thm:algebraic-flow-line} Let $E$ be a complex vector bundle over a compact Riemann surface $X$, and let $(\bar{\partial}_A, \phi)$ be a Higgs bundle on $E$. Suppose that $E$ admits a filtration $(E^{(1)}, \phi^{(1)}) \subset \cdots \subset (E^{(n)}, \phi^{(n)}) = (E, \phi)$ by Higgs subbundles such that the quotients $(Q_k, \phi_k) := (E^{(k)}, \phi^{(k)}) / (E^{(k-1)}, \phi^{(k-1)})$ are Higgs polystable and $\slope(Q_k) < \slope(Q_j)$ for all $k < j$. Then there exists $g \in \mathcal{G}^\mathbb C$ and a solution to the reverse Yang-Mills-Higgs heat flow equation with initial condition $g \cdot (\bar{\partial}_A, \phi)$ which converges to a critical point isomorphic to $(Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$.
Conversely, if there exists a solution of the reverse heat flow from the initial condition $(\bar{\partial}_A, \phi)$ converging to a critical point $(Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$ then $(\bar{\partial}_A, \phi)$ admits a filtration $(E^{(1)}, \phi^{(1)}) \subset \cdots \subset (E^{(n)}, \phi^{(n)}) = (E, \phi)$ whose graded object is isomorphic to $(Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$. \end{theorem}
\begin{proof} Suppose first that $(\bar{\partial}_A, \phi)$ admits a filtration $(E^{(1)}, \phi^{(1)}) \subset \cdots \subset (E^{(n)}, \phi^{(n)}) = (E, \phi)$ by Higgs subbundles such that the quotients $(Q_k, \phi_k) := (E^{(k)}, \phi^{(k)}) / (E^{(k-1)}, \phi^{(k-1)})$ are Higgs polystable and $\slope(Q_k) < \slope(Q_j)$ for all $k < j$. Let $x$ be a critical point isomorphic to $(Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$, and let $U$ be the neighbourhood of $x$ from Lemma \ref{lem:classify-neg-slice}. Then by applying the isomorphism $x \cong (Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$ and scaling the extension classes there exists a complex gauge transformation such that $g \cdot (\bar{\partial}_A, \phi)$ is in $U$. Applying Lemma \ref{lem:classify-neg-slice} shows that $(\bar{\partial}_A, \phi)$ is isomorphic to a point in $S_x^-$, and therefore Proposition \ref{prop:convergence-group-action} shows that $(\bar{\partial}_A, \phi)$ is isomorphic to a point in $W_x^-$.
Conversely, if $x = (Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$ is a critical point and $(\bar{\partial}_A, \phi) \in W_x^-$, then Proposition \ref{prop:unstable-maps-to-slice} shows that there exists $g \in \mathcal{G}^\mathbb C$ such that $g \cdot (\bar{\partial}_A, \phi) \in S_x^-$. Therefore Lemma \ref{lem:classify-neg-slice} shows that $(\bar{\partial}_A, \phi)$ admits a filtration whose graded object is isomorphic to $(Q_1, \phi_1) \oplus \cdots \oplus (Q_n, \phi_n)$. \end{proof}
\section{The Hecke correspondence via Yang-Mills-Higgs flow lines}\label{sec:hecke}
Let $(E, \phi)$ be a polystable Higgs bundle of rank $r$ and degree $d$, and let $(L_u, \phi_u)$ be a Higgs line bundle with $\deg L_u < \slope E$. Let $F$ be a smooth complex vector bundle $C^\infty$ isomorphic to $E \oplus L_u$ and choose a metric on $F$ such that the Higgs structure on $(E, \phi) \oplus (L_u,\phi_u)$ is a Yang-Mills-Higgs critical point in the space $\mathcal{B}(F)$ of Higgs bundles on $F$. The goal of this section is to show that Hecke modifications of the Higgs bundle $(E, \phi)$ correspond to Yang-Mills-Higgs flow lines in $\mathcal{B}(F)$ connecting the critical point $(E, \phi) \oplus (L_u, \phi_u)$ to lower critical points.
In Section \ref{sec:Higgs-hecke-review} we review Hecke modifications of Higgs bundles. Section \ref{sec:canonical-map} describes how the space of Hecke modifications relates to the geometry of the negative slice and Section \ref{sec:YMH-flow-hecke} contains the proof of Theorem \ref{thm:flow-hecke} which shows that Hecke modifications correspond to $\mathop{\rm YMH}\nolimits$ flow lines. In Section \ref{sec:secant-criterion} we give a geometric criterion for points to be connected by unbroken flow lines in terms of the secant varieties of the space of Hecke modifications inside the negative slice. In particular, this gives a complete classification of the $\mathop{\rm YMH}\nolimits$ flow lines for rank $2$ (cf. Corollary \ref{cor:rank-2-classification}). Throughout this section the notation $\mathcal{E}$ is used to denote the sheaf of holomorphic sections of the bundle $E$.
\subsection{Hecke modifications of Higgs bundles}\label{sec:Higgs-hecke-review}
The purpose of this section is to derive some basic results for Hecke modifications of Higgs bundles which will be used in Section \ref{sec:YMH-flow-hecke} to prove Theorem \ref{thm:flow-hecke}. In Section \ref{sec:secant-criterion} we extend these results to study unbroken YMH flow lines.
First recall that a Hecke modification of a holomorphic bundle $E$ over a Riemann surface $X$ is determined by points $p_1, \ldots, p_n \in X$ (not necessarily distinct) and nonzero elements $v_j \in E_{p_j}^*$ for $j = 1, \ldots, n$. This data determines a sheaf homomorphism $\mathcal{E} \rightarrow \oplus_{j=1}^n \mathbb C_{p_j}$ to the skyscraper sheaf supported at $p_1, \ldots, p_n$ with kernel a locally free sheaf $\mathcal{E}'$. This determines a holomorphic bundle $E' \rightarrow X$ which we call the \emph{Hecke modification of $E$ determined by $v = (v_1, \ldots, v_n)$}. \begin{equation*} 0 \rightarrow \mathcal{E}' \rightarrow \mathcal{E} \stackrel{v}{\longrightarrow} \bigoplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0 \end{equation*} Since the kernel sheaf $\mathcal{E}'$ only depends on the equivalence class of each $v_j$ in $\mathbb{P} E_{p_j}^*$ then from now on we abuse the notation slightly and also use $v_j \in \mathbb{P} E_{p_j}^*$ to denote the equivalence class of $v_j \in E_{p_j}^*$.
As explained in \cite[Sec. 4.5]{witten-hecke}, if $(E, \phi)$ is a Higgs bundle, then a Hecke modification of $(E, \phi)$ may introduce poles into the Higgs field and so there are restrictions on the allowable modifications which preserve holomorphicity of the Higgs field.
\begin{definition} Let $(E, \phi)$ be a Higgs bundle. A Hecke modification $E'$ of $E$ is \emph{compatible} with $\phi$ if the induced Higgs field on $E'$ is holomorphic. \end{definition}
The next result describes a basic condition for the modification to be compatible with the Higgs field.
\begin{lemma} Let $(E, \phi)$ be a Higgs bundle, and $0 \rightarrow \mathcal{E}' \rightarrow \mathcal{E} \stackrel{v}{\longrightarrow} \mathbb{C}_p \rightarrow 0$ a Hecke modification of $E$ induced by $v \in E_p^*$. Then the induced Higgs field $\phi'$ on $E'$ is holomorphic if and only if there exists an eigenvalue $\mu$ of $\phi(p)$ such that the composition $\mathcal{E} \otimes K^{-1} \stackrel{\phi - \mu \cdot \id}{\ensuremath{\relbar\mathrel{\mkern-4mu}\relbar\mathrel{\mkern-4mu}\longrightarrow}} \mathcal{E} \stackrel{v}{\rightarrow} \mathbb C_p$ is zero. \end{lemma}
\begin{proof} Let $\phi \in H^0(\mathop{\rm End}\nolimits(E) \otimes K)$. Then $\phi$ pulls back to a holomorphic Higgs field $\phi' \in H^0(\mathop{\rm End}\nolimits(E') \otimes K)$ if and only if for any open set $U \subset X$ and any section $s \in \mathcal{E}(U)$, the condition $s \in \ker (\mathcal{E}(U) \stackrel{v}{\rightarrow} \mathbb C_p(U))$ implies that $\phi(s) \in \ker ((\mathcal{E} \otimes K)(U) \stackrel{v}{\rightarrow} \mathbb C_p(U))$. After choosing a trivialisation of $K$ in a neighbourhood of $p$, we can decompose the Higgs field $\phi(p)$ on the fibre $E_p$ as follows \begin{equation}\label{eqn:fibre-extension} \xymatrix{
0 \ar[r] & \ker v \ar[r] \ar[d]^{\left. \phi(p) \right|_{\ker v}} & E_p \ar[r] \ar[d]^{\phi(p)} & \mathbb{C}_p \ar[r] \ar[d]^\mu & 0 \\ 0 \ar[r] & \ker v \ar[r] & E_p \ar[r] & \mathbb{C}_p \ar[r] & 0 } \end{equation} where scalar multiplication by $\mu$ is induced from the action of $\phi(p)$ on the quotient $\mathbb{C}_p = E_p / \ker v$. Therefore the endomorphism $\left( \phi(p) - \mu \cdot \id \right)$ maps $E_p$ into the subspace $\ker v$ and so $v \in E_p^*$ descends to a well-defined homomorphism $v' : \mathop{\rm coker}\nolimits \left( \phi(p) - \mu \cdot \id \right) \rightarrow \mathbb C$.
Conversely, given an eigenvalue $\mu$ of $\phi(p)$ and an element $v' \in \mathop{\rm coker}\nolimits(\phi(p) - \mu \cdot \id)^*$, one can choose a basis of $E_p$ and extend $v'$ to an element $v \in E_p^*$ such that $\im (\phi(p) - \mu \cdot \id) \subset \ker v$. Equivalently, $\phi(p)$ preserves $\ker v$ and so $v \in E_p^*$ defines a Hecke modification $E'$ of $E$ such that the induced Higgs field on $E'$ is holomorphic. \end{proof}
\begin{corollary}\label{cor:Higgs-compatible} Let $(E, \phi)$ be a Higgs bundle and let $0 \rightarrow \mathcal{E}' \rightarrow \mathcal{E} \stackrel{v}{\rightarrow} \mathbb C_p \rightarrow 0$ be a Hecke modification of $E$ induced by $v \in \mathbb{P} E_p^*$. The following conditions are equivalent \begin{enumerate} \item The induced Higgs field $\phi'$ on $E'$ is holomorphic.
\item There exists an eigenvalue $\mu$ of $\phi(p)$ such that $v(\phi(s)) = \mu v(s)$ for all sections $s$ of $E$.
\item There exists an eigenvalue $\mu$ of $\phi(p)$ such that $v$ descends to a well-defined $v' \in (\mathop{\rm coker}\nolimits (\phi(p) - \mu \cdot \id))^*$.
\end{enumerate} \end{corollary}
\begin{lemma}\label{lem:resolve-higgs-subsheaf} Let $(E, \phi)$ be a Higgs bundle and $(G, \varphi)$ a Higgs subsheaf. Then there exists a Higgs subbundle $(G', \varphi') \subset (E, \phi)$ such that $\rank(G) = \rank (G')$ and $(G, \varphi)$ is a Higgs subsheaf of $(G', \varphi')$. \end{lemma}
\begin{proof} Since $\dim_\mathbb C X = 1$ then a standard procedure shows that there is a holomorphic subbundle $G' \subset E$ with $\rank(G) = \rank (G')$ and $G$ is a subsheaf of $G'$, and so it only remains to show that this is a Higgs subbundle. The reverse of the construction above shows that the Higgs field $\varphi$ preserving $G$ extends to a meromorphic Higgs field $\varphi'$ preserving $G'$, and since this is the restriction of a holomorphic Higgs field $\phi$ on $E$ to the holomorphic subbundle $G'$, then $\varphi'$ must be holomorphic on $G'$. Therefore $G'$ is $\phi$-invariant. \end{proof}
\begin{definition}\label{def:m-n-stable} A Higgs bundle $(E, \phi)$ is \emph{$(m,n)$-stable (resp. $(m,n)$-semistable)} if for every proper $\phi$-invariant holomorphic subbundle $F \subset E$ we have \begin{equation*} \frac{\deg F + m}{\rank F} < \frac{\deg E - n}{\rank E} \quad \text{(resp. $\leq$)} . \end{equation*} \end{definition}
If $(E, \phi)$ is $(0,n)$-semistable then any Hecke modification $0 \rightarrow (\mathcal{E}', \phi') \rightarrow (\mathcal{E}, \phi) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$ is semistable.
\begin{definition} Then \emph{space of admissible Hecke modifications} is the subset $\mathcal{N}_{\phi} \subset \mathbb{P} E^*$ corresponding to the Hecke modifications which are compatible with the Higgs field. \end{definition}
\begin{remark}\label{rem:hecke-explicit} \begin{enumerate}
\item If $\phi = 0$ then $\mathcal{N}_{0} = \mathbb{P} E^*$. If $E$ is $(0,1)$-stable then there is a well-defined map $\mathbb{P} E^* \rightarrow \mathbb{P} H^1(E^*)$. The construction of the next section generalises this to a map $\mathcal{N}_{\phi} \rightarrow \mathbb{P} \mathcal{H}^1(E^*)$ (cf. Remark \ref{rem:canonical-map}).
\item Note that the construction above is the reverse of that described in \cite{witten-hecke}, which begins with $E'$ and modifies the bundle to produce a bundle $E$ with $\deg E = \deg E' + 1$. Here we begin with $E$ and construct $E'$ via a modification $0 \rightarrow \mathcal{E}' \rightarrow \mathcal{E} \rightarrow \mathbb C_p \rightarrow 0$ since we want to interpret the compatible modifications in terms of the geometry of the negative slice (see Section \ref{sec:canonical-map}) in order to draw a connection with the results on gradient flow lines for the Yang-Mills-Higgs flow functional from Section \ref{sec:filtration-criterion}.
\item One can also see the above construction more explicitly in local coordinates as in \cite{witten-hecke} by choosing a local frame $\{ s_1, \ldots, s_n \}$ for $E$ in a neighbourhood $U$ of $p$ with local coordinate $z$ centred at $p$ and for which the evaluation map $\mathcal{E} \stackrel{v}{\rightarrow} \mathbb C_p$ satisfies $v(s_1) = s_1(0)$ and $v(s_j) = 0$ for all $j=2, \ldots, n$. Then over $U \setminus \{ p \}$, the functions $\{ \frac{1}{z} s_1(z), s_2(z) \ldots, s_n(z) \}$ form a local frame for $E'$. Equivalently, the transition function $g = \left( \begin{matrix} \frac{1}{z} & 0 \\ 0 & \id \end{matrix} \right)$ maps the trivialisation for $E$ to a trivialisation for $E'$ (note that this is the inverse of the transition function from \cite[Sec. 4.5.2]{witten-hecke} for the reason explained in the previous paragraph). In this local frame field on $E$ we write $\phi(z) = \left( \begin{matrix} A(z) & B(z) \\ C(z) & D(z) \end{matrix} \right)$. The action on the Higgs field is then \begin{equation*} g \left( \begin{matrix} A(z) & B(z) \\ C(z) & D(z) \end{matrix} \right) g^{-1} = \left( \begin{matrix} A(z) & \frac{1}{z} B(z) \\ z C(z) & D(z) \end{matrix} \right) \end{equation*} Therefore the induced Higgs field on $E'$ will have a pole at $p$ unless $B(0) = 0$. The scalar $A(0)$ in this local picture is the same as the scalar $\mu$ from \eqref{eqn:fibre-extension}, and we see that \begin{equation*} \phi(p) - \mu \cdot \id = \left( \begin{matrix} 0 & 0 \\ C(0) & D(0) - \mu \cdot \id \end{matrix} \right) \end{equation*} With respect to the basis of $E_p$ given by the choice of local frame, $v(\phi(p) - \mu \cdot \id) = 0$. Moreover, via this local frame $\mathop{\rm coker}\nolimits ( \phi(p) - \mu \cdot \id)$ is identified with a subspace of $E_p$ which contains the linear span of $s_1(0)$. Therefore we see in the local coordinate picture that $v \in E_p^*$ descends to an element of $(\mathop{\rm coker}\nolimits (\phi(p) - \mu \cdot \id))^*$. \end{enumerate} \end{remark}
The next result shows that the admissible Hecke modifications have an interpretation in terms of the spectral curve associated to the Higgs field. This extends the results of \cite{witten-hecke} to include the possibility that $p$ is a branch point of the spectral cover.
First recall Hitchin's construction of the spectral curve from \cite{Hitchin87-2}. Let $(E, \phi)$ be a Higgs pair. Then there is a projection map $\pi : K \rightarrow X$ and a bundle $\pi^* E$ over the total space of the canonical bundle together with a tautological section $\lambda$ of $\pi^* E$. The zero set of the characteristic polynomial of $\pi^* \phi$ defines a subvariety $S$ inside the total space of $K$. The projection $\pi$ restricts to a map $\pi : S \rightarrow X$, where for each $p \in X$ the fibre $\pi^{-1}(p)$ consists of the eigenvalues of the Higgs field $\phi(p)$. As explained in \cite{Hitchin87-2}, generically the discriminant of the Higgs field has simple zeros and in this case $S$ is a smooth curve called the \emph{spectral curve}. The induced projection $\pi : S \rightarrow X$ is then a ramified covering map with ramification divisor denoted $\mathcal{R} \subset S$.
The pullback of the Higgs field to the spectral curve is a bundle homomorphism $\pi^* E \rightarrow \pi^*(E \otimes K)$, and the eigenspaces correspond to $\ker (\pi^* \phi - \lambda \cdot \id)$, where $\lambda$ is the tautological section defined above. When the discriminant of the Higgs field has simple zeros then Hitchin shows in \cite{Hitchin87-2} that the eigenspaces form a line bundle $\mathcal{N} \rightarrow S$ and that the original bundle $E$ can be reconstructed as $\pi_* \mathcal{L}$, where the line bundle $\mathcal{L} \rightarrow S$ is formed by modifying $\mathcal{N}$ at the ramification points $0 \rightarrow \mathcal{N} \rightarrow \mathcal{L} \rightarrow \bigoplus_{p \in \mathcal{R}} \mathbb C_p \rightarrow 0$. One can reconstruct the Higgs field $\phi$ by pushing forward the endomorphism defined by the tautological section $\lambda : \mathcal{L} \rightarrow \mathcal{L} \otimes \pi^* K$.
\begin{lemma} If the discriminant of $\phi$ has simple zeros then an admissible Hecke modification of $(E, \phi)$ corresponds to a Hecke modification of the line bundle $\mathcal{L}$ over the spectral curve. \end{lemma}
\begin{proof} Consider the pullback bundle $\pi^* E \rightarrow S$. The pullback of the Higgs field induces a sheaf homomorphism $(\pi^* \phi - \lambda \cdot \id) : \pi^* \mathcal{E} \otimes (\pi^* K)^{-1} \rightarrow \pi^* \mathcal{E}$. As explained in \cite[Sec. 2.6]{witten-hecke}, when the discriminant of $\phi$ has simple zeros then the cokernel of this homomorphism is the line bundle $\mathcal{L} \rightarrow S$ such that $\mathcal{E} \cong \pi_* \mathcal{L}$.
For $\mu \in S$ such that $p = \pi(\mu)$, there is an isomorphism of the stalks of the skyscraper sheaves $\mathbb C_p \cong \pi_* (\mathbb C_\mu)$. Then a Hecke modification $\mathcal{L} \stackrel{v'}{\rightarrow} \mathbb C_\mu$ given by nonzero $v' \in \mathcal{L}_\mu^*$ induces a Hecke modification $v = v' \circ q \circ \pi^* : \mathcal{E} \rightarrow \mathbb C_p$, defined by the commutative diagram below. \begin{equation*} \xymatrix{ \pi^* \mathcal{E} \otimes (\pi^* K)^{-1} \ar[rr]^(0.6){\pi^*\phi - \lambda \cdot \id} & & \pi^* \mathcal{E} \ar[r]^(0.3)q & \mathop{\rm coker}\nolimits(\pi^* \phi - \lambda \cdot \id) \ar[dr]^(0.6){v'} \ar[r] & 0 \\ & & \mathcal{E} \ar[u]^{\pi^*} \ar[rr]^v & & \mathbb C_p \ar[r] & 0 } \end{equation*} The definition of $v$ implies that for any open set $U \subset X$ with a trivialisation of $K$ in a neighbourhood of $p$, and all $s \in \mathcal{E}(U)$ we have \begin{equation*} v(\phi s) = v' \circ q(\pi^* (\phi s)) = v' \circ q(\mu \, \pi^* (s)) = \mu \, v' \circ q \circ \pi^* (s) = \mu \, v(s) \end{equation*} and so $v$ is compatible with the Higgs field by Corollary \ref{cor:Higgs-compatible}.
Conversely, let $v \in E_p^*$ be compatible with the Higgs field $\phi$. Corollary \ref{cor:Higgs-compatible} shows that this induces a well-defined element of $\mathop{\rm coker}\nolimits(\phi - \mu \cdot \id)^*$. Consider the endomorphisms $\phi(p) - \mu \cdot \id$ on the fibre of $E$ over $p \in X$ and $\pi^* \phi(\mu) - \mu \cdot \id$ on the fibre of $\pi^* E$ over $\mu \in S$.
\begin{equation*} \xymatrix{ (\pi^* E \otimes \pi^* K^{-1})_\mu \ar[rr]^(0.6){\pi^* \phi - \mu \cdot \id} & & (\pi^* E)_\mu \ar[r] & \mathop{\rm coker}\nolimits (\pi^* \phi - \mu \cdot \id)_\mu \ar[r] & 0\\ (E \otimes K^{-1})_p \ar[rr]^(0.6){\phi - \mu \cdot \id} \ar[u] & & E_p \ar[r] \ar[u] & \mathop{\rm coker}\nolimits(\phi - \mu \cdot \id)_p \ar[r] \ar@{-->}[u] & 0 } \end{equation*}
The universal property of cokernel defines a map $\mathop{\rm coker}\nolimits(\phi - \mu \cdot \id)_p \rightarrow \mathop{\rm coker}\nolimits(\pi^* \phi - \mu \cdot \id)_\mu$. Since the discriminant of the Higgs field has simple zeros then both fibres are one-dimensional and so this map becomes an isomorphism. Therefore $v$ induces a well-defined homomorphism on the fibre $\mathop{\rm coker}\nolimits(\pi^* \phi - \mu \cdot \id)_\mu \rightarrow \mathbb C$, and hence a Hecke modification of $\mathcal{L}$ at $\mu \in S$. \end{proof}
\begin{remark} When $p \in X$ is not a branch point of $\pi : S \rightarrow X$ then this result is contained in \cite{witten-hecke}. \end{remark}
\begin{corollary} If the discriminant of $\phi$ has simple zeros then the space of Hecke modifications is $\mathcal{N}_{\phi} = S$. \end{corollary}
\subsection{Secant varieties associated to the space of Hecke modifications}\label{sec:canonical-map}
The purpose of this section is to connect the geometry of the space of Hecke modifications with the geometry of the negative slice at a critical point in order to prepare for the proof of Theorem \ref{thm:flow-hecke} in the next section.
Let $(E_1, \phi_1)$ and $(E_2, \phi_2)$ be Higgs bundles and let $\bar{\partial}_A$ denote the induced holomorphic structure on $E_1^* E_2$. Then there is an elliptic complex \begin{equation*} \Omega^0(E_1^* E_2) \stackrel{L_1}{\longrightarrow} \Omega^{0,1}(E_1^* E_2) \oplus \Omega^{1,0}(E_1^* E_2) \stackrel{L_2}{\longrightarrow} \Omega^{1,1}(E_1^* E_2) , \end{equation*} where $L_1(u) = (\bar{\partial}_A u, \phi_2 u - u \phi_1)$ and $L_2(a, \varphi) = (\bar{\partial}_A \varphi + [a, \phi])$. Let $\mathcal{H}^0 = \ker L_1$, $\mathcal{H}^1 = \ker L_1^* \cap \ker L_2$ and $\mathcal{H}^2 = \ker L_2^*$ denote the spaces of harmonic forms. Recall that if $(E_1, \phi_1)$ and $(E_2, \phi_2)$ are both Higgs stable and $\slope(E_2) < \slope(E_1)$ then $\mathcal{H}^0(E_1^* E_2) = 0$.
Now consider the special case where $(E_1, \phi_1)$ is $(0, n)$-stable and $(E_2, \phi_2)$ is a Higgs line bundle. Let $\mathcal{B}$ denote the space of Higgs bundles on the smooth bundle $E_1 \oplus E_2$ and choose a metric such that $(E_1, \phi_1) \oplus (E_2, \phi_2)$ is a critical point of $\mathop{\rm YMH}\nolimits : \mathcal{B} \rightarrow \mathbb R$. Definition \ref{def:slice} shows that $\mathcal{H}^1(E_1^* E_2) \cong S_x^-$ is the negative slice at this critical point.
Let $0 \rightarrow (\mathcal{E}', \phi') \rightarrow (\mathcal{E}_1, \phi_1) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$ be a Hecke modification defined by $v_1, \ldots, v_n \in \mathbb{P} E_1^*$. Applying the functor $\mathop{\rm Hom}\nolimits(\cdot, \mathcal{E}_2)$ to the short exact sequence $0 \rightarrow \mathcal{E}' \rightarrow \mathcal{E}_1 \rightarrow \oplus_j \mathbb C_{p_j} \rightarrow 0$ gives us an exact sequence of sheaves $0 \rightarrow \mathop{\rm Hom}\nolimits(\mathcal{E}_1, \mathcal{E}_2) \rightarrow \mathop{\rm Hom}\nolimits(\mathcal{E}', \mathcal{E}_2) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j}^* \rightarrow 0$, where the final term comes from the isomorphism $\mathop{\rm Ext}\nolimits^1(\oplus_j \mathbb C_{p_j}, \mathcal{E}_2) \cong \mathop{\rm Hom}\nolimits(\mathcal{E}_2, \oplus_j \mathbb C_{p_j} \otimes K)^* \cong \oplus_j \mathbb C_{p_j}^*$. Note that this depends on a choice of trivialisations of $E_2$ and $K$, however the kernel of the map $\mathop{\rm Hom}\nolimits(\mathcal{E}', \mathcal{E}_2) \rightarrow \oplus_j \mathbb C_{p_j}$ is independent of these choices. This gives us the following short exact sequence of Higgs sheaves \begin{equation}\label{eqn:dual-short-exact} 0 \rightarrow \mathcal{E}_1^* \mathcal{E}_2 \rightarrow (\mathcal{E}')^* \mathcal{E}_2 \rightarrow \bigoplus_{j=1}^n \mathbb C_{p_j}^* \rightarrow 0 \end{equation}
There is an induced map $\Omega^0((E')^* E_2) \rightarrow \Omega^{1,0}((E')^* E_2)$ given by $s \mapsto \phi_2 s - s \phi'$. Recall from Corollary \ref{cor:Higgs-compatible} that there exists an eigenvalue $\mu_j$ for $\phi_1(p_j)$ such that $v(\phi_1(p_j) - \mu_j \cdot \id) = 0$ for each $j=1, \ldots, n$. From the above exact sequence there is an induced homomorphism $\Omega^{1,0}((E')^* E_2) \stackrel{ev^1}{\longrightarrow} \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$. The component of $ev^1(\phi_2 s - s \phi')$ in $\mathbb C_{p_j}$ is $(\phi_2(p_j) - \mu_j ) s$. In particular, $\phi_2 s - s \phi' \in \ker(ev^1)$ iff $\phi_2(p_j) = \mu_j$ for all $j=1, \ldots, n$.
\begin{definition}\label{def:Hecke-compatible} Let $(E_1, \phi_1)$ be a Higgs bundle, and $(E_2, \phi_2)$ a Higgs line bundle. The \emph{space of Hecke modifications compatible with $\phi_1$ and $\phi_2$}, denoted $\mathcal{N}_{\phi_1, \phi_2} \subset \mathcal{N}_{\phi_1}$, is the set of Hecke modifications compatible with $\phi_1$ such that $ev^1(\phi_2 s - s \phi') = 0$ for all $s \in \Omega^0((E')^* E_2)$. \end{definition}
\begin{remark}\label{rem:miniscule-compatible} Note that if $n = 1$ and $v \in \mathbb{P} E_1^*$ is a Hecke modification compatible with $\phi_1$, then the requirement that $v \in \mathcal{N}_{\phi_1, \phi_2}$ reduces to $\phi_2(p) = \mu$, where $\mu$ is the eigenvalue of $\phi_1(p)$ from Corollary \ref{cor:Higgs-compatible}. Such a $\phi_2 \in H^0(\mathop{\rm End}\nolimits(E_2) \otimes K) = H^0(K)$ always exists since the canonical linear system is basepoint free and therefore $\bigcup_{\phi_2 \in H^0(K)} \mathcal{N}_{\phi_1, \phi_2} = \mathcal{N}_{\phi_1}$. If $n > 1$ then $\phi_2$ with these properties may not exist for some choices of $\phi_1 \in H^0(\mathop{\rm End}\nolimits(E_1) \otimes K)$ and $v_1, \ldots, v_n \in \mathbb{P} E_1^*$ (the existence of $\phi_2$ depends on the complex structure of the surface $X$). If $\phi_1 = 0$, then we can choose $\phi_2 = 0$ and in this case $\mathcal{N}_{\phi_1, \phi_2} = \mathcal{N}_{\phi_1} = \mathbb{P} E_1^*$ (this corresponds to the case of the Yang-Mills flow in Theorem \ref{thm:flow-hecke}). \end{remark}
\begin{lemma} Let $(E_1, \phi_1)$ be Higgs polystable and $(E_2, \phi_2)$ be a Higgs line bundle. Let $0 \rightarrow (\mathcal{E}', \phi') \rightarrow (\mathcal{E}, \phi) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$ be a Hecke modification defined by distinct $v_1, \ldots, v_n \in \mathcal{N}_{\phi_1, \phi_2}$.
Then there is an exact sequence \begin{equation}\label{eqn:hyper-exact-sequence} 0 \rightarrow \mathcal{H}^0(E_1^* E_2) \rightarrow \mathcal{H}^0((E')^* E_2) \rightarrow \mathbb C^n \rightarrow \mathcal{H}^1(E_1^* E_2) \rightarrow \mathcal{H}^1((E')^* E_2) \end{equation} \end{lemma}
\begin{proof}
The short exact sequence \eqref{eqn:dual-short-exact} leads to the following commutative diagram of spaces of smooth sections \begin{equation*} \xymatrix@C=1.6em{ 0 \ar[r] & \Omega^0(E_1^* E_2) \ar[r]^{i^*} \ar[d]^{L_1} & \Omega^0((E')^* E_2) \ar[r]^{ev^0} \ar[d]^{L_1} & \bigoplus_{j=1}^n \mathbb C_{p_j} \ar[r] & 0 \\ 0 \ar[r] & \Omega^{0,1}(E_1^* E_2) \oplus \Omega^{1,0}(E_1^* E_2) \ar[r]^(0.47){i^*} & \Omega^{0,1}((E')^* E_2) \oplus \Omega^{1,0}((E')^* E_2) \ar[r]^(0.62){ev^1} & \bigoplus_{j=1}^n \mathbb C_{p_j} \oplus \mathbb C_{p_j} \ar[r] & 0 } \end{equation*} Since $\bar{\partial}_A s$ depends on the germ of a section around a point, then there is no well-defined map $\bigoplus_{j=1}^n \mathbb C_{p_j} \rightarrow \bigoplus_{j=1}^n \mathbb C_{p_j} \oplus \mathbb C_{p_j}$ making the diagram commute, so the exact sequence \eqref{eqn:hyper-exact-sequence} does not follow immediately from the standard construction, and therefore we give an explicit construction below.
First construct a map $\mathbb C^n \rightarrow \mathcal{H}^1(E_1^* E_2)$ as follows. Given $z \in \mathbb C^n$, choose a smooth section $s' \in \Omega^0((E')^* E_2)$ such that $ev^0(s') = z$ and $ev^1(\bar{\partial}_A s') = 0$. Since $\phi_2(p_j) = \mu_j$, then $ev^1(\phi_2 s' - s' \phi') = 0$ and so $ev^1(L_1 s') = 0$. Therefore $(\bar{\partial}_A s', \phi_2 s' - s' \phi') = i^*(a, \varphi)$ for some $(a, \varphi) \in \Omega^{0,1}(E_1^* E_2) \oplus \Omega^{1,0}(E_1^* E_2)$. Let $[(a, \varphi)] \in \mathcal{H}^1(E_1^* E_2)$ denote the harmonic representative of $(a, \varphi)$. Define the map $\mathbb C^n \rightarrow \mathcal{H}^1(E_1^* E_2)$ by $z \mapsto [(a, \varphi)]$.
To see that this is well-defined independent of the choice of $s' \in \Omega^0((E')^* E_2)$, note that if $s'' \in \Omega^0((E')^* E_2)$ is another section such that $ev^0(s'') = z$ and $ev^1(\bar{\partial}_A s'') = 0$, then $ev^0(s'' - s') = 0$, and so $s'' - s' = i^*(s)$ for some $s \in \Omega^0(E_1^* E_2)$. Therefore $L_1(s'' - s') = i^* L_1(s)$ with $[L_1(s)] = 0 \in \mathcal{H}^1(E_1^* E_2)$, and so $s'$ and $s''$ determine the same harmonic representative in $\mathcal{H}^1(E_1^* E_2)$.
To check exactness of \eqref{eqn:hyper-exact-sequence} at the term $\mathbb C^n$, note that if $z = ev^0(s')$ for some harmonic $s' \in \mathcal{H}^0((E')^* E_2)$, then $L_1(s') = 0 = i^* (0,0)$, and so $z \in \mathbb C^n$ maps to $0 \in \mathcal{H}^1(E_1^* E_2)$. Moreover, if $z$ maps to $0 \in \mathcal{H}^1(E_1^*E_2)$, then there exists $s' \in \Omega^0((E')^* E_2)$ such that $L_1(s') = i^*(a, \varphi)$ where $(a, \varphi) \in \Omega^{0,1}(E_1^* E_2) \oplus \Omega^{1,0}((E')^* E_2)$ and $(a, \varphi) = L_1(s)$ for some $s \in \Omega^0(E_1^* E_2)$. Therefore $s'$ and $i^* s$ differ by a harmonic section of $\mathcal{H}^0((E')^* E_2)$. Since $ev^0(i^* s) = 0$ then $z$ is the image of this harmonic section under the map $\mathcal{H}^0((E')^* E_2) \rightarrow \mathbb C^n$.
To check exactness at $\mathcal{H}^1(E_1^* E_2)$, given $z \in \mathbb C^n$ construct $(a, \varphi)$ as above and note that $i^*(a, \varphi) = L_1 s'$ for some $s' \in \Omega^0((E')^* E_2)$. Therefore $i^*[(a, \varphi)] = 0 \in \mathcal{H}^1((E')^* E_2)$ and so the image of $\mathbb C^n \rightarrow \mathcal{H}^1(E_1^* E_2)$ is contained in the kernel of $\mathcal{H}^1(E_1^* E_2) \rightarrow \mathcal{H}^1((E')^* E_2)$. Now suppose that the image of $[(a, \varphi)]$ is zero in $\mathcal{H}^1((E')^* E_2)$, i.e. $i^*(a, \varphi) = L_1 s'$ for some $s' \in \Omega^0((E')^* E_2)$. Let $z = ev^0(s')$. Note that $z = 0$ implies that $s' = i^* s$ for some $s \in \Omega^0(E_1^* E_2)$, and so $[(a, \varphi)] = 0$. If $z \neq 0$ then there exists $s'' \in \Omega^0((E')^* E_2)$ such that $ev^1(L_1(s'')) = 0$ and $ev^0(s'') = z$. Then $L_1(s'') = i^*(a'', \varphi'')$ for some $(a'', \varphi'') \in \Omega^{0,1}(E_1^* E_2) \oplus \Omega^{1,0}(E_1^* E_2)$. Moreover, $ev^0(s'' - s') = 0$, so $s'' - s' = i^* s$ for some $s \in \Omega^0(E_1^* E_2)$. Commutativity implies that $L_1 s = (a'', \varphi'') - (a, \varphi)$, and so the harmonic representatives $[(a, \varphi)]$ and $[(a'', \varphi'')]$ are equal. Therefore $[(a, \varphi)]$ is the image of $z$ by the map $\mathbb C^n \rightarrow \mathcal{H}^1(E_1^* E_2)$, which completes the proof of exactness at $\mathcal{H}^1(E_1^* E_2)$.
Exactness at the rest of the terms in the sequence \eqref{eqn:hyper-exact-sequence} then follows from standard methods. \end{proof}
For any stable Higgs bundle $(E, \phi)$ with $d = \deg E$ and $r = \rank E$, define the \emph{generalised Segre invariant} by \begin{equation*} s_k(E, \phi) := k d - r \left( \max_{F \subset E, \rank F = k} \deg F \right) . \end{equation*} where the maximum is taken over all $\phi$-invariant holomorphic subbundles of rank $k$. Note that $s_k(E, \phi) \geq s_k(E, 0) =: s_k(E)$ and \begin{equation*} \frac{1}{rk} s_k(E, \phi) = \min_{F \subset E, \rank F = k} \left( \slope(E) - \slope(F) \right) \end{equation*} Note that any Hecke modification $(E', \phi') \hookrightarrow (E, \phi)$ with $\deg E - \deg E' = n$ has Segre invariant $s_k(E', \phi') \geq s_k(E, \phi) - nk$. As a special case, $(E', \phi')$ is stable if $n < \frac{1}{k} s_k(E, \phi)$ for all $k = 1, \ldots, r-1$.
A theorem of Lange \cite[Satz 2.2]{Lange83} shows that a general stable holomorphic bundle $E$ satisfies $s_k(E) \geq k(r - k)(g-1)$ for all $k = 1, \ldots, r - 1$. Since there is an dense open subset of stable Higgs bundles whose underlying holomorphic bundle is stable, then Lange's theorem also gives the same lower bound on the Segre invariant for a general stable Higgs bundle.
\begin{lemma}\label{lem:segre-bound} Let $0 \rightarrow (E', \phi') \rightarrow (E, \phi) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$ be a Hecke modification defined by distinct points $v_1, \ldots, v_n \in \mathbb{P} E^*$ such that $n < \frac{1}{k} s_k(E, \phi)$ for all $k = 1, \ldots, r-1$. Then $\slope(G) < \slope (E')$ for any proper non-zero Higgs subbundle $(G, \phi_G) \subset (E, \phi)$. In particular, this condition is satisfied if $(E, \phi)$ is a general stable Higgs bundle and $n < g-1$. \end{lemma}
\begin{proof} Let $k = \rank G$ and $h = \deg G$. Then the lower bound on the Segre invariant implies that \begin{align*} \slope(E') - \slope(G) = \frac{d - n}{r} - \frac{h}{k} & = \frac{1}{rk} \left(kd - kn - rh \right) \\
& \geq \frac{1}{rk} \left( s_k(E, \phi) - kn \right) \end{align*} Therefore if $n < \frac{1}{k} s_k(E, \phi)$ then $\slope(E') - \slope(G) > 0$ for any Higgs subbundle of rank $k$. If $n < g-1$ then \cite[Satz 2.2]{Lange83} shows that this condition is satisfied for general stable Higgs bundles. \end{proof}
\begin{corollary}\label{cor:n-dim-kernel} Let $(E_1, \phi_1)$ be a stable Higgs bundle, let $n < \frac{1}{k} s_k(E_1, \phi_1)$ for all $k=1, \ldots, \rank(E_1)-1$ and let $(E_2, \phi_2)$ be a Higgs line bundle such that $\deg E_2 < \frac{\deg E_1 - n}{\rank E_1}$. Then given any set of $n$ distinct points $\{ v_1, \ldots, v_n \} \subset \mathcal{N}_{\phi_1, \phi_2}$ there is a well-defined $n$-dimensional subspace $\ker (\mathcal{H}^1(E_1^* E_2) \rightarrow \mathcal{H}^1((E')^* E_2))$. \end{corollary}
\begin{proof} Let $(E', \phi')$ be the Hecke modification of $(E_1, \phi_1)$ determined by $\{ v_1, \ldots v_n \} \subset \mathbb{P} E_1^*$. The lower bound on the Segre invariant implies that $(E', \phi')$ is Higgs stable, and therefore $\mathcal{H}^0((E')^* E_2) = 0$ since $\slope(E_2) < \slope(E') = \frac{\deg E_1 - n}{\rank E_1}$. The exact sequence \eqref{eqn:hyper-exact-sequence} then reduces to \begin{equation*} 0 \rightarrow \mathbb C^n \rightarrow \mathcal{H}^1(E_1^* E_2) \rightarrow \mathcal{H}^1((E')^* E_2) \end{equation*} and so $\ker (\mathcal{H}^1(E_1^* E_2) \rightarrow \mathcal{H}^1((E')^* E_2))$ is a well-defined $n$-dimensional subspace of $\mathcal{H}^1(E_1^* E_2)$ associated to $\{ v_1, \ldots, v_n \}$. \end{proof}
\begin{remark}\label{rem:canonical-map} As noted above, the maps $\mathbb C^n \rightarrow \mathcal{H}^1(E_1^* E_2)$ depend on choosing trivialisations, but different choices lead to the same map up to a change of basis of $\mathbb C^n$, and so the subspace $\ker (\mathcal{H}^1((E')^* E_2) \rightarrow \mathcal{H}^1(E_1^* E_2))$ is independent of these choices.
In the special case where $n=1$, then this construction gives a well-defined map $\mathcal{N}_{\phi_1, \phi_2} \rightarrow \mathbb{P} \mathcal{H}^1(E_1^* E_2)$. When $n < \frac{1}{k} s_k(E_1, \phi_1)$ for all $k$, then Corollary \ref{cor:n-dim-kernel} shows that any $n$ distinct points $v_1, \ldots, v_n$ span a nondegenerate copy of $\mathbb{P}^{n-1}$ in $\mathbb{P} \mathcal{H}^1(E_1^* E_2)$.
In the special case where $\phi_1 = \phi_2 = 0$ and $E_2$ is trivial, then $\mathcal{N}_{\phi_1, \phi_2} = \mathbb{P} E^*$ and $\mathcal{H}^1(E_1^*) \cong H^{0,1}(E_1^*) \oplus H^{1,0}(E_1^*)$. Then the map $\mathbb{P} E^* \rightarrow \mathcal{H}^1(E_1^*) \rightarrow H^{0,1}(E_1^*) \cong H^0(E_1 \otimes K)^*$ is the usual map defined for holomorphic bundles (cf. \cite[p804]{HwangRamanan04}). \end{remark}
\begin{definition}\label{def:secant-variety} The \emph{$n^{th}$ secant variety}, denoted $\Sec^n(\mathcal{N}_{\phi_1, \phi_2}) \subset \mathbb{P} \mathcal{H}^1(E_1^* E_2)$, is the union of the subspaces $\vecspan \{ v_1, \ldots, v_n \} \subset \mathbb{P} \mathcal{H}^1(E_1^* E_2)$ taken over all $n$-tuples of distinct points $v_1, \ldots, v_n \in \mathcal{N}_{\phi_1, \phi_2}$. \end{definition}
The next lemma is a Higgs bundle version of \cite[Lemma 3.1]{NarasimhanRamanan69}. Since the proof is similar to that in \cite{NarasimhanRamanan69} then it is omitted.
\begin{lemma}\label{lem:nr-higgs} Let $0 \rightarrow (E_2, \phi_2) \rightarrow (F, \tilde{\phi}) \rightarrow (E_1, \phi_1) \rightarrow 0$ be an extension of Higgs bundles defined by the extension class $[(a, \varphi)] \in \mathcal{H}^1(E_1^* E_2)$. Let $(E', \phi') \stackrel{i}{\longrightarrow} (E_1, \phi_1)$ be a Higgs subsheaf such that $i^*[(a, \varphi)] = 0 \in \mathcal{H}^1((E')^* E_2)$. Then $(E', \phi')$ is a Higgs subsheaf of $(F, \tilde{\phi})$. \end{lemma}
\begin{equation*} \xymatrix{
& & & (E', \phi') \ar[d]^i \ar@{-->}[dl] \\ 0 \ar[r] & (E_2, \phi_2) \ar[r] & (F, \tilde{\phi}) \ar[r] & (E_1, \phi_1) \ar[r] & 0 \\ }\qedhere \end{equation*}
\begin{corollary}\label{cor:linear-span} Let $(E_1, \phi_1)$ be stable, let $n < \frac{1}{k} s_k(E_1, \phi_1)$ for all $k=1, \ldots, \rank(E_1)-1$, let $(E_2, \phi_2)$ be a Higgs line bundle and suppose that $\deg E_2 < \frac{\deg E_1 - n}{\rank E_1}$. Let $0 \rightarrow (E_2, \phi_2) \rightarrow (F, \tilde{\phi}) \rightarrow (E_1, \phi_1) \rightarrow 0$ be an extension of Higgs bundles with extension class $[(a, \varphi)] \in \mathcal{H}^1(E_1^* E_2)$. Let $0 \rightarrow (E', \phi') \stackrel{i}{\hookrightarrow} (E_1, \phi_1) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$ be a Hecke modification determined by distinct points $\{v_1, \ldots, v_n\} \in \mathcal{N}_{\phi_1, \phi_2}$.
Then $(E', \phi')$ is a subsheaf of $(F, \tilde{\phi})$ if $[(a, \varphi)] \in \vecspan\{ v_1, \ldots, v_n \} \subset \mathcal{H}^1(E_1^* E_2)$. \end{corollary}
\begin{proof} If $[(a, \varphi)] \in \vecspan\{ v_1, \ldots, v_n \}$ then $[(a, \varphi)] \in \ker (\mathcal{H}^1(E_1^* E_2) \rightarrow \mathcal{H}^1((E')^* E_2))$ by Corollary \ref{cor:n-dim-kernel}, and therefore $(E', \phi')$ is a subsheaf of $(F, \tilde{\phi})$ by Lemma \ref{lem:nr-higgs}. \end{proof}
The next lemma gives a condition on the extension class $[(a, \varphi)] \in \mathcal{H}^1(E_1^* E_2)$ for $(E', \phi')$ to be the subsheaf of largest degree which lifts to a subsheaf of $(F, \tilde{\phi})$. This is used to study unbroken flow lines in Section \ref{sec:secant-criterion}.
\begin{lemma}\label{lem:nondegenerate-maximal} Let $(E_1, \phi_1)$ be a stable Higgs bundle, choose $n$ such that $2n-1 < \frac{1}{k} s_k(E_1, \phi_1)$ for all $k=1, \ldots, \rank(E_1)$, let $(E_2, \phi_2)$ be a Higgs line bundle and suppose that $\deg E_2 < \frac{\deg E_1 - (2n-1)}{\rank E_1}$. Let $0 \rightarrow (E_2, \phi_2) \rightarrow (F, \tilde{\phi}) \rightarrow (E_1, \phi_1) \rightarrow 0$ be an extension of Higgs bundles with extension class $[(a, \varphi)] \in \Sec^n(\mathcal{N}_{\phi_1, \phi_2}) \setminus \Sec^{n-1}(\mathcal{N}_{\phi_1, \phi_2}) \subset \mathbb{P} \mathcal{H}^1(E_1^* E_2)$ and let $0 \rightarrow (E', \phi') \stackrel{i}{\hookrightarrow} (E_1, \phi_1) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$ be a Hecke modification determined by distinct points $v_1, \ldots, v_n \in \mathcal{N}_{\phi_1, \phi_2}$ such that $i^* [(a, \varphi)] = 0$.
Let $(\mathcal{E}'', \phi'') \stackrel{i''}{\hookrightarrow} (\mathcal{E}, \phi)$ be a subsheaf such that $(i'')^* [(a, \varphi)] = 0 \in \mathcal{H}^1((E'')^* E_2)$ and $\rank E'' = \rank E$. Then $\deg(E'') \leq \deg(E')$. \end{lemma}
\begin{proof} Let $\{ v_1'', \ldots, v_m'' \} \subset \mathcal{N}_{\phi_1, \phi_2}$ be the set of distinct points defining the Hecke modification $(\mathcal{E}'', \phi'') \stackrel{i''}{\hookrightarrow} (\mathcal{E}_1, \phi_1)$. Then $i^* [(a, \varphi)] = 0$ and $(i'')^*[(a, \varphi)] = 0$ together imply that $[(a, \varphi)] \in \vecspan\{ v_1, \ldots, v_n \} \cap \vecspan \{ v_1'', \ldots, v_m''\}$. Either $m + n > 2n-1$ (and so $\deg E'' \leq \deg E'$) or $m + n \leq 2n-1$ in which case Corollary \ref{cor:n-dim-kernel} together with the lower bound $2n-1 < \frac{1}{k} s_k(E_1, \phi_1)$ implies that $\vecspan \{ v_1, \ldots, v_n \} \cap \vecspan \{v_1'', \ldots, v_m'' \}$ is the linear span of $\{ v_1, \ldots, v_n \} \cap \{ v_1'', \ldots, v_m'' \}$. Since $m+n \leq 2n-1$ then $\{ v_1, \ldots, v_n \} \cap \{ v_1'', \ldots, v_m'' \}$ is a strict subset of $\{ v_1, \ldots, v_n\}$, which is not possible since $[(a, \varphi)] \notin \Sec^{n-1}(\mathcal{N}_{\phi_1, \phi_2})$. Therefore $\deg E'' \leq \deg E'$. \end{proof}
\subsection{Constructing Hecke modifications of Higgs bundles via the Yang-Mills-Higgs flow.}\label{sec:YMH-flow-hecke}
Let $(E, \phi)$ be a stable Higgs bundle and $L_u$ a line bundle with $\deg L_u < \frac{\deg E - 1}{\rank E}$, and let $E'$ be a Hecke modification of $E$ which is compatible with the Higgs field \begin{equation*} 0 \rightarrow (\mathcal{E}', \phi') \stackrel{i}{\hookrightarrow} (\mathcal{E}, \phi) \stackrel{v}{\rightarrow} \mathbb C_p \rightarrow 0 . \end{equation*}
The goal of this section is to construct critical points $x_u = (L_u, \phi_u) \oplus (E, \phi)$ and $x_\ell = (L_\ell, \phi_\ell) \oplus (E', \phi')$ together with a broken flow line connecting $x_u$ and $x_\ell$. The result of Theorem \ref{thm:algebraic-flow-line} shows that this amounts to constructing a Higgs field $\phi_u \in H^0(K)$, a Higgs pair $(F, \tilde{\phi})$ in the unstable set of $x_u$ and a complex gauge transformation $g \in \mathcal{G}^\mathbb C$ such that $(E', \phi')$ is a Higgs subbundle of $g \cdot (F, \tilde{\phi})$.
\begin{lemma}\label{lem:construct-Higgs-extension} Let $0 \rightarrow (\mathcal{E}', \phi') \rightarrow (\mathcal{E}, \phi) \stackrel{v}{\rightarrow} \mathbb C_p \rightarrow 0$ be a Hecke modification such that $(E, \phi)$ and $(E', \phi')$ are both Higgs semistable, and let $L_u$ be a line bundle with $\deg L_u < \slope(E') < \slope(E)$. Then there exists a Higgs field $\phi_u \in H^0(K)$ and a non-trivial Higgs extension $(F, \tilde{\phi})$ of $(L_u, \phi_u)$ by $(E, \phi)$ such that $(E', \phi')$ is a Higgs subsheaf of $(F, \tilde{\phi})$. \end{lemma}
\begin{proof} By Remark \ref{rem:miniscule-compatible}, there exists $\phi_u \in H^0(K)$ such that $v \in \mathcal{N}_{\phi, \phi_u}$. Since $(E', \phi')$ is semistable with $\slope(E') > \slope(L_u)$ then $\mathcal{H}^0((E')^* L_u) = 0$ and so the exact sequence \eqref{eqn:hyper-exact-sequence} shows that the Hecke modification $v \in \mathbb{P} E^*$ determines a one-dimensional subspace of $\mathcal{H}^1(E^* L_u)$, and that any non-trivial extension class in this subspace is in the kernel of the map $\mathcal{H}^1(E^* L_u) \rightarrow \mathcal{H}^1((E')^* L_u)$. Let $0 \rightarrow (L_u, \phi_u) \rightarrow (F, \tilde{\phi}) \rightarrow (E, \phi) \rightarrow 0$ be such an extension. Then Lemma \ref{lem:nr-higgs} shows that $(E', \phi')$ is a Higgs subsheaf of $(F, \tilde{\phi})$. \end{proof}
We can now use this result to relate Hecke modifications at a single point with $\mathop{\rm YMH}\nolimits$ flow lines.
\begin{theorem}\label{thm:flow-hecke}
\begin{enumerate}
\item Let $0 \rightarrow (E', \phi') \rightarrow (E, \phi) \stackrel{v}{\rightarrow} \mathbb C_p \rightarrow 0$ be a Hecke modification such that $(E, \phi)$ is stable and $(E', \phi')$ is semistable, and let $L_u$ be a line bundle with $\deg L_u + 1 < \slope(E') < \slope(E)$. Then there exist sections $\phi_u, \phi_\ell \in H^0(K)$, a line bundle $L_\ell$ with $\deg L_\ell = \deg L_u + 1$ and a metric on $E \oplus L_u$ such that $x_u = (E, \phi) \oplus (L_u, \phi_u)$ and $x_\ell = (E_{gr}', \phi_{gr}') \oplus (L_\ell, \phi_\ell)$ are critical points connected by a $\mathop{\rm YMH}\nolimits$ flow line, where $(E_{gr}', \phi_{gr}')$ is isomorphic to the graded object of the Seshadri filtration of $(E', \phi')$.
\item Let $x_u = (E, \phi) \oplus (L_u, \phi_u)$ and $x_\ell = (E', \phi') \oplus (L_\ell, \phi_\ell)$ be critical points connected by a $\mathop{\rm YMH}\nolimits$ flow line such that $L_u, L_\ell$ are line bundles with $\deg L_u = \deg L_\ell + 1$, $(E, \phi)$ is stable and $(E', \phi')$ is polystable with $\deg L_u + 1 < \slope(E') < \slope(E)$. If $(E', \phi')$ is Higgs stable then it is a Hecke modification of $(E, \phi)$. If $(E', \phi')$ is Higgs polystable then it is the graded object of the Seshadri filtration of a Hecke modification of $(E, \phi)$.
\end{enumerate}
\end{theorem}
\begin{proof}[Proof of Theorem \ref{thm:flow-hecke}] Given a Hecke modification $0 \rightarrow (\mathcal{E}', \phi') \rightarrow (\mathcal{E}, \phi) \rightarrow \mathbb C_p \rightarrow 0$ as in Lemma \ref{lem:construct-Higgs-extension}, choose $\phi_u \in H^0(K)$ such that $v \in \mathcal{N}_{\phi, \phi_u}$ and apply a gauge transformation to $E \oplus L_u$ such that $x_u = (E, \phi) \oplus (L_u, \phi_u)$ is a critical point of $\mathop{\rm YMH}\nolimits$. The harmonic representative of the extension class $[(a, \varphi)] \in \mathcal{H}^1(E^* L_u)$ from Lemma \ref{lem:construct-Higgs-extension} defines an extension $0 \rightarrow (L_u, \phi_u) \rightarrow (F, \tilde{\phi}) \rightarrow (E, \phi) \rightarrow 0$ such that $y = (F, \tilde{\phi})$ is in the negative slice of $x_u$, and therefore flows down to a limit isomorphic to the graded object of the Harder-Narasimhan-Seshadri filtration of $(F, \tilde{\phi})$.
Lemma \ref{lem:construct-Higgs-extension} also shows that $(E', \phi')$ is a Higgs subsheaf of $(F, \tilde{\phi})$. Lemma \ref{lem:resolve-higgs-subsheaf} shows that this has a resolution as a Higgs subbundle of $(F, \tilde{\phi})$, however since the Harder-Narasimhan type of $(F, \tilde{\phi})$ is strictly less than that of $(E, \phi) \oplus (L_u, \phi_u)$, $\rank(E') = \rank(F) - 1$ and $\deg E' = \deg E - 1$, then $(E', \phi')$ already has the maximal possible slope for a semistable Higgs subbundle of $(F, \tilde{\phi})$, and therefore $(E', \phi')$ must be the maximal semistable Higgs subbundle. Since $\rank(E') = \rank(F) - 1$, then the graded object of the Harder-Narasimhan-Seshadri filtration of $(F, \tilde{\phi})$ is $(E_{gr}', \phi_{gr}') \oplus (L_\ell, \phi_\ell)$, where $(L_\ell, \phi_\ell) = (F, \tilde{\phi}) / (E', \phi')$. Theorem \ref{thm:algebraic-flow-line} then shows that $(E, \phi) \oplus (L_u, \phi_u)$ and $(E_{gr}', \phi_{gr}') \oplus (L_\ell, \phi_\ell)$ are connected by a flow line.
Conversely, if $x_u = (E, \phi) \oplus (L_u, \phi_u)$ and $x_\ell = (E', \phi') \oplus (L_\ell, \phi_\ell)$ are critical points connected by a flow line, then Theorem \ref{thm:algebraic-flow-line} shows that there exists a Higgs pair $(F, \tilde{\phi})$ in the negative slice of $x_u$ such that $(E', \phi')$ is the graded object of the Seshadri filtration of the maximal semistable Higgs subbundle of $(F, \tilde{\phi})$. If $(E', \phi')$ is Higgs stable, then since $\slope(E') > \slope (L_u)$ we see $(E', \phi')$ is a Higgs subsheaf of $(E, \phi)$ with $\rank(E) = \rank(E')$ and $\deg(E') = \deg(E) - 1$. Therefore $(E', \phi')$ is a Hecke modification of $(E, \phi)$. If $(E', \phi')$ is Higgs polystable then the same argument shows that $(E', \phi')$ is the graded object of the Seshadri filtration of a Hecke modification of $(E, \phi)$. \end{proof}
In general, for any flow one can define the space $\mathcal{F}_{\ell, u}$ of flow lines connecting upper and lower critical sets $C_u$ and $C_\ell$, and the space $\mathcal{P}_{\ell, u} \subset C_u \times C_\ell$ of pairs of critical points connected by a flow line. These spaces are equipped with projection maps to the critical sets defined by the canonical projection taking a flow line to its endpoints. \begin{equation} \xymatrix{
& \mathcal{F}_{\ell, u} \ar[d] \ar@/_/[ddl] \ar@/^/[ddr] & \\
& \mathcal{P}_{\ell, u} \ar[dl] \ar[dr] & \\ C_\ell & & C_u } \end{equation}
For the Yang-Mills-Higgs flow, given critical sets $C_u$ and $C_\ell$ of respective Harder-Narasimhan types $(\frac{d}{r}, \deg L_u)$ and $(\frac{d-1}{r}, \deg L_u + 1)$ as in Theorem \ref{thm:flow-hecke} above, there are natural projection maps to the moduli space $C_u \rightarrow \mathcal{M}_{ss}^{Higgs}(r, d)$ and $C_\ell \rightarrow \mathcal{M}_{ss}^{Higgs}(r, d-1)$. Since the flow is $\mathcal{G}$-equivariant, then there is an induced correspondence variety $\mathcal{M}_{\ell, u} \subset \mathcal{M}_{ss}^{Higgs}(r, d-1) \times \mathcal{M}_{ss}^{Higgs}(r, d)$. \begin{equation}\label{eqn:flow-hecke-diagram} \xymatrix{
& \mathcal{P}_{\ell, u} \ar[dl] \ar[dr] \ar[d] & \\ C_\ell \ar[d] & \mathcal{M}_{\ell,u} \ar[dl] \ar[dr] & C_u \ar[d] \\ \mathcal{M}_{ss}^{Higgs}(r, d-1) & & \mathcal{M}_{ss}^{Higgs}(r,d) } \end{equation}
Theorem \ref{thm:flow-hecke} shows that $\left( (E', \phi'), (E, \phi) \right) \in \mathcal{M}_{\ell, u}$ if and only if $(E', \phi')$ is a Hecke modification of $(E, \phi)$ and both Higgs pairs are semistable. If $r$ and $d$ are coprime then $\mathcal{M}_{ss}^{Higgs}(r,d)$ consists of stable Higgs pairs and so every Hecke modification of $(E, \phi)$ is semistable. Therefore we have proved \begin{corollary} $\mathcal{M}_{\ell,u}$ is the Hecke correspondence. \end{corollary}
For Hecke modifications defined at multiple points (non-miniscule Hecke modifications in the terminology of \cite{witten-hecke}), we immediately have the following result.
\begin{corollary}\label{cor:broken-hecke} Let $(E, \phi)$ be a $(0,n)$-stable Higgs bundle and consider a Hecke modification $0 \rightarrow (\mathcal{E}', \phi') \rightarrow (\mathcal{E}, \phi) \rightarrow \oplus_{j=1}^n \mathbb C_{p_n} \rightarrow 0$ defined by $n > 1$ distinct points $\{ v_1, \ldots, v_n \} \in \mathbb{P} E^*$. If there exists $\phi_u \in H^0(K)$ such that $v_1, \ldots, v_n \in \mathcal{N}_{\phi, \phi_u}$, then there is a broken flow line connecting $x_u = (E, \phi) \oplus (L_u, \phi_u)$ and $x_\ell = (E_{gr}', \phi_{gr}') \oplus (L_\ell, \phi_\ell)$, where $(E_{gr}', \phi_{gr}')$ is the graded object of the Seshadri filtration of the semistable Higgs bundle $(E', \phi')$. \end{corollary}
\begin{proof} Inductively apply Theorem \ref{thm:flow-hecke}. \end{proof}
\subsection{A geometric criterion for unbroken $\mathop{\rm YMH}\nolimits$ flow lines}\label{sec:secant-criterion}
Corollary \ref{cor:broken-hecke} gives a criterion for two $\mathop{\rm YMH}\nolimits$ critical points $x_u = (E, \phi) \oplus (L_u, \phi_u)$ and $x_\ell = (E', \phi') \oplus (L_\ell, \phi_\ell)$ to be connected by a broken flow line. It is natural to ask whether they are also connected by an \emph{unbroken} flow line. The goal of this section is to answer this question by giving a geometric construction for points in the negative slice of $x_u$ which correspond to unbroken flow lines connecting $x_u$ and $x_\ell$ in terms of the secant varieties $\Sec^n(\mathcal{N}_{\phi, \phi_u})$. For holomorphic bundles, the connection between secant varieties and Hecke modifications has been studied in \cite{LangeNarasimhan83}, \cite{ChoeHitching10} and \cite{Hitching13}.
Given a $\mathop{\rm YMH}\nolimits$ critical point $x_u = (E, \phi) \oplus (L_u, \phi_u)$ with $(E, \phi)$ stable and $\rank L_u =1$, consider an extension $0 \rightarrow (L_u, \phi_u) \rightarrow (F, \tilde{\phi}) \rightarrow (E, \phi) \rightarrow 0$ with extension class $[(a, \varphi)] \in \mathcal{H}^1(E^* L_u) = S_{x_u}^-$. Let $0 \rightarrow (E', \phi') \rightarrow (E, \phi) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$ be a Hecke modification of $(E, \phi)$ as in the previous lemma, such that $\deg L_u < \slope(E')$.
\begin{lemma}\label{lem:subbundle-slope-bound} If $(G, \phi_G)$ is a semistable Higgs subbundle of $(F, \tilde{\phi})$ with $\slope(G) > \deg L_u$ and $\rank(G) < \rank (E)$, then there is a Higgs subbundle $(G', \phi_G') \subset (E, \phi)$ with $\slope(G') \geq \slope(G)$ and $\rank(G) = \rank(G')$. \end{lemma}
\begin{proof} If $(G, \phi_G)$ is a semistable Higgs subbundle of $(F, \tilde{\phi})$ with $\slope(G) > \deg L_u$, then $\mathcal{H}^0(G^* L_u) = 0$, and so $(G, \phi_G)$ is a Higgs subsheaf of $(E, \phi)$. \begin{equation*} \xymatrix{
& & & (G, \phi_G) \ar[dl] \ar@{-->}[d] \\ 0 \ar[r] & (L_u, \phi_u) \ar[r] & (F, \tilde{\phi}) \ar[r] & (E, \phi) \ar[r] & 0 } \end{equation*} Lemma \ref{lem:resolve-higgs-subsheaf} shows that the subsheaf $(G, \phi_G)$ can be resolved to form a Higgs subbundle $(G', \phi_G')$ of $(E, \phi)$ with $\slope(G') \geq \slope(G)$. \end{proof}
\begin{theorem}\label{thm:unbroken-criterion} Let $(E, \phi)$ be a stable Higgs bundle with Segre invariant $s_k(E, \phi)$ and choose $n$ such that $0 < 2n-1 < \min_{1 \leq k \leq r-1} \left( \frac{1}{k} s_k(E, \phi) \right)$. Let $0 \rightarrow (\mathcal{E}', \phi') \rightarrow (\mathcal{E}, \phi) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0$ be a Hecke modification of $(E, \phi)$ defined by distinct points $v_1, \ldots, v_n \in \mathbb{P} E^*$, and let $(L_u, \phi_u)$ be a Higgs line bundle such that $v_1, \ldots, v_n \in \mathcal{N}_{\phi, \phi_u}$. Choose a metric such that $x_u = (E, \phi) \oplus (L_u, \phi_u)$ is a $\mathop{\rm YMH}\nolimits$ critical point.
Then any extension class $[(a, \varphi)] \in \vecspan \{ v_1, \ldots, v_n \} \cap \left( \Sec^n(\mathcal{N}_{\phi, \phi_u}) \setminus \Sec^{n-1}(\mathcal{N}_{\phi, \phi_u}) \right) \subset \mathbb{P} \mathcal{H}^1(E^* L_u)$ is isomorphic to an unbroken flow line connecting $x_u = (E, \phi) \oplus (L_u, \phi_u)$ and $x_\ell = (E', \phi') \oplus (L_\ell, \phi_\ell)$. \end{theorem}
\begin{proof} Let $(F, \tilde{\phi})$ be a Higgs bundle determined by the extension class $[(a, \varphi)] \in \mathbb{P} \mathcal{H}^1(E^* L_u)$. The choice of bundle is not unique, but the isomorphism class of $(F, \tilde{\phi})$ is unique. The proof reduces to showing that $(E', \phi')$ is the maximal semistable Higgs subbundle of $(F, \tilde{\phi})$.
Since $[(a, \varphi)] \notin \Sec^{n-1}(\mathcal{N}_{\phi, \phi_u})$, then Lemma \ref{lem:nondegenerate-maximal} shows that $(E', \phi')$ is the subsheaf of $(E, \phi)$ with maximal degree among those that lift to a subsheaf of $(F, \tilde{\phi})$. Any semistable Higgs subbundle $(E'', \phi'')$ of $(F, \tilde{\phi})$ with $\rank(E'') = \rank(E)$ either has $\slope(E'') \leq \deg L_u < \slope(E')$, or it is a subsheaf of $(E, \phi)$ and so must have $\slope(E'') \leq \slope(E')$.
The previous lemma shows that if $(G, \phi_G)$ is any semistable Higgs subbundle of $(F, \tilde{\phi})$ with $\slope(G) > \deg L_u$ and $\rank(G) < \rank(E)$, then there is a Higgs subbundle $(G', \phi_G')$ of $(E, \phi)$ with $\slope(G') \geq \slope(G)$. The upper bound on $n = \deg E - \deg E'$ in terms of the Segre invariant then implies that $\slope(E') > \slope(G') \geq \slope(G)$ by Lemma \ref{lem:segre-bound}.
Therefore the subbundle $(\tilde{E}', \tilde{\phi}')$ resolving the subsheaf $(E', \phi') \subset (F, \tilde{\phi})$ is the maximal semistable Higgs subbundle of $(F, \tilde{\phi})$. Since $(\tilde{E}', \tilde{\phi}')$ is semistable and $\slope(\tilde{E}') \geq \slope(E') > \deg L_u$, then $\mathcal{H}^0((\tilde{E}')^* L_u) = 0$, and so $(\tilde{E}', \tilde{\phi}')$ is a Higgs subsheaf of $(E, \phi)$ that lifts to a subbundle of $(F, \tilde{\phi})$. Since $\deg E'$ is maximal among all such subsheaves, then we must have $(E', \phi') = (\tilde{E}', \tilde{\phi}')$ and so $(E', \phi')$ is the maximal semistable subbundle of $(F, \tilde{\phi})$. Therefore Theorem \ref{thm:algebraic-flow-line} shows that $x_u$ and $x_\ell$ are connected by an unbroken flow line. \end{proof}
If $\rank(F) = 2$ (so that $E$ is a line bundle), then the condition on the Segre invariant $s_k(E, \phi)$ becomes vacuous. Moreover, $\mathbb{P} E^* \cong X$ and so Hecke modifications of $E$ are determined by a subset $\{ v_1, \ldots, v_n \} \subset X$. Therefore in the case $\rank(F) = 2$, we have a complete classification of the $\mathop{\rm YMH}\nolimits$ flow lines on the space of Higgs bundles $\mathcal{B}(F)$.
\begin{corollary}\label{cor:rank-2-classification} Let $F \rightarrow X$ be a $C^\infty$ Hermitian vector bundle with $\rank(F) = 2$. Let $x_u = (L_1^u, \phi_1^u) \oplus (L_2^u, \phi_2^u)$ and $x_\ell = (L_1^\ell, \phi_1^\ell) \oplus (L_2^\ell, \phi_2^\ell)$ be non-minimal critical points with $\mathop{\rm YMH}\nolimits(x_u) > \mathop{\rm YMH}\nolimits(x_\ell)$. Suppose without loss of generality that $\deg L_1^u > \deg L_1^\ell > \deg L_2^\ell > \deg L_2^u$. Let $n = \deg L_1^u - \deg L_1^\ell$.
Then $x_u$ and $x_\ell$ are connected by a broken flow line if and only if there exists $\{ v_1, \ldots, v_n \} \in \mathcal{N}_{\phi_1^u, \phi_2^u}$ such that \begin{align*} 0 \rightarrow (L_1^\ell, \phi_1^\ell) \rightarrow (L_1^u, \phi_1^u) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0 \\ 0 \rightarrow (L_2^u, \phi_2^u) \rightarrow (L_2^\ell, \phi_2^\ell) \rightarrow \oplus_{j=1}^n \mathbb C_{p_j} \rightarrow 0 \end{align*} are both Hecke modifications determined by $\{ v_1, \ldots, v_n \}$. They are connected by an unbroken flow line if the previous condition holds and $\{ v_1, \ldots, v_n \} \in \Sec^n(\mathcal{N}_{\phi_1^u, \phi_2^u}) \setminus \Sec^{n-1}(\mathcal{N}_{\phi_1^u, \phi_2^u})$. \end{corollary}
\appendix
\section{Uniqueness for the reverse Yang-Mills-Higgs flow}\label{sec:uniqueness}
The methods of Donaldson \cite{Donaldson85} and Simpson \cite{Simpson88} show that the Yang-Mills-Higgs flow resembles a nonlinear heat equation, and therefore the backwards flow is ill-posed. In Section \ref{sec:scattering-convergence} we prove existence of solutions to the backwards heat flow that converge to a critical point. To show that these solutions are well-defined we prove in this section that if a solution to the reverse $\mathop{\rm YMH}\nolimits$ flow exists then it must be unique.
Using the Hermitian metric, let $d_A$ be the Chern connection associated to $\bar{\partial}_A$ and let $\psi = \phi + \phi^* \in \Omega^1(i \mathop{\rm ad}\nolimits(E))$. The holomorphicity condition $\bar{\partial}_A \phi = 0$ becomes the pair of equations $d_A \psi = 0$, $d_A^* \psi = 0$ which also imply that $[F_A, \psi] = d_A^2 \psi = 0$, and the Yang-Mills-Higgs functional is $\| F_A + \psi \wedge \psi \|_{L^2}^2$.
\begin{proposition}\label{prop:backwards-uniqueness} Let $(d_{A_1}, \psi_1)(t)$, $(d_{A_2}, \psi_2)(t)$ be two solutions of the Yang-Mills-Higgs flow \eqref{eqn:YMH-flow-general} on a compact Riemann surface with respective initial conditions $(d_{A_1}, \psi_1)(0)$ and $(d_{A_2}, \psi_2)(0)$. If there exists a finite $T > 0$ such that $(d_{A_1}, \psi_1)(T) = (d_{A_2}, \psi_2)(T)$ then $(d_{A_1}, \psi_1)(t) = (d_{A_2}, \psi_2)(t)$ for all $t \in [0, T]$. \end{proposition}
The result of Proposition \ref{prop:backwards-uniqueness} is valid when the base manifold is a compact Riemann surface, since we use the estimates of \cite[Sec. 3.2]{Wilkin08} to prove that the constant $C$ in Lemma \ref{lem:heat-inequalities} is uniform. In the case of the Yang-Mills flow on a compact K\"ahler manifold the estimates of Donaldson in \cite{Donaldson85} show that we can make this constant uniform on a finite time interval $[0,T]$ and so the result also applies in this setting. The setup described in the previous paragraph consisting of Higgs pairs $(d_A, \psi)$ satisfying $d_A \psi = 0$, $d_A^* \psi = 0$ is valid on any Riemannian manifold, and so the result of Proposition \ref{prop:backwards-uniqueness} will also apply to any class of solutions for which one can prove that the connection, Higgs field, the curvature and all of their derivatives are uniformly bounded on the given finite time interval $[0, T]$.
Let $\nabla_A$ denote the covariant derivative associated to the connection $d_A$. The complex connection associated to the pair $(d_A, \psi)$ is $D_{(A, \psi)} \eta = d_A \eta + [\psi, \eta]$ and the Laplacian is $\Delta_{(A, \psi)} \eta = D_{(A, \psi)}^* D_{(A, \psi)} \eta + D_{(A, \psi)} D_{(A, \psi)}^* \eta$ for any form $\eta \in \Omega^p(\mathop{\rm End}\nolimits(E))$. The equation $d_A \psi = 0$ implies that the curvature of the complex connection is $D_{(A, \psi)} D_{(A, \psi)} \eta = [F_A + \psi \wedge \psi, \eta]$.
We have the following identities which will be useful in what follows. The notation $a \times b$ is used to denote various bilinear expressions with constant coefficients.
\begin{align} 0 & = d_A(F_A + \psi \wedge \psi) , \quad 0 = [\psi, F_A + \psi \wedge \psi] \label{eqn:Higgs-Bianchi} \\ \Delta_{(A, \psi)} \eta & = \nabla_A^* \nabla_A \eta + (F_A + \psi \wedge \psi) \times \eta + R_M \times \eta + \psi \times \psi \times \eta + \nabla_A \psi \times \psi \times \eta \quad \label{eqn:Higgs-Weitzenbock} \\ 0 & = D_{(A, \psi)}^* D_{(A, \psi)}^* (F_A + \psi \wedge \psi) \label{eqn:compose-adjoint} \end{align} The first identity follows from the Bianchi identity and the equation $d_A \psi = 0$. Equation \eqref{eqn:Higgs-Weitzenbock} is the Weitzenbock identity for a Higgs pair which follows from the usual identity for $\nabla_A$ (see for example \cite{BourguignonLawson81}) together with the fact that $(\psi \wedge \psi) \times \eta$ and the remaining terms in the Laplacian are of the form $\psi \times \psi \times \eta + \nabla_A \psi \times \psi \times \eta$. To see the identity \eqref{eqn:compose-adjoint}, take the inner product of the right hand side with an arbitrary $\eta \in \Omega^0(\mathop{\rm End}\nolimits(E))$. We have (cf. \cite[(2.2)]{Rade92} for the case $\psi = 0$) \begin{align*} \left< D_{(A, \psi)}^* D_{(A, \psi)}^* (F_A + \psi \wedge \psi), \eta \right> & = \left< F_A + \psi \wedge \psi, D_{(A, \psi)} D_{(A, \psi)} \eta \right> \\
& = \left< F_A + \psi \wedge \psi, [F_A + \psi \wedge \psi, \eta] \right> = 0 \end{align*}
Consider the Yang-Mills-Higgs flow equations \begin{equation}\label{eqn:YMH-flow-general} \frac{\partial A}{\partial t} = - d_A^* (F_A + \psi \wedge \psi), \quad \frac{\partial \psi}{\partial t} = * [\psi, *(F_A + \psi \wedge \psi)] \end{equation} After using the metric to decompose $\Omega^1(\mathop{\rm End}\nolimits(E)) \cong \Omega^1(\mathop{\rm ad}\nolimits(E)) \oplus \Omega^1(i \mathop{\rm ad}\nolimits(E))$, the flow equation can be written more compactly as \begin{equation*} \frac{\partial}{\partial t} (d_A + \psi) = - D_{(A, \psi)}^* (F_A + \psi \wedge \psi) \end{equation*} We then have \begin{align*} \frac{\partial}{\partial t} (F_A + \psi \wedge \psi) & = d_A \left( \frac{\partial A}{\partial t} \right) + \frac{\partial \psi}{\partial t} \wedge \psi + \psi \wedge \frac{\partial \psi}{\partial t} \\
& = - d_A d_A^* (F_A + \psi \wedge \psi) + \left[ \psi, *[\psi, *(F_A + \psi \wedge \psi)] \right] \\
& = - \Delta_{(A, \psi)} (F_A + \psi \wedge \psi) - d_A*[\psi, *(F_A + \psi \wedge \psi)] + [\psi, d_A^*(F_A + \psi \wedge \psi)] \end{align*} where in the last step we use the Bianchi identity \eqref{eqn:Higgs-Bianchi}. We also have \begin{align*} \frac{\partial}{\partial t} \left( d_A^* (F_A + \psi \wedge \psi) \right) & = -*\left[ \frac{\partial A}{\partial t}, *(F_A + \psi \wedge \psi) \right] + d_A^* \left( \frac{\partial}{\partial t} (F_A + \psi \wedge \psi) \right) \\
& = * \left[ d_A^*(F_A + \psi \wedge \psi), F_A + \psi \wedge \psi \right] - d_A^* d_A d_A^* (F_A + \psi \wedge \psi) \\
& \quad \quad + d_A^* [\psi, *[\psi, *(F_A + \psi \wedge \psi)]] \end{align*} and \begin{align*} \frac{\partial}{\partial t} \left( -*[\psi, *(F_A + \psi \wedge \psi)] \right) & = -* \left[ \frac{\partial \psi}{\partial t}, *(F_A + \psi \wedge \psi) \right] - *\left[ \psi, \frac{\partial}{\partial t} *(F_A + \psi \wedge \psi) \right] \\
& = * \left[ - *[\psi, *(F_A + \psi \wedge \psi)], *(F_A + \psi \wedge \psi) \right] + * \left[ \psi, * d_A d_A^* (F_A + \psi \wedge \psi) \right] \\
& \quad \quad - * \left[ \psi, *[\psi, *[\psi, *(F_A + \psi \wedge \psi)]] \right] \end{align*} Adding these two results gives us \begin{align*} \frac{\partial}{\partial t} \left( D_{(A, \psi)}^* (F_A + \psi \wedge \psi) \right) & = * \left[ D_{(A, \psi)}^* (F_A + \psi \wedge \psi), F_A + \psi \wedge \psi \right] - D_{(A, \psi)}^* D_{(A, \psi)} D_{(A, \psi)}^* (F_A + \psi \wedge \psi) \\
& = * \left[ D_{(A, \psi)}^* (F_A + \psi \wedge \psi), F_A + \psi \wedge \psi \right] - \Delta_{(A, \psi)} D_{(A, \psi)}^* (F_A + \psi \wedge \psi) \end{align*} where the last step uses \eqref{eqn:compose-adjoint}. Let $\mu_{(A, \psi)} = F_A + \psi \wedge \psi$ and $\nu_{(A, \psi)} = D_{(A, \psi)}^* (F_A + \psi \wedge \psi)$. The above equations become \begin{align} \left( \frac{\partial}{\partial t} + \Delta_{(A, \psi)} \right) \mu_{(A, \psi)} & = -d_A*[\psi, *(F_A + \psi \wedge \psi)] + [\psi, d_A^*(F_A + \psi \wedge \psi)] \label{eqn:mu-evolution} \\ \left( \frac{\partial}{\partial t} + \Delta_{(A, \psi)} \right) \nu_{(A, \psi)} & = *[\nu_{(A, \psi)}, *\mu_{(A, \psi)}] \label{eqn:nu-evolution} \end{align}
Now consider two solutions $(A_1, \psi_1)(t)$ and $(A_2, \psi_2)(t)$ to the Yang-Mills-Higgs flow equations \eqref{eqn:YMH-flow-general} on the time interval $[0,T]$ such that $(A_1, \psi_1)(T) = (A_2, \psi_2)(T)$. We will show below that this implies $(A_1, \psi_1)(0) = (A_2, \psi_2)(0)$.
Define $(a_t, \varphi_t) = (A_2, \psi_2)(t) - (A_1, \psi_1)(t)$, $m_t = \mu_{(A_2, \psi_2)} - \mu_{(A_1, \psi_1)}$ and $n_t = \nu_{(A_2, \psi_2)} - \nu_{(A_1, \psi_1)}$. In terms of $(a_t, \varphi_t)$ we can write \begin{align*} m_t = \mu_{(A_2, \psi_2)} - \mu_{(A_1, \psi_1)} = d_{A_1} a_t + a_t \wedge a_t + [\psi, \varphi_t] + \varphi_t \wedge \varphi_t \end{align*} and for any $\eta \in \Omega^p(\mathop{\rm End}\nolimits(E))$ the difference of the associated Laplacians has the form \begin{equation}\label{eqn:laplacian-difference} \left(\Delta_{(A_2, \psi_2)} - \Delta_{(A_1, \psi_1)} \right) \eta = \nabla_A a \times \eta + a \times \nabla_A \eta + a \times a \times \eta + \psi \times \varphi \times \eta + \varphi \times \varphi \times \eta \end{equation} where again $\omega_1 \times \omega_2$ is used to denote a bilinear expression in $\omega_1$ and $\omega_2$ with constant coefficients. By definition of $\nu_{(A, \psi)}$ as the gradient of the Yang-Mills-Higgs functional at $(d_A, \psi)$ we immediately have \begin{equation*} \frac{\partial}{\partial t} (a_t + \varphi_t) = n_t , \quad \text{and} \quad \frac{\partial}{\partial t} (\nabla_A a_t + \nabla_A \varphi_t) = \left( \frac{\partial A}{\partial t} \times a_t, \frac{\partial A}{\partial t} \times \varphi_t \right) + \nabla_A n_t \end{equation*} Equation \eqref{eqn:mu-evolution} then becomes \begin{align*} \left( \frac{\partial}{\partial t} + \Delta_{(A_1, \psi_1)} \right) m_t & = - \left( \Delta_{(A_2, \psi_2)} - \Delta_{(A_1, \psi_1)} \right) \mu_{(A_2, \psi_2)} \\
& \quad \quad + a_t \times \psi_1 \times (F_{A_1} + \psi_1 \wedge \psi_1) + \nabla_{A_1} \varphi_t \times (F_{A_1} + \psi_1 \wedge \psi_1) \\
& \quad \quad + \nabla_{A_1} \psi_1 \times m_t + \psi_1 \times n_t \end{align*} and equation \eqref{eqn:nu-evolution} becomes \begin{align*} \left( \frac{\partial}{\partial t} + \Delta_{(A_1, \psi_1)} \right) n_t & = *[\nu_{(A_2, \psi_2)}, * \mu_{(A_2, \psi_2)}] - *[\nu_{(A_1, \psi_1)}, *\mu_{(A_1, \psi_1)}] - \left( \Delta_{(A_2, \psi_2)} - \Delta_{(A_1, \psi_1)} \right) \nu_{(A_2, \psi_2)} \\
& = *[n_t, *\mu_{(A_2, \psi_2)}] +*[\nu_{(A_1, \psi_1)}, *m_t] - \left( \Delta_{(A_2, \psi_2)} - \Delta_{(A_1, \psi_1)} \right) \nu_{(A_2, \psi_2)} \\ \end{align*} Using \eqref{eqn:laplacian-difference} and the Weitzenbock formula \eqref{eqn:Higgs-Weitzenbock}, we then have the following inequalities. In the case where $X$ is a compact Riemann surface then the estimates of \cite[Sec. 2.2]{Wilkin08} show that all of the derivatives of the connection, the Higgs field and the curvature $F_A$ are uniformly bounded along the flow and so the constant can be chosen uniformly on the interval $[0,T]$. \begin{lemma}\label{lem:heat-inequalities} For any pair of solutions $(d_{A_1}, \psi_1)(t)$ and $(d_{A_2}, \psi_2)(t)$ to the Yang-Mills-Higgs flow \eqref{eqn:YMH-flow-general} there exists a positive constant $C$ (possibly depending on $t$) such that the following inequalities hold \begin{align}
\left| \left( \frac{\partial}{\partial t} + \nabla_{A_1}^* \nabla_{A_1} \right) m_t \right| & \leq C \left( |a_t| + |\varphi_t| + | \nabla_{A_1} a_t| +| \nabla_{A_1} \varphi_t |+ | m_t | + |n_t| \right) \label{eqn:m-evolution} \\
\left| \left( \frac{\partial}{\partial t} + \nabla_{A_1}^* \nabla_{A_1} \right) n_t \right| & \leq C \left( |a_t| + |\varphi_t| + | \nabla_{A_1} a_t| +| \nabla_{A_1} \varphi_t |+ | m_t | + |n_t| \right) \label{eqn:n-evolution} \\
\left| \frac{\partial}{\partial t} (a_t + \varphi_t) \right| & = | n_t | \label{eqn:a-evolution} \\
\left| \frac{\partial}{\partial t} (\nabla_A a_t + \nabla_A \varphi_t) \right| & \leq C \left( |a_t| + |\varphi_t| + | \nabla_A n_t | \right) \label{eqn:nabla-a-evolution} \end{align} Moreover, if $X$ is a compact Riemann surface then the constant $C$ can be chosen uniformly on any finite time interval $[0, T]$. \end{lemma}
For simplicity of notation, in the following we use $\nabla := \nabla_{A_1}$ and $\square := \nabla_{A_1}^* \nabla_{A_1}$. Let $X := (m_t, n_t)$ and $Y := (a_t, \varphi_t, \nabla a_t, \nabla \varphi_t)$. The previous lemma implies that there exists a positive constant $C$ such that the following inequalities hold \begin{align}\label{eqn:coupled-system} \begin{split}
\left| \frac{\partial X}{\partial t} + \square X \right| & \leq C \left( | X | + | \nabla X| + | Y | \right) \\
\left| \frac{\partial Y}{\partial t} \right| & \leq C \left( |X| + |\nabla X| + |Y| \right) \end{split} \end{align} A general result of Kotschwar in \cite[Thm 3]{kotschwar-uniqueness} shows that any system satisfying \eqref{eqn:coupled-system} on the time interval $[0,T]$ for which $X(T) = 0$, $Y(T)=0$, must also satisfy $X(t) = 0$, $Y(t) = 0$ for all $t \in [0, T]$. In the context of the Yang-Mills-Higgs flow \eqref{eqn:YMH-flow-general}, this gives us the proof of Proposition \ref{prop:backwards-uniqueness}.
\end{document}
|
arXiv
|
{
"id": "1605.05970.tex",
"language_detection_score": 0.5760983228683472,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{DropCompute: simple and more robust distributed synchronous training via compute variance reduction}
\begin{abstract} \textbf{Background.} Distributed training is essential for large scale training of deep neural networks (DNNs). The dominant methods for large scale DNN training are synchronous (e.g. \textit{All-Reduce}), but these require waiting for all workers in each step. Thus, these methods are limited by the delays caused by straggling workers.\\ \textbf{Results.} We study a typical scenario in which workers are straggling due to variability in compute time. We find an analytical relation between compute time properties and scalability limitations, caused by such straggling workers. With these findings, we propose a simple yet effective decentralized method to reduce the variation among workers and thus improve the robustness of synchronous training. This method can be integrated with the widely used \textit{All-Reduce}. Our findings are validated on large-scale training tasks using 200 Gaudi Accelerators. A reference implementation\footnote[2]{\url{https://github.com/paper-submissions/dropcompute}} is provided. \end{abstract}
\section{Introduction}
Deep Neural Networks (DNNs) training continues to scale over size and computational footprint, as a result of a higher number of trainable parameters, wider and deeper models, and growing amounts of training data. As improvements in model quality (as measured by test loss, for example) \citep{scalingnlp} lead over hardware capabilities \citep{hwlottery}, this scale-up translates into a need for a growing number of training devices working in tandem \citep{chowdhery2022palm}, turning distributed training to the standard approach for training DNNs on a large scale.
\begin{figure}\label{fig:abstract}
\end{figure}
Distributed training typically refers to three parallelism paradigms --- data parallel, model parallel and layer pipelining \citep{DemystifyingDistributed}. Several variants and hybrid solutions exist in modern implementations such as tensor parallel \citep{Megatron2021} and parameter sharding \citep{rajbhandari2020zero, rasley2020deepspeed}. These can be used separately or combined as they are orthogonal to each other. Mainly, data parallelism is straightforward, where the data is sharded among workers, and all workers share the same global model state. At each step, workers compute gradients locally and then aggregate them before taking an optimization step. When training synchronously, workers update their parameters in lockstep. This ensures that all workers hold a consensus on the same model and that gradients are averaged over all workers before being applied to the model. This approach is easy to implement and allows for good convergence properties, and correspondingly is the prevalent optimization method.
Although state-of-the-art models use synchronous optimization for training, synchronous methods scale poorly, as stragglers and communication overhead might severely deteriorate system utilization. These issues are exacerbated as the required scale grows, even in homogeneous high-performance computing clusters. We are interested in cases where significant computing variance between the workers exists. This includes (but is not limited to) straggling workers. For instance, certain learning tasks entail heterogeneity in the required computation of data, such as varying sentence lengths in language processing, or different image sizes and frame numbers in computer vision. In addition, recent state-of-the-art models use all three parallelism paradigms (data (DP), tensor (TP), and pipeline (PP) parallelism), thus each data parallel node is a set of processing units (accelerators) communicating between them (via TP and PP) to calculate the model gradients collectively. This could potentially intensify compute variance between data parallel workers. Furthermore, slowdowns due to changes over time in hardware state or load might occur and cause variance in compute and specifically lead to straggling workers. As compute variance grows, the utilization deteriorates, such that more workers remain idle waiting for slower workers to complete calculating their gradients \cite{chen2016revisiting,chen2019round,ji2022ep4ddl}.
In this paper, we suggest a simple, yet effective method called \textit{DropCompute} to improve the robustness and scalability of synchronous optimization in the face of compute variance. We model the compute time as a random variable and show that under reasonable assumptions, a tail of straggling workers slows down the system at a rate that is not proportional to the contributed compute by these straggling workers. We harness the gradient accumulation method widely used in Large Language Models (LLMs) \citep{ott-etal-2018-scaling, roberta} to implement the method in a few lines of code on a relevant large-scale learning task.
The contributions of our work include: \begin{itemize}
\item \textit{DropCompute}: a novel, decentralized method to better handle heterogeneity or stragglers without additional hyper-parameters. \textit{DropCompute} is hardware and framework agnostic, runs on top of existing optimizers, and can also be combined with other methods that improve other aspects of robustness such as communication overhead.
\item A theoretical convergence proof of SGD with stochastic batch size, as in \textit{DropCompute}.
\item A theoretical runtime analysis on standard synchronous training and the proposed method, with an approximation of the expected speedup using \textit{DropCompute}.
\item Empirical evaluation of the proposed method on a relevant large scale task, using up to 200 accelerators connected with high bandwidth communication. \end{itemize}
\section{Related Work}
The challenge of training deep neural networks on a large scale has been extensively explored. With rapidly growing models and data sizes, numerous works tackled the weaknesses in synchronous DNN training on a large scale and suggested methods to alleviate these weaknesses.
\textbf{Redundancy methods.} This line of work addresses the straggling worker problem using a redundancy mechanism. Redundant workers or redundant data are used such that straggling workers will not slow down the entire system \citep{chen2016revisiting, bitar2020stochastic}. These methods provide better robustness to synchronous training, even in the event of a complete failure of a subset of the workers or considerable communication slowdown. However, the robustness is limited by the redundancy factor, and more generally, more compute resources are required and full utilization cannot be achieved. In addition, some coordination method is required to keep the training synchronous, i.e., keeping consensus between the model replicas of the workers. In particular, \citet{chen2016revisiting, bitar2020stochastic} use a centralized approach of a parameter server to determine which workers are left out at each iteration. Modern large-scale systems use decentralized variants of \textit{All-Reduce} \citep{baidu_ringreduce, patarasuk2009bandwidth}, so it is not trivial to determine which workers should be considered at each step, given that each worker can see a different subset of straggling workers. Moreover, combining redundancy with communication primitive collectives (e.g., \textit{All-Reduce}) requires adaptation to existing underlying frameworks \citep{sanders2019sequential}.
\textbf{Asynchronous optimization.} Another approach is introducing asynchrony to the optimization. Asynchronous training is inherently more scalable than synchronous training by being robust to all kinds of workers and communication faults. This includes periodic synchronization by exchanging parameters every $\tau$ optimization steps \citep{stich2018local, Lin2020Don't, wang2021cooperative, zhang2015deep, Wang2020SlowMo, li2020federated}, approximate distributed averaging where each worker communicates with a subset of workers each step \citep{jiang2017collaborative, lian2017can, assran2019stochastic, yang2020mitigating}, and many more. These works provide better scale-up properties and improve time performance. The main drawback is asynchronous optimization itself. In practice, the convergence is less stable, and more optimization steps are needed to generalize as well as synchronous training. In addition, hyperparameters should be chosen more precisely to guarantee convergence and generalization properties \citep{giladi2019stability, mitliagkas2016asynchrony}. Due to these issues, asynchronous methods are less commonly used on a large scale.
\textbf{Sparse and compressed communication.} Alternatively, several works addressed only the communication overhead. A common technique is to reduce the amount of data exchanged by the workers at each step. This can be done by gradient pruning \citep{xu2021deepreduce}, gradient compression \citep{1bitsgd, chen2020scalecom, 1bitadam} or low-rank approximation \citep{vogels2019powersgd}. These works reduce the communication overhead in a deterministic form while ignoring any compute variance across processing workers. This makes these works orthogonal to ours and potentially can be combined with our method.
\section {Reducing Compute Variance} \label{section:distributed_training} This paper proposes a method called \textit{DropCompute} that improves the robustness of synchronous training by reducing compute variance. First, we describe the vanilla synchronous training framework. Then, we introduce the proposed approach.
\subsection{Problem setup} \label{subsection:problem_setup}
We start with a model formulation for data-parallel synchronous SGD training with $N$ workers, where each worker holds a replica of the model parameters $\theta$. Given a dataset $\mathcal{D}$, we are interested in minimizing the empirical loss \begin{equation*}
\mathcal{L}(\mathcal{D},\theta)=\frac{1}{\abs{\mathcal{D}}}\sum_{z\in\mathcal{D}}\ell(z,\theta), \end{equation*} where $\ell(z,\theta)$ is the loss with respect to data-point $z$ and the model parameters $\theta$. At each step, the workers calculate gradients based on a local batch and then aggregate the gradients before taking an optimization step. An equivalent strategy would be that the workers aggregate the parameters after taking an optimization step based on the local gradients. The aggregation is done in a decentralized fashion, such as \textit{AllReduce}.
We also consider the use of gradient accumulation to increase the size of the local batch. Gradient accumulation breaks down each worker's local batch into $M$ micro-batches for computation. This enables reaching a large global batch size beyond hardware capacity. In each iteration $i$ and accumulation $m$, the gradients of the loss function with respect to the model parameters are computed, denoted by \begin{equation*}
g_{n}^{(m)}(\theta_i) = \nabla \mathcal{L}(\mathcal{D}_{i,n}^{(m)}, \theta_i)\,, \end{equation*}
where $\mathcal{D}_{i,n}^{(m)}\subseteq \mathcal{D}$ is the micro-batch $m$ of worker $n$, sampled without replacement from $\mathcal{D}$, and we assume a constant micro-batch size $|\mathcal{D}_{i,n}^{m}|$.
The gradients accumulated by worker $n$ at step $i$ are \begin{equation*}
g_{n}(\theta_i) = \frac{1}{M} \sum_{m=1}^{M} g_{n}^{(m)}(\theta_i) \,. \end{equation*} Finally, the workers aggregate and average the computed gradients to update the model parameters \begin{equation}
\theta_{i+1} = \theta_i - \eta g(\theta_i);\quad g(\theta_i) = \frac{1}{N} \sum_{n=1}^{N} g_{n}(\theta_i) \,,
\label{eq:theta} \end{equation} where $\eta$ is the learning rate. Equation \ref{eq:theta} requires all workers to receive all gradients before updating the parameters. This communication restriction is what makes the training synchronous and ensures the model parameters are the same for all workers. However, due to this restriction, the slowest worker at each step dictates the iteration time. More generally, any variation in computation time among workers will lead to workers with idle time. Therefore, to improve the efficiency and robustness of the synchronous training process, we need to reduce the computation time variance among workers (namely, the compute variance).
\subsection{Our method: \textit{DropCompute}} \label{sec:our_method} To mitigate the effect of compute variance and to improve the robustness of synchronous training, we propose a simple yet effective method called \textit{DropCompute}. \textit{DropCompute} reduces the compute variance by introducing a \textit{compute threshold} after which workers have to stop calculating local gradients and move directly to the communication phase, i.e., Equation \ref{eq:theta}.
In each iteration $i$, each worker $n$ measures the time while calculating its local batch, and compares it against a given threshold $\tau$, which is set as described in section \ref{section:automatic_threshold}. If that time exceeds the \textit{compute threshold}, the worker stops and sends the gradients it has calculated so far. The rest of the training remains unchanged, thus synchronous training is intact. The pseudo-code of \textit{DropCompute} is shown in Algorithm \ref{algorithm:drop_compute}.
This method maintains a decentralized approach while improving its robustness to compute variance and stragglers in particular. Since our method drops samples, the batch size is no longer deterministic, Specifically, the total batch size (the total number of samples used in one iteration by all workers) can be smaller than the maximal total batch size $b_{\max}=NM|\mathcal{D}_{i,n}^{(m)}|$. Nevertheless, we show both theoretically and empirically that this makes no difference to convergence or generalization. Next, we analyze this method in section \ref{section:analysis} and evaluate it in section \ref{section:experiments}.
\begin{algorithm}[h!]
\caption{DropCompute on worker $n$ at iteration $i$}
\label{algorithm:drop_compute}
\begin{algorithmic}[1]
\Input
\State model parameters $\theta_i$; total number of micro-batches $M$; compute threshold $\tau$
\State local batch data $\mathcal{D}_{i,n}=\{\mathcal{D}_{i,n}^{(1)}, \mathcal{D}_{i,n}^{(2)}, \cdots, \mathcal{D}_{i,n}^{(M)}\}$; step time $T_n$ to compute local batch $\mathcal{D}_{i,n}$
\EndInput
\State \textbf{Initialize step time:} $T_n=0$
\State \textbf{Initialize accumulated gradients:} $g_{n}(\theta_i)=0$
\InParallel
\NumFor{1}{$m=1, \dots, M$}
\State \hskip1.5em $g_{n}^{(m)}(\theta_i) = \nabla \mathcal{L}(\mathcal{D}_{i,n}^{(m)}, \theta_i)$ \Comment{Compute gradient}
\State \hskip1.5em $g_{n}(\theta_i) \leftarrow g_{n}(\theta_i) + g_{n}^{(m)}(\theta_i)/M$ \Comment{Accumulate gradients}
\EndNumFor
\State \textbf{(2) wait for $T_n>\tau$ and break for loop (1)}
\EndInParallel
\State \textbf{Output: }$g_{n}(\theta_i)$ \Comment{\textit{AllReduce}}
\end{algorithmic} \end{algorithm}
\section{Method Analysis} \label{section:analysis}
After establishing the notion of compute variance and the formulation of the proposed method, we analyze synchronous training and the potential value of the proposed method on the training time performance. To do so, we start with convergence guarantees when using \textit{DropCompute}. Then, we theoretically analyze the computation time with synchronous training and when using \textit{DropCompute}. Through this analysis, we estimate the potential speedup over vanilla synchronous training.
\subsection{Convergence analysis of \textit{DropCompute}} \label{subsec:convergence} Using \textit{DropCompute}, the batch size is no longer fixed, but a random variable, which can also potentially depends on the data samples. To the best of our knowledge, this setting is somewhat different from existing convergence guarantees. Therefore, in this section, we provide convergence guarantees for \textit{DropCompute} by analyzing the convergence of SGD with stochastic batch size.
\begin{assumption} \label{assumption:convergence}
Following the notations in section \ref{subsection:problem_setup}, consider a possibly non-convex smooth loss function $\mathcal{L}(\mathcal{D},\theta)$, with a global minimum $\theta^*\in{\mathbb R}^d$, and the following assumptions
\begin{enumerate}
\item \textbf{L-smooth}: All functions $\mathcal{L}(\cdot,\theta)$ are with L-Lipschitzian gradients.
\item \textbf{Unbiased estimation}: Each $\nabla\ell(z,\theta_i) ,z\in\mathcal{D}$ is an unbiased estimate of $\nabla \mathcal{L}(\mathcal{D},\theta_i)$ which is the true (full batch) gradient at $\theta_i$\footnote{This is easily satisfied when all workers can access all data.}. Namely,
$\forall i: \E[\nabla\ell(z,\theta_i)\vert \theta_i] = \nabla\mathcal{L}(\mathcal{D},\theta_i)\,.$
\item \textbf{Bounded variance}: The variance of the stochastic gradient is bounded by a constant $\sigma$,
$$\forall i: \E[\|\nabla\ell(z,\theta_i)-\nabla\mathcal{L}(\mathcal{D},\theta_i)\|^2\vert \theta_i] \leq \sigma^2\,.$$
\end{enumerate} \end{assumption}
\begin{theorem} \label{theorem:non_convex}
Under the above assumption, applying SGD with \textit{DropCompute} (Algorithm \ref{algorithm:drop_compute}), ensures
\begin{equation} \label{eq:NonCvxSGD_GUARANTEES_final}
\mathbb{E}\|\nabla\mathcal{L}(\mathcal{D},\bar{\theta})\|^2 \leq \frac{2L b_{\max} (\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*))}{K} + \frac{2\sigma\sqrt{L(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)) }}{\sqrt{K}}~,
\end{equation}
where $b_{\max}$ is the maximal total batch size, $K$ is the total number of samples that are used throughout the training, $\theta_1$ is the initial model, and $\bar{\theta}$ is a random sample of $\theta_i$ from the trajectory obtained by Algorithm \ref{algorithm:drop_compute}, where $i$ is selected with probability proportional to the total batch size at iteration $i$. The expectation is with respect to the randomization introduced due to sampling from $\mathcal{D}$ throughout the optimization process and with respect to choosing $\bar{\theta}$. \end{theorem}
The bound in Theorem~\ref{theorem:non_convex} is in a similar spirit to existing fixed batch-size guarantees~\citep{dekel2012optimal}. The second term in the bound does not degrade with the batch sizes, and behaves like $O(1/\sqrt{K})$, while the first term behaves like $O(b_{\max} /{K})$. This implies that as long as $b_{\max} \leq O(\sqrt{K})$, the second term is dominant and we are in the regime of linear speedup. Thus, we show that one can attain a linear speedup despite using changing batch sizes, as long as the maximal batch size is bounded.
Similarly, in Theorem \ref{theorem:convex_case} in appendix \ref{appendix:convex_proof} we show that the loss itself converges, in the convex case along with a proof. The proof of Theorem \ref{theorem:non_convex} is provided in appendix \ref{appendix:non_convex_proof}. Lastly, we discuss the impact of the stochastic batch size on generalization in appendix \ref{app:generalization_discussion}.
\subsection{Iteration time in standard synchronous distributed training} \label{sec:analysis_iteration_time}
We start with finding a closed-form expression for the cumulative distribution function (CDF) of the iteration time, denoted as $T$, defined as \begin{equation*}
T=\max(T_1,T_2,...,T_N)\,, \end{equation*} where $T_n$ represents the time taken by worker $n$ to compute its local batch, which follows some cumulative probability function $F_{T_n}(x)=\mathbb{P}(T_n<x),$ which we assume is independent for each worker. Let $F_T(x)$ represent the cumulative distribution function and $f_T(x)$ represent the probability density function of the maximum iteration time $T$. The relation between $F_T(x)$ and $F_{T_n}(x)$ is
\begin{equation*}
F_T(x) = \mathbb{P}\left(\max \left(T_1, \ldots, T_N\right) \leq x\right) =\prod_{n=1}^N F_{T_n}(x)\,. \end{equation*}
Differentiating with respect to $x$ and applying the chain rule gives: \begin{equation*}
f_T(x)=\frac{d F_T}{d x}(x)=\sum_{n=1}^N f_{T_n}(x)\prod_{n'\neq n}^{N}F_{T_{n'}}(x) \,. \end{equation*}
In the special case where all of the workers' iteration time distributions are identically and independently distributed (i.i.d.), this reduces to the well-known formula: \begin{equation}
\label{pdf}
f_T(x)=N \cdot f_{T_n}(x) \cdot F_{T_n}(x)^{N-1} \end{equation}
If the iteration time of each worker is distributed normally ($\sim \mathcal{N}(\mu, \sigma^2)$), the expected value of $T$ can be approximated as shown by \citet{bailey2014pseudomathematics}: \begin{equation} \mathbb{E}(T) \approx \sigma\cdot \left( (1-\gamma)\cdot \Phi^{-1}\left(1-\frac{1}{N} \right) + \gamma\cdot \Phi^{-1}\left(1-\frac{1}{e\cdot N}\right)\right) + \mu \label{eq:estimate_T} \end{equation} where $\Phi$ is the CDF of the standard normal distribution, and $\gamma$ is the euler-mascheroni constant. When the number of micro-batches $M\gg1$, we can make a similar approximation under Central Limit Theorem (CLT) conditions. More details are in appendix \ref{appendix:speedup_analysis}.
\subsection{Iteration time and number of micro-batches with \textit{DropCompute}}
When using \textit{DropCompute} with a constant threshold $\tau$, each worker is preempted at $\tilde{T}_n=\min\left\{\tau,T_n\right\}$ and joins the \textit{AllReduce}. Therefore, the total iteration time with \textit{DropCompute} is $$ \tilde{T} + T^c = \min\left\{\tau,T\right\}+T^c \, ,$$ where $T^c$ is a serial latency present in each iteration, which includes the \textit{AllReduce} step. This upper limit serves to clip extreme values of $T_n$, effectively constraining the range of potential outcomes for $\tilde{T}$. As a result, the compute time variability decreases, leading to a narrower distribution and enhanced compute efficiency. These effects are illustrated in Figure \ref{fig:experiement_vs_theory}.
As a consequence of preempting each worker at $\tilde{T}_n$, the number of micro-batches computed in each step varies. Denote as $t_n^{(m)}$ the compute latency of a single micro-batch $m$ for worker $n$, and $T_n^{(m)}=\sum_{j=1}^mt_n^{(j)}$. We can define the average number of micro-batches computed by each worker before reaching threshold $\tau$ as $$\tilde{M}(\tau)=\frac{1}{N}\sum_{n=1}^N\sum_{m=1}^M \left\{
\begin{array}{lr}
1, & \text{if } T_n^{(m)}<\tau\\
0, & \text{otherwise }
\end{array}\right\}~.$$ Under CLT conditions, the expected value for $\tilde{M}(\tau)$ can be approximated in a closed form: \begin{equation}
\mathbb{E}[\tilde{M}(\tau)] \approx \sum_{m=1}^M \Phi\left(\frac{\tau - m\cdot \mu}{\sqrt{m\cdot \sigma^2}} \right)
\label{eq:estimate_M} \end{equation} where $\mu, \sigma^2$ are the mean and variance for a single micro-batch $t_n^{(m)}$ compute latency, and $\Phi$ is the CDF of the standard normal distribution. This approximation closely fits the real value of $\tilde{M}$ and can be used to analyze the expected gain from \textit{DropCompute}. More details in appendix \ref{appendix:speedup_analysis}. \begin{figure}\label{fig:experiement_vs_theory}
\end{figure} \subsection{Choosing the threshold} \label{section:automatic_threshold} The throughput of the system can be seen as the number of micro-batches computed per second. For $N$ workers, this can be written as $NM/(T+T^c)$. To evaluate the effectivness of \textit{DropCompute}, we consider the difference in throughput between the baseline and when using \textit{DropCompute}. Doing so, we can define the effective speedup for $\tau$ as: \begin{equation}
S_{\mathrm{eff}}(\tau)= \frac{\text{DropCompute Throughput}}{\text{Baseline Throughput}} = \frac{N\tilde{M}(\tau)/ (\min\left\{\tau,T\right\}+T^c)}{NM/(T+T^c)} = \frac{\tilde{M}(\tau)(T+T^c)}{M(\min\left\{\tau,T\right\}+T^c)}
\label{eq:auto_eff} \end{equation} Given the statistical characteristics of the training setting, it is possible to estimate analytically the expected value of the effective speedup $\mathbb{E}[S_\mathrm{eff}]$ by using Equations \ref{eq:estimate_M} and \ref{eq:estimate_T}.
As shown in \ref{fig:estimate_s_eff_b}, Equation \ref{eq:estimate_T} is less accurate when samples deviate from a normal distribution. \begin{figure}\label{fig:estimate_s_eff_a}
\label{fig:estimate_s_eff_b}
\label{threshold_choosing}
\label{fig:estimate_s_eff}
\end{figure} To find the optimal compute threshold, we synchronize the empirical distribution of micro-batch compute latency between all workers after a few iterations. Given this distribution, we find $T$ and $\tilde{M}(\tau)$, and search in a decentralized way for the threshold $\tau^*$ that maximizes the effective speedup $S_{\mathrm{eff}}(\tau)$ defined in Equation \ref{eq:auto_eff}.
Lowering the threshold leads to reduced compute time but higher compute drop rates. Figure \ref{threshold_choosing} highlights this trade-off and the optimal $\tau^*$ is marked.
\section{Experiments} \label{section:experiments} To be useful, \textit{DropCompute} must possess two properties. First, it should not compromise the accuracy of the trained model. This property is put to test in section \ref{sec:generalization_perf} where we fully train BERT-Large and ResNet-50 \citep{devlin2018bert, resnet}, each on a different task, with different drop rates to compare accuracy. Second, \textit{DropCompute} should maintain a high level of runtime performance, especially when compute variance or straggling workers exist and vanilla synchronous training time deteriorates. Section \ref{sec:runtime_perf} tests runtime performance of \textit{DropCompute} by training a 1.5 billion parameter language model, BERT1.5B \citep{devlin2018bert} with additive noise to the compute time of each worker.
\textbf{Experimental setup.} The analysis of all BERT models is performed on the same dataset as \citet{devlin2018bert}, which is a concatenation of Wikipedia and BooksCorpus with 2.5B and 800M words respectively. The finetuning of the pretrained models is performed on SQuAD-v1 \citep{rajpurkar-etal-2016-squad}. We verify the generality of \textit{DropCompute} by additional evaluation of a ResNet-50 model for image classification on ImageNet \citep{imagenet_cvpr09}. The experiments depicted in section \ref{sec:runtime_perf} and section \ref{sec:generalization_perf} are executed on Habana Gaudi-1 and Gaudi-2 accelerators, respectively, with high performance network \citep{habana2020-whitepaper}.
\subsection{Generalization performance} \label{sec:generalization_perf} The sole difference in the optimization when \textit{DropCompute} is applied is that the batch size is not deterministic, but stochastic, as explained in section \ref{sec:our_method}. To complement theorem \ref{theorem:non_convex}, we examine the generalization performance achieved with a stochastic batch size on two popular tasks.
\textbf{Image classification.} To evaluate the generality of stochastic batch size and \textit{DropCompute} in particular, we evaluate the Top-1 accuracy of a ResNet-50 model on the Imagenet dataset using our method. Since it is not common to use gradient accumulation in large scale training of this task, we simulate the drops such that each worker randomly drops its local batch, so the total batch size is stochastic. This simulated environment enables us to examine the extent of drop rate we can use without compromising accuracy. Figure \ref{fig:resnet50_accuracy} in appendix \ref{appendix:image_classification} shows that up to $10\%$ drop rate, which is more than what \textit{DropCompute} operates on, there is a negligible deterioration in accuracy.
\begin{wraptable}{r}{0.4\textwidth}
\centering
\caption{\textbf{Maintaining accuracy of BERT-Large pretraining.} Fine-tuning results on SqUAD v1.1, where The F1 score is obtained by the pretrained model, each with a different drop rate during pretraining.}
\begin{tabular}{| c c |}
\hline
\% Drop rate & F1 score on dev set \\ [0.5ex]
\hline\hline
0\% & 91.32 $\pm$ 0.15 \\
\hline
2.5-3\% & 91.34 $\pm$ 0.04 \\
\hline
5.5-6\% & 91.44 $\pm$ 0.02 \\
\hline
10-11\% & 91.13 $\pm$ 0.02 \\
\hline
\end{tabular}
\label{table:accuracy}
\end{wraptable}
\textbf{Large language model.} Training LLMs is resource intensive, typically using large batch sizes, which makes \textit{DropCompute} appealing. We evaluate \textit{DropCompute} method on this task by fully pretraining BERT-Large model several times, each with a different drop rate. We follow the optimization regime described in \citet{you2019large} with a batch size of 64K for phase-1 and 32K for phase-2 (more details are provided in appendix \ref{appendix:generalization}). Each of the pretrained models is fine-tuned on the SQuAD task 3 times with different initializations. Fine-tuning is performed without drops, as it is not a large scale resource consuming task. Table \ref{table:accuracy} shows the average accuracy ($\pm$ standard deviation) obtained for each drop rate. As shown, \textit{DropCompute} at drop rates of up to $10\%$ have negligible accuracy difference. Higher values measured up to $20\%$ of dropped gradients provide acceleration with a small yet discernible hit on accuracy. We note that these results are for a fixed budget of steps. In the presence of compute variance, the effective speedup indicates that additional steps can be executed while still maintaining competitive runtime performance. This notion is demonstrated in section \ref{sec:runtime_perf}.
\subsection{Runtime performance} \label{sec:runtime_perf} The main purpose of our proposed method is to maintain runtime performance when compute variance is present. We examine this by measuring the speedup of \textit{DropCompute} over standard synchronous training in several settings. First, we measure the potential speedup for different drop rates and training settings by post analysis of synchronous training without drops. In addition, we introduce compute variance by training with additive noise, and measure actual speedups using \textit{DropCompute}. The experiments in this section are performed on BERT1.5B. Details are provided in appendix \ref{appendix:runtime_performace}.
\textbf{Training with different number of workers and micro-batches.} We evaluate the potential speedup of \textit{DropCompute} on several training settings with natural heterogeneity and no drops. For each setting, we post analyze what would have been the speedup for different drop rates. As can be seen in Figure \ref{fig:speedup}, \textit{DropCompute} exhibits increasing benefits with a growing number of workers and compute requirements. However, there are diminishing returns in terms of speedup with more accumulations. This could possibly be explained by the amortization time of a large number of micro-batches.
\begin{figure}\label{fig:speedup}
\end{figure}
\textbf{Simulated delay environment.} Although \textit{DropCompute} may have value when the workers' compute latency variance is low, its significance becomes crucial when the workers' compute latency exhibits high variability. To evaluate our method, we introduce a delay environment where random latency is added to each micro-batch computation. This additive noise follows a bounded log-normal distribution. Detailed information and motivation regarding the additive noise are in appendix \ref{appendix:runtime_performace}. The experiments are executed with 12 gradient accumulations and a local batch size of 192. In Figure \ref{fig:abstract}, the negative impact of compute variance on scalability is demonstrated and mitigated using \textit{DropCompute}. The results in Figure \ref{fig:abstract} also correspond to section \ref{section:analysis} and Equation \ref{eq:eff_speedup}, where a theoretical extrapolation follows the same trend line. When utilizing \textit{DropCompute} in this setup, achieving the same training loss as the baseline might requires additional training steps, however, it leads to a notable reduction in overall training time. Figure \ref{fig:time_to_train} demonstrates it in a training session with 64 workers, where approximately 3\% more steps is needed to reach the same loss, in 13\% less time.
\begin{figure}\label{fig:time_to_train}
\end{figure}
\section{Discussion} \textbf{Summary.} Efficient scalable systems are a key component to enable the continued development of deep learning models. To this day, state-of-the-art models rely on synchronous distributed optimization. The challenge to maintain synchronous training as an efficient solution grows larger with the quickly growing model sizes and data. Therefore, improving the robustness and scalability of distributed synchronous training is an important endeavor. This paper tackles the challenge of maintaining synchronous training scalable in the face of compute variance. We propose \textit{DropCompute} to improve the robustness of synchronous training. Workers drop their remaining compute when they reach a compute threshold, determined by exchanging and analyzing the compute latency distribution. We find that for a small percentage of dropped data, a much larger percentage of time can be saved, depending on the compute latency distribution of the workers. In addition, we provide theoretical convergence guarantees and runtime predictions.
\textbf{Limitations.} While \textit{DropCompute} is simple and straightforward, it deals with system efficiency, and as such, the user-level implementation provided is not optimal. Most of all, the provided implementation is limited by using gradient accumulations and integrating compute timeout in between accumulations. This can be solved by integrating this method in low-level accelerator code to enable execution more efficiently.
\textbf{Future directions.} \textit{DropCompute} is described and analyzed in this paper as a method built on top of synchronous training. However, this method can be integrated with other possibly asynchronous methods such as periodic synchronization. In appendix \ref{app:local_sgd}, we implement \textit{DropCompute} on top of Local-SGD \citep{Lin2020Don't} and show that \textit{DropCompute} can also improve the robustness of Local-SGD to stragglers.
\section*{Acknowledgments} We thank Itay Hubara for technical advising and valuable comments on the manuscript. The research of DS was Funded by the European Union (ERC, A-B-C-Deep, 101039436). Views and opinions expressed are however those of the author only and do not necessarily reflect those of the European Union or the European Research Council Executive Agency (ERCEA). Neither the European Union nor the granting authority can be held responsible for them. DS also acknowledges the support of Schmidt Career Advancement Chair in AI.
\appendix
\part*{Appendix}
\section{Experiments} \label{appendix:experiments}
\subsection{Runtime performance experiments} \label{appendix:runtime_performace} In this section, we provide details for the experiments of section \ref{sec:runtime_perf}.
\textbf{Experiment details.} As mentioned in the paper in section \ref{sec:runtime_perf}, we pre-train BERT1.5B following \citet{habana-deepspeed-bert}. The experiments in this section use up to 200 Gaudi accelerators with high bandwidth inter-connectivity. The training is done with a maximum input sequence length of 128 and 80 maximum predictions per sequence. The training regime consists of a local batch size of 196, 12 gradient accumulations, LANS optimizer \citep{zheng2020accelerated}, and a learning rate of 0.0015. Due to the large capacity of the model, we used ZeRO optimizer stage 1 to fit the model in memory \citep{rajbhandari2020zero}.
\textbf{Simulated delay.} Many frameworks use padding to allow for constant input length which improves hardware efficiency \citep{kosec2021packing}. However, some learning tasks inherently involve dynamic shapes, such as translation \citep{ott-etal-2018-scaling} and multi-task sequences \citep{2020t5}. These use cases motivate us to explore scenarios of dynamic length via simulation. To demonstrate the value of \textit{DropCompute} in dealing with compute variance we added to each micro-batch compute time an additional random waiting time. The additive noise is based on a Log-normal distribution since it is typical for user post lengths in internet discussions \citep{sobkowicz2013lognormal}, which are used as training data in recent language models \citep{radford2019language}. To make this setting more realistic, we scale down and bound the noise so that each accumulation takes $\times1.5$ longer on average, and, in extreme cases, can take up to 6 times longer. This allows us to simulate stragglers and high compute variance while keeping a conservative limit on iteration time. Thus, the additive noise takes the form of $$\epsilon=\min\left(\frac{1}{\alpha}Z, \beta\right) ,\;\;\;\;\;\;\;\;\; Z\sim \mathrm{LogNormal}(4,1) \,.$$ This noise was added to each accumulation $$t_n^{(m)}\gets t_n^{(m)} + \mu\cdot \epsilon \,,$$ where $\mu$ is the mean value for $t_n^{(m)}$, $\alpha=2\exp(4.5)$ and $\beta=5.5$ are the scaling and bounding constants, and the log-normal parameters (4,1) fit user post lengths, as seen in \citet{sobkowicz2013lognormal}. As illustrated in Figure \ref{fig:epsilon_dist}, the noise distribution leads to each micro-batch latency increased by up to $6\mu$, while the majority of accumulations have low latency. Further analysis on the affect of noise properties is discussed in \ref{appendix:noise_analysis}.
\begin{figure}
\caption{\textbf{The latency distribution in a simulated delay environment.} (left) The distribution of the additive noise $\epsilon$, added to each accumulation. (right) The distribution for iteration time $T_n$, with 12 accumulations, each with added noise, in BERT1.5B training.}
\label{fig:epsilon_dist}
\end{figure}
\subsection{Generalization experiments} \label{appendix:generalization} In this section, we provide details for the experiments of section \ref{sec:generalization_perf}.
\subsubsection{Large language models}
Here we provide more details about how the LLM experiment was executed as well as additional graphs related to the LLM experiment described in section \ref{sec:generalization_perf}.
\textbf{Experiment details.} As mentioned in the paper, in section\ref{sec:generalization_perf} we follow \citet{you2019large} optimization regime with LAMB optimizer. Specifically, for phase-1 where the sequence length is 128 tokens per sample, we use a batch size of 64K, the learning rate is 0.006, the warmup ratio is 0.2843, and the steps number is 7038. For phase-2 where the sequence length is 512, we use a batch size of 32K, the learning rate is 0.004, the warmup ratio is 0.128 and the steps number is 1563. The experiments were executed on 64 workers.
\textbf{Batch size distribution.} As explained in section \ref{sec:generalization_perf} we fully pretrain a BERT-Large model with \textit{DropCompute} several times, each with a different drop rate. Figure \ref{fig:batch_dist} shows the empirical batch distribution of each of the drop rates in phase-1.
\begin{figure}
\caption{\textbf{Batch size distribution.} BERT-Large phase-1 pretraining batch size distribution when using \textit{DropCompute} and drop rate of (a) 2.5\% , (b) 5.5\%,
and (c) 11.5\%}
\label{fig:batch_dist}
\end{figure}
\textbf{Convergence loss.} In addition to the results depicted in Table \ref{table:accuracy}, we show the convergence of the training loss with the different drop rates in Figure \ref{fig:loss_1563}.
\begin{figure}
\caption{\textbf{Train loss convergence.} BERT-Large phase-1 (left) and phase-2 (right) pretraining train loss for different drop rates.}
\label{fig:loss_1563}
\end{figure}
\subsubsection{Image classification} \label{appendix:image_classification} This section provides the details of the image classification experiment described in section \ref{sec:generalization_perf} as well as Figure \ref{fig:resnet50_accuracy} which is referenced from the paper.
\textbf{Experiment details.} To simulate \textit{DropCompute}, at each training step, the gradients of each worker are set to zero with a probability of $P_{\text {drop }}$. We repeat each training process 3 times with different initializations. To examine the generalization of \textit{DropCompute} over different optimizers, we implement our method on two popular training regimes of ResNet50. First, we follow the optimization regime described in \citet{goyal2017accurate} that uses SGD with 32 workers and a global batch size of 4096. Second, we follow \citet{mlperf} that uses LARS \citep{you2017large} with 8 workers and a global batch size of 2048.
\begin{figure}
\caption{\textbf{Generalization over varying drop rates.} Top-1 validation accuracy of ResNet50 trained on ImageNet with varying simulated drop rates. The dashed line is the baseline accuracy without drops. The solid line is the average over 3 runs, and the blue area is the standard deviation. (left) Training regime with SGD \citep{goyal2017accurate}. (right) Training regime with LARS \citep{mlperf}. Up to a 10\% drop rate, there is a negligible accuracy deterioration.}
\label{fig:resnet50_accuracy}
\end{figure}
\textbf{Learning rate correction.} Previous works showed that the learning rate should be scaled with respect to the batch size \citep{hoffer2017train, goyal2017accurate}. With a stochastic batch size and specifically \textit{DropCompute}, it is possible that a learning rate correction should be considered to maintain accuracy with the same number of steps. We examine such corrections when training with stochastic batch size. First, we decrease the learning rate by a constant factor, equal to the average drop rate. Specifically, for an average drop rate $P_{\text {drop}}\in[0,1]$ we multiply the learning rate by $(1-P_{\text {drop}})$. A different correction we consider is a stochastic correction, such that in each step we divide the gradients by the computed batch size, instead of the original batch size. This result in a different normalization in each step depending on the actual dropped samples. We note that for the latter, the workers have to synchronize the computed batch of each worker at each step. This is generally can be done during the \textit{AllReduce}, with negligible overhead. We repeat the training of ResNet50 on ImageNet as described in \citet{goyal2017accurate} to evaluate the generalization without correction and with the two suggested corrections. We use 128 workers, batch size 8192, and use ghost batch norm (GBN) \citep{hoffer2017train} to match batch normalization of 32 samples and restore the results in \citet{goyal2017accurate}. As can be seen in Figure \ref{fig:resnet50_lr_correction}, for low drop rates, there is no superior correction method, and no correction generally achieves the same generalization. Yet, it is possible that a learning rate correction could potentially improve generalization on a different task or with a different optimizer.
\begin{figure}
\caption{\textbf{Learning rate correction is not necessary for low drop rates.} Top-1 validation accuracy of ResNet50 trained on ImageNet with varying simulated drop rates, using the learning-rate correction methods described in \ref{appendix:image_classification}. The dashed line is the baseline accuracy without drops. Each solid line is the average over 3 runs, and the area around it is the standard deviation. Up to a 10\% drop rate, there is a negligible accuracy deterioration regardless of the correction applied.}
\label{fig:resnet50_lr_correction}
\end{figure}
\subsection{Local-SGD} \label{app:local_sgd} Periodic synchronization methods, such as Local-SGD, provide better scalability properties than synchronous methods. By exchanging parameters less frequently, communication overhead is mitigated. For compute variance and straggling workers in particular, the robustness of these methods greatly depends on the distribution of the compute time between workers. For example, when straggling workers appear randomly with homogeneous distribution, Local-SGD can mitigate the straggling workers slowdowns to some extent; this is because of the amortization effect in synchronizing periodically once every several steps. On the other hand, if straggling workers appear from a small set of workers such as a single server, a realistic scenario, Local-SGD acts more closely to synchronous training as the worst-case scenario is when a single worker always straggling behind. \textit{DropCompute} can be easily integrated with Local-SGD by leveraging periodic synchronization instead of gradient accumulations. We implement \textit{DropCompute} on top of Local-SGD by comparing the compute time with a threshold at each local step. We show that when stragglers are apparent, \textit{DropCompute} can improve the robustness of Local-SGD. We randomly slow down workers to simulate stragglers in two scenarios as described in Figure \ref{fig:local_sgd}. The experiment setting is 32 workers training on ResNet50 and ImageNet. At each local step, each worker is selected to be straggler with a $4\%$ chance. This way, there is at least 1 straggler for each local step on average. We measure relative speedup compared to synchronous training in terms of step time, both for Local-SGD and with \textit{DropCompute} on top of Local-SGD. As can be seen, with \textit{DropCompute} (set to $~6.2\%$ drop rate in this experiment) we improve the robustness of Local-SGD. \ \begin{figure}\label{fig:local_sgd}
\end{figure}
\section{Analyzing the effective speedup using \textit{DropCompute}} In this section, we provide more details on the process of choosing the threshold $\tau^*$ that will maximize the effective speedup. We begin by giving technical details on the process used during training, given samples drawn from the empirical distribution of the compute latency. Next, we continue to explore and establish the analytic connection between the latency statistics and the effective speedup.
\subsection{Automatic selection of the drop threshold} \label{appendix:auto_threshold}
In Algorithm \ref{algorithm:optimal_threshold} below we present the algorithm used for automatic selection of the drop threshold $\tau$. In this algorithm, $t_{i,n}^{(m)}$ is the time it takes worker $n$ to process micro-batch $m$ at step $i$ and $T_i^c$ is the time spent on communication for at step $i$. This data is measured by each worker, and then synchronized between all workers after $I$ iterations. After the synchronization step each worker will have the same drop threshold $\tau$, which depends on both his own speed and the compute latency of the other workers. Since $T_i^c$ is used in the normalization of the effective speedup, the chosen threshold takes into account both compute and communication time.
\begin{algorithm}[h!]
\caption{Automatic choice for the optimal threshold $\tau$}
\label{algorithm:optimal_threshold}
\begin{algorithmic}
\Statex \textbf{Input}:
\State number of workers $N$
\State number of iterations $I$
\State number of micro-batches per step $M$
\State micro-batch time samples $\{\{t\}_{i,n}^{(m)}\}^{i\in[1:I], n\in[1:N], m\in[1:M]}$
\State communication time for each iteration $T_i^c$
\State potential thresholds $[\tau_0, \tau_1, ...]$
\For{$\tau \in [\tau_0,\tau_1,...]$}
\For{$i = 1$ to $I$}
\State Initialize completed micro-batch count: $\tilde{M}_i(\tau)=0$
\State Initialize compute step latency for all workers: $T_i=0$
\For{$n = 1$ to $N$}
\State Initialize single worker step compute latency: $T_{i,n}=0$
\For{$m = 1$ to $M$}
\State $T_{i,n} \gets T_{i,n} + t_{i,n}^{(m)} $
\State $\tilde{M}_i(\tau) \gets \tilde{M}_i(\tau) + \frac{1}{N}\cdot\left\{
\begin{array}{lr}
1, & \text{if } T_{i,n}<\tau\\
0, & \text{otherwise }
\end{array}\right\}$
\EndFor
\State $T_i \gets \text{max}(T_i, T_{i,n})$ \Comment{compute time of the slowest worker at step (i)}
\EndFor
\State $S_i(\tau) = \frac{T_i + T_i^c}{\min(\tau,T_i)+T_i^c}\cdot \frac{\tilde{M}_i(\tau)}{M}$ \Comment{Effective speedup for step (i)}
\EndFor
\State $S_{\mathrm{eff}}(\tau) = \frac{1}{I}\sum_{i=1}^{I}{S_i(\tau)}$ \Comment{Mean speedup for threshold ($\tau$)}
\EndFor
\State $\tau^* \gets \text{argmax}_\tau\left(S_{\mathrm{eff}}(\tau) \right)$
\end{algorithmic} \end{algorithm}
\subsection{DropCompute speedup analytic analysis} \label{appendix:speedup_analysis} In this section we further explore the relation between the compute latency distribution and the effective speedup. We will derive a closed-form representation of the effective speedup by making certain assumptions. First we assume that \begin{assumption} \label{assu: iid} $t_n^{(m)}$ is i.i.d with finite mean $\mu$ and finite variance $\sigma^2$.\end{assumption} Note that in the assumption above, for simplicity of presentation we assumed all workers as identical and so $\mu$ and $\sigma^2$ are identical. However, it is possible to derive similar properties with nonidentical workers, each with their own $\mu_n$, $\sigma_n$. Next, denote the time for completion of micro-batch $m$ as $T_n^{(m)}=\sum_{j=1}^m t_n^{(j)}$. Then, we assume \begin{assumption} \label{assume:gaussian} $T_n^{(m)}\text{ is Gaussian }\sim\mathcal{N}(m\mu,m\sigma^2)$ for $m>\sqrt{M}\,.$\end{assumption} This assumption holds in the limit $M\to\infty$ given Assumption \ref{assu: iid}, from the Central Limit Theorem (CLT). Lastly, denoting $\tau$ as the threshold used, we assume \begin{assumption} \label{assume:tau_limit} $\tau>\frac{M\mu}{2}\,.$\end{assumption} This bound can be considered as the minimum threshold allowed, in order for \textit{DropCompute} to be effective. Taking a lower threshold will result in unacceptable high drop rate.
Using these assumptions we first derive analytical expression for the iteration time $\mathbb{E}[T]$ and the mean completed number of gradient accumulations $\mathbb{E}[\tilde{M}]$ with \textit{DropCompute}. Then, we combine these expressions to obtain an expression for the mean effective speed $\mathbb{E}[S_{\mathrm{eff}}]$.
Figure \ref{fig:estimate_s_eff_b} shows an example of how close is the derived expression of $\mathbb{E}[S_{\mathrm{eff}}]$ to the value calculated by using the algorithm described in section \ref{appendix:auto_threshold}. The `analytical' curve is slightly off due to the inaccuracy of the Gaussian Assumption \ref{assume:gaussian} in calculating $\mathbb{E}[T]$, as we discuss below. We therefore added another curve `analytical given $\mathbb{E}(T)$', which uses same the derived expression for $\mathbb{E}[S_{\mathrm{eff}}]$ but replacing value of $\mathbb{E}[T]$, with the empiric mean: $\overline{T}=\frac{1}{I}\sum_{i=1}^I \max_n\left\{T_{n,i}^{(M)} \right\}$ where: $i\in[1:I]$ are the iterations measured.
\textbf{Iteration time.} as written in section \ref{sec:analysis_iteration_time}, the iteration time for all workers is $$T=\max_n \left(T_n^{(M)}\right) = T^c + \max_n \left( \sum_{m=1}^M t_n^{(m)} \right)\,.$$ When $T_n^{(M)}\sim\mathcal{N}(M\mu, M\sigma^2)\,.$ the expected value of $T$ can be approximated as \citep{bailey2014pseudomathematics}: \begin{equation} \label{eq:approximate_t}
\mathbb{E}[T] \approx \sqrt{M\sigma^2}\cdot\left( (1-\gamma)\cdot \Phi^{-1}\left(1-\frac{1}{N} \right) + \gamma\cdot \Phi^{-1}\left(1-\frac{1}{e\cdot N} \right)\right) + M\mu + T^c\,. \end{equation} It is worth noting that the distribution of $T$ is mostly affected by the tail of distribution of $T_n=T_n^{(M)}$ (as a consequence of Equation \ref{pdf}). Therefore, in practice the Gaussian Assumption \ref{assume:gaussian}, and therefore Equation \ref{eq:approximate_t}, can be inaccurate, especially for small $M$ and large $N$. It is therefore more accurate to use the real value of $\mathbb{E}[T]$, measured without \textit{DropCompute}, to estimate the potential effective speedup. An example for the inaccuracy of this approximation can be seen in Figure \ref{fig:estimate_s_eff_b}, when $T_n^{(M)}$ does not follow a normal distribution.
\textbf{Completed micro-batches.} The average number of micro-batch computed by a single worker $n$ when using \textit{DropCompute} with threshold $\tau$, is: $$\tilde{M}(\tau)=\frac{1}{N}\sum_{n=1}^N\sum_{m=1}^M \left\{
\begin{array}{lr}
1, & \text{if } T_n^{(m)}<\tau\\
0, & \text{otherwise }
\end{array}\right\} \,.$$ Its expected value can be written as: $$ \mathbb{E}[\tilde{M}(\tau)]=\sum_{m=1}^M P\left(T_n^{(m)}<\tau\right) \,.$$
In order to use assumption \ref{assume:gaussian}, we can split $\tilde{M}$ into 2 sums and derive a closed-form formula for the expected value: \begin{equation} \label{eq:tilde_m_split}
\tilde{M}(\tau)=\sum_{m=1}^{ \lfloor{{\sqrt{M}}}\rfloor} P\left(T_n^{(m)}<\tau\right) + \sum_{m=\lceil{{\sqrt{M}}}\rceil}^M P\left(T_n^{(m)}<\tau\right)\,. \end{equation} For the right term we can use assumption \ref{assume:gaussian} so that $P(T_n^{(m)}<\tau)=\Phi\left(\frac{\tau - m\cdot \mu}{\sqrt{m\cdot \sigma^2}} \right)$. For the left term, when $m<\sqrt{M}$ we use Markov inequality and assumption \ref{assume:tau_limit} to show that $$ 0\leq P(T_n^{(m)}>\tau)\leq \frac{\mathbb{E}[T_n^{(m)}]}{\tau}=\frac{m\mu}{\tau}\leq \frac{2m}{M}\,.$$
In other words, when using \textit{DropCompute} with low drop rates, $P(T_n^{(m)}<\tau)$ is very high for $m<\sqrt{M}$. The Gaussian approximation for $m<\sqrt{M}$ diminishes exponentially when increasing $M$, as seen by applying Chernoff bound: $$ 0\leq P(Z^{(m)}>\tau)\leq e^{-\frac{(\tau-m\mu)^2}{2m\sigma^2}}\leq e^{-\frac{(M/2-m)^2\mu^2}{2m\sigma^2}}$$ where $Z^{(m)}\sim\mathcal{N}(m\mu, m\sigma^2)$. Therefore, the error resulting in replacing $T_n^{(m)}$ with a Gaussian approximation, is bounded: $$ P(T_n^{(m)}<\tau) - P(Z^{(m)}<\tau) = P(Z^{(m)}>\tau) -P(T_n^{(m)}>\tau) $$ $$\Downarrow$$ $$ -\frac{2m}{M} \leq -P(T_n^{(m)}>\tau) \leq P(T_n^{(m)}<\tau) - P(Z^{(m)}<\tau) \leq P(Z^{(m)}>\tau)\leq e^{-\frac{(M/2-m)^2\mu^2}{2m\sigma^2}}$$ $$\Downarrow$$ $$ -\frac{2m}{M} + P(Z^{(m)}<\tau) \leq P(T_n^{(m)}<\tau) \leq P(Z^{(m)}<\tau) + e^{-\frac{(M/2-m)^2\mu^2}{2m\sigma^2}}$$ Plugging these inequalities into the left term in equation \ref{eq:tilde_m_split} gives us:
\begin{equation} \label{eq:tilde_m_split_ineq} \begin{split} \sum_{m=1}^{ \lfloor{{\sqrt{M}}}\rfloor} P\left(T_n^{(m)}<\tau\right) \leq& \sum_{m=1}^{ \lfloor{{\sqrt{M}}}\rfloor} \left(P\left(Z^{(m)}<\tau\right)+e^{-\frac{(M/2-m)^2\mu^2}{2m\sigma^2}} \right) \\ \\ \sum_{m=1}^{ \lfloor{{\sqrt{M}}}\rfloor} P\left(T_n^{(m)}<\tau\right) \geq& \sum_{m=1}^{ \lfloor{{\sqrt{M}}}\rfloor}\left( P\left(Z^{(m)}<\tau\right) -\frac{2m}{M}\right) = O(1) +\sum_{m=1}^{ \lfloor{{\sqrt{M}}}\rfloor} P\left( Z^{(m)}<\tau\right) \end{split} \end{equation}
Combining equations \ref{eq:tilde_m_split},\ref{eq:tilde_m_split_ineq} we can write the expected value as \begin{equation} \label{eq:approximate_m} \mathbb{E}[\tilde{M}(\tau)]=\sum_{m=1}^M P\left(T_n^{(m)}<\tau\right) = O(1)+ \sum_{m=1}^M \Phi\left(\frac{\tau - m\cdot \mu}{\sqrt{m\cdot \sigma^2}} \right)\,. \end{equation}
\textbf{Effective speedup.} As seen in section \ref{section:automatic_threshold}, we define the effective speedup as $$ S_{\mathrm{eff}}(\tau)= \frac{\tilde{M}(\tau)(T+T^c)}{M\cdot(\min\left\{\tau,T\right\}+T^c)}\,. $$ We are interested in calculating the expected value for the effective speedup, and in order to use the formulations in equations \ref{eq:approximate_t},\ref{eq:approximate_m} we first need to show that $\mathbb{E}[\tilde{M}(\tau)\cdot T] \approx \mathbb{E}[\tilde{M}(\tau)]\cdot \mathbb{E}[T]$. We examine $$ \mathbb{E}[\tilde{M}T]=\mathbb{E}[T(\mathbb{E}[\tilde{M}] + \tilde{M}- \mathbb{E}[\tilde{M}])]=\mathbb{E}[\tilde{M}]\mathbb{E}[T]+\mathbb{E}[T(\tilde{M}-\mathbb{E}[\tilde{M}])]$$ Applying Cauchy–Schwarz inequality we get
$$|\mathbb{E}[T(\tilde{M}-\mathbb{E}[\tilde{M}])]| \leq \sqrt{\mathbb{E}[T^2]\mathbb{E}[(\tilde{M}-\mathbb{E}[\tilde{M}])^2]}= \sqrt{\mathbb{E}[T^2]\frac{\sigma_{\tilde{M}_n}^2}{N}}=O(N^{-\frac{1}{2}}) \,,$$
where $\sigma_{\tilde{M}_n}^2$ denotes the variance of $\tilde{M}_n(\tau)=\sum_{m=1}^M \left\{ \begin{array}{lr} 1, & \text{if } T_n^{(m)}<\tau\\ 0, & \text{otherwise } \end{array}\right\}$. Hence: $$ \mathbb{E}[\tilde{M}T] = \mathbb{E}[\tilde{M}]\mathbb{E}[T] + O(N^{-\frac{1}{2}})$$ We can now write the expected value for the effective speedup as: \begin{equation} \label{eq:eff_speedup}
\mathbb{E}[S_\mathrm{eff}] =
\frac{\sum_{m=1}^M \Phi\left(\frac{\tau - m\cdot \mu}{\sqrt{m\sigma^2}} \right) }{M} \cdot
\frac{\mathbb{E}[T]}{\min(\tau, \mathbb{E}[T])+T^c}+ O(M^{-1}+M^{-1}N^{-\frac{1}{2}}) \end{equation} \begin{equation} \label{eq:T_approximation}
\mathbb{E}[T]\approx\sqrt{M\sigma^2} \left( (1-\gamma)\cdot \Phi^{-1}\left(1-\frac{1}{N} \right) + \gamma\cdot \Phi^{-1}\left(1-\frac{1}{e N}\right)\right) + M\mu + T^c \end{equation} As mentioned above, when the Gaussian Assumption \ref{assume:gaussian} is inaccurate it may be useful to plug instead in the empirical value for $\mathbb{E}[T]$ in equation \ref{eq:eff_speedup} in order to get a more accurate estimation of $\mathbb{E}[S_\mathrm{eff}]$.
\textbf{Finding $\tau^*$.} The optimal threshold $\tau^*$ can be chosen as: $$ \tau^* = \text{argmax}_\tau \mathbb{E}[S_\mathrm{eff}(\tau)] =\text{argmax}_\tau \left(\frac{1}{\tau + T^c}\cdot \sum_{m=1}^M \Phi\left(\frac{\tau - m\cdot \mu}{\sqrt{m\cdot \sigma^2}} \right)\right)$$
By using the above derivations, we can utilize $\mu$, $\sigma$, $T^c$ to understand the potential value of \textit{DropCompute}. This can be done without actually training and measuring the empiric distribution of $t_n^{(m)}$ as done in appendix section \ref{appendix:auto_threshold}. We note that finding $\tau^*$ does not require any estimation of $T$ and can be done without any statistics that originate from a large scale training session.
\subsection{Additive noise analysis} \label{appendix:noise_analysis} As a conclusion of the previous sections, we understand that the effectiveness of \textit{DropCompute} is mostly due to the behavior of the stochastic latency of each worker. To analyze this phenomenon we simulate a training of multiple workers using the scheme presented in section \ref{appendix:runtime_performace} with various additive noise types. As shown in figures \ref{fig:noise_type}, \ref{fig:noise_scale}, the ratio $\mathbb{E}[T]/\mathbb{E}[T_i]$ is good indicator for determining the potential of \textit{DropCompute} on a given training setting. High ratios indicate a gap between the step time for a single worker and the step time for multiple workers, that can be compensated by using \textit{DropCompute}.
\begin{figure}\label{fig:noise_type}
\end{figure}
\begin{figure}\label{fig:noise_scale}
\end{figure}
\section{Convergence with stochastic batch size} \label{appendix:proof} In this section, we provide proof of theorem \ref{theorem:non_convex} (\ref{appendix:non_convex_proof}), as well as convergence proof for the loss itself in the convex case (\ref{appendix:convex_proof}). We also discuss the generalization properties with a stochastic batch in section \ref{app:generalization_discussion}.
\subsection{Proof for convex case} \label{appendix:convex_proof}
\begin{theorem} \label{theorem:convex_case} Under assumption \ref{assumption:convergence} and the specific case where $f$ is a convex function, for SGD with \textit{DropCompute} (Algorithm \ref{algorithm:drop_compute}), given $N$ workers and a local batch size $b_{local}$, we have that \al \label{eq:SGD_GUARANTEES_final} \E[\mathcal{L}(\mathcal{D},\bar{\theta}) - \mathcal{L}(\mathcal{D},\theta^*)] \leq
\frac{8Lb_\mathrm{max}\|\theta_1-\theta^*\|^2}{K} + \frac{6\sigma \|\theta_1-\theta^*\|}{\sqrt{K}}~. \eal where $K$ is the total number of samples used throughout the algorithm \footnote{We assume that the batch sizes may be stochastic but $K$ is predefined and deterministic}, and the expectation is with respect to the randomization introduced due to sampling from $\mathcal{D}$ throughout the optimization process. \end{theorem}
\textbf{Proof of theorem \ref{theorem:convex_case}}
\textbf{Notation:} During the proof we denote by $b_i$ the total batch size (summed over all workers) that we employ at iteration $i$. $K = \sum_{i=1}^S b_i$ is the total number of samples along all $S$ iterations. At iteration $i$ we maintain a weight vector $\theta_i\in{\mathbb R}^d$ and query a gradient estimate $g_i$ based on a batch size of $b_i$ samples. Thus, we set $$ g_i = \frac{1}{b_i} \sum_{s=1}^{b_i} g_i^s $$
where $g_i^s=\nabla\ell(z_i^s,\theta_i)$, and $z_i^s$ is randomly sampled from $\mathcal{D}$. We also maintain importance weights, $\alpha_i = b_i$.
Note that there exists $b_{\max}$ such that $\alpha_i\leq b_{\max}~;\forall i$
Now, the update rule: \al \label{eq:Update} \theta_{i+1} = \theta_i - \eta \alpha_i g_i \eal Eventually, we output $$ \bar{\theta} = \frac{1}{\alpha_{1:S}}\sum_{i=1}^S \alpha_i \theta_i $$ where $\alpha_{1:S}:=\sum_{i=1}^S \alpha_i.$
We assume that the batch sizes $b_i$ are stopping times w.r.t.~the natural filtration induced by the samples we draw during the optimization process. Informally, this means that the value of $b_i$ depends only on the history of the samples that we have seen prior to setting $b_i$.
We can therefore prove the following lemma, \begin{lemma} \label{lem:Unbiased}
Upon choosing $\alpha_i = b_i$ the following holds,
\als
\E [\alpha_i (g_i - \nabla \mathcal{L}(\mathcal{D},\theta_i)\vert \theta_i]
= \E [\sum_{s=1}^{b_i} (g_i^s - \nabla\mathcal{L}(\mathcal{D},\theta_i))\vert \theta_i] = 0~.
\eals \end{lemma}
Now, using standard analysis for Online Gradient Descent \citep{hazan2016introduction} with the update rule of \eqref{eq:Update} gives, $$
\sum_{i=1}^S \alpha_i g_i\cdot(\theta_i-\theta^*) \leq \frac{\|\theta_1-\theta^*\|^2}{\eta} + \eta \sum_{i=1}^S \alpha_i^2 \|g_i\|^2 $$ Taking expectation and using Lemma~\ref{lem:Unbiased} gives, $$
\E\sum_{i=1}^S \alpha_i \nabla\mathcal{L}(\mathcal{D},\theta_i)\cdot(\theta_i-\theta^*) \leq \frac{\|\theta_1-\theta^*\|^2}{\eta} + \eta \E\sum_{i=1}^S \alpha_i^2 \|g_i\|^2 $$ From convexity we know that $0\leq \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*)\leq \nabla \mathcal{L}(\mathcal{D},\theta_i)\cdot(\theta_i-\theta^*)$, therefore the above implies, \al \label{eq:SGD_GUARANTEES}
\E\sum_{i=1}^S \alpha_i( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*) ) \leq \frac{\|\theta_1-\theta^*\|^2}{\eta} + \eta \E\sum_{i=1}^S \alpha_i^2 \|g_i\|^2 \eal
Now, we write $g_i = \nabla \mathcal{L}(\mathcal{D},\theta_i) + \xi_i$ where $\xi_i = g_i - \nabla \mathcal{L}(\mathcal{D},\theta_i)$ and note that $$\alpha_i \xi_i = \sum_{s=1}^{b_i}(g_i^s - \nabla \mathcal{L}(\mathcal{D},\theta_i)) = \sum_{s=1}^{b_i} \xi_i^i$$ where we denote $\xi_i^s: = g_i^s - \nabla \mathcal{L}(\mathcal{D},\theta_i)$. Next, we shall use the following lemma, \begin{lemma}\label{lem:BoundG1} The following holds, \al \label{eq:VarStep2_FinalA}
\E \alpha_i^2 \|g_i\|^2
&\leq 2 b_{\max} \E \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 + 2\sigma^2 \E b_i~. \eal Moreover, due to the $L$-smoothness of $f(\cdot)$, and global optimality of $\theta^*$, the following holds, \al \label{eq:VarStep2_FinalB}
\E \alpha_i^2 \|g_i\|^2 &\leq 4 b_{\max}L \E \alpha_i ( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*)) + 2\sigma^2 \E b_i~. \eal \end{lemma}
\paragraph{Final Bound:} Plugging the above lemma back into Eq.~\eqref{eq:SGD_GUARANTEES} gives, \al \label{eq:SGD_GUARANTEES2} \E\sum_{i=1}^S \alpha_i( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*) ) &\leq
\frac{\|\theta_1-\theta^*\|^2}{\eta} + 4\eta b_{\max} \E\sum_{i=1}^S\alpha_i L( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*)) + 8\eta\sigma^2 \E\sum_{i=1}^S b_i \nonumber\\ &\leq
\frac{\|\theta_1-\theta^*\|^2}{\eta} + 4\eta b_{\max}L \E\sum_{i=1}^S\alpha_i ( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*)) + 8\eta\sigma^2 K ~, \eal where we used $K = \sum_{i=1}^S b_i$.
Now if we pick $\eta$ such that $4\eta b_{\max}L \leq 1/2$ then we can move the second term in the RHS to the LHS and obtain, \al \label{eq:SGD_GUARANTEES3} \frac{1}{2}\E\sum_{i=1}^S \alpha_i( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*) ) &\leq
\frac{\|\theta_1-\theta^*\|^2}{\eta} + 8\eta\sigma^2 K ~,
\eal Thus, choosing $\eta = \min\left\{ \frac{\|\theta_1-\theta^*\|}{\sigma\sqrt{8K}} , \frac{1}{8L b_{\max}} \right\}$ gives the following bound, \al \label{eq:SGD_GUARANTEES4} \E\sum_{i=1}^S \alpha_i( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*) ) &\leq
8Lb_{\max}\|\theta_1-\theta^*\|^2 + 6\sigma \|\theta_1-\theta^*\|\sqrt{K} \eal Now, recalling that $K : = \sum_{i=1}^S b_i= \sum_{i=1}^S \alpha_i$ and using Jensen's inequality together with $\bar{\theta} = \frac{1}{\alpha_{1:S}}\sum_{i=1}^S \alpha_i \theta_i$ yields, \al \label{eq:SGD_GUARANTEES5} \E[\mathcal{L}(\mathcal{D},\bar{\theta}) - \mathcal{L}(\mathcal{D},\theta^*)] &\leq \E\sum_{i=1}^S \frac{\alpha_i}{\alpha_{1:S}}( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*) ) \leq
\frac{8Lb_{\max}\|\theta_1-\theta^*\|^2 + 6\sigma \|\theta_1-\theta^*\|\sqrt{K}}{\alpha_{1:S}}\nonumber\\ &\leq
\frac{8Lb_{\max}\|\theta_1-\theta^*\|^2}{K} + \frac{6\sigma \|\theta_1-\theta^*\|}{\sqrt{K}} \eal where we used $\alpha_{1:S} = K$. \hspace{0.5cm}\qedsymbol{}
\subsection{Proof for non-convex case} \label{appendix:non_convex_proof}
\textbf{Proof of theorem \ref{theorem:non_convex}}
We use the same notation for $b_i$ and $g_i$ as before. And again used weights, $$ \alpha_i = b_i $$ We also assume that $b_i\leq b_{\max}~,\forall i$.
The update rule is the following, $$ \theta_{i+1} = \theta_i - \eta \alpha_i g_i $$ And the output is $\bar{\theta}$, where we define, $$ \bar{\theta} = \theta_i~;\quad \text{w.p.}~~ \frac{\alpha_i}{\alpha_{1:S}} $$ Thus,
$$ \E \|\nabla \mathcal{L}(\mathcal{D},\bar{\theta})\|^2=\frac{1}{\alpha_{1:S}} \sum_{i=1}^S \alpha_i \E\|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 $$
Using smoothness, \als \mathcal{L}(\mathcal{D},\theta_{i+1}) &\leq
\mathcal{L}(\mathcal{D},\theta_i) - \nabla \mathcal{L}(\mathcal{D},\theta_i)\cdot (\theta_{i+1}-\theta_i) + \frac{L}{2}\|\theta_{i+1}-\theta_i \|^2 \nonumber\\ &\leq
\mathcal{L}(\mathcal{D},\theta_i) - \eta \alpha_i \nabla \mathcal{L}(\mathcal{D},\theta_i)\cdot g_i + \frac{L\eta^2}{2}\|\alpha_i g_i \|^2 \nonumber\\ &\leq
\mathcal{L}(\mathcal{D},\theta_i) - \eta \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2-\eta \alpha_i \nabla \mathcal{L}(\mathcal{D},\theta_i)\cdot (g_i-\nabla \mathcal{L}(\mathcal{D},\theta_i)) + \frac{L\eta^2}{2}\|\alpha_i g_i \|^2 \eals Re-arranging the above yields, \als
\eta \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 &\leq
\mathcal{L}(\mathcal{D},\theta_i) - \mathcal{L}(\mathcal{D},\theta_{i+1}) -\eta \alpha_i \nabla \mathcal{L}(\mathcal{D},\theta_i)\cdot (g_i-\nabla \mathcal{L}(\mathcal{D},\theta_i)) + \frac{L\eta^2}{2}\|\alpha_i g_i \|^2 \eals Summing the above, and dividing by $\eta$, we obtain, \al \label{eq:NonCvx1}
\sum_{i=1}^S \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 &\leq
\frac{1}{\eta}(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta_{S+1})) - \sum_{i=1}^S\alpha_i \nabla \mathcal{L}(\mathcal{D},\theta_i)\cdot (g_i-\nabla \mathcal{L}(\mathcal{D},\theta_i)) + \frac{L\eta}{2}\sum_{i=1}^S\|\alpha_i g_i \|^2 \nonumber\\ &\leq
\frac{1}{\eta}(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)) - \sum_{i=1}^S\alpha_i \nabla \mathcal{L}(\mathcal{D},\theta_i)\cdot (g_i-\nabla \mathcal{L}(\mathcal{D},\theta_i)) + \frac{L\eta}{2}\sum_{i=1}^S\|\alpha_i g_i \|^2 \eal where we uses $\mathcal{L}(\mathcal{D},\theta^*) \leq \mathcal{L}(\mathcal{D},\theta_{S+1})$ since $\theta^*$ is the global minimum of $\mathcal{L}(\mathcal{D},\cdot)$.
Now recall that from Lemma~\ref{lem:Unbiased} we have, \al \label{eq:Doob2B} \E[\alpha_i (g_i-\nabla \mathcal{L}(\mathcal{D},\theta_i))\vert \theta_i] = 0~. \eal And from Lemma~\ref{lem:BoundG1} we have, \al \label{eq:VarStep2B}
\E \alpha_i^2 \|g_i\|^2
&\leq 2 b_{\max} \E \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 + 2\sigma^2 \E b_i~. \eal
Thus, taking expectation in Eq.~\eqref{eq:NonCvx1}, and plugging Eq.~\eqref{eq:Doob2B} and \eqref{eq:VarStep2B}, yields, \al \label{eq:NonCvx2}
\E \sum_{i=1}^S \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 &\leq
\frac{1}{\eta}(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)) - \sum_{i=1}^S\E \alpha_i \nabla \mathcal{L}(\mathcal{D},\theta_i)\cdot (g_i-\nabla \mathcal{L}(\mathcal{D},\theta_i)) + \frac{L\eta}{2}\sum_{i=1}^S \E\|\alpha_i g_i \|^2 \nonumber\\ &\leq
\frac{1}{\eta}(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)) + L\eta b_{\max} \sum_{i=1}^S \E\alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 +L\eta \sigma^2 \E \sum_{i=1}^S b_i \nonumber\\ &\leq
\frac{1}{\eta}(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)) + L\eta b_{\max} \sum_{i=1}^S \E\alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 +L\eta \sigma^2 \cdot K \eal where the last line uses $K:= \sum_{i=1}^S b_i$.
Now if we pick $\eta$ such that $\eta b_{\max}L \leq 1/2$ then we can move the second term in the RHS to the LHS and obtain, \al \label{eq:NonCvx3}
\frac{1}{2}\E \sum_{i=1}^S \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 &\leq \frac{1}{\eta}(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)) +L\eta \sigma^2 \cdot K \eal Thus, choosing $\eta = \min\left\{ \frac{\sqrt{\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)}}{\sigma\sqrt{LK}} , \frac{1}{2L b_{\max}} \right\}$ gives the following bound, \al \label{eq:NonCvx4}
\E \sum_{i=1}^S \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 &\leq 2L b_{\max}(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)) +2\sigma\sqrt{L(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)) } \cdot \sqrt{K} \eal Dividing by $K:=\alpha_{1:S}$ and using the definition of $\bar{\theta}$ yields, \al \label{eq:NonCvx5}
\E\|\nabla \mathcal{L}(\mathcal{D},\bar{\theta})\|^2 =
\E \frac{1}{\alpha_{1:S}}\sum_{i=1}^S \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 &\leq \frac{2L b_{\max}(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*))}{K} +\frac{2\sigma\sqrt{L(\mathcal{L}(\mathcal{D},\theta_1) - \mathcal{L}(\mathcal{D},\theta^*)) }}{\sqrt{K}} \eal
\subsection{Remaining Proofs} \subsubsection{Proof of Lemma~\ref{lem:BoundG1}}
\begin{proof}[Proof of Lemma~\ref{lem:BoundG1}] We can write, \al \label{eq:VarStep1}
\alpha_i^2 \|g_i\|^2 &=
\|b_i \nabla \mathcal{L}(\mathcal{D},\theta_i) + \sum_{s=1}^{b_i}\xi_i^s\|^2 \nonumber\\ &\leq
2\|b_i \nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 + 2\|\sum_{s=1}^{b_i}\xi_i^s\|^2 \nonumber\\ &=
2 b_i^2 \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 + 2\|\sum_{s=1}^{b_i}\xi_i^s\|^2 \nonumber\\
&\leq 2 b_{\max} \alpha_i \|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 + 2\|\sum_{s=1}^{b_i}\xi_i^s\|^2 \nonumber\\
\eal where the second line uses $\|a+b\|^2 \leq 2\|a\|^2+2\|b\|^2$; the fourth line uses $b_i\leq b_{\max}$ as well as $\alpha_i = b_i$ implying that $b_i^2 \leq b_{\max}\alpha_i$.
\paragraph{Bounding $E\|\sum_{s=1}^{b_i}\xi_i^s\|^2$:}
Given $i$ and $\theta_i$, Let us define the following sequence, $Q_0=0$, $Q_1 = \|\xi_i^1\|^2-\sigma$, and for any $k>1$ $$
Q_k = \sum_{s=1}^k\|\xi_i^k\|^2-\sigma^2 \cdot k + 2\sum_{s=1}^j \sum_{n=s+1}^j \xi_i^s\cdot \xi_i^n $$ It can be directly shown that $\{Q_k\}_k$ is a Supermartingale sequence, and that $$
\|\sum_{s=1}^{b_i}\xi_i^s\|^2 = Q_{b_i}+ \sigma^2 \cdot b_i $$ Thus, since $b_i$ is a bounded stopping time, we can use Doob's optional stopping theorem which implies that, $$
\E \|\sum_{s=1}^{b_i}\xi_i^s\|^2 = \E Q_{b_i}+ \sigma^2 \E\cdot b_i \leq EQ_0 + \sigma^2 \E\cdot b_i = 0+\sigma^2 \E b_i $$ Plugging the above back into Eq.~\eqref{eq:VarStep1} yields, \al \label{eq:VarStep2_Pre}
\E \alpha_i^2 \|g_i\|^2
&\leq 2 b_{\max} \E \alpha_i \| \nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 + 2\sigma^2 \E b_i~. \eal
Now, since $\mathcal{L}(\mathcal{D},\cdot)$ is $L$-smooth and $\theta^*$ is its global minima, then the following holds: $\|\nabla \mathcal{L}(\mathcal{D},\theta_i)\|^2 \leq 2L( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*))$; See e.g.~\citet{levy2017online} for the proof. Plugging this into the above equation we obtain, \al \label{eq:VarStep2_Post}
\E \alpha_i^2 \|g_i\|^2 &\leq 4 b_{\max} \E \alpha_i L( \mathcal{L}(\mathcal{D},\theta_i)-\mathcal{L}(\mathcal{D},\theta^*)) + 2\sigma^2 \E b_i~. \eal \end{proof}
\subsubsection{Proof of Lemma~\ref{lem:Unbiased}} \begin{proof}[Proof of Lemma~\ref{lem:Unbiased}] We can define the following Martingale sequence for each step $s$: $M_0 = 0$, and $M_j = \sum_{s=1}^j (g_i^s - \nabla \mathcal{L}(\mathcal{D},\theta_i))$ for any $j=1,2,\ldots$.
Thus, since the mixing time $b_i$ is bounded by $b_{\max}$, then according to Doob's optional stopping theorem \citep{levin2017markov} we have that, \al \label{eq:Doobs} \E[M_{b_i}\vert \theta_i] = \E [\sum_{s=1}^{b_i} (g_i^s - \nabla \mathcal{L}(\mathcal{D},\theta_i))\vert \theta_i] = \E [M_0\vert \theta_i] = 0~. \eal
Now, notice that for any $i$, we have $\alpha_i (g_i-\nabla \mathcal{L}(\mathcal{D},\theta_i)) = M_{b_i}$, and therefore, \al \label{eq:Doob2} \E[\alpha_i (g_i-\nabla \mathcal{L}(\mathcal{D},\theta_i))\vert \theta_i] =\E[ M_{b_i}\vert \theta_i] = 0~. \eal
\end{proof}
\subsection{Generalization discussion} \label{app:generalization_discussion} An interesting observation arising from our results is the small impact of gradient dropping as measured in final test accuracy. One explanation for this can be based on viewing \textit{DropCompute} as noise induced over gradients. Optimization using variants of SGD is inherently noisy due to the use of data samples used to evaluate the intermediate error. The stochastic nature of computed weight gradients was previously found to provide generalization benefits for the final trained model, although this is still part of ongoing debate \citep{geiping2022stochastic}. Nevertheless, several works found generalization benefits with the \emph{injection} of noise into the weights' gradients \citep{gradnoise} or their use in computed update rule \citep{lrdropout}.
\end{document}
|
arXiv
|
{
"id": "2306.10598.tex",
"language_detection_score": 0.7458488345146179,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Key polynomials and pseudo-convergent sequences} \author{Josnei Novacoski} \author{Mark Spivakovsky}
\keywords{Key polynomials, Pseudo-convergent sequences, Valuations} \subjclass[2010]{Primary 13A18} \begin{abstract} In this paper we introduce a new concept of key polynomials for a given valuation $\nu$ on $K[x]$. We prove that such polynomials have many of the expected properties of key polynomials as those defined by MacLane and Vaqui\'e, for instance, that they are irreducible and that the truncation of $\nu$ associated to each key polynomial is a valuation. Moreover, we prove that every valuation $\nu$ on $K[x]$ admits a sequence of key polynomials that completely determines $\nu$ (in the sense which we make precise in the paper). We also establish the relation between these key polynomials and pseudo-convergent sequences defined by Kaplansky. \end{abstract}
\maketitle \section{Introduction} Given a valuation $\nu$ of a field $K$, it is important to understand what are the possible extensions of $\nu$ to $K[x]$. Many different theories have been developed in order to understand such extensions. For instance, in \cite{Mac}, MacLane develops the theory of key polynomials. He proves that given a discrete valuation $\nu$ of $K$, every extention of $\nu$ to $K[x]$ is uniquely determined by a sequence (with order type at most $\omega$) of key polynomials. Recently, M. Vaqui\'e developed a more general theory of key polynomials (see \cite{Vaq}), which extends the results of MacLane for a general valued field (that is, the given valuation of $K$ is no longer assumed to be discrete). At the same time, F.H. Herrera Govantes, W. Mahboub, M.A. Olalla Acosta and M. Spivakovsky developed another definition of key polynomials (see \cite{HOS}). This definition is an adaptation of the concept of generating sequences introduced by Spivakovsky in \cite{Spi1}. A comparison between this two definitions of key polynomials is presented in \cite{Mahboud}.
Roughly speaking, for a given valuation $\mu$ of $K[x]$, a MacLane -- Vaqui\'e key polynomial $\phi\in K[x]$ for $\mu$ is a polynomial that allows us to obtain a new valuation $\mu_1$ of $K[x]$ with $\mu_1(\phi)=\gamma_1>\mu(\phi)$ and $\mu(p)=\mu_1(p)$ for every $p\in K[x]$ with $\deg(p)<\deg(\phi)$ (in this case we denote $\mu_1$ by $[\mu;\mu_1(\phi)=\gamma_1]$). Then, for any valuation $\nu$ of $K[x]$ one tries to obtain a sequence of valuations $\mu_0,\mu_1,\ldots,\mu_n,\ldots$ with $\mu_0$ a monomial valuation and $\mu_{i+1}=[\mu_i;\mu_{i+1}(\phi_{i+1})=\gamma_{i+1}]$ for a key polynomial $\phi_{i+1}$ for $\mu_i$, such that \begin{equation} \nu=\lim \mu_i\label{eq:nu=lim} \end{equation} (in the sense that will be defined precisely below). This process does not work in general, that is, the equality (\ref{eq:nu=lim}) may not hold even after one constructs an infinite sequence $\{\mu_i\}$. This leads one to introduce the concept of ``limit key polynomial".
It is known that valuations which admit limit key polynomials are more difficult to handle. For instance, it was proved by J.-C. San Saturnino (see Theorem 6.5 of \cite{JCSS}), that if a valuation $\nu$ is centered on a noetherian local domain and $\nu$ does not admit limit key polynomials (on any sub-extension $R'\subseteq R'[x]\subseteq R$ with $\dim\ R'=\dim\ R-1$), then it has the local uniformization property (where we assume, inductively, that local uniformization holds for $R'$).
In this paper, we introduce a new concept of key polynomials. Let $K$ be a field and $\nu$ a valuation on $K[x]$. Let $\Gamma$ denote the value group of $K$ and $\Gamma'$ the value group of $K[x]$. For a positive integer $b$, let $\partial_b:=\frac{1}{b!}\frac{\partial^b}{\partial x^b}$ (this differential operator of order $b$ is sometiems called \textbf{the $b$-th formal derivative}). For a polynomial $f\in K[x]$ let \[ \epsilon(f)=\max_{b\in \mathbb{N}}\left\{\frac{\nu(f)-\nu(\partial_bf)}{b}\right\}. \] A monic polynomial $Q\in K[x]$ is said to be a \textbf{key polynomial} (of level $\epsilon (Q)$) if for every $f\in K[x]$ if $\epsilon(f)\geq \epsilon(Q)$, then $\deg(f)\geq\deg(Q)$.
This new definition offers many advantages. For instance, it gives a criterion to determine, for a given valuation $\nu$ of $K[x]$, whether any given polynomial is a key polynomial for $\nu$. This has a different meaning than in the approach of MacLane-Vaqui\'e. In their approach, a key polynomial allows us to ``extend the given valuation" and here a key polynomial allows us to ``truncate the given valuation". For instance, our definition of key polynomials treats the limit key polynomials on the same footing as the non-limit ones. Moreover, we present a characterization of key polynomial (Theorem \ref{definofkeypol}) which allows us to determine whether a given key polynomial is a limit key polynomial. A more precise comparison between the concept of key polynomial introduced here and that of MacLane -- Vaqui\'e will be explored in a forthcoming paper by Decoup, Mahboub and Spivakovsky.
Given two polynomials $f,q\in K[x]$ with $q$ monic, we call the \textbf{$q$-standard expansion of $f$} the expression \[ f(x)=f_0(x)+f_1(x)q(x)+\ldots+f_n(x)q^n(x) \] where for each $i$, $0\leq i\leq n$, $f_i=0$ or $\deg(f_i)<\deg(q)$. For a polynomial $q(x)\in K[x]$, the \textbf{$q$-truncation of $\nu$} is defined as \[ \nu_q(f):=\min_{0\leq i\leq n}\{\nu(f_iq^i)\} \] where $f=f_0+f_1q+\ldots+f_nq^n$ is the $q$-standard expansion of $f$. In Section 2, we present an example that shows that $\nu_q$ does need to be a valuation. We also prove (Theorem \ref{proptruncakeypolval}) that if $Q$ is a key polynomial, then $\nu_Q$ is a valuation. A set $\Lambda$ of key polynomials is said to be a \textbf{complete set of key polynomials for $\nu$} if for every $f\in K[x]$, there exists $Q\in \Lambda$ such that $\nu_Q(f)=\nu(f)$. One of the main results of this paper is the following:
\begin{Teo}\label{Theoremexistencecompleteseqkpol} Every valuation $\nu$ on $K[x]$ admits a complete set of key polynomials. \end{Teo}
Another way of describing extensions of valuations from $K$ to $K[x]$ is the theory of pseudo-convergent sequences developed by Kaplansky in \cite{Kap}. He uses this theory to determine whether a maximal immediate extension of the valued field $(K,\nu)$ is unique (up to isomorphism). For a valued field $(K,\nu)$, a \textbf{pseudo-convergent sequence} is a well-ordered subset $\{a_{\rho}\}_{\rho<\lambda}$ of $K$, without last element, such that \[ \nu(a_\sigma-a_\rho)<\nu(a_\tau-a_\sigma)\mbox{ for all }\rho<\sigma<\tau<\lambda. \] For a given pseudo-convergent sequence $\{a_{\rho}\}_{\rho<\lambda}$ it is easy to show that either $\nu(a_\rho)<\nu(a_\sigma)$ for all $\rho<\sigma<\lambda$ or there is $\rho<\lambda$ such that $\nu(a_\sigma)=\nu(a_\rho)$ for every $\rho<\sigma<\lambda$. If we set $\gamma_\rho:=\nu(a_{\rho+1}-a_\rho)$, then $\nu(a_\sigma-a_\rho)=\gamma_\rho$ for every $\rho<\sigma<\lambda$. Hence, the sequence $\{\gamma_\rho\}_{\rho<\lambda}$ is an increasing subset of $\Gamma$. An element $a\in K$ is said to be a \textbf{limit} of the pseudo-convergent sequence $\{a_\rho\}_{\rho<\lambda}$ if $\nu(a-a_\rho)=\gamma_\rho$ for every $\rho<\lambda$.
One can prove that for every polynomial $f(x)\in K[x]$, there exists $\rho_f<\lambda$ such that either \begin{equation}\label{condforpscstotra} \nu(f(a_\sigma))=\nu(f(a_{\rho_f}))\mbox{ for every }\rho_f\leq \sigma<\lambda, \end{equation} or \begin{equation}\label{condforpscstoalg} \nu(f(a_\sigma))>\nu(f(a_{\rho}))\mbox{ for every }\rho_f\leq \rho< \sigma<\lambda. \end{equation} If case (\ref{condforpscstotra}) happens, we say that the value of $f$ is fixed by $\{a_\rho\}_{\rho<\lambda}$ (or that $\{a_\rho\}_{\rho<\lambda}$ fixes the value of $f$). A pseudo-convergent sequence $\{a_\rho\}_{\rho<\lambda}$ is said to be of \textbf{transcendental type} if for every polynomial $f(x)\in K[x]$ the condition (\ref{condforpscstotra}) holds. Otherwise, $\{a_\rho\}_{\rho<\lambda}$ is said to be of \textbf{algebraic type}, i.e., if there exists at least one polynomial for which condition (\ref{condforpscstoalg}) holds.
The concept of key polynomials appears in the approach to local uniformization by Spivakovsky.
On the other hand, the concept of pseudo-convergent sequence plays an important role in the work of Knaf and Kuhlmann (see \cite{KK_1}) on the local uniformization problem. In this paper, we present a comparison between the concepts of key polynomials and pseudo-convergent sequences. More specifically, we prove the following:
\begin{Teo}\label{compthemkppsc} Let $\nu$ be a valuation on $K[x]$ and let $\{a_\rho\}_{\rho<\lambda}\subset K$ be a pseudo-convergent sequence, without a limit in $K$, for which $x$ is a limit. If $\{a_\rho\}_{\rho<\lambda}$ is of transcendental type, then $\Lambda:=\{x-a_\rho\mid \rho<\lambda\}$ is a complete set of key polynomials for $\nu$. On the other hand, if $\{a_\rho\}_{\rho<\lambda}$ is of algebraic type, then every polynomial $q(x)$ of minimal degree among the polynomials not fixed by $\{a_\rho\}_{\rho<\lambda}$ is a limit key polynomial for $\nu$. \end{Teo}
\section{Key polynomials} We will assume throughout this paper that $K$ is a field, $\nu$ a valuation of $K[x]$, non-trivial on $K$ with $\nu(x)\geq 0$. We begin by making some remarks. \begin{Obs}\label{exlinearnonlinepolarepsk} \begin{description} \item[(i)] Every linear polynomial $x-a$ is a key polynomial (of level $\epsilon(x-a)=\nu(x-a)$).
\item[(ii)] Take a polynomial $f(x)\in K[x]$ of degree greater than one and assume that there exists $a\in K$ such that $\nu(\partial_bf(a))=\nu(\partial_bf(x))$ for every $b\in\mathbb{N}$ (note that such an $a$ always exists if the assumptions of Theorem \ref{compthemkppsc} hold and the pseudo-convergent sequence is transcendental or is algebraic and $\deg(f)\leq\deg(q)$). Write \[ f(x)=f(a)+\sum_{i=1}^n\partial_if(a)(x-a)^i \] and take $h\in\{1,\ldots,n\}$ such that \[ \nu(\partial_hf(x))+h\nu(x-a)=\min_{1\leq i\leq n}\{\nu(\partial_if(x))+i\nu(x-a)\}. \] If $\nu(f(a))<\nu(\partial_hf(x))+h\nu(x-a)$, then $\nu(f(x))=\nu(f(a))$ and hence \[ \frac{\nu(f(x))-\nu(\partial_if(x))}{i}<\nu(x-a) \] for every $i$, $1\leq i\leq n$. Consequently, $\epsilon(f)<\nu(x-a)=\epsilon(x-a)$ and hence $f$ is not a key polynomial. On the other hand, if $$ \nu(\partial_hf(x))+h\nu(x-a)\leq\nu(f(a)), $$ then \begin{equation}\label{eqtaylorexpwithpcs} \nu(f(x))\geq\nu(\partial_hf(x))+h\nu(x-a) \end{equation} and if the equality holds in (\ref{eqtaylorexpwithpcs}), then \[ \epsilon(f)=\frac{\nu(f(x))-\nu(\partial_hf(x))}{h}=\nu(x-a)=\epsilon(x-a) \] and hence $f$ is not a key polynomial. In other words, the only situation when $f$ may be a key polynomial is when $$ f(x)>\min_{1\leq i\leq n}\{f(a),\nu(\partial_if(x))+i\nu(x-a)\}. $$ \end{description} \end{Obs}
\begin{Obs} We observe that if $Q$ is a key polynomial of level $\epsilon:=\epsilon(Q)$, then for every polynomial $f\in K[x]$ with $\deg(f)<\deg(Q)$ and every $b\in\mathbb{N}$ we have \begin{equation}\label{eqpolyndegsmallkeypol} \nu(\partial_b(f))>\nu(f)-b\epsilon. \end{equation} Indeed, from the definition of key polynomial we have that $\epsilon>\epsilon(f)$. Hence, for every $b\in\mathbb{N}$ we have \[ \frac{\nu(f)-\nu(\partial_b(f))}{b}\leq\epsilon(f)<\epsilon \] and this implies (\ref{eqpolyndegsmallkeypol}). \end{Obs} Let \[
I(f)=\left\{b\in\mathbb{N}\left|\frac{\nu(f)-\nu(\partial_bf)}{b}=\epsilon(f)\right.\right\} \] and $b(f)=\min I(f)$. \begin{Lema} Let $Q$ be a key polynomial and take $f,g\in K[x]$ such that $$ \deg(f)<\deg(Q) $$ and $$ \deg(g)<\deg(Q). $$ Then for $\epsilon:=\epsilon(Q)$ and any $b\in\mathbb{N}$ we have the following: \begin{description}\label{lemaonkeypollder} \item[(i)] $\nu(\partial_b(fg))>\nu(fg)-b\epsilon$ \item[(ii)] If $\nu_Q(fQ+g)<\nu(fQ+g)$ and $b\in I(Q)$, then $\nu(\partial_b(fQ+g))=\nu(fQ)-b\epsilon$; \item[(iii)] If $h_1,\ldots,h_s$ are polynomials such that $\deg(h_i)<\deg(Q)$ for every $i=1,\ldots, s$ and $\displaystyle\prod_{i=1}^sh_i=qQ+r$ with $\deg(r)<\deg(Q)$, then \[ \nu(r)=\nu\left(\prod_{i=1}^sh_i\right)<\nu(qQ). \] \end{description} \end{Lema}
\begin{proof} \textbf{(i)} Since $\deg(f)<\deg(Q)$ and $\deg(g)<\deg(Q)$, for each $j\in \mathbb{N}$, we have \[ \nu(\partial_jf)>\nu(f)-j\epsilon\mbox{ and }\nu(\partial_jg)>\nu(g)-j\epsilon. \] This, and the fact that \[ \partial_b(fg)=\sum_{j=0}^b\partial_jf\partial_{b-j}g, \] imply that \[ \nu(\partial_b(fg))\geq\min_{0\leq j\leq b}\{\nu(\partial_jf)+\nu(\partial_{b-j}g)\}>\nu(fg)-b\epsilon. \]
\textbf{(ii)} If $\nu_Q(fQ+g)<\nu(fQ+g)$, then $\nu(fQ)=\nu(g)$. Hence, \[ \nu(\partial_bg)>\nu(g)-b\epsilon=\nu(fQ)-b\epsilon. \] Moreover, for every $j\in\mathbb{N}$, we have \[ \nu(\partial_j f\partial_{b-j}Q)=\nu(\partial_jf)+\nu(\partial_{b-j}Q)>\nu(f)-j\epsilon+\nu(Q)-(b-j)\epsilon=\nu(fQ)-b\epsilon. \] Therefore, \[ \nu(\partial_b(fQ+g))=\nu\left(f\partial_bQ+\sum_{j=1}^b\partial_jf\partial_{b-j}Q+\partial_bg\right)=\nu(fQ)-b\epsilon. \]
\textbf{(iii)} We proceed by induction on $s$. If $s=1$, then $h_1=qQ+r$ with $$ \deg(h_1)<\deg(Q), $$ which implies that $h_1=r$ and $q=0$. Our result follows immediately.
Next, consider the case $s=2$. Take $f,g\in K[x]$ such that $\deg(f)<\deg(Q)$, $\deg(g)<\deg(Q)$ and write $fg=qQ+r$ with $\deg(r)<\deg(Q)$. Then $$ \deg(q)<\deg(Q) $$ and for $b\in I(Q)$ we have \[ \nu\left(\partial_b(qQ)\right)=\nu\left(\sum_{j=0}^b\partial_jq\partial_{b-j}Q\right)=\nu(qQ)-b\epsilon. \] This and part \textbf{(i)} imply that \begin{displaymath} \begin{array}{rcl} \nu(qQ)-b\epsilon &=& \nu\left(\partial_b(qQ)\right)= \nu(\partial_b(fg)-\partial_b(r))\\
&\geq &\min\{\nu\left(\partial_b(fg)\right),\nu\left(\partial_b(r)\right)\}\\
&>&\min\{\nu(fg),\nu(r)\}-b\epsilon.
\end{array} \end{displaymath} and consequently \begin{equation}\label{equationwithepsilon} \nu(r)=\nu(fg)<\nu(qQ). \end{equation}
Assume now that $s>2$ and define $\displaystyle h:=\prod_{i=1}^{s-1}h_i$. Write $h=q_1Q+r_1$ with $\deg(r_1)<\deg(Q)$. Then by the induction hypothesis we have $$ \nu(r_1)=\nu(h)<\nu(q_1Q) $$ and hence \[ \nu\left(\prod_{i=1}^sh_i\right)=\nu(r_1h_s)<\nu(q_1h_sQ). \] Write $r_1h_s=q_2Q+r_2$. Then, by equation (\ref{equationwithepsilon}) we have \[ \nu(r_2)=\nu(r_1h_s)<\nu(q_2Q). \] If $\displaystyle \prod_{i=1}^sh_i=qQ+r$ with $\deg(r)<\deg(Q)$, then \[ qQ+r=\prod_{i=1}^sh_i=hh_s=(q_1Q+r_1)h_s=q_1h_sQ+r_1h_s=q_1h_sQ+q_2Q+r_2 \] and hence $q=q_1h_s+q_2$ and $r=r_2$. Therefore, \[ \nu(qQ)\geq\min\{\nu(q_1h_sQ),\nu(q_2Q)\}>\nu(r_1h_s)=\nu(r)=\nu\left(\prod_{i=1}^sh_i\right). \] This is what we wanted to prove. \end{proof}
We denote by $p$ the \textbf{exponent characteristic} of $K$, that is, $p=1$ if $\mbox{\rm char}(K)=0$ and $p=\mbox{\rm char}(K)$ if $\mbox{\rm char}(K)>0$. \begin{Prop}\label{propaboutpseudkeyool} Let $Q\in K[x]$ be a key polynomial and set $\epsilon:=\epsilon(Q)$. Then the following hold: \begin{description} \item[(i)] Every element in $I(Q)$ is a power of $p$; \item[(ii)] $Q$ is irreducible. \end{description} \end{Prop} \begin{proof} \textbf{(i)} Take $b\in I(Q)$ and assume, aiming for contradiction, that $b$ is not a power of $p$. Write $b=p^tr$ where $r>1$ is prime to $p$. Then, by Lemma 6 of \cite{Kap}, $\binom{b}{p^t}$ is prime to $p$ and hence $\nu\binom{b}{p^t}=0$. Since $\binom{b}{p^t}\partial_b=\partial_{p^t}\circ\partial_{b'}$ for $b'=b-p^t$, we have \[ \nu(\partial_{b'}Q)-\nu(\partial_bQ)=\nu(\partial_{b'}Q)-\nu(\partial_{p^t}(\partial_{b'}Q))\leq p^t\epsilon(\partial_{b'}(Q))<p^t\epsilon \] because $\deg(\partial_{b'}Q)<\deg(Q)$ and $Q$ is a key polynomial. Hence, \[ b\epsilon=\nu(Q)-\nu(\partial_bQ)=\nu(Q)-\nu(\partial_{b'}Q)+\nu(\partial_{b'}Q)-\nu(\partial_bQ)< b'\epsilon+p^t\epsilon=b\epsilon, \] which gives the desired contradiction.
\textbf{(ii)} If $Q=gh$ for non-constant polynomials $g,h\in K[x]$, then by Lemma \ref{lemaonkeypollder} \textbf{(i)}, we would have for $b\in I(Q)$ that \[ \nu(\partial_bQ)>\nu(Q)-b\epsilon, \] which is a contradiction to the definition of $b$ and $\epsilon$. \end{proof}
We present an example to show that $\nu_q$ does not need to be a valuation for a general polynomial $q(x)\in K[x]$. \begin{Exa} Consider a valuation $\nu$ in $K[x]$ such that $\nu(x)=\nu(a)=1$ for some $a\in K$. Take $q(x)=x^2+1$ (which can be irreducible, for instance, if $K=\mathbb{R}$ or $K=\mathbb F_p$ and $-1$ is not a quadratic residue $\mod p$). Since $x^2-a^2=(x^2+1)-(a^2+1)$ we have \[ \nu_q(x^2-a^2)=\min\{\nu(x^2+1),\nu(a^2+1)\}=0. \] On the other hand, $\nu_q(x+a)=\nu(x+a)\geq\min\{\nu(a),\nu(x)\}=1$ (and the same holds for $\nu_q(x-a)$). Hence \[ \nu_q(x^2-a^2)=0<1+1\leq\nu_q(x-a)+\nu_q(x+a) \] which shows that $\nu_q$ is not a valuation. \end{Exa}
If $f=f_0+f_1q+\ldots+f_nq^n$ is the $q$-standard decomposition of $f$ we set \[ S_q(f):=\{i\in\{0,\ldots, n\}\mid \nu(f_iq^i)=\nu_q(f)\}\mbox{ and }\delta_q(f)=\max S_q(f). \] \begin{Prop}\label{proptruncakeypolval} If $Q$ is a key polynomial, then $\nu_Q$ is a valuation of $K[x]$. \end{Prop} \begin{proof} One can easily see that $\nu_Q(f+g)\geq\min\{\nu_Q(f),\nu_Q(g)\}$ for every $f,g\in K[x]$. It remains to prove that $\nu_Q(fg)=\nu_Q(f)+\nu_Q(g)$ for every $f,g\in K[x]$. Assume first that $\deg(f)<\deg(Q)$ and $\deg(g)<\deg(Q)$ and let $fg=aQ+c$ be the $Q$-standard expansion of $fg$. By Lemma \ref{lemaonkeypollder} \textbf{(iii)} we have \[ \nu(fg)=\nu(c)<\nu(aQ) \] and hence \[ \nu_Q(fg)=\min\{\nu(aQ),\nu(c)\}=\nu(c)=\nu(fg)=\nu(f)+\nu(g)=\nu_Q(f)+\nu_Q(g). \]
Now assume that $f,g\in K[x]$ are any polynomials and consider the $Q$-expansions \[ f=f_0+\ldots+f_nQ^n\mbox{ and }g=g_0+\ldots+g_mQ^m \] of $f$ and $g$. Then, using the first part of the proof, we obtain \[ \nu_Q(fg)\geq\min_{i,j}\{\nu_Q(f_ig_jQ^{i+j})\}=\min_{i,j}\{\nu_Q(f_iQ^i)+\nu_Q(g_jQ^j)\}=\nu_Q(f)+\nu_Q(g). \] For each $i\in\{0,\ldots,n\}$ and $j\in\{0,\ldots,m\}$, let $f_ig_j=a_{ij}Q+c_{ij}$ be the $Q$-standard expansion of $f_ig_j$. Then, by Lemma \ref{lemaonkeypollder} \textbf{(iii)}, we have \[ \nu(f_iQ^i)+\nu(g_jQ^j)=\nu(f_ig_j)+\nu(Q^{i+j})=\nu(c_{ij})+\nu(Q^{i+j})=\nu(c_{ij}Q^{i+j}). \] Let \[ i_0=\min\{i\mid\nu_Q(f)=\nu(f_iQ^i)\}\mbox{ and }j_0=\min\{j\mid\nu_Q(g)=\nu(g_jQ^j)\}, \] and set $k_0:=i_0+j_0$. Then for every $i<i_0$ or $j<j_0$ we have \begin{equation}\label{eqnatnaksdjs} \min\left\{\nu(a_{ij}Q^{i+j+1}),\nu(c_{ij}Q^{i+j})\right\}=\nu(f_iQ^i)+\nu(g_jQ^j)>\nu(c_{i_0j_0}Q^{k_0}). \end{equation} Let $fg=a_0+a_1Q+\ldots+a_rQ^r$ be the $Q$-standard expansion of $fg$. Then \[ a_{k_0}=\sum_{i+j+1=k_0}a_{ij}+\sum_{i+j=k_0}c_{ij}. \] This and equation (\ref{eqnatnaksdjs}) give us that \[ \nu(a_{k_0}Q^{k_0})=\nu(c_{i_0j_0}Q^{k_0})=\nu(f_{i_0}Q^{i_0})+\nu(g_{j_0}Q^{j_0})=\nu_Q(f)+\nu_Q(g). \] Therefore, \[ \nu_Q(fg)=\min_{0\leq k\leq r}\{\nu(a_kQ^k)\}\leq \nu_Q(f)+\nu_Q(g), \] which completes the proof.
\end{proof}
\begin{Prop}\label{Propdificil} Let $Q\in K[x]$ be a key polynomial and set $\epsilon:=\epsilon(Q)$. For any $f\in K[x]$ the following hold: \begin{description} \item[(i)] For any $b\in\mathbb{N}$ we have \begin{equation}\label{eqthatcompvalutrunc} \frac{\nu_Q(f)-\nu_Q(\partial_bf)}{b}\leq \epsilon; \end{equation} \item[(ii)] If $S_Q(f)\neq\{0\}$, then the equality in (\ref{eqthatcompvalutrunc}) holds for some $b\in\mathbb{N}$;
\item[(iii)] If for some $b\in\mathbb{N}$, the equality in (\ref{eqthatcompvalutrunc}) holds and $\nu_Q(\partial_bf)=\nu(\partial_bf)$, then $\epsilon(f)\geq\epsilon$. If in addition, $\nu(f)>\nu_Q(f)$, then $\epsilon(f)>\epsilon$. \end{description} \end{Prop} Fix a key polynomial $Q$ and $h\in K[x]$ with $\deg(h)<\deg(Q)$. Then, for every $b\in\mathbb{N}$ the Leibnitz rule for derivation gives us that \begin{equation} \partial_b(hQ^n)=\sum_{b_0+\ldots+b_r=b}T_b(b_0,\ldots,b_r) \end{equation} where \[ T_b(b_0,\ldots, b_r):=\partial_{b_0}h\left(\prod_{i=1}^r\partial_{b_i}Q\right)Q^{n-r}. \]
In order to prove Proposition \ref{Propdificil}, we will need the following result: \begin{Lema}\label{Lemamagic3} Let $Q$ be a key polynomial, $h\in K[x]$ with $\deg(h)<\deg(Q)$ and set $\epsilon:=\epsilon(Q)$. For any $b\in\mathbb{N}$ we have \[ \nu_Q(T_b(b_0,\ldots,b_r))\geq \nu(hQ^n)-b\epsilon. \] Moreover, if either $b_0>0$ or $b_i\notin I(Q)$ for some $i=1,\ldots, r$, then \[ \nu_Q(T_b(b_0,\ldots,b_r))> \nu(hQ^n)-b\epsilon. \] \end{Lema} \begin{proof} Since $\deg(h)<\deg(Q)$ and $Q$ is a key polynomial we have $\epsilon(h)<\epsilon$. Hence, if $b_0>0$ we have \[ \nu(\partial_{b_0}h)\geq \nu(h)-b_0\epsilon(h)>\nu(h)-b_0\epsilon. \] On the other hand, for every $i=1,\ldots, r$, by definition of $\epsilon$ we have \[ \nu(\partial_{b_i}Q)\geq \nu(Q)-b_i\epsilon, \] and if $b_i\notin I(Q)$ we have \[ \nu(\partial_{b_i}Q)> \nu(Q)-b_i\epsilon. \] Since $\nu_Q(\partial_{b_0}h)=\nu(\partial_{b_0}h)$ and $\nu_Q(\partial_{b_i}Q)=\nu(\partial_{b_i}Q)$, we have \begin{displaymath} \begin{array}{rcl} \nu_Q(T_b(b_0,\ldots,b_r))&=&\displaystyle\nu_Q\left(\partial_{b_0}h\left(\prod_{i=1}^r\partial_{b_i}Q\right)Q^{n-r}\right)\\
&=&\displaystyle\nu_Q(\partial_{b_0}h)+\sum_{i=1}^r\nu_Q(\partial_{b_i}Q)+(n-r)\nu_Q(Q)\\
&\geq &\displaystyle\nu(h)-b_0\epsilon+\sum_{i=1}^r\left(\nu(Q)-b_i\epsilon\right)+(n-r)\nu(Q)\\
&\geq &\nu(hQ^n)-b\epsilon. \end{array} \end{displaymath} Moreover, if $b_0>0$ or $b_i\notin I(Q)$ for some $i=1,\ldots, r$, then the inequality above is strict. \end{proof}
\begin{Cor}\label{Coroaboutderib} For every $b\in\mathbb{N}$ we have $\nu_Q\left(\partial_b(aQ^n)\right)\geq\nu(aQ^n)-b\epsilon$. \end{Cor}
\begin{proof}[Proof of Proposition \ref{Propdificil}] \textbf{(i)} Take any $f\in K[x]$ and consider its $Q$-standard expansion $f=f_0+f_1Q+\ldots+f_nQ^n$. For each $i=0,\ldots,n$, Corollary \ref{Coroaboutderib} gives us that \[ \nu_Q\left(\partial_b(f_iQ^i)\right)\geq \nu (f_iQ^i)-b\epsilon. \] Hence, \[ \nu_Q\left(\partial_b(f)\right)\geq\min_{0\leq i\leq n}\{\nu_Q(f_iQ^i)\}\geq\min_{0\leq i\leq n}\{\nu(f_iQ^i)-b\epsilon\}=\nu_Q(f)-b\epsilon. \] \textbf{(ii)} Assume that $S_Q(f)\neq \{0\}$ and set $j_0=\min S_Q(f)$. Then $j_0=p^er$ for some $e\in\mathbb{N}\cup\{0\}$ and some $r\in\mathbb{N}$ with $(r,p)=1$. We set $b:=p^eb(Q)$ and will prove that $\nu_Q(\partial_b(f))=\nu_Q(f)-b\epsilon$.
Write \[ f_{j_0}\left(\partial_{b(Q)}Q\right)^{p^e}=rQ+h \] for some $r,h\in K[x]$ and $\deg(h)<\deg(Q)$ (note that $h\ne0$ because $Q$ is irreducible and $Q\nmid f_{j_0}$ and $Q\nmid \partial_{b(Q)}Q$). Then Lemma \ref{lemaonkeypollder} \textbf{(iii)} gives us that \[ \nu(h)=\nu\left(f_{j_0}(\partial_{b(Q)}Q)^{p^e}\right). \] This implies that \begin{equation}\label{equationboa} \nu\left(hQ^{j_0-p^e}\right)=\nu_Q(f)-b\epsilon. \end{equation} Indeed, we have \begin{displaymath} \begin{array}{rcl} \nu\left(hQ^{j_0-p^e}\right)&=& \nu(h)+\nu\left(Q^{j_0-p^e}\right)=\nu\left(f_{j_0}(\partial_{b(Q)}Q)^{p^e}\right)+ \nu\left(Q^{j_0-p^e}\right)\\
&=&\nu(f_{j_0})+p^e\nu\left(\partial_{b(Q)}Q\right)+(j_0-p^e)\nu(Q)\\
&=&\nu(f_{j_0})+p^e\left(\nu\left(Q\right)-b(Q)\epsilon\right)+(j_0-p^e)\nu(Q)\\
&=&\nu(f_{j_0})+j_0\nu(Q)-p^eb(Q)\epsilon\\
&=&\nu(f_{j_0}Q^{j_0})-p^eb(Q)\epsilon=\nu_Q(f)-b\epsilon.\\
\end{array} \end{displaymath}
Since $f=f_0+f_1Q+\ldots+f_nQ^n$, we have $\partial_b(f)=\partial_b(f_0)+\partial_b(f_1Q)\ldots+\partial_b(f_nQ^n)$. For each $j=0,\ldots, n$, if $j\notin S_Q(f)$, then \[ \nu_Q\left(\partial_b(f_jQ^j)\right)\geq \nu_Q(f_jQ^j)-b\epsilon>\nu_Q(f)-b\epsilon. \] We set \[ h_1=\sum_{j\notin S_Q(f)}f_jQ^j. \] Then $\nu_Q(h_1)>\nu_Q(f)-b\epsilon$.
For each $j\in S_Q(f)$ the term $\partial_b(f_jQ^j)$ can be written as a sum of terms of the form $T_b(b_0,\ldots,b_r)$. For each $T_b(b_0,\ldots,b_r)$ we have the following cases:
\textbf{Case 1:} $b_0>0$ or $b_i\notin I(Q)$ for some $i$.\\ In this case, by Lemma \ref{Lemamagic3} we have $\nu_Q(T_b(b_0,\ldots,b_r))>\nu_Q(f)-b\epsilon$. In particular, if $h_2$ is the sum of all these terms, then $\nu_Q(h_2)>\nu_Q(f)-b\epsilon$.
\textbf{Case 2:} $b_0=0$ and $b_i\in I(Q)$ for every $i=1,\ldots,r$ but $b_{i_0}\neq b(Q)$ for some $i_0=1,\ldots,r$.\\ This implies, in particular, that $j\geq j_0$ and since $b=p^eb(Q)$ we must have $r<p^e$. Hence \[ T_b(b_0,b_1,\ldots,b_r)=\partial_{b_0}f_j\left(\prod_{i=1}^r\partial_{b_i}Q\right)Q^{j-r}=sQ^{j_0-p^e+1} \] for some $s\in K[x]$.
\textbf{Case 3:} $b_0=0$, $j>j_0$ and $b_i=b(Q)$ for every $i=1,\ldots,r$.\\ Since $b=p^eb(Q)$, $b_i=b(Q)$ and $\displaystyle\sum_{i=1}^rb_i=b$ we must have $r=p^e$. Hence \[ T_b(b_0,b_1,\ldots,b_r)=f_j\left(\partial_{b(Q)}Q\right)^{p^e}Q^{j-p^e}=s'Q^{j_0-p^e+1} \] for some $s'\in K[x]$.\\
\textbf{Case 4:} $b_0=0$, $j=j_0$ and $b_i=b(Q)$ for every $i=1,\ldots,r$.\\ In this case we have \begin{equation}\label{caseintport} \begin{array}{rcl} T_b(b_0,b_1,\ldots,b_r)&=&f_{j_0}\left(\partial_{b(Q)}Q\right)^{p^e}Q^{j_0-p^e}\\
&=&\left(h-rQ\right)Q^{j_0-p^e}\\
&=&hQ^{j_0-p^e}-rQ^{j_0-p^e+1}. \end{array} \end{equation}
Observe that the number of times that the term (\ref{caseintport}) appears in $\partial_b(f_{j_0}Q^{j_0})$ is $\binom{j_0}{p^e}$, that is, the number of ways that one can choose a subset with $p^e$ elements in a set of $j_0$ elements.
Therefore, we can write \[ \partial_b(f)=\binom{j_0}{p^e}hQ^{j_0-p^e}+\left(s+s'-\binom{j_0}{p^e}r\right)Q^{j_0-p^e+1}+h_1+h_2 \] Since $p\nmid \binom{j_0}{p^e}$ the equation (\ref{equationboa}) gives us that \[ \nu\left(\binom{j_0}{p^e}hQ^{j_0-p^e}\right)=\nu_Q(f)-b\epsilon. \] Then \[ \nu_Q\left(\binom{j_0}{p^e}hQ^{j_0-p^e}+\left(s+s'-\binom{j_0}{p^e}r\right)Q^{j_0-p^e+1}\right)\leq \nu_Q(f)-b\epsilon. \] This and the fact that $\nu_Q(h_1+h_2)>\nu_Q(f)-b\epsilon$ imply that $\nu_Q\left(\partial_b(f)\right)\leq\nu_Q(f)-b\epsilon$. This concludes the proof of \textbf{(ii)}.
\textbf{(iii)} The assumptions on $b$ give us \[ \frac{\nu_Q(f)-\nu_Q(\partial_bf)}{b}= \epsilon \] and \[ \nu_Q(\partial_bf)=\nu(\partial_bf). \] Consequently, \[ \epsilon(f)\geq \frac{\nu(f)-\nu(\partial_bf)}{b}\geq\frac{\nu_Q(f)-\nu_Q(\partial_bf)}{b}= \epsilon. \] In the inequality above, one can see that if $\nu(f)>\nu_Q(f)$, then $\epsilon(f)>\epsilon$.
\end{proof}
\begin{Prop}\label{Propcompkeypol} For two key polynomials $Q,Q'\in K[x]$ we have the following: \begin{description} \item[(i)] If $\deg(Q)<\deg(Q')$, then $\epsilon(Q)<\epsilon(Q')$; \item[(ii)] If $\epsilon(Q)<\epsilon(Q')$, then $\nu_Q(Q')<\nu(Q')$; \item[(iii)] If $\deg(Q)=\deg(Q')$, then \begin{equation}\label{eqwhdegsame} \nu(Q)<\nu(Q')\Longleftrightarrow \nu_Q(Q')<\nu(Q')\Longleftrightarrow \epsilon(Q)<\epsilon(Q'). \end{equation} \end{description} \end{Prop} \begin{proof} Item \textbf{(i)} follows immediately from the the definition of key polynomial (in fact, the same holds if we substitute $Q$ for any $f\in K[x]$).
In order to prove \textbf{(ii)} we set $\epsilon:=\epsilon(Q)$ and $b':=b(Q')$. By \textbf{(i)} of Proposition \ref{Propdificil}, we have \[ \nu_Q(Q')\leq \nu_Q(\partial_{b'}Q')+b'\epsilon. \] Since $\epsilon(Q)<\epsilon(Q')$, we also have \[ \nu(\partial_{b'}Q')+b'\epsilon< \nu(\partial_{b'}Q')+b'\epsilon(Q')=\nu(Q'). \] This, and the fact that $\nu_Q(\partial_{b'}Q')\leq \nu(\partial_{b'}Q')$, imply that $\nu_Q(Q')<\nu(Q')$.
Now assume that $\deg(Q)=\deg(Q')$ and let us prove (\ref{eqwhdegsame}). Since $$ \deg(Q)=\deg(Q') $$ and both $Q$ and $Q'$ are monic, the $Q$-standard expansion of $Q'$ is given by $$ Q'=Q+(Q-Q'). $$ Hence \[ \nu_Q(Q')=\min\{\nu(Q),\nu(Q-Q')\}. \] The first equivalence follows immediately from this. In view of part \textbf{(ii)}, it remains to prove that if $\nu_Q(Q')<\nu(Q')$, then $\epsilon(Q)<\epsilon(Q')$. Since $\nu_Q(Q')<\nu(Q')$ we have $S_Q(Q')\neq \{0\}$. Hence, by Proposition \ref{Propdificil} \textbf{(ii)}, the equality holds in (\ref{eqthatcompvalutrunc}) (for $f=Q'$) for some $b\in\mathbb{N}$. Moreover, since $\deg(Q)=\deg(Q')$, we have $\deg(\partial_bQ')<\deg(Q)$ and consequently $\nu_Q(\partial_bQ')=\nu(\partial_bQ')$. Then Proposition \ref{Propdificil} \textbf{(iii)} implies that $\epsilon(Q)<\epsilon(Q')$. \end{proof}
For a key polynomial $Q\in K[x]$, let \[ \alpha(Q):=\min\{\deg(f)\mid \nu_Q(f)< \nu(f)\} \] (if $\nu_Q=\nu$, then set $\alpha(Q)=\infty$) and \[ \Psi(Q):=\{f\in K[x]\mid f\mbox{ is monic},\nu_Q(f)< \nu(f)\mbox{ and }\alpha(Q)=\deg (f)\}. \] \begin{Lema}\label{lemmapsikeypoly} If $Q$ is a key polynomial, then every element $Q'\in\Psi(Q)$ is also a key polynomial. Moreover, $\epsilon(Q)<\epsilon(Q')$. \end{Lema} \begin{proof} By assumption, we have $\nu_Q(Q')<\nu(Q')$, hence $S_{Q}(Q')\neq \{0\}$. This implies, by Proposition \ref{Propdificil} \textbf{(ii)}, that there exists $b\in\mathbb{N}$ such that \[ \nu_Q(Q')-\nu_Q(\partial_b Q')=b\epsilon(Q). \] Since $\deg(\partial_b Q')<\deg(Q')=\alpha(Q)$, we have $\nu_Q(\partial_b Q')=\nu(\partial_b Q')$. Consequently, by Proposition \ref{Propdificil} \textbf{(iii)}, $\epsilon(Q)<\epsilon(Q')$.
Now take any polynomial $f\in K[x]$ such that $\deg(f)<\deg(Q')=\alpha(Q)$. In particular, $\nu_Q(f)=\nu(f)$. Moreover, for every $b\in\mathbb{N}$, $\deg(\partial_b f)<\deg(Q')= \alpha(Q)$ which implies that $\nu_Q(\partial_bf)=\nu(\partial_b f)$. Then, for every $b\in\mathbb{N}$, \[ \frac{\nu(f)-\nu(\partial_bf)}{b}=\frac{\nu_Q(f)-\nu_Q(\partial_bf)}{b}\leq \epsilon(Q)<\epsilon(Q'). \] This implies that $\epsilon(f)<\epsilon(Q')$, which shows that $Q'$ is a key polynomial.
\end{proof} \begin{Teo}\label{definofkeypol} A polynomial $Q$ is a key polynomial if and only if there exists a key polynomial $Q_-\in K[x]$ such that $Q\in \Psi(Q_-)$ or the following conditions hold: \begin{description} \item[(K1)] $\alpha(Q_-)=\deg (Q_-)$ \item[(K2)] the set $\{\nu(Q')\mid Q'\in\Psi(Q_-)\}$ does not contain a maximal element \item[(K3)] $\nu_{Q'}(Q)<\nu(Q)$ for every $Q'\in \Psi(Q_-)$ \item[(K4)] $Q$ has the smallest degree among polynomials satisfying \textbf{(K3)}. \end{description} \end{Teo}
\begin{proof} We will prove first that if such $Q_-$ exists, then $Q$ is a key polynomial. The case when $Q\in \Psi(Q_-)$ follows from Lemma \ref{lemmapsikeypoly}. Assume now that \textbf{(K1) - (K4)} hold. Take $f\in K[x]$ such that $\deg(f)<\deg(Q)$. This implies that $\deg(\partial_bQ)<\deg(Q)$ and $\deg(\partial_bf)<\deg(Q)$ for every $b\in\mathbb{N}$. Hence, by \textbf{(K4)}, there exists $Q'\in\Psi(Q_-)$ such that \[ \nu_{Q'}(f)=\nu(f), \nu_{Q'}(\partial_bf)=\nu(\partial_bf) \mbox{ and }\nu_{Q'}(\partial_bQ)=\nu(\partial_bQ)\mbox{ for every }b\in\mathbb{N}. \] We claim that $\epsilon(Q')<\epsilon(Q)$. If not, by Proposition \ref{Propcompkeypol} \textbf{(i)}, we would have $\deg(Q)\leq\deg(Q')$. Since $\nu_{Q'}(Q)<\nu(Q)$, this implies that $\deg(Q)=\deg(Q')$. This and Proposition \ref{Propcompkeypol} \textbf{(iii)} give us that $\epsilon(Q')<\epsilon(Q)$ which is a contradiction.
Now, \[ \epsilon(f)\leq \frac{\nu(f)-\nu(\partial_bf)}{b}=\frac{\nu_{Q'}(f)-\nu_{Q'}(\partial_bf)}{b}\leq\epsilon(Q')<\epsilon(Q). \] Hence $Q$ is a key polynomial.
For the converse, take a key polynomial $Q\in K[x]$ and consider the set \[ \mathcal S:=\{Q'\in K[x]\mid Q'\mbox{ is a key polynomial and }\nu_{Q'}(Q)<\nu(Q)\}. \] Observe that $\mathcal S\neq\emptyset$. Indeed, if $\deg(Q)>1$, then every key polynomial $x-a\in\mathcal S$. If $Q=x-a$, then there exits $b\in K$ such that $\nu(b)<\min\{\nu(a),\nu(x)\}$. Therefore, $x-b\in \mathcal S$.
If there exists a key polynomial $Q_-\in \mathcal S$ such that $\deg(Q)=\deg(Q_-)$ , then we have $Q\in \Psi(Q_-)$ and we are done. Hence, assume that every polynomial $Q'\in \mathcal S$ has degree smaller that $\deg(Q)$.
Assume that there exists $Q_-\in \mathcal S$ such that for every $Q'\in \mathcal S$ we have \begin{equation}\label{eqmaxphi} (\deg(Q_-),\nu(Q_-))\geq ((\deg(Q'),\nu(Q')) \end{equation} in the lexicographical ordering. We claim that $Q\in \Psi(Q_-)$. If not, there would exist a key polynomial $Q''$ such that $\nu_{Q_-}(Q'')<\nu(Q'')$ and $\deg(Q'')<\deg(Q)$. Since $\deg(Q'')<\deg(Q)$ Proposition \ref{Propcompkeypol} \textbf{(i)} and \textbf{(ii)} give us that $\nu_{Q''}(Q)<\nu(Q)$. Hence $Q''\in \mathcal S$. The inequality (\ref{eqmaxphi}) gives us that $\deg(Q'')\leq\deg(Q_-)$. On the other hand, since $\nu_{Q_-}(Q'')<\nu(Q'')$ we must have $\deg(Q_-)=\deg(Q'')$. Hence, Proposition \ref{Propcompkeypol} \textbf{(iii)} gives us that $\nu(Q_-)<\nu(Q'')$ and this is a contradiction to the inequality (\ref{eqmaxphi}).
Now assume that for every $Q'\in \mathcal S$, there exists $Q''\in \mathcal S$ such that \begin{equation}\label{eqmaxphihas} (\deg(Q'),\nu(Q'))<(\deg(Q''),\nu(Q'')) \end{equation} in the lexicographical ordering. Take $Q_-\in\mathcal S$ such that $\deg(Q_-)\geq \deg(Q')$ for every $Q'\in\mathcal S$. We will show that the conditions \textbf{(K1) - (K4)} are satisfied. By (\ref{eqmaxphihas}), there exists $Q''\in \mathcal S$ such that \begin{equation}\label{eqbanolimit} (\deg(Q_-),\nu(Q_-))<(\deg(Q''),\nu(Q'')). \end{equation} In particular, $\deg(Q_-)=\deg(Q'')$ and $\nu(Q_-)<\nu(Q'')$. Proposition \ref{Propcompkeypol} \textbf{(iii)} gives us that $\nu_{Q_-}(Q'')<\nu(Q'')$. Hence $\alpha(Q_-)=\deg(Q_-)$ and we have proved \textbf{(K1)}. If $Q'\in\Psi(Q_-)$, then $\deg(Q')=\deg(Q_-)<\deg(Q)$ and hence $\nu_{Q'}(Q)<\nu(Q)$. This implies that $Q'\in\mathcal S$. The equation (\ref{eqbanolimit}) tells us that $\{\nu(Q')\mid Q'\in\Psi(Q_-)\}$ has no maximum, so we have proved \textbf{(K2)}. Now take any element $Q'\in\Psi(Q_-)$. Then $\deg(Q')<\deg(Q)$ and Proposition \ref{Propcompkeypol} \textbf{(i)} and \textbf{(ii)} give us that $\nu_{Q'}(Q)<\nu(Q)$. This proves \textbf{(K3)}. Take a polynomial $\widetilde{Q}$ with $\nu_{Q'}(\widetilde{Q})<\nu(\widetilde{Q})$ for every $Q'\in\Psi(Q_-)$ with minimal degree possible. We want to prove that $\deg(\widetilde Q)=\deg(Q)$. Assume, aiming for a contradiction, that $\deg(\widetilde{Q})<\deg(Q)$. The first part of the proof gives us that $\widetilde Q$ is a key polynomial. Fix $Q'\in\Psi(Q_-)$. Then $\nu_{Q'}(\widetilde Q)<\nu(\widetilde Q)$ and consequently $\deg(\widetilde{Q})=\deg(Q')=\deg(Q_-)$. Therefore $\nu(Q')<\nu(\widetilde Q)$ for every $Q'\in \Psi(Q_-)$, which is a contradiction to (\ref{eqmaxphihas}). This concludes our proof. \end{proof}
\begin{Def} When conditions \textbf{(K1) - (K4)} of Theorem \ref{definofkeypol} are satisfied, we say that $Q$ is a \textbf{limit key polynomial}. \end{Def}
\begin{Obs} Observe that as a consequence of the proof we obtain that $$ \epsilon(Q_-)<\epsilon(Q). $$ \end{Obs}
\begin{proof}[Proof of Theorem \ref{Theoremexistencecompleteseqkpol}] Consider the set \[ \Gamma_0:=\{\nu(x-a)\mid a\in K\}. \] We have two possibilities:
\begin{itemize} \item $\Gamma_0$ has a maximal element \end{itemize} Set $Q_0:=x-a_0$ where $a_0\in K$ is such that $\nu(x-a_0)$ is a maximum of $\Gamma_0$. If $\nu=\nu_{Q_0}$ we are done, so assume that $\nu\neq\nu_{Q_0}$. If the set \[ \{\nu(Q)\mid Q\in \Psi(Q_0)\} \] has a maximum, choose $Q_1\in \Psi(Q_0)$ such that $\nu(Q_1)$ is this maximum. If not, choose $Q_1$ as any polynomial in $\Psi(Q_0)$. Set $\Lambda_1:=\{Q_0,Q_1\}$ (ordered by $Q_0<Q_1$).
\begin{itemize} \item $\Gamma_0$ does not have a maximal element \end{itemize} For every $\gamma\in \Gamma_0$ set $Q_\gamma:=x-a_\gamma$ for some $a_\gamma\in K$ such that $\nu(x-a_\gamma)=\gamma$. If for every $f\in K[x]$, there exists $\gamma\in \Gamma_0$ such that $\nu(f)=\nu_{Q_\gamma}(f)$ we are done. If not, let $Q$ be a polynomial of minimal degree among all the polynomials for which $\nu_{Q_\gamma}(Q)<\nu(Q)$ for every $\gamma\in \Gamma_0$. If $\alpha(Q)=\deg(Q)$ and the set $\{\nu(Q')\mid Q'\in \Psi(Q)\}$ contains a maximal element, choose $Q_1\in \Psi(Q)$ such that $\nu(Q_1)\geq \nu(Q')$ for every $Q'\in \Psi(Q)$. If not, set $Q_1:=Q$. Set $\Lambda_1:=\{Q_\gamma\mid \gamma\in \Gamma_0\}\cup\{Q_1\}$ (ordered by $Q_1>Q_\gamma$ for every $\gamma\in \Gamma$ and $Q_\gamma>Q_{\gamma'}$ if $\gamma>\gamma'$).
Observe that in either case, $\deg(Q_1)>\deg(Q_0)$ and for $Q,Q'\in \Lambda_1$, $Q<Q'$ if and only if $\epsilon(Q)<\epsilon(Q')$. Moreover, if $\alpha(Q_1)=\deg(Q_1)$, then $\{\nu(Q)\mid Q\in\Psi(Q_1)\}$ does not have a maximum.
Assume that for some $i\in\mathbb{N}$, there exists a totally ordered set $\Lambda_i$ consisting of key polynomials with the following properties: \begin{description} \item[(i)] there exist $Q_0,Q_1,\ldots,Q_i\in \Lambda_i$ such that $Q_i$ is the last element of $\Lambda_i$ and $\deg(Q_0)<\deg(Q_1)<\ldots<\deg(Q_i)$. \item[(ii)] if $\alpha(Q_i)=\deg(Q_i)$, then $\Gamma_i:=\{\nu(Q)\mid Q\in \Psi(Q_i)\}$ does not have a maximum. \item[(iii)] for $Q,Q'\in \Lambda_i$, $Q<Q'$ if and only if $\epsilon(Q)<\epsilon(Q')$. \end{description} If $\nu_{Q_i}\neq \nu$, then we will construct a set $\Lambda_{i+1}$ of key polynomials having the same properties (changing $i$ by $i+1$).
Since $\nu_{Q_i}\neq \nu$, the set $\Psi(Q_i)$ is not empty. We have two cases: \begin{itemize} \item $\alpha(Q_i)>\deg(Q_i)$. \end{itemize} If $\Gamma_i$ has a maximum, take $Q_{i+1}\in\Psi(Q_i)$ such that $\nu(Q_{i+1})\geq \Gamma_i$. Otherwise, choose $Q_{i+1}$ to be any element of $\Psi(Q_i)$. Observe that if $\alpha(Q_{i+1})=\deg(Q_{i+1})$, then $\Gamma_{i+1}$ does not have a maximum. Set $\Lambda_{i+1}=\Lambda_i\cup\{Q_{i+1}\}$ with the extension of the order in $\Lambda_i$ obtained by setting $Q_{i+1}>Q$ for every $Q\in \Lambda_i$. \begin{itemize} \item $\alpha(Q_i)=\deg(Q_i)$. \end{itemize} By assumption, the set $\Gamma_i$ does not have a maximum. For each $\gamma\in\Gamma_i$, choose a polynomial $Q_\gamma\in \Psi(Q_i)$ such that $\nu(Q_\gamma)=\gamma$. If for every $f\in K[x]$, there exists $\gamma\in \Gamma_i$ such that $\nu_{Q_\gamma}(f)=\nu(f)$, then we are done. Otherwise, choose a monic polynomial $Q$, of smallest degree possible, such that $\nu_{Q'}(Q)<\nu(Q)$ for every $Q'\in\Psi(Q_i)$. If $\alpha(Q)=\deg(Q)$ and $\{\nu(Q')\mid Q'\in\Psi(Q)\}$ has a maximum, we choose $Q_{i+1}$ such that $\nu(Q_{i+1})\geq\{\nu(Q')\mid Q'\in\Psi(Q)\}$. Otherwise we set $Q_{i+1}=Q$. Then set \[ \Lambda_{i+1}:=\Lambda_i\cup\{Q_\gamma\mid \gamma\in\Gamma_i\}\cup\{Q_{i+1}\}, \] with the extension of the order of $\Lambda_i$ given by $$ Q_{i+1}>Q'\mbox{ for every }Q'\in \Lambda_{i+1}\setminus\{Q_{i+1}\}, $$ $Q_\gamma> Q'$ for every $\gamma\in \Gamma_i$ and $Q'\in\Lambda_i$ and $Q_\gamma>Q_{\gamma'}$ for $\gamma,\gamma'\in \Gamma_i$ with $\gamma>\gamma'$.
In all cases, the set $\Lambda_{i+1}$ has the properties \textbf{(i)}, \textbf{(ii)} and \textbf{(iii)}.
Assume now that for every $i\in\mathbb{N}$ the sets $\Lambda_i$ and $\Lambda_{i+1}$ can be constructed. Then we can construct a set \[ \Lambda_\infty:=\bigcup_{i=1}^\infty\Lambda_i \] of key polynomials having the property that for $Q,Q'\in \Lambda_\infty$, $Q<Q'$ if and only if $\epsilon(Q)<\epsilon(Q')$ and there are polynomials $Q_0,\ldots,Q_i,\ldots\in \Lambda_\infty$ such that $$ \deg(Q_{i+1})>\deg(Q_i) $$ for every $i\in\mathbb{N}$. This means that for every $f\in K[x]$ there exists $i\in\mathbb{N}$ such that $\deg(f)<\deg(Q_i)$, which implies that $\nu_{Q_i}(f)=\nu(f)$. Therefore, $\Lambda_\infty$ is a complete set of key polynomials for $\nu$. \end{proof}
Observe that at each stage, the same construction would work if we replaced $\Gamma_i$ by any cofinal set $\Gamma_i'$ of $\Gamma_i$. Hence, if the rank of $\nu$ is equal to 1, then we can choose $\Gamma_i'$ to have order type at most $\omega$. Then, from the construction of the sets $\Lambda_i$ and $\Lambda_\infty$, we can conclude the following: \begin{Cor} If the rank of $\nu$ is equal to one, then there exists a complete sequence of key polynomials of $\nu$ with order type at most $\omega\times\omega$. \end{Cor}
\section{Pseudo-convergent sequences} The next two theorems justify the definitions of algebraic and transcendental pseudo-convergent sequences. \begin{Teo}[Theorem 2 of \cite{Kap}] If $\{a_\rho\}_{\rho<\lambda}$ is a pseudo-convergent sequence of transcendental type, without a limit in $K$, then there exists an immediate transcendental extension $K(z)$ of $K$ defined by setting $\nu(f(z))$ to be the value $\nu(f(a_{\rho_f}))$ as in condition (\ref{condforpscstotra}). Moreover, for every valuation $\mu$ in some extension $K(u)$ of $K$, if $u$ is a pseudo-limit of $\{a_\rho\}_{\rho<\lambda}$, then there exists a value preserving $K$-isomorphism from $K(u)$ to $K(z)$ taking $u$ to $z$. \end{Teo}
\begin{Teo}[Theorem 3 of \cite{Kap}]\label{thmonalgimmext} Let $\{a_\rho\}_{\rho<\lambda}$ be a pseudo-convergent sequence of algebraic type, without a limit in $K$, $q(x)$ a polynomial of smallest degree for which (\ref{condforpscstoalg}) holds and $z$ a root of $q(x)$. Then there exists an immediate algebraic extension of $K$ to $K(z)$ defined as follows: for every polynomial $f(x)\in K[x]$, with $\deg f<\deg q$ we set $\nu(f(z))$ to be the value $\nu(f(a_{\rho_f}))$ as in condition (\ref{condforpscstotra}). Moreover, if $u$ is a root of $q(x)$ and $\mu$ is some extension $K(u)$ of $K$ making $u$ a pseudo-limit of $\{a_\rho\}_{\rho<\lambda}$, then there exists a value preserving $K$-isomorphism from $K(u)$ to $K(z)$ taking $u$ to $z$. \end{Teo}
For the rest of this paper, let $\{a_\rho\}_{\rho<\lambda}$ be a pseudo-convergent sequence for the valued field $(K,\nu)$, without a limit in $K$. For each $\rho<\lambda$, we denote $\nu_\rho=\nu_{x-a_\rho}$. For a polynomial $f(x)\in K[x]$ and $a\in K$ we consider the Taylor expansion of $f$ at $a$ given by \[ f(x)=f(a)+\partial_1f(a)(x-a)+\ldots+\partial_nf(a)(x-a)^n. \] Assume that $\{a_\rho\}_{\rho<\lambda}$ fixes the value of the polynomials $\partial_if(x)$ for every $1\leq i\leq n$. We denote by $\beta_i$ this fixed value. \begin{Lema}[Lemma 8 of \cite{Kap}]\label{lemmakaplvalpol} There is an integer $h$, which is a power of $p$, such that for sufficiently large $\rho$ \[ \beta_i+i\gamma_\rho>\beta_h+h\gamma_\rho\mbox{ whenever }i\ne h\mbox{ and } \nu(f(a_\rho))=\beta_h+h\gamma_\rho. \] \end{Lema}
\begin{Cor}\label{correlanurhowithnu} If $\{a_\rho\}_{\rho<\lambda}$ fixes the value of $f(x)$, then $\nu_\rho(f(x))=\nu(f(x))$. On the other hand, if $\{a_\rho\}_{\rho<\lambda}$ does not fix the value of $f(x)$, then $\nu_\rho(f(x))<\nu(f(x))$ for every $\rho<\lambda$. \end{Cor} \begin{proof} By definition of $\nu_\rho$ we have \[ \nu_\rho(f(x))=\min_{0\leq i\leq n}\{\nu(\partial_if(a_\rho)(x-a_\rho)^i)\}=\min_{0\leq i\leq n}\{\beta_i+i\gamma_\rho\}, \] where $\beta_0:=\nu(f(a_\rho))$. This implies, using the lemma above, that \[ \nu_\rho(f(x))=\nu(f(a_\rho)). \] If $\{a_\rho\}_{\rho<\lambda}$ fixes the value of $f(x)$, then $\nu(f(a_\rho))=\nu(f(x))$ for $\rho$ sufficiently large. Thus $\nu_\rho(f(x))=\nu(f(x))$. On the other hand, if $\{a_\rho\}_{\rho<\lambda}$ does not fix the value of $f(x)$, then $\nu(f(x))>\nu(f(a_\rho))=\nu_\rho(f(x))$ for every $\rho<\lambda$. \end{proof}
\begin{proof}[Proof of Theorem \ref{compthemkppsc}] If $\{a_\rho\}_{\rho<\lambda}$ is of transcendental type it fixes, for any polynomial $f(x)\in K[x]$, the values of the polynomials $\partial_if(x)$ for every $0\leq i\leq n$ (here $\partial_0f:=f$). Hence, Corollary \ref{correlanurhowithnu} implies that $\nu_\rho(f(x))=\nu(f(x))$ for sufficiently large $\rho<\lambda$, which is what we wanted to prove.
Now assume that $\{a_\rho\}_{\rho<\lambda}$ is of algebraic type. Take $\rho<\lambda$ such that $$ \nu(q(a_\tau))>\nu(q(a_\sigma)) $$ for every $\rho<\sigma<\tau<\lambda$ and set $Q_-=x-a_\rho$. Then \[ \nu_{Q_-}(x-a_\sigma)=\nu_{Q_-}(x-a_\rho+a_\rho-a_\sigma)=\nu(x-a_\rho)<\nu(x-a_\sigma) \] for every $\rho<\sigma<\lambda$. This implies that $\alpha(Q-)=1$ and then $\alpha(Q_-)=\deg (Q_-)$. Consequently, \textbf{(K1)} is satisfied. Moreover, \[ \Psi(Q_-)=\{x-a\mid \nu_{Q_-}(x-a)<\nu(x-a)\}. \] In order to prove \textbf{(K2)} assume, aiming for a contradiction, that $\nu(\Psi(Q_-))$ has a maximum, let us say $\nu(x-a)$. Then, in particular, $\nu(x-a)>\nu(x-a_\sigma)$ for every $\rho<\sigma<\lambda$. This implies that $a\in K$ is a limit of $\{a_\rho\}_{\rho<\lambda}$, which is a contradiction. Condition \textbf{(K3)} and \textbf{(K4)} follow immediately from Corollary \ref{correlanurhowithnu} and the fact that $\{\nu(x-a_\rho)\mid \rho<\lambda\}$ is cofinal in $\nu(\Psi(Q_-))$. \end{proof}
\end{document}
|
arXiv
|
{
"id": "1611.05679.tex",
"language_detection_score": 0.6528453230857849,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{ \bf Inverse problem on a tree-shaped network}
\maketitle
\footnotetext[1]{CNRS ; LAAS ; 7 avenue du colonel Roche, F-31077 Toulouse, France ;\\ Universit\'e de Toulouse ; UPS, INSA, INP, ISAE, UT1, UTM, LAAS ; F-31077 Toulouse, France.\\
E-mail: {\tt [email protected]}} \footnotetext[2]{Graduate School of Mathematical Sciences, University of Tokyo, 3-8-1 Komaba, Tokyo, 153-8914 Japan.\\
E-mail: {\tt [email protected]}}
\abstract{ In this article, we prove a uniqueness result for a coefficient inverse problems regarding a wave, a heat or a Schr\"odinger equation set on a tree-shaped network, as well as the corresponding stability result of the inverse problem for the wave equation. The objective is the determination of the potential on each edge of the network from the additional measurement of the solution at all but one external end-points. Our idea for proving the uniqueness is to use a traditional approach in coefficient inverse problem by Carleman estimate. Afterwards, using an observability estimate on the whole network, we apply a compactness-uniqueness argument and prove the stability for the wave inverse problem. }
{\bf Keywords:} networks, inverse problem, Carleman estimate.
{\bf AMS subject classifications:} 35R30, 93C20, 34B45
\renewcommand{\fnsymbol{footnote}}{\fnsymbol{footnote}}
\section{Introduction and main results}
Let $\Lambda$ be a tree-shaped network composed of $ N+1$ open segments $(e_j)_{j=0, 1, ..., N}$ of length~$\ell_j$, linked by $N_1$ internal node points belonging to the set $\Pi_1$ and let us denote by $\Pi_2$ the set of $N_2$ exterior end-points where only one segment starts. Here we note that $N+1 = N_1+N_2$. By ``tree-shaped network'', we mean that $\Lambda$ does not contain any closed loop.
For any function $f: \Lambda\rightarrow\mathbb{R}$ and any internal node $P\in\Pi_1$ where $n_P$ segments, say $e_1, ..., e_{n_P}$, meet, we set $$ f_j=f\vert_{e_j}: \mbox{the restriction of $f$ to the edge $e_j$}, \mbox{ and} \quad \left[ f \right]_P := \sum_{j=1}^{n_P}f_{j}(P). $$ We consider on this plane $1$-d tree-shaped network $\Lambda$ either wave or heat or even Schr\"odinger equations, with a different potential term $x\mapsto p_j(x)$ on each segment.
Our first, and main, system of interest is the following $1$-d wave equation on the network $\Lambda$: \begin{equation}\label{NW} \left\{ \begin{array}{lll} \partial_t^2 u_j - \partial_x^2 u_j + p_j(x)u_j = 0\quad &\quad\forall j\in\{0,1,..., N\}, (x,t)\in e_j \times (0,T),\\ u(Q,t)=h(t),&\quad \,\forall Q\in \Pi_2, t\in(0,T),\\ u(x,0) = u^0(x), \partial_t u(x,0)=u^1(x),&\quad x\in\Lambda, \end{array} \right. \end{equation} assuming some compatibility condition between the boundary and initial data. Moreover we assume the continuity and what is called the Kirchhoff law at any internal node $P\in\Pi_1$, which are given by \begin{equation}\label{C} u_j(P,t)=u_i(P,t)=:u(P,t),\quad\forall i,j\in\left\{1,...,n_P\right\},\, 0<t<T,\\ \end{equation} \begin{equation}\label{K} \left[ u_x (t)\right]_P :=\sum_{j=1}^{n_P} \partial_x u_{j}(P,t)=0,\quad 0<t<T. \end{equation} Henceforth we choose an orientation of $\Lambda$ such that to two endpoints of each segment $e$, correspond an initial node $I(e)$ and a terminal node $T(e)$. We further define the outward normal derivative $\partial_{n_e}u_j$ at a node $P$ of $e_j$ by $$ \partial_{n_e}u_j(P,t) = \left\{ \begin{array}{lll} -\partial_xu_j(P,T), \quad &\mbox{if $P\in I(e_j)$}, \\ \partial_xu_j(P,T), \quad &\mbox{if $P\in T(e_j)$}. \\ \end{array} \right. $$ Henceforth we set $$ u = (u_0, ..., u_N), \quad u_j = u\vert_{e_j}, \quad\hbox{ and } \quad p = (p_0, ..., p_N), \quad p_j = p\vert_{e_j} \quad \mbox{for $j \in \{0, 1, ..., N\}$}. $$ Let us also mention that at a node point, at least three segments $e_j$ meet. If only two segments, say $e_1, e_2$, meet at a node point, then by \eqref{C} and \eqref{K}, setting $u = u_1$ and $p=p_1$ in $e_1$ and $u = u_2$, $p=p_2$ in $e_2$, we have $\partial_t^2u - \partial_x^2 u + p u$ in $e_1 \cup e_2$. Therefore we can regard $e_1 \cup e_2$ as one open segment.\\ Since one can prove the unique existence of solution to (1) - (3) in a suitable function space (e.g., Lions and Magenes \cite{LionsMagenesBook}), we denote the solution by $u[p](x,t)$, and we set $u[p] = (u[p]_0, ..., u[p]_N)$.
Moreover we consider the following heat system on the same network~$\Lambda$ \begin{equation}\label{NH} \left\{ \begin{array}{lll} \partial_t u_j - \partial_x^2 u_j + p_j(x)u_j = 0\quad &\quad\forall j\in\{0,1,..., N\}, \forall (x,t)\in e_j \times (0,T),\\ \partial_xu(Q,t)=0,&\quad \forall Q\in \Pi_2, \forall t\in(0,T),\\ u(x,0) = u^0(x), &\quad \forall x\in\Lambda, \end{array} \right. \end{equation} and the Schr\"odinger system on the network $\Lambda$ \begin{equation}\label{NS} \left\{ \begin{array}{lll} i\partial_t u_j - \partial_x^2 u_j + p_j(x)u_j = 0\quad &\quad\forall j\in\{0,1,..., N\}, \forall (x,t)\in e_j \times (0,T),\\ u(Q,t)=h(t),&\quad \forall Q\in \Pi_2, \forall t\in(0,T),\\ u(x,0) = u^0(x), &\quad \forall x\in\Lambda, \end{array} \right. \end{equation} both under the same node conditions \eqref{C} and \eqref{K}. Here and henceforth we set $i = \sqrt{-1}$. If there is no possible confusion, by the same notation $u[p]$ we denote the solution to \eqref{NH} or \eqref{NS}, under \eqref{C} and \eqref{K}.\\
\noindent\textbf{Inverse Problem}: Is it possible to retrieve the potential $p$ everywhere in the whole network~$\Lambda$ from measurements at all external nodes except one?\\
In our article, we address the following two fundamental theoretical questions concerning coefficient inverse problems: \\
\noindent\textbf{Uniqueness}: Do the equalities of the measurements $\partial_x u[p] (Q,t) = \partial_x u[q] (Q,t)$ for all $t\in(0,T)$ and $Q\in \Pi_2\setminus {\{Q_{N_2}\}}$ imply $p = q$ on $\Lambda$?\\
\noindent\textbf{Stability}: Can we estimate, in appropriate norms, the difference of two potentials $p - q$ on $\Lambda$ by the difference of the corresponding measurements $\partial_x u[p] (Q,t) - \partial_x u[q] (Q,t)$ for all $t\in(0,T)$ and $Q\in \Pi_2\setminus {\{Q_{N_2}\}}$ ?\\
\begin{figure}
\caption{A star-shaped network with $10$ edges ($N=9$, $N_1 = 4$, $N_2= 7$).
}
\label{fig}
\end{figure}
This inverse problem is nonlinear and we will give here the proof of the uniqueness of the solution with an argument which do not use a global Carleman estimate. Very recent papers on coefficient inverse problems on networks, as Baudouin, Cr\'epeau and Valein \cite{BaudouinCrepeauValein11} for the wave equation, and Ignat, Pazoto and Rosier \cite{IgnatPazRosier12} for the heat and the Schr\"odinger equations, give indeed \textit{stability} and therefore \textit{uniqueness} from appropriate global Carleman estimates. Our first goal is to prove the uniqueness of the potential on the tree-shaped network from measurements only at all the exterior end-points of the network, except one. The argument for the uniqueness will work for either the wave or the heat or the Schr\"odinger equations on the network. The question of the proof of the Lipschitz stability in the case of the wave equation will be addressed afterwards, using a compacteness-uniqueness argument, and relies on the observability estimate on the whole network which was already proved in the literature in several situations.
Concerning the precise topic which we are considering, the bibliography lies in two different domains, namely coefficient inverse problems for partial differential equation on the one hand and control and stabilization in networks on the other hand.
Therefore one can begin by mentioning the book of Isakov \cite{Isakov} which adresses some techniques linked to the study of inverse problem for several partial differential equations. Actually, as the first answer to the uniqueness for a coefficient inverse problem with a single measurement, we refer to Bukhgeim and Klibanov \cite{BuKli81}, and see also Klibanov \cite{Klibanov92} and Yamamoto \cite{Yam99} for example. Here we do not intend to give an exhaustive list of references. After the proof of uniqueness using the basic $1$-d result on the basis of local Carleman estimates, the idea beneath this article is to take advantage of an observability estimate to obtain the Lipschitz stability of the inverse problem with a compactness-uniqueness argument. Nowadays, many results on the stability of inverse problems are derived directly from global Carleman estimates, and see e.g., \cite{BaudouinCrepeauValein11} and \cite{IgnatPazRosier12}. One should also know that studies on inverse problems and controllability of partial differential equations share some technical materials such as Carleman estimates and observability inequalities. In the particular network setting, we would like to make use of classical results such as well-known $1$-d local Carleman estimates, observability estimates on the network borrowed from control studies, in order to obtain uniqueness and stability results. We can also give some more references on inverse problems for hyperbolic equations such as Baudouin, Mercado and Osses \cite{BMO07}, Imanuvilov and Yamamoto \cite{ImYamIP01}, \cite{ImYamCom01}, Puel and Yamamoto \cite{PuelYam97}, Yamamoto and Zhang \cite{YamZhang03}, which are all based upon local or global Carleman estimates.
Besides, the control, observation and stabilization problems of networks have been the object of recent and intensive researches such as e.g., D\'ager and Zuazua \cite{DagZua06}, Lagnese, Leugering and Schmidt \cite{LaLeSch94}, Zuazua \cite{ZuaSurveyNetworks}. More specifically, the control being only applied at one single end of the network, the articles D\'ager \cite{Dag04}, D\'ager and Zuazua \cite{DagZua00,DagZua06} prove controllability results for the wave equation on networks, using observability inequalities under assumptions about the irrationality properties
of the ratios of the lengths of the strings. We can also underline that many results of controllability on networks concern only the wave equation without lower order terms (see \cite{LaLeSch94}, Schmidt \cite{Schmidt92} for instance). However it is difficult to consider such measurements at more limited nodes for the inverse problem and we do not consider the measurements at less external nodes. \\
In the sequel, we shall use the following notations: \begin{eqnarray*} L^{\gamma}(\Lambda)&=&\left\{f; \thinspace f_j\in L^{\gamma}(e_j), \,\forall j\in\{0,1,..., N\}\right\}, \quad \gamma \ge 1,\\ H^1_0(\Lambda)&=& \Big\{f; \thinspace f_j\in H^1(e_j),\,\forall j\in\{0,1,..., N\}, \, f_j(P)=f_k(P) \thinspace \mbox{if $e_j$ and $e_k$ meet at $P$},\\
&&\, \forall P \in\Pi_1,\, \textnormal{and } f(Q)=0, \, \forall Q\in \Pi_2 \Big\}. \end{eqnarray*} For shortness, for $f\in L^1(\Lambda)$, we often write, \begin{equation*} \int_{\Lambda}f dx=\sum_{j=0}^N \int_{e_j} f_j(x)dx, \end{equation*} where the integral on $e_j$ is oriented from $I(e_j)$ to $T(e_j)$. Then the norms of the Hilbert spaces $L^2(\Lambda)$ and $H_0^1(\Lambda)$ are defined by $$
\left\|f\right\|_{L^2(\Lambda)}^2=\int_{\Lambda}\left|f\right|^2dx
\hbox{ and }\left\|f\right\|_{H_0^1(\Lambda)}^2=\int_{\Lambda}\left|\partial_x f\right|^2dx. $$ For $M\ge 0$, we introduce the set $$ L^\infty_M(\Lambda) = \left\{q=(q_0,..., q_N); \thinspace q_j\in L^\infty(e_j),\, \forall j\in\{0,1,..., N\}\thinspace
\mbox{such that $\|q\|_{L^\infty(\Lambda)} \leq M$} \right\}. $$
We are ready to state our first main result:
\begin{theorem}[\bf Uniqueness]\label{Thm1}
Let $r>0$ be an arbitrary constant. Assume that $p, q \in L^{\infty}(\Lambda)$ and the initial value $u^0$ satisfies $$
|u^0(x)|\geq r> 0, \quad \mbox{a.e. in $\Lambda$}. $$ Assume further that the solutions $u[p], u[q]$ of \eqref{NW}-\eqref{C}-\eqref{K} belong to $$H^3(0,T; L^{\infty}(\Lambda)) \cap H^1(0,T; H^2(\Lambda)).$$ Then there exists $T_0>0$ such that for all $T\geq T_0$, if $$ \partial_x u[p] (Q,t) = \partial_x u[q] (Q,t) \quad \mbox{for each $t\in(0,T)$ and $Q\in\Pi_2\setminus\{Q_{N_2}\}$}, $$ then we have $p=q$ in $\Lambda$. \end{theorem}
The proof of this result in Section 2 relies on a 1-d result of uniqueness for the determination of potential in the wave equation and an ``undressing'' argument.\\
It is worth mentioning that our argument gives the uniqueness for the inverse problems of determination of potentials on tree-shaped networks also for the heat and the Schr\"odinger equations using only measurements at $N_2 - 1$ exterior end-points. In fact, our arguments in proving the uniqueness for the wave and the Schr\"odinger equations are essentially the same and are based on local Carleman estimates, while the uniqueness for the inverse heat problem is reduced to the uniqueness for the corresponding inverse wave problem (in a sense to be detailed later).
\begin{theorem}[\bf Uniqueness for the heat inverse problem]\label{Thm1-1}
Assume that $p, q \in L^{\infty}(\Lambda)$, the initial value $u^0$ satisfies $$
|u^0(x)|\geq r> 0, \quad \mbox{a.e. in $\Lambda$} $$ for some constant $r$, and the solutions $u[p]$ and $u[q]$ to \eqref{NH}-\eqref{C}-\eqref{K}, belong to $$ H^2(0,T; L^{\infty}(\Lambda)) \cap H^1(0,T;H^2(\Lambda)).$$ Then there exists $T>0$ such that if $$ u[p] (Q,t) = u[q] (Q,t) \quad \mbox{for each $t\in(0,T)$ and $Q\in\Pi_2\setminus\{Q_{N_2}\}$}, $$ then we have $p=q$ in $\Lambda$. \end{theorem}
\begin{theorem}[\bf Uniqueness for the Schr\"odinger inverse problem]\label{Thm1-2}
Assume that $p, q \in L^{\infty}(\Lambda)$, the initial value $u^0$ satisfies $$
|u^0(x)|\geq r> 0, \quad \mbox{a.e. in $\Lambda$} $$ for some constant $r$, and the solutions $u[p]$ and $u[q]$ to \eqref{NS}-\eqref{C}-\eqref{K}, belong to $$H^2(0,T; L^{\infty}(\Lambda)) \cap H^1(0,T;H^2(\Lambda)).$$ Then there exists $T>0$ such that $$ \partial_xu[p] (Q,t) = \partial_xu[q] (Q,t) \quad \mbox{for each $t\in(0,T)$ and $Q\in\Pi_2\setminus\{Q_{N_2}\}$}, $$ then we have $p=q$ in $\Lambda$. \end{theorem}
One can refer to \cite{BaudouinCrepeauValein11} for the same inverse problem in the wave equation on a network where the proof is detailed in a star-shaped network but is actually generalizable to tree-shaped networks. Reference \cite{IgnatPazRosier12} discusses the inverse heat problem on tree-shaped network. Moreover the paper \cite{IgnatPazRosier12} treats the Schr\"odinger case in a star-shaped network and needs measurements at all external nodes. We do not know any uniqueness result for non-tree graphs, which are graphs containing a closed cycle. For observability inequality on general graph, see e.g., \cite{DagZua06}.\\
For the inverse problem in the wave equation case, we state
\begin{theorem}[\bf Stability]\label{Thm2} Let $M>0$ and $r>0$. Assume that $p\in L^\infty_M(\Lambda)$ and the solutions $u[p]$ and $u[q]$ to \eqref{NW}-\eqref{C}-\eqref{K} satisfy $$ u[p], u[q] \in H^3(0,T;L^{\infty}(\Lambda)) \cap H^1(0,T;H^2(\Lambda)). $$ Assume also that the initial value $u^0$ satisfies $$
|u^0(x)|\geq r> 0, \quad \mbox{a.e. in $\Lambda$}. $$ Then there exists $T_0>0$ such that for all $T\geq T_0$, there exists $C=C(T,r,M, \ell_0,..., \ell_N)>0$ such that \begin{equation}\label{stabpi}
|| q-p||_{L^2(\Lambda)}\leq C \sum_{j=1}^{N_2-1}\left\| \partial_xu_{j}[p](Q_j)
- \partial_xu_{j}[q](Q_j)\right\|_{H^1(0,T)}. \end{equation} \end{theorem}
This paper is composed of five sections. The proof of uniqueness in the inverse problem in the wave equation case (Theorem~\ref{Thm1}) is presented in Section 2. The cases of Schr\"odinger and heat equations are studied in Section 3, devoted to the proofs of Theorems~\ref{Thm1-1} and \ref{Thm1-2}. Theorem~\ref{Thm2} is finally proven in Section 5 by a compactness-uniqueness argument and an observability estimate on the whole network.\\
We conclude this section with a classical result on the existence and regularity of solutions of the wave system and provide the corresponding energy estimates for the solution which we will need later.
\begin{lemma}\label{Energy} Let $\Lambda$ be a tree-shaped network and assume that $p\in L^{\infty}_M(\Lambda)$, $g \in L^1(0,T;L^2(\Lambda))$, $u^0\in H_0^1(\Lambda)$ and $u^1\in L^2(\Lambda)$. We consider the 1-d wave equation on the network with the conditions \eqref{C} and \eqref{K}: \begin{equation}\label{e} \left\{\begin{array}{lll} \partial_t^2u-\partial_x^2 u+p(x)u=g(x,t),&\quad\mbox{in $\Lambda\times(0,T)$},\\
u(Q,t)=0,&\quad \mbox{in $(0,T),\, Q\in\Pi_2$},\\ u_j(P,t)=u_k(P,t),&\quad \mbox{in $(0,T),\, P\in\Pi_1,\, j,k\in \{ 1, ..., n_P \}$}, \\ \left[ \partial_x u (t)\right]_P=0,&\quad \mbox{in $(0,T)$, $P\in\Pi_1$},\\ u(0) = u^0, \quad \partial_t u(0)=u^1,&\quad \mbox{in $\Lambda$}. \end{array}\right. \end{equation} The Cauchy problem is well-posed and equation \eqref{e} admits a unique weak solution $$ u\in C([0,T],H_0^1(\Lambda)) \cap C^1([0,T],L^2(\Lambda)). $$ Moreover there exists a constant $C=C(\Lambda,T,M)>0$ such that for all $t\in(0,T)$, the energy $$
E(t) = ||\partial_t u(t)||^2_{L^2(\Lambda)}+||\partial_x u(t)||^2_{L^2(\Lambda)} $$ of the system \eqref{e} satisfies \begin{equation}\label{estimeenergy}
E(t) \leq C\left(||u^0||^2_{H_0^1(\Lambda)} + ||u^1||^2_{L^2(\Lambda)} +
||g||^2_{L^1(0,T,L^2(\Lambda))} \right) \end{equation} and we also have the following trace estimate \begin{equation}\label{hiddenregularity}
\sum_{j=1}^{N_2}\left\|\partial_x u_j(Q_j)\right\|_{L^2(0,T)}^2
\leq C \left(||u^0||^2_{H_0^1(\Lambda)}
+ ||u^1||^2_{L^2(\Lambda)} + ||g||^2_{L^1(0,T,L^2(\Lambda))} \right). \end{equation} \end{lemma}
The proof of the unique existence of solution to equation~\eqref{e} can be read in \cite[Chap. 3]{LionsMagenesBook}. Estimate \eqref{estimeenergy} is a classical result which can be formally obtained by multiplying the main equation in \eqref{e} by $\partial_t u_{j}$, summing up for $j\in\left\{0,...,N\right\}$ the integral of this equality on $(0,T)\times e_j$ and using some integrations by parts. Estimate~\eqref{hiddenregularity} is a hidden regularity result which can be obtained by multipliers technique (we refer to \cite[Chapter 1]{Lions}). Formally, for the particular case of a star-shaped network of vertex $P=0$ for example, it comes from the multiplication of \eqref{e} by $m(x)\partial_x u_{j}$, where $m \in C^1(\bar{\Lambda})$ with $m(0)=0$ and $m_j(l_j) = 1$, summing up the integrals of this equality on $(0,T)\times (0,l_j)$ over $j\in \{0,...,N\}$ and using integrations by parts.
\section{Uniqueness of the inverse problem - wave network case}
As already evoked in the introduction, the proof of Theorem~\ref{Thm1} will use a well-known 1-d result of uniqueness for the inverse problem. We recall it in the following lemma.
\begin{lemma}\label{1d} Let $r>0$, $p\in L^\infty(0,\ell)$ and $T > 2\ell$. Consider the 1-d wave equation in $[0,\ell]$ with homogeneous Dirichlet boundary data as follows: \begin{equation}\label{1D} \left\{ \begin{array}{lll} \partial_t^2 y - \partial_x^2 y + p(x)y = f(x)R(x,t), \qquad& (x,t) \in (0,\ell) \times (0,T),\\ y(\ell,t)=0,& t\in(0,T),\\ y(x,0) = 0, \partial_{t}y(x,0)=0 ,& x\in(0,\ell), \end{array} \right. \end{equation} where $f\in L^2(0,\ell)$ and $R\in H^1(0,T;L^\infty(0,\ell))$ satisfies
$ |R(x,0)|\geq r> 0$ a.e. in $(0,\ell)$.\\ If $\, \partial_x y (\ell,t) = 0 \,$ for all $t\in(0,T)$, then we have $f\equiv 0$ in $(0,\ell)$ and $y\equiv 0$ in $(0,\ell) \times (0,T)$. \end{lemma}
This lemma is a classical uniqueness result for the inverse source problem in a wave equation and the proof can be done by the method in \cite{BuKli81} on the basis of a 1-d Carleman estimate and the even extension of $y$ to negative times $t$. We further refer to Imanuvilov and Yamamoto \cite{ImYamIP01}, \cite{ImYamCom01}, Klibanov \cite{Klibanov92}, Klibanov and Timonov \cite{KliTiBook} for example, and we omit details of the proof. \\
\noindent {\bf Proof of Theorem~\ref{Thm1}.} We define the following operation of ``removing'' segments from the tree-shaped network $\Lambda$, starting from all the external nodes where we make measurements, except one. We divide the proof into several steps. \\
{\bf Step 1.} From Lemma~\ref{1d}, we can easily prove that if $e_j$ is a segment of $\Lambda$ which ends at an external node $Q_j\in \Pi_2$, and if the solutions $u[p]$ and $u[q]$ to \eqref{NW} satisfy $\partial_x u[p](Q_j,t) = \partial_x u[q](Q_j,t)$ for all $t\in(0,T)$, then $p= q$ on the segment~$e_j$ and $u[p](x,t) = u[q](x,t)$ for all $x\in e_j$ and for all $ t\in(0,T)$. Indeed, if we set $y = u_j[p_j] - u_j[q_j]$, then \begin{equation}\label{1Dj} \left\{ \begin{array}{lll} \partial_t^2 y - \partial_x^2 y + p_j(x)y = (q_j-p_j)(x)u_j[q_j](x,t)\qquad &\quad (x,t)\in (0,\ell) \times (0,T),\\ y(Q_j,t)=0,&\quad t\in(0,T),\\ y(x,0) = 0, \partial_{t}y(x,0)=0 ,&\quad x\in(0,\ell), \end{array} \right. \end{equation} and noting that $T>0$ is sufficiently large, we can apply Lemma~\ref{1d} since $\partial_x y (Q_j,t) = 0$ for
all $t\in(0,T)$, $u_j[q_j] \in H^1(0,T;L^{\infty}(\Lambda))$ and $|u_j^0(x)| \geq r> 0$ on $e_j$. We obtain that $p_j \equiv q_j$ on $e_j$ and consequently $u_j[p_j](x,t) = u_j[p_j](x,t)$ in $e_j\times(0,T_1)$, where $T_1 \in (0,T)$ is some constant. \\ Therefore, for any segment $e$ with the end-points $P$ and $Q$ such that $Q \in \Pi_2 \setminus \{ Q_{N_2}\}$, we see that $p=q$ on $e$ and $(u[p]\vert_e)(P,t) = (u[q]\vert_e)(P,t)$, $(\partial_xu[p]\vert_e)(P,t) = (\partial_xu[q]\vert_e)(P,t)$ for $0 < t < T_1$. Let $\Pi^2_1$ be all the interior node points $P$ of segments of $\Lambda$ having their other end-point in $\Pi_2 \setminus \{Q_{N_2}\}$. We note that $\Pi_1^2 \subset \Pi_1$. Applying the above argument to all the exterior end-points except for $Q_{N_2}$, we have $$ u[p]_j(P,t) = u[q]_j(P,t), \quad \partial_xu[p]_j(P,t) = \partial_xu[q]_j(P,t) $$ for each $P \in \Pi_1^2$, $0 < t < T_1$ and $j \in \{1, ..., N_3\}$. Here by $e_1, ..., e_{N_3}$, we enumerate the segments connecting a point in $\Pi_1^2$ and a point in $\Pi_2 \setminus \{Q_{N_2}\}$. \\
{\bf Step 2.} Let $P \in \Pi_1$ be a given node such that $n_P$ segments, say, $e_1, ..., e_{n_P}$ meet at $P$ and $e_1, ..., e_{n_P-1}$ connect $P$ with exterior end-points, say, $Q_1, ..., Q_{n_P-1} \in \Pi_2$ and \begin{equation}\label{star} \begin{array}{c} u[p]_j(P,t) = u[q]_j(P,t), \\ \partial_x u[p]_j(P,t) = \partial_x u[q]_j(P,t), \quad j \in \{1, ..., n_P-1\}, \thinspace 0<t<T. \end{array} \end{equation} Using the continuity \eqref{C} and the Kirchhoff law \eqref{K} at node $P$, we can deduce that $$ \begin{array}{c} u[p]_{n_P}(P,t) = u[q]_{n_P}(P,t), \\ \partial_x u[p]_{n_P}(P,t) = \partial_x u[q]_{n_P}(P,t), \quad 0 < t < T. \end{array} $$
{\bf Step 3.} Let $\Lambda^2$ be the graph generated from $\Lambda$ by removing $e_1, ..., e_{N_3}$. Therefore, since $T_1>0$ is still sufficiently large, we can apply the same argument as in Step 1 to the graph $\Lambda^2$.
We repeat this operation to obtain the sets $\Lambda^3$, then $\Lambda^4$,..., $\Lambda^n$. Hence, let $L^k$ be the set of all the open segments of $\Lambda_k$, $\Pi^k_1$ the set of the interior node points of $\Lambda_k$, $\Pi^k_2$ the set of external endpoints of $\Lambda_k$. Setting $\Lambda^1 = \Lambda$, we note that $L^1 = \{ e_0, ..., e_{N} \}$, $\Pi^1_1 = \{ P_1, ..., P_{N_1}\}$, $\Pi^1_2 = \{ Q_1, ..., Q_{N_2}\}$.
By \eqref{C} and \eqref{K}, we see that $$ \Pi^{k-1}_1 \supset \Pi^k_1, \qquad \forall k \in \mathbb N $$ and $$ \Lambda_k = L^k \cup \Pi^k_1 \cup \Pi^k_2, \quad L^k \cap \Pi^k_1 = L^k \cap \Pi^k_2 = \Pi^k_1 \cap \Pi^k_2 = \emptyset, \quad \forall k\in \mathbb N. $$ In order ro complete the proof, it is sufficient to prove there exists $n \in \mathbb N$ such that \begin{equation}\label{emptylambda} \Lambda_n = \emptyset. \end{equation} Assume contrarily that $\Lambda_n \ne \emptyset$ for all $n \in \mathbb N$. Since every segment with exterior end-point in $\Pi_2 \setminus \{ Q_{N_2}\}$, can be removed (meaning that $u[p] = u[q]$ on the segment) by the above operation, we obtain that there exists $n_0 \in N$ such that $\Lambda_{n_0} = L^{n_0} \cup \Pi_1^{n_0}$, \textit{i.e.,} $\Pi^{n_0}_2 = \emptyset$. Then $\Lambda_{n_0}$ must be a closed cycle since it possesses no external endpoint. By assumption, there exist no closed cycles in a tree-shape network. This is a contradiction and thus the proof of \eqref{emptylambda}, and therefore, the one of Theorem~\ref{Thm1} is completed.
{}$\square$
\section{Uniqueness for the inverse problem - Schr\"odinger and heat network cases}
\subsection{Proof of Theorem \ref{Thm1-1} - Heat case.}
We apply an argument similar to the proof of Theorem 4.7 in \cite{Klibanov92} which is based on the reduction of the inverse heat problem to an inverse wave problem by a kind of Laplace transform called the Reznitzkaya transform (e.g., \cite{Isakov}, \cite{LRS}, \cite{RomBook}).
First we define an operator $\Delta_{\Lambda}$ in $L^2(\Lambda)$ by $\Delta_{\Lambda}u = \partial_x^2 u_j$ in $e_j$, for all $j\in \{0, 1, ...., N\}$ with $$ \mathcal{D}(\Delta_{\Lambda}) = \big\{ u = (u_0, ..., u_N); \, \mbox{$u_j \in H^2(e_j)$, $\partial_xu(Q) = 0$ for $Q \in \Pi_2$, $u_j$ satisfying \eqref{C} and \eqref{K}}\big\}. $$ Here, $e_j$ is oriented from $I(e_j)$ to $T(e_j)$ when defining $\partial_x^2$. Then, similarly to \cite{IgnatPazRosier12}, we can prove that $\Delta_{\Lambda}$ is self-adjoint and $(\Delta_{\Lambda}u, u)_{L^2(\Lambda)} := \sum_{j=0}^N (\partial_x^2u_j,u_j)_{L^2(e_j)} \ge 0$. Therefore $\Delta_{\Lambda}$ generates an analytic semigroup $e^{t\Delta_{\Lambda}}$, $t>0$ (e.g., Pazy \cite{Pazy}, Tanabe \cite{TanabeBook}). Since $p \in L^{\infty}(\Lambda)$, the perturbed operator $\Delta_{\Lambda} + p$ generates an analytic semigroup (e.g., Theorem 2.1 in \cite{Pazy}, p.80). Therefore by the semigroup theory (e.g. \cite{Pazy}, \cite{TanabeBook}), we know that the solutions $u[p](x,t)$ and $u[q](x,t)$ of equation (4) are analytic in $t$ for any fixed $x\in \Lambda$. More precisely, $u[p], u[q]: (0,\infty) \longrightarrow H^2(\Lambda)$ are analytic in $t>0$.
By $u^H[p]$ we denote the solution of the heat system \eqref{NH} and by $u^H[q]$ the corresponding solution when the potential is $q$. By the analyticity in $t$ and the assumption in the theorem, we have \begin{equation}\label{assumHeat} u^H[p](Q,t) = u^H[q](Q,t), \quad \forall Q \in\Pi_2\setminus \{Q_{N_2}\}, \thinspace \forall t > 0. \end{equation} On the other hand, denote by $\widetilde u[p]$ the solution of the wave system \begin{equation}\label{NWH} \left\{ \begin{array}{lll} \partial_t^2 u_j - \partial_x^2 u_j + p_j(x)u_j = 0,\quad &\quad\forall j\in\{0,1,..., N\}, \forall (x,t)\in e_j \times (0,\infty),\\ \partial_x u[p](Q,t)=0,&\quad \forall Q\in \Pi_2, \forall t\in(0,\infty),\\ u[p](x,0) = 0, \thinspace \partial_t u(x,0)=u^0(x),&\quad \forall x\in\Lambda \end{array} \right. \end{equation} and by $\widetilde u[q]$ the corresponding solution when the potential is $q$. Then we obtain (e.g., \cite[pp.251-252]{LRS}) that $$ \frac{1}{2\sqrt{\pi t^3}}\int^{\infty}_0 \tau e^{-\frac{\tau^2}{4t}} \widetilde u[p](x,\tau) d\tau $$ satisfies (4). The uniqueness of solution to equation (4) implies $$ u^H[p](x,t) = \frac{1}{2\sqrt{\pi t^3}}\int^{\infty}_0 \tau e^{-\frac{\tau^2}{4t}}\widetilde u[p](x,\tau) d\tau, \quad \forall x \in \Lambda, \forall t > 0 $$ and the same equality with $q$. By assumption \eqref{assumHeat}, we obtain $$ \frac{1}{2\sqrt{\pi t^3}}\int^{\infty}_0 \tau e^{-\frac{\tau^2}{4t}}(\widetilde u[p] - \widetilde u[q])(Q,\tau) d\tau = 0, \quad \forall Q\in \Pi_2 \setminus \{Q_{N_2}\}, \forall t>0. $$ By the change of variables $s = \frac{1}{4t}$ and $\tau^2 = \eta$, we obtain $$ \int^{\infty}_0 e^{-s\eta}(\widetilde{u}[p] - \widetilde{u}[q])(Q,\sqrt{\eta}) d\eta = 0, \quad \forall Q\in \Pi_2 \setminus \{Q_{N_2}\}, \forall s>0 $$ and the injectivity of the Laplace transform yields \begin{equation} (\widetilde{u}[p] - \widetilde{u}[q])(Q,\sqrt\eta) = 0, \quad \forall Q\in \Pi_2 \setminus \{Q_{N_2}\}, \forall \eta>0. \end{equation} Applying the same argument as in Section 2 for the wave system, we prove $p=q$ in $\Lambda$. Thus the proof of Theorem~\ref{Thm1-1} is completed.
\subsection{Proof of Theorem~\ref{Thm1-2} - Schr\"odinger case.}
It is sufficient to prove the following lemma. \\ \begin{lemma}\label{schrodi} Let $r>0$ and $p \in L^{\infty}(0,\ell)$, $f\in L^2(0,\ell)$ be real-valued, and $T>0$ be arbitrarily fixed. We consider a 1-d Schr\"odinger equation: $$ \left\{ \begin{array}{lll} i\partial_t y - \partial_x^2 y + p(x)y = f(x)R(x,t), \qquad &\forall (x,t) \in (0,\ell) \times (0,T), \\ y(\ell,t) = 0, &\forall t \in (0, T),\\ y(x,0) = 0, &\forall x \in (0, \ell), \end{array} \right. $$ where $R \in H^1(0,T; L^{\infty}(0,\ell))$ satisfies $\vert R(x,0)\vert \ge r > 0$ a.e. in $(0,\ell)$. \\
If $\, \partial_xy(\ell,t) = 0\,$ for all $ t \in (0, T)$, then we have $f=0$ in $(0,\ell)$ and $y=0$ in $(0,\ell) \times (0,T)$. \end{lemma}
Using the same method as the one for the proof of Lemma~\ref{1d}, this lemma is proved by means of the following Carleman estimate: \\ \begin{lemma} For $x_0 \not\in [0,\ell]$ and $\beta > 0$ arbitrarily fixed, we set $$ Sv = i\partial_tv - \partial_x^2v, \quad \varphi(x,t) = e^{\gamma(\vert x-x_0\vert^2 - \beta t^2)}, \quad (x,t) \in (0,\ell) \times (0,T). $$ Then there exists a constant $\gamma_0 > 0$ such that for arbitrary $\gamma\ge \gamma_0$ we can choose $s_0 > 0$ satisfying, for a constant $C>0$, $$ \int^T_0\int^{\ell}_0 (s\vert \partial_xv\vert^2 + s^3\vert v\vert^2) e^{2s\varphi} dxdt \le C\int^T_0\int^{\ell}_0 \vert Sv\vert^2 e^{2s\varphi} dxdt $$ for all $s > s_0$ and all $v \in L^2(0,T;H^2_0(0,\ell)) \cap H^1_0(0,T;L^2(0,\ell))$. \end{lemma}
This is a Carleman estimate with regular weight function $\gamma(\vert x-x_0\vert^2 - \beta t^2)$ and for the proof, we refer to e.g. \cite[Lemma 2.1]{YuYam-AA07} (see also \cite{YuYam-CAM10}). Concerning a Carleman estimate for Schr\"odinger equation in a bounded domain $\Omega \subset \mathbb R^n$ with singular weight function $\varphi$, we can refer for example to \cite{BaudouinPuel02,MOR08}.
On the basis of this lemma, the proof of Lemma~\ref{schrodi} is done by a usual method by Bukhgeim and Klibanov \cite{BuKli81} by using the extension of $y$ to $-T<t<0$ by $y(\cdot,t) = \overline{y(\cdot,-t)}$ and a cut-off argument. We omit the details of the proof.
\section{Observability in the wave network}
The proof of the stability result will rely strongly on the classical result of observability that we are now presenting and proving. One should specifically mention the survey \cite{ZuaSurveyNetworks} and the books \cite{DagZua06}, \cite{LaLeSch94}, where the question of observability in networks of strings (or wave equations) is widely explored in different cases.
We concentrate here on the case where the observation available comes from all but one external nodes, in a setting with a system of wave equations with potential. Since most of the literature on string networks focus only on the wave equation without lower order terms (see \cite{LaLeSch94} or \cite{DagZua06} for instance), we detail here how to obtain the observability result for the wave equation with potential. In some other cases, we can prove the observability inequality directly by a global Carleman estimate (e.g. \cite{BaudouinCrepeauValein11}).
\begin{theorem}[\bf Observability inequality]\label{Thm3} On the tree-shaped network $\Lambda$, assuming $p\in L^\infty(\Lambda)$, let us consider the system of 1-d wave equations under the continuity and Kirchhoff law's assumptions \eqref{C} and \eqref{K}: \begin{equation}\label{eqobs}
\left\{\begin{array}{lll} \partial_t^2u-\partial_x^2 u+p(x)u=0,&\quad\mbox{in $\Lambda\times(0,T)$},\\
u(Q,t)=0,&\quad \mbox{in $(0,T),\, \forall Q\in\Pi_2$},\\ u_j(P,t)=u_k(P,t),&\quad \mbox{in $(0,T),\,\forall P\in\Pi_1,\, \forall j,k\in \{1, ..., n_P\}$}, \\ \left[ \partial_x u (t)\right]_P=0,&\quad \mbox{in $(0,T), \forall P\in\Pi_1$},\\ u(x,0) = 0, \quad \partial_t u(x,0)=a(x),&\quad \mbox{in $\Lambda$}, \end{array}\right. \end{equation} Then there exists a minimal time $T_0$ such that for all $T > T_0$, the observability estimate \begin{equation}\label{obs}
\int_\Lambda | a(x)|^2 dx
\le C\sum_{j=1}^{N_2-1} \int^T_0 | \partial_xu_j(Q_j,t)|^2 dt \end{equation} holds for a solution $u$ of \eqref{eqobs}. \end{theorem}
\noindent {\bf Proof of Theorem~\ref{Thm3}.} Let $v$ be the solution of the system $$ \left\{ \begin{array}{lll} \partial_t^2 v - \partial_x^2 v= -pu \quad &\quad\forall (x,t)\in \Lambda \times (0,T),\\ v(Q,t)=0,&\quad \forall Q\in \Pi_2, t\in(0,T),\\ v_j(x,0) = 0, \partial_t v_j(x,0)=0,&\quad\forall j\in\{0,1,..., N\}, \quad x\in e_j, \end{array} \right. $$ under conditions \eqref{C} and \eqref{K}. Then \eqref{hiddenregularity} in Lemma~\ref{Energy} and $p\in L^\infty(\Lambda)$ yields \begin{equation}\label{eqeq}
\sum_{j=1}^{N_2} \int^T_0 | \partial_xv_j(Q_j,t)|^2dt
\le C \int^T_0\int_\Lambda | pu|^2 dxdt
\le C \int^T_0\int_\Lambda | u|^2 dxdt. \end{equation} Setting $w = u - v$, we still have \eqref{C} and \eqref{K} satisfied by $w$, along with the following equation $$ \left\{ \begin{array}{lll} \partial_t^2 w - \partial_x^2 w= 0 \quad &\quad\forall (x,t)\in \Lambda \times (0,T),\\ w(Q,t)=0,&\quad \forall Q\in \Pi_2, t\in(0,T),\\ w_j(x,0) = 0, \partial_t w_j(x,0)=a(x),&\quad\forall x\in \Lambda. \end{array} \right. $$ Therefore, using a classical observability inequality in the case where $p=0$ (e.g., \cite{DagZua06,LaLeSch94}), we have $$
\int_\Lambda | a(x)|^2 dx
\le C \sum_{j=1}^{N_2-1} \int^T_0 | \partial_xw_j(Q_j,t)|^2 dt. $$ Hence, by \eqref{eqeq}, we have \begin{eqnarray}
&& \int_\Lambda | a(x)|^2 dx
\le C\sum_{j=1}^{N_2-1} \int^T_0 | \partial_xu_j(Q_j,t)|^2 dt
+ C\sum_{j=1}^{N_2-1} \int^T_0 | \partial_xv_j(Q_j,t)|^2 dt\nonumber\\
&&\le C\sum_{j=1}^{N_2-1} \int^T_0 | \partial_xu_j(Q_j,t)|^2 dt
+ C \int^T_0\int_\Lambda | u|^2dxdt.\label{unicompact} \end{eqnarray} Therefore a usual compactness-uniqueness argument yields the observability inequality \eqref{obs}. Indeed, if \eqref{obs} is not satisfied, then we can assume that there exists $a^n\in L^2(\Lambda), n\in \mathbb N$ such that \begin{equation}\label{hypcontr}
\|a^n\|_{L^2(\Lambda)} = 1, \, \forall n\in\mathbb N \quad \hbox{ and } \quad
\lim_{n\to + \infty}\sum_{j=1}^{N_2-1} \int^T_0 | \partial_xu_j^n(Q_j,t)|^2 dt = 0. \end{equation} Using the energy estimate \eqref{estimeenergy} of Lemma~\ref{Energy} on the solution $u^n$ of system \eqref{eqobs} with initial data $a^n$, we obtain $$
|| u^n(t)||^2_{H^1_0(\Lambda)} = ||\partial_x u^n(t)||^2_{L^2(\Lambda)}
\leq C||a^n||^2_{L^2(\Lambda)} \leq C. $$ Since the embedding $H^1_0(\Lambda) \subset L^2(\Lambda)$ is compact, we can extract a subsequence, denoted again by the same notation and we have $(u^n)_{n\in\mathbb N^*}$ convergent in $L^2(\Lambda)$. Therefore, using \eqref{unicompact}, we obtain \begin{align*}
\int_\Lambda | a^n-a^m|^2 dx
\le & ~C\sum_{j=1}^{N_2-1} \int^T_0 | \partial_xu_j^n(Q_j,t) |^2 dt
+ C\sum_{j=1}^{N_2-1} \int^T_0 |\partial_xu_{j}^m(Q_j,t)|^2 dt \\ &
+~ C \int^T_0\int_\Lambda | u^n - u^m|^2dxdt \end{align*} so that \eqref{hypcontr} and $
\displaystyle\lim_{n,m\to \infty} \|u^n - u^m\|_{L^2(\Lambda)} = 0 $ imply $
\displaystyle\lim_{n,m\to \infty} ||a^n-a^m||^2_{L^2(\Lambda)} = 0. $ Consequently, there exists a limit $a_0$ such that $\displaystyle\lim_{n\to +\infty}a^n = a_0$ in $L^2(\Lambda)$ and from
\eqref{hypcontr} , we have $\|a_0\|_{L^2(\Lambda)} = 1$. Moreover, the solution $u[a_0]$ of system \eqref{eqobs} with initial data $a_0$ is such that $$ \partial_xu_{j}^m[a_0](Q,t) = 0 , \quad \forall t\in(0,T), \forall Q\in\Pi_2. $$ Hence we apply a classical unique continuation result for a wave equation to obtain that $u[a_0]$ vanishes everywhere so that $a_0 = 0$, which contradicts
$\|a_0\|_{L^2(\Lambda)} = 1$. Here, the unique continuation can be proved for instance by a Carleman estimate (e.g. \cite{Isakov}, \cite{KliTiBook}). This ends the proof of Theorem~\ref{Thm3}.
\section{Proof of the stability for the wave network inverse problem}
This section is devoted to the proof of Theorem~\ref{Thm2}. The proof relies on a compactness-uniqueness argument and the observability estimate (Theorem~\ref{Thm3}) on the whole network. \\
Let us denote by $u[p]$ the solution of \eqref{NW} under the assumptions \eqref{C} and \eqref{K}. Henceforth we always assume the conditions \eqref{C} and \eqref{K}. We consider $y = \partial_t\left(u[p] - u[q] \right)$ that satisfy \begin{equation}\label{eqy} \left\{ \begin{array}{lll} \partial_t^2 y - \partial_x^2 y + q(x)y = (q-p) \partial_t u[p]\quad &\quad\forall (x,t)\in \Lambda \times (-T,T),\\ y(Q,t)=0,&\quad \forall Q\in \Pi_2, t\in(0,T),\\ y(x,0) = 0, \partial_t y(x,0)=(q-p)u^0(x),&\quad \forall x\in\Lambda, \end{array} \right. \end{equation} We define $\psi$ and $\phi$ as the solutions of \begin{equation}\label{eqpsi} \left\{ \begin{array}{lll} \partial_t^2 \psi - \partial_x^2 \psi + q(x)\psi = (q-p) \partial_t u[p]\quad &\quad\forall (x,t)\in \Lambda \times (-T,T),\\ \psi(Q,t)=0,&\quad \forall Q\in \Pi_2, t\in(0,T),\\ \psi(x,0) = 0, \partial_t \psi(x,0)=0,&\quad \forall x\in\Lambda, \end{array} \right. \end{equation} and \begin{equation}\label{eqphi} \left\{ \begin{array}{lll} \partial_t^2 \phi - \partial_x^2 \phi + q(x)\phi = 0\quad &\quad \forall (x,t)\in \Lambda \times (-T,T),\\ \phi(Q,t)=0,&\quad \forall Q\in \Pi_2, t\in(0,T),\\ \phi(x,0) = 0, \partial_t \phi(x,0)=(q-p)u^0(x),&\quad \forall x\in\Lambda. \end{array} \right. \end{equation} such that $y = \psi + \phi$. We can apply Theorem~\ref{Thm3} to equation \eqref{eqphi} so that \begin{equation}\label{obsphi}
\int_\Lambda | (q-p)u^0|^2 dx
\le C\sum_{j=1}^{N_2-1} \int^T_0 | \partial_x\phi_j(Q_j,t)|^2 dt. \end{equation} On the other hand, a regularity result of Lemma~\ref{Energy} applied to a time derivative of equation \eqref{eqpsi} gives \begin{eqnarray}
\sum_{j=1}^{N_2}\left\|\partial_x \psi_j(Q_j)\right\|_{H^1(0,T)}^2
&\leq& C \left( ||(q-p) \partial_{t}^2 u[p]||^2_{L^1(0,T,L^2(\Lambda))}
+ ||(q-p) u^1||^2_{L^2(\Lambda)}\right)\nonumber\\
&\leq& 2CK^2 ||q-p||^2_{L^2(\Lambda)} \label{hiddenregularitypsi} \end{eqnarray} as soon as we have $ u[p]\in H^2(0,T,L^\infty(\Lambda))$ which yields $\partial_tu[p]\in C([0,T];L^\infty(\Lambda))$ so that
$u^1\in L^\infty(\Lambda)$) with $\|u[p]\|_{H^2(0,T,L^\infty(\Lambda))} \leq K$. The compact embedding $H^1 (0,T) \subset L^2(0,T)$ allows then to write that the operator $\Psi : L^2(\Lambda) \to L^2(0,T)$ defined by $$ \Psi(p-q)(t) = \sum_{j=1}^{N_2} \partial_x \psi_j(Q_j,t) , \qquad 0<t<T $$ is compact.
Therefore, since we have $|u^0(x)|\geq r> 0$ almost everywhere in $\Lambda$, by \eqref{obsphi} and \eqref{hiddenregularitypsi}, we obtain \begin{eqnarray}
|| q-p||_{L^2(\Lambda)}
&\leq& C \int_\Lambda | (q-p)u^0|^2 dx~ \leq ~ C\sum_{j=1}^{N_2-1} \int^T_0 |
\partial_x\phi_j(Q_j,t)|^2 dt \nonumber\\
&\leq& C\sum_{j=1}^{N_2-1} \int^T_0 | \partial_x y_j(Q_j,t)|^2 dt
+ C\sum_{j=1}^{N_2} \int^T_0 | \partial_x\psi_j(Q_j,t)|^2 dt
\nonumber\\
&\leq& C\sum_{j=1}^{N_2-1} \int^T_0 | \partial_x y_j(Q_j,t)|^2 dt
+ C ||\Psi(q-p)||^2_{L^2(0,T)} \label{estimfinale}\\
&\leq& C \sum_{j=1}^{N_2-1}\left\| \partial_xu_{j}[p](Q_j)- \partial_xu_{j}[q](Q_j)
\right\|_{H^1(0,T)} + C ||\Psi(q-p)||^2_{L^2(0,T)}.
\nonumber \end{eqnarray} We aim at proving that we can get rid of the second term on the right-hand side of the last estimate in order to obtain \eqref{stabpi}. Again, a compactness-uniqueness argument will be the key and it relies here on the compactness of $\Psi$ and the uniqueness result of Theorem~\ref{Thm1}.
Indeed, we set $f= q-p$. We assume that $$
|| f||_{L^2(\Lambda)}\leq C \sum_{j=1}^{N_2-1}\left\| \partial_x y_j(Q_j)
\right\|_{L^2(0,T)}, $$ which is equivalent to \eqref{stabpi}, does not hold. Then one can assume that there exists $f^n\in L^2(\Lambda), n\in\mathbb N$ such that \begin{equation}\label{hypcontr1}
\|f^n\|_{L^2(\Lambda)} = 1, \, \forall n\in\mathbb N
\quad \hbox{ and } \quad
\lim_{n\to + \infty} \sum_{j=1}^{N_2-1}\left\| \partial_x y_j^n(Q_j)
\right\|_{L^2(0,T)} = 0. \end{equation}
First, since the sequence $(f^n)_{n\in\mathbb N}$ is bounded in $L^2(\Lambda)$, we can extract a subsequence denoted again by $(f^n)_{n\in\mathbb N}$ such that it converges towards some $f^0\in L^2(\Lambda)$ weakly in $L^2(\Lambda)$. Since $\Psi$ is a compact operator, we obtain therefore the strong convergence result \begin{equation}\label{strongcv}
\lim_{n,m\to \infty} \|\Psi(f^n) - \Psi(f^m)\|_{L^2(0,T)} = 0. \end{equation} Then, from \eqref{estimfinale} we can write $$
|| f^n - f^m||_{L^2(\Lambda)}
\leq C \sum_{j=1}^{N_2-1}\left\| \partial_x y_j^n(Q_j)\right\|_{L^2(0,T)}
+ C \sum_{j=1}^{N_2-1}\left\| \partial_x y_j^m(Q_j)\right\|_{L^2(0,T)}
+ C ||\Psi(f^n) - \Psi(f^m)||^2_{L^2(\Lambda)} $$ and deduce from \eqref{hypcontr1} and \eqref{strongcv} that $\displaystyle\lim_{n,m\to \infty}
\|f^n - f^m\|_{L^2(\Lambda)} = 0$, so that $\displaystyle\lim_{n\to \infty}
\|f^n - f^0\|_{L^2(\Lambda)} = 0$ with \begin{equation}\label{f0}
\|f^0\|_{L^2(\Lambda)} = 1. \end{equation}
Moreover, using the trace estimate \eqref{hiddenregularity} of Lemma~\ref{Energy} for the solution $y^n$ of system \eqref{eqy} with initial data $f^n u^0$ and source term $f^n\partial_tu[p]$, we obtain $$
\sum_{j=1}^{N_2-1}\left\|\partial_x y_j^n(Q_j)\right\|_{L^2(0,T)}^2
\leq C\left( ||f^n u^0||^2_{L^2(\Lambda)}
+ ||f^n\partial_tu[p]||^2_{L^1(0,T,L^2(\Lambda))} \right)
\leq 2CK^2 \|f^n\|_{L^2(\Lambda)}. $$ Thus we can write $$
\lim_{n\to \infty} \sum_{j=1}^{N_2-1}\left\|\partial_x y_j^n(Q_j)
- \partial_x y_j^0(Q_j)\right\|_{L^2(0,T)}^2
\leq 2CK^2 \lim_{n\to \infty} \|f^n - f^0\|_{L^2(\Lambda)} = 0, $$ which, combined with \eqref{hypcontr1}, gives $$ \partial_x y_j^0(Q,t) = 0, \qquad \forall Q\in \Pi_2\setminus {\{Q_{N_2}\}}, \forall t\in(0,T). $$ We finally apply Theorem~\ref{Thm1} and obtain $f^0 = 0$ in $L^2(\Lambda)$, which contradicts \eqref{f0}. Thus the proof of Theorem~\ref{Thm2} is complete.
\providecommand{\MR}[1]{}
\end{document}
|
arXiv
|
{
"id": "1407.5566.tex",
"language_detection_score": 0.6581422090530396,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\title{Computation of isotopisms of algebras over finite fields by means of graph invariants} \author{O. J. Falc\'on$^{1}$} \ead{[email protected]} \author{R. M. Falc\'on$^2$} \ead{[email protected]} \author{J. N\'u\~nez$^1$} \ead{[email protected]} \author{A. M. Pacheco$^{3}$} \ead{[email protected]} \author{M. T. Villar$^1$} \ead{[email protected]} \address{$^1$ Department of Geometry and Topology. University of Seville, Spain.\\ $^2$ Department of Applied Mathematics I. University of Seville, Spain.\\ $^3$ Department of Quantitative Methods. Loyola University Andalusia, Spain. }
\begin{abstract} In this paper we define a pair of faithful functors that map isomorphic and isotopic finite-dimensional algebras over finite fields to isomorphic graphs. These functors reduce the cost of computation that is usually required to determine whether two algebras are isomorphic. In order to illustrate their efficiency, we determine explicitly the classification of two- and three-dimensional partial quasigroup rings. \end{abstract}
\begin{keyword} Graph theory \sep finite field \sep isomorphism \sep Latin square. \MSC 05C25 \sep 05C30 \sep 05B15. \end{keyword} \end{frontmatter}
\section{Introduction}
Graph invariants constitute an interesting tool in Chemistry, Communication or Engineering \cite{Dobrynin2001, Khalifeh2010, Yousefi2011}. In Mathematics, one of the topics for which graph invariants have revealed to play an important role is the classical problem of deciding whether two algebras are isomorphic. This problem is usually dealt with by computing the reduced Gr\"obner basis of the system of polynomial equations that is uniquely related to the structure constants of both algebras. This computation is, however, very sensitive to the number of variables \cite{Gao2009} and gives rise to distinct problems of computation time and memory usage even for low-dimensional algebras \cite{Falcon2016a, Graaf2005}. This paper deals with Graph Theory in order to reduce this cost of computation.
Graph invariants have been proposed in the last years as an efficient alternative to study isomorphisms of distinct types of algebras \cite{Bocian2014, Ceballos2016, Kaveh2011}. Nevertheless, the problem of identifying a functor that relates the category of algebras with that of graphs remains still open. Based on a proposal of McKay et al. \cite{McKay2007} for identifying isotopisms of Latin squares with isomorphisms of vertex-colored graphs, we describe in Section 3 a pair of graphs that enable us to find faithful functors between finite-dimensional algebras over finite fields and these types of graphs. These functors map isomorphic and isotopic algebras to isomorphic graphs. Reciprocally, any pair of isomorphic graphs is uniquely related to a pair of algebras so that there exists a multiplicative map between them. The main advantage of our proposal, apart from the reduction of the mentioned cost of computation, is the feasibility of studying the possible isomorphism between two given finite-dimensional algebras defined over the same field, whatever the types of both algebras are. As an illustrative example, we focus in Section 4 on the classification of partial quasigroup rings according to the known isotopism classes of partial Latin squares on which they are based.
\section{Preliminaries}
In this section we expose some basic concepts and results on Graph Theory, isotopisms of algebras, partial Latin squares and Computational Algebraic Geometry that we use throughout the paper. For more details about these topics we refer, respectively, to the manuscripts \cite{Harary1969, Albert1942, Denes1974, Cox1998}.
\subsection{Graph Theory}
A {\em graph} is a pair $G=(V,E)$ formed by a set $V$ of {\em vertices} and a set $E$ of $2$-subsets of $V$ called {\em edges}. Two vertices defining an edge are said to be {\em adjacent}. The {\em degree} of a vertex $v$ is the number $d(v)$ of edges containing $v$. The graph $G$ is {\em vertex-colored} if there exists a partition of $V$ into color sets. The color of a vertex $v$ is denoted as $\mathrm{color}(v)$. An {\em isomorphism} between two vertex-colored graphs $G$ and $G'$ is any bijective map $f$ between their sets of vertices that preserves collinearity and color sets, that is, such that it maps edges to edges and $\mathrm{color}(f(v))=\mathrm{color}(v)$, for all vertex $v$ in $G$.
\subsection{Isotopisms of algebras}
Two algebras $A$ and $A'$ over a field $\mathbb{K}$ are said to be {\em isotopic} if there exist three non-singular linear transformations $f$, $g$ and $h$ from $A$ to $A'$ such that $f(u)g(v)=h(uv)$, for all $u,v\in A$. The triple $(f,g,h)$ is an {\em isotopism} between $A$ and $A'$. If $f=g=h$, then this constitutes an {\em isomorphism}.
The {\em structure constants} of an $n$-dimensional algebra $A$ over a field $\mathbb{K}$ of basis $\{e_1,\ldots,e_n\}$ are the numbers $c_{ij}^k\in \mathbb{K}$ such that $e_ie_j = \sum_{k=1}^n c_{ij}^k e_k$, for all $i, j \leq n$. If all of them are zeros, then $A$ is {\em abelian}. In particular, the $n$-dimensional abelian algebra is not isotopic to any other $n$-dimensional algebra.
The {\em left annihilator} of a vector subspace $S$ of the algebra $A$ is the set $\mathrm{Ann}_{A^-}(S)=\{u\in A\mid\, uv=0, \text { for all } v\in S\}$. Its {\em right annihilator} is the set $\mathrm{Ann}_{A^+}(S)=\{u\in A\mid\, vu=0, \text { for all } v\in S\}$. The intersection of both sets is the {\em annihilator} $\mathrm{Ann}_A(S)$.
\begin{lemm} \label{lemm_annihilator} Let $(f,g,h)$ be an isotopism between two $n$-dimensional algebras $A$ and $A'$, and let $S$ be a vector subspace of $A$. Then, \begin{enumerate}[a)] \item $f(\mathrm{Ann}_{A^-}(S)) = \mathrm{Ann}_{{A'}^-}(g(S))$. \item $g(\mathrm{Ann}_{A^+}(S)) = \mathrm{Ann}_{{A'}^+}(f(S))$. \item $f(\mathrm{Ann}_{A^-}(S))\cap g(\mathrm{Ann}_{A^+}(S)) = \mathrm{Ann}_{A'}(f(S)\cap g(S)).$ \end{enumerate} \end{lemm}
\begin{proof} Let us prove assertion (a). Assertion (b) follows similarly and assertion (c) is a consequence of (a) and (b). Let $u\in g(S)$ and $v\in f(\mathrm{Ann}_{A^-}(S))$. Then, $vu=f(f^{-1}(v))g(g^{-1}(u))=h(f^{-1}(v)g^{-1}(u))=h(0)=0$, because $g^{-1}(u)\in S$ and $f^{-1}(v)\in \mathrm{Ann}_{A^-}(S)$. Hence, $f(\mathrm{Ann}_{A^-}(S))$ $\subseteq \mathrm{Ann}_{{A'}^-}(g(S))$. Now, let $u\in \mathrm{Ann}_{{A'}^-}(g(S))$ and $v\in S$. From the regularity of $f$, we have that $h(f^{-1}(u)v)=ug(v)=0$. The regularity of $h$ involves that $f^{-1}(u)v=0$. Thus, $u\in f(\mathrm{Ann}_{A^-}(S))$ and hence, $\mathrm{Ann}_{{A'}^-}(g(S))\subseteq f(\mathrm{Ann}_{A^-}(S))$. \end{proof}
The {\em derived algebra} of $A$ is the subalgebra $A^2=\{uv\mid\, u,v\in A\}\subseteq A$.
\begin{lemm} \label{lemm_M1} Let $(f,g,h)$ be an isotopism between two $n$-dimensional algebras $A$ and $A'$. Then, $h(A^2)=A'^2$. \end{lemm}
\begin{proof} The regularity of $f$ and $g$ involves that $f(A)=g(A)=A'$ and hence, $A'^2=f(A)g(A)=h(A^2)$. \end{proof}
Let $\cdot$ be a partial binary operation over the set $[n]=\{1,\ldots,n\}$. The pair $([n],\cdot)$ is called a {\em partial magma} of {\em order} $n$. It is {\em isotopic} to a partial magma $([n],\circ)$ if there exist three permutations $\alpha$, $\beta$ and $\gamma$ in the symmetric group $S_n$ such that $\alpha(i)\circ\beta(j)=\gamma(i\cdot j)$, for all $i,j\leq n$ such that $i\cdot j$ exists. If $\alpha=\beta=\gamma$, then the partial magmas are said to be {\em isomorphic}. The triple $(\alpha,\beta,\gamma)$ is an {\em isotopism} of partial magmas (an {\em isomorphism} if $\alpha=\beta=\gamma$).
A {\em partial magma algebra} $A^{\cdot}$ {\em based on} a partial magma $([n],\cdot)$ is an $n$-dimensional algebra over a field $\mathbb{K}$ such that there exists a basis $\{e_1,\ldots,e_n\}$ satisfying that, if $i\cdot j$ exists for some pair of elements $i,j\leq n$, then $e_ie_j=c_{ij}e_{i\cdot j}$ for some non-zero structure constant $c_{ij}\in \mathbb{K}\setminus\{0\}$. If all the structure constants are equal to $1$, then this is called a {\em partial magma ring}.
\begin{lemm}\label{lemm_partial_magma} Two partial magma rings are isotopic (isomorphic, respectively) if their respective partial magmas on which they are based are isotopic (isomorphic, respectively). \end{lemm}
\begin{proof} Let $A^{\cdot}$ and $A^{\circ}$ be two partial magma rings based, respectively, on two isotopic partial magmas $([n],\cdot)$ and $([n],\circ)$. Let $\{e_1,\ldots,e_n\}$ and $\{e'_1,\ldots,e'_n\}$ be the respective bases of these two algebras and let $(f,g,h)$ be an isotopism between their corresponding partial magmas. For each $\alpha\in\{f,g,h\}$, let us define the map $\overline{\alpha}(e_i)=e'_{\alpha(i)}$. Then, $\overline{f}(e_i)\overline{g}(e_j)=e'_{f(i)}e'_{g(j)}= e'_{f(i)\circ g(j)}=e'_{h(i\cdot j)} =\overline{h}(e_{i\cdot j})= \overline{h}(e_ie_j)$. From linearity, the triple $(\overline{f},\overline{g},\overline{h})$ determines an isotopism between $A^{\cdot}$ and $A^{\circ}$. If $f=g=h$, then this constitutes an isomorphism. \end{proof}
The reciprocal of Lemma \ref{lemm_partial_magma} is not true in general. Thus, for instance, the two partial magmas $([2],\cdot)$ and $([2],\circ)$ that are respectively described by the non-zero products $1\cdot 1=1$ and $1\circ 1 = 1 = 2\circ 1$ are not isotopic. Nevertheless, the partial magma rings $A^{\cdot}$ and $A^{\circ}$, with respective bases $\{e_1,e_2\}$ and $\{e'_1,e'_2\}$, are isotopic by means of the isotopism $(f,\mathrm{Id},\mathrm{Id})$, where the linear transformation $f$ is described by $f(e_1)=e'_1$ and $f(e_2) = e'_2-e'_1$.
\subsection{Partial Latin squares}
A {\em partial quasigroup} is a partial magma $([n],\cdot)$ such that if the equations $ix=j$ and $yi=j$, with $i,j\in [n]$, have solutions for $x$ and $y$ in $[n]$, then these solutions are unique. The concepts of {\em partial quasigroup algebras} and {\em partial quasigroup rings} arise similarly to those of partial magma algebras and rings. Lemma \ref{lemm_partial_magma} also holds analogously for partial quasigroup rings. Every partial quasigroup of order $n$ constitutes the multiplication table of a {\em partial Latin square} of order $n$, that is, an $n \times n$ array in which each cell is either empty or contains one element chosen from the set $[n]$, such that each symbol occurs at most once in each row and in each column. Every isotopism of a partial quasigroup is uniquely related to a permutation of the rows, columns and symbols of the corresponding partial Latin square. The distribution of partial Latin squares into isotopism classes is known for order up six \cite{Falcon2013, Falcon2015a}. In this paper we make use of graph invariants to study which ones of the known non-isotopic classes of partial Latin squares of order $n\leq 3$ give rise to isotopic classes of partial quasigroup rings over the finite fields $\mathbb{F}_2$ and $\mathbb{F}_3$. In this regard, it is straightforwardly verified that there exists only two one-dimensional partial quasigroup rings: the abelian and that one described by the product $e_1e_1=e_1$. They constitute distinct isotopism classes.
Let $L=(l_{ij})$ be a partial Latin square of order $n$ without empty cells (that is, a {\em Latin square}). McKay et al. \cite{McKay2007} defined the vertex-colored graph $G(L)$ with $n^2+3n$ vertices $\{r_i\mid\, i\leq n\}\cup\{c_i\mid\, i\leq n\}\cup\{s_i\mid\, i\leq n\}\cup \{t_{ij}\mid\, i,j\leq n\}$, where each of the four subsets (related to the rows ($r_i$), columns ($c_i$), symbols ($s_i$) and cells ($t_{ij}$) of the Latin square $L$) has a different color, and $3n^2$ edges $\{r_it_{ij},c_jt_{ij},s_{l_{ij}}t_{ij}\mid\, i,j\leq n\}\}$ (see Figure \ref{Fig_LS}, where we have used distinct styles ($\circ$, $\blacktriangle$, $\blacktriangleright$, $\blacktriangleleft$ and $\bullet$) to represent the colors of the vertices). Two Latin squares $L_1$ and $L_2$ of the same order are isotopic if and only if the graphs $G(L_1)$ and $G(L_2)$ are isomorphic (see Theorem 6 in \cite{McKay2007}).
\begin{figure}
\caption{Graph related to a Latin square of order $2$.}
\label{Fig_LS}
\end{figure}
\subsection{Computational Algebraic Geometry}
Let $\mathbb{K}[X]$ be a multivariate polynomial ring over a field $\mathbb{K}$. The {\em algebraic set} defined by an ideal $I$ of $\mathbb{K}[X]$ is the set $\mathcal{V}(I)$ of common zeros of all the polynomials in $I$. If this set is finite, then the ideal $I$ is {\em zero-dimensional}. This is {\em radical} if every polynomial $f\in \mathbb{K}[X]$ belongs to $I$ whenever there exists a natural number $m$ such that $f^m\in I$. The largest monomial of a polynomial in $I$ with respect to a given monomial term ordering is its {\em leading monomial}. The ideal generated by all the leading monomials of $I$ is its {\em initial ideal}. A {\em standard monomial} of $I$ is any monomial that is not contained in its initial ideal. Regardless of the monomial term ordering, if the ideal $I$ is zero-dimensional and radical, then the number of standard monomials in $I$ coincides with the Krull dimension of the quotient ring $\mathbb{K}[X]/I$ and with the number of points of the algebraic set $\mathcal{V}(I)$. This is computed from the reduced Gr\"obner basis of the ideal. Specifically, a {\em Gr\"obner basis} of the ideal $I$ is any subset $G$ of polynomials in $I$ whose leading monomials generate its initial ideal. This is {\em reduced} if all its polynomials are monic and no monomial of a polynomial in $G$ is generated by the leading monomials of the rest of polynomials in the basis. There exists only one reduced Gr\"obner basis, which can always be computed from Buchberger's algorithm \cite{Buchberger2006}. The computation that is required to this end is extremely sensitive to the number of variables.
\begin{thm}[\cite{Gao2009}, Proposition 4.1.1]\label{Gao} Let $\mathbb{F}_q$ be a finite field, with $q$ a power prime. The complexity time that Buchberger's algorithm requires to compute the reduced Gr\"obner bases of an ideal $\langle\, p_1,\ldots, p_m,p_1^q-p_1,\ldots,p^q_m-p_m\,\rangle$ defined over a polynomial ring $\mathbb{F}_q[x_1,\ldots,x_n]$, where $p_1,\ldots,p_m$ are polynomials given in sparse form and have longest length $l$, is $q^{O(n)}+O(m^2l)$. Here, sparsity refers to the number of monomials. \end{thm}
Gr\"obner bases can be used to determine the isomorphisms and isotopisms between two $n$-dimensional algebras $A$ and $A'$ over a finite field $\mathbb{F}_q$, with $q$ a prime power, respective basis $\{e_1,\ldots,e_n\}$ and $\{e'_1,\ldots,e'_n\}$, and respective structure constants $c_{ij}^k$ and ${c'}_{ij}^k$. To this end, let us define the sets of variables $\mathfrak{F}_n=\{\mathfrak{f}_{ij}\mid\, i,j\leq n\}$, $\mathfrak{G}_n=\{\mathfrak{g}_{ij}\mid\, i,j\leq n\}$ and $\mathfrak{H}_n=\{\mathfrak{h}_{ij}\mid\, i,j\leq n\}$. These variables play the respective role of the entries in the regular matrices related to a possible isotopism $(f,g,h)$ between the algebras $A$ and $A'$. Here, $\alpha(e_i)=\sum_{j=1}^n \alpha_{ij}e'_j$, for each $\alpha\in\{f,g,h\}$. From the coefficients of each basis vector $e_m$ in the expression $f(e_i)g(e_j)=h(e_ie_j)$, we have that $$\sum_{k,l=1}^n \mathfrak{f}_{ik}\mathfrak{g}_{jl}{c'}_{kl}^m = \sum_{s=1}^n c_{ij}^s\mathfrak{h}_{sm}, \text{ for all } i,j,m\leq n.$$
\begin{thm}\label{thm_CAG_Isom} The next two assertions hold. \begin{enumerate}[a)] \item The isotopism group between the algebras $A$ and $A'$ is identified with the algebraic set of the ideal $I^{\mathrm{Isot}}_{A,A'}$ of $\mathbb{F}_q[\mathfrak{F}_n\cup\mathfrak{G}_n\cup\mathfrak{H}_n]$, which is defined as
{\small $$\langle\, \sum_{k,l=1}^n \mathfrak{f}_{ik}\mathfrak{g}_{jl}{c'}_{kl}^m - \sum_{s=1}^n c_{ij}^s\mathfrak{h}_{sm}\mid\, i,j,m\leq n\,\rangle + \langle\,\det(M)^{q-1}-1\mid\, M\in\{F,G,H\}\,\rangle,$$}
\noindent where $F$, $G$ and $H$ denote, respectively, the matrices of entries in $\mathfrak{F}_n$, $\mathfrak{G}_n$ and $\mathfrak{H}_n$. Besides, $|\mathcal{V}(I^{\mathrm{Isot}}_{A,A'})|= \mathrm{dim}_{\mathbb{F}_q} (\mathbb{F}_q[\mathfrak{F}_n\cup\mathfrak{G}_n\cup\mathfrak{H}_n]/ I^{\mathrm{Isot}}_{A,A'})$. \item The isomorphism group between the algebras $A$ and $A'$ is identified with the algebraic set of the ideal $I^{\mathrm{Isom}}_{A,A'}$ of $\mathbb{F}_q[\mathfrak{F}_n]$, which is defined as $$\langle\, \sum_{k,l=1}^n \mathfrak{f}_{ik}\mathfrak{f}_{jl}{c'}_{kl}^m - \sum_{s=1}^n c_{ij}^s\mathfrak{f}_{sm}\mid\, i,j,m\leq n \,\rangle + \langle\,\det(F)^{q-1}-1\,\rangle,$$ where $F$ denotes the matrix of entries in $\mathfrak{F}_n$. Besides,
$|\mathcal{V}(I^{\mathrm{Isom}}_{A,A'})|= \mathrm{dim}_{\mathbb{F}_q}(\mathbb{F}_q[\mathfrak{F}_n]/ I^{\mathrm{Isom}}_{A,A'})$. \end{enumerate} \end{thm}
\begin{proof} Let us prove the second assertion, being analogous the reasoning for assertion (a). The generators of the ideal $I^{\mathrm{Isom}}_{A,A'}$ involve each zero $(f_{11},\ldots,$ $f_{nn})$ of its algebraic set to constitute the entries of the regular matrix of an isomorphism $f$ between the algebras $A$ and $A'$. The result follows from the fact of being this ideal zero-dimensional and radical. Particularly, the ideal $I^{\mathrm{Isom}}_{A,A'}$ is zero-dimensional because its algebraic set is a finite subset of $\mathbb{F}_q^{n^2}$. Besides, from Proposition 2.7 of \cite{Cox1998}, the ideal $I$ is also radical, because, for each $i,j\leq n$, the unique monic generator of $I\cap \mathbb{F}_q[\mathfrak{f}_{ij}]$ is the polynomial $(\mathfrak{f}_{ij})^q-\mathfrak{f}_{ij}$, which is intrinsically included in each ideal of $\mathbb{F}_q[\mathfrak{F}_n]$ and is square-free. \end{proof}
\begin{corollary}\label{coro_CAG_Isom} The complexity times that Buchberger's algorithm requires to compute the reduced Gr\"obner bases of the ideals $I^{\mathrm{Isot}}_{A,A'}$ and $I^{\mathrm{Isom}}_{A,A'}$ in Theorem \ref{thm_CAG_Isom} are, respectively, $q^{O(3n^2)}+O(n^6n!)$ and $q^{O(n^2)}+O(n^6n!)$. \end{corollary}
\begin{proof} We prove the result for the second ideal, being analogous the reasoning for the first one. The result follows straightforwardly from Theorem \ref{Gao} once we observe that all the generators of the ideal in Theorem \ref{thm_CAG_Isom} are sparse in $\mathbb{F}_q[\mathfrak{F}_n]$. More specifically, the number of variables is $n^2$, the number of generators of the ideal under consideration that are not of the form $(\mathfrak{f}_{ij})^q-\mathfrak{f}_{ij}$ is $n^3+1$ and the maximal length of these generators is $n!$. \end{proof}
Theorem \ref{thm_CAG_Isom} has been implemented as a procedure called {\em isoAlg} in the open computer algebra system for polynomial computations {\sc Singular} \cite{Decker2016}. This has been included in the library {\em GraphAlg.lib}, which is available online at {\texttt{http://personales.us.es/raufalgan/LS/GraphAlg.lib}}. Let us illustrate the use of this procedure with an example related to the distribution of the set $\mathcal{P}_2(\mathbb{F}_2)$ of two-dimensional partial quasigroup rings over the finite field $\mathbb{F}_2$ into isotopism and isomorphism classes. All the computations that are exposed throughout this paper are implemented in a system with an {\em Intel Core i7-2600, with a 3.4 GHz processor and 16 GB of RAM}.
\begin{example}\label{ejemplo_PQ2} Let us consider the pair of partial quasigroup rings in $\mathcal{P}_2(\mathbb{F}_2)$ that are respectively related to the partial Latin squares
$$\begin{array}{|c|c|} \hline 1 & 2\\ \hline 2 & \ \\ \hline \end{array} \hspace{0.25cm} \text { and } \hspace{0.25cm} \begin{array}{|c|c|} \hline 1 & 2\\ \hline 2 & 1 \\ \hline \end{array}$$ These two partial Latin squares are not isotopic because isotopisms preserve the number of filled cells. Nevertheless, their related partial quasigroup rings over $\mathbb{F}_2$, with respective bases $\{e_1,e_2\}$ and $\{e'_1,e'_2\}$, and which are respectively described by the products $$\begin{cases}e_1e_1=e_1,\\ e_1e_2=e_2=e_2e_1. \end{cases} \hspace{0.5cm} \text{ and } \hspace{0.5cm} \begin{cases}e'_1e'_1=e'_1=e'_2e'_2,\\ e'_1e'_2=e'_2=e'_2e'_1. \end{cases}$$ are isotopic. Specifically, by implementing the procedure {\em isoAlg}, our system computes in $0$ seconds the existence of four isotopisms between these two partial quasigroup rings. One of this isotopisms is, for instance, the isomorphism $f$ such that $f(e_1)=e'_1$ and $f(e_2)=e'_1+e'_2$. The procedure {\em isoAlg} also ensures us the existence of $f$ as the unique possible isomorphism.
$\lhd$ \end{example}
In practice, in those cases in which the run time required for the computations involved in Theorem \ref{thm_CAG_Isom} becomes excessive, it is recommendable to eliminate the generators of the corresponding ideal that are referred to the determinants of the matrices $F$, $G$ and $H$. This reduces the complexity time in Corollary \ref{coro_CAG_Isom} to $q^{O(3n^2)} + O(n^8)$ and $q^{O(n^2)} + O(n^8)$, respectively, and gives enough information to analyze a case study on which base the possible isomorphisms and isotopisms between two given algebras, whatever the base field is. The next example illustrates this fact by focusing on the possible isotopisms that there exist over any field between the two partial quasigroup rings that appear in Example \ref{ejemplo_PQ2}.
\begin{example}\label{ejemplo_PQ2a} The implementation of the procedure {\em isoAlg} enables us to ensure that, whatever the base field is, the reduced Gr\"obner basis of the ideal $I^{\mathrm{Isot}}_{A,A'}$ in Theorem \ref{thm_CAG_Isom} related to the isotopism group between the two partial quasigroup rings of Example \ref{ejemplo_PQ2} holds that $2\mathfrak{h}_{22}^3=0$ and $\mathfrak{h}_{21}^2+\mathfrak{h}_{22}^2=0$. If the characteristic of the base field is not two, then $\mathfrak{h}_{21}=\mathfrak{h}_{22}=0$. This involves $H$ to be singular and hence, these two partial quasigroup rings are not isotopic. Otherwise, it is straightforwardly verified that the linear transformation $f$ that is indicated in Example \ref{ejemplo_PQ2} constitutes an isomorphism between both rings for every base field of characteristic two.
$\lhd$ \end{example}
\section{Description of faithful functors between algebras and graphs}
Based on the proposal of McKay et al. \cite{McKay2007} for Latin squares, we describe now a pair of graphs that are uniquely related to a finite-dimensional algebra $A$ over a finite field $\mathbb{K}$. Firstly, we define the vertex-colored graph $G_1(A)$ with four maximal monochromatic subsets $R_{A}=\{r_u\mid\, u\in A\setminus \mathrm{Ann}_{A^-}(A)\}$, $C_{A}=\{c_u\mid\, u\in A\setminus \mathrm{Ann}_{A^+}(A)\}$, $S_{A}=\{s_u\mid\, u\in A^2\setminus \{0\}\}$ and $T_{A}=\{t_{u,v}\mid\, u,v\in A, uv\neq 0\}$, and edges $\{r_ut_{u,v}, c_vt_{u,v}, s_{uv}t_{u,v}\mid u,v\in A, uv\neq 0\}$. From this graph we also define the vertex-colored graph $G_2(A)$ by adding the edges $\{r_uc_u,\mid\, u\in A\setminus \mathrm{Ann}_A(A)\}\, \cup \{c_us_u\mid\, u \in A^2\setminus \mathrm{Ann}_{A^+}(A)\} \, \cup \{r_us_u\mid\, u \in A^2\setminus \mathrm{Ann}_{A^-}(A)\}$. As an illustrative example, Figure \ref{Fig_1} shows the two graphs that are related to any $n$-dimensional algebra over the finite field $\mathbb{F}_2$, with basis $\{e_1,\ldots,e_n\}$, that is described as $e_1e_2=e_2e_1=e_1$.
\begin{figure}
\caption{Graphs related to the algebra $e_1e_2=e_2e_1=e_1$ over $\mathbb{F}_2$.}
\label{Fig_1}
\end{figure}
\begin{lemm}\label{lemm_graph0} The next assertions hold. \begin{enumerate}[a)] \item If the algebra $A$ is abelian, then $G_1(A)$ and $G_2(A)$ have no vertices. \item The graph $G_1(A)$ does not contain triangles. \item In both graphs $G_1(A)$ and $G_2(A)$, \begin{itemize} \item The number of vertices is
{\small \[|A\setminus \mathrm{Ann}_{A^-}(A)|+|A\setminus \mathrm{Ann}_{A^+}(A)|+
|A^2|+|\{(u,v)\in A\times A\mid\, uv\neq 0\}| - 1. \]} \item The degree of the vertex $t_{u,v}$ is \[d(t_{u,v})=3, \text{ for all } u,v\in A \text{ such that } uv\neq 0.\] \end{itemize} \item In the graph $G_1(A)$, \begin{itemize}
\item $d(r_u)=|A\setminus \mathrm{Ann}_{A^+}(\{u\})|$, for all $u\not\in \mathrm{Ann}_{A^-}(A)$.
\item $d(c_u)=|A\setminus \mathrm{Ann}_{A^-}(\{u\})|$, for all $u\not\in \mathrm{Ann}_{A^+}(A)$.
\item $d(s_u)=\sum_{v\in A}|\mathrm{ad}^{-1}_v(u)|$, for all $u\in A^2\setminus\{0\}$, where $\mathrm{ad}_v:A\rightarrow A^2$ is the {\em adjoint action} of $v$ in $A$ such that $\mathrm{ad}_v(w)=vw$, for all $w\in A$. \end{itemize} \item Let $\mathbf{1}$ denotes the characteristic function. Then, in the graph $G_2(A)$, \begin{itemize}
\item $d(r_u)=|A\setminus \mathrm{Ann}_{A^+}(\{u\})|+ \mathbf{1}_{A\setminus\mathrm{Ann}_{A^+}(A)}(u) + \mathbf{1}_{A^2}(u)$, for all $u\not\in \mathrm{Ann}_{A^-}(A)$.
\item $d(c_u)=|A\setminus \mathrm{Ann}_{A^-}(\{u\})|+ \mathbf{1}_{A\setminus\mathrm{Ann}_{A^-}(A)}(u) + \mathbf{1}_{A^2}(u)$, for all $u\not\in \mathrm{Ann}_{A^+}(A)$.
\item $d(s_u)=\mathbf{1}_{A\setminus\mathrm{Ann}_{A^-}(A)}(u)+ \mathbf{1}_{A\setminus\mathrm{Ann}_{A^+}(A)}(u)+\sum_{v\in A}|\mathrm{ad}^{-1}_v(u)|$, for all $u\in A^2\setminus\{0\}$. \end{itemize} \end{enumerate} \end{lemm}
\begin{proof} The result follows straightforwardly from the definition of the graphs $G_1(A)$ and $G_2(A)$. \end{proof}
\begin{prp}\label{prop_graph0_a} The next assertions hold. \begin{enumerate}[a)] \item The number of edges of the graph $G_1(A)$ is
$\sum_{u\not\in\mathrm{Ann}_{A^-}(A)} |A\setminus \mathrm{Ann}_{A^+}(\{u\})|$ $+\sum_{u\not\in\mathrm{Ann}_{A^+}(A)}|A\setminus \mathrm{Ann}_{A^-}(\{u\})|+\sum_{u\in A^2\setminus\{0\}}\sum_{v\in A}|\mathrm{ad}_v^{-1}(u)|$.
\item The number of edges of the graph $G_2(A)$ coincides with those of $G_1(A)$ plus $|A\setminus\mathrm{Ann}_A(A)| + |A^2\setminus \mathrm{Ann}_{A^-}(A)| + |A^2\setminus\mathrm{Ann}_{A^+}(A)|$. \end{enumerate} \end{prp}
\begin{proof} The result follows from the first theorem of Graph Theory \cite{Harary1969}, which enables us to ensure that the number of edges of a graph is the half of the summation of degrees of its vertices. Now, for each pair of vectors $u,v\in A$ such that $uv\neq 0$, the vertex $t_{u,v}\in T_A$ is the only vertex in $T_A$ that is adjacent to the vertices $r_u\in R_A$, $c_v\in C_A$ and $s_{uv}\in S_A$. They constitute indeed the three vertices related to the degree of $t_{u,v}$ that is indicated in assertion (c) of Lemma \ref{lemm_graph0}. As a consequence, the summation of degrees of all the vertices in $T_A$ coincides with $\sum_{u\in R_A}d(r_u) + \sum_{u\in C_A}d(c_u) + \sum_{u\in S_A}d(s_u)$. The result follows then from assertions (d) and (e) in Lemma \ref{lemm_graph0}. \end{proof}
\begin{thm}\label{theo_graph0} Let $A$ and $A'$ be two finite-dimensional algebras over a finite field $\mathbb{K}$. Then, \begin{enumerate}[a)] \item If both algebras are isotopic, then their corresponding graphs $G_1(A)$ and $G_1(A')$ are isomorphic. Reciprocally, if the graphs $G_1(A)$ and $G_1(A')$ are isomorphic, then there exist three bijective maps $f$, $g$ and $h$ between $A$ and $A'$ such that $f(u)g(v)=h(uv)$. \item If both algebras are isomorphic, then their corresponding graphs $G_2(A)$ and $G_2(A')$ are also isomorphic. Reciprocally, if the graphs $G_2(A)$ and $G_2(A')$ are isomorphic, then there exists a multiplicative bijective map between the algebras $A$ and $A'$, that is, a bijective map $f:A\rightarrow A'$ so that $f(u)f(v)=f(uv)$, for all $u,v\in A$. \end{enumerate} \end{thm}
\begin{proof} Let $(f,g,h)$ be an isotopism between the algebras $A$ and $A'$. We define the map $\alpha$ between $G_1(A)$ and $G_1(A')$ such that $$\begin{cases}\alpha(r_u)=r_{f(u)}, \text{ for all } u\in A\setminus \mathrm{Ann}_{A^-}(A),\\ \alpha(c_u)=c_{g(u)}, \text{ for all } u\in A\setminus \mathrm{Ann}_{A^+}(A),\\ \alpha(s_u)=s_{h(u)}, \text{ for all } u\in A^2\setminus \{0\},\\ \alpha(t_{u,v})=t_{f(u),g(v)}, \text{ for all } u,v\in A \text { such that } uv\neq 0.\end{cases}$$
The description of $G_1(A)$ and $G_1(A')$, together with Lemmas \ref{lemm_annihilator} and \ref{lemm_M1}, and the regularity of $f$, $g$ and $h$, involves $\alpha$ to be an isomorphism between these two graphs. The same map $\alpha$ constitutes an isomorphism between the graphs $G_2(A)$ and $G_2(A')$ in case of being $f=g=h$, that is, if the algebras $A$ and $A'$ are isomorphic. Reciprocally, let $\alpha$ be an isomorphism between the graphs $G_1(A)$ and $G_1(A')$. Collinearity involves this isomorphism to be uniquely determined by its restriction to $R_A\cup C_A\cup S_A$. Specifically, the image of each vertex $t_{u,v}\in T_A$ by means of $\alpha$ is uniquely determined by the corresponding images of $r_u$, $c_v$ and $s_{uv}$. Let $\beta$ and $\beta'$ be the respective bases of the algebras $A$ and $A'$ and let $\pi: A\rightarrow A'$ be the natural map that preserves the components of each vector with respect to the mentioned bases. That is, $\pi((u_1,\ldots,u_n)_{\beta})=(u_1,\ldots,u_n)_{\beta'}$, for all $u_1,\ldots,u_n\in\mathbb{K}$. Let us define three maps $f$, $g$ and $h$ from $A$ to $A'$ such that $$f(u)=\begin{cases}\pi(u), \text{ for all } u\in \mathrm{Ann}_{A^-}(A),\\ v, \text{ otherwise, where } v\in A \text{ is such that } \alpha(r_u)=r_v. \end{cases}$$ $$g(u)=\begin{cases}\pi(u), \text{ for all } u\in \mathrm{Ann}_{A^+}(A),\\ v, \text{ otherwise, where } v\in A \text{ is such that } \alpha(c_u)=c_v. \end{cases}$$ $$h(u)=\begin{cases}\pi(u), \text{ for all } u\in (A\setminus A^2)\cup \{0\},\\ v, \text{ otherwise, where } v\in A \text{ is such that } \alpha(s_u)=s_v. \end{cases}$$
From Lemmas \ref{lemm_annihilator} and \ref{lemm_M1}, these three maps are bijective. Let $u,v\in A$. If $u\in \mathrm{Ann}_{A^-}(A)$ or $v\in \mathrm{Ann}_{A^+}(A)$, then there does not exist the vertex $t_{u,v}$ in the graph $G_1(A)$. Since $\alpha$ preserves collinearity, there does not exist the vertex $t_{f(u),g(v)}$ in the graph $G_1(A')$, which means that $f(u)\in \mathrm{Ann}_{{A'}^-}(A')$ or $g(v)\in \mathrm{Ann}_{{A'}^+}(A')$. In any case, we have that $f(u)g(v)=0=h(uv)$. Finally, if $u\not\in\mathrm{Ann}_{A^-}(A)$ and $v\not \in \mathrm{Ann}_{A^+}(A)$, then the vertex $t_{u,v}$ connects the vertices $r_u$, $c_v$ and $s_{uv}$ in the graph $G_1(A)$. Now, the isomorphism $\alpha$ maps this vertex $t_{u,v}$ in $G_1(A)$ to a vertex $t_{u',v'}$ in $G_2(A)$ that is connected to the vertices $r_{u'}$, $c_{v'}$ and $s_{u'v'}$. Again, since $\alpha$ preserves collinearity, it is $f(u)=u'$, $g(v)=v'$ and, finally, $h(uv)=f(u)g(v)$.
In case of being $\alpha$ an isomorphism between the graphs $G_2(A)$ and $G_2(A')$ it is enough to consider $f=g=h$ in the previous description. This is well-defined because of the new edges that are included to the graphs $G_1(A)$ and $G_1(A')$ in order to define, respectively, the graphs $G_2(A)$ and $G_2(A')$. These edges involve the multiplicative character of the bijective map $f$, that is, $f(u)g(v)=h(uv)$, for all $u,v\in A$. \end{proof}
Theorem \ref{theo_graph0} enables us to ensure that graph invariants reduce the cost of computation that is required to distribute a set of finite-dimensional algebras over finite fields into isotopism and isomorphism classes. It is only necessary to compute those reduced Gr\"obner bases in Theorem \ref{thm_CAG_Isom} that are related to a pair of algebras whose associated graphs have equal invariants. The complexity to compute these invariants is always much less than that related to the calculus of a reduced Gr\"obner basis. Thus, for instance, the complexity to compute the number of vertices, edges and triangles of the graphs related to any $n$-dimensional algebra over the finite field $\mathbb{F}_q$ is $q^{O(2n)}$. This corresponds to the computation of the adjacency matrices of both graphs by means of all the possible products among the $q^n$ distinct vectors of any such an algebra. This enables us in particular to implement the formulas exposed in Lemma \ref{lemm_graph0} and Proposition \ref{prop_graph0_a}. Besides, recall that the trace of the adjacency matrix of a graph raised to the third power coincides with the number of triangles of such a graph. All this computation has been implemented in the procedure {\em isoGraph}, which has been included in the mentioned library {\em GraphAlg.lib}. In order to illustrate the efficiency of these invariants, we focus on the set $\mathcal{L}_n(\mathbb{F}_q)$ of $n$-dimensional Lie algebras over the finite field $\mathbb{F}_q$, with $q$ a power prime. Recall that a {\em Lie algebra} is an anti-commutative algebra $A$ that holds the {\em Jacobi identity} $u(vw)+v(wu)+w(uv)=0$, for all $u,v,w\in A$. For $n=3$, it is known \cite{Falcon2016a, Graaf2005, Strade2007} that there are $32$ distinct Lie algebras in $\mathcal{L}_3(\mathbb{F}_2)$, which are distributed into four isotopism classes and six isomorphism classes, and $123$ Lie algebras in $\mathcal{L}_3(\mathbb{F}_3)$, which are distributed into four isotopism classes and seven isomorphism classes. Table \ref{Table_0} shows the run time and memory usage that our computer system requires to determine the mentioned classification depending on whether graph invariants are considered ({\em Graph}) or not ({\em Alg}). Further, Tables \ref{Table_1} and \ref{Table_1a} show, respectively, the invariants of the graphs $G_1$ and $G_2$ related to the isomorphism classes of $\mathcal{L}_3(\mathbb{F}_q)$, for $q\in\{2,3\}$. The components of the $4$-tuples that are indicated in the corresponding columns of vertices in both tables refer, respectively, to the number of vertices in $R_A$, $C_A$, $S_A$ and $T_A$.
\begin{table}[ht] \begin{center} \resizebox{\textwidth}{!}{
\begin{tabular}{c|cc|cc|cc}
\ & \multicolumn{2}{c|}{Graph} & \multicolumn{4}{c}{Alg}\\
\ & & & \multicolumn{2}{c|}{Isomorphisms} & \multicolumn{2}{c}{Isotopisms}\\ $q$ & Run time & Memory usage & Run time & Memory usage & Run time & Memory usage\\ \hline 2 & 1 s & 0 Mb & 1 s & 0 Mb & 34 s & 384 Mb\\ 3 & 47 s & 3 Mb & 4 s & 6 Mb & \multicolumn{2}{c}{Run out of memory}\\ \end{tabular}} \caption{Cost of computation to distribute $\mathcal{L}_3(\mathbb{F}_q)$ into isotopism and isomorphism classes.} \label{Table_0} \end{center} \end{table}
\begin{table}[ht] \resizebox{\textwidth}{!}{
\begin{tabular}{c|cc|cc}
\ & \multicolumn{2}{c|}{$\mathbb{F}_2$} & \multicolumn{2}{c}{$\mathbb{F}_3$} \\ $ A$ & Vertices & Edges & Vertices & Edges \\ \hline Abelian & (0,0,0,0) & 0 & (0,0,0,0) & 0\\ $e_1e_2=e_3$ & (6,6,1,24) & 72 & (24,24,2,432) & 1296\\ $e_1e_2=e_1$ & (6,6,1,24) & 72 & (24,24,2,432) & 1296\\ $e_1e_2=e_3, e_1e_3=-e_2$ & - & - & (26,26,8,576) & 1728 \\ $e_1e_2=e_3, e_1e_3=e_2$ & (7,7,3,36) & 108 & (26,26,8,576) & 1728\\ $e_1e_2=e_2, e_1e_3=e_3$ & (7,7,3,36) & 108 & (26,26,8,576) & 1728\\ $e_1e_2=e_2, e_1e_3=-e_3, e_2e_3=-e_1$ & (7,7,7,42) & 126 & - & -\\ $e_1e_2=e_2, e_1e_3=-e_3, e_2e_3=2e_1$ & - & - & (26,26,26,624) & 1872\\ \end{tabular}} \caption{Invariants of the graph $G_1$ related to $\mathcal{L}_3(\mathbb{F}_q)$, for $q\in\{2,3\}$.} \label{Table_1} \end{table}
\begin{table}[ht] \resizebox{\textwidth}{!}{
\begin{tabular}{c|ccc|ccc}
\ & \multicolumn{3}{c|}{$\mathbb{F}_2$} & \multicolumn{3}{c}{$\mathbb{F}_3$} \\ $A$ & Vertices & Edges & Triangles & Vertices & Edges & Triangles\\ \hline Abelian & (0,0,0,0) & 0 & 0 & (0,0,0,0) & 0 & 0\\ $e_1e_2=e_3$ & (6,6,1,24) & 78 & 0 & (24,24,2,432) & 1320 & 0\\ $e_1e_2=e_1$ & (6,6,1,24) & 80 & 9 & (24,24,2,432) & 1324 & 38\\ $e_1e_2=e_3, e_1e_3=-e_2$ & - & - & - & (26,26,8,576) & 1770 & 8\\ $e_1e_2=e_3, e_1e_3=e_2$ & (7,7,3,36) & 121 & 11 & (26,26,8,576) & 1770 & 80\\ $e_1e_2=e_2, e_1e_3=e_3$ & (7,7,3,36) & 121 & 27 & (26,26,8,576) & 1770 & 152\\ $e_1e_2=e_2, e_1e_3=-e_3, e_2e_3=-e_1$ & (7,7,7,42) & 147 & 19 & - & - & -\\ $e_1e_2=e_2, e_1e_3=-e_3, e_2e_3=2e_1$ & - & - & - & (26,26,26,624) & 1950 & 74\\ \end{tabular}} \caption{Invariants of the graph $G_2$ related to $\mathcal{L}_3(\mathbb{F}_q)$, for $q\in\{2,3\}$.} \label{Table_1a} \end{table}
\section{Graphs and partial quasigroup rings}
Let us finish with an illustrative example that focuses on those graphs $G_1$ and $G_2$ related to the set $\mathcal{P}_n(\mathbb{K})$ of $n$-dimensional partial quasigroup rings over a finite field $\mathbb{K}$ that are based on the known distribution of partial Latin squares of order $n\leq 3$ into isotopism classes. In this regard, Table \ref{Table_PQ} shows several graph invariants that are related to the isotopism classes of $\mathcal{P}_2(\mathbb{F}_q)$, for $q\in\{2,3\}$. Partial Latin squares are written row after row in a single line, with empty cells represented by zeros.
\begin{table}[ht] \begin{center} \resizebox{\textwidth}{!}{
\begin{tabular}{c|c|c|cc|c|c|cc}
\ & \multicolumn{4}{c|}{$\mathbb{F}_2$} & \multicolumn{4}{c}{$\mathbb{F}_3$}\\
\ & $G_1\, \& \, G_2$ & $G_1$ & \multicolumn{2}{c|}{$G_2$} & $G_1\, \& \, G_2$ & $G_1$ & \multicolumn{2}{c}{$G_2$}\\ Partial Latin square & Vertices & Edges & Edges & Triangles & Vertices & Edges & Edges & Triangles\\ \hline 00 00 & (0,0,0,0) & 0 & 0 & 0 & (0,0,0,0) & 0 & 0 & 0\\ 10 00 & (2,2,1,4) & 12 & 16 & 7 & (6,6,2,36) & 108 & 118 & 20\\ 10 01 & (3,3,1,6) & 18 & 23 & 7 & (8,8,2,48) & 144 & 156 & 22\\ 10 02 & (3,3,3,7) & 21 & 30 & 16 & (8,8,8,56) & 168 & 192 & 48\\ 10 20 & (3,2,3,6) & 18 & 25 & 12 & (8,6,8,48) & 144 & 164 & 42\\ 12 00 & (2,3,3,6) & 18 & 25 & 12 & (6,8,8,48) & 144 & 164 & 42\\ 12 20 & (3,3,3,8) & 24 & 33 & 13 & (8,8,8,60) & 180 & 204 & 38\\ 12 21 & (3,3,3,8) & 24 & 33 & 13 & (8,8,8,56) & 168 & 192 & 48\\ \end{tabular}} \caption{Invariants of the graphs $G_1$ and $G_2$ related to $\mathcal{P}_2(\mathbb{F}_q)$, for $q\in\{2,3\}$.} \label{Table_PQ} \end{center} \end{table}
\begin{thm}\label{theo_PQ2a} The set $\mathcal{P}_2(\mathbb{K})$ is distributed into seven isotopism classes, whatever the base field is. \end{thm}
\begin{proof} A computational case study based on a similar reasoning to that exposed in Example \ref{ejemplo_PQ2a} enables us to ensure the result. If the characteristic of the base field is distinct from two, then the seven isotopism classes under consideration are those related to the next partial Latin squares of order $2$
$$\begin{array}{|c|c|} \hline \ & \ \\ \hline \ & \ \\ \hline \end{array} \hspace{0.5cm}
\begin{array}{|c|c|} \hline 1 & \ \\ \hline \ & \ \\ \hline \end{array} \hspace{0.5cm}
\begin{array}{|c|c|} \hline 1 & \ \\ \hline \ & 1 \\ \hline \end{array} \hspace{0.5cm}
\begin{array}{|c|c|} \hline 1 & \ \\ \hline 2 & \ \\ \hline \end{array} \hspace{0.5cm}
\begin{array}{|c|c|} \hline 1 & 2 \\ \hline \ & \ \\ \hline \end{array} \hspace{0.5cm}
\begin{array}{|c|c|} \hline 1 & 2 \\ \hline 2 & \ \\ \hline \end{array}\hspace{0.5cm}
\begin{array}{|c|c|} \hline 1 & 2 \\ \hline 2 & 1 \\ \hline \end{array}$$
Otherwise, if the characteristic of the base field is two, then the isotopism classes related to the last two partial Latin squares coincide. In this case, the next partial Latin square corresponds to the seventh isotopism class
$$\begin{array}{|c|c|} \hline 1 & \ \\ \hline \ & 2 \\ \hline \end{array}$$
If the characteristic of the base field is distinct from two, the partial quasigroup ring related to this partial Latin square is isotopic to that related to the unique Latin square of the previous list. \end{proof}
It is known \cite{Falcon2013} that there are $2$, $8$ and $81$ distinct isotopism classes of partial Latin squares of respective orders $1$ to $3$. In order to determine those distinct isotopism classes that give rise to isotopic partial quasigroup rings over the finite field $\mathbb{F}_2$, we have implemented the procedure {\em isoAlg} in our previously mentioned computer system. With a total run time of $761$ seconds, we have obtained that there exist $2$, $7$ and $72$ distinct isotopism classes of partial quasigroup rings of respective dimensions $1$ to $3$. Particularly, the existence of two classes for the one-dimensional case agrees with that exposed in Subsection 2.3. Besides, the seven isotopism classes for the two-dimensional case agrees with Theorem \ref{theo_PQ2a}. For the three-dimensional case, the next nine pairs of non-isotopic partial Latin squares give rise to isotopic partial quasigroup rings $${\scriptsize\begin{array}{ccccc}
\begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & \ & \ \\ \hline
\ & \ & \ \\ \hline \end{array} \text { and } \begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & 1 & \ \\ \hline \ & \ & \ \\ \hline \end{array} & , &
\begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & \ & \ \\ \hline
\ & \ & 1 \\ \hline \end{array} \text { and } \begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & 1 & \ \\ \hline
\ & \ & 1 \\ \hline \end{array} & , & \begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & \ & \ \\ \hline
\ & \ & 3 \\ \hline \end{array} \text { and } \begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & 1 & \ \\ \hline \ & \ & 3 \\ \hline \end{array}\\ \\
\begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline \ & 1 & \ \\ \hline 3 & \ & \ \\ \hline \end{array} \text { and }
\begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & 1 & \ \\ \hline 3 & \ & \ \\ \hline \end{array} &,&\begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline \ & 1 & 3 \\ \hline \ & \ & \ \\ \hline \end{array} \text { and }
\begin{array}{|c|c|c|} \hline 1 & 2 & 3 \\ \hline 2 & 1 & \ \\ \hline \ & \ & \ \\ \hline \end{array} &,&
\begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline
\ & 1 & \ \\ \hline 3 & \ & 2 \\ \hline \end{array} \text { and } \begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & 1 & \ \\ \hline 3 & \ & 1 \\ \hline \end{array}\\ \\
\begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & \ & 3 \\ \hline
\ & \ & 1 \\ \hline \end{array} \text { and } \begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & 1 & 3 \\ \hline \ & \ & 1 \\ \hline \end{array} &,&
\begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline
\ & 1 & 3 \\ \hline 3 & \ & \ \\ \hline \end{array} \text { and } \begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline 2 & 1 & 3 \\ \hline 3 & \ & \ \\ \hline \end{array}&,&
\begin{array}{|c|c|c|} \hline 1 & 2 & \ \\ \hline
\ & 1 & 3 \\ \hline 3 & \ & 2 \\ \hline \end{array} \text { and } \begin{array}{|c|c|c|} \hline 1 & 2 & 3 \\ \hline 2 & 1 & \ \\ \hline 3 & \ & 1 \\ \hline \end{array} \end{array}}$$
The run time of 761 seconds that are required to determine the mentioned distribution of partial quasigroup rings reduces to only 30 seconds in the same computer system if the invariants that we have just exposed in Table \ref{Table_PQ} and those exposed in Table \ref{Table_PQ3} are previously computed. The new run time includes indeed the extra 9 seconds of computation that is required for computing such invariants.
\begin{table}[ht] \resizebox{\textwidth}{!}{
\begin{tabular}{ccc|ccc|ccc} Partial Latin square & Vertices & Edges & Partial Latin square & Vertices & Edges & Partial Latin square & Vertices & Edges\\ \hline 100 000 000 & (4,4,1,16) & 48 & 100 010 002 & (7,7,3,34) & 120 & 031 302 & (7,7,7,42) & 126 \\ 120 000 000 & (4,6,3,24) & 72 & 120 001 002 & (7,7,3,36) & 108 & 120 210 301 & (7,7,7,42) & 126 \\ 123 000 000 & (4,7,7,28) & 84 & 120 200 002 & (7,7,3,36) & 108 & 120 213 001 & (7,7,7,42) & 126 \\ 100 200 000 & (6,4,3,24) & 72 & 120 200 001 & (7,7,3,38) & 114 & 120 213 300 & (7,7,7,42) & 126 \\ 100 010 000 & (6,6,1,24) & 72 & 120 210 001 & (7,7,3,38) & 114 & 120 001 312 & (7,7,7,43) & 129 \\ 100 020 000 & (6,6,3,28) & 84 & 120 201 010 & (7,7,3,40) & 120 & 120 201 302 & (7,7,7,43) & 129 \\ 120 200 000 & (6,6,3,32) & 96 & 120 201 012 & (7,7,3,40) & 120 & 120 231 300 & (7,7,7,43) & 129 \\ 120 210 000 & (6,6,3,32) & 96 & 100 020 003 & (7,7,7,37) & 111 & 123 231 312 & (7,7,7,43) & 129 \\ 120 000 300 & (6,6,6,32) & 96 & 120 002 003 & (7,7,7,38) & 114 & 120 003 312 & (7,7,7,44) & 132 \\ 120 000 310 & (6,6,6,36) & 108 & 120 002 300 & (7,7,7,38) & 114 & 120 013 301 & (7,7,7,44) & 132 \\ 120 001 000 & (6,7,3,32) & 96 & 120 003 300 & (7,7,7,38) & 114 & 120 013 302 & (7,7,7,44) & 132 \\ 120 012 000 & (6,7,3,36) & 108 & 120 001 300 & (7,7,7,39) & 117 & 120 200 312 & (7,7,7,44) & 132 \\ 120 003 000 & (6,7,7,34) & 102 & 120 200 003 & (7,7,7,40) & 120 & 120 203 301 & (7,7,7,44) & 132 \\ 120 000 302 & (6,7,7,36) & 108 & 120 200 302 & (7,7,7,40) & 120 & 123 210 301 & (7,7,7,44) & 132 \\ 123 200 000 & (6,7,7,36) & 108 & 120 210 003 & (7,7,7,40) & 120 & 123 031 310 & (7,7,7,45) & 135 \\ 120 013 000 & (6,7,7,38) & 114 & 123 010 001 & (7,7,7,40) & 120 & 123 200 312 & (7,7,7,45) & 135 \\ 123 210 000 & (6,7,7,38) & 114 & 123 200 300 & (7,7,7,40) & 120 & 123 230 310 & (7,7,7,45) & 135 \\ 123 230 000 & (6,7,7,40) & 120 & 120 001 302 & (7,7,7,41) & 123 & 123 012 230 & (7,7,7,46) & 138 \\ 123 231 000 & (6,7,7,40) & 120 & 120 001 310 & (7,7,7,41) & 123 & 123 210 031 & (7,7,7,46) & 138 \\ 100 200 300 & (7,4,7,28) & 84 & 120 201 300 & (7,7,7,41) & 123 & 123 201 312 & (7,7,7,46) & 138 \\ 100 200 010 & (7,6,3,32) & 96 & 123 200 010 & (7,7,7,41) & 123 \\ 120 200 010 & (7,6,3,36) & 108 & 120 003 310 & (7,7,7,42) & 126 \\ 100 200 030 & (7,6,7,34) & 102 & 120 010 301 & (7,7,7,42) & 126 \\ 120 030 300 & (7,6,7,36) & 108 & 120 010 302 & (7,7,7,42) & 126 \\ 120 200 300 & (7,6,7,36) & 108 & 120 012 300 & (7,7,7,42) & 126 \\ 120 010 300 & (7,6,7,38) & 114 & 120 013 300 & (7,7,7,42) & 126 \\ 120 210 300 & (7,6,7,38) & 114 & 120 200 013 & (7,7,7,42) & 126 \\ 120 230 300 & (7,6,7,40) & 120 & 120 203 001 & (7,7,7,42) & 126 \\ 120 230 310 & (7,6,7,40) & 120 & 120 203 300 & (7,7,7,42) & 126 \\ 100 010 001 & (7,7,1,28) & 84 & 123 010 300 & (7,7,7,42) & 126 \\ \end{tabular}} \caption{Invariants of the graph $G_1$ related to non-abelian partial algebras in $\mathcal{P}_3(\mathbb{F}_2)$.} \label{Table_PQ3} \end{table}
\section{Conclusion and further studies}
We have described in this paper a pair of graphs that enable us to define faithful functors between finite-dimensional algebras over finite fields and these graphs. The computation of related graph invariants plays a remarkable role in the distribution of distinct families of algebras into isotopism and isomorphism classes. Some preliminary results have been exposed in this regard, particularly on the distribution of partial quasigroup rings over finite fields. Based on the known classification of partial Latin squares into isotopism classes, further work is required to determine completely this distribution.
\end{document}
|
arXiv
|
{
"id": "1609.01061.tex",
"language_detection_score": 0.6584783792495728,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[Loose Hamiltonian cycles forced by $(k-2)$-degree]{Loose Hamiltonian cycles forced by large $(k-2)$-degree \\ -- sharp version --}
\author[J.~de~O.~Bastos]{Josefran de Oliveira Bastos} \author[G.~O.~Mota]{Guilherme Oliveira Mota} \address{Instituto de Matem\'atica e Estat\'{\i}stica, Universidade de
S\~ao Paulo, S\~ao Paulo, Brazil}
\email{\{josefran|mota\}@ime.usp.br}
\author[M.~Schacht]{Mathias Schacht} \author[J.~Schnitzer]{Jakob Schnitzer} \author[F.~Schulenburg]{Fabian Schulenburg} \address{Fachbereich Mathematik, Universit\"at Hamburg, Hamburg, Germany} \email{[email protected]}
\email{\{jakob.schnitzer|fabian.schulenburg\}@uni-hamburg.de}
\thanks{The first author was supported by CAPES\@. The second author was supported by FAPESP (Proc. 2013/11431-2 and 2013/20733-2) and CNPq (Proc. 477203/2012-4 and {456792/2014-7}). The cooperation was supported by a joint CAPES/DAAD PROBRAL (Proc. 430/15).}
\begin{abstract} We prove for all $k\geq 4$ and $1\leq\ifmmode\ell\else\polishlcross\fi<k/2$ the sharp minimum $(k-2)$-degree bound for a $k$-uniform hypergraph~$\cch$ on~$n$ vertices to contain a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle if $k-\ifmmode\ell\else\polishlcross\fi$ divides~$n$ and~$n$ is sufficiently large. This extends a result of Han and Zhao for $3$-uniform hypegraphs. \end{abstract}
\keywords{hypergraphs, Hamiltonian cycles, degree conditions} \subjclass[2010]{05C65 (primary), 05C45 (secondary)}
\maketitle
\@ifstar{\origsection*}{\@startsection{section}{1}\z@{.7\linespacing\@plus\linespacing}{.5\linespacing}{\normalfont\scshape\centering\S}}{Introduction}
Given $k\geq 2$, a $k$-uniform hypergraph $\cch$ is a pair $(V,E)$ with vertex set $V$ and edge set $E\subseteq V^{(k)}$, where $V^{(k)}$ denotes the set of all $k$-element subsets of $V$. Given a $k$-uniform hypergraph $\cch=(V,E)$ and a subset $S \in V^{(s)}$, we denote by $d(S)$ the number of edges in $E$ containing~$S$ and we denote by $N(S)$ the $(k-s)$-element sets $T\in V^{(k-s)}$ such that $T\dcup S\in E$, so $d(S)=|N(S)|$. The \emph{minimum $s$-degree} of $\cch$ is denoted by $\delta_s(\cch)$ and it is defined as the minimum of $d(S)$ over all sets $S\in V^{(s)}$. We denote by the \textit{size} of a hypergraph the number of its edges.
We say that a $k$-uniform hypergraph $\ccc$ is an \emph{$\ifmmode\ell\else\polishlcross\fi$-cycle} if there exists a cyclic ordering of its vertices such that every edge of $\ccc$ is composed of $k$ consecutive vertices, two (vertex-wise) consecutive edges share exactly $\ifmmode\ell\else\polishlcross\fi$ vertices, and every vertex is contained in an edge. Moreover, if the ordering is not cyclic, then $\ccc$ is an \emph{$\ifmmode\ell\else\polishlcross\fi$-path} and we say that the first and last~$\ifmmode\ell\else\polishlcross\fi$ vertices are the ends of the path. The problem of finding minimum degree conditions that ensure the existence of Hamiltonian cycles, i.e.\ cycles that contain all vertices of a given hypergraph, has been extensively studied over the last years (see, e.g., the surveys~\cites{RRsurv,Zhao-survey}). Katona and Kierstead~\cite{KaKi99} started the study of this problem, posing a conjecture that was confirmed by R\"odl, Ruci\'nski, and Szemer\'edi~\cites{RoRuSz06,RoRuSz08}, who proved the following result: For every $k\geq 3$, if $\cch$ is a $k$-uniform $n$-vertex hypergraph with $\delta_{k-1}(\cch)\geq {(1/2+o(1))}n$, then $\cch$ contains a Hamiltonian $(k-1)$-cycle. K\"uhn and Osthus proved that $3$-uniform hypergraphs~$\cch$ with $\delta_2(\cch)\geq {(1/4 +o(1))}n$ contain a Hamiltonian $1$-cycle~\cite{KuOs06}, and H\`an and Schacht~\cite{HaSc10} (see also~\cite{KeKuMyOs11}) generalized this result to arbitrary $k$ and $\ifmmode\ell\else\polishlcross\fi$-cycles with $1\leq \ifmmode\ell\else\polishlcross\fi <k/2$. In~\cite{KuMyOs10}, K\"uhn, Mycroft, and Osthus generalized this result to $1\leq \ifmmode\ell\else\polishlcross\fi<k$, settling the problem of the existence of Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycles in $k$-uniform hypergraphs with large minimum $(k-1)$-degree. In Theorem~\ref{theorem:asymp} below (see~\cites{BuHaSc13,BaMoScScSc16+}) we have minimum $(k-2)$-degree conditions that ensure the existence of Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycles for $1\leq \ifmmode\ell\else\polishlcross\fi<k/2$.
\begin{theorem}\label{theorem:asymp}
For all integers $k\geq 3$ and $1\leq \ifmmode\ell\else\polishlcross\fi<k/2$ and every $\gamma>0$ there exists an $n_0$ such that every $k$-uniform hypergraph $\cch=(V,E)$ on $|V|=n\geq n_0$ vertices with $n\in(k-\ifmmode\ell\else\polishlcross\fi)\bbn$ and \begin{equation*}
\delta_{k-2}(\cch)\geq\left(\frac{4(k-\ifmmode\ell\else\polishlcross\fi)-1}{4{(k-\ifmmode\ell\else\polishlcross\fi)}^2}+\gamma\right)\binom{n}{2} \end{equation*} contains a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle. \qed \end{theorem}
The minimum degree condition in Theorem~\ref{theorem:asymp} is asymptotically optimal as the following well-known example confirms. The construction of the example varies slightly depending on whether~$n$ is an odd or an even multiple of~$k-\ifmmode\ell\else\polishlcross\fi$. We first consider the case that $n = (2m + 1)(k-\ifmmode\ell\else\polishlcross\fi)$ for some integer~$m$. Let $\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)=(V,E)$ be a $k$-uniform hypergraph on $n$ vertices such that an edge belongs to $E$ if and only if it contains at least one vertex from $A \subset V$, where $|A|=\left\lfloor \frac{n}{2(k-\ifmmode\ell\else\polishlcross\fi)} \right\rfloor$. It is easy to see that $\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)$ contains no Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle, as it would have to contain $\frac{n}{k-\ifmmode\ell\else\polishlcross\fi}$ edges and each vertex in~$A$ is contained in at most two of them. Indeed any maximal $\ifmmode\ell\else\polishlcross\fi$-cycle includes all but $k-\ifmmode\ell\else\polishlcross\fi$ vertices and adding any additional edge to the hypergraph would imply a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle. Let us now consider the case that $n= 2m(k-\ifmmode\ell\else\polishlcross\fi)$ for some integer~$m$. Similarly, let $\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)=(V,E)$ be a $k$-uniform hypergraph on $n$ vertices that contains all edges incident to $A \subset V$, where $|A|=\frac{n}{2(k-\ifmmode\ell\else\polishlcross\fi)}-1$. Additionally, fix some $\ell+1$ vertices of $B = V\smallsetminus A$ and let $\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)$ contain all edges on $B$ that contain all of these vertices, i.e., an $(\ifmmode\ell\else\polishlcross\fi+1)$-star. Again, of the $\frac{n}{k-\ifmmode\ell\else\polishlcross\fi}$ edges that a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle would have to contain, at most $\frac{n}{k-\ifmmode\ell\else\polishlcross\fi} - 2$ can be incident to $A$. So two edges would have to be completely contained in $B$ and be disjoint or intersect in exactly $\ifmmode\ell\else\polishlcross\fi$ vertices, which is impossible since the induced subhypergraph on $B$ only contains an $(\ifmmode\ell\else\polishlcross\fi+1)$-star. Note that for the minimum $(k-2)$-degree the $(\ifmmode\ell\else\polishlcross\fi+1)$-star on $B$ is only relevant if $\ifmmode\ell\else\polishlcross\fi=1$, in which case this star increases the minimum $(k-2)$-degree by one.
In~\cite{HaZh15b}, Han and Zhao proved the exact version of Theorem~\ref{theorem:asymp} when $k=3$, i.e., they obtained a sharp bound for $\delta_{1}(\cch)$. We extend this result to $k$-uniform hypergraphs.
\begin{theorem}[Main Result]\label{theorem:main}
For all integers $k\geq 4$ and $1\leq \ifmmode\ell\else\polishlcross\fi<k/2$ there exists $n_0$ such that every $k$-uniform hypergraph $\cch=(V,E)$ on $|V|=n\geq n_0$ vertices with $n\in(k-\ifmmode\ell\else\polishlcross\fi)\bbn$ and \begin{equation}\label{eq:sharp_minimum_degree}
\delta_{k-2}(\cch)
>
\delta_{k-2}(\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)) \end{equation} contains a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle. In particular, if \begin{equation*}
\delta_{k-2}(\cch)
\geq
\frac{4(k-\ifmmode\ell\else\polishlcross\fi)-1}{4{(k-\ifmmode\ell\else\polishlcross\fi)}^2} \binom{n}{2}, \end{equation*} then $\cch$ contains a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle. \end{theorem}
The following notion of extremality is motivated by the hypergraph $\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)$. A $k$-uniform hypergraph $\cch=(V,E)$ is called \emph{$(\ifmmode\ell\else\polishlcross\fi,\xi)$-extremal} if there exists a partition $V=A\dcup B$ such that $|A|=\left\lceil \frac{n}{2(k-\ell)} - 1 \right\rceil$, $|B|=\left\lfloor \frac{2(k-\ifmmode\ell\else\polishlcross\fi)-1}{2(k-\ifmmode\ell\else\polishlcross\fi)}n + 1 \right\rfloor$ and $e(B)=|E\cap B^{(k)}|\leq \xi \binom{n}{k}$. We say that $A\dcup B$ is an \emph{$(\ifmmode\ell\else\polishlcross\fi,\xi)$-extremal partition} of $V$. Theorem~\ref{theorem:main} follows easily from the next two results, the so-called \emph{extremal case} (see Theorem~\ref{theorem:extremal} below) and the \emph{non-extremal case} (see Theorem~\ref{theorem:non-extremal}).
\begin{theorem}[Non-extremal Case]\label{theorem:non-extremal} For any $0<\xi<1$ and all integers $k\geq 4$ and $1\leq \ifmmode\ell\else\polishlcross\fi<k/2$, there exists $\gamma>0$ such that the following holds for sufficiently large $n$. Suppose $\cch$ is a $k$-uniform hypergraph on $n$ vertices with $n\in(k-\ifmmode\ell\else\polishlcross\fi)\bbn$ such that $\cch$ is not $(\ifmmode\ell\else\polishlcross\fi,\xi)$-extremal and \begin{equation*}
\delta_{k-2}(\cch)\geq\left(\frac{4(k-\ifmmode\ell\else\polishlcross\fi)-1}{4{(k-\ifmmode\ell\else\polishlcross\fi)}^2}-\gamma\right)\binom{n}{2}. \end{equation*} Then $\cch$ contains a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle. \qed \end{theorem} The non-extremal case was the main result of~\cite{BaMoScScSc16+}. \begin{theorem}[Extremal Case]\label{theorem:extremal} For any integers $k\geq 3$ and $1\leq \ifmmode\ell\else\polishlcross\fi<k/2$, there exists $\xi>0$ such that the following holds for sufficiently large $n$. Suppose $\cch$ is a $k$-uniform hypergraph on $n$ vertices with $n\in(k-\ifmmode\ell\else\polishlcross\fi)\bbn$ such that $\cch$ is $(\ifmmode\ell\else\polishlcross\fi,\xi)$-extremal and \begin{equation*}
\delta_{k-2}(\cch)
>
\delta_{k-2}(\ccx_{k,\ifmmode\ell\else\polishlcross\fi}). \end{equation*} Then $\cch$ contains a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle. \end{theorem}
In Section~\ref{sec:overview} we give an overview of the proof of Theorem~\ref{theorem:extremal} and state Lemma~\ref{lem:mainlemma}, the main result required for the proof. In Section~\ref{sec:mainproof} we first prove some auxiliary lemmas and then we prove Lemma~\ref{lem:mainlemma}.
\@ifstar{\origsection*}{\@startsection{section}{1}\z@{.7\linespacing\@plus\linespacing}{.5\linespacing}{\normalfont\scshape\centering\S}}{Overview}\label{sec:overview}
Let $\cch=(V,E)$ be a $k$-uniform hypergraph and let $X,Y\subset V$ be disjoint subsets. Given a vertex set $L\subset V$ we denote by $d(L,X^{(i)} Y^{(j)})$ the number of edges of the form $L \cup I \cup J$, where $I \in X^{(i)}$, $J \in Y^{(j)}$, and $|L| + i + j = k$. We allow for $Y^{(j)}$ to be omitted when $j$ is zero and write $d(v,X^{(i)} Y^{(j)})$ for $d(\{ v \},X^{(i)} Y^{(j)})$.
The proof of Theorem~\ref{theorem:extremal} follows ideas from~\cite{HaZh15}, where a corresponding result with a $(k-1)$-degree condition is proved. Let $\cch=(V,E)$ be an extremal hypergraph satisfying~\eqref{eq:sharp_minimum_degree}. We first construct an $\ifmmode\ell\else\polishlcross\fi$-path~$\ccq$ in $\cch$ (see~Lemma~\ref{lem:mainlemma} below) with ends $L_0$ and $L_1$ such that there is a partition $A_{\ast}\dcup B_{\ast}$ of $(V\smallsetminus \ccq) \cup L_0 \cup L_1$ composed only of ``typical'' vertices (see~\ref{it:2-mainlemma} and~\ref{it:3-mainlemma} below). The set $A_{\ast}\cup B_{\ast}$ is suitable for an application of Lemma~\ref{lem:3.10} below, which ensures the existence of an $\ifmmode\ell\else\polishlcross\fi$-path $\ccq'$ on $A_{\ast}\cup B_{\ast}$ with $L_0$ and $L_1$ as ends. Note that the existence of a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-cycle in $\cch$ is guaranteed by $\ccq$ and $\ccq'$. So, in order to prove Theorem~\ref{theorem:extremal}, we only need to prove the following lemma.
\begin{lemma}[Main Lemma]\label{lem:mainlemma}
For any $\varrho > 0$ and all integers $k\geq 3$ and $1\leq \ifmmode\ell\else\polishlcross\fi <k/2$, there exists a positive~$\xi$ such that the following holds for sufficiently large~$n\in (k-\ifmmode\ell\else\polishlcross\fi)\bbn$.
Suppose that~$\cch=(V,E)$ is an $(\ifmmode\ell\else\polishlcross\fi,\xi)$-extremal $k$-uniform hypergraph on~$n$ vertices and
\[
\delta_{k-2}(\cch)
>
\delta_{k-2}(\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)).
\]
Then there exists a non-empty $\ifmmode\ell\else\polishlcross\fi$-path $\ccq$ in $\cch$ with ends $L_0$ and $L_1$ and a partition $A_{\ast}\dcup B_{\ast}=(V\smallsetminus \ccq) \cup L_0 \cup L_1$ where $L_0, L_1\subset B_{\ast}$ such that the following hold:
\begin{enumerate}[label=\rmlabel]
\item\label{it:1-mainlemma} $|B_{\ast}|=(2k-2\ifmmode\ell\else\polishlcross\fi-1)|A_{\ast}|+\ifmmode\ell\else\polishlcross\fi$,
\item\label{it:2-mainlemma} $d(v,B_{\ast}^{(k-1)})\geq (1-\varrho) \binom{|B_{\ast}|}{k-1}$ for any vertex $v\in A_{\ast}$,
\item\label{it:3-mainlemma} $d(v,A_{\ast}^{(1)}B_{\ast}^{(k-2)})\geq (1-\varrho) |A_{\ast}| \binom{|B_{\ast}|}{k-2}$ for any vertex $v\in B_{\ast}$,
\item\label{it:4-mainlemma} $d(L_0,A_{\ast}^{(1)} B_{\ast}^{(k - \ifmmode\ell\else\polishlcross\fi - 1)}), d(L_1,A_{\ast}^{(1)}B_{\ast}^{(k - \ifmmode\ell\else\polishlcross\fi - 1)}) \geq (1-\varrho)|A_{\ast}|\binom{|B_{\ast}|}{k - \ifmmode\ell\else\polishlcross\fi - 1}$.
\end{enumerate} \end{lemma}
The next result, which we will use to conclude the proof of Theorem~\ref{theorem:extremal}, was obtained by Han and Zhao (see~\cite{HaZh15}*{Lemma~3.10}).
\begin{lemma}\label{lem:3.10}
For any integers $k\geq 3$ and $1\leq \ifmmode\ell\else\polishlcross\fi < k/2$ there exists $\varrho > 0$ such that the following holds.
If $\cch$ is a sufficiently large $k$-uniform hypergraph with a partition $V(\cch)=A_{\ast}\dcup B_{\ast}$ and there exist two disjoint $\ifmmode\ell\else\polishlcross\fi$-sets $L_0,L_1\subset B_{\ast}$ such that~\ref{it:1-mainlemma}--\ref{it:4-mainlemma} hold, then $\cch$ contains a Hamiltonian $\ifmmode\ell\else\polishlcross\fi$-path $\ccq'$ with $L_0$ and $L_1$ as ends.
\qed \end{lemma}
\@ifstar{\origsection*}{\@startsection{section}{1}\z@{.7\linespacing\@plus\linespacing}{.5\linespacing}{\normalfont\scshape\centering\S}}{Proof of the Main Lemma}\label{sec:mainproof}
We will start this section by describing the setup for the proof, which will be fixed for the rest of the paper. Then we will prove some auxiliary lemmas and finally prove Lemma~\ref{lem:mainlemma}. Let $\varrho > 0$ and integers $k\geq 3$ and $1\leq \ifmmode\ell\else\polishlcross\fi<k/2$ be given. Fix constants \[
\frac{1}{k}, \frac{1}{\ifmmode\ell\else\polishlcross\fi}, \varrho \gg \delta \gg \varepsilon \gg \eps' \gg \vartheta \gg \xi. \] Let $n\in (k-\ifmmode\ell\else\polishlcross\fi)\bbn$ be sufficiently large and let $\cch$ be an $(\ifmmode\ell\else\polishlcross\fi,\xi)$-extremal $k$-uniform hypergraph on $n$ vertices that satisfies the $(k-2)$-degree condition \begin{equation*}
\delta_{k-2}(\cch)
>
\delta_{k-2}(\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)). \end{equation*} Let $A \dcup B=V(\cch)$ be a minimal extremal partition of~$V(\cch)$, i.e.\ a partition satisfying \begin{equation}
\label{eq:partition-sizes}
a = |A| = \left\lceil \frac{n}{2(k-\ifmmode\ell\else\polishlcross\fi)} \right\rceil - 1, \quad b = |B| = n - a, \qand e(B) \leq \xi \binom{n}{k}, \end{equation} which minimises $e(B)$. Recall that the extremal example $\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)$ implies \begin{equation}
\label{eq:minimumDegree}
\delta_{k-2}(\cch)
>
\binom{a}{2} + a(b - k + 2). \end{equation} Since $e(B)\leq \xi\binom{n}{k}$, we expect most vertices $v\in B$ to have low degree $d(v,B^{(k-1)})$ into~$B$. Also, most $v\in A$ must have high degree $d(v,B^{(k-1)})$ into $B$ such that the degree condition for $(k-2)$-sets in~$B$ can be satisfied. Thus, we define the sets $A_{\eps}$ and $B_{\eps}$ to consist of vertices of high respectively low degree into~$B$ by \begin{align*}
A_{\eps} &= \left\lbrace v\in V\colon d(v,B^{(k-1)})\geq (1-\varepsilon)\binom{|B|}{k-1}\right\rbrace,\\
B_{\eps} &=\left\lbrace v\in V\colon d(v,B^{(k-1)})\leq \varepsilon\binom{|B|}{k-1}\right\rbrace, \end{align*}
and set $V_{\eps}=V\smallsetminus (A_{\eps} \cup B_{\eps} )$. We will write $a_{\eps} = |A_{\eps}|$, $b_{\eps} = |B_{\eps}|$, and $v_{\eps} = |V_{\eps}|$. It follows from these definitions that \begin{equation}\label{eq:Aeps-Beps-inclusion}
\text{if } A\cap B_{\eps} \neq \varnothing,
\quad \text{then} \quad
B \subset B_{\eps} ,
\quad \text{while otherwise} \quad
A \subset A_{\eps}. \end{equation} For the first inclusion, consider a vertex $v \in A \cap B_{\eps}$ and a vertex $w \in B \smallsetminus B_{\eps}$. Exchanging~$v$ and~$w$ would create a minimal partition with fewer edges in $e(B)$, a contradiction to the minimality of the extremal partition. The other inclusion is similarly implied by the minimality.
Actually, as we shall show below, the sets $A_{\eps}$ and $B_{\eps}$ are not too different from $A$ and $B$ respectively: \begin{equation}\label{eq:Aeps-Beps-sizes}
|A\smallsetminus A_{\eps} |, |B\smallsetminus B_{\eps} |, |A_{\eps} \smallsetminus A|, |B_{\eps} \smallsetminus B|\leq \vartheta b \qand |V_{\eps}|\leq 2\vartheta b. \end{equation} Note that by the minimum $(k-2)$-degree \[
\binom{a}{2}\binom{b}{k-2}+a\binom{b}{k-1}(k-1)
<
\binom{b}{k-2}\delta_{k-2}(\cch)
\leq
\sum_{S\in B^{(k-2)}} d(S). \]
Every vertex $v \in |A\smallsetminus A_{\eps} |$ satisfies $d(v, B^{(k-1)}) < (1 - \varepsilon)\binom{b}{k-1}$, so we have \begin{align*}
\sum_{S\in B^{(k-2)}} d(S)
\leq&
\binom{a}{2}\binom{b}{k-2}+a\binom{b}{k-1}(k-1) \\
&+ e(B)\binom{k}{2} - |A \smallsetminus A_{\eps}| \varepsilon \binom{b}{k-1}(k-1). \end{align*}
Consequently $|A\smallsetminus A_{\eps} |\leq \vartheta b$, as $e(B) < \xi \binom{n}{k}$ and $\xi \ll \vartheta, \varepsilon$.
Moreover, $|B\smallsetminus B_{\eps} | \leq \vartheta b$ holds as a high number of vertices in $B \smallsetminus B_{\eps} $ would contradict $e(B) < \xi \binom{b}{k}$. The other three inequalities~\eqref{eq:Aeps-Beps-sizes} follow from the already shown ones, for example for $|A_{\eps} \smallsetminus A| < \vartheta b$ observe that \[
A_{\eps} \smallsetminus A = A_{\eps} \cap B \subset B \smallsetminus B_{\eps}. \] Although the vertices in $B_{\eps}$ were defined by their low degree into $B$, they also have low degree into the set $B_{\eps}$ itself; for any $v \in B_{\eps}$ we get \begin{align*}
d(v, B_{\eps}^{(k - 1)})
&\leq
d(v, B^{(k - 1)}) + |B_{\eps}\smallsetminus B| \binom{|B_{\eps}| - 1}{k - 2}\\
&\leq
\varepsilon \binom{b}{k -1 } + \vartheta b |B_{\eps}|^{k -1}\\
&<
2 \varepsilon \binom{|B_{\eps}|}{k - 1}. \end{align*}
Since we are interested in $\ifmmode\ell\else\polishlcross\fi$-paths, the degree of $\ifmmode\ell\else\polishlcross\fi$-tuples in $B_{\eps}$ will be of interest, which motivates the following definition. An $\ifmmode\ell\else\polishlcross\fi$-set $L\subset B_{\eps}$ is called $\varepsilon$-\emph{typical} if \[
d(L,B^{(k - \ifmmode\ell\else\polishlcross\fi)})\leq \varepsilon\binom{|B|}{k-\ifmmode\ell\else\polishlcross\fi}. \] If $L$ is not $\varepsilon$-typical, then it is called $\varepsilon$-\emph{atypical}. Indeed, most $\ifmmode\ell\else\polishlcross\fi$-sets in $B_{\eps}$ are $\varepsilon$-typical; denote by $x$ the number of $\varepsilon$-atypical sets in~$B_{\eps}$. We have \begin{equation}\label{eq:num-typical-sets}
\frac{x\cdot\varepsilon \binom{|B|}{k-\ifmmode\ell\else\polishlcross\fi}}{\binom{k}{\ifmmode\ell\else\polishlcross\fi}} \leq e(B \cup B_{\eps}) \leq \xi \binom{n}{k} + \vartheta {|B|}^k,
\quad \text{implying} \quad
x\leq \eps' \binom{|B_{\eps}|}{\ifmmode\ell\else\polishlcross\fi}. \end{equation}
\begin{lemma}\label{lem:typical-degree}
The following holds for any $B_{\eps}^{(m)}$-set $M$ if $m \leq k-2$.
\[
d(M,A_{\eps}^{(1)} B_{\eps}^{(k - m - 1)}) + \frac{k-m}{2} d(M, B_{\eps}^{(k - m)})
\geq
\left(1-\delta\right)|A_{\eps}|\binom{|B_{\eps}| - m}{k-m-1}.
\]
In particular, the following holds for any $\varepsilon$-typical $B^{(\ifmmode\ell\else\polishlcross\fi)}$-set $L$.
\[
d(L,A_{\eps}^{(1)} B_{\eps}^{(k - \ifmmode\ell\else\polishlcross\fi - 1)})\geq (1-2\delta)|A_{\eps}|\binom{|B_{\eps}|-\ifmmode\ell\else\polishlcross\fi}{k-\ifmmode\ell\else\polishlcross\fi-1}.
\] \end{lemma}
In the proof of the main lemma we will connect two $\varepsilon$-typical sets only using vertices that are unused so far. Even more, we want to connect two $\varepsilon$-typical sets using exactly one vertex from $A$. The following corollary of Lemma~\ref{lem:typical-degree} allows us to do this.
\begin{corollary}\label{corr:connect-extend-typical}
Let $L$ and $L'$ be two disjoint $\varepsilon$-typical sets in $B_{\eps}$ and $U\subset V$ with~$|U| \leq \varepsilon n$.
Then the following holds.
\begin{enumerate}[label=\alabel]
\item\label{it:connect-typical} There exists an $\ifmmode\ell\else\polishlcross\fi$-path disjoint from $U$ of size two with ends $L$ and $L'$ that contains exactly one vertex from $ A_{\eps}$.
\item\label{it:extend-typical} There exist $a \in A_{\eps} \smallsetminus U$ and a set $(k - \ifmmode\ell\else\polishlcross\fi -1)$-set $C \subset B_{\eps} \smallsetminus U$ such that $L \cup a \cup C$ is an edge in $\cch$ and every $\ifmmode\ell\else\polishlcross\fi$-subset of $C$ is $\varepsilon$-typical.
\end{enumerate} \end{corollary} \begin{proof}[Proof of Corollary~\ref{corr:connect-extend-typical}]
For~\ref{it:connect-typical}, the second part of Lemma~\ref{lem:typical-degree} for $L$ and $L'$ implies that they both extend to an edge with at least $(1 - 2\delta)|A_{\eps}|\binom{|B_{\eps}| - \ifmmode\ell\else\polishlcross\fi}{k - \ifmmode\ell\else\polishlcross\fi - 1}$ sets in $A_{\eps}^{(1)}B_{\eps}^{(k - \ifmmode\ell\else\polishlcross\fi - 1)}$.
Only few of those intersect $U$ and by an averaging argument we obtain two sets $C, C' \in A_{\eps}^{(1)}B_{\eps}^{(k - \ifmmode\ell\else\polishlcross\fi - 1)}$ such that $|C \cap C'| = \ifmmode\ell\else\polishlcross\fi$ and $L \cup C$ as well as $L' \cup C'$ are edges in $\cch$, which yields the required $\ifmmode\ell\else\polishlcross\fi$-path.
In view of~\eqref{eq:num-typical-sets},~\ref{it:extend-typical} is a trivial consequence of the second part of Lemma~\ref{lem:typical-degree}. \end{proof} \begin{proof}[Proof of Lemma~\ref{lem:typical-degree}]
Let $m \leq k-2$ and let $M \in B_{\eps}^{(m)}$ be an $m$-set.
We will make use of the following sum over all $(k-2)$-sets $D \subset B_{\eps}$ that contain $M$.
\begin{equation}\label{eq:sumdeg1}
\begin{split}
\sum_{\substack{M \subset D \subset B_{\eps}\\|D| = k - 2}} d(D)
=
\sum_{\substack{M \subset D \subset B_{\eps} \\ |D| = k - 2}} \Big(& d(D, A_{\eps}^{(1)} B_{\eps}^{(1)}) + d(D, {(A_{\eps} \cup V_{\eps})}^{(2)}) \\
& \qquad + d(D,B_{\eps}^{(2)}) + d(D, V_{\eps}^{(1)} B_{\eps}^{(1)})\Big)
\end{split}
\end{equation}
Note that we can relate the sums $\sum d(D, A_{\eps}^{(1)}B_{\eps}^{(1)})$ and $\sum d(D,B_{\eps}^{(2)})$ in~\eqref{eq:sumdeg1} to the terms in question as follows.
\begin{equation}\label{eq:sumdeg2}
\begin{split}
d(M, A_{\eps}^{(1)} B_{\eps}^{(k - m - 1)})
&=
\frac{1}{k - m - 1}\sum_{\substack{M \subset D \subset B_{\eps}\\|D| = k - 2}} d(D, A_{\eps}^{(1)} B_{\eps}^{(1)}),
\\
d(M, B_{\eps}^{(k - m)})
&=
\frac{1}{\binom{k-m}{2}}\sum_{\substack{M \subset D \subset B_{\eps}\\|D| = k - 2}} d(D, B_{\eps}^{(2)}).
\end{split}
\end{equation}
We will bound some of the terms on the right-hand side of~\eqref{eq:sumdeg1}.
It directly follows from~\eqref{eq:Aeps-Beps-sizes} that $d(D, {(A_{\eps}\cup V_{\eps})}^{(2)})\leq \binom{a+3\vartheta b}{2}$; moreover, $d(D, V_{\eps}^{(1)} B_{\eps}^{(1)}) \leq 2 \vartheta bb_{\eps}$.
Using the minimum $(k-2)$-degree condition~\eqref{eq:minimumDegree} we obtain
\begin{equation*}
\sum_{\substack{M \subset D \subset B_{\eps}\\|D| = k - 2}} d(D)
>
\binom{b_{\eps} - m}{k - m - 2}\left(\binom{a}{2} + a(b - k + 2)\right).
\end{equation*}
Combining these estimates with~\eqref{eq:sumdeg1}~and~\eqref{eq:sumdeg2} yields
\begin{align*}
& d(M, A_{\eps}^{(1)} B_{\eps}^{(k - m - 1)})
+ \frac{k-m}{2} d(M, B_{\eps}^{(k - m)})
\\
& \quad \geq
\frac{1}{k - m - 1}\binom{b_{\eps} - m}{k - m - 2}
\left(\binom{a}{2} + a(b-k+2) - \binom{a+3\vartheta b}{2} - 2\vartheta bb_{\eps} \right)
\\
& \quad \geq
\left(1-\delta\right)a_{\eps}\binom{b_{\eps} - m}{k - m - 1}.
\end{align*}
For the second part of the lemma, note that the definition of $\varepsilon$-typicality and $\varepsilon \ll \delta$ imply that $\frac{k-\ifmmode\ell\else\polishlcross\fi}{2} d(L, B_{\eps}^{(k - \ifmmode\ell\else\polishlcross\fi)})$ is smaller than $\delta a_{\eps} \binom{b_{\eps} - \ifmmode\ell\else\polishlcross\fi}{k - \ifmmode\ell\else\polishlcross\fi - 1}$ for any $\varepsilon$-typical $\ifmmode\ell\else\polishlcross\fi$-set $L$, which concludes the proof. \end{proof}
For Lemma~\ref{lem:mainlemma}, we want to construct an $\ifmmode\ell\else\polishlcross\fi$-path $\ccq$, such that $V_{\eps} \subset V(\ccq)$ and the remaining sets $A_{\eps}\smallsetminus \ccq$ and $B_{\eps}\smallsetminus \ccq$ have the right relative proportion of vertices, i.e., their sizes are in a ratio of one to $(2k - 2\ifmmode\ell\else\polishlcross\fi - 1)$. If $|A \cap B_{\eps}| > 0$, then $B \subset B_{\eps}$ (see~\eqref{eq:Aeps-Beps-inclusion}) and so $\ccq$ should cover $V_{\eps}$ and contain the right number of vertices from $B_{\eps}$. For this, we have to find suitable edges inside $B_{\eps}$, which the following lemma ensures.
\begin{lemma}\label{lem:2q-path}
Suppose that $q = |A \cap B_{\eps}| > 0$.
Then there exist $2q + 2$ disjoint paths of size three, each of which contains exactly one vertex from $A_{\eps}$ and has two $\varepsilon$-typical sets as its ends. \end{lemma} \begin{proof}
We say that an $(\ifmmode\ell\else\polishlcross\fi - 1)$-set $M \subset B_{\eps}$ is $\emph{good}$ if it is a subset of at least $(1 - \sqrt{\eps' })b_{\eps}$ $\varepsilon$-typical sets, otherwise we say that the set is \emph{bad}.
We will first show that there are $2q+2$ edges in $B_{\eps}$, each containing one $\varepsilon$-typical and one \emph{good} $(\ifmmode\ell\else\polishlcross\fi - 1)$-set.
Then we will connect pairs of these edges to $\ifmmode\ell\else\polishlcross\fi$-paths of size three.
Suppose that $q = |A \cap B_{\eps}| > 0$.
So $B \subset B_{\eps}$ by~\eqref{eq:Aeps-Beps-inclusion} and consequently $|B_{\eps}| = |B| + q$ and $q \leq \vartheta |B|$.
It is not hard to see from~\eqref{eq:num-typical-sets} that at most a $\sqrt{\eps'}$ fraction of the $(\ifmmode\ell\else\polishlcross\fi - 1)$-sets in~$B_{\eps}^{(l-1)}$ are bad.
Hence, at least
\[
\left( 1 - \binom{k - 2}{\ifmmode\ell\else\polishlcross\fi}\eps' - \binom{k - 2}{\ifmmode\ell\else\polishlcross\fi - 1}\sqrt{\eps'} \right) \binom{b}{k-2}
\]
$(k -2)$-sets in $B_{\eps}$ contain no $\varepsilon$-atypical or bad subset.
Let $\ccb \subset B_{\eps}^{(k)}$ be the set of edges inside $B_{\eps}$ that contain such a $(k-2)$-set.
For all $M \in B_{\eps}^{(k - 2)}$, by the minimum degree condition, we have $d(M, B_{\eps}^{(2)}) \geq q(b-k+2) + \binom{q}{2}$ and, with the above, we have
\begin{align*}
|\ccb| & \geq \left(1 - \binom{k - 2}{\ifmmode\ell\else\polishlcross\fi}\eps' - \binom{k - 2}{\ifmmode\ell\else\polishlcross\fi - 1}\sqrt{\eps' }\right)\binom{b}{k-2}\frac{q(b-k+2)}{\binom{k}{2}} \nonumber \\
& = \left(1 - \binom{k - 2}{\ifmmode\ell\else\polishlcross\fi}\eps' - \binom{k - 2}{\ifmmode\ell\else\polishlcross\fi - 1}\sqrt{\eps' }\right)\binom{b}{k-1}\frac{2q}{k} \geq \frac{q}{k} \binom{b}{k-1}.
\end{align*}
On the other hand, for any $v \in B_{\eps}$ we have $d(v, B_{\eps}^{(k - 1)}) < 2 \varepsilon \binom{b_{\eps}}{k - 1}$ which implies that any edge in $\ccb$ intersects at most $2k \varepsilon \binom{b_{\eps}}{k - 1}$ other edges in $\ccb$.
So, in view of $\varepsilon \ll \frac{1}{k}$ we may pick a set $\ccb'$ of $2q+2$ disjoint edges in $\ccb$.
We will connect each of the edges in $\ccb'$ to an $\varepsilon$-typical set.
Assume we have picked the first $i-1$ desired $\ifmmode\ell\else\polishlcross\fi$-paths, say $\ccp_1, \dots, \ccp_{i-1}$, and denote by $U$ the set of vertices contained in one of the paths or one of the edges in $\ccb'$.
For the rest of this proof, when we pick vertices and edges, they shall always be disjoint from $U$ and everything chosen before.
Let $e$ be an edge in $\ccb'$ we have not considered yet and pick an arbitrary $\varepsilon$-typical set $L' \subset B_{\eps} \smallsetminus U$.
We will first handle the cases that $2\ifmmode\ell\else\polishlcross\fi + 1 < k$ or that $\ifmmode\ell\else\polishlcross\fi=1$, $k=3$.
In the first case, a $(k-2)$-set that contains no $\varepsilon$-atypical set already contains two disjoint $\varepsilon$-typical sets.
In the second case, an $\ifmmode\ell\else\polishlcross\fi$-set $\{v\}$ is $\varepsilon$-typical for any vertex $v$ in $B_{\eps}$ by the definition of $\varepsilon$-typicality.
Hence in both cases $e$ contains two disjoint $\varepsilon$-typical sets, say $L_0$ and $L_1$.
We can use Corollary~\ref{corr:connect-extend-typical}\,\ref{it:connect-typical}, as $|U| \leq 6kq$, to connect $L_1$ to $L'$ and obtain an $\ifmmode\ell\else\polishlcross\fi$-path $\ccp_i$ of size three that contains one vertex in $A_{\eps}$ and has $\varepsilon$-typical ends $L_0$ and $L'$.
So now assume that $2\ifmmode\ell\else\polishlcross\fi + 1 = k$ and $k > 3$, in particular $k - 2 = 2\ifmmode\ell\else\polishlcross\fi - 1$ and we may split the $(k-2)$-set considered in the definition of $\ccb$ into an $\varepsilon$-typical $\ifmmode\ell\else\polishlcross\fi$-set $L$ and a good $(\ifmmode\ell\else\polishlcross\fi-1)$-set $G$.
Moreover, let $w \in e \smallsetminus (L \cup G)$ be one of the remaining two vertices and set~$N = G \cup {w}$.
First assume that $d(N, A_{\eps}^{(1)} B_{\eps}^{(\ifmmode\ell\else\polishlcross\fi)}) \geq \frac{\delta}{3} a_{\eps}\binom{b_{\eps}}{\ifmmode\ell\else\polishlcross\fi}$.
As $\vartheta \ll \delta$, at most $\frac{\delta}{3} a_{\eps} \binom{b}{\ifmmode\ell\else\polishlcross\fi}$ sets in $A_{\eps}^{(1)} B_{\eps}^{(\ifmmode\ell\else\polishlcross\fi)}$ intersect $U$.
So it follows from Lemma~\ref{lem:typical-degree} that there exist $A_{\eps}^{(1)} B_{\eps}^{(\ifmmode\ell\else\polishlcross\fi)}$-sets $C$, $C'$ such that $N \cup C$ and $L' \cup C'$ are edges, $|C \cap C'| = \ifmmode\ell\else\polishlcross\fi$ and~$|C \cap C' \cap A_{\eps}|=1$.
Now assume that $d(N, A_{\eps}^{(1)} B_{\eps}^{(\ifmmode\ell\else\polishlcross\fi)}) < \frac{\delta}{3} a_{\eps}\binom{b_{\eps}}{\ifmmode\ell\else\polishlcross\fi}$.
As the good set $G$ forms an $\varepsilon$-typical set with most vertices in $B_{\eps}$, there exists $v \in B_{\eps}\smallsetminus U$ such that
\begin{equation*}
d(N \cup \{v\}, A_{\eps}^{(1)} B_{\eps}^{(\ifmmode\ell\else\polishlcross\fi - 1)}) < \delta a_{\eps}\binom{b_{\eps}}{\ifmmode\ell\else\polishlcross\fi - 1}
\end{equation*}
and $G \cup \{ v \}$ is an $\varepsilon$-typical set.
Lemma~\ref{lem:typical-degree} implies that
\begin{align*}
d(N \cup \{v\}, B_{\eps}^{(\ifmmode\ell\else\polishlcross\fi)})
&\geq
\frac{2}{\ifmmode\ell\else\polishlcross\fi} \left((1-\delta) a_{\eps} \binom{b_{\eps}-(\ell+1)}{\ifmmode\ell\else\polishlcross\fi - 1} - \deltaa_{\eps}\binom{b_{\eps}}{\ifmmode\ell\else\polishlcross\fi - 1}\right)\\
&\geq
\frac{2}{\ifmmode\ell\else\polishlcross\fi} \left(\frac{1}{2} - 2\delta\right) a_{\eps} \binom{b_{\eps}}{\ifmmode\ell\else\polishlcross\fi - 1}\\
&\geq
\delta \binom{b_{\eps}}{\ifmmode\ell\else\polishlcross\fi}.
\end{align*}
So there exists an $\varepsilon$-typical $\ifmmode\ell\else\polishlcross\fi$-set $L^* \subset (B_{\eps} \smallsetminus U)$ such that $N \cup L^* \cup \{ v \}$ is an edge in $\cch$.
Use Lemma~\ref{corr:connect-extend-typical}\,\ref{it:connect-typical} to connect $L^*$ to $L'$ and obtain an $\ifmmode\ell\else\polishlcross\fi$-path $\ccp_i$ of size three that contains one vertex in $A_{\eps}$ and has $\varepsilon$-typical ends $G \cup \{ v \}$ and $L'$. \end{proof}
If the hypergraph we consider is very close to the extremal example then Lemma~\ref{lem:2q-path} does not apply and we will need the following lemma.
\begin{lemma}\label{lem:one-or-two-edges}
Suppose that $B = B_{\eps}$.
If $n$ is an odd multiple of $k-\ell$ then there exists a single edge on $B_{\eps}$ containing two $\varepsilon$-typical $\ell$-sets.
If $n$ is an even multiple of $k-\ell$ then there either exist two disjoint edges on $B_{\eps}$ each containing two $\varepsilon$-typical $\ell$-sets or an $\ell$-path of size two with $\varepsilon$-typical ends. \end{lemma} \begin{proof}
For the proof of this lemma all vertices and edges we consider will always be completely contained in $B_{\eps}$.
First assume that there exists an $\varepsilon$-atypical $\ell$-set $L$.
Recall that this means that $d(L,B^{(k - \ifmmode\ell\else\polishlcross\fi)}) > \varepsilon\binom{|B|}{k-\ifmmode\ell\else\polishlcross\fi}$ so in view of~\eqref{eq:num-typical-sets} and $\eps' \ll \varepsilon$ we can find two disjoint $(k-\ifmmode\ell\else\polishlcross\fi)$-sets extending it to an edge, each containing an $\varepsilon$-typical set, which would prove the lemma.
So we may assume that all $\ell$-sets in $B_{\eps}^{(\ifmmode\ell\else\polishlcross\fi)}$ are $\varepsilon$-typical.
We infer from the minimum degree condition that $B_{\eps}$ contains a single edge, which proves the lemma in the case that $n$ is an odd multiple of $k-\ifmmode\ell\else\polishlcross\fi$ and for the rest of the proof we assume that $n$ is an even multiple of $k-\ifmmode\ell\else\polishlcross\fi$.
Assume for a moment that $\ifmmode\ell\else\polishlcross\fi = 1$.
Recall that in this case any $(k-2)$-set in $B$ in the extremal hypegraph $\ccx_{k,\ifmmode\ell\else\polishlcross\fi}(n)$ is contained in one edge.
Consequently, the minimum degree condition implies that any $(k-2)$-set in $B_{\eps}$ extends to at least two edges on $B_{\eps}$.
Fix some edge $e$ in $B_{\eps}$; any other edge on $B_{\eps}$ has to intersect $e$ in at least two vertices or the lemma would hold.
Consider any pair of disjoint $(k-2)$-sets $K$ and $M$ in $B_{\eps} \smallsetminus e$ to see that of the four edges they extend to, there is a pair which is either disjoint or intersect in one vertex, proving the lemma for the case $\ifmmode\ell\else\polishlcross\fi=1$.
Now assume that $\ifmmode\ell\else\polishlcross\fi > 1$.
In this case the minimum degree condition implies that any $(k-2)$-set in $B_{\eps}$ extends to at least one edge on $B_{\eps}$.
Again, fix some edge $e$ in $B_{\eps}$; any other edge on $B_{\eps}$ has to intersect $e$ in at least one vertex or the lemma would hold.
Applying the minimum degree condition to all $(k-2)$-sets disjoint from $e$ implies that one vertex $v \in e$ is contained in at least $\frac{1}{2k^2} \binom{|B_{\eps}|}{k-2}$ edges.
We now consider the $(k-1)$-uniform link hypergraph of $v$ on $B_{\eps}$.
Since any two edges intersecting in $\ifmmode\ell\else\polishlcross\fi-1$ vertices would finish the proof of the lemma, we may assume that there are no such pair of edges.
However, a result of Frankl and Füredi~\cite[Theorem 2.2]{FranklFuredi} guarantees that this $(k-1)$-uniform hypergraph without an intersection of size $\ifmmode\ell\else\polishlcross\fi-1$ contains at most $\binom{|B_{\eps}|}{k-\ifmmode\ell\else\polishlcross\fi-1}$ edges, a contradiction. \end{proof}
The following lemma will allow us to handle the vertices in $V_{\eps}$.
\begin{lemma}\label{lemma:pathV0}
Let $U \subset B_{\eps}$ with $|U| \leq 4k\vartheta$.
There exists a family $\ccp_1, \ldots, \ccp_{v_{\eps}}$ of disjoint $\ifmmode\ell\else\polishlcross\fi$-paths of size two, each of which is disjoint from $U$ such that for all $i \in [v_{\eps}]$
\[
|V(\ccp_i) \cap V_{\eps}| = 1 \qand |V(\ccp_i) \cap B_{\eps}| = 2k - \ifmmode\ell\else\polishlcross\fi - 1,
\]
and both ends of $\ccp_i$ are $\varepsilon$-typical sets. \end{lemma} \begin{proof}
Let $V_{\eps} = \{ x_1, \dots, x_{v_{\eps}} \}$.
We will iteratively pick the paths.
Assume we have already chosen $\ifmmode\ell\else\polishlcross\fi$-paths $\ccp_1, \dots, \ccp_{i-1}$ containing the vertices $v_1, \dots, v_{i-1}$ and satisfying the lemma.
Let $U'$ be the set of all vertices in $U$ or in one of those $\ifmmode\ell\else\polishlcross\fi$-paths.
From $v_i \notin B_{\eps}$ we get
\[
d(v_i, B_{\eps}^{(k-1)}) \geq d(v_i, B) - |B \smallsetminus B_{\eps}| \cdot \binom{|B|}{k-2} \geq \frac{\varepsilon}{2} \binom{b_{\eps}}{k-1}.
\]
From~\eqref{eq:num-typical-sets} we get that at most $k^\ifmmode\ell\else\polishlcross\fi \eps' \binom{b_{\eps}}{k-1}$ sets in $B_{\eps}^{(k-1)}$ contain at least one $\varepsilon$-atypical $\ifmmode\ell\else\polishlcross\fi$-set.
Also, less than~$\frac{\varepsilon}{8} \binom{b_{\eps}}{k-1}$ sets in $B_{\eps}^{(k-1)}$ contain one of the vertices of $U'$.
In total, at least $\frac{\varepsilon}{4} \binom{b_{\eps}}{k-1}$ of the $B_{\eps}^{(k-1)}$-sets form an edge with $v_i$.
So we may pick two edges $e$ and $f$ in $V_{\eps}^{(1)} B_{\eps}^{(k-1)}$ that contain the vertex $v_i$ and intersect in $\ifmmode\ell\else\polishlcross\fi$ vertices.
In particular, these edges form an $\ifmmode\ell\else\polishlcross\fi$-path of size two as required by the lemma. \end{proof}
We can now proceed with the proof of Lemma~\ref{lem:mainlemma}. Recall that we want to prove the existence of an $\ifmmode\ell\else\polishlcross\fi$-path $\ccq$ in $\cch$ with ends $L_0$ and $L_1$ and a partition \[
A_{\ast}\dcup B_{\ast}=(V\smallsetminus \ccq)\dcup L_0\dcup L_1 \]
satisfying properties~\ref{it:1-mainlemma}--\ref{it:4-mainlemma} of Lemma~\ref{lem:mainlemma}. Set $q = |A \cap B_{\eps}|$. We will split the construction of the $\ifmmode\ell\else\polishlcross\fi$-path $\ccq$ into two cases, depending on whether $q=0$ or not.
First, suppose that $q > 0$. In the following, we denote by $U$ the set of vertices of all edges and $\ifmmode\ell\else\polishlcross\fi$-paths chosen so far. Note that we will always have $|U| \leq 20 k \vartheta n$ and hence we will be in position to apply Corollary~\ref{corr:connect-extend-typical}. We use Lemma~\ref{lem:2q-path} to obtain paths $\ccq_1, \ldots, \ccq_{2q+2}$ and then we apply Lemma~\ref{lemma:pathV0} to obtain $\ifmmode\ell\else\polishlcross\fi$-paths $\ccp_1, \ldots, \ccp_{v_{\eps}}$. Every path $\ccq_i$, for $i \in [2q+2]$, contains $3k - 2\ifmmode\ell\else\polishlcross\fi - 1$ vertices from $B_{\eps}$ and one from $A_{\eps}$, while every $\ccp_j$, for $j \in [v_{\eps}]$, contains $2k - \ifmmode\ell\else\polishlcross\fi - 1$ from $B_{\eps}$ and one from $V_{\eps}$.
As the ends of all these paths are $\varepsilon$-typical, we apply Corollary~\ref{corr:connect-extend-typical}\,\ref{it:connect-typical} repeatedly to connect them to one $\ifmmode\ell\else\polishlcross\fi$-path $\ccp$. In each of the $v_{\eps} + 2q + 1$ steps of connecting two $\ifmmode\ell\else\polishlcross\fi$-paths, we used one vertex from $A_{\eps}$ and $2k - 3\ifmmode\ell\else\polishlcross\fi - 1$ vertices from $B_{\eps}$. Overall, we have that \[
|V(\ccp) \cap A_{\eps}| = v_{\eps} + 4q + 3, \] as well as \[
|V(\ccp) \cap B_{\eps}| = (4k - 4\ifmmode\ell\else\polishlcross\fi - 2)v_{\eps} + (5k - 5\ifmmode\ell\else\polishlcross\fi - 2)(2q+2) - (2k - 3\ifmmode\ell\else\polishlcross\fi - 1). \]
Furthermore $|V(\ccp)| \leq 10 k \vartheta b$.
Using the identities $a_{\eps} + b_{\eps} + v_{\eps} = n$ and $a_{\eps} + q + v_{\eps} = a$, we will now establish property~\ref{it:1-mainlemma} of Lemma~\ref{lem:mainlemma}. Set $s(\ccp) = (2k - 2\ifmmode\ell\else\polishlcross\fi - 1)|A_{\eps}\smallsetminus V(\ccp)| - |B_{\eps}\smallsetminus V(\ccp)| - 2\ifmmode\ell\else\polishlcross\fi$, so \begin{align*}
s(\ccp) & = (2k - 2\ifmmode\ell\else\polishlcross\fi - 1)|A_{\eps} \smallsetminus V(\ccp)| - |B_{\eps} \smallsetminus V(\ccp)| - 2\ifmmode\ell\else\polishlcross\fi\\
& = (2k - 2\ifmmode\ell\else\polishlcross\fi - 1)(a_{\eps} - (v_{\eps} + 4q + 3)) - b_{\eps} \\
& \phantom{{} = {}} {} + (4k - 4\ifmmode\ell\else\polishlcross\fi -2)v_{\eps} + (5k - 5\ifmmode\ell\else\polishlcross\fi - 2)(2q+2) - (2k - 3\ifmmode\ell\else\polishlcross\fi - 1) - 2\ifmmode\ell\else\polishlcross\fi \\
& = (2k - 2\ifmmode\ell\else\polishlcross\fi - 1)a_{\eps} - b_{\eps} + (2k - 2\ifmmode\ell\else\polishlcross\fi - 1)v_{\eps} + 2(k - \ifmmode\ell\else\polishlcross\fi)q + 2k - 3\ifmmode\ell\else\polishlcross\fi \\
& = 2(k - \ifmmode\ell\else\polishlcross\fi)(a_{\eps} + v_{\eps} + q + 1) - n - \ifmmode\ell\else\polishlcross\fi \\
& = 2(k - \ifmmode\ell\else\polishlcross\fi)(a + 1) - n - \ifmmode\ell\else\polishlcross\fi. \end{align*} If $n/(k - \ifmmode\ell\else\polishlcross\fi)$ is even, $s(\ccp) = -\ifmmode\ell\else\polishlcross\fi$ (see~\eqref{eq:partition-sizes}) and we set $\ccq = \ccp$. Otherwise $s(\ccp) = k - 2\ifmmode\ell\else\polishlcross\fi$ and we use Corollary~\ref{corr:connect-extend-typical}\,\ref{it:extend-typical} to append one edge to $\ccp$ to obtain $\ccq$. It is easy to see that one application of Corollary~\ref{corr:connect-extend-typical}\,\ref{it:extend-typical} decreases $s(\ccp)$ by $k - \ifmmode\ell\else\polishlcross\fi$. Setting $A_{\ast} = A_{\eps}\smallsetminus V(\ccq)$ and $B_{\ast} = (B_{\eps} \smallsetminus V(\ccq)) \cup L_0 \cup L_1$ we get from $s(\ccq) = -\ifmmode\ell\else\polishlcross\fi$ that $A_{\ast}$ and $B_{\ast}$ satisfy~\ref{it:1-mainlemma}.
Now, suppose that $q = 0$. Apply Lemma~\ref{lemma:pathV0} to obtain $\ifmmode\ell\else\polishlcross\fi$-paths $\ccp_1, \ldots, \ccp_{v_{\eps}}$. If $B=B_{\eps}$, apply Lemma~\ref{lem:one-or-two-edges} to obtain one or two more $\ifmmode\ell\else\polishlcross\fi$-paths contained in $B_{\eps}$. We apply Corollary~\ref{corr:connect-extend-typical}\,\ref{it:connect-typical} repeatedly to connect them to one $\ifmmode\ell\else\polishlcross\fi$-path $\ccp$.
Since $q = 0$, we have that $B_{\eps} \subset B$ and $a_{\eps} + v_{\eps} = |V\smallsetminus B_{\eps}| = a + |B\smallsetminus B_{\eps}|$. We can assume without loss of generality that $V_{\eps} \neq \varnothing$, otherwise just take $V_{\eps} = \{ v \}$ for an arbitrary $v \in V (\cch)$. If $B=B_{\eps}$ let $x$ be $2(k-\ifmmode\ell\else\polishlcross\fi)$ or $k-\ifmmode\ell\else\polishlcross\fi$ depending on whether $n$ is an odd or even multiple of $k-\ifmmode\ell\else\polishlcross\fi$; otherwise let $x=0$. With similar calculations as before and the same definition of $s(\ccp)$ we get that \[
s(\ccp) = 2(k-\ifmmode\ell\else\polishlcross\fi)a + x + 2(k-\ifmmode\ell\else\polishlcross\fi)|B \smallsetminus B_{\eps}| - n - \ifmmode\ell\else\polishlcross\fi \equiv -\ifmmode\ell\else\polishlcross\fi \mod(k - \ifmmode\ell\else\polishlcross\fi). \] Extend the $\ifmmode\ell\else\polishlcross\fi$-path $\ccp$ to an $\ifmmode\ell\else\polishlcross\fi$-path $\ccq$ by adding $\frac{s(\ccp) + \ifmmode\ell\else\polishlcross\fi}{k - l}$ edges using Corollary~\ref{corr:connect-extend-typical}\,\ref{it:extend-typical}. Thus $s(\ccq) = -\ifmmode\ell\else\polishlcross\fi$, and we get~\ref{it:1-mainlemma} as in the previous case.
In both cases, we will now use the properties of the constructed $\ifmmode\ell\else\polishlcross\fi$-path $\ccq$ to show~\ref{it:2-mainlemma}-\ref{it:4-mainlemma}. We will use that $v(\ccq) \leq 20 k \vartheta b$, which follows from the construction. Since $A_{\ast} \subset A_{\eps}$, for all $v \in A_{\ast}$ we have $d(v, B^{(k-1)}) \geq (1 - \varepsilon)B^{(k-1)}$. Thus \[
d(v, B_{\ast}^{(k-1)}) \geq d(v, B^{(k-1)}) - |B_{\ast}\smallsetminus B|\binom{|B_{\ast}| - 1}{k-2} \geq (1 - 2\varepsilon)\binom{|B_{\ast}|}{k-1}, \] which shows~\ref{it:2-mainlemma}.
For~\ref{it:3-mainlemma}, Lemma~\ref{lem:typical-degree} yields for all vertices $v \in B_{\ast} \subset B_{\eps}$ that \[
d(v,A_{\eps}^{(1)} B_{\eps}^{(k-2)}) + \frac{k-1}{2} d(v, B_{\eps}^{(k-1)})
\geq
\left(1-\delta\right)|A_{\eps}|\binom{|B_{\eps}| - 1}{k-2}. \]
The second term on the left can be bounded from above by $2k\varepsilon\binom{b_{\eps}}{k-1}$. So, as $\delta, \varepsilon \ll \varrho$ and $a_{\eps} - |A_{\ast}| \ll \varrho |A_{\ast}| $ as well as $b_{\eps} - |B_{\ast}| \ll \varrho |B_{\ast}|$, we can conclude~\ref{it:3-mainlemma}.
By Lemma~\ref{lem:typical-degree}, we know that \[
d(L_{0}, A_{\eps}^{(1)}B_{\eps}^{(k - 1)}),d(L_{1}, A_{\eps}^{(1)}B_{\eps}^{(k - 1)}) \geq (1 - \delta) a_{\eps}\binom{b_{\eps} - \ifmmode\ell\else\polishlcross\fi}{k - \ifmmode\ell\else\polishlcross\fi - 1}. \]
As $\delta \ll \varrho$ and $a_{\eps} - |A_{\ast}| \ll \varrho |A_{\ast}| $ as well as $b_{\eps} - |B_{\ast}| \ll \varrho |B_{\ast}|$, we can conclude~\ref{it:4-mainlemma}.
\begin{bibdiv} \begin{biblist}
\bib{BaMoScScSc16+}{article}{ author={Bastos, {J. de O.}},
author={Mota, G. O.},
author={Schacht, M.},
author={Schnitzer, J.},
author={Schulenburg, F.},
title={Loose Hamiltonian cycles forced by large (k-2)-degree -- approximate version},
journal={SIAM J. Discrete Math.},
volume= {31},
number={4},
year={2017},
pages={ 2328--2347},
doi={10.1137/16M1065732}, }
\bib{BuHaSc13}{article}{ author={Bu{\ss}, Enno},
author={H{\`a}n, Hi\d{\^e}p},
author={Schacht, Mathias},
title={Minimum vertex degree conditions for loose Hamilton cycles in
3-uniform hypergraphs},
journal={J. Combin. Theory Ser. B},
volume={103},
date={2013},
number={6},
pages={658--678},
issn={0095-8956},
review={\MR{3127586}},
doi={10.1016/j.jctb.2013.07.004}, }
\bib{FranklFuredi}{article}{
author={Frankl, Peter},
author={F\"uredi, Zolt\'an},
title={Forbidding just one intersection},
journal={J. Combin. Theory Ser. A},
volume={39},
date={1985},
number={2},
pages={160--176},
issn={0097-3165},
review={\MR{793269}},
doi={10.1016/0097-3165(85)90035-4}, }
\bib{HaSc10}{article}{
author={H{\`a}n, Hi\d{\^e}p},
author={Schacht, Mathias},
title={Dirac-type results for loose Hamilton cycles in uniform
hypergraphs},
journal={J. Combin. Theory Ser. B},
volume={100},
date={2010},
number={3},
pages={332--346},
issn={0095-8956},
review={\MR{2595675}},
doi={10.1016/j.jctb.2009.10.002}, }
\bib{HaZh15}{article}{ author={Han, Jie},
author={Zhao, Yi},
title={Minimum codegree threshold for Hamilton $\ell$-cycles in $k$-uniform hypergraphs},
journal={J. Combin. Theory Ser. A},
volume={132},
date={2015},
pages={194--223},
issn={0097-3165},
review={\MR{3311344}},
doi={10.1016/j.jcta.2015.01.004}, }
\bib{HaZh15b}{article}{
author={Han, Jie},
author={Zhao, Yi},
title={Minimum vertex degree threshold for loose Hamilton cycles in
3-uniform hypergraphs},
journal={J. Combin. Theory Ser. B},
volume={114},
date={2015},
pages={70--96},
issn={0095-8956},
review={\MR{3354291}},
doi={10.1016/j.jctb.2015.03.007}, }
\bib{KaKi99}{article}{
author={Katona, Gyula Y.},
author={Kierstead, H. A.},
title={Hamiltonian chains in hypergraphs},
journal={J. Graph Theory},
volume={30},
date={1999},
number={3},
pages={205--212},
issn={0364-9024},
review={\MR{1671170}},
doi={10.1002/(SICI)1097-0118(199903)30:3<205::AID-JGT5>3.3.CO;2-F}, }
\bib{KeKuMyOs11}{article}{
author={Keevash, Peter},
author={K{\"u}hn, Daniela},
author={Mycroft, Richard},
author={Osthus, Deryk},
title={Loose Hamilton cycles in hypergraphs},
journal={Discrete Math.},
volume={311},
date={2011},
number={7},
pages={544--559},
issn={0012-365X},
review={\MR{2765622}},
doi={10.1016/j.disc.2010.11.013}, }
\bib{KuMyOs10}{article}{
author={K{\"u}hn, Daniela},
author={Mycroft, Richard},
author={Osthus, Deryk},
title={Hamilton $\ell$-cycles in uniform hypergraphs},
journal={J. Combin. Theory Ser. A},
volume={117},
date={2010},
number={7},
pages={910--927},
issn={0097-3165},
review={\MR{2652102}},
doi={10.1016/j.jcta.2010.02.010}, }
\bib{KuOs06}{article}{
author={K{\"u}hn, Daniela},
author={Osthus, Deryk},
title={Loose Hamilton cycles in 3-uniform hypergraphs of high minimum
degree},
journal={J. Combin. Theory Ser. B},
volume={96},
date={2006},
number={6},
pages={767--821},
issn={0095-8956},
review={\MR{2274077}},
doi={10.1016/j.jctb.2006.02.004}, }
\bib{RRsurv}{article}{
author={R{\"o}dl, Vojtech},
author={Ruci{\'n}ski, Andrzej},
title={Dirac-type questions for hypergraphs---a survey (or more problems
for Endre to solve)},
conference={
title={An irregular mind},
},
book={
series={Bolyai Soc. Math. Stud.},
volume={21},
publisher={J\'anos Bolyai Math. Soc., Budapest},
},
date={2010},
pages={561--590},
review={\MR{2815614}},
doi={10.1007/978-3-642-14444-8\_16}, }
\bib{RoRuSz06}{article}{
author={R{\"o}dl, Vojt{\v{e}}ch},
author={Ruci{\'n}ski, Andrzej},
author={Szemer{\'e}di, Endre},
title={A Dirac-type theorem for 3-uniform hypergraphs},
journal={Combin. Probab. Comput.},
volume={15},
date={2006},
number={1-2},
pages={229--251},
issn={0963-5483},
review={\MR{2195584}},
doi={10.1017/S0963548305007042}, }
\bib{RoRuSz08}{article}{
author={R{\"o}dl, Vojt{\v{e}}ch},
author={Ruci{\'n}ski, Andrzej},
author={Szemer{\'e}di, Endre},
title={An approximate Dirac-type theorem for $k$-uniform hypergraphs},
journal={Combinatorica},
volume={28},
date={2008},
number={2},
pages={229--260},
issn={0209-9683},
review={\MR{2399020}},
doi={10.1007/s00493-008-2295-z}, }
\bib{Zhao-survey}{article}{
author={Zhao, Yi},
title={Recent advances on Dirac-type problems for hypergraphs},
conference={
title={Recent trends in combinatorics},
},
book={
series={IMA Vol. Math. Appl.},
volume={159},
publisher={Springer},
},
date={2016},
pages={145--165},
review={\MR{3526407}},
doi={10.1007/978-3-319-24298-9\_6}, }
\end{biblist} \end{bibdiv}
\end{document}
|
arXiv
|
{
"id": "1705.03707.tex",
"language_detection_score": 0.6299752593040466,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Characterization of Extreme Copulas}
\begin{abstract}
In this paper our aim is to characterize the set of extreme points of the set of all $n$-dimensional copulas $(n>1)$. We have shown that a copula must induce a singular measure with respect to Lebesgue measure in order to be an extreme point in the set of $n$-dimensional copulas. We also have discovered some sufficient conditions for a copula to be an extreme copula. We have presented a construction of a small subset of $n$-dimensional extreme copulas such that any $n$-dimensional copula is a limit point of that subset with respect to weak convergence. The applications of such a theory are widespread, finding use in many facets of current mathematical research, such as distribution theory, survival analysis, reliability theory and optimization purposes. To illustrate the point further, examples of how such extremal representations can help in optimization have also been included.
\end{abstract}
\section{Introduction}
\hspace*{3mm} Copula models are popular in high dimensional statistical applications due to their ability to describe the dependence structure among random variables; see, e.g. \hyperlink{ref9}{[9]}, \hyperlink{ref10}{[10]}. In a situation where we need to study the influence of dependence structure on a statistical problem with given marginals of some random vector, it would be helpful if we do not have to study dependence structure over the class of all copulas- only a small class of copulas would do the job. In \hyperlink{ref6}{[6]}, \hyperlink{ref7}{[7]}
a special case has been considered. For a bi-variate random vector $(X,Y)$ with continuous and strictly increasing marginal distribution functions they have found a copula under which the probability of the event $\left\{X=Y\right\}$ is maximal. However, our work is applicable in more general scenarios. Using Krein-Milman theorem (see e.g. \hyperlink{ref2}{[2]}) we can see that study of dependence structure over only the convex hull of the extreme copulas is enough. This motivates us to study and characterize the set of extreme copulas. It is also interesting from mathematical point of view. In this paper we aim to provide some properties such that any copula satisfying these properties will be an extreme copula. We also have shown that the probability induced by any extreme copula has to be singular with respect to Lebesgue measure. This shows how small the class of extreme copulas is. Finally, we have shown a particularly strong result that says we do not need to consider even the convex hull of the extreme points, but only a small subset of these extreme points in order to study the influence of dependence structure.
\section{Preliminaries}
\hspace*{3mm} First, we recall the definition of copula. see, e.g. \hyperlink{ref5}{[5]}.\\
\begin{defn}
An ${n}${-dimensional copula} is an $n$-dimensional distribution function concentrated on $\left[0,1\right]^n$ whose univariate marginals are uniformly distributed on $\left[0,1\right]$.
\end{defn}
\hspace*{3mm} Clearly the set of all copulas is a convex set and it is closed (with respect to $ d_{\infty}$ metric), uniformly bounded by 1 and every copula is 1-Lipschitz continuous. Hence by Arzel\`{a}-Ascoli theorem (see, e.g. \hyperlink{ref3}{[3]}) the set of all copulas is compact. Again, as it is convex and compact by Krein-Milman theorem (see, e.g. \hyperlink{ref4}{[4]}) the set of all copulas is the closure (with respect to $d_{\infty}$ metric) of the convex hull of its extreme points. So we will study the set of extreme points.
\section{Notations}
\hspace*{3mm} For a Borel-measurable function $f : \left[0,1\right]\mapsto\left[0,1\right]^{n-1}$ denote the graph of $f$ as
\[
\mathcal{G}_f=\left\{ \left(x,f(x)\right) \mid x\in \left[0,1\right] \right\}
\]
Write $f$ as $(g_1,g_2,\ldots,g_{n-1})$. We denote \[\mathcal{G}_f^{(i)}=\left\{(g_1(x),\ldots,g_{i-1}(x),x,g_i(x),\ldots,g_{n-1}(x) ) \mid x\in \left[0,1\right] \right\}\]
We denote the projection on $i$th co-ordinate by $\pi_i$ i.e. $\pi_i:\left[0,1\right]^n\mapsto\left[0,1\right]$ such that $\pi_i\left((x_1,x_2,\ldots,x_n)\right)=x_i$. Now for any probability $P$ on $\left[0,1\right]^n$ we denote $P\circ\pi_i^{-1}$, the $i$th marginal of $P$ by $P_i$. Also, we denote the Lebesgue measure by $\lambda$.\\
Define the $n$-dimensional square $S_{\utilde{a},\varepsilon}$ as $\prod_{i=1}^{n}\left[a_i,a_i+\varepsilon\right]$.
Now denote $R_{\utilde{a},\utilde{b}}$ as the $n$-dimensional rectangle formed by the two points $\utilde{a}$ and $\utilde{b}$ i.e.
${R_{\utilde{a},\utilde{b}}}=\{\utilde{x}\in \mathbb{R}^n \mid \utilde{a}\leq \utilde{x}\leq\utilde{b} \}$.\\
We denote $u_j(t)=(x_1,x_2,\ldots,x_n)$ such that $x_j=t$ and $x_k=1$ for all $k\neq j$ and $v_j(t)=(y_1,y_2,\ldots,y_n)$ such that $y_j=t$ and $y_k=0$ for all $k\neq j$.
\section{Sufficient Conditions}
\hspace*{3mm} In this section we provide certain sufficient conditions for a copula to be an extreme copula. \\
\begin{thm}
\hypertarget{theo.4.1}{}
Let $f : \left[0,1\right]\mapsto\left[0,1\right]^{n-1}$ be a Borel measurable function and $\mu$ be a probability measure on $\left(\left[0,1\right],\mathcal{B}(\left[0,1\right])\right)$. Then there exists a unique $n$-dimensional probability supported on $\mathcal{G}_f$ whose first marginal is $\mu$.
\end{thm}
\begin{proof}
At first we will show that there exists an $n$-dimensional probability supported on $\mathcal{G}_f$ whose first marginal is $\mu$. It will be enough to get such an $n$-dimensional probability $P$ on $\{S_1\times S_2 \mid S_1 \in \mathcal{B}(\left[0,1\right]), S_2 \in \mathcal{B}(\left[0,1\right]^{n-1}) \}$ as it is a semi-field for $\mathcal{B}\left(\left[0,1\right]^{n}\right)$. Define P on this semi-field as
\[P(S_1\times S_2)=\mu(S_1\cap f^{-1}(S_2))\]
for all $S_1 \in \mathcal{B}(\left[0,1\right])$ and $S_2 \in \mathcal{B}(\left[0,1\right]^{n-1}) $.
Clearly it is a probability and by Caratheodory Extension Theorem $P$ can be defined on $\mathcal{B}\left(\left[0,1\right]^{n}\right)$. Observe that $P$ is supported on $\mathcal{G}_f$ and its first marginal is $\mu$. \\
\hspace*{3mm} Now it remains to show is that $P$ is the unique probability with the required property. Let $\bar{P}$ be another $n$-dimensional probability supported on $\mathcal{G}_f$ whose first marginal is $\mu$. Then for all $S_1 \in \mathcal{B}(\left[0,1\right])$ and $S_2 \in \mathcal{B}(\left[0,1\right]^{n-1} ) $ we have\\
\begin{equation*}
\begin{split}
\bar{P}(S_1\times S_2) &= \bar{P}\left((S_1\times S_2)\cap\mathcal{G}_f \right)\\
&= \bar{P}\left( \left\{(x,f(x)) \mid x \in S_1, f(x) \in S_2 \right\} \right)\\
&= \bar{P}\left( \left\{(x,f(x)) \mid x \in S_1\cap f^{-1}(S_2) \right\} \right)\\
&= \bar{P}\left( \left\{(x,y) \mid x \in S_1\cap f^{-1}(S_2), y\in \left[0,1\right]^{n-1} \right\} \right)\\
&= \mu(S_1\cap f^{-1}(S_2))\\
&= P(S_1\times S_2)
\end{split}
\end{equation*}
As $\{S_1\times S_2 \mid S_1 \in \mathcal{B}(\left[0,1\right]), S_2 \in \mathcal{B}(\left[0,1\right]^{n-1}) \}$ is a semi-field for $\mathcal{B}\left(\left[0,1\right]^{n}\right)$, by Caratheodory extension theorem $\bar{P}=P$. Hence $P$ is the unique probability supported on $\mathcal{G}_f$ whose first marginal is $\mu$.
\end{proof}
\begin{cor}
For any Borel measurable function $f : \left[0,1\right]\mapsto\left[0,1\right]^{n-1}$ any copula supported on $\mathcal{G}_f$ is an extreme copula.
\end{cor}
\begin{proof}
Let $f : \left[0,1\right]\mapsto\left[0,1\right]^{n-1}$ be a Borel measurable function. Suppose there exists a copula $C$ supported on $\mathcal{G}_f$. As the first marginal of every copula is $\mathbb{U}\left[0,1\right]$, by using \hyperlink{theo.4.1}{Theorem 4.1} we can conclude that it is the unique copula supported on $\mathcal{G}_f$. If there exist any two copulas $C_1$ and $C_2$ such that $C$ can be written as a convex combination of them, then both $C_1$ and $C_2$ need to give zero measure on ${\mathcal{G}_f}^c$. Hence $C_1=C_2=C$, which implies $C$ is an extreme copula.
\end{proof}
\hspace*{3mm} For an extreme copula $C$ and for any permutation $\sigma$ on $\left\{1,2,\ldots,n\right\}$ the copula $C_\sigma$ defined as $C_\sigma(x_1,x_2,\ldots,x_n)=C(x_{\sigma(1)},x_{\sigma(2)},\ldots,x_{\sigma(n)})$ is also an extreme copula. Clearly there exists a copula supported on $\mathcal{G}_f^{(i)}$ iff there exists a copula supported on $\mathcal{G}_f^{(j)}$
for any $i,j \in \left\{1,2,\ldots,n\right\}$.\\
\begin{cor}
There are uncountably many extreme copulas on $\left[0,1\right]^n$.
\end{cor}
\begin{proof}
Fix $t\in (0,1)$. Let $L_1$ be the line joining $\utilde{0}$ and $(t,1,1,\ldots,1)$ and $L_2$ be the line joining $(t,1,1,\ldots,1)$ and $(1,0,0,\ldots,0)$. See the figure for $n=2$.\\
\begin{center}
\scalebox{.45}{ \includegraphics{cor03}}\\
\end{center}
Let $\mathbb{U}(L_i)$ denote the uniform distribution on $L_i$. Then c.d.f. of $t\mathbb{U}(L_1)+(1-t)\mathbb{U}(L_2)$ is a copula on $L_1\cup L_2$ and hence by previous corollary it is an extreme copula. Thus for every distinct $t\in(0,1)$ we get a distinct extreme copula. And hence there are uncountably many extreme copulas.
\end{proof}
\hspace*{3mm}We can see the above theorem talks about a small class of copulas. For example it can not say whether the following 2-dimensional curve can be a support of an extreme copula or not.
\begin{center}
\scalebox{.35}{\includegraphics{thm1_n}}
\end{center}
Therefore we need to have a more general theorem.\\
\begin{thm}
Let $D$ be a support of a copula such that there exist $B_1, B_2,\ldots, B_n\in \mathcal{B}(\left[0,1\right]) $ and Borel measurable functions $f_i : \left[0,1\right]\mapsto\left[0,1\right]^{n-1}$ for $i\in \{1,2,\ldots,n\}$ such that $\bigcup_{i=1}^n\pi_i^{-1}(B_i) \supseteq D $ and $\pi_i^{-1}(B_i)\cap D=\pi_i^{-1}(B_i)\cap \mathcal{G}_{f_i}^{(i)}$. Then the copula will be an extreme copula. In particular it will be the unique copula supported on $D$.
\end{thm}
\begin{proof}
We only need to show that there exists at most one copula supported on $D$. Let $C$, $\bar{C}$ be two copulas supported on $D$. Let $P$, $\bar{P}$ be the corresponding induced probabilities. Observe for any $S_i\in \mathcal{B} \left( \left[0,1\right]^n \right) $, $S_i\subseteq\pi_i^{-1}(B_i)$
\begin{equation*}
\begin{split}
P(S_i)= P(S_i\cap D)=P\left(S_i\cap \mathcal{G}_{f_i}^{(i)} \right)
&= P\left(\pi_i^{-1}\left( \pi_i \left(S_i\cap \mathcal{G}_{f_i}^{(i)} \right) \right) \right)\\
&= \lambda \left( \pi_i \left(S_i\cap \mathcal{G}_{f_i}^{(i)}\right) \right)
\end{split}
\end{equation*}
Similarly
\begin{equation*}
\begin{split}
\bar{P}(S_i)= \bar{P}(S_i\cap D)=\bar{P}\left(S_i\cap \mathcal{G}_{f_i}^{(i)} \right)
&= \bar{P}\left(\pi_i^{-1}\left( \pi_i \left(S_i\cap \mathcal{G}_{f_i}^{(i)} \right) \right) \right)\\
&= \lambda \left( \pi_i \left(S_i\cap \mathcal{G}_{f_i}^{(i)} \right) \right)
\end{split}
\end{equation*}
Therefore for all $i$, ${P}(S_i)=\bar{P}(S_i)$ whenever $S_i\in \mathcal{B} \left( \left[0,1\right]^n \right) $ and $S_i\subseteq\pi_i^{-1}(B_i)$. Now for any $A\in \mathcal{B} \left( \left[0,1\right]^n \right) $ there exist Borel sets $A_0,A_1,\ldots,A_n$ disjoint such that $A=\bigcup_{i=0}^n A_i$ where $A_0\subseteq D^C$ and for $i \geq 1$, $A_i\subseteq\pi_i^{-1}(B_i)$. So we have
\[\bar{P}(A)=\sum\limits_{i=0}^n \bar{P}(A_i)=\sum\limits_{i=0}^n {P}(A_i) = P(A) \] Which implies $P=\bar{P}$ and hence $C=\bar{C}$. So, we have proved that there exists at most one copula supported on $D$ and hence it is an extreme copula.
\end{proof}
\hspace*{3mm} This theorem overcomes some of the drawbacks of the previous theorem. For example it can say that if there exists a copula supported on the following 2-dimensional curve then it has to be an extreme copula.\\
\begin{center}
\includegraphics[scale=.5]{gen_thm_n}
\end{center}
Now we will give some examples of extreme copulas.\\
\begin{exmp}
Let $0=t_1<t_2<\cdots<t_m<t_{m+1}=1$.
Let $L_i$ be any interior diagonal of the $n$-dimensional rectangle $R_{v_1(t_i),u_1(t_{i+1})}$. See the figure for $n=2$.\\
\begin{center}
\scalebox{.5}{\includegraphics{ex1}}
\end{center}
Then c.d.f. of the probability measure $P=\sum_{i=1}^{m}(t_{i+1}-t_i)\mathbb{U}(L_i)$ is an extreme copula. To see this let
\[
f(x)=\sum\limits_{i=1}^m L_i(x)\cdot 1_{\left\{x\in \left[ t_i,t_{i+1} \right) \right\}}
\]
Observe
\[
\left( \bigcup\limits_{i=1}^m L_i \right)\Delta\mathcal{G}_f \subseteq \left\{ \left( t_{i+1}, L_i\left(t_{i+1}\right) \right) \mid i \in \mathbb{N} \right\}
\]
But $P\left(\left\{ \left( t_{i+1}, L_i\left(t_{i+1}\right) \right) \mid i \in \mathbb{N} \right\}\right)=0$. So $P$ is supported on $\mathcal{G}_f$, $f$ measurable.
Observe $P_1=\sum_{i=1}^{m} (t_{i+1}-t_i)\mathbb{U}\left[t_i,t_{i+1}\right]=\mathbb{U}\left[0,1\right]$ and for $j\neq 1$ we have $P_j=\sum_{i=1}^{m} (t_{i+1}-t_i)\mathbb{U}\left[0,1\right]=\mathbb{U}\left[0,1\right]$. Therefore c.d.f. of $P$ is a copula and hence is an extreme copula.\\
\end{exmp}
\begin{exmp}
Let $\{t_m\}$ be any non-decreasing sequence of real numbers in $\left[0,1\right]$ with $t_1=0$ and $t_0=\lim_{m\rightarrow\infty} t_m$. Let $L_i$ be any interior diagonal of the $n$-dimensional rectangle $R_{v_1(t_i),u_1(t_{i+1})}$ and $L_0$ be any interior diagonal of the $n$-dimensional rectangle $R_{v_1(t_0),\utilde{1}}$. Then by similar approach one can get that the c.d.f. of the probability measure $P=(1-t_0)\mathbb{U}(L_0)+\sum_{i=1}^{\infty}(t_{i+1}-t_i)\mathbb{U}(L_i)$ is an extreme copula.\\
\end{exmp}
\begin{exmp}[\textbf{Permutation copula of order} $\mathbf{m}$]
Draw grids of size $\frac{1}{m}$ in $\left[0,1\right]^n$. Let $\sigma_k$ for $k=2,3,\ldots,n$ are permutations on $\{0,1,\ldots,m-1\}$. Define $\sigma_1$ as the identity permutation.\\
Let $L_i$ be an interior diagonal of $S_{(\frac{\sigma_1(i)}{m},\frac{\sigma_2(i)}{m},\ldots,\frac{\sigma_n(i)}{m}),\frac{1}{m} }$ for all $i=0,1,\ldots,m-1$. See the figure for $n=2$.\\
\begin{center}
\scalebox{.5}{\includegraphics{ex4}}
\end{center}
Let us define a probability $P$ as
\[
P=\sum_{i=0}^{m-1} \frac{1}{m}\cdot\mathbb{U}(L_i)
\]
Observe for all $k$, the $k$th marginal of $S_{(\frac{\sigma_1(i)}{m},\frac{\sigma_2(i)}{m},\ldots,\frac{\sigma_n(i)}{m}),\frac{1}{m} }$ with respect to $P$ is\linebreak $\frac{1}{m}\cdot\mathbb{U}\left(\left[\frac{\sigma_k(i)}{m},\frac{\sigma_{k}(i)+1}{m} \right]\right)$. Therefore for all $k$
\[
P_k= \sum_{i=0}^{m-1} \frac{1}{m}\cdot\mathbb{U}\left(\left[\frac{\sigma_k(i)}{m},\frac{\sigma_{k}(i)+1}{m} \right]\right)=\mathbb{U}\left(\left[0,1\right]\right)
\]
Hence c.d.f. of $P$ is a copula. Such a copula will be called permutation copula. Let
\[
f(x)=\sum\limits_{i=0}^{m-1} L_i(x)\cdot 1_{\left\{x\in \left[ \frac{i}{m},\frac{i+1}{m} \right) \right\}}
\]
Clearly $f$ is measurable. Observe
\[
P\left[ \left( \bigcup\limits_{i=0}^{m-1} L_i \right)\Delta\mathcal{G}_f \right]\leq P\left[ \left\{ \left. \left( \frac{i+1}{m}, L_i\left(\frac{i+1}{m}\right) \right)\,\,\,\right| \,\,\, i \in \{0,1,\ldots,m-1\} \right\} \right]=0
\]
Therefore $P$ is supported on $\mathcal{G}_f$ and hence c.d.f. of $P$ is an extreme copula.\\
\end{exmp}
\begin{exmp}
Let $C$ be a copula and $P$ be the corresponding induced probability For any $\utilde{\alpha}\in \mathbb{R}^n$ for any $A\subseteq \left[0,1\right]^n$ define
\[
A+\utilde{\alpha}=\left\{ \left( \{x_1+\alpha_1\},\{x_2+\alpha_2\},\ldots,\{x_n+\alpha_n\} \right) \mid \utilde{x}\in A \right\}
\]
where $\{.\}$ denotes the fractional part. See the figure for $n=2$.\\
\begin{center}
\includegraphics[scale=1]{ex_n4}
\end{center} \vspace*{.5cm}
Define $P_{\utilde{\alpha}}$ as $P_{\utilde{\alpha}}(B)=P(B+\utilde{\alpha})$ for all $B\in \mathcal{B}\left(\left[0,1\right]^n\right)$. Clearly $C_{\utilde{\alpha}}$, the c.d.f. of $P_{\utilde{\alpha}}$ is a copula. If $C$ is not an extreme copula i.e. $\exists$ copulas $C_1\neq C_2 \neq C$ such that $C=\frac{1}{2}\left(C_1+ C_2\right)$. This implies $C_{\utilde{\alpha}}=\frac{1}{2}\left((C_1)_{\utilde{\alpha}}+ (C_2)_{\utilde{\alpha}}\right)$. Hence $C_{\utilde{\alpha}}$ is not an extreme copula. Again, as $C=\left(C_{\utilde{\alpha}}\right)_{-\utilde{\alpha}}$. In a similar vein we can say $C_{\utilde{\alpha}}$ is not an extreme copula implies $C$ is not an extreme copula. Hence $C$ is an extreme copula iff $C_{\utilde{\alpha}}$ is an extreme copula.\\
\end{exmp}
\begin{exmp}
Let $C$ be a copula and $P$ be the induced probability. Fix a co-ordinate $i$ and two disjoint intervals $\left[a,a+\delta\right],\,\left[b,b+\delta\right]\subseteq\left[0,1\right]$. Define a transformation $T$ as
\[
T(\utilde{x})=\left\{
\begin{array}{ll}
\left(x_1,\ldots,x_{i-1},x_i-a+b,x_{i+1},\ldots,x_n\right)& \mbox{, if } \utilde{x}\in \pi_i^{-1}\left(\left[a,a+\delta\right] \right)\\
\left(x_1,\ldots,x_{i-1},x_i-b+a,x_{i+1},\ldots,x_n\right)& \mbox{, if } \utilde{x}\in \pi_i^{-1}\left(\left[b,b+\delta\right] \right)\\
\utilde{x}&\mbox{, otherwise}
\end{array}
\right.
\]
\vspace*{.5cm}
\begin{center}
\includegraphics[scale=1]{ex_n5}
\end{center}
Let $P'$ be the probability induced by $T$. As $T=T^{-1}$ we can say $P'=P\circ T$. Let the $C'$ be c.d.f. of $P'$. Clearly $C'$ is a copula and it can be easily shown that $C$ is an extreme copula iff $C'$ is an extreme copula.\\
\end{exmp}
\hspace*{3mm}Starting from the copula supported on any interior diagonal of $\left[0,1\right]^n$ if one applies the transformations mentioned in the above examples, he will always get a copula supported on a graph of a Borel measurable function. Now we are going to show that if a copula is supported on graph of a Borel measurable function then the function has to be measure preserving in each co-ordinate with respect to Lebesgue measure.
Conversely for any Borel measurable function from $\left[0,1\right]$ to $\left[0,1\right]^{n-1}$, there exists a copula supported on its graph if it is measure preserving in each co-ordinate with respect to Lebesgue measure.\\
\begin{thm}
For any Borel measurable function $f : \left[0,1\right]\mapsto\left[0,1\right]^{n-1}$, $\exists$ a copula supported on $\mathcal{G}_f$ iff $f$ is measure preserving in each co-ordinate with respect to $\lambda$. In that case the copula will be an extreme copula
\end{thm}
\begin{proof}
Let $f : \left[0,1\right]\mapsto\left[0,1\right]^{n-1}$ be a Borel measurable function. Suppose there is a copula $C$ supported on $\mathcal{G}_f$. Let $P$ be the probability induced by $C$. write $f$ as $(f_2,f_3,\ldots,f_n)$. Then for all $i$, for all $B\in\mathcal{B}(\left[0,1\right])$
\[
\begin{split}
\lambda\left[f_i^{-1}(B)\right] &= P\left[\pi_1^{-1}\left(f_i^{-1}(B)\right)\right]\\
&= P\left[\pi_1^{-1}\left(f_i^{-1}(B)\right)\cap \mathcal{G}_f \right]\\
&= P\left[ \left\{ (x,f(x)) \mid x\in f_i^{-1}(B) \right\} \right]\\
&= P\left[ \left\{ (x,f(x)) \mid f_i(x)\in B \right\} \right]\\
&= P\left[\pi_i^{-1}(B)\cap \mathcal{G}_f \right]\\
&= P\left[ \pi_i^{-1}(B) \right]\\
&= \lambda\left[B\right]
\end{split}
\]
Hence $f$ is measure preserving in each co-ordinate with respect to $\lambda$.\\
\begin{center}
\includegraphics[scale=.4]{theo_4_3_n}
\end{center}
\hspace*{3mm}Now suppose $f : \left[0,1\right]\mapsto\left[0,1\right]^{n-1}$ be a Borel measurable function which is measure preserving in each co-ordinate with respect to $\lambda$. Define a probability $P$ as \[P(B) = \lambda\left[\pi_1(B\cap\mathcal{G}_f)\right]\] for all $B\in\mathcal{B}(\left[0,1\right])$. Clearly $P$ is supported on $\mathcal{G}_f$. Let $P_i$ be its $i$th marginal. Observe
\[
\begin{split}
P_1 (B)= P\left[\pi_1^{-1}(B)\right] &=P\left[\pi_1^{-1}(B)\cap \mathcal{G}_f \right]\\
&= \lambda \left[ \pi_1\left( \pi_1^{-1}(B)\cap \mathcal{G}_f \right) \right]\\
&= \lambda \left[B\right]
\end{split}
\]
For all $i\neq 1$
\[
\begin{split}
P_i (B)= P\left[\pi_i^{-1}(B)\right] &=P\left[\pi_i^{-1}(B)\cap \mathcal{G}_f \right]\\
&= P\left[\pi_1^{-1}\left(f_i^{-1}(B)\right)\cap \mathcal{G}_f \right]\\
&= P\left[\pi_1^{-1}\left(f_i^{-1}(B)\right)\right]\\
&= P_1\left[f_i^{-1}(B)\right]\\
&= \lambda\left[f_i^{-1}(B)\right]\\
&= \lambda \left[B\right]
\end{split}
\]
Therefore c.d.f. of $P$ is a copula.\\
\hspace*{3mm}Clearly as the copula is supported on a graph of a Borel measurable function it is an extreme copula.
\end{proof}
\section{Necessary Conditions}
\hspace*{3mm}A trivial necessary condition for a set $D\subseteq \left[0,1\right]^n$ to be the support of a copula is that for any $B\in \mathcal{B} \left( \left[0,1\right]^n \right)$ satisfying $B\cap D= \pi_i^{-1}\left(\pi_i(B)\right)\cap D=\pi_j^{-1}\left(\pi_j(B)\right)\cap D $ we must have $\lambda\left(\pi_i(B)\right)=\lambda\left(\pi_j(B)\right)$.
Here using Lebesgue decomposition theorem (see, e.g. \hyperlink{ref1}{[1]}) we are going to show that probability induced by any extreme copula has to be singular with respect to Lebesgue measure. To prove this we need to prove the following lemma.\\
\begin{lem}
\hypertarget{lemma.5.1}{}
Let $F$ be a copula and $P_F$ be the induced probability. Let the Lebesgue decomposition be \[ P_F (A)=\int_A f\,d\lambda^n\, +\mu(A)\]
Where $f$ is non-negative, $A\in \mathcal{B}\left(\left[0,1\right]^n\right)$ and $\mu\perp\lambda^n$. Suppose there exists a closed square $S_{\utilde{a},\varepsilon}=\prod_{i=1}^{n}\left[a_i,a_i+\varepsilon\right]$ $ \subseteq \left[0,1\right]^n$ such that
\[\lambda^n\left[f^{-1}\left( \{0\} \right)\cap S_{\utilde{a},\varepsilon} \right] <\frac{1}{4}\cdot\lambda^n\left[ S_{\utilde{a},\varepsilon} \right] =\frac{\varepsilon^n}{4} \]
Then $F$ can not be an extreme copula.
\end{lem}
\begin{proof}
We will get a $g: \left[0,\frac{\varepsilon}{2}\right]\times\left[0,\frac{\varepsilon}{2}\right] \times \left[0,\varepsilon\right]^{n-2}\mapsto\mathbb{R}$ which is not 0 almost everywhere such that $h_1$ and $h_2$ defined below are non-negative. Define for $i=1,2$
\[
h_i(\utilde{x})=\left\{ \begin{array}{l}
f(\utilde{x}) \mbox{\hspace*{1.7cm} if } \utilde{x}\in S_{\utilde{a},\varepsilon}^C\\[.35cm]
f(\utilde{x}) +(-1)^ig(\utilde{x}-\utilde{a})\\
\mbox{\hspace*{2.5cm} if } \utilde{x}\in \left[a_1,a_1+\frac{\varepsilon}{2}\right]\times\left[a_2,a_2+\frac{\varepsilon}{2}\right]\times\prod\limits_{i=3}^n\left[a_i,a_i+\varepsilon\right]\\[.35cm]
f(\utilde{x}) +(-1)^{i+1}g(\utilde{x}-(a_1+\frac{\varepsilon}{2},a_2,a_3,\ldots,a_n)) \\
\mbox{\hspace*{2.5cm} if } \utilde{x}\in \left[a_1+\frac{\varepsilon}{2},a_1+\varepsilon\right]\times\left[a_2,a_2+\frac{\varepsilon}{2}\right]\times\prod\limits_{i=3}^n\left[a_i,a_i+\varepsilon\right]\\[.35cm]
f(\utilde{x}) +(-1)^{i+1}g(\utilde{x}-(a_1,a_2+\frac{\varepsilon}{2},a_3,\ldots,a_n))\\
\mbox{\hspace*{2.5cm} if } \utilde{x}\in \left[a_1,a_1+\frac{\varepsilon}{2}\right]\times\left[a_2+\frac{\varepsilon}{2},a_2+\varepsilon\right]\times\prod\limits_{i=3}^n\left[a_i,a_i+\varepsilon\right]\\[.35cm]
f(\utilde{x}) +(-1)^{i}g(\utilde{x}-(a_1+\frac{\varepsilon}{2},a_2+\frac{\varepsilon}{2},a_3,\ldots,a_n))\\
\mbox{\hspace*{2.5cm} if } \utilde{x}\in \left[a_1+\frac{\varepsilon}{2},a_1+\varepsilon\right]\times\left[a_2+\frac{\varepsilon}{2},a_2+\varepsilon\right]\times\prod\limits_{i=3}^n\left[a_i,a_i+\varepsilon\right]
\end{array}
\right.
\]
Then we will define the copulas $H_1$ and $H_2$ such that
\[P_{H_1}(A)=\int_A h_1\,d\lambda^n\, +\mu(A) \mbox{\hspace{2cm}} P_{H_2}(A)=\int_A h_2\,d\lambda^n\, +\mu(A)\]
where $P_{H_1}$ and $P_{H_2}$ are probabilities induced by $H_1$ and $H_2$ respectively. Clearly $H_1$ and $H_2$ are copulas. As $f=\frac{1}{2}(h_1+h_2)$, hence $P_F=\frac{1}{2}(P_{H_1}+P_{H_2})$. Therefore $F=\frac{1}{2}(H_1+H_2)$. As $g$ in not 0 almost everywhere $F\neq {H_1}\neq {H_2}$. Hence $F$ will not be
an extreme copula. So it is enough to get $g$ which is not 0 almost everywhere such that both $h_1$ and $h_2$ are non-negative. For this we will define functions $f_1,\,f_2,\,f_3,\,f_4$ on $\left[0,\frac{\varepsilon}{2}\right]\times\left[0,\frac{\varepsilon}{2}\right] \times \left[0,\varepsilon\right]^{n-2}$ such that
\[
\begin{split}
f_1(\utilde{x})&= f(\utilde{x}+\utilde{a})\\
f_2(\utilde{x})&= f\left(\utilde{x}+\left(a_1+\frac{\varepsilon}{2},a_2,a_3,\ldots,a_n\right)\right)\\
f_3(\utilde{x})&= f\left(\utilde{x}+\left(a_1,a_2+\frac{\varepsilon}{2},a_3,\ldots,a_n\right)\right)\\
f_4(\utilde{x})&= f\left(\utilde{x}+\left(a_1+\frac{\varepsilon}{2},a_2+\frac{\varepsilon}{2},a_3,\ldots,a_n\right)\right)
\end{split}
\]
Take $g$ to be $\min\{f_1,\,f_2,\,f_3,\,f_4\}$. This ensures that both $h_1$ and $h_2$ are non-negative almost everywhere. Now we need to show $g$ is not 0 almost everywhere. Observe $g^{-1}\left(\{0\}\right)=\bigcup_{i=1}^4f_i^{-1}\left(\{0\}\right)$.
Therefore
\[
\begin{split}
\lambda^n\left(g^{-1}\left(\{0\}\right)\right)
&= \lambda^n \left(\bigcup_{i=1}^4f_i^{-1}\left(\{0\}\right)\right)\\[.25cm]
&\leq \lambda^n\left(f^{-1}\left(\{0\}\right)\cap S_{\utilde{a},\varepsilon} \right)\\[.25cm]
&< \frac{1}{4}\cdot\lambda^n\left( S_{\utilde{a},\varepsilon} \right) \mbox{\hspace{1cm}(by assumption)}\\[.25cm]
&= \frac{\varepsilon^n}{4}\\[.25cm]
&= \lambda^n\left(\left[0,\frac{\varepsilon}{2}\right]\times\left[0,\frac{\varepsilon}{2}\right] \times \left[0,\varepsilon\right]^{n-2}\right)\\[.25cm]
\end{split}
\]
Therefore $g$ is not $0$ almost everywhere. Hence $F$ can not be an extreme copula.
\end{proof}
\hspace{.25cm}
\begin{thm}
Probability induced by any extreme copula has to be singular with respect to Lebesgue measure.
\end{thm}
\begin{proof}
Let $C$ be a copula and $P$ be the probability induced by $C$. Suppose $P$ is not singular with respect to $\lambda^n$. Then there exists non-negative Borel measurable $f$ which is not 0 almost everywhere and a measure $\mu\perp\lambda^n$ such that the Lebesgue decomposition of $P$ is given by:
\[
P(A)=\int_A f\,d\lambda^n\, +\mu(A)
\]
So, there exists $ a,b $ with $0<a<b$ such that
\[
\alpha=\lambda^n\left(f^{-1}\left([a,b]\right)\right)>0
\]
Fix $\varepsilon\in (0,\alpha)$.
By Lusin's theorem, there exists a compact set $K$ with $\lambda^n(K)>1-\varepsilon$ such that $f$ restricted to $K$ is continuous. Therefore the set
$ f^{-1}\left([a,b]\right)\cap K $ is closed. Denote it by $K_0$. So, we have $\lambda^n\left(K_0\right) \geq\alpha-\varepsilon>0$ \\
Now draw grids of size $\frac{1}{m}$. Let $S_{\utilde{c_1},\frac{1}{m}},\, S_{\utilde{c_2},\frac{1}{m}},\,\ldots,\, S_{\utilde{c_r},\frac{1}{m}}$ be the minimum possible closed squares in the grid required to cover $K_0$.\\
Define $A_\xi=\left\{ \utilde{x} \mid d( \utilde{x},A )\leq \xi \right\}$ where $d$ is the Euclidean distance. Then it can be easily shown that ${K_0}_{\frac{\sqrt{n}}{m}}\supseteq \bigcup_{p=1}^r S_{\utilde{c_p},\frac{1}{m}}$. Observe as $m\rightarrow \infty$, ${K_0}_{\frac{\sqrt{n}}{m}}\downarrow K_0$ as $K_0$ is closed. Therefore
\[\lim\limits_{m\rightarrow\infty}\lambda^n\left({K_0}_{\frac{\sqrt{n}}{m}}\cap f^{-1}\left\{0\right\}\right)=\lambda^n\left( K_0\cap f^{-1}\left\{0\right\}\right)\Rightarrow \lim\limits_{m\rightarrow\infty}\lambda^n\left(\bigcup_{p=1}^r S_{\utilde{c_p},\frac{1}{m}}\cap f^{-1}\left\{0\right\}\right)=0
\]
So we can choose $m$ large enough such that
\[
\begin{split}
\lambda^n\left[f^{-1}\left( \{0\} \right)\cap \left( \bigcup_{p=1}^r S_{\utilde{c_p},\frac{1}{m}} \right) \right]
&< \frac{\alpha-\varepsilon}{4}\\[.25cm]
&\leq \frac{1}{4}\cdot \lambda^n \left(K_0\right)\\[.25cm]
&\leq \frac{1}{4}\cdot \lambda^n\left( \bigcup_{p=1}^r S_{\utilde{c_p},\frac{1}{m}} \right)
\end{split}
\]
So, $\exists\, q\in\left\{1,2,\ldots,r\right\}$ such that
\[
\lambda^n\left[f^{-1}\left( \{0\} \right)\cap S_{\utilde{c_q},\varepsilon} \right] \geq\frac{1}{4}\cdot\lambda^n\left[ S_{\utilde{c_q},\varepsilon} \right]
\]
And hence by \hyperlink{lemma.5.1}{Lemma 5.1}, $C$ can not be an extreme copula.
\end{proof}
\hspace*{3mm} In the following 3-dimensional figure denote $L_1$ to be the line joining $ (0,0,\frac{1}{2}) $ and $ (\frac{1}{2},\frac{1}{2},1) $, $L_2$ to be the line joining $ (\frac{1}{2},\frac{1}{2},\frac{1}{2}) $ and $ (1,1,1) $, $L_3$ to be the line joining $ (0,\frac{1}{2},0) $ and $ (\frac{1}{2},1,\frac{1}{2}) $ and $L_4$ to be the line joining $ (\frac{1}{2},0,0) $ and $ (1,\frac{1}{2},\frac{1}{2}) $. Define a probability $P$ as $P=\sum_{i=1}^{4}\mathbb{U}(L_i)$. It can be easily shown that c.d.f. of $P$ is an extreme copula. Clearly it satisfies the necessary condition but there does not exist any $i\in\{1,2,3\}$, $B\in \mathcal{B}(\left[0,1\right]) $ and Borel measurable functions $f : \left[0,1\right]\mapsto\left[0,1\right]^2$ such that $\pi_i^{-1}(B)\cap(\cup_{i=1}^4 L_i)=\pi_i^{-1}(B)\cap \mathcal{G}_f^{(i)}$. Hence it does not satisfy the sufficient condition.
\begin{center}
\includegraphics[scale=.8]{nes_suff}
\end{center}
\hspace*{3mm}As it has been shown above, the necessary condition and sufficient condition that we have postulated do not necessarily hold together and hence cannot be melded to give a complete characterization of extreme copulas. The initiative to find such a result that combine both might be undertaken as a future research problem.
\section{A construction of dense set using extreme copulas}
\hspace*{3mm}We are going to show that any copula can be uniformly approximated by extreme copulas. In fact, any copula can be uniformly approximated by permutation copulas. For this we will need the following definitions.\\
\begin{defn}
A $n$-dimensional matrix of order $m_1\times m_2\times\cdots\times m_n$ is a real valued function on $\{0,1,\ldots,m_1-1\}\times\{0,1,\ldots,m_2-1\}\times\cdots\times\{0,1,\ldots,m_n-1\}$
\end{defn}
\begin{defn}
A $n$-dimensional $m\times m\times\cdots\times m$ matrix $M$ is called stochastic in $k$th co-ordinate if for all $i_k\in\{0,1,\ldots,m-1\}$ \[\mathop{\sum\limits_{(r_1,r_2,\ldots,r_n)}}_{r_k=i_k} M(r_1,r_2,\ldots,r_n) =\frac{1}{m} \]
\end{defn}
\hspace*{3mm}Let us denote by $\mathcal{M}_m^n$ the set of all $n$-dimensional $m\times m\times\cdots\times m$ matrices with non-negative entries which are stochastic in every co-ordinate.\\
\begin{thm}
Any copula can be uniformly approximated by permutation copulas.
\end{thm}
\begin{proof}
Let $C$ be any copula and $P$ be the probability induced by $C$. Draw grids of size $\frac{1}{m}$ in $\left[0,1\right]^n$. Define a $n$-dimensional $m\times m\times\cdots\times m$ matrix $M$ as
\[ M(i_1,i_2,\ldots,i_n) = P\left(S_{\left(\frac{i_1}{m},\frac{i_2}{m},\ldots,\frac{i_n}{m}\right),\frac{1}{m}}\right) \]
Clearly $M\in\mathcal{M}_m^n $. Define a metric $\rho$ on $\mathcal{M}_m^n$ as for all $M_1,M_2\in\mathcal{M}_m^n$
\[\rho\left(M_1,M_2 \right)= \sup_{\left(i_1,i_2,\ldots,i_n\right)}
\abs{ M_1\left(i_1,i_2,\ldots,i_n\right)- M_2\left(i_1,i_2,\ldots,i_n\right)} \]
\hspace*{3mm}Now get $N\in\mathcal{M}_m^n $ with all rational entries such that $\rho\left(N,M\right)<\varepsilon=m^{-(n+1)}$. We are going to construct a permutation copula $\bar{C}_m$ with induced probability $\bar{P}_m$ such that \\
\[ \bar{P}_m\left(S_{\left(\frac{i_1}{m},\frac{i_2}{m},\ldots,\frac{i_n}{m}\right),\frac{1}{m}}\right) = N(i_1,i_2,\ldots,i_n) \]\\
If we construct it then we will have \\
\[
\begin{split}
&\abs{C\left(\frac{i_1}{m},\frac{i_2}{m},\ldots,\frac{i_n}{m}\right) - \bar{C}_m\left(\frac{i_1}{m},\frac{i_2}{m},\ldots,\frac{i_n}{m}\right) }\\[.25cm]
&\leq \sum\limits_{(r_1,r_2,\ldots,r_n)} \abs{ P\left(S_{\left(\frac{r_1}{m},\frac{r_2}{m},\ldots,\frac{r_n}{m}\right),\frac{1}{m}}\right) - \bar{P}_m\left(S_{\left(\frac{r_1}{m},\frac{r_2}{m},\ldots,\frac{r_n}{m}\right),\frac{1}{m}}\right) }\\[.25cm]
&= \sum\limits_{(r_1,r_2,\ldots,r_n)} \abs{ M(r_1,r_2,\ldots,r_n) - N (r_1,r_2,\ldots,r_n) }\\[.25cm]
&< m^n\varepsilon\\[.25cm]
&= \frac{1}{m}
\end{split}
\]\\
and therefore for all $\utilde{t}\in \left[0,1\right]^n$\\
\[
\begin{split}
& \abs{C(\utilde{t})-\bar{C}_m(\utilde{t})}\\[.25cm]
&\leq \,\abs{ C(\utilde{t}) - C\left(\frac{\lfloor mt_1 \rfloor}{m},\ldots,\frac{\lfloor mt_n \rfloor}{m} \right) }+
\abs{ \bar{C}_m(\utilde{t}) - \bar{C}_m\left(\frac{\lfloor mt_1 \rfloor}{m},\ldots,\frac{\lfloor mt_n \rfloor}{m} \right) }\\[.25cm]
& \mbox{\hspace{2.6cm}}+ \abs{ {C}\left(\frac{\lfloor mt_1 \rfloor}{m},\ldots,\frac{\lfloor mt_n \rfloor}{m} \right) - \bar{C}_m\left(\frac{\lfloor mt_1 \rfloor}{m},\ldots,\frac{\lfloor mt_n \rfloor}{m} \right) }\\[.25cm]
&\leq \frac{n}{m}+\frac{n}{m}+\frac{1}{m}\\[.25cm]
&= \frac{2n+1}{m}
\end{split}
\]\\
This will imply $\bar{C}_m$ converges to $C$ uniformly as $m\rightarrow\infty$. Hence $C$ can be uniformly approximated by permutation copulas. So, the only thing remains is to construct such a permutation copula $\bar{C}_m$.\\
\hspace*{3mm}To do this, for all $(i_1,i_2,\ldots,i_n)$ we want to have an $n$-dimensional cube which is a subset of $S_{\left(\frac{i_1}{m},\frac{i_2}{m},\ldots,\frac{i_n}{m}\right),\frac{1}{m}}$ and length of whose edges is same as $N(i_1,i_2,\ldots,i_n)$ such that the co-ordinate projections of those cubes will be disjoint. Then if we have a permutation copula which gives probability $N(i_1,i_2,\ldots,i_n)$ to the cube that is subset of $S_{\left(\frac{i_1}{m},\frac{i_2}{m},\ldots,\frac{i_n}{m}\right),\frac{1}{m}}$ we are done. See the figure for $n=2$\\
\begin{center}
\scalebox{.5}{\includegraphics{dense}}
\end{center}
\hspace{.25cm}
Observe for all $k$
\[ \lambda\left(\left[ \frac{j}{m},\frac{j+1}{m} \right) \right) = \frac{1}{m}= \mathop{\sum\limits_{(r_1,r_2,\ldots,r_n)}}_{r_k=j} N(r_1,r_2,\ldots,r_n) \]
So we can divide $\left[ \frac{j}{m},\frac{j+1}{m} \right)$ into $m^{n-1}$ disjoint intervals as
\[\left[ \frac{j}{m},\frac{j+1}{m} \right) = \mathop{\bigcup\limits_{(r_1,r_2,\ldots,r_n)}}_{r_k=j} I^{(k)}_{(r_1,r_2,\ldots,r_n)} \]
such that \[\lambda\left(I^{(k)}_{(r_1,r_2,\ldots,r_n)} \right) =N(r_1,r_2,\ldots,r_n) \] Observe \[S_{\left(\frac{i_1}{m},\frac{i_2}{m},\ldots,\frac{i_n}{m}\right),\frac{1}{m}}\supseteq \prod_{k=1}^n I^{(k)}_{(i_1,i_2,\ldots,i_n)} \]
Let $L_{(i_1,i_2,\ldots,i_n)}$ be any one of the interior diagonals of $\overline{\left(\prod_{k=1}^{n}I^{(k)}_{(i_1,i_2,\ldots,i_n)}\right)} $. Define $\bar{P}_m$ as
\[ \bar{P}_m=\sum_{(i_1,i_2,\ldots,i_n)} N(i_1,i_2,\ldots,i_n)\cdot \mathbb{U}\left( L_{(i_1,i_2,\ldots,i_n)} \right) \]
Clearly
\[ \bar{P}_m\left(S_{\left(\frac{i_1}{m},\frac{i_2}{m},\ldots,\frac{i_n}{m}\right),\frac{1}{m}}\right) = N(i_1,i_2,\ldots,i_n) \]
and $\bar{C}_m$ is a copula. As $N(i_1,i_2,\ldots,i_n)$ are rationals then $\exists q\in \mathbb{N}$ such that for all $k$ for all $(i_1,i_2,\ldots,i_n) \in \{0,1,\ldots,m-1\}^n$, $\lambda\left(I^{(k)}_{(r_1,r_2,\ldots,r_n)} \right)$ is a integer multiple of $\frac{1}{q}$. Hence $\bar{C}_m$ is a permutation copula of order $q$.
\end{proof}
\hspace*{3mm}So we obtain that permutation copulas are dense in the space of copulas. This is an example of a convex set containing more than one point whose extreme points are dense in that set. As a corollary we got that the set of extreme copulas are not closed under $L_{\infty}$ norm.
\section{Application}
\hspace*{3mm} In a situation where we need to study the influence of the dependence structure on a statistical problem with given marginals of some random vector, we consider an optimization problem over the Fr\'{e}chet class $\mathcal{F}(F_1, F_2,\ldots,F_n)$ of all joint distributions with the marginals $F_1, F_2,\ldots,F_n$ (see, e.g. \hyperlink{ref8}{[8]}). For a given bounded continuous function $g:\mathbb{R}^n\mapsto\mathbb{R}$, an optimization problem over the Fr\'{e}chet class $\mathcal{F}(F_1, F_2,\ldots,F_n)$ looks like
\[ m(g):= \sup\left\{\int g\,dF : F\in \mathcal{F}(F_1, F_2,\ldots,F_n) \right\} \]
Clearly we have
\[
\begin{split}
m(g)&= \sup\left\{\mathbb{E}\left( g(\utilde{X}) \right) : \utilde{X}\sim F\in \mathcal{F}(F_1, F_2,\ldots,F_n) \right\}\\
&= \sup\left\{\mathbb{E}\left( g\circ (F_1, F_2,\ldots,F_n)^{-1}(\utilde{U}) \right) : \utilde{U}\sim C \mbox{, where $C$ is a copula} \right\}\\
&= \lim\limits_{k\rightarrow\infty} \max\left\{\mathbb{E}\left( g\circ (F_1, F_2,\ldots,F_n)^{-1}(\utilde{U}) \right) : \utilde{U}\sim C_k \mbox{, a permutation copula of order $k$} \right\}
\end{split}
\]
where $(F_1, F_2,\ldots,F_n)^{-1}(\utilde{U}) = (F_1^{-1}(U_1), F_2^{-1}(U_2), \ldots,F_n^{-1}(U_n))$.\\
\hspace*{3mm} In special cases in which we need to maximize probability of the event $\left\{X=Y\right\}$ (i.e., $g=1_{\left\{x=y\right\}}$) where $X,\,Y$ are random variables with distribution functions $F_X,\, F_Y$, we can construct a function $g_{\varepsilon}$ as
\[ g_{\varepsilon}(x,y)=\left\{ \begin{array}{ll}
1& \mbox{, if $x=y$}\\
1-\frac{|x-y|}{\varepsilon}&\mbox{, if $|x-y|<\varepsilon$}\\
0& \mbox{, otherwise}
\end{array} \right. \]
If $P_F$ is the probability induced by the joint distribution function $F$ we have
\[
\begin{split}
m(g)&=\sup\left\{P_F(X=Y) : F\in \mathcal{F}(F_X, F_Y) \right\}\\
&= \lim\limits_{\varepsilon\rightarrow 0}\lim\limits_{k\rightarrow\infty} \max\left\{\mathbb{E}\left( g_{\varepsilon}\circ (F_X, F_Y)^{-1}(\utilde{U}) \right) : \utilde{U}\sim C_k \mbox{, a permutation copula of order $k$} \right\}
\end{split}
\]
\hspace*{3mm} We can use these results to simulate in computer the approximate value of $m(g)$ to solve the optimization problem.
\begin{description}
\item[[\hypertarget{ref1}{1}]] \textsc{R. B. Ash}: Probability and Measure Theory, 2nd edn. Harcourt Academic Press, 2000.
\item[[\hypertarget{ref2}{2}]] \textsc{Dunford, Schwartz}: Linear Operators, Part I: General Theory, Wiley, 1958.
\item[[\hypertarget{ref3}{3}]] \textsc{George F. Simmons}: Introduction to Topology and Modern Analysis, Robert E Krieger Publishing Company, 1983.
\item[[\hypertarget{ref4}{4}]] \textsc{Barry Simon}: Convexity: an Analytic Viewpoint (Cambridge Tracts in Mathematics 187), Cambridge University Press, 2011.
\item[[\hypertarget{ref5}{5}]] \textsc{Fabrizio Durante, Carlo Sempi}: Principles of Copula Theory. CRC Press, 2015.
\item[[\hypertarget{ref6}{6}]] \textsc{Fabrizio Durante, J. Fern\'{a}ndez S\'{a}nchez, W. Trutschnig}: On the Singular Components of a Copula, J. Appl. Prob. 52, 1175-1182, 2015.
\item[[\hypertarget{ref7}{7}]] \textsc{J.-F. Mai, M. Scherer}: Simulating from the copula that generates the maximal probability for a
joint default under given (inhomogeneous) marginals. In Topics in Statistical Simulation, eds V. B. Melas et al.,
Springer, New York, pp. 333-341, 2014.
\item[[\hypertarget{ref8}{8}]] \textsc{Giovanni Puccetti, Ruodu Wang}: Extremal Dependence Concepts. Statistical Science, vol.30, No.4, 485-517, 2015.
\item[[\hypertarget{ref9}{9}]] \textsc{P. Jaworski, F. Durante, W. K. H\"{a}rdle}: Copulae in Mathematical and Quantitative Finance (Lecture Notes in Statistics 213), Springer, 2013.
\item[[\hypertarget{ref10}{10}]] \textsc{P. Jaworski, F. Durante, W. K. H\"{a}rdle, T. Rychlik}: Copula Theory and Its Applications (Lecture Notes in Statistics 198), Springer, 2010.
\end{description}
\end{document}
|
arXiv
|
{
"id": "1709.02472.tex",
"language_detection_score": 0.5972839593887329,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{frontmatter}
\title{Boundary Lipschitz Regularity and the Hopf Lemma on Reifenberg Domains for Fully Nonlinear Elliptic Equations \tnoteref{t1}}
\author[rvt]{Yuanyuan Lian} \ead{[email protected]; [email protected]} \author[rvt]{Wenxiu Xu\corref{cor1}} \ead{[email protected]} \author[rvt]{Kai Zhang} \ead{zhang\[email protected]; [email protected]} \tnotetext[t1]{This research is supported by the National Natural Science Foundation of China (Grant No. 11701454) and the Natural Science Basic Research Plan in Shaanxi Province of China (Program No. 2018JQ1039).}
\cortext[cor1]{Corresponding author}
\address[rvt]{Department of Applied Mathematics, Northwestern Polytechnical University, Xi'an, Shaanxi, 710129, PR China}
\begin{abstract} In this paper, we prove the boundary Lipschitz regularity and the Hopf Lemma by a unified method on Reifenberg domains for fully nonlinear elliptic equations. Precisely, if the domain $\Omega$ satisfies the exterior Reifenberg $C^{1,\mathrm{Dini}}$ condition at $x_0\in \partial \Omega$ (see \Cref{d-re}), the solution is Lipschitz continuous at $x_0$; if $\Omega$ satisfies the interior Reifenberg $C^{1,\mathrm{Dini}}$ condition at $x_0$ (see \Cref{d-H-re}), the Hopf lemma holds at $x_0$. Our paper extends the results under the usual $C^{1,\mathrm{Dini}}$ condition. \end{abstract}
\begin{keyword} Boundary regularity \sep Lipschitz continuity \sep Hopf lemma \sep Fully nonlinear elliptic equation \sep Reifenberg domain
\end{keyword}
\end{frontmatter}
\section{Introduction}\label{S1}
In this paper, we intend to obtain the pointwise boundary Lipschitz regularity and prove the Hopf Lemma for the viscosity solutions of the following fully nonlinear elliptic equations \begin{equation}\label{r-e} \left\{\begin{aligned} &u\in S(\lambda,\Lambda,f)&& ~~\mbox{in}~~\Omega;\\ &u=g&& ~~\mbox{on}~~\partial \Omega, \end{aligned}\right. \end{equation} where $\Omega$ is a bounded domain and $S\left( {\lambda ,\Lambda ,f} \right)$ denotes the Pucci class with uniform constants $\lambda$ and $\Lambda$ (see \cite{MR1351007} for the definition and its basic properties).
It is well known that the exterior sphere condition and the interior sphere condition imply the boundary Lipschitz regularity and the Hopf lemma respectively. In recent decades, sphere condtion has been extended to a more general geometrical condition, i.e., the $C^{1,\mathrm{Dini}}$ condition (see \Cref{r-2}). With aid of the boundary Harnack inequality, Safonov \cite{Safonov2008} proved the boundary Lipschitz regularity under the exterior $C^{1,\mathrm{Dini}}$ condition, and the Hopf lemma under the interior $C^{1,\mathrm{Dini}}$ condition for classical solutions of linear elliptic equations in nondivergence form. Huang, Li and Wang \cite{MR3167627} also obtained the boundary Lipschitz regularity for linear elliptic equations under the exterior $C^{1,\mathrm{Dini}}$ condition. They used an auxiliary barrier function and the iteration technique, without using the boundary Harnack inequality. Lieberman \cite{MR779924} proved the Hopf lemma for linear elliptic equations under the interior $C^{1,\mathrm{Dini}}$ condition by applying the regularized distance. Recently, Lian and Zhang \cite{lian2018boundary} extend above results to fully nonlinear elliptic equations by a unified method. Moreover, the proof is simple. In that paper, curved boundaries were regarded as the perturbation of a hyperplane. Then the desired regularity can be obtained by a perturbation argument.
In this paper, we prove the boundary Lipschitz regularity and the Hopf lemma for viscosity solutions of fully nonlinear elliptic equations under the Reifenberg $C^{1,\mathrm{Dini}}$ condition which extends the $C^{1,\mathrm{Dini}}$ condition. We use an improved technique of \cite{lian2018boundary} to derive our results. It can be shown that the directions at different scales converges to a direction, say $e_n$. The main difficulty is that we can not estimate the derivative of the solution along $e_n$ directly, which is carried out in and the difference between two adjoint scales has to be estimated. The Reifenberg $C^{1,\mathrm{Dini}}$ condition was introduced by Ma, Moreira and Wang \cite{MR1351008}where the boundary $C^{1}$ regularity was obtained for fully nonlinear parabolic equations. Note that the Reifenberg ${C^{1,\mathrm{Dini}}}$ condition is more general than the ${C^{1,\mathrm{Dini}}}$ condition (see \Cref{d-re} and \Cref{d-H-re}).
Before the statement of our main results, we introduce some standard notations and definitions. Let $B_r(x_0)$ denote the open ball in $R^n$ with center $x_0$ and radius $r$. Set $B_r=B_r(0)$, $B_r^+=B_r\cap \left\{x|x_n>0\right\}$ and $T_r=B_r\cap \left\{x|x_n=0\right\}$. Denote by $Q_r(x_0)$ the open cube in $R^n$ with center $x_0$ and side-length $r$. Set $Q_r=Q_r(0)$ and $Q_r^+=Q_r\cap \left\{x|x_n>0\right\}$. In this paper, $\left\{e_1,...,e_n\right\}$ stands for the standard basis in $R^n$.
\begin{definition}\label{d-Dini} The function $\omega:[0,+\infty)\rightarrow [0,+\infty)$ is called a Dini function if $\omega$ is nondecreasing and satisfies the following Dini condition for some $r_0>0$ \begin{equation}\label{e-dini}
\int_{0}^{r_0}\frac{\omega(r)}{r}dr<\infty. \end{equation} \end{definition}
\begin{definition}\label{d-Dini-f} Let $\Omega \subset R^{n}$ be a bounded domain and $f$ be a function defined on $\bar{\Omega}$. We say that $f$ is Lipschitz at $x_0\in \bar{\Omega}$ or $f\in C^{0,1}(x_0)$ if there exists a constant $C$ such that \begin{equation*}
|f(x)-f(x_0)|\leq C|x-x_0|,~~\forall~x\in \bar{\Omega}. \end{equation*}
Then, define $[f]_{C^{0,1}(x_0)}=\inf C$ where $C$ satisfies above equation, and $\|f\|_{C^{0,1}(x_0)}=\|f\|_{L^{\infty}(\Omega)}+[f]_{C^{0,1}(x_0)}$.
Similarly, we call that $f$ is $C^{1,\mathrm{Dini}}$ at $x_0$ or $f\in C^{1,\mathrm{Dini}}(x_0)$ if there exist a vector $l$ and a constant $C$ such that \begin{equation*}
|f(x)-f(x_0)-l\cdot (x-x_0)|\leq C|x-x_0|\omega(|x-x_0|),~~\forall~x\in \bar{\Omega}, \end{equation*}
where $\omega$ is a Dini function. Then we denote $l$ by $\nabla f(x_0)$. We define $[f]_{C^{1,\mathrm{Dini}}(x_0)}=\inf C$ where $C$ satisfies above equation, and $\|f\|_{C^{1,\mathrm{Dini}}(x_0)}=\|f\|_{L^{\infty}(\Omega)}+|l|+[f]_{C^{1,\mathrm{Dini}}(x_0)}$. \end{definition}
Now, we give the definitions of the Reifenberg $C^{1,\mathrm{Dini}}$ condition. \begin{definition}[\textbf{Exterior Reifenberg $C^{1,\mathrm{Dini}}$ condition}]\label{d-re}
We say that $\Omega$ satisfies the exterior Reifenberg ${C^{1,\mathrm{Dini}}}$ condition at $0 \in \partial \Omega$ if there exist a positive constant $R$ and a Dini function $\omega_{\Omega}(r)$ such that
a) for any $0 < r < R,$ there exists a hyperplane ${\Gamma _r}$ and its unit normal vector $n_{\Gamma _r}$ such that \begin{equation}\label{e-re} {B_r} \cap \Omega \subset {B_r} \cap \left\{ {{x \cdot {n_{{\Gamma _r}}}} > - r\omega_{\Omega}(r)} \right\}. \end{equation}
b) $\left| n_{\Gamma _r}-n_{\Gamma _{\theta r}} \right| \leq {K(\theta) }\omega_{\Omega}(r)$ for each $ 0<\theta <1$ and $0<r<R$, where the nonnegative function $K$ depends only on $\theta$ and is bounded on $[\theta _0,1]$ for any $0<{\theta _0}<1$. \end{definition}
\begin{definition}[\textbf{Interior Reifenberg $C^{1,\mathrm{Dini}}$ condition}]\label{d-H-re}
We say that $\Omega$ satisfies the interior Reifenberg ${C^{1,\mathrm{Dini}}}$ condition at $0 \in \partial \Omega$ if there exist a positive constant $R$ and a Dini functions $\omega_{\Omega}(r)$ such that
a) for any $0 < r < R,$ there exists a hyperplane ${\Gamma _r}$ and its unit normal vector $n_{\Gamma _r}$ such that \begin{equation}\label{e-re-2} {B_r} \cap \Omega^c\subset {B_r} \cap \left\{ {{x \cdot {n_{{\Gamma _r}}}} < r\omega_{\Omega}(r)} \right\} . \end{equation}
b) $\left| n_{\Gamma _r}-n_{\Gamma _{\theta r}} \right| \leq {K(\theta) }\omega_{\Omega}(r)$ for each $ 0<\theta <1$ and $0<r<R$. \end{definition}
\begin{remark}\label{r-1} If $\Omega$ satisfies both the exterior and interior Reifenberg ${C^{1,\mathrm{Dini}}}$ condition at $0$, we call that $\partial\Omega$ is Reifenberg ${C^{1,\mathrm{Dini}}}$ at $0$. Without loss of generality, we always assume that $K\geq 1$. \end{remark}
\begin{remark}\label{r-2} If $\Gamma_r$ and $n_{\Gamma_r}$ are the same for different $r$, we arrive at the definition of the usual ${C^{1,\mathrm{Dini}}}$ conditions (see \cite[Definition 1.2 and Definition 1.3]{lian2018boundary}). The Reifenberg ${C^{1,\mathrm{Dini}}}$ condition is more general than the usual ${C^{1,\mathrm{Dini}}}$ condition, which is shown clearly by the following example adopted from \cite{MR1351008}.
Let $\Omega = {B_1} \cap \left\{ {y > f\left( x \right)} \right\}\subset R^2$ where $f\left( x \right) ={x}/ {\ln \left| x \right|},x \in \left(-1/2,1/2 \right)$. Clearly, $\partial\Omega$ is not ${C^{1,\mathrm{Dini}}}$ at $0$ since $1/|\ln r|$ is not a Dini function. Now we show that it is Reifenberg ${C^{1,\mathrm{Dini}}}$ at $0$.
For any $0<r<1/2$, let $\Gamma_r$ be the line $\left\{(x,y)| y=x/\ln r\right\}$ and take $n_{\Gamma_r}=(-1/\ln r, 1)$. It is easy to see that there exists a unique $-r<x^*<0$ satisfying $f'\left( {{x^*}} \right) = {1}/{\ln r}$, and for any $(x,y)\in B_r\cap \Omega$,
\begin{equation*}
(x,y)\cdot n_{\Gamma_r}> (x^*,f(x^*))\cdot n_{\Gamma_r}= \frac{x^*}{\ln |x^*|} - \frac{x^*}{\ln r}. \end{equation*} Note that \begin{equation*}
f'(x^*)=\frac{1}{\ln |x^*|}-\frac{1}{\ln^2 |x^*|}=\frac{1}{\ln r}. \end{equation*} I.e., \begin{equation*}
\frac{x^*}{\ln r}=\frac{x^*}{\ln |x^*|}-\frac{x^*}{\ln^2 |x^*|}. \end{equation*} Hence, \begin{equation*}
(x,y)\cdot n_{\Gamma_r}> \frac{x^*}{\ln^2 |x^*|}\geq \frac{-r}{\ln^2r}. \end{equation*} In addition, \begin{equation*}
|n_{\Gamma_r}-n_{\Gamma_{\theta r}}| \leq \left| {\frac{1}{{\ln r}} - \frac{1}{{\ln \left( {\theta r} \right)}}} \right| \leq \left| {\frac{{\ln \theta }}{{\ln r\ln \left( {\theta r} \right)}}} \right| \leq \left| \frac{\ln \theta }{ \ln^2 r} \right|. \end{equation*}
Since $1/\ln^2r$ is a Dini function, $\partial\Omega$ satisfies the exterior Reifenberg ${C^{1,\mathrm{Dini}}}$ condition at $0$. Similarly, it can be verified that $\partial\Omega$ also satisfies the interior Reifenberg ${C^{1,\mathrm{Dini}}}$ condition at $0$. \end{remark}
Now, we state our main results. For the boundary Lipschitz regularity, we have \begin{theorem}\label{t-2} Suppose that $\Omega$ satisfies the exterior Reifenberg $C^{1,\mathrm{Dini}}$ condition at $0\in \partial \Omega$ for some Dini function $\omega_{\Omega}(r)$ and $R>0$. Let $u$ be a viscosity solution of \begin{equation} \left\{\begin{aligned} &u\in S(\lambda,\Lambda,f)&& ~~\mbox{in}~~\Omega;\\ &u=g&& ~~\mbox{on}~~\partial \Omega, \end{aligned}\right. \end{equation} where $g$ is $C^{1,\mathrm{Dini}}$ at $0$ with a Dini function $\omega_g$ and $f\in L^{n}(\Omega)$ satisfies \begin{equation}\label{e-dini-f}
\int_{0}^{R}\frac{\omega_f(r)}{r}dr:= \int_{0}^{R}\frac{\|f\|_{L^{n}(\Omega\cap B_r)}}{r\|f\|_{L^{n}(\Omega)}}dr<\infty. \end{equation}
Then $u$ is $C^{0,1}$ at $0$ and \begin{equation*}
|u(x)-u(0)|\leq C |x|\left(\|u\|_{L^{\infty }(\Omega)}+\|f\|_{L^{n}(\Omega)}+[g]_{C^{1,\mathrm{Dini}}(0)}\right), ~~\forall ~x\in \Omega\cap B_{R}, \end{equation*} where $C$ depends only on $n, \lambda, \Lambda, K,\omega_{\Omega},\omega_{f},\omega_{g}$ and $R$. \end{theorem}
For the Hopf lemma, we have \begin{theorem}[\textbf{Hopf lemma}]\label{t-H-2} Suppose that $\Omega$ satisfies the interior Reifenberg $C^{1,\mathrm{Dini}}$ condition at $0\in \partial \Omega$ for some Dini function $\omega_{\Omega}(r)$ and $R>0$. Let $u\in C(\bar{\Omega})$ satisfy \begin{equation} M^{-}(D^2u,\lambda,\Lambda)\leq 0 ~~\mbox{in}~~\Omega~~ (\mathrm{i.e.,}~~ u\in \bar{S}(\lambda,\Lambda,0)) \end{equation} with $u(0)=0$ and $u\geq 0$ in $\Omega$.
Then there exists a unit vector, say $e_n$, such that for any $l\in R^n$ with $|l|=1$ and $l\cdot {e_n}>0$, \begin{equation}\label{e-H-main}
u(tl)\geq cu\left(\frac{e_n}{2}\cdot R \right)t, ~~\forall~ 0<t<\delta, \end{equation} where $c>0$ and $\delta>0$ depend only on $n, \lambda, \Lambda,K,\omega_{\Omega},R$ and $l$. \end{theorem}
\section{Proofs of the main results} In this section, we give the detailed proofs of the main results.
Now, we clarify the idea briefly. Firstly, note that if the $\Omega$ satisfies the Reifenberg ${C^{1,\mathrm{Dini}}}$ condition from the exterior (or the interior) at $0\in \partial\Omega$, the normal vectors in different scales converges. In addition, the difference of unit normal vectors in different scales is controlled by the Dini function. Next, we use solutions with flat boundaries (i.e., $v$ in the proofs) to approximate the solution $u$. Then the error between $u$ and $v$ (i.e., $w$ in the proofs) can be estimated by maximum principles. By an iteration argument, the boundary regularity for $u$ is obtained. For the boundary Lipschitz regularity, the right hand function $f$, the boundary value $g$ and the curved boundary $\partial \Omega$ are regarded perturbations of $0$, $0$ and a hyperplane (see the definition of $v$ in the proof) which are inspired directly by \cite{MR3780142}. For the Hopf lemma, since the solution is nonnegative and the equation has the right hand zero, it is easier to prove.
First, we prove the following simple lemma. \begin{lemma}\label{t-1}
Suppose that $\left| n_{\Gamma _r}-n_{\Gamma _{\theta r}} \right| \leq {K(\theta) }\omega_{\Omega}(r)$ for each $0 < \theta < 1$ and $0<r<R$. Then there exists a unique unit vector $n_0$ such that \begin{equation} \mathop {\lim }\limits_{r \to 0 } {{n_{{\Gamma _r}}}} = {n_0} . \end{equation} \end{lemma} \proof For any $0<\eta<1$ and $l,m\in N^+$, we have \begin{equation*}
| {{n_{m+l}} - {n_{m}}}| \leq {K(\eta) }\sum_{i=m}^{m+l-1}\omega_{\Omega}(\eta^i) , \end{equation*} where $n_k$ denotes ${n_{{\Gamma _{{\eta ^k}}}}}$ for convenience. Since $\omega_{\Omega}\left( r \right)$ is a Dini function, there exists a unit vector $n_0$ satisfying \begin{equation*} \mathop {\lim }\limits_{k \to \infty } {{n_{k}} = n_0}. \end{equation*}
Now, for any $\varepsilon>0$, there exists $k_0\geq 0$ such that $|n_k-n_0|\leq \varepsilon/2$ for any $k\geq k_0$ and $\sup_{\theta\in[\eta,1]}K(\theta)\cdot \omega_{\Omega}(\eta^{k_0})\leq \varepsilon/2$. Then for any $0<r<R\eta^{k_0}$, there exists $k\geq k_0$ such that $R\eta^{k+1}\leq r<R\eta^{k}$. Then for some $\eta\leq \theta<1$, \begin{equation*}
|n_{\Gamma_r}-n_0|\leq |n_{\Gamma_r}-n_k+n_k-n_0|\leq K(\theta)\omega_{\Omega}(\eta^{k})+\varepsilon/2\leq \varepsilon. \end{equation*} Therefore, $\lim_{r \to 0}n_{\Gamma_r} = n_0$.~\qed ~\\
\begin{remark}\label{r-3} Without loss of generality, we always assume that $n_0=e_n$ throughout this paper. \end{remark}
Next, we introduce the following lemma, which concerns the boundary $C^{1,\alpha}$ regularity for solutions with flat boundaries. It was first proved by Krylov \cite{MR688919} and further simplified by Caffarelli (see \cite[Theorem 9.31]{MR1814364} and \cite[Theorem 4.28]{MR787227}). We will use the solutions in this lemma to approximate the solutions in \Cref{t-2} and \Cref{t-H-2}. \begin{lemma}\label{l-1} Let $u$ satisfy \begin{equation} \left\{\begin{aligned} &u\in S(\lambda,\Lambda,0)&& ~~\mbox{in}~~B_1^+;\\ &u=0&& ~~\mbox{on}~~T_1. \end{aligned}\right. \end{equation}
Then $u$ is $C^{1,\alpha}$ at $0$ and \begin{equation*}
|u(x)-u(0)-ax_n|\leq C |x|^{1+\alpha}\|u\|_{L^{\infty }(B_1^+)}, ~~\forall ~x\in B_{1/2}^+ \end{equation*} with \begin{equation*}
|a|\leq C\|u\|_{L^{\infty }(B_1^+)}, \end{equation*} where $\alpha$ and $C$ depend only on $n, \lambda$ and $\Lambda$. \end{lemma}
Now, we give the
\noindent\textbf{Proof of \Cref{t-2}.} Let $\omega(r)=\max \left\{\omega_{\Omega}(r),\omega_g(r),\omega_f(r)\right\}$. From the Dini condition, there exists $r_1>0$ such that for any $0<r\leq r_1$, \begin{equation}\label{e-H-dini-2}
\omega(r)\leq c_0 ~~\mbox{and}~~ \int_{0}^{r}\frac{\omega(s)}{s}ds\leq c_0, \end{equation} where $c_0\leq 1/4$ is a small constant to be specified later and depends only on $n,\lambda,\Lambda$ and $K$. By a proper scaling, we assume that $r_1=1$. Furthermore, we assume that $u(0)=g(0)=0$ and $\nabla g(0)=0$. Otherwise, we may consider $v:=u-g(0)-\nabla g(0)\cdot x$, which satisfies the same equation.
Let $M=\|u\|_{L^{\infty }(\Omega)}+\|f\|_{L^{n}(\Omega)}+[g]_{C^{1,\mathrm{Dini}}(0)}$ and $\Omega _{r}=\Omega \cap B_{r}$. To prove that $u$ is $C^{0,1}$ at $0$, we only need to prove the following:
There exist constants $0<\alpha_{0}, \eta < 1,\bar{C}$ (depending only on $n$, $\lambda$, $\Lambda$), $\hat{C}$ (depending only on $n,\lambda,\Lambda$ and $K$) and a nonnegative sequence $\{a_k\}$ ($k\geq -1$) such that for all $k\geq 0$ \begin{equation}\label{e1.16} \sup_{\Omega _{\eta^k}}(u-a_k{n_{k}} \cdot x)\leq \hat{C} M \eta ^{k}A_k \end{equation} and \begin{equation}\label{e1.17}
|a_k-a_{k-1}|\leq \bar{C}\hat{C}MA_k, \end{equation} where \begin{equation}\label{e-Ak} n_k=n_{\Gamma_{\eta^k}},A_0=c_0, A_k=\max(\omega(\eta^{k}),\eta^{\alpha_0} A_{k-1}) (k\geq 1). \end{equation}
Indeed, from \cref{e-Ak}, we have for any $k\geq 1$, \begin{equation*}
A_k\leq \omega(\eta^k)+\eta^{\alpha_0}A_{k-1}. \end{equation*} Hence, \begin{equation*}
\sum_{i=0}^{k} A_i\leq \sum_{i=1}^{k}\omega(\eta^i)+\eta^{\alpha_0}\sum_{i=0}^{k} A_i+c_0, \end{equation*} which indicates \begin{equation}\label{e.n1} \begin{aligned} \sum_{i=0}^{k} A_i\leq \frac{1}{1-\eta^{\alpha_0}}\left(\sum_{i=1}^{\infty}\omega(\eta^i)+c_0\right)\leq C \end{aligned} \end{equation} for some constant $C$ independent of $k$. That is, $\sum_{i=0}^{\infty} A_i$ converges. Thus, $a_k$ converges to some constant $a$.
Then for any $r>0$, there exists $k\geq 0$ such that $\eta^{k+1}<r\leq \eta^{k}$. From \cref{e1.16} and \cref{e1.17}, we have \begin{equation*} \sup_{\Omega _{r}}u\leq \sup_{\Omega _{\eta^k}}u\leq \hat{C} M \eta ^{k}A_k+CM\eta^k\leq CM\eta^k\leq CMr, \end{equation*} where $C$ depends only on $n,\lambda,\Lambda$ and $K$. In addition, \begin{equation*} \inf_{\Omega _{r}}u\geq -C M r \end{equation*} can be proved similarly. Therefore, \begin{equation*}
\|u\|_{L^{\infty}(\Omega_{r})}\leq CMr. \end{equation*} That is, $u$ is $C^{0, 1}$ at $0$.
Now, we prove \cref{e1.16} and \cref{e1.17} by induction. For $k=0$, by setting $a_{-1}=a_0=0$, they hold clearly provided \begin{equation}\label{e.21} \hat{C}c_0\geq 1. \end{equation} Suppose that they hold for $k$. We need to prove that they hold for $k+1$.
For convenience, we use the following notations. Let $r=\eta ^{k}$, $B_{{\Gamma _r}}^ + =B_r\cap \left\{x|x\cdot{n_{{\Gamma _r}}}>0\right\}$ and ${T_{{\Gamma _r}}} = {B_r} \cap \left\{ {x\left| {x \cdot {n_{{\Gamma _r}}} = 0} \right.} \right\}$ where ${{\Gamma _r}}$ denotes a hyperplane depending only on $r$ and ${n_{{\Gamma _r}}}$ is the unit normal vector of ${{\Gamma _r}}$. We may also denote $n_{\Gamma _{\eta^k}}(n_{\Gamma _r})$ by $n_k$.
Since $\partial\Omega$ satisfies the exterior Reifenberg ${C^{1,\mathrm{Dini}}}$ condition at $0$, there exist a hyperplane ${\Gamma _r}$ and its unit normal vector $n_{\Gamma _r}$ such that $B_r\cap\Omega\subset B_r\cap\{x\cdot {n_{\Gamma _r}}>-r\omega(r)\}$. Let $\tilde B_{{\Gamma _r}}^ + = B_{{\Gamma _r}}^ + -K(\eta) r\omega \left( r \right) n_{\Gamma _r},$ $\tilde T_{{\Gamma _r}}^ += T_{{\Gamma _r}}^ + - K(\eta)r\omega \left( r \right) n_{\Gamma _r}$ and $\tilde{\Omega }_{r}=\Omega \cap \tilde B_{{\Gamma _r}}^ + $. Take \begin{equation}\label{e.K} c_0K(\eta)<\frac{1}{4}. \end{equation} Then $\omega(r) \leq \omega(1)\leq c_0\leq 1/(4K(\eta))$ and $\Omega _{\eta r}\subset\tilde{\Omega }_{r}$.
Let $v$ solve \begin{equation*} \left\{\begin{aligned}
&M^{+}(D^2v,\lambda,\Lambda)=0 &&\mbox{in}~~\tilde{B}^{+}_{{\Gamma _r}}; \\
&v=0 &&\mbox{on}~~\tilde{T}_{{\Gamma _r}};\\
&v=\hat{C} M \eta ^{k}A_k &&\mbox{on}~~\partial \tilde{B}^{+}_{{\Gamma _r}}\backslash \tilde{T}_{{\Gamma _r}}. \end{aligned} \right. \end{equation*} Let $w=u-a_k{n_{k}} \cdot x-v$. Then $w$ satisfies (note that $v\geq 0$ in $\tilde{B}^{+}_{{\Gamma _r}}$) \begin{equation*}
\left\{
\begin{aligned}
&w\in \underline{S}(\lambda /n,\Lambda , f) &&\mbox{in}~~ \Omega \cap \tilde{B}^{+}_{{\Gamma _r}}; \\
&w\leq g-a_k {n_{k}} \cdot x &&\mbox{on}~~\partial \Omega \cap \tilde{B}^{+}_{{\Gamma _r}};\\
&w\leq 0 &&\mbox{on}~~\partial \tilde{B}^{+}_{{\Gamma _r}}\cap \bar{\Omega}.
\end{aligned}
\right. \end{equation*}
In the following arguments, we estimate $v$ and $w$ respectively. By the boundary $C^{1,\alpha}$ estimate for $v$ (see \Cref{l-1}) and the maximum principle, there exist $0<\alpha<1$ (depending only on $n,\lambda$ and $\Lambda$) and $\bar{a}\geq 0$ such that \begin{equation*} \begin{aligned}
\|v-\bar{a}({n_{k}}\cdot x+K(\eta)r\omega \left( r \right){n_{k}})\|_{L^{\infty }(\Omega _{\eta r})}&\leq C_1\frac{(\eta r)^{1+ \alpha}}{r^{1+ \alpha}}\|v\|_{L^{\infty }( \tilde{B}^{+}_{\Gamma_r})}\\
&\leq C_1\eta ^{\alpha-\alpha_0 }\cdot \hat{C}M\eta ^{(k+1)}\eta^{\alpha_0}A_k\\
&\leq C_1\eta ^{\alpha-\alpha_0 }\cdot \hat{C}M\eta ^{(k+1)}A_{k+1} \end{aligned} \end{equation*} and \begin{equation}\label{e.19} 0\leq\bar{a}\leq C_2\hat{C}MA_k, \end{equation} where $C_1$ and $C_2$ depend only on $n,\lambda$ and $\Lambda$. Take $\alpha_0=\alpha/2$ and then \begin{equation}\label{e.20} \begin{aligned}
\|v-\bar{a} n_{k}\cdot x\|_{L^{\infty }(\Omega _{\eta^{k+1}})}=&\|v-\bar{a} n_{k}\cdot x\|_{L^{\infty }(\Omega _{\eta r})}\\ \leq & C_1\eta ^{\alpha_0 }\cdot \hat{C}M\eta ^{(k+1)}A_{k+1}+\bar{a}K(\eta)r\omega(r)\\ \leq &\left( C_1\eta ^{\alpha_{0} }+\frac{C_2K(\eta)\omega(\eta^{k})}{\eta^{1+\alpha_{0}}}\right)\cdot \hat{C}M\eta ^{(k+1)}A_{k+1}\\ \leq &\left( C_1\eta ^{\alpha_{0} }+\frac{C_2K(\eta)c_0}{\eta^{1+\alpha_{0}}}\right)\cdot \hat{C}M\eta ^{(k+1)}A_{k+1}. \end{aligned} \end{equation}
For $w$, by the Alexandrov-Bakel'man-Pucci maximum principle, we have \begin{equation*}
\begin{aligned}
\sup_{\Omega_{\eta^{k+1}}}w\leq \sup_{\tilde{\Omega} _{r}}w& \leq\|g\|_{L^{\infty }(\partial \Omega \cap \tilde{B}^{+}_{\Gamma _{ r}})}+\sup_{\partial \Omega \cap \tilde{B}^{+}_{\Gamma _r}}(-a_k n_{k} \cdot x)+C_3r\|f\|_{L^n(\tilde{\Omega}_{r})}\\
&\leq Mr\omega_g(r) +\sum_{i=0}^{k}|a_i-a_{i-1}|\eta^k\omega(\eta^k)+C_3r\|f\|_{L^{n}(\Omega)}\omega_f(r)\\
&\leq M \eta^k \omega(\eta^k)+\bar{C}\hat{C}M\sum_{i=0}^{k}A_i\eta^k\omega( \eta^k)+C_3M\eta^k \omega(\eta^k),
\end{aligned} \end{equation*} where $C_3$ depends only on $n,\lambda$ and $\Lambda$.
From \cref{e.n1}, we have \begin{equation}\label{e.22} \begin{aligned}
\sum_{i=0}^{\infty} A_i&\leq \frac{1}{1-\eta^{\alpha_0}}\sum_{i=1}^{\infty}\omega(\eta^i)+c_0\\ &=\frac{1}{1-\eta^{\alpha_0}}\sum_{i=1}^{\infty}\frac{\omega(\eta^i) \left(\eta^{i-1}-\eta^i\right)}{\eta^{i-1}-\eta^i}+c_0\\ &=\frac{1}{\left(1-\eta^{\alpha_0}\right)\left(1-\eta\right)}\sum_{i=1}^{\infty} \frac{\omega(\eta^i)\left(\eta^{i-1}-\eta^i\right)}{\eta^{i-1}}+c_0\\ &\leq \frac{1}{\left(1-\eta^{\alpha_0}\right)\left(1-\eta\right)}\int_{0}^{1} \frac{\omega(r)dr}{r}+c_0\\ &\leq \frac{c_0}{\left(1-\eta^{\alpha_0}\right)\left(1-\eta\right)}+c_0\leq 3c_0, \end{aligned} \end{equation} provided \begin{equation}\label{e-w} \left(1-\eta^{\alpha_0}\right)\left(1-\eta\right)\geq 1/2. \end{equation} From the definition of $A_k$ again, \begin{equation*}
\omega(\eta^{k})\leq A_k\leq \frac{A_{k+1}}{\eta^{\alpha_0}}. \end{equation*} Hence, \begin{equation}\label{e1.22}
\begin{aligned} \sup_{\Omega_{k+1}}w &\leq \frac{1}{\eta^{1+\alpha_0}} M\eta^{k+1}A_{k+1}+\frac{3c_0\bar{C}}{\eta^{1+\alpha_0}}\hat{C}M\eta^{k+1}A_{k+1}+\frac{C_3 }{\eta^{1+\alpha_0}} M\eta^{k+1}A_{k+1}\\ &\leq \frac{C_3+1}{\eta^{1+\alpha_0}} M\eta^{k+1}A_{k+1}+\frac{3c_0\bar{C}}{\eta^{1+\alpha_0}}\hat{C}M\eta^{k+1}A_{k+1}\\ &\leq\left(\frac{C_3+1}{\hat{C}\eta^{1+\alpha_0}}+\frac{3c_0\bar{C}}{\eta^{1+\alpha_0}}\right) \hat{C}M\eta^{k+1}A_{k+1}. \end{aligned} \end{equation} Since \begin{equation*}
\begin{aligned}
\left| n_{k} - n_{k+1}\right| \leq {K(\eta) }\omega \left( {{\eta ^k}} \right), \end{aligned} \end{equation*} we have for $x\in B_{\eta^{k+1}}$, \begin{equation}\label{e-H-1.25}
\begin{aligned}
\left| {{a_{k + 1}}\left(n_k-n_{k+1} \right) \cdot x} \right| &\leq {a_{k + 1}}{K(\eta) }\omega ( \eta ^k){\eta ^{k + 1}}\\
&\leq \frac{{{a_{k + 1}}{K(\eta) }}}{{\hat CM{\eta ^{{\alpha _0}}}}}\hat C M{\eta ^{\left( {k + 1} \right)}}{A_{k + 1}}\\
&\leq \frac{3c_0 K(\eta)\bar{C}}{\eta^{\alpha_0}}\hat C M{\eta ^{\left( {k + 1} \right)}}{A_{k + 1}}. \end{aligned} \end{equation}
Let $\bar{C}=C_2/{\eta ^{{\alpha _0}}}$. Take $\eta $ small enough such that \cref{e-w} holds and \begin{equation*}
C_1\eta ^{\alpha_0 }\leq \frac{1}{6}. \end{equation*} Take $c_0$ small enough such that \begin{equation*} \frac{C_2K(\eta)c_0}{\eta^{1+\alpha_0}}\leq\frac{1}{6}, \frac{3c_0\bar{C}}{\eta^{1+\alpha_0}}\leq\frac{1}{6}~\mbox{and}~ \frac{3c_0 K(\eta)\bar{C}}{\eta^{\alpha_0}}\leq \frac{1}{3}. \end{equation*} Finally, take $\hat{C}$ large enough such that \cref{e.21} holds and \begin{equation*}
\frac{C_3+1}{\hat{C} \eta ^{1+\alpha_0}}\leq \frac{1}{6}. \end{equation*}
Let $a_{k+1}=a_k+\bar{a}$. Then combining \cref{e.20}, \cref{e1.22} and \cref{e-H-1.25}, we have for $x\in \Omega_{\eta^{k+1}}$, \[\begin{array}{l} \quad u - {a_{k + 1}}{n_{k + 1}} \cdot x = u - {a_{k + 1}}{n_{k}} \cdot x + {a_{k + 1}}{n_{k}} \cdot x - {a_{k + 1}}{n_{k + 1}} \cdot x\\
= u - \left( {{a_k} + \bar a} \right){n_{k}} \cdot x + {a_{k + 1}}\left( {{n_{k}} - {n_{k + 1}}} \right) \cdot x\\
= u - {a_k}{n_{k}} \cdot x - v + v - \bar a{n_k} \cdot x + {a_{k + 1}}\left( {{n_{k}} - {n_{k + 1}}} \right) \cdot x\\
\le \hat CM{\eta ^{\left( {k + 1} \right)}}{A_{k + 1}}. \end{array}\] By induction, the proof is completed. \qed~\\
The proof of the Hopf lemma is similar to that of the boundary Lipschitz regularity. Here, we focus on the curved boundary toward the interior of the domain. We need the following lemma, which can be easily proved by constructing a proper barrier. \begin{lemma}\label{le-H-1} Let $u\geq 0$ satisfy \begin{equation*}
\left\{\begin{aligned}
M^{-}(D^2u,\lambda,\Lambda)&=0~~\mbox{in}~~Q_1^+;\\
u&=0~~\mbox{on}~~T_1;\\
u&\geq 1~~\mbox{on}~~T_1+e_n. \end{aligned}\right. \end{equation*} Then \begin{equation*}\label{e-H-l} u(x)\geq c_1x_n ~~\mbox{in}~~B_{\delta_1}^+, \end{equation*} where $\delta_1>0$ and $c_1>0$ depend only on $n,\lambda$ and $\Lambda$. \end{lemma}
Now, we give the~\\ \noindent\textbf{Proof of \Cref{t-H-2}.} As before, from the Dini condition, there exists $0<r_1<R$ such that for any $0<r\leq r_1$, \begin{equation}\label{e-H-dini-4}
\omega(r)\leq c_0 ~~\mbox{and}~~ \int_{0}^{r}\frac{\omega(s)}{s}ds\leq c_0, \end{equation} where $c_0\leq 1/4$ is a small constant to be specified later and depends only on $n,\lambda,\Lambda$ and $K$. Moreover, since $n_{\Gamma_r}$ converges as $r\to 0$ (see \Cref{t-1}), we assume that $n_{\Gamma_r} \to e_n$ without loss of generality. Then there exists $0<r_1<R$ such that for any $0<r\leq r_2$, \begin{equation}\label{e-H-dini-3}
\left| {{n_{\Gamma_r}} - {e_n}} \right| \leq c_0. \end{equation} By a proper scaling, we assume that $\min\{r_1,r_2\}=1$ and $u(e_n/2)=1$. Let $\Omega _r^ + = \Omega \cap B_{{\Gamma _r}}^ +$. To prove \cref{e-H-main}, we only need to prove the following:
There exist constants $0<\alpha_{0}, \eta < 1$, $\bar{C}$ and $\tilde{a}>0$ (depending only on $n,\lambda$ and $\Lambda$), $\hat{C}$ (depending only on $n,\lambda,\Lambda$ and $K$) and a nonnegative sequence $\{a_k\}$ such that for $k\geq0$, \begin{equation}\label{e-H1.16} \inf_{\Omega _{{\eta ^{k + 1}}}^ + }(u-\tilde{a}{n_{k}}\cdot x+a_k{n_{k}}\cdot x)\geq -\hat{C} \eta ^{k}A_k, \end{equation} \begin{equation}\label{e-H1.17}
|a_k-a_{k-1}|\leq \bar{C}\hat{C}A_k \end{equation} and \begin{equation}\label{e-H-atilde} a_k\leq \frac{\tilde{a}}{2}, \end{equation} where $A_k$ and $n_k$ are defined as before.
Indeed, for any unit vector $l\in R^n$ with $l\cdot {e_n}=\tau>0$, there exists $k_0\geq1$ such that $\omega(\eta^{k})\leq \tau/4$ and $l\cdot n_k\geq \tau/2$ for any $k\geq k_0$. Then $tl\in \Omega$ for any $0<t<1$. Note that $A_k\to 0$ as $k\to \infty$ and then there exits $k_1\geq k_0$ such that \begin{equation*}
A_k\cdot \frac{\hat{C}}{ l\cdot{n_{k}}\eta^2}\leq \frac{\tilde{a}}{4}. \end{equation*} Take $\delta=\eta^{k_1}$. For $0<t<\delta$, there exists $k\geq k_1$ such that \begin{equation*}
\eta^{k+2}\leq t \leq \eta^{k+1}. \end{equation*} Then by \cref{e-H1.16}, \begin{equation*} \begin{aligned} u(tl)&\geq \tilde{a}({n_{k}}\cdot l)t-a_k({n_{k}}\cdot l)t-\hat{C}\eta^kA_k &\geq \frac{\tilde{a}({n_{k}}\cdot l)t}{2}-\frac{\hat{C}A_kt}{\eta^2} &\geq \frac{\tilde{a}({n_{k}}\cdot l)t}{4} &\geq \frac{\tilde{a}\tau t}{4}. \end{aligned} \end{equation*} That is, \cref{e-H-main} holds.
Now, we prove \crefrange{e-H1.16}{e-H-atilde} by induction. Let $\tilde{Q}=Q^{+}_{1/2}+c_0{e_n},\tilde{T}=\partial \tilde{Q}\cap \left\{x_n=1/4+c_0\right\}$. That is, $\tilde{T}$ is the top boundary of $\tilde{Q}$. Note that $c_0\leq 1/4$. Thus, $\tilde{Q}\subset \Omega\cap B^+_{1}$.
By the Harnack inequality, \begin{equation*}
\inf_{\tilde{T}} u\geq \tilde{c} u({e_n}/2)=\tilde{c}, \end{equation*} where $\tilde{c}$ depends only on $n,\lambda$ and $\Lambda$. Let $\tilde{u}$ solve \begin{equation*}
\left\{\begin{aligned} &M^{-}(D^2\tilde{u},\lambda,\Lambda)= 0 &&\mbox{in}~~\tilde{Q}; \\ &\tilde{u}=\tilde{c}&&\mbox{on}~~\tilde{T};\\ &\tilde{u}=0 &&\mbox{on}~~\partial \tilde{Q}\backslash \tilde{T}.
\end{aligned}\right. \end{equation*} From \Cref{le-H-1}, there exist $\delta_1>0$ and $0<c_2<1/2$ such that \begin{equation*} u(x)\geq \tilde{u}(x) \geq c_2(x_n-c_0) ~~\mbox{in}~~B^+_{\delta_1}+c_0{e_n}. \end{equation*} Note that $u\geq 0$ and hence, \begin{equation*}
u(x)\geq c_2(x_n-c_0) ~~\mbox{in}~~\Omega\cap B_{\delta_1}. \end{equation*} Therefore, by noting \cref{e-H-dini-3}, \begin{equation*}
u(x)\geq c_2x\cdot n_0-2c_2c_0\geq c_2x\cdot n_0-c_0 ~~\mbox{in}~~\Omega\cap B_{\delta_1}. \end{equation*}
Set $\tilde{a}=c_2$ and $a_{-1}=a_0=0$. Take \begin{equation}\label{e-H-1} \eta\leq \delta_1. \end{equation} Then \crefrange{e-H1.16}{e-H-atilde} hold for $k=0$. Suppose that they hold for $k$. We need to prove that they hold for $k+1$.
Let $r=\eta ^{k+1},\tilde{B}^+=B^{+}_{\Gamma _r}-K(\eta)r\omega(r)n_{\Gamma_r}$ and $\tilde{T}=T_{\Gamma _r}-K(\eta)r\omega(r)n_{\Gamma_r}$. Then $B^{+}_{\Gamma _{\eta r}}\subset \tilde{B}^+$. Let $v$ solve \begin{equation*} \left\{\begin{aligned}
&M^{-}(D^2v,\lambda,\Lambda)=0 &&\mbox{in}~~\tilde{B}^+; \\
&v=0 &&\mbox{on}~~\tilde{T};\\
&v=-\hat{C}\eta ^{k}A_k &&\mbox{on}~~\partial \tilde{B}^{+}\backslash T_{\Gamma _r}. \end{aligned} \right. \end{equation*} Let $w=u-\tilde{a}{n_{k}}\cdot x+a_k{n_{k}}\cdot x-v$. Then $w$ satisfies (note that $u\geq 0$ and $v\leq 0$) \begin{equation*}
\left\{
\begin{aligned}
&M^{-}(D^2w,\lambda/n,\Lambda)\leq 0 &&\mbox{in}~~ \Omega\cap \tilde{B}^{+}; \\
&w\geq -\tilde{a}{n_{k}}\cdot x+a_k{n_{k}}\cdot x&&\mbox{on}~~\partial \Omega \cap \tilde{B}^{+};\\
&w\geq 0 &&\mbox{on}~~\partial \tilde{B}^{+}\cap \bar{\Omega}.
\end{aligned}
\right. \end{equation*}
In the following arguments, we estimate $v$ and $w$ respectively. By the boundary $C^{1,\alpha}$ estimate for $v$ (see \Cref{l-1}) and the maximum principle, there exist $0<\alpha<1$ (depending only on $n,\lambda$ and $\Lambda$) and $\bar{a}\geq 0$ such that (note that $A_k\leq A_{k+1}/\eta^{\alpha_0}$) \begin{equation*} \begin{aligned}
\|v&+\bar{a}{n_{k+1}}\cdot (x-K(\eta)r\omega(r)n_{k+1})\|_{L^{\infty }(B^+ _{\Gamma _{\eta r}})}\leq C_1 \eta ^{1+ \alpha}\|v\|_{L^{\infty }(\tilde{B}^{+})}\\
&\leq C_1\eta ^{\alpha}\cdot \hat{C}\eta ^{(k+1)}A_k
\leq C_1\eta ^{\alpha-\alpha_0 }\cdot \hat{C}\eta ^{(k+1)}A_{k+1} \end{aligned} \end{equation*} and \begin{equation}\label{e.H-19} \bar{a}\leq C_2\hat{C}A_k/\eta, \end{equation} where $C_1$ and $C_2$ depend only on $n,\lambda$ and $\Lambda$. Take $\alpha_0=\alpha/2$. Then \begin{equation}\label{e.H-20} \begin{aligned}
\|v+\bar{a}{n_{k+1}}\cdot x\|_{L^{\infty }(\Omega^+_{k+2})}&\leq\|v+\bar{a}{n_{k+1}}\cdot x\|_{L^{\infty }(B^+_{\Gamma _{\eta r}})}\\ &\leq \left(C_1\eta ^{\alpha_0 }+\frac{C_2A_kK(\eta)}{\eta}\right)\cdot \hat{C}\eta ^{(k+1)}A_{k+1}\\ &\leq \left(C_1\eta ^{\alpha_0 }+\frac{C_2c_0K(\eta)}{\eta}\right)\cdot \hat{C}\eta ^{(k+1)}A_{k+1} \end{aligned} \end{equation}
For $w$, by the maximum principle, we have \begin{equation*}
\begin{aligned} \inf_{\Omega^+_{k+2}}w&\geq \inf_{\Omega\cap\tilde{B}^+}w \geq \inf_{\partial \Omega \cap \tilde{B}^{+}}\left(-\tilde{a}{n_{k}}\cdot x+a_k n_{k}\cdot x\right)\\
&\geq \inf_{\partial \Omega \cap \tilde{B}^{+}}\left(-\tilde{a}|n_{k}-n_{k+1}|\cdot|x|-\tilde{a}(n_{k+1}\cdot x)\right) \\ &\geq -\tilde{a}K(\eta)\eta^{k + 1}\omega(\eta^k)-\tilde{a}{\eta ^{k + 1}}\omega(\eta ^{k + 1}).
\end{aligned} \end{equation*} As before, from the definition of $A_k$ (see \cref{e-Ak}), \begin{equation*}
\omega(\eta^{k+1})\leq A_{k+1}~\mbox{and}~\omega(\eta^{k})\leq A_{k}\leq \frac{A_{k+1}}{\eta^{\alpha_0}}. \end{equation*} Hence, \begin{equation}\label{e-H-1.21}
\begin{aligned} \inf_{\Omega^+_{k+2}}w &\geq -\frac{K(\eta)+1}{\hat{C}\eta^{\alpha_0}}\cdot \hat{C}\eta^{k+1}A_{k+1}. \end{aligned} \end{equation}
Since $\left| {{n_{k}} - {n_{k + 1}}} \right| \le {K(\eta) }\omega (\eta ^k)$, for $x\in B_{\eta^{k+2}}$, \begin{equation}\label{e-H-1.22}
\begin{aligned}
\left| {\tilde a\left( {{n_{k}} - {n_{k + 1}}} \right) \cdot x} \right|
\leq \tilde a{K(\eta) }\omega \left( {{\eta ^k}} \right)\eta^{k+2}
\leq \frac{K(\eta) }{\hat C}\hat C{\eta ^{\left( {k + 1} \right)}}{A_{k + 1}}. \end{aligned} \end{equation}
Take $\eta $ small enough such that \cref{e-H-1} holds, \begin{equation*}
C_1\eta ^{\alpha_0 }\leq 1/6 \end{equation*} and \begin{equation}\label{e-H-2} \left(1-\eta^{\alpha_0}\right)\left(1-\eta\right)\geq 1/2. \end{equation}
Let $\bar{C}=C_2/{{\eta ^{1 + {\alpha _0}}}}$. Take $\hat{C}$ large enough such that \begin{equation*} \frac{K(\eta)+1}{\hat C\eta ^{\alpha _0}} \leq \frac{1}{3}. \end{equation*} As before, by noting that \cref{e.22} and \cref{e-H-2}, we have \begin{equation*}
\sum_{i=0}^{\infty} A_i\leq 3c_0. \end{equation*} Finally, take $c_0$ small enough such that \begin{equation*} 3c_0 \bar{C}\hat{C}\leq\frac{\tilde{a}}{2}~\mbox{and}~\frac{C_2c_0K(\eta)}{\eta} \leq \frac{1}{6}. \end{equation*}
Let $a_{k+1}=a_k+\bar{a}$. Then combining \crefrange{e.H-20}{e-H-1.22}, we have for $x\in \Omega^+_{k+2}$ \begin{equation*}
\begin{aligned} u &-\tilde a{n_{k+1}} \cdot x + {a_{k + 1}}{n_{k+1}} \cdot x\\ &=u-\tilde a n_{k} \cdot x + a_{k} n_{k} \cdot x-\tilde a(n_{k+1}-n_{k}) \cdot x + a_k(n_{k+1}-n_{k}) \cdot x+ \bar{a}n_{k+1} \cdot x\\ &= u - \tilde a{n_{k}} \cdot x + {a_k}{n_{k}} \cdot x - v + \left( {v + \bar{a}n_{k+1} \cdot x} \right) + (a_k-\tilde a)\left( {{n_{k}} - {n_{k+1}}} \right) \cdot x\\ &= \omega + \left( {v + \bar{a}n_{k+1} \cdot x} \right)+ (a_k-\tilde a)\left( {{n_{k}} - {n_{k+1}}} \right) \cdot x\\ &\geq - \hat C{\eta ^{\left( {k + 1} \right)}}{A_{k + 1}}
\end{aligned} \end{equation*} and \begin{equation*}
a_{k+1}\leq \sum_{i=0}^{k}|a_i-a_{i-1}|\leq\bar{C}\hat C\sum\limits_{i = 0}^\infty {{A_i}}\leq 3c_0 \bar{C}\hat{C}\leq \frac{\tilde{a}}{2}. \end{equation*} By induction, the proof is completed. \qed~\\
\end{document}
|
arXiv
|
{
"id": "1912.11958.tex",
"language_detection_score": 0.5630010962486267,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\begin{abstract}
We survey the development of the notion of Lazarsfeld-Mukai bundles together with various applications, from the classification of Mukai manifolds to Brill-Noether theory and syzygies of $K3$ sections. To see these techniques at work, we present a short proof of a result of M. Reid on the existence of elliptic pencils. \end{abstract} \maketitle
\section*{Introduction}
Lazarsfeld--Mukai bundles appeared naturally in connection with two completely different important problems in algebraic geometry from the 1980s. The first problem, solved by Lazarsfeld, was to find explicit examples of smooth curves which are generic in the sense of Brill-Noether-Petri \cite{LazarsfeldJDG}. The second problem was the classification of prime Fano manifolds of coindex 3 \cite{MukaiPNAS}. More recently, Lazarsfeld--Mukai bundles have found applications to syzygies and higher-rank Brill--Noether theory.
The common feature of all these research topics is the central role played by $K3$ surfaces and their hyperplane sections. For the Brill--Noether--Petri genericity, Lazarsfeld proves that a general curve in a linear system that generates the Picard group of a $K3$ surface satisfies this condition. For the classification of prime Fano manifolds of coindex 3, after having proved the existence of smooth fundamental divisors, one uses the geometry of a two-dimensional linear section which is a very general $K3$ surface.
The idea behind this definition is that the Brill--Noether theory of smooth curves~on a $K3$ surface, also called \textit{$K3$ sections}, is governed by higher-rank vector bundles on the surface. To be more precise, consider $S$ a $K3$ surface (considered always to be smooth, complex, projective), $C$ a smooth curve on $S$ of genus $\ge 2$, and $|A|$ a base-point-free pencil on $C$. If we attempt to lift the linear system
$|A|$ to the surface $S$, in~most cases, we will fail. For instance, $|A|$ cannot lift to a pencil on $S$ if $C$ generates
$\mathrm{Pic}(S)$ or if $S$ does not contain any elliptic curve at all. However, interpreting a general divisor in $|A|$ as a zero-dimensional subscheme of $S$, it is natural to try and find a rank-two bundle $E$ on $S$ and a global section of $E$ whose scheme of zeros coincides with the divisor in question. Varying the divisor, one~should exhibit in fact a two-dimensional space of global sections of $E$. The effective construction of $E$ is realized through elementary modifications, see Sect.\,\ref{section: Def}, and this is precisely a Lazarsfeld--Mukai bundle of rank two. The passage to higher ranks is natural, if we start with a complete, higher-dimensional, base-point-free linear system on $C$. At the end, we obtain vector bundles with unusually high number of global sections, which provide us with a rich geometric environment.
The structure of this chapter is as follows. In the first section, we recall the definition of Lazarsfeld--Mukai bundles and its first properties. We note equivalent conditions for a bundle to be Lazarsfeld--Mukai in Sect.\,\ref{subsection: Definition LM}, and we discuss simplicity in the rank-two case in Sect.\,\ref{subsection: Simple LM}. The relation with the Petri conjecture and the classification of Mukai manifolds, the original motivating problems for the~definition, are considered in Sects.\,\ref{subsection: BNP} and \ref{subsection: Mukai manifolds}, respectively. In Sect.\,\ref{section: Constancy} we treat the problem of constancy of invariants in a given linear system. For small gonalities, Saint-Donat and Reid proved that minimal pencils on $K3$ sections are induced from elliptic pencils on the $K3$ surface; we present a short proof using Lazarsfeld--Mukai bundles in Sect.\,\ref{subsection Gonality I}. Harris and Mumford conjectured that the gonality should always be constant. We discuss the evolution of this conjecture, from Donagi--Morrison's counterexample, Sect.\,\ref{subsection Gonality I}, to Green--Lazarsfeld's reformulation in terms of Clifford index, Sect.\,\ref{subsection: Cliff} and to Ciliberto--Pareschi's results on the subject, Sect.\,\ref{subsection: Gonality II}. The works around this problem emphasized the importance of parameter spaces of Lazarsfeld--Mukai bundles. We conclude the section with a discussion of dimension calculations of these spaces, Sect.\,\ref{subsection: Parameter}, which are applied afterwards to Green's conjecture. Sect.\,\ref{section: Green} is devoted to Koszul cohomology and notably to Green's conjecture for $K3$ sections. After recalling the definition and the motivations that led to the definition, we discuss the statement of Green's conjecture, and we sketch the proof for $K3$ sections. Voisin's approach using punctual Hilbert schemes, which is an essential ingredient, is examined in Sect.\,\ref{subsection: Hilbert}. Lazarsfeld--Mukai bundles are fundamental objects in this topic, and their role is outlined in Sect.\,\ref{subsection: Role of LM}. The final step in the solution of Green's conjecture for $K3$ sections is tackled in Sect.\,\ref{subsection: Green for K3}. We conclude this chapter with a short discussion on Farkas--Ortega's new applications of Lazarsfeld--Mukai bundles to Mercat's conjecture (which belongs to the rapidly developing higher-dimensional Brill--Noether theory), Sect.\,\ref{section: Higher BN}.
\noindent{\it Notation.} The additive and the multiplicative notation for divisors and line bundles will be mixed sometimes. If $E$ is a vector bundle on $X$ and $L\in \mbox{Pic}(X)$, we set $E(-L):=E\otimes L^*$; this notation will be used especially when $E$ is replaced by the canonical bundle $K_C$ of a curve $C$.
\section{Definition, Properties, the First Applications} \label{section: Def}
\subsection{Definition and First Properties} \label{subsection: Definition LM}
We fix $S$ a smooth, complex, projective $K3$ surface and $L$ a globally generated line bundle on $S$ with
$L^2=2g-2$. Let $C\in|L|$ be a smooth curve and $A$ be a base-point-free line bundle in $W^r_d(C)\setminus W^{r+1}_d(C)$. As mentioned in the Introduction, the definition of Lazarsfeld--Mukai bundles emerged from the attempt to lift the linear system
$A$ to the surface $S$. Since it is virtually impossible to lift it to another linear system, a higher-rank vector bundle is constructed such that $H^0(C,A)$ corresponds to an $(r+1)$-dimensional space of global sections. Hence $|A|$ lifts to a higher-rank analogue of a linear system.
The kernel of the evaluation of sections of $A$\vspace*{-3pt} \begin{equation} \label{eqn: F}
0\to F_{C,A}\to H^0(C,A)\otimes \mathcal{O}_S \buildrel{\mathrm{ev}}\over{\to} A\to 0\vspace*{-3pt} \end{equation} is a vector bundle of rank $(r+1)$.
\begin{defn}[Lazarsfeld \cite{LazarsfeldJDG}, Mukai \cite{MukaiPNAS}] The \textit{Lazarsfeld--Mukai bundle} $E_{C,A}$ associated to the pair $(C,A)$ is the dual of $F_{C,A}$. \end{defn}
By dualizing the sequence~(\ref{eqn: F}) we obtain the short exact sequence\vspace*{-3pt} \begin{equation} \label{eqn: E}
0\to H^0(C,A)^*\otimes \mathcal{O}_S \to E_{C,A}\to K_C(-A)\to 0,\vspace*{-3pt} \end{equation} and hence $E_{C,A}$ is obtained from the trivial bundle by modifying it along the curve $C$ and comes equipped with a natural $(r+1)$-dimensional space of global sections as planned.
We note here the first properties of $E_{C,A}$:
\begin{prop}[Lazarsfeld] \label{prop: E_{C,A}} The invariants of $E$ are the following: \begin{itemize}
\item[(1)] $\det (E_{C,A})= L$.
\item[(2)] $c_2(E_{C,A})=d$.
\item[(3)] $h^0 (S,E_{C,A})= h^0(C,A)+h^1(C,A)= 2r-d+1+g$.
\item[(4)] $h^1(S,E_{C,A})=h^2(S,E_{C,A})=0$.
\item[(5)] $\chi(S,E_{C,A}\otimes F_{C,A})=2(1-\rho(g,r,d))$.
where $\rho(g,r,d)=g-(r+1)(g-d+r)$.
\item[(6)] $E_{C,A}$ is globally generated off the base locus of $K_C(-A)$;
in particular, $E_{C,A}$ is globally generated if $K_C(-A)$
is globally generated. \end{itemize} \end{prop} It is natural to ask conversely if given $E$ a vector bundle on $S$ with $\mbox{rk}(E)=r+1$, $h^1(S,E)=h^2(S,E)=0$, and $\det (E)= L$, $E$ is the Lazarsfeld--Mukai bundle associated to a pair $(C,A)$. To this end, note that there is a rational map\vspace*{-3pt} \[
h_E : G(r+1,H^0(S,E)) \dashrightarrow |L|\vspace*{-3pt} \] defined in the following way. A general subspace $\varLambda\in G(r+1,H^0(S,E))$ is mapped to the degeneracy locus of the evaluation map: $ \mathrm{ev} _{\varLambda} : \varLambda \otimes \mathcal{O}_S \to E. $
If~the image $h_E(\varLambda)$ is a smooth curve $C\in |L|$, we set $\mbox{Coker}(\mathrm{ev}_{\varLambda}):=K_C(-A)$, where $A\in \mbox{Pic}(C)$ and $\mbox{deg}(A)=c_2(E)$, and observe that $E=E_{C,A}$. Indeed, since $h^1(S,E)=0$, $A$ is globally generated, and from $h^2(S,E)=0$ it follows that $\varLambda\cong H^0(C,A)^*$. The conclusion is that:
\begin{prop} \label{prop: caracterizare LM} A rank-$(r+1)$ vector bundle $E$ on $S$ is a Lazarsfeld--Mukai bundle if and only if $H^1(S,E)=H^2(S,E)=0$ and there exists an $(r+1)$-dimensional subspace of sections $\varLambda\subset H^0(S,E)$, such that the degeneracy locus of the morphism $\mathrm{ev}_{\varLambda}$ is a smooth curve. In particular, being a Lazarsfeld--Mukai vector bundle is an \textit{open condition}. \end{prop}
Note that there might be different pairs with the same Lazarsfeld--Mukai bundles, the difference being given by the corresponding spaces of global sections.
\subsection{Simple and Non-simple Lazarsfeld--Mukai Bundles} \label{subsection: Simple LM}
We keep the notation from the previous subsection. In the original situation, the bundles used by Lazarsfeld \cite{LazarsfeldJDG} and Mukai \cite{MukaiPNAS} are simple. The non-simple Lazarsfeld--Mukai bundles are, however, equally useful \cite{Aprodu-FarkasCOMP,Ciliberto-PareschiCRELLE}. For instance, Lazarsfeld's argument is partly based on an analysis of the non-simple bundles.
Proposition~\ref{prop: E_{C,A}} already shows that for $\rho(g,r,d)<0$ the associated Lazarsfeld--Mukai bundle cannot be simple. The necessity of making a distinction between simple and non-simple bundles for nonnegative $\rho$ will become more evident in the next sections.
In the rank-two case, one can give a precise description \cite{Donagi-MorrisonJDG} of non-simple Lazarsfeld--Mukai bundles, see also \cite{Ciliberto-PareschiCRELLE} Lemma 2.1:
\begin{lem}[Donagi--Morrison] \label{lemma: DM} Let $E_{C,A}$ be a non-simple Lazarsfeld--Mukai bundle. Then there exist line bundles $M,N\in \mathrm{Pic}(S)$ such that $h^0(S,M)$, $h^0(S,N)\ge 2$, $N$ is globally generated, and there exists a locally complete intersection subscheme $\xi$ of $S$, either of dimension zero or the empty set, such that $E_{C,A}$ is expressed as an extension \begin{equation} \label{eq: DM} 0\to M\to E_{C,A}\to N\otimes I_{\xi} \to 0. \end{equation} Moreover, if $h^0(S,M\otimes N^*)=0$, then $\xi=\emptyset$ and the extension splits. \end{lem}
One can prove furthermore that $h^1(S,N)=0$, \cite{Aprodu-FarkasCOMP} Remark 3.6.
We say that~(\ref{eq: DM}) is the \emph{Donagi--Morrison extension} associated to $E_{C, A}$. This notion makes perfect sense as this extension is uniquely determined by the vector bundle, if it is indecomposable \cite{Aprodu-FarkasCOMP}. Actually, a \textit{decomposable} Lazarsfeld--Mukai bundle $E$ cannot be expressed as an extension~(\ref{eq: DM}) with $\xi\ne\emptyset$, and hence a Donagi--Morrison extension is always unique, up to a permutation of factors in the decomposable case. Moreover, a Lazarsfeld--Mukai bundle is decomposable if and only if the corresponding Donagi--Morrison extension is trivial.
In the higher-rank case, we do not have such a precise description.\footnote{In fact, we do have a Harder--Narasimhan filtration, but we cannot control all the factors.} However, a similar sufficiently strong statement is still valid \cite{LazarsfeldJDG,LazarsfeldICTP,PareschiJAG}.
\begin{prop}[Lazarsfeld]
Notation as above. If $E_{C,A}$ is not simple, then
the linear system $|L|$ contains a reducible or a multiple curve. \end{prop}
In the rank-two case, this statement comes from the decomposition $L\cong M\otimes N$.
\subsection{The Petri Conjecture Without Degenerations} \label{subsection: BNP}
A smooth curve of genus $g$ is said to satisfy \textit{Petri's condition}, or to be \textit{Brill--Noether--Petri generic}, if the multiplication map (the Petri map) \[
\mu_{0,A}: H^0(C, A) \otimes H^0(C, K_C(-A))
\to H^0(C, K_C), \] is injective for any line bundle $A$ on $C$. One consequence of this condition is that all the Brill--Noether loci $W^r_d(C)$ have the expected dimension and are smooth away from $W^{r+1}_d(C)$; recall that the tangent space at the point $[A]$ to $W^r_d(C)$ is naturally isomorphic to the dual of $\mathrm{Coker}(\mu_{0,A})$ \cite{Arbarello-Cornalba-Griffiths-Harris}. The Petri conjecture, proved by degenerations by Gieseker, states that a general curve satisfies Petri's condition. Lazarsfeld \cite{LazarsfeldJDG} found a simpler and elegant proof without degenerations by analyzing curves on very general $K3$ surfaces.
Lazarsfeld's idea is to relate the Petri maps to the Lazarsfeld--Mukai bundles; this relation is valid in general and has many other applications. Suppose, as in the previous subsections, that $S$ is a $K3$ surface and $L$ is a globally generated line bundle on $S$. For the moment, we do not need to assume that $L$ generates the Picard group. E. Arbarello and M. Cornalba constructed a scheme $\mathcal{W}^r_d(|L|)$
parameterizing pairs $(C,A)$ with $C\in |L|$ smooth and $A\in W^r_d(C)$ and a morphism\vspace*{-3pt} \[
\pi_S:\mathcal{W}^r_d(|L|)\to |L|.\vspace*{-3pt} \]
Assume that $A\in W^r_d(C)\setminus W^{r+1}_d(C)$ is globally generated, and consider $M_A$ the vector bundle of rank $r$ on $C$ defined as the kernel of the evaluation map \begin{equation}\label{MA}
0\to M_A \to H^0(C,A)\otimes \mathcal{O}_C \buildrel{\mathrm{ev}}\over{\to} A\to 0. \end{equation} Twisting~(\ref{MA}) with $K_C\otimes A^*$, we obtain the following description of the kernel of the Petri map:\footnote{This ingenious procedure is an efficient replacement of the base-point-free pencil trick; ``it has killed the base-point-free pencil trick,'' to quote Enrico Arbarello.}\vspace*{4pt} \[
\mbox{Ker}(\mu_{0,A})= H^0(C, M_A\otimes K_C\otimes A^*). \]
There is another exact sequence on $C$\vspace*{-4pt} \[
0\to \mathcal{O}_C\to F_{C,A}|_C\otimes K_C\otimes A^*\to M_A\otimes K_C\otimes A^*\to 0,\vspace*{-4pt} \] and from the defining sequence of $E_{C,A}$ one obtains the exact sequence on $S$\vspace*{-4pt} \[
0\to H^0(C, A)^*\otimes F_{C,A}\to E_{C,A}\otimes F_{C,A}\to F_{C, A}|_C\otimes K_C\otimes A^*\to 0.\vspace*{-4pt} \]
From the vanishing of $h^0(C,F_{C,A})$ and of $h^1(C, F_{C, A})$, we obtain\vspace*{-4pt} \[
H^0(C,E_{C,A}\otimes F_{C,A})=H^0(C,F_{C,A}|_C\otimes K_C\otimes A^*).\vspace*{-4pt} \]
Suppose that $\mathcal{W}\subset \mathcal{W}^r_d(|L|)$ is a dominating component and $(C,A)\in\mathcal{W}$ is an element such that $A$ is globally generated and $h^0(C,A)=r+1$. A deformation-theoretic argument shows that if the Lazarsfeld--Mukai bundle $E_{C,A}$ is simple, then the coboundary map $H^0(C,M_A\otimes K_C\otimes A^*)\to H^1(C,\mathcal{O}_C)$ is zero \cite{PareschiJAG}, which eventually implies the injectivity of $\mu_{0,A}$.
By reduction to complete base-point-free bundles on the curve \cite{LazarsfeldJDG,PareschiJAG} this analysis yields:
\begin{thm}[Lazarsfeld] \label{thm: Lazarsfeld} Let $C$ be a smooth curve of genus $g\ge 2$ on a $K3$
surface~$S$, and assume that any divisor in the linear system $|C|$ is reduced and irreducible. Then a generic element in the linear system $|C|$ is Brill--Noether--Petri generic. \end{thm}
A particularly interesting case is when the Picard group of $S$ is generated by $L$ and $\rho(g,r,d)=0$. Obviously, the condition $\rho=0$ can be realized only for composite genera, as $g=(r+1)(g-d+r)$, for example, $r=1$ and $g$ even. Under these assumptions, there is a unique Lazarsfeld--Mukai bundle $E$ with $c_1(E)=L$ and $c_2(E)=d$, and different pairs $(C,A)$ correspond to different \hbox{$\varLambda\in G(r+1,$} $H^0(S,E))$; in other words the natural rational map $G(r+1,H^0(S,E))\dashrightarrow
\mathcal{W}^r_d(|L|)$ is dominating. Note that $E$ must be stable and globally generated.
\subsection{Mukai Manifolds of Picard Number One} \label{subsection: Mukai manifolds}
A Fano manifold $X$ of dimension $n\ge 3$ and index $n-2$ (i.e., of coindex 3) is called a \textit{Mukai manifold}.\footnote{Some authors consider that Mukai manifolds have dimension four or more.} In the classification, special attention is given to prime Fano manifolds: note that if $n\ge 7$, $X$ is automatically prime as shown by Wisniewski; see, for example, \cite{Iskovskih-Prokhorov}.
Assume that the Picard group of $X$ is generated by an ample line bundle $L$, and let the sectional genus $g$ be the integer $(L^n)/2+1$. Mukai and Gushel used vector bundle techniques to obtain a complete classification of these manifolds. A~first major obstacle is to prove that the fundamental linear system contains indeed a smooth element, aspect which is settled by Shokurov and Mella; see, for example, \cite{Iskovskih-Prokhorov}. Then the $(g+n-2)$-dimensional linear system
$|L|$ is base-point-free, and a general linear section with respect to the generator of the Picard group is a $K3$ surface. More precisely, if $\mathrm{Pic}(X) =\mathbb{Z}\cdot L$, then for $H_1,\cdots,H_{n-2}$ general elements in the fundamental linear system
$|L|$, $S:=H_1\cap\cdots\cap H_{n-2}$ is scheme-theoretically a $K3$ surface. Note that if $n\ge 4$ and $i\ge 3$, the intersection $H_1\cap\cdots\cap H_{n-i}$ is again a Fano manifold of coindex 3.
Mukai noticed that the fundamental linear system either is very ample, and the image of $X$ is projectively normal or is associated to a double covering of $\mathbb{P}^n$ ($g=2$) or of the hyper-quadric $Q^n\subset\mathbb{P}^{n+1}$ ($g=3$). The difficulty of the problem is thus to classify all the possible cases where
$|L|$ is normally generated, called \textit{of the first species}. Taking linear sections one reduces (not quite immediately) to the case $n=3$ \cite{Iskovskih-Prokhorov} p.110.
For simplicity, let us assume that $X$ is a prime Fano 3-fold of index 1. If
$g=4$ and $g=5$, $X$ is a complete intersection; hence the hard cases begin with genus $6$. A hyperplane section $S$ is a $K3$ surface, and, by a result of Moishezon, $\mathrm{Pic}(S)$ is generated by $L|_S$.
Let us denote by $\mathcal{F}_g$ the moduli space of polarized $K3$ surfaces of degree $2g-2$, by $\mathcal{P}_g$ the moduli space of pairs $(K3\ \mbox{surface},\mbox{curve})$ and $\mathcal{M}_g$ the moduli space of genus-$g$ curves. There are two nice facts in Mukai's proof involving these two moduli spaces. His first observation is that if there exists a prime Fano 3-fold $X$ of the first species of genus $g\ge 6$ and index $1$, the rational map $\phi_g:\mathcal{P}_g\dashrightarrow \mathcal{M}_g$ is \textit{not} generically finite \cite{MukaiLMS}. The second nice fact is that $\phi_g$ is generically finite if and only if $g=11$ or $g\ge 13$ \cite{MukaiLMS}.\footnote{In genus $11$, it is actually birational \cite{MukaiLNPAM}.} Hence, one is reduced to study the genera $6\le g\le 12$ with $g\ne 11$. At this point, Lazarsfeld--Mukai bundles are employed. By the discussion from Sect.\,\ref{subsection: BNP}, for any decomposition $g=(r+1)(g-d+r)$, with $r\ge 1$, $d\le g-1$, there exists a unique Lazarsfeld--Mukai bundle $E$ of rank $(r+1)$. It has already been noticed that the bundle $E$ is stable and globally generated. Moreover, the determinant map\vspace*{-4pt} \[
\mathrm{det}:\wedge^{r+1}H^0(S,E)\to H^0(S,L)\vspace*{-4pt} \] is surjective \cite{MukaiPNAS}, and hence it induces a linear embedding\vspace*{-4pt} \[
\mathbb{P}H^0(S,L)^*\hookrightarrow \mathbb{P}(\wedge^{r+1}H^0(S,E)^*).\vspace*{-4pt} \]
Following \cite{MukaiPNAS}, we have a commutative diagram \begin{center}
\mbox{\xymatrix{S \ar[r]^{\phi_E}\ar@{^(->}[d]^{\phi_{|L|}} & G \ar@{^(->}[d]^{\mathrm{Pluecker}}\\ \mathbb{P}H^0(L)^*\ar@{^(->}[r] & \mathbb{P}(\wedge^{r+1}H^0(E)^*)}} \end{center} where $G:=G(r+1,H^0(S,E)^*)$ and $\phi_E$ is given by $E$. This diagram shows that $S$ is embedded in a suitable linear section of the Grassmannian $G$. Moreover, this diagram extends over $X$: by a result of Fujita, $E$ extends to a stable vector bundle on $X$, and the diagram over $X$ is obtained for similar reasons. Hence $X$ is a linear section of a Grassmannian. By induction on the dimension, $X$ is contained in a \textit{maximal} Mukai manifold, which is also a linear section of the Grassmannian. A complete list of maximal Mukai manifolds is given in \cite{MukaiPNAS}. Notice that in genus $12$, the maximal Mukai manifolds are threefold already.
\setcounter{thm}{0} \section{Constancy of Invariants of $K3$ Sections} \label{section: Constancy}
\subsection{Constancy of the Gonality. I} \label{subsection Gonality I}
In his analysis of linear systems on $K3$ surfaces Saint--Donat \cite{Saint-DonatAJM} shows that any smooth curve which is linearly equivalent to a hyperelliptic or trigonal curve is also hyperelliptic, respectively trigonal. The idea was to prove that the minimal pencils are induced by elliptic pencils defined on the surface. This result was sensibly extended by Reid
\cite{ReidJLM} who proved the following existence result:
\begin{thm}[Reid] Let $C$ be a smooth curve of genus $g$ on a $K3$ surface $S$ and $A$ be a complete, base-point-free $g^1_d$ on $C$. If\vspace*{-3pt} \[ \frac{d^2}{4}+d+2<g,\vspace*{-3pt} \] then $A$ is the restriction of an elliptic pencil on $S$. \end{thm}
It is a good occasion to present here, as a direct application of techniques involving Lazarsfeld--Mukai bundles, an alternate shorter proof of Reid's theorem.
\begin{proof} We use the notation of previous sections. By the hypothesis, the Lazarsfeld--Mukai bundle $E$ is not simple, and hence we have a unique Donagi--Morrison extension\vspace*{3pt} \[ 0\to M\to E\to N\otimes I_{\xi}\to 0,\vspace*{2pt} \] with $\xi$ of length $\ell$. Note that $M\cdot N=d-\ell\le d$. By the Hodge index theorem, we have $(M^2)\cdot(N^2)\le (M\cdot N)^2\le d^2$, whereas from $M+N=C$ we obtain $(M^2)=2(g-1-d)-(N^2)$, hence\vspace*{-4pt} \[ (N^2)\le\frac{d^2}{2(g-1-d)-(N^2)}.\vspace*{-4pt} \]
Therefore, the even integer $x:=(N^2)$ satisfies the following inequality $x^2-2x(g-1-d)+d^2\ge 0.$ The hypothesis shows that the above inequality fails for $x\ge 2$, and hence $N$ must be an elliptic pencil. \end{proof}
In conclusion, for small values, the gonality\footnote{The gonality $\mathrm{gon}(C)$ of a curve $C$ is the minimal degree of a morphism from $C$ to the projective~line.} is constant in the linear system. Motivated by these facts, Harris and Mumford conjectured that \textit{the gonality of \hbox{$K3$-sections} should always be constant} \cite{Harris-MumfordINVENTIONES}.
This conjecture is unfortunately wrong as stated: Donagi and Morrison \cite{Donagi-MorrisonJDG} gave the following counterexample:
\begin{ex} \label{ex: DM}
Let $S\to \mathbb{P}^2$ be a double cover branched
along a smooth sextic and $L$ be the pull-back of
$\mathcal{O}_{\mathbb{P}^2}(3)$. The curves in $|L|$
have all genus $10$. The general curve $C\in|L|$
is isomorphic to a smooth plane sextic, and hence
it is pentagonal. On the other hand, the pull-back of
a general smooth plane cubic $\varGamma$ is a double
cover of $\varGamma$, and thus it is tetragonal. \end{ex}
\subsection{Constancy of the Clifford Index} \label{subsection: Cliff}
Building on his work on Koszul cohomology and its relations with geometry, M. Green proposed a reformulation of the Harris-Mumford conjecture replacing the gonality by the Clifford index.
Recall that the \textit{Clifford index} of a nonempty linear system $|A|$ on a smooth curve $C$
is the codimension of the image of the natural addition map $|A|\times|K_C(-A)|\to |K_C|$. This definition is nontrivial only for relevant linear systems $|A|$, i.e., such that both $|A|$ and $|K_C(-A)|$ are at least one-dimensional; such an $A$ is said to \textit{contribute to the Clifford index}. The \textit{Clifford index of $C$} is the minimum of all the Clifford indices taken over the linear systems that contribute to the Clifford index and is denoted by $\mathrm{Cliff}(C)$. The Clifford index is related to the gonality by the following inequalities\vspace*{3pt} \[ \mathrm{gon}(C)-3\le \mathrm{Cliff}(C)\le \mathrm{gon}(C)-2,\vspace*{3pt} \] and curves with $\mathrm{gon}(C)-3 = \mathrm{Cliff}(C)$ are very rare: typical examples are plane curves and Eisenbud--Lange--Martens--Schreyer curves \cite{ELMS,KnutsenIJM}.\footnote{It is conjectured that the only other examples should be some half-canonical curves of even genus and maximal gonality \cite{ELMS}; however, this conjecture seems to be very difficult.}
From the Brill--Noether theory, we obtain the bound $\mathrm{Cliff}(C)\le\left[(g-1)/2\right]$ (and, likewise, $\mathrm{gon}(C)\le\left[(g+3)/2\right]$), and it is known that the equality is achieved for general curves. The Clifford index is in fact a measure of how special a curve is in the moduli space.
The precise statement obtained by Green and Lazarsfeld is the following \cite{Green-LazarsfeldINVENTIONES}:
\begin{thm}[Green--Lazarsfeld] \label{thm: GL Cliff}
Let $S$ be a $K3$ surface and $C\subset S$ be a smooth
irreducible curve of genus $g\ge 2$. Then $\mathrm{Cliff}(C')
=\mathrm{Cliff}(C)$ for every smooth curve $C'\in|C|$. Furthermore,
if $\mathrm{Cliff}(C)$ is strictly less than the generic value
$ \left[(g-1)/2\right]$, then there exists a line bundle
$M$ on $S$ whose restriction to any smooth curve $C'\in|C|$ computes
the Clifford index of~$C'$. \end{thm}
The proof strategy is based on a reduction method of the associated Lazarsfeld--Mukai bundles. The bundle $M$ is obtained from the properties of the reductions; we refer to \cite{Green-LazarsfeldINVENTIONES} for details.
From the Clifford index viewpoint, Donagi--Morrison's example is not different from the other cases. Indeed, all smooth curves in $|L|$ have Clifford index $2$. We shall see in the next subsection that Donagi--Morrison's example is truly an isolated exception for the constancy of the gonality.
\subsection{Constancy of the Gonality. II} \label{subsection: Gonality II}
As discussed above, the Green--Lazarsfeld proof of the constancy of the Clifford index was mainly based on the analysis of Lazarsfeld--Mukai bundles. It is natural to try and explain the peculiarity of Donagi--Morrison's example from this point of view. This was done in \cite{Ciliberto-PareschiCRELLE}. The surprising answer found by Ciliberto and Pareschi \cite{Ciliberto-PareschiCRELLE} (see also \cite{Donagi-MorrisonJDG}) is the following:
\begin{thm}[Ciliberto--Pareschi] \label{thm: Ciliberto-Pareschi}
Let $S$ be a $K3$ surface and $L$ be an ample line bundle on $S$. If the gonality of the smooth curves in $|L|$ is not constant, then $S$ and $L$ are as in Donagi--Morrison's example. \end{thm}
Theorem~\ref{thm: Ciliberto-Pareschi} was refined by Knutsen \cite{KnutsenIJM} who replaced ampleness by the more general condition that $L$ be globally generated. The extended setup covers also the case of exceptional curves, as introduced by Eisenbud, Lange, Martens, and Schreyer \cite{ELMS}.
The proof of Theorem~\ref{thm: Ciliberto-Pareschi}
consists of a thorough analysis of the loci $\mathcal{W}^1_d(|L|)$, where $d$ is the minimal gonality of smooth curves in $|L|$, through the associated Lazarsfeld--Mukai bundles. The authors identify Donagi--Morrison's example in the following way:
\begin{thm}[Ciliberto--Pareschi] \label{thm: Ciliberto-Pareschi2}
Let $S$ be a $K3$ surface and $L$ be an ample line bundle on $S$. If the gonality of smooth curves in $|L|$ is not constant and if there is a pair $(C,A)\in\mathcal{W}^1_d(|L|)$ such that $h^1(S,E_{C,A}\otimes F_{C,A})=0$, then $S$ and $L$ are as in Donagi--Morrison's example. \end{thm}
To conclude the proof of Theorem~\ref{thm: Ciliberto-Pareschi}, Ciliberto and Pareschi prove that non-constancy of the gonality implies the existence of a pair $(C,A)$ with $h^1(S,E_{C,A}\otimes F_{C,A})=0$; see \cite{Ciliberto-PareschiCRELLE} Proposition 2.4.
It is worth to notice that, in Example~\ref{ex: DM}, if $C$ is the inverse image of a plane cubic and $A$ is a $g^1_4$ (the pull-back of an involution), then $E_{C,A}$ is the pull-back of $\mathcal{O}_{\mathbb{P}^2}(1)\oplus\mathcal{O}_{\mathbb{P}^2}(2)$ \cite{Ciliberto-PareschiCRELLE}, and hence the vanishing of $h^1(S,E_{C,A}\otimes F_{C,A})$ is guaranteed in this case.
\subsection{Parameter Spaces of Lazarsfeld--Mukai Bundles and~Dimension of Brill--Noether Loci} \label{subsection: Parameter}
We have already seen that the Brill--Noether loci are smooth of expected dimension at pairs corresponding to simple Lazarsfeld--Mukai bundles. It is interesting to know what is the dimension of these loci at other points as well. Precisely, we look for a uniform bound on the dimension of Brill--Noether loci of general curves in a linear system.
A first step was made by Ciliberto and Pareschi \cite{Ciliberto-PareschiCRELLE} who proved, as a necessary step in Theorem~\ref{thm: Ciliberto-Pareschi}, that an ample curve of gonality strictly less than the generic value, general in its linear system, carries finitely many minimal pencils. This result was extended to other Brill--Noether loci \cite{Aprodu-FarkasCOMP}, proving a phenomenon of \textit{linear growth} with the degree; see below. Let us mention that, for the moment, the only results in this direction are known to hold for pencils \cite{Aprodu-FarkasCOMP} and nets \cite{Lelli-Chiesa}.
As before, we consider $S$ a $K3$ surface and $L$ a globally generated line bundle on $S$. In order to parameterize all pairs $(C, A)$ with non-simple Lazarsfeld--Mukai bundles, we need a global construction. We fix a nontrivial globally generated line bundle $N$ on $S$ with $H^0(L(-2N))\neq 0$ and an integer $\ell\ge 0$. We set $M:=L(-N)$ and $g:=1+L^2/2$. Define $\widetilde{\mathcal{P}}_{N,\ell}$ to be the family of \textit{vector bundles} of rank $2$ on
$S$ given by nontrivial extensions\vspace*{-3pt} \begin{equation} \label{eq: extension} 0\to M\to E\to N\otimes I_{\xi}\to 0,\vspace*{-3pt} \end{equation} where $\xi$ is a zero-dimensional locally complete intersection subscheme (or the empty set) of $S$ of length $\ell$, and set \[ \mathcal{P}_{N,\ell}:=\{[E]\in\widetilde{\mathcal{P}}_{N, \ell}:\ h^1(S,E)=h^2(S,E)=0\}. \] Equivalently (by Riemann--Roch), $[E]\in \mathcal{P}_{N, \ell}$ if and only if $h^0(S,E)=$\break $g-c_2(E)+3$ and $h^1(S,E)=0$. Note that any non-simple Lazarsfeld--Mukai bundle on $S$ with determinant $L$ belongs to some family $\mathcal{P}_{N,\ell}$, from Lemma~\ref{lemma: DM}. The family $\mathcal{P}_{N,\ell}$, which, a priori, might be the empty set, is an open Zariski subset of a projective bundle of the Hilbert scheme~$S^{[\ell]}$.
Assuming that $\mathcal{P}_{N,\ell}\ne \emptyset$, we consider the Grassmann bundle $\mathcal{G}_{N,\ell}$ over $\mathcal{P}_{N,\ell}$ classifying pairs $(E,\varLambda)$ with $[E]\in\mathcal{P}_{N,\ell}$ and $\varLambda\in \mathrm{G}(2,H^0(S,E))$. If $d:=c_2(E)$ we define the rational map $h_{N, \ell}: \mathcal{G}_{N, \ell}
\dashrightarrow \mathcal{W}^1_d(|L|)$, by setting $h_{N, \ell}(E, \varLambda):=(C_{\varLambda}, A_{\varLambda})$, where $A_{\varLambda}\in \mbox{Pic}^d(C_{\varLambda})$ is such that the following exact sequence on $S$ holds:\vspace*{4pt} \[ 0\to \varLambda\otimes \mathcal{O}_S\stackrel{\mathrm{ev}_{\varLambda}}\to E\to K_{C_{\varLambda}}\otimes A_{\varLambda}^*\to 0.\vspace*{3pt} \]
One computes $\dim\ \mathcal{G}_{N,\ell}=g+\ell+h^0(S,M\otimes N^*)$. If we assume furthermore that $\mathcal{P}_{N,\ell}$ contains a Lazarsfeld--Mukai vector bundle $E$ on $S$ with $c_2(E)=d$
and consider $\mathcal{W}\subset \mathcal{W}^1_d(|L|)$ the closure of the image of the rational map $h_{N,\ell}:\mathcal{G}_{N,\ell}\dashrightarrow \mathcal{W}^1_d(|L|)$, then we find $\dim\ \mathcal{W} =g+d-M\cdot N=g+\ell$.
On the other hand, if $C\in|L|$ has Clifford dimension one and $A$ is a globally generated line bundle on $C$ with $h^0(C,A)=2$ and $[E_{C,A}]\in\mathcal{P}_{N,\ell}$, then $M\cdot N\ge \mathrm{gon}(C)$.
These considerations on the indecomposable case, together with a simpler analysis of decomposable bundles, yield finally \cite{Aprodu-FarkasCOMP}:
\begin{thm} \label{thm: Green Cliffdim 1}
Let $S$ be a $K3$ surface and $L$ a globally generated line bundle on $S$, such that general curves in $|L|$ are of Clifford dimension one. Suppose that
$\rho(g,1,k)\leq 0$, where $L^2=2g-2$ and $k$ is the (constant) gonality of all smooth curves in $|L|$. Then for a general curve $C\in|L|$, we have \begin{equation} \label{lgc}
\mathrm{dim}\ W^1_{k+d}(C)=d\mbox{ for all } 0\le d\le g-2k+2. \end{equation} \end{thm}
The condition~(\ref{lgc}) is called the \textit{linear growth condition}. It is equivalent to \[ \mathrm{dim}\ W^1_{g-k+2}(C)=\rho(g,1,g-k+2)=g-2k+2. \]
Note that the condition that $C$ carry finitely many minimal pencils, which is a part of~(\ref{lgc}), appears explicitly in~\cite{Ciliberto-PareschiCRELLE}. It is directly related to the constancy of the gonality discussed before.
\setcounter{thm}{0} \section{Green's Conjecture for Curves on $K3$ Surfaces} \label{section: Green}
\subsection{Koszul Cohomology}\label{ch1:sec3.1}
Let $X$ be a (not necessarily smooth) complex, irreducible, projective variety and $L\in\mathrm{Pic}(X)$ globally generated. The Euler sequence on the projective space $\mathbb{P}(H^0(X,L)^*)$ pulls back to a short exact sequence of vector bundles on~$X$ \begin{equation} \label{eqn: Euler} 0 \to M_L\to H^0(X,L)\otimes \mathcal{O}_X\to L\to 0. \end{equation}
After taking exterior powers in the sequence~(\ref{eqn: Euler}), twisting with multiples of $L$ and going to global sections, we obtain an exact sequence for any nonnegative $p$ and $q$:\vspace*{-5pt} \begin{align} \label{eqn: WedgeEuler} 0\to H^0(\wedge^{p+1}M_L\otimes L^{q-1}) \to \wedge^{p+1}H^0(L)\otimes H^0(L^{q-1}) \stackrel{\delta}{\to} H^0(\wedge^pM_L\otimes L^q).\nonumber\\ \vspace*{-5pt} \end{align}
The finite-dimensional vector space $K_{p,q}(X,L):=\mathrm{Coker}(\delta)$ is called the \textit{Koszul cohomology space}\footnote{The indices $p$ and $q$ are usually forgotten when defining Koszul cohomology.} of $X$ with values in $L$ \cite{LazarsfeldICTP,GreenJDG,GreenICTP}. Observe that $K_{p,q}$ can be defined alternatively as: \[ K_{p,q}(X,L)=\mathrm{Ker}\left(H^1(\wedge^{p+1}M_L\otimes L^{q-1}) \to \wedge^{p+1}H^0(L)\otimes H^1(L^{q-1})\right),\vspace*{-3pt} \] description which is particularly useful when $X$ is a curve.
Several versions are used in practice, for example, replace $H^0(L)$ in~(\ref{eqn: Euler}) by a subspace that generates $L$ or twist~(\ref{eqn: WedgeEuler}) by $\mathcal{F}\otimes L^{q-1}$ where $\mathcal{F}$ is a coherent sheaf. For our presentation, however, we do not need to discuss these natural generalizations.
Composing the maps\vspace*{-3pt} \[ \wedge^{p+1}H^0(L)\otimes H^0(L^{q-1}) \stackrel{\delta}{\to} H^0(\wedge^pM_L\otimes L^q) \hookrightarrow \wedge^pH^0(L)\otimes H^0(L^q)\vspace*{-3pt} \] we obtain, by iteration, a complex\vspace*{-3pt} \[ \wedge^{p+1}H^0(L)\otimes H^0(L^{q-1})\to \wedge^pH^0(L)\otimes H^0(L^q)\to \wedge^{p-1}H^0(L)\otimes H^0(L^{q+1})\vspace*{-3pt} \] whose cohomology at the middle is $K_{p,q}(X,L)$, and this is the definition given by Green \cite{GreenJDG}.
An important property of Koszul cohomology is upper-semicontinuity in flat families with constant cohomology; in particular, vanishing of Koszul cohomology is an open property in such families. For curves, constancy of $h^1$ is a consequence of flatness and of constancy of $h^0$, as shown by the Riemann--Roch theorem.
The original motivation for studying Koszul cohomology spaces was given by the relation with minimal resolutions over the polynomial ring. More precisely, if $L$ is very ample, then the Koszul cohomology computes the minimal resolution of the graded module\vspace*{4pt} \[ R(X,L):=\bigoplus_qH^0(X,L^q)\vspace*{3pt} \] over the polynomial ring \cite{GreenJDG,GreenICTP}; see also \cite{EisenbudBOOK,Aprodu-NagelULECT}, in the sense that any graded piece that appears in the minimal resolution is (non-canonically) isomorphic to a $K_{p,q}$. If the image of $X$ is projectively normal, this module coincides with the homogeneous coordinate ring of $X$. The projective normality of $X$ can also be read off Koszul cohomology, being characterized by the vanishing condition $K_{0,q}(X,L)=0$ for all $q\ge 2$. Furthermore, for a projectively normal $X$, the homogeneous ideal is generated by quadrics if and only if $K_{1,q}(X,L)=0$ for all $q\ge 2$.\footnote{The dimension of $K_{1,q}$ indicates the number of generators of degree $(q+1)$ in the homogeneous~ideal.} The phenomenon continues as follows: if $X$ is projectively normal and the homogeneous ideal is generated by quadrics, then the relations between the generators are linear if and only if $K_{2,q}(X,L)=0$ for all $q\ge 2$, whence the relation with syzygies ~\cite{GreenJDG}.
Other notable application of Koszul cohomology is the description of Castelnuovo--Mumford regularity, which coincides with, \cite{GreenJDG,Aprodu-NagelULECT} \[ \mathop{\mathrm{min}}_q\{K_{p,q}(X,L)=0,\mbox{ for all }p\}. \]
Perhaps the most striking property of Koszul cohomology, discovered by Green and Lazarsfeld \cite[Appendix]{GreenJDG}, is a consequence of a nonvanishing result:
\begin{thm}[Green--Lazarsfeld] \label{thm: GL nonvan}
Suppose $X$ is smooth and $L=L_1\otimes L_2$ with $r_i:=h^0(X,L_i)-1\ge 1$.
Then $K_{r_1+r_2-1,1}(X,L)\ne 0$. \end{thm}
Note that the spaces $K_{p,1}$ have the following particular attribute: if $K_{p,1}\ne 0$ for some $p\ge 1$ then $K_{p',1}\ne 0$ for all $1\le p'\le p$. This is obviously false for $K_{p,q}$ with $q\ge 2$.
Theorem~\ref{thm: GL nonvan} shows that the existence of nontrivial decompositions of $L$ reflects onto the existence of nontrivial Koszul classes in some space $K_{p,1}$. Its most important applications are for curves, in particular for canonical curves, case which is discussed in the next subsection. In the higher-dimensional cases, for surfaces, for instance, the meaning of Theorem~\ref{thm: GL nonvan} becomes more transparent if it is accompanied by a restriction theorem which compares the Koszul cohomology of $X$ with the Koszul cohomology of the linear sections \cite{GreenJDG}:
\begin{thm}[Green] \label{thm: Lefschetz}
Suppose $X$ is smooth and $h^1(X,L^q)=0$ for all $q\ge 1$.
Then for any connected reduced divisor $Y\in|L|$,
the restriction map induces an isomorphism\vspace*{4pt}
\[
K_{p,q}(X,L)\stackrel{\sim}{\to}K_{p,q}(Y,L|_Y),\vspace*{3pt}
\]
for all $p$ and $q$. \end{thm}
The vanishing of $h^1(X,\mathcal{O}_X)$ suffices to prove that the restriction is an isomorphism between the spaces $K_{p,1}$ \cite{Aprodu-NagelULECT}.
In the next subsections, we shall apply Theorem~\ref{thm: Lefschetz} for $K3$ sections.
\begin{cor} \label{cor: Lefschetz K3} Let $C$ be a smooth connected curve on a $K3$ surface $S$. Then\vspace*{-4pt} \[ K_{p,q}(S,\mathcal{O}_S(C))\cong K_{p,q}(C,K_C)\vspace*{-4pt} \] for all $p$ and $q$. \end{cor}
One direct consequence is a duality theorem for Koszul cohomology of $K3$ surfaces.\footnote{Duality for Koszul cohomology of curves follows from Serre's duality. For higher-dimensional manifolds, some supplementary vanishing conditions are required \cite{GreenJDG,GreenICTP}.} It shows the symmetry of the table containing the dimensions of the spaces $K_{p,q}$, called \textit{the Betti table}.
\vspace*{-4pt} \subsection{Statement of Green's Conjecture}\label{ch1:sec3.2}
Let us particularize Theorem~\ref{thm: GL nonvan} for a canonical curve. Consider $C$ a smooth curve and choose a decomposition $K_C=A\otimes K_C(-A)$. Theorem~\ref{thm: GL nonvan} applies only if $h^0(C,A)\ge 2$ and $h^1(C,A)\ge 2$, i.e., if $A$ contributes to the Clifford index. The~quantity $r_1+r_2-1$ which appears in the statement equals $g-\mathrm{Cliff}(A)-2$, and hence, if $A$ \textit{computes} the Clifford index, we obtain the following:
\begin{thm}[Green--Lazarsfeld]
For any smooth curve $C$ of genus $g$ Clifford index $c$
we have $K_{g-c-2,1}(C,K_C)\ne 0$. \end{thm}
It is natural to determine whether or not this result is sharp, question which is addressed in the statement Green's conjecture:
\begin{con}[Green]
Let $C$ be a smooth curve. For all $p\ge g-c-1$,
we have $K_{p,1}(C,K_C) = 0$. \end{con}
For the moment, Green's conjecture remains a hard open problem. At the same time, strong evidence has been discovered. For instance, it is known to hold for general curves \cite{VoisinJEMS,VoisinCOMP}, for curves of odd genus and maximal Clifford index \cite{VoisinCOMP,Hirschowitz-RamananAENS}, for general curves of given gonality \cite{VoisinJEMS,TeixidorDUKE},\footnote{Voisin's and Teixidor's cases complete each other quite remarkably.} \cite{SchreyerLNM}, for curves with small Brill--Noether loci \cite{AproduMRL}, for plane curves \cite{LooseMANUSCRIPTA}, for curves on $K3$ surfaces \cite{VoisinJEMS,VoisinCOMP,Aprodu-FarkasCOMP},~etc.; see also \cite{Aprodu-NagelULECT} for a discussion.
We shall consider in the sequel the case of curves on $K3$ surfaces with emphasis on Voisin's approach to the problem and the role played by Lazarsfeld--Mukai bundles. It is interesting to notice that Green's conjecture for $K3$ sections can be formulated directly in the $K3$ setup, as a vanishing result on the moduli space $\mathcal{F}_g$ of polarized $K3$ surfaces. However, in the proof of this statement, as it usually happens in mathematics, we have to exit the $K3$ world, prove a more general result in the extended setup, and return to $K3$ surfaces. The steps we have to take, ordered logically and not chronologically, are the following. In the first, most elaborated step, one finds an example for odd genus \cite{VoisinCOMP,VoisinJEMS}. At this stage, we are placed in the moduli space $\mathcal{F}_{2k+1}$. Secondly, we exit the $K3$ world, land in $\mathcal{M}_{2k+1}$, and prove the equality of two divisors \cite{Hirschowitz-RamananAENS,VoisinJEMS}. The first step is used, and the identification of~the divisors extends to their closure over the component $\varDelta_0$ of the boundary \cite{AproduMRL}. In~the third step, we jump from a gonality stratum $\mathcal{M}^1_{g,d}$ in a moduli space $\mathcal{M}_g$ to the~boundary of another moduli space of stable curves $\overline{\mathcal{M}}_{2k+1}$, where $k=g-d+1$ \cite{AproduMRL}. The second step reflects into a vanishing result on an explicit open subset of $\mathcal{M}^1_{g,d}$. Finally one goes back to $K3$ surfaces and applies the latter vanishing result \cite{Aprodu-FarkasCOMP} on $\mathcal{F}_g$. In the steps concerned with $K3$ surfaces (first and last), the Lazarsfeld--Mukai bundles are central objects.
\subsection{Voisin's Approach} \label{subsection: Hilbert}
The proof of the generic Green conjecture was achieved by Voisin in two papers \cite{VoisinJEMS,VoisinCOMP}, using a completely different approach to Koszul cohomology via Hilbert scheme of points.
Let $X$ be a complex connected projective manifold and $L$ a line bundle on $X$. It is obvious that any global section $\sigma$ is uniquely determined by the collection $\{\sigma(x)\}_x$, where $\sigma(x)\in L|_x\cong\mathbb{C}$ and $x$ belongs to a nonempty open subset of $X$. One tries to find a similar fact for multisections in $\wedge^nH^0(X,L)$.
Let $\sigma_1\wedge\cdots\wedge\sigma_n$ be a decomposable element in $\wedge^nH^0(X,L)$ with $n\ge 1$. By analogy with the case $n=1$, we have to look at the restriction
$\sigma_1|_{\xi}\wedge\cdots\wedge\sigma_n|_{\xi}\in \wedge^nL|_{\xi}$ where $\xi$ is now a zero-dimensional subscheme, and it is clear that we need $n$ points for otherwise this restriction would be zero. Note that a zero-dimensional subscheme of length $n$ defines a point in the punctual Hilbert scheme $X^{[n]}$. For technical reasons, we shall restrict to curvilinear subschemes\footnote{A curvilinear subscheme is defined locally, in the classical topology, by $x_1=\cdots=x_{s-1}=x_s^k=0$; equivalently, it is locally embedded in a smooth curve.} which form a large open subset $X^{[n]}_c$ in a connected component of the Hilbert scheme.\footnote{The connectedness of $X^{[n]}_c$ follows from the observation that a curvilinear subscheme is a deformation of a reduced subscheme.} Varying $\xi\in X^{[n]}_c$, the collection
$\{\sigma_1|_{\xi}\wedge\cdots\wedge\sigma_n|_{\xi}\}_{\xi}$ represents a section in a line bundle described as follows. Put $\varXi_n\subset X^{[n]}_c\times X$ the incidence variety and denote by $q$ and $p$ the projections on the two factors; note that $q$ is finite of degree $n$. Then $L^{[n]}:=q_*p^*(L)$ is a vector bundle of rank $n$ on $X^{[n]}_c$, and the fibre at a point $\xi\in X^{[n]}_c$
is $L^{[n]}|_{\xi}\cong L|_{\xi}$. In conclusion, the collection $\{\sigma|_{\xi}\wedge\cdots\wedge\sigma|_{\xi}\}_{\xi}$ defines a section in the line bundle $\mathrm{det}(L^{[n]})$. The map we are looking at $\wedge^nH^0(L)\to H^0(\mathrm{det}(L^{[n]}))$ is deduced from the evaluation map $\mathrm{ev}_n:H^0(L)\otimes \mathcal{O}_{X^{[n]}_c}\to L^{[n]}$, taking $\wedge^n\mathrm{ev}_n$ and applying $H^0$. It is remarkable that \cite{VoisinJEMS,VoisinCOMP,Ellingsrud-Goettsche-LehnJAG}: \begin{thm}[Voisin, Ellingsrud--G\"ottsche--Lehn]
The map\vspace*{-3pt}
\[
H^0(\wedge^n\mathrm{ev}_n):\wedge^nH^0(X,L)\to
H^0\left(X^{[n]}_c,\mathrm{det}(L^{[n]})\right)\vspace*{-3pt}
\]
is an isomorphism. \end{thm}
Since the exterior powers of $H^0(L)$ are building blocks for Koszul cohomology, it is natural to believe that the isomorphism above yields a relation between the Koszul cohomology and the Hilbert scheme. To this end, the Koszul differentials must be reinterpreted in the new context.
There is a natural birational morphism\footnote{We see one advantage of working on $X^{[n]}_c$: subtraction makes sense only for curvilinear subschemes.}\vspace*{-3pt} \[ \tau:\varXi_{n+1}\to X^{[n]}_c\times X,\ (\xi,x)\mapsto(\xi-x,x)\vspace*{-3pt} \] presenting $\varXi_{n+1}$ as the blowup of $X^{[n]}_c\times X$ along $\varXi_n$. If we denote by $D_{\tau}$ the exceptional locus, we obtain an inclusion \cite{VoisinJEMS} \[
q^*\mathrm{det}(L^{[n+1]})\cong\tau^*(\mathrm{det}(L^{[n]})\boxtimes
L)(-D_{\tau})\hookrightarrow \tau^*(\mathrm{det}(L^{[n]})\boxtimes
L) \] whence \[ H^0\left(X^{[n+1]}_c,\mathrm{det}(L^{[n+1]})\right)\hookrightarrow H^0(X^{[n]}_c\times X,\mathrm{det}(L^{[n]})\boxtimes L), \] identifying the left-hand member with the kernel of a Koszul differential \cite{VoisinJEMS}. A~version of this identification leads us to \cite{VoisinJEMS,VoisinCOMP}:
\begin{thm}[Voisin] \label{thm: Hilbert scheme}
For any integers $m$ and $n$, $K_{n,m}(X,L)$ is isomorphic to the
cokernel of the restriction map:
\[
H^0\left(X^{[n+1]}_c\times X,\mathrm{det}(L^{[n+1]})\boxtimes L^{m-1}\right)\to
H^0\left(\varXi_{n+1},\mathrm{det}(L^{[n+1]})\boxtimes L^{m-1}|_{\varXi_{n+1}}\right).
\] \end{thm}
The vanishing of Koszul cohomology is thus reduced to proving surjectivity of the restriction map above. In general, it is very hard to prove surjectivity directly, and one has to make a suitable base-change \cite{VoisinJEMS}.
\subsection{The Role of Lazarsfeld--Mukai Bundles in the Generic Green Conjecture and Consequences} \label{subsection: Role of LM}
In order to prove Green's conjecture for general curves, it suffices to exhibit one example of a curve of maximal Clifford index, which verifies the predicted vanishing. Afterwards, the vanishing of Koszul cohomology propagates by semicontinuity. Even so, finding one single example is a task of major difficulty. The curves used by Voisin in \cite{VoisinJEMS,VoisinCOMP} are $K3$ sections, and the setups change slightly, according to the parity of the genus. For even genus, we have \cite{VoisinJEMS}:
\begin{thm}[Voisin]
\label{thm: Voisin even} Suppose that $g=2k$. Consider $S$ a $K3$ surface with $\mathrm{Pic}(S)\cong \mathbb{Z}\cdot L$, $L^2=2g-2$, and
$C\in|L|$ a smooth curve. Then $K_{k,1}(C,K_C)=0$. \end{thm}
For odd genus, the result is \cite{VoisinCOMP}:
\begin{thm}[Voisin]
\label{thm: Voisin odd} Suppose that $g=2k+1$. Consider $S$ a $K3$ surface with $\mathrm{Pic}(S)\cong \mathbb{Z}\cdot L\oplus \mathbb{Z}\cdot \varGamma$, $L^2=2g-2$, $\varGamma$ a smooth rational curve. $L\cdot \varGamma =2$ and
$C\in|L|$ a smooth curve. Then $K_{k,1}(C,K_C)=0$. \end{thm}
Note that the generic value for the Clifford index in genus $g$ is $[(g-1)/2]$, and hence, in both cases, the prediction made by Green's conjecture for general curve $C$ is precisely $K_{k,1}(C,K_C)=0$.
There are several reasons for making these choices: the curves have maximal Clifford index, by Theorem~\ref{thm: GL Cliff} (and the Clifford dimension is one), the Lazarsfeld--Mukai bundles associated to minimal pencils are $L$-stable, the hyperplane section theorem applies, etc.
We outline here the role played by Lazarsfeld--Mukai bundles in Voisin's proof and, for simplicity, we restrict to the even-genus case. By the \hbox{hyperplane} section Theorem~\ref{thm: Lefschetz}, the required vanishing on the curve is equivalent to \hbox{$K_{k,1}(S,L)=0$.} From the description of Koszul cohomology in terms of Hilbert schemes, Theorem \ref{thm: Hilbert scheme}, adapting the notation from the previous subsection, one has to prove the surjectivity of the map \[ q^*: H^0\left(S_c^{[n+1]},\mathrm{det}(L^{[n+1]})\right)\to
H^0\left(\varXi_{n+1},q^*\mathrm{det}(L^{[n+1]})|_{\varXi_{n+1}}\right).
\] The surjectivity is proved after performing a suitable base-change.
We are in the case $\rho(g,1,k+1)=0$; hence there is a unique Lazarsfeld--Mukai bundle $E$ on $S$ associated to all $g^1_{k+1}$
on curves in $|L|$. The uniqueness yields an alternate description of $E$ as extension \[ 0\to \mathcal{O}_S\to E\to L\otimes I_{\xi}\to 0, \] where $\xi$ varies in $S_c^{[k+1]}$.
There exists a morphism $\mathbb{P} H^0(S,E)\rightarrow S^{[k+1]}$ that sends a global section $s\in H^0(S,E)$ to its zero set $Z(s)$. By restriction to an open subset $\mathbb{P}\subset\mathbb{P} H^0(S,E)$, we obtain a morphism $\mathbb{P}\rightarrow S_c^{[k+1]}$, inducing a commutative diagram \begin{center}
\mbox{\xymatrix{ \mathbb{P}^{\prime} = \mathbb{P}\times_{S_c^{[k+1]}}\varXi_{k+1}\ar[r]\ar[d]^{q^{\prime}} &\varXi_{k+1} \ar[d]^q \\ \mathbb{P} \ar[r] & S_c^{[k+1]}. }}
\end{center} Set-theoretically \[
\mathbb{P}^{\prime} = \{(Z(s),x)|s\in H^0(S,E),x\in Z(s)\}. \]
Unfortunately, this very natural base-change does not satisfy the necessary conditions that imply the surjectivity of $q^*$, \cite{VoisinJEMS}. Voisin modifies slightly this construction and replaces $\mathbb{P}$ with another variety related to $\mathbb{P}$ which parameterizes zero-cycles of the form $Z(s)-x+y$ with $[s]\in\mathbb{P}$, $x\in\mathrm{Supp}(Z(s))$ and $y\in S$. It turns out, after numerous elaborated calculations using the rich geometric framework provided by the Lazarsfeld--Mukai bundle, that the new base-change is suitable and the surjectivity of $q^*$ follows from vanishing results on the Grassmannian \cite{VoisinJEMS}.
In the odd-genus case, Voisin proves first Green's conjecture for smooth curves in $|L+\varGamma|$, which are easily seen to be of maximal Clifford index. The situation on $|L+\varGamma|$ is somewhat close to the setup of Theorem~\ref{thm: Voisin even}, and the proof is similar. The next hard part is to descend from the vanishing of $K_{k+1,1}(S,L\otimes\mathcal{O}_S(\varGamma))$
to the vanishing of $K_{k,1}(S,L)$. This step uses again intensively the unique Lazarsfeld--Mukai bundle associated to any $g^1_{k+2}$ on curves in $|L+\varGamma|$.
The odd-genus case is of maximal interest: mixed with Hirschowitz-Ramanan result \cite{Hirschowitz-RamananAENS}, Theorem~\ref{thm: Voisin odd} gives a solution to Green's conjecture for \textit{any} curve of odd genus and maximal Clifford index:
\begin{thm}[Hirschowitz--Ramanan, Voisin] \label{thm: HRV}
Let $C$ be a smooth curve of odd genus $2k+1\ge 5$
and Clifford index $k$. Then $K_{k,1}(C,K_C)=0$. \end{thm}
Note that Theorem~\ref{thm: HRV} implies the following statement:
\begin{cor} A smooth curve of odd genus and maximal Clifford index has Clifford dimension one. \end{cor}
The proof of Theorem~\ref{thm: HRV} relies on the comparison of two effective divisors on the moduli space of curves $\mathcal{M}_{2k+1}$, one given by the condition $\mathrm{gon}(C)\le k+1$, which is known to be a divisor from \cite{Harris-MumfordINVENTIONES}, and the second given by $K_{k,1}(C,K_C)\ne 0$. By duality $K_{k,1}(C,K_C)\cong K_{k-2,2}(C,K_C)$. Note that $K_{k-2,2}(C,K_C)$ is isomorphic~to\vspace*{-3pt} \[ \mathrm{Coker} \left(\wedge^kH^0(K_C)\otimes H^0(K_C)/\wedge^{k+1}H^0(K_C) \to H^0(\wedge^{k-1}M_{K_C}\otimes K_C^2)\right)\vspace*{-3pt} \] and the two members have the same dimension. The locus of curves with $K_{k,1}\ne 0$ can be described as the degeneracy locus of a morphism between vector bundles of the same dimension, and hence it is a virtual divisor. Theorem~\ref{thm: Voisin odd} implies that this locus is not the whole space, and in conclusion it must be an effective divisor. \hbox{Theorem \ref{thm: GL nonvan}} already gives an inclusion between the supports of two divisors in question, and the set-theoretic equality is obtained from a divisor class calculation~\cite{Hirschowitz-RamananAENS}.
\vspace*{-6pt} \subsection{Green's Conjecture for Curves on $K3$ Surfaces} \label{subsection: Green for K3}
We have already seen that general $K3$ sections have a mild behavior from the Brill--Noether theory viewpoint. In some sense, they behave like general curves in any gonality stratum of the moduli space of curves.
As in the previous subsections, fix a $K3$ surface $S$ and a globally generated line bundle $L$ with $L^2=2g-2$ on $S$, and denote by $k$ the gonality of a general smooth curve in the linear system $|L|$. Suppose that $\rho(g, 1, k)\le 0$ to exclude the case $g=2k-3$
(when $\rho(g, 1, k)=1$). If in addition the curves in $|L|$ have Clifford dimension one, Theorem~\ref{thm: Green Cliffdim 1} shows that\vspace*{-3pt} \[ \mathrm{dim}\ W^1_{g-k+2}(C)=\rho(g,1,g-k+2)=g-2k+2,\vspace*{-3pt} \] property which was called the \textit{linear growth condition}.
This property appears in connection with Green's conjecture~\cite{AproduMRL} for a much larger class of curves:
\begin{thm} \label{thm: Aprodu}
If $C$ is any smooth curve of genus $g\ge 6$ and gonality $3\le k<[g/2]+2$
with $\mathrm{dim}\ W^1_{g-k+2}(C)=\rho(g,1,g-k+2)$,
then $K_{g-k+1,1}(C,K_C)=0$. \end{thm}
One effect of Theorems~\ref{thm: Aprodu} and~\ref{thm: GL nonvan} is that an arbitrary curve that satisfies the linear growth condition is automatically of Clifford dimension one and verifies Green's conjecture.
Theorem~\ref{thm: Aprodu} is a consequence of Theorem~\ref{thm: HRV} extended over the boundary of the moduli space. Starting from a $k$-gonal smooth curve $[C]\in \mathcal{M}_g$, by identifying pairs of general points $\{x_i,y_i\}\subset C$ for $i=0, \dots, g-2k+2$ we produce a stable irreducible curve \[ \left[X:=C/(x_0\sim y_0, \ldots, x_{g-2k+2}\sim y_{g-2k+2})\right] \in \overline{\mathcal{M}}_{2(g-k+1)+1}, \] and the Koszul cohomology of $C$ and of $X$ are related by the inclusion $K_{p,1}(C,K_C)$ $\subset K_{p,1}(X,\omega_X)$ for all $p\ge 1$, \cite{VoisinJEMS}. If $C$ satisfies the linear growth condition then $X$ has maximal gonality\footnote{The gonality for a singular stable curve is defined in terms of admissible covers \cite{Harris-MumfordINVENTIONES}.} $\mathrm{gon}(X)=g-k+3$, i.e., $X$ lies outside the closure of the divisor $\mathcal{M}_{2(g-k+1)+1, g-k+2}^1$ consisting of curves with a pencil $g^1_{g-k+2}$. The class of the failure locus of Green's conjecture on $\overline{\mathcal{M}}_{2(g-k+1)+1}$ is a multiple of the divisor $\overline{\mathcal{M}}_{2(g-k+1)+1, g-k+2}^1$; hence Theorem~\ref{thm: HRV} extends to irreducible stable curves of genus $2(g-k+1)+1$ of maximal gonality $(g-k+3)$. In particular, $K_{g-k+1,1}(X,\omega_X)=0$, implying $K_{g-k+1,1}(C,K_C)=0$.
Coming back to the original situation, we conclude from Theorems~\ref{thm: Aprodu} and~\ref{thm: Green Cliffdim 1} and Corollary~\ref{cor: Lefschetz K3} that Green's conjecture holds for a $K3$ section $C$ having Clifford dimension one. If $\mbox{Cliff}(C)=\mathrm{gon}(C)-3$, either $C$ is a smooth plane curve or else there exist smooth curves $D, \varGamma \subset S$, with $\varGamma^2=-2, \varGamma\cdot D=1$ and $D^2\geq 2$, such that $C\equiv 2D+ \varGamma$ and $\mbox{Cliff}(C)=\mbox{Cliff}(\mathcal{O}_C(D))$ \cite{Ciliberto-PareschiCRELLE,KnutsenIJM}. The linear growth condition is no longer satisfied, and this case is treated differently, by degeneration to a reduced curve with two irreducible components \cite{Aprodu-FarkasCOMP}.
The outcome of this analysis of the Brill--Noether loci is the following \cite{VoisinJEMS,VoisinCOMP,Aprodu-FarkasCOMP}:
\begin{thm} \label{thm: Green on K3} Green's conjecture is valid for any smooth curve on a $K3$ surface. \end{thm}
Applying Theorem~\ref{thm: Green on K3}, Theorem~\ref{thm: Lefschetz}, and the duality, we obtain a full description of the situations when Koszul cohomology of a $K3$ surface is zero \cite{Aprodu-FarkasCOMP}:
\begin{thm} \label{thm: K3} Let $S$ be a $K3$ surface and $L$ a globally generated line bundle with $L^2=2g-2\ge 2$. The Koszul cohomology group $K_{p,q}(S,L)$ is nonzero if and only if one of the following cases occurs: \begin{itemize} \item[(1)] $q=0$ and $p=0$, or \item[(2)] $q=1$, $1\le p\le g-c-2$, or \item[(3)] $q=2$ and $c\le p\le g-1$, or \item[(4)] $q=3$ and $p=g-2$. \end{itemize} \end{thm}
The moral is that the shape of the Betti table, i.e., the distribution of zeros in the table, of a polarized $K3$ surface is completely determined by the geometry of hyperplane sections; this is one of the many situations where algebra and geometry are intricately related.
\section{Counterexamples to Mercat's Conjecture in Rank Two} \label{section: Higher BN}
Starting from Mukai's works, experts tried to generalize the classical Brill--Noether theory to higher-rank vector bundles on curves. Within these extended theories,\footnote{Higher-rank Brill--Noether theory is a major, rapidly growing research field, and it deserves a separate dedicated survey.} we note the attempt to find a proper generalization of the Clifford index. H. Lange and P. Newstead proposed the following definition. Let $E$ be a semistable vector bundle of rank $n$ of degree $d$ on a smooth curve $C$. Put \[
\gamma(E):=\mu(E)-2\frac{h^0(E)}{n}+2. \]
\begin{defn}[Lange--Newstead]
The \textit{Clifford index} of rank $n$ of $C$ is \[
\mathrm{Cliff}_n(C):=\mathrm{min}\{\gamma(E):\ \mu(E)\le g-1,\ h^0(E)\ge 2n\}. \] \end{defn}
From the definition, it is clear that $\mathrm{Cliff}_1(C)=\mathrm{Cliff}(C)$ and $\mathrm{Cliff}_n(C)\le\mathrm{Cliff}(C)$ for all $n$.\footnote{For any line bundle $A$, we have $\gamma(A^{\oplus n})=\mathrm{Cliff}(A)$.}
Mercat conjectured \cite{MercatIJM} that $\mathrm{Cliff}_n(C)=\mathrm{Cliff}(C)$. In rank two, the conjecture is known to hold in a number of cases: for general curves of small gonality, i.e., corresponding to a general point in a gonality stratum $\mathcal{M}^1_{g,k}$ for small $k$ (Lange-Newstead), for plane curves (Lange--Newstead), for general curves of genus $\le 16$ (Farkas--Ortega), etc. However, even in rank two, the conjecture is false. It is remarkable that counterexamples are found for curves of maximal Clifford index~\cite{Farkas-OrtegaIJM}:
\begin{thm}[Farkas--Ortega]
Fix $p\ge 1$, $a\ge 2p+3$. Then there exists a smooth curve of genus $2a+1$ of maximal Clifford index lying on a smooth $K3$ surface $S$ with $\mathrm{Pic}(S)=\mathbb{Z}\cdot C\oplus \mathbb{Z}\cdot H$, $H^2=2p+2$, $C^2=2g-2$, $C\cdot H=2a+2p+1$, and there exists a stable rank-two vector bundle $E$ with $\mathrm{det}(E)=\mathcal{O}_S(H)$ with $h^0(E)=p+3$, $\gamma(E)=a-\frac{1}{2}<a=\mathrm{Cliff}(A)$, and hence Mercat's conjecture in rank two fails for~$C$. \end{thm}
The proof uses restriction of Lazarsfeld--Mukai bundles. However, it is interesting that the bundles are not restricted to the same curves to which they are associated. More precisely, the genus of $H$ is $2p+2$ and $H$ has maximal gonality $p+2$. Consider $A$ a minimal pencil on $H$, and take $E=E_{H,A}$ the associated Lazarsfeld--Mukai bundle. The restriction of $E$ to $C$ is stable and verifies all the required properties.
A particularly interesting case is $g=11$. In this case, as shown by Mukai \cite{MukaiLNPAM}, a general curve $C$ lies on a unique $K3$ surface $S$ such that $C$ generates $\mathrm{Pic}(S)$.\vadjust{\pagebreak} It~is~remarkable that the failure locus of Mercat's conjecture in rank two \textit{coincides} with the Noether-Lefschetz divisor \[ \mathcal{NL}^4_{11,13}:=\left\{[C]\in \mathcal{M}_{11}: \begin{array}{l} C\mbox{ lies on a } K3 \mbox{ surface } S, \ \mathrm{Pic}(S)\supset \mathbb{Z}\cdot C\oplus \mathbb{Z}\cdot H,\\ H\in \mathrm{Pic}(S) \mbox{ is nef},
H^2=6, \ C\cdot H=13, \ C^2=20 \end{array} \right\} \] inside the moduli space $\mathcal{M}_{11}$. We refer to \cite{Farkas-OrtegaIJM} for details.
\end{document}
|
arXiv
|
{
"id": "1205.4415.tex",
"language_detection_score": 0.7767780423164368,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{An alternative framework for quantifying coherence of quantum channels}
\author{Shi-Yun Kong$^1$} \author{Ya-Juan Wu$^2$}
\author{Qiao-Qiao Lv$^1$} \author{Zhi-Xi Wang$^1$}
\author{Shao-Ming Fei$^{1,3}$}
\affiliation{ {\footnotesize $^1$School of Mathematical Sciences, Capital Normal University, Beijing 100048, China}\\ {\footnotesize $^2$School of Mathematics, Zhengzhou University of Aeronautics, Zhengzhou 450046, China}\\ {\footnotesize $^3$Max-Planck-Institute for Mathematics in the Sciences, 04103 Leipzig, Germany} }
\begin{abstract} We present an alternative framework for quantifying the coherence of quantum channels, which contains three conditions: the faithfulness, nonincreasing under sets of all the incoherent superchannels and the additivity. Based on the characteristics of the coherence of quantum channels and the additivity of independent channels, our framework provides an easier way to certify whether a function is a bona fide coherence measure of quantum channels. Detailed example is given to illustrate the effectiveness and advantages of our framework. \end{abstract}
\maketitle
\section{Introduction} As one of the characteristic features that marks the departure of quantum mechanics from the classical realm, the quantum coherence plays a central role in quantum optics \cite{glauber,scully,gyongyosi}, thermodynamics \cite{brandao,gour,narasimhachar,aberg2,lostaglio1,lostaglio2,gentaro,mayuhan,xushengzhi}, nanoscale physics \cite{plenio,rebentrost,licheming,huelga} and quantum measurements \cite{napoli,mateng,longguilu}. Quantum coherence is also the origin of many quantum phenomena such as entanglement and multiparticle interference. Recently the coherence of quantum states, concerning the quantifications \cite{baumgratz,girolami}, interconversion \cite{bromley} and applications \cite{luo,aberg,monras}, has been extensively investigated.
Similar to the resource theory of quantum entanglement, Baumgratz, Cramer and Plenio presented a rigorous framework (BCP framework) for quantifying the coherence of quantum states and introduced several measures of coherence including the relative entropy of coherence, the $l_1-$norm of coherence and fidelity \cite{baumgratz}. The BCP framework is widely used in quantifying coherence. Inspired by the BCP framework, Yu $et.~al.$ put forward another equivalent framework \cite{yuxiaodong}, which can be more conveniently used in some cases and is applicable to various physical contexts.
The quantum state transfer depends on quantum channels. The coherence of quantum channels characterizes the ability to optimize the coherence of all output states of the channels \cite{xueyuanhu,bendana,korzekawa,theurer,chandan}. Similar to the resource theory of coherence for quantum states, the resource theory of coherence for quantum channels has attracted much attention. With respect to the BCP framework for coherence of quantum states, Xu established a framework for quantifying the coherence of quantum channels \cite{xujianwei}.
In this paper, similar to the alternative framework for quantum states given in \cite{yuxiaodong} which simplifies the BCP framework in particular cases, we establish a new framework for quantifying the coherence of quantum channels, which improves the applications of the previous framework given in \cite{xujianwei}. Detailed examples are presented to illustrate the advantages of our framework.
\maketitle
\section{An alternative framework for coherence of quantum channels}
Let $H_A$ and $H_B$ be Hilbert spaces with dimensions ${\rm dim} H_A=|A|$ and ${\rm dim} H_B=|B|$, and $\{|j\rangle\}_j,~\{|k\rangle\}_k$ and $\{|\alpha\rangle\}_\alpha,~\{|\beta\rangle\}_\beta$ be the fixed orthonormal bases of $H_A$ and $H_B$, respectively. Denote $\mathcal{D}_A$ ($\mathcal{D}_B$) the set of all density operators on $H_A$ ($H_B$). Let $\mathcal{C}_{AB}$ be the set of all channels from $\mathcal{D}_A$ to $\mathcal{D}_B$.
A quantum channel $\phi\in \mathcal{C}_{AB}$ is a linear completely positive and trace preserving (CPTP) map with Kraus operators $\{K_n\}_n$ satisfying $\sum_n K_n^\dagger K_n=I$ such that $\phi(\rho)=\sum_n K_n\rho K_n^\dagger$. The corresponding Choi matrix with respect to the channel $\phi\in\mathcal{C}_{AB}$ has the following form, \begin{equation}\label{W1}
J_{\phi}=\sum_{jk}|j\rangle\langle k|\otimes\phi(|j\rangle\langle k|)=\sum_{jk,\alpha\beta}\phi_{jk\alpha\beta}|j\rangle\langle k|\otimes|\alpha\rangle\langle\beta|, \end{equation}
where $\phi_{jk\alpha\beta}=\langle\alpha|\phi(|j\rangle\langle k|)|\beta\rangle$ are complex numbers and $\sum_{\alpha}\phi_{jk,\alpha\alpha}=\delta_{jk}$. If $C$ is a coherence measure of quantum states, then $C(\phi)=C(\frac{J_\phi}{|A|})$ is a corresponding coherence measure for quantum channels \cite{xujianwei}, which quantifies the coherence of the channel by the coherence of the state $\frac{J_\phi}{|A|}$. The Choi matrix of a channel gives a relation between the coherence measures for quantum channels and for that quantum states.
Let $\mathcal{I}$ be the set of all incoherent states whose density matrices are diagonal in the given basis. An incoherent state $\rho\in H_A$ satisfies $\Delta_A(\rho)=\rho$ with $\Delta_A(\rho)=\sum_j\langle j|\rho|j\rangle|j\rangle\langle j|$ being a completely dephasing channel. $\phi\in\mathcal{C}_{AB}$ is called an incoherent quantum channel if $\Upsilon(\phi)=\phi$, where $\Upsilon(\phi)=\Delta_B \phi \Delta_A$, $\Delta_A$ and $\Delta_B$ are resource destroying maps\cite{liuziwen}. We denote $\mathcal{IC}$ the set of all the incoherent channels. Let $\mathcal{SC}_{ABA'B'}$ be the superchannels that are linear maps from $\mathcal{C}_{AB}$ to $\mathcal{C}_{A'B'}$, we define the Choi matrix of the superchannel $\Theta\in\mathcal{SC}_{ABA'B'}$ as $J_\Theta=\sum_{jk\alpha\beta}|j\alpha\rangle\langle k\beta|\otimes\Theta(|j\alpha\rangle\langle k\beta|)$. Any superchannel $\Theta$ is an incoherent superchannel ($\mathcal{ISC}$) if there exists an expression of Kraus operators $\Theta=\{M_m\}_m$ such that for each $m$, $M_m=\sum_{j\alpha}M_{mj\alpha}|f(j\alpha)\rangle\langle j\alpha|$ with $f(j\alpha)=f(j,\alpha)\in \{(j',\alpha')|_{j'=1}^{|A'|},_{\alpha'=1}^{|B'|}\}$.
In \cite{xujianwei}, the author presented a framework for quantifying the coherence of quantum channels. A proper measure $C$ of the coherence for quantum channels must satisfy the following conditions: \begin{enumerate} \item[$\mathrm{(B1)}$] For any quantum channel $C(\phi)\geqslant0$, $C(\phi)=0$ if and only if $\phi\in\mathcal{IC}_{AB}$; \item[$\mathrm{(B2)}$] $C(\phi)\geqslant C[\Theta(\phi)]$ for any incoherent superchannel $\Theta$;
\item[$\mathrm{(B3)}$] $C(\phi)\geqslant\sum_m p_mC(\phi_m)$ for any incoherent superchannel $\Theta$, with $\{M_m\}_m$ an incoherent Kraus operator of $\Theta$, $p_m=\frac{\mathrm{tr}(M_m J_\phi M_m^\dagger)}{|A'|}$, and $J_{\phi_m}=|A'|\frac{M_m J_\phi M_m^\dagger}{\mathrm{tr}(M_m J_\phi M_m^\dagger)}$; \item[$\mathrm{(B4)}$] $C(\sum_m p_m \phi_m)\leqslant\sum_m p_m C(\phi_m)$ for any set of quantum channels $\{\phi_m\}$ and any probability distribution $\{p_m\}_m$. \end{enumerate} The items (B1)-(B4) give necessary conditions for a bona fide measure of coherence for quantum channels, from which the quantification of the coherence for quantum channels has been further investigated.
Nevertheless, similar to the case of coherence measures for quantum states, the last two conditions (B3) and (B4) are rather difficult to be verified for a given measure of coherence for quantum channels. In the following we present an alternative framework consisting of three conditions, which is equivalent to the above framework but can be easily applied. A function $C$ is a well defined coherence measure of quantum channels if it satisfies the following three conditions:
\begin{enumerate} \item[$\mathrm{(C1)}$] $\mathit{Faithfulness}$. $C(\phi)\geqslant0$ for any $\phi\in \mathcal{C}_{AB}$, and $C(\phi)=0$ if and only if $\phi\in\mathcal{IC}_{AB}$; \item[$\mathrm{(C2)}$] $\mathit{Nonincreasing\ under\ ISCs}.$ $C(\phi)\geqslant C[\Theta(\phi)]$ for any $\Theta\in\mathcal{ISC}_{ABA'B'}$;
\item[$\mathrm{(C3)}$] $\mathit{Additivity}$. $C(\Phi)=p_1C(\phi_1)+p_2C(\phi_2)$ for $p_1+p_2=1$, $\phi_1\in \mathcal{C}_{AB_1}$ and $\phi_2\in \mathcal{C}_{AB_2}$, where $\Phi(|j\rangle\langle k|)=p_1\phi_1(|j\rangle\langle k|)\oplus p_2\phi_2(|j\rangle\langle k|)$, $\Phi\in \mathcal{C}_{AB}$,\ and $|B|=|B_1|+|B_2|$. \end{enumerate}
In the following, we prove that the framework given by (B1)-(B4) is equivalent to the one given by (C1)-(C3). We first prove that (B1)-(B4) give rise to (C1)-(C3), namely, (B3)(B4) give rise to (C3) since (C1) and (C2) are the same as (B1) and (B2). Consider a CPTP map $\Theta_1\in\mathcal{ISC}_{ABAB}$, $\Theta_1(\cdot)=Q_1\cdot Q_1^\dagger+Q_2\cdot Q_2^\dagger$, where \begin{equation} \begin{aligned}
Q_1=&|0\rangle\la0|+\cdots+||B_1|-1\rangle\langle |B_1|-1|+||B_1|+|B_2|\rangle\langle |B_1|+|B_2||+\cdots+|2|B_1|+|B_2|-1\rangle\\
&\langle 2|B_1|+|B_2|-1|+\cdots+|(|A|-1)(|B_1|+|B_2|)\rangle\langle(|A|-1)(|B_1|+|B_2|)|+\cdots\\
&+|(|A|-1)(|B_1|+|B_2|)+|B_1|-1\rangle\langle(|A|-1)(|B_1|+|B_2|)+|B_1|-1| \end{aligned} \end{equation} and \begin{equation} \begin{aligned}
Q_2=&||B_1|\rangle\langle |B_1||+\cdots+||B_1|+|B_2|-1\rangle\langle |B_1|+|B_2|-1|+|2|B_1|+|B_2|\rangle\la2|B_1|+|B_2||+\cdots\\
&+|2(|B_1|+|B_2|)-1\rangle\la2(|B_1|+|B_2|)-1|+\cdots+||A||B_1|+(|A|-1)|B_2|\rangle\\
&\langle |A||B_1|+(|A|-1)|B_2||+\cdots+||A|(|B_1|+|B_2|)-1\rangle\langle |A|(|B_1|+|B_2|)-1|. \end{aligned} \end{equation} Note that $Q_1$ and $Q_2$ are just projectors onto $\mathcal{C}_{AB}$. Obviously, one sees that $Q_i\mathcal{IC}Q_i^\dagger\subset\mathcal{IC}$. The Choi matrix of $\Phi$ in (C3) is given by \begin{equation}\label{W3}
J_\Phi=\sum_{j,k}|j\rangle\langle k|\otimes[p_1\phi_1(|j\rangle\langle k|)\oplus p_2\phi_2(|j\rangle\langle k|)]. \end{equation} Then \begin{equation} \Theta_1(\Phi)=Q_1J_\Phi Q_1^\dagger+Q_2J_\Phi Q_2^\dagger=p_1J_{\tilde{\phi}_1}+p_2J_{\tilde{\phi}_2}, \end{equation}
where $p_1=\frac{\mathrm{tr}(Q_1J_\Phi Q_1^\dagger)}{|A|}$, $p_2=\frac{\mathrm{tr}(Q_2J_\Phi Q_2^\dagger)}{|A|}$ and $J_{\tilde{\phi}_1}=|A|\frac{Q_1J_\Phi Q_1^\dagger}{\mathrm{tr}(Q_1J_\Phi Q_1^\dagger)}$, $J_{\tilde{\phi}_2}=|A|\frac{Q_2J_\Phi Q_2^\dagger}{\mathrm{tr}(Q_2J_\Phi Q_2^\dagger)}$. From (B2) and (B3) we have \begin{equation}\label{W4} C(\Phi)\geqslant p_1C(\tilde{\phi_1})+p_2C(\tilde{\phi_2}), \end{equation}
where $\tilde{\phi}_1,\,\tilde{\phi}_2\in\mathcal{C}_{AB}$, $\tilde{\phi}_1(|j\rangle\langle k|)=\phi_1(|j\rangle\langle k|)\oplus{\bf0}(|j\rangle\langle k|)$ and $\tilde{\phi}_2(|j\rangle\langle k|)={\bf0}(|j\rangle\langle k|)\oplus\phi_2(|j\rangle\langle k|)$ with ${\bf0}$ a zero map. From (B4) we have \begin{equation}\label{W5} C(\Phi)\leqslant p_1C(\tilde{\phi_1})+p_2C(\tilde{\phi_2}). \end{equation} Combining \eqref{W4} with \eqref{W5}, we get \begin{equation}\label{W6} C(\Phi)=p_1C(\tilde{\phi_1})+p_2C(\tilde{\phi_2}). \end{equation}
To obtain (C3), we need to certify $C(\tilde{\phi_1})=C(\phi_1)$ further. Under an incoherent superchannel, any $\phi$ and $\overline{\Theta}(\phi)=\sum P_nJ_\phi P_n^\dagger$ can be transformed into each other, where the Kraus operators $\{P_n\}_n$ are permutation matrices. By (B2) we have $C(\overline{\Theta}(\phi))=C(\phi)$. Based on this fact, we define an incoherent superchannel $\overline{\Theta}_1\in\mathcal{ISC}_{ABAB}$ with Kraus operators $\{P_{nl}^{(1)}\}_{nl}$ as the permutation matrices, \begin{equation} P_{nl}^{(1)}(i,j)=\left\{ \begin{aligned} &1,~{\rm if}~(i, j)=(i_{nl}, j_{nl}) ~{\rm or} ~(i, j)=(j_{nl}, i_{nl}),\\ &1,~{\rm if}~i=j,\\ &0,~\rm{other wise}, \end{aligned}\right. \end{equation}
where $i_{nl}=(n-1)(|B_1|+|B_2|)+l$, $j_{nl}=(n-1)|B_1|+l$, $i,j=1,\cdots,|A||B|$, $n=1,\cdots,|A|$ and $l=1,\cdots,|B_1|$. Then, \begin{equation}\overline{\Theta}_1(\tilde{\phi}_1)=\sum_{n,l}P_{nl}^{(1)}J_{\tilde{\phi}_1}P_{nl}^{(1)\dagger}=J_{\phi_1}\oplus O_{AB_2},\end{equation}
where $O_{AB_2}$ is a $|A||B_2|\times|A||B_2|$ null matrix. It is easily seen that \begin{equation}\label{W7} C(\tilde{\phi}_1)=C(\overline{\Theta}_1(\tilde{\phi}_1)). \end{equation}
Next, we need to prove $C(\overline{\Theta}_1(\tilde{\phi}_1))=C(\phi_1)$. For this, we define two incoherent superchannels: $\Theta_2\in\mathcal{ISC}_{AB_1AB}$ with Kraus operator $M_0$ satisfying $\langle j|M_0|k\rangle=\delta_{jk}$ and $\Theta_3\in\mathcal{ISC}_{ABAB_1}$ with Kraus operators $\{M_n\}_{n=0}^{\lceil\frac{|B_2|}{|B_1|}\rceil}$ satisfying $\langle j|M_n|k\rangle=\delta_{j,k-n|B_1|}$. Then we get, \begin{equation}~\Theta_2(\phi_1)=M_0J_{\phi_1}M_0^\dagger=J_{\phi_1}\oplus O_{AB_2}\end{equation} and \begin{equation} \Theta_3[(\overline{\Theta}_1(\tilde{\phi}_1)]
=\sum_{n=0}^{\lceil\frac{|B_2|}{|B_1|}\rceil}M_n(J_{\phi_1}\oplus O_{AB_2})M_n^\dagger=J_{\phi_1}. \end{equation} From (B2) we obtain \begin{equation}\label{6jia} C(\overline{\Theta}_1(\tilde{\phi}_1))=C(\phi_1). \end{equation} Combining \eqref{6jia} with \eqref{W6} and \eqref{W7}, we get the condition (C3).
We have shown that any $C$ satisfying (B1)-(B4) also satisfies (C1)-(C3). Next, we prove that any $C$ satisfying (C1)-(C3) must satisfy (B3) and (B4). First, we prove (B3), i.e., $C$ is convex. Define $\Phi_1\in\mathcal{C}_{AB'}$ as,
\begin{equation}\Phi_1(|j\rangle\langle k|)=\phi(|j\rangle\langle k|)\oplus {\bf 0}(|j\rangle\langle k|)\oplus\cdots\oplus {\bf 0}(|j\rangle\langle k|),\end{equation} where $\phi\in\mathcal{C}_{AB},\ H_{B'}=\underbrace{H_B\otimes H_B\otimes\cdots\otimes H_B}_{M}.$ From (C3), we have \begin{equation}\label{W8} C(\Phi_1)=C(\phi). \end{equation} Consider $\overline{\Theta}_2\in\mathcal{ISC}_{AB'AB'}$, with its Kraus operators $\{P_n^{(2)}\}_n$ being the permutation matrices, such that $\overline{\Theta}_2(\Phi_1)=\sum_n P_n^{(2)}J_{\Phi_1}P_n^{(2)\dagger}=J_\phi\oplus \underbrace{O_{AB}\oplus\cdots\oplus O_{AB}}_{M-1}$. Apply an incoherent superchannel $\Theta_4\in \mathcal{SC}_{ABA_1B'_{1}}$ with Kraus operators $\{U_m\otimes M_m\}_m$ such that \begin{equation} \Theta_4[\overline{\Theta}_2(\Phi_1)]=\sum_{m=0}^{M-1}(U_m\otimes M_m)(\sum_nP_n^{(2)}J_{\Phi_1}P_n^{(2)\dagger})(U_m\otimes M_m)^\dagger, \end{equation}
where $U_m=\sum_{k=0}^{M-1}|(k+m)~\mathrm{mod}~M\rangle\langle k|,~H_{B'_1}=\underbrace{H_{B_1}\otimes H_{B_1}\otimes\cdots\otimes H_{B_1}}_{M}$, and $\{M_m\}$ are incoherent Kraus operators of the superchannel in $\mathcal{ISC}_{ABA_1B_1}$. One can easily see that $(U_m\otimes M_m)\mathcal{IC}(U_m\otimes M_m)^\dagger \subset\mathcal{IC}$, \begin{equation}
\Theta_4[\overline{\Theta}_2(\Phi_1)]=\sum_{m=0}^{M-1}p_m |m\rangle\langle m|\otimes J_{\phi_m}, \end{equation}
where $p_m=\frac{\rm{tr}(M_mJ_\phi M_m^\dagger)}{|A_1|}$ and $J_{\phi_m}=|A_1|\frac{M_mJ_\phi M_m^\dagger}{\rm{tr}(M_mJ_\phi M_m^\dagger)}$.
Similarly, there exists $\overline{\Theta}_3\in\mathcal{ISC}_{A_1B_1'A_1B_1'}$, with its Kraus operators $\{P_n^{(3)}\}_n$ being the permutation matrices, such that $\overline{\Theta}_3[\Theta_4(\overline{\Theta}_2(\Phi_1))]=\sum P_n^{(3)}(\sum_{m=0}^{M-1}p_m |m\rangle\langle m|\otimes J_{\phi_m})P_n^{(3)\dagger}=\sum_{i,j=0}^{M-1}|i\rangle\langle j|\otimes[p_0\phi_0(|i\rangle\langle j|)\oplus\cdots\oplus p_{M-1}\phi_{M-1}(|i\rangle\langle j|)]$. $\overline{\Theta}_3[\Theta_4(\overline{\Theta}_2(\Phi_1))]$ matches to a channel $\Phi_2\in C_{A_1{B_1}'},~\Phi_2(|i\rangle\langle j|)=p_0\phi_0(|i\rangle\langle j|)\oplus\cdots\oplus p_{M-1}\phi_{M-1}(|i\rangle\langle j|)$. \\Following (C3), we have \begin{equation}\label{W9} C(\Phi_2)=\sum_{m=0}^{M-1}p_m C(\phi_m). \end{equation} By (C2), \eqref{W8} and \eqref{W9} we can have\begin{equation}C(\phi)\geqslant\sum_{m=0}^{M-1}p_m C(\phi_m), \end{equation} which proves that (B3) holds.
We now prove (B4). We first define an initial channel $\Phi_3\in C_{AB'}$ satisfying \begin{equation}\Phi_3(|j\rangle\langle k|)=\oplus_{m=0}^{M-1}p_m\phi_m(|j\rangle\langle k|),\end{equation} where $\phi_m\in\mathcal{C}_{AB},~\{p_m\}$ are the probability distribution of $\{\phi_m\}$ and $\sum_{m=0}^{M-1}p_m=1$. According to (C3), one has \begin{equation}\label{W10} C(\Phi_3)=\sum_{m=0}^{M-1}p_m C(\phi_m). \end{equation} Apply $\overline{\Theta}_4\in\mathcal{ISC}_{AB'AB'}$, with its Kraus operators being the permutation matrices $\{P_n^{(4)}\}$, such that $\overline{\Theta}_4(\Phi_3)=\sum_n P_n^{(4)} J_{\Phi_3} P_n^{(4)\dagger}=\oplus_{m=0}^{M-1}p_m J_{\phi_m}$. Let $\Theta_5\in\mathcal{ISC}_{AB'AB'}$ be an incoherent super channel such that \begin{equation}\begin{aligned}
\Theta_5\overline{\Theta}_4(\Phi_3)&=\sum_{m=0}^{M-1}(|0\rangle\langle m|\otimes I)(\sum_n P_n^{(4)} J_{\Phi_3} P_n^{(4)\dagger})(|0\rangle\langle m|\otimes I)^\dagger\\ &=\sum_{m=0}^{M-1}p_mJ_{\phi_m}\oplus O_{A_1B_1}\oplus\cdots\oplus O_{A_1B_1}.\end{aligned} \end{equation} Apply $\overline{\Theta}_5\in\mathcal{ISC}_{AB'AB'}$, with Kraus operators $\{P_n^{(5)}\}$ as permutation matrices, such that \begin{equation}\begin{aligned} \overline{\Theta}_5[\Theta_5\overline{\Theta}_4(\Phi_3)]&=\sum_n P_n^{(5)}(\sum_{m=0}^{M-1}p_mJ_{\phi_m}\oplus O_{A_1B_1}\oplus\cdots\oplus O_{A_1B_1})P_n^{(5)\dagger}\\
&=\sum_{j,k=0}^{|A|-1}|j\rangle\langle k|\otimes[\sum_{m=0}^{M-1}p_m\phi_m(|j\rangle\langle k|)\oplus{\bf 0}(|j\rangle\langle k|)\oplus\cdots\oplus{\bf0}(|j\rangle\langle k|)]. \end{aligned}\end{equation}
Thus, $\overline{\Theta}_5[\Theta_5\overline{\Theta}_4(\Phi_3)]$ corresponds to $\Phi_4\in\mathcal{C}_{AB'}$ with $\Phi_4(|j\rangle\langle k|)=\sum_{m=0}^{M-1}p_m\phi_m(|j\rangle\langle k|)\oplus{\bf0}(|j\rangle\langle k|)\oplus\cdots\oplus{\bf0}(|j\rangle\langle k|)$. From (C3) we have \begin{equation}\label{W11} C(\Phi_4)=C(\sum_{m=0}^{M-1}p_m\phi_m). \end{equation} Combining (C2) with \eqref{W10} and \eqref{W11}, we get \begin{equation} \sum_{m=0}^{M-1}p_mC(\phi_m)\geqslant C(\sum_{m=0}^{M-1}p_m\phi_m), \end{equation} namely, (B4) holds.
We usually get coherence measures for quantum channels from corresponding coherence measures for quantum states. For instance, the $l_1-$norm of coherence $C_{l_1}(\rho)=\sum_{i\ne j}|\rho_{i,j}|$\cite{baumgratz} and the relative entropy of coherence $C_{\rm rel.}(\rho)=S(\rho_{\rm diag})-S(\rho),$ where $S$ is the von Neumann entropy and $\rho_{\rm diag}$ denotes the state obtained from $\rho$ by deleting all off-diagonal elements\cite{baumgratz}, are coherence measures for quantum states, on this basis, $C_{l_1}(\phi)=\sum_{i\ne j}|\frac{J_\phi}{|A|}|$ and $C_{\rm rel.}(\phi)=S(\phi_{\rm diag})-S(\phi)=S(\frac{J_{\phi_{\rm diag}}}{|A|})-S(\frac{J_\phi}{|A|})$ both are coherence measures for quantum channels\cite{xujianwei}.
The above proof shows that our new framework is equivalent to the framework given by (B1)-(B4) for quantum channels. In determining whether a function $C$ can be used as a coherence measure for channels, in some cases, it is not easy to verify whether $C$ satisfies (B3). The condition (C3) in our framework provides a new way to solve the problem. We give an example to show the efficiency of our framework.
{\bf Example}
The trace distance measure of coherence defined by $C_{\rm tr}(\rho):=\min_{\delta \in \cal{I}} \|\rho-\delta\|_{\rm tr}=\min_{\delta \in \cal{I}}{\rm tr}|\rho-\delta|$ is not a well defined coherence measure for quantum states \cite{yuxiaodong}. Let us check whether $C_{\rm tr}(\phi)=C_{\rm tr}(\frac{J_\phi}{|A|})$ is a bona fide coherence measure for quantum channels or not. Here we define \begin{equation}
C_{\rm tr}(\phi):=\min_{{\tilde{\phi}}\in \cal{IC}}\|\phi-\tilde{\phi}\|_{\rm tr}, \end{equation}
where $\|\phi-\tilde{\phi}\|_{\rm tr}=\|\frac{J_\phi}{|A|}-\frac{J_{\tilde{\phi}}}{|A|}\|_{\rm tr}={\rm tr}|\frac{J_\phi}{|A|}-\frac{J_{\tilde{\phi}}}{|A|}|$ is the trace norm between $\phi$ and $\tilde{\phi}$ with $\phi,\,\tilde{\phi}\in \mathcal{C}_{AB}$. We need to verify that $C_{\rm tr}(\phi)$ satisfies either (B1)-(B4) or (C1)-(C3). It has been already proved in previous works \cite{baumgratz,bromley} that $C_{\rm tr}$ satisfies (B1), (B2) and (B4). However, the verification of (B3) is rather difficult. The inequality can only be fulfilled for qubit and $X$ quantum states \cite{shaolianhe,Rana}.
We use condition (C3) to verify the validity of $C_{\rm tr}(\phi)$. For the isometry channel $\phi_{\rm max}\in \mathcal{C}_{AB}$ \cite{xujianwei},
\begin{equation}\phi_{\rm max}(|j\rangle\langle k|)=\frac{1}{|B|}\sum_{\alpha, \beta=0}^{|B|-1}e^{i(\theta_{j\alpha}-\theta_{k\beta})}|\alpha\rangle\langle\beta|,\end{equation} we have
\begin{equation}C_{\rm tr}(\phi_{\rm max})=\min_{\tilde{\phi}\in\mathcal{IC}}\|\frac{J_{\phi_{\rm max}}}{|A|}-\frac{J_{\tilde{\phi}}}{|A|}\|_{\rm tr},\end{equation} where
\begin{equation}\frac{J_{\phi_{\rm max}}}{|A|}=\frac{1}{|A||B|}\sum_{j,k=0}^{|A|-1}|j\rangle\langle k|\otimes(\sum_{\alpha,~\beta=0}^{|B|-1}e^{i(\theta_{j\alpha}-\theta_{k\beta})}|\alpha\rangle\langle\beta|)=|\psi\rangle\langle\psi|, \end{equation}
with $|\psi\rangle=\frac{1}{\sqrt{|A||B|}}\sum_{j=0}^{|A|-1}\sum_{\alpha=0}^{|B|-1}e^{i\theta_{j\alpha}}|j\alpha\rangle$.
Set $U_n=\sum_{k=0}^{|A||B|-1}e^{i(\theta_{(k+n)~\rm mod~|A||B|}-\theta_{k})}|(k+n)~{\rm mod}~|A||B|\rangle\langle k|$. Then we have $U_n|\psi\rangle=|\psi\rangle$. Since $\|A\|_{\rm tr}+\|B\|_{\rm tr}\geqslant\|A+B\|_{\rm tr}$ and $\|U_n|\psi\rangle\|_{\rm tr}=~\||\psi\rangle\|_{\rm tr}$ for the unitary operation $U_n$, we obtain \begin{equation} \begin{array}{rcl}
\|\frac{J_{\phi_{\rm max}}}{|A|}-\frac{J_{\tilde{\phi}}}{|A|}\|_{\rm tr}&=&\frac{1}{|A||B|}\sum_{n=0}^{|A||B|-1}\|U_n(\frac{J_{\phi_{\rm max}}}{|A|}-\frac{J_{\tilde{\phi}}}{|A|}){U_n}^\dagger\|_{\rm tr}\\[1mm]
&\geqslant&\frac{1}{|A||B|}\|\sum_{n=0}^{|A||B|-1}(U_n(\frac{J_{\phi_{\rm max}}}{|A|}-\frac{J_{\tilde{\phi}}}{|A|}){U_n}^\dagger)\|_{\rm tr}. \end{array} \end{equation} As \begin{equation}
U_n\frac{J_{\phi_{\rm max}}}{|A|}{U_n}^\dagger=U_n|\psi\rangle\langle\psi|{U_n}^\dagger=\frac{J_{\phi_{\rm max}}}{|A|} \end{equation} and
\begin{equation}\sum_{n=0}^{|A||B|-1}U_n\frac{J_{\tilde{\phi}}}{|A|}{U_n}^\dagger=I_{|A||B|}, \end{equation} we have \begin{equation}
\|\frac{J_\phi}{|A|}-\frac{J_{\tilde{\phi}}}{|A|}\|_{\rm tr}\geqslant\|\frac{J_\phi}{|A|}-\frac{1}{|A||B|}I_{|A||B|}\|_{\rm tr}. \end{equation} Therefore, \begin{equation}\label{12}
\min_{\tilde{\phi}\in\mathcal{IC}}\|\frac{J_{\phi_{\rm max}}}{|A|}-\frac{J_{\tilde{\phi}}}{|A|}\|_{\rm tr}=\|\frac{J_{\phi_{\rm max}}}{|A|}-\frac{1}{|A||B|}I_{|A||B|}\|_{\rm tr}=\frac{2(|A||B|-1)}{|A||B|}. \end{equation}
Next, we consider a specific channel $\phi\in \mathcal{C}_{AB}$, \begin{equation}
\phi(|j\rangle\langle k|)=\frac{1}{2}\phi_1(|j\rangle\langle k|)\oplus\frac{1}{2}\phi_2(|j\rangle\langle k|), \end{equation} where \begin{equation}
\phi_1(|j\rangle)=\frac{1}{\sqrt{2}}\sum_{\alpha=0}^{1}e^{i\theta_{j\alpha}}|\alpha\rangle,
~~~\phi_2(|j\rangle)=\frac{1}{\sqrt{3}}\sum_{\beta=0}^{2}e^{i\theta_{j\beta}}|\beta\rangle, \end{equation}
with $\phi_1\in\mathcal{C}_{AB_1}$ and $\phi_2\in\mathcal{C}_{AB_2}$ the isometry channels, $|A|=2,~|B_1|=2,~|B_2|=3$ and $|B|=5$. In particular, we take the incoherent channel $\phi_0\in C_{AB},~\phi_0(|j\rangle\langle k|)=\frac{1}{4}\delta_{jk}(|j\rangle\langle k|)\oplus \mathbf{0}(|j\rangle\langle k|)$. Then \begin{equation}
C_{\rm tr}(\phi)=\displaystyle\min_{\tilde{\phi}\in\mathcal{IC}}\|\frac{J_\phi}{2}-\frac{J_{\tilde{\phi}}}{2}\|_{\rm tr}\leqslant\|\frac{J_\phi}{2}-\frac{J_{\phi_0}}{2}\|_{\rm tr}. \end{equation}
From (10) we get $\frac{1}{2}C_{\rm tr}(\phi_1)+\frac{1}{2}C_{\rm tr}(\phi_2)=\frac{19}{12}$. However, $C_{\rm tr}(\phi)\leqslant\|\frac{J_\phi}{2}-\frac{J_{\phi_0}}{2}\|_{\rm tr}=1$. Obviously, $C_{\rm tr}(\phi)\ne\frac{1}{2}C_{\rm tr}(\phi_1)+\frac{1}{2}C_{\rm tr}(\phi_2)$. Therefore, the trace norm of coherence is not a well defined coherence measure of quantum channels. In other words, it also violates (B3). Here, inspired by the definition of trace norm, one may propose a similar trace norm function ${C_{\rm tr}}'(\phi)=\displaystyle\min_{\lambda\geqslant0,
~\tilde{\phi}\in\mathcal{IC}}\|\phi-\lambda\tilde{\phi}\|_{\rm tr}$, which can be shown to be a legal coherent measure for quantum channels \cite{yuxiaodong,xujianwei}.
We have studied the coherence of quantum channels based on the corresponding Choi matrices to the quantum channels. Note that $p_1J_{\phi_1}\oplus p_2J_{\phi_2}$ is not necessarily a Choi matrix for arbitrary channels $\phi_1\in \mathcal{C}_{AB_1}$ and $\phi_2\in \mathcal{C}_{AB_2}$. From (1) the Choi matrix corresponding to a channel $\phi\in \mathcal{C}_{AB}$ is a $|A||B|\times|A||B|$ positive definite matrix where each $\phi(|j\rangle\langle k|)$ is a $|B|\times|B|$ block matrix and ${\rm tr}(\phi(|j\rangle\langle j|))=1$. Assuming that there is a channel $\Phi\in \mathcal{C}_{AB}$ such that $J_{\Phi}=p_1J_{\phi_1}\oplus p_2J_{\phi_2}$ with $|B|=|B_1|+|B_2|$. With respect to the matrix $p_iJ_{\phi_i}$, each $p_i\phi_i(|j\rangle\langle k|)$ is a $|B_i|\times|B_i|$ block matrix and ${\rm tr}[p_i\phi_i(|j\rangle\langle j|)]=p_i$, $i=1,2$. One can see that the trace of the $|B|\times|B|$ block matrix on all diagonals cannot always be 1 for arbitrary probability $p_1$ and $p_2$. In other words, $p_1J_{\phi_1}\oplus p_2J_{\phi_2}$ is not necessarily a Choi matrix, as it does not satisfy the structure of the Choi matrix corresponding to the channels.
\section{Conclusions}
We have presented an alternative framework to quantify the coherence of quantum channels. Our framework and the framework given by (B1)-(B4) for quantum channels are equivalent. We have used this framework to certify the validity of the trace norm coherence measure for quantum channels. Similar to the case for the coherence measure of quantum states \cite{yuxiaodong}, our framework has the similar unique superiorities and may significantly simplify the quantification for coherence of quantum channels. Our results may highlight further investigations on the resource theory of quantum channels.
\noindent{\bf Acknowledgments}\, \, This work is supported by NSFC (Grant Nos. 12075159 and 12171044), Beijing Natural Science Foundation (Z190005), Academy for Multidisciplinary Studies, Capital Normal University, the Academician Innovation Platform of Hainan Province, and Shenzhen Institute for Quantum Science and Engineering, Southern University of Science and Technology (No. SIQSE202001); Academician Innovation Platform of Hainan Province.
\end{document}
|
arXiv
|
{
"id": "2204.08640.tex",
"language_detection_score": 0.5582510828971863,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Pre-fixed Threshold Real Time Selection Method in Free-space Quantum Key Distribution}
\author{Wenyuan Wang} \affiliation{Centre for Quantum Information and Quantum Control (CQIQC), Dept. of Electrical \& Computer Engineering and Dept. of Physics, University of Toronto, Toronto, Ontario, M5S 3G4, Canada}
\author{Feihu Xu}
\affiliation{Shanghai Branch, National Laboratory for Physical Sciences at Microscale, University of Science and Technology of China, Shanghai, 201315, China}
\author{Hoi-Kwong Lo} \affiliation{Centre for Quantum Information and Quantum Control (CQIQC), Dept. of Electrical \& Computer Engineering and Dept. of Physics, University of Toronto, Toronto, Ontario, M5S 3G4, Canada}
\begin{abstract}
Free-space Quantum key distribution (QKD) allows two parties to share a random key with unconditional security, between ground stations, between mobile platforms, and even in satellite-ground quantum communications. Atmospheric turbulence causes fluctuations in transmittance, which further affect the quantum bit error rate (QBER) and the secure key rate. Previous post-selection methods to combat atmospheric turbulence require a threshold value determined after all quantum transmission. In contrast, here we propose a new method where we pre-determine the optimal threshold value even before quantum transmission. Therefore, the receiver can discard useless data immediately, thus greatly reducing data storage requirement and computing resource. Furthermore, our method can be applied to a variety of protocols, including, for example, not only single-photon BB84, but also asymptotic and finite-size decoy-state BB84, which can greatly increase its practicality. \end{abstract}
\date{\today} \maketitle
\section{Introduction}
Quantum key distribution (QKD), first proposed by Bennett and Brassard\cite{bb84} in 1984, allows two parties to securely share a random secret key, which can be further combined with cryptographic protocols, such as one-time pad\cite{onetimepad}, to encode messages with unconditional security unachievable by classical cryptography.
There has been increasing interest in implementing QKD through free-space channels. A major attraction for free-space QKD is that, performed efficiently, it could potentially be applied to airborne or maritime quantum communications where participating parties are on mobile platforms. Furthermore, it could even enable applications for ground to satellite quantum communications, and eventually, global quantum communication network.
Free-space quantum communication has seen great advances over the past 25 years. The first demonstration of free-space QKD was published by Bennett et al. from IBM research in 1992 \cite{freespace_IBM} over 32cm of free-space channel, which was also the first successful demonstration of experimental QKD. Over the next two decades, numerous demonstrations for free-space QKD have been made. In 1998, Buttler and Hughes et al. \cite{freespace_Hughes} have performed QKD over 1km of free-space channel outdoors at nighttime. In 2005, Peng et al. \cite{freespace_2005_Peng} performed distribution of entangled photons over 13km. In 2007, two successful experimental ground-to-ground free-space QKD experiments based on BB84 and E91 protocol \cite{free2007,freespace_144km_E91} were implemented over a 144km link between the Canary Island of La Palma and Tenerife. Ling et al. \cite{freespace_urban_E91} performed another entanglement-based QKD Experiment with modified E91 protocol over 1.4km in urban area in 2008. In 2012, Yin et al. and Ma et al. \cite{freespace_2012_100km, freespace_2012_143km} respectively performed quantum teleportation over 100km and 143 km.
In recent years, free-space QKD has also seen much development over rapidly moving platforms, with an air-to-ground experiment in 2013 by Nauerth et al. \cite{freespace_plane} over a plane 20km from ground, a demonstration of QKD with devices on moving turntable and floating hot ballon over 20km and 40km by J-Y Wang et al. \cite{freespace_2013_satellite} in 2013, a very recent report on drone-to-drone QKD experiment in 2017 by D. Hill et al. \cite{freespace_drone}, and notably, satellite-based quantum communication experiments in 2017 \cite{freespace_satellite1, freespace_satellite2, freespace_satellite3}, including a QKD experiment from a quantum satellite to the ground over a 1200km channel. Meanwhile, there is a lot of interest in doing QKD in a maritime environment either over sea water\cite{freespace_maritime_data} or through sea water \cite{freespace_underwater}. A study on quantum communication performance over a turbulent free-space maritime channel using real atmospheric data can be found in Ref.\cite{freespace_maritime_data}.
A major characteristic of a free-space channel is its time-dependent transmittance. This is due to the temporal fluctuations of local refractive index in the free-space channel, i.e. \textit{atmospheric turbulence}, which causes scintillation and beam wandering \cite{freespacethesis}, and result in fluctuations in the channel transmittance, which in turn affect QKD performance. Therefore, addressing turbulence is a major challenge for QKD over free-space. This fluctuation due to turbulence can be modeled as a probability distribution, called the Probability Distribution of Transmission Coefficient (PDTC). Hence the real time transmittance $\eta$ is a random time-dependent variable that can be described by the PDTC.
\begin{table*}[t]
\caption{Comparison of transmittance post-selection methods in QKD through turbulence channel}
\begin{center}
\begin{tabular}{cccc}
Method & Threshold choice & Model of signals & Sampling of transmittance\\
\hline
ARTS\cite{probetest} & post-determined & single-photon & secondary probe laser\\
SNRF\cite{SNRF} & post-determined & single-photon & detector count (coincidence) rate \\
P-RTS & pre-determined & universal & universal
\end{tabular}
\end{center} \end{table*}
As free-space channels have time-varying transmittance due to turbulence, the QBER (and hence the secure key rate) for QKD changes with time. In previous literature discussing free-space QKD, such as \cite{free2007,Hughes}, the time variance of the channel is ignored, i.e. the secure key rate is calculated based on the time-average of channel transmittance. Having knowledge of the PDTC, however, Vallone et al. proposed a method named Adaptive Real-Time Selection (ARTS)\cite{probetest} that acquires information about real-time transmittance fluctuation due to turbulence, and makes use of this information to perform post-selection and improve the key rate of QKD.
However, ARTS method needs to "adaptively" choose an optimal threshold by performing numerical optimization after collecting all the data. A similar proposal by Erven et al. \cite{SNRF} called "signal-to-noise-ratio-filter (SNRF)" also discusses the idea of using a threshold to post-select high-transmittance periods, but uses the quantum data itself rather than a secondary classical channel. However, it needs to numerically optimize the threshold after collecting all experiment data, too.
Here we ask the question, is scanning through all acquired data after experiment and finding such an "adaptive" threshold really necessary? The answer is in fact no. In this paper, we propose a new method called "pre-fixed threshold real-time selection (P-RTS)", and show the important observation that, for post-selection based on transmittance in a turbulent channel, the optimal post-selection threshold is independent of the channel, and can be directly calculated from experimental parameters of the devices beforehand - thus simplifying the implementation and enabling post-selection of signals in real time, which can also reduce the data storage requirements and computational resources in Bob's system. This is because, instead of having to wait until all data is collected to optimize the threshold, Bob can immediately discard the data that are obtained below the pre-fixed threshold and doesn't need to store all collected data. Moreover, he doesn't need to have a model for the PDTC of the channel, and no longer need to run a numerical optimization to find optimal threshold, thus we can additionally save software development effort and computing resource for Bob, too.
Furthermore, both ARTS and SNRF are limited to single photon model only, while decoy-state must be used for QKD with practical weak coherent pulse (WCP) source. Here we also propose an universal framework for QKD through a channel with fluctuating transmittance, for not only single-photon BB84, but also practical decoy-state BB84 with WCP source, and decoy-state BB84 with finite-size effects (both of which we are the first to apply threshold post-selection to), thus greatly improving its usefulness in practice. We also propose a model to estimate the maximum improvement in key rate from using threshold post-selection, and show that with P-RTS method we can achieve a key rate very close to the maximum performance with an optimal threshold.
A comparison of P-RTS with post-selection methods in previous literature can be seen in Table I. As shown here, P-RTS has the great advantage of being able to predict the optimal threshold independently of the channel. This means that one no longer needs to store all the data after experiment and optimize the threshold, but can perform real-time selection with a single threshold, regardless of the actual channel turbulence and loss condition. Moreover, our result is valid not only for BB84 with single photons, but for any general protocol that has a fluctuating transmittance. It is also not restricted to transmittance sampling with a secondary laser as in ARTS, but for instance can also use observed photon count rates in a given time interval as in SNRF.
Lastly, we have performed a computer simulation to show the actual advantage of using P-RTS in practical decoy-state BB84, with up to 170\% improvement in decoy-state BB84 key rate for certain scenarios, or 5.1dB increase in maximum tolerant loss at $R=10^{-7}$, under medium-level turbulence. We also include a numerical demonstration for applying P-RTS to BB84 with finite-size effects, which still shows significant increase in rate even when total number of signals is limited, e.g. maximum tolerant loss at $R=10^{-7}$ gains an increase of 1.4dB to 5.2dB, for $N=10^{11}-10^{13}$ under high turbulence.
The organization of the paper is listed as follows: in section 2 we first present a brief recapitulation of ARTS method, and proceed to propose a universal framework for QKD key rate in turbulent channel. We then propose P-RTS method, and discuss how and why an optimal threshold can be pre-fixed, and show an upper bound for the rate of P-RTS. We also present numerical results from simulations to show how P-RTS behaves compared to no post-selection. Lastly, we discuss P-RTS in decoy-state BB84, for the asymptotic case in Section III and for finite-size regime in Section IV, and show with simulation results that P-RTS works effectively for both of them, too.
\section{Methods} \subsection{The ARTS Method}
In Ref.\cite{probetheory}, Capraro et al performed an experiment to study the impact of turbulence on a quantum laser transmitted through an 143km channel on Canary Islands, and proposed the idea of improving SNR with a threshold at the expense of number of signals. Subsequently, in Ref.\cite{probetest}, Vallone et al from the same group performed an experiment of free-space single-photon B92 QKD through the same channel, and showed the effectiveness of using real-time transmittance information in a turbulent channel to improve secure key rate, by performing post-selection on signals with a threshold, hence naming the method adaptive real-time selection (ARTS).
This is realized by using a classical probe signal (a strong laser beam) alongside the quantum channel. In the quantum channel, the bits are polarization-encoded into quantum signals, which are detected by single-photon avalanche diodes (SPADs) that return click events. Meanwhile, the laser passing through the classical channel is detected by an avalanche photodetector that returns a voltage proportional to received light intensity, which is also proportional to the channel transmittance at that moment. An illustration of the setup can be seen in Fig. \ref{fig:ARTS}.
The key idea is that the transmittance of the classical channel will correspond to that of the quantum channel. Therefore, by reading voltage from the classical detector (defined as $V$), one can gain information of the periods of high transmittance. Combined with a threshold on the classical signal (defined as $V_T$), this information can be used to post-select only those quantum signals received by Bob during high transmittance periods (only when $V \geq V_T$), thus increasing the overall average transmittance, at the expense of a smaller number of signals due to post-selection.
This post-selection increases the signal-to-noise ratio among the selected signals, and hence reduces the QBER, which subsequently increases the key rate. However, post-selection also reduces the total number of signals. Therefore, this becomes an optimization problem, and the choice of threshold value becomes critical. By numerically choosing an optimal threshold that maximizes the rate, it is possible to acquire a secure key rate much higher than before applying post-selection. This, as defined in Ref.\cite{probetest}, is called the adaptive real time selection (ARTS) method.
\subsection{Universal Framework for QKD Key Rate in a Turbulent Channel}
In this section, we will expand upon this threshold post-selection idea from ARTS, and apply it to a general framework of post-selection upon transmittance. We will then discuss the effects of threshold post-selection based on transmittance on the secure key rate. Our following discussions will be based on the channel transmittance $\eta$ only, and they are not limited to the secondary-laser transmittance sampling as in ARTS, but can be applied to any sampling method of transmittance, including photon count rate such as in SNRF.
As mentioned in Section I, an important characteristic of a turbulent channel is the time-dependent transmittance, which follows a probability distribution called the PDTC. There have been multiple efforts to accurately characterize the PDTC, and a widely accepted model is the log-normal distribution\cite{distribution,laser} (a plot of which is shown in Fig. \ref{fig:PDTC} (a)): \begin{equation} p(\eta)_{\eta_0,\sigma}={1 \over {\sqrt{2\pi \sigma}}\eta}e^{-{{[ln({\eta \over \eta_0})+{1\over 2}{\sigma}^2]^2}\over{2\sigma^2}}} \end{equation}
\begin{figure}
\caption{The ARTS setup by Vallone et al., where Alice and Bob are linked by a quantum channel and a classical channel. One can post-select quantum signals passing through the channel with high-transmittance, using a threshold on the corresponding classical channel signal}
\label{fig:ARTS}
\end{figure}
\noindent where p is the probability density, $\eta$ the transmittance, and $\eta_0$ and $\sigma$ the mean and variance. The distribution is solely determined by the two parameters $\eta_0$ and $\sigma$, which are inherent to the channel itself. $\eta_0$ is the expected atmospheric transmittance, with a typical value of $10^{-3}$ to $10^{-4}$ (corresponding to 30-40 dB of loss) for a 100km channel, while $\sigma$, typically taking a value between 0 and 1, is determined by the amount of turbulence - the larger the amount of turbulence, the larger the variance. The pair $(\eta_0,\sigma)$ hence contains all information of the PDTC.
Now, we make an important observation: For any given protocol implementation (say, single-photon BB84, or decoy-state BB84), if all experimental parameters in the system except $\eta$ are fixed - i.e. the device parameters including background and dark count rate, detector efficiency, laser intensities, and optical misalignment are all fixed - then the key rate solely depends upon the transmittance $\eta$, and can be written as a single-variable function of $\eta$, i.e. $R(\eta)$.
To estimate secure key rate of QKD through turbulent channel, the question therefore becomes studying how the function $R(\eta)$ changes, when $\eta$ is a random variable following a probability distribution that we know, the PDTC.
Here, we will propose two models for $R(\eta)$ under turbulence:
\begin{enumerate}
\item \textbf{Rate-wise integration model}, $R^{\text{Rate-wise}}$, which is the case where we integrate the rate over PDTC, thus making use of all information of the PDTC. This rate only depends on the rate function and the PDTC, and is independent of what actual threshold we choose.
\item \textbf{Simplified model}, $R^{\text{Simplified}}(\eta_T)$, which estimates the performance of decoy-state QKD with P-RTS, using post-selection with a threshold $\eta_T$ on channel transmittance. It is a function of the threshold $\eta_T$ that one uses. \end{enumerate}
Let us first start with the rate-wise integration model. We can begin by considering an ideal case, where we assume that we have complete knowledge of the channel transmittance $\eta$ when each single signal passes through the channel. Moreover, here we discuss the asymptotic case where there is an infinite number of signals sent. Then, it is possible to order all signals from low to high transmittance $\eta$ when they pass through the channel, and divide the signals into bins of $[\eta,\eta+\Delta \eta)$ (which ranges from 0 to 1), as shown in Fig. \ref{fig:PDTC} (b).
Therefore, within the bin $[\eta,\eta+\Delta \eta)$, we can assume that all signals pass through the channel with the same transmittance $\eta$, given that the bin is sufficiently small, i.e. $\Delta \eta \rightarrow 0$. That is, the signals in the same bin can be considered as in a "static channel", and enjoy the same rate formula $R(\eta)$ and security analysis as if $\eta$ is a static constant.
Then, we can calculate the number of secure key bits from each bin, according to their respective $\eta$, and add all bins together. In the limit of $\Delta \eta \rightarrow 0$, this is an integration of $R(\eta)$ over $\eta$, with $p_{\eta_0,\sigma}(\eta)$ being the weight (i.e. the proportion of signals in each bin). We call this model the \textit{"rate-wise integration model"}. Its rate $R^{\text{Rate-wise}}$ satisfies:
\begin{equation} R^{Rate-wise}=\int_{0}^{1}R(\eta)p_{\eta_0,\sigma}(\eta)d\eta \end{equation}
$R^{\text{Rate-wise}}$ makes use of all PDTC information from turbulence. Since all bins have either zero or positive rate, using a threshold $\eta_T$ in the rate-wise integration model will always result in either same or smaller rate. i.e.
\begin{equation} \begin{aligned} R^{Rate-wise}(0)&=\int_{0}^{1}R(\eta)p_{\eta_0,\sigma}(\eta)d\eta \\ &\geq \int_{\eta_T}^{1}R(\eta)p_{\eta_0,\sigma}(\eta)d\eta \\ & = R^{\text{Rate-wise}}(\eta_T) \end{aligned} \end{equation}
\noindent Hence, from here on if $\eta_T$ is not specified, by $R^{\text{Rate-wise}}$ we will always mean $R^{\text{Rate-wise}}(0)$, which is a constant-value "max possible performance" of key rate that is only dependent on the PDTC of the channel and the experimental device parameters.\\
\begin{figure}
\caption{(a) The Probability Distribution of Transmittance Coefficient (PDTC), where $\eta$ is the transmittance, taking a value between [0,1], while P is the probability density function of $\eta$. Here we are showing a plot generated from the log-normal model of the PDTC; (b) Dividing signals into bins according to their respective $\eta$, in the rate-wise integration model}
\label{fig:PDTC}
\end{figure}
Now, let us consider applying post-selection to free-space QKD. Using a similar model as in ARTS method (instead of using classical detector voltage V, here we will directly use $\eta$, which is proportional to V). We can set a threshold $\eta_T$ and perform post-selection: we select quantum signals received when transmittance $\eta\geq \eta_T$, and discard all signals received when $\eta<\eta_T$.
Unlike the ideal case of the rate-wise integration model, in reality we do not have infinite resolution from the classical detector, nor do we have an infinite number of signals. In practice, we are post-selecting signals with only two status: "pass" or "fail". To make use of this "pass/fail" information, here we propose a practical model that estimates the rate with only the mean transmittance of the post-selected signals. We name it the \textit{"simplified model"}. First, with no post-selection applied, the rate is:
\begin{equation} R^{\text{Simplified}}(0)=R(\eta_0) \end{equation}
\noindent which means that we simply use the mean value of transmittance $\eta_0$ in the channel for all calculations and assume a "static channel", using the same rate formula for a static channel, too. This is, in fact, what has been done in most literature for free-space QKD that don't consider fluctuations due to turbulence, such as in \cite{free2007,Hughes}.
Now, when a threshold is used and post-selection is performed, $R^{\text{Simplified}}$ is written as:
\begin{equation} R^{\text{Simplified}}(\eta_T)=\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta \times R(\langle \eta \rangle) \end{equation}
\noindent here we again treat all post-selected signals as having passed through a "static channel", and use the same rate expression for static case. But the difference is that we use the new mean transmittance among only the post-selected signals, denoted as $\langle \eta \rangle$, as the transmittance of the channel. $\langle \eta \rangle$ satisfies (using expected value formula for a truncated distribution):
\begin{equation} \langle \eta \rangle={{\int_{\eta_T}^{1}\eta p_{\eta_0,\sigma}(\eta)d\eta}\over{\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta}} \end{equation}
When we apply post-selection (like the case with ARTS), in the rate formula for $R^{\text{Simplified}}$, we take into account the loss of signals due to post-selection, and only a portion of $\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta$ remains. This portion is always no larger than 1, and strictly decreases with $\eta_T$. On the other hand, $\langle \eta \rangle$ is always increasing with $\eta_T$, because we are post-selecting only the signals with higher transmittance. So, just like for the single photon case discussed in Section II.A, we have an optimization problem, where the choice of $\eta_T$ is crucial to the rate we acquire. Using optimal threshold and applying post-selection, as we will later show in the numerical results in the next sections, can dramatically increase the rate over using no post-selection at all.
Therefore, using the simplified model, we can effectively treat the static channel QKD protocol as a "black-box". We enjoy the same rate formula and security analysis as a static channel, while the only difference is that we use a higher $\langle \eta \rangle$ after post-selection as the input, and multiply a reduced portion $\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta$ to the output.\\
Now, let us compare the performance of the two models. From an information theory perspective, the rate-wise integration model makes use of all possible information on fluctuating transmittance (i.e. the whole PDTC), while the simplified model discards all distribution information and only acknowledges "pass or fail", and keeps only the single mean transmittance after post-selection. Therefore, we expect that the rate-wise integration model, which makes use of the most information, would have a higher rate than the simplified model. We can write the relation as:
\begin{equation} R^{\text{Rate-wise}} \geq R^{\text{Simplified}} \end{equation}
This relation suggestions that the Rate-wise integration model is an upper bound for the Simplified model key rate. This result can be shown rigorously by Jensen's Inequality (we include the detailed proof in the Appendix B), under the condition that the rate function $R(\eta)$ is convex. Numerically, we show that (in next section) the rate for single-photon BB84 and decoy-state BB84 are both convex. Therefore, the relation Eq. 7 always holds true.
The next question is, naturally, what is the optimal threshold to choose, such that $R^{\text{Simplified}}$ approaches the upper bound $R^{\text{Rate-wise}}$? Moreover, how closely can it approach the upper bound? We will discuss this optimal threshold in the next section, and show that it is only dependent upon $R(\eta)$ and independent of the PDTC.
\subsection{Optimal Threshold and Near-Tightness of Upper Bound}
In this section, we propose the "Pre-fixed threshold Real Time Selection" (P-RTS) method, and show that the optimal threshold is independent of the PDTC and can be pre-fixed based on experimental parameters only. We also show that with this pre-fixed threshold the simplified model can approach its upper bound very closely.
Here, to describe the key rate function, we have to bring it into the context of an actual protocol model. We will first discuss single-photon BB84, using the Shor-Preskill \cite{Preskill} rate:
\begin{equation} R=1-2h_2[QBER] \end{equation}
\noindent here to keep the consistency of notations with following discussions, we will use parameters from Table II (which is also used as the channel model for decoy-state discussion), where detector dark count/background count rate is $Y_0$, basis misalignment is $e_d$, and total system transmittance is $\eta_{sys}=\eta\eta_d$:
\begin{equation} R_{S-P}=(Y_0+\eta_{sys})\{1-2h_2[e(\eta_{sys})]\} \end{equation}
\noindent while the single-photon QBER is
\begin{equation} e(\eta_{sys})={{{1\over 2}Y_0+e_d \eta_{sys}}\over{Y_0+\eta_{sys}}} \end{equation}
\begin{figure}
\caption{Single-photon rate and PDTC vs Transmittance $\eta$. As can be seen, there is an $\eta_{critical}$ such that $R_{S-P}(\eta)=0$ for all $\eta \leq \eta_{critical}$. For this example, we have plotted the single-photon rate R, using experimental parameters are listed in Table II. We acquire an $\eta_{critical}=0.00020$ for single-photon case. Note that $\eta_{critical}$ is only determined by the experimental parameters of our devices (e.g. dark count rate, and misalignment, and the chosen intensities), and is independent of the actual PDTC. Linear interpolation of the asymptotic $\eta \gg Y_0$ case shows that the function is very close to linear. Here an instance of P, the PDTC function, is also plotted for comparison.}
\label{fig:critical}
\end{figure}
A point worth noting is that $R_{S-P}(\eta)$ has the unique property of having an $\eta_{critical}$ such that $R_{S-P}(\eta)=0$ for all $\eta<\eta_{critical}$, and $R_{S-P}(\eta) \geq 0$ for $\eta\geq \eta_{critical}$. This critical position can be expressed as:
\begin{equation} \eta_{critical}={Y_0 \over \eta_d}{{{1\over 2}-e_{critical}}\over{e_{critical}-e_d}} \end{equation}
\noindent where $e_{critical}$ is the threshold QBER satisfying \begin{equation} 1-2h_2(e_{critical})=0 \end{equation}
\noindent that returns zero rate. For Shor-Preskill rate, this threshold is $e_{critical}=11\%$. More details can be seen in Appendix D.
As can be shown in Fig.\ref{fig:critical}, we plot out the single-photon rate $R_{S-P}(\eta)$, where a sharp turning point $\eta_{critical}$ exists. Moreover, within the $[\eta_{critical},1]$ region, numerical results show that $R(\eta)$ is slightly convex but very close to linear. (For larger $\eta \gg Y_0$, the rate is completely linear. Using this approximation we can make an interpolation of the "linear" part of the rate. As shown in the plot, this linear interpolation is very close to the rate function itself.)
This can lead to a very interesting result: We showed in Section II.B that $R^{\text{Rate-wise}}$ predicts the maximum possible performance of QKD with threshold post-selection in a turbulence channel. If we choose the threshold $\eta_T=\eta_{critical}$ for the simplified model, we can apply Jensen's Inequality for the truncated $p(\eta)$ distribution within region $[\eta_{critical},1]$, and acquire \begin{equation}
R^{\text{Simplified}}(\eta_{critical}) \approx R^{\text{Rate-wise}}(\eta_{critical}) \end{equation}
\noindent given that $R_{S-P}(\eta)$ is very close to linear within the region (but still convex, so $R^{\text{Simplified}}$ is still slightly smaller), since Jensen's Inequality takes equal sign for a linear function. There is also no loss in $R^{\text{Rate-wise}}$ from truncating $[0,\eta_{critical})$, as $R(\eta)=0$ for all $\eta<\eta_{critical}$.
\begin{equation} R^{\text{Rate-wise}}(\eta_{critical}) = R^{\text{Rate-wise}}(0) \end{equation}
\noindent Therefore, $R^{\text{Simplified}}$ can approximately reach the upper bound with $\eta_T=\eta_{critical}$, and the upper bound given by $R^{\text{Rate-wise}}$ is near-tight, due to the near-linearity of $R(\eta)$. A more rigorous proof showing that the optimal threshold for the simplified model is indeed $\eta_{critical}$ can be found in Appendix C.
Also, despite that there is no explicit analytical expression for $\eta_{critical}$, we can show that it depends more heavily on the background/dark count rate (approximately proportional to $Y_0$, if $\eta \ll 1$, and $Y_0 \ll \eta$). Details can be seen in Appendix D.
This result for optimal threshold has two significant implications for using threshold post-selection and applying the simplified model:
\begin{itemize}
\item Since $R(\eta)$ is only a function of $\eta$, and not $(\eta_0, \sigma)$, this optimal threshold position $\eta_{critical}$ is only determined by the experimental parameters of the devices (e.g. detector efficiency, dark count rate, misalignment, and Alice's intensities - although here we make an assumption that the misalignment is independent of $\eta$), and thus $\eta_{critical}$ is \textit{\textbf{independent}} of the channel itself and its PDTC. This means that, regardless of the turbulence level, we can use the same threshold to get optimized performance - although the actual amount of performance improvement over not using post-selection \textit{will} be determined by the average loss and the amount of turbulence (i.e. the actual PDTC), as will also be shown in numerical results in the next section.
\item Given that we choose the optimal threshold and apply P-RTS, not only are we optimizing the rate for the simplified model, but we are also achieving the maximum possible performance for the turbulent channel, even if we make use of all information on transmittance fluctuations. This is because, at $\eta_{critical}$, the max value for $R^{\text{Simplified}}$ can almost reach the upper bound given by the rate-wise integration model - meaning that the upper bound is nearly tight. We will illustrate this point further with numerical results in the next section. \end{itemize}
The significant implication is that, as long as we know the experimental parameters, we can determine the optimal threshold in advance, without the need to know any information about the channel (such as to measure the turbulence level), and perform post-selection in real time using the fixed threshold.
Therefore, we show that it is possible to perform post-selection on the channel transmittance with a pre-fixed threshold - which we will call "Pre-fixed threshold Real Time Selection" (P-RTS). This is significantly more convenient than protocols which perform optimization of threshold after the experiment is done. It will substantially reduce the amount of data storage requirements in the experiment, since Bob doesn't need to store all data until after experiment for optimization of threshold, and will also save the computational resource since Bob no longer needs to perform optimization of threshold.
\subsection{Numerical Results}
In this section we put the above models into a simulation program for single-photon BB84 in a turbulent channel. We use the experimental parameters from Ref.\cite{free2007}, as listed in Table II. One note is that, though the dark count and stray light contribution is reported to be as high as 1700/s in the paper, because of the gated behavior of the detector and the post-selection, only the counts within a 5.9ns time window (in 100ns period between two pulses, for the 10MHz source used) will affect the result. Therefore, here we take dark count rate as $Y_0=1\times 10^{-5}$ in the simulations.
\begin{table*}[t]
\caption{Experimental parameters for free-space QKD over an 144km channel in Ref.\cite{free2007}}
\begin{center}
\begin{tabular}{ccccccc}
dark count rate $Y_0$ & pulse rate & detector efficiency $\eta_d$ & misalignment $e_d$ & error-correction efficiency $f$\\
\hline
$1\times 10^{-5}$ (per signal) & 10MHz & 25\% & 3\% & 1.22\\
\end{tabular}
\end{center} \end{table*}
\begin{figure}\label{fig:threshold}
\end{figure}
Here, we first take a turbulence level of $\sigma=0.9$, and compare the performance of the two models plus the static model (which is a simplified model with no post-selection, i.e. $R_{S-P}(\eta_0)$) at a fixed loss of 37dB. We plot the results in Fig.\ref{fig:threshold}. As shown in the figure, $R^{\text{Simplified}}(\eta_T)$ first increases with threshold $\eta_T$ (because of post-selecting high-transmittance signals) and then decreases when threshold is further increased (because the decrease in rate due to loss of signals starts to dominate).
Just as predicted in Section II.C, the simplified model can achieve a very similar performance as the upper bound given by the rate-wise integration model, when the optimal threshold is chosen. For this case, at the optimal threshold $\eta_T=0.00020$, which, as we predicted, is the same as $\eta_{critical}=0.00020$ in Fig.\ref{fig:critical}, we get $R^{\text{Simplified}}=1.18 \times 10^{-5}$, very close to the upper bound $R^{\text{Rate-wise}}=1.22 \times 10^{-5}$ (only by $3\%$ difference - which is due to the rate above $\eta_{critical}$ not perfectly linear), and with dramatic increase in key rate compared with the default static model (using mean transmittance) $R^{\text{Static}}=3.5 \times 10^{-6}$, demonstrating the significant performance gain from using P-RTS in turbulence channel.
Furthermore, we compare the rate-wise integration model $R^{\text{Rate-wise}}$, the optimized $R^{\text{Simplified}}(\eta_T)$ with $\eta_T=\eta_{critical}$, and the non-post-selected model (whose rate is equivalent to static model, i.e. $R(\eta_0)$, as in Eq. 4) , by generating the rate vs loss relation for different average loss in the channel. Results can be seen in Fig. \ref{fig:turbulence2}. We see that indeed the rate for simplified model with fixed threshold is extremely close to its upper bound (as suggested in Eq. 13), the rate-wise integration model. Comparing with the static case, we see that the P-RTS method works best for high-loss regions, where post-selection can "salvage" some rate where static case would fail entirely, hence "getting something out of practically nothing". Therefore, one of the major improvements we acquire from using P-RTS in free-space QKD is a dramatically increased maximum tolerant loss (which would mean longer maximum distance).
\begin{figure}
\caption{Comparison of the rate-wise integration model, simplified model with optimal threshold, and no post-selection (static model) under $\sigma=0.9$, for the single-photon case. Parameters are from Table II. We can see that the simplified model, with optimized threshold, approaches the rate-wise integration model extremely closely, and both cases have significant improvement in key rate over static (no post-selection) model, especially in high-loss region.}
\label{fig:turbulence2}
\end{figure}
\section{Decoy-State BB84}
On the other hand, for decoy-state BB84 QKD, we follow decoy-state BB84 QKD theory from Ref. \cite{decoystate_Hwang,decoystate_LMC,decoystate_Wang}, and adopt the notations as in Lo, Ma, and Chen's Paper in 2005 \cite{decoystate_LMC}. Using the GLLP formula \cite{GLLP}, in the asymptotic limit of infinitely many data, we can calculate the secure key rate as:
\begin{equation}
\begin{aligned}
R_{GLLP} = q\{-f(E_\mu)Q_\mu h_2(E_\mu)+Q_1[1-h_2(e_1)]\}
\end{aligned} \end{equation}
\noindent where $h_2$ is the binary entropy function, $q={1\over 2}$ or $q\approx 1$ depending on whether efficient BB84 is used, and $f$ is the error-correction efficiency. $Q_\mu$ and $E_\mu$ are the observed Gain and QBER, while $Q_1$ and $e_1$ are the single-photon Gain and QBER contributions estimated using decoy-state. (For a more detailed recapitulation of decoy-state, see Appendix A.1. We have also discussed the channel model that we use for P-RTS in Appendix A.2).
\begin{figure}
\caption{Rate and PDTC vs Transmittance $\eta$ for (asymptotic) decoy-state BB84 with infinite data size. Intensities are $\mu=0.3$, $\nu=0.05$, and experimental parameters are from Table II. As can be seen, there is also an $\eta_{critical}=0.0012$ such that $R_{GLLP}(\eta)=0$ for all $\eta \leq \eta_{critical}$, just like for single photons.}
\label{fig:critical_decoy}
\end{figure}
Here for free-space decoy-state QKD. We fix the signal and decoy-state intensities as $\mu=0.3$, $\nu=0.05$, and the vacuum state $\omega=0$, and use the vacuum+weak method to estimate single-photon contribution, as in Ma et al.'s 2005 paper \cite{decoypractical} for practical decoy-state QKD.
Like for the single-photon case, again we generate the rate vs $\eta$ function. As can be observed in Fig. \ref{fig:critical_decoy}, the decoy-state rate function $R_{GLLP}(\eta)$ behaves similarly as the single-photon rate $R_{S-P}(\eta)$, with a critical transmittance $\eta_{critical}$ ($\eta_{critical}=0.0012$ for this parameter set) such that all $\eta$ below it returns zero rate, and a nearly linear rate-transmittance relation for $\eta \geq \eta_{critical}$. Therefore, using the same proof from section II, we can conclude that $\eta_{critical}$ is the optimal (and fixed) threshold for decoy-state BB84 with post-selection too.
\begin{figure}
\caption{Comparison of the optimized Simplified model vs no post-selection (static model) under different levels of turbulence, for (asymptotic) decoy-state BB84 with infinite data size. Here we use $\sigma=0.3, 0.6, 0.9$ and $\eta_T=0.0012$. Intensities are $\mu=0.3$, $\nu=0.05$, and experimental parameters are from Table II. We see that the improvement in rate from using P-RTS increases with the level of turbulence, and has a significant improvement over static model even under medium-level turbulence of $\sigma=0.6$.}
\label{fig:turbulence}
\end{figure}
Using the fixed threshold $\eta_T=\eta_{critical}$ to get the optimized rate $R^{\text{Simplified}}(\eta_{critical})$, we generate the rate vs loss relation for different levels or turbulence, as shown in Fig.\ref{fig:turbulence}. As can be seen, the P-RTS method works in the same way with decoy-sate BB84. We can also see that the higher the turbulence level is, the larger performance gain from applying P-RTS will we be able to achieve. As described in Section II.C, the optimal threshold is only determined by the parameters of the equipment, but the actual optimal \textit{performance} is determined by the amount of turbulence present in the channel that we can utilize. As can be seen in the plot, even for a medium-level turbulence of $\sigma=0.6$: for the same loss=29dB, $R^{\text{Simplified}}=8.453 \times 10^{-6}$, a 170\% increase over $R^{\text{Static}}=3.119 \times 10^{-6}$ at loss=29dB. Also, for a minimum rate of $R=10^{-7}$, simplified model has a maximum tolerant loss of 34.4dB, versus 29.5dB for Static Model, with 5.1dB increase in tolerant loss.
\begin{figure}
\caption{Comparison of the optimized simplified model vs static model, for decoy-state BB84 with finite-size effects. We test different data sizes $N=10^{11},10^{12},10^{13}$, and the near-asymptotic case $N=10^{99}$. Here we use a high turbulence of $\sigma=0.9$. The experimental parameters also follow Table II, and intensities and probabilities used are $\mu=0.31$, $\nu=0.165$, $\omega=2\times 10^{-4}$, $p_{\mu}=0.5$, $p_{\nu}=0.36$, and the probability of sending X basis $q_x=0.75$. The dotted lines are the cases where no post-selection is applied, while the solid lines all have post-selection applied with $\eta_T=0.0012$. We can see from the figure that the improvement in rate from using P-RTS increases with the data size N, and still increases maximum tolerant loss by 3.5dB and 1.4dB when $N=10^{12}$ and $10^{11}$.}
\label{fig:finite}
\end{figure}
\section{Decoy-State BB84 with Finite-size Effects}
We now turn to the case with finite data size, and apply simplified model and P-RTS to decoy-state BB84 under finite-size effects. We also use simulations to numerically demonstrate the improvements in key rate for finite-size case. The protocol is based on C. Lim et al.'s finite-size decoy-state BB84 paper \cite{finitebb84}, and we have adopted the same channel model as in Ref. \cite{decoypractical}. Here we use the same experimental parameters (including dark count rate, detector efficiency and misalignment) as in Table II, same as the ones used in our previous asymptotic-case simulations. Also, we fix the signal and decoy intensities to $\mu=0.31$, $\nu=0.165$ (in addition to the vacuum intensity $\omega=2\times 10^{-4}$), the probabilities of sending them $p_{\mu}=0.5$, $p_{\nu}=0.36$, and the probability of sending X basis $q_x=0.75$. Unlike in Ref. \cite{finitebb84}, however, we do not scan through the decoy-state intensities and probabilities to perform optimization. Instead, since we only concentrate on high-loss region, we use fixed parameters that are already very close to optimal (while changing them with distance does not provide much improvement in performance). Using intensities that do not change with channel loss also avoids changing the expression for $R_{GLLP}(\eta)$ (which depends on intensities $\mu$, $\nu$), and ensures that $\eta_{critical}$ is independent of the actual loss of the channel.
As described for simplified model, we can use the same "black box" idea, and simply substitute $R_{GLLP}$ for asymptotic BB84 with rate for finite-size BB84. However, one difference from the asymptotic case is that N, the number of signals sent by Alice, matters when calculating the rate, i.e. the rate becomes $R_{Finite-Size}(\eta,N)$ instead of $R_{GLLP}(\eta)$. Then, instead of using Eq. 5 for $R^{\text{Simplified}}$, we use
\begin{equation} \begin{aligned} R^{\text{Simplified}}&=\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta \\ &\times R_{Finite-Size}(\langle \eta \rangle, N\times \int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta) \end{aligned} \end{equation}
\noindent which means that, the post-selection not only affects the overall rate due to the portion of lost signals, but also affects the rate for the \textit{selected} signals, too, since fewer signals than N are used to actually perform the protocol, and higher statistical fluctuations will be present among the selected signals. This means that we need to be even more prudent with post-selection when treating finite-size BB84.
The numerical results are shown in Fig.\ref{fig:finite}. As can be seen, P-RTS has a similar effect on finite-size BB84 as on the asymptotic case: we gain a significant advantage in the high-loss region, and have an improved maximum tolerant loss, when a minimum acceptable Rate is required. For instance, at $\sigma=0.9$ and for a minimum $R=10^{-7}$, the maximum loss increases by 1.4dB, 3.5dB, 5.2dB, and 6.2dB, respectively, for the cases with $N=10^{11}, 10^{12}, 10^{13}$, and near-asymptotic case, while not much improvement can be gained from P-RTS with N smaller than $10^{10}$. As shown, the improvement increases with the size of N (which is understandable, since the smaller N is, the more sensitive the rate will be to post-selection - because we are cutting off a portion from the already-insufficient number of signals and further aggravating the statistical fluctuations - while for the asymptotic case, for instance, the performance of selected signals does not depend upon how big is the selected portion of signals, and the only negative effect post-selection has is the lost portion of signals). For a free-space QKD system with 100MHz repetition rate, $N=10^{11}$ would require about 17 minutes of communication.
\section{Conclusion}
In this paper we have proposed a post-selection method with prefixed threshold for QKD through turbulent channel, and have also proposed a universal framework for determining the optimal threshold beforehand and predicting the maximum possible performance. By choosing the threshold in advance, we can perform post-selection in real time regardless of the channel condition. This real-time post-selection also provides an additional benefit of reducing the amount of data that is required to be stored in the detector system on Bob's side. We also performed simulations to show the method's effectiveness in not only single-photon BB84, but also practical decoy-state QKD in both the asymptotic case and the case with finite-size effects.
This method is especially effective for regions of high turbulence and high loss, and can even "salvage something out of nothing", when the secure key rate could have been zero without P-RTS method. In order to sample the real-time transmittance condition, the P-RTS method can use only an additional classical channel for each quantum channel, which would be easily implemented (or may even be already implemented as a beacon laser is often required for alignment in free-space QKD). Moreover, since our results only depend on post-selection of $\eta$, in essence our method is even possible without an additional classical channel, such as in Erven et al.'s SNRF setup \cite{SNRF} (which samples transmittance by observing quantum signal count rate). The thresholding, on the other hand, is purely implemented in post-processing, therefore does not require any additional resource, and could be readily deployed into existent infrastructure, and gain a ready increase in secure key rate performance over existing implementation for free-space QKD.
\begin{thebibliography}{1}
\bibitem{bb84} C Bennett, G Brassard {\em Quantum cryptography: Public key distribution and coin tossing.} International Conference on Computer System and Signal Processing, IEEE (1984)
\bibitem{onetimepad} F Miller {\em Telegraphic code to insure privacy and secrecy in the transmission of telegrams} CM Cornwell (1882)
\bibitem{freespace_IBM} CH Bennett, F Bessette, G Brassard, L Salvail, and J Smolin {\em Experimental quantum cryptography} Journal of cryptology 5.1 (1992): 3-28.
\bibitem{freespace_Hughes} WT Buttler et al. {\em Practical free-space quantum key distribution over 1 km}. Physical Review Letters, 1998, 81(15): 3283.
\bibitem{freespace_2005_Peng} C-Z Peng et al. {\em Experimental free-space distribution of entangled photon pairs over 13 km: towards satellite-based global quantum communication.} Physical Review Letters 94.15 (2005): 150501.
\bibitem{freespace_144km_E91} R Ursin et al. {\em Entanglement-based quantum communication over 144 km.} Nature physics 3.7 (2007): 481-486.
\bibitem{free2007} T Schmitt-Manderbach et al. {\em Experimental demonstration of free-space decoy-state quantum key distribution over 144 km} Physical Review Letters 98.1 (2007): 010504.
\bibitem{freespace_urban_E91} A Ling et al. {\em Experimental quantum key distribution based on a Bell test} Physical Review A 78.2 (2008): 020301.
\bibitem{freespace_2012_100km} J Yin et al. {\em Quantum teleportation and entanglement distribution over 100-kilometre free-space channels} Nature 488.7410 (2012): 185-188.
\bibitem{freespace_2012_143km} X-S Ma et al. {\em Quantum teleportation over 143 kilometres using active feed-forward} Nature 489.7415 (2012): 269-273.
\bibitem{freespace_plane} S Nauerth et al. {\em Air-to-ground quantum communication} Nature Photonics 7.5 (2013): 382-386.
\bibitem{freespace_2013_satellite} J-Y Wang et al. {\em Direct and full-scale experimental verifications towards ground-satellite quantum key distribution} Nature Photonics 7.5 (2013): 387-393.
\bibitem{freespace_drone} AD Hill, et al. {\em Drone-based Quantum Key Distribution} Urbana 51 (2017): 61801-3003.
\bibitem{freespace_satellite1} S-K Liao et al. {\em Satellite-to-ground quantum key distribution} Nature 549.7670 (2017): 43-47.
\bibitem{freespace_satellite2} J Yin et al. {\em Satellite-based entanglement distribution over 1200 kilometers} Science 356.6343 (2017): 1140-1144.
\bibitem{freespace_satellite3} J-G Ren et al. {\em Ground-to-satellite quantum teleportation} Nature 549.7670 (2017): 70-73.
\bibitem{freespace_maritime_data} J Gariano, M Neifeld, and I Djordjevic. {\em Engineering trade studies for a quantum key distribution system over a 30 km free-space maritime channel.} Applied Optics 56.3 (2017): 543-557.
\bibitem{freespace_underwater} L Ji et al. {\em Towards quantum communications in free-space seawater.} Optics Express 25.17 (2017): 19795-19806.
\bibitem{freespacethesis} J-P Bourgoin {\em Experimental and theoretical demonstration of the feasibility of global quantum cryptography using satellites}, University of Waterloo (2014)
\bibitem{Hughes} RJ Hughes, JE Nordholt, D Derkacs, and CG Peterson {\em Practical free-space quantum key distribution over 10 km in daylight and at night} New Journal of Physics, Vol. 4 (2002)
\bibitem{probetest} G Vallone et al {\em Adaptive real time selection for quantum key distribution in lossy and turbulent free-space channels} Phys. Rev. A 91, 042320 (2015)
\bibitem{SNRF} C Erven et al. {\em Studying free-space transmission statistics and improving free-space quantum key distribution in the turbulent atmosphere.} New Journal of Physics 14.12 (2012): 123018.
\bibitem{probetheory} I Capraro et al {\em Impact of turbulence in long range quantum and classical communications} Phys. Rev. Lett. 109, 200502 (2012)
\bibitem{distribution} P Milonni, JH Carter, CG Peterson, and RJ Hughes {\em Effects of propagation through atmospheric turbulence on photon statistics} J. Opt. B: Quantum Semiclass., Vol. 6, No. 8 (2004)
\bibitem{laser} LC Andrews, RL Phillips, and CY Hopen. {\em Laser beam scintillation with applications} Vol. 99. SPIE press (2001)
\bibitem{Preskill} P Shor, J Preskill {\em Simple Proof of Security of the BB84 Quantum Key Distribution Protocol} Phys. Rev. Lett. 85, 441 (2000)
\bibitem{decoystate_Hwang} W-Y Hwang {\em Quantum key distribution with high loss: toward global secure communication} Physical Review Letters 91.5 (2003): 057901.
\bibitem{decoystate_LMC} HK Lo, XF Ma, and K Chen {\em Decoy state quantum key distribution.} Physical review letters 94.23 (2005): 230504.
\bibitem{decoystate_Wang} X-B Wang {\em Beating the photon-number-splitting attack in practical quantum cryptography} Physical review letters 94.23 (2005): 230503.
\bibitem{GLLP} D Gottesman, HK Lo, N Lutkenhaus, and J Preskill {\em Security of quantum key distribution with imperfect devices} Quantum Info. and Comp., 4, No.5 (2004) 325-360
\bibitem{decoypractical} X Ma, B Qi, Y Zhao, and HK Lo {\em Practical decoy state for quantum key distribution} Phys. Rev. A 72, 012326 (2005)
\bibitem{finitebb84} C Lim, M Curty, N Walenta, F Xu, and H Zbinden {\em Concise security bounds for practical decoy-state quantum key distribution} Physical Review A 89.2 (2014): 022307.
\bibitem{rytov} M E Gracheva, and A S Gurvich {\em Strong fluctuations in the intensity of light propagated through the atmosphere close to the earth} Radiophysics and Quantum Electronics 8.4 (1965): 511-515.
\bibitem{modtran} A Berk et al. {\em MODTRAN6: a major upgrade of the MODTRAN radiative transfer code} SPIE Defense+ Security. International Society for Optics and Photonics (2014)
\end{thebibliography}
\appendix
\section{Decoy-State BB84 Rate Function}
\subsection{Standard Channel Model}
Here we present a brief recapitulation of the decoy-state BB84 model we used. We follow the notations as in Lo, Ma, and Chen's Paper in 2005 \cite{decoystate_LMC}.
Alice uses a WCP source at intensity $\mu$, which sends pulses with a Poissonian photon number distribution: $P_i={\mu^i \over i!}e^{-\mu}$. We will first consider using the standard channel model (as in the original paper Ref.\cite{decoystate_LMC}), where for each i-photon pulse $\ket{i}$, the transmittance, yield $Y_i$, gain $Q_i$, and QBER $e_i$ are:
\begin{equation}
\begin{aligned}
\eta_i&=1-(1-\eta)^i\\
Y_i&\approx Y_0+\eta_i=Y_0+1-(1-\eta)^i\\
Q_i&=Y_i{\mu^i \over i!}e^{-\mu}\\
e_i&={{e_0Y_0+e_d \eta _i}\over Y_i}\\
\end{aligned}
\end{equation}
\noindent where $Y_0$, $e_d$ are the dark count rate and misalignment, respectively, and $e_0={1\over 2}$. The overall Gain $Q_{\mu}$ and QBER ${E_{\mu}}$ for this intensity $\mu$ are:
\begin{equation}
\begin{aligned}
Q_{\mu}&=\sum_{i=0}^{\infty}Y_i{\mu^i \over i!} e^{-\mu}=\sum_{i=0}^{\infty}Y_iP_i \\
E_\mu &= {1 \over Q_\mu} \sum_{i=0}^{\infty}e_i Y_i{\mu^i \over i!} e^{-\mu}={1 \over Q_\mu} \sum_{i=0}^{\infty}e_iY_iP_i
\end{aligned}
\end{equation}
\noindent where $Q_{\mu}$ and $E_{\mu}$ are simulated here for rate estimation using known channel transmittance $\eta$, while in experiment they will be measured observables.
For this standard channel model, we assume that the photon number distribution after passing through the channel would still be Gaussian. Using decoy-state technique to combine $Q_{\mu}$ and $E_{\mu}$ for different intensities, we can estimate the single-photon contributions $Q_1$ and $e_1$. The achievable secure key rate is at least
\begin{equation}
\begin{aligned}
R_{GLLP} = q\{-f(E_\mu)Q_\mu h_2(E_\mu)+Q_1[1-h_2(e_1)]\}
\end{aligned}
\end{equation}
\noindent as given by the GLLP formula\cite{GLLP}, where $h_2$ is the binary entropy function, $q={1\over 2}$ or $q\approx 1$ depending on whether efficient BB84 is used, and $f$ is the error-correction efficiency.
\subsection{Channel Model after Post-Selection}
However, one thing worth noting is that although photon number distribution is Gaussian after the signals pass through the standard channel model, it is no longer necessarily so if we perform post-selection, in which case the photon number distribution might change, and thus the decoy-state key rate form in Eq. A3 (which depends on a Gaussian distribution model) might no longer be adequate.
To show that this will not be a concern for us, we will explicitly discuss how the post-selection from P-RTS will affect the yield for each photon number. From Eq. A1, before post-selection, the yield for pulses with a given photon number $i$ is
\begin{equation}
\begin{aligned}
Y_i(\eta)=Y_0+1-(1-\eta)^i
\end{aligned}
\end{equation}
For simplified model, among the post-selected signals, we have replaced $\eta$ in Eq. A4 with
\begin{equation}
\begin{aligned}
\langle \eta \rangle={{\int_{\eta_T}^{1}\eta p_{\eta_0,\sigma}(\eta)d\eta}\over{\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta}}
\end{aligned}
\end{equation}
\noindent thus the yield for i-photon pulse is assumed to be:
\begin{equation}
\begin{aligned}
Y_i^{Simplified}(\eta_T)=Y_0+1-(1-\langle \eta \rangle)^i
\end{aligned}
\end{equation}
\noindent in which case, we are simply replacing the $\eta$ with a higher expected value $\langle \eta \rangle$, but the expression is in the same form as Eq. A4, and the received photon number distribution is still Gaussian. (Hence decoy-state analysis and key rate expression still hold).
However, if we consider the more realistic case, post-selection might have a different effect on pulses with different photon number $i$. Therefore, to estimate the yield for each photon number, and analyze the photon number distribution after the channel and the post-selection, we should group up pulses with the same given photon number, and calculate the expected value of the yield for each given $i$. We can call this the "pulse-wise integration" model.
\begin{equation}
\begin{aligned}
Y_i^{Pulse-wise}(\eta_T)={{\int_{\eta_T}^{1}Y_i(\eta) p_{\eta_0,\sigma}(\eta)d\eta}\over{\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta}}=\langle Y_i(\eta) \rangle
\end{aligned}
\end{equation}
\noindent and the Gain $Q_{\mu}$ and QBER ${E_{\mu}}$ would become:
\begin{equation}
\begin{aligned}
Q_{\mu}&=\sum_{i=0}^{\infty}\langle Y_i(\eta) \rangle{\mu^i \over i!} e^{-\mu}=\sum_{i=0}^{\infty}\langle Y_i(\eta) \rangle P_i \\
E_\mu &= {1 \over Q_\mu} \sum_{i=0}^{\infty}e_i \langle Y_i(\eta) \rangle{\mu^i \over i!} e^{-\mu}={1 \over Q_\mu} \sum_{i=0}^{\infty}e_i \langle Y_i(\eta) \rangle P_i
\end{aligned}
\end{equation}
In the case of this "pulse-wise integration" model, $Q_\mu$ and $E_\mu$ can no longer be considered as from a Gaussian distribution with intensity $\eta\mu$, which is seemingly warning us that the decoy-state analysis might not hold true anymore. However, here we make the observation that for $i=0$, trivially,
\begin{equation}
\begin{aligned}
Y_0^{Simplified}=Y_0=Y_0^{Pulse-wise}
\end{aligned}
\end{equation}
\noindent and for $i=1$, the yield is a linear function of $\eta$, hence
\begin{equation}
\begin{aligned}
Y_1^{Simplified}=Y_0+\langle \eta \rangle = \langle(Y_0+\eta)\rangle=Y_1^{Pulse-wise}
\end{aligned}
\end{equation}
While for all multi-photon cases where $i\geq 2$, the function
\begin{equation}
\begin{aligned}
Y_i(\eta)=Y_0+1-(1-\eta)^i
\end{aligned}
\end{equation}
is a strictly concave function on the domain $[0,1]$. Therefore, from Jensen's Inequality, the expected value of a concave function is strictly smaller than the function ($Y_i$) of the expected value, i.e.
\begin{equation}
\begin{aligned}
Y_i^{Pulse-wise}=\langle Y_i(\eta)\rangle < Y_i(\langle \eta \rangle) = Y_i^{Simplified}, i\geq 2
\end{aligned}
\end{equation}
This means that, with the simplified model, with the Gaussian photon number distribution assumption and the standard decoy-state key rate analysis, we are correctly estimating the vacuum and single-photon contributions, but always \textit{over-estimating} the multi-photon contributions. This will in fact result in an \textit{under-estimated} key rate for simplified model than the realistic case (yield-wise integration model). Therefore, we make the "validity argument" here that, despite post-selection will result in a non-Gaussian photon number distribution, by using simplified model and the same decoy-state analysis, we will never incorrectly over-estimate the key rate, and can be confident in the improvement in performance from using P-RTS.
\section{Proof of Rate-Wise Integration Model as Upper Bound}
To better compare the models, let us first simply the notations, and define $\langle f(\eta)\rangle$ operator as taking the expected value of $f(\eta)$ over $p_{\eta_0,\sigma}(\eta)$ (in the case of using post-selection, the distribution is truncated, and will be normalized by dividing by $\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta$). The expected value $\langle f(\eta) \rangle$ can be expressed as:
\begin{equation}
\langle f(\eta) \rangle={{\int_{\eta_T}^{1}f(\eta) p_{\eta_0,\sigma}(\eta)d\eta}\over{\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta}}
\end{equation}
Then, we can easily see that, mathematically, the two models we proposed so far, the rate-wise integration model and the simplified model, are only different in that they apply the "expected value" operator at different levels of the function. We can simply write $R^{\text{Rate-wise}}$ and $R^{\text{Simplified}}$ as:
\begin{equation}
\begin{aligned}
R^{\text{Rate-wise}} (\eta_T)&=\langle R(\eta) \rangle\\
R^{\text{Simplified}} (\eta_T)&=R(\langle \eta \rangle)
\end{aligned}
\end{equation}
Now, we introduce the \textit{Jensen's Inequality}: \\
\textit {For a random variable X following a probability distribution p(X), and for any given convex function f(x), we always have}
\begin{equation}
\langle f(X) \rangle \geq f(\langle X \rangle)
\end{equation}
\noindent the equal sign is taken when the function $f(x)$ is linear.
For decoy-state BB84, $R_{GLLP}(\eta)$ is a convex (and increasing) function of $\eta$, therefore we have $\langle R(\eta) \rangle \geq R(\langle \eta \rangle)$, i.e.
\begin{equation}
R^{\text{Rate-wise}}(0) \geq R^{\text{Simplified}}(0)
\end{equation}
\noindent This holds true even after a threshold is applied, too, since we can simply replace the distribution $p(\eta)$ with the truncated distribution on domain $[\eta_T,1]$, and normalize it by dividing by the constant $\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta$. Since $R(\eta)$ is non-concave on all sections of $[0,1]$, the Jensen's Inequality always holds true, regardless of the threshold. i.e.
\begin{equation}
R^{\text{Rate-wise}}(0) \geq R^{\text{Rate-wise}}(\eta_T) \geq R^{\text{Simplified}}(\eta_T)
\end{equation}
\noindent here we also include Eq. 3's result that $R^{\text{Rate-wise}}(\eta_T)$ is non-increasing with $\eta_T$.
Therefore, we see that $R^{\text{Rate-wise}}$ serves as an upper bound for the possible rate in a turbulent channel, as it is the maximum achievable rate when we know all transmittance information and make use of the entire PDTC. simplified model always has no higher rate than this upper bound. This means that, when we use $R^{\text{Simplified}}$ to calculate the rate, we \textit{never overestimate} the performance of the protocol. When we demonstrate the improvements we gain by using P-RTS in decoy-state BB84, the actual possible rate will be even higher, thus the validity argument for the usage of the simplified model in estimating the rate.
\section{Proof of Optimality of Critical Transmittance as Threshold for Simplified Model}
Following the argument in Section II.C, here we give a rigorous proof that $\eta_T=\eta_{critical}$ is indeed the optimal threshold for the simplified model, given that $R_{S-P}(\eta)$ (and similarly for $R_{GLLP}(\eta)$) is nearly linear. For simplified model, we showed that
\begin{equation}
R^{\text{Simplified}}(\eta_T)=\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta \times R_{S-P}(\langle \eta \rangle)
\end{equation}
\noindent where $\langle \eta \rangle$ satisfies:
\begin{equation}
\langle \eta \rangle={{\int_{\eta_T}^{1}\eta p_{\eta_0,\sigma}(\eta)d\eta}\over{\int_{\eta_T}^{1}p_{\eta_0,\sigma}(\eta)d\eta}}
\end{equation}
\noindent Then, using the Leibniz Integration Rule, and taking derivative with respect to $\eta_T$ (here we omit the subscript of $\eta_0,\sigma$ for the PDTC, and $S-P$ (or $GLLP$) for the rate), we have
\begin{equation}
\begin{aligned}
{d \over d\eta_T}{\langle \eta \rangle}={{p(\eta_T)}\over{\int_{\eta_T}^{1}p(\eta)d\eta}}({\langle \eta \rangle} - \eta_T)
\end{aligned}
\end{equation}
\noindent using the chain rule,
\begin{equation}
\begin{aligned}
{d \over d\eta_T}R(\langle \eta \rangle) &= {dR(\eta) \over {d\langle \eta \rangle}} {{d\langle \eta \rangle}\over d\eta_T}\\
&= R'(\langle \eta \rangle) {{p(\eta_T)}\over{\int_{\eta_T}^{1}p(\eta)d\eta}}({\langle \eta \rangle} - \eta_T)\\
\end{aligned}
\end{equation}
\noindent Maximizing $R^{\text{Simplified}}$ requires that
\begin{equation}
{d \over d\eta_T}R^{\text{Simplified}}(\eta_T)=0
\end{equation}
\noindent expanding the derivative using Eq. C1 gives us
\begin{equation}
\begin{aligned}
&{d \over d\eta_T}R^{\text{Simplified}}(\eta_T)\\
&= \left( \int_{\eta_T}^{1}p(\eta)d\eta \right) \times {d \over d\eta_T}R(\langle \eta \rangle) - p(\eta_T)R(\langle \eta \rangle) \\
&= R'(\langle \eta \rangle) {p(\eta_T)}({\langle \eta \rangle} - \eta_T) - p(\eta_T)R(\langle \eta \rangle) \\
&= p(\eta_T)[({\langle \eta \rangle} - \eta_T) R'(\langle \eta \rangle) - R(\langle \eta \rangle)]
\end{aligned}
\end{equation}
\noindent Therefore, the optimal threshold requires that
\begin{equation}
({\langle \eta \rangle} - \eta_T) R'(\langle \eta \rangle) = R(\langle \eta \rangle)
\end{equation}\\
When $R(\eta)$ is a linear function on the domain $[\eta_{critical}, 1]$ and $R(\eta_{critical})=0$, there is
\begin{equation}
R'(\langle \eta \rangle) = {{R(\langle \eta \rangle) - R(\eta_T)}\over{({\langle \eta \rangle} - \eta_T)}}
\end{equation}
\noindent combined with Eq. C7, we have
\begin{equation}
R(\eta_T)=0
\end{equation}
\noindent for $\eta \in [\eta_{critical}, 1]$, there is one and only one point satisfying $R(\eta_T)=0$, that is
\begin{equation}
\eta_T=\eta_{critical}
\end{equation}
\noindent For $\eta \in [0, \eta_{critical})$, on the other hand,
\begin{equation}
R(\langle \eta \rangle) - R(\eta_T) < R'(\langle \eta \rangle) ({\langle \eta \rangle} - \eta_T) = R(\langle \eta \rangle)
\end{equation}
\noindent which becomes
\begin{equation}
R(\eta_T) > 0
\end{equation}
\noindent but $R(\eta)=0$ for all $\eta \leq \eta_{critical}$, so no $\eta_T\in [0, \eta_{critical})$ satisfies the zero derivative requirement. Which means that, when $R_{GLLP}(\eta)$ is near linear on $[\eta_{critical}, 1]$, we have
\begin{equation}
\eta_T=\eta_{critical}
\end{equation}
\noindent as the one and only optimal threshold for $R^{\text{Simplified}}$.\\
Additionally, if we do not ignore the convexity of $R(\eta)$, consider the tangent line for $R(\eta)$ at $\langle \eta \rangle$, since $R(\eta)$ is a convex function of $\eta$,
\begin{equation}
({\langle \eta \rangle} - \eta_T) R'(\langle \eta \rangle) > R(\langle \eta \rangle) - R(\eta_T)
\end{equation}
\noindent optimal threshold requires that
\begin{equation}
R(\langle \eta \rangle) > R(\langle \eta \rangle) - R(\eta_T)
\end{equation}
\noindent i.e. $R(\eta_T) > 0$, which means that the optimal threshold position will be shifted rightward from $\eta_{critical}$, the actual amount of shift depends on how much $R$ deviates from linearity (in numerical simulations, we see that since $R(\eta)$ is very close to linear, this shift is very small). Also, although $R^{\text{Rate-wise}}$ is not affected for a threshold no larger than $\eta_{critical}$, using a threshold larger than $\eta_{critical}$ will cause $R^{\text{Rate-wise}}$ to decrease, since "bins" with positive rate are discarded. Therefore, the maximum point for $R^{\text{Simplified}}$ is no longer the maximum $R^{\text{Rate-wise}}$, but slightly smaller than it. This also explains why in the numerical results, the optimal $R^{\text{Simplified}}$ is always slightly lower than upper bound, due to non-linearity of $R(\eta)$.
Also, a small note is that, the Jensen's Inequality asks the function to be differentiable at every point, while the turning point of $R$ at $\eta_{critical}$ is a sharp point. To address this, we can construct another $R_2$ with an infinitesimally small yet smooth "turn" at $\eta_{critical}$ to replace the sharp point, but as the "turn" is infinitely small, integrating $R$ and $R_2$ over any region will yield infinitely close results. Therefore the turning point's structure does not affect the above results.
\section{Analytical Expression for Optimal Threshold}
\subsection{Single-Photon Case}
Let $\eta_{sys}=\eta\times \eta_d$. The single photon Shor-Preskill rate is
\begin{equation}
R_{S-P}=(Y_0+\eta_{sys})\{1-2h_2[e(\eta_{sys})]\}
\end{equation}
where the single-photon QBER is
\begin{equation}
e(\eta_{sys})={{{1\over 2}Y_0+e_d \eta_{sys}}\over{Y_0+\eta_{sys}}}
\end{equation}
For the rate to be zero, we require:
\begin{equation}
R_{S-P}=0
\end{equation}
hence
\begin{equation}
1-2h_2[e(\eta_{sys})]=0
\end{equation}
or, $h_2[e(\eta_{sys})]={1\over 2}$. This numerically corresponds to $e(\eta_{sys})=11\%=e_{critical}$ (which is the QBER threshold for Shor-Preskill rate). Therefore, substituting into Eq. D2, we have
\begin{equation}
\eta_{sys}={{{1\over 2}-e_{critical}}\over{e_{critical}-e_d}}Y_0
\end{equation}
expressing it in channel transmittance $\eta$
\begin{equation}
\eta_{critical}={Y_0 \over \eta_d}{{{1\over 2}-e_{critical}}\over{e_{critical}-e_d}}
\end{equation}
\noindent or, if we substitute $\eta_{critical}=11\%$ into the equation, we have
\begin{equation}
\eta_{critical}={Y_0 \over \eta_d}{0.39\over{0.11-e_d}}
\end{equation}
This is the analytical expression for the critical transmittance for the single-photon case. Also, we can see that the critical transmittance is proportional to the background count (i.e. noise) in the system. i.e.
\begin{equation}
\eta_{critical} \propto{Y_0 \over \eta_d}
\end{equation}
\subsection{Decoy-State BB84}
Consider the asymptotic case of decoy-state BB84, with infinite number of decoys (i.e. the only significant intensity is the signal intensity $\mu$). Using the GLLP rate,
\begin{equation}
\begin{aligned}
R_{GLLP} = q\{-fQ_\mu h_2(E_\mu)+Q_1[1-h_2(e_1)]\}
\end{aligned}
\end{equation}
\noindent we would like to find $\eta_{critical}$ such that
\begin{equation}
R(\eta_{critical})=0
\end{equation}
\noindent hence
\begin{equation}
fQ_\mu h_2(E_\mu)=Q_1[1-h_2(e_1)]
\end{equation}
\noindent or
\begin{equation}
h_2(e_1)+f{Q_\mu \over Q_{1}} h_2(E_\mu)=1
\end{equation}
Let $\eta_{sys}=\eta\times \eta_d$, the observables and single-photon contributions can be written as:
\begin{equation}
\begin{aligned}
Q_\mu &= Y_0+1-exp(-\mu\eta_{sys})\\
E_\mu &= {{{1\over 2}Y_0+e_d(1-exp(-\mu\eta_{sys}))}\over{Y_0+1-exp(-\mu\eta_{sys})}}\\
Q_1 &=\mu exp(-\mu)(Y_0+\eta_{sys})\\
e_1 &={{{1\over 2}Y_0+e_d\eta_{sys}}\over{Y_0+\eta_{sys}}}
\end{aligned}
\end{equation}
Now, if $\eta \ll 1$, we can use the approximation $1-exp(-\mu\eta_{sys})=\mu\eta_{sys}$. If the dark/background count rate $Y_0$ also satisfies $Y_0 \ll \eta_{sys}$ (which is a reasonable approximation, since with parameters in Table II, $Y_0$ is at the order of $10^{-5}$, while $\eta_d \eta_{critical}$ is at the order of $10^{-3}$), we can write
\begin{equation}
\begin{aligned}
{Q_\mu \over Q_1 }&\approx {{Y_0+\mu\eta_{sys}}\over{\mu exp(-\mu)(Y_0+\eta_{sys})}} \approx exp(\mu)\\
e_1 &\approx {1\over 2}{Y_0 \over \eta_{sys}}+e_d \\
E_\mu &\approx {1\over {2\mu}}{Y_0 \over \eta_{sys}}+e_d
\end{aligned}
\end{equation}
\noindent substituting back into Eq. D12, and defining
\begin{equation}
x={Y_0 \over \eta_{sys}}={Y_0\over{\eta_d\eta}}
\end{equation}
\noindent we can have
\begin{equation}
h_2({1\over 2}x+e_d)+fe^\mu h_2({1\over {2\mu}}x+e_d)=1
\end{equation}
\noindent this is a function that is only determined by $e_d$ and $\mu$. We can write its solution for x as
\begin{equation}
x_{critical}=\mathcal{F}(e_d,\mu)
\end{equation}
Then the critical transmittance (i.e. optimal threshold position) can be written as
\begin{equation}
\eta_{critical}={{Y_0}\over{\eta_d}} [{1\over\mathcal{F}(e_d,\mu)}]
\end{equation}
\noindent where $\mathcal{F}(e_d,\mu)$ does not have an explicit analytical expression, because $h_2$ function cannot be analytically expanded. (One can, however, numerically use linear fit to expand $h_2$, if given the approximate range of the experimental parameters $e_d$ and $\mu$). The important observation here, however, is that for the decoy-state case, we can still have:
\begin{equation}
\eta_{critical} \propto{Y_0 \over \eta_d}
\end{equation}
\noindent which points out that the critical threshold is directly proportional to the dark (or background) count rate of the experimental devices, and inversely proportional to the detector efficiency.
\section{PDTC parameters}
In our simulations, we have fixed several typical values for $\sigma$ for free-space QKD, corresponding to the case of weak-to-medium level turbulence, and have considered the PDTC to be a fixed distribution for a given $\sigma$ regardless of the channel loss. In reality, though, $\sigma$ is distance-dependent, too. A commonly used estimation for $\sigma$ is the "Rytov Approximation"\cite{rytov}
\begin{equation}
\sigma^2 = 1.23 C_n^2 k^{7/6} L^{11/6}
\end{equation}
\noindent which relates $\sigma$ both to the distance $L$ and the refractive index structure constant $C_n^2$ (which is determined by atmospheric conditions).
Also, with simulation software such as MODTRAN\cite{modtran}, it is possible to simulate the relationship between $\eta_0$ and $L$ for a given free-space channel. Therefore, one necessary next step would also be to estimate performance for cases with realistic values for $\eta_0$ and $\sigma$, both from literature and from simulations, as well as to study the possible correlation $\sigma$ and $\eta_0$ (both related to $L$) in simulations.
\end{document}
|
arXiv
|
{
"id": "1712.08949.tex",
"language_detection_score": 0.8177902102470398,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{\bf Frames of translates with prescribed fine structure \\ in shift invariant spaces} \author{Mar\'\i a J. Benac $^{*}$, Pedro G. Massey $^{*}$, and Demetrio Stojanoff \footnote{Partially supported by CONICET (PIP 0435/10) and Universidad Nacional de La PLata (UNLP 11X681) } \
\footnote{ e-mail addresses: [email protected] , [email protected] , [email protected]} \\ {\small Depto. de Matem\'atica, FCE-UNLP and IAM-CONICET, Argentina }} \date{} \maketitle
\begin{abstract}
For a given finitely generated shift invariant (FSI) subspace ${\cal W}\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ we obtain a simple criterion for the existence of shift generated (SG) Bessel sequences $E(\mathcal F)$ induced by finite sequences of vectors $\mathcal F\in {\cal W}^n$ that have a prescribed fine structure i.e., such that the norms of the vectors in $\mathcal F$ and the spectra of $S_{E(\mathcal F)}$ is prescribed in each fiber of $\text{Spec}({\cal W})\subset \mathbb{T}^k$. We complement this result by developing an analogue of the so-called sequences of eigensteps from finite frame theory in the context of SG Bessel sequences, that allows for a detailed description of all sequences with prescribed fine structure. Then, given $0<\alpha_1\leq \ldots\leq \alpha_n$ we characterize the finite sequences $\mathcal F\in{\cal W}^n$ such that $\|f_i\|^2=\alpha_i$, for $1\leq i\leq n$, and such that the fine spectral structure of the shift generated Bessel sequences $E(\mathcal F)$ have minimal spread (i.e. we show the existence of optimal SG Bessel sequences with prescribed norms); in this context the spread of the spectra is measured in terms of the convex potential $P^{\cal W}_\varphi$ induced by ${\cal W}$ and an arbitrary convex function $\varphi:\mathbb{R}}\def\C{\mathbb{C}_+\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$. \end{abstract}
\noindent AMS subject classification: 42C15.
\noindent Keywords: frames of translates, shift invariant subspaces, Schur-Horn theorem, frame design problems, convex potentials.
\tableofcontents
\section{Introduction} Let ${\cal W}$ be a closed subspace of a separable complex Hilbert space $\mathcal{H}$ and let $\mathbb I$ be a finite or countable infinite set. A sequence $\mathcal F=\{f_i\}_{i\in \mathbb I}$ in ${\cal W}$ is a frame for ${\cal W}$ if there exist positive constants $0<a\leq b$ such that $$
a \, \|f\|^2\leq \sum_{i\in \mathbb I}|\langle f,f_i\rangle |^2\leq b\,\|f\|^2 \peso{for every} f\in {\cal W}\, . $$ If we can choose $a=b$ then we say that $\mathcal F$ is a tight frame for ${\cal W}$. A frame $\mathcal F$ for ${\cal W}$ allows for linear (typically redundant) and stable encoding-decoding schemes of vectors (signals) in ${\cal W}$. Indeed, if ${\cal V}$ is a closed subspace of $\mathcal{H}$ such that ${\cal V}\oplus{\cal W}^\perp=\mathcal{H}$ (e.g. ${\cal V}={\cal W}$) then it is possible to find frames $\mathcal G=\{g_i\}_{i\in \mathbb I}$ for ${\cal V}$ such that \begin{equation}\label{eq: intro duals} f=\sum_{i\in \mathbb I}\langle f,g_i\rangle \ f_i \ , \quad \text{ for } f\in{\cal W}\,. \end{equation} The representation above lies within the theory of oblique duality (see \cite{YEldar3,CE06,YEldar1,YEldar2}). In applied situations, it is usually desired to develop encoding-decoding schemes as above, with some additional features related with stability of the scheme.
In some cases, we search for schemes such that the sequence of norms $\{\|f_i\|^2\}_{i\in \mathbb I}$ as well as the spectral properties of the family $\mathcal F$ are given in advance, leading to what is known in the literature as frame design problem (see \cite{AMRS,BF,CFMPS,CasLeo,MR08,Pot} and the papers \cite{FMP,MRS13,MRS14,MRS13b} for the more general frame completions problem with prescribed norms). It is well known that both the spread of the sequences of norms as well as the spread of the spectra of the frame $\mathcal F$ are linked with numerical properties of $\mathcal F$. Once we have constructed a frame $\mathcal F$ for ${\cal W}$ with the desired properties, we turn our attention to the construction of frames $\mathcal G$ for ${\cal V}$ satisfying Eq.\eqref{eq: intro duals} and having some prescribed features related with their numerical stability (see \cite{BMS14,BMS15,CE06,MRS13,MRS13b}).
\noi It is well known that the frame design problem has an equivalent formulation in terms of the relation between the main diagonal of a positive semi-definite operator and its spectra; in the finite dimensional setting this relation is characterized in the Schur-Horn theorem from matrix analysis. There has been recent important advances in both the frame design problems as well as the Schur-Horn theorems in infinite dimensions, mainly due to the interactions of these problems (see \cite{AMRS,BoJa,BoJa2,BoJa3,Jas,KaWe}). There are also complete parametrizations of all finite frames with prescribed norms and eigenvalues (of their frame operators) in terms of the so-called eigensteps sequences \cite{CFMPS}. On the other hand, the spectral structure of oblique duals (that include classical duals) of a fixed frame can be described in terms of the relations between the spectra of a positive semi-definite operator and the spectra of its compressions to subspaces. In the finite dimensional context (see \cite{BMS14,MRS13}) these relations are known as the Fan-Pall inequalities (that include the so-called interlacing inequalities as a particular case). Yet, in general, the corresponding results in frame theory do not take into consideration any additional structure of the frame. For example, regarding the frame design problem, it seems natural to wonder whether we can construct a structured frame (e.g., wavelet, Gabor or a shift generated frame) with prescribed structure; similarly, in case we fix a structured frame $\mathcal F$ for ${\cal W}$ it seems natural to wonder whether we can construct structured oblique dual frames with further prescribed properties.
\noi In \cite{BMS15}, as a first step towards a detailed study of the spectral properties of structured oblique duals of shift generated systems induced by finite families of vectors in $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$, we extended the Fan-Pall theory to the context of measurable fields of positive semi-definite matrices and their compressions by measurable selections of subspaces; this allowed us to give an explicit description of what we called {\it fine spectral structure} of the shift generated duals of a fixed shift generated (SG) frame for a finitely generated shift invariant (FSI) subspace ${\cal W}$ of $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$. Given a convex function $\varphi:\mathbb{R}}\def\C{\mathbb{C}_+\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ we also introduced the convex potential associated to pair the $(\varphi,{\cal W})$, that is a functional on SG Bessel sequences that measures the spread of the fine spectral structure of the sequence; there we showed that these convex potentials detect tight frames as their minimizers (under some normalization conditions). Yet, our analysis was based on the fine spectral structure of a given SG Bessel sequence in a FSI subspace ${\cal W}\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$.
\noi
In this paper, building on an extension of the Schur-Horn theorem for measurable fields of positive semi-definite matrices, we characterize the possible {\it fine structures} of SG Bessel sequences in FSI subspaces (see Section \ref{SI cosas} for preliminaries on SG Bessel sequences, Remark \ref{rem sobre estruc fina} and Theorem \ref{teo sobre disenio de marcos});
thus, we solve a frame design problem, where the prescribed features of the SG Bessel sequences are described in terms of some internal (or fine) structure, relative to a finitely generated shift invariant subspace ${\cal W}$. We also show that the Fan-Pall theory for fields of positive semi-definite matrices can be used to obtain a detailed description of SG Bessel sequences with prescribed fine structure, similar to that obtained in terms of the eigensteps in \cite{CFMPS}. In turn, we use these results to show that
given a FSI subspace ${\cal W}$, a convex function $\varphi:\mathbb{R}}\def\C{\mathbb{C}_+\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ and a finite sequence of positive numbers $\alpha_1\geq \ldots\geq \alpha_n>0$, there exist vectors $f_i\in{\cal W}$ such that $\|f_i\|^2=\alpha_i$, for $1\leq i\leq n$, and such that the SG Bessel sequence induced by these vectors minimizes the convex potential associated to the pair $(\varphi,{\cal W})$, among all such SG Bessel sequences (for other optimal design problems in shift invariant spaces see \cite{AlCHM1,AlCHM2}). The existence of these $(\varphi,{\cal W})$-optimal shift generated frame designs with prescribed norms is not derived using a direct ``continuity + compactness'' argument. Actually, their existence follows from a discrete nature of their spectral structure; we make use of the this fact to reduce the problem of describing the structure of optimal designs, to an optimization problem in a finite dimensional setting. As a tool, we consider the waterfilling construction in terms of majorization in general probability spaces. It is worth pointing out that there has been interest in the structure of finite sequences of vectors that minimize convex potentials in the finite dimensional context (see \cite{CKFT,FMP,MR08,MR10}), originating from the seminal paper \cite{BF}; our present situation is more involved and, although we reduce the problem to a finite dimensional setting, this reduction is not related with the techniques nor the results of the previous works on finite families of vectors.
\noi The paper is organized as follows. In Section \ref{sec prelim}, after fixing the general notations used in the paper, we present some preliminary material on frames, shift invariant subspaces and shift generated Bessel sequences; we end this section with the general notion of majorization in probability spaces. In Section \ref{subsec exac charac} we obtain an exact characterization of the existence of shift generated Bessel sequences with prescribed fine structure in terms of majorization relations; this result is based on a version of the Schur-Horn theorem for measurable fields of positive semi-definite matrices (defined on measure spaces) that is developed in the appendix (see Section \ref{Appendixity}). In Section \ref{subsec eigensteps}, building on the Fan-Pall inequalities from \cite{BMS15}, we obtain a detailed description of all shift generated Bessel sequences with prescribed fine structure that generalizes the so-called eigensteps construction in the finite dimensional setting. In Section \ref{sec opti frames with prescribed norms} we show that for a fixed sequence of positive numbers $\alpha_1\geq \ldots\geq \alpha_n>0$, a
convex function $\varphi:\mathbb{R}}\def\C{\mathbb{C}_+\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ and a FSI subspace ${\cal W}\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ there exist vectors $f_i\in{\cal W}$ such that $\|f_i\|^2=\alpha_i$, for $1\leq i\leq n$, and such that $\mathcal F$ minimizes the convex potential associated to the pair $(\varphi,{\cal W})$ among all such finite sequences; in order to do this, we first consider in Section \ref{subse uniform} the uniform case in which the dimensions of the fibers of ${\cal W}$ are constant on the spectrum of ${\cal W}$. The general case of the optimal design problem with prescribed norms in a FSI is studied in Section \ref{subsec gral mi gral}; our approach is based on a reduction of the problem to an optimization procedure in the finite dimensional setting. The paper ends with an Appendix, in which we consider a measurable version of the Schur-Horn theorem needed in Section \ref{subsec exac charac} as well as some technical aspects of an optimization problem needed in Section \ref{subsec gral mi gral}.
\section{Preliminaries} \label{sec prelim}
In this section we recall some basic facts related with frames for subspaces and shift generated frames for shift invariant (SI) subspaces of $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$. At the end of this section we describe majorization between functions in arbitrary probability spaces.
\def\mathbb {I} _m{\mathbb {I} _m} \def\mathbb {I} _d{\mathbb {I} _d} \def\mathds{1}{\mathds{1}}
\noi {\bf General Notations}
\noi Throughout this work we shall use the following notation: the space of complex $d\times d$ matrices is denoted by $\mathcal{M}_d(\mathbb{C})$, the real subspace of self-adjoint matrices is denoted $\mathcal{H}(d)$ and $\mat^+$ denotes the set of positive semi-definite matrices; $\mathcal{G}\textit{l}\,(d)$ is the group of invertible elements of $\mathcal{M}_d(\mathbb{C})$, $\mathcal{U}(d)$ is the subgroup of unitary matrices and $\mathcal{G}\textit{l}\,(d)^+ = \mat^+ \cap \mathcal{G}\textit{l}\,(d)$. If $T\in \mathcal{M}_d(\mathbb{C})$, we denote by
$\|T\|$ its spectral norm, by $\text{\rm rk}\, T= \dim R(T) $ the rank of $T$, and by $\tr T$ the trace of $T$.
\noi Given $d \in \mathbb{N}$ we denote by $\mathbb {I} _d = \{1, \dots , d\} \subseteq \mathbb{N}$ and we set $\mathbb{I}_0=\emptyset$. For a vector $x \in \mathbb{R}}\def\C{\mathbb{C}^d$ we denote by $x^\downarrow\in \mathbb{R}}\def\C{\mathbb{C}^d$
the rearrangement of $x$ in non-increasing order. We denote by $(\mathbb{R}}\def\C{\mathbb{C}^d)^\downarrow = \{ x\in \mathbb{R}}\def\C{\mathbb{C}^d : x = x^\downarrow\}$ the set of downwards ordered vectors. Given $S\in \mathcal{H}(d)$, we write $\lambda(S) = \lambda^\downarrow(S)= (\lambda_1(S) \, , \, \dots \, , \, \lambda_d(S)\,) \in (\mathbb{R}}\def\C{\mathbb{C}^d)^\downarrow$ for the vector of eigenvalues of $S$ - counting multiplicities - arranged in decreasing order.
\noi If $W\subseteq \C^d$ is a subspace we denote by $P_W \in \mat^+$ the orthogonal projection onto $W$. Given $x\, , \, y \in \C^d$ we denote by $x\otimes y \in \mathcal{M}_d(\mathbb{C})$ the rank one matrix given by \begin{equation} \label{tensores} x\otimes y \, (z) = \langle z\, , \, y\rangle \, x \peso{for every} z\in \C^d \ .
\end{equation} Note that, if $x\neq 0$, then the projection $P_x \ \stackrel{\mbox{\tiny{def}}}{=}\ P_{\gen\{x\}}= \|x\|^{-2} \, x\otimes x \,$.
\subsection{Frames for subspaces}\label{sec defi frames subespacios}
In what follows $\mathcal{H}$ denotes a separable complex Hilbert space and $\mathbb I$ denotes a finite or countable infinite set. Let ${\cal W}$ be a closed subspace of $\mathcal{H}$: recall that a sequence $\mathcal F=\{f_i\}_{i\in \mathbb I}$ in ${\cal W}$ is a {\it frame} for ${\cal W}$ if there exist positive constants $0<a\leq b$ such that \begin{equation}\label{defi frame}
a \, \|f\|^2\leq \sum_{i\in \mathbb I}|\langle f,f_i\rangle |^2\leq b\,\|f\|^2 \peso{for every} f\in {\cal W}\, . \end{equation} In general, if $\mathcal F$ satisfies the inequality to the right in Eq. \eqref{defi frame} we say that $\mathcal F$ is a $b$-Bessel sequence for ${\cal W}$. Moreover, we shall say that a sequence $\mathcal G=\{g_i\}_{i\in\mathbb I}$ in $\mathcal{H}$ is a Bessel sequence - without explicit reference to a closed subspace - whenever
$\mathcal G$ is a Bessel sequence for its closed linear span; notice that this is equivalent to the fact that $\mathcal G$ is a Bessel sequence for $\mathcal{H}$.
\noi Given a Bessel sequence $\mathcal F=\{f_i\}_{i\in \mathbb I}$ we consider its {\it synthesis operator} $T_\mathcal F\in L(\ell^2(\mathbb I),\mathcal{H})$ given by $T_\mathcal F((a_i)_{i\in \mathbb I})=\sum_{i\in \mathbb I} a_i\ f_i$ which, by hypothesis on $\mathcal F$, is a bounded linear transformation. We also consider $T_\mathcal F^*\in L(\mathcal{H},\ell^2(\mathbb I))$ called the {\it analysis operator} of $\mathcal F$, given by $T_\mathcal F^*(f)=(\langle f,f_i\rangle )_{i\in \mathbb I}$ and the {\it frame operator} of $\mathcal F$ defined by $S_\mathcal F=T_\mathcal F\,T_\mathcal F^*$. It is straightforward to check that $$
\langle S_\mathcal F f,f \rangle =\sum_{i\in \mathbb I}|\langle f,f_i\rangle |^2 \peso{for every}
f\in \mathcal{H}\ .
$$ Hence, $S_\mathcal F$ is a positive semi-definite bounded operator; moreover, a Bessel sequence $\mathcal F$ in ${\cal W}$ is a frame for ${\cal W}$ if and only if $S_\mathcal F$ is an invertible operator when restricted to ${\cal W}$ or equivalently, if the range of $T_\mathcal F$ coincides with ${\cal W}$.
\noi If ${\cal V}$ is a closed subspace of $\mathcal{H}$ such that ${\cal V}\oplus{\cal W}^\perp=\mathcal{H}$ (e.g. ${\cal V}={\cal W}$) then it is possible to find frames $\mathcal G=\{g_i\}_{i\in \mathbb I}$ for ${\cal V}$ such that $$f=\sum_{i\in \mathbb I}\langle f,g_i\rangle \ f_i \ , \quad \text{ for } f\in{\cal W}\,. $$ The representation above lies within the theory of oblique duality (see \cite{YEldar3,CE06,YEldar1,YEldar2}). In this note we shall not be concerned with oblique duals; nevertheless, notice that the numerical stability of the encoding-decoding scheme above depends both on the numerical stability corresponding to $\mathcal F$ and $\mathcal G$ as above. One way to measure stability of the encoding or decoding algorithms is to measure the spread of the spectra of the frame operators corresponding to $\mathcal F$ and $\mathcal G$. Therefore both the task of constructing optimally stable $\mathcal F$ together with obtaining optimally stable duals $\mathcal G$ of $\mathcal F$ are of fundamental interest in frame theory.
\subsection{SI subspaces, frames of translates and their convex potentials}\label{SI cosas}
In what follows we consider $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ (with respect to Lebesgue measure) as a separable and complex Hilbert space. Recall that a closed subspace ${\cal V}\subseteq L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ is {\it shift-invariant} (SI) if $f\in {\cal V}$ implies $T_\ell f \in {\cal V}$ for any $\ell\in \mathbb{Z}^k$, where $T_yf(x)=f(x-y)$ is the translation by $y \in \mathbb{R}}\def\C{\mathbb{C}^k$. For example, take a subset $\mathcal{A} \subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ and set $$
{\cal S}(\mathcal{A})= \overline{\text{span}}\,\{T_\ell f:\ f\in\mathcal{A}\, , \ \ell\in\mathbb Z^k\} \,. $$ Then, ${\cal S}(\mathcal{A})$ is a shift-invariant subspace called the {\it SI subspace generated by $\mathcal{A}$}; indeed, ${\cal S}(\mathcal{A})$ is the smallest SI subspace that contains $\mathcal{A}$. We say that a SI subspace ${\cal V}$ is {\it finitely generated} (FSI) if there exists a finite set $\mathcal{A}\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ such that ${\cal V}={\cal S}(\mathcal{A})$. We further say that ${\cal W}$ is a principal SI subspace if there exists $f\in L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ such that ${\cal W}={\cal S}(f)$.
\noi In order to describe the fine structure of a SI subspace we consider the following representation of $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ (see \cite{BDR,Bo,RS95} and \cite{CabPat} for extensions of these notions to the more general context of actions of locally compact abelian groups). Let $\mathbb{T}=[-1/2,1/2)$ endowed with the Lebesgue measure and let $L^2(\mathbb{T}^k, \ell^2(\mathbb{Z}^k))$ be the Hilbert space of square integrable $\ell^2(\mathbb{Z}^k)$-valued functions that consists of all vector valued measurable functions $\phi: \mathbb{T}^k \to \ell^2(\mathbb{Z}^k)$ with the norm
$$\| \phi\|^2= \int_{\mathbb{T}^k} \| \phi(x)\|_{\ell^2(\mathbb{Z}^k)}^{2} \ dx< \infty.$$ Then, $\Gamma: L^2(\mathbb{R}}\def\C{\mathbb{C}^k)\to L^2(\mathbb{T}^k, \ell^2(\mathbb{Z}^k))$ defined for $f\in L^1(\mathbb{R}}\def\C{\mathbb{C}^k)\cap L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ by \begin{equation}\label{def: iso} \Gamma f: \mathbb{T}^k \to \ell^2(\mathbb{Z}^k)\ ,\quad\Gamma f(x)= (\hat{f}(x+\ell))_{\ell\in \mathbb{Z}^k}, \end{equation} extends uniquely to an isometric isomorphism between $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ and $L^2(\mathbb{T}^k, \ell^2(\mathbb{Z}^k))$; here $$\hat f(x)= \int_{\mathbb{R}}\def\C{\mathbb{C}^k} f(y) \ e^{-2\pi\, i\,\langle y,\,x\rangle} \ dy \quad \text{ for } \quad x\in\mathbb{R}}\def\C{\mathbb{C}^k\, , $$ denotes the Fourier transform of $f\in L^1(\mathbb{R}}\def\C{\mathbb{C}^k)\cap L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$.
\noi Let ${\cal V}\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ be a SI subspace. Then, there exists a function $J_{\cal V}:\mathbb{T}^k\rightarrow\{$ closed subspaces of $\ell^2(\mathbb{Z}^k)\}$ such that: if $ P_{J_{\cal V}(x)}$ denotes the orthogonal projection onto $J_{\cal V}(x)$ for $x\in\mathbb{T}^k$, then for every $\xi,\,\eta\in \ell^2(\mathbb{Z}^k)$ the function $x\mapsto \langle P_{J_{\cal V}(x)} \,\xi\, , \, \eta\rangle$ is measurable and \begin{equation}\label{pro: V y J} {\cal V}=\{ f\in L^2(\mathbb{R}}\def\C{\mathbb{C}^k): \Gamma f(x) \in J_{\cal V}(x) \,\ \text{for a.e.}\,\ x\in \mathbb{T}^k\}. \end{equation} The function $J_{\cal V}$ is the so-called {\it measurable range function} associated with ${\cal V}$. By \cite[Prop.1.5]{Bo}, Eq. \eqref{pro: V y J} establishes a bijection between SI subspaces of $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ and measurable range functions. In case ${\cal V}=S(\mathcal A) \subseteq L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ is the SI subspace generated by $\mathcal A=\{h_i:i\in \mathbb I\}\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$, where $\mathbb I$ is a finite or countable infinite set, then for a.e. $x\in\mathbb{T}^k$ we have that \begin{equation}\label{eq Jv} J_{\cal V}(x)=\overline{\text{span}}\,\{\Gamma h_i(x): \ i\in \mathbb I\}\,. \end{equation}
\noi Recall that a bounded linear operator $S\in L(L^2(\mathbb{R}}\def\C{\mathbb{C}^k))$ is {\it shift preserving} (SP) if $T_\ell \, S=S\,T_\ell$ for every $\ell\in\mathbb{Z}^k$. In this case (see \cite[Thm 4.5]{Bo}) there exists a (weakly) measurable field of operators $[S]_{(\cdot)}:\mathbb{T}^k\rightarrow \ell^2(\mathbb{Z}^k)$ (i.e. such that for every $\xi,\,\eta\in \ell^2(\mathbb{Z}^k)$ the function $\mathbb{T}^k\ni x\mapsto \langle [S]_x\, \xi\, , \, \eta \rangle $
is measurable) and essentially bounded (i.e. the function $\mathbb{T}^k\ni x\mapsto \|\,[S]_{x}\,\|$ is essentially bounded) such that \begin{equation}\label{defi hatS}
[S]_x \big(\Gamma f(x)\,\big)=\Gamma (Sf) (x) \quad \text{ for a.e. }x\in\mathbb{T}^k\ , \ \ f\in L^2(\mathbb{R}}\def\C{\mathbb{C}^k)\,.
\end{equation} Moreover, $\|S\|={\mathrm{ess}\sup}_{x\in\mathbb{T}^k} \|\, [S]_x\, \|$. Conversely, if $s:\mathbb{T}^k\rightarrow L(\ell^2(\mathbb{Z}^k))$ is a weakly measurable and essentially bounded field of operators then,
there exists a unique bounded operator $S\in L(L^2(\mathbb{R}}\def\C{\mathbb{C}^k))$ that is SP and such that $[S]=s$. For example, let ${\cal V}$ be a SI subspace and consider $P_{\cal V}\in L(L^2(\mathbb{R}}\def\C{\mathbb{C}^k))$, the orthogonal projection onto ${\cal V}$; then, $P_{\cal V}$ is SP so that $[P_{\cal V}]{} :\mathbb{T}^k\rightarrow L(\ell^2(\mathbb{Z}^k))$ is given by $[P_{\cal V}]{}_x=P_{J_{\cal V}(x)}$ i.e., the orthogonal projection onto $J_{\cal V}(x)$, for a.e. $x\in\mathbb{T}^k$.
\noi The previous notions associated with SI subspaces and SP operators allow to develop a detailed study of frames of translates. Indeed, let $\mathcal F=\{f_i\}_{i\in \mathbb I}$ be a (possibly finite) sequence in $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$. In what follows we consider the sequence of integer translates of $\mathcal F$, denoted $E(\mathcal F)$ and given by
$$E(\mathcal F)=\{T_\ell \, f_i\}_{(\ell,\, i)\,\in\, \mathbb{Z}^{k}\times \mathbb I}\,.$$
For $x\in \mathbb{T}^k$, let $\Gamma\mathcal F(x)=\{\Gamma f_i(x)\}_{i\in \mathbb I}$ which is a (possibly finite) sequence in $\ell^2(\mathbb{Z}^k)$. Then $E(\mathcal F)$ is a $b$-Bessel sequence if and only if $\Gamma\mathcal F(x)$ is a $b$-Bessel sequence for a.e. $x\in \mathbb{T}^k$ (see \cite{Bo,RS95}). In this case, we consider the synthesis operator $T_{\Gamma\mathcal F(x)}:\ell^2(\mathbb I)\rightarrow \ell^2(\mathbb{Z}^k)$ and frame operator $S_{\Gamma\mathcal F(x)}:\ell^2(\mathbb{Z}^k)\rightarrow \ell^2(\mathbb{Z}^k)$ of $\Gamma\mathcal F(x)$, for $x\in\mathbb{T}^k$. It is straightforward to check that $S_{E(\mathcal F)}$ is a SP operator.
\noi If $\mathcal F=\{f_i\}_{i\in \mathbb I}$ and $\mathcal G=\{g_i\}_{i\in \mathbb I}$ are such that $E(\mathcal F)$ and $E(\mathcal G)$ are Bessel sequences then (see \cite{HG07,RS95}) the following fundamental relation holds: \begin{equation}\label{eq:fourier} [T_{E(\mathcal G)}\,T^*_{E(\mathcal F)}]_x = T_{\Gamma\mathcal G(x)}\,T^*_{\Gamma\mathcal F(x)}\ , \quad \text{for a.e }\, x \in \mathbb{T}^k \,. \end{equation} These equalities have several consequences. For example, if ${\cal W}$ is a SI subspace of $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ and we assume further that $\mathcal F,\,\mathcal G\in{\cal W}^n$ then, for every $f,\,g\in L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$, $$ \langle S_{E(\mathcal F)}\, f,\,g\rangle =\int_{\mathbb{T}^k} \langle S_{\Gamma\mathcal F(x)} \ \Gamma f(x),\,\Gamma g(x)\rangle_{\ell^2(\mathbb{Z}^k)}\ dx\ . $$ This last fact implies that $[S_{E(\mathcal F)}]_x=S_{\Gamma \mathcal F(x)}$ for a.e. $x\in\mathbb{T}^k$. Moreover, $E(\mathcal F)$ is a frame for ${\cal W}$ with frame bounds $0<a\leq b$ if and only if $\Gamma\mathcal F(x)$ is a frame for $J_{\cal W}(x)$ with frame bounds $0<a\leq b$ for a.e. $x\in \mathbb{T}^k$ (see \cite{Bo}).
\noi We end this section with the notion of convex potentials in FSI introduced in \cite{BMS15}\,; in order to describe these potentials we consider the sets \begin{equation}\label{def convf} \convf = \{ \varphi:\mathbb{R}}\def\C{\mathbb{C}_+\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+\ , \ \varphi \ \mbox{ is a convex function} \ \} \end{equation} and $\convfs = \{\varphi\in \convf \ , \ \varphi$ is strictly convex $\}$.
\begin{fed} \label{defi pot conv}\rm Let ${\cal W}$ be a FSI subspace in $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$, let $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$ be such that $E(\mathcal F)$ is a Bessel sequence and consider $\varphi\in \convf$. The convex potential associated to $(\varphi,{\cal W})$ on $E(\mathcal F)$, denoted $P_\varphi^{\cal W}(E(\mathcal F))$, is given by \begin{equation}\label{eq defi pot} P_\varphi^{\cal W}(E(\mathcal F))=\int_{\mathbb{T}^k} \tr(\varphi(S_{\Gamma \mathcal F(x)})\, [P_{\cal W}]_x) \ dx \end{equation} where $\varphi(S_{\Gamma \mathcal F(x)})$ denotes the functional calculus of the positive and finite rank operator $S_{\Gamma \mathcal F(x)}\in L(\ell^2(\mathbb{Z}^k))^+$ and $\tr(\cdot)$ denotes the usual semi-finite trace in $L(\ell^2(\mathbb{Z}^k))$.
$\triangle$ \end{fed}
\begin{exa} Let ${\cal W}$ be a FSI subspace of $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ and let $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$. If we set $\varphi(x)=x^2$ for $x\in \mathbb{R}}\def\C{\mathbb{C}_+$ then, the corresponding potential on $E(\mathcal F)$, that we shall denote $\FP(E(\mathcal F))$, is given by $$ \FP(E(\mathcal F))= \int_{\mathbb{T}^k}
\tr( S_{\Gamma \mathcal F(x)}^2) \ dx= \int_{\mathbb{T}^k} \ \sum_{i,\,j\in \mathbb {I} _n} |\langle \Gamma f_i(x),\Gamma f_j(x)\rangle|^2 \ dx\,, $$ where we have used the fact that $\varphi(0)=0$ in this case. Hence, $\FP(E(\mathcal F))$ is a natural extension of the Benedetto-Fickus frame potential (see \cite{BF}).
$\triangle$ \end{exa}
\noi With the notation of Definition \ref{defi pot conv}, it is shown in \cite{BMS15} that $P_\varphi^{\cal W}(E(\mathcal F))$ is a well defined functional on the class of Bessel sequences $E(\mathcal F)$ induced by a finite sequence $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$ as above. The main motivation for considering convex potentials is that, under some natural normalization hypothesis, they detect tight frames as their minimizers (see \cite[Theorem 3.9.]{BMS15} or Corollary \ref{cororo1} below); that is, convex potentials provide simple scalar measures of stability that can be used to compare shift generated frames. Therefore, the convex potentials for FSI are natural extensions of the convex potentials in finite dimensions introduced in \cite{MR10}. In what follows, we shall consider the existence of tight frames $E(\mathcal F)$ for the FSI ${\cal W}$ with prescribed norms. It turns out that there are natural restrictions for the existence of such frames (see Theorem \ref{teo sobre disenio de marcos} below). In case these restrictions are not fulfilled then, the previous remarks show that minimizers of convex potentials associated to a pair $(\varphi,\,{\cal W})$ within the class of frames with prescribed norms are natural substitutes of tight frames.
\subsection{Majorization in probability spaces}\label{2.3}
Majorization between vectors (see \cite{Bhat,MaOl}) has played a key role in frame theory. On the one hand, majorization allows to characterize the existence of frames with prescribed properties (see \cite{AMRS,CFMPS,CasLeo}). On the other hand, majorization is a preorder relation that implies a family of tracial inequalities; this last fact can be used to explain the structure of minimizers of general convex potentials, that include the Benedetto-Fickus' frame potential (see \cite{BF,CKFT,MR08,MR10,MRS13,MRS14,MRS13b}). We will be dealing with convex potentials in the context of Bessel families of integer translates of finite sequences; accordingly, we will need the following general notion of majorization between functions in probability spaces.
\noi Throughout this section the triple $(X,\mathcal{X},\mu)$ denotes a probability space i.e.
$\mathcal{X}$ is a $\sigma$-algebra of sets in $X$ and $\mu$ is a probability measure defined on $\mathcal{X}$. We shall denote by $L^\infty(X,\mu)^+ = \{f\in L^\infty(X,\mu): f\ge 0\}$. For $f\in L^\infty(X, \mu)^+$, the {\it decreasing rearrangement} of $f$ (see \cite{MaOl}), denoted $f^*:[0,1)\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$, is given by \begin{equation}\label{eq:reord} f^*(s ) \ \stackrel{\mbox{\tiny{def}}}{=}\ \sup \,\{ t\in \mathbb{R}}\def\C{\mathbb{C}_+ : \ \mu \{x\in X:\ f(x)>t\} >s\} \peso{for every} s\in [0,1)\, . \end{equation}
\begin{rem} \label{rem:prop rear elem}We mention some elementary facts related with the decreasing rearrangement of functions that we shall need in the sequel. Let $f\in L^\infty(X,\mu)^+$, then: \begin{enumerate} \item $f^*$ is a right-continuous and non-increasing function.
\item $f$ and $f^*$ are equimeasurable i.e. for every Borel set $A\subset \mathbb{R}}\def\C{\mathbb{C}$ then $\mu(f^{-1}(A))=|(f^*)^{-1}(A)|$, where $|B|$ denotes the Lebesgue measure of the Borel set $B\subset \mathbb{R}}\def\C{\mathbb{C}$. In turn, this implies that for every continuous $\varphi:\mathbb{R}}\def\C{\mathbb{C}_+\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ then: $\varphi\circ f\in L^\infty(X,\mu)$ iff $\varphi\circ f^*\in L^\infty([0,1])$ and in this case \begin{equation}\label{reor int} \int_X \varphi\circ f\ d\mu =\int_{0}^1 \varphi\circ f^*\ dx\ . \end{equation} \item If $g\in L^\infty(X,\mu)$ is such that $f\leq g$ then $0\leq f^*\leq g^*$; moreover, in case $f^*=g^*$ then $f=g$.
$\triangle$ \end{enumerate} \end{rem}
\begin{fed}\rm Let $f, g\in L^\infty (X, \mu)^+$ and let $f^*,\,g^*$ denote their decreasing rearrangements. We say that $f$ \textit{submajorizes} $g$ (in $(X,\mathcal{X},\mu)$), denoted $g \prec_w f$, if \begin{eqnarray*}\label{eq: mayo de func} \int_{0}^{s} g^*(t) \,\ dt &\leq& \int_{0}^{s} f^*(t)\,\ dt \peso{for every} 0\leq s\leq 1 \,. \end{eqnarray*} If in addition $\int_{0}^{1} g^*(t)\,\ dt = \int_{0}^{1} f^*(t)\,\ dt$
we say that $f$ \textit{majorizes} $g$ and write $g \prec f$.
$\triangle$ \end{fed}
\noi In order to check that majorization holds between functions in probability spaces, we can consider the so-called {\it doubly stochastic maps}. Recall that a linear operator $D$ acting on $L^\infty(X,\mu)$ is a doubly-stochastic map if $D$ is unital, positive and trace preserving i.e. $$ D(1_X)=1_X \ , \ \ D\big(\, L^\infty (X, \mu)^+ \, \big)\subseteq L^\infty (X, \mu)^+ \peso{and} \int_X D(f)(x)\ d\mu(x) =\int_X f(x)\ d\mu(x) $$ for every $f\in L^\infty(X,\mu)$. It is worth pointing out that $D$ is necessarily a contractive map.
\noi Our interest in majorization relies in its relation with integral inequalities in terms of convex functions. The following result summarizes this relation as well as the role of the doubly stochastic maps (see for example \cite{Chong,Ryff}). Recall that $\convf$ and $\convfs$ (see Eq. \eqref{def convf}) denote the sets of convex and strictly convex functions $\varphi:\mathbb{R}}\def\C{\mathbb{C}_+\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$, respectively.
\begin{teo}\label{teo porque mayo} \rm Let $f,\,g\in L^\infty (X, \mu)^+$. Then the following conditions are equivalent: \begin{enumerate} \item $g\prec f$; \item There is a doubly stochastic map $D$ acting on $L^\infty(X,\mu)$ such that $D(f)=g$; \item For every $\varphi \in \convf$ we have that \begin{equation}\label{eq teo:desi mayo}
\int_X \varphi(g(x)) \ d\mu(x)\leq \int_X \varphi(f(x))\ d\mu(x)\ . \end{equation} \end{enumerate}
Similarly, $g\prec_w f \iff $ Eq. \eqref{eq teo:desi mayo} holds for every {non-decreasing} convex function $\varphi$. \qed \end{teo}
\noi The following result plays a key role in the study of the structure of minimizers of $\prec_w$ within (appropriate) sets of functions.
\begin{pro}[\cite{Chong}]\label{pro int y reo} \rm Let $f,\,g\in L^\infty(X,\mu)^+$ such that $g\prec_w f$. If there exists $\varphi\in\convfs$ such that \begin{equation} \int_X \varphi(f(x))\ d\mu(x) =\int_X \varphi(g(x)) \ d\mu(x) \peso{then} g^*=f^* \ . \tag*{
$\square$} \end{equation} \end{pro}
\noi
\section{Existence of shift generated frames with prescribed fine struture}
In this section we characterize the fine structure of a Bessel sequence $E(\mathcal F)$, where $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$. By the fine (or relative) structure of $E(\mathcal F)$ we mean the sequence of norms of the vectors $\Gamma \mathcal F(x)=(\Gamma f_i(x)) _{i\in\mathbb {I} _n}$ and the sequence of eigenvalues of $[S_{E(\mathcal F)}]_x$ for $x\in\mathbb{T}^k$ (see Remark \ref{rem sobre estruc fina} for a precise description). As we shall see, the possible fine structure of $E(\mathcal F)$ can be described in terms of majorization relations.
\subsection{A complete characterization in terms of majorization relations}\label{subsec exac charac}
\noi We begin by showing the existence of measurable spectral representations of self-adjoint SP operators with range lying in a FSI subspace (see Lemma \ref{lem spect represent ese}), which follow from results from \cite{RS95} regarding the existence of measurable fields of eigenvectors and eigenvalues (counting multiplicities and arranged in non-increasing order) of measurable fields $M:\mathbb{T}^k \rightarrow \mathcal{H}(d)$ of selfadjoint matrices. In order to do that, we first recall some notions and results from \cite{Bo}.
\noi Given ${\cal W}\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ a FSI subspace, we say that $f\in {\cal W}$ is a quasi-orthogonal generator of ${\cal W}$ if \begin{equation}\label{eq:}
\|g\|^2=\sum_{\ell\in \mathbb{Z}^k} |\langle T_\ell f, g \rangle|^2\ , \peso{for every} g\in {\cal W}\, . \end{equation} The next theorem, which is a consequence of results from \cite{Bo}, provides a decomposition of any FSI subspace of $L^2(\mathbb{R}}\def\C{\mathbb{C}^n)$ into a finite orthogonal sum of principal SI subspaces with quasi-orthogonal generators.
\begin{teo}[\cite{Bo}]\label{teo:la descom de Bo} \rm Let ${\cal W}$ be a FSI subspace of $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$, with $d=\supes_{x\in \mathbb{T}^k} d(x)$, where $d(x)=\dim J_{{\cal W}}(x)$ for $x\in\mathbb{T}^k$. Then there exist $h_1,\ldots,h_d\in{\cal W}$ such that ${\cal W}$ can be decomposed as an orthogonal sum \begin{eqnarray} {\cal W}=\bigoplus_{j\in\mathbb{I}_{d}} {\cal S}(h_j), \end{eqnarray} where $h_j$ is a quasi orthogonal generator of ${\cal S}(h_j)$ for $j\in\mathbb{I}_{d}\,$, and $\text{Spec}({\cal S}(h_{j+1}))\subseteq \text{Spec}({\cal S}(h_j))$ for $j\in \mathbb{I}_{d-1}\,$. Moreover, in this case $\{\Gamma h_j(x)\}_{j\in\mathbb{I}_{d(x)}}$ is a ONB of $J_{\cal W}(x)$ for a.e. $x\in\mathbb{T}^k$. \qed \end{teo}
\begin{lem}\label{lem spect represent ese} Let ${\cal W}$ be a FSI subspace in $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ with $d=\supes_{x\in \mathbb{T}^k} d(x)$, where $d(x)=\dim J_{{\cal W}}(x)$ for $x\in\mathbb{T}^k$. Let $S\in L(L^2(\mathbb{R}}\def\C{\mathbb{C}^k))$ be a SP self-adjoint operator such that $R(S)\subseteq{\cal W}$. Then, there exist: \begin{enumerate} \item measurable vector fields $v_j:\mathbb{T}^k\rightarrow \ell^2(\mathbb{Z}^k)$ for $j\in\mathbb{I}_{d}$ such that $v_j(x)=0$ if $j>d(x)$ and $\{v_j(x)\}_{j\in\mathbb{I}_{d(x)}}$ is an ONB for $J_{\cal W}(x)$ for a.e. $x\in\mathbb{T}^k$; \item bounded, measurable functions $\lambda_j:\mathbb{T}^k\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ for $j\in\mathbb{I}_{d}\,$, such that $\lambda_1\geq \ldots\geq\lambda_d$, $\lambda_j(x)=0$ if $j>d(x)$ and \begin{equation}\label{lem repre espec S} [S]_x =\sum_{j\in\mathbb{I}_{d(x)}}\lambda_j(x)\ v_j(x)\otimes v_j(x) \ , \peso{for a.e.}\ x\in\mathbb{T}^k\,. \end{equation} \end{enumerate} \end{lem} \begin{proof} By considering a convenient finite partition of $\mathbb{T}^k$ into measurable sets we can assume, without loss of generality, that $d(x)=d$ for a.e. $x\in \mathbb{T}^k$. In this case, by Theorem \ref{teo:la descom de Bo} we have that $$ {\cal W}=\bigoplus_{j\in \mathbb{I}_{d}} {\cal S}(h_j) \ , $$ where $h_j\in {\cal W}$, for $j\in\mathbb{I}_{d}$, are such that $\{\Gamma h_j(x)\}_{j\in\mathbb{I}_{d}}$ is a ONB of $J_{{\cal W}}(x)$ for a.e. $x\in \mathbb{T}^k$. Consider the measurable field of self-adjoint matrices $M(\cdot):\mathbb{T}^k\to \mathcal{H}(d)$ given by $$ M(x)=\big(\langle [S]_{x} \,\ \Gamma h_j(x) ,\,\ \Gamma h_i(x)\rangle\big)_{i,j\,\in \mathbb{I}_{d}}\ . $$ By \cite{RS95}, we can consider measurable functions $\lambda_j:\mathbb{T}^k\to \mathbb{R}}\def\C{\mathbb{C}_+$ for $j\in \mathbb{I}_d\,$, such that $\lambda_1\geq \ldots\geq \lambda_d$ and measurable vector fields $w_j:\mathbb{T}^k\to \C^d$ for $j\in \mathbb{I}_{d}\,$, such that $\{w_j(x)\}_{j\in \mathbb{I}_{d}}$ is a ONB of $\C^d$ and \begin{equation}\label{ecuac agreg1} M(x)=\sum_{j\in \mathbb{I}_{d}} \lambda_j(x)\,\ w_j(x) \otimes w_j(x) \peso{for a.e.} \ x\in \mathbb{T}^k\, . \end{equation} If $w_j(x)=(w_{ij}(x))_{i\in\mathbb{I}_{d}}$ for $j\in\mathbb{I}_{d}\,$, consider the measurable vector fields $v_j:\mathbb{T}^k \to \ell^2(\mathbb{Z}^k)$ for $j\in \mathbb{I}_{d}\,$, given by $$ v_j(x) =\sum_{i\in \mathbb{I}_{d}} w_{ij}(x)\,\ \Gamma h_i(x)\,\ \text{for}\,\ x\in \mathbb{T}^k\ . $$ Then, it is easy to see that $\{v_j(x)\}_{j\in \mathbb{I}_{d}}$ is ONB of $J_{\cal W}(x)$ for a.e. $x\in\mathbb{T}^k$; moreover, Eq. \eqref{ecuac agreg1} implies that Eq. \eqref{lem repre espec S} holds in this case. \end{proof}
\begin{rem}\label{rem sobre estruc fina} Let ${\cal W}$ be a FSI subspace with $d(x)=\dim J_{\cal W}(x)$ for $x\in\mathbb{T}^k$, and let $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$ be a finite sequence in $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ such that $E(\mathcal F)$
is a Bessel sequence. In what follows we consider: \begin{enumerate} \item the {\it fine spectral structure of} $E(\mathcal F)$, that is the weakly measurable function $$ \mathbb{T}^k\ni x\mapsto (\lambda_j([S_{E(\mathcal F)}]_x )\,)_{j\in\mathbb{N}}\in \ell^1_+(\mathbb{Z}^k) \ , $$ with $\lambda_j([S_{E(\mathcal F)}]_x )=\lambda_j(x)$ as in Lemma \ref{lem spect represent ese} for $j\in\mathbb{I}_{d(x)}\,$, and $\lambda_j([S_{E(\mathcal F)}]_x )=0$ for $j\geq d(x)+1$
and $x\in \mathbb{T}^k$. Thus, the fine spectral structure of $\mathcal F$ describes the eigenvalues of the positive finite rank operator $[S_{E(\mathcal F)}]_x =S_{\Gamma \mathcal F(x)}\in L(\ell^2(\mathbb{Z}^k))$, counting multiplicities and arranged in non-increasing order.
\item The {\it fine structure of} $E(\mathcal F)$ given by the fine spectral structure together with the measurable vector valued function $\mathbb{T}^k\ni x\mapsto (\|\Gamma f_i(x)\|^2)_{i\in\mathbb {I} _n}\in\mathbb{R}}\def\C{\mathbb{C}_+^n\,$.
$\triangle$ \end{enumerate} \end{rem}
\noi In order to state our main result of this section we shall need the notion of vector majorization from matrix analysis. Recall that given $a=(a_i)_{i\in \mathbb {I} _n}\in \mathbb{R}}\def\C{\mathbb{C}^n$ and $b=(b_i)_{i\in \mathbb {I} _n}\in \mathbb{R}}\def\C{\mathbb{C}^m$ we say that $a$ is majorized by $b$, denoted $a\prec b$, if \begin{equation}\label{eq: mayo dif} \sum_{i\in \mathbb{I}_k} a_i \leq \sum_{i\in \mathbb{I}_k} b_i\ , \ \ 1\leq k \leq \min\{n, m\} \ \ \text{and} \ \ \sum_{i\in \mathbb{I}_n} a_i = \sum_{i\in \mathbb{I}_m} b_i\,. \end{equation}
\begin{teo}[Existence of shift generated sequences with prescribed fine structure] \label{teo sobre disenio de marcos}
Let ${\cal W}$ be a FSI subspace in $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ and let $d(x)=\dim J_{{\cal W}}(x)$ for $x\in\mathbb{T}^k$. Given measurable functions $\alpha_j:\mathbb{T}^k \to \mathbb{R}}\def\C{\mathbb{C}_+$ for $j\in \mathbb {I} _n$ and $\lambda_j:\mathbb{T}^k \to \mathbb{R}}\def\C{\mathbb{C}_+$ for $j\in \mathbb{N}$,
the following conditions are equivalent: \begin{enumerate} \item There exists $\mathcal F=\{f_j\}_{j\in \mathbb {I} _n} \in {\cal W}^n$ such that $E(\mathcal F)$ is a Bessel sequence and: \begin{enumerate}
\item $\|\Gamma f_j(x)\|^2=\alpha_j(x)$ for $j\in \mathbb {I} _n$ and a.e. $x\in\mathbb{T}^k$; \item $\lambda_j([S_{E(\mathcal F)}]_x ) =\lambda_j(x)$ for $j\in\mathbb{N}$ and a.e. $x\in\mathbb{T}^k$. \end{enumerate} \item The following admissibility conditions hold: \begin{enumerate} \item $\lambda_j(x)=0$ for a.e. $x\in\mathbb{T}^k$ such that $j\geq\min\{d(x),n\}+1$. \item $(\alpha_j(x))_{j\in\mathbb {I} _n}\prec (\lambda_j(x))_{j\in\mathbb{I}_{d(x)}}$ for a.e. $x\in\mathbb{T}^k$. \end{enumerate} \end{enumerate} \end{teo}
\noi Our proof of Theorem \ref{teo sobre disenio de marcos} is based on the following extension of a basic result in matrix analysis related with the Schur-Horn theorem (for its proof, see section \ref{Appendixity} - Appendix). In what follows we let $D_b\in \mathcal{M}_d(\mathbb{C})$ be the diagonal matrix with main diagonal $b\in\C^d$.
\begin{teo}\label{teo:mayo equiv} \rm Let $b:\mathbb{T}^k\rightarrow (\mathbb{R}}\def\C{\mathbb{C}_+)^d$ and $c:\mathbb{T}^k\rightarrow (\mathbb{R}}\def\C{\mathbb{C}_+)^n$ be measurable vector fields. The following statements are equivalent: \begin{enumerate} \item For a.e. $x\in \mathbb{T}^k$ we have that $c(x)\prec b(x)$.
\item There exist measurable vector fields $u_j: \mathbb{T}^k\to \C^d$ for $j\in\mathbb {I} _n$ such that $\|u_j(x)\|=1$ for a.e. $x\in \mathbb{T}^k$ and $j\in \mathbb{I}_n\,$, and such that \begin{equation} D_{b(x)}=\sum_{j\in \mathbb{I}_n} c_j(x)\,\ u_j(x) \otimes u_j(x) \ , \peso{for a.e.} \ x\in \mathbb{T}^k\ . \tag*{
$\square$} \end{equation} \end{enumerate} \end{teo}
\begin{proof}[Proof of Theorem \ref{teo sobre disenio de marcos}]
Assume that there exists $\mathcal F=\{f_j\}_{j\in \mathbb {I} _n} \in {\cal W}^n$ such that $\|\Gamma f_j(x)\|^2=\alpha_j(x)$, for $j\in \mathbb {I} _n\,$, and $\lambda_j([S_{E(\mathcal F)}]_x ) =\lambda_j(x)$ for $j\in\mathbb{N}$ and a.e. $x\in\mathbb{T}^k$. Consider the measurable field of positive semi-definite matrices $G:\mathbb{T}^k\to {\cal M}_n(\C)^+$ given by the Gramian $$ G(x)=\Big(\, \big\langle \Gamma f_i(x)\, , \, \Gamma f_j(x) \big\rangle\, \Big)_{i,j\in \mathbb {I} _n} \ , \peso{for} x\in \mathbb{T}^k \ . $$ Notice that $G(x)$ is the matrix representation of $T^*_{\Gamma \mathcal F(x)}T_{\Gamma \mathcal F(x)}\in L(\C^n)$ with respect to the canonical basis of $\C^n$ for $x\in \mathbb{T}^k$; using the fact that the finite rank operators $T^*_{\Gamma \mathcal F(x)}T_{\Gamma \mathcal F(x)}$ and $T_{\Gamma \mathcal F(x)}T^*_{\Gamma \mathcal F(x)}=[S_{E(\mathcal F)}]_x $ have the same positive eigenvalues (counting multiplicities) we see that
$$ \lambda_j(G(x)) = \begin{cases} \lambda_j(x) & \peso{for} 1\leq j\leq \min\{d(x),n\} \\ \ \ \ 0 & \peso{for} \min\{d(x),n\}+1\leq j\leq n \end{cases} \peso{for a.e.} x\in\mathbb{T}^k \ . $$
On the other hand, the main diagonal of $G(x)$ is $(\|\Gamma f_j(x)\|^2)_{j\in\mathbb {I} _n}=(\alpha_j(x))_{j\in\mathbb {I} _n}\,$; hence, by the classical Schur-Horn theorem (see \cite{HJ13}) we see that $$ (\alpha_j(x))_{j\in\mathbb {I} _n}\prec \lambda(G(x))\in \mathbb{R}}\def\C{\mathbb{C}_+ ^n \implies (\alpha_j(x))_{j\in\mathbb {I} _n}\prec (\lambda_j(x))_{j\in \mathbb{I}_{d(x)}} \peso{for a.e.} x\in\mathbb{T}^k \ . $$
\noi Conversely, assume that $(\alpha_j(x))_{j\in \mathbb {I} _n} \prec (\lambda_i(x))_{i\in \mathbb{I}_{d(x)}}$ for a.e. $x\in \mathbb{T}^k$.
By considering a convenient finite partition of $\mathbb{T}^k$ into measurable subsets we can assume, without loss of generality, that $d(x)=d$ for $x\in\mathbb{T}^k$. Therefore, by Theorem \ref{teo:mayo equiv}, there exist measurable vector fields $u_j: \mathbb{T}^k\to \C^d$ for $j\in\mathbb {I} _n$ such that $\|u_j(x)\|=1$ for a.e. $x\in \mathbb{T}^k$ and $j\in \mathbb{I}_n$, and such that \begin{equation} \label{diago} D_{\lambda(x)}=\sum_{j\in \mathbb{I}_n} \alpha_j(x)\,\ u_j(x) \otimes u_j(x) \ , \peso{for a.e.} \ x\in \mathbb{T}^k\ , \end{equation} where $\lambda(x)=(\lambda_j(x))_{j\in\mathbb{I}_d}$ for $x\in \mathbb{T}^k$. Now, by Theorem \ref{teo:la descom de Bo} there exist measurable vector fields $v_j:\mathbb{T}^k\rightarrow \ell^2(\mathbb{Z}^k)$ for $j\in\mathbb{I}_d$ such that $\{v_j(x)\}_{j\in\mathbb{I}_d}$ is a ONB of $J_{\cal W}(x)$ for a.e. $x\in\mathbb{T}^k$. Let $u_j(x)=(u_{ij}(x))_{i\in\mathbb{I}_d}$ for $j\in\mathbb {I} _n$ and $x\in\mathbb{T}^k$; then we consider the finite sequence $\mathcal F=\{f_j\}_{j\in\mathbb {I} _n}\in{\cal W}^n$ determined by $\Gamma f_j(x)=\alpha_j^{1/2}(x) \sum_{i\in\mathbb{I}_d} u_{ij}(x) \ v_i(x)$ for $j\in\mathbb {I} _n$ and $x\in\mathbb{T}^k$. It is clear that $$
\|\Gamma f_j(x)\|^2=\|\alpha_j^{1/2}(x)\ u_j(x)\|^2=\alpha_j(x) \peso{ for a.e. } x\in\mathbb{T}^k,\quad j\in\mathbb {I} _n\ . $$ Moreover, using Eq. \eqref{diago} it is easy to see that $$\left(\sum_{j\in\mathbb {I} _n} \Gamma f_j(x)\otimes \Gamma f_j(x) \right)\, v_i(x)= \lambda_i(x)\,v_i(x) \peso{for} i\in\mathbb{I}_d \ \ \text { and \ a.e. }\ x\in\mathbb{T}^k\,. $$ Hence, $\lambda_j([S_{E(\mathcal F)}]_x ) =\lambda_j(x)$ for $j\in\mathbb{N}$ and a.e. $x\in\mathbb{T}^k$ \end{proof}
\begin{rem}\label{se puede incluso con op SP}
Let ${\cal W}$ be a FSI subspace in $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ and let $d(x)=\dim J_{{\cal W}}(x)$ for $x\in\mathbb{T}^k$. Let $\alpha_j:\mathbb{T}^k \to \mathbb{R}}\def\C{\mathbb{C}_+$, $j\in \mathbb {I} _n\,$, be measurable functions and let $S\in L(L^2(\mathbb{R}}\def\C{\mathbb{C}^k))^+$ be a positive SP operator such that $R(S)\subseteq {\cal W}$. Let $\mathbb{T}^k\ni x\mapsto (\lambda_j([S]_x) )_{j\in\mathbb{N}}$ be the fine spectral structure of $S$ (which is well defined by Lemma \ref{lem spect represent ese}). Assume that for a.e. $x\in\mathbb{T}^k$ we have that $$ (\alpha_j(x))_{j\in\mathbb {I} _n}\prec (\lambda_j([S]_x) )_{j\in\mathbb{I}_{d(x)}}\ . $$ Then, there exists $\mathcal F=\{f_j\}_{j\in \mathbb {I} _n} \in {\cal W}^n$ such that $E(\mathcal F)$ is a Bessel sequence, $$
S_{E(\mathcal F)}=S \peso{and} \|\Gamma f_j(x)\|^2=\alpha_j(x) \peso{for a.e.} x\in\mathbb{T}^k \, , \, j\in\mathbb {I} _n \ . $$ Indeed, if in the proof of Theorem \ref{teo sobre disenio de marcos} above we take the measurable vector fields $v_j:\mathbb{T}^k\rightarrow \ell^2(\mathbb{Z}^k)$ such that $\{v_j(x)\}_{j\in\mathbb{I}_d}$ is a ONB of $J_{\cal W}(x)$ and such that $[S]_x \, v_j(x)=\lambda_j([S]_x)\ v_j$ for a.e. $x\in\mathbb{T}^k$ (notice that this can always be done by Lemma \ref{lem spect represent ese}) then we conclude, as before, that \begin{equation} [S_{E(\mathcal F)}]_x \ v_j(x)= \lambda_j([S]_x)\,v_j(x) \peso{for} j\in\mathbb{I}_d \implies [S_{E(\mathcal F)}]_x = [S]_x \peso{for a.e.} x\in\mathbb{T}^k\ . \tag*{\EOE} \end{equation} \end{rem}
\noi As a first application of Theorem \ref{teo sobre disenio de marcos} we show the existence of shift generated uniform tight frames for an arbitrary FSI. In turn, this allows us to strengthen some results from \cite{BMS15} (see also Corollary \ref{coro tight 2}).
\begin{cor}\label{cororo1} Let $\{0\}\neq{\cal W}\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ be a FSI subspace and let $d(x)=\dim J_{\cal W}(x)$ for $x\in\mathbb{T}^k$. Assume that $n\geq \supes_{x\in\mathbb{T}^k}d(x)$,
let $Z_i=d^{-1}(i)$ for $i\in\mathbb {I} _n\,$ and set $C_{\cal W}= \sum_{i\in\mathbb {I} _n} i\cdot |Z_i|>0$. Then: \begin{enumerate}
\item There exists a sequence $\mathcal F=\{f_j\}_{j\in\mathbb {I} _n}\in{\cal W}^n$ such that $\|f_j\|^2=n^{-1}$ for $j\in\mathbb {I} _n$ and such that $E(\mathcal F)$ is a uniform tight frame for ${\cal W}$.
\item For any sequence $\mathcal G=\{g_j\}_{j\in\mathbb {I} _n}\in{\cal W}^n$ such that $E(\mathcal G)$ is a Bessel sequence and such that $\sum_{j\in\mathbb {I} _n} \|g_j\|^2=1$, and for every $\varphi\in\convf$ we get that: \begin{equation}\label{del item 2} P^{\cal W}_\varphi(E(\mathcal G))\geq C_{\cal W} \ \varphi(C_{\cal W}^{-1})=P^{\cal W}_\varphi(E(\mathcal F))\, . \end{equation} Moreover, if we assume that $\varphi\in\convfs$ then $P^{\cal W}_\varphi(E(\mathcal G))=C_{\cal W} \ \varphi(C_{\cal W}^{-1})$ if and only if $E(\mathcal G)$ is a tight frame for ${\cal W}$. \end{enumerate} \end{cor} \begin{proof}
Let $p_i=|Z_i|$ (where $|A|$ denotes the Lebesgue measure of $A\subset \mathbb{T}^k$ and $Z_i=d^{-1}(i)$) for $1\leq i\leq n$; then $C_{\cal W}= \sum_{i\in\mathbb {I} _n} i\cdot p_i$. Notice that by hypothesis Spec ${\cal W}=\cup_{i\in\mathbb {I} _n} Z_i\,$. For $x\in\mathbb{T}^k$ set $\alpha(x)=0$ if $x\in \mathbb{T}^k\setminus\text{Spec}({\cal W})$ and: \begin{equation} \label{defi alfas} \alpha(x):=\left\{
\begin{array}{ccc}
\frac{j\cdot C_{\cal W}^{-1} }{n}
& {\rm if} & x\in Z_j \ \text{ and } \ p_j>0\,; \\
0 & {\rm if} & x\in Z_j \ \text{ and } \ p_j=0 \, .\\
\end{array}
\right. \end{equation}
Then, it is easy to see that $(\alpha(x))_{i\in\mathbb {I} _n}\prec (C_{\cal W}^{-1})_{i\in\mathbb{I}_{d(x)}}$ for every $x\in \mathbb{T}^k$; hence, by Theorem \ref{teo sobre disenio de marcos} we see that there exists $\mathcal F=\{f_j\}_{j\in\mathbb {I} _n}\in{\cal W}^n$ such that $\|\Gamma f_j(x)\|^2=\alpha(x)$ for $j\in\mathbb {I} _n$ and such that $S_{\Gamma \mathcal F(x)}=C_{\cal W}^{-1}\ P_{J_{\cal W}(x)}$ for a.e. $x\in\mathbb{T}^k$. Therefore, $S_{E(\mathcal F)}=C_{\cal W}^{-1}\, P_{\cal W}$ and $$
\|f_j\|^2=\int_{\mathbb{T}^k} \alpha(x)\ dx= \frac{C_{\cal W}^{-1}}{n} \, \sum_{i\in\mathbb {I} _n} \int_{Z_i} i\ dx=\frac{C_{\cal W}^{-1}}{n}\ \sum_{i\in\mathbb {I} _n} i \cdot p_i=\frac{1}{n}\ . $$ If $\mathcal G$ is as in item 2. then, by \cite{BMS15}, we get the inequality \eqref{del item 2}. Notice that the lower bound is attained at $\mathcal F$ (since it is tight); the last part of the statement was already shown in \cite{BMS15}. \end{proof}
\subsection{Generalized (measurable) eigensteps}\label{subsec eigensteps}
In this section we derive a natural extension of the notion of eigensteps introduced in \cite{CFMPS}, that allows us to describe a procedure to inductively construct finite sequences $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$ such that the fine structure of $E(\mathcal F)$ (that is, the fine spectral structure of $E(\mathcal F)$ and the finite sequence of measurable functions
$\|\Gamma f_i(\cdot)\|^2:\mathbb{T}^k\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$, $i\in\mathbb{I}_n$) are prescribed. Hence, we obtain an in-depth description of a step-by-step construction of Bessel sequences $E(\mathcal F)$ with prescribed fine structure. We point out that our techniques are not based on those from \cite{CFMPS}; indeed, our approach is based on an additive model developed in \cite{BMS15}.
\begin{rem}\label{rem hay eigensteps}Let ${\cal W}$ be a FSI subspace and let $d(x)=\dim J_{\cal W}(x)$ for $x\in\mathbb{T}^k$; let $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$ be such that $E(\mathcal F)$ is a Bessel sequence and set: \begin{enumerate}
\item $\alpha_i(x)=\|\Gamma f_i(x)\|^2$, for $x\in\mathbb{T}^k$ and $i\in\mathbb {I} _n\,$. \item $\lambda_i(x)=\lambda_i([S_{E(\mathcal F)}]_x )$ , for $x\in\mathbb{T}^k$ and $i\in\mathbb {I} _n\,$, where $\mathbb{T}^k\ni x\mapsto (\lambda_i([S_{E(\mathcal F)}]_x )\,)_{i\in\mathbb{N}}$ denotes the fine spectral structure of $E(\mathcal F)$. \end{enumerate} By Theorem \ref{teo sobre disenio de marcos} these functions satisfy the following admissibility conditions: \begin{enumerate} \item[Ad.1] $\lambda_i(x)=0$ for a.e. $x\in\mathbb{T}^k$ such that $i\geq\min\{n,\,d(x)\}+1$. \item[Ad.2] $(\alpha_i(x))_{i\in\mathbb {I} _n}\prec (\lambda_i(x))_{i\in\mathbb{I}_{d(x)}}$ for a.e. $x\in\mathbb{T}^k$. \end{enumerate} For $j\in\mathbb {I} _n$ consider the sequence $\mathcal F_j=\{f_i\}_{i\in\mathbb{I}_{j}}\in{\cal W}^j$. In this case $E(\mathcal F_j) =\{T_\ell f_i\}_{(\ell,i)\in\mathbb{Z}^k\times \mathbb{I}_{j}}$ is a Bessel sequence and $S_j=S_{E(\mathcal F_j)}$ is a SP operator such that $$ [S_j]_x=S_{\Gamma \mathcal F_j(x)} =\sum_{i\in\mathbb{I}_{j}} \Gamma f_i(x)\otimes \Gamma f_i(x)\in L(\ell^2(\mathbb{Z}^k))^+ \peso{for a.e. } x\in\mathbb{T}^k\ , \ \ j\in\mathbb {I} _n\ . $$ For $j\in\mathbb {I} _n$ and $i\in\mathbb{I}_{j}\,$, consider the measurable function $\lambda_{i,j}:\mathbb{T}^k\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ given by $$ \lambda_{i,j}(x)=\lambda_i([S_j]_x)\peso{for} x\in\mathbb{T}^k\ , $$ where $\mathbb{T}^k\ni x\mapsto (\lambda_i([S_j]_x))_{i\in\mathbb{N}}$ denotes the fine spectral structure of $E(\mathcal F_j)$ (notice that by construction $\lambda_i([S_j]_x)=0$ for $i\geq j+1$). Then, it is well known (see \cite{CFMPS}) that
$(\lambda_{i,j}(x))_{i\in\mathbb{I}_{j}}$ interlaces $(\lambda_{i,(j+1)}(x))_{i\in\mathbb{I}_{j+1}}$ i.e. $$ \lambda_{i,(j+1)}(x)\geq \lambda_{i,j}(x)\geq \lambda_{(i+1),(j+1)}(x) \peso{for} i\in\mathbb{I}_{j}\ ,\ j\in\mathbb{I}_{n-1}\ , \ \text{ and a.e. } x\in\mathbb{T}^k\ . $$ Notice that for a.e. $x\in\mathbb{T}^k$, $$ \sum_{i\in\mathbb{I}_{j}} \lambda_{i,j}(x)= \tr([S_j]_x)
= \sum_{i\in\mathbb{I}_{j}}\|\Gamma f_i(x)\|^2 =\sum_{i\in\mathbb{I}_{j}}\alpha_i(x) \peso{for} j\in\mathbb {I} _n\ . $$ Finally notice that by construction $S_n=S_{E(\mathcal F)}$ and hence, $\lambda_{i,n}(x)=\lambda_i(x)$ for $i\in\mathbb {I} _n$ and $x\in \mathbb{T}^k$. These facts motivate the following extension of the notion of eigensteps introduced in \cite{CFMPS}.
$\triangle$ \end{rem}
\begin{fed}\label{defi measiegen} \rm Let ${\cal W}$ be a FSI subspace and let $\lambda_i\, , \, \alpha_i:\mathbb{T}^k\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ for $i\in\mathbb {I} _n$ be measurable functions satisfying the admissibility assumptions Ad.1 and Ad.2 in Remark \ref{rem hay eigensteps}.
A sequence of eigensteps for $(\lambda,\alpha)$ is a doubly-indexed sequence of measurable functions $\lambda_{i,j}:\mathbb{T}^k\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ for $i\in\mathbb{I}_{j}$ and $j\in\mathbb {I} _n$ such that: \begin{enumerate} \item $\lambda_{i,(j+1)}(x)\geq \lambda_{i,j}(x)\geq \lambda_{(i+1),(j+1)}(x)$ for $i\in\mathbb{I}_{j} \, , \, j\in\mathbb{I}_{n-1}\, , \, $
and a.e. $x\in\mathbb{T}^k$; \item $\sum_{i\in\mathbb{I}_{j}} \lambda_{i,j}(x)= \sum_{i\in\mathbb{I}_{j}}\alpha_i(x)$ for $j\in\mathbb {I} _n$ and a.e. $x\in \mathbb{T}^k$; \item $\lambda_{i,n}(x)=\lambda_i(x)$ for $i\in\mathbb {I} _n$ and a.e. $x\in \mathbb{T}^k$.
$\triangle$ \end{enumerate} \end{fed}
\begin{rem}\label{rem hay eigensteps2} Consider the notations and terminology from Remark \ref{rem hay eigensteps}. Then $((\lambda_{i,j}(\cdot))_{i\in\mathbb{I}_{j}})_{j\in\mathbb {I} _n}$ is a sequence of eigensteps for $(\lambda,\alpha)$. We say that $((\lambda_{i,j}(\cdot))_{i\in\mathbb{I}_{j}})_{j\in\mathbb {I} _n}$ is the sequence of eigensteps for $(\lambda,\alpha)$ associated to $\mathcal F$.
$\triangle$ \end{rem}
\noi In what follows we show that every sequence of eigensteps is associated to some $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$ such that $E(\mathcal F)$ is a Bessel sequence (see Theorem \ref{teo todo eigen echachi} below). In order to show this, we recall an additive (operator) model from \cite{BMS15}.
\begin{fed}\label{el conjunto U}\rm Let ${\cal W}$ be a FSI subspace and let $d:\mathbb{T}^k\rightarrow \mathbb{N}_{\geq 0}$ be the measurable function given by $d(x)=\dim J_{{\cal W}}(x)$ for $x\in\mathbb{T}^k$. Let $S\in L(L^2(\mathbb{R}}\def\C{\mathbb{C}^k))^+$ be SP and such that $R(S)\subset {\cal W}$. Given a measurable function $m:\mathbb{T}^k\rightarrow \mathbb{Z}$ such that $m(x)\leq d(x)$ for a.e. $x\in\mathbb{T}^k$ we consider $$U^{\cal W}_m(S) = \Big\{S+ B:B\in L(L^2(\mathbb{R}}\def\C{\mathbb{C}^k))^+ \text{ is SP},\, R(B)\subset {\cal W},\, \text{\rm rk}([B]_x)\leq d(x)-m(x) \ \text{for a.e. } x\in \mathbb{T}^k \Big\} \,. $$
$\triangle$ \end{fed}
\begin{teo}[Appendix of \cite{BMS15}]\label{estructdelU} \rm Consider the notations from Definition \ref{el conjunto U}. Given a measurable function $\mu:\mathbb{T}^k\rightarrow \ell^1(\mathbb{N})^+$ the following are equivalent:
\begin{enumerate}
\item There exists $C\in U^{\cal W}_m(S)$ such that $\lambda(\hat C_x)=\mu(x)$, for a.e. $x\in \mathbb{T}^k$;
\item For a.e. $x\in\mathbb{T}^k\setminus \text{Spec}({\cal W})$ then $\mu(x)=0$; for a.e. $x\in \text{Spec}({\cal W})$ we have that $\mu_i(x)=0$ for $i\geq d(x)+1$ and
\begin{enumerate}
\item in case $m(x)\leq 0$, $\mu_i(x)\geq \lambda_i([S]_ x)$ for $i\in\mathbb{I}_{d(x)}\,$;
\item in case $m(x)\in\mathbb{I}_{d(x)}\,$, $\mu_i(x)\geq \lambda_i([S] _x)$ for $i\in\mathbb{I}_{d(x)}\,$ and \begin{equation} \lambda_i([S]_x)\geq \mu_{(d(x)-m(x))+i}(x) \peso{for} i\in\mathbb{I}_{m(x)}\ . \tag*{
$\square$} \end{equation}
\end{enumerate}
\end{enumerate} \end{teo}
\begin{rem} We point out that Theorem \ref{estructdelU} is obtained in terms of a natural extension of the Fan-Pall interlacing theory from matrix theory, to the context of measurable fields of positive matrices (see \cite[Appendix]{BMS15}); we also notice that the result is still valid for fields (of vectors and operators) defined in measurable subsets of $\mathbb{T}^k$. The original motivation for considering the additive model above was the fact that it describes the set of frame operators of oblique duals of a fixed frame. In the present setting, this additive model will also allow us to link the sequences of eigensteps with the construction of SG Bessel sequences with prescribed fine structure.
$\triangle$ \end{rem}
\begin{teo}\label{teo todo eigen echachi} Let ${\cal W}$ be a FSI subspace and let $\lambda_i,\, \alpha_i:\mathbb{T}^k\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ for $i\in\mathbb {I} _n$ be measurable functions satisfying the admissibility conditions Ad.1 and Ad.2 in Remark \ref{rem hay eigensteps}. Consider a sequence of eigensteps $((\lambda_{i,j}(\cdot))_{i\in\mathbb{I}_{j}})_{j\in\mathbb {I} _n}$ for $(\lambda,\alpha)$. Then, there exists $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$ such that $E(\mathcal F)$ is a Bessel sequence and $((\lambda_{i,j}(\cdot))_{i\in\mathbb{I}_{j}})_{j\in\mathbb {I} _n}$ is the sequence of eigensteps associated to $\mathcal F$. \end{teo}
\begin{proof} First notice that both the assumptions as well as the properties of the objects that we want to construct are checked point-wise; hence, by considering a convenient partition of $\mathbb{T}^k$ into measurable sets we can assume (without loss of generality) that $d(x)=d\geq 1$, for $x\in\mathbb{T}^k$. Now, we argue by induction on $j$. Notice that by hypothesis for $i=j=1$, we see that $\lambda_{1,1}(x)=\alpha_1(x)$ for a.e. $x\in\mathbb{T}^k$. Let $f_1\in{\cal W}$ be such that $\|\Gamma f_1(x)\|^2=\alpha_1(x)$ for a.e. $x\in\mathbb{T}^k$; indeed, we can take $f_1\in{\cal W}$ determined by the condition $\Gamma f_1(x)=\alpha^{1/2}(x)\ \Gamma h_1(x)$, where $\{h_i\}_{i\in\mathbb{I}_{\ell}}$ are the quasi orthogonal generators for the orthogonal sum decomposition of ${\cal W}$ as in Theorem \ref{teo:la descom de Bo}. Then, by construction $\|\Gamma f_1(x)\|^2=\alpha_1(x)$ and $\lambda_{1,1}(x)=\|\Gamma f_1(x)\|^2=\lambda_1([S_{E(f_1)}]_x)$ for a.e. $x\in \mathbb{T}^k$.
\noi Assume that for $j\in\mathbb{I}_{n-1}$ we have constructed $\mathcal F_j=\{f_i\}_{i\in\mathbb{I}_{j}}\in{\cal W}^j$ such that
\begin{equation}\label{eq induc} \lambda_{i,\ell}(x)=\lambda_i([S_{E(\mathcal F_\ell)}]_x) \peso{for} i\in\mathbb{I}_{\ell} \ , \ \ \ell\in\mathbb{I}_{j}\ \ \text{and a.e. } x\in\mathbb{T}^k\,. \end{equation} We now construct $f_{j+1}$ as follows: set $\mu_i=\lambda_{i,j+1}$ for $i\in\mathbb{I}_{j+1}$ and $\mu_i=0$ for $i>j+1$; set $\mu=(\mu_i)_{i\in\mathbb{N}}:\mathbb{T}^k\rightarrow \ell^1(\mathbb{N})^+$ which is a measurable function. Further, set $S=S_{E(\mathcal F_j)}\in L(L^2(\mathbb{R}}\def\C{\mathbb{C}^k))^+$ which is a SP operator with $R(S)\subset {\cal W}$ and set $m(x)=m=d-1$. Moreover, by taking $\ell=j$ in Eq. \eqref{eq induc} above we see that $\lambda_i([S]_x)=\lambda_{i,j}(x)$ for $i\in\mathbb{I}_{j}$ and $\lambda_i([S]_x)=0$ for $i\geq j+1$, for a.e. $x\in\mathbb{T}^k$ .
\noi By hypothesis, we see that $\lambda_{i,j+1}\leq \lambda_{i,j+2}\leq \ldots\leq \lambda_{i,n}=\lambda_i$; since the admissibility conditions in Remark \ref{rem hay eigensteps} hold, we conclude that $\mu_i=\lambda_{i,j+1}=0$ whenever $i\geq d+1$. On the other hand, since $d-m=1$ we see that the conditions in item 2. in Theorem \ref{estructdelU} can be put together as the interlacing relations $$ \mu_i(x)\geq \lambda_i([S]_x)\geq\mu_{i+1}(x) \peso{for} i\in\mathbb{I}_{j} \ \text{ and a.e. }\ x\in\mathbb{T}^k\,, $$ which hold by hypothesis (see condition 1. in Definition \ref{defi measiegen}); therefore, by Definition \ref{el conjunto U} and Theorem \ref{estructdelU}, there exists a SP operator $B\in L(L^2(\mathbb{R}}\def\C{\mathbb{C}^k))^+$ such that $R(B)\subset {\cal W}$, $\text{\rm rk}([B] _x )\leq 1$ for a.e. $x\in\mathbb{T}^k$ and such that $\lambda_i([S+B]_x)=\mu_i(x)=\lambda_{i,j+1}(x)$ for $i\in\mathbb{I}_{j+1}\,$, for a.e. $x\in\mathbb{T}^k$. The previous conditions on $B$ imply that there exists $f_{j+1}\in{\cal W}$ such that $B=S_{E(f_{j+1})}$; indeed, $f_{j+1}$ is such that it satisfies: $\Gamma f_{j+1}(x)\otimes \Gamma f_{j+1}(x)=[B] _x$ for a.e. $x\in\mathbb{T}^k$.
Finally, if we set $\mathcal F_{j+1}=\{f_i\}_{i\in\mathbb{I}_{j+1}}$ then $S_{E(\mathcal F_{j+1})}=S_{E(\mathcal F_{j})}+S_{E(f_{j+1})}=S+B$ and hence $\lambda_{i,j+1}(x)= \lambda_i([S_{E(\mathcal F_{j+1})}]_x)$ for $i\in\mathbb{I}_{j+1}$ and a.e. $x\in\mathbb{T}^k$. This completes the inductive step. \end{proof}
\noi We end this section with the following remark. With the notations and terminology in Theorem \ref{teo todo eigen echachi}, notice that the constructed sequence $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}$ is such that its fine structure is prescribed by $(\lambda,\,\alpha)$: indeed, $\lambda_i([S_{E(\mathcal F)}]_x)=\lambda_{i,\,n}(x)=\lambda_i(x)$ and
$\|\Gamma f_i(x)\|^2=\alpha_i(x)$ for $i\in\mathbb {I} _n$ and a.e. $x\in\mathbb{T}^k$ (this last fact can be checked using induction and item 2. in Definition \ref{defi measiegen}). That is, the measurable eigensteps provide a detailed description of Bessel sequences $E(\mathcal F)$ with prescribed fine structure.
\section{An application: optimal frames with prescribed norms for FSI subspaces}\label{sec opti frames with prescribed norms}
In order to describe the main problem of this section we consider the following: \begin{fed}\label{defi bal} \rm Let ${\cal W}$ be a FSI subspace of $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ and let $\alpha=(\alpha_i)_{i\in\mathbb {I} _n}\in (\mathbb{R}}\def\C{\mathbb{C}_{>0}^n)^\downarrow$. We let \begin{equation}
\mathfrak {B}_\alpha({\cal W})=\{\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n:\ E(\mathcal F) \ \text{is a Bessel sequence }, \ \|f_i\|^2=\alpha_i\,,\ i\in\mathbb {I} _n\}\ , \end{equation} the set of SG Bessel sequences in ${\cal W}$ with norms prescribed by $\alpha$.
$\triangle$ \end{fed}
\noi
Notice that the restrictions on the families $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in \mathfrak {B}_\alpha({\cal W})$
(namely $\|f_i\|^2=\alpha_i$ for $i\in\mathbb {I} _n$) are of a {\it global} nature. Our problem is to describe those $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$ such that the encoding schemes associated to their corresponding Bessel sequences $E(\mathcal F)$ are as stable as possible. Ideally, we would search for sequences $\mathcal F$ such that $E(\mathcal F)$ are tight frames for ${\cal W}$; yet, Theorem \ref{teo sobre disenio de marcos} shows that there are obstructions for the existence of such sequences (see Corollary \ref{coro tight 2} below).
\noi By a simple re-scaling argument, we can assume that $\sum_{i\in\mathbb {I} _n}\alpha_i=1$; then Corollary \ref{cororo1} (see also \cite[Theorem 3.9.]{BMS15}) shows that if there exists $\mathcal F_0\in \mathfrak {B}_\alpha({\cal W})$ such that $E(\mathcal F_0)$ is a tight frame for ${\cal W}$ then $E(\mathcal F_0)$ is a minimizer in $\mathfrak {B}_\alpha({\cal W})$ of every frame potential $P_\varphi^{\cal W}$ for any convex function $\varphi\in\convf$ and $P_\varphi^{\cal W} (E(\mathcal F_0))= C_{\cal W} \ \varphi(C_{\cal W}^{-1})$; moreover, in case $\varphi\in\convfs$ is a strictly convex function, then every such $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$ for which $P_\varphi^{\cal W}(E(\mathcal F))=C_{\cal W} \ \varphi(C_{\cal W}^{-1})$ is a tight frame. This suggests that in the general case, in order to search for $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$ such that the encoding schemes associated to their corresponding Bessel sequences $E(\mathcal F)$ are as stable as possible, we could study the minimizers in $\mathfrak {B}_\alpha({\cal W})$ of the convex potential $P_\varphi^{\cal W}$ associated to a strictly convex function $\varphi\in\convfs$.
\noi Therefore, given $\varphi\in\convf$, in what follows we show the existence of finite sequences $\mathcal F^{\rm op}\in \mathfrak {B}_\alpha({\cal W})$ such that $$ P_\varphi^{\cal W}(E(\mathcal F^{\rm op}))=\min\{P_\varphi^{\cal W}(E(\mathcal F)): \ \mathcal F\in \mathfrak {B}_\alpha({\cal W})\}\,. $$ Moreover, in case $\varphi\in\convfs$ then we describe the fine spectral structure of the frame operator of $E(\mathcal F^{\rm op})$. In case $\varphi(x)=x^2$, our results extend some results from \cite{BF,CKFT,MR10} for the frame potential to the context of SG Bessel sequences lying in a FSI subspace ${\cal W}$.
\noi Let us fix some general notions and notation for future reference: \begin{notas}\label{nota impor2}
In what follows we consider: \begin{enumerate} \item A FSI subspace ${\cal W}\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$; \item $d(x)=\dim J_{\cal W}(x)\leq \ell\in\mathbb{N}$, for a.e. $x \in \mathbb{T}^k$;
\item The Lebesgue measure on $\mathbb{R}}\def\C{\mathbb{C}^k$, denoted $|\cdot|$ ; $Z_i=d^{-1}(i)\subseteq \mathbb{T}^k$ and $p_i=|Z_i|$, $i\in\mathbb{I}_{\ell}\,$.
\item The spectrum of ${\cal W}$ is the measurable set $\text{Spec}({\cal W}) = \bigcup_{i\in\mathbb{I}_{\ell}} Z_i = \{x\in \mathbb{T}^k: d(x)\neq 0\}$. \end{enumerate}
$\triangle$ \end{notas}
\subsection{The uniform dimension case}\label{subse uniform}
Consider the Notations \ref{nota impor2}. In this section we obtain the fine spectral structure of minimizers of convex potentials in $\mathfrak B_\alpha({\cal W})$ under the assumption that $d(x)=d$ for a.e. $x\in\text{Spec}({\cal W})$. In order to deal with this particular case, we recall some notions and constructions from \cite{BMS15}.
\begin{rem}[Waterfilling in measure spaces]\label{recordando waterfilling} Let $(X,\mathcal{X},\mu)$ denote a probability space and let $L^\infty(X,\mu)^+ = \{g\in L^\infty(X,\mu): g\ge 0\}$. Recall that for $f\in L^\infty(X,\mu)^+ $ and $c \geq \infes f\geq 0$ we consider the {\it waterfilling} of $f$ at level $c$, denoted $f_c\in L^\infty(X,\mu)^+$, given by $f_c= \max\{f,\,c\}=f + (c-f)^+$, where $g^+$ denotes the positive part of a real function $g$. Recall the decreasing rearrangement of non-negative functions defined in Eq. \eqref{eq:reord}. It is straightforward to check that if \begin{equation}\label{eq lem reord} s_0= \mu\{x\in X:\ f(x)>c\} \peso{then} f_c^*(s)=\left\{
\begin{array}{ccc}
f^*(s) & if & 0\leq s <s_0 \,; \\
c & if & s_0\leq s \leq 1\, .
\end{array}
\right.
\end{equation} We further consider $\phi_f: [\infes f, \infty)\to \mathbb{R}}\def\C{\mathbb{C}_+$ given by $$ \phi_f(c)=\int_X f_c\ d\mu= \int_X f(x) + (c-f(x))^+\ d\mu(x)\,. $$ Then, it is easy to see that: \begin{enumerate} \item $\phi_f(\infes f)=\int_X f\ d\mu$ and $\lim _{c\to +\infty} \phi_f(c)= +\infty$; \item $\phi_f$ is continuous and strictly increasing. \end{enumerate} Hence, for every $v\geq \int_X f\ d\mu$ there exists a unique $c=c(v)\geq \infes f$ such that $\phi_f(c)=v$. With the previous notations then, by \cite[Theorem 5.5.]{BMS15} we get that if $h\in L^\infty(X,\mu)^+ $ is such that \begin{equation}\label{eq teo 5.5} f\leq h \peso{and} v\leq \int_X h\ d\mu \peso{then} f_{c(v)}\prec_w h\ .\end{equation}
$\triangle$ \end{rem}
\begin{lem}\label{lem: wat er filling} Let $(X,\mathcal{X},\mu)$ denote a probability space and let $f,\,g\in L^\infty(X,\mu)^+$ be such that $f\prec_w g$. Let $c,\,d\geq 0$ be such that $\int_X f_c\ d\mu=\int_X g_d\ d\mu$, where $f_c$ and $g_d$ denote the waterfillings of $f$ and $g$ at levels $c$ and $d$ respectively. Then $f_c\prec g_d$ in $(X,\mu)$. \end{lem} \begin{proof} Set $s_0=\mu\{x\in X:\ f(x)>c\}\in[0,1]$; notice that by construction $g\leq g_d$ in $X$ so that, by Remark \ref{rem:prop rear elem}, $g^*\leq (g_d)^*$ in $[0,1]$. Hence, for every $s\in[0,s_0]$ we have \begin{equation}\label{eq desi11} \int_0^s(f_c)^*\ dt=\int_0^s f^*\ dt\leq \int_0^s g^*\ dt\leq \int_0^s (g_d)^*\ dt\,. \end{equation} On the other hand, $$ \int_0^{s_0} (g_d)^* \ dt\geq \int_0^{s_0} g^* \ dt\geq \int_0^{s_0} f^* \ dt \implies \omega \ \stackrel{\mbox{\tiny{def}}}{=}\ \int_0^{s_0} (g_d)^* \ dt-\int_0^{s_0} f^* \ dt\geq 0\, .$$ Using Remark \ref{recordando waterfilling} and the hypothesis we get that $$ \int_0^{s_0} f^* \ dt + (1-s_0) \, c = \int_0^1 f_c^* \ dt \stackrel{\eqref{reor int}}{=}\int_X f_c\ d\mu=\int_X g_d\ d\mu \stackrel{\eqref{reor int}}{=} \int_0^{s_0} (g_d)^* \ dt + \int_{s_0}^1 (g_d)^* \ dt $$ $$ \implies \quad \quad (1-s_0) \, c=\int_{s_0}^1 \Big[ \ (g_d)^* + \frac{\omega}{1-s_0} \ \Big]\ dt\,.$$ Thus, by \cite[Lemma 5.3.]{BMS15} we get that for $s\in[s_0\, , \, 1]$: $$ (s-s_0)\,c\leq \int_{s_0}^s \Big[ \ (g_d)^* + \frac{\omega}{1-s_0} \ \Big]\ dt
\leq \int_{s_0}^s (g_d)^* \ dt+ \omega\ . $$ This last identity and Remark \ref{recordando waterfilling} show that for $s\in[s_0\, , \, 1]$, \begin{equation} \label{eq desi22} \int_0^s (f_c)^*\ dt= \int_0^{s_0} (g_d)^*\ dt-\omega + (s-s_0)\, c\leq \int_0^{s_0} (g_d)^*\ dt + \int_{s_0}^s (g_d)^*\ dt\,. \end{equation} The lemma is a consequence of Eqs. \eqref{eq desi11} and \eqref{eq desi22}. \end{proof}
\begin{rem}\label{rem: sobre medidas}
Let $(Z\, , \, \mathcal Z\, , \, |\, \cdot \,|)$ be a (non-zero) measure subspace of $(\mathbb{T}^k\, , \, \mathcal B\, , \, |\, \cdot \,|)$ and consider $(\mathbb{I}_r, \mathcal P(\mathbb{I}_r),\#(\cdot))$ i.e, $\mathbb{I}_r$ endowed with the counting measure. In what follows we consider the product
space $X\ \stackrel{\mbox{\tiny{def}}}{=}\ Z\times \mathbb{I}_r$ endowed with the product measure $\mu\ \stackrel{\mbox{\tiny{def}}}{=}\ |\cdot |\times \#(\cdot)$.
$\triangle$ \end{rem}
\begin{lem}\label{lem utilisima} Consider the notations in Remark \ref{rem: sobre medidas} and let $\alpha: Z\rightarrow \mathbb{R}}\def\C{\mathbb{C}^r$ be a measurable function. Let $\breve\alpha:X\rightarrow \mathbb{R}}\def\C{\mathbb{C}$ be given by $$ \breve \alpha(x,i)=\alpha_i(x) \peso{for} x\in Z \peso{and} i\in\mathbb{I}_r\ . $$ Then $\breve \alpha$ is a measurable function and we have that: \begin{enumerate} \item If $\varphi\in\convf$ then
$\int_X \varphi\circ \breve \alpha \ d\mu = \sum\limits_{i\in\mathbb{I}_r} \int_{Z} \varphi(\alpha_i(x))\ dx \ .$ \item Let $\beta: Z\rightarrow \mathbb{R}}\def\C{\mathbb{C}^r$ be a measurable function and let $\breve \beta:X\rightarrow \mathbb{R}}\def\C{\mathbb{C}$ be constructed analogously. If $$ \alpha(x)\prec\beta(x) \peso{for a.e.} x\in Z \ \implies \ \ \breve \alpha\prec \breve\beta $$
in the probability space $(X,\mathcal X,\tilde \mu)$, where $\tilde \mu=(r\cdot|Z|)^{-1}\,\mu$. \item Similarly, $\alpha(x)\prec_w\beta(x)$ for a.e. $x\in Z$ implies that $\breve \alpha\prec_w \breve \beta$ in $(X,\mathcal X,\tilde \mu)$. \end{enumerate} \end{lem} \begin{proof} The proof of the first part of the statement is straightforward. In order to see item 2., notice that if $\varphi\in\convf$ then $\alpha(x)\prec\beta(x)$ implies that $\sum_{i\in\mathbb{I}_r}\varphi(\alpha_i(x))\leq \sum_{i\in\mathbb{I}_r}\varphi(\beta_i(x))$ for a.e. $x\in Z$. Then, using item 1. we get that $$
\int_X \varphi\circ \breve \alpha\ d\tilde \mu=(r\cdot |Z|)^{-1} \int_{Z} \sum_{i\in\mathbb{I}_r}\varphi(\alpha_i(x)) \ dx
\leq (r\cdot |Z|)^{-1} \int_{Z} \sum_{i\in\mathbb{I}_r}\varphi(\beta_i(x))\ dx = \int_X \varphi\circ \breve \beta\ d\tilde \mu\ . $$ Since $\varphi\in\convf$ is arbitrary, Theorem \ref{teo porque mayo} shows that $\breve \alpha\prec \breve \beta$. Item 3. follows using similar arguments, based on the characterization of submajorization in terms of integral inequalities involving non-decreasing convex functions given in Theorem \ref{teo porque mayo} (see also \cite{Chong}). \end{proof}
\noi The following is the first main result of this section.
\begin{teo}[Existence of optimal sequences in $\mathfrak {B}_\alpha({\cal W})$]\label{teo dim unif}
Consider the Notations \ref{nota impor2}. Let $\alpha=(\alpha_i)_{i\in\mathbb {I} _n}\in (\mathbb{R}}\def\C{\mathbb{C}_{>0}^n)^\downarrow$ and assume that ${\cal W}$ is such that $d(x)=d$ for a.e. $x\in \text{Spec}({\cal W})$; set $r=\min\{n,d\}$. Let $p=p_d=|\text{Spec}({\cal W})|$. Then there exist $c=c(\alpha,\,d,\,p)\geq 0$ and $\mathcal F^{\rm op}\in \mathfrak {B}_\alpha({\cal W})$ such that: \begin{enumerate} \item For a.e. $x\in \text{Spec}({\cal W})$ we have that \begin{equation}\label{eq defi la op unif dim} \lambda_j([S_{E(\mathcal F^{\rm op})}]_x) =\left\{
\begin{array}{ccc}
\max\{\frac{\alpha_j}{p}\, , \, c\} & if & j\in\mathbb{I}_{r} \ ; \\
0 & if & r+1 \le j\le d\ .
\end{array}
\right.
\end{equation}
In particular, if $d\leq n$ (i.e. $r=d$) then $E(\mathcal F^{\rm op})$ is a frame for ${\cal W}$. \item For every $\varphi\in\convf$ and every $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$ then \begin{equation}\label{eq prop c}
p\cdot \sum_{j\in\mathbb{I}_{r}} \varphi (\max\{\frac{\alpha_j}{p}\, , \, c\}) + p \, (d-r)\, \varphi(0) =P_\varphi^{\cal W}(E(\mathcal F^{\rm op}))\leq P_\varphi^{\cal W}(E(\mathcal F)) \,. \end{equation} \end{enumerate} \end{teo} \begin{proof} Consider $\text{Spec}({\cal W})$ as a (non-zero, otherwise the result is trivial) measure subspace of the $k$-torus endowed with Lebesgue measure. Then, we consider
$X=\text{Spec}({\cal W})\times \mathbb{I}_r$ endowed with the product measure $\mu=|\cdot|\times \#(\cdot)$, where $\#(\cdot)$ denotes the counting measure on $\mathbb{I}_r$ (as in Remark \ref{rem: sobre medidas}). We also consider the normalized measure $\tilde \mu=\frac{1}{p\cdot r}\ \mu$ on $X$.
Let $\mathcal F=\{f_j\}_{j\in\mathbb {I} _n}\in \mathfrak {B}_\alpha({\cal W})$ and set $\beta_j(x)=\|\Gamma f_j(x)\|^2$ for $x\in \text{Spec}({\cal W})$ and $j\in\mathbb {I} _n\, $. Notice that \begin{equation}\label{ecua betaj}
\int_{\text{Spec}({\cal W})} \beta_j(x)\ dx=\|f_j\|^2=\alpha_j\ , \peso{for} j\in\mathbb {I} _n\ . \end{equation} Let $\breve\gamma\, , \, \breve\beta:X\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ be the measurable functions determined by $$ \breve \gamma (x,j)=\frac{\alpha_j}{p} \peso{and} \breve \beta(x,j)=\beta_j(x) \peso{for} x\in \text{Spec}({\cal W}) \peso{and} j\in\mathbb{I}_{r} \ . $$ Consider the map $D:L^\infty(X,\tilde \mu) \rightarrow L^\infty(X,\tilde \mu)$ given by $$ D(h)(x,j)=r\cdot\int_{\text{Spec}({\cal W})\times \{j\}} h \ d\tilde \mu = \frac 1p \ \int_{\text{Spec}({\cal W})} h(x,j) \ dx \peso{for} x\in \text{Spec}({\cal W}) \peso{and} j \in \mathbb{I}_{r}\ . $$ Then, it is easy to see that $D$ is positive, unital and trace preserving i.e. $D$ is a doubly stochastic map; moreover, by Eq. \eqref{ecua betaj}, $D(\breve \beta)=\breve \gamma$ and by Theorem \ref{teo porque mayo} we conclude that $\breve \gamma\prec \breve \beta\,$.
\noi Now, consider the measurable vector-valued function $\beta^\downarrow(x)=(\beta^\downarrow_j(x))_{j\in\mathbb {I} _n}$ obtained by re-arrangement of the entries of the vector $\beta(x)=(\beta_j(x))_{j\in\mathbb {I} _n}$, for $x\in Z$ independently. By construction we get the submajorization relations $(\beta_j(x))_{j\in\mathbb{I}_{r}}\prec_w (\beta^\downarrow_j(x))_{j\in\mathbb{I}_{r}}$ for every $x\in Z$ (notice that we are considering just the first $r$ entries of these $n$-tuples).
\noi Thus, if we consider the measurable function $\breve {\beta^\downarrow} :X\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ determined by $\breve{\beta^\downarrow}(x,j)=\beta^\downarrow_j(x)$ if $x\in \text{Spec}({\cal W})$ and $j\in\mathbb{I}_{r}\,$, then Lemma \ref{lem utilisima} shows that $\breve\beta \prec_w \breve{\beta ^\downarrow}$ in $(X,\tilde \mu)$. By transitivity, we conclude that $\breve \gamma\prec_w \breve{\beta^\downarrow}$. By Remark \ref{recordando waterfilling} there exists a unique $b\geq \text{ess-}\inf\limits_{x\in X} \breve{\beta^\downarrow} (x)$ such that the waterfilling of $\breve{\beta^\downarrow}$ at level $b$, denoted $\breve{\beta^\downarrow}_b$, satisfies $$ \int_X \breve{\beta^\downarrow} _b \ d\tilde \mu=(r\cdot p)^{-1}\, \sum_{i\in\mathbb {I} _n} \alpha_i \geq \int_X \breve{\beta^\downarrow} \ d\tilde \mu \ . $$ Similarly, let $c\geq \text{ess-}\inf\limits_{x\in X} \,\breve \gamma(x)$ be such that the waterfilling of $\breve \gamma$ at level $c$, denoted $\breve \gamma_c\,$, satisfies $$ \int_X \breve\gamma_c \ d\tilde \mu=(r\cdot p)^{-1}\, \sum_{i\in\mathbb {I} _n} \alpha_i\geq \int_X \breve\gamma \ d\tilde \mu\ . $$ Therefore, by Lemma \ref{lem: wat er filling}, we see that \begin{equation}\label{eq relac fc fbetaparaabajo} \breve\gamma_c\prec \breve{\beta^\downarrow} _b\peso{in} (X,\tilde \mu)\ . \end{equation} By Lemma \ref{lem spect represent ese} there exist measurable functions $\lambda_j:\mathbb{T}^k\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ for $j\in\mathbb{I}_{d}$ such that we have a representation of $[S_{E(\mathcal F)}]_x=S_{\Gamma \mathcal F(x)}$ as in Eq. \eqref{lem repre espec S}, in terms of some measurable vector fields $v_j:\mathbb{T}^k\rightarrow \ell^2(\mathbb{Z}^k)$ for $j\in\mathbb{I}_d$, such that $\{v_j(x)\}_{j\in\mathbb{I}_d}$ is a ONB of $J_{\cal W}(x)$ for a.e. $x\in \text{Spec}({\cal W})$; indeed, in this case $\lambda_j(x)=0$ for $j\geq r+1$ and a.e. $x\in \text{Spec}({\cal W})$.
\noi
If we let $e(x)\geq 0$ be determined by the condition $$ \sum_{i\in\mathbb{I}_r}\max\{\beta^\downarrow _i(x),e(x)\}=\sum_{i\in\mathbb{I}_{r}}\lambda_i(x)\ \Big(\, =\sum_{i\in\mathbb{I}_{d}}\lambda_i(x)\, \Big) \ , \peso{for a.e.} x\in \text{Spec}({\cal W}) $$ then by \cite{MR10} (also see \cite{MRS13,MRS14b,MRS14}) we have that \begin{equation}\label{eq rel mayo MR} (\delta_i(x))_{i\in\mathbb{I}_{r}}\ \stackrel{\mbox{\tiny{def}}}{=}\ (\max\{ \beta^\downarrow_i(x),\, e(x)\} )_{i\in\mathbb{I}_{r}}\prec (\lambda_i(x))_{i\in\mathbb{I}_{r}} \ , \peso{for a.e.} x\in \text{Spec}({\cal W})\,. \end{equation} Notice that the vector $(\delta_i(x))_{i\in\mathbb{I}_{r}}$ can be considered as the (discrete) waterfilling of the vector $(\beta^\downarrow_j(x))_{j\in\mathbb{I}_{r}}$ at level $e(x)$, for $x\in \text{Spec}({\cal W})$. If $\breve\delta \, , \, \breve\lambda:X\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+$ are the measurable functions given by $$ \breve\delta(x,j)=\delta_j(x) \peso{and} \breve\lambda(x,j)=\lambda_j(x) \peso{for} x\in \text{Spec}({\cal W}) \peso{and} j\in\mathbb{I}_{r} $$ then, by Lemma \ref{lem utilisima}, we get that $\breve\delta\prec \breve\lambda$ in $(X,\tilde\mu)$. Notice that by construction, $\breve\delta\geq \breve{\beta^\downarrow} $ and $$\int_X \breve\delta\ d\tilde \mu =(r\cdot p)^{-1}\,\sum_{i\in\mathbb {I} _n}\alpha_i \,.$$ Hence, by Remark \ref{recordando waterfilling}, we get that $\breve{\beta^\downarrow}_b\prec \breve\delta\,$. Putting all the pieces together, we now see that \begin{equation}\label{eq relac major func} \breve \gamma_c\prec \breve{\beta^\downarrow}_b\prec \breve\delta\prec \breve\lambda\ , \peso{in} (X,\tilde\mu)\ . \end{equation} Recall that by construction, we have that \begin{equation}\label{eq la pinta de fc}
\breve \gamma_c(x)=\max\{ \frac{\alpha_j}{p} \, , \, c\} \ , \peso{for} x\in \text{Spec}({\cal W})\times \{j\}\subset X \, , \ j\in\mathbb{I}_{r}\ . \end{equation} Then, it is straightforward to check that \begin{equation}\label{eq. c es el correcto} (r\cdot p)^{-1}\,\sum_{i\in\mathbb {I} _n}\alpha_i=\int_X \breve\gamma_c\ d\tilde\mu= r^{-1} \cdot \sum_{j\in\mathbb{I}_{r}}\max\{ \frac{\alpha_j}{p} \, , \, c\} \implies (\frac{\alpha_j}{p})_{j\in\mathbb {I} _n}\prec (\max\{\frac{\alpha_j}{p} \, , \, c\})_{j\in\mathbb{I}_{r}}\ . \end{equation} Thus, by Theorem \ref{teo sobre disenio de marcos}, there exists a Bessel sequence $\mathcal F^{\rm op}=\{f^{\rm op}_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n$ such that the fine spectral structure $(\lambda_j([S_{E(\mathcal F^{\rm op})}]_x)\,)_{j\in\mathbb{N}}$ satisfies Eq. \eqref{eq defi la op unif dim} and such that
$\|\Gamma f^{\rm op}_i(x)\|^2=\frac{\alpha_i}{p}\,$, for $i\in\mathbb {I} _n\,$, and $x\in \text{Spec}({\cal W})$. In particular, $\|f^{\rm op}_i\|^2=\alpha_i$ for $i\in\mathbb {I} _n\,$, so $\mathcal F^{\rm op}\in \mathfrak {B}_\alpha({\cal W})$. If $\varphi\in\convf$ then, by the majorization relations in Eq. \eqref{eq relac major func} and Lemma \ref{lem utilisima}, \begin{eqnarray*}\label{eq desi poten} P_\varphi^{\cal W}(E(\mathcal F^{\rm op}))&=&\int_{\text{Spec}({\cal W})} [ \sum_{j\in\mathbb{I}_{r}}\varphi(\max\{\frac{\alpha_j}{p} \, , \, c\}) + (d-r)\, \varphi(0) ]\ dx= \int_X \varphi\circ \breve \gamma_c\ d\mu + p\,(d-r)\, \varphi(0) \\ &\leq & \int_X \varphi\circ \breve\lambda\ d\mu + p\,(d-r)\, \varphi(0)
=P_\varphi^{\cal W}(E(\mathcal F))\,. \end{eqnarray*} Hence, $\mathcal F^{\rm op}$ satisfies items 1. and 2. in the statement. \end{proof}
\noi The previous result shows that there are indeed structural optimal frames with prescribed norms in the sense that these frames minimize any frame potential within $\mathfrak {B}_\alpha({\cal W})$; along its proof we showed several majorization relations that allow us to prove that the spectral structure of any such structural optimal frame is described by Eq. \eqref{eq defi la op unif dim}.
\begin{teo}[Fine spectral structure of optimal sequences in $\mathfrak {B}_\alpha({\cal W})$] \label{teo struct fina dim hom} With the hypothesis and notations from Theorem \ref{teo dim unif}, assume that $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$ is such that there exists $\varphi\in\convfs$ with $P_\varphi^{\cal W}(E(\mathcal F))=P_\varphi^{\cal W}(E(\mathcal F^{\rm op}))$. Then, for a.e. $x\in \text{Spec}({\cal W})$ we have that \begin{equation}\label{eq defi la op unif dim2} \lambda_j([S_{E(\mathcal F)}]_x) =\left\{
\begin{array}{ccc}
\max\{\frac{\alpha_j}{p} \, , \, c\} = \max\{ \beta^\downarrow_j(x)\, ,\, c\}& if & j\in\mathbb{I}_{r} \ ; \\
0 & if & r+1 \le j\le d\ ,
\end{array}
\right.
\end{equation} where $\beta^\downarrow_1(x)\geq \ldots\beta^\downarrow_n(x)\geq 0$ are obtained by re-arranging the sequence $$ \beta(x) = \big( \, \beta_1(x) \, , \, \ldots \, , \, \beta_n(x)\,\big)
=\big(\, \|\Gamma f_1(x)\|^2 \, , \, \ldots \, , \, \|\Gamma f_n(x)\|^2 \,\big) \in \mathbb{R}}\def\C{\mathbb{C}^n $$ in non-increasing order, independently for each $x\in\text{Spec}({\cal W})$. \end{teo} \begin{proof} We continue to use the notations and terminology from the proof of Theorem \ref{teo dim unif}. Assume further that $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$ is such that there exists $\varphi\in\convfs$ with $$ p\cdot \sum_{j\in\mathbb{I}_{r}} \varphi (\max\{\frac{\alpha_j}{p}\, , \, c\}) + p\,(d-r)\,\varphi(0) = P_\varphi^{\cal W}(E(\mathcal F))\ . $$ Then, using this last fact and Lemma \ref{lem utilisima} we see that $$ (r\cdot p)\,\int_X \varphi\circ \breve \gamma_c\ d\tilde\mu= (r\cdot p)\,\int_X \varphi\circ \breve\lambda\ d\tilde\mu\ . $$ Hence, by Eq. \eqref{eq relac major func} we have that $$ \int_X \varphi\circ \breve \gamma_c\ d\tilde\mu= \int_X \varphi\circ \breve{\beta^\downarrow }_b\ d\tilde\mu = \int_X \varphi\circ \breve\delta\ d\tilde\mu= \int_X \varphi\circ \breve\lambda\ d\tilde\mu\ . $$ Thus, by Proposition \ref{pro int y reo} the functions $\breve \gamma_c,\,\breve{\beta^\downarrow}_b,\,\breve\delta,\,\breve\lambda$ are equimeasurable. On the one hand, Eq. \eqref{eq rel mayo MR}, together with the equality above imply that $\max\{ \beta^\downarrow_j(x)\, , \, e(x)\} =\lambda_j(x)$, for $j\in\mathbb{I}_{r}$ and a.e. $x\in \text{Spec}({\cal W})$ and hence, by construction, $\breve\delta=\breve\lambda\,$. On the other hand, by \cite[Corollary 5.6]{BMS15} we also get that $\breve{\beta^\downarrow} _b=\breve\delta\,$. Therefore, $\breve{\beta^\downarrow} _b=\breve\delta=\breve\lambda\,$; in particular, we get that $\max\{ \beta^\downarrow_j(x)\, , \, b\} =\lambda_j(x)$, for $j\in\mathbb{I}_{r}$ and a.e. $x\in \text{Spec}({\cal W})$.
\noi Notice that, since $\breve \gamma_c$ and $\breve\lambda$ are equi-measurable, then
$|\breve\lambda^{-1}(\max\{\frac{\alpha_j}{p}\, , \, c\})|=|{\breve \gamma_c} ^{-1}(\max\{\frac{\alpha_j}{p} \, , \, c\})|$ for $j\in\mathbb{I}_{r}\,$; thus, $\breve\lambda$ takes the values $\max\{\frac{\alpha_j}{p},\,c\}$ for $j\in\mathbb{I}_{r}$ (off a zero-measure set). As $\breve\lambda$ and $\breve \gamma_c$ are both induced by the vector-valued functions $$ \text{Spec}({\cal W})\ni x\mapsto (\max\{\frac{\alpha_j}{p} \, , \, c\})_{j\in\mathbb{I}_{r}}\in(\mathbb{R}}\def\C{\mathbb{C}_+^r)^\downarrow \peso{and} \text{Spec}({\cal W})\ni x\mapsto (\lambda_j(x))_{j\in\mathbb{I}_{r}}\in(\mathbb{R}}\def\C{\mathbb{C}_+^r)^\downarrow $$ respectively, we conclude that $$ (\max\{\frac{\alpha_j}{p}\, , \, c\})_{j\in\mathbb{I}_{r}}=(\lambda_j(x))_{j\in\mathbb{I}_{r}} =(\max\{ \beta^\downarrow_j(x)\, , \, b\} )_{j\in\mathbb{I}_{r}} \ , \peso{for} x\in \text{Spec}({\cal W})\ . $$ From this last fact, we see that we can set $b=c$ and the result follows. \end{proof}
\begin{rem}\label{interpret de prop dim unif} Consider the notations and terminology from Theorem \ref{teo dim unif}. We point that there is a simple formula for the constant $c$. Indeed, notice that if $\mathcal F^{\rm op}\in \mathfrak {B}_\alpha({\cal W})$ is the structural solution of the optimization problem considered in Theorem \ref{teo dim unif} then $$ \sum_{j\in\mathbb{I}_r}\lambda_j([S_{E(\mathcal F^{\rm op})}]_ x)
=\tr([S_{E(\mathcal F^{\rm op})}]_ x)=\sum_{j\in\mathbb {I} _n}\|\Gamma f^{\rm op}_j(x)\|^2\peso{for a.e.} x\in\mathbb{T}^k $$ Therefore, \begin{equation}\label{formu c}
\sum_{i\in\mathbb{I}_r}\max\{\frac{\alpha_i}{p} \, , \, c\}=\frac 1p \ \sum_{j\in\mathbb {I} _n}\alpha_j\ , \end{equation} which shows that $c$ is obtained by the previous discrete waterfilling condition.
$\triangle$ \end{rem}
\noi Tight frames play a central role in applications. On the one hand, they give raise to simple reconstruction formulas; on the other hand, they have several robustness properties related with numerical stability of the encoding-decoding scheme that they induce. It is therefore important to have conditions that assure the existence of tight frames with prescribed norms: in the finite dimensional context (i.e. finite frame theory) this problem is solved in \cite{CKFT} in terms of the so-called fundamental inequality. As a consequence of Remark \ref{interpret de prop dim unif}, we obtain conditions for the existence of tight SG frames with norms given by a finite sequence of positive numbers, in the uniform dimensional case.
\begin{cor} \label{coro tight 2} Consider the notations and hypothesis of Theorem \ref {teo dim unif}. In the uniform dimensional case (so in particular, $d(x)=d$ for a.e. $x\in \text{Spec}({\cal W})\,$), we have that $$ \text{\rm there exist {\bf tight} frames in $\mathfrak {B}_\alpha({\cal W})$ } \ \iff \ \ d=r\le n \peso{ \rm and} d \cdot\alpha_1 \le \sum_{j\in\mathbb {I} _n}\alpha_j\ . $$
\end{cor} \proof It is a direct consequence of Eqs. \eqref{eq defi la op unif dim} and \eqref{formu c}. \qed \subsection{Existence and structure of $P^{\cal W}_\varphi$-minimizers in $\mathfrak {B}_\alpha({\cal W})$: the general case}\label{subsec gral mi gral}
It turns out that Theorem \ref{teo dim unif} allows to reduce the study of the spectral structure of minimizers of convex potentials in FSI subspaces with norm restrictions to a finite dimensional model. Indeed, consider the Notations \ref{nota impor2} and, for the sake of simplicity, assume that $p_i>0$ for every $i\in\mathbb{I}_{\ell}\,$. Consider $\alpha\in (\mathbb{R}}\def\C{\mathbb{C}_{>0}^n)^\downarrow$ and let $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$. For each $i\in\mathbb{I}_\ell$ let ${\cal W}_i\subset L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ be the closed FSI subspace whose fibers coincide with those of ${\cal W}$ in $Z_i=d^{-1}(i)$ and are the zero subspace elsewhere, and let $\mathcal F_i=\{f_{i,j}\}_{j\in\mathbb {I} _n}\in{\cal W}_i^n$ be determined by $$ \Gamma f_{i,j}(x)=\chi_{Z_i}(x)\ \Gamma f_j(x) \peso{for a.e.} x\in\mathbb{T}^k \peso{and} j\in\mathbb {I} _n\ , $$ where $\chi_Z$ denotes the characteristic function of a measurable set $Z\subset\mathbb{T}^k$. Fix a convex function $\varphi\in\convf$. Since each ${\cal W}_i$ is also a uniform FSI, it satisfies the hypothesis of Theorem \ref{teo dim unif}. Then we conclude that for each $i\in\mathbb{I}_\ell$ there exists $\mathcal F_i^{\rm dis}=\{f_{i,j}^{\rm dis}\}_{j\in\mathbb {I} _n}\in{\cal W}_i ^n$ such that $$
\|f_{i,j}^{\rm dis}\|^2=\|f_{i,j}\|^2 \peso{for} j\in\mathbb{I}_n \peso{and} P^{{\cal W}_i}_\varphi (E(\mathcal F_i^{\rm dis}))\leq P^{{\cal W}_i}_\varphi (E(\mathcal F_i)) \peso{for} i\in\mathbb{I}_\ell\ . $$ We can recover the initial family $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}$ by gluing together the families $\mathcal F_i$ for $i\in\mathbb{I}_\ell\,$. Similarly, if we glue the families $\mathcal F_i^{\rm dis}$ we get a family $\mathcal F^{\rm dis}$ (in such a way that $(\mathcal F^{\rm dis})_i=\mathcal F_i^{\rm dis}\in {\cal W}_i^n$ as before, for $i\in\mathbb{I}_\ell$). Notice that $\mathcal F^{\rm dis}\in \mathfrak {B}_\alpha({\cal W})$ since $$
\|f_i^{\rm dis}\|^2=\sum_{j\in\mathbb{I}_n}\|f_{i,j}^{\rm dis}\|^2= \|f_i\|^2=\alpha_i \peso{for} i\in\mathbb{I}_n \ , $$ using the fact that the subspaces $\{{\cal W}_i\}_{i\in\mathbb{I}_\ell}$ are mutually orthogonal. Also $$ P^{{\cal W}}_\varphi (E(\mathcal F^{\rm dis}))= \sum_{i\in\mathbb{I}_\ell} P^{{\cal W}_i}_\varphi (E(\mathcal F_i^{\rm dis}))\leq \sum_{i\in\mathbb{I}_\ell} P^{{\cal W}_i}_\varphi (E(\mathcal F_i))= P^{{\cal W}}_\varphi (E(\mathcal F))\ . $$ Now, the fine spectral structure of $\mathcal F_i^{\rm dis}$ is of a discrete nature (as described in Theorem \ref{teo dim unif}). Moreover, this fine structure is explicitly determined in terms of the matrix \begin{equation}\label{las B}
B=(p_i^{-1}\, \|f_{i,j}\|^2)_{i\in \mathbb{I}_\ell,\,j\in\mathbb {I} _n} \in \mathbb{R}}\def\C{\mathbb{C}_{+}^{\ell\times n} \peso{fulfilling the identity} p^T\,B=\alpha \ , \end{equation} where $p=(p_i)_{i\in\mathbb{I}_\ell}$ and $\alpha=(\alpha_i)_{i\in\mathbb {I} _n}\,$. Notice that the set of all such matrices form a convex compact subset of $\mathbb{R}}\def\C{\mathbb{C}_{+}^{m\times n}$. The advantage of this approach is that we can use simple tools such as convexity, compactness and continuity in a finite dimensional context, to show existence of optimal spectral structure within our reduced model. Nevertheless, the reduced model has a rather combinatorial nature (see the definition of $\Lambda_{\alpha,\,p}^{\rm op}(\delta)$ below), so we build it in steps.
\begin{notas}\label{muchas nots} In order to simplify the exposition of the next result, we introduce the following notations that are motivated by the remarks above. Let $ m \, , \, n\in\mathbb{N}$: \begin{enumerate} \item Inspired in Eq. \eqref{las B}, for finite sequences $\alpha\in (\mathbb{R}}\def\C{\mathbb{C}_{>0}^n)^\downarrow$ and $p=(p_i)_{i\in\mathbb{I}_{m}}\in \mathbb{R}}\def\C{\mathbb{C}_{>0}^m$ we consider the set of weighted partitions $$ \begin{array}{rl} W_{\alpha,\,p} &=\{ B\in \mathbb{R}}\def\C{\mathbb{C}_+^{m\times n } \ : \ p^T\, B = \alpha \, \} \ . \end{array} $$ It is straightforward to check that $W_{\alpha,\,p}$ is a convex compact set. \item \label{item2} Given $d\in \mathbb{N}$ we define the map $L_d : \mathbb{R}}\def\C{\mathbb{C}_+^n \to (\mathbb{R}}\def\C{\mathbb{C}_+^d)^\downarrow$ given by \begin{equation}\label{eq defi gammacd} L_d (\gamma ) \ \stackrel{\mbox{\tiny{def}}}{=}\ \left\{
\begin{array}{ccc}
(\max\{\gamma^\downarrow_i\, , \, c_d(\gamma) \})_{i\in\mathbb{I}_{d}} &
& \text{if } \ d\leq n \\
(\gamma^\downarrow,0_{d-n}) & & \text{if } \ d> n
\end{array} \peso{for every} \gamma \in \mathbb{R}}\def\C{\mathbb{C}_+^n \ ,
\right. \end{equation} where the constant $c_d(\gamma) \in \mathbb{R}}\def\C{\mathbb{C}_+$ is uniquely determined by $\tr \, L_d (\gamma ) = \tr \, \gamma$, in case $d\le n$. By \cite[Prop. 2.3]{MR10} we know that $\gamma\prec L_d (\gamma )\,$, and $L_d (\gamma )\prec \beta$ for every $\beta\in\mathbb{R}}\def\C{\mathbb{C}^d$ such that $\gamma\prec \beta$. \item\label{item3} Let $\delta=(d_i)_{i\in\mathbb{I}_{m}}\in\mathbb{N}^m$ be such that $1 \le d_1< \ldots< d_m$. For each $B \in W_{\alpha,\,p}$ consider \begin{equation}\label{Bdelta} B_{\delta}= \big[ \, L_{d_i} (R_i(B)\,)\, \big]_{i\in\mathbb{I}_{m}} \in \prod_{i\in\mathbb{I}_{m}} (\mathbb{R}}\def\C{\mathbb{C}_+^{d_i})^\downarrow \ , \end{equation} where $R_i(B)\in \mathbb{R}}\def\C{\mathbb{C}_+^n$ denotes the $i$-th row of $B$. Moreover, using the previous notations we introduce the {\it reduced model (for optimal spectra)} $$ \Lambda^{\rm op}_{\alpha,\,p}(\delta)\ \stackrel{\mbox{\tiny{def}}}{=}\ \{ B_\delta :\ B\in W_{\alpha,\,p}\} \subset \prod_{i\in\mathbb{I}_{m}} (\mathbb{R}}\def\C{\mathbb{C}_+^{d_i})^\downarrow\, . $$ In general, $\Lambda^{\rm op}_{\alpha,\,p}(\delta)$ is not a convex set and indeed, the structure of this set seems rather involved; notice that item 2 above shows that the elements of $\Lambda^{\rm op}_{\alpha,\,p}(\delta)$ are $\prec$-minimizers within appropriate sets.
$\triangle$ \end{enumerate} \end{notas}
\noi The following result describes the existence and uniqueness of the solution to an optimization problem in the reduced model for a fixed $\varphi\in\convfs$, which corresponds to the minimization of the convex potential $P^{\cal W}_\varphi$ in $\mathfrak {B}_\alpha({\cal W})$ for a FSI subspace ${\cal W}$ and a sequence of weights $\alpha\in (\mathbb{R}}\def\C{\mathbb{C}_{>0}^n)^\downarrow$. The proof of this result is presented in section \ref{subsec reduced} (Appendix).
\begin{teo}\label{teo estruc prob reducido unificado} Let $ m,\, n\in\mathbb{N}$, $\alpha\in (\mathbb{R}}\def\C{\mathbb{C}_{>0}^n)^\downarrow$, $p=(p_i)_{i\in\mathbb{I}_{m}}\in \mathbb{R}}\def\C{\mathbb{C}_{>0}^m$ and $\delta=(d_i)_{i\in\mathbb{I}_{m}}\in\mathbb{N}^m$ be such that $1\leq d_1< \ldots< d_m$. If $\varphi\in\convf$ then there exists $\Psi^{\rm op}=[\psi_i^{\rm op}]_{i\in\mathbb{I}_{m}}\in \Lambda^{\rm op}_{\alpha,\,p}(\delta)$ such that $$ \sum_{i\in\mathbb{I}_{m}} {p_i}\,\tr(\varphi(\psi_i^{\rm op}) ) \leq \sum_{i\in\mathbb{I}_{m}} {p_i}\,\tr(\varphi(\psi_i) )
\peso{for every} \Psi=[\psi_i]_{i\in\mathbb{I}_{m}}\in \Lambda^{\rm op}_{\alpha,\,p}(\delta)\,.$$ Moreover: \begin{enumerate} \item If $\varphi\in\convfs$ then such $\Psi^{\rm op}$ is unique; \item If $n\geq d_m$ and $\varphi\in\convfs$ is differentiable in $\mathbb{R}}\def\C{\mathbb{C}_+\,$ then $\Psi^{\rm op}\in \prod_{i\in\mathbb{I}_m} (\mathbb{R}}\def\C{\mathbb{C}_{>0}^{d_i})^\downarrow$. \qed \end{enumerate} \end{teo}
\noi We now turn to the statement and proof of our main result in this section (Theorem \ref{teo min pot fsi generales} below). Hence, we let ${\cal W}$ be an arbitrary FSI subspace of $L^2(\mathbb{R}}\def\C{\mathbb{C}^k)$ and let $\alpha=(\alpha_i)_{i\in\mathbb {I} _n}\in (\mathbb{R}}\def\C{\mathbb{C}_{>0}^n)^\downarrow$. Recall that $$
\mathfrak {B}_\alpha({\cal W})=\{\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in{\cal W}^n:\ E(\mathcal F) \ \text{is a Bessel sequence }, \ \|f_i\|^2=\alpha_i\,,\ i\in\mathbb {I} _n\}\,. $$ Given $\varphi\in\convf$, in what follows we show the existence of finite sequences $\mathcal F^{\rm op}\in \mathfrak {B}_\alpha({\cal W})$ such that $$ P_\varphi^{\cal W}(E(\mathcal F^{\rm op}))=\min\{P_\varphi^{\cal W}(E(\mathcal F)): \ \mathcal F\in \mathfrak {B}_\alpha({\cal W})\}\,. $$ Moreover, in case $\varphi\in\convfs$ then we describe the fine spectral structure of the frame operator of $E(\mathcal F^{\rm op})$ of any such $\mathcal F^{\rm op}$.
\begin{teo}\label{teo min pot fsi generales}
Let $\alpha=(\alpha_i)_{i\in\mathbb {I} _n}\in (\mathbb{R}}\def\C{\mathbb{C}_{>0}^n)^\downarrow$, consider the Notations \ref{nota impor2}
and fix $\varphi\in\convf$. Then, there exists $\mathcal F^{\rm op}\in \mathfrak {B}_\alpha({\cal W})$ such that: \begin{enumerate} \item $\lambda_j([S_{E(\mathcal F^{\rm op})}]_x)=:\psi^{\rm op}_{i,j}\in\mathbb{R}}\def\C{\mathbb{C}_+$ is a.e. constant for $x\in Z_i$, $j\in \mathbb{I}_{i}$ and $i\in\mathbb{I}_\ell$;
\item For every $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$ we have that $$ \sum_{i\in\mathbb{I}_{\ell}} {p_i}\left( \sum_{j\in\mathbb{I}_i}\varphi(\psi^{\rm op}_{i,j})\right)=P_\varphi^{{\cal W}}(E(\mathcal F^{\rm op}))\leq P_\varphi^{{\cal W}}(E(\mathcal F))\ . $$ \end{enumerate}
If we assume that $\varphi\in\convfs$ then: \begin{enumerate} \item[a)] If $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$ is such that $P_\varphi^{{\cal W}}(E(\mathcal F))=P_\varphi^{{\cal W}}(E(\mathcal F^{\rm op}))$ then $S_{E(\mathcal F)}$ has the same fine spectral structure as $S_{E(\mathcal F^{\rm op})}$. \item[b)] If we assume further that $\varphi$ is differentiable in $\mathbb{R}}\def\C{\mathbb{C}_+$ and that
$n\geq i$ for every $i\in\mathbb{I}_\ell$ such that $p_i=|Z_i|>0$, then $E(\mathcal F)$ is a frame for ${\cal W}$. \end{enumerate} \end{teo} \begin{proof}
Without loss of generality, we can assume that there exists an $m\leq \ell $ such that $p_i=|Z_i|>0$ for $i\in\mathbb{I}_{m}$ and $p_i=|Z_i|=0$ for $m+1\leq i\leq \ell$
(indeed, the general case follows by restricting the argument given below to the set of indexes $i\in\mathbb{I}_\ell$ for which $p_i=|Z_i|>0$). We set $p=(p_i)_{i\in\mathbb{I}_m}\in\mathbb{R}}\def\C{\mathbb{C}^m_{>0}$ and consider $\mathcal F=\{f_i\}_{i\in\mathbb {I} _n}\in \mathfrak {B}_\alpha({\cal W})$. For $i\in\mathbb{I}_{m}$ and $j\in\mathbb {I} _n$ set $$
B_{i,j}\ \stackrel{\mbox{\tiny{def}}}{=}\ \frac{1}{ p_i}\,\int_{ Z_i} \|\Gamma f_j(x)\|^2\ dx \implies \sum_{i\in\mathbb{I}_{m}} p_i\,B_{i,j}
=\int_{\text{Spec}({\cal W})}\|\Gamma f_j(x)\|^2\ dx=\|f_j\|^2=\alpha_j \ , $$ for every $j\in\mathbb {I} _n\,$, since $\text{Spec}({\cal W})=\cup_{i\in\mathbb{I}_m}Z_i$. Then $p^T\, B=\alpha $ so using Notations \ref{muchas nots}, $B\in W_{\alpha,\, p}\,$.
\noi Now, fix $i\in\mathbb{I}_{m}$ and consider the weights $\beta^i= p_i\, R_{i}(B)^\downarrow\in\mathbb{R}}\def\C{\mathbb{C}_+^n\,$. For the sake of simplicity we assume, without loss of generality, that $\beta^i=p_i\, R_{i}(B)$. For $i\in\mathbb{I}_m\,$, let ${\cal W}_i$ be the FSI subspace whose fibers coincide with those of ${\cal W}$ inside $Z_i$ and that are the zero subspace elsewhere; hence, Spec$({\cal W}_i)= Z_i$ and $\dim J_{{\cal W}_i}(x)=i$ for $x\in\text{Spec}({\cal W}_i)$. For $i\in\mathbb{I}_m\,$, set $\mathcal F_i=\{f_{i,j}\}_{j\in\mathbb {I} _n}$ where $\Gamma f_{i,j}(x)=\Gamma f_{j}(x)$ for $x\in Z_i$ and $\Gamma f_{i,j}(x)=0$ elsewhere; then $\mathcal F_i\in \mathfrak B_{\beta^i}({\cal W}_i)$ and $$ [S_{E(\mathcal F_i)}]_x=S_{\Gamma \mathcal F_i(x)}=[S_{E(\mathcal F)}]_x \peso{for} x\in Z_i=\text{Spec}({\cal W}_i) \ , \quad i\in\mathbb{I}_m\, .$$ If we consider the minimization of $P_\varphi^{{\cal W}_i}$ in $\mathfrak B_{\beta^i}({\cal W}_i)$ then, Theorem \ref{teo dim unif} and Remark \ref{interpret de prop dim unif} imply that there exists $c_i\geq 0$ such that \begin{equation} \label{desi caso uniforme}
p_i\, \sum_{j\in\mathbb{I}_{i}} \varphi (\max\{B_{i,j}\, , \, c_i\}) \leq P_\varphi^{{\cal W}_i}(E(\mathcal F_i)) \peso{and} \sum_{j\in I_{i}} \max\{B_{i,j}\, , \, c_i\}=\sum_{i\in\mathbb {I} _n} B_{i,j} \,. \end{equation} Using Notations \ref{muchas nots} and Eq. \eqref{eq. c es el correcto}, we get that for $i\in\mathbb{I}_m$ $$ L_{i}(R_{i}(B))= (\max\{B_{i,j}\, , \, c_i\})_{j\in I_{i}} \implies B_{\delta}=[(\max\{B_{i,j}\, , \, c_i\})_{j\in \mathbb{I}_{i}}]_{i\in\mathbb{I}_{m}}\in \Lambda^{\rm op}_{\alpha,\,p}(\delta)\,,$$
where $\delta=(i)_{i\in\mathbb{I}_m}$. Notice that ${\cal W}=\oplus_{i\in\mathbb{I}_{m}}{\cal W}_i$ (orthogonal sum) and hence $$ \sum_{i\in\mathbb{I}_{m}} p_i\, \sum_{j\in\mathbb{I}_{i}} \varphi (\max\{B_{i,j}\, , \, c_i\}) \le \sum_{i\in\mathbb{I}_{m}} P_\varphi^{{\cal W}_i}(E(\mathcal F_i))=P_\varphi^{\cal W}(E(\mathcal F))\,. $$
\noi Let $[\psi^{\rm op}_i]_{i\in\mathbb{I}_{m}}=\Psi^{\rm op}
\,\in \Lambda_{\alpha,\,p}^{\rm op}(\delta)$ be as in
Theorem \ref{teo estruc prob reducido unificado}. Then \begin{equation}\label{desi potop} \sum_{i\in\mathbb{I}_{m}} { p_i}\, \tr(\varphi(\psi^{\rm op}_i)) = \sum_{i\in\mathbb{I}_{m}} {p_i}\left( \sum_{j\in\mathbb{I}_i}\varphi(\psi^{\rm op}_{i,j})\right) \leq \sum_{i\in\mathbb{I}_{m}} p_i\, \sum_{j\in\mathbb{I}_{i}} \varphi (\max\{B_{i,j}\, , \, c_i\})\leq P_\varphi^{\cal W}(E(\mathcal F))\ . \end{equation} Recall that by construction, there exists $B^{\rm op}=(\gamma_{i,j})_{(i\, , \, j)\in\mathbb{I}_{m}\times\mathbb {I} _n}\in W_{\alpha,\,p}$ such that $B^{\rm op}_{\delta}=\Psi^{\rm op}$ (see item \ref{item3} in Notations \ref{muchas nots}). In this case, $$ \psi^{\rm op}_i=L_{i}(\,(\gamma_{i,j})_{j\in\mathbb {I} _n}) \implies (\gamma_{i,j})_{j\in\mathbb {I} _n}\prec \psi^{\rm op}_i \peso{for} i\in\mathbb{I}_{m}\ . $$ Let $\gamma:\text{Spec}({\cal W})\rightarrow \mathbb{R}}\def\C{\mathbb{C}^n$ be given by $\gamma(x) =R_i(B^{\rm op})= (\gamma_{i,j})_{j\in\mathbb {I} _n}$ if $x\in Z_i$, for $i\in\mathbb{I}_{m}\,$; similarly, let $\lambda:\text{Spec}({\cal W})\rightarrow \coprod_{i\in\mathbb{I}_{m}}\mathbb{R}}\def\C{\mathbb{C}^{i}$, $\lambda(x)=\psi^{\rm op}_i$ if $x\in Z_i$, for $i\in\mathbb{I}_{m}\,$. Then, by the previous remarks we get that $\gamma(x)\prec \lambda(x)$ for $x\in\text{Spec}({\cal W})$.
\noi Hence, by Theorem \ref{teo sobre disenio de marcos} there exists $\mathcal F^{\rm op}=\{f_j^{\rm op}\}_{j\in\mathbb {I} _n}$ such that $$
\|\Gamma f_j^{\rm op}(x)\|^2=\gamma_{i,j} \peso{and} \lambda_j([S_{E(\mathcal F^{\rm op})}]_x)=\psi_{i\, , \, j}^{\rm op} \peso{for} x\in Z_i\, , \ \ j\in\mathbb{I}_i \peso{and} i\in\mathbb{I}_{m}\ . $$ Since $B^{\rm op}\in W_{\alpha,\,p}$ then
$$\|f_j^{\rm op}\|^2=\int_{\text{Spec}({\cal W})}\|\Gamma f_j^{\rm op}(x)\|^2\ dx=\sum_{i\in\mathbb{I}_{m}} p_i \ \gamma_{i,j}=\alpha_j \peso{for} j\in\mathbb {I} _n\implies \mathcal F^{\rm op}\in \mathfrak B_{\alpha}({\cal W})$$ and \begin{equation}\label{eq casi estamos} P_\varphi^{{\cal W}}(E(\mathcal F^{\rm op}))=\int_{\text{Spec}({\cal W})} \tr(\varphi(\lambda(x)))\ dx=\sum_{i\in\mathbb{I}_{m}} p_i \ \tr(\varphi(\psi_i^{\rm op})),
\end{equation} then by Eq. \eqref{desi potop} we see that $P_\varphi^{{\cal W}}(E(\mathcal F^{\rm op}))\leq P_\varphi^{{\cal W}}(E(\mathcal F))\,.$ Since $\mathcal F\in \mathfrak B_{\alpha}({\cal W})$ was arbitrary, the previous facts show that $\mathcal F^{\rm op}$ satisfies items 1. and 2. in the statement.
\noi Assume further that $\varphi\in\convfs$ and $\mathcal F\in \mathfrak {B}_\alpha({\cal W})$ is such that $P_\varphi^{{\cal W}}(E(\mathcal F))=P_\varphi^{{\cal W}}(E(\mathcal F^{\rm op}))$. Then, by Eqs. \eqref{desi caso uniforme}, \eqref{desi potop} and \eqref{eq casi estamos} we see that $$
p_i\, \sum_{j\in\mathbb{I}_{i}} \varphi (\max\{B_{i,j}\, , \, c_i\}) = P_\varphi^{{\cal W}_i}(E(\mathcal F_i)) \peso{for} i\in\mathbb{I}_{m}\ . $$ Therefore, by the case of equality in Theorem \ref{teo struct fina dim hom} and the uniqueness of $\Psi^{\rm op}$ from Theorem \ref{teo estruc prob reducido unificado} we conclude that $$ \lambda_j([S_{E(\mathcal F)}]_x) =\lambda_j([S_{E(\mathcal F_i)}]_x)=\psi^{\rm op}_{i,\,j} \peso{for} x\in Z_i\, , \ j\in\mathbb{I}_{i}\, , \ i\in\mathbb{I}_{m}\ . $$ Finally, in case $\varphi\in\convfs$ is differentiable in $\mathbb{R}}\def\C{\mathbb{C}_+$ and $n\geq m$ then, again by Theorem \ref{teo estruc prob reducido unificado}, we see that $S_{E(\mathcal F)}$ is bounded from below in ${\cal W}$ (since the vectors in $\Psi^{\rm op}$ have no zero entries) and hence $E(\mathcal F)$ is a frame for ${\cal W}$. \end{proof}
\noi We end this section with the following remarks. With the notations of Theorem \ref{teo min pot fsi generales}, notice that the optimal Bessel sequence $\mathcal F^{\rm op}\in \mathfrak {B}_\alpha({\cal W})$ depends on the convex function $\varphi\in\convf$, which was fixed in advance. That is, unlike the uniform case, we are not able to show that there exists $\mathcal F^{\rm univ}\in \mathfrak {B}_\alpha({\cal W})$ such that $\mathcal F^{\rm univ}$ is a $P^{\cal W}_\varphi$-minimizer in $\mathfrak {B}_\alpha({\cal W})$ for every $\varphi\in\convf$. It is natural to wonder whether there exists such a universal solution $\mathcal F^{\rm univ}\in \mathfrak {B}_\alpha({\cal W})$; we conjecture that this is always the case.
\section{Appendix}\label{Appendixity}
\subsection{The Schur-Horn theorem for measurable fields of self-adjoint matrices and applications}
The simple notion of majorization between real vectors has played an important role in finite frame theory in finite dimensions. In particular, it is well known that the existence of finite sequences with prescribed norms and frame operator can be characterized in terms of majorization, applying the Schur-Horn theorem.
\noi Next we develop a Schur-Horn type theorem for measurable fields of self-adjoint matrices and use this result to prove Theorem \ref{teo:mayo equiv}.
Our proof is an adaptation of that given in \cite{HJ13} for the classical Schur-Horn theorem. We will use the existence of measurable eigenvalues and eigenvectors (i.e. diagonalization by measurable fields of unitary matrices) of measurable fields of self-adjoint matrices from \cite{RS95}. In what follows we consider a measure subspace $(X,\, \mathcal X,\, |\,\cdot|)$ of the measure space $(\mathbb{T}^k,\,\mathcal B(\mathbb{T}^k),\,|\,\cdot|)$ of the $k$-torus with Lebesgue measure on Borel sets.
\begin{teo} \label{teo:mayo y el unitario} Let $A(\cdot): X \to \mathcal{H}(n)$ be a measurable field of self-adjoint matrices with associated measurable eigenvalues $b_j:X\to \mathbb{R}}\def\C{\mathbb{C}$ for $j\in \mathbb{I}_n$ such that $b_1\geq \cdots \geq b_n\,$. Let $c_j:X\to \mathbb{R}}\def\C{\mathbb{C}$ be measurable functions for $j\in \mathbb{I}_n\,$. The following statements are equivalent: \begin{enumerate} \item $c(x)=(c_1(x)\, , \, \cdots \, , \, c_n(x))\prec b(x)=(b_1(x)\, , \, \cdots\, , \, b_n(x))$, for a.e. $x\in X$. \item There exists a measurable field of unitary matrices $U(\cdot):X\to {\cal U}(n)$, such that \begin{equation}\label{eq SH} d(U(x)^*\,A(x)\ U(x))=c(x)\, , \peso{for a.e.} x\in X\,, \end{equation} where $d(B)\in \C^n$ denotes the main diagonal of the matrix $B\in {\cal M}_n(\C)$. \end{enumerate} \end{teo} \begin{proof} First notice that the implication $2.\implies 1.$ follows from the classical Schur theorem.
\noi $1.\implies 2.\,$: By considering a convenient measurable field of permutation matrices, we can (and will) assume that the entries of the vector $c(x)$ is also arranged in non-increasing order: $c_1(x)\geq c_2(x)\geq \ldots c_n(x)$. By the results from \cite{RS95} showing the existence of a measurable field of unitary matrices diagonalizing the field $A$, we can assume without loss of generality that $A(x)=D_{b(x)}$ where $D_{b(x)}$ is the diagonal matrix with main diagonal $(b_1(x)\, , \, \ldots \, , \, b_n(x))$ for a.e. $x\in X$.
\noi We will argue by induction on $n$. For $n=1$ the result is trivial.
Hence, we may assume that $n\geq 2$. Since $c(x)\prec b(x)$, we have $b_1(x)\geq c_1(x)\geq c_n(x)\geq b_n(x)$, so if $b_1(x)=b_n(x)$ it follows that all the entries of $c(x)$ and $b(x)$ coincide, $A(x)=c_1(x) I_n\,$, and we can take $U(x)=I_n\,$ for every such $x \in X$. By considering a convenient partition of $X$ we may therefore assume that $b_1(x)>b_n(x)$ in $X$. Similarly, in case $c_1(x)=c_n(x)$ then the unitary matrix $U(x)=n^{-1/2}\, (w^{j\,k})_{j,k\in\mathbb {I} _n}\,$, where $w=e^{\frac{-2\pi i}{ \ n}}$, satisfies that $U(x)^*\, D_{b(x)}\ U(x)=(c_1(x)\, , \, \ldots \, , \, c_n(x))$. Therefore, by considering a convenient partition of $X$ we may therefore assume that $c_1(x)>c_n(x)$ in $X$
\noi For $n=2$, we have $b_1(x)>b_2(x)$ and $b_1(x)\geq c_1(x)\geq c_2(x)= (b_1(x)- c_1(x))+b_2(x)\geq b_2(x).$ Consider the matrix
$$U(x)=\frac{1}{\sqrt{b_1(x)-b_2(x)}} \begin{pmatrix} \sqrt{b_1(x)-c_2(x)} &-\sqrt{c_2(x)-b_2(x)} \\ \sqrt{b_2(x)-c_2(x)} &\sqrt{b_1(x)-c_2(x)} \end{pmatrix} \peso{for a.e.} x\in X\,. $$ Notice that $U(x):X \to M_2(\C)^+$ is a measurable function and an easy computation reveals that $U(x)^*\, U(x)=I_2$, so $U(x)$ is unitary for a.e. $x\in X$. A further computation shows that $$U(x)^*\, A(x)\ U(x)= \begin{pmatrix} &c_1(x)& &*& \\ &*& &c_2(x)& \end{pmatrix} \peso{for a.e.} x\in X\,. $$ That is, $d(U^*(x)\, A(x)\, U(x))=(c_1(x),\,\ c_2(x))$ and $U(\cdot)$ has the desired properties.
\noi Suppose that $n\geq 3$ and asssume that the theorem is true if the vectors $c(x)$ and $b(x)$ have size at most $n-1$. For each $x\in X$ let $k(x)$ be the largest integer $k\in\mathbb {I} _n$ such that $b_k(x)\geq c_1(x)$. Since $b_1(x)\geq c_1(x)>c_n(x)\geq b_n(x)$, we see that $1\leq k\leq n-1$. Then, by considering a convenient partition of $X$ into measurable sets we can assume that $k(x)=k$ for $x\in X$. Therefore, by definition of $k$ we get that $b_k(x)\geq c_1(x)>b_{k+1}(x)$ for $x\in X$. Let $\eta(x)=b_k(x)+b_{k+1}(x)-c_1(x)$ and observe that $\eta(x)=(b_k(x)-c_1(x))+b_{k+1}(x)\geq b_{k+1}(x)$. Then, the measurable vector $(b_k(x), b_{k+1}(x))$ majorizes the measurable vector $(c_1(x), \eta(x))$ and $b_k(x)>b_{k+1}(x)$ for a.e. $x\in X$.
Let $$D_1(x)=\begin{pmatrix} &b_k(x)& &0& \\ &0& &b_{k+1}(x)& \end{pmatrix} \peso{for a.e.} x\in X\,.$$ By the case $n=2$ we obtain a measurable field of unitary matrices $U_1(\cdot):X\rightarrow {\cal U}(2)$ such that $$d(U_1(x)^*\, D_1(x)\, U_1(x))= (c_1(x), \eta(x)) \peso{for a.e.} x\in X\,.$$ Since $b_k(x)=\eta(x)+(c_1(x)-b_{k+1}(x))>\eta(x)$, we have:
\noi If $k=1$ then $b_1(x)>\eta(x)\geq b_2(x)\geq \cdots \geq b_n(x)$; if we let $D_2(x)\in \mathbb {M}_{n-2}(\C)$ be the diagonal matrix with main diagonal $(b_3(x)\, , \, \ldots\, , \, b_n(x))$ then $D_{b(x)}=D_1(x)\oplus D_2(x)$ and $$\begin{pmatrix} U_1(x)& 0 \\ 0 &I_{n-2} \end{pmatrix}^* \begin{pmatrix} D_1(x)& 0 \\ 0 &D_2(x) \end{pmatrix} \begin{pmatrix} U_1(x) &0 \\ 0 &I_{n-2} \end{pmatrix}=\begin{pmatrix}c_1(x) &Z(x)^*\\ Z(x) &V_1(x)\end{pmatrix}$$ where $Z(x)^*=(\overline{z(x)}\, , \, 0\, , \, \ldots\, , \, 0)\in M_{1,(n-1)}(\C)$, $z(\cdot):X\rightarrow \C$ is a measurable function and $V_1(x)\in\mathbb {M}_{n-1}(\C)$ is the diagonal matrix with main diagonal $(\eta(x)\, , \, b_3(x)\, , \, \ldots\, , \, b_n(x))$. Moreover, in this case it turns out that $(\eta(x)\, , \, b_3(x)\, , \, \cdots \, , \, b_n(x))$ majorizes $(c_2(x)\, , \, \cdots\, , \, c_n(x))$ for a.e. $x\in X$ (see \cite{HJ13}). By the inductive hypothesis there exists a measurable field $U_2(\cdot):X\rightarrow {\cal U}(n-1)$ such that $d(U_2(x)^* V_1(x) U_2(x))=(c_2(x)\, , \, \cdots\, , \, c_n(x))$. Hence, if we set $U(x)=(U_1(x)\oplus I_{n-2})\cdot (1\oplus U_2(x))$ for $x\in X$ then $U(\cdot):X\rightarrow {\cal U}(n)$ has the desired properties.
\noi If $k>1$ then $b_1(x)\geq\ldots \geq b_{k-1}(x)\geq b_{k}(x)>\eta(x)\geq b_{k+1}(x)\geq \ldots \geq b_n(x)$. Let $D_2(x)\in \mathbb {M}_{n-2}(\C)$ be the diagonal matrix with main diagonal $$\beta(x)\ \stackrel{\mbox{\tiny{def}}}{=}\ (b_1(x)\, , \, \ldots\, , \, b_{k-1}(x)\, , \, b_{k+2}(x)\, , \, \ldots\, , \, b_n(x))\in\mathbb{R}}\def\C{\mathbb{C}^{n-2}.$$ Notice that in this case $$\begin{pmatrix} U_1(x)& 0 \\ 0 &I_{n-2} \end{pmatrix}^* \begin{pmatrix} D_1(x)& 0 \\ 0 &D_2(x) \end{pmatrix} \begin{pmatrix} U_1(x) &0 \\ 0 &I_{n-2} \end{pmatrix}=\begin{pmatrix}c_1(x) &W(x)^*\\ W(x) &V_2(x)\end{pmatrix}$$ where $W(x)^*=(\overline{w(x)}\, , \, 0\, , \, \ldots\, , \, 0)\in M_{1,(n-1)}(\C)$, $w(\cdot):X\rightarrow \C$ is a measurable function and $V_2(x)\in M_{n-1}(\C)$ is the diagonal matrix with main diagonal $$ \gamma(x)\ \stackrel{\mbox{\tiny{def}}}{=}\ (\eta(x)\, , \, b_1(x)\, , \, \ldots\, , \, b_{k-1}(x)\, , \, b_{k+2}(x)\, , \, \ldots\, , \, b_n(x)) \peso{for a.e.} x\in X \ . $$ It turns out that $(c_2(x)\, , \, \ldots\, , \, c_n(x))\prec \gamma(x)$ for a.e. $x\in X$; by the inductive hypothesis there exists a measurable field $U_2(\cdot):X\rightarrow {\cal U}(n-1)$ such that $d(U_2(x)^* V_2(x) U_2(x))=(c_2(x)\, , \, \ldots\, , \, c_n(x))$ for a.e. $x\in X$. Notice that there exists a permutation matrix $P\in{\cal U}(n)$ such that $P^*(x) D_{b(x)} P=D_1\oplus D_2\,$. Hence, if we set $U(x)=P\cdot (U_1(x)\oplus I_{n-2})\cdot (1\oplus U_2(x))$ for a.e. $x\in X$ then, $U(\cdot):X\rightarrow {\cal U}(n)$ has the desired properties. \end{proof}
\noi Next we prove Theorem \ref{teo:mayo equiv}, based on the Schur-Horn theorem for measurable field i.e. Theorem \ref{teo:mayo y el unitario} above. Our approach is an adaptation of some known results in finite frame theory (see \cite{AMRS}).
\noi {\bf Theorem \ref{teo:mayo equiv}} \it Let $b:\mathbb{T}^k\rightarrow (\mathbb{R}}\def\C{\mathbb{C}_+)^d$ and $c:\mathbb{T}^k\rightarrow (\mathbb{R}}\def\C{\mathbb{C}_+)^n$ be measurable vector fields. The following statements are equivalent: \begin{enumerate} \item For a.e. $x\in \mathbb{T}^k$ we have that $c(x)\prec b(x)$.
\item There exist measurable vector fields $u_j: \mathbb{T}^k\to \C^d$ for $j\in\mathbb {I} _n$ such that $\|u_j(x)\|=1$ for a.e. $x\in \mathbb{T}^k$ and $j\in \mathbb{I}_n\,$, and such that $$ D_{b(x)}=\sum_{j\in \mathbb{I}_n} c_j(x)\,\ u_j(x) \otimes u_j(x) \ , \peso{for a.e.} \ x\in \mathbb{T}^k\ . $$ \end{enumerate} \rm \begin{proof} First notice that the implication $2.\implies 1.$ follows from well known results in finite frame theory (see \cite{AMRS}) in each point $x\in \mathbb{T}^k$. Hence, we show $1.\implies 2.$ We assume, without loss of generality, that the entries of the vectors $b(x)$ and $c(x)$ are arranged in non-increasing order. We now consider the following two cases:
\noi {\bf Case 1:} assume that $n<d$. We let $\tilde c:\mathbb{T}^k\rightarrow \C^d$ be given by $\tilde c(x)=(c(x)\, , \, 0_{d-n})$ for $x\in \mathbb{T}^k$. Then, $\tilde c(x)\prec b(x)$ for $x\in\mathbb{T}^k$ and therefore, by Theorem \ref{teo:mayo y el unitario} there exists a measurable field $U(\cdot):\mathbb{T}^k\rightarrow {\cal U}(d)$ such that \begin{equation} \label{eq aplic SH11} d(U(x)^* D_{b(x)} \, U(x))=(c_1(x)\, , \, \ldots\, , \, c_n(x)\, , \, 0_{d-n})\peso{for a.e.} x\in\mathbb{T}^k\,. \end{equation} Let $v_1(x)\, , \, \ldots\, , \, v_d(x)\in\C^d$ denote the columns of $C(x)=D_{b(x)}^{1/2}\,U(x)$, for $x\in\mathbb{T}^k$. Then, Eq. \eqref{eq aplic SH11} implies that:
$$ \|v_j(x)\|^2= c_j(x) \peso{for} j\in\mathbb{I}_n \ , \quad v_j=0 \peso{for} n+1\leq j\leq d $$ $$ \peso{and} D_{b(x)}= C(x)\, C(x)^*=\sum_{j\in\mathbb{I}_n} v_j(x)\otimes v_j(x) \peso{for a.e.} x\in\mathbb{T}^k \,.$$ Thus, the vectors $u_j(x)$ are obtained from $v_j(x)$ by normalization, for a.e. $x\in\mathbb{T}^k$ and $j\in\mathbb{I}_n\,$.
\noi {\bf Case 2:} assume that $n\geq d$. We let $\tilde b:\mathbb{T}^k\rightarrow \C^n$ be given by $\tilde b(x)=(b(x)\, , \, 0_{n-d})$ for $x\in \mathbb{T}^k$. Then, $c(x)\prec \tilde b(x)$ for $x\in\mathbb{T}^k$ and therefore, by Theorem \ref{teo:mayo y el unitario} there exists a measurable field $U(\cdot):\mathbb{T}^k\rightarrow {\cal U}(n)$ such that \begin{equation} \label{eq aplic SH1} d(U(x)^* D_{\tilde b(x)} \, U(x))=(c_1(x)\, , \, \ldots\, , \, c_n(x))\peso{for a.e.} x\in\mathbb{T}^k\ . \end{equation} Let $\tilde v_1(x)\, , \, \ldots\, , \, \tilde v_n(x)\in\C^n$ denote the columns of $C(x)=D_{\tilde b(x)}^{1/2}U(x)$, for $x\in\mathbb{T}^k$. As before, Eq. \eqref{eq aplic SH1} implies that $$
\|\tilde v_j(x)\|^2= c_j(x) \peso{for} j\in\mathbb{I}_n \peso{and} D_{\tilde b(x)}= \sum_{j\in\mathbb{I}_n} \tilde v_j(x)\otimes \tilde v_j(x) \peso{for a.e.} x\in\mathbb{T}^k \ . $$ If we let $\tilde v_j(x)=(v_{i,j}(x))_{i\in\mathbb {I} _n}$ then, the second identity above implies that $\tilde v_{i,j}(x)=0$ for a.e. $x\in\mathbb{T}^k$ and every $d+1\leq i\leq n$. If we let $v_j(x)=(v_{i,j}(x))_{i\in\mathbb{I}_d}$ for a.e. $x\in\mathbb{T}^k$ and $j\in\mathbb {I} _n\,$, we get that
$$ \|v_j(x)\|^2= c_j(x) \peso{for} j\in\mathbb{I}_n \peso{and} D_{b(x)}= \sum_{j\in\mathbb{I}_n} v_j(x)\otimes v_j(x) \peso{for a.e.} x\in\mathbb{T}^k \,.$$ Thus, the vectors $u_j(x)$ are obtained from $v_j(x)$ by normalization, for a.e. $x\in\mathbb{T}^k$ and $j\in\mathbb{I}_n\,$. \end{proof}
\subsection{The reduced finite-dimensional model: proof of Theorem \ref{teo estruc prob reducido unificado}}\label{subsec reduced}
In this section we present the proof of Theorem \ref{teo estruc prob reducido unificado}, divided into two parts (namely, Propositions \ref{teo estruc prob reducido} and \ref{era facilongo nomas} below).
\begin{pro}\label{teo estruc prob reducido} Let $ m,\, n\in\mathbb{N}$, $\alpha\in (\mathbb{R}}\def\C{\mathbb{C}_{>0}^n)^\downarrow$, $p=(p_i)_{i\in\mathbb{I}_{m}}\in \mathbb{R}}\def\C{\mathbb{C}_{>0}^m$ and $\delta=(d_i)_{i\in\mathbb{I}_{m}}\in\mathbb{N}^m$ be such that $1\leq d_1< \ldots< d_m$. If $\varphi\in\convf$ then there exists $\Psi^{\rm op}=[\psi_i^{\rm op}]_{i\in\mathbb{I}_{m}}\in \Lambda^{\rm op}_{\alpha,\,p}(\delta)$ such that $$ \sum_{i\in\mathbb{I}_{m}} {p_i}\,\tr(\varphi(\psi_i^{\rm op}) ) \leq \sum_{i\in\mathbb{I}_{m}} {p_i}\,\tr(\varphi(\psi_i) )
\peso{for every} \Psi=[\psi_i]_{i\in\mathbb{I}_{m}}\in \Lambda^{\rm op}_{\alpha,\,p}(\delta)\,.$$ Moreover, if $\varphi\in\convfs$ then such $\Psi^{\rm op}$ is unique. \end{pro} \begin{proof} Let us consider the set $$ \Lambda_{\alpha,\,p}(\delta)\ \stackrel{\mbox{\tiny{def}}}{=}\ \bigcup_{B\in W_{\alpha,\,p}} M(B) \subseteq \prod_{i\in\mathbb{I}_{m}} (\mathbb{R}}\def\C{\mathbb{C}_+^{d_i})^\downarrow \ , $$ where $$ M(B)\ \stackrel{\mbox{\tiny{def}}}{=}\ \{ [\lambda_i]_{i\in\mathbb{I}_{m}}\in \prod_{i\in\mathbb{I}_{m}} (\mathbb{R}}\def\C{\mathbb{C}_+^{d_i})^\downarrow:\ R_i(B)\prec \lambda_i\ , \ i\in\mathbb{I}_{m}\} \ . $$ Notice that by construction $\Lambda_{\alpha,\,p}^{\rm op}(\delta)\subseteq \Lambda_{\alpha,\,p}(\delta)$.
\noi We claim that $\Lambda_{\alpha,\,p}(\delta)$ is a convex set. Indeed, let $[\lambda_i]_{i\in\mathbb{I}_{m}}\in M(B_1)$, $[\mu_i]_{i\in\mathbb{I}_{m}}\in M(B_2)$ for $B_1 $, $B_2\in W_{\alpha,\,p}$ and $t\in [0,1]$. Take the matrix $B = t\, B_1 + (1-t)\,B_2\in W_{\alpha,\,p}\,$ (since $ W_{\alpha,\,p}$ is a convex set). Then $$ [\,\gamma_i\,]_{i\in\mathbb{I}_{m}} = [\,t\,\lambda_i+(1-t)\, \mu_i\,]_{i\in\mathbb{I}_{m}}\in M(B) \subseteq \Lambda_{\alpha,\,p}(\delta) \ : $$ on the one hand, $\gamma_i\in(\mathbb{R}}\def\C{\mathbb{C}_+^{d_i})^\downarrow$, $i\in\mathbb{I}_m$; on the other hand, by Lidskii's additive inequality (see \cite{Bhat}) we have that, for each $i\in\mathbb{I}_{m}\,$ $$ R_i(B)= t\,R_i(B_1)+ (1-t)\,R_i(B_2)\prec t\, R_i(B_1)^\downarrow + (1-t)\, R_i(B_2)^\downarrow \in (\mathbb{R}}\def\C{\mathbb{C}_+)^\downarrow \ . $$ On the other hand, by the hypothesis (and the definition of majorization) one deduces that $$ R_i(B_1)^\downarrow\prec \lambda_i \peso{and} R_i(B_2)^\downarrow\prec \mu_i \implies R_i(B) \prec t\, \lambda_i + (1-t)\, \mu_i= \gamma_i $$ for every $i\in \mathbb{I}_{m}\,$. This proves the claim, so $\Lambda_{\alpha,\,p}(\delta)$ is a convex set. Moreover, by the compactness of $W_{\alpha,\,p}$ and by the conditions defining $M(B)$ for $B\in W_{\alpha,\,p}\,$, it follows that $\Lambda_{\alpha,\,p}(\delta)$ is a compact set. Let $$ \varphi_p:\Lambda_{\alpha,\,p}(\delta)\rightarrow \mathbb{R}}\def\C{\mathbb{C}_+ \peso{given by} \varphi_p(\Psi)\ \stackrel{\mbox{\tiny{def}}}{=}\ \sum_{i\in\mathbb{I}_{m}} {p_i}\,\tr \,\varphi(\psi_i) \ , $$ for $\Psi=[\psi_i]_{i\in\mathbb{I}_{m}} \in \Lambda_{\alpha,\,p}(\delta)\,$. It is easy to see that $\varphi_p$ is a convex function, which is strictly convex whenever $\varphi\in\convfs$. Using this last fact it follows that there exists $ \Psi_0\in \Lambda_{\alpha,\,p}(\delta)$ that satisfies $$ \varphi_p(\Psi_0)\leq \varphi_p(\Psi) \peso{for every} \Psi\in \Lambda_{\alpha,\,p}(\delta)\ , $$ and such $ \Psi_0$ is unique whenever $\varphi\in\convfs$. Notice that by construction there exists some $B\in W_{\alpha,\,p}$ such that $\Psi_0=[\psi_i^0]_{i\in\mathbb{I}_{m}} \in M(B)$. Then, by item \ref{item2} of Notation \ref{muchas nots}, $$ R_i(B)\prec \psi_i^0 \implies L_{d_i}(R_i(B))\prec \psi_i^0 \implies \tr \, \varphi(L_{d_i}(R_i(B)))\leq \tr \,\varphi(\psi_i^0) \peso{for} i\in\mathbb{I}_{m}\ . $$ Hence, the sequence $B_\delta$ defined in Eq. \eqref{Bdelta} using this matrix $B$ satisfies that $\varphi_p(B_\delta)\le \varphi_p(\Psi_0)$. So we define $\Psi^{\rm op}\ \stackrel{\mbox{\tiny{def}}}{=}\ B_\delta\in \Lambda_{\alpha,\,p}^{\rm op}(\delta)\subset \Lambda_{\alpha,\,p}(\delta)$, that has the desired properties. Finally, the previous remarks show that $\Psi_0= \Psi^{\rm op}\in \Lambda_{\alpha,\,p}^{\rm op}(\delta)$ whenever $\varphi\in\convfs$. \end{proof}
\begin{pro}\label{era facilongo nomas} With the notations and terminology of Proposition \ref{teo estruc prob reducido}, assume further that $n\geq d_m$ and that $\varphi\in\convfs$ is differentiable in $\mathbb{R}}\def\C{\mathbb{C}_+\,$. Then $$ \Psi^{\rm op}\in \prod_{i\in\mathbb{I}_m} (\mathbb{R}}\def\C{\mathbb{C}_{>0}^{d_i})^\downarrow\,.$$ \end{pro} \begin{proof} Let $\Psi^{\rm op}=[\psi_i^{\rm op}]_{i\in\mathbb{I}_{m}}$ where each vector $\psi_i^{\rm op} \in (\mathbb{R}}\def\C{\mathbb{C}_{+}^{d_i})^\downarrow \,$,
and assume that there exists $i_0\in\mathbb{I}_m$ such that $\psi_{i_0}^{\rm op}=(\psi_{i_0,j}^{\rm op})_{j\in \mathbb{I}_{d_{i_0}}}$ satisfies that $\psi^{\rm op}_{i_0,k}=0$ for some $1\leq k\leq d_{i_0}$; let $1\leq k_0\leq d_{i_0}$ be the smallest such index. Let $B\in W_{\alpha,\,p}$ be such that $B_\delta=\Psi^{\rm op}$. Recall from Eq. \eqref{eq defi gammacd} that, if we denote $c_i = c_{d_{i}}(R_{i}(B))$ for every $i \in \mathbb{I}_{m}\,$, then $$ \psi^{\rm op}_{i_0,j}= L_{d_{i_0}} (R_{i_0}(B))_j = \max\{R_{i_0}(B)^\downarrow_j\, , \, c_{i_0}\} \peso{for} j\in \mathbb{I}_{d_{i_0}} \ , $$ since $n\geq d_{i_0}$ by hypothesis. Hence, in this case $c_{i_0}=
0$ and $R_{i_0}(B)^\downarrow_{k_0}=0$. Let $j_0\in \mathbb {I} _n $ such that $0=R_{i_0}(B)^\downarrow_{k_0} = B_{i_0\, , \, j_0}\,$. By construction $\sum_{i\in\mathbb{I}_m}p_i\ B_{i\, , \, j_0}=\alpha_{j_0}>0$ so that there exists $i_1\in\mathbb{I}_m$ such that $B_{i_1,\,j_0}>0$. Let $\{e_j\}_{i\in\mathbb {I} _n}$ denote the canonical basis of $\mathbb{R}}\def\C{\mathbb{C}^n$. For every $t\in I=[0,\frac{\beta_{i_1,\,j_0} \ p_{i_1}}{p_{i_0}}]$ consider the matrix $B(t)$ defined by its rows as follows: \begin{itemize} \item $R_{i_0}(B(t))=R_{i_0}(B)+t\, e_{j_0} $ \item $R_{i_1}(B(t))= R_{i_1}(B)- \frac{p_{i_0}\, t}{p_{i_1}}\ e_{j_0}\,$ \item $R_{i}(B(t))=R_{i}(B)$ for $i\in\mathbb{I}_m\setminus\{i_0,\,i_1\}$. \end{itemize} It is straightforward to check that $B(t)\in W_{\alpha,\,p}$ for $t\in I$ and that $B(0)=B$. Set $\Psi(t)=[\psi_i(t)]_{i\in\mathbb{I}_m}=B(t)_\delta\in \Lambda_{\alpha,\, p}^{\rm op}(\delta)$ for $t\in I$ and notice that $\Psi(0)=\Psi^{\rm op}$. We now consider two cases:
\noi {\bf Case 1:\ }\ $B_{i_1,\,j_0}> c_{i_1}$ (recall that $\psi^{\rm op}_{i_1\, , \, j}= L_{d_{i_1}} (R_{i_1}(B))_j = \max\{R_{i_1}(B)^\downarrow_j\, , \, c_{i_1}\} $). Therefore $B_{i_1,\,j_0} =R_{i_1}(B)^\downarrow_{k}$ for some $1\leq k\leq d_{i_1}$ and we let $1\leq k_1\leq d_{i_1}$ be the largest such $k$. It is straightforward to check that in this case there exists $\varepsilon >0$ such that $$\psi_{i_0}(t)=\psi^{\rm op}_{i_0}+t\, e_{k_0} \peso{and} \psi_{i_1}(t)=\psi^{\rm op}_{i_1}-\frac{p_{i_0}}{p_{i_1}}\, t\, e_{k_1} \peso{for} t\in [0,\epsilon]\,.$$ Therefore, for $t\in [0,\epsilon]$ we have that $$ f(t)=\varphi_p(\Psi(t))- \varphi_p(\Psi^{\rm op})=p_{i_0}\ (\varphi(t)-\varphi(0))+p_{i_1}\ (\varphi(B_{i_1,\, j_0} - \frac{p_{i_0}}{p_{i_1}} t ) - \varphi(B_{i_1,\, j_0}))\ . $$ Hence $f(0)=0$ and by hypothesis $f(t)\geq 0$ for $t\in[0,\epsilon]$. On the other hand, $$ f'(0)= p_{i_0}\ (\varphi'(0) - \varphi'(B_{i_1,\, j_0}) )<0 $$ since by the hypothesis $\varphi'$ is strictly increasing and $B_{i_1,\, j_0}>0$. This condition contradicts the previous facts about $f$. From this we see that the vectors in $\Psi^{\rm op}$ have no zero entries.
\noi {\bf Case 2:\ }\ $B_{i_1,\,j_0}\le c_{i_1}$. Hence, in this case $0<c_{i_1}$ and there exists $0\leq r\leq d_{i_1}-1$ such that $$ \psi^{\rm op}_{i_1}=(R_{i_1}(B)^\downarrow_1\, , \, \ldots \, , \, R_{i_1}(B)^\downarrow_r \, , \, c_{i_1}\, , \, \ldots\, , \, c_{i_1}) $$ so that there exists $\varepsilon>0$ such that for $t\in[0,\varepsilon]$ we have that $$ \psi_{i_1}(t)=(R_{i_1}(B)^\downarrow_1\, , \, \ldots \, , \, R_{i_1}(B)^\downarrow_r \, , \, c_{i_1}\, , \, \ldots\, , \, c_{i_1})-\frac{p_{i_0}\ t}{(d-r)\ p_{i_1}}\sum_{j=r+1}^{d_1} e_j\,.
$$ Therefore, for $t\in [0,\epsilon]$ we have that $$f(t)=\varphi_p(\Psi(t))- \varphi_p(\Psi^{\rm op})= p_{i_0}\ (\varphi(t)-\varphi(0))+p_{i_1} \ (d-r)\ (\varphi(c_{i_1}- \frac{p_{i_0} \, t}{(d-r)\,p_{i_1} } ) - \varphi(c_{i_1}))\, .$$ As before, $f(0)=0$ and $f(t)\geq 0$ for $t\in[0,\epsilon]$; a simple computation shows that in this case we also have that $f'(0)<0$, which contradicts the previous facts; thus, the vectors in $\Psi^{\rm op}$ have no zero entries. \end{proof}
{\scriptsize
}
\end{document}
|
arXiv
|
{
"id": "1512.07122.tex",
"language_detection_score": 0.5759743452072144,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title[Center of gravity of the associahedron]{The centers of gravity of the associahedron and of the permutahedron are the same}
\author[C. Hohlweg]{Christophe~Hohlweg} \address[Christophe Hohlweg]{Universit\'e du Qu\'ebec \`a Montr\'eal\\ LaCIM et D\'epartement de Math\'ematiques\\ CP 8888 Succ. Centre-Ville\\ Montr\'eal, Qu\'ebec, H3C 3P8\\ CANADA} \email{[email protected]} \urladdr{http://www.lacim.uqam.ca/\~{}hohlweg}
\author[J. Lortie]{Jonathan~Lortie} \address[Jonathan Lortie]{Universit\'e du Qu\'ebec \`a Montr\'eal\\ LaCIM et D\'epartement de Math\'ematiques\\ CP 8888 Succ. Centre-Ville\\ Montr\'eal, Qu\'ebec, H3C 3P8\\ CANADA} \email{[email protected]}
\author[A. Raymond]{Annie~Raymond} \address[Annie Raymond]{Berlin Mathematical School\\ Strasse des 17. Juni 136\\ Berlin, 10623, Germany} \email{[email protected]}
\date{\today}
\thanks{$^*$ This work is supported by FQRNT and NSERC. It is the result of a summer undergraduate research internship supported by LaCIM}
\begin{abstract} \noindent In this article, we show that Loday's realization of the associahedron has the the same center of gravity than the permutahedron. This proves an observation made by F.~Chapoton.
We also prove that this result holds for the associahedron and the cyclohedron as realized by the first author and C.~Lange.
\end{abstract}
\maketitle
\section{Introduction.}\label{se:Intro}
In 1963, J.~Stasheff discovered the associahedron~\cite{stasheff,stasheff2}, a polytope of great importance in algebraic topology.
The associahedron in $\mathbb R^n$ is a simple $n-1$-dimensional convex polytope. The classical realization of the associahedron given by
S.~Shnider and S.~Sternberg in \cite{shnider_sternberg} was
completed by J.~L.~Loday in 2004~\cite{loday}. Loday gave a
combinatorial algorithm to compute the integer coordinates of the
vertices of the associahedron, and showed that it can be obtained
naturally from the classical permutahedron of dimension $n-1$.
F.~Chapoton observed that the centers of gravity of the
associahedron and of the permutahedron are the same \cite[Section 2.11]{loday}.
As far as we know, this property of Loday's realization has never been proved.
In 2007, the first author and C.~Lange gave a family of realizations of the associahedron that contains the classical realization of the associahedron. Each of these realizations is also obtained naturally from the classical permutahedron \cite{realisation1}. They conjectured that for any of these realizations, the center of gravity coincide with the center of gravity of the permutahedron. In this article, we prove this conjecture to be true.
The associahedron fits in a larger family of polytopes, {\em generalized associahedra}, introduced by S.~Fomin and A.~Zelevinsky in \cite{fomin_zelevinsky} within the framework of cluster algebras (see \cite{chapoton_fomin_zelevinsky,realisation2} for their realizations).
In 1994, R.~Bott and C.~Taubes discovered the cyclohedron~\cite{bott_taubes} in connection with knot theory. It was rediscovered independently by R. Simion \cite{simion}. In \cite{realisation1}, the first author and C.~Lange also gave a family of realizations for the cyclohedron, starting with the permutahedron of type $B$.
We also show that the centers of gravity of the cyclohedron and of the permutahedron of type $B$ are the same.
The article is organized as follows. In \S\ref{se:1}, we first recall the realization of the permutahedron and how to compute its center of gravity. Then we compute the center of gravity of Loday's realization of the associahedron. In order to do this, we partition its vertices into isometry classes of triangulations, which parameterize the vertices, and we show that the center of gravity for each of those classes is the center of gravity of the
permutahedron.
In \S\ref{se:2}, we show that the computation of the center of
gravity of any of the realizations given by the first author and
C.~Lange is reduced to the computation of the center of gravity of the classical
realization of the associahedron. We do the same for the cyclohedron in \S\ref{se:3}.
We are grateful to Carsten Lange for allowing us to use some of the pictures he made in~\cite{realisation1}.
\section{Center of gravity of the classical permutahedron and associahedron}\label{se:1}
\subsection{The permutahedron} Let $S_n$ be the symmetric group acting on the set $[n]=\{1,2,\dots,n\}$. The {\em permutahedron} ${\mathsf{Perm}}(S_n)$ is the classical $n-1$-dimensional simple convex polytope defined as the convex hull of the points $$ M(\sigma)=(\sigma(1),\sigma(2),\dots, \sigma (n))\in\mathbb R^n,\qquad \forall \sigma\in S_n. $$ The {\em center of gravity} (or {\em isobarycenter}) is the unique point $G$ of $\mathbb R^n$ such that $$ \sum_{\sigma\in S_n} \vect{GM(\sigma)}=\vect 0. $$ Since the permutation $w_0:i\mapsto n+1-i$ preserves ${\mathsf{Perm}}(S_n)$, we see, by sending $M(\sigma)$ to $$ M(w_0\sigma)=(n+1-\sigma(1),n+1-\sigma(2),\dots, n+1-\sigma (n)), $$ that the center of gravity is $ G=(\frac{n+1}{2},\frac{n+1}{2},\dots,\frac{n+1}{2}). $
\subsection{Loday's realization} We present here the realization of the associahedron given by J.~L.~Loday \cite{loday}. However, instead of using planar binary trees, we use triangulations of a regular polygon to parameterize the vertices of the associahedron (see \cite[Remark 1.2]{realisation1}).
\subsubsection{Triangulations of a regular polygon} Let $P$ be a regular $(n+2)$-gon in the Euclidean plane with vertices $A_0,A_1,\dots,A_{n+1}$ in counterclockwise direction. A {\em triangulation of $P$} is a set of $n$ noncrossing diagonals of $P$.
Let us be more explicit. A {\em triangle of $P$} is a triangle whose vertices are vertices of $P$. Therefore a side of a triangle of $P$ is either an edge or a diagonal of $P$. A triangulation of $P$ is then a collection of $n$ distinct triangles of $P$ with noncrossing sides. Any of the triangles in $T$ can be described as $A_i A_j A_k$ with $0\leq i<j<k\leq n+1$. Each $1\leq j\leq n$ corresponds to a unique triangle $\Delta_j(T)$ in $T$ because the sides of triangles in $T$ are noncrossing.
Therefore we write $T=\{\Delta_1(T),\dots, \Delta_n(T)\}$ for a triangulation $T$, where $\Delta_j(T)$ is the unique triangle in $T$ with vertex $A_j$ and the two other vertices $A_i$ and $A_k$ satisfying the inequation $0\leq i<j<k\leq n+1$.
Denote by ${\mathcal T}_{n+2}$ the set of triangulations of $P$.
\subsubsection{Loday's realization of the associahedron}
Let $T$ be a triangulation of $P$. The {\em weight} $\delta_j(T)$ of the triangle $\Delta_j(T)=A_i A_jA_k$, where $i<j<k$, is the positive number $$ \delta_j(T)=(j-i)(k-j). $$ The weight $\delta_j(T)$ of $\Delta_j(T)$ represents the product of the number of boundary edges of $P$ between $A_i$ and $A_j$ passing through vertices indexed by smaller numbers than $j$ with the number of boundary edges of $P$ between $A_j$ and $A_k$ passing through vertices indexed by larger numbers than $j$.
The {\em classical associahedron} ${\mathsf{Asso}}(S_n)$ is obtained as the convex hull of the points $$ M(T)=(\delta_1(T),\delta_2(T),\dots, \delta_n(T))\in \mathbb R^n,\quad\forall T\in{\mathcal T}_{n+2}. $$ We are now able to state our first result.
\begin{thm}\label{thm:Main} The center of gravity of ${\mathsf{Asso}}(S_n)$ is $G=(\frac{n+1}{2},\frac{n+1}{2},\dots,\frac{n+1}{2})$.
\end{thm}
In order to prove this theorem, we need to study closely a certain partition of the vertices of $P$.
\subsection{Isometry classes of triangulations}\label{se:centergravity} As $P$ is a regular $(n+2)$-gon, its isometry group is the dihedral group ${\mathcal D}_{n+2}$ of order $2(n+2)$. So ${\mathcal D}_{n+2}$ acts on the set ${\mathcal T}_{n+2}$ of all triangulations of $P$: for $f\in{\mathcal D}_{n+2}$ and $T\in{\mathcal T}_{n+2}$, we have $f\cdot T\in{\mathcal T}_{n+2}$. We denote by $\mathcal O (T)$ the orbit of $T\in{\mathcal T}_{n+2}$ under the action of ${\mathcal D}_{n+2}$.
We know that $G$ is the center of gravity of ${\mathsf{Asso}}(S_n)$ if and only if $$ \sum_{T\in{\mathcal T}_{n+2}} \vect{GM(T)} =\vect 0. $$ As the orbits of the action of ${\mathcal D}_{n+2}$ on ${\mathcal T}_{n+2}$ form a partition of the set ${\mathcal T}_{n+2}$, it is sufficient to compute $$ \sum_{T\in\mathcal O} \vect{GM(T)} $$ for any orbit $\mathcal O$. The following key observation implies directly Theorem~\ref{thm:Main}.
\begin{thm}\label{thm:key} Let $\mathcal O$ be an orbit of the action of ${\mathcal D}_{n+2}$ on ${\mathcal T}_{n+2}$, then $G$ is the center of gravity of $\{M(T)\,|\, T\in\mathcal O\}$. In particular, $ \sum_{T\in\mathcal O} \vect{GM(T)}=\vect 0. $ \end{thm}
Before proving this theorem, we need to prove the following result.
\begin{prop}\label{prop:canonique} Let $T\in{\mathcal T}_{n+2}$ and $j\in [n]$, then $\displaystyle{\sum_{f\in {\mathcal D}_{n+2}} \delta_j(f\cdot T) = (n+1)(n+2)}$. \end{prop}
\begin{proof}
We prove this proposition by induction on $j\in [n]$. For any triangulation $T'$, we denote by
$a_j(T')<j<b_j(T')$ the indices of the vertices of $\Delta_j(T')$. Let $H$ be the group of rotations
in ${\mathcal D}_{n+2}$. It is well-known that for any reflection $s\in {\mathcal D}_{n+2}$, the classes $H$ and
$sH$ form a partition of ${\mathcal D}_{n+2}$ and that $|H|=n+2$. We consider also the unique reflection $s_k\in{\mathcal D}_{n+2}$ which maps $A_x$ to $A_{n+3+k-x}$, where the values of the indices are taken in modulo $n+2$. In particular, $s_k(A_0)=A_{n+3+k}=A_{k+1}$, $s_k(A_1)=A_k$, $s_k(A_{k+1})=A_{n+2}=A_0$, and so on.
\noindent {\bf Basic step $j=1$:} We know that $a_1(T')=0$ for any triangulation $T'$, hence the weight of $\Delta_1(T')$ is $\delta_1(T')=(1-0)(b_1(T')-1)=b_1(T')-1$.
The reflection $s_0\in {\mathcal D}_{n+2}$ maps $A_x$ to $A_{n+3-x}$ (where $A_{n+2}=A_0$ and $A_{n+3}=A_1$). In other words, $s_0(A_0)=A_1$ and $s_0(\Delta_1(T'))$ is a triangle in $s_0\cdot T'$. Since $$ s_0(\Delta_1( T'))= s_0(A_0A_1A_{b_1(T')})= A_0A_1A_{n+3-b_1(T')} $$ and $0<1<n+3-b_1(T')$, $s_0(\Delta_1(T'))$ has to be $\Delta_1(s_0\cdot T')$. In consequence, we obtain that $$ \delta_1(T')+\delta_1(s_0\cdot T')= (b_1(T')-1)+(n+3-b_1(T')-1)=n+1, $$ for any triangulation $T'$. Therefore $$ \sum_{f\in {\mathcal D}_{n+2}} \delta_1(f\cdot T) = \sum_{g\in H}\big(
(\delta_1(g\cdot T)+\delta_1(s_0\cdot (g\cdot T))\big)= |H| (n+1)=(n+1)(n+2), $$ proving the initial case of the induction.
\noindent {\bf Inductive step:} Assume that, for a given $1\leq j<n$, we have $$ \sum_{f\in {\mathcal D}_{n+2}}\delta_j(f\cdot T) = (n+1)(n+2). $$ We will show that $$ \sum_{f\in {\mathcal D}_{n+2}}\delta_{j+1}(f\cdot T) = \sum_{f\in {\mathcal D}_{n+2}}\delta_j(f\cdot T). $$ Let $r\in H\subseteq {\mathcal D}_{n+2}$ be the unique rotation mapping $A_{j+1}$ to $A_{j}$. In particular, $r(A_0)=A_{n+1}$. Let $T'$ be a triangulation of $P$. We have two cases:
\noindent {\bf Case 1.} If $a_{j+1}(T')>0$ then $a_{j+1}(T')-1<j<b_{j+1}(T')-1$ are the indices of the vertices of the triangle $r(\Delta_{j+1}(T'))$ in $r\cdot T'$. Therefore, by unicity, $r(\Delta_{j+1}(T'))$ must be $\Delta_j(r\cdot T')$. Thus
\begin{eqnarray*}
\delta_{j+1}(T')&=&(b_{j+1}(T')-(j+1))(j+1-a_{j+1}(T'))\\
&=&\big((b_{j+1}(T')-1)-j\big)(j-(a_{i+1}(T')-1))\\
&=&\delta_j(r\cdot T'). \end{eqnarray*} In other words: \begin{eqnarray}\label{equ:1} \sum_{{f\in {\mathcal D}_{n+2},\atop a_{j+1}(f\cdot T)\not = 0}}\delta_{j+1}(f\cdot T) & =& \sum_{{f\in {\mathcal D}_{n+2},\atop a_{j+1}(f\cdot T)\not = 0}}\delta_j(r\cdot(f\cdot T))\\\nonumber & =& \sum_{{g\in {\mathcal D}_{n+2},\atop b_{j}(g\cdot T)\not = n+1}}\delta_j(g\cdot T). \end{eqnarray}
\noindent {\bf Case 2.} If $a_{j+1}(T')=0$, then $j<b_{j+1}(T')-1<n+1$ are the indices of the vertices of $r(\Delta_{j+1}(T'))$, which is therefore not $\Delta_j(r\cdot T')$: it is $\Delta_{b_{j+1}(T')-1}(r\cdot T')$. To handle this, we need to use the reflections $s_j$ and $s_{j-2}$.
On one hand, observe that $j+1<n+3+j-b_{j+1}(T')$ because $b_{j+1}(T')<n+1$.
Therefore $$ s_j(\Delta_{j+1}(T'))=A_{j+1}A_0 A_{n+3+j-b_{j+1}(T')}=\Delta_{j+1}(s_j\cdot T'). $$ Hence \begin{eqnarray*} \delta_{j+1}(T')+\delta_{j+1}(s_j\cdot T')&=&(j+1)(b_{j+1}(T')-(j+1))\\ &&+(j+1)(n+3+j-b_{j+1}(T')-(j+1))\\ &=&(j+1)(n+1-j). \end{eqnarray*}
On the other hand, consider the triangle $\Delta_j(r\cdot T')$ in $r\cdot T'$. Since $$ r(\Delta_{j+1}(T'))=A_{j}A_{b_{j+1}(T')-1}A_{n+1}=\Delta_{b_{j+1}(T')-1}(r\cdot T') $$ is in $r\cdot T'$, $[j,n+1]$ is a diagonal in $r\cdot T'$. Hence $b_j(r\cdot T')=n+1$. Thus $\Delta_j(r\cdot T')=A_{a_j(r\cdot T')}A_j A_{n+1}$ and $\delta_j(r\cdot T')=(j-a_j(r\cdot T'))(n+1-j)$. We have $s_{j-2}(A_j)=A_{n+1}$, $s_{j-2}(A_{n+2})=A_j$ and $s_{j-2}(A_{a_j(r\cdot T')})=A_{n+1+j-a_j(r\cdot T')}=A_{j-a_j(r\cdot T')-1}$ since $a_j(r\cdot T')<j$. Therefore $s_{j-2}(\Delta_j(r\cdot T'))=A_{j-a_j(r\cdot T')-1}A_jA_{n+1}=\Delta_j(s_{j-2}r\cdot T')$ and $\delta_j(s_{j-2}r\cdot T')=(a_j(r\cdot T')+1)(n+1-j)$. Finally we obtain that
\begin{eqnarray*}
\delta_{j}(r\cdot T')+\delta_{j}(s_{j-2}r\cdot T')&=&(j-a_j(r\cdot T'))(n+1-j)+(a_j(r\cdot T')+1)(n+1-j)\\ &=&(j+1)(n+1-j).
\end{eqnarray*}
Since $\{H,s_k H\}$ forms a partition of ${\mathcal D}_{n+2}$ for any $k$, we have \begin{eqnarray}\label{equ:2}
\sum_{{f\in {\mathcal D}_{n+2},\atop a_{j+1}(f\cdot T)=0}}\delta_{j+1}(f\cdot T) & =& \sum_{{f\in H,\atop a_{j+1}(f\cdot T)=0}}\big(\delta_{j+1}(f\cdot T) +\delta_{j+1}(s_j f\cdot T)\big)\\ \nonumber
&=& \sum_{{f\in H,\atop a_{j+1}(f\cdot T)=0}} (j+1)(n+1-j)\\ \nonumber
&=& \sum_{{rf\in H,\atop b_{j}(rf\cdot T)=n+1}}\big(\delta_{j}(rf\cdot T) +\delta_{j}(s_{j-2} rf\cdot T)\big),\ \textrm{since }r\in H\\ \nonumber
&=& \sum_{{g\in H,\atop b_{j}(g\cdot T)=n+1}}\delta_{j}(g\cdot T). \end{eqnarray}
\noindent We conclude the induction by adding Equations~(\ref{equ:1}) and (\ref{equ:2}). \end{proof}
\begin{proof}[Proof of Theorem~\ref{thm:key}] We have to prove that $$ \vect u=\sum_{T'\in\mathcal O(T)} \vect{GM(T')}=\vect 0. $$
Denote by ${\textnormal{Stab}}(T')=\{f\in{\mathcal D}_{n+2}\,|\, f\cdot T'=T'\}$ the stabilizer of $T'$, then $$
\sum_{f\in {\mathcal D}_{n+2}} M(f\cdot T) = \sum_{T'\in \mathcal O(T)} |{\textnormal{Stab}}(T')| M(T'). $$
Since $T'\in\mathcal O(T)$, $|{\textnormal{Stab}}(T')|=|{\textnormal{Stab}}(T)|=\frac{2(n+2)}{|\mathcal O(T)|}$, we have $$
\sum_{f\in {\mathcal D}_{n+2}} M(f\cdot T) = \frac{2(n+2)}{|\mathcal O(T)|} \sum_{T'\in \mathcal O(T)} M(T') . $$ Therefore by Proposition~\ref{prop:canonique} we have for any $i\in [n]$ \begin{equation}\label{equ:3}
\sum_{T'\in \mathcal O(T)} \delta_i(T')= \frac{|\mathcal O(T)|}{2(n+2)}(n+1)(n+2)=\frac{|\mathcal O(T)|(n+1)}{2}. \end{equation}
Denote by $O$ the point of origin of $\mathbb R^n$. Then $\vect{OM}=M$ for any point $M$ of $\mathbb R^n$. By Chasles' relation we have finally
$$
\vect u=\sum_{T'\in\mathcal O(T)} \vect{GM(T')}= \sum_{T'\in\mathcal O(T)} (M(T')-G) =\sum_{T'\in\mathcal O(T)} M(T') - |\mathcal O(T)| G. $$
So the $i^{th}$ coordinate of $\vect u$ is $ \sum_{T'\in \mathcal O(T)} \delta_i(T')- \frac{|\mathcal O(T)|(n+1)}{2}=0 $, hence $\vect u =\vect 0$ by (\ref{equ:3}). \end{proof}
\section{Center of gravity of generalized associahedra of type $A$ and $B$}\label{se:2}
\subsection{Realizations of associahedra} As a Coxeter group (of type $A$), $S_n$ is generated by the simple transpositions $\tau_i=(i,\, i+1)$, $i\in [n-1]$. The Coxeter graph $\Gamma_{n-1}$ is then
Let ${\mathscr A}$ be an orientation of $\Gamma_{n-1}$. We distinguish between {\em up} and {\em down} elements of $[n]$~: an element $i\in [n]$ is {\em up} if the edge $\{\tau_{i-1}, \tau_i\}$ is directed from $\tau_i$ to $\tau_{i-1}$ and {\em down} otherwise (we set $1$ and $n$ to be down). Let ${\mathsf D}_{\mathscr A}$ be the set of down elements and let ${\mathsf U}_{\mathscr A}$ be the set of up elements (possibly empty).
The notion of up and down induces a labeling of the $(n+2)$-gon $P$ as follows. Label $A_0$ by $0$. Then the vertices of $P$ are, in counterclockwise direction, labeled by the down elements in increasing order, then by $n+1$, and finally by the up elements in decreasing order. An example is given in Figure~\ref{fig:example_labelling}. \begin{figure}\label{fig:example_labelling}
\end{figure}
We recall here a construction due to Hohlweg and Lange~\cite{realisation1}. Consider $P$ labeled according to a fixed orientation~${\mathscr A}$ of $\Gamma_{n-1}$. For each $l\in [n]$ and any triangulation $T$ of $P$, there is a unique triangle $\Delta^{\mathscr A}_l(T)$ whose vertices are labeled by $k<l<m$. Now, count the number of edges of $P$ between $i$ and $k$, whose vertices are labeled by smaller numbers than $l$. Then multiply it by the number of edges of $P$ between $l$ and $m$, whose vertices are labeled by greater numbers than $l$. The result $\omega_l^{\mathscr A}(T)$ is called the {\em weight} of $\Delta_l^{\mathscr A}(T)$. The injective map \begin{align*}
M_{{\mathscr A}}: {\mathcal T}_{n+2} &\longrightarrow {\mathbb R}^n \\
T &\longmapsto (x^{\mathscr A}_1(T),x^{\mathscr A}_2(T),\dots,x^{\mathscr A}_n(T)) \end{align*} that assigns explicit coordinates to a triangulation is defined as follows: \[
x^{\mathscr A}_j(T) := \begin{cases}
\omega_j^{\mathscr A} (T) & \textrm{if } j\in{\mathsf D}_{\mathscr A}\\
n+1-\omega_j^{\mathscr A}(T) & \textrm{if } j\in{\mathsf U}_{\mathscr A}.
\end{cases} \]
Hohlweg and Lange showed that the convex hull ${\mathsf{Asso}}_{\mathscr A}(S_n)$
of~$\{M_{{\mathscr A}}(T)\,|\,T\in {\mathcal T}_{n+2}\}$ is a realization of the associahedron with integer coordinates \cite[Theorem 1.1]{realisation1}. Observe that if the orientation ${\mathscr A}$ is {\em canonic}, that is, if ${\mathsf U}_{\mathscr A}=\emptyset$, then ${\mathsf{Asso}}_{\mathscr A}(S_n)={\mathsf{Asso}}(S_n)$.
The key is now to observe that the weight of $\Delta_{j}^{\mathscr A}(T)$ in $T$ is precisely the weight of $\Delta_j (T')$ where $T'$ is a triangulation in the orbit of $T$ under the action of ${\mathcal D}_{n+2}$, as stated in the next proposition.
\begin{prop}\label{prop:weight} Let ${\mathscr A}$ be an orientation of $\Gamma_{n-1}$. Let $j\in [n]$ and let $A_l$ be the vertex of $P$ labeled by $j$. There is an isometry $r_j^{\mathscr A}\in \mathcal D_{n+2}$ such that: \begin{enumerate} \item[(i)] $r_j^{\mathscr A}(A_l)=A_j$;
\item[(ii)] the label of the vertex $A_k$ is smaller than $j$ if and only if the index $i$ of the vertex $A_i=r_j^{\mathscr A}(A_k)$ is smaller than $j$.
\end{enumerate} Moreover, for any triangulation $T$ of $P$ we have $\omega_j^{\mathscr A}(T)=\delta_j(r_j^{\mathscr A}\cdot T).$ \end{prop} \begin{proof} If ${\mathscr A}$ is the canonical orientation, then $r_j^{\mathscr A}$ is the identity, and the proposition is straightforward. In the following proof, we suppose therefore that ${\mathsf U}_{\mathscr A}\not=\emptyset$.
\noindent Case 1: Assume that $j\in{\mathsf D}_{\mathscr A}$. Let $\alpha$ be the greatest up element smaller than $j$ and let $A_{\alpha+1}$ be the vertex of $P$ labeled by $\alpha$. Then by construction of the labeling, $A_{\alpha}$ is labeled by a larger number than $j$, and $[A_{\alpha},A_{\alpha+1}]$ is the
unique edge of $P$ such that $A_{\alpha+1}$ is labeled by a smaller number than $j$. Denote by $\Lambda_{\mathscr A}$ the path from $A_l$ to $A_{\alpha+1}$ passing through vertices of $P$ labeled by smaller numbers than $j$. This is the path going from $A_l$ to $A_{\alpha+1}$ in clockwise direction on the boundary of $P$.
By construction, $A_k\in \Lambda_{\mathscr A}$ if and only if the label of $A_k$ is smaller than $j$. In other words, the path $\Lambda_{\mathscr A}$ consists of {\em all} vertices of $P$ labeled by smaller numbers than $j$. Therefore the cardinality of $\Lambda_{\mathscr A}$ is $j+1$.
Consider $r_j^{\mathscr A}$ to be the rotation mapping $A_l$ to $A_j$. Recall that a rotation is an isometry preserving the orientation of the plane. Then the path $\Lambda_{\mathscr A}$, which is obtained by walking on the boundary of $P$ from $A_l$ to $A_{\alpha+1}$ in clockwise direction, is sent to the path $\Lambda$ obtained by walking on the boundary of $P$ in clockwise direction from $A_j$ and going through $j+1=|\Lambda_{\mathscr A}|$ vertices of $P$. Therefore $\Lambda=\{A_0,A_1,\dots, A_j\}$, thus proving the first claim of our proposition in this case.
\noindent Case 2: assume that $j\in {\mathsf U}_{\mathscr A}$. The proof is almost the same as in the case of a down element. Let $\alpha$ be the greatest down element smaller than $j$ and let $A_{\alpha}$ be the vertex of $P$ labeled by $\alpha$. Then by construction of the labeling, $A_{\alpha+1}$ is labeled by a larger number than $j$, and $[A_{\alpha},A_{\alpha+1}]$ is the unique edge of $P$ such that $A_{\alpha}$ is labeled by a smaller number than $j$. Denote by $\Lambda_{\mathscr A}$ the path from $A_l$ to $A_{\alpha}$ passing through vertices of $P$ labeled by smaller numbers than $j$. This is the path going from $A_{\alpha}$ to $A_l$
in clockwise direction on the boundary of $P$.
As above, $A_k\in \Lambda_{\mathscr A}$ if and only if the label of $A_k$ is smaller than $j$. In other words, the path $\Lambda_{\mathscr A}$ consists of all the vertices of $P$ labeled by smaller numbers than $j$.
Therefore, again, the cardinality of $\Lambda_{\mathscr A}$ is $j+1$.
Let $r_j^{\mathscr A}$ be the reflection mapping $A_\alpha$ to $A_0$ and $A_{\alpha+1}$ to $A_{n+1}$. Recall that a reflection is an isometry reversing the orientation of the plane. Then the path $\Lambda_{\mathscr A}$, which is obtained by walking on the boundary of $P$ from $A_\alpha$ to $A_{l}$ in clockwise direction, is sent to the path $\Lambda$ obtained by walking on the boundary of $P$ in clockwise direction from $A_\alpha$ and going through
$j+1=|\Lambda_{\mathscr A}|$ vertices of $P$. Therefore $\Lambda=\{A_0,A_1,\dots, A_j\}$. Hence $r_j^{\mathscr A}(A_l)$ is sent on the final vertex of the path $\Lambda$ which is $A_j$, proving the first claim of our proposition.
Thus it remains to show that for a triangulation $T$ of $P$ we have $\omega_j^{\mathscr A}(T)=\delta_j(r_j^{\mathscr A}\cdot T).$ We know that $\Delta_j^{\mathscr A}(T)=A_k A_l A_m$ such that the label of $A_k$ is smaller than $j$, which is smaller than the label of $A_m$. Write
$A_a=r_j^{\mathscr A}(A_k)$ and $A_b=r_j^{\mathscr A}(A_m)$. Because of Proposition~\ref{prop:weight}, $a<j<b$ and therefore $$ r_j^{\mathscr A}(\Delta_j^{\mathscr A}(T))= A_a A_jA_b=\Delta_j(r_j^{\mathscr A}\cdot T). $$ So $(j-a)$ is the number of edges of $P$ between $A_l$ and $A_k$, whose vertices are labeled by smaller numbers than $j$. Similarly, $(b-j)$ is the number of edges between $A_l$ and $A_m$, whose vertices are labeled by smaller numbers than $j$, and $(b-j)$ is the number of edges of $P$ between $A_l$ and $A_m$ and whose vertices are labeled by larger numbers than $j$. So $\omega_l^{\mathscr A}(T)=(j-a)(b-j)=\delta_j(r_j^{\mathscr A}\cdot T)$. \end{proof}
\begin{cor}\label{cor:Canon} For any orientation ${\mathscr A}$ of the Coxeter graph of $S_n$ and for any $j\in [n]$, we have $$ \sum_{f\in {\mathcal D}_{n+2}} x^{\mathscr A}_j(f\cdot T) = (n+1)(n+2). $$ \end{cor} \begin{proof} Let $r_j^{\mathscr A}\in \mathcal D_{n+2}$ be as in Proposition~\ref{prop:weight}.
Suppose first that $j\in {\mathsf U}_{\mathscr A}$, then \begin{eqnarray*} \sum_{f\in {\mathcal D}_{n+2}} x^{\mathscr A}_i(f\cdot T) &=&2(n+2)(n+1)-\sum_{f\in {\mathcal D}_{n+2}} \omega_i^{\mathscr A}(f\cdot T)\\ &=&2(n+2)(n+1)-\sum_{f\in {\mathcal D}_{n+2}} \delta_j(fr_j^{\mathscr A}\cdot T),\ \textrm{by Proposition~\ref{prop:weight}} \\ &=&2(n+2)(n+1)-\sum_{g\in {\mathcal D}_{n+2}} \delta_j(g^{\mathscr A}\cdot T),\ \textrm{since $r_j^{\mathscr A}\in\mathcal D_{n+2}$} \\ &=& (n+1)(n+2),\ \textrm{by Proposition~\ref{prop:canonique}} \end{eqnarray*}
If $i\in {\mathsf D}_{\mathscr A}$, the result follows from a similar calculation. \end{proof}
\subsection{Center of gravity of associahedra}
\begin{thm}\label{thm:Main2} The center of gravity of ${\mathsf{Asso}}_{\mathscr A}(S_n)$ is $G=(\frac{n+1}{2},\frac{n+1}{2},\dots,\frac{n+1}{2})$ for any orientation ${\mathscr A}$.
\end{thm}
By following precisely the same arguments as in \S\ref{se:centergravity}, we just have to show the following generalization of Theorem~\ref{thm:key}.
\begin{thm}\label{thm:keyGenAss} Let $\mathcal O$ be an orbit of the action of ${\mathcal D}_{n+2}$ on ${\mathcal T}_{n+2}$, then $G$ is the center of gravity of $\{M_{\mathscr A}(T)\,|\, T\in\mathcal O\}$. In particular, $\sum_{T\in\mathcal O} \vect{GM_{\mathscr A}(T)}=\vect 0. $ \end{thm} \begin{proof} The proof is entirely similar to the proof of Theorem~\ref{thm:key}, using Corollary~\ref{cor:Canon} instead of Proposition~\ref{prop:canonique}. \end{proof}
\section{Center of gravity of the cyclohedron}\label{se:3}
\subsection{The type $B$-permutahedron}
The hyperoctahedral group $W_n$ is defined by $W_n=\{\sigma\in S_{2n}\,|\, \sigma(i)+\sigma(2n+1-i)=2n+1,\ \forall i\in[n]\}$. The {\em type $B$-permutahedron} ${\mathsf{Perm}}(W_n)$ is the simple $n$-dimensional convex polytope defined as the convex hull of the points $$ M(\sigma)=(\sigma(1),\sigma(2),\dots, \sigma (n))\in\mathbb R^{2n},\qquad \forall \sigma\in W_n. $$ As $w_0=(2n,2n-1,\dots,3,2,1)\in W_n$, we deduce from the same argument as in the case of ${\mathsf{Perm}}(S_n)$ that the center of gravity of ${\mathsf{Perm}}(W_n)$ is $$G=(\frac{2n+1}{2},\frac{2n+1}{2},\dots,\frac{2n+1}{2}).$$
\subsection{Realizations of the associahedron} An orientation~${\mathscr A}$ of~$\Gamma_{2n-1}$ is {\em symmetric} if the edges $\{\tau_i,\tau_{i+1}\}$ and $\{\tau_{2n-i-1},\tau_{2n-i}\}$ are oriented in \emph{opposite directions} for all~$i\in [2n-2]$. There is a bijection between symmetric orientations of~$\Gamma_{2n-1}$ and orientations of the Coxeter graph of $W_n$ (see \cite[\S1.2]{realisation1}). A triangulation $T\in {\mathcal T}_{2n+2}$ is {\em centrally symmetric} if~$T$, viewed as a triangulation of $P$, is centrally symmetric. Let ${\mathcal T}_{2n+2}^B$ be the set of the centrally symmetric triangulations of $P$. In \cite[Theorem 1.5]{realisation1} the authors show that for any symmetric orientation ${\mathscr A}$ of $\Gamma_{2n-1}$. The convex hull
${\mathsf{Asso}}_{\mathscr A}(W_{n})$ of $\{M_{{\mathscr A}}(T)\,|\,T\in {\mathcal T}^B_{2n+2}\}$ is a realization of the cyclohedron with integer coordinates.
Since the full orbit of symmetric triangulations under the action of ${\mathcal D}_{2n+2}$ on triangulations provides vertices of ${\mathsf{Asso}}_{\mathscr A}(W_{n})$, and vice-versa, Theorem~\ref{thm:keyGenAss} implies the following corollary.
\begin{cor}\label{cor:Main} Let ${\mathscr A}$ be a symmetric orientation of $\Gamma_{2n-1}$, then the center of gravity of ${\mathsf{Asso}}_{\mathscr A}(W_n)$ is $G=(\frac{2n+1}{2},\frac{2n+1}{2},\dots,\frac{2n+1}{2})$.
\end{cor}
\end{document}
|
arXiv
|
{
"id": "0910.0265.tex",
"language_detection_score": 0.7485657334327698,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{{f Critical ($P_5$,bull)-free graphs}
\begin{abstract}
Given two graphs $H_1$ and $H_2$, a graph is $(H_1,H_2)$-free if it contains no induced subgraph isomorphic to $H_1$ or $H_2$. Let $P_t$ and $C_t$ be the path and the cycle on $t$ vertices, respectively. A bull is the graph obtained from a triangle with two disjoint pendant edges. In this paper, we show that there are finitely many 5-vertex-critical ($P_5$,bull)-free graphs.
{\bf Keywords.} coloring; critical graphs; forbidden induced subgraphs; strong perfect graph theorem; polynomial-time algorithms.
\end{abstract}
\section{Introduction}
All graphs in this paper are finite and simple. We say that a graph $G$ {\em contains} a graph $H$ if $ H $ is isomorphic to an induced subgraph of $G$. A graph $G$ is {\em H-free} if it does not contain $H$. For a family of graphs $\mathcal{H}$, $G$ is {\em $\mathcal{H}$-free} if $G$ is $H$-free for every $H\in \mathcal{H}$. When $\mathcal{H}$ consists of two graphs, we write $(H_1,H_2)$-free instead of $\{H_1,H_2\}$-free.
A $k$-{\em coloring} of a graph $G$ is a function $\phi:V(G)\rightarrow\{1,...,k\}$ such that $\phi(u)\neq\phi(v)$ whenever $u$ and $v$ are adjacent in $G$. Equivalently, a $k$-coloring of $G$ is a partition of $V(G)$ into $k$ independent sets. We call a graph $k$-{\em colorable} if it admits a $k$-coloring. The {\em chromatic number} of $G$, denoted by $\chi(G)$, is the minimum number $k$ for which $G$ is $k$-colorable. The {\em clique number} of $G$, denoted by $\omega(G)$, is the size of a largest clique in $G$.
A graph $G$ is said to be $k$-{\em chromatic} if $\chi(G)=k$. We say that $G$ is {\em critical} if $\chi(H)<\chi(G)$ for every proper subgraph $H$ of $G$. A $k$-{\em critical} graph is one that is $k$-chromatic and critical. An easy consequence of the definition is that every critical graph is connected. Critical graphs were first investigated by Dirac \cite{Di51,Di52,Di52i} in 1951, and then by Lattanzio and Jensen \cite{L02,J02} among others, and by Goedgebeur \cite{GS18} in recent years.
Vertex-criticality is a weaker notion. Suppose that $G$ is a graph. Then $G$ is said to be $k$-{\em vertex-critical} if $G$ has chromatic number $k$ and removing any vertex from $G$ results in a graph that is $(k-1)$-colorable. For a set $\mathcal{H}$ of graphs, we say that $ G $ is {\em k-vertex-critical $\mathcal{H}$-free} if it is $k$-vertex-critical and $\mathcal{H}$-free. The following problem arouses our interest.
{\noindent} \textbf{The finiteness problem.} Given a set $\mathcal{H}$ of graphs and an integer $k\ge 1$, are there only finitely many $ k $-vertex-critical $\mathcal{H}$-free graphs?
This problem is meaningful because the finiteness of the set has a fundamental algorithmic implication.
\begin{theorem}[Folklore]\label{Folklore}
If the set of all $k$-vertex-critical $\mathcal{H}$-free graphs is finite, then there is a polynomial-time algorithm to determine whether an $\mathcal{H}$-free graph is $(k-1)$-colorable. \qed
\end{theorem}
Let $K_n$ be the complete graph on $n$ vertices. Let $ P_t $ and $ C_t $ denote the path and the cycle on $t$ vertices, respectively. The {\em complement} of $G$ is denoted by $\overline{G}$. For $s,r\ge 1$, let $K_{r,s}$ be the complete bipartite graph with one part of size $r$ and the other part of size $s$. A class of graphs that has been extensively studied recently is the class of $P_t$-free graphs. In \cite{BHS09}, it was shown that there are finite many 4-vertex-critical $P_5$-free graphs. This result was later generalized to $P_6$-free graphs \cite{CGSZ16}. In the same paper, an infinite family of 4-vertex-critical $P_7$-free graphs was constructed. Moreover, for every $k\ge 5$, an infinite family of $k$-vertex-critical $P_5$-free graphs was constructed in \cite{HMRSV15}. This implies that the finiteness of $k$-vertex-critical $P_t$-free graphs for $t\ge 1$ and $k\ge 4$ has been completely solved by researchers. We summarize the results in the following table.
\begin{table}[!ht]
\centering
\caption{The finiteness of $k$-vertex-critical $P_t$-free graphs.}
\renewcommand\arraystretch{1.5}
\setlength{\tabcolsep}{4mm}{}
\begin{tabular}{|c|p{1.6cm}<{\centering}|p{1.9cm}<{\centering}|p{1.6cm}<{\centering}|p{1.7cm}<{\centering}|}
\hline
\diagbox{$k$}{$t$} & $\le4$ & 5 & 6 & $\ge7$\\
\hline
4 & finite & finite \cite{BHS09}& finite \cite{CGSZ16}& infinite \cite{CGSZ16}\\
\hline
$\ge 5$ & finite & infinite \cite{HMRSV15}& infinite & infinite \\
\hline
\end{tabular}
\end{table}
Because there are infinitely many 5-vertex-critical $P_5$-free graphs, many researchers have investigated the finiteness problem of $k$-vertex-critical $(P_5,H)$-free graphs. Our research is mainly motivated by the following dichotomy result.
\begin{theorem}[\cite{CGHS21}]
Let $H$ be a graph of order 4 and $k\ge 5$ be a fixed integer. Then there are infinitely many k-vertex-critical $(P_5,H)$-free graphs if and only if $H$ is $2P_2$ or $P_1+K_3$.
\end{theorem}
This theorem completely solves the finiteness problem of $k$-vertex-critical $(P_5,H)$-free graphs for graphs of order 4. In \cite{CGHS21}, the authors also posed the natural question of which five-vertex graphs $H$ lead to finitely many $k$-vertex-critical $(P_5,H)$-free graphs. It is known that there are exactly 13 5-vertex-critical $(P_5,C_5)$-free graphs \cite{HMRSV15}, and that there are finitely many 5-vertex-critical ($P_5$,banner)-free graphs \cite{CHLS19,HLS19}, and finitely many $k$-vertex-critical $(P_5,\overline{P_5})$-free graphs for every fixed $k$ \cite{DHHMMP17}. In \cite{CGS}, Cai, Goedgebeur and Huang show that there are finitely many $k$-vertex-critical ($P_5$,gem)-free graphs and finitely many $k$-vertex-critical ($P_5,\overline{P_3+P_2}$)-free graphs. Hell and Huang proved that there are finitely many $k$-vertex-critical $(P_6,C_4)$-free graphs \cite{HH17}. This was later generalized to $(P_5,K_{r,s})$-free graphs in the context of $H$-coloring \cite{KP17}. This gives an affirmative answer for $H=K_{2,3}$.
\noindent {\bf Our contributions.} We continue to study the finiteness of vertex-critical $(P_5,H)$-free graphs when $H$ has order 5. The {\em bull} graph (see \autoref{bull}) is the graph obtained from a triangle with two disjoint pendant edges. In this paper, we prove that there are only finitely many 5-vertex-critical ($P_5$,bull)-free graphs.
\begin{figure}
\caption{The bull graph.}
\label{bull}
\end{figure}
To prove the result on bull-free graphs, we performed a careful structural analysis combined with the pigeonhole principle based on the properties of 5-vertex-critical graphs.
The remainder of the paper is organized as follows. We present some preliminaries in Section \ref{Preliminarlies} and give structural properties around an induced $C_5$ in a ($P_5$,bull)-free graph in Section \ref{structure}. We then show that there are finitely many 5-vertex-critical ($P_5$,bull)-free graphs in Section \ref{bull-free}.
\section{Preliminaries}\label{Preliminarlies}
For general graph theory notation we follow \cite{BM08}. For $k\ge 4$, an induced cycle of length $k$ is called a {\em $k$-hole}. A $k$-hole is an {\em odd hole} (respectively {\em even hole}) if $k$ is odd (respectively even). A {\em $k$-antihole} is the complement of a $k$-hole. Odd and even antiholes are defined analogously.
Let $G=(V,E)$ be a graph. For $S\subseteq V$ and $u\in V\setminus S$, let $d(u,S)={min}_{v\in S}d(u,v)$, where $d(u,v)$ denotes the length of the shortest path from $u$ to $v$. If $uv\in E$, we say that $u$ and $v$ are {\em neighbors} or {\em adjacent}, otherwise $u$ and $v$ are {\em nonneighbors} or {\em nonadjacent}. The {\em neighborhood} of a vertex $v$, denoted by $N_G(v)$, is the set of neighbors of $v$. For a set $X\subseteq V$, let $N_G(X)=\cup_{v\in X}N_G(v)\setminus X$. We shall omit the subscript whenever the context is clear. For $x\in V$ and $S\subseteq V$, we denote by $N_S(x)$ the set of neighbors of $x$ that are in $S$, i.e., $N_S(x)=N_G(x)\cap S$. For two sets $X,S\subseteq V(G)$, let $N_S(X)=\cup_{v\in X}N_S(v)\setminus X$. For $X,Y\subseteq V$, we say that $X$ is {\em complete} (resp. {\em anticomplete}) to $Y$ if every vertex in $X$ is adjacent (resp. nonadjacent) to every vertex in $Y$. If $X=\{x\}$, we write ``$x$ is complete (resp. anticomplete) to $Y$'' instead of ``$\{x\}$ is complete (resp. anticomplete) to $Y$''. If a vertex $v$ is neither complete nor anticomplete to a set $S$, we say that $v$ is {\em mixed} on $S$. For a vertex $v\in V$ and an edge $xy\in E$, if $v$ is mixed on $\{x,y\}$, we say that $v$ is {\em mixed} on $xy$. For a set $H\subseteq V$, if no vertex in $V-H$ is mixed on $H$, we say that $H$ is a {\em homogeneous set}, otherwise $H$ is a {\em nonhomogeneous set}. A vertex subset $S\subseteq V$ is {\em independent} if no two vertices in $S$ are adjacent. A {\em clique} is the complement of an independent set. Two nonadjacent vertices $u$ and $v$ are said to be {\em comparable} if $N(v)\subseteq N(u)$ or $N(u)\subseteq N(v)$. A vertex subset $K\subseteq V$ is a {\em clique cutset} if $G-K$ has more connected components than $G$ and $K$ is a clique. For an induced subgraph $A$ of $G$, we write $G-A$ instead of $G-V(A)$. For $S\subseteq V$, the subgraph \emph{induced} by $S$ is denoted by $G[S]$. For $S\subseteq V$ and an induced subgraph $A$ of $G$, we may write $S$ instead of $G[S]$ and $A$ instead of $V(A)$ for the convenience of writing whenever the context is clear.
We proceed with a few useful results that will be needed later. The first one is well-known in the study of $k$-vertex-critical graphs.
\begin{lemma}[Folklore]\label{lem:xy}
A $k$-vertex-critical graph contains no clique cutsets.
\end{lemma}
Another folklore property of vertex-critical graphs is that such graphs contain no comparable vertices. In \cite{CGHS21}, a generalization of this property was presented.
\begin{lemma}[\cite{CGHS21}]\label{lem:XY}
Let $G$ be a $k$-vertex-critical graph. Then $G$ has no two nonempty disjoint subsets $X$ and $Y$ of $V(G)$ that satisfy all the following conditions.
\begin{itemize}
\item $X$ and $Y$ are anticomplete to each other.
\item $\chi(G[X])\le\chi(G[Y])$.
\item Y is complete to $N(X)$.
\end{itemize}
\end{lemma}
A property on bipartite graphs is shown as follows.
\begin{lemma}[\cite{F93}]\label{2K2}
Let $G$ be a connected bipartite graph. If $G$ contains a $2K_2$, then $G$ must contain a $P_5$.
\end{lemma}
As we mentioned earlier, there are finitely many 4-vertex-critical $P_5$-free graphs.
\begin{theorem}[\cite{BHS09,MM12}]\label{thm:finite4Critical}
If $G=(V,E)$ is a 4-vertex-critical $P_5$-free graph, then $|V|\le 13$.
\end{theorem}
A graph $G$ is {\em perfect} if $\chi(H)=\omega(H)$ for every induced subgraph $H$ of $G$. Another result we use is the well-known Strong Perfect Graph Theorem.
\begin{theorem}[The Strong Perfect Graph Theorem\cite{CRST06}]\label{thm:SPGT}
A graph is perfect if and only if it contains no odd holes or odd antiholes.
\end{theorem}
Moreover, we prove a property about homogeneous sets, which will be used frequently in the proof of our results.
\begin{lemma}\label{lem:homogeneous}
Let $G$ be a 5-vertex-critical $P_5$-free graph and $S$ be a homogeneous set of $V(G)$. For each component $A$ of $G[S]$,
\begin{enumerate}[(i)]
\item if $\chi(A)=1$, then $A$ is a $K_1$;
\item if $\chi(A)=2$, then $A$ is a $K_2$;
\item if $\chi(A)=3$, then $A$ is a $K_3$ or a $C_5$.
\end{enumerate} \end{lemma}
\begin{proof}
(i) is clearly true. Moreover, since $V(A)\subseteq S$, $V(A)$ is also a homogeneous set. Next we prove (ii) and (iii).
(ii)Since $\chi(A)=2$, let $\{x,y\}\subseteq V(A)$ induce a $K_2$. Suppose that there is another vertex $z$ in $A$. Because $G$ is 5-vertex-critical, $G-z$ is 4-colorable. Since $\chi(A)=2$, let $\{V_1,V_2,V_3,V_4\}$ be a 4-coloring of $G-z$ where $V(A)\setminus\{z\}\subseteq V_1\cup V_2$. Since $A$ is homogeneous, $\{V_1\cup \{z\},V_2,V_3,V_4\}$ or $\{V_1,V_2\cup \{z\},V_3,V_4\}$ is a 4-coloring of $G$, a contradiction. Thus $A$ is a $K_2$.
(iii)We first show that $G$ must contain a $K_3$ or a $C_5$. If $A$ is $K_3$-free, then $\omega(A)<\chi(A)=3$ and so $A$ is imperfect. Since $A$ is $P_5$-free, $A$ must contain a $C_5$ by \autoref{thm:SPGT}. Thus $A$ contains either a $K_3$ or a $C_5$.
If $A$ contains a $K_3$ induced by $\{x,y,z\}$, suppose that there is another vertex $s$ in $A$. Because $G$ is 5-vertex-critical, $G-s$ is 4-colorable. Since $\chi(A)=3$, let $\{V_1,V_2,V_3,V_4\}$ be a 4-coloring of $G-s$ where $V(A)\setminus\{s\}\subseteq V_1\cup V_2\cup V_3$. Since $A$ is homogeneous, $\{V_1\cup \{s\},V_2,V_3,V_4\}$, $\{V_1,V_2\cup \{s\},V_3,V_4\}$ or $\{V_1,V_2,V_3\cup \{s\},V_4\}$ is a 4-coloring of $G$, a contradiction. Thus $A$ is a $K_3$. Similarly, $A$ is a $C_5$ if $A$ contains a $C_5$. \end{proof}
\section{Structure around a 5-hole}\label{structure}
Let $G=(V,E)$ be a graph and $H$ be an induced subgraph of $G$. We partition $V\setminus V(H)$ into subsets with respect to $H$ as follows: for any $X\subseteq V(H)$, we denote by $S(X)$ the set of vertices in $V\setminus V(H)$ that have $X$ as their neighborhood among $V(H)$, i.e.,
$$S(X)=\{v\in V\setminus V(H): N_{V(H)}(v)=X\}.$$
\noindent For $0\le m\le|V(H)|$, we denote by $S_m$ the set of vertices in $V\setminus V(H)$ that have exactly $m$ neighbors in $V(H)$. Note that $S_m=\cup_{X\subseteq V(H):|X|=m}S(X)$.
Let $G$ be a ($P_5$,bull)-free graph and $C=v_1,v_2,v_3,v_4,v_5$ be an induced $C_5$ in $G$. We partition $V\setminus C$ with respect to $C$ as above. All subscripts below are modulo five. Clearly, $S_1=\emptyset$ and so $V(G)=V(C)\cup S_0\cup S_2\cup S_3\cup S_4\cup S_5$. Since $G$ is ($P_5$,bull)-free, it is easy to verify that $S(v_i,v_{i+1})=S(v_{i-2},v_i,v_{i+2})=\emptyset$. So $S_2=\cup_{1\le i\le 5}S(v_{i-1},v_{i+1})$ and $S_3=\cup_{1\le i\le 5}S(v_{i-1},v_{i},v_{i+1})$. Note that $S_4=\cup_{1\le i\le 5}S(v_{i-2},v_{i-1},v_{i+1},v_{i+2})$. In the following, we write $S_2(i)$ for $S(v_{i-1},v_{i+1})$, $S_3(i)$ for $S(v_{i-1},v_{i},v_{i+1})$ and $S_4(i)$ for $S(v_{i-2},v_{i-1},v_{i+1},v_{i+2})$. We now prove a number of useful properties of $S(X)$ using the fact that $G$ is ($P_5$,bull)-free. All properties are proved for $i=1$ due to symmetry. In the following, if we say that $\{r,s,t,u,v\}$ induces a bull, it means that $r,v$ are two pendant vertices, $s$ is the neighbor of $r$, $u$ is the neighbor of $v$, and $stu$ is a triangle. If we say that $\{r,s,t,u,v\}$ induces a $P_5$, it means that any two consecutive vertices are adjacent.
\begin{enumerate}[label=\bfseries (\arabic*)]
\item {$S_2(i)$ is complete to $S_2(i+1)\cup S_3(i+1)$.}\label{S2(i)S2(i+1)}
Let $x\in S_2(1)$ and $y\in S_2(2)\cup S_3(2)$. If $xy\notin E$, then $\{x,v_5,v_4,v_3,y\}$ induces a $P_5$.
\item {$S_2(i)$ is anticomplete to $S_2(i+2)$.}\label{S2(i)S2(i+2)}
Let $x\in S_2(1)$ and $y\in S_2(3)$. If $xy\in E$, then $\{v_3,v_2,y,x,v_5\}$ induces a bull.
\item {$S_2(i)$ is anticomplete to $S_3(i+2)$.}\label{S2(i)S3(i+2)}
Let $x\in S_2(1)$ and $y\in S_3(3)$. If $xy\in E$, then $\{v_1,v_2,x,y,v_4\}$ induces a bull.
\item {$S_2(i)$ is anticomplete to $S_4(i)$.}\label{S2(i)S4(i)}
Let $x\in S_2(1)$ and $y\in S_4(1)$. If $xy\in E$, then $\{v_1,v_2,x,y,v_4\}$ induces a bull.
\item {$S_2(i)\cup S_3(i)$ is complete to $S_4(i+2)$.}\label{S2(i)S4(i+2)}
Let $x\in S_2(1)\cup S_3(1)$ and $y\in S_4(3)$. If $xy\notin E$, then $\{v_3,v_4,y,v_5,x\}$ induces a bull.
\item {$S_2(i)$ is complete to $S_4(i+1)\cup S_5$.}\label{S2S5}
Let $x\in S_2(1)$ and $y\in S_4(2)\cup S_5$. If $xy\notin E$, then $\{v_3,y,v_1,v_5,x\}$ induces a bull.
\item {$S_3(i)$ is complete to $S_3(i+1)$.}\label{S3(i)S3(i+1)}
Let $x\in S_3(1)$ and $y\in S_3(2)$. If $xy\notin E$, then $\{x,v_5,v_4,v_3,y\}$ induces a $P_5$.
\end{enumerate}
\section{The main result}\label{bull-free}
Let $\mathcal{F}$ be the set of graphs shown in \autoref{fig:5vertexcritical}. It is easy to verify that all graphs in $\mathcal{F}$ are 5-vertex-critical.
\begin{figure}
\caption{$F_1$.}
\caption{$F_2$.}
\caption{$F_3$.}
\caption{$F4$.}
\caption{$F_5$.}
\caption{$F_6$.}
\caption{$F_7$.}
\caption{$F_8$.}
\caption{$F_9$.}
\caption{Some 5-vertex-critical graphs.}
\label{K5}
\label{F1}
\label{F2}
\label{F3}
\label{F4}
\label{F5}
\label{F6}
\label{F7}
\label{F8}
\label{F9}
\label{fig:5vertexcritical}
\end{figure}
\begin{theorem}\label{th}
There are finitely many 5-vertex-critical $(P_5,bull)$-free graphs.
\end{theorem}
\begin{proof}
Let $G=(V,E)$ be a 5-vertex-critical ($P_5$,bull)-free graph. We show that $|G|$ is bounded. If $G$ has a subgraph isomorphic to a member $F\in\mathcal{F}$, then $|V(G)|=|V(F)|$ by the definition of vertex-critical graph and so we are done. Hence, we assume in the following that $G$ has no subgraph isomorphic to a member in $\mathcal{F}$. Since there are exactly 13 5-vertex-critical $(P_5,C_5$)-free graphs \cite{HMRSV15}, the proof is completed if $G$ is $C_5$-free. So assume that $G$ contains an induced $C_5$ in the following. Let $C={v_1,v_2,v_3,v_4,v_5}$ be an induced $C_5$. We partition $V(G)$ with respect to $C$.
\begin{claim}\label{S5}
$S_5$ is an independent set.
\end{claim}
\begin{proof}
Suppose that $x,y\in S_5$ and $xy\in E$. Then $G$ contains $F_1$, a contradiction.
\end{proof}
\begin{claim}\label{coloring number}
For each $1\le i\le 5$, some properties of $G$ are as follows:
\begin{itemize}
\item $\chi(G[S_3(i)])\le 2$.
\item $\chi(G[S_2(i)\cup S_3(i)])\le 3$.
\item $\chi(G[S_4(i)])\le 2$.
\item $\chi(G[S_5\cup S_0])\le 4$.
\end{itemize}
\end{claim}
\begin{proof}
It suffices to prove for $i=1$. Suppose that $\chi(G[S_3(1)])\ge3$. Then $\chi(G-v_3)\ge 5$, contradicting that $G$ is 5-vertex-critical. So $\chi(G[S_3(1)])\le2$. Similarly, We can prove the other three properties.
\end{proof}
We first bound $S_0$.
\begin{claim}\label{S0}
{$N(S_0)\subseteq S_5$}.
\end{claim}
\begin{proof}
Let $x\in N(S_0)$ and $y\in S_0$ be a neighbor of $x$. Then we show that $x\in S_5$. Let $1\le i\le5$. If $x\in S_2(i)\cup S_3(i)$, then $\{y,x,v_{i+1},v_{i+2},v_{i+3}\}$ induces a $P_5$. If $x\in S_4(i)$, then $\{v_i,v_{i+1},v_{i+2},x,y\}$ induces a bull. Therefore, $y\notin S_2\cup S_3\cup S_4$. It follows that $y\in S_5$.
\end{proof}
\begin{claim}\label{S0 color}
If A is a component of $G[S_0]$, then $\chi(A)=4$.
\end{claim}
\begin{proof}
By \autoref{coloring number}, $\chi(A)\le4$. Suppose that $\chi(A)\le 3$. So $\chi(C)\ge \chi(A)$. Combined with the fact that $C$ is anticomplete to $A$, we know that $C$ is not complete to $N(A)$ by \autoref{lem:XY}. This contradicts the facts that $C$ is complete to $S_5$ and $N(A)\subseteq S_5$. Thus $\chi(A)=4$.
\end{proof}
\begin{claim}\label{S0 connected}
$G[S_0]$ is connected.
\end{claim}
\begin{proof}
Suppose that there are two components $A_1$ and $A_2$ in $G[S_0]$. Since $G$ is connected, there must exist $w_1\in N(A_1)$ and so $w_1\in S_5$ by \autoref{S0}. By \autoref{coloring number}, $w_1$ cannot be complete to $A_1$ and $A_2$. So $w_1$ is mixed on an edge $x_1y_1\in E(A_1)$. Similarly, there exists $w_2\in S_5$ mixed on an edge $x_2y_2\in E(A_2)$ and not complete to $A_1$. So $w_2$ is anticomplete to $A_1$, otherwise if $w_2$ is mixed on an edge $z_1z_2\in E(A_1)$, then $\{z_1,z_2,w_2,x_2,y_2\}$ induces a $P_5$. It follows that $w_2$ is anticomplete to $\{x_1,y_1\}$. Then $\{y_1,x_1,w_1,v_1,w_2\}$ induces a $P_5$, a contradiction.
\end{proof}
By Claims \ref{S0 color}-\ref{S0 connected}, we obtain the following claim.
\begin{claim}\label{S0 4-chromatic}
$G[S_0]$ is a connected 4-chromatic graph.
\end{claim}
\begin{claim}
$N(S_0)=S_5$.
\end{claim}
\begin{proof}
Suppose that $w_1\in S_5$ is anticomplete to $S_0$. Since $G$ is connected, there must exist $w_2\in S_5$, which is a neighbor of $S_0$. By \autoref{coloring number}, $w_2$ is not complete to $S_0$ and so mixed on an edge $xy$ in $G[S_0]$. Thus, $\{w_1,v_1,w_2,x,y\}$ induces a $P_5$, a contradiction.
\end{proof}
To bound $S_0$, we partition $S_0$ into two parts. Let $L=S_0\cap N(S_5)$ and $R=S_0\setminus L$.
\begin{claim}\label{L S5}
If $R\neq\emptyset$, then (i)$L$ is complete to $S_5$; (ii)$N(R)=L$.
\end{claim}
\begin{proof}
Let $L_i=\{l\in L|d(l,R)=i\}$, where $i\ge 1$. Let $l\in L_1$. There exists $r\in R$, which is adjacent to $l$. Let $u\in S_5$ be a neighbor of $l$. Note that if $|S_5|=1$, $S_5$ is a clique cutset of $G$, contradicting \autoref{lem:xy}. So $|S_5|\ge 2$. For each $u'\in S_5\setminus\{u\}$, $u'$ is adjacent to $l$, otherwise $\{r,l,u,v_1,u'\}$ induces a $P_5$. Hence, $L_1$ is complete to $S_5$. Let $l_2\in L_2$. By the definition of $L_2$, there must exist $l_1\in L_1$, $l_2$ is adjacent to $l_1$. Let $r_1\in R$ and $u_2\in S_5$ be the neighbor of $l_1$ and $l_2$, respectively. Since $d(l_2,R)=2$, $l_2r_1\notin E$. Since $L_1$ is complete to $S_5$, $l_1u_2\in E$. Thus $\{v_1,u_2,l_2,l_1,r_1\}$ induces a bull, a contradiction. So $L_2=\emptyset$ and thus $L_i=\emptyset$ for each $i\ge 3$. Then $L=L_1$. Therefore, $L$ is complete to $S_5$ and $N(R)=L$.
\end{proof}
\begin{claim}\label{LR components}
Let $L'$ and $R'$ be components of $G[L]$ and $G[R]$, respectively. Then $L'$ is complete or anticomplete to $R'$.
\end{claim}
\begin{proof}
Let $u\in S_5$. By \autoref{L S5}, $u$ is complete to $L'$. Assume $L'$ is not anticomplete to $R'$. We show that $L'$ is complete to $R'$ in the following. Let $l_1\in V(L')$ and $r_1\in V(R')$ be adjacent. If $l_1$ is mixed on $R'$, then $l_1$ must be mixed on an edge $x_1y_1$ in $R'$ and so $\{v_1,u,l_1,x_1,y_1\}$ induces a $P_5$, a contradiction. So $l_1$ is complete to $R'$. Suppose that $l_2\in V(L')$ is not complete to $R'$, then there exists $r_2\in V(R')$ not adjacent to $l_2$. Since $l_1r_2\in E$, $r_2$ is mixed on $L'$ and so mixed on an edge $x_2y_2$ in $L'$. Thus $\{v_1,u,x_2,y_2,r_2\}$ induces a bull, a contradiction. It follows that $L'$ is complete to $R'$.
\end{proof}
\begin{claim}\label{R finite}
$|R|\le 8$.
\end{claim}
\begin{proof}
Let $R'$ and $R''$ be two arbitrary components of $G[R]$. Let $u_1\in S_5$. If there exists $l_1,l_2\in L$ such that $l_1\in N(R')\setminus N(R'')$ and $l_2\in N(R'')\setminus N(R')$, then $\{u_1,l_1,l_2\}\cup R'\cup R''$ contains an induced bull or an induced $P_5$, depending on whether $l_1l_2\in E$. So $N(R')\subseteq N(R'')$ or $N(R'')\subseteq N(R')$. We may assume $N(R')\subseteq N(R'')$. By \autoref{LR components}, $R''$ is complete to $N(R')$. It follows from \autoref{lem:XY} that $\chi(R'')<\chi(R')$. By \autoref{S0 4-chromatic} and \autoref{LR components}, for each component of $G[R]$, there must exist a vertex in $L$ complete to this component. Since $G[S_0]$ is 4-chromatic, the chromatic number of components of $G[R]$ is at most 3. So there are at most three components $R_1,R_2$ and $R_3$ in $G[R]$. Assume that $\chi(R_1)=1,\chi(R_2)=2$ and $\chi(R_3)=3$. By \autoref{LR components} and the definition of $R$, we know that $R_1,R_2$ and $R_3$ are all homogeneous. By \autoref{lem:homogeneous}, we know that $|R_1|=1$, $|R_2|=2$ and $|R_3|\le 5$. Therefore, $|R|\le 8$.
\end{proof}
\begin{claim}\label{L finite}
If $R\neq \emptyset$, then $|L|\le 8$.
\end{claim}
\begin{proof}
Let $L'$ and $L''$ be two arbitrary components of $G[L]$. By \autoref{L S5}, $L',L''\subseteq N(R)$. Let $u_1\in S_5$. By \autoref{L S5}, \autoref{LR components} and \autoref{coloring number}, each component of $G[L]$ must be complete to some component of $G[R]$ and so $\chi(G[L])\le 3$. Suppose that there exists $r_1,r_2\in R$ such that $r_1\in N(L')\setminus N(L'')$ and $r_2\in N(L'')\setminus N(L')$. Then $r_1$ and $r_2$ belong to different components of $R$ by \autoref{LR components}. So $r_1r_2\notin E$. Then $\{u_1,r_1,r_2\}\cup L'\cup L''$ contains an induced $P_5$, a contradiction. Combined with \autoref{L S5}, we know that $N(L')\subseteq N(L'')$ or $N(L'')\subseteq N(L')$. We may assume $N(L')\subseteq N(L'')$. By \autoref{LR components}, $L''$ is complete to $N(L')$. It follows from \autoref{lem:XY} that $\chi(L'')<\chi(L')$. Note that $\chi(G[L])\le 3$. So there are at most three components $L_1,L_2$ and $L_3$ in $G[L]$. Assume that $\chi(L_1)=1,\chi(L_2)=2$ and $\chi(L_3)=3$. By \autoref{LR components} and \autoref{L S5}, we know that $L_1,L_2$ and $L_3$ are all homogeneous. By \autoref{lem:homogeneous}, we know that $|L_1|=1$, $|L_2|=2$ and $|L_3|\le 5$. Therefore, $|L|\le 8$.
\end{proof}
By Claims \ref{R finite}-\ref{L finite}, we obtain the following claim.
\begin{claim}\label{cla:S0 1}
If $R\neq\emptyset$, $|S_0|\le 16$.
\end{claim}
Next, we bound $S_0$ when $R=\emptyset$.
\begin{claim}\label{cla:S0 2}
If $R=\emptyset$, then $|S_0|\le 13$.
\end{claim}
\begin{proof}
Since $R=\emptyset$, $S_0\subseteq N(S_5)$. For each $v\in S_0$, $\chi(G-v)=4$ since $G$ is 5-vertex-critical. Let $\pi$ be a 4-coloring of $G-v$. By the fact that $\chi(C)=3$ and $S_5$ is complete to $C$, all vertices in $S_5$ must be colored with the same color in $\pi$. Since $S_0\subseteq N(S_5)$, the vertices in $S_0\setminus\{v\}$ must be colored with the remaining three colors, i.e., $\chi(G[S_0]-v)\le 3$. Combined with \autoref{S0 4-chromatic}, $G[S_0]$ is a $P_5$-free 4-vertex-critical graph. By \autoref{thm:finite4Critical}, $|S_0|\le 13$.
\end{proof}
By Claims \ref{cla:S0 1}-\ref{cla:S0 2}, $|S_0|\le 16$. Next, we bound $S_5$.
\begin{claim}\label{S4(i)S5}
For at most one value of $i$, where $1\le i\le 5$, $S_4(i)$ is not anticomplete to $S_5$.
\end{claim}
\begin{proof}
Suppose that $S_4(i)$ and $S_4(j)$ are not anticomplete to $S_5$, where $1\le i<j\le5$. Then $G$ must have a subgraph isomorphic to $F_2,F_3,F_4$ or $F_5$, a contradiction.
\end{proof}
\begin{claim}
$|S_5|\le 2^{16}$.
\end{claim}
\begin{proof}
Suppose that $|S_5|> 2^{|S_0|}$. By the pigeonhole principle, there are two vertices $u,v\in S_5$ that have the same neighborhood in $S_0$. Since $u$ and $v$ are not comparable, there exists $x\in N(u)\setminus N(v)$ and $y\in N(v)\setminus N(u)$. Clearly, $x,y\in S_3\cup S_4(i)$ by \autoref{S4(i)S5} and \ref{S2S5}, for some $1\le i\le 5$. By symmetry, we assume $i=1$.
Suppose that $x,y\in S_4(1)$. Then $xy\notin E$, otherwise $G$ has a subgraph isomorphic to $F_8$. So $\{x,u,v_1,v,y\}$ induces a $P_5$, a contradiction.
Suppose that $x,y\in S_3$. Without loss of generality, we assume $x\in S_3(1)$. If $y\in S_3(3)\cup S_3(4)$, $G$ must have a subgraph isomorphic to $F_7$, a contradiction. If $y\in S_3(2)\cup S_3(5)$, then $xy\in E$ by \ref{S3(i)S3(i+1)} and so $G$ contains $F_8$, a contradiction. If $y\in S_3(1)$, then $xy\notin E$, otherwise $G$ has a subgraph isomorphic to $F_6$. Then $\{x,u,v_3,v,y\}$ induces a $P_5$, a contradiction.
So we assume that $x\in S_4(1)$ and $y\in S_3$. If $y\in S_3(1)\cup S_3(2)\cup S_3(5)$, then $G$ has a subgraph isomorphic to $F_7$, a contradiction. Thus $y\in S_3(3)\cup S_3(4)$. From \ref{S2(i)S4(i+2)} we know that $xy\in E$. Note that $G$ has a subgraph isomorphic to $F_8$, a contradiction.
Therefore, $|S_5|\le 2^{|S_0|}\le 2^{16}$.
\end{proof}
Next, we bound $S_2$. By \ref{S2(i)S2(i+1)}-\ref{S2S5} and \autoref{S0}, for each $1\le i\le 5$, all vertices in $V\setminus S_2(i)$ are complete or anticomplete to $S_2(i)$, except those in $S_3(i)$. So we divide $S_2(i)$ into two parts. Let $R(i)=S_2(i)\cap N(S_3(i))$ and $L(i)=S_2(i)\setminus R(i)$.
\begin{claim}\label{P3 conclusion1}
If $G[R(i)]$ contains a $P_3$, then the two endpoints of the $P_3$ have the same neighborhood in $S_3(i)$.
\end{claim}
\begin{proof}
Let $uvw$ be a $P_3$ contained in $R(i)$. Let $u'\in S_3(i)$ be a neighbor of $w$. Then $uu'\in E$, otherwise $\{u,v,w,u',v_i\}$ induces a bull or a $P_5$, depending on whether $vu'\in E$. So $N_{S_3(i)}(w)\subseteq N_{S_3(i)}(u)$. Similarly, $N_{S_3(i)}(u)\subseteq N_{S_3(i)}(w)$. Therefore, $u$ and $w$ have the same neighborhood in $S_3(i)$.
\end{proof}
\begin{claim}\label{L(i)}
$|L(i)|\le 8$.
\end{claim}
\begin{proof}
If $S_3(i)=\emptyset$ or $R(i)=\emptyset$, then $S_2(i)$ is homogeneous. If there are two components $X$ and $Y$ in $G[S_2(i)]$, then $Y$ is complete to $N(X)$ and $X$ is complete to $N(Y)$, contradicting \autoref{lem:XY}. So $G[S_2(i)]$ is connected. By \autoref{coloring number} and \autoref{lem:homogeneous}, $G[S(i)]$ is a $K_1$, a $K_2$, a $K_3$ or a $C_5$. Thus $|L(i)|\le 5$.
So we assume that $S_3(i)\neq \emptyset$ and $R(i)\neq \emptyset$. Let $u$ be an arbitrary vertex in $R(i)$ and $u'$ be its neighbor in $S_3(i)$. Then $u$ is not mixed on any edge $xy$ in $L(i)$, otherwise $\{y,x,u,u',v_i\}$ induces a $P_5$. Then $u$ is complete or anticomplete to any component of $L(i)$ and so all components of $L(i)$ are homogeneous. By \autoref{lem:homogeneous}, each component of $L(i)$ is a $K_1$, a $K_2$, a $K_3$ or a $C_5$.
We show that there is at most one 3-chromatic component in $L(i)$. Suppose that $X_1$ and $Y_1$ are two 3-chromatic components in $L(i)$. Note that $X_1$ and $Y_1$ are homogeneous. Since $\chi(G[S_2(i)])\le 3$, $X_1$ and $Y_1$ are anticomplete to $R(i)$. So $Y_1$ is complete to $N(X_1)$ and $X_1$ is complete to $N(Y_1)$, which contradicts \autoref{lem:XY}. So, there is at most one 3-chromatic component in $L(i)$.
Then we show that there is at most one $K_2$-component in $L(i)$. Suppose that $X_2=x_1y_1$ and $Y_2=x_2y_2$ are two $K_2$-components in $L(i)$. Note that $X_2$ and $Y_2$ are homogeneous. By \autoref{lem:XY}, there must exist $u_1,u_2\in R(i)$ such that $u_1$ is complete to $X_2$ and anticomplete to $Y_2$ and $u_2$ is complete to $Y_2$ and anticomplete to $X_2$ . Let $u_1',u_2'\in S_3(i)$ be the neighbor of $u_1$ and $u_2$, respectively. Clearly, $u_1'$ and $u_2'$ are not the same vertex, otherwise $\{x_1,u_1,u_1',u_2,x_2\}$ induces a bull or a $P_5$, depending on whether $u_1u_2\in E$. So $u_1'u_2\notin E$ and $u_2'u_1\notin E$. It follows that $u_1u_2\notin E$, otherwise $\{x_2,u_2,u_1,u_1',v_i\}$ induces a $P_5$. Then $\{u_1,u_1',v_i,u_2',u_2\}$ induces a bull or a $P_5$, depending on whether $u_1'u_2'\in E$, a contradiction. So, there is at most one $K_2$-component in $L(i)$.
Similarly, there is at most one $K_1$-component in $L(i)$. It follows that $|L(i)|\le 8$. The proof is completed.
\end{proof}
\begin{figure}
\caption{The graph contained in $G[R(i)]$.}
\label{fig:uvwst}
\end{figure}
\begin{claim}\label{P3 conclusion2}
If $G[R(i)]$ contains $P_3=uvw$, then $G[R(i)]$ must contain the graph induced by $\{u,v,w,s,t\}$ in \autoref{fig:uvwst}. Moreover, $u,w,s$ and $t$ have the same neighborhood in $S_3(i)$ and $N_{S_3(i)}(u)\cap N_{S_3(i)}(v)=\emptyset$.
\end{claim}
\begin{proof}
Let $u'$ be an arbitrary neighbor of $w$ in $S_3(i)$. By \autoref{P3 conclusion1} we know that $N_{S_3(i)}(u)=N_{S_3(i)}(w)$ and so $uu'\in E$. Since $u$ and $w$ are not comparable, there must exist $s\in N(u)\setminus N(w)$ and $t\in N(w)\setminus N(u)$. Clearly, $s,t\in L(i)\cup R(i)$.
\noindent{\bf Case 1. }$s,t\in L(i)$. Then $st\notin E$, otherwise $\{s,t,w,u',v_i\}$ induces a $P_5$. Moreover, $sv\notin E$, otherwise $\{s,v,w,u',v_i\}$ induces a bull or a $P_5$, depending on whether $vu'\in E$. Similarly, $tv\notin E$. So $\{s,u,v,w,t\}$ induces a $P_5$, a contradiction.
\noindent{\bf Case 2. }One vertex of $\{s,t\}$ belongs to $L(i)$ and the other belongs to $R(i)$. We assume that $s\in L(i)$ and $t\in R(i)$. Then $sv\notin E$, otherwise $\{s,v,w,u',v_i\}$ induces a bull or a $P_5$, depending on whether $vu'\in E$. So $vu'\notin E$, otherwise $\{s,u,v,u',v_i\}$ induces a bull. Let $z'$ be a neighbor of $v$ in $S_3(i)$. Clearly, $\{s,u,v,z',v_i\}$ induces a bull or a $P_5$, depending on whether $uz'\in E$, a contradiction.
\noindent{\bf Case 3. }$s,t\in R(i)$. Suppose that $sv\notin E$. Then $suv$ is a $P_3$ and so $u'$ is complete or anticomplete to $\{s,v\}$ by \autoref{P3 conclusion1}. Suppose that $u'$ is complete to $\{s,v\}$. If $vt\in E$, then $uvt$ is a $P_3$ and so $tu'\in E$ by \autoref{P3 conclusion1}. Then $\{t,v,w,u'\}$ induces a $K_4$, contradicting that $\chi(G[S_2(i)\cup S_3(i)])\le 3$. So $vt\notin E$. Hence $vwt$ is a $P_3$ and then $tu'\in E$ by \autoref{P3 conclusion1}. Then $st\in E$, otherwise $\{s,u,v,w,t\}$ induces a $P_5$. It is easy to verify that $\{s,u,v,w,t,u'\}$ induces a 4-chromatic subgraph, contradicting that $\chi(G[S_2(i)\cup S_3(i)])\le 3$. So $u'$ must be anticomplete to $\{s,v\}$. Then $st\notin E$, otherwise $\{s,t,w,u',v_i\}$ induces a bull or a $P_5$, depending on whether $tu'\in E$. Hence $tv\in E$, otherwise $\{s,u,v,w,t\}$ induces a $P_5$. Let $z'$ be an arbitrary neighbor of $v$ in $S_3(i)$. Since $suv$ is a $P_3$, $sz'\in E$ by \autoref{P3 conclusion1}. Note that $uvt$ and $uvw$ are all $P_3$ and so $N_{S_3(i)}(u)=N_{S_3(i)}(w)=N_{S_3(i)}(t)$. Then $tz'\notin E$, otherwise $\{t,v,z',w\}$ induces a $K_4$. Note that $\{s,z',v_i,u',w\}$ induces a bull or a $P_5$, depending on whether $u'z'\in E$, a contradiction. Thus $sv\in E$. By symmetry, $tv\in E$.
Since $svw$ and $uvt$ are all $P_3$, we know that $u,w,s,t$ have the same neighborhood in $S_3(i)$ by \autoref{P3 conclusion1} and so $su',tu'\in E$. Then $vu'\notin E$, otherwise $\{v,w,t,u'\}$ induces a $K_4$. Since $u'$ is an arbitrary neighbor of $w$ in $S_3(i)$, $v$ is anticomplete to $N_{S_3(i)}(u)$. Thus $N_{S_3(i)}(u)\cap N_{S_3(i)}(v)=\emptyset$.
If $st\in E$, then $ust$ is a $P_3$. From the above proof we know that $s$ is anticomplete to $N_{S_3(i)}(u)$, which contradicts the fact that $su'\in E$. So $st\notin E$. It follows that $\{u,v,w,s,t\}$ induces the graph in \autoref{fig:uvwst}. This completes the proof of the claim.
\end{proof}
\begin{claim}\label{P3-free}
$G[R(i)]$ is $P_3$-free.
\end{claim}
\begin{proof}
Suppose that $G[R(i)]$ contains a $P_3=uvw$. By \autoref{P3 conclusion2}, $G[R(i)]$ contains a subgraph in \autoref{fig:uvwst} induced by $\{u,v,w,s,t\}$. Moreover, $u,w,s,t$ have the same neighborhood in $S_3(i)$ and $v$ is anticomplete to $N_{S_3(i)}(u)$. Let $u'$ and $v'$ be arbitrary neighbor of $u$ and $v$ in $S_3(i)$, respectively. Then $u'$ is complete to $\{u,w,s,t\}$ and nonadjacent to $v$ and $v'$ is anticomplete to $\{u,w,s,t\}$. It follows from \autoref{lem:XY} that $\{w,t\}$ is not complete to $N\{u,s\}$. So there exists $a\in N\{u,s\}$ such that $a$ is not complete to $\{w,t\}$. Clearly, $a\in L(i)\cup R(i)$.
Suppose $a\in L(i)$. Assume that $as\in E$. So $au\in E$, otherwise $\{a,s,u,u',v_i\}$ induces a bull. Then $av\in E$, otherwise $\{a,u,v,v',v_i\}$ induces a $P_5$. Note that $\{a,s,v,u\}$ induces a $K_4$, a contradiction. Thus $a\in R(i)$.
If $a$ is adjacent to only one vertex in $\{s,u\}$, then either $usa$ or $sua$ is a $P_3$ and so $N_{S_3(i)}(s)\cap N_{S_3(i)}(u)=\emptyset$ by \autoref{P3 conclusion2}, contradicting that $su',uu'\in E$. Thus $a$ is complete to $\{s,u\}$. Then $av\notin E$, otherwise $\{s,u,a,v\}$ induces a $K_4$. Because $auv$ is a $P_3$, we know that $au'\notin E$ and $av'\in E$ by \autoref{P3 conclusion2}. Since $a$ is not complete to $\{w,t\}$, we assume that $at\notin E$ by symmetry. Note that $\{t,u',v_i,v',a\}$ induces a bull or a $P_5$, depending on whether $u'v'\in E$, a contradiction.
Therefore, $G[R(i)]$ is $P_3$-free.
\end{proof}
Since $G[R(i)]$ is $P_3$-free, $G[R(i)]$ is a disjoint union of cliques. By \autoref{coloring number}, each component of $G[R(i)]$ is a $K_1$, a $K_2$ or a $K_3$. We next prove that the number of them is finite.
\begin{claim}\label{R(i) 1}
There are at most $2^{|L(i)|}$ $K_1$-components and 5 $K_2$-components in $G[R(i)]$.
\end{claim}
\begin{proof}
We first show that there are at most $2^{|L(i)|}$ $K_1$-components in $G[R(i)]$. Suppose there are more than $2^{|L(i)|}$ $K_1$-components in $G[R(i)]$. By the pigeonhole principle, there exists $u,v\in R(i)$ and they have the same neighborhood in $L(i)$. Since $u$ and $v$ are not comparable, there exists $u',v'\in S_3(i)$ such that $u'\in N(u)\setminus N(v)$ and $v'\in N(v)\setminus N(u)$. Then $\{u,u',v_i,v',v\}$ induces a bull or a $P_5$, depending on whether $u'v'\in E$, a contradiction. So there are at most $2^{|L(i)|}$ $K_1$-components in $G[R(i)]$.
Next we show that there are at most 5 $K_2$-components in $G[R(i)]$.
Suppose that $A_1$ and $A_2$ are two homogeneous $K_2$-components of $G[R(i)]$. By \autoref{lem:XY}, there exists $x_1\in N(A_1)\setminus N(A_2)$ and $y_1\in N(A_2)\setminus N(A_1)$. Clearly, $x_1,y_1\in S_3(i)\cup L(i)$. Suppose that $x_1,y_1\in L(i)$. Let $w_1,w_2\in S_3(i)$ be the neighbor of $A_1$ and $A_2$, respectively. If $x_1y_1\in E$, then $\{y_1,x_1,w_1,v_i\}\cup A_1$ contains an induced $P_5$. So $x_1y_1\notin E$. Note that $w_2\notin N(A_1)$, otherwise $\{w_2,x_1,y_1\}\cup A_1\cup A_2$ contains an induced $P_5$. Similarly, $w_1\notin N(A_2)$. Then $\{v_i,w_1,w_2\}\cup A_1\cup A_2$ contains an induced bull or an induced $P_5$, depending on whether $w_1w_2\in E$, a contradiction. Suppose that $x_1\in L(i)$ and $y_1\in S_3(i)$. Let $w_3$ be the neighbor of $A_1$ in $S_3(i)$. Note that $w_3\in N(A_2)$, otherwise $\{v_i,w_3,y_1\}\cup A_1\cup A_2$ contains an induced bull or an induced $P_5$, depending on whether $w_3y_1\in E$. Then $w_3y_1\in E$, otherwise $\{x_1,y_1,w_3\}\cup A_1\cup A_2$ contains an induced $P_5$. Then $\{w_3,y_1\}\cup A_2$ induces a $K_4$, contradicting that $\chi(G[S_2(i)\cup S_3(i)])\le 3$. So $x_1,y_1\in S_3(i)$ and then $\{v_i,x_1,y_1\}\cup A_1\cup A_2$ contains an induced bull or an induced $P_5$, depending on whether $x_1y_1\in E$, a contradiction. Thus there is at most one homogeneous $K_2$-component in $G[R(i)]$.
Let $B_1=x_3y_3$ and $B_2=x_4y_4$ be two arbitrary nonhomogeneous $K_2$-components of $G[R(i)]$ and the vertices mixed on $B_1$ or $B_2$ are clearly in $L(i)\cup S_3(i)$. Suppose that each vertex in $S_3(i)$ is complete or anticomplete to $B_1$, then there exists $z'\in L(i)$ mixed on $B_1$. Let $t\in S_3(i)$ be complete to $B_1$, then $\{z',x_3,y_3,t,v_i\}$ induces a bull, a contradiction. So there must exist $z_3\in S_3(i)$ mixed on $B_1$. Similarly, there exists $z_4\in S_3(i)$ mixed on $B_2$. By symmetry, we assume $z_3x_3,z_4x_4\in E$ and $z_3y_3,z_4y_4\notin E$. Then $z_3$ is complete or anticomplete to $B_2$, otherwise $\{y_3,x_3,z_3,x_4,y_4\}$ induces a $P_5$. Similarly, $z_4$ is complete or anticomplete to $B_1$. If $z_3$ is anticomplete to $B_2$ and $z_4$ is anticomplete to $B_1$, then $\{x_3,z_3,v_i,z_4,x_4\}$ induces a bull or a $P_5$, depending on whether $z_3z_4\in E$. If $z_3$ is complete to $B_2$ and $z_4$ is complete to $B_1$, then $\{y_3,z_4,v_i,z_3,y_4\}$ induces a bull or a $P_5$, depending on whether $z_3z_4\in E$. So we assume $z_3$ is anticomplete to $B_2$ and $z_4$ is complete to $B_1$. It follows that $z_3z_4\in E$, otherwise $\{y_4,x_4,z_4,v_i,z_3\}$ induces a $P_5$. So there are at most 4 nonhomogeneous $K_2$-components in $R(i)$, otherwise the vertices in $S_3(i)$ mixed on them respectively can induce a $K_5$, a contradiction.
The above proof shows that there are at most $2^{|L(i)|}$ $K_1$-components and 5 $K_2$-components in $G[R(i)]$.
\end{proof}
\begin{claim}\label{R(i) 2}
There is at most one $K_3$-component in $G[R(i)]$.
\end{claim}
\begin{proof}
Suppose that $T_1=x_1y_1z_1,T_2=x_2y_2z_2$ are two arbitrary $K_3$-components of $G[R(i)]$. Let $x',y'\in S_3(i)$ be the neighbor of $T_1$ and $T_2$, respectively. Since $\chi(G[S_2(i)\cup S_3(i)])\le 3$, $x'$ is mixed on $T_1$ and $y'$ is mixed on $T_2$. By symmetry, we assume that $x'x_1,y'x_2\in E$ and $x'y_1,y'y_2\notin E$. So $x'$ is not mixed on $T_2$, otherwise $\{y_1,x_1,x'\}\cup T_2$ contains an induced $P_5$. Moreover, since $\chi(G[S_2(i)\cup S_3(i)])\le 3$, $x'$ is not complete to $T_2$. Thus $x'$ is anticomplete to $T_2$. Similarly, $y'$ is anticomplete to $T_1$. Then $\{x_1,x',v_i,y',x_2\}$ induces a bull or a $P_5$, depending on whether $x'y'\in E$, a contradiction.
Therefore, there is at most one $K_3$-component in $G[R(i)]$.
\end{proof}
By Claims \ref{L(i)}, \ref{R(i) 1} and \ref{R(i) 2}, $|L(i)|\le 8$ and $|R(i)|\le 2^{|L(i)|}+13$. So $|S_2|\le 5\times(2^8+21)$.
Finally, we bound $S_3$ and $S_4$.
\begin{claim}\label{S3 trivial}
For each $1\le i\le 5$, the number of $K_1$-components in $G[S_3(i)]$ is not more than $2^{|S_2(i)\cup S_5|}$.
\end{claim}
\begin{proof}
It suffices to prove for $i=1$. Suppose that the number of $K_1$-components in $G[S_3(1)]$ is more than $2^{|S_2(1)\cup S_5|}$. The pigeonhole principle shows that there are two $K_1$-components $u,v$ having the same neighborhood in $S_2(1)\cup S_5$. Since $u$ and $v$ are not comparable, there must exist $u'\in N(u)\setminus N(v)$ and $v'\in N(v)\setminus N(u)$. By \ref{S2(i)S2(i+1)}, \ref{S2(i)S3(i+2)}, \ref{S3(i)S3(i+1)} and \ref{S2(i)S4(i+2)}, $u',v'\in S_3(3)\cup S_3(4)\cup S_4(1)\cup S_4(2)\cup S_4(5)$. So $\{u,u',v_3,v',v\}$ induces a bull or a $P_5$, depending on whether $u'v'\in E$, a contradiction.
\end{proof}
\begin{claim}\label{S4 trivial}
For each $1\le i\le 5$, the number of $K_1$-components in $G[S_4(i)]$ is not more than $2^{|S_5|}$.
\end{claim}
\begin{proof}
It suffices to prove for $i=1$. Suppose that the number of $K_1$-components in $G[S_4(1)]$ is more than $2^{|S_5|}$. The pigeonhole principle shows that there are two $K_1$-components $u,v$ having the same neighborhood in $S_5$. Since $u$ and $v$ are not comparable, there must exist $u'\in N(u)\setminus N(v)$ and $v'\in N(v)\setminus N(u)$. By \ref{S2(i)S4(i)}, \ref{S2(i)S4(i+2)} and \ref{S2S5}, $u',v'\in (\cup_{i=1,2,5}S_3(i))\cup (\cup_{2\le i\le 5}S_4(i))$. So $\{u,u',v_1,v',v\}$ induces a bull or a $P_5$, depending on whether $u'v'\in E$, a contradiction.
\end{proof}
\begin{claim}\label{S4(i) 2-chromatic}
If $\chi(S_4(i))=2$ for some $1\le i\le 5$, then $S_3\cup S_4$ is bounded.
\end{claim}
\begin{proof}
Without loss of generality, we assume $\chi(S_4(1))=2$. It follows from \ref{S2(i)S4(i+2)} that $S_3(3)=S_3(4)=\emptyset$, otherwise $S_4(1)\cup S_3(3)\cup \{v_3,v_4\}$ contains an induced $K_5$. Since $G$ has no subgraph isomorphic to $F_9$, $\chi(S_4(i))\le 1$ for each $2\le i\le 5$ and $\chi(S_3(j))\le 1$ for each $j=1,2,5$. By Claims \ref{S3 trivial}-\ref{S4 trivial}, $S_3\cup(\cup_{2\le i\le 5}S_4(i))$ is bounded and the number of $K_1$-components in $G[S_4(1)]$ is also bounded.
We now show that the number of vertices in a 2-chromatic component of $G[S_4(1)]$ is bounded. Let $A$ be a 2-chromatic component of $G[S_4(1)]$ and so $A$ is bipartite. Let the bipartition of $A$ be $(X,Y)$. Suppose that $|X|>2^{|S_3\cup (\cup_{2\le i\le 5}S_4(i))\cup S_5|}$. By the pigeonhole principle, there exists two vertices $x_1,x_2\in X$ which have the same neighborhood in $S_3\cup (\cup_{2\le i\le 5}S_4(i))\cup S_5$. Since $x_1$ and $x_2$ are not comparable, there must exist $y_1\in N(x_1)\setminus N(x_2),y_2\in N(x_2)\setminus N(x_1)$. Clearly, $y_1,y_2\in Y$ and so $\{x_1,x_2,y_1,y_2\}$ induces a $2K_2$ in $A$. Since $A$ is connected and bipartite, $A$ contains a $P_5$ by \autoref{2K2}, a contradiction. Thus $|X|\le 2^{|S_3\cup (\cup_{2\le i\le 5}S_4(i))\cup S_5|}$. Similarly, $|Y|\le 2^{|S_3\cup (\cup_{2\le i\le 5}S_4(i))\cup S_5|}$. Thus the number of vertices in $A$ is bounded.
Then we show that there are at most five 2-chromatic components in $G[S_4(1)]$.
Suppose that $A_1$ and $A_2$ are two homogeneous 2-chromatic components of $G[S_4(1)]$. By \autoref{lem:XY}, $A_1$ is not complete to $N(A_2)$ and $A_2$ is not complete to $N(A_1)$. So there must exist $z_1\in N(A_1)\setminus N(A_2)$ and $z_2\in N(A_2)\setminus N(A_1)$. Clearly, $z_1,z_2\in (\cup_{i=1,2,5}S_3(i))\cup (\cup_{2\le i\le 5}S_4(i))\cup S_5$. Then $\{v_1,z_1,z_2\}\cup A_1\cup A_2$ contains an induced bull or an induced $P_5$, depending on whether $z_1z_2\in E$, a contradiction. Thus there is at most one homogeneous 2-chromatic component in $G[S_4(1)]$.
Let $B_1,B_2$ be two nonhomogeneous 2-chromatic components of $G[S_4(1)]$. So there exists $x'$ mixed on $B_1$ and $y'$ mixed on $B_2$. Let $x'$ be mixed on edge $x_3y_3$ in $B_1$ and $y'$ be mixed on edge $x_4y_4$ in $B_2$. By symmetry, assume that $x'x_3,y'x_4\in E$ and $x'y_3,y'y_4\notin E$. It is evident that $x'$ and $y'$ are not the same vertex, otherwise $\{y_3,x_3,x',x_4,y_4\}$ induces a $P_5$. Similarly, $x'$ is not mixed on $x_4y_4$ and $y'$ is not mixed on $x_3y_3$. Clearly, $x',y'\in (\cup_{i=1,2,5}S_3(i))\cup (\cup_{2\le i\le 5}S_4(i))\cup S_5$. If $x'$ is anticomplete to $\{x_4,y_4\}$ and $y'$ is anticomplete to $\{x_3,y_3\}$, then $\{x_3,x',v_1,y',x_4\}$ induces a bull or a $P_5$, depending on whether $x'y'\in E$. If $x'$ is complete to $\{x_4,y_4\}$ and $y'$ is complete to $\{x_3,y_3\}$, then $\{y_4,x',v_1,y',y_3\}$ induces a bull or a $P_5$, depending on whether $x'y'\in E$. So we assume that $x'$ is complete to $\{x_4,y_4\}$ and $y'$ is anticomplete to $\{x_3,y_3\}$. Then $x'y'\in E$, otherwise $\{y',x_4,y_4,x',x_3\}$ induces a bull. So the number of nonhomogeneous 2-chromatic components of $G[S_4(1)]$ is not more than 4, otherwise the vertices mixed on them respectively can induce a $K_5$.
So there are at most five 2-chromatic components in $G[S_4(1)]$. It follows that $S_3\cup S_4$ is bounded.
\end{proof}
\begin{claim}\label{S3(i) 2-chromatic}
If $\chi(S_3(i))=2$ for some $1\le i\le 5$, then $S_3\cup S_4$ is bounded.
\end{claim}
\begin{proof}
Without loss of generality, we assume $\chi(S_3(3))=2$. It follows from \ref{S3(i)S3(i+1)} that $S_3(2)=S_3(4)=\emptyset$, otherwise $S_3(3)\cup S_3(2)\cup \{v_2,v_3\}$ or $S_3(3)\cup S_3(4)\cup \{v_4,v_3\}$ contains an induced $K_5$. Similarly, it follows from \ref{S2(i)S4(i+2)} that $S_4(1)=S_4(5)=\emptyset$. Since $G$ has no subgraph isomorphic to $F_9$, $\chi(S_4(i))\le 1$ for each $2\le i\le 4$ and $\chi(S_3(j))\le 1$ for each $j=1,5$. By Claims \ref{S3 trivial}-\ref{S4 trivial}, $(\cup_{i=1,5}S_3(i))\cup S_4$ is bounded and the number of $K_1$-components in $G[S_3(3)]$ is also bounded.
We now show that the number of vertices in a 2-chromatic component of $G[S_3(3)]$ is bounded. Let $A$ be a 2-chromatic component of $G[S_3(3)]$ and so $A$ is bipartite. Let the bipartition of $A$ be $(X,Y)$. Suppose that $|X|>2^{|S_2(3)\cup S_5\cup (\cup_{i=1,5}S_3(i))\cup (\cup_{2\le i\le 4}S_4(i))|}$. By the pigeonhole principle, there exists two vertices $x_1,x_2\in X$ which have the same neighborhood in $S_2(3)\cup S_5\cup (\cup_{i=1,5}S_3(i))\cup (\cup_{2\le i\le 4}S_4(i))$. Since $x_1$ and $x_2$ are not comparable, there must exist $y_1\in N(x_1)\setminus N(x_2),y_2\in N(x_2)\setminus N(x_1)$. Clearly, $y_1,y_2\in Y$ and so $\{x_1,x_2,y_1,y_2\}$ induces a $2K_2$ in $A$. Since $A$ is connected and bipartite, $A$ contains a $P_5$ by \autoref{2K2}, a contradiction. Thus $|X|\le 2^{|S_2(3)\cup S_5\cup (\cup_{i=1,5}S_3(i))\cup (\cup_{2\le i\le 4}S_4(i))|}$. Similarly,
\[|Y|\le 2^{|S_2(3)\cup S_5\cup (\cup_{i=1,5}S_3(i))\cup (\cup_{2\le i\le 4}S_4(i))|}.\] Thus the number of vertices in $A$ is bounded.
Then we show that there are at most $(2^{|S_2(3)|}+4)$ 2-chromatic components in $G[S_3(3)]$.
Suppose that the number of homogeneous 2-chromatic components of $G[S_3(3)]$ is more than $2^{|S_2(3)|}$. By the pigeonhole principle, there are two 2-chromatic components $A_1,A_2$ such that $N_{S_2(3)}(A_1)=N_{S_2(3)}(A_2)$. By \autoref{lem:XY}, $A_1$ is not complete to $N(A_2)$ and $A_2$ is not complete to $N(A_1)$. So there must exist $z_1\in N(A_1)\setminus N(A_2)$ and $z_2\in N(A_2)\setminus N(A_1)$. Clearly, $z_1,z_2\in (\cup_{i=1,5}S_3(i))\cup (\cup_{2\le i\le 4}S_4(i))\cup S_5$. Then $\{v_1,z_1,z_2\}\cup A_1\cup A_2$ contains an induced bull or an induced $P_5$, depending on whether $z_1z_2\in E$, a contradiction. Thus there are at most $2^{|S_2(3)|}$ homogeneous 2-chromatic components in $G[S_3(3)]$.
Let $B_1,B_2$ be two nonhomogeneous 2-chromatic components of $G[S_3(3)]$. So there exists $x'$ mixed on $B_1$ and $y'$ mixed on $B_2$. Let $x'$ be mixed on edge $x_3y_3$ in $B_1$ and $y'$ be mixed on edge $x_4y_4$ in $B_2$. By symmetry, assume that $x'x_3,y'x_4\in E$ and $x'y_3,y'y_4\notin E$. It is evident that $x'$ and $y'$ are not the same vertex, otherwise $\{y_3,x_3,x',x_4,y_4\}$ induces a $P_5$. Similarly, $x'$ is not mixed on $x_4y_4$ and $y'$ is not mixed on $x_3y_3$. Clearly, $x',y'\in S_2(3)\cup S_5\cup (\cup_{i=1,5}S_3(i))\cup (\cup_{2\le i\le 4}S_4(i))$.
\noindent{\bf Case 1.} $x'$ is anticomplete to $\{x_4,y_4\}$ and $y'$ is anticomplete to $\{x_3,y_3\}$. Then $x'$ is nonadjacent to $y'$, otherwise $\{y_3,x_3,x',y',x_4,y_4\}$ induces a $P_6$. If $x',y'\notin S_2(3)$, then $\{x_3,x',v_1,y',x_4\}$ induces a $P_5$. If $x',y'\in S_2(3)$, then $\{x',x_3,v_3,x_4,y'\}$ induces a $P_5$. So assume $x'\in S_2(3)$ and $y'\notin S_2(3)$. Then $\{x_4,v_3,y_3,x_3,x'\}$ induces a bull, a contradiction.
\noindent{\bf Case 2.} $x'$ is complete to $\{x_4,y_4\}$ and $y'$ is anticomplete to $\{x_3,y_3\}$. Then $x'y'\in E$, otherwise $\{y',x_4,y_4,x',x_3\}$ induces a bull. So as the case when $x'$ is anticomplete to $\{x_4,y_4\}$ and $y'$ is complete to $\{x_3,y_3\}$.
\noindent{\bf Case 3.} $x'$ is complete to $\{x_4,y_4\}$ and $y'$ is complete to $\{x_3,y_3\}$. Suppose that $x',y'\notin S_2(3)$ and so $\{y_4,x',v_1,y',y_3\}$ induces a bull or a $P_5$, depending on whether $x'y'\in E$, a contradiction. If $x',y'\in S_2(3)$, then $x'y'\in E$, otherwise $\{x',y_4,v_3,y_3,y'\}$ induces a $P_5$. If $x'\in S_2(3)$ and $y'\notin S_2(3)$, then $x'y'\in E$, otherwise $\{v_1,y',y_3,x_3,x'\}$ induces a bull.
We now know that $x'$ must be adjacent to $y'$. So the number of nonhomogeneous 2-chromatic components of $G[S_3(3)]$ is not more than 4, otherwise the vertices mixed on them respectively can induce a $K_5$, a contradiction. It follows that there are at most $(2^{|S_2(3)|}+4)$ 2-chromatic components in $G[S_3(3)]$.
Therefore, $S_3\cup S_4$ is bounded.
\end{proof}
By Claims \ref{S3 trivial}-\ref{S3(i) 2-chromatic}, $S_3\cup S_4$ is bounded and so is $|G|$. This completes the proof of \autoref{th}.
\end{proof}
\end{document}
|
arXiv
|
{
"id": "2211.04179.tex",
"language_detection_score": 0.6690574288368225,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Decidability of the Monadic Shallow Linear First-Order Fragment with Straight Dismatching Constraints}
\author{Andreas Teucke\inst{1,2} \and Christoph Weidenbach\inst{1}}
\institute{Max-Planck Institut f\"ur Informatik, Saarland Informatics Campus,
66123 Saarbr\"ucken
Germany \and Graduate School of Computer Science, Saarbr\"ucken, Germany }
\maketitle
\begin{abstract} The monadic shallow linear Horn fragment is well-known to be decidable and has many application, e.g., in security protocol analysis, tree automata, or abstraction refinement. It was a long standing open problem how to extend the fragment to the non-Horn case, preserving decidability, that would, e.g., enable to express non-determinism in protocols. We prove decidability of the non-Horn monadic shallow linear fragment via ordered resolution further extended with dismatching constraints and discuss some applications of the new decidable fragment. \end{abstract}
\section{Introduction} \label{sec:intro}
Motivated by the automatic analysis of security protocols, the monadic shallow linear Horn (MSLH) fragment was shown to be decidable in \cite{Weidenbach99cade}. In addition to the restriction to monadic Horn clauses, the main restriction of the fragment is positive literals of the form $S(f(x_1,\ldots,x_n))$ or $S(x)$ where all $x_i$ are different, i.e., all terms are shallow and linear. The fragment can be finitely saturated by superposition (ordered resolution) where negative literals with non-variable arguments are always selected. As a result, productive clauses with respect to the superposition model operator $\I{N}$ have the form $S_1(x_1),\ldots,S_n(x_n) \rightarrow S(f(x_1,\ldots,x_n))$. Therefore, the models of saturated MSLH clause sets can both be represented by tree automata \cite{tata2007} and shallow linear sort theories \cite{JacquemardMeyerEtAl98}. The models are typically infinite. The decidability result of MSLH clauses was rediscovered in the context of tree automata research \cite{GoubaultLarrecq05IPL} where in addition DEXPTIME-completeness of the MSLH fragment was shown. The fragment was further extended by disequality constraints \cite{journalsiplSeidlR11,Seidl12F} still motivated by security protocol analysis \cite{SeidlVerma2007}. Although from a complexity point of view, the difference between Horn clause fragments and the respective non-Horn clause fragments is typically reflected by membership in the deterministic vs.\ the non-deterministic respective complexity fragment, for monadic shallow linear clauses so far there was no decidability result for the non-Horn case.
The results of this paper close this gap. We show the monadic shallow linear non-Horn (MSL) clause fragment to be decidable by superposition (ordered resolution). From a security protocol application point of view, non-Horn clauses enable a natural representation of non-determinism. Our second extension to the fragment are unit clauses with disequations of the form $s\not\approx t$, where $s$ and $t$ are not unifiable. Due to the employed superposition calculus, such disequations do not influence saturation of an MSL clause set, but have an effect on potential models. They can rule out identification of syntactically different ground terms as it is, e.g., desired in the security protocol context for syntactically different messages or nonces. Our third extension to the fragment are straight dismatching constraints. These constraints are incomparable to the disequality constraints mentioned above \cite{journalsiplSeidlR11,Seidl12F}. They do not strictly increase the expressiveness of the MSL theory, but enable up to exponentially more compact saturations. For example, the constrained clause\newline \centerline{$(S(x), T(y) \rightarrow S(f(x,y)); y\neq f(x',f(a,y')))$} over constants $a, b$ describes the same set of ground clauses as the six unconstrained clauses\newline \centerline{$S(x), T(a) \rightarrow S(f(x,a)) \qquad S(x), T(b) \rightarrow S(f(x,b)) \qquad \ldots$} \centerline{$S(x), T(f(b,y')) \rightarrow S(f(x,f(b,y')))$} \centerline{$S(x), T(f(f(x'',y''),y')) \rightarrow S(f(x,f(f(x'',y''),y'))$.} Furthermore, for a satisfiability equivalent transformation into MSL clauses, the nested terms in the positive literals would have to be factored out by the introduction of further predicates and clauses. E.g., the first clause is replaced by the two MSL clauses $S(x), T(a), R(y) \rightarrow S(f(x,y))$ and $R(a)$ where $R$ is a fresh monadic predicate. The constrained clause belongs to the MSL(SDC) fragment. Altogether, the resulting MSL(SDC) fragment is shown to be decidable in Section~\ref{sec:decide}.
The introduction of straight dismatching constraints (SDCs) enables an improved refinement step of our approximation refinement calculus~\cite{Teucke2015}. Before, several clauses were needed to rule out a specific instance of a clause in an unsatisfiable core. For example, if due to a linearity approximation from clause $S(x), T(x) \rightarrow S(f(x,x))$ to $S(x), T(x), S(y), T(y) \rightarrow S(f(x,y))$ an instance $\{x\mapsto f(a,x')$, $y\mapsto f(b,y')\}$ is used in the proof, before \cite{Teucke2015} several clauses were needed to replace $S(x), T(x) \rightarrow S(f(x,x))$ in a refinement step in order to rule out this instance. With straight dismatching constraints the clause $S(x), T(x) \rightarrow S(f(x,x))$ is replaced by the two clauses $S(f(a,x)), T(f(a,x)) \rightarrow S(f(f(a,x),f(a,x)))$ and $(S(x), T(x) \rightarrow S(f(x,x)); x\neq f(a,y))$. For the improved approximation refinement approach (FO-AR) presented in this paper, any refinement step results in just two clauses, see Section~\ref{sec:approx}. The additional expressiveness of constraint clauses comes almost for free, because necessary computations, like, e.g., checking emptiness of SDCs, can all be done in polynomial time, see Section~\ref{sec:prelim}.
In addition to the extension of the known MSLH decidability result and the improved approximation refinement calculus FO-AR, we discuss in Section~\ref{sec:experiments} the potential of the MSL(SDC) fragment in the context of FO-AR, Theorem~\ref{theo:refinement:scfoar}, and its prototypical implementation in SPASS-AR (\url{http://www.mpi-inf.mpg.de/fileadmin/inf/rg1/spass-ar.tgz}). It turns out that for clause sets containing certain structures, FO-AR is superior to ordered resolution/superposition~\cite{BachmairGanzinger94b} and instance generating methods~\cite{Korovin13ganzinger}. The paper ends with a discussion on challenges and future research directions, Section~\ref{sec:conclusion}.
\section{First-Order Clauses with Straight Dismatching Constraints: MSL(SDC)} \label{sec:prelim}
We consider a standard first-order language where letters $v,w,x,$ $y,z$ denote variables, $f,g,h$ functions, $a,b,c$ constants, $s,t$ terms, $p,q,r$ positions and Greek letters $\sigma,\tau,\rho,\delta$ are used for substitutions. $S,P,Q,R$ denote predicates, $\approx$ denotes equality, $A,B$ atoms, $E,L$ literals, $C,D$ clauses, $N$ clause sets and $\mathcal{V}$ sets of variables. $\overline L$ is the complement of $L$. The signature $\Sigma=(\mathcal{F},\mathcal{P})$ consists of two disjoint, non-empty, in general infinite sets of function and predicate symbols $\mathcal{F}$ and $\mathcal{P}$, respectively. The set of all \emph{terms} over variables $\mathcal{V}$ is $\mathcal{T}(\mathcal{F},\mathcal{V})$. If there are no variables, then terms, literals and clauses are called \emph{ground}, respectively. A \emph{substitution} $\sigma$ is denoted by pairs $\{x \mapsto t\}$ and its update at $x$ by $\sigma[ x \mapsto t]$. A substitution $\sigma$ is a \emph{grounding} substitution for $\mathcal{V}$ if $x\sigma$ is ground for every variable $x \in \mathcal{V}$.
The set of \emph{free} variables of an atom $A$ (term $t$) denoted by $\operatorname{vars}(A)$ ($\operatorname{vars}(t)$). A \emph{position} is a sequence of positive integers, where $\varepsilon$ denotes the empty position. As usual $t\vert_p = s$ denotes the subterm $s$ of $t$ at position $p$, which we also write as $t[s]_p$, and $t[p/s']$ then denotes the replacement of $s$ with $s'$ in $t$ at position $p$. These notions are extended to literals and multiple positions.
A predicate with exactly one argument is called \emph{monadic}. A term is \emph{complex} if it is not a variable and \emph{shallow} if it has at most depth one. It is called \emph{linear} if there are no duplicate variable occurrences. A literal, where every argument term is shallow, is also called \emph{shallow}. A variable and a constant are called \emph{straight}. A term $f(s_1,\ldots,s_n)$ is called \emph{straight}, if $s_1,\ldots,s_n$ are different variables except for at most one straight term $s_i$.
A \emph{clause} is a multiset of literals which we write as an implication $\Gamma \rightarrow \Delta$ where the atoms in the multiset $\Delta$ (the\emph{ succedent}) denote the positive literals and the atoms in the multiset $\Gamma$ (the \emph{antecedent}) the negative literals. We write $\square$ for the empty clause. If $\Gamma$ is empty we omit $\rightarrow$, e.g., we can write $P(x)$ as an alternative of $\rightarrow P(x)$. We abbreviate disjoint set union with sequencing, for example, we write $\Gamma,\Gamma' \rightarrow \Delta,L$ instead of $\Gamma \cup \Gamma' \rightarrow \Delta \cup \{L\}$. A clause $E,E,\Gamma \rightarrow \Delta$ is equivalent to $E,\Gamma \rightarrow \Delta$ and we call them equal \emph{modulo duplicate literal elimination}. If every term in $\Delta$ is shallow, the clause is called \emph{positive shallow}. If all atoms in $\Delta$ are linear and variable disjoint, the clause is called \emph{positive linear}. A clause $\Gamma \rightarrow \Delta$ is called an \emph{MSL} clause, if it is (i)~positive shallow and linear, (ii)~all occurring predicates are monadic, (iii)~no equations occur in $\Delta$, and (iv)~no equations occur in $\Gamma$ or $\Gamma = \{s\approx t\}$ and $\Delta$ is empty where $s$ and $t$ are not unifiable. \emph{MSL} is the first-order clause fragment consisting of MSL clauses. Clauses $\Gamma,s\approx t \rightarrow \Delta$ where $\Gamma$, $\Delta$ are non-empty and $s,t$ are not unifiable could be added to the MSL fragment without changing any of our results. Considering the superposition calculus, it will select $s\approx t$. Since the two terms are not unifiable, no inference will take place on such a clause and the clause will not contribute to the model operator. In this sense such clauses do not increase the expressiveness of the fragment.
An \emph{atom ordering} $\prec$ is an irreflexive, well-founded, total ordering on ground atoms. It is lifted to literals by representing $A$ and $\neg A$ as multisets $\{A\}$ and $\{A,A\}$, respectively. The multiset extension of the literal ordering induces an ordering on ground clauses. The clause ordering is compatible with the atom ordering; if the maximal atom in $C$ is greater than the maximal atom in $D$ then $D \prec C$. We use $\prec$ simultaneously to denote an atom ordering and its multiset, literal, and clause extensions. For a ground clause set $N$ and clause $C$, the set $N^{\prec C}=\{D \in N \mid D \prec C\}$ denotes the clauses of $N$ smaller than $C$.
A \emph{Herbrand interpretation} $\I{}$ is a - possibly infinite - set of ground atoms. A ground atom $A$ is called \emph{true} in $\I{}$ if $A\in\I{}$ and \emph{false}, otherwise. $\I{}$ is said to \emph{satisfy} a ground clause $C= \Gamma \rightarrow \Delta$, denoted by $\I{}\vDash C$, if $\Delta \cap \I{} \neq \emptyset$ or $\Gamma \not\subseteq \I{}$. A non-ground clause $C$ is satisfied by $\I{}$ if $\I{}\vDash C\sigma$ for every grounding substitution $\sigma$. An interpretation $\I{}$ is called a \emph{model} of $N$, $\I{}\vDash N$, if $\I{}\vDash C$ for every $C\in N$. A model $\I{}$ of $N$ is considered \emph{minimal} with respect to set inclusion, i.e., if there is no model $\I{}'$ with $\I{}'\subset \I{}$ and $\I{}'\vDash N$. A set of clauses $N$ is \emph{satisfiable}, if there exists a model that satisfies $N$. Otherwise, the set is \emph{unsatisfiable}.
A disequation $t \neq s$ is an \emph{atomic straight dismatching constraint} if $s$ and $t$ are variable disjoint terms and $s$ is straight. A straight dismatching constraint $\pi$ is a conjunction of atomic straight dismatching constraints. Given a substitution $\sigma$, $\pi\sigma= \bwedge{i\in I} ~~ t_i\sigma \neq s_i $. $\mathrm{lvar}(\pi) := \bigcup_{i \in I} \operatorname{vars}(t_i)$ are the left-hand variables of $\pi$ and the depth of $\pi$ is the maximal term depth of the $s_i$. A \emph{solution} of $\pi$ is a grounding substitution $\delta$ such that for all $i\in I$, $t_i\delta$ is not an instance of $s_i$, i.e., there exists no $\sigma$ such that $t_i\delta = s_i\sigma$. A dismatching constraint is solvable if it has a solution and unsolvable, otherwise. Whether a straight dismatching constraint is solvable, is decidable in linear-logarithmic time \cite{DBLP:conf/cade/TeuckeW16}. $\top$ and $\bot$ represent the true and false dismatching constraint, respectively.
We define constraint normalization $\norm{\pi}$ as the normal form of the following rewriting rules over straight dismatching constraints.
\shortrules{\ }{$ ~\pi \wedge f(t_1,\ldots,t_n)\neq y $~~~~~~~~~~~~~~~}{$\bot$}{}{}{10}
\shortrules{\ }{$ \pi \wedge f(t_1,\ldots,t_n)\neq f(y_1,\ldots,y_n) $}{$\bot$}{}{}{10}
\shortrules{\ }{$ \pi \wedge f(t_1,\ldots,t_n)\neq f(s_1,\ldots,s_n) $}{$\pi \wedge t_i \neq s_i$ ~~if $s_i$ is complex}{}{}{10}
\shortrules{\ }{$ \pi \wedge f(t_1,\ldots,t_n)\neq g(s_1,\ldots,s_m) $}{$\pi$}{}{}{10}
\shortrules{\ }{$\pi \wedge x\neq s \wedge x\neq s\sigma $~~~~~~~~~~~~~~~~~\;}{$\pi \wedge x\neq s $ }{}{}{10} Note that $f(t_1,\ldots,t_n)\neq f(s_1,\ldots,s_n)$ normalizes to $t_i \neq s_i$ for some $i$,
where $s_i$ is the one straight complex argument of $f(s_1,\ldots,s_n)$. Furthermore, the depth of $\norm{\pi}$ is less or equal to the depth of $\pi$ and both have the same solutions.
A pair of a clause and a constraint $(C;\pi)$ is called a \emph{constrained clause}. Given a substitution $\sigma$, $(C;\pi)\sigma = (C\sigma;\pi\sigma)$. $C\delta$ is called a ground clause of $(C;\pi)$ if $\delta$ is a solution of $\pi$. $\ground{(C;\pi)}$ is the set of ground instances of $(C;\pi)$. If $\ground{(C;\pi)}\subseteq \ground{(C';\pi')}$, then $(C;\pi)$ is an instance of $(C';\pi')$. If $\ground{(C;\pi)} = \ground{(C';\pi')}$, then $(C;\pi)$ and $(C';\pi')$ are called variants. A Herbrand interpretation $\I{}$ satisfies $(C;\pi)$, if $\I{}\vDash \ground{(C;\pi)}$. A constrained clause $(C;\pi)$ is called \emph{redundant} in $N$ if for every $D \in \ground{(C;\pi)}$, there exist $D_1,\ldots,D_n$ in $\ground{N}^{\prec D}$ such that $D_1,\ldots,D_n \vDash D$. A constrained clause $(C';\pi')$ is called a \emph{condensation} of $(C;\pi)$ if $C' \subset C $ and there exists a substitution $\sigma$ such that, $\pi\sigma= \pi'$, $\pi' \subseteq \pi$, and for all $L \in C$ there is an $L'\in C'$ with $L\sigma=L'$. A finite unsatisfiable subset of $\ground{N}$ is called an unsatisfiable core of $N$.
An MSL clause with straight dismatching constraints is called an \emph{MSL(SDC)} clause with MSL(SDC) being the respective first-order fragment. Note that any clause set $N$ can be transformed into an equivalent constrained clause set by changing each $C\in N$ to $(C;\top)$.
\section{Decidability of the MSL(SDC) fragment}\label{sec:decide}
In the following we will show that the satisfiability of the MSL(SDC) fragment is decidable. For this purpose we will define ordered resolution with selection on constrained clauses \cite{DBLP:conf/cade/TeuckeW16} and show that with an appropriate ordering and selection function, saturation of an MSL(SDC) clause set terminates.
For the rest of this section we assume an atom ordering $\prec$ such that a literal $\neg Q(s)$ is not greater than a literal $P(t[s]_p)$, where $p \neq \varepsilon$. For example, a KBO where all symbols have weight one has this property.
\begin{definition}[sel]\label{def:decide:sel} Given an MSL(SDC) clause $(C;\pi) = (S_1(t_1),\dots,S_n(t_n) \rightarrow P_1(s_1),\dots, P_m(s_m);\pi)$. The Superposition Selection function $\mathrm{sel}$ is defined by $S_i(t_i)\in \mathrm{sel}(C)$ if (1) $t_i$ is not a variable or (2) $t_1,\ldots,t_n$ are variables and $t_i \notin \operatorname{vars}(s_1,\dots,s_m)$ or (3) $\{ t_1,\ldots,t_n\} \subseteq \operatorname{vars}(s_1,\dots,s_m)$ and for some $1 \leq j \leq m$, $s_j=t_i$. \end{definition}
The selection function $\mathrm{sel}$ (Definition \ref{def:decide:sel}) ensures that a clause $\Gamma \rightarrow \Delta$ can only be resolved on a positive literal if $\Gamma$ contains only variables, which also appear in $\Delta$ at a non-top position. For example:\newline \centerline{$\begin{array}{r@{\,=\,}l} \mathrm{sel}(P(f(x)),P(x), Q(z) \rightarrow Q(x),R(f(y)) & \{P(f(x))\}\\ \mathrm{sel}(P(x), Q(z) \rightarrow Q(x),R(f(y))) & \{Q(z)\}\\ \mathrm{sel}(P(x), Q(y) \rightarrow Q(x),R(f(y))) & \{P(x)\}\\ \mathrm{sel}(P(x), Q(y) \rightarrow Q(f(x)),R(f(y))) & \emptyset.\\ \end{array}$} Note that given an MSL(SDC) clause $(C;\pi) = (S_1(t_1),\dots,S_n(t_n)$ $ \rightarrow P_1(s_1),\dots P_m(s_m);\pi)$, if some $S_i(t_i)$ is maximal in $C$, then at least one literal is selected.
\begin{definition} A literal $A$ is called \emph{[strictly] maximal} in a constrained clause $(C \vee A;\pi)$ if and only if there exists a solution $\delta$ of $\pi$ such that for all literals $B$ in $C$, $B\delta \preceq A\delta~ [ B\delta \prec A\delta ]$. \end{definition}
\begin{definition}[SDC-Resolution]\label{def:decide:resolution} $$ { \frac{( \Gamma_1 \rightarrow \Delta_1, A ~;~\pi_1) \qquad ( \Gamma_2 , B \rightarrow \Delta_2 ~;~\pi_2)}{ ((\Gamma_1,\Gamma_2 \rightarrow \Delta_1,\Delta_2)\sigma~; ~\norm{(\pi_1 \wedge \pi_2)\sigma}) }}~~~, \text{if} $$ \begin{tabular}{rlrl}
1. & $\sigma=\mathrm{mgu}(A,B)$ & 2. & $\norm{(\pi_1 \wedge \pi_2)\sigma}$ is solvable\\
3. & \multicolumn{3}{l}{$A\sigma$ is strictly maximal in $(\Gamma_1 \rightarrow \Delta_1, A;\pi_1)\sigma$ and $\mathrm{sel}(\Gamma_1 \rightarrow \Delta_1, A)=\emptyset$}\\
4. & $B \in \mathrm{sel}(\Gamma_2 , B \rightarrow \Delta_2)$ & & \\
5. & \multicolumn{3}{l}{$\mathrm{sel}(\Gamma_2 , B \rightarrow \Delta_2)=\emptyset$ and $\neg B\sigma$ maximal in $(\Gamma_2 , B \rightarrow \Delta_2;\pi_2)\sigma$}\\ \end{tabular} \end{definition}
\begin{definition}[SDC-Factoring]\label{def:decide:factoring} $$ {\frac{(\Gamma\rightarrow \Delta, A, B ~;~ \pi) }{ ((\Gamma\rightarrow \Delta, A)\sigma; \norm{\pi\sigma})}} ~~~, \text{if}$$ \begin{tabular}{rl@{$\quad$}rl}
1. & $\sigma=\mathrm{mgu}(A,B)$ &
2. & $\mathrm{sel}(\Gamma\rightarrow \Delta, A, B)=\emptyset$\\ 3. & $A\sigma$ is maximal in $(\Gamma\rightarrow \Delta, A, B;\pi)\sigma$ & 4. & $\norm{\pi\sigma}$ is solvable\\ \end{tabular} \end{definition}
Note that while the above rules do not operate on equations, we can actually allow unit clauses that consist of non-unifiable disequations, i.e., clauses $s \approx t \rightarrow$ where $s$ and $t$ are not unifiable. There are no potential superposition inferences on such clauses as long as there are no positive equations. So resolution and factoring suffice for completeness. Nevertheless, clauses such as $s \approx t \rightarrow$ affect the models of satisfiable problems. Constrained Resolution and Factoring are sound.
\begin{lemma}[Soundness]
SDC-Resolution and SDC-Factoring are sound. \end{lemma} \begin{proof}
Let $(\Gamma_1,\Gamma_2 \rightarrow \Delta_1,\Delta_2)\sigma\delta$ be a ground instance of $((\Gamma_1,\Gamma_2 \rightarrow \Delta_1,\Delta_2)\sigma; (\pi_1 \wedge \pi_2)\sigma)$.
Then, $\delta$ is a solution of $(\pi_1 \wedge \pi_2)\sigma$ and $\sigma\delta$ is a solution of $\pi_1$ and $\pi_2$.
Hence, $(\Gamma_1 \rightarrow \Delta_1, A)\sigma\delta$ and $(\Gamma_2 , B \rightarrow \Delta_2)\sigma\delta$ are ground instances of $(\Gamma_1 \rightarrow \Delta_1, A;\pi_1)$ and $(\Gamma_2 , B \rightarrow \Delta_2;\pi_2)$, respectively.
Because $A\sigma\delta= B \sigma\delta $, if $(\Gamma_1 \rightarrow \Delta_1, A)\sigma\delta$ and $(\Gamma_2 , B \rightarrow \Delta_2)\sigma\delta$ are satisfied, then $(\Gamma_1,\Gamma_2 \rightarrow \Delta_1,\Delta_2)\sigma\delta$ is also satisfied.
Therefore, SDC-Resolution is sound.
Let $(\Gamma\rightarrow \Delta, A)\sigma\delta$ be a ground instance of $((\Gamma\rightarrow \Delta, A)\sigma; \pi\sigma)$.
Then, $\delta$ is a solution of $\pi\sigma$ and $\sigma\delta$ is a solution of $\pi$.
Hence, $(\Gamma\rightarrow \Delta, A, B )\sigma\delta$ is a ground instance of $(\Gamma\rightarrow \Delta, A, B ;\pi)$.
Because $A\sigma\delta= B \sigma\delta $, if $(\Gamma\rightarrow \Delta, A, B )\sigma\delta$ is satisfied, then $(\Gamma\rightarrow \Delta, A)\sigma\delta$ is also satisfied.
Therefore, SDC-Factoring is sound.\qed \end{proof}
\begin{definition}[Saturation]\label{def:decide:saturated} A constrained clause set $N$ is called saturated up to redundancy, if for every inference between clauses in $N$ the result $(R;\pi)$ is either redundant in $N$ or $\ground{(R;\pi)} \subseteq \ground{N}$. \end{definition}
Note that our redundancy notion includes condensation and the condition $\ground{(R;\pi)} \subseteq \ground{N}$ allows ignoring variants of clauses.
\begin{lemma}\label{lem:decide:condensation}
Let constrained clause $(C';\pi')$ be a condensation of constrained clause $(C;\pi)$.
Then, (i)$(C;\pi) \vDash (C';\pi')$ and (ii)$(C;\pi)$ is redundant in $\{(C';\pi')\}$. \end{lemma}
\begin{proof}
Let $\sigma$ be a substitution such that $C' \subset C$, $\pi\sigma= \pi'$,
$\pi' \subseteq \pi$, and for all $L \in C$ there is a $L'\in C'$ with $L\sigma=L'$.
(i) Let $C'\delta \in \ground{(C';\pi')}$.
Then $\sigma\delta$ is a solution of $\pi$ and hence $C\sigma\delta \in \ground{(C;\pi)}$.
Let $\I{} \vDash C\sigma\delta$. Hence, there is a $L\sigma\delta \in \I{}$ for some $L \in C$ and thus $L'\delta \in \I{}$ for some $L' \in C'$ with $L\sigma=L'$.
Therefore, $\I{} \vDash C'\delta$. Since $\I{}$ and $C'\delta$ were arbitrary, $(C;\pi) \vDash (C';\pi')$.
(ii) Let $C\delta \in \ground{(C;\pi)}$. Because $\pi' \subseteq \pi$, $\delta$ is a solution of $\pi'$ and hence, $C'\delta \in \ground{(C';\pi')} $.
Therefore, since $C'\delta \subset C\delta$, $C'\delta \in \ground{\{(C';\pi')\}}^{\prec C\delta}$ and $C'\delta\vDash C\delta$.\qed \end{proof}
\begin{definition}[Partial Minimal Model Construction]\label{def:decide:model} Given a constrained clause set $N$, an ordering $\prec$ and the selection function $\mathrm{sel}$, we construct an interpretation $\I{N}$ for $N$, called a partial model, inductively as follows: \begin{align*}
\I{C} &:= \bigcup_{D\prec C}^{D \in \ground{N}} \delta_D, \text{ where }C \in \ground{N}\\
\delta_D &:= \left\{
\begin{array}{l l}
\{A\} & \quad \text{if $D=\Gamma \rightarrow \Delta, A$ }\\
& \quad \text{$A$ strictly maximal, $\mathrm{sel}(D)=\emptyset$ and $\I{D}\not \vDash D$}\\
\emptyset & \quad \text{otherwise}
\end{array} \right. \\ \I{N} &:= \bigcup_{C \in \ground{N}} \delta_C \end{align*} Clauses $D$ with $\delta_D \neq \emptyset$ are called productive. \end{definition}
\begin{lemma}[Ordered SDC Resolution Completeness]\label{lem:decide:complete}
Let $N$ be a constrained clause set saturated up to redundancy by ordered SDC-resolution with selection. Then $N$ is unsatisfiable, if and only if $\square\in \ground{N}$. If $\square\not\in \ground{N}$ then $\I{N}\models N$. \end{lemma}
\begin{proof}
Assume $N$ is unsatisfiable but $\square\not\in \ground{N}$.
For the partial model $\I{N}$, there exists a minimal false clause $C\sigma\in \ground{(C;\pi)}$ for some $(C;\pi)\in N$.
$C\sigma$ is not productive, because otherwise $\I{N} \vDash C\sigma$.
Hence, either $\mathrm{sel}(C)\neq \emptyset$ or no positive literal in $C\sigma$ is strictly maximal.
Assume $C= \Gamma_2, B \rightarrow \Delta_2$ with $B \in \mathrm{sel}(C)$ or $\neg B\sigma$ maximal.
Then, $B\sigma \in \I{C\sigma}$ and there exists a ground instance $(\Gamma_1 \rightarrow \Delta_1 , A)\tau= D\tau \prec C\sigma$ of some clause $(D;\pi')\in N$,
which produces $A\tau=B\sigma$.
Therefore, there exists a $\rho=\mathrm{mgu}(A,B)$ and ground substitution $\delta$
such that $C\sigma=C\rho\delta$, $D\tau=D\rho\delta$.
Since $\rho\delta=\sigma$ is a solution of $\pi$ and $\pi'$, $\delta$ is a solution of $(\pi \wedge \pi')\rho$.
Under these conditions, SDC-Resolution can be applied to $(\Gamma_1 \rightarrow \Delta_1 , A;\pi')$ and $(\Gamma_2, B \rightarrow \Delta_2;\pi)$.
Their resolvent $(R;\pi_R)=((\Gamma_1, \Gamma_2 \rightarrow \Delta_1, \Delta_2 )\rho;(\pi \wedge \pi')\rho)$ is either redundant in $N$ or $\ground{(R;\pi_R)} \subseteq \ground{N}$.
Its ground instance $R\delta$ is false in $\I{N}$ and $R\delta \prec C\sigma$.
If $(R;\pi_R)$ is redundant in $N$, there exist $C_1,\ldots,C_n$ in $\ground{N}^{\prec R\delta}$ with $C_1,\ldots,C_n \vDash R\delta$.
Because $C_i \prec R\delta \prec C\sigma$, $\I{N} \vDash C_i$ and hence $\I{N} \vDash R\delta$, which contradicts $\I{N} \not \vDash R\delta$.
Otherwise, if $\ground{(R;\pi_R)} \subseteq \ground{N}$, then $R\delta\in \ground{N}$, which contradicts $C\sigma$ being minimal false.
Now, assume $\mathrm{sel}(C)= \emptyset$ and $C= \Gamma \rightarrow \Delta, B$ with $B\sigma$ maximal.
Then, $C= \Gamma \rightarrow \Delta', A, B$ with $A\sigma=B\sigma$.
Therefore, there exists a $\rho=\mathrm{mgu}(A,B)$ and ground substitution $\delta$
such that $C\sigma=C\rho\delta$ and $\rho\delta$ is a solution of $\pi$.
Hence, $\delta$ is a solution of $\pi\rho$.
Under these conditions, SDC-Factoring can be applied to $(\Gamma \rightarrow \Delta', A, B;\pi)$.
The result $(R;\pi_R)=((\Gamma \rightarrow \Delta', A)\rho;\pi\rho)$
is either redundant in $N$ or $\ground{(R;\pi_R)} \subseteq \ground{N}$.
Its ground instance $R\delta$ is false in $\I{N}$ and $R\delta \prec C\sigma$.
If $(R;\pi_R)$ is redundant in $N$, there exist $C_1,\ldots,C_n$ in $\ground{N}^{\prec R\delta}$ with $C_1,\ldots,C_n \vDash R\delta$.
Because $C_i \prec R\delta \prec C\sigma$, $\I{N} \vDash C_i$ and hence $\I{N} \vDash R\delta$, which contradicts $\I{N} \not \vDash R\delta$.
Otherwise, if $\ground{(R;\pi_R)} \subseteq \ground{N}$, then $R\delta\in \ground{N}$, which contradicts $C\sigma$ being minimal false.
Therefore, if $\square\not\in \ground{N}$, no minimal false clause exists and $\I{N}\models N$. \qed \end{proof}
\begin{lemma}\label{lem:decide:finite_base} Let $N$ be a set of MSL(SDC) clauses without variants or uncondensed clauses over a finite signature $\Sigma$. $N$ is finite if there exists an integer $d$ such that for every $(C;\pi) \in N$, depth($\pi$)$\leq d$ and\\ (1) $C= S_1(x_1),\dots,S_n(x_n),S'_1(t),\dots,S'_m(t) \rightarrow \Delta$ or \\ (2) $C= S_1(x_1),\dots,S_n(x_n),S'_1(t),\dots,S'_m(t) \rightarrow S(t),\Delta$\\ with $t$ shallow and linear, and $\operatorname{vars}(t) \cap \operatorname{vars}(\Delta) = \emptyset$. \end{lemma}
\begin{proof}
Let $(C;\pi) \in N$.
$(C;\pi)$ can be separated into variable disjoint components $(\Gamma_1,\ldots,\Gamma_n \rightarrow \Delta_1,\ldots,\Delta_n; \pi_1 \wedge \ldots \wedge \pi_n)$,
where $\vert \Delta_i \vert \leq 1$ and $\mathrm{lvar}(\pi_i) \subseteq \operatorname{vars}(\Gamma_i \rightarrow \Delta_i)$.
For each positive literal $P(s) \in \Delta$ there is a fragment
\begin{align*}
(A)~~ & (S_1(x_1),\dots,S_k(x_k) \rightarrow P(s);\pi') \\
\intertext{with $\{x_1,\dots,x_k\}\subseteq \operatorname{vars}(s)$. If $m > 0$, there is another fragment}
(B)~~ & (S_1(x_1),\dots,S_k(x_k),S'_1(t),\dots,S'_m(t) \rightarrow ;\pi') \\
\intertext{ or }
(C)~~ & (S_1(x_1),\dots,S_k(x_k),S'_1(t),\dots,S'_m(t) \rightarrow S(t);\pi') \\
\intertext{ with $\{x_1,\dots,x_k\}\subseteq \operatorname{vars}(t)$, respectively.
Lastly, for each variable $x\in \operatorname{vars}(C)$ with $x \notin \operatorname{vars}(t) \cup \operatorname{vars}(\Delta)$ there is a fragment }
(D)~~ & (S_1(x),\dots,S_k(x)\rightarrow;\pi').
\end{align*}
Since there are only finitely many terms $s$ with depth($s$)$\leq d$ modulo renaming,
there are only finitely many atomic constraints $x \neq s$ for a given variable $x$ different up to renaming $s$.
Thus, a normal constraint can only contain finitely many combinations of subconstraints $\bwedge{i\in \mathcal{I}} ~ x \neq s_i$ without some $s_i$ being an instance of another $s_j$.
Therefore, for a fixed set of variables $x_1,\dots,x_k$, there are only finitely many constraints $\pi=\bwedge{i\in \mathcal{I}} ~~ z_i \neq s_i$ with $\mathrm{lvar}(\pi)\subseteq \{x_1,\dots,x_k\}$ up to variants.
Since the number of predicates, function symbols, and their ranks is finite,
the number of possible shallow and linear atoms $S(t)$ different up to variants is finite.
For a given shallow and linear $t$, there exist only finitely many clauses of the form $({S_{1}(t),\dots,S_{n}(t) \rightarrow S(t)}; \pi)$ or $(S_{1}(t),\dots,S_{n}(t) \rightarrow;\pi)$ with $\mathrm{lvar}(\pi)\subseteq \operatorname{vars}(t)$ modulo condensation and variants.
For a fixed set of variables $x_1,\dots,x_k$, there exist only finitely many clauses of the form $(S_1(y_1),\dots,S_k(y_l) \rightarrow; \pi ) $ with $\{y_1,\dots,y_l\} \cup \mathrm{lvar}(\pi)\subseteq \{x_1,\dots,x_k\}$ modulo condensation and variants.
Therefore, there are only finitely many distinct clauses of each form (A)-(D) without variants or condensations.
If in the clause $(C;\pi)=(\Gamma_1,\ldots,\Gamma_n \rightarrow \Delta_1,\ldots,\Delta_n; \pi_1 \wedge \ldots \wedge \pi_n)$ for some $i\neq j$, $(\Gamma_i \rightarrow \Delta_i;\pi_i)$ is a variant of $(\Gamma_j \rightarrow \Delta_j;\pi_j)$,
then $(C;\pi)$ has a condensation and is therefore not part of $N$.
Hence, there can be only finitely many different $(C;\pi)$ without variants or condensations and thus $N$ is finite. \qed \end{proof}
\begin{lemma}[Finite Saturation]\label{lem:decide:saturation}
Let $N$ be an MSL(SDC) clause set. Then $N$ can be finitely saturated up to redundancy by SDC-resolution with selection function $\mathrm{sel}$. \end{lemma}
\begin{proof}
The general idea is that given the way $\mathrm{sel}$ is defined the clauses involved in constrained resolution and factoring can only fall into certain patterns.
Any result of such inferences then is either strictly smaller than one of its parents by some terminating measure
or falls into a set of clauses that is bounded by Lemma~\ref{lem:decide:finite_base}.
Thus, there can be only finitely many inferences before $N$ is saturated.
Let $d$ be an upper bound on the depth of constraints found in $N$ and $\Sigma$ be the finite signature consisting of the function and predicate symbols occurring in $N$.
Let $(\Gamma_1 \rightarrow \Delta_1, S(t);\pi_1)$ and $(\Gamma_2 , S(t') \rightarrow \Delta_2;\pi_2)$ be clauses in $N$ where sdc-resolution applies with $\sigma=\mathrm{mgu}(S(t),S(t'))$ and resolvent $R=((\Gamma_1,\Gamma_2\rightarrow \Delta_1,\Delta_2)\sigma;\norm{(\pi_1 \wedge \pi_2)\sigma})$.
Because no literal is selected by $\mathrm{sel}$, $\Gamma_1 \rightarrow \Delta_1, S(t)$ can match only one of two patterns:
\begin{align*}
(A)~~ & S_1(x_1),\dots, S_n(x_n) \rightarrow S(f(y_1,\dots,y_k)),\Delta
\intertext{where $t=f(y_1,\dots,y_k)$ and $\{x_1,\dots,x_n\}\subseteq\{y_1,\dots,y_k\}\cup \operatorname{vars}(\Delta)$.}
(B) ~~ & S_1(x_1),\dots, S_n(x_n) \rightarrow S(y),\Delta
\intertext{where $t=y$ and $x_1,\dots,x_n$ are variables in $\operatorname{vars}(\Delta)$, i.e., $y$ occurs only once.}
\end{align*}
The literal $S(t')$ is selected by $\mathrm{sel}$ in $\Gamma_2 , S(t') \rightarrow \Delta_2$, and therefore $\Gamma_2 , S(t') \rightarrow \Delta_2$ can match only one of the following three patterns:
\begin{align*}
(1)~~ &S(f(t_1,\dots,t_k)),\Gamma' \rightarrow \Delta' \\
(2)~~ & S(y'),\Gamma' \rightarrow \Delta' \text{ where $\Gamma'$ has no function terms and $y\notin \operatorname{vars}(\Delta')$. }\\
(3)~~ & S(y'),\Gamma' \rightarrow S'(y'),\Delta' \text{ where $\Gamma'$ has no function terms. }
\end{align*}
This means that the clausal part $(\Gamma_1,\Gamma_2\rightarrow \Delta_1,\Delta_2)\sigma$ of $R$ has one of six forms:
\begin{align*}
(A1)~~& S_1(x_1)\sigma,\dots, S_n(x_n)\sigma, \Gamma' \rightarrow \Delta,\Delta' \text{ with $\sigma=\{{y_1 \mapsto t_1},\dots \}$.}\\
\intertext{ $\Delta\sigma = \Delta$ because $S(f(y_1,\dots,y_k))$ and $\Delta$ do not share variables.}
(B1)~~& S_1(x_1),\dots, S_n(x_n), \Gamma' \rightarrow \Delta,\Delta'.\\
\intertext {The substitution $\{y \mapsto f(t_1,\dots,t_k) \}$ is irrelevant since $S(y)$ is the only literal with variable $y$.}
(A2)~~& S_1(x_1),\dots, S_n(x_n), \Gamma'\tau \rightarrow \Delta,\Delta' \text{ with $\tau=\{{y' \mapsto f(y_1,\dots,y_k)} \}$.}\\
\intertext{ $\Delta'\tau=\Delta'$ because $y' \notin \operatorname{vars}(\Delta')$.}
(B2)~~& S_1(x_1),\dots, S_n(x_n), \Gamma' \rightarrow \Delta,\Delta'. \\
(A3)~~& S_1(x_1),\dots, S_n(x_n), \Gamma'\tau \rightarrow S'(f(y_1,\dots,y_k)),\Delta,\Delta' \text{ with $\tau=\{y \mapsto f(y_1,\dots,y_k) \}$.}\\
\intertext{ $\Delta'\tau=\Delta'$ because $y' \notin \operatorname{vars}(\Delta')$.}
(B3)~~& S_1(x_1),\dots, S_n(x_n), \Gamma' \rightarrow S'(y'),\Delta,\Delta'.
\end{align*}
In the constraint $\norm{(\pi_1 \wedge \pi_2) \sigma}$ the maximal depth of the subconstraints is less or equal to the maximal depth of $\pi_1$ or $\pi_2$.
Hence, $d$ is also an upper bound on the constraint of the resolvent.
In each case, the resolvent is again an MSL(SDC) clause.
In the first and second case, the multiset of term depths of the negative literals in $R$ is strictly smaller than for the right parent.
In both, the $\Gamma$ is the same between the right parent and the resolvent.
Only the $f(t_1,\dots,t_k)$ term is replaced by $x_1\sigma,\dots, x_n\sigma$ and $x_1,\dots, x_n$ respectively.
In the first case, the depth of the $x_i\sigma$ is either zero if $x_i\notin \{y_1,\dots,y_k\}$ or at least one less than $f(t_1,\dots,t_k)$ since $x_i\sigma=t_i$.
In the second case, the $x_i$ have depth zero which is strictly smaller than the depth of $f(t_1,\dots,t_k)$.
Since the multiset ordering on natural numbers is terminating, the first and second case can only be applied finitely many times by constrained resolution.
In the third to sixth case $R$ is of the form
$(S_1(x_1),\dots,S_l(x_l),S'_1(t),\dots,S'_m(t) \rightarrow \Delta;\pi)$ or
$(S_1(x_1),\dots,S_l(x_l),S'_1(t),\dots,S'_m(t) \rightarrow S(t)),\Delta;\pi)$ with $t=f(y_1,\dots,y_k)$.
By Lemma \ref{lem:decide:finite_base}, there are only finitely many such clauses after condensation and removal of variants.
Therefore, these four cases can apply only finitely many times during saturation.
Let $(\Gamma \rightarrow \Delta, S(t), S(t');\pi)$ be a clause in $N$ where sdc-factoring applies with $\sigma=\mathrm{mgu}(S(t),S(t'))$ and $R=((\Gamma \rightarrow \Delta, S(t))\sigma;\norm{\pi\sigma})$.
Because in $\Gamma \rightarrow \Delta, S(t), S(t')$ no literal is selected, $\Gamma \rightarrow \Delta, S(t), S(t')$ and $(\Gamma \rightarrow \Delta, S(t))\sigma$ can only match one of three patterns.
\begin{align*}
(A)~~& S_1(x_1),\dots, S_n(x_n) \rightarrow S(f(y_1,\dots,y_k)),S(f(z_1,\dots,z_l)),\Delta \\
\intertext {where $t=f(y_1,\dots,y_k)$, $t'=f(z_1,\dots,z_k)$, and $\{x_1,\dots,x_n\}\subseteq\{y_1,\dots,y_k\}\cup\{z_1,\dots,z_l\}\cup \operatorname{vars}(\Delta)$. The result is}
& S_1(x_1)\sigma,\dots, S_n(x_n)\sigma \rightarrow S(f(y_1,\dots,y_k)),\Delta \text{ with $\sigma=\{{z_1 \mapsto y_1},\dots \}$.}\\
(B)~~& S_1(x_1),\dots, S_n(x_n) \rightarrow S(f(y_1,\dots,y_k)),S(z),\Delta \\
\intertext{ where $t=f(y_1,\dots,y_k)$, $t'=z$ and $\{x_1,\dots,x_n\}\subseteq\{y_1,\dots,y_k\}\cup\operatorname{vars}(\Delta)$, i.e., $z$ occurs only once. The result is}
& S_1(x_1),\dots, S_n(x_n) \rightarrow S(f(y_1,\dots,y_k)),\Delta.\\
(C)~~& S_1(x_1),\dots, S_n(x_n) \rightarrow S(y),S(z),\Delta \\
\intertext{ where $t=y$, $t'=z$ and $\{x_1,\dots,x_n\}\subseteq\operatorname{vars}(\Delta)$, i.e., $y$ and $z$ occur only once. The result is }
& S_1(x_1),\dots, S_n(x_n) \rightarrow S(y),\Delta.
\end{align*}
In the new constraint $\norm{\pi \sigma}$ the maximal depth of the subconstraints is less or equal to the maximal depth of $\pi$.
Hence $d$ is also an upper bound on the constraint of the resolvent.
In each case, the resolvent is again an MSL(SDC) clause.
Furthermore, in each case the clause is of the form $(S_1(x_1),\dots,S_l(x_l) \rightarrow \Delta; \pi)$.
By Lemma \ref{lem:decide:finite_base}, there are only finitely many such clauses after condensation and removal of variants.
Therefore, these three cases can apply only finitely many times during saturation.\qed \end{proof}
\begin{theorem}[MSL(SDC) Decidability]\label{theo:decide:main} Satisfiability of the MSL(SDC) first-order fragment is decidable. \end{theorem}
\begin{proof}
Follows from Lemma \ref{lem:decide:saturation} and \ref{lem:decide:complete}. \end{proof}
\section{Approximation and Refinement}\label{sec:approx}
In the following, we show how decidability of the MSL(SDC) fragment can be used to improve the approximation refinement calculus presented in \cite{Teucke2015}.
Our approach is based on a counter-example guided abstraction refinement (CEGAR) idea. The procedure loops trough four steps: approximation, testing (un)satisfiability, lifting, and refinement. The approximation step transforms any first-order logic clause set into the decidable MSL(SDC) fragment while preserving unsatisfiability. The second step employs the decidability result for MSL(SDC), Section~\ref{sec:decide}, to test satisfiability of the approximated clause set. If the approximation is satisfiable, the original problem is satisfiable as well and we are done. Otherwise, the third step, lifting, tests whether the proof of unsatisfiability found for the approximated clause set can be lifted to a proof of the original clause set. If so, the original clause set is unsatisfiable and we are again done. If not, we extract a cause for the lifting failure that always amounts to two different
instantiations of the same variable in a clause from the original clause set. This is resolved by the fourth step, the refinement. The crucial clause in the original problem is replaced and instantiated in a satisfiability preserving way such that the different instantiations do not reoccur anymore in subsequent iterations of the loop.
As mentioned before, our motivation to use dismatching constraints is that for an unconstrained clause the refinement adds quadratically many new clauses to the clause set. In contrast, with constrained clauses the same can be accomplished with adding just a single new clause. This extension is rather simple as constraints are treated the same as the antecedent literals in the clause. Furthermore we present refinement as a separate transformation rule.
The second change compared to the previous version is the removal of the Horn approximation rule, where we have now shown in Section~\ref{sec:decide} that a restriction to Horn clauses is not required for decidability anymore. Instead, the linear and shallow approximations are extended to apply to non-Horn clauses instead.
The approximation consists of individual transformation rules $N \Rightarrow N'$ that are non-deterministically applied. They transform a clause that is not in the MSL(SDC) fragment in finite steps into MSL(SDC) clauses. Each specific property of MSL(SDC) clauses, i.e, monadic predicates, shallow and linear positive literals, is generated by a corresponding rule: the Monadic transformation encodes non-Monadic predicates as functions, the shallow transformation extracts non-shallow subterms by introducing fresh predicates and the linear transformation renames non-linear variable occurrences.
Starting from a constrained clause set $N$ the transformation is parameterized by a single monadic projection predicate $T$, fresh to $N$ and for each non-monadic predicate $P$ a separate projection function $f_P$ fresh to $N$. The clauses in $N$ are called the original clauses while the clauses in $N'$ are the approximated clauses. We assume all clauses in $N$ to be variable disjoint.
\begin{definition}\label{def:approx:termencod} Given a predicate $P$, projection predicate $T$, and projection function $f_P$, define the injective function $\Proj{P}^T(P(\args{t})) := T(f_p(\args{t}))$ and $\Proj{P}^T(Q(\args{s})) := Q(\args{s})$ for $P \neq Q$. The function is extended to [constrained] clauses, clause sets and interpretations. Given a signature $\Sigma$ with non-monadic predicates $P_1,\ldots,P_n$, define $\Proj{\Sigma}^T(N):=\Proj{P_1}^T(\ldots(\Proj{P_n}^T(N))\ldots)$ and $\Proj{\Sigma}^T(\I{}):=\Proj{P_1}^T(\ldots(\Proj{P_n}^T(\I{}))\ldots)$. \end{definition}
\shortrules{Monadic}{$N$}{$\Proj{P}^T(N)$}{provided $P$ is a non-monadic predicate in the signature of $N$.}{MO}{15}
\shortrules{Shallow}{$N~\dot{\cup}~\{(\Gamma \rightarrow E[s]_{p},\Delta;\pi)\}$}{\\ $~~~~~~~~~~~~~~~~~~~~~~~ N\cup\{(S(x),\Gamma_l \rightarrow E[p/x],\Delta_l;\pi)$; $(\Gamma_r \rightarrow S(s),\Delta_r;\pi)\}$} {provided $s$ is complex, $\vert p\vert=2$, $x$ and $S$ fresh, $\Gamma_l\{x \mapsto s\} \cup \Gamma_r = \Gamma$, $\Delta_l {\cup} \Delta_r = \Delta$, $\{Q(y)\in \Gamma \mid {y \in \operatorname{vars}(E[p/x],\Delta_l)\}} \subseteq \Gamma_l$, $\{Q(y)\in \Gamma \mid {y \in \operatorname{vars}(s,\Delta_r) \}} \subseteq \Gamma_r$.}{SH}{15}
\shortrules{Linear 1}{$N~\dot{\cup}~\{(\Gamma \rightarrow \Delta, E'[x]_{p},E[x]_q;\pi)\}$}{\\ $~~~~~~~~~~~~~~~~~~~~~~\;N\cup\{(\Gamma\sigma,\Gamma \rightarrow \Delta, E'[x]_{p},E[q/x'];\pi \wedge \pi\sigma)\}$} {provided $x'$ is fresh and $\sigma= \{x \mapsto x'\}$.}{LI}{15}
\shortrules{Linear 2}{$N~\dot{\cup}~\{(\Gamma \rightarrow \Delta, E[x]_{p,q};\pi)\}$}{\\ $~~~~~~~~~~~~~~~~~~~~~~\;N\cup\{(\Gamma\sigma,\Gamma \rightarrow \Delta, E[q/x'];\pi \wedge \pi\sigma)\}$} {provided $x'$ is fresh, $p \neq q$ and $\sigma= \{x \mapsto x'\}$.}{LI}{15}
\shortrules{Refinement}{$N~\dot{\cup}~\{(C,\pi)\} $} {$N \cup \{(C;\pi \wedge x \neq t),(C;\pi)\{x \mapsto t\}\}$} {provided $x\in \operatorname{vars}(C)$, $t$ straight and $\operatorname{vars}(t) \cap \operatorname{vars}((C,\pi))=\emptyset$. }{ \operatorname{Ref}}{15}
Note that variables are not renamed unless explicitly stated in the rule. This means that original clauses and their approximated counterparts share variable names. We use this to trace the origin of variables in the approximation.
The refinement transformation $ \Rightarrow_{\operatorname{Ref}}$ is not needed to eventually generate MSL(SDC) clauses, but can be used to achieve a more fine-grained approximation of $N$, see below.
In the shallow transformation, $\Gamma$ and $\Delta$ are separated into $\Gamma_l$, $\Gamma_r$, $\Delta_l$, and $\Delta_r$, respectively. The separation can be almost arbitrarily chosen as long as no atom from $\Gamma$, $\Delta$ is skipped. However, the goal is to minimize the set of shared variables, i.e., the variables of $(\Gamma \rightarrow E[s]_{p},\Delta;\pi)$ that are inherited by both approximation clauses, $\operatorname{vars}(\Gamma_r,s,\Delta_r) \cap \operatorname{vars}(\Gamma_l,E[p/x],\Delta_l)$. If there are no shared variables, the shallow transformation is satisfiability equivalent. The conditions on $\Gamma_l$ and $\Gamma_r$ ensure that $S(x)$ atoms are not separated from the respective positive occurrence of $x$ in subsequent shallow transformation applications.
Consider the clause $Q(f(x),y) \rightarrow P(g(f(x),y))$. The simple shallow transformation $S(x'),Q(f(x),y) \rightarrow P(g(x',y)); S(f(x))$ is not satisfiability equivalent -- nor with any alternative partitioning of $\Gamma$. However, by replacing the occurrence of the extraction term $f(x)$ in $Q(f(x),y)$ with the fresh variable $x'$, the approximation $S(x'),Q(x',y) \rightarrow P(g(x',y)); S(f(x))$ is satisfiability equivalent. Therefore, we allow the extraction of $s$ from the terms in $\Gamma_l$ and require $\Gamma_l\{x \mapsto s\} \cup \Gamma_r = \Gamma$.
We consider Linear~1 and Linear~2 as two cases of the same linear transformation rule. Their only difference is whether the two occurrences of $x$ are in the same literal or not. The duplication of literals and constraints in $\Gamma$ and $\pi$ is not needed if $x$ does not occur in $\Gamma$ or $\pi$.
Further, consider a linear transformation $N \cup \{(C;\pi)\} \Rightarrow_{\operatorname{LI}} N \cup \{(C_a;\pi_a)\}$, where a fresh variable $x'$ replaces an occurrence of a non-linear variable $x$ in $(C;\pi)$. Then, $(C_a;\pi_a)\{x' \mapsto x\}$ is equal to $(C;\pi)$ modulo duplicate literal elimination. A similar property can be observed of a resolvent of $(C_l;\pi)$ and $(C_r;\pi)$ resulting from a shallow transformation $N \cup \{(C;\pi)\} \Rightarrow_{\operatorname{SH}} N \cup \{(C_l;\pi), (C_r;\pi)\}$. Note that by construction, $(C_l;\pi)$ and $(C_r;\pi)$ are not necessarily variable disjoint. To simulate standard resolution, we need to rename at least the shared variables in one of them.
\begin{definition}[$ \Rightarrow_{\operatorname{AP}}$] \label{def:approx:apr} We define $ \Rightarrow_{\operatorname{AP}}$ as the priority rewrite system~\cite{Baeten89tcs} consisting of $ \Rightarrow_{\operatorname{Ref}}$, $\Rightarrow_{\operatorname{MO}}$, $\Rightarrow_{\operatorname{SH}}$ and $\Rightarrow_{\operatorname{LI}}$ with priority $ \Rightarrow_{\operatorname{Ref}} \,>\, \Rightarrow_{\operatorname{MO}} \,>\, \Rightarrow_{\operatorname{SH}} \,>\, \Rightarrow_{\operatorname{LI}}$, where $ \Rightarrow_{\operatorname{Ref}}$ is only applied finitely many times. \end{definition}
\begin{lemma}[$ \Rightarrow_{\operatorname{AP}}$ is a Terminating Over-Approximation]\label{lem:approx:sound} (i)~$ \Rightarrow_{\operatorname{AP}}^*$ terminates, (ii) if $N \Rightarrow_{\operatorname{AP}} N'$ and $N'$ is satisfiable, then $N$ is also satisfiable. \end{lemma}
\begin{proof}
(i)~The transformations can be considered sequentially, because of the imposed rule priority.
There are, by definition, only finitely many refinements at the beginning of an approximation $ \Rightarrow_{\operatorname{AP}}^*$.
The monadic transformation strictly reduces the number of non-monadic atoms.
The shallow transformation strictly reduces the multiset of term depths of the newly introduced clauses compared
to the removed parent clause.
The linear transformation strictly reduces the number of duplicate variable occurrences in positive literals.
Hence $ \Rightarrow_{\operatorname{AP}}$ terminates.
(ii) Let $N \cup \{ (C;\pi)\} \Rightarrow_{\operatorname{LI}} N \cup \{ (C_a;\pi_a)\}$ where an occurrence of a variable $x$ in $(C;\pi)$ is replaced by a fresh $x'$.
As $(C_a;\pi_a)\{x' \mapsto x\}$ is equal to $(C;\pi)$ modulo duplicate literal elimination,
$\I{} \models (C;\pi)$ if $\I{} \models (C_a;\pi_a)$.
Therefore, the linear transformation is an over-approximation.
Let $N \cup \{ (C;\pi)\} \Rightarrow_{\operatorname{SH}} N \cup \{ (C_l;\pi_l),(C_r;\pi_r)\}$
and $(C_a;\pi_a)$
be the shallow $\rho$-resolvent.
As $(C_a;\pi_a)\rho^{-1}$ equals $(C;\pi)$ modulo duplicate literal elimination,
$\I{} \models (C;\pi)$ if $\I{} \models (C_l;\pi_l), (C_r;\pi_r)$.
Therefore, the shallow transformation is an over-approximation.
Let $N \Rightarrow_{\operatorname{MO}} \Proj{P}(N)=N'$.
Then, $N=\R{P}(N') $.
Let $\I{}$ be a model of $N'$ and $(C;\pi) \in N$.
Since $ \Proj{P}((C;\pi)) \in N'$ , $\I{} \vDash \Proj{P}((C;\pi))$ and thus, $\R{P}(\I{})\vDash (C;\pi)$.
Hence, $\R{P}(\I{})$ is a model of $N$.
Therefore, the monadic transformation is an over-approximation. Actually, it
is a satisfiability preserving transformation.
Let $N \cup \{(C;\pi)\} \Rightarrow_{\operatorname{Ref}} N \cup \{(C;\pi \wedge x \neq t),(C;\pi)\{x \mapsto t\}\}$.
Let $C\delta \in \ground{(C;\pi)} $.
If $x\delta$ is not an instance of $t$, then $\delta$ is a solution of $\pi \wedge x \neq t$ and $C\delta \in \ground{(C;\pi \wedge x \neq t)}$.
Otherwise, $\delta=\{x\mapsto t\}\delta'$ for some substitution $\delta'$.
Then, $\delta$ is a solution of $\pi\{x\mapsto t\}$ and thus, $C\delta=C\{x\mapsto t\}\delta' \in \ground{(C\{x \mapsto t\};\pi\{x \mapsto t\})}$.
Hence, $ \ground{(C;\pi)} \subseteq \ground{(C;\pi \wedge x \neq t)} \cup \ground{(C;\pi)\{x \mapsto t\}}.$
Therefore, if $\I{}$ is a model of $N \cup \{(C;\pi \wedge x \neq t),(C;\pi)\{x \mapsto t\}\}$, then $\I{}$ is also a model of $N \cup \{(C;\pi)\}$. \qed \end{proof}
Note that $ \Rightarrow_{\operatorname{Ref}}$ and $ \Rightarrow_{\operatorname{MO}}$ are also satisfiability preserving transformations.
\begin{corollary}\label{cor:approx:sound} If $N \Rightarrow_{\operatorname{AP}}^* N'$ and $N'$ is satisfied by a model $\I{}$, then $\R{\Sigma}(\I{})$ is a model of $N$. \end{corollary} \begin{proof}
Follows from Lemma~\ref{lem:approx:sound} (ii)-(v).\qed \end{proof}
On the basis of $ \Rightarrow_{\operatorname{AP}}$ we can define an ancestor relation $ \Rightarrow_{\operatorname{A}}$ that relates clauses, literal occurrences, and variables with respect to approximation. This relation is needed in order to figure out the exact clause, literal, variable for refinement.
\begin{definition}[The Shallow Resolvent]\label{def:approx:resolvent}
Let $N \cup \{(C;\pi)\} \Rightarrow_{\operatorname{SH}} N \cup \{(C_l;\pi), (C_r;\pi)\}$ with $C=\Gamma \rightarrow E[s]_{p},\Delta$, $C_l=S(x),\Gamma_l \rightarrow E[p/x],\Delta_l$ and $C_r= \Gamma_r \rightarrow S(s),\Delta_r$.
Let $x_1,\ldots,x_n$ be the variables shared between $C_l$ and $C_r$ and $\rho=\{x_1 \mapsto x'_1, \ldots, x_n \mapsto x'_n\}$ be a variable renaming with $x'_1,\ldots,x'_n$ fresh in $C_l$ and $C_r$.
We define $(\Gamma_l\{x \mapsto s\rho \},\Gamma_r\rho \rightarrow E[p/s\rho],\Delta_l,\Delta_r\rho;\pi \wedge \pi\rho)$ as the shallow $\rho$-resolvent. \end{definition}
Let $(C_a;\pi_a)$ be the shallow $\rho$-resolvent of $N \cup \{(C;\pi)\} \Rightarrow_{\operatorname{SH}} N \cup \{(C_l;\pi), (C_r;\pi)\}$. Note that for any two ground instances $C_l\delta_l$ and $C_r\delta_r$, their resolvent is a ground instance of $(C_a;\pi_a)$. Furthermore, using the reverse substitution $\rho^{-1}= \{x'_1 \mapsto x_1, \ldots, x'_n \mapsto x_n\}$, $(C_a;\pi_a)\rho^{-1}= (\Gamma_l\{x \mapsto s \},\Gamma_r \rightarrow E[s]_{p},\Delta_l,\Delta_r;\pi \wedge \pi)$ is equal to $(C;\pi)$ modulo duplicate literal elimination. This is because, $\Delta_l \cup \Delta_r = \Delta$ and $\Gamma_l\{x \mapsto s \} \cup \Gamma_r = \Gamma$ by definition of $ \Rightarrow_{\operatorname{SH}}$ and $\pi \wedge \pi$ is equivalent to $ \pi$.
Next, we establish parent relations that link original and approximated clauses, as well as their variables and literals. Together the parent, variable and literal relations will allow us to not only trace any approximated clause back to their origin, but also predict what consequences changes to the original set will have on its approximations.
For the following definitions, we assume that clause and literal sets are lists and that $\Proj{P}^T$ and substitutions act as mappings. This means we can uniquely identify clauses and literals by their position in those lists. Further, for every shallow transformation $N \Rightarrow_{\operatorname{SH}} N'$, we will also include the shallow resolvent in the parent relation as if it were a member of $N'$.
\begin{definition}[Parent Clause]\label{def:approx:pclause}
For an approximation step $N \Rightarrow_{\operatorname{AP}} N'$ and two clauses $(C;\pi)\in N$ and $(C';\pi')\in N'$,
we define $[(C;\pi), N] \Rightarrow_{\operatorname{A}} [(C';\pi'), N']$ expressing that $(C;\pi)$ in $N$ is the parent clause of $(C';\pi')$ in $N'$:\\
If $N \Rightarrow_{\operatorname{MO}} \Proj{P}^T(N)$, then
$[(C;\pi),N] \Rightarrow_{\operatorname{A}} [\Proj{P}^T((C;\pi)), \Proj{P}^T(N)]$ for all $(C;\pi) \in N$.\\
If $N=N'' \cup \{(C;\pi)\} \Rightarrow_{\operatorname{SH}} N'' \cup \{(C_l;\pi_l),(C_r;\pi_r)\}=N'$, then
$[(D,\pi'),N] \Rightarrow_{\operatorname{A}} [(D,\pi'),N']$ for all $(D,\pi') \in N''$ and
$[(C,\pi),N] \Rightarrow_{\operatorname{A}} [(C_l;\pi_l),N']$,
$[(C,\pi),N] \Rightarrow_{\operatorname{A}} [(C_r;\pi_r),N']$ and
$[(C,\pi),N] \Rightarrow_{\operatorname{A}} [(C_a;\pi_a),N']$ for any shallow resolvent $(C_a;\pi_a)$.\\
If $N=N'' \cup \{(C;\pi)\} \Rightarrow_{\operatorname{LI}} N'' \cup \{(C_a;\pi_a)\}=N'$, then
$[(D,\pi'),N] \Rightarrow_{\operatorname{A}} [(D,\pi'),N']$ for all $(D,\pi') \in N''$ and
$[(C,\pi),N] \Rightarrow_{\operatorname{A}} [(C_a,\pi_a),N']$. \\
If $N=N'' \cup \{(C;\pi)\} \Rightarrow_{\operatorname{Ref}} N'' \cup \{(C;\pi \wedge x \neq t),(C;\pi)\{x \mapsto t\}\}=N'$, then
$[(D,\pi'),N] \Rightarrow_{\operatorname{A}} [(D,\pi'),N']$ for all $(D,\pi') \in N''$ ,
$[(C,\pi),N] \Rightarrow_{\operatorname{A}} [(C;\pi \wedge x \neq t),N']$ and
$[(C,\pi),N] \Rightarrow_{\operatorname{A}} [(C;\pi)\{x \mapsto t\}, N'] $. \end{definition}
\begin{definition}[Parent Variable]\label{def:approx:pvar}
Let $N \Rightarrow_{\operatorname{AP}} N'$ be an approximation step and $[(C;\pi),N] \Rightarrow_{\operatorname{A}} [(C';\pi'),N']$.
For two variables $x$ and $y$,
we define $[x,(C;\pi), N] \Rightarrow_{\operatorname{A}} [y,(C';\pi'), N']$ expressing that $x \in \operatorname{vars}(C)$ is the parent variable of $y \in \operatorname{vars}(C')$:\\
If $x\in \operatorname{vars}((C;\pi))\cap \operatorname{vars}((C';\pi'))$, then
$[x,(C;\pi),N] \Rightarrow_{\operatorname{A}} [x,(C';\pi'),N']$.\\
If $N \Rightarrow_{\operatorname{SH}} N'$ and $(C',\pi')$ is the shallow $\rho$-resolvent,
$[x_i,(C;\pi),N] \Rightarrow_{\operatorname{A}} [x_i\rho,(C';\pi'),N']$ for each $x_i$ in the domain of $\rho$.\\
If $N \Rightarrow_{\operatorname{LI}} N'$, $C= \Gamma \rightarrow \Delta[x]_{p,q}$ and $C'=\Gamma\{x \mapsto x'\},\Gamma \rightarrow \Delta[q/x']$, then
$[x,(C;\pi),N] \Rightarrow_{\operatorname{A}} [x',(C';\pi'),N']$. \end{definition}
Note that if $N \Rightarrow_{\operatorname{SH}} N'$ and $x$ is the fresh extraction variable in $(C_l;\pi_l)$, then $x$ has no parent variable. For literals, we actually further specify the relation on the positions within literals of a clause $(C;\pi)$ using pairs $(L,r)$ of literals and positions. We write $(L,r)\in C$ to denote that $(L,r)$ is a literal position in $(C;\pi)$ if $L\in C$ and $r\in \mathrm{pos}(L)$. Note that a literal position $(L,r)$ in $(C;\pi)$ corresponds to the term $L\vert_r$.
\begin{definition}[Parent literal position]\label{def:approx:pterm}
Let $N \Rightarrow_{\operatorname{AP}} N'$ be an approximation step and $[(C;\pi),N] \Rightarrow_{\operatorname{A}} [(C';\pi'),N']$.
For two literal positions $(L,r)$ and $(L',r')$,
we define $[r,L,(C;\pi), N] \Rightarrow_{\operatorname{A}} [r',L',(C';\pi'), N']$ expressing that $(L,r)$ in $(C;\pi)$ is the parent literal position of $(L',r')$ in $(C';\pi')$:\\
If $(C;\pi)=(C';\pi')$, then
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in C$.\\
If $N \Rightarrow_{\operatorname{Ref}} N'$ and $(C',\pi')=(C;\pi \wedge x \neq t)$, then
$[r,L,(C;\pi),N ] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in C$.\\
If $N \Rightarrow_{\operatorname{Ref}} N'$ and $(C',\pi')=(C;\pi)\{x \mapsto t\}$, then
$[r,L,(C;\pi),N ] \Rightarrow_{\operatorname{A}} [ r,L\{x \mapsto t\},(C';\pi'),N']$ for all $(L,r)\in C$.\\
If $N \Rightarrow_{\operatorname{MO}} \Proj{P}^T(N)=N'$, then
$[\varepsilon,P(\args{t}),(C;\pi),N] \Rightarrow_{\operatorname{A}} [ \varepsilon,T(f_p(\args{t})),(C';\pi'),N']$ for all $P(\args{t})\in C$ and
$[r,P(\args{t}),(C;\pi),N] \Rightarrow_{\operatorname{A}} [ 1.r,T(f_p(\args{t})),(C';\pi'),N']$ for all $(P(\args{t}),r)\in C$.\\
If $N \Rightarrow_{\operatorname{SH}} N'$, $C= \Gamma \rightarrow E[s]_{p},\Delta$ and $C'=S(x),\Gamma_l \rightarrow E[p/x],\Delta_l$, then
$[r,E[s]_{p},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,E[p/x],(C';\pi'),N']$ for all $r\in \mathrm{pos}(E[p/x])$,
$[p,E[s]_{p},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,S(x),(C';\pi'),N']$ for all $r\in \mathrm{pos}(S(x))$,
$[r,L\{x \mapsto s\},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in \Gamma_l$,
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in \Delta_l$.\\
If $N \Rightarrow_{\operatorname{SH}} N'$, $C= \Gamma \rightarrow E[s]_{p},\Delta$ and $C'=\Gamma_r \rightarrow S(s),\Delta_r$, then
$[p,E[s]_{p},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ \varepsilon ,S(s),(C';\pi'),N']$,
$[pr,E[s]_{p},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ 1.r,S(s),(C';\pi'),N']$ for all $r\in \mathrm{pos}(s)$, and
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in \Gamma_r\cup\Delta_r$.\\
If $N \Rightarrow_{\operatorname{SH}} N'$, $C= \Gamma \rightarrow E[s]_{p},\Delta$ and $(C',\pi')$ is the shallow $\rho$-resolvent, then
$[r,E[s]_{p},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,E[p/s\rho],(C';\pi'),N']$ for all $r\in \mathrm{pos}(E[p/s\rho])$,
$[r,L\{x \mapsto s\},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L\{x \mapsto s\rho \},(C';\pi'),N']$ for all $(L,r)\in \Gamma_l$,
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L\rho,(C';\pi'),N']$ for all $(L,r)\in \Gamma_r\cup\Delta_r$, and
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in \Delta_l$.\\
If $N \Rightarrow_{\operatorname{LI}} N'$, $C= \Gamma \rightarrow \Delta, E'[x]_{p},E[x]_q$ and $C'=\Gamma\{x \mapsto x'\},\Gamma \rightarrow \Delta, E'[x]_{p},E[q/x']$,
$[r,E'[x]_{p},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,E'[x]_p,(C';\pi'),N']$ for all $r\in \mathrm{pos}(E'[x]_p)$,
$[r,E[x]_{q},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,E[q/x'],(C';\pi'),N']$ for all $r\in \mathrm{pos}(E[q/x'])$,,
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L\{x \mapsto x'\},(C';\pi'),N']$ for all $(L,r)\in \Gamma$,
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in \Gamma$, and
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in \Delta$.\\
If $N \Rightarrow_{\operatorname{LI}} N'$, $C= \Gamma \rightarrow \Delta, E[x]_{p,q}$ and $C'=\Gamma\{x \mapsto x'\},\Gamma \rightarrow \Delta, E[q/x']$, then
$[r,E[x]_{p,q},(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,E[q/x'],(C';\pi'),N']$ for all $r\in \mathrm{pos}(E[q/x'])$,
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L\{x \mapsto x'\},(C';\pi'),N']$ for all $(L,r)\in \Gamma$,
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in \Gamma$, and
$[r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}} [ r,L,(C';\pi'),N']$ for all $(L,r)\in \Delta$.\\ \end{definition}
\begin{figure}
\caption{Visual representation of the parent literal position relation (Definition~\ref{def:approx:pterm})}
\end{figure}
The transitive closures of each parent relation are called ancestor relations.
The over-approximation of a clause set $N$ can introduce resolution refutations that have no corresponding equivalent in $N$ which we consider a lifting failure. Compared to our previous calculus~\cite{Teucke2015}, the lifting process is identical with the exception that there is no case for the removed Horn transformation. We only update the definition of conflicting cores to consider constrained clauses.
\begin{definition}[Conflicting Core]\label{def:lifting:core}
A finite set of unconstrained clauses and a solvable constraint $(N^\bot;\pi)$ are a conflicting core if
$N^\bot\delta$ is unsatisfiable for all solutions $\delta$ of $\pi$ over $\operatorname{vars}(N^\bot)\cup \mathrm{lvar}(\pi)$.
A conflicting core $(N^\bot;\pi)$ is a conflicting core of the constrained clause set $N$ if for every $C\in N^\bot$ there is a clause $(C',\pi')\in N$
such that $(C;\pi)$ is an instance of $(C';\pi')$ modulo duplicate literal elimination.
The clause $(C';\pi')$ is then called the instantiated clause of $(C;\pi)$ in $(N^\bot;\pi)$.
We call $(N^\bot;\pi)$ complete if for every clause $C \in N^\bot$ and literal $L\in C$, there exists a clause $D\in N^\bot$ with $\overline L\in D$. \end{definition}
A conflicting core is a generalization of a ground unsatisfiability core that allows global variables to act as parameters. This enables more efficient lifting and refinement compared to a simple ground unsatisfiable core. We show some examples at the end of this section.
We discuss the potential lifting failures and the corresponding refinements only for the linear and shallow case because lifting the satisfiability equivalent monadic and refinement transformations always succeeds. To reiterate from our previous work: in the linear case, there exists a clause in the conflicting core that is not an instance of the original clauses. In the shallow case, there exists a pair of clauses whose resolvent is not an instance of the original clauses. We combine these two cases by introducing the notion of a lift-conflict.
\begin{definition}[Conflict]\label{def:refine:conflict} Let $N \cup \{(C,\pi)\} \Rightarrow_{\operatorname{LI}} N\cup \{(C_a,\pi_a)\}$ and $N^\bot$ be a complete ground conflicting core of $N\cup \{(C_a,\pi_a)\}$. We call a conflict clause $C_c \in N^\bot$ with the instantiated clause $(C_a,\pi_a)$ a lift-conflict if $C_c$ is not an instance of $(C,\pi)$ modulo duplicate literal elimination. Then, $C_c$ is an instance of $(C_a,\pi_a)$, which we call the conflict clause of $C_c$.
Let $N \cup \{(C,\pi)\} \Rightarrow_{\operatorname{SH}} N \cup \{(C_l,\pi_l),(C_r,\pi_r)\}$, $(C_a;\pi_a)$ be the shallow resolvent and $N^\bot$ be a complete ground conflicting core of $N\cup \{(C_l,\pi_l),(C_r,\pi_r)\}$. We call the resolvent $C_c$ of $C_l\delta_l\in N^\bot$ and $C_r\delta_r \in N^\bot$ a lift-conflict if $C_c$ is not an instance of $(C,\pi)$ modulo duplicate literal elimination. Then, $C_c$ is an instance of $(C_a;\pi_a)$, which we call the conflict clause of $C_c$. \end{definition}
The goal of refinement is to instantiate the original parent clause in such a way that is both satisfiability equivalent and prevents the lift-conflict after approximation. Solving the refined approximation will then either necessarily produce a complete saturation or a new refutation proof, because its conflicting core has to be different. For this purpose, we use the refinement transformation to segment the original parent clause $(C;\pi)$ into two parts $(C;\pi \wedge x \neq t)$ and $(C;\pi)\{x \mapsto t\}$.
For example, consider $N$ and its linear transformation $N'$.\newline \centerline{$\begin{array}{r@{\,\rightarrow\,}lcr@{\,\rightarrow\,}l}
& P(x,x) & \;\Rightarrow_{\operatorname{LI}}\; & &P(x,x')\\
P(a,b) & & \; \Rightarrow_{\operatorname{AP}}^0\; & P(a,b)&\\
\end{array}$} The ground conflicting core of $N'$ is\newline \centerline{$\begin{array}{r@{\,\rightarrow\,}l}
& P(a,b) \\
P(a,b)& \\
\end{array}$} Because $P(a,b)$ is not an instance of $P(x,x)$, lifting fails. $P(a,b)$ is the lift-conflict. Specifically, $\{x\mapsto a\}$ and $\{x \mapsto b\}$ are conflicting substitutions for the parent variable $x$. We pick $\{x\mapsto a\}$ to segment $P(x,x)$ into $ (P(x,x);x \neq a)$ and $P(x,x)\{x \mapsto a\}$. Now, any descendant of $(P(x,x);x \neq a)$ cannot have $a$ at the position of the first $x$, and any descendant of $P(x,x)\{x \mapsto a\}$ must have an $a$ at the position of the second $x$. Thus, $P(a,b)$ is excluded in both cases and no longer appears as a lift-conflict.
To show that the lift-conflict will not reappear in the general case, we use that the conflict clause and its ancestors have strong ties between their term structures and constraints.
\begin{definition}[Constrained Term Skeleton]\label{def:refine:termskel} The constrained term skeleton of a term $t$ under constraint $\pi$, $\mathrm{skt}(t,\pi)$, is defined as the normal form of the following transformation: \newline \centerline{$\begin{array}{c}
(t[x]_{p,q} ;\pi) \Rightarrow_{\mathrm{skt}} (t[q/x'] ;\pi \wedge \pi\{x \mapsto x'\}), \text{ where } p \neq q \text{ and $x'$ is fresh}.
\end{array}$} \end{definition}
The constrained term skeleton of a term $t$ is essentially a linear version of $t$ where the restrictions on each variable position imposed by $\pi$ are preserved. For $(t,\pi)$ and a solution $\delta$ of $\pi$, $t\delta$ is called a ground instance of $(t,\pi)$.
\begin{lemma}\label{lem:refinement:ancestor_skel} Let $N_0 \Rightarrow_{\operatorname{AP}}^* N_k$, $(C_k;\pi_k)$ in $N$ with the ancestor clause $(C_0;\pi_0)\in N_0$ and $N^\bot_k$ be a complete ground conflicting core of $N_k$. Let $\delta$ be a solution of $\pi_k$ such that $C_k\delta$ is in $N^\bot_k$. If $(L',q')$ is a literal position in $(C_k;\pi_k)$ with the ancestor $(L,q)$ in $(C_0,\pi_0)$, then (i) $L'\delta\vert_{q'}$ is an instance of $\mathrm{skt}(L\vert_q,\pi_0)$, (ii) $q=q'$ if $L$ and $L'$ have the same predicate, and (iii) if $L'\vert_{q'}=x$ and there exists an ancestor variable $y$ of $x$ in $(C_0,\pi_0)$, then $L\vert_{q}=y$. \end{lemma}
\begin{proof}
By induction on the length of the approximation $N_0 \Rightarrow_{\operatorname{AP}}^* N_k$.
The base case $N_k=N_0$, is trivial.
Let $N_0= N \cup \{ (C;\pi)\} \Rightarrow_{\operatorname{SH}} N \cup \{ (C_l;\pi_l),(C_r;\pi_r)\}=N_k$, $(C_k;\pi_k)$ be the shallow $\rho$-resolvent and $C_k\delta$ be the resolvent of two instances of $(C_l;\pi_l)$ and $(C_r;\pi_r)$ in $N^\bot_k$.
Then, $(C_k;\pi_k)\rho^{-1}$ is equal to $ (C;\pi)$ modulo duplicate literal elimination.
Thus, by definition $(L,q)=(L',q')\rho^{-1}$.
Therefore, (i) $L'\delta\vert_{q'}$ is an instance of $\mathrm{skt}(L\vert_q,\pi_0)$,
(ii) $q=q'$ if $L$ and $L'$ have the same predicate, and
(iii) if $L'\vert_{q'}=x$ and there exists an ancestor variable $y$ of $x$ in $(C_0,\pi_0)$, then $L\vert_{q}=y$.
Now, let $ N_0 \Rightarrow_{\operatorname{AP}} N_{1} \Rightarrow_{\operatorname{AP}}^* N_k$.
Since $(L',p)$ has an ancestor literal position in $(C_0,\pi_0)$,
the ancestor clause of $(C_k;\pi_k)$ in $N_1$, $(C_1,\pi_1)$, contains the the ancestor literal position $(L_1,{q_1})$, which has $(L,q)$ as its parent literal position.
By the induction hypothesis on $N_{1} \Rightarrow_{\operatorname{AP}}^* N_k$,
(i) $L'\delta\vert_{q'}$ is an instance of $\mathrm{skt}(L_1\vert_{q_1},\pi_1)$,
(ii) $q_1=q'$ if $L_1$ and $L'$ have the same predicate, and
(iii) if $L'\vert_{q'}=x$ and there is an ancestor variable $y_1$ of $x$ in $(C_1,\pi_1)$, then $L_1\vert_{q_1}=y_1$.
Let $N_0= N \cup \{ (C;\pi)\} \Rightarrow_{\operatorname{Ref}} N \cup \{ (C;\pi \wedge x \neq t),(C;\pi)\{x \mapsto t\}\}=N_1$.
If $(C_1,\pi_1)$ is neither $(C;\pi \wedge x \neq t)$ nor $(C;\pi)\{x \mapsto t\}$, then trivially $(C_0,\pi_0)=(C_1,\pi_1)$.
Otherwise, $(C_1,\pi_1)= (C;\pi \wedge x \neq t)$ or $(C_1,\pi_1)= (C;\pi)\{x \mapsto t\}$.
Then $(L_1,{q_1})= (L,q)$ or $(L_1,{q_1})=(L,q)\{x \mapsto t\}$.
In either case,(i) $L'\delta\vert_{q'}$ is an instance of $\mathrm{skt}(L\vert_q,\pi_0)$,
(ii) $q=q'$ if $L$ and $L'$ have the same predicate, and
(iii) if $L'\vert_{q'}=x$ and there exists an ancestor variable $y$ of $x$ in $(C_0,\pi_0)$, then $L\vert_{q}=y$.
Let $N_0 \Rightarrow_{\operatorname{MO}} \Proj{P}(N)=N_1$.
If $P$ is not the predicate of $L$, then trivially $(L,q)=(L_1,{q_1})$.
If $P$ is the predicate of $L$, then $(L,q)=(P(t_1,\ldots,t_n),q)$ and $(L_1,{q_1})=(T(f_p(t_1,\ldots,t_n)),1.q)$.
Thus, (i) $L'\delta\vert_{q'}$ is an instance of $\mathrm{skt}(L\vert_{q},\pi_0) =$ $\mathrm{skt}(T(f_p(t_1,\ldots,t_n)\vert_{1.q},\pi_0)$.
(ii) The predicate of $L'$ is not $P$ by definition.
(iii) Let $L'\vert_{q'}=x$ and $y$ be the ancestor variable of $x$ in $(C_0,\pi_0)$.
Then, $y$ is also the ancestor variable of $x$ in $(C_1,\pi_1)$ and $L_1\vert_{q_1}=y$.
Therefore, $L\vert_q=P(t_1,\ldots,t_n)\vert_{q}=T(f_p(t_1,\ldots,t_n)\vert_{1.q}=L_1\vert_{q_1}=y$.
Let $N_0= N \cup \{ (C;\pi)\} \Rightarrow_{\operatorname{LI}} N \cup \{ (C_a;\pi_a)\}=N_1 $ where an occurrence of a variable $x$ is replaced by a fresh $x'$.
If $(C_1,\pi_1) \neq (C_a;\pi_a)$, then trivially $(C_0,\pi_0)=(C_1,\pi_1)$.
Otherwise, $(C_1,\pi_1) = (C_a;\pi_a)$, $(C_0,\pi_0)=(C,\pi)$.
By definition, $(L,q)=(L_1\{x' \mapsto x\},q_1)$ and $\pi_0=\pi_1\{x' \mapsto x\}$.
Thus, $\mathrm{skt}(L\vert_{q},\pi_0)= \mathrm{skt}(L_1\vert_{q_1},\pi_1)$.
Therefore, $L'\delta\vert_{q'}$ is an instance of $\mathrm{skt}(L\vert_{q},\pi_0)$.
Since $L$ and $L_1$ have the same predicate and $q=q_1$, $q=q'$ if $L$ and $L'$ have the same predicate.
Let $L'\vert_{q'}=z$ and $y$ be the ancestor variable of $z$ in $(C_1,\pi_1)$.
If $y \neq x'$, then $y$ is the ancestor variable of $z$ in $(C_0,\pi_0)$ and $L\vert_{q}=L_1\{x' \mapsto x\}\vert_{q_1}=y_1$.
Otherwise, $x$ is the ancestor variable of $z$ in $(C_0,\pi_0)$ and $L\vert_{q}=L_1\{x' \mapsto x\}\vert_{q_1}=x$.
Let $N_0= N \cup \{ (C;\pi)\} \Rightarrow_{\operatorname{SH}} N \cup \{ (C_l;\pi_l),(C_r;\pi_r)\}=N_1$ where a term $s$ is extracted from a positive literal $Q(s'[s]_p)$ via introduction of fresh predicate $S$ and variable $x$.
If $(C_1,\pi_1)$ is neither $(C_l;\pi_l)$ nor $(C_r;\pi_r)$, then trivially $(C_0,\pi_0)=(C_1,\pi_1)$.
If $(C_1,\pi_1)= (C_l;\pi_l)$ and $L_1 = S(x)$,
then $(C_0,\pi_0)=(C;\pi)$, $q_1=1$, $(L',q')=(S(x),1)$ and $(Q(s'[s]_p),1.p)$ is the parent literal position of $(S(x),1)$.
Let $L'\delta=S(t)$.
Because $N^\bot_k$ is complete and ground, there is a clause $C'_k\delta'\in N^\bot_k$ that contains the positive literal $S(t)$.
The ancestor of $(C'_k,\pi'_k)\in N_k$ in $N_1$ is $(C_r;\pi_r)$ because it is the only clause in $N_1$ with a positive $S$-literal.
Then, by the inductive hypothesis, $(S(s),1)$ in $(C_r;\pi_r)$ is the ancestor literal position of $(S(x),1)$ in $(C'_k,\pi'_k)$.
Thus, $t$ is an instance of $\mathrm{skt}(S(s)\vert_1,\pi_r)=\mathrm{skt}(s,\pi_r)$.
Therefore, $t=L'\delta\vert_{q'}$ is an instance of $\mathrm{skt}(Q(s'[s]_p)\vert_{1.p},\pi)=\mathrm{skt}(s,\pi_r)$.
Further, $Q$ and $S$ are not the same predicate because $S$ is fresh.
Since $x$ has no parent variable, $L'\vert_{q'}=x$ has no ancestor variable in $(C_0,\pi_0)$.
If $(C_1,\pi_1)= (C_l;\pi_l)$ and $L_1 = Q(s'[p/x])$,
then $(C_0,\pi_0)=(C;\pi)$ and $(Q(s'[s]_p),q_1)$ in $(C;\pi)$ is the parent literal position of $(L_1,q_1)$ in $(C_1,\pi_1)$ and ancestor literal position of $(L',q')$ in $(C_k,\pi_k)$.
If $q_1$ is not a position at or above $p$, the subterm at $p$ is irrelevant and thus $\mathrm{skt}(Q(s'[s]_p)\vert_{q_1},\pi)=\mathrm{skt}(Q(s'[p/x])\vert_{q_1},\pi_l)$.
Otherwise, let $r$ be a position such that $q_1r=1.p$.
Since $\vert p \vert=2$, no following shallow transformation step extracts a subterm of $s'[p/x]$ containing $x$.
Thus by definition of $ \Rightarrow_{\operatorname{AP}}$, $L'=Q(t'[x]_p)$ and $C_k$ also contains the negative literal $S(x)$.
Let $S(x)\delta=S(t)$.
Analogously to the previous case, $t$ is an instance of $\mathrm{skt}(s,\pi_r)$.
Combined with $L'\delta\vert_{q'}$ being an instance of $\mathrm{skt}(L_1\vert_{q_1},\pi_1)=\mathrm{skt}(Q(s'[p/x])\vert_{q_1},\pi_l)$ and $L'\delta\vert_{1.p}=t$,
$L'\delta\vert_{q'}$ is an instance of $\mathrm{skt}(Q(s'[s]_p)\vert_{q},\pi)$.
Since $L$ and $L_1$ have the same predicate and $q=q_1$, $q=q'$ if $L$ and $L'$ have the same predicate.
Let $L'\vert_{q'}=z$ and $y$ in $(C_1,\pi_1)$ be the ancestor variable of $z$ in $(C_k,\pi_k)$.
Since $x$ has no parent, $y\neq x$ and $y$ in $(C_0,\pi_0)$ is the ancestor variable of $z$.
Therefore, $Q(s'[s]_p)\vert_{q_1}=y$ because $Q(s'[p/x])\vert_{q_1}=y$.
If $(C_1,\pi_1)= (C_r;\pi_r)$ and $L_1 = S(s)$, let $q_1=1.q'_1$.
Then, $(C_0,\pi_0)=(C;\pi)$ and $(L,q)=(Q(s'[s]_p),1.pq'_1)$ in $(C_0,\pi_0)$ is the parent literal position of $(L_1,q_1)$ in $(C_1,\pi_1)$.
Thus, $L'\delta\vert_{q'}$ is an instance of $\mathrm{skt}((Q(s'[s]_p)\vert_{1.pq'_1},\pi)=\mathrm{skt}(s\vert_{q'_1},\pi)=\mathrm{skt}(L_1\vert_{q_1},\pi_r)$.
Because $S$ is fresh, $Q$ is not the predicate of $L'$.
Let $L'\vert_{q'}=z$ and $y$ in $(C_1,\pi_1)$ be the ancestor variable of $z$ in $(C_k,\pi_k)$.
Then, $y$ in $(C_0,\pi_0)$ is the ancestor variable of $z$
and $Q(s'[s]_p)\vert_{q}=s\vert_{q'_1}=y$ because $s\vert_{q'_1}=L_1\vert_{q_1}=y$.
Otherwise, $(L_1,q_1)$ in $(C_0,\pi_0)$ is the parent literal position of $(L_1,q_1)$ in $(C_1,\pi_1)$, by definition.
Then, $\mathrm{skt}(L_1,\pi)=\mathrm{skt}(L_1,\pi_l)$ or $\mathrm{skt}(L_1,\pi)=\mathrm{skt}(L_1,\pi_r)$, respectively.\qed \end{proof}
Next, we define the notion of descendants and descendant relations to connect lift-conflicts in ground conflicting cores with their corresponding ancestor clauses. The goal, hereby, is that if a ground clause $D$ is not a descendant of a clause in $N$, then it can never appear in a conflicting core of an approximation of $N$.
\begin{definition}[Descendants]\label{def:refine:descendant} Let $N \Rightarrow_{\operatorname{AP}}^* N'$, $[(C;\pi),N] \Rightarrow_{\operatorname{A}}^* [(C';\pi'),N']$ and $D$ be a ground instance of $(C';\pi')$. Then, we call $D$ a \emph{descendant} of $(C;\pi)$ and define the $[(C;\pi),N] \Rightarrow_{\operatorname{A}}^*[(C';\pi'),N']$-descendant relation $\des{}{D}{}$ that maps literals in $D$ to literal positions in $(C;\pi)$ using the following rule: $$ L'\delta \des{}{D}{} (L,r) \text{ if } L'\delta\in D \text{ and } [r,L,(C;\pi),N] \Rightarrow_{\operatorname{A}}^* [\varepsilon,L',(C';\pi'),N'] $$ \end{definition}
For the descendant relations it is of importance to note that while there are potentially infinite ways that a lift-conflict $C_c$ can be a descendant of an original clause $(C;\pi)$, there are only finitely many distinct descendant relations over $C_c$ and $(C;\pi)$. This means, if a refinement transformation can prevent one distinct descendant relation without generating new distinct descendant relations (Lemma~\ref{lem:refinement:descendants}), a finite number of refinement steps can remove the lift-conflict $C_c$ from the descendants of $(C;\pi)$ (Lemma~\ref{lem:refinement:refine}). Thereby, preventing any conflicting cores containing $C_c$ from being found again.
A clause $(C;\pi)$ can have two descendants that are the same except for the names of the $S$-predicates introduced by shallow transformations. Because the used approximation $N \Rightarrow_{\operatorname{AP}}^* N'$ is arbitrary and therefore also the choice of fresh $S$-predicates, if $D$ is a descendant of $(C;\pi)$, then any clause $D'$ equal to $D$ up to a renaming of $S$-predicates is also a descendant of $(C;\pi)$. On the other hand, the actual important information about an $S$-predicate is which term it extracts. Two descendants of $(C;\pi)$ might be identical but their $S$-predicate extract different terms in $(C;\pi)$. For example, $P(a)\rightarrow S(f(a))$ is a descendant of $P(x),P(y)\rightarrow Q(f(x),g(f(x)))$ but might extract either occurrence of $f(x)$. These cases are distinguished by their respective descendant relations. In the example, we have either $S(f(a)) \des{}{D}{} (Q(f(x),g(f(x))),1)$ or $S(f(a)) \des{}{D}{} (Q(f(x),g(f(x))),2.1)$.
\begin{lemma}\label{lem:refinement:descendants} Let $N_0=N \cup \{(C;\pi)\} \Rightarrow_{\operatorname{Ref}} N \cup \{(C;\pi \wedge x \neq t),(C;\pi)\{x \mapsto t\}\}=N_1$ be a refinement transformation and $D$ a ground clause. If there is a $[(C;\pi \wedge x \neq t),N_1] \Rightarrow_{\operatorname{A}}^*[(C';\pi'),N_2]$- or $[(C;\pi)\{x \mapsto t\},N_1] \Rightarrow_{\operatorname{A}}^*[(C';\pi'),N_2]$-descendant relation $\des{}{D}{1}$, then there is an equal $[(C;\pi),N_0] \Rightarrow_{\operatorname{A}}^*[(C';\pi'),N_2]$-descendant relation $\des{}{D}{0}$. \end{lemma}
\begin{proof}
Let $L_D $ be a literal of $D$ and $L' \des{}{D}{1} (L,r)$.
If $D$ is a descendant of $(C;\pi \wedge x \neq t)$, then $[r,L,(C;\pi \wedge x \neq t),N_1] \Rightarrow_{\operatorname{A}}^* [\varepsilon,L',(C';\pi'),N_2]$.
Because $[r,L,(C;\pi),N_0] \Rightarrow_{\operatorname{A}} [r,L,(C;\pi \wedge x \neq t),N_1]$, $L' \des{}{D}{0} (L,r)$.
If $D$ is a descendant of $(C;\pi)\{x \mapsto t\}$, the proof is analogous.\qed \end{proof}
\begin{lemma}[Refinement]\label{lem:refinement:refine} Let $N \Rightarrow_{\operatorname{AP}} N'$ and $N^\bot$ be a complete ground conflicting core of $N'$. If $C_c\in N^\bot$ is a lift-conflict, then there exists a finite refinement $N \Rightarrow_{\operatorname{Ref}}^* N_{R}$ such that for any approximation $N_{R} \Rightarrow_{\operatorname{AP}}^* N'_{R}$ and ground conflicting core $N^\bot_{R}$ of $N'_{R}$, $C_c$ is not a lift-conflict in $N^\bot_{R}$ modulo duplicate literal elimination. \end{lemma}
\begin{proof}
Let $(C_a,\pi_a)$ be the conflict clause of $C_c$ and $(C;\pi)\in N$ be the parent clause of $(C_a,\pi_a)$.
$C_c$ is a descendant of $(C;\pi)$ with the corresponding $[(C;\pi),N] \Rightarrow_{\operatorname{A}} [(C_a;\pi_a),N']$-descendant relation $\des{}{C_c}{0}$.
We apply induction on the number of distinct $[(C;\pi),N] \Rightarrow_{\operatorname{A}}^* [(C';\pi'),N'']$-descendant relations $\des{}{C_c}{}$ for arbitrary approximations $N \Rightarrow_{\operatorname{AP}}^* N''$.
Since only the shallow and linear transformations can produce lift-conflicts,
the clause $(C;\pi)$ is replaced by either a linearized clause $(C';\pi')$ or
two shallow clauses $(C_l;\pi)$ and $(C_r;\pi)$.
Then, the conflict clause $(C_a;\pi_a)$ of $C_c$ is either the linearized $(C';\pi')$ or the resolvent of $(C_l;\pi)$ and $(C_r;\pi)$.
In either case, $C_c=C_a\delta$ for some solution $\delta$ of $\pi_a$.
Furthermore, there exists a substitution $\tau=\{x'_1 \mapsto x_1,\ldots,x'_n \mapsto x_n\}$ such that $(C;\pi)$ and $(C_a;\pi_a)\tau$ are equal modulo duplicate literal elimination.
That is, $\tau= \{x'\mapsto x\}$ for a linear transformation and $\tau=\rho^{-1}$ for shallow transformation (Definition~\ref{def:approx:resolvent}).
Assume $C_c=C_a\tau\sigma$ for some grounding substitution $\sigma$, where $\tau\sigma$ is a solution of $\pi_a$.
Thus, $\sigma$ is a solution of $\pi_a\tau$, which is equivalent to $\pi$.
Then, $C_c$ is equal to $C\sigma$ modulo duplicate literal elimination an instance of $(C;\pi)$,
which contradicts with $C_c$ being a lift-conflict.
Hence, $C_c=C_a\delta$ is not an instance of $C_a\tau$ and thus, $x_i\delta \neq x'_i\delta$ for some $x_i$ in the domain of $\tau$.
Because $x_i\delta $ and $ x'_i\delta$ are ground, there is a position $p$ where $x_i\delta\vert_p $ and $x'_i\delta\vert_p$ have different function symbols.
We construct the straight term $t$ using the path from the root to $p$ on $x_i\delta$ with variables that are fresh in $(C,\pi)$.
Then, we can use $x_i$ and $t$ to segment $(C;\pi)$ into $(C;\pi \wedge x_i \neq t)$ and $(C;\pi)\{x_i \mapsto t\}$ for the refinement $N \Rightarrow_{\operatorname{Ref}} N_{R}$.
Note, that $x_i\delta$ is a ground instance of $t$, while $x'_i\delta$ is not.
Let $(L'_1,r'_1)$ and $(L'_2,r'_2)$ in $(C_a,\pi_a)$ be literal positions of the variables $x_i$ and $x'_i$ in $C_a$,
and $(L_1,r_1)$ and $(L_2,r_2)$ in $(C,\pi)$ be the parent literal positions of $(L'_1,r'_1)$ and $(L'_2,r'_2)$, respectively.
Because $(C_a,\pi_a)\tau$ is equal to $(C;\pi)$ modulo duplicate literal elimination, $L_1\vert_{r_1} = L_2\vert_{r_2}=x_i$.
Let $N \Rightarrow_{\operatorname{Ref}} N_1$ be the refinement where $(C;\pi)$ is segmented into $(C;\pi \wedge x_i \neq t)$ and $(C;\pi)\{x_i \mapsto t\}$.
By Lemma~\ref{lem:refinement:descendants}, all $[(C;\pi \wedge x_i \neq t),N_1] \Rightarrow_{\operatorname{A}}^* [(C'_a;\pi'_a),N_2]$- or $[(C;\pi)\{x_i \mapsto t\},N_1] \Rightarrow_{\operatorname{A}}^* [(C'_a;\pi'_a),N_2]$-descendant relations
correspond to an equal $[(C;\pi),N] \Rightarrow_{\operatorname{A}} [(C'_a;\pi'_a),N_2]$-descendant relation.
Assume there is a $[(C;\pi \wedge x_i \neq t),N_1] \Rightarrow_{\operatorname{A}}^* [(C'_a;\pi'_a),N_2]$-descendant relation $\des{}{C_c}{1}$ that is not distinct from $\des{}{C_c}{0}$.
Because $L'_1\delta \des{}{C_c}{0} (L_1,r)$ for some literal position $(L_1,r)$ in $(C;\pi)$, which is the parent literal position of $(L_1,r)$ in $(C;\pi \wedge x_i \neq t)$,
$L'_1\delta \des{}{C_c}{1} (L_1,r)$.
However, this contradicts Lemma~\ref{lem:refinement:ancestor_skel} because $x_i\delta $ is not an instance of $\mathrm{skt}(L_1\vert_{r_1},\pi \wedge x_i \neq t)=\mathrm{skt}(x_i,\pi \wedge x_i \neq t)$.
The case that there is a $[(C;\pi)\{x_i \mapsto t\},N_1] \Rightarrow_{\operatorname{A}}^* [(C'_a;\pi'_a),N_2]$-descendant relation that is not distinct from $\des{}{C_c}{0}$ is analogous using the argument that $x'_i\delta $ is not an instance of $\mathrm{skt}(L_2\{x_i \mapsto t\}\vert_{r_2},\pi)=\mathrm{skt}(t,\pi)$.
Hence, there are strictly less distinct descendant relations over $C_c$ and $(C;\pi \wedge x \neq t)$ or $(C;\pi)\{x \mapsto t\}$
than there are distinct descendant relations over $C_c$ and $(C,\pi)$.
If there are no descendant relations, then $C_c$ can no longer appear as a lift conflict.
Otherwise, by the inductive hypothesis, there exists a finite refinement $N \Rightarrow_{\operatorname{Ref}} N_1 \Rightarrow_{\operatorname{Ref}}^* N_{R}$
such that for any approximation $N_{R} \Rightarrow_{\operatorname{AP}} N'_{R}$ and ground conflicting core $N^\bot_{R}$ of $N'_{R}$,
$C_c$ is not a lift-conflict in $N^\bot_{R}$ modulo duplicate literal elimination.\qed \end{proof}
\begin{theorem}[Soundness and Completeness of FO-AR]\label{theo:refinement:scfoar} Let $N$ be an unsatisfiable clause set and $N'$ its MSL(SDC) approximation: (i)~if $N$ is unsatisfiable then there exists a conflicting core of $N'$ that can be lifted to a refutation in $N$, (ii)~if $N'$ is satisfiable, then $N$ is satisfiable too. \end{theorem} \begin{proof}(Idea)
By Lemma~\ref{lem:approx:sound} and Lemma~\ref{lem:refinement:refine}, where the latter can be used to show that a core of $N'$ that cannot
be lifted also excludes the respective instance for unsatisfiability of $N$.
Let $(C_a,\pi_a)$ be the conflict clause of $C_c$ and $(C;\pi)\in N$ be the parent clause of $(C_a,\pi_a)$.
$C_c$ is a descendant of $(C;\pi)$ with the corresponding $[(C;\pi),N] \Rightarrow_{\operatorname{A}} [(C_a;\pi_a),N']$-descendant relation $\des{}{C_c}{0}$.
We apply induction on the number of distinct $[(C;\pi),N] \Rightarrow_{\operatorname{A}}^* [(C';\pi'),N'']$-descendant relations $\des{}{C_c}{}$ for arbitrary approximations $N \Rightarrow_{\operatorname{AP}}^* N''$.
Since only the shallow and linear transformations can produce lift-conflicts,
the clause $(C;\pi)$ is replaced by either a linearized clause $(C';\pi')$ or
two shallow clauses $(C_l;\pi)$ and $(C_r;\pi)$.
Then, the conflict clause $(C_a;\pi_a)$ of $C_c$ is either the linearized $(C';\pi')$ or the resolvent of $(C_l;\pi)$ and $(C_r;\pi)$.
In either case, $C_c=C_a\delta$ for some solution $\delta$ of $\pi_a$.
Furthermore, there exists a substitution $\tau=\{x'_1 \mapsto x_1,\ldots,x'_n \mapsto x_n\}$ such that $(C;\pi)$ and $(C_a;\pi_a)\tau$ are equal modulo duplicate literal elimination.
That is, $\tau= \{x'\mapsto x\}$ for a linear transformation and $\tau=\rho^{-1}$ for shallow transformation (Definition~\ref{def:approx:resolvent}).
Assume $C_c=C_a\tau\sigma$ for some grounding substitution $\sigma$, where $\tau\sigma$ is a solution of $\pi_a$.
Thus, $\sigma$ is a solution of $\pi_a\tau$, which is equivalent to $\pi$.
Then, $C_c$ is equal to $C\sigma$ modulo duplicate literal elimination an instance of $(C;\pi)$,
which contradicts with $C_c$ being a lift-conflict.
Hence, $C_c=C_a\delta$ is not an instance of $C_a\tau$ and thus, $x_i\delta \neq x'_i\delta$ for some $x_i$ in the domain of $\tau$.
Because $x_i\delta $ and $ x'_i\delta$ are ground, there is a position $p$ where $x_i\delta\vert_p $ and $x'_i\delta\vert_p$ have different function symbols.
We construct the straight term $t$ using the path from the root to $p$ on $x_i\delta$ with variables that are fresh in $(C,\pi)$.
Then, we can use $x_i$ and $t$ to segment $(C;\pi)$ into $(C;\pi \wedge x_i \neq t)$ and $(C;\pi)\{x_i \mapsto t\}$ for the refinement $N \Rightarrow_{\operatorname{Ref}} N_{R}$.
Note, that $x_i\delta$ is a ground instance of $t$, while $x'_i\delta$ is not.
Let $(L'_1,r'_1)$ and $(L'_2,r'_2)$ in $(C_a,\pi_a)$ be literal positions of the variables $x_i$ and $x'_i$ in $C_a$,
and $(L_1,r_1)$ and $(L_2,r_2)$ in $(C,\pi)$ be the parent literal positions of $(L'_1,r'_1)$ and $(L'_2,r'_2)$, respectively.
Because $(C_a,\pi_a)\tau$ is equal to $(C;\pi)$ modulo duplicate literal elimination, $L_1\vert_{r_1} = L_2\vert_{r_2}=x_i$.
Let $N \Rightarrow_{\operatorname{Ref}} N_1$ be the refinement where $(C;\pi)$ is segmented into $(C;\pi \wedge x_i \neq t)$ and $(C;\pi)\{x_i \mapsto t\}$.
By Lemma~\ref{lem:refinement:descendants}, all $[(C;\pi \wedge x_i \neq t),N_1] \Rightarrow_{\operatorname{A}}^* [(C'_a;\pi'_a),N_2]$- or $[(C;\pi)\{x_i \mapsto t\},N_1] \Rightarrow_{\operatorname{A}}^* [(C'_a;\pi'_a),N_2]$-descendant relations
correspond to an equal $[(C;\pi),N] \Rightarrow_{\operatorname{A}} [(C'_a;\pi'_a),N_2]$-descendant relation.
Assume there is a $[(C;\pi \wedge x_i \neq t),N_1] \Rightarrow_{\operatorname{A}}^* [(C'_a;\pi'_a),N_2]$-descendant relation $\des{}{C_c}{1}$ that is not distinct from $\des{}{C_c}{0}$.
Because $L'_1\delta \des{}{C_c}{0} (L_1,r)$ for some literal position $(L_1,r)$ in $(C;\pi)$, which is the parent literal position of $(L_1,r)$ in $(C;\pi \wedge x_i \neq t)$,
$L'_1\delta \des{}{C_c}{1} (L_1,r)$.
However, this contradicts Lemma~\ref{lem:refinement:ancestor_skel} because $x_i\delta $ is not an instance of $\mathrm{skt}(L_1\vert_{r_1},\pi \wedge x_i \neq t)=\mathrm{skt}(x_i,\pi \wedge x_i \neq t)$.
The case that there is a $[(C;\pi)\{x_i \mapsto t\},N_1] \Rightarrow_{\operatorname{A}}^* [(C'_a;\pi'_a),N_2]$-descendant relation that is not distinct from $\des{}{C_c}{0}$ is analogous using the argument that $x'_i\delta $ is not an instance of $\mathrm{skt}(L_2\{x_i \mapsto t\}\vert_{r_2},\pi)=\mathrm{skt}(t,\pi)$.
Hence, there are strictly less distinct descendant relations over $C_c$ and $(C;\pi \wedge x \neq t)$ or $(C;\pi)\{x \mapsto t\}$
than there are distinct descendant relations over $C_c$ and $(C,\pi)$.
If there are no descendant relations, then $C_c$ can no longer appear as a lift conflict.
Otherwise, by the inductive hypothesis, there exists a finite refinement $N \Rightarrow_{\operatorname{Ref}} N_1 \Rightarrow_{\operatorname{Ref}}^* N_{R}$
such that for any approximation $N_{R} \Rightarrow_{\operatorname{AP}} N'_{R}$ and ground conflicting core $N^\bot_{R}$ of $N'_{R}$,
$C_c$ is not a lift-conflict in $N^\bot_{R}$ modulo duplicate literal elimination.\qed \end{proof}
Actually, Lemma~\ref{lem:refinement:refine} can be used to define a fair strategy on refutations in $N'$ in order to receive also a dynamically complete FO-AR calculus, following the ideas presented in \cite{Teucke2015}.
In Lemma~\ref{lem:refinement:refine}, we segment the conflict clause's immediate parent clause. If the lifting later successfully passes this point, the refinement is lost and will be possibly repeated. Instead, we can refine any ancestor of the conflict clause as long as it contains the ancestor of the variable used in the refinement. By Lemma~\ref{lem:refinement:ancestor_skel}-(iii), such an ancestor will contain the ancestor variable at the same positions. If we refine the ancestor in the original clause set, the refinement is permanent because lifting the refinement steps always succeeds. Only variables introduced by shallow transformation cannot be traced to the original clause set. However, these shallow variables are already linear and the partitioning in the shallow transformation can be chosen such that they are not shared variables. Assume a shallow, shared variable $y$, that is used to extract the term $t$, in the shallow transformation of $\Gamma\rightarrow E[s]_p,\Delta$ into $S(x),\Gamma_l\rightarrow E[p/x],\Delta_l$ and $\Gamma_r\rightarrow S(s),\Delta_r$. Since $\Delta_l$ $\dot{\cup}$ $\Delta_r=\Delta$ is a partitioning, $y$ can only appear in either $E[p/x],\Delta_l$ or $S(s),\Delta_r$. If $y \in \operatorname{vars}(E[p/x],\Delta_l)$ we instantiate $\Gamma_r$ with $\{y \mapsto t\}$ and $\Gamma_l$, otherwise. Now, $y$ is no longer a shared variable.\\
The refinement Lemmas only guarantee a refinement for a given ground conflicting core. In practice, however, conflicting cores contain free variables. We can always generate a ground conflicting core by instantiating the free variables with ground terms. However, if we only exclude a single ground case via refinement, next time the new conflicting core will likely have overlaps with the previous one. Instead, we can often remove all ground instances of a given conflict clause at once.
The simplest case is when unifying the conflict clause with the original clause fails because their instantiations differ at some equivalent positions. For example, consider $N= \{ P(x,x); P(f(x,a), f(y,b)) \rightarrow\}$. $N$ is satisfiable but the linear transformation is unsatisfiable with conflict clause $P(f(x,a), f(y,b))$ which is not unifiable with $P(x,x)$, because the two terms $f(x,a)$ and $f(y,b)$ have different constants at the second argument. A refinement of $ P(x,x)$ is \newline \centerline{$\begin{array}{r@{\,;\,}l}
( P(x,x) & x \neq f(v,a)) \\ (P(f(x,a),f(x,a)) & \top) \\ \end{array}$}
$ P(f(x,a), f(y,b))$ shares no ground instances with the approximations of the refined clauses.
Next, assume that again unification fails due to structural difference, but this time the differences lie at different positions. For example, consider $N= {\{ P(x,x); P(f(a,b), f(x,x)) \rightarrow\}}$. $N$ is satisfiable but the linear transformation of $N$ is unsatisfiable with conflict clause $ P(f(a,b), f(x,x))$
which is not unifiable with $ P(x,x)$ because in $f(a,b)$ the first an second argument are different but the same in $f(x,x)$. A refinement of $ P(x,x)$ is \newline \centerline{$\begin{array}{r@{\,;\,}l}
( P(x,x) & x \neq f(a,v)) \\
(P(f(a,x), f(a,x))) & x \neq a) \\
( P(f(a,a), f(a,a))) & \top) \\
\end{array}$}
$ P(f(a,b), f(x,x))$ shares no ground instances with the approximations of the refined clauses.
It is also possible that the conflict clause and original clause are unifiable by themselves, but the resulting constraint has no solutions. For example, consider $N= {\{ P(x,x); (P(x, y) \rightarrow; x \neq a \wedge x \neq b \wedge y \neq c \wedge y \neq d )\}}$ with signature $\Sigma=\{a,b,c,d\}$. $N$ is satisfiable but the linear transformation of $N$ is unsatisfiable with conflict clause $ (\rightarrow P(x,y); x \neq a \wedge x \neq b \wedge y \neq c \wedge y \neq d)$. While $P(x,x)$ and $P(x,y)$ are unifiable, the resulting constraint $ x \neq a \wedge x \neq b \wedge x \neq c \wedge x \neq d$ has no solutions. A refinement of $P(x,x)$ is \newline \centerline{$\begin{array}{r@{\,;\,}l}
(P(x,x) & x \neq a \wedge x \neq b) \\
( P(a,a) & \top) \\
(P(b,b) & \top) \\
\end{array}$}
$ ( P(x,y); x \neq a \wedge x \neq b \wedge y \neq c \wedge y \neq d)$ shares no ground instances with the approximations of the refined clauses.\\
Lastly, we should mention that there are cases where the refinement process does not terminate. For example, consider the clause set $N= {\{ P(x,x) ; P(y,g(y)) \rightarrow \}}$. $N$ is satisfiable but the linear transformation of $N$ is unsatisfiable with conflict clause $ P(y,g(y))$, which is not unifiable with $ P(x,x)$. A refinement of $ P(x,x)$ based on the ground instance $P(a,g(a))$ is \newline \centerline{$\begin{array}{r@{\,;\,}l}
( P(x,x) & x \neq g(v)) \\ ( P(g(x),g(x)) & \top) \\ \end{array}$} While $ P(y,g(y))$ is not an instance of the refined approximation, it shares ground instances with $ P(g(x),g(x'))$. The new conflict clause is $ P(g(y),g(g(y)))$ and the refinement will continue to enumerate all $ P(g^i(x),g^i(x))$ instances of $ P(x,x)$ without ever reaching a satisfiable approximation. Satisfiability of first-order clause sets is undecidable, so termination cannot be expected by any calculus, in general.
\section{Experiments}\label{sec:experiments}
In the following we discuss several first-order clause classes for which FO-AR implemented in SPASS-AR immediately decides satisfiability but superposition and instantiation-based methods fail. We argue both according to the respective calculi and state-of-the-art implementations, in particular SPASS~3.9~\cite{DBLP:conf/cade/WeidenbachDFKSW09}, Vampire~4.1~\cite{KovacsVoronkov13,DBLP:conf/cav/Voronkov14}, for ordered-resolution/superposition, iProver~2.5~\cite{DBLP:conf/cade/Korovin08} an implementation of Inst-Gen \cite{Korovin13ganzinger}, and Darwin~v1.4.5 \cite{DBLP:journals/ijait/BaumgartnerFT06} an implementation of the model evolution calculus \cite{DBLP:conf/cade/BaumgartnerT03}. All experiments were run on a 64-Bit Linux computer (Xeon(R) E5-2680, 2.70GHz, 256GB main memory). For Vampire and Darwin we chose the CASC-sat and CASC settings, respectively. For iProver we set the schedule to ``sat'' and SPASS, SPASS-AR were used in default mode. Please note that Vampire and iProver are portfolio solvers including implementations of several different calculi including superposition (ordered resolution), instance generation, and finite model finding. SPASS, SPASS-AR, and Darwin only implement superposition, FO-AR, and model evolution, respectively.
For the first example\newline \centerline{$P(x,y) \rightarrow P(x,z) , P(z,y);\quad P(a,a)$} and second example,\newline \centerline{$Q(x,x);\quad
Q(v,w) , P(x,y) \rightarrow P(x,v) , P(w,y);\quad
P(a,a)$} the superposition calculus produces independently of the selection strategy and ordering an infinite number of clauses of form\newline \centerline{$\begin{array}{r@{\,\rightarrow\,}l} &P(a,z_1),\; P(z_1,z_2),\;\ldots,\;P(z_n,a).\\ \end{array}$}
Using linear approximation, however, FO-AR replaces $P(x,y) \rightarrow P(x,z) , P(z,y)$ and $\rightarrow Q(x,x)$ with $P(x,y) \rightarrow P(x,z) , P(z',y)$ and $\rightarrow Q(x,x')$, respectively. Consequently, ordered resolution derives $\rightarrow P(a,z_1) , P(z_2,a)$ which subsumes any further inferences $\rightarrow P(a,z_1) , P(z_2,z_3) , P(z_4,a)$. Hence, saturation of the approximation terminates immediately. Both examples belong to the Bernays-Sch\"onfinkel fragment, so model evolution (Darwin) and Inst-Gen (iProver) can decide them as well. Note that the concrete behavior of superposition is not limited to the above examples but potentially occurs whenever there are variable chains in clauses.
On the third problem\newline \centerline{$P(x,y) \rightarrow P(g(x),z);\quad P(a,a)$} superposition derives all clauses of the form $\rightarrow P(g(\ldots g(a)\ldots),z)$. With a shallow approximation of $ P(x,y) \rightarrow P(g(x),z)$ into $ S(v) \rightarrow P(v,z)$ and $ P(x,y) \rightarrow S(g(x))$, FO-AR (SPASS-AR) terminates after deriving $ \rightarrow S(g(a))$ and $S(x) \rightarrow S(g(x))$. Again, model evolution (Darwin) and Inst-Gen (iProver) can also solve this example.
The next example\newline \centerline{$P(a);\quad P(f(a))\rightarrow;\quad
P(f(f(x))) \rightarrow P(x);\quad
P(x) \rightarrow P(f(f(x)))$} is already saturated under superposition. For FO-AR, the clause $P(x) \rightarrow P(f(f(x)))$ is replaced by $S(x) \rightarrow P(f(x))$ and $P(x) \rightarrow S(f(x))$. Then ordered resolution terminates after inferring $S(a) \rightarrow$ and $S(f(x)) \rightarrow P(x)$.
The Inst-Gen and model evolution calculi, however, fail. In either, a satisfying model is represented by a finite set of literals, i.e, a model of the propositional approximation for Inst-Gen and the trail of literals in case of model evolution. Therefore, there necessarily exists a literal $P(f^n(x))$ or $\neg P(f^n(x))$ with a maximal $n$ in these models. This contradicts the actual model where either $P(f^n(a))$ or $P(f^n(f(a)))$ is true. However, iProver can solve this problem using its built-in ordered resolution solver whereas Darwin does not terminate on this problem.
Lastly consider an example of the form\newline \centerline{$f(x) \approx x \rightarrow;\,\; f(f(x)) \approx x \rightarrow;\,\, \ldots; f^n(x) \approx x \rightarrow$} which is trivially satisfiable, e.g., saturated by superposition, but any model has at least $n+1$ domain elements. Therefore, adding these clauses to any satisfiable clause set containing $f$ forces calculi that explicitly consider finite models to consider at least $n+1$ elements. The performance of final model finders~\cite{SlaneySurendonk96} typically degrades in the number of different domain elements to be considered.
Combining each of these examples into one problem is then solvable by neither superposition, Inst-Gen, or model evolution and not practically solvable with increasing $n$ via testing finite models. For example, we tested\newline \centerline{$\begin{array}{c}
P(x,y) \rightarrow P(x,z) , P(z,y); \quad P(a,a); \quad P(f(a),y) \rightarrow; \\
P(f(f(x)),y) \rightarrow P(x,y);\quad P(x,y) \rightarrow P(f(f(x)),y); \\
f(x) \approx x \rightarrow;, \ldots, f^{n}(x) \approx x \rightarrow;\\
\end{array}$} for $n=20$ against SPASS, Vampire, iProver, and Darwin for more than one hour each without success. Only SPASS-AR solved it in less than one second.
For iProver we added an artificial positive equation $b\approx c$. For otherwise, iProver throws away all disequations while preprocessing. This is a satisfiability preserving operation, however, the afterwards found (finite) models are not models of the above clause set due to the collapsing of ground terms.
\section{Conclusion} \label{sec:conclusion}
The previous section showed FO-AR is superior to superposition, instantiation-based methods on certain classes of clause sets. Of course, there are also classes of clause sets where superposition and instantiation-based methods are superior to FO-AR, e.g., for unsatisfiable clause sets where the structure of the clause set forces FO-AR to enumerate failing ground instances due to the approximation in a bottom-up way.
Our prototypical implementation SPASS-AR cannot compete with systems such as iProver or Vampire on the respective CASC categories of the TPTP~\cite{Sut09}. This is already due to the fact that they are all meanwhile portfolio solvers. For example, iProver contains an implementation of ordered resolution and Vampire an implementation of Inst-Gen. Our results, Section~\ref{sec:experiments}, however, show that these systems may benefit from FO-AR by adding it to their portfolio.
The DEXPTIME-completeness result for MSLH strongly suggest that both the MSLH and also our MSL(SDC) fragment have the finite model property. However, we are not aware of any proof. If MSL(DSC) has the finite model property, the finite model finding approaches are complete on MSL(SDC). The models generated by FO-AR and superposition are typically infinite. It remains an open problem, even for fragments enjoying the finite model property, e.g., the first-order monadic fragment, to design a calculus that combines explicit finite model finding with a structural representation of infinite models. For classes that have no finite models this problem seems to become even more difficult. To the best of our knowledge, SPASS is currently the only prover that can show satisfiability of the clauses $R(x,x)\rightarrow$; $R(x,y), R(y,z)\rightarrow R(x,z)$; $R(x,g(x))$ due to an implementation of chaining~\cite{BachmairGanzinger98,SudaEtAl10}. Apart from the superposition calculus, it is unknown to us how the specific inferences for transitivity can be combined with any of the other discussed calculi, including the abstraction refinement calculus introduced in this paper.
Finally, there are not many results on calculi that operate with respect to models containing positive equations. Even for fragments that are decidable with equality, such as the Bernays-Schoenfinkel-Ramsey fragment or the monadic fragment with equality, there seem currently no convincing suggestions compared to the great amount of techniques for these fragments without equality. Adding positive equations to MSL(SDC) while keeping decidability is, to the best of our current knowledge, only possible for at most linear, shallow equations $f(x_1,\ldots,x_n) \approx h(y_1,\ldots,y_n)$~\cite{JacquemardMeyerEtAl98}. However, approximation into such equations from an equational theory with nested term occurrences typically results in an almost trivial equational theory. So this does not seem to be a very promising research direction.
\end{document}
|
arXiv
|
{
"id": "1703.02837.tex",
"language_detection_score": 0.6570768356323242,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Description of the minimizers of least squares regularized~ with~ $\bm{\ell_0} \newcommand{\slugmaster}{ \slugger{siims}{xxxx}{xx}{x}{x--x}}
\vspace*{-4cm}
\renewcommand{\arabic{footnote}}{\alph{footnote}} \hspace*{5cm}\textcolor[rgb]{0.1,0.1,0.1}{\sc (in press\footnote {Received by the editors November 10, 2011; accepted for publication (in revised form) January 14, 2013; published electronically DATE. })}
\vspace*{4cm}
\renewcommand{\arabic{footnote}}{\arabic{footnote}} \setcounter{footnote}{0} \begin{abstract} We have an $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\n$ real-valued arbitrary matrix $A$ (e.g. a dictionary) with $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}<\n$ and data $d$ describing the sought-after object with the help of $A$. This work provides an in-depth analysis of the (local and global) minimizers of an objective function $\mathcal{F}_d$ combining a quadratic data-fidelity term and an $\ell_0$ penalty applied to each entry of the sought-after solution, weighted by a regularization parameter $\beta>0$. For several decades, this objective has attracted a ceaseless effort to conceive algorithms approaching a good minimizer. Our theoretical contributions, summarized below, shed new light on the existing algorithms and can help the conception of innovative numerical schemes. To solve the normal equation associated with any $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$-row submatrix of $A$ is equivalent to compute a local minimizer $\hat u$ of $\mathcal{F}_d$. (Local) minimizers $\hat u$ of $\mathcal{F}_d$ are strict if and only if the submatrix, composed of those columns of $A$ whose indexes form the support of $\hat u$, has full column rank. An outcome is that strict local minimizers of $\mathcal{F}_d$ are easily computed without knowing the value of $\beta$. Each strict local minimizer is linear in data. It is proved that $\mathcal{F}_d$ has global minimizers and that they are always strict. They are studied in more details under the (standard) assumption that $\mathrm{rank}(A)=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}<\n$. The global minimizers with $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$-length support are seen to be impractical. Given $d$, critical values $\beta_\k$ for any $\k\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1$ are exhibited such that if $\beta>\beta_\k$, all global minimizers of $\mathcal{F}_d$ are $\k$-sparse. An assumption on $A$ is adopted and proved to fail only on a closed negligible subset. Then for all data $d$ beyond a closed negligible subset, the objective $\mathcal{F}_d$ for $\beta>\beta_\k$, $\k\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1$, has a unique global minimizer and this minimizer is $\k$-sparse. Instructive small-size ($5\times 10$) numerical illustrations confirm the main theoretical results. \end{abstract}
\begin{keywords}asymptotically level stable functions, global minimizers, local minimizers, $\ell_0$ regularization, nonconvex nonsmooth minimization, perturbation analysis, quadratic programming, solution analysis, sparse recovery, strict minimizers, underdetermined linear systems, uniqueness of the solution, variational methods \end{keywords}
\begin{AMS}
15A03,
15A29, 15A99, 26E35, 28A75, 46N99, 47J07, 49J99, 49K40,
90C26, 90C27, 94A08, 94A12. ltering, etc.) Communication, information - INFORMATION AND COMMUNICATION, CIRCUITS \end{AMS}
\paragraph{\hspace*{-6mm}\textbf{\small DOI}} {\small 10.1137/11085476X}
\begin{center}} \def\ec{\end{center}\textcolor[rgb]{0.1,0.1,0.1}{--------------------------------------------------------------------------------------------------------------------}\ec
\pagestyle{myheadings} \thispagestyle{plain} \markboth{MILA NIKOLOVA}{THE MINIMIZERS OF LEAST SQUARES REGULARIZED~ WITH~ $\bm{\ell_0}$-NORM}
\section{Introduction}
Let $A$ be an arbitrary matrix (e.g., a dictionary) such that \[~A\in\RR^{\m\x\n}~~~\mbox{for}~~~\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}<\n~,\] where the positive integers $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ and $\n$ are fixed. Given a data vector $d\in\RR^{\m}$, we consider an objective function $\mathcal{F}_d:\RR^{\n}\rightarrow\mathbb R$ of the form
\begin{eqnarray}} \def\eeqn{\end{eqnarray}\mathcal{F}_d(u)&=&\|Au-d\|^2_2+\beta\|u\|_0~,~~\beta>0~,\label{fd}\\
\|u\|_0&=&\,\sharp\,\sigma(u)~, \nonumber\label{rj} \eeqn where $u\in\RR^{\n}$ contains the coefficients describing the sought-after object,
$\beta$ is a regularization parameter, $\,\sharp\,$ stands for cardinality and $\sigma(u)$ is the support of $u$ (i.e., the set of all $i\in\{1,\cdots,\n\}$ for which the $i$th entry of $u$ satisfies $u[i]\neq0$). By an abuse of language, the penalty in~\eqref{fd} is called the $\ell_0$-norm. Define $\phi:\mathbb R\rightarrow\mathbb R$ by \begin{equation}} \def\eeq{\end{equation}
\phi(t)\stackrel{\rm def}{=}\left\{\begin{array}} \def\earr{\end{array}{lll}0&\mbox{\rm if}&t=0~,\\1&\mbox{\rm if}&t\neq0~.\earr\right.\label{phi}\eeq Then $\disp{\|u\|_0=\sum_{i=1}^\n\phi(u[i])=\sum_{i\in\sigma(u)}\phi(u[i])}$, so $\mathcal{F}_d$ in~\eqref{fd} equivalently reads \begin{equation}} \def\eeq{\end{equation}
\mathcal{F}_d(u)=\|Au-d\|^2_2+\beta\sum_{i=1}^\n\phi(u[i]) =
\|Au-d\|^2_2+\beta\sum_{i\in\sigma(u)}\phi(u[i])~.\label{fds} \eeq We focus on all (local and global) minimizers $\hat u$ of an objective $\mathcal{F}_d$ of the form~\eqref{fd}: \begin{equation}} \def\eeq{\end{equation}\hat u\in\RR^{\n}~~\mbox{such that}~~ \mathcal{F}_d(\hat u)=\min_{u\in\O}\mathcal{F}_d(u)~,\label{P0}\eeq where $\O\,$ is an open neighborhood of $\hat u$~. We note that finding a global minimizer of $\mathcal{F}_d\,$ must be an {\em NP-hard} computational problem \cite{DavisMallatAvellaneda,Tropp06}.
The function $\phi$ in~\eqref{phi} served as a regularizer for a long time. In the context of Markov random fields it was used by Geman and Geman in 1984 \cite{Geman84} and Besag in 1986 \cite{Besag86}
as a prior in MAP energies to restore labeled images. The MAP objective reads as
\begin{equation}} \def\eeq{\end{equation}\mathcal{F}_d(u)=\|Au-d\|^2_2+\beta\sum_{k}\phi(D_ku)~,\label{map}\eeq where $D_k$ is a finite difference operator and $\phi$ is given by~\eqref{phi}. This label-designed form is known as the Potts prior model, or as the multi-level logistic model \cite{Besag89,Li95}. Various stochastic and deterministic algorithms have been considered to minimize~\eqref{map}. Leclerc \cite{Leclerc89} proposed in 1989 a deterministic continuation method to restore piecewise constant images. Robini, Lachal and Magnin \cite{Robini07} introduced the stochastic continuation approach and successfully used it to reconstruct 3D tomographic images. Robini and Magnin refined the method and the theory in \cite{Robini10}. Very recently, Robini and Reissman \cite{Robini12} gave theoretical results relating the probability for global convergence and the computation speed.
The problem stated in~\eqref{fd} and~\eqref{P0}---to (locally) minimize $\mathcal{F}_d$---arises when {\em sparse\,} solutions are desired. Typical application fields are signal and image processing, morphologic component analysis, compression, dictionary building, inverse problems, compressive sensing, machine learning, model selection, classification, and subset selection, among others. The original hard-thresholding method proposed by Donoho and Johnstone \cite{Donoho94} amounts to\footnote{As a reminder, if $d$ are some noisy coefficients, the restored coefficients $\hat u$
minimize $\|u-d\|^2+\beta\|u\|_0$ and read $\hat u[i]=0$ if $\left|d[i]\right|\leqslant\sqrt{\beta}$ and $\hat u[i]=d[i]$ otherwise. }
minimizing~$\mathcal{F}_d$, where $d$ contains the coefficients of a signal or an image expanded in a wavelet basis ($\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}=\n$). When $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}<\n$, various (usually strong) restrictions on $\|u\|_0$ (often $\|u\|_0$ is replaced by a less irregular function)
and on $A$ (e.g., RIP-like criteria, conditions on $\|A\|$, etc.) are needed to conceive numerical schemes approximating a minimizer of $\mathcal{F}_d$, to establish local convergence and derive the asymptotic of the obtained solution. In statistics the problem has been widely considered for subset selection, and numerous algorithms have been designed, with limited theoretical production, as explained in the book by Miller~\cite{Miller02}. More recently, Haupt and Nowak \cite{Haupt06} investigate the statistical performances of the global minimizer of~$\mathcal{F}_d$ and propose an iterative bound-optimization procedure. Fan and Li \cite{Fan06} discuss a variable splitting and penalty decomposition minimization technique for \eqref{fd}, along with other approximations of the $\ell_0$-norm. Liu and Wu \cite{Liu07} mix the $\ell_0$ and $\ell_1$ penalties, establish some asymptotic properties of the new estimator and use mixed integer programming aimed at global minimization. For model selection, Lv and Fan \cite{Lv09} approximate the $\ell_0$ penalty using functions that are concave on $\mathbb R_+$ and prove a nonasymptotic nearly oracle property of the resultant estimator. Thiao, Dinh, and Thi~\cite{Thiao08} reformulate the problem so that an approximate solution can be found using difference-of-convex-functions programming. Blumensath and Davies \cite{Davies08} propose an iterative thresholding scheme to approximate a solution and prove convergence to a local minimizer of $\mathcal{F}_d$. Lu and Zhang \cite{LuZhang10} suggest a penalty decomposition method to minimize $\mathcal{F}_d$. Fornasier and Ward \cite{Fornasier10} propose an iterative thresholding algorithm for minimizing
$\mathcal{F}_d$ where $\ell_0$ is replaced by a reasonable sparsity-promoting relaxation given by $\phi(t)=\min\{|t|,1\}$; then convergence to a local minimizer is established. In a recent paper by Chouzenoux et al.~\cite{Pesquet12}, a mixed $\ell_2-\ell_0$ regularization is considered: a slightly smoothed version of the objective is analyzed and
a majorize-minimize subspace approach, satisfying a finite length property, converges to a critical point. Since the submission of our paper, image reconstruction methods have been designed where $\ell_0$ regularization is applied to the coefficients of the expansion of the sought-after image in a wavelet frame~\cite{Zhang12,Dong12}: the provided numerical results outperform $\ell_1$ regularization for a reasonable computational cost achieved using penalty decomposition techniques. In a general study on the convergence of descent methods for nonconvex objectives, Attouch, Bolte, and Svaiter~\cite{Attouch11} apply an inexact forward-backward splitting scheme to find a critical point of $\mathcal{F}_d$. Several other references can be evoked, e.g., \cite{Schorr05,Gasso09}.
Even though overlooked for several decades, the objective $\mathcal{F}_d$ was essentially considered from a numerical standpoint. The motivation naturally comes from the promising applications and the intrinsic difficulty of minimizing $\mathcal{F}_d$.
{\em The goal of this work is to analyze the (local and global) minimizers $\hat u$ of objectives $\mathcal{F}_d$ of the form~\eqref{fd}. \begin{itemize}} \def\eit{\end{itemize} \item We provide detailed results on the minimization problem. \item The uniqueness of the global minimizer of $\mathcal{F}_d$ is examined as well. \eit} \noindent We do not propose an algorithm. However, our theoretical results raise salient questions about the existing algorithms and can help the conception of innovative numerical schemes.
The minimization of $\mathcal{F}_d$ in~\eqref{fd} might seem close to its constraint variants: \begin{equation}} \def\eeq{\end{equation} \label{ka}\begin{array}} \def\earr{\end{array}{lll}
\mbox{given } \varepsilon\geq0,~&\mbox{minimize}~~\|u\|_0&~\mbox{subject to}~~\|Au-d\|^2\leqslant \varepsilon~,\\
\mbox{given } K\in\II_{\m},~&\mbox{minimize}~\|Au-d\|^2&~\mbox{subject to}~~\|u\|_0\leqslant K~. \earr\eeq The latter problems are abundantly studied in the context of sparse recovery in different fields. An excellent account is given in \cite{DonBruckElad09}, see also the book \cite{Mallat08}. For recent achievements, we refer the reader to~\cite{CSR-URL}. It is worth emphasizing that in general, {\em there is no equivalence between the problems stated in~\eqref{ka} and the minimization of $\mathcal{F}_d$ in~\eqref{fd}} because all of these problems are nonconvex.
\subsection{Main notation and definitions}\label{ntd}
We recall that if $\hat u$ is a {\em (local) minimizer} of $\mathcal{F}_d$, the value $\mathcal{F}_d(\hat u)$ is a (local) minimum\footnote{These two terms are often confused in the literature.} of $\mathcal{F}_d$ reached at (possibly numerous) points $\hat u$. Saying that a (local) minimizer $\hat u$ is {\em strict} means that there is a neighborhood $\O\subset\RR^{\n}$, containing $\hat u$, such that $\mathcal{F}_d(\hat u)<\mathcal{F}_d(v)$ for any $v\in\O\,\mathbf{\setminus}\,\{\hat u\}$. So $\hat u$ is an isolated minimizer.
Let $\k$ be any positive integer. The expression $\big\{u\in\mathbb R^\k~:~u~~\mbox{satisfying property}~~\mathfrak{P}\big\}$ designates the subset of $\mathbb R^\k$ formed from all elements $u$ that meet $\mathfrak{P}$. The identity operator on $\mathbb R^\k$ is denoted by $I_\k$. The entries of a vector $u\in\mathbb R^\k$ read as $u[i]$, for any $i$. The $i$th vector of the canonical basis\footnote{More precisely, for any $i\in\II_{\k}$, the vector $e_i\in\mathbb R^\k$ is defined by $e_i[i]=1$ and $e_i[j]=0,~\forall\; j\in\II_{\k}\,\mathbf{\setminus}\,\{i\}$. }
of $\mathbb R^\k$ is denoted by $e_i\in\mathbb R^\k$. Given $u\in\mathbb R^\k$ and $\rho>0$, the {\em open} ball at $u$ of radius $\rho$ with respect to the $\ell_p$-norm for $1\leqslant p\leqslant\infty$ reads as
\[ \Bm_p(u,\rho)\stackrel{\rm def}{=}\{v\in\mathbb R^\k~:~\|v-u\|_p<\rho\}~.\] To simplify the notation, the $\ell_2$-norm is {\em systematically} denoted by
$$\|\cdot\|\stackrel{\rm def}{=}\|\cdot\|_2~.$$ We denote by $\II_{\k}$ the {\em totally and strictly ordered} index set\footnote{E.g. without strict order we have $\omega=\{1,2,3\}=\{2,1,1,3\}$ in which case the notation in \eqref{as}-\eqref{us} below is ambiguous.} \begin{equation}} \def\eeq{\end{equation}\II_{\k}\stackrel{\rm def}{=}\big(\{1,\cdots,\k\},<\big)~,\label{ik}\eeq where the symbol $<$ stands for the natural order of the positive integers. Accordingly, {\em any subset $\omega\subseteq\II_{\k}$ inherits the property of being totally and strictly ordered.}
We shall often consider the index set $\II_{\n}$. The complement of $\omega\subseteq\II_{\n}$ in $\II_{\n}$ is denoted by \[\omega^c=\II_{\n}\,\mathbf{\setminus}\,\omega\subseteq \II_{\n}~.\]
\begin{definition}} \def\ED{\end{definition}\label{sp} For any $u\in\mathbb R^\n$, the {\em support} $\sigma(u)$ of $u$ is defined by \begin{equation}} \def\eeq{\end{equation}\sigma(u)=\Big\{i\in\II_{\n}~:~ u[i]\neq 0\Big\}\subseteq \II_{\n}~.\nonumber\label{s}\eeq \ED If $u=0$, clearly $\sigma(u)=\varnothing$.
The $i$th column in a matrix $A\in\RR^{\m\x\n}$ is denoted by $a_i$. It is {\em systematically} assumed that \begin{equation}} \def\eeq{\end{equation}\mbox{\framebox{$\disp{~a_i\neq0~~~~\forall\; i\in\II_{\n}~.}$}}\label{ao}\eeq For a matrix $A\in\RR^{\m\x\n}$ and a vector $u\in\RR^{\n}$, with any $\omega\subseteq\II_{\n}$, we associate the {\em submatrix} $A_{\vt}} \def\Pvt{\Pi_{\vt}$ and the {\em subvector} $u_\omega$ given by \begin{eqnarray}} \def\eeqn{\end{eqnarray} A_{\vt}} \def\Pvt{\Pi_{\vt}&\stackrel{\rm def}{=}&\big(a_{\omega[1]},\cdots,a_{\omega[\,\sharp\,\omega]}\big)\in\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times \,\sharp\,\omega}~,\label{as}\\ u_\omega&\stackrel{\rm def}{=}&\Big(u\big[\omega[1]\big],\cdots,u\big[\omega[\,\sharp\,\omega]\big]\Big)\in\mathbb R^{\,\sharp\,\omega}\label{us}~, \eeqn respectively, as well as the zero padding operator $Z_\omega~:~\mathbb R^{\,\sharp\,\omega}\rightarrow\RR^{\n}$ that inverts~\eqref{us}: \begin{equation}} \def\eeq{\end{equation} u=Z_\omega\left(u_\omega\right)~,~~u[i]=\left\{ \begin{array}} \def\earr{\end{array}{ll}0&\mbox{\rm if}~~i\not\in\omega~, \\ u_\omega[k]&\mbox{for the unique $k$ such that}~~\omega[k]=i.\earr \right.\label{zo}\eeq Thus for $\omega=\varnothing$ one finds $u_{\varnothing}=\varnothing~~~\mbox{\rm and}~~~u=Z_{\varnothing}\left(u_{\varnothing}\right)=0\in\RR^{\n}~.$
Using Definition~\ref{sp} and the notation in \eqref{as}-\eqref{us}, for any $u\in\RR^{\n}\,\mathbf{\setminus}\,\{0\}$ we have \begin{equation}} \def\eeq{\end{equation}\label{sd} \omega\in\II_{\n}~~\mbox{\rm and}~~\omega\supseteq\sigma(u)~~~\Rightarrow~~~Au=A_{\vt}} \def\Pvt{\Pi_{\vt} u_{\omega}~.\eeq
To simplify the presentation, we adopt the following {\em definitions} \footnote{Note that $(a)$ corresponds to the zero mapping on $\mathbb R^0$ and that $(b)$ is the usual definition for the rank of an empty matrix.}: \begin{equation}} \def\eeq{\end{equation}\begin{array}} \def\earr{\end{array}{lll} (a)&& A_{\varnothing}=[\ ]\in\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times 0}~,\\ (b)&&\mathrm{rank}\left(A_{\varnothing}\right)=0~.\earr\label{cw}\eeq In order to avoid possible ambiguities\footnote{In the light of~\eqref{as}, $A_\omega^T$ could also mean $\left(A^T\right)_\omega$.}, we set \[A_{\vt}} \def\Pvt{\Pi_{\vt}^T\stackrel{\rm def}{=}\left(A_{\vt}} \def\Pvt{\Pi_{\vt}\right)^T~,\] where the superscript $^T$ stands for transposed. If $A_{\vt}} \def\Pvt{\Pi_{\vt}$ is invertible, similarly $A_{\vt}} \def\Pvt{\Pi_{\vt}^{-1}\stackrel{\rm def}{=}\left(A_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}$.
In the course of this work, we shall frequently refer to the constrained quadratic optimization problem stated next.
{\em Given $d\in\RR^{\m}$ and $\omega\subseteq\II_{\n}$, problem {\rm{($\,\P_{\vt}\,$)}~} reads as: \begin{equation}} \def\eeq{\end{equation}~~~~~~~~~~~~~~~~~~~~~~~~~~~\framebox{\mbox{$\left\{\begin{array}} \def\earr{\end{array}{lll}
&&\disp{\min_{u\in\RR^{\n} }\|Au-d\|^2}~,\\~\\ \mbox{\rm subject to}&& u[i]=0,~~\forall\; i\in \omega^c~.\\ \earr\right.$}}~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\mbox{{\rm{($\,\P_{\vt}\,$)}~}}\label{cpa}\eeq } {\em Clearly, problem {\rm{($\,\P_{\vt}\,$)}~} always admits a solution.}
The definition below will be used to evaluate the extent of some subsets and assumptions. \begin{definition}} \def\ED{\end{definition}\label{ps} A property (an assumption) is called {\em generic on} $\mathbb R^\k$ if it holds true on a {\em dense open} subset of $\mathbb R^\k$. \ED
As usual, a subset ${\mathcal{S}}\subset\mathbb R^\k$ is said to be {\em negligible} in $\mathbb R^\k$ if there exists $\Z\subset\mathbb R^\k$ whose Lebesgue measure in $\mathbb R^\k$ is $\mathbb{L}^\k(\Z)=0$ and ${\mathcal{S}}\subseteq\Z$~. If a property fails only on a negligible set, it is said to hold
{\em almost everywhere}, meaning ``with probability one''. Definition~\ref{ps} requires much more than {\em almost everywhere\,}. Let us explain.
{\em If a property holds true for all $v\in\mathbb R^\k\,\mathbf{\setminus}\,{\mathcal{S}}$, where ${\mathcal{S}}\subseteq\Z\subset\mathbb R^\k$, $\Z$ is {\em closed in $\mathbb R^\k$} and $\mathbb{L}^\k(\Z)=0$, then this property is {\em generic} on $\mathbb R^\k$.} Indeed, $\mathbb R^\k\,\mathbf{\setminus}\,\Z$ contains a {\em dense open} subset of $\mathbb R^\k$. So if a property is generic on $\mathbb R^\k$, then it holds true almost everywhere on $\mathbb R^\k$. But the converse is false: an almost everywhere true property is not generic if the closure of its negligible subset has a positive measure,\footnote{ There are many examples---e.g. $\Z=\{x\in[0,1]~:~x ~\mbox{is rational}\}$, then $\mathbb{L}^1(\Z)=0$ and $\mathbb{L}^1(\mbox{closure}(\Z))=1$.} because then $\mathbb R^\k\,\mathbf{\setminus}\,\Z$ does not contains an open subset of $\mathbb R^\k$. In this sense, a generic property is stable with respect to the objects to which it applies.
{\em The elements of a set ${\mathcal{S}}\subset\mathbb R^\k$ where a generic property fails are highly exceptional in~$\mathbb R^\k$.} The chance that a truly random $v\in\mathbb R^\k$---i.e., a $v$ following a non singular probability distribution on $\mathbb R^\k$---comes across such an ${\mathcal{S}}$ can be ignored in practice.
\subsection{Content of the paper} The main result in section~\ref{LM} tells us that finding a solution of {\rm{($\,\P_{\vt}\,$)}~} for $\omega\subset\II_{\n}$ is {\em equivalent} to computing a (local) minimizer of $\mathcal{F}_d$. In section~\ref{sSM} we prove that the (local) minimizers $\hat u$ of $\mathcal{F}_d$ are {\em strict} if and only if the submatrix $A_{\sigma(\hat u)}$ has full column rank. The strict minimizers of $\mathcal{F}_d$ are shown to be linear in data $d$. The importance of the $(\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1)$-sparse strict minimizers is emphasized. The global minimizers of $\mathcal{F}_d$ are studied in section~\ref{sgm}. Their existence is proved. They are shown to be strict for any $d$ and for any $\beta>0$. More details are provided under the standard assumption that $\mathrm{rank}(A)=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}<\n$. Given $d\in\RR^{\m}$, critical values $\beta_\k$ for $\k\in\II_{\m-1}$ are exhibited such that all global minimizers of $\mathcal{F}_d$ are $\k$-sparse\footnote{
As usual, a vector $u$ is said to be $\k$-sparse if $\|u\|_0\leqslant\k$.}
if $\beta>\beta_\k$.
In section~\ref{kb}, a gentle assumption on $A$ is shown to be {\em generic} for all $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\n$ real matrices. Under this assumption, for all data $d\in\RR^{\m}$ beyond a closed negligible subset, the objective $\mathcal{F}_d$ for $\beta>\beta_\k$, $\k\in\II_{\m-1}$, has a unique global minimizer and this minimizer is $\k$-sparse.
Small size ($A$ is $5\x10$) numerical tests in section~\ref{nm} illustrate the main theoretical results.
\section{All minimizers of $\bm{\mathcal{F}_d}$}\label{LM}
\subsection{Preliminary results}\label{pr}
First, we give some basic facts on problem {\rm{($\,\P_{\vt}\,$)}~} as defined in \eqref{cpa} that are needed for later use. If $\omega=\varnothing$, then $\omega^c=\II_{\n}$, so the unique solution of {\rm{($\,\P_{\vt}\,$)}~} is $\hat u=0$. For an arbitrary $\omega\subset\II_{\n}$ meeting $\,\sharp\,\omega\geq1$, {\rm{($\,\P_{\vt}\,$)}~} amounts to minimizing a quadratic term with respect to only $\,\sharp\,\omega$ components of $u$, the remaining entries being null. This {\em quadratic} problem {\rm{($\,\Q_{\vt}\,$)}~} reads as \begin{equation}} \def\eeq{\end{equation}\!~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
\disp{\min_{v\in\mathbb R^{\,\sharp\,\omega}}\big\|A_{\vt}} \def\Pvt{\Pi_{\vt} v-d\big\|^2},~~\,\sharp\,\omega\geq1~,~\label{cpr} ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\mbox{{\rm{($\,\Q_{\vt}\,$)}~}} \eeq and it always admits a solution. Using the zero-padding operator $Z_\omega$ in \eqref{zo}, we have \[\Big[~\hu_{\vt}\in\mathbb R^{\,\sharp\,\omega}~\mbox{\rm solves {\rm{($\,\Q_{\vt}\,$)}~} and}~~\hat u=Z_\omega\left(\hu_{\vt}\right)~\Big]~~~\Leftrightarrow~~~ \Big[~\mbox{$\hat u\in\RR^{\n}$ solves {\rm{($\,\P_{\vt}\,$)}~}},~~\,\sharp\,\omega\geq1~\Big]~.\] The optimality conditions for {\rm{($\,\Q_{\vt}\,$)}~}, combined with the definition in~\eqref{cw}(a), give rise to the following equivalence, which holds true for any $\omega\subseteq\II_{\n}$: \begin{equation}} \def\eeq{\end{equation} \Big[~\mbox{$\hat u\in\RR^{\n}$ solves {\rm{($\,\P_{\vt}\,$)}~}}\Big]~~\Leftrightarrow~~ \Big[~\hu_{\vt}\in\mathbb R^{\,\sharp\,\omega}~~\mbox{\rm solves $~\disp{A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\, v=A_{\vt}} \def\Pvt{\Pi_{\vt}^Td~}$ and}~~\hat u=Z_\omega\left(\hu_{\vt}\right)~\Big]. \label{eu}\eeq Note that $\disp{A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\, v=A_{\vt}} \def\Pvt{\Pi_{\vt}^Td}$ in~\eqref{eu} is the normal equation associated with $A_{\vt}} \def\Pvt{\Pi_{\vt}\,v=d$. The remark below shows that the optimal value of {\rm{($\,\P_{\vt}\,$)}~} in \eqref{cpa} can also be seen as an orthogonal projection problem.
\begin{remark} \label{ref} \rm Let $r\stackrel{\rm def}{=}\mathrm{rank}(A_{\vt}} \def\Pvt{\Pi_{\vt})$ and ${B_{\omega}}\in\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times r}$ be an orthonormal basis for $\ran(A_{\vt}} \def\Pvt{\Pi_{\vt})$. Then $A_{\vt}} \def\Pvt{\Pi_{\vt}=B_\omega H_\omega$ for a uniquely defined matrix ${H_{\omega}}\in\mathbb R^{r\times\,\sharp\,\omega}$ with $\mathrm{rank}({H_{\omega}})=r$. Using~\eqref{eu}, we have \begin{eqnarray*}} \def\eeqnn{\end{eqnarray*} A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\hu_{\vt}=A_{\vt}} \def\Pvt{\Pi_{\vt}^Td~~\Leftrightarrow~~{H_{\omega}}^T{H_{\omega}}\hu_{\vt}={H_{\omega}}^T{B_{\omega}}^Td ~~\Leftrightarrow~~{H_{\omega}}\hu_{\vt}={B_{\omega}}^Td~~\Leftrightarrow~~A_{\vt}} \def\Pvt{\Pi_{\vt}\hu_{\vt}={B_{\omega}}{B_{\omega}}^Td~. \eeqnn In addition, $\Pi_{\ran(A_{\vt}} \def\Pvt{\Pi_{\vt})}={B_{\omega}}{B_{\omega}}^T$ is the orthogonal projector onto the subspace spanned by the columns of $A_{\vt}} \def\Pvt{\Pi_{\vt}$, see e.g.~\cite{Meyer00}. The expression above combined with~\eqref{eu} shows that \[\Big[\,\mbox{$\hat u\in\RR^{\n}$ solves }(\P_{\omega})\,\Big]~~\Leftrightarrow~~ \Big[~\hu_{\vt}\in\mathbb R^{\,\sharp\,\omega}~~\mbox{\rm meets}~~A_{\vt}} \def\Pvt{\Pi_{\vt}\hu_{\vt}=\Pi_{\ran(A_{\vt}} \def\Pvt{\Pi_{\vt})}d~~~\mbox{and}~~\hat u=Z_\varpi(\hu_{\vs})\,\Big]. \label{pref}\] \em Obviously, $A\hat u=A_{\vt}} \def\Pvt{\Pi_{\vt}\hu_{\vt}$ is the orthogonal projection of $d$ onto the basis~$B_\omega$.\end{remark}
For $\omega\subseteq\II_{\n}$, let $\Km_\omega$ denote the vector subspace \begin{equation}} \def\eeq{\end{equation}\Km_\omega\stackrel{\rm def}{=}\big\{v\in\RR^{\n}~:~v[i]=0,~\forall\; i\in\omega^c\big\}~.\label{km}\eeq This notation enables problem {\rm{($\,\P_{\vt}\,$)}~} in \eqref{cpa} to be rewritten as
\begin{equation}} \def\eeq{\end{equation}\disp{\min_{u\in\Km_\omega}\|Au-d\|^2~.}\label{fa}\eeq
The technical lemma below will be used in what follows. We emphasize that its statement is {\em independent} of the vector $\hat u\in\RR^{\n}\,\mathbf{\setminus}\,\{0\}$.
\begin{lemma}} \def\EL{\end{lemma}\label{tl} Let $d\in\RR^{\m}$, $\beta>0$, and $\hat u\in\RR^{\n}\,\mathbf{\setminus}\,\{0\}$ be {\em arbitrary}. For ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)$, set \begin{equation}} \def\eeq{\end{equation}
\rho\stackrel{\rm def}{=}\min\left\{\min_{i\in{\hat\sigma}}\big|\,\hat u[i]\,\big|,~\frac{\beta}{2\Big(\|A^T(A\hat u-d)\|_1+1\Big)}\right\}. \label{rho}\eeq Then $\rho>0$. \begin{itemize}} \def\eit{\end{itemize} \item[\rm(i)] For $\phi$ as defined in~\eqref{phi}, we have \begin{equation}} \def\eeq{\end{equation}\nonumber\label{tlo} v\in \Bm_\infty(0,\rho)~~\Rightarrow~~\sum_{i\in\II_{\n}}\phi\big(\,\hat u[i]+v[i]\,\big)= \sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]\right)+\sum_{i\in{\hat\sigma}^c}\phi\left(v[i]\right)~. \eeq \item[\rm(ii)] For $\Km_{{\hat\sigma}}$ defined according to \eqref{km}, $\mathcal{F}_d$ satisfies \begin{equation}} \def\eeq{\end{equation} v\in\Bm_\infty(0,\rho)\cap\big(\RR^{\n}\,\mathbf{\setminus}\,\Km_{{\hat\sigma}}\big)~~~\Rightarrow~~~\mathcal{F}_d(\hat u+v)\geqslant\mathcal{F}_d(\hat u) ~, \nonumber\label{pip}\eeq where the inequality is {\em strict} whenever ${\hat\sigma}^c\neq\varnothing$. \eit \EL
The proof is outlined in Appendix~\ref{dtl}.
\subsection{The (local) minimizers of $\bm{\mathcal{F}_d}$ solve quadratic problems}\label{fip}~~
{\em It is worth emphasizing that no special assumptions on the matrix $A$ are adopted.}
We begin with an easy but cautionary result.
\begin{lemma}} \def\EL{\end{lemma}\label{uz} For any $d\in\RR^{\m}$ and for all $\beta>0$, $\mathcal{F}_d$ has a {\em strict} (local) minimum at $\hat u=0\in\RR^{\n}$. \EL
\par\paragraph{Proof} \ignorespaces Using the fact that $\mathcal{F}_d(0)=\|d\|^2\geq0$, we have \begin{eqnarray}} \def\eeqn{\end{eqnarray}
\mathcal{F}_d(v)&=&\|Av-d\|^2+\beta\|v\|_0=\mathcal{F}_d(0)+\mathcal{R}_d(v)~,\label{wq}\\ \mbox{where}\quad
\mathcal{R}_d(v)&=&\|Av\|^2-2\<v,A^Td\>+\beta\|v\|_0~.\eeqn Noticing that $\beta\|v\|_0\geqslant\beta>0$ for $v\neq0$ leads to
\[v\in \Bm_2\left(0,\frac{\beta}{2\|A^Td\|+1}\right)\setminus\{0\}~~~\Rightarrow~~~
\mathcal{R}_d(v)\geqslant-2\|v\|\;\|A^Td\|+\beta >0~.\] Inserting this implication into~\eqref{wq} proves the lemma.
$\Box$\newline
{\em For any $\beta>0$ and $d\in\RR^{\m}$, the sparsest strict local minimizer of $\mathcal{F}_d$ reads $\hat u=0$. Initialization with zero of a suboptimal algorithm should generally be a bad choice.} Indeed, experiments have shown that such an initialization can be harmful; see, e.g., \cite{Miller02,Davies08}.
The next proposition states a result that is often evoked in this work.
\begin{proposition}} \def\EP{\end{proposition}\label{cp} Let $d\in\RR^{\m}$. Given an $\omega\subseteq\II_{\n}$, let $\hat u$ solve problem {\rm{($\,\P_{\vt}\,$)}~} as formulated in~\eqref{cpa}. Then for any $\beta>0$, the objective $\mathcal{F}_d$ in~\eqref{fd} reaches a (local) minimum at $\hat u$ and \begin{equation}} \def\eeq{\end{equation}\sigma(\hat u)\subseteq\omega~,\label{cop}\eeq where $\sigma(\hat u)$ is given in Definition~\ref{sp}. \EP
\par\paragraph{Proof} \ignorespaces Let $\hat u$ solve problem {\rm{($\,\P_{\vt}\,$)}~}, and let $\beta>0$. The constraint in {\rm{($\,\P_{\vt}\,$)}~} entails~\eqref{cop}.
Consider that $\hat u\neq0$, in which case for ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)$ we have $1\leqslant\,\sharp\,{\hat\sigma}\leqslant\,\sharp\,\omega$. Using the equivalent formulation of {\rm{($\,\P_{\vt}\,$)}~} given in~\eqref{km}-\eqref{fa}, yields
\begin{equation}} \def\eeq{\end{equation} v\in\Km_\omega~~~~\Rightarrow~~~\|A(\hat u+v)-d\|^2\geqslant\|A\hat u-d\|^2~.\label{sqi}\eeq The inclusion in \eqref{cop} is equivalent to $\omega^c\subseteq{\hat\sigma}^c~.$ Let $\Km_{{\hat\sigma}}$ be defined according to~\eqref{km} as well. Then \[\hat u\in\Km_{{\hat\sigma}}\subseteq\Km_\omega.\] Combining~the latter relation with~\eqref{sqi} leads to \begin{equation}} \def\eeq{\end{equation}\label{yui}
v\in\Km_{{\hat\sigma}}~~~\Rightarrow~~~\|A(\hat u+v)-d\|^2\geqslant\|A\hat u-d\|^2~. \eeq Let $\rho$ be defined as in~\eqref{rho} Lemma~\ref{tl}. Noticing that by \eqref{phi} and \eqref{km} \begin{equation}} \def\eeq{\end{equation} v\in\Km_{{\hat\sigma}}~~~\Rightarrow~~~\phi\left(v[i]\right)=0~~~\forall\; i\in{\hat\sigma}^c~, \label{rer}\eeq the following inequality chain is derived: \begin{eqnarray*}} \def\eeqnn{\end{eqnarray*} v\in \Bm_\infty(0,\rho)\cap\Km_{{\hat\sigma}}~~~\Rightarrow~~~
\mathcal{F}_d(\hat u+v)&=&\|A(\hat u+v)-d\|^2+\beta\sum_{i\in\II_{\n}}\phi\left(\hat u[i]+v[i]\right)\nonumber\\ \Big[\mbox{\rm by Lemma~\ref{tl}(i)}\Big]~~~~
&=&\|A(\hat u+v)-d\|^2+\beta\sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]\right)+\beta\sum_{i\in{\hat\sigma}^c}\phi\left(v[i]\right)\nonumber\\ \Big[\mbox{\rm by~\eqref{rer}}\Big]~~~~
&=&\|A(\hat u+v)-d\|^2+\beta\sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]\right)\nonumber\\
\Big[\mbox{\rm by~\eqref{yui}}\Big]~~~~&\geqslant&\|A\hat u-d\|^2+\beta\sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]\right)\nonumber\\ \Big[\mbox{\rm by~\eqref{fds}}\Big]~~~~&=&\mathcal{F}_d(\hat u) ~.\label{4v}\eeqnn Combining the obtained implication with~Lemma~\ref{tl}(ii) shows that \[\mathcal{F}_d(\hat u+v)\geqslant\mathcal{F}_d(\hat u)~~~\forall\; v\in \Bm_\infty(0,\rho)~.\]
If $\hat u=0$, this is a (local) minimizer of $\mathcal{F}_d$ by Lemma~\ref{uz}.
$\Box$\newline
Many authors mention that initialization is paramount for the success of approximate algorithms minimizing $\mathcal{F}_d$. In view of Proposition~\ref{cp}, if one already has a well-elaborated initialization, it could be enough to solve the relevant problem {\rm{($\,\P_{\vt}\,$)}~}.
The statement reciprocal to Proposition~\ref{cp} is obvious but it helps the presentation.
\begin{lemma}} \def\EL{\end{lemma}\label{ob} For $d\in\RR^{\m}$ and $\beta>0$, let $\mathcal{F}_d$ have a (local) minimum at $\hat u$. Then $\hat u$ solves {\rm{($\,\P_{\hsi}\,$)}~} for ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)$. \EL
\par\paragraph{Proof} \ignorespaces Let $\hat u$ be a (local) minimizer of $\mathcal{F}_d$. Denote ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)$. Then $\hat u$ solves the problem
\[\min_{u\in\RR^{\n}}\Big\{\|Au-d\|^2+\beta\,\sharp\,{\hat\sigma}\Big\}~~\mbox{subject to } u[i]=0~~\forall\; i\in {\hat\sigma}^c.\] Since $\,\sharp\,{\hat\sigma}$ is a constant, $\hat u$ solves {\rm{($\,\P_{\hsi}\,$)}~}.
$\Box$\newline
\begin{remark}} \def\ER{\end{remark} ~~ By Proposition~\ref{cp} and Lemma~\ref{ob}, solving {\rm{($\,\P_{\vt}\,$)}~} for some $\omega\subseteq\II_{\n}$ \\ is {\em equivalent} to finding a (local) minimizer of~$\mathcal{F}_d$. \label{x1}\ER
This equivalence underlies most of the theory developed in this work.
\begin{corollary}} \def\EC{\end{corollary}\label{nor} For $d\in\RR^{\m}$ and $\beta>0$, let $\hat u$ be a (local) minimizer of $\mathcal{F}_d$. Set ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)$. Then \begin{equation}} \def\eeq{\end{equation}\label{ee}\hat u=Z_{{\hat\sigma}}(\hu_{\hsi})~,~~~\mbox{\rm where $~\hu_{\hsi}~$ satisfies}~~~A_{\hsi}} \def\Phs{\Pi_{\hsi}^TA_{\hsi}} \def\Phs{\Pi_{\hsi}\hu_{\hsi}=A_{\hsi}} \def\Phs{\Pi_{\hsi}^Td~. \eeq Conversely, if $\hat u\in\RR^{\n}$ satisfies \eqref{ee} for ${\hat\sigma}=\sigma(\hat u)$, then $\hat u$ is a (local) minimizer of $\mathcal{F}_d$. \EC \par\paragraph{Proof} \ignorespaces By Lemma~\ref{ob}, $\hat u$ solves {\rm{($\,\P_{\hsi}\,$)}~}. The equation in~\eqref{ee} follows directly from \eqref{eu}. The last claim is a straightforward consequence of \eqref{eu} and Proposition~\ref{cp}.
$\Box$\newline
\begin{remark}} \def\ER{\end{remark} \label{x2} \rm Equation \eqref{ee} shows that a (local) minimizer $\hat u$ of $\mathcal{F}_d$ follows a {\em pseudo}-hard thresholding scheme\footnote{In a Bayesian setting, the quadratic data fidelity term in~$\mathcal{F}_d$ models data corrupted with Gaussian i.i.d. noise.}: the nonzero part $\hu_{\hsi}$ of $\hat u$ is the least squares solution with respect to the submatrix $A_{\hsi}} \def\Phs{\Pi_{\hsi}$ and the whole data vector $d$ is involved in its computation. Unlike the hard thresholding scheme in \cite{Donoho94}, unsignificant or purely noisy data entries can hardly be discarded from $\hat u$ and they threaten to pollute its nonzero part $\hu_{\hsi}$. See also Remark~\ref{ifn}. \ER
Noisy data $d$ should degrade $\hu_{\hsi}$ and this effect is stronger if $A_{\hsi}} \def\Phs{\Pi_{\hsi}^TA_{\hsi}} \def\Phs{\Pi_{\hsi}$ is ill-conditioned \cite{Demoment89}. The quality of the outcome critically depends on the selected (local) minimizer and on the pertinence of $A$.
It may be interesting to evoke another consequence of Proposition~\ref{cp}: \begin{remark}} \def\ER{\end{remark} \label{x3} Given $d\in\RR^{\m}$, for any $\omega\subseteq\II_{\n}$, $\mathcal{F}_d$ has a (local) minimizer $\hat u$ defined by \eqref{ee} and obeying $\sigma(\hat u)\subseteq\omega$. \ER
\section{The strict minimizers of $\bm{\mathcal{F}_d}$}\label{sSM}~~
{\em We remind, yet again, that no special assumptions on $A\in\RR^{\m\x\n}$ are adopted.}
Strict minimizers of an objective function enable unambiguous solutions of inverse problems.
The definition below is useful in characterizing the strict minimizers of $\mathcal{F}_d$.
\begin{definition}} \def\ED{\end{definition}\label{om} Given a matrix $A\in\RR^{\m\x\n}$, for any $r\in\II_{\m}$ we define $\Omega_r$ as the subset of {\em all} $r$-length supports that correspond to full column rank $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times r$ submatrices of $A$, i.e., \begin{equation}} \def\eeq{\end{equation}\mbox{\framebox{$\disp{~ \Omega_r=\Big\{~\omega\subset\II_{\n}~:~~\,\sharp\,\omega=r=\mathrm{rank}\left(A_{\vt}} \def\Pvt{\Pi_{\vt}\right)~\Big\}~.}$}} \nonumber\label{sm}\eeq Set $\Omega_0=\varnothing$ and define as well \begin{equation}} \def\eeq{\end{equation}\mbox{\framebox{$\disp{~ \Omega\stackrel{\rm def}{=}\bigcup_{r=0}^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}\Omega_r~~\mbox{\rm and}~~\Om_{\max}\stackrel{\rm def}{=}\Omega\cup\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}~.}$}}\nonumber\label{si}\eeq \ED Definition~\ref{om} shows that for any $r\in\II_{\m}$, \[\mathrm{rank}(A)=r\geq1~~~\Leftrightarrow~~~\Omega_r\neq\varnothing~~\mbox{\rm and}~~\Omega_t=\varnothing~~\forall\; t\geqslant r+1~.\]
\subsection{How to recognize a strict minimizer of $\bm{\mathcal{F}_d}$?}\label{rec}
The theorem below gives an exhaustive answer to this question. \begin{theorem}} \def\ET{\end{theorem}\label{ra} Given $d\in\RR^{\m}$ and $\beta>0$, let $\hat u$ be a (local) minimizer of $\mathcal{F}_d$. Define \[{\hat\sigma}\stackrel{\rm def}{=} \sigma(\hat u)~.\] The following statements are {\em equivalent}: \begin{itemize}} \def\eit{\end{itemize} \item[\rm (i)] The (local) minimum that $\mathcal{F}_d$ has at $\hat u$ is {\em strict}; \item[\rm (ii)] \framebox{$\disp{\mathrm{rank}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)=\,\sharp\,{\hat\sigma}}$}~; \item[\rm(iii)] ${\hat\sigma}\in\Om_{\max}$~. \eit If $\hat u$ is a strict (local) minimizer of $\mathcal{F}_d$, then it reads \begin{equation}} \def\eeq{\end{equation} \hat u=Z_{{\hat\sigma}}\left(\hu_{\hsi}\right)~~~\mbox{for}~~~ \hu_{\hsi}=\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}^TA_{\hsi}} \def\Phs{\Pi_{\hsi}\right)^{-1}A_{\hsi}} \def\Phs{\Pi_{\hsi}^Td~\label{hu}
\eeq and satisfies $\,\sharp\,{\hat\sigma}= \|\hat u\|_0\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}~.$ \ET
\par\paragraph{Proof} \ignorespaces We break the proof into four parts.
\paragraph{\rm[(i)$\Rightarrow$(ii)]}~ We recall that by the rank-nullity theorem \cite{Golub96,Meyer00} \begin{equation}} \def\eeq{\end{equation}\mathrm{dim}\,\mathrm{ker}} \def\ran{\mathrm{range}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)=\,\sharp\,{\hat\sigma}-\mathrm{rank}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)~.\label{rn}\eeq Let\footnote{This part can alternatively be proven using Remark~\ref{ref}.}
$\hat u\neq0$ be a {\em strict} (local) minimizer of $\mathcal{F}_d$. Assume that (ii) fails. Then \eqref{rn} implies that \begin{equation}} \def\eeq{\end{equation} \mathrm{dim}\,\mathrm{ker}} \def\ran{\mathrm{range}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)\geq1~.\label{qa}\eeq By Lemma~\ref{ob}, $\hat u$ solves {\rm{($\,\P_{\hsi}\,$)}~}. Let $\rho$ read as in~\eqref{rho} and let $\Km_{{\hat\sigma}}$ be defined according to~\eqref{km}. Noticing that \begin{equation}} \def\eeq{\end{equation} v\in\Km_{{\hat\sigma}},~~{\hat\sigma}\neq\varnothing~~~\Rightarrow~~~Av=A_{\hsi}} \def\Phs{\Pi_{\hsi} v_{{\hat\sigma}}~,\label{ui} \eeq Lemma~\ref{tl}(i) shows that \begin{eqnarray*}} \def\eeqnn{\end{eqnarray*} \left\{\begin{array}} \def\earr{\end{array}{c}v\in \Bm_\infty(0,\rho)\cap\Km_{{\hat\sigma}}~,\\~~\\ v_{{\hat\sigma}}\in\mathrm{ker}} \def\ran{\mathrm{range}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)\earr\right.~~~\Rightarrow~~~
\mathcal{F}_d(\hat u+v)&=&\|A_{\hsi}} \def\Phs{\Pi_{\hsi}\left(\hu_{\hsi}+v_{{\hat\sigma}}\right)-d\|^2+\beta\sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]+v[i]\right)\\ \Big[\mbox{\rm by Lemma~\ref{tl}(i)}\Big]~~~~
&=&\|A_{\hsi}} \def\Phs{\Pi_{\hsi}\hu_{\hsi}-d\|^2+\beta\sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]\right)+\beta\sum_{i\in{\hat\sigma}^c}\phi\left(v[i]\right)\\
\Big[\mbox{\rm by \eqref{rer}~}\Big]~~~~&=&\|A_{\hsi}} \def\Phs{\Pi_{\hsi}\hu_{\hsi}-d\|^2+\beta\sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]\right)\\ \Big[\mbox{\rm by \eqref{fds}~}\Big]~~~~&=&\mathcal{F}_d(\hat u)~, \eeqnn i.e., that $\hat u$ is not a strict minimizer, which contradicts (i). Hence the assumption in~\eqref{qa} is false. Therefore (ii) holds true.
If $\hat u=0$, then ${\hat\sigma}=\varnothing$; hence $A_{\hsi}} \def\Phs{\Pi_{\hsi}\in\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\x0}$ and $\mathrm{rank}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)=0=\,\sharp\,{\hat\sigma}$ according to \eqref{cw}.
\paragraph{\rm$\mathrm{[(ii)\Rightarrow(i)]}$} Let $\hat u$ be a minimizer of $\mathcal{F}_d$ that satisfies (ii). To have $\,\sharp\,{\hat\sigma}=0$ is equivalent to $\hat u=0$. By Lemma~\ref{uz}, $\hat u$ is a strict minimizer. Focus on $\,\sharp\,{\hat\sigma}\geq1$. Since $\mathrm{rank}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)=\,\sharp\,{\hat\sigma}\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ and problem~{\rm{($\,\Q_{\vt}\,$)}~} in~\eqref{cpr} is strictly convex for $\omega={\hat\sigma}$, its unique solution $\hu_{\hsi}$ satisfies
\[v\in\mathbb R^{\,\sharp\,{\hat\sigma}}\,\mathbf{\setminus}\,\{0\}~~~\Rightarrow~~~\|A_{\hsi}} \def\Phs{\Pi_{\hsi}\left(\hu_{\hsi}+v\right)-d\|^2>\|A_{\hsi}} \def\Phs{\Pi_{\hsi}\hu_{\hsi}-d\|^2~.\] Using \eqref{ui}, this is equivalent to
\begin{equation}} \def\eeq{\end{equation} v\in\Km_{{\hat\sigma}}\,\mathbf{\setminus}\,\{0\}~~~\Rightarrow~~~\|A(\hat u+v)-d\|^2=\|A_{\hsi}} \def\Phs{\Pi_{\hsi}\left(\hu_{\hsi}+v_{{\hat\sigma}}\right)-d\|^2>\|A_{\hsi}} \def\Phs{\Pi_{\hsi}\hu_{\hsi}-d\|^2
=\|A\hat u-d\|^2~.\label{$}\eeq
Lemma~\ref{tl}(i), along with \eqref{rer}, yields \begin{eqnarray*}} \def\eeqnn{\end{eqnarray*} v\in\Bm_\infty(0,\rho)\cap\Km_{{\hat\sigma}}\,\mathbf{\setminus}\,\{0\}~~~\Rightarrow~~~\mathcal{F}_d(\hat u+v)
&=&\|A(u+v)-d\|^2+\beta\sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]\right)\\
\Big[\mbox{\rm by~\eqref{$}}\Big]~~~~~~~~~&>&\|A\hat u-d\|^2+\beta\sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]\right)\\ &=&\mathcal{F}_d(\hat u)~. \eeqnn Since $\,\sharp\,{\hat\sigma}\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\leqslant\n-1$, we have ${\hat\sigma}^c\neq\varnothing$. So Lemma~\ref{tl}(ii) tells us that \[v\in\Bm_\infty(0,\rho)\,\mathbf{\setminus}\,\Km_{{\hat\sigma}}~~~\Rightarrow~~~\mathcal{F}_d(\hat u+v)>\mathcal{F}_d(\hat u)~.\] Combining the last two implications proves (i).
\paragraph{\rm$\mathrm{[(ii)\Rightarrow(iii)]}$} Comparing (iii) with Definitions \ref{sp} and \ref{om} proves the claim.
\paragraph{\rm[Equation~\eqref{hu}]} The proof follows from equation~\eqref{ee} in Corollary~\ref{nor} where\footnote{For ${\hat\sigma}=\varnothing$, \eqref{zo} and~\eqref{cw}$(a)$ show that~\eqref{hu} yields $\hat u=0$.} $A_{\hsi}} \def\Phs{\Pi_{\hsi}^TA_{\hsi}} \def\Phs{\Pi_{\hsi}$ is invertible.
$\Box$\newline
{\em Theorem~\ref{ra} provides a simple rule enabling one to verify whether or not a numerical scheme has reached a strict (local) minimizer of $\mathcal{F}_d$.}
The notations $\Omega_r$, $\Omega$ and $\Om_{\max}$ are frequently used in this paper. Their interpretation is obvious in the light of Theorem~\ref{ra}. For any $d\in\RR^{\m}$ and for all $\beta>0$, the set $\Om_{\max}$ is composed of the supports of all possible strict (local) minimizers of $\mathcal{F}_d$, while $\Omega$ is is the subset of those that are $(\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1)$-sparse.
An easy and useful corollary is presented next.
\begin{corollary}} \def\EC{\end{corollary}\label{blg} Let $d\in\RR^{\m}$. Given an arbitrary $\omega\in\Om_{\max}$, let $\hat u$ solve {\rm{($\,\P_{\vt}\,$)}~}. Then \begin{itemize}} \def\eit{\end{itemize} \item[\rm(i)] $\hat u$ reads as \begin{equation}} \def\eeq{\end{equation}\hat u=Z_\omega\left(\hu_{\vt}\right)~,~~\mbox{where}~~~\hu_{\vt}=\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}A_{\vt}} \def\Pvt{\Pi_{\vt}^Td~,\label{bla}\eeq and obeys ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)\subseteq\omega$ and ${\hat\sigma}\in\Om_{\max}$~; \item[\rm(ii)] for any $\beta>0$, $\hat u$ is a {\em strict} (local) minimizer of $\mathcal{F}_d$; \item[\rm(iii)] $\hat u$ solves {\rm{($\,\P_{\hsi}\,$)}~}. \eit \EC
\par\paragraph{Proof} \ignorespaces Using~\eqref{eu}, $\hat u$ fulfills (i) since $A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}$ is invertible and $\sigma(\hat u)\subseteq\omega$ by the constraint in {\rm{($\,\P_{\vt}\,$)}~}. If ${\hat\sigma}=\varnothing$, (ii) follows from~Lemma~\ref{uz}. For $\,\sharp\,{\hat\sigma}\geq1$, $A_{\hsi}} \def\Phs{\Pi_{\hsi}$ is an $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\,\sharp\,{\hat\sigma}$ submatrix of $A_{\vt}} \def\Pvt{\Pi_{\vt}$. Since $\mathrm{rank}\left(A_{\vt}} \def\Pvt{\Pi_{\vt}\right)=\,\sharp\,\omega$, we have $\mathrm{rank}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)=\,\sharp\,{\hat\sigma}$ and so ${\hat\sigma}\in\Om_{\max}$. By~Proposition~\ref{cp} $\hat u$ is a (local) minimizer of $\mathcal{F}_d$, and~Theorem~\ref{ra} leads to (ii).
Lemma~\ref{ob} and Corollary~\ref{blg}(ii) yield (iii).
$\Box$\newline
\begin{remark}} \def\ER{\end{remark} \label{x4} One can easily compute a strict (local) minimizer $\hat u$ of $\mathcal{F}_d$ without knowing the value of the regularization parameter $\beta$. Just data $d$ and an $\omega\in\Om_{\max}$ are needed. \ER
This consequence of~Corollary~\ref{blg} might be striking.
}\def\bs{
}\def\ss{
Clearly, the support $\sigma(\bar u)$ of a nonstrict local minimizer $\bar{u}$ of $\mathcal{F}_d$ contains some subsupports yielding strict (local) minimizers of $\mathcal{F}_d$. It is easy to see that among them, there is ${\hat\sigma}\subsetneqq\sigma(\bar{u})$ such that the corresponding $\hat u$ given by \eqref{hu} {\em strictly} decreases the value of $\mathcal{F}_d$; i.e., $\mathcal{F}_d(\hat u)<\mathcal{F}_d(\bar{u}).$
\subsection{Every strict (local) minimizer of $\bm{\mathcal{F}_d}$ is linear in $\bm{d}$}\label{els} Here we explore the behavior of the strict (local) minimizers of $\mathcal{F}_d$ with respect to variations of~$d$. An interesting sequel of Theorem~\ref{ra} is presented in the following corollary.
\begin{corollary}} \def\EC{\end{corollary}\label{sk} For $d\in\RR^{\m}$ and $\beta>0$, let $\hat u$ be a (local) minimizer of $\mathcal{F}_d$ satisfying ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)\in\Omega~.$ Define \[\Nm_{{\hat\sigma}}\stackrel{\rm def}{=}\mathrm{ker}} \def\ran{\mathrm{range}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}^T\right)\subset\RR^{\m}~.\] We have $\,\mathrm{dim}\,\Nm_{{\hat\sigma}}=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-\,\sharp\,{\hat\sigma}\geq1$ and \[{d^{\,\prime}}\in\Nm_{{\hat\sigma}}~~~\Rightarrow~~~\mathcal{F}}\def\G{\mathcal{G}}\def\H{{\mathcal H}}\def\I{{\mathcal I}}\def\J{{\mathcal J}}\def\K{\mathcal{K}_{d+{d^{\,\prime}}}~~\mbox{has a strict (local) minimum at}~~\hat u~.\] \EC
\par\paragraph{Proof} \ignorespaces Since ${\hat\sigma}\in\Omega$, the minimizer $\hat u$ is strict by Theorem~\ref{ra}. By ${d^{\,\prime}}\in\mathrm{ker}} \def\ran{\mathrm{range}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}^T\right)$ we find $A_{\hsi}} \def\Phs{\Pi_{\hsi}^T(d+{d^{\,\prime}})=A_{\hsi}} \def\Phs{\Pi_{\hsi}^Td$ for any $\,{d^{\,\prime}}\in\Nm_{{\hat\sigma}}$. Inserting this into~\eqref{hu} in Theorem~\ref{ra} yields the result.
$\Box$\newline
All data located in the vector subspace $\Nm_{{\hat\sigma}}\supsetneqq\{0\}$ yield the same strict (local) minimizer $\hat u$.
\begin{remark} \label{ifn} {\rm If data contain noise $n$, it can be decomposed in a unique way as $n=n_{\Nm_{{\hat\sigma}}}+n_{\Nm_{{\hat\sigma}}^\bot}$ where $n_{\Nm_{{\hat\sigma}}}\in\Nm_{{\hat\sigma}}$ and $n_{\Nm_{{\hat\sigma}}^\bot}\in\Nm_{{\hat\sigma}}^\bot$. The component $n_{\Nm_{{\hat\sigma}}}$ is removed (Corollary~\ref{sk}), while $n_{\Nm_{{\hat\sigma}}^\bot}$ is transformed according to~\eqref{hu} and added to $\hu_{\hsi}$. } \end{remark}
We shall use the following definition.
\begin{definition}} \def\ED{\end{definition} Let ${\O}\subseteq\RR^{\m}$ be an open domain. We say that $\U:{\O}\rightarrow\RR^{\n}$ is a {\em local minimizer function} for the family of objectives $\mathfrak{F}_{\O}\stackrel{\rm def}{=}\{\mathcal{F}_d~:~d\in {\O}\}$ if for any $d\in {\O}$, the function $\mathcal{F}_d$ reaches a {\em strict} (local) minimum at $\U(d)$. \label{mf} \ED
Corollary~\ref{blg} shows that for any $d\in\RR^{\m}$, each strict (local) minimizer of $\mathcal{F}_d$ is entirely described by an $\omega\in\Om_{\max}$ via equation~\eqref{bla} in the same corollary. Consequently, a local minimizer function $\U$ is associated with every $\omega\in\Om_{\max}$.
\begin{lemma}} \def\EL{\end{lemma}\label{ct} For some arbitrarily fixed $\omega\in\Om_{\max}$ and $\beta>0$, the family of functions $\mathfrak{F}_{\RR^{\m}}$ has a {\em linear} (local) minimizer function $\U:\RR^{\m}\rightarrow\RR^{\n}$ that reads as \begin{equation}} \def\eeq{\end{equation} \forall\; d\in\RR^{\m},~~~\U(d)=Z_\omega\left(U_\omega \,d\right)~,~~ \mbox{where}~~
U_\omega=\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}A_{\vt}} \def\Pvt{\Pi_{\vt}^T ~\in\mathbb R^{\,\sharp\,\omega\times\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}}~. \label{bu}\eeq \EL
\par\paragraph{Proof} \ignorespaces The function $\U$ in~\eqref{bu} is linear in $d$. From~Corollary~\ref{blg}, for any $\beta>0$ and for any $d\in\RR^{\m}$, $\mathcal{F}_d$ has a strict (local) minimum at $\hat u=\U(d)$. Hence $\U$ fits Definition~\ref{mf}.
$\Box$\newline
{\em Thus, even if $\mathcal{F}_d$ has many strict (local) minimizers, each is linear in $d$.}
Next we exhibit a {\em closed negligible} subset of $\RR^{\m}$, associated with a nonempty $\omega\in\Om_{\max}$, whose elements are data $d$ leading to $\|\U(d)\|_0<\,\sharp\,\omega$.
\begin{lemma}} \def\EL{\end{lemma}\label{dom} For any $\omega\in\Om_{\max}$, define the subset $\Dm_\omega\subset\RR^{\m}$ by \begin{equation}} \def\eeq{\end{equation}\Dm_\omega\stackrel{\rm def}{=}\bigcup_{i=1}^{\,\sharp\,\omega} \Big\{~g\in\RR^{\m}~:~\left\<e_i\,,\,\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}A_{\vt}} \def\Pvt{\Pi_{\vt}^T \;g\right\>=0 ~\Big\}~.\label{can}\eeq Then $\Dm_\omega$ is closed in $\RR^{\m}$ and $\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\left(\Dm_\omega\right)=0$. \EL
\par\paragraph{Proof} \ignorespaces If $\omega=\varnothing$ then $\Dm_\omega=\varnothing$, hence the claim. Let $\,\sharp\,\omega\geq1$. For some $i\in\mathbb{I}_{\,\sharp\,\omega}$, set \begin{eqnarray*}} \def\eeqnn{\end{eqnarray*}\Dm&\stackrel{\rm def}{=}&\Big\{~g\in\RR^{\m}~:~\left\<e_i\,,\,\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}A_{\vt}} \def\Pvt{\Pi_{\vt}^T \;g\right\>=0 ~\Big\}\\&=& \Big\{~g\in\RR^{\m}~:~\Big\langle} \def\>{\rangleA_{\vt}} \def\Pvt{\Pi_{\vt}\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^T A_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}e_i,\;g\,\Big\>=0\Big\}~. \eeqnn Since $\mathrm{rank}\!\left(A_{\vt}} \def\Pvt{\Pi_{\vt}\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^T A_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}\right)\!=\!\,\sharp\,\omega$, $\mathrm{ker}} \def\ran{\mathrm{range}\!\left(A_{\vt}} \def\Pvt{\Pi_{\vt}\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^T A_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}\right)\!=\!\{0\}$. Hence $A_\omega\left(A_\omega^T A_\omega\right)^{-1}\!e_i\neq0$. Therefore $\Dm$ is a vector subspace of $\RR^{\m}$ of dimension $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1$ and so $\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\left(\Dm\right)=0$. The conclusion follows from the fact that $\Dm_\omega$ in~\eqref{can} is the union of $\,\sharp\,\omega$ subsets like $\Dm$ (see, e.g., \cite{Rudin76,Evans92}).
$\Box$\newline
\begin{proposition}} \def\EP{\end{proposition}\label{ctp} For some arbitrarily fixed $\omega\in\Om_{\max}$ and $\beta>0$, let $\,\U:\RR^{\m}\rightarrow\RR^{\n}$ be the relevant (local) minimizer function for $\mathfrak{F}_{\RR^{\m}}$ as given in~\eqref{bu} (Lemma~\ref{ct}). Let $\Dm_\omega$ read as in \eqref{can}. Then the function $d\mapsto\mathcal{F}_d\big(\U(d)\big)$ is $\C^\infty$ on $\RR^{\m}\,\mathbf{\setminus}\,\Dm_\omega$ and \begin{equation}} \def\eeq{\end{equation} d\in\RR^{\m}\,\mathbf{\setminus}\,\Dm_\omega~~~\Rightarrow~~~ \sigma\left(\U(d)\right)=\omega~,\nonumber\label{mp}\eeq where the set $\RR^{\m}\,\mathbf{\setminus}\,\Dm_\omega$ contains an open and dense subset of $\RR^{\m}$. \EP
\par\paragraph{Proof} \ignorespaces The statement about $\RR^{\m}\,\mathbf{\setminus}\,\Dm_\omega$ is a direct consequence of Lemma~\ref{dom}.
If $\omega=\varnothing$, then $\U(d)=0$ for all $d\in\RR^{\m}$, so all claims in the proposition are trivial. Consider that $\,\sharp\,\omega\geq1$. For any $i\in\mathbb{I}_{\,\sharp\,\omega}$, the $\omega[i]$th component of $\,\U(d)$ reads as (see Lemma~\ref{ct}) \[\U_{\omega[i]}(d)=\left\<e_i\,,\,\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}A_{\vt}} \def\Pvt{\Pi_{\vt}^T \;d\right\>~.\] The definition of $\Dm_\omega$ shows that \[d\in\RR^{\m}\,\mathbf{\setminus}\,\Dm_\omega~~\mbox{\rm and}~~i\in\mathbb{I}_{\,\sharp\,\omega}~~~\Rightarrow~~~ \U_{\omega[i]}(d)\neq0~,\] whereas $\U_i(d)=0$ for all $i\in\omega^c$. Consequently, \[\omega\in\Om_{\max}~~\mbox{\rm and}~~d\in\RR^{\m}\,\mathbf{\setminus}\,\Dm_\omega~~~\Rightarrow~~~\sigma\left(\U(d)\right)=\omega~.\]
Then $\|\U(d)\|_0$ is constant on $\RR^{\m}\,\mathbf{\setminus}\,\Dm_\omega$ and
\[\omega\in\Om_{\max}~~\mbox{\rm and}~~d\in\RR^{\m}\,\mathbf{\setminus}\,\Dm_\omega~~~\Rightarrow~~~\mathcal{F}_d(\U(d))=\big\|A\U(d)-d\big\|^2+\beta\,\sharp\,\omega~.\]
We infer from \eqref{bu} that $d\mapsto\big\|A\U(d)-d\big\|^2$ is $\C^\infty$ on $\RR^{\m}$, so $d\mapsto\mathcal{F}_d\big(\U(d)\big)$ is $\C^\infty$ on $\RR^{\m}\,\mathbf{\setminus}\,\Dm_\omega$.
$\Box$\newline
{\em A generic property is that a local minimizer function corresponding to $\mathcal{F}_d$ produces solutions sharing the same support.} The application $d\mapsto\mathcal{F}_d\big(\U(d)\big)$ is discontinuous on the {\em closed negligible} subset $\Dm_\omega$, where the support of $\U(d)$ is shrunk.
\subsection{Strict minimizers with an $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$-length support} \label{smm}
Here we explain why minimizers with an $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$-length support are useless in general.
\begin{proposition}} \def\EP{\end{proposition}\label{bad} Let $\mathrm{rank}(A)=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$, $\beta>0$ and for $d\in\RR^{\m}$ set
\[\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\stackrel{\rm def}{=}\big\{\hat u\in\RR^{\n}~:~ \hat u~~\mbox{is a strict (local) minimizer of ~$\mathcal{F}_d$~ meeting}~~\|\hat u\|_0=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}~\big\}~.\] Put \begin{equation}} \def\eeq{\end{equation}\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\stackrel{\rm def}{=}\bigcup_{\omega\in\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}}~\bigcup_{i\in\II_{\m}} \big\{g\in\RR^{\m}~:~\left\<e_i\,,\,A_{\vt}} \def\Pvt{\Pi_{\vt}^{-1}g\right\>=0\,\big\}~. \label{qm}\eeq Then $\RR^{\m}\,\mathbf{\setminus}\,\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ contains a dense open subset of $\RR^{\m}$ and \[d\in\RR^{\m}\,\mathbf{\setminus}\,\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\quad\Rightarrow\quad\,\sharp\,\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}=\,\sharp\,\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\quad\mbox{\rm and}\quad\mathcal{F}_d(\hat u)=\beta\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}~~\forall\;\hat u\in\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}~.\] \EP
\par\paragraph{Proof} \ignorespaces Using the notation in~\eqref{can}, $\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ reads as \[\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}=\bigcup_{\omega\in\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}}\Dm_\omega~.\] The claim on $\RR^{\m}\,\mathbf{\setminus}\,\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ follows from~Lemma~\ref{dom}. Since $\mathrm{rank}(A)=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$, we have $\,\sharp\,\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\geq1$.
Consider that $d\in\RR^{\m}\,\mathbf{\setminus}\,\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$. By Proposition~\ref{ctp} \[d\in\RR^{\m}\,\mathbf{\setminus}\,\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\quad\mbox{\rm and}\quad\omega\in\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\quad\Rightarrow\quad\mathcal{F}_d~ \mbox{has a strict (local) minimizer ~$\hat u$~ obeying}~\sigma(\hat u)=\omega~.\] Hence $\hat u\in\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$. Therefore, we have a mapping $b:\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\rightarrow\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ such that $\hat u=b(\omega)\in\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$. Using Lemma~\ref{ob} and Corollary~\ref{blg}, it reads as \[b(\omega)=Z_\omega(A_{\vt}} \def\Pvt{\Pi_{\vt}^{-1}d)~.\] For $(\omega,\varpi)\in\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ with $\varpi\neq\omega$ one obtains $\hat u=b(\omega)\in\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$, $\bar{u}=b(\varpi)\in\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ and $\hat u\neq\bar{u}$, hence $b$ is one-to-one. Conversely, for any $\hat u\in\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ there is $\omega\in\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ such that $\hat u=b(\omega)$ and $\sigma(\hat u)=\omega$ (because $d\not\in\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$). It follows that $b$ maps $\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ onto $\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$. Therefore, $\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ are $\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ in one-to-one correspondence, i.e. $\,\sharp\,\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}=\,\sharp\,\Um_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$.
Last, it is clear that $\omega\in\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ and $d\not\in\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ lead to
$\|A\hat u-d\|^2=0$ and $\mathcal{F}_d(\hat u)=\beta\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$.
$\Box$\newline
{\em For any $\beta>0$, a generic property of $\mathcal{F}_d$ is that it has $\,\sharp\,\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$
strict minimizers $\hat u$ obeying $\|\hat u\|_0=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ and $\mathcal{F}_d(\hat u)=\beta\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$. It is hard to discriminate between all these minimizers. Hence the interest in minimizers with supports located in $\Omega$, i.e., strict $(\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1)$-sparse minimizers of $\mathcal{F}_d$.}
\section{On the global minimizers of $\bm{\mathcal{F}_d}$}\label{sgm}
The next proposition gives a {\em necessary condition} for a global minimizer of $\mathcal{F}_d$. It follows directly from~\cite[Proposition 3.4]{Nikolova04a} where\footnote{ Just set $g_i=e_i$, $P=I_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ and $H=I_\n$ in \cite[Proposition 3.4]{Nikolova04a}.}
the regularization term is $\|Du\|_0$ for a full row rank matrix $D$. For $\mathcal{F}_d$ in~\eqref{fd} with $\|a_i\|_2=1$, $\forall\; i\in\II_{\n}$, a simpler condition was derived later in \cite[Theorem 12]{Tropp06}, using different tools. For completeness, the proof for a general $A$ is outlined in Appendix~\ref{pou}.
\begin{proposition}} \def\EP{\end{proposition}\label{ou} For $d\in\RR^{\m}$ and $\beta>0$, let $\mathcal{F}_d$ have a global minimum at $\hat u$. Then\footnote{Recall that
$a_i\neq0$ for all $i\in\II_{\n}$ by \eqref{ao} and that $\|\cdot\|=\|\cdot\|_2$.}
\begin{equation}} \def\eeq{\end{equation} i\in\sigma(\hat u)~~~\Rightarrow~~~\big|\,\hat u[i]\,\big|\geqslant \frac{\sqrt{\beta}}{\|a_i\|}~. \label{rt}\eeq \EP
Observe that the lower bound on
$\disp{\left\{\big|\,\hat u[i]\,\big|~:~i\in\sigma(\hat u)\right\}}$ given in \eqref{rt} is independent of~$d$. This means that in general, \eqref{rt} provides a pessimistic bound.
The proof of the statement shows that~\eqref{rt} is met also by all (local) minimizers of $\mathcal{F}_d$~satisfying $$\mathcal{F}_d(\hat u)\leqslant\mathcal{F}_d(\hat u+\rho e_i)~~~\forall\;\rho\in\mathbb R,~~\forall\; i\in\II_{\n}~.$$
\subsection{The global minimizers of $\bm{\mathcal{F}_d}$ are strict}\label{sGM}~~
\begin{remark} \label{nu} \rm Let $d\in\RR^{\m}$ and $\beta>\|d\|^2$. Then $\mathcal{F}_d$ has a strict global minimum at $\hat u=0$. {\rm Indeed,}
\[v\neq 0~~\Rightarrow~~\|v\|_0\geq1~~\Rightarrow~~
\mathcal{F}_d(0)=\|d\|^2<\beta\leqslant\|Av-d\|^2+\beta\|v\|_0~.\] \end{remark}
For least-squares regularized with a more regular $\phi$, one usually gets $\hat u=0$ asymptotically as $\beta\rightarrow+\infty$ but $\hat u\neq0$ for finite values of $\beta$. This does not hold for $\mathcal{F}_d$ by Remark~\ref{nu}.
Some theoretical results on the global minimizers of $\mathcal{F}_d$ have been obtained \cite{Nikolova04a,Haupt06,Tropp06,Davies08}. Surprisingly, the question about the {\em existence of global minimizers of $\mathcal{F}_d$} has never been raised. We answer this question using the notion of {\em asymptotically level stable functions} introduced by Auslender~\cite{Auslender00} in 2000.
As usual, \[\mathrm{lev}\,(\mathcal{F}_d,\lambda)\stackrel{\rm def}{=}\{v\in\RR^{\n}~:~\mathcal{F}_d(v)\leqslant\lambda\}\quad\mbox{\rm for}\quad \lambda>\inf \mathcal{F}_d~.\] The following definition is taken from \cite[p.~94]{Auslender03}.
\begin{definition}} \def\ED{\end{definition}\label{als} Let $\mathcal{F}_d:\RR^{\n}\rightarrow\mathbb R\cup\{+\infty\}$ be lower semicontinuous and proper. Then $\mathcal{F}_d$ is said to be {\em asymptotically level stable} if for each $\rho>0$, each bounded sequence $\{\lambda_k\}\in\mathbb R$ and each sequence $\{v_k\}\in\RR^{\n}$ satisfying
\begin{equation}} \def\eeq{\end{equation} v_k\in\mathrm{lev}\,(\mathcal{F}_d,\lambda_k),\quad\|v_k\|\rightarrow+\infty, \quad v_k\;\|v_k\|^{-1}\rightarrow \bar v\in\mathrm{ker}} \def\ran{\mathrm{range}\left((\mathcal{F}_d)_\infty\right)~, \label{aal}\eeq where $(\mathcal{F}_d)_\infty$ denotes the asymptotic (or recession) function of $\mathcal{F}_d$, there exists $k_0$ such that \begin{equation}} \def\eeq{\end{equation} v_k-\rho\bar v\in\mathrm{lev}\,(\mathcal{F}_d,\lambda_k)\quad\forall\; k\geqslant k_0~.\label{lsa}\eeq \ED
One can note that a coercive function is asymptotically level stable, since \eqref{aal} is empty. We prove that our discontinuous noncoercive objective $\mathcal{F}_d$ is asymptotically level stable as well.
\begin{proposition}} \def\EP{\end{proposition}\label{pal} Let $\mathcal{F}_d:\RR^{\n}\rightarrow\mathbb R$ be of the form~\eqref{fd}. Then $\mathrm{ker}} \def\ran{\mathrm{range}\left((\mathcal{F}_d)_\infty\right)=\mathrm{ker}} \def\ran{\mathrm{range}(A)$ and $\mathcal{F}_d$ is asymptotically level stable. \EP
The proof is outlined in Appendix~\ref{palp}.
\begin{theorem}} \def\ET{\end{theorem}\label{gs}Let $d\in\RR^{\m}$ and $\beta>0$. Then \begin{itemize}} \def\eit{\end{itemize} \item[\rm(i)] the set of the global minimizers of $\mathcal{F}_d$ \begin{equation}} \def\eeq{\end{equation}\disp{\hat U\stackrel{\rm def}{=} \left\{\hat u\in\RR^{\n}~:~\hat u=\min_{u\in\RR^{\n}}\mathcal{F}_d(u)\right\}}\label{ops}\eeq is nonempty; \item[\rm(ii)] every $\hat u\in \hat U$ is a {\em strict} minimizer of $\mathcal{F}_d$, i.e., \[\sigma(\hat u)\in\Om_{\max}~,\]
hence $\|\hat u\|_0\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$. \eit \ET
\par\paragraph{Proof} \ignorespaces For claim (i), we use the following statement\footnote{This result was originally exhibited in \cite{Baiocchi98} (without the notion of asymptotically level stable functions).}, whose proof can be found in the monograph by Ausleneder and Teboulle~\cite{Auslender03}:
\noindent{\cite[Corollary 3.4.2]{Auslender03}} {\it Let $\mathcal{F}_d:\RR^{\n}\rightarrow\mathbb R\cup\{+\infty\}$ be asymptotically level stable with $\inf\mathcal{F}_d>-\infty$. Then the optimal set $\hat U$---as given in \eqref{ops}---is nonempty .}
\noindent From Proposition~\ref{pal}, $\mathcal{F}_d$ is asymptotically level stable and $\inf\mathcal{F}_d\geq0$ from~\eqref{fd}. Hence $\hat U\neq\varnothing$.
(ii).~ Let $\hat u$ be a global minimizer of $\mathcal{F}_d$. Set ${\hat\sigma}=\sigma(\hat u)$.
If $\hat u=0$, (ii) follows from Lemma~\ref{uz}. Suppose that the global minimizer $\hat u\neq0$ of $\mathcal{F}_d$ is {\em nonstrict}. Then Theorem~\ref{ra}(ii) fails to hold; i.e., \begin{equation}} \def\eeq{\end{equation}\mathrm{dim}\,\mathrm{ker}} \def\ran{\mathrm{range}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)\geq1~. \label{sit}\eeq Choose $v_{{\hat\sigma}}\in \mathrm{ker}} \def\ran{\mathrm{range}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)\,\mathbf{\setminus}\,\{0\}$ and set $v=Z_{{\hat\sigma}}\left(v_{{\hat\sigma}}\right)$. Select an $i\in{\hat\sigma}$ obeying $v[i]\neq0$. Define $\widetilde u$ by \begin{equation}} \def\eeq{\end{equation} \widetilde u\stackrel{\rm def}{=} \hat u -\hat u [i]\frac{v}{v[i]}~.\eeq We have $\widetilde u[i]=0$ and $\hat u [i]\neq0~.$
Set $\widetilde\sigma\stackrel{\rm def}{=}\sigma\left(\widetilde u\right)$. Then \begin{equation}} \def\eeq{\end{equation}\widetilde\sigma\subsetneqq{\hat\sigma}~~~\mbox{hence}~~~\,\sharp\,\widetilde\sigma\leqslant\,\sharp\,{\hat\sigma}-1~.\eeq From $\disp{v_{{\hat\sigma}}\frac{\hat u [i]}{v[i]}\in\mathrm{ker}} \def\ran{\mathrm{range}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi} \right)}$, using~\eqref{sd} and Remark~\ref{ref} shows that\footnote{In detail we have $A\hat u =A_{\hsi}} \def\Phs{\Pi_{\hsi} \hu_{\hsi} =A_{\hsi}} \def\Phs{\Pi_{\hsi} \left(\hu_{\hsi} -v_{{\hat\sigma}}\frac{\hat u [i]}{v[i]}\right)=A_{\hsi}} \def\Phs{\Pi_{\hsi} \widetilde u_{{\hat\sigma}}= A_{\widetilde\sigma}\widetilde u_{\widetilde\sigma}=A\widetilde u~.$} $A\hat u =A_{\hsi}} \def\Phs{\Pi_{\hsi} \hu_{\hsi} =A_{\widetilde\sigma}\widetilde u_{\widetilde\sigma}=A\widetilde u$. Then
\begin{eqnarray}} \def\eeqn{\end{eqnarray}\mathcal{F}_d\left(\widetilde u\right)&=&\|A\widetilde u-d\|^2+\beta\,\sharp\,\widetilde\sigma\nonumber\\
&\leqslant&\mathcal{F}_d\left(\hat u \right)-\beta=\|A\hat u -d\|^2+\beta\left(\,\sharp\,{\hat\sigma}-1\right)~.\nonumber\label{tr}\eeqn It follows that $\hat u$ is not a global minimizer, hence \eqref{sit} is false. Therefore $\mathrm{rank}\left(A_{\hsi}} \def\Phs{\Pi_{\hsi}\right)=\,\sharp\,{\hat\sigma}$ and $\hat u$ is a strict minimizer of $\mathcal{F}_d$ (Theorem~\ref{ra}).
$\Box$\newline
One can note that if $\mathrm{rank}(A)=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$, any global minimizer $\hat u$ of $\mathcal{F}_d$ obeys $\,\mathcal{F}_d(\hat u)\leqslant\beta\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}~.$
According to Theorem~\ref{gs}, {\em the global minimizers of $\mathcal{F}_d$ are strict and their number is finite: this is a nice property that fails for many convex nonsmooth optimization problems.}
\subsection{$\bm\k$-sparse global minimizers for $\bm{\k\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}$}
In order to simplify the presentation, in what follows we consider that \[\mbox{\framebox{~$\mathrm{rank}(A)=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}<\n~.$}}\]
Since $\mathcal{F}_d$ has a large number (typically equal to $\,\sharp\,\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$)
of strict minimizers with $\|\hat u\|_0=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ yielding the same value $\mathcal{F}_d(\hat u)=\beta\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ (see Proposition~\ref{bad} and the comments given after its proof), it is important to be sure that the global minimizers of $\mathcal{F}_d$ are $(\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1)$-sparse.
We introduce a notation which is used in the rest of this paper. For any $\k\in\II_{\m-1}$, put \begin{equation}} \def\eeq{\end{equation} \mbox{\framebox{$\disp{~~\overline{\Omega}_\k\stackrel{\rm def}{=}\bigcup_{r=0}^\k\Omega_r~}$~}}\label{ga}\eeq where $\Omega_r$ was set up in~Definition~\ref{om}. Theorem~\ref{ra} gives a clear meaning of the sets\footnote{Clearly, $\overline{\Omega}_{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}=\Omega~.$}~$\overline{\Omega}_\k$. {\em For any $d\in\RR^{\m}$ and any $\beta>0$, for any fixed $\k\in\II_{\m-1}$, the set $\overline{\Omega}_\k$ is composed of the supports of all possible $\k$-sparse strict (local) minimizers of $\mathcal{F}_d$.}
The next propositions checks the existence of $\beta>0$ ensuring that all the global minimizers of $\mathcal{F}_d$ are $\k$-sparse, for some $\k\in\II_{\m-1}$.
\begin{proposition}} \def\EP{\end{proposition}\label{ie} Let $d\in\RR^{\m}$.
For any $\k\in\II_{\m-1}$, there exists $\beta_\k\geq0$ such that if $\beta>\beta_\k$, then each global minimizer $\hat u$ of $\mathcal{F}_d$ satisfies
\begin{equation}} \def\eeq{\end{equation}\|\hat u\|_0\leqslant\k~~~\mbox{\rm and}~~~\sigma(\hat u)\in\overline{\Omega}_\k~ .\label{spo}\eeq One can choose $\beta_\k=\|A\widetilde u-d\|^2$ where $\widetilde u$ solves {\rm{($\,\P_{\vt}\,$)}~} for some $\omega\in\Omega_{\k}$. \EP
The proof is given in Appendix~\ref{pie}. The value of $\beta_\k$ in the statement is easy to compute, but in general it is not sharp\footnote{ For $\beta\gtrapprox\beta_\k$ the global minimizers of $\mathcal{F}_d$ might be $k$-sparse for $k\ll\k$. A sharper $\beta_\k$ can be obtained~as
$\beta_\k=\min_{\omega\in\Omega_\k}\left\{\|A\widetilde u-d\|^2~:~\widetilde u~~\mbox{solve {\rm{($\,\P_{\vt}\,$)}~} for}~~ \omega\in\Omega_\k\right\}~.$ }.
\section{Uniqueness of the global minimizer of $\bm{\mathcal{F}_d}$}\label{kb}
The presentation is simplified using the notation introduced next. Given a matrix $A\in\RR^{\m\x\n}$, with any $\omega\in\Omega$ (see Definition~\ref{om}), we associate the $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ matrix $\Pvt$ that yields the orthogonal projection\footnote{ If $\omega=\varnothing$, we have $A_{\vt}} \def\Pvt{\Pi_{\vt}\in\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\x0}$ and so $\Pi_\omega$ is an $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ null matrix.} onto the subspace spanned by the columns of $A_{\vt}} \def\Pvt{\Pi_{\vt}$: \begin{equation}} \def\eeq{\end{equation}\mbox{\framebox{$\disp{~\Pvt\stackrel{\rm def}{=}A_{\vt}} \def\Pvt{\Pi_{\vt}\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}A_{\vt}} \def\Pvt{\Pi_{\vt}^T~.}$}}\label{pro}\eeq \noindent For $\omega\in\Omega$, the projector in Remark~\ref{ref} reads $\Pi_{\ran(A_{\vt}} \def\Pvt{\Pi_{\vt})}= \Pvt$.
Checking whether a global minimizer $\hat u$ of $\mathcal{F}_d$ is unique requires us to compare its value $\mathcal{F}_d(\hat u)$ with the values $\mathcal{F}_d(\bar{u})$ of the concurrent strict minimizers~$\bar{u}$. Let $\hat u$ be an $(\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1)$-sparse strict (local) minimizer of $\mathcal{F}_d$. Then ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)\in\Omega$. Using~Remark~\ref{ref} shows that
\begin{eqnarray}} \def\eeqn{\end{eqnarray} \mathcal{F}_d(\hat u)&=&\|A_{\hsi}} \def\Phs{\Pi_{\hsi}\hu_{\hsi}-d\|^2+\beta\,\sharp\,{\hat\sigma}
=\|\Phs d-d\|^2+\beta\,\sharp\,{\hat\sigma}\nonumber\\&=&d^T\left(I_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-\Phs\right)d+\beta\,\sharp\,{\hat\sigma}~.\label{sx}\eeqn Let $\bar{u}$ be another $(\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1)$-sparse strict minimizer of $\mathcal{F}_d$; set $\bar{\sigma}\stackrel{\rm def}{=}\sigma(\bar{u})$. Then \[\mathcal{F}_d(\hat u)-\mathcal{F}_d(\bar{u})=d^T\left(\Pbs-\Phs\right)d+\beta(\,\sharp\,{\hat\sigma}-\,\sharp\,\bar{\sigma})~.\] If both $\hat u$ and $\bar{u}\neq\hat u$ are global minimizers of $\mathcal{F}_d$, the previous equality yields \begin{equation}} \def\eeq{\end{equation} \mathcal{F}_d(\hat u)=\mathcal{F}_d(\bar{u})~~~\Leftrightarrow~~~ d^T\left(\Pbs-\Phs\right)d=-\beta(\,\sharp\,{\hat\sigma}-\,\sharp\,\bar{\sigma})~.\label{hq}\eeq {\em Equation~\eqref{hq} reveals that the uniqueness of the global minimizer of $\mathcal{F}_d$ cannot be guaranteed without suitable assumptions on $A$ and on~$d$.}
\subsection{A generic assumption on $\bm{A}$}\label{ru}
We adopt an assumption on the matrix $A$ in $\mathcal{F}_d$ in order to restrict the cases when~\eqref{hq} takes place for some supports ${\hat\sigma}\neq\bar{\sigma}$ obeying $\,\sharp\,{\hat\sigma}=\,\sharp\,\bar{\sigma}$. \begin{hyp}} \def\EH{\end{hyp}\label{aa} The matrix $A\in\RR^{\m\x\n}$, where $\mathsf{M}<\mathsf{N}$, is such that for some given $\k\in\II_{\m-1}$, \begin{equation}} \def\eeq{\end{equation} \mbox{\framebox{$ ~ r\in\mathbb{I}_{\mathsf{K}}~~\mbox{\rm and}~~ (\omega,\varpi)\in(\Omega_r\times \Omega_r),~~\omega\neq\varpi~~~\Rightarrow~~~\Pi_{\omega}\neq \Pi_{\varpi}~.$}}\eeq \EH
Assumption H\ref{aa} means that the angle (or the gap) between the equidimensional subspaces $\ran\left(A_{\vt}} \def\Pvt{\Pi_{\vt}\right)$ and $\ran\left(A_{\vs}} \def\Pvs{\Pi_{\vs}\right)$ must be nonzero~\cite{Meyer00}. For instance, if $(i,j)\in\II_{\n}\times\II_{\n}$ satisfy $i\neq j$, H\ref{aa} implies that $a_i\neq\kappa\,a_j$ for any $\kappa\in\mathbb R\,\mathbf{\setminus}\,\{0\}$
since $\Pi_{\{i\}}=a_ia_i^T/\|a_i\|^2~.$
Checking whether H\ref{aa} holds for a given matrix $A$ requires a combinatorial search over all possible couples $(\varpi,\omega)\in(\Omega_r\times \Omega_r)$ satisfying $\varpi\neq\omega$, $\forall\; r\in\II_{\k}$. This is hard to do. Instead, we wish to know whether or not H\ref{aa} is a {\em practical} limitation. Using some auxiliary claims, we shall show that H\ref{aa} {\em fails only for a closed negligible subset of matrices} in the space of all $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\n$ real matrices.
\begin{lemma}} \def\EL{\end{lemma}\label{bb} Given $r\in\II_{\m-1}$ and $\varpi\in\Omega_r$, define the following set of submatrices of $A$: \[\H_\varpi=\Big\{A_{\vt}} \def\Pvt{\Pi_{\vt}~:~\omega\in\Omega_r~~\mbox{\rm and}~~\Pvt=\Pvs\Big\}~.\] Then $\H_\varpi$ belongs to an $(r\times r)$-dimensional subspace of the space of all $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times r$ matrices. \EL
\par\paragraph{Proof} \ignorespaces Using the fact that $\varpi\in\Omega_r$ and $\omega\in\Omega_r$, we have\footnote{Using \eqref{pro}, as well as the fact that $A_{\vt}} \def\Pvt{\Pi_{\vt}=\PvtA_{\vt}} \def\Pvt{\Pi_{\vt}$, $\forall\;\omega\in\Omega_r$, one easily derives \eqref{fo} since \[\Pvt=\Pvs~~\Leftrightarrow\left\{\begin{array}} \def\earr{\end{array}{l} A_{\vt}} \def\Pvt{\Pi_{\vt}\left(A_{\vt}} \def\Pvt{\Pi_{\vt}^TA_{\vt}} \def\Pvt{\Pi_{\vt}\right)^{-1}A_{\vt}} \def\Pvt{\Pi_{\vt}^T=\Pvs\\A_{\vs}} \def\Pvs{\Pi_{\vs}\left(A_{\vs}} \def\Pvs{\Pi_{\vs}^TA_{\vs}} \def\Pvs{\Pi_{\vs}\right)^{-1}A_{\vs}} \def\Pvs{\Pi_{\vs}^T=\Pvt\earr\right.\Rightarrow~ \left\{\begin{array}} \def\earr{\end{array}{l}A_{\vt}} \def\Pvt{\Pi_{\vt}=\PvsA_{\vt}} \def\Pvt{\Pi_{\vt}\\A_{\vs}} \def\Pvs{\Pi_{\vs}=\PvtA_{\vs}} \def\Pvs{\Pi_{\vs}\earr\right.\Rightarrow~ \left\{\begin{array}} \def\earr{\end{array}{l}\Pvt=\Pvs\Pvt\\\Pvs=\Pvt\Pvs\earr\right.\Rightarrow~\Pvt=\Pvs.\]} \begin{equation}} \def\eeq{\end{equation}\Pvt=\Pvs~~~\Leftrightarrow~~~A_{\vt}} \def\Pvt{\Pi_{\vt}=\PvsA_{\vt}} \def\Pvt{\Pi_{\vt}~~.\label{fo}\eeq Therefore $\H_\varpi$ equivalently reads \begin{equation}} \def\eeq{\end{equation}\H_\varpi=\Big\{A_{\vt}} \def\Pvt{\Pi_{\vt}~:~\omega\in\Omega_r~~\mbox{\rm and}~~A_{\vt}} \def\Pvt{\Pi_{\vt}=\PvsA_{\vt}} \def\Pvt{\Pi_{\vt}\Big\}~.\label{gr}\eeq Let $A_{\vt}} \def\Pvt{\Pi_{\vt}\in\H_\varpi$. Denote the columns of $A_{\vt}} \def\Pvt{\Pi_{\vt}$ by $\tilde a_i$ for $i\in\II_r$. Then \eqref{gr} yields \[ \Pvs\tilde a_i=\tilde a_i,~~\forall\; i\in\II_r\quad\Rightarrow\quad\tilde a_i\in\ran(A_{\vs}} \def\Pvs{\Pi_{\vs}), ~~\forall\; i\in\II_r~.\] Hence all $\tilde a_i$, $i\in\II_r$, live in the $r$-dimensional vector subspace $\ran(A_{\vs}} \def\Pvs{\Pi_{\vs})$. All the columns of each matrix $A_{\vt}} \def\Pvt{\Pi_{\vt}\in\H_\varpi$ belong to $\ran(A_{\vs}} \def\Pvs{\Pi_{\vs})$ as well. It follows that $\H_\varpi$ belongs to a (closed) subspace of dimension $r\times r$ in the space of all $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times r$ matrices, where $r\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1$.
$\Box$\newline
More details on the submatrices of $A$ living in $\H_\varpi$ are given next. \begin{remark} \label{P} {\rm The {\em closed negligible} subset $\H_\varpi$ in~Lemma~\ref{bb} is formed from all the submatrices of $A$ that are column equivalent to $A_{\vs}} \def\Pvs{\Pi_{\vs}$ (see \cite[p.~171]{Meyer00}), that is, \begin{equation}} \def\eeq{\end{equation}A_{\vt}} \def\Pvt{\Pi_{\vt}\in\H_\varpi~~~\Leftrightarrow~~~\exists\,P\in\mathbb R^{r\times r}~~\mbox{such that}~~ \mathrm{rank}(P)=r~~\mbox{\rm and}~~A_{\vt}} \def\Pvt{\Pi_{\vt}=A_{\vs}} \def\Pvs{\Pi_{\vs} P~.\label{yr} \eeq Observe that $P$ has $r^2$ unknowns that must satisfy $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K} r$ equations and that $P$ must be invertible. It should be quite unlikely that such a matrix $P$ does exist.} \end{remark}
This remark can help to discern whether or not structured dictionaries satisfy H\ref{aa}.
Next we inspect the set of all matrices $A$ failing assumption H\ref{aa}. \begin{lemma}} \def\EL{\end{lemma}\label{ug} Consider the set $\H$ formed from $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\n$ real matrices described next: \[\H\stackrel{\rm def}{=}\left\{A\in\RR^{\m\x\n}~:\exists\; r\in\II_{\m-1},~ \exists\;(\varpi,\omega)\in\Omega_r\times\Omega_r,~\varpi\neq\omega~~\mbox{\rm and}~~\Pvs=\Pvt\right\}~.\] Then $\H$ belongs to a finite union of vector subspaces in $\RR^{\m\x\n}$ whose Lebesgue measure in $\mathbb{R}^{\mathsf{M}\times\mathsf{N}}$ is null. \EL
\par\paragraph{Proof} \ignorespaces Let $A\in\H$. Then there exist at least one integer $r\in\II_{\m-1}$ and at least one pair $(\varpi,\omega)\in\Omega_r\times\Omega_r$ such that $\varpi\neq\omega$ and $\Pvs=\Pvt$. Using Lemma~\ref{bb}, $A$ contains (at least) one $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times r$ submatrix $A_{\vs}} \def\Pvs{\Pi_{\vs}$ belonging to an $r\times r$ vector subspace in the space of all $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times r$ real matrices. Identifying $A$ with an $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\n$-length vector, its entries are included in a vector subspace of $\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\n}$ of dimension no larger than $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\n-1$. The claim of the lemma is straightforward.
$\Box$\newline
We can now clarify assumption H\ref{aa} and show that it is really good.
\begin{theorem}} \def\ET{\end{theorem}\label{usg} Given an arbitrary $\k\in\II_{\m-1}$, consider the set of $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\n$ real matrices below \[\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_\k\stackrel{\rm def}{=}\Big\{A\in\RR^{\m\x\n}~:~~A~~\mbox{\rm satisfies~~ H\ref{aa}~~for~~}\k~\Big\}~.\] Then $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_\k$ contains an open and dense subset in the space of all $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\n$ real-valued matrices. \ET
\par\paragraph{Proof} \ignorespaces The complement of $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_\k$ in the space of all $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\n$ real matrices reads as \[\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_\k^c=\Big\{A\in\RR^{\m\x\n}~:~\mbox{\rm H\ref{aa} fails for }~A~~\mbox{\rm and}~~\k~\Big\}~.\] It is clear that $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_\k^c\subset\H~,$ where $\H$ is described in Lemma~\ref{ug}. By the same lemma, $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_\k^c$ is included in a closed subset of vector subspaces in $\RR^{\m\x\n}$ whose Lebesgue measure in $\mathbb{R}^{\mathsf{M}\times\mathsf{N}}$ is null. Consequently, $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_\k$ satisfies the statement of the theorem.
$\Box$\newline
{\em For any $\k\in\II_{\m-1}$,
H\ref{aa} is a {\em generic} property of all $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\times\n$ real matrices meeting $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}<\n$.}
This is the meaning of Theorem~\ref{usg} in terms of Definition~\ref{ps}.
We can note that \[\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_{\k+1}\subseteq\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_\k,~~~\forall\;\k\in\mathbb{I}_{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-2}~.\] One can hence presume that H\ref{aa} is weakened as $\k$ decreases. This issue is illustrated in section~\ref{nm}.
\subsection{A generic assumption on $\bm{d}$}\label{rs}
A preliminary result is stated next. \begin{lemma}} \def\EL{\end{lemma}\label{H} Let $(\omega,\varpi)\in\overline{\Omega}_\k\times \overline{\Omega}_\k$ for $\omega\neq\varpi$ and let H\ref{aa} hold for $\k\in\II_{\m-1}$. Given $\kappa\in\mathbb R$, define \begin{equation}} \def\eeq{\end{equation}\Tm_\kappa\stackrel{\rm def}{=}\{g\in\RR^{\m}~:~g^T\left(\Pvt-\Pvs\right)g=\kappa\}~.\nonumber\label{setS}\eeq Then $\Tm_\kappa$ is a closed subset of $\RR^{\m}$ and $\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\left(\Tm_\kappa\right)=0$. \EL
\par\paragraph{Proof} \ignorespaces Define $f:\RR^{\m}\rightarrow\mathbb R$ by $f(g)=g^T\left(\Pvt-\Pvs\right)g~.$ Then \begin{equation}} \def\eeq{\end{equation}\Tm_\kappa=\{g\in\RR^{\m}~:~f(g)=\kappa\}~.\label{tyk}\eeq Using H\ref{aa}, $\Tm_\kappa$ is closed in $\RR^{\m}$. Set \[\Qm=\{g\in\RR^{\m}~:~\nabla f(g)\neq0\}\quad\mbox{and}\quad\Qm^c=\RR^{\m}\,\mathbf{\setminus}\,\Qm.\]
Consider an arbitrary $g\in\Tm_\kappa\cap\Qm$. From H\ref{aa}, $\mathrm{rank} (\nabla f(g))=1$. For simplicity, assume that \[\nabla f(g)[\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}]=\frac{df(g)}{dg[\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}]}\neq0~.\label{asm}\] By the implicit functions theorem, there are open neighborhoods $\O_g\subset\Qm\subset\RR^{\m}$ and $\V\subset\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}$ of $g$ and $g_{\II_{\m-1}}$, respectively,
and a unique $\C^1$-function $h_g:\V\rightarrow \mathbb R$ with $\nabla h_g$ bounded, such that \begin{equation}} \def\eeq{\end{equation} \gamma=(\gamma_{\II_{\m-1}},\gamma[\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}])\in\O_g~~\mbox{and}~~f(\gamma)=\kappa\quad\Leftrightarrow\quad \gamma_{\II_{\m-1}}\in\V ~~\mbox{and}~~\gamma[\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}]=h_g(\gamma_{\II_{\m-1}})~.\label{imp} \eeq From~\eqref{tyk} and \eqref{imp} it follows that\footnote{From~\eqref{imp}, $\V$ is the restriction of $\O_g$ to $\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}$.} \[\O_g\cap\Tm_k=\psi^g\left(\O_g\cap(\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}\times\{0\})\right),~\] where $\psi^g$ is a diffeomorphism on $\O_g $ given by \[\psi^g_i(\gamma)=\gamma[i],\quad 1\leqslant i\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1\quad\mbox{and}\quad\psi^g_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}(\gamma)=h_g(\gamma_{\II_{\m-1}})+\gamma[\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}]~.\label{nap}\] Since $\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\left(\O_g\cap(\mathbb R^{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}\times\{0\})\right)=0$ and $\nabla\psi^g$ is bounded on $\O_g$, it follows from \cite[Lemma 7.25]{Rudin87} that\footnote{The same result follows from the change-of-variables theorem for the Lebesgue integral, see e.g. \cite{Rudin87}.}
$\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\left(\V_g\cap\Tm_k\right)=0$. We have thus obtained that \begin{equation}} \def\eeq{\end{equation} S\subset Q~~\mbox{\rm and}~~S~\mbox{bounded}\quad\Rightarrow\quad\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}(S\cap\Tm_k)=0~.\label{sq} \eeq Using that every open subset of $\RR^{\m}$ can be written as a countable union\footnote{From \eqref{sq}, adjacent cubes can also intersect in our case.}
of cubes in $\RR^{\m}$ \cite{Rudin76,Evans92,Stein05}, the result in \eqref{sq} entails that $\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}(\Tm_\kappa\cap Q)=0$.
Next, $Q^c\!=\mathrm{ker}} \def\ran{\mathrm{range}\left(\Pvt-\Pvs\right)$. By H\ref{aa}, $\mathrm{dim}\,\mathrm{ker}} \def\ran{\mathrm{range}\left(\Pvt\!-\Pvs\right)\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1$. Hence $\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}(\Tm_\kappa\cap Q^c)=0$.
The proof follows from the equality $\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}(\Tm_\kappa)=\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}(\Tm_\kappa\cap Q)+\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}(\Tm_\kappa\cap Q^c)$.
$\Box$\newline
We exhibit a {\em closed negligible} subset of data in $\RR^{\m}$ that can still meet the equality in \eqref{hq}.
\begin{proposition}} \def\EP{\end{proposition}\label{cor} For $\beta>0$ and $\k\in\II_{\m-1}$, put \begin{equation}} \def\eeq{\end{equation}\Sigma_\k\stackrel{\rm def}{=}\bigcup_{n=-\k}^{\k}~\bigcup_{\omega\in\overline{\Omega}_\k}~ \bigcup_{\varpi\in\overline{\Omega}_\k}\left\{g\in\RR^{\m}~:~ \omega\neq\varpi~~\mbox{\rm and}~~g^T\Big(\Pvt-\Pvs\Big)g=n\beta\right\}~,\label{S} \eeq where $\overline{\Omega}_\k$ is given in~\eqref{ga}. Let H\ref{aa} hold for $\k$. Then $\Sigma_\k$ is closed in $\RR^{\m}$ and $\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\left(\Sigma_\k\right)=0$. \EP
\par\paragraph{Proof} \ignorespaces For some $n\in\{-\k,\cdots,\k\}$ and $(\omega,\varpi)\in (\overline{\Omega}_\k\times\overline{\Omega}_\k)$ such that $\omega\neq\varpi$, put \[\Sigma\stackrel{\rm def}{=}\left\{g\in\RR^{\m}:g^T\Big(\Pvt-\Pvs\Big)g=n\beta\right\}~. \] If $\,\sharp\,\omega\neq\,\sharp\,\varpi$, then $\mathrm{rank}\big(\Pvt-\Pvs\big)\geq1$. If $\,\sharp\,\omega=\,\sharp\,\varpi$, H\ref{aa} guarantees that $\mathrm{rank}\big(\Pvt-\Pvs\big)\geq1$, yet again. The number $n\beta\in\mathbb R$ is given. According to Lemma~\ref{H}, $\Sigma$ is a closed subset of $\RR^{\m}$ and $\mathbb{L}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\left(\Sigma\right)=0$. The conclusion follows from the fact that $\Sigma_\k$ is a finite union of subsets like~$\Sigma$.
$\Box$\newline
{\em We assume hereafter that if H\ref{aa} holds for some $\k\in\II_{\m-1}$, data $d$ satisfy \[d\in \{g\in\RR^{\m}~:~g\not\in\Sigma_\k\}=\RR^{\m}\,\mathbf{\setminus}\,\Sigma_\k~.\]}
\subsection{The unique global minimizer of $\bm{\mathcal{F}_d}$ is $\bm\k$-sparse for $\bm{\k\leqslant(\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1)}$}\label{kr}
We are looking for guarantees that $\mathcal{F}_d$ has a {\em unique} global minimizer $\hat u$ obeying
\[ \|\hat u\|_0\leqslant\k~~\mbox{for some fixed}~~ \k\in\II_{\m-1}~.\] This is the aim of the next theorem. \begin{theorem}} \def\ET{\end{theorem} \label{Uglob} Given $\k\in\II_{\m-1}$, let H\ref{aa} hold for $\k$, $\beta>\beta_\k$ where $\beta_\k$ meets Proposition~\ref{ie} and $\Sigma_\k\subset\RR^{\m}$ reads as in \eqref{S}. Consider that \[d\in\RR^{\m}\,\mathbf{\setminus}\,\Sigma_\k~.\] Then \begin{itemize}} \def\eit{\end{itemize}\item[\rm(i)] the set $\RR^{\m}\,\mathbf{\setminus}\,\Sigma_\k$ is open and dense in $\RR^{\m}$;
\item[\rm(ii)] $\mathcal{F}}\def\G{\mathcal{G}}\def\H{{\mathcal H}}\def\I{{\mathcal I}}\def\J{{\mathcal J}}\def\K{\mathcal{K}_d$ has a {\em unique global minimizer} $\hat u$, and $\|\hat u\|_0\leqslant\k$. \eit \ET
\par\paragraph{Proof} \ignorespaces Statement (i) follows from Proposition~\ref{cor}.
Since $\beta>\beta_\k$, all global minimizers of $\mathcal{F}_d$ have their support in $\overline{\Omega}_\k$ (Proposition~\ref{ie}). Using the fact that $d\in\RR^{\m}\,\mathbf{\setminus}\,\Sigma_\k~$, the definition of $\Sigma_\k$ in \eqref{S} shows that \begin{equation}} \def\eeq{\end{equation}-\k\leqslant n\leqslant\k~~\mbox{\rm and}~~(\omega,\varpi)\in (\overline{\Omega}_\k\times\overline{\Omega}_\k),~\omega\neq\varpi~~~\Rightarrow~~~ d^T\Big(\Pvt-\Pvs\Big)d\neq n\beta~.\label{gc}\eeq The proof is conducted by contradiction. Let $\hat u$ and $\bar{u}\neq\hat u$ be two global minimizers of $\mathcal{F}_d$. Then \[{\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)\in\overline{\Omega}_\k~~~\mbox{\rm and}~~~\bar{\sigma}\stackrel{\rm def}{=}\sigma(\bar{u})\in\overline{\Omega}_\k~,\] and ${\hat\sigma}~\neq~\bar{\sigma}$. By $\mathcal{F}}\def\G{\mathcal{G}}\def\H{{\mathcal H}}\def\I{{\mathcal I}}\def\J{{\mathcal J}}\def\K{\mathcal{K}_d(\hat u)=\mathcal{F}}\def\G{\mathcal{G}}\def\H{{\mathcal H}}\def\I{{\mathcal I}}\def\J{{\mathcal J}}\def\K{\mathcal{K}_d(\bar{u})$,~\eqref{hq} yields \begin{equation}} \def\eeq{\end{equation} d^T\left(\Phs-\Pbs\right)d=\beta(\,\sharp\,{\hat\sigma}-\,\sharp\,\bar{\sigma})~.\label{qw}\eeq An enumeration of all possible values of $\,\sharp\,{\hat\sigma}-\,\sharp\,\bar{\sigma}$ shows that \[\beta\left(\,\sharp\,{\hat\sigma}-\,\sharp\,\bar{\sigma}\right)=n\beta~~~\mbox{for some}~~~n\in\{-\k,\cdots,\k\}~.\] Inserting this equation into~\eqref{qw} leads to \[d^T\left(\Phs-\Pbs\right)d=n\beta~~~\mbox{for some}~~~n\in\{-\k,\cdots,\k\}~.\] The last result contradicts~\eqref{gc}; hence it violates the assumptions H\ref{aa} and $d\in\RR^{\m}\,\mathbf{\setminus}\,\Sigma_\k$. Consequently, $\mathcal{F}_d$ cannot have two global minimizers. Since $\mathcal{F}_d$ always has global minimizers (Theorem~\ref{gs}(i)), it follows that
$\mathcal{F}_d$ has a unique global minimizer, say $\hat u$. And $\|\hat u\|_0\leqslant \mathsf{K}$ because $\sigma(\hat u)\in\overline{\Omega}_\mathsf{K}$.
$\Box$\newline
{\em For $\beta>\beta_\k$, the objective $\mathcal{F}_d$ in~\eqref{fd} has a unique global minimizer and it is $\k$-sparse for $\k\leqslant\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1$. For all $\;\k\in\II_{\m-1}$, the claim holds true in a generic sense.} This is the message of Theorem~\ref{Uglob} using Definition~\ref{ps}.
\section{Numerical illustrations}\label{nm} \subsection{On assumption H\ref{aa}} Assumption H\ref{aa} requires that $\Pvt\neq\Pvs$ when $(\omega,\varpi)\in\Omega_r\times\Omega_r$, $\omega\neq\varpi$ for all $r\leqslant\k\in\II_{\m-1}$. From a practical viewpoint, the magnitude of $\left(\Pvt-\Pvs\right)$ should be discernible. One way to assess the viability of H\ref{aa} for a matrix $A$ and $\k\in\II_{\m-1}$ is to calculate \begin{eqnarray}} \def\eeqn{\end{eqnarray}\xi_\k(A)&\stackrel{\rm def}{=}&\min_{r\in\II_{\k}}~~\mu_r(A)~,\label{xi}\\ \mbox{where}~~~~~
\mu_r(A)&=&\disp{\min_{\small\begin{array}} \def\earr{\end{array}{c}(\omega,\varpi)\in\Omega_r\times\Omega_r\\\omega\neq\varpi\earr}
\left\|\Pvt-\Pvs\right\|_2,~~~\forall\; r\in\II_{\k}~.}\nonumber\eeqn In fact, $\left\|\Pvt-\Pvs\right\|_2=\sin(\theta)$, where $\theta\in[0,\pi/2]$ is the maximum angle between $\ran\left(A_{\vt}} \def\Pvt{\Pi_{\vt}\right)$ and $\ran\left(A_{\vs}} \def\Pvs{\Pi_{\vs}\right)$; see \cite[p. 456]{Meyer00}. These subspaces have the same dimension and $\Pvt\neq\Pvs$ when $(\omega,\varpi)\in\Omega_r\times\Omega_r$, $\omega\neq\varpi$ and $r\in\II_{\k}$, hence $\theta\in(0,\pi/2]$. Consequently, \[\mbox{H\ref{aa} holds for}~~\k\in\II_{\m-1}~~~\Rightarrow~~~\mu_r(A)\in(0,1]~~\forall\; r\in\II_{\k}~~~\Rightarrow~~~\xi_\k(A)\in(0,1]~.\] According to \eqref{xi}, we have $\xi_\k\geqslant\xi_{\k+1}$, $\forall\; \k\in\mathbb{I}_{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-2}$. Our guess that {\em assumption~H\ref{aa} is lightened when $\k$ decreases} (see the comments following the proof of Theorem~\ref{usg}) means that \begin{equation}} \def\eeq{\end{equation}\xi_1(A)>\cdots>\xi_{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}(A)~.\label{tya}\eeq
We provide numerical tests on two subsets of real-valued random matrices for $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}=5$ and $\n=10$, denoted by $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^N_{20}$ and $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^U_{1000}$. The values of $\xi_\k(\cdot)$, $\k\in\II_{\m-1}=\mathbb{I}_4$, for every matrix in $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^N_{20}$ and in $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^U_{1000}$, were calculated using an {\em exhaustive combinatorial search}. {\em All tested matrices satisfy assumption H\ref{aa}}, which confirms Theorem~\ref{usg} and its consequences. In order to evaluate the extent of H\ref{aa}, we computed the {\em worst} and the {\em best} values of $\xi_\k(\cdot)$ over these sets: \begin{equation}} \def\eeq{\end{equation}\left\{\begin{array}} \def\earr{\end{array}{rcl}\xi_\k^{\mbox{\footnotesize\it worst}}&=&\disp{\min_{A\in\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}}\xi_\k(A)}~\\~\\ \xi_\k^{\mbox{\footnotesize\it best}}&=&\disp{\max_{A\in\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}}\xi_\k(A)}~\earr\right.~~~~~~\forall\;\k\in\II_{\m-1}~,~~~ \mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}\in\left\{\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^N_{20},~\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^U_{1000}\right\}~.\label{be}\eeq
\paragraph{Set $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^N_{20}$} This set was formed from 20 matrices $A^n$, ~$n\in\mathbb{I}_{20}$ of size $5\times 10$. The components of each matrix $A^n$ were independent and uniformly drawn
from the standard normal distribution with mean zero and variance one. The values of $\xi_\k(\cdot)$ are depicted in Fig.~\ref{TH}. We have\footnote{ This is why on the figure, in columns 10 and 17, the green ``$\circ$'' and the red ``$\lozenge$'' overlap.} $\xi_1(A^{10})=\xi_2(A^{10})$ and $\xi_1(A^{17})=\xi_2(A^{17})$. In all other cases \eqref{tya} is satisfied. Fig.~\ref{TH} clearly shows that $\xi_\k(\cdot)$ increases as $\k$ decreases (from $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1$ to $1$).
\setlength{\unitlength}{1mm} [h!] \begin{center}} \def\ec{\end{center} \begin{picture}} \def\epic{\end{picture}(143,81) \put(0,1){\epsfig{figure=testH_20_N.eps,width=10.0cm,height=8cm}} \put(20,-1.4){(a)~ $\xi_\k(\!A^n\!)$ for the matrices in $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}_{20}$} \put(89,0.5){$n$} \put(-5.6,61){\small$\xi_\k(\!A^n\!)$} \put(101,1.2){\epsfig{figure=testH_20_NZ.eps,width=4.7cm,height=5cm}} \put(101.5,30){z} \put(101.5,28){o} \put(101.5,26){o} \put(100.8,24){m} \put(107,-1.3){\small (b) Zoom of (a) -- $y$-axis} \put(107,70){$\k=1$: \textcolor[rgb]{0,0.5,0}{\bf green $\bm{\circ}$}} \put(107,65){$\k=2$: \textcolor[rgb]{0.75,0,0}{\bf red $\bm{\lozenge}$}} \put(107,60){$\k=3$: \textcolor{blue}{\bf blue $\bm{\square}$}} \put(107,55){$\k=4$: \textcolor[rgb]{0.8,0,0.6}{\bf magenta $\bm{\vartriangle}$}} \epic \ec \caption{$x$-axis: the list of the $20$ random matrices in $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^N_{20}$. (a) $y$-axis: the value $\xi_\k(A^n)$ according to~\eqref{xi} for all $\k\in\II_{\m-1}$ and for all $n\in\mathbb{I}_{20}$. The plot in (b) is a zoom of~(a) along the $y$-axis.} \label{TH} \efig
The worst and the best values of $\xi_\k(\cdot)$ over the whole set $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^N_{20}$ are displayed in Table~\ref{mmu}. \begin{table}} \def\etabe{\end{table}[h!] \caption{The worst and the best values of $\xi_\k(A)$, for $\k\in\II_{\m-1}$, over the set $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^N_{20}$, see \eqref{be}.}
\centering\mbox{\begin{tabular}{l||c|c|c|c|} \hline & $\k=1$ & $\k=2$ & $\k=3$ & $\k=4$\\ \hline
$\disp{\xi_\k^{\mbox{\footnotesize\it worst}}}$ & 0.3519 & 0.1467 & 0.0676 & 0.0072\\ \hline
$\disp{\xi_\k^{\mbox{\footnotesize\it best}}}$ &0.8666 & 0.5881 & 0.3966 & 0.0785\\
\hline \end{tabular}}\label{mmu} \etabe
\paragraph{Set $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^U_{1000}$} The set $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^U_{1000}$ was composed of one thousand $5\times 10$ matrices $A^n$, ~$n\in\mathbb{I}_{1000}$. The entries of each matrix $A^n$ were independent and uniformly sampled on $[-1,1]$. The obtained values for $\xi_\k^{\mbox{\footnotesize\it worst}}$ and $\xi_\k^{\mbox{\footnotesize\it best}}$, calculated according to~\eqref{be}, are shown in Table~\ref{xmu}.
\begin{table}} \def\etabe{\end{table}[h!] \caption{The worst and the best values of $\xi_\k(A)$, for $\k\in\II_{\m-1}$, over the set $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^U_{1000}$.}
\centering\mbox{\begin{tabular}{l||c|c|c|c|} \hline & $\k=1$ & $\k=2$ & $\k=3$ & $\k=4$\\ \hline
$\disp{\xi_\k^{\mbox{\footnotesize\it worst}}}$ & 0.1085 & 0.0235 & 0.0045 & 0.0001\\ \hline
$\disp{\xi_\k^{\mbox{\footnotesize\it best}}}$ &0.9526 & 0.8625 & 0.5379 & 0.1152\\ \hline \end{tabular}}\label{xmu} \etabe For $\k\in\mathbb{I}_3$, the {\em best} values of $\xi_\k(\cdot)$ were obtained for the same matrix, $A^{964}$. Note that $\xi_4(A^{964})= 0.0425\gg\disp{\xi_4^{\mbox{\footnotesize\it worst}}}$. The {\em worst} values in Table~\ref{xmu} are smaller than those in Table~\ref{mmu}, while the {\em best} values in Table~\ref{xmu} are larger than those in Table~\ref{mmu}; one credible reason is that $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^U_{1000}$ is much larger than $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^N_{20}$.
\begin{table}} \def\etabe{\end{table}[h!] \caption{Percentage of the cases in $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^U_{1000}$ when \eqref{tya} fails to hold.}
\centering\mbox{\begin{tabular}{c||c|c|c|} \hline & $\xi_1(A^n)=\xi_2(A^n)$&$\xi_2(A^n)=\xi_3(A^n)$&$\xi_3(A^n)=\xi_4(A^n)$\\ \hline occurrences $\{n\}$ ~ & 5 $\%$& 1.6 $\%$& 0.1 $\%$\\ \hline \end{tabular}} \label{kx} \etabe
Overall, \eqref{tya} is satisfied on $\mathcal{A}}\def\B{\mathcal{B}}\def\C{{\mathcal C}}\def\D{\mathcal{D}}\def\E{{\mathcal E}^U_{1000}$---the percentages in Table~\ref{kx} are pretty small. All three tables and Figure~\ref{TH} agree with our guess that~H\ref{aa} is more viable for smaller values of~$\k$.
{\em Based on the magnitudes for $\xi_\k^{\mbox{\footnotesize\it best}}$ in Tables~\ref{mmu} and~\ref{xmu}, one can expect that there are some classes of matrices (random or not) that fit~H\ref{aa} for larger values of $\xi_\k(\cdot)$.}
\subsection{On the global minimizers of $\bm{\mathcal{F}_d}$}\label{ong}
Here we summarize the outcome of a series of experiments corresponding to several matrices $A\in\RR^{\m\x\n}$ where $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}=5$ and $\n=10$, satisfying H\ref{aa} for $\k=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1$, different original vectors ${\ddot{u}}\in\RR^{\n}$ and data samples $d=A{\ddot{u}}+\mathrm{noise}$, for various values of $\beta>0$. In each experiment, we computed the complete list of all different strict (local) minimizers of $\mathcal{F}_d$, say $\left(\hat u^i\right)_{i=1}^n$. Then the sequence of values $\left(\mathcal{F}_d(\hat u^i)\right)_{i=1}^n$ was sorted in increasing order, $\mathcal{F}_d\left(\hat u^{i_1}\right)\leqslant\mathcal{F}_d\left(\hat u^{i_2}\right)\leqslant\cdots\leqslant\mathcal{F}_d\left(\hat u^{i_n}\right)~.$ A global minimizer $\hat u^{i_1}$ is unique provided that
$\mathcal{F}_d\left(\hat u^{i_1}\right)<\mathcal{F}_d\left(\hat u^{i_2}\right)$. In order to discard numerical errors, we also checked whether $\left|\mathcal{F}_d\left(\hat u^{i_1}\right)-\mathcal{F}_d\left(\hat u^{i_2}\right)\right|$ is easy to detect.
{\em In all experiments we carried out, the following facts were observed: \begin{itemize}} \def\eit{\end{itemize} \item The global minimizer of $\mathcal{F}_d$ was unique---manifestly data $d$ never did belong to the {\em closed negligible} subset $\Sigma_\k$ in~Proposition~\ref{cor}. This confirms Theorem~\ref{Uglob}. \item The global minimizers of $\mathcal{F}_d$ remained unchanged under large variations of~$\beta$. \item The necessary condition for a global minimizer in Proposition~\ref{ou} was met. \eit}
Next we present in detail two of these experiments where $\mathcal{F}_d$ is defined using \begin{equation} A=\left[\begin{array}{c} 7~~~2~~~4~~~9~~~0~~~3~~~3~~~6~~~6~~~7\\ 3~~~4~~~9~~~3~~~3~~~9~~~1~~~3~~~1~~~5\\ 5~~~4~~~2~~~4~~~0~~~7~~~1~~~9~~~2~~~9\\ 8~~~4~~~0~~~9~~~6~~~0~~~4~~~2~~~3~~~7\\ 6~~~3~~~6~~~5~~~0~~~9~~~0~~~0~~~3~~~8 \end{array}\right] ~~\begin{array}{l}d=A{\ddot{u}}+n~,\\~\\ \mbox{~~~where $n$ is noise and}\\~\\ {\ddot{u}}=\big(\,0\,,~1\,,~8\,,~0\,,~3\,,~0\,,~0\,,~0\,,~0\,,~9\,\big)^T.\end{array} \label{num}\end{equation} Only integers appear in~\eqref{num} for better readability. We have $\mathrm{rank}(A)=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}=5$. An exhaustive combinatorial test shows that the arbitrary matrix $A$ in~\eqref{num} satisfies~H\ref{aa} for $\k=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1$. The values of $\xi_\k(A)$ are seen in Table~\ref{D}. One notes that $\mu_2(A)>\mu_1(A)$; hence $\xi_1(A)=\xi_2(A)$.
\begin{table}} \def\etabe{\end{table}[h!] \caption{The values of $\xi_\k(A)$ and $\mu_\k(A)$, $\forall\;\k\in\II_{\m-1}$, for the matrix $A$ in~\eqref{num}.}
\centering\mbox{\begin{tabular}{l||c|c|c|c|} \hline & $\k=1$ & $\k=2$ & $\k=3$ & $\k=4$\\ \hline
$\xi_\k(A)$ & 0.2737 & 0.2737 & 0.2008 & 0.0564\\ \hline \hline
$\mu_\k(A)$ & 0.2737 & 0.2799 & 0.2008 & 0.0564\\ \hline \end{tabular}} \label{D} \etabe
One {\em expects} (at least when data are noise-free) that the global minimizer $\hat u$ of $\mathcal{F}_d$ obeys $\hat\sigma\subseteq\sigma({\ddot{u}})$, where ${\ddot{u}}$ is the original in~\eqref{num}, and that the vanished entries of $\hat u$ correspond to the least entries of ${\ddot{u}}$. This inclusion provides a partial way to rate the quality of the solution provided by a global minimizer $\hat u$ of $\mathcal{F}_d$.
The experiments described hereafter correspond to two data samples relevant to \eqref{num}---without and with noise---and to several values of $\beta>0$.
\paragraph{Noise-free data} The noise-free data in \eqref{num} read as: \begin{equation}} \def\eeq{\end{equation} d=A{\ddot{u}}=\big(~97\,,~~130\,,~~101\,,~~85\,,~~123~\big)^T.\label{kli}\eeq For different values of $\beta$, the global minimizer $\hat u$ is given in Table~\ref{bem}. \begin{table}[h!] \caption{The global minimizer $\hat u$ of $\mathcal{F}_d$ and its value $\mathcal{F}_d(\hat u)$ for the noise-free data $d$ in~\eqref{kli} for different values of $\beta$. Last row: the original ${\ddot{u}}$ in \eqref{num}.} \begin{tabular}{l} \hline
\begin{tabular}{c|c|c|c|c}
$\beta$& The global minimizer $\hat u$ of $\mathcal{F}_d$ {\it ~~(row vector)}&$\|\hat u\|_0$ & $\mathcal{F}_d(\hat u)$ \\ \hline $\begin{array}{c} 1\\ 10^2\\ 10^3\\ 10^4\\ 7\!\cdot\! 10^4 \end{array}$ & $\begin{array}{cccccccccc} 0 & \mathbf{ 1} & \mathbf{8} & 0 & \mathbf{3} & 0 & 0 & 0 & 0 & \mathbf{9}\\
0 & 0 & \mathbf{ 8.12} & 0 & \mathbf{3.31} & 0 & 0 & 0 & 0 & \mathbf{9.33}\\
0 & 0 & 0 & 0 & 0 & \mathbf{12.58} & \mathbf{20.28} & 0 & 0 & 0\\ 0 & \mathbf{29.95} & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0\\ 0&0&0&0&0&0&0&0&0&0 \end{array}$& $\begin{array}{c}4\\3\\2\\1\\0\end{array}$ & $\begin{array}{c} 4\\ 301.52\\ 2179.3\\ 14144\\ 58864 \end{array}$ \\ \hline \hline \end{tabular}\\ $~~~~~~{\ddot{u}}~~~=~\,0~~~~~\;\mathbf{1}~~~~~~~~ \mathbf{8}~~~~~ 0~~~~\,\mathbf{3}~~~~~~~~ 0~~~~~~~~~ 0~~~~~~ 0~~~ 0~~~~~ \mathbf{9}$\\
\hline \end{tabular} \label{bem} \end{table} Since $\sigma({\ddot{u}})\in\Omega$ and the data are noise-free,
$\mathcal{F}_d$ does not have global minimizers with $\|\hat u\|_0=5$. Actually, applying Proposition~\ref{ie} for $\widetilde u={\ddot{u}}$ yields $\beta_{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}=0$, hence for any $\beta>0$ all global minimizers of $\mathcal{F}_d$ have a support in $\Omega=\overline{\Omega}_{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}$ (see Definition~\ref{om} and~\eqref{ga}). The global minimizer $\hat u$ for $\beta=1$ meets $\hat u={\ddot{u}}$. For $\beta=100$, the global minimizer $\hat u$ obeys ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)=\{3,5,10\}\subsetneqq\sigma({\ddot{u}})$ and
$\|\hat u\|_0=3$---the least nonzero entry of the original ${\ddot{u}}$ is canceled, which is reasonable. The global minimizers corresponding to $\beta\geqslant 300$ are meaningless. We could not find any positive value of $\beta$ giving better $2$-sparse global minimizers. Recalling that data are noise-free, this confirms Remark~\ref{x2}:
the global minimizers of $\mathcal{F}_d$ realize a only {\em pseudo}-hard thresholding. For $\beta\geqslant 7\cdot 10^4>\|d\|^2$, the global minimizer of $\mathcal{F}_d$ is $\hat u=0$ which confirms~Remark~\ref{nu}.
\paragraph{Noisy data} Now we consider {\em noisy data} in \eqref{num} for \begin{equation}} \def\eeq{\end{equation} n=\big(~4\,,~~ -1\,,~~ 2\,,~~ -3\,,~~ 5~\big)^T~.\label{noy}\eeq This arbitrary noise yields a signal-to-noise ratio\footnote{ Let us denote $\ddot{d}=A{\ddot{u}}$ and $d=\ddot{d}+n$. The SNR reads \cite{Vetterli95} $\mathrm{SNR}(\ddot{d},d)=10\,\log_{10}\frac{\sum_{i=1}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\left(\ddot{d}[i]-\frac{1}{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}}\sum_{i=1}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K} \ddot{d}[i]\right)^2} {\sum_{i=1}^\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}\left(d[i]-\ddot{d}[i]\right)^2}~.$ } (SNR) equal to $14.07$ dB. If $\beta\leq0.04$, $\mathcal{F}_d$ has $252$ different strict global minimizers $\hat u$
obeying $\|\hat u\|_0=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ and $\mathcal{F}_d(\hat u)=\beta\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ (recall Proposition~\ref{bad}). For $\beta\geq0.05$, the global minimizer $\hat u$ of $\mathcal{F}_d$ is unique and satisfies $\sigma(\hat u)\in\Omega$. It is given in Table~\ref{ben} for several values of $\beta\geq0.05$. \begin{table}[h!] \caption{The global minimizer $\hat u$ of $\mathcal{F}_d$ and its value $\mathcal{F}_d(\hat u)$ for noisy data given by \eqref{num} and \eqref{noy}, for different values of $\beta$. Last row: the original ${\ddot{u}}$.} \begin{tabular}{l} \hline
\begin{tabular}{c|c|c|c}
$\beta$& The global minimizer $\hat u$ of $\mathcal{F}_d$ {\it ~~(row vector)}&$\|\hat u\|_0$ & $\mathcal{F}_d(\hat u)$\\ \hline $\begin{array}{c} 1\\ 10^2\\ 10^3\\ 10^4\\ 7\!\cdot\!10^4 \end{array}$ & $\begin{array}{cccccccccc} 0 & \mathbf{6.02} & \mathbf{2.66} & \mathbf{6.43} & 0 & \mathbf{6.85} & 0 & 0 & 0 & 0\\ 0 & 0 & \mathbf{8.23} & 0 & \mathbf{2.3} & 0 & 0 & 0 & 0 & \mathbf{9.71}\\
0 & 0 & \mathbf{8.14} & 0 & 0 & 0 & 0 & 0 & 0 & \mathbf{10.25}\\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 & \mathbf{14.47}\\ 0&0&0&0&0&0&0&0&0&0 \end{array}$ &$\begin{array}{c}4\\3\\2\\1\\0\end{array}$ & $\begin{array}{c} 4.0436\\301.94\\2174.8\\ 14473\\ 60559 \end{array}$ \\ \hline \hline \end{tabular}\\ $~~~~~~{\ddot{u}}~~~=\,~0~~~~~\mathbf{1}~~~~~~~ \mathbf{8}~~~~~~~\, 0~~~~~~ \mathbf{3}~~~~~~
0~~~~~ 0~~~ 0~~\, 0~~~~~~ \mathbf{9}$\\ \hline \end{tabular} \label{ben} \end{table} For $\beta=1$, the global minimizer is meaningless. We could not find any positive value of $\beta$ yielding a better global minimizer with a $4$-length support. For the other values of $\beta$, the
global minimizer $\hat u$ meets ${\hat\sigma}\stackrel{\rm def}{=}\sigma(\hat u)\varsubsetneqq\sigma({\ddot{u}})$, and its vanished entries correspond to the least entries in the original ${\ddot{u}}$. For $\beta=100$, the global minimizer seems to furnish a good approximation to ${\ddot{u}}$. Observe that the last entry of the global minimizer $\hat u[10]$, corresponding to the largest magnitude in ${\ddot{u}}$, freely increases when
$\beta$ increases from $10^2$ to $10^4$. We tested a tight sequence of intermediate values of $\beta$ without finding better results. Yet again, $\beta\geqslant 7\cdot 10^4>\|d\|^2$ leads to a unique null global minimizer (see Remark~\ref{nu}).
\setlength{\unitlength}{1mm} [h!] \begin{center}} \def\ec{\end{center} \begin{picture}} \def\epic{\end{picture}(160,83) \put(5,7){\epsfig{figure=allmin.eps,width=7.4cm,height=7.6cm}} \put(85,8){\epsfig{figure=allminZ.eps,width=6.9cm,height=7.5cm}} \put(6,0){\footnotesize (a) The values of all strict (local) minimizers of $\mathcal{F}_d$} \put(94,1){\footnotesize (b) Zoom of (a) along the $y$-axis} \put(-2,79){\footnotesize $4\;10^4$} \put(0.5,4.5){\footnotesize $(0,0)$} \put(75.3,4.5){\footnotesize $637$} \put(65.3,4.5){\footnotesize $\{\hat u\}$} \put(0.0,7.5){\footnotesize $302$} \put(-3,68){\footnotesize $\mathcal{F}_d(\hat u)$} \put(50,75){\footnotesize $\mathcal{F}_d(\hat u)$ for}
\put(50,70){\footnotesize $\|\hat u\|_0\!=\!1$: \textcolor[rgb]{0,0.5,0}{\bf green}}
\put(50,65){\footnotesize $\|\hat u\|_0\!=\!2$: \textcolor[rgb]{0.75,0,0}{\bf red}}
\put(50,60){\footnotesize $\|\hat u\|_0\!=\!3$: \textcolor{blue}{\bf blue}}
\put(50,55){\footnotesize $\|\hat u\|_0\!=\!4$: \textcolor[rgb]{0.8,0,0.6}{\bf magenta}}
\put(50,50){\footnotesize $\|\hat u\|_0\!=\!5$: {\bf black}} \put(79.5,68){\footnotesize $\mathcal{F}_d(\hat u)$} \put(80.0,12){\footnotesize $302$} \put(80.0,77.1){\footnotesize $500$} \put(79.8,68){\footnotesize $\mathcal{F}_d(\hat u)$} \put(149.2,5){\footnotesize $637$} \put(140.5,5){\footnotesize $\{\hat u\}$} \put(87.5,5){\footnotesize $0$}
\put(91.9,18){\footnotesize $\|\hat u\|_0\!=\!3$}
\put(109.7,18){\footnotesize $\|\hat u\|_0=4$}
\put(133.3,18){\footnotesize $\|\hat u\|_0=5$} \epic \ec
\caption{All $\bm{638}$ strict (local) minima of $\mathcal{F}_d$ in \eqref{num} for $\beta=100$ and data $d$ corrupted with the arbitrary noise in \eqref{noy}. The $x$-axis lists all strict (local) minimizers $\{\hat u\}$ of $\mathcal{F}_d$ sorted according to their $\ell_0$-norm $\|\hat u\|_0$ in increasing order.
(a) The $y$-axis shows the value $\mathcal{F}_d(\hat u)$ of these minimizers marked with a star. The value of $\mathcal{F}_d$ for $\hat u=0$ is not shown because it is too large ($\mathcal{F}_d(0)=60559=\|d\|^2$). (b) A zoom of (a) along the $y$-axis. It clearly shows that $\mathcal{F}_d$ has a very recognizable unique global minimizer} \label{all} \efig
Figure~\ref{all} shows the value $\mathcal{F}_d(\hat u)$ of all the strict local minimizers of $\mathcal{F}_d$ for $\beta=100$. In the zoom in Figure~\ref{all}(b) it is easily seen that the global minimizer is unique (remember Theorem~\ref{Uglob}). It obeys $\|\hat u\|_0=3$ and $\mathcal{F}_d(\hat u)=301.94$. One observes that $\mathcal{F}_d$ has $252=\,\sharp\,\Omega_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ different strict local minimizers $\hat u$
with $\|\hat u\|_0=5=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ and $\mathcal{F}_d(\hat u)=500=\beta\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$. This confirms Proposition~\ref{bad}---obviously $d$ does not belong to the closed negligible subset $\Qm_\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}$ described in the proposition.
\section{Conclusions and perspectives}\label{Con}
We provided a detailed analysis of the (local and global) minimizers of a regularized objective $\mathcal{F}_d$ composed of a quadratic data fidelity term and an $\ell_0$ penalty weighted by a parameter $\beta>0$, as given in \eqref{fd}. We exhibited easy necessary and sufficient conditions ensuring that a (local) minimizer $\hat u$ of $\mathcal{F}_d$ is strict (Theorem~\ref{ra}). The global minimizers of $\mathcal{F}_d$ (whose existence was proved) were shown to be strict as well (Theorem~\ref{gs}). Under very mild conditions, $\mathcal{F}_d$ was shown to have a unique global minimizer (Theorem~\ref{Uglob}). Other interesting results were listed in the abstract. Below we pose some perspectives and open questions raised by this work. \begin{itemize}} \def\eit{\end{itemize} \item The relationship between the value of the regularization parameter $\beta$ and the sparsity of the global minimizers of $\mathcal{F}_d$ (Proposition~\ref{ie}) can be improved.
\item The {\em generic} linearity in data $d$ of each strict (local) minimizer of $\mathcal{F}_d$ (subsection~\ref{els}) should be exploited to better characterize the global minimizers of $\mathcal{F}_d$.
\item Is there a simple way to check whether assumption H\ref{aa} is satisfied by a given matrix $A\in\RR^{\m\x\n}$ when $\mathsf{N}$ and $\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}<\n$ are large? Remark~\ref{P} and in particular~\eqref{yr} could help to discard some nonrandom matrices. Conversely, one can ask whether there is a systematic way to construct matrices $A$ that satisfy H\ref{aa}.
An alternative would be to exhibit families of matrices that satisfy H\ref{aa} for
large values of $\xi_\k(\cdot)$, where the latter quantifiers are defined in equation~\eqref{xi}.
\item A proper adaptation of the results to matrices $A$ and data $d$ with complex entries should not present inherent difficulties.
\item The theory developed here can be extended to MAP energies of the form evoked in~\eqref{map}. This is important for the imaging applications mentioned there.
\item Based on Corollary~\ref{nor}, and Remarks~\ref{x2} and \ref{ifn}, and the numerical tests in subsection \ref{ong}, one is justified in asking for conditions ensuring that the global minimizers of $\mathcal{F}_d$ perform a valid work. Given the high quality of the numerical results provided in many papers (see e.g.,~\cite{Robini07,Robini10}), the question deserves attention. \eit
There exist numerous algorithms aimed at approximating a (local) minimizer of $\mathcal{F}_d$. As a by-product of our research, we obtained simple rules to verify whether or not an algorithm could find \begin{itemize}} \def\eit{\end{itemize}\item[-] a (local) minimizer $\hat u$ of $\mathcal{F}_d$---by checking whether $\hat u$ satisfies \eqref{ee} in Corollary~\ref{nor};
\item[-] and whether this local minimizer is strict by testing whether the submatrix whose columns are indexed by the support of $\hat u$ (i.e., $A_{\sigma(\hat u)}$) has full column rank (Theorem~\ref{ra}). \eit Some properties of the minimizers of $\mathcal{F}_d$ given in this work can be inserted in numerical schemes in order to quickly escape from shallow local minimizers.
Many existing numerical methods involve a studious choice of the regularization parameter $\beta$, and some of them are proved to converge to a local minimizer of $\mathcal{F}_d$. {\em We have seen that finding a (strict or nonstrict) local minimizer of $\mathcal{F}_d$ is easy and that it is independent of the value of $\beta$ (Corollaries~\ref{nor} and \ref{blg}). {\em It is therefore obscure what meaning to attach to ``choosing a good $\beta$ and proving (local) convergence''.}}
Other successful algorithms are not guaranteed to converge to a local minimizer of $\mathcal{F}_d$. Whenever algorithms do a good job, the choice of $\beta$, the assumptions on $A$ and on $\|\hat u\|_0$, and the iterative scheme and its initialization {\em obviously} provide a tool for selecting a meaningful solution by minimizing $\mathcal{F}_d$. {\em There is a theoretical gap that needs clarification.}
{\em The connection between the existing algorithms and the description of the minimizers exposed in this paper deserves deep exploration. What conditions ensure that an algorithm minimizing $\mathcal{F}_d$ yields meaningful solutions? Clearly, showing local convergence does not answer this important question.}
One can expect such research to give rise to innovative and more efficient algorithms enabling one to compute relevant solutions by minimizing the tricky objective $\mathcal{F}_d$.
\section{Appendix}
\subsection{Proof of Lemma~\ref{tl}}\label{dtl}
Since $\hat u\neq0$, the definition of ${\hat\sigma}$ shows that $\disp{\min_{i\in{\hat\sigma}}\big|\,\hat u[i]\,\big|>0}$. Then $\rho$ in~\eqref{rho} fulfills $\rho>0$.
\paragraph{\rm(i)}~ Since $\,\sharp\,{\hat\sigma}\geq1$, we have
\begin{eqnarray}} \def\eeqn{\end{eqnarray} i\in{\hat\sigma}~,~~v\in\Bm_\infty(0,\rho)~~&\Rightarrow&~~\max_{j\in{\hat\sigma}}\big|\,v[j]\,\big|<\rho\nonumber\\
&\Rightarrow&~~\max_{j\in{\hat\sigma}}\big|\,v[j]\,\big|<\min_{j\in{\hat\sigma}}\big|\,\hat u[j]\,\big|\nonumber\\
&\Rightarrow&~~|\hat u[i]+v[i]|\geqslant |\hat u[i]|-|v[i]|\nonumber\\
&&~~\geqslant\min_{j\in{\hat\sigma}} |\hat u[j]|-\max_{j\in{\hat\sigma}}|v[j]|\geqslant\rho-\max_{j\in{\hat\sigma}}|v[j]|>0\nonumber\\ &\Rightarrow&~~ \hat u[i]+v[i]\neq0~\nonumber\\ \Big[\mbox{\rm by~\eqref{phi}~}\Big]~~~~~~~~&\Rightarrow&~~\phi\left(\hat u[i]+v[i]\right)=\phi\left(\hat u[i]\right)=1~. \label{ry}\eeqn If ${\hat\sigma}^c=\varnothing$ the result is proved. Let ${\hat\sigma}^c\neq\varnothing$. Then $\hat u[i]=0=\phi\left(\hat u[i]\right)$, $\forall\; i\in{\hat\sigma}^c$. Inserting this and~\eqref{ry} into \[\sum_{i\in\II_{\n}}\phi\big(\,\hat u[i]+v[i]\,\big)= \sum_{i\in{\hat\sigma}}\phi\big(\,\hat u[i]+v[i]\,\big)+\sum_{i\in{\hat\sigma}^c}\phi\big(\,\hat u[i]+v[i]\,\big)~\] proves claim (i).
\paragraph{\rm(ii)}~
Using the fact that $\|A(\hat u+v)-d\|^2=\|A\hat u-d\|^2+\|Av\|^2+2\<Av,\,A\hat u-d\>$, one obtains \begin{eqnarray}} \def\eeqn{\end{eqnarray} v\in\Bm_\infty(0,\rho)\,\mathbf{\setminus}\,\Km_{{\hat\sigma}}~~~\Rightarrow~~~
\mathcal{F}_d(\hat u+v)&=&\|A\hat u-d\|^2+\|Av\|^2+2\<Av,\,A\hat u-d\>\nonumber\\ \Big[\mbox{\rm by Lemma~\ref{tl}(i)}\Big]~~~~&&+\beta\sum_{i\in{\hat\sigma}}\phi\left(\hat u[i]\right)+\beta\sum_{i\in{\hat\sigma}^c}\phi\left(v[i]\right)\nonumber\\
\Big[\mbox{\rm using~\eqref{fds}~}\Big]~~~~&=&\mathcal{F}_d(\hat u)+\|Av\|^2+2\<Av,\,A\hat u-d\>+\beta\sum_{i\in{\hat\sigma}^c}\phi\left(v[i]\right)\nonumber\\
&\geqslant&\mathcal{F}_d(\hat u)-\big|2\<v,\,A^T(A\hat u-d)\>\big|+\beta\|v_{{\hat\sigma}^c}\|_0\nonumber\\
\Big[\mbox{\rm by H\"{o}lder's inequality}\Big]~~~~&\geqslant&\mathcal{F}_d(\hat u)-2\|v\|_\infty\,\|A^T(A\hat u-d)\|_1+\beta \|v_{{\hat\sigma}^c}\|_0~.\label{pop}
\eeqn If $\,\sharp\,{\hat\sigma}^c=0$, then $\Km_{{\hat\sigma}}=\RR^{\n}$, so $v\in\mathbb R^0$ and $\|v\|_\infty=0$; hence we have the inequality.
Let $\,\sharp\,{\hat\sigma}^c\geq1$. For $v\not\in\Km_{{\hat\sigma}}$, there at least one index $i\in{\hat\sigma}^c$ such that $v[i]\neq0$; hence $\|v_{{\hat\sigma}^c}\|_0\geq1$. The definition of $\rho$ in~\eqref{rho} shows that \begin{eqnarray*}} \def\eeqnn{\end{eqnarray*} v\in\Bm_\infty(0,\rho)\,\mathbf{\setminus}\,\Km_{{\hat\sigma}}~&\Rightarrow&~
-\|v\|_\infty>-\rho\geqslant-\,\frac{\beta}{2\Big(\|A^T(A\hat u-d)\|_1+1\Big)}\\
~&\Rightarrow&~-2\|v\|_\infty\,\|A^T(A\hat u-d)\|_1+\beta\|v_{{\hat\sigma}^c}\|_0
>-\,\frac{2\beta\|A^T(A\hat u-d)\|_1}{2\Big(\|A^T(A\hat u-d)\|_1+1\Big)}+\beta>0~. \eeqnn Introducing the last inequality into~\eqref{pop} shows that for $\,\sharp\,{\hat\sigma}^c\geq1$, the inequality in (ii) is strict.
\subsection{Proof of Proposition~\ref{ou}}\label{pou} If $\hat u=0$, the statement is obvious. We focus on $\hat u\neq0$. For an arbitrary $i\in\II_{\n}$, define \[\hat u^{(i)}\stackrel{\rm def}{=}\big(\hat u[1],\cdots,\hat u[i-1],\,0,\,\hat u[i+1],\cdots,\hat u[\n]\big)\in\RR^{\n}~.\] We shall use the equivalent formulation of $\mathcal{F}_d$ given in~\eqref{fds}. Clearly\footnote{Using the definition of $\hat u^{(i)}$, we have $\hat u^{(i)}=A_{(\II_{\n}\,\mathbf{\setminus}\,\{i\})}\hat u_{(\II_{\n}\,\mathbf{\setminus}\,\{i\})}~,$ hence $A\hat u^{(i)}$ is independent of $\hat u[i]$. },
\[\mathcal{F}_d(\hat u)=\mathcal{F}_d\big(\hat u^{(i)}+e_i\hat u[i]\big)=\|A\hat u^{(i)}+a_i\hat u[i]-d\|^2+
\beta \sum_{j\in\II_{\n}}\phi\left(\hat u^{(i)}[j]\right)+\phi\left(\hat u[i]\right)~.\] Consider $f:\mathbb R\rightarrow\mathbb R$ as given below \begin{equation}} \def\eeq{\end{equation} f(t)\stackrel{\rm def}{=}\mathcal{F}_d\left(\hat u^{(i)}+e_it\right). \label{fi}\eeq Since $\hat u$ is a global minimizer of $\mathcal{F}_d$, for any $i\in\II_{\n}$, we have \begin{eqnarray*}} \def\eeqnn{\end{eqnarray*} f\left(\hat u[i]\right)&=&\mathcal{F}_d\big(\hat u^{(i)}+e_i\hat u[i]\big)\\
&\leqslant&\mathcal{F}_d\big(\hat u^{(i)}+e_i\,t\big)=f(t)~~~\forall\; t\in\mathbb R~.\eeqnn Equivalently, for any $i\in\II_{\n}\,$, $\,f\left(\hat u[i]\right)\,$ is the global minimum of $f(t)$ on $\mathbb R$. Below we will determine the global minimizer(s) $\hat t=\hat u[i]$ of $f$ as given in~\eqref{fi}, i.e., \[\hat t=\hat u[i]=\arg\min_{t\in\mathbb R}f(t)~.\] In detail, the function $f$ reads as
\begin{eqnarray}} \def\eeqn{\end{eqnarray} f(t)&=&\|A\hat u^{(i)}+a_it-d\|^2+
\beta \sum_{j\in\II_{\n}}\phi\left(\hat u^{(i)}[j]\right)+\beta\phi(t)\nonumber\\
&=&\|A\hat u^{(i)}-d\|^2+\|a_i\|^2t^2+2t\<a_i,A\hat u^{(i)}-d\> +\beta \sum_{j\in\II_{\n}}\phi\left(\hat u^{(i)}[j]\right)+\beta\phi(t)\nonumber\\
&=&\|a_i\|^2t^2+2t\<a_i,A\hat u^{(i)}-d\>+\beta\phi(t)+C\label{fin0}~, \eeqn where
\[C=\|A\hat u^{(i)}-d\|^2+\beta \sum_{j\in\II_{\n}}\phi\left(\hat u^{(i)}[j]\right)~.\] Note that $C$ does not depend on $t$. The function $f$ has two local minimizers denoted by $\hat t_0$ and $\hat t_1$. The first is \begin{equation}} \def\eeq{\end{equation}\hat t_0=0~~~\Rightarrow~~~f(\hat t_0)=C~.\label{min1}\eeq The other one, $\hat t_1\neq0$, corresponds to $\phi(t)=1$. From~\eqref{fin0}, $\hat t_1$ solves
\[2\|a_i\|^2t+2\<a_i,A\hat u^{(i)}-d\>=0~.\] Recalling that $a_i\neq0$, $\forall\; i\in\II_{\n}$ (see \eqref{ao}), it follows that
\begin{equation}} \def\eeq{\end{equation}\hat t_1=-\frac{\<a_i,A\hat u^{(i)}-d\>}{\|a_i\|^2}~~~\Rightarrow~~~
f(\hat t_1)=-\frac{\<a_i,A\hat u^{(i)}-d\>^2}{\|a_i\|^2}+\beta+C~. \label{min2}\eeq Next we check whether $\hat t_0$ or $\hat t_1$ is a global minimizer of $f$. From~\eqref{min1} and~\eqref{min2} we get
\[ f(\hat t_0)-f(\hat t_1)=\frac{\<a_i,A\hat u^{(i)}-d\>^2}{\|a_i\|^2}-\beta~. \] Furthermore, \begin{eqnarray}} \def\eeqn{\end{eqnarray} f(\hat t_0)<f(\hat t_1)~~&\Rightarrow&~~\hat u[i]=\hat t_0=0~,\nonumber\\
f(\hat t_1)<f(\hat t_0)~~&\Rightarrow&~~\hat u[i]=\hat t_1=-\frac{\<a_i,A\hat u^{(i)}-d\>}{\|a_i\|^2}~,\label{min3}\\ f(\hat t_0)=f(\hat t_1)~~&\Rightarrow&~~\mbox{$\hat t_0$ and $\hat t_1$ are global minimizers of $f$.}\nonumber \eeqn In particular, we have
\begin{eqnarray}} \def\eeqn{\end{eqnarray} f(\hat t_1)\leqslant f(\hat t_0)~~&\Leftrightarrow&~~\<a_i,A\hat u^{(i)}-d\>^2\geqslant\beta\|a_i\|^2\label{Delta1}\\ \Big[\mbox{\rm by~\eqref{min3}}\Big]~~~~~~~~&\Rightarrow&~
|\hat u[i]|=\frac{\left|\<a_i,A\hat u^{(i)}-d\>\right|}{\|a_i\|^2}\nonumber\\ \Big[\mbox{\rm by~\eqref{Delta1}}\Big]~~~~~~~~
&&~~~~~~~~\geqslant\frac{\sqrt{\beta}\|a_i\|}{\|a_i\|^2}=\frac{\sqrt{\beta}}{\|a_i\|}~.\nonumber \eeqn It is clear that the conclusion holds true for any $i\in\II_{\n}$.
\subsection{ Proof of Proposition~\ref{pal}}\label{palp} The asymptotic function $\left(\mathcal{F}_d\right)_\infty(v)$ of $\mathcal{F}_d$ can be calculated according to\footnote{In the nonconvex case, the notion of asymptotic functions and the representation formula were first given by J.P. Dedieu \cite{Dedieu77}.} ~\cite[Theorem 2.5.1]{Auslender03} \[ \left(\mathcal{F}_d\right)_\infty(v)=\liminf_{\begin{array}} \def\earr{\end{array}{c}v'\rightarrow v\\ t\rightarrow\infty\earr}\frac{\mathcal{F}_d(tv')}{t}~.\] Then
\begin{eqnarray*}} \def\eeqnn{\end{eqnarray*} \left(\mathcal{F}_d\right)_\infty(v)&=&\liminf_{\begin{array}} \def\earr{\end{array}{c}v'\rightarrow v\\ t\rightarrow\infty\earr}\frac{\|Av'-d\|^2+\beta\|v'\|_0}{t}\nonumber\\
&=&\liminf_{\begin{array}} \def\earr{\end{array}{c}v'\rightarrow v\\ t\rightarrow\infty\earr}\left(t\|Av'\|^2-2\<d,Av'\>+\frac{\|d\|^2+\beta\|v'\|_0}{t}\right)\nonumber\\~\nonumber\\ &=&\left\{\begin{array}} \def\earr{\end{array}{lll} 0 & \mbox{if} & v\in \mathrm{ker}} \def\ran{\mathrm{range}(A)~,\\+\infty & \mbox{if} & v\not\in \mathrm{ker}} \def\ran{\mathrm{range}(A)~.\earr \right.\label{auv} \eeqnn Hence \begin{equation}} \def\eeq{\end{equation}\mathrm{ker}} \def\ran{\mathrm{range}\left((\mathcal{F}_d)_\infty\right)=\mathrm{ker}} \def\ran{\mathrm{range}(A)~,\label{wd}\eeq where $\mathrm{ker}} \def\ran{\mathrm{range}\left((\mathcal{F}_d)_\infty\right)=\{v\in\RR^{\n}~:~(\mathcal{F}_d)_\infty(v)=0\}$.
Let $\{v_k\}$ satisfy~\eqref{aal} with $v_k\,\|v_k\|^{-1}\rightarrow \bar v\in\mathrm{ker}} \def\ran{\mathrm{range}(A)$. Below we compare the numbers $\|v_k\|_0$ and $\|v_k-\rho \bar v\|_0$ where $\rho>0$. There are two options. \begin{enumerate}} \def\een{\end{enumerate}
\item Consider that $i\in\sigma(\bar v)$, that is, $\disp{\bar v[i]=\lim_{k\rightarrow\infty} v_k[i] \, \|v_k\|^{-1}\neq 0}$. Then $|\,v_k[i]\,|>0$ for all but finitely many $k$ as otherwise, $v_k[i] \, \|v_k\|^{-1}$ would converge to 0. Therefore, there exists $k_i$ such that
\begin{equation}} \def\eeq{\end{equation} \left|\,v_k[i]-\rho\,\bar v[i]\,\right|\geq0\quad\mbox{and}\quad |\,v_k[i]\,|>0 \quad\forall\; k\geqslant k_i~.\label{wb}\eeq
\item If $i\in(\sigma(\bar v))^c$, i.e. $\bar v[i]=0$, then clearly \begin{equation}} \def\eeq{\end{equation} v_k[i]-\rho\, \bar v[i]=v_k[i] ~.\label{wc}\eeq
\een Combining \eqref{wb} and \eqref{wc}, the definition of $\|\cdot\|_0$ using $\phi$ in \eqref{phi} shows that
\begin{equation}} \def\eeq{\end{equation}\|v_k-\rho\,\bar v\|_0\leqslant\|v_k\|_0 \quad\forall\; k\geqslant k_0\stackrel{\rm def}{=}\max_{i\in\sigma(\bar v)}k_i~. \label{we}\eeq
By \eqref{wd}, $A\bar v=0$. This fact, jointly with~\eqref{we}, entails that
\begin{eqnarray}} \def\eeqn{\end{eqnarray}\mathcal{F}_d(v_k-\rho\bar v)&=&\|A(v_k+\rho\bar v)-d\|^2+\beta\|v_k-\rho\bar v\|_0\nonumber\\
&=&\|Av_k-d\|^2+\beta\|v_k-\rho\bar v\|_0\nonumber\\
&\leqslant& \|Av_k-d\|^2+\beta\|v_k\|_0=\mathcal{F}_d(v_k)\quad\forall\; k\geqslant k_0~.\nonumber\label{hb}\eeqn It follows that for any $k\geqslant k_0$ we have \[v_k\in\mathrm{lev}\,(\mathcal{F}_d,\lambda_k)\quad \Rightarrow\quad v_k-\rho\bar v\in\mathrm{lev}\,(\mathcal{F}_d,\lambda_k)~,\] and thus $\mathcal{F}_d$ satisfies Definition~\ref{als}.
\subsection{Proof of Proposition~\ref{ie}}\label{pie} Given $\k\in\mathbb{I}_{\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1}$, set \begin{equation}} \def\eeq{\end{equation}\Um_{\k+1}\stackrel{\rm def}{=}\bigcup_{\omega\subset\II_{\n}}\left\{\bar{u}~:~\bar{u}~~\mbox{solves {\rm{($\,\P_{\vt}\,$)}~}}~~\mbox{\rm and}~~
\|\bar{u}\|_0\geqslant\k+1\right\}~.\label{ws}\eeq \begin{itemize}} \def\eit{\end{itemize}\item Let $\Um_{\k+1}\neq\varnothing$. By Proposition~\ref{cp}, for any $\beta>0$, $\mathcal{F}_d$ has a (local) minimum at each $\bar{u}\in\Um_{\k+1}$. Thus
\begin{equation}} \def\eeq{\end{equation}\bar{u}~~\mbox{is a (local) minimizer of $\mathcal{F}_d$ and}~~\|\bar{u}\|_0\geqslant\k+1 ~~~\Leftrightarrow~~~\bar{u}\in\Um_{\k+1}~.\label{koi}\eeq Then for any $\beta>0$ \begin{equation}} \def\eeq{\end{equation}\mathcal{F}_d(\bar{u})\geqslant\beta(\k+1)\quad\forall\;\bar{u}\in\Um_{\k+1}~.\label{kf}\eeq
Let $\widetilde u$ be defined by\footnote{Such a $\widetilde u$ always exists; see subsection~\ref{ntd}. By Proposition~\ref{cp} and Theorem~\ref{ra}, it is uniquely defined.}: \[\widetilde u~~\mbox{solves~{\rm{($\,\P_{\vt}\,$)}~}~for~some}~~\omega\in\Omega_\k~.\] Then
\begin{equation}} \def\eeq{\end{equation} \|\widetilde u\|_0\leqslant\k~\label{ku}~.\eeq Set $\beta$ and $\beta_\k$ according to
\begin{equation}} \def\eeq{\end{equation}\beta>\beta_\k\stackrel{\rm def}{=}\|A\widetilde u-d\|^2~.\label{kn}\eeq For such a $\beta$ we have
\begin{eqnarray*}} \def\eeqnn{\end{eqnarray*}\mathcal{F}_d(\widetilde u)&=&\|A\widetilde u-d\|^2+\beta\|\widetilde u\|_0\\ \Big[\mbox{\rm by~\eqref{ku}~and~\eqref{kn}~}\Big]&<&\beta+\beta\k=\beta(\k+1)\\ \Big[\mbox{\rm by~\eqref{kf}~}\Big]&\leqslant&\mathcal{F}_d(\bar{u})~~~~\forall\;\bar{u}\in\Um_{\k+1}~. \eeqnn Let $\hat u$ be a global minimizer of $\mathcal{F}_d$. Then \[\mathcal{F}_d(\hat u)\leqslant\mathcal{F}_d(\widetilde u)<\mathcal{F}_d(\bar{u})~~~~\forall\;\bar{u}\in\Um_{\k+1}~.\]
Using \eqref{ws}-\eqref{koi}, we find $$\|\hat u\|_0\leqslant\k~.$$ \item $\Um_{\k+1}=\varnothing$ entails that\footnote{Let $A=(e_1,e_2,e_3,e_4,e_1)\in\mathbb R^{4\times 5}$ and $d=e_1\in\mathbb R^4$. For $\k=\mathsf{M}} \def\n{\mathsf{N}} \def\l{\mathsf{L}} \def\k{\mathsf{K}-1=3$ one can check that $\Um_{\k+1}=\varnothing$.}
\begin{equation}} \def\eeq{\end{equation}\bar{u}~~\mbox{solves~{\rm{($\,\P_{\vt}\,$)}~}~for}~~\omega\subset\II_{\n},~~\,\sharp\,\omega\geqslant\k+1~~~\Rightarrow~~~\|\bar{u}\|_0\leqslant\k~.\label{ha}\eeq Let $\hat u$ be a global minimizer of $\mathcal{F}_d$. By \eqref{ha} we have
\[\|\hat u\|_0\leqslant\k~.\] \eit According to Theorem~\ref{gs}(ii), any global minimizer of $\mathcal{F}_d$ is strict, hence $\disp{\sigma(\hat u)\in\overline{\Omega}_\k}~.$ \section*{Acknowledgments} The author would like to thank the anonymous Reviewer $2$ for very helpful remarks and suggestions.
\defIEEE Transactions on Acoustics Speech and Signal Processing{IEEE Transactions on Acoustics Speech and Signal Processing} \defIEEE Transactions on Image Processing{IEEE Transactions on Image Processing} \defIEEE Transactions on Information Theory{IEEE Transactions on Information Theory} \defIEEE Transactions on Pattern Analysis and Machine Intelligence{IEEE Transactions on Pattern Analysis and Machine Intelligence} \defIEEE Transactions on Signal Processing{IEEE Transactions on Signal Processing} \defJournal of the Royal Statistical Society B{Journal of the Royal Statistical Society B} \defCompte-rendus de l'acad\'emie des sciences{Compte-rendus de l'acad\'emie des sciences} \defBiometrika{Biometrika} \defInternational Journal of Computer Vision{International Journal of Computer Vision} \defSIAM Journal on Multiscale Modeling and Simulation{SIAM Journal on Multiscale Modeling and Simulation} \defJournal of Applied Statistics{Journal of Applied Statistics} \defSIAM Review{SIAM Review} \defSIAM Journal on Imaging Sciences{SIAM Journal on Imaging Sciences} \defJournal of Global Optimization{Journal of Global Optimization} \defJournal of Scientific Computing{Journal of Scientific Computing} \defMathematics of Computation{Mathematics of Computation} \defMathematical Programming{Mathematical Programming} \defJournal of Fourier Analysis and Applications{Journal of Fourier Analysis and Applications}
\def\uppercase{\uppercase}
\end{document}
|
arXiv
|
{
"id": "1304.5218.tex",
"language_detection_score": 0.47840768098831177,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Sampling of globally depolarized random quantum circuit} \begin{flushright} YITP-19-90 \end{flushright} \author{Tomoyuki Morimae} \email{[email protected]} \affiliation{Yukawa Institute for Theoretical Physics, Kyoto University, Kitashirakawa Oiwakecho, Sakyoku, Kyoto 606-8502, Japan} \affiliation{JST, PRESTO, 4-1-8 Honcho, Kawaguchi, Saitama, 332-0012, Japan} \author{Yuki Takeuchi} \email{[email protected]} \affiliation{NTT Communication Science Laboratories, NTT Corporation, 3-1 Morinosato Wakamiya, Atsugi, Kanagawa 243-0198, Japan} \author{Seiichiro Tani} \email{[email protected]} \affiliation{NTT Communication Science Laboratories, NTT Corporation, 3-1 Morinosato Wakamiya, Atsugi, Kanagawa 243-0198, Japan}
\date{\today} \begin{abstract} The recent paper [F. Arute et al. Nature {\bf 574}, 505 (2019)] considered exact classical sampling of the output probability distribution of the globally depolarized random quantum circuit. In this paper, we show three results. First, we consider the case when the fidelity $F$ is constant. We show that if the distribution is classically sampled in polynomial time within a constant multiplicative error, then ${\rm BQP}\subseteq{\rm SBP}$, which means that BQP is in the second level of the polynomial-time hierarchy. We next show that for any $F\le1/2$, the distribution is classically trivially sampled by the uniform distribution within the multiplicative error $F2^{n+2}$, where $n$ is the number of qubits. We finally show that for any $F$, the distribution is classically trivially sampled by the uniform distribution within the additive error $2F$. These last two results show that if we consider realistic cases, both $F\sim2^{-m}$ and $m\gg n$, or at least $F\sim2^{-m}$, where $m$ is the number of gates, quantum supremacy does not exist for approximate sampling even with the exponentially-small errors. We also argue that if $F\sim2^{-m}$ and $m\gg n$, the standard approach will not work to show quantum supremacy even for exact sampling. \end{abstract}
\maketitle
\section{Introduction} Several sub-universal quantum computing models are shown to be hard to classically simulate. For example, output probability distributions of the depth-four model~\cite{TD}, the Boson Sampling model~\cite{BS}, the IQP model~\cite{IQP1,IQP2}, the one-clean qubit model~\cite{KL,MFF,M,Kobayashi,KobayashiICALP}, and the random circuit model~\cite{random1,random2} cannot be classically sampled in polynomial time unless some conjectures in classical complexity theory (such as the infiniteness of the polynomial-time hierarchy) are refuted. Impossibilities of exponential-time classical simulations of sub-universal quantum computing models have also been shown recently based on classical fine-grained complexity theory~\cite{Dalzell,DalzellPhD,Huang,Huang2,MorimaeTamaki}.
Let $p_z$ be the probability that an $n$-qubit ideal random quantum circuit outputs the $n$-bit string $z\in\{0,1\}^n$. Ref.~\cite{Google} considered the globally depolarized version where the probability $p_z'$ that the output is $z\in\{0,1\}^n$ is written as \begin{eqnarray*} p_z'=Fp_z+\frac{1-F}{2^n}, \end{eqnarray*} where $0< F<1$ is the fidelity.
In this paper, we show the following three results:
\begin{theorem} \label{theorem:result} Assume that $F$ is constant. Then, if the probability distribution $\{p_z'\}_{z\in\{0,1\}^n}$ is sampled in classical $poly(n)$ time within a constant multiplicative error $\epsilon<1$, then ${\rm BQP}\subseteq{\rm SBP}$. \end{theorem}
\begin{theorem} \label{theorem:result2} For any $F\le\frac{1}{2}$, $\{p_z'\}_{z\in\{0,1\}^n}$ is classically sampled by the uniform distribution within the multiplicative error $F2^{n+2}$. \end{theorem}
\begin{theorem} \label{theorem:result3} For any $F$, $\{p_z'\}_{z\in\{0,1\}^n}$ is classically sampled by the uniform distribution within the additive error $2F$. \end{theorem}
Proofs are given in the later sections. In the rest of this section, we provide several remarks.
First, the class SBP~\cite{SBP} is defined as follows.
\begin{definition} A language $L$ is in SBP if and only if there exist a polynomial $s$ and a classical polynomial-time probabilistic algorithm such that if
$x\in L$ then $p_{acc}\ge 2^{-s(|x|)}$, and if $x\notin L$ then $p_{acc}\le 2^{-s(|x|)-1}$. Here, $p_{acc}$ is the acceptance probability. \end{definition}
Note that the class SBP remains unchanged even when the two thresholds, $2^{-s(|x|)}$ and $2^{-s(|x|)-1}$, are replaced with
$\alpha 2^{-s(|x|)}$ and $\beta2^{-s(|x|)}$, respectively, for any constants $\alpha$ and $\beta$ satisfying $0\le\beta<\alpha\le1$. It is known that SBP is in AM, and therefore ${\rm BQP}\subseteq{\rm SBP}$ means that BQP is in the second level of the polynomial-time hierarchy. The containment of BQP in the polynomial-time hierarchy is not believed. (For example, there is an oracle separation~\cite{Raz}.)
Second, note that quantum supremacy for the globally depolarized circuits was previously studied in Ref.~\cite{nonclean} for the one-clean qubit model.
Third, Theorem~\ref{theorem:result} holds for a broader class of quantum circuits than the globally depolarized random circuit. In particular, we can replace our random gate application with the coherent one. In this paper, however, we concentrate on the globally depolarized random circuit for the simplicity. Theorem~\ref{theorem:result2} and Theorem~\ref{theorem:result3} hold for the output probability distribution $\{p_z\}_z$ of any quantum circuit.
Finally, in Ref.~\cite{Google}, it was claimed that if the exact polynomial-time classical sampling of $\{p_z'\}_z$ is possible, then estimating
$|\langle0^n|U|0^n\rangle|^2$ for an $n$-qubit unitary $U$ can be done by an Arthur-Merlin protocol with the $F^{-1}$-time Arthur. However, if we consider the realistic case, $F\sim2^{-m}$ and $m\gg n$, where $m$ is the number of gates, the time-complexity of Arthur is $\sim2^m$.
(On the other hand, the exact computation of $|\langle0^n|U|0^n\rangle|^2$ can be done in time $\sim2^n$.) Moreover, although Ref.~\cite{Google} considered exact sampling of $\{p_z'\}_z$, what a realistic quantum computer can do is approximately sampling $\{p_z'\}_z$. Theorem~\ref{theorem:result2} shows that if we consider the realistic case, $F\sim2^{-m}$ and $m\gg n$, quantum supremacy does not exist for approximate sampling of $\{p_z'\}_z$ even with the exponentially-small multiplicative error $\sim 2^{-(m-n)}$. Theorem~\ref{theorem:result3} shows that if we consider the realistic case, $F\sim2^{-m}$, quantum supremacy does not exist for approximate sampling of $\{p_z'\}_z$ even with the exponentially-small additive error $\sim2^{-m}$.
\section{Discussion}
Our theorems show that if $F\sim2^{-m}$ and $m\gg n$, or at least $F\sim2^{-m}$, quantum supremacy does not exist for approximate sampling of $\{p_z'\}_z$. In this section, we argue that if $F\sim2^{-m}$ and $m\gg n$, the standard approach will not work to show quantum supremacy for exact sampling of $\{p_z'\}_z$.
In the standard proof of quantum supremacy~\cite{TD,BS,IQP1,MFF,Kobayashi,KobayashiICALP}, we first consider the following promise problem: given the classical description of an $n$-qubit $m$-size quantum circuit $U$, and parameters $a$ and $b$, decide $p_{acc}\ge a$ or $p_{acc}\le b$, where $p_{acc}$ is the acceptance probability. In the standard proof of quantum supremacy, we take the promise problem as the complete problem of a ``strong" quantum class, such as postBQP, SBQP, or NQP.
We next assume that $p_{acc}'\equiv Fp_{acc}+\frac{1-F}{2^n}$ is exactly classically sampled. It means that there exists a polynomial-time classical probabilistic algorithm that accepts with probability $q_{acc}$ such that $q_{acc}=p_{acc}'$.
If the answer of the promise problem is yes, then $q_{acc}\ge Fa+\frac{1-F}{2^n}\equiv\alpha$. If the answer of the promise problem is no, then $q_{acc}\le Fb+\frac{1-F}{2^n}\equiv\beta$. In the standard proof of quantum supremacy, we then conclude that the promise problem is in a ``weaker" class (such as postBPP, SBP, or NP) that leads to an unlikely consequence in complexity theory, such as ${\rm postBQP}\subseteq{\rm postBPP}$, ${\rm SBQP}\subseteq{\rm SBP}$, or ${\rm NQP}\subseteq{\rm NP}$. However, deciding $q_{acc}\ge \alpha$ or $q_{acc}\le \beta$ seems to be ``more difficult" than the original promise problem: the original promise problem can be solved in time $\sim2^n$, while deciding $q_{acc}\ge\alpha$ or $q_{acc}\le\beta$ will not be solved in that time because $\alpha-\beta=F(a-b)=O(2^{-m})$, and $m\gg n$. Therefore we will not have any unlikely consequence in this approach.
Although the above argument does not exclude the existence of a completely new supremacy proof for the exact sampling of $\{p_z'\}_z$ that works even when $F\sim2^{-m}$ and $m\gg n$, we can also argue that even if the realistic quantum computer exactly samples $\{p_z'\}_z$, it is ``effectively" classically samplable by the uniform distribution when $F\sim2^{-m}$ unless we can access exponentially many samples.
To see this, let us consider the task of distinguishing $\rho_0\equiv\frac{I^{\otimes n}}{2^n}$ and $\rho_1\equiv F\rho+(1-F)\frac{I^{\otimes n}}{2^n}$, where $\rho$ is any $n$-qubit state. Assume that we can measure $k$ copies of $\rho_0$ or $\rho_1$. Let $\Pi_0$ $(\Pi_1)$ be the POVM element that we conclude that the actual state is $\rho_0^{\otimes k}$ ($\rho_1^{\otimes k}$), where $\Pi_0+\Pi_1=I^{\otimes nk}$. The probability $p_{correct}$ that we make the correct decision is \begin{eqnarray*} p_{correct}&\equiv& \frac{1}{2}{\rm Tr}(\Pi_0\rho_0^{\otimes k}) +\frac{1}{2}{\rm Tr}(\Pi_1\rho_1^{\otimes k})\\ &=& \frac{1}{2}+\frac{1}{2}\Big[ {\rm Tr}(\Pi_0\rho_0^{\otimes k}) -{\rm Tr}(\Pi_0\rho_1^{\otimes k})\Big]\\ &\le& \frac{1}{2}+
\frac{1}{4}\big\|\rho_0^{\otimes k}
-\rho_1^{\otimes k}\big\|_1\\ &\le& \frac{1}{2} +\frac{k}{4}
\big\|\rho_0-\rho_1\big\|_1\\ &=& \frac{1}{2}+
\frac{k}{4}\Big\|\frac{I^{\otimes n}}{2^n}
-\Big[F\rho+(1-F)\frac{I^{\otimes n}}{2^n}\Big]\Big\|_1\\ &=& \frac{1}{2}+
\frac{kF}{4}\Big\|
\rho-\frac{I^{\otimes n}}{2^n}\Big\|_1\\ &\le&\frac{1}{2}+\frac{kF}{2}. \end{eqnarray*} If $F\sim2^{-m}$ and $k=o(2^m)$, $p_{correct}\to \frac{1}{2}$.
\section{Proof of Theorem~\ref{theorem:result}}
Assume that a language $L$ is in BQP. Then for any polynomial $r$, there exists a polynomial-time uniformly generated family $\{V_x\}_x$ of quantum circuits such that if $x\in L$ then $p_{acc}\ge 1-2^{-r(|x|)}$, and if $x\notin L$ then $p_{acc}\le 2^{-r(|x|)}$. Here \begin{eqnarray*}
p_{acc}\equiv\langle0^w|V_x|0^w\rangle \end{eqnarray*}
with $w=poly(|x|)$ is the acceptance probability.
Let $m$ be the number of elementary gates in $V_x$, i.e., $V_x=w_mw_{m-1}...w_2w_1$, where each $w_j$ is an elementary gate (such as $H$, $CNOT$, and $T$, etc.). Let us consider the following random quantum circuit on $n\equiv w+m$ qubits: \begin{itemize} \item[1.]
The initial state is $|0^w\rangle\otimes|0^m\rangle$, where we call the first $w$-qubit register the main register, and the second $m$-qubit register the ancilla register. \item[2.] For each $j=1,2,...,m$, apply $w_j\otimes I$ or $\eta_j\otimes X$ with probability 1/2, where $\eta_j$ is any elementary gate, $w_j$ and $\eta_j$ act on the main register, and $I$ and $X$ act on the $j$th qubit of the ancilla register. Thus obtained the final state is \begin{eqnarray} \frac{1}{2^m}\sum_{\alpha\in\{0,1\}^m}
\xi_m^{\alpha_m}...\xi_1^{\alpha_1}|0^w\rangle\langle0^w| (\xi_1^{\alpha_1})^\dagger...(\xi_m^{\alpha_m})^\dagger
\otimes|\alpha\rangle\langle\alpha|, \label{final} \end{eqnarray} where $\alpha\equiv(\alpha_1,...,\alpha_m)\in\{0,1\}^m$ is an $m$-bit string, $\xi_j^0=w_j$, and $\xi_j^1=\eta_j$. \item[3.] Measure all $n$ qubits in the computational basis. If all results are 0, accept. Otherwise, reject. \end{itemize}
If we consider the globally depolarized version, the state of Eq.~(\ref{final}) is replaced with \begin{eqnarray*} \frac{F}{2^m}\sum_{\alpha\in\{0,1\}^m}
\xi_m^{\alpha_m}...\xi_1^{\alpha_1}|0^w\rangle\langle0^w| (\xi_1^{\alpha_1})^\dagger...(\xi_m^{\alpha_m})^\dagger
\otimes|\alpha\rangle\langle\alpha| +(1-F)\frac{I^{\otimes n}}{2^n}. \end{eqnarray*} The acceptance probability $p_{acc}'$ is \begin{eqnarray*} p_{acc}'=\frac{Fp_{acc}^2}{2^m}+\frac{1-F}{2^n}. \end{eqnarray*}
Assume that there exists a classical $poly(n)$-time probabilistic algorithm that accepts with probability $q_{acc}$ such that
$|p_{acc}'-q_{acc}|\le \epsilon p_{acc}'$, where $\epsilon<1$ is a constant. Then, if $x\in L$, \begin{eqnarray*} q_{acc}&\ge&(1-\epsilon)p_{acc}'\\ &=&(1-\epsilon)\Big(\frac{Fp_{acc}^2}{2^m}+\frac{1-F}{2^n}\Big)\\ &\ge&(1-\epsilon)F2^{-m}(1-2^{-r})^2, \end{eqnarray*} and if $x\notin L$, \begin{eqnarray*} q_{acc}&\le&(1+\epsilon)p_{acc}'\\ &=&(1+\epsilon)\Big(\frac{Fp_{acc}^2}{2^m}+\frac{1-F}{2^n}\Big)\\ &\le&(1+\epsilon)\Big(F2^{-2r-m}+\frac{1-F}{2^n}\Big)\\ &=&2^{-m}(1+\epsilon)F\Big(2^{-2r}+\frac{1-F}{F2^w}\Big). \end{eqnarray*} If $r$ and $w$ are sufficiently large, $L$ is in SBP. \fbox
Note that although here we have considered constant $F$, the same result also holds for other ``not so small" $F$ such as $F=\frac{1}{poly(m)}$.
\section{Proof of Theorem~\ref{theorem:result2}} Let us take $\epsilon=F2^{n+2}$. For any $z\in\{0,1\}^n$, \begin{eqnarray*}
\Big|p_z'-\frac{1}{2^n}\Big|
=\Big|\Big(Fp_z+\frac{1-F}{2^n}\Big)
-\frac{1}{2^n}\Big|\le F\Big(1+\frac{1}{2^n}\Big)< \epsilon p_z', \end{eqnarray*} where in the last inequality, we have used \begin{eqnarray*} \epsilon p_z'-F\Big(1+\frac{1}{2^n}\Big) &=& \epsilon\Big(Fp_z+\frac{1-F}{2^n}\Big)-F\Big(1+\frac{1}{2^n}\Big)\\ &\ge&\frac{\epsilon(1-F)}{2^n}-F\Big(1+\frac{1}{2^n}\Big)\\ &=&\frac{F2^{n+2}(1-F)}{2^n}-F\Big(1+\frac{1}{2^n}\Big)\\ &=&4F(1-F)-F\Big(1+\frac{1}{2^n}\Big)\\ &>0&. \end{eqnarray*} \fbox
\section{Proof of Theorem~\ref{theorem:result3}} \begin{eqnarray*}
\sum_{z\in\{0,1\}^n}\Big|p_z'-\frac{1}{2^n}\Big|
= F\sum_{z\in\{0,1\}^n}\Big|p_z-\frac{1}{2^n}\Big| \le 2F. \end{eqnarray*} \fbox
\acknowledgements
TM is supported by MEXT Quantum Leap Flagship Program (MEXT Q-LEAP) Grant Number JPMXS0118067394, JST PRESTO No.JPMJPR176A, and the Grant-in-Aid for Young Scientists (B) No.JP17K12637 of JSPS. YT is supported by MEXT Quantum Leap Flagship Program (MEXT Q-LEAP) Grant Number JPMXS0118067394.
\end{document}
|
arXiv
|
{
"id": "1911.02220.tex",
"language_detection_score": 0.6854119896888733,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
\begin{document}
\title{Hopf bifurcation and heteroclinic cycles in a class of $\mathbb{D}
\begin{abstract} In this paper we analyze a generic dynamical system with $\mathbb{D}_2$ constructed via a Cayley graph. We study the Hopf bifurcation and find conditions for obtaining a unique branch of periodic solutions. Our main result comes from analyzing the system under weak coupling, where we identify the conditions for heteroclinic cycle between four equilibria in the two-dimensional fixed point subspace of some of the isotropy subgroups of $\mathbb{D}_2\times\mathbb{S}^1.$ We also analyze the stability of the heteroclinic cycle. \end{abstract}
\begin{keywords} equivariant dynamical system, Cayley graph, Hopf bifurcation, heteroclinic cycle. \end{keywords}
\begin{AMS} 37C80, 37G40, 34C15, 34D06, 34C15 \end{AMS}
\pagestyle{myheadings} \thispagestyle{plain} \markboth{Adrian C. Murza}{Hopf bifurcation and heteroclinic cycles in a class of $\mathbb{D}_2-$equivariant systems} \section{Introduction} The global dynamics of networks of $n$ coupled oscillators with different types of coupling has been studied in \cite{Ashwin_Swift}. In their formalism it is shown that the symmetry group of the network can be considered a subgroup of $\mathbb{S}_n$, as long as the oscillators taken individually have no internal symmetries. These ideas have been investigated in the above cited paper, for general $\mathbb{D}_n$ and $\mathbb{Z}_n$ cases. However, it is a routine considering $n\geqslant3$ so the case $\mathbb{D}_2$ is sometimes not explicitly taken into account. Another reason for carrying out our study is the fact that among the dihedral groups, $\mathbb{D}_2$ is the only abelian group, which makes its action and therefore its analysis slightly different.
In this paper we are concerned with two properties of networks with $\mathbb{D}_2$ symmetry: Hopf bifurcation and low coupling case leading to heteroclinic cycles. Firstly we use the methodology developed by Ashwin and Stork \cite{Stork} to construct a network of differential systems with $\mathbb{D}_2$ symmetry. Our case is a particular situation of a network of $n$ coupled oscillators with symmetry to be a subgroup of $\mathbb{S}_n,$ that analyzes different types of coupling between the oscillators, as shown in \cite{Stork}. While this approach has been successfully used for the subgroups $\mathbb{Z}_n$ and $\mathbb{D}_n,$ our approach is interesting not only because of the particularities already mentioned of $\mathbb{D}_2,$ but also because it offers the possibility of analyzing the weak coupling limit, where dynamics of the network is governed only by the phases of the oscillators.\\
A large variety of subgroups of $\mathbb{S}_n$ can be easily generated by the method described by Ashwin and Stork \cite{Stork} based on graph theory. The automorphism group of the colored graph can be defined to be its symmetry group. Therefore, it is possible, via Cayley graphs, to design oscillatory networks with the prescribed symmetry of a subgroup of $\mathbb{S}_n.$ While the Hopf bifurcation is in fact a simple adaptation of the theory developed by Golubitsky, Stewart and Schaeffer in \cite{GS85},\cite{GS86} and \cite{GS88}, the low coupling case is much more interesting since it allows the existence of heteroclinic cycles in systems with $\mathbb{D}_2$ symmetry. In this case it is possible to reduce the asymptotic dynamics to a flow on an four-dimensional torus $\mathbb{T}^4;$ by assuming a weak coupling we average the whole network and introduce an extra $\mathbb{S}^1$ symmetry.
We determine the two-dimensional invariant subspaces on this torus and show that a heteroclinic cycle consisting of four one-dimensional routes among the four zero-dimensional equilibria can appear. We then apply the general stability theory for heteroclinic cycles developed by Krupa and Melbourne in \cite{Krupa}, to analyze the stability of the heteroclinic cycles.
\section{The Cayley graph of the $\mathbb{D}_2$ group}\label{section Cayley} Our aim is to construct an oscillatory system with the $\mathbb{D}_2$ symmetry. To achieve this goal we need first to represent the group by a Cayley diagram. A Cayley diagram is a graph (that is, a set of nodes and the arrows between them) to represent a group. Vertex or nodes of the graph are the group elements and the arrows show how the generators act on the elements of the group. At any vertex there are arrows pointing towards it, others pointing away. Proceeding as in \cite{ADSW} or \cite{Stork}, let $J\subset\mathbb{D}_2$ be the generating set of $\mathbb{D}_2.$ This implies that \begin{itemize} \item[(a)] $J$ generates $\mathbb{D}_2,$ \item[(b)] $J$ is finite and \item[(c)] $J=J^{-1}.$ \end{itemize} Since $\mathbb{D}_2$ is finite we can forget assumptions $(b)$ and $(c).$ Then the Cayley graph of $\mathbb{D}_2$, is a directed colored graph whose arrows indicate the action the group elements have on each other and the vertices are the group elements. As shown in \cite{ADSW} the generating set $J$ is a set of "colors" of the directed edges and the group elements $\sigma_i$ and $\sigma_j$ are connected through an edge from $\sigma_i$ to $\sigma_j$ of color $c\in J$ if and only if $\sigma_i=c\sigma_j.$
\begin{figure}
\caption{A Cayley graph of the $\mathbb{D}_2$ group. Solid arrows represent left-multiplication with $\kappa,$ dot-and-dashed arrows left multiplication with $\zeta,$ the two generators of this group.}
\label{second_figure}
\end{figure}
The Cayley graphs for $\mathbb{D}_2$ is shown in Figure \eqref{second_figure}. Following the approach by Stork et al. \cite{Stork}, we may identify the vertices as cells with a certain dynamics and the edges as couplings between the cells. In this way we can construct an ODE system which has the symmetry of $\mathbb{D}_2$. The action of the group $\mathbb{D}_2$ on the cells can be written as \begin{equation}\label{elements} \begin{array}{l} \kappa=(1~2)(3~4)\in\mathbb{S}_4,\\ \zeta=(1~3)(2~4)\in\mathbb{S}_4, \end{array} \end{equation} where the two generators and their commuting product $\kappa\zeta=\zeta\kappa$ act on $(x_1,x_2,x_3,x_4)\in\mathbb{R}^4$ as \begin{equation}\label{commutators} \begin{array}{l} \kappa:(x_1,x_2,x_3,x_4)\rightarrow(x_2,x_1,x_4,x_3)\\ \zeta:(x_1,x_2,x_3,x_4)\rightarrow(x_3,x_4,x_1,x_2)\\ \kappa\zeta:(x_1,x_2,x_3,x_4)\rightarrow(x_4,x_3,x_2,x_1) \end{array} \end{equation}
If we assign coupling between cells related by the permutations in \eqref{elements}, we can build the following pairwise system in with the $\mathbb{D}_2$ symmetry.
\begin{equation}\label{array 4 eq} \begin{array}{l} \dot{x}_1=f(x_1)+g(x_2,x_1)+h(x_3,x_1)\\ \dot{x}_2=f(x_2)+g(x_1,x_2)+h(x_4,x_2)\\ \dot{x}_3=f(x_3)+g(x_4,x_3)+h(x_1,x_3)\\ \dot{x}_4=f(x_4)+g(x_3,x_4)+h(x_2,x_4)\\ \end{array} \end{equation} where $f:\mathbb{R}\rightarrow\mathbb{R}$ and $g,~h:\mathbb{R}^2\rightarrow\mathbb{R}$. As shown by Ashwin and Stork \cite{Stork} we can think of $f,~g,~h$ as being generic functions that assure that the isotropy of this vector field under the action of $\mathbb{O}_4$ is generically $\mathbb{D}_2$.
\section{Hopf bifurcation}\label{section Hopf bifurcation}
The group $\mathbb{D}_2$ has order $4$, the generators being $(\kappa,~\zeta).$ Since all irreducible representations of the $\mathbb{D}_2$ group are one dimensional, we restrict our study to the actions of the generators on $z\in\mathbb{C}\equiv\mathbb{R}^2.$ The orthogonal representation of $\mathbb{D}_2$ $$R_{\mathbb{D}_2}:\mathbb{D}_2\times W\rightarrow W$$ on the complex vector space $W$ is irreducible if and only if the only $\mathbb{D}_2-$invariant subspaces of $W$ are the trivial ones. In fact it can be shown that $$\mathbb{D}_2\times\mathbb{S}^1/\mathrm{ker}R_{\mathbb{D}_2\times\mathbb{S}^1}\equiv\mathbb{Z}_2\times\mathbb{Z}_2\times\mathbb{S}^1/\mathrm{ker}R_{\mathbb{Z}_2\times\mathbb{Z}_2\times\mathbb{S}^1}.$$ The group element $(\kappa,~\zeta)$ acts on the point $z\in\mathbb{C}$ by \begin{equation}\label{action on C} \begin{array}{l} \kappa\cdot z=\bar{z}\\ \zeta\cdot z=\pm z. \end{array} \end{equation} The orbit of a point $(x,y)\in\mathbb{R}^2$ under the action of $\mathbb{D}_2$ is
$$\{(\kappa,\zeta)\cdot z|(\kappa,\zeta)\in\mathbb{D}_2\}.$$ Therefore the orbits are \begin{itemize} \item[(a)] The origin, $(0,0),$ \item[(b)] Points on the $x$-axis, $\pm x,0$ with $(x\neq0),$ \item[(c)] Points on the $y$-axis, $0,\pm y$ with $(y\neq0),$ \item[(d)] Points off the axes $\pm x,\pm y$ with $(x\neq0,~y\neq0).$ \end{itemize}
The isotropy subgroup of a point $(x,y)\in\mathbb{R}^2$ under the action of $\mathbb{D}_2$ is
$$\{(\kappa,~\zeta)\in\mathbb{D}_2|(\kappa,~\zeta)\cdot(x,y)=(x,y)\}.$$ $\mathbb{D}_2$ has four isotropy subgroups: \begin{itemize} \item[(a)] $\mathbb{D}_2$ corresponding to the origin \item[(a)] $\mathbb{Z}_2=\{(1,\zeta)\}$ corresponding to $(x,0)$ with $x\neq0,$ \item[(a)] $\mathbb{Z}_2=\{(\kappa,1)\}$ corresponding to $(0,y)$ with $y\neq0,$ \item[(a)] $\mathbbm{1}$ corresponding to $(x,y)$ with $x\neq0,~y\neq0.$ \end{itemize}
In this section we basically recall the theory of Golubitsky and Stewart \cite{GS85} and \cite{GS88} on one parameter Hopf bifurcations with symmetry.
\subsection{General considerations on Hopf bifurcation with $\mathbb{D}_2\times\mathbb{S}^1-$symmetry}
The group $\mathbb{D}_2$ only has four $1-$dimensional real irreducible representations and so they are absolutely irreducible. One of them is the trivial representation and the other three are non-isomorphic each with a nontrivial $\mathbb{Z}_2$ action. Therefore, the only possible way to have a Hopf bifurcation is for the linearization at the equilibrium to be from a $\Gamma-$simple subspace of the form $V\oplus V$ where $V$ is an absolutely irreducible representation which has either a trivial or a $\mathbb{Z}_2$ action. Generically, as a parameter is varied, the critical eigenspace with purely imaginary eigenvalues is two-dimensional of the form $V\oplus V$ described above. The periodic solution bifurcating is unique and has the same symmetry has the $\mathbb{Z}_2$ action on V, or has no symmetry.
In fact, the $\mathbb{D}_2$ symmetric coupled system does not support any periodic solution. One can easily show that the isotypic decomposition at a trivial steady-state is made up of the trivial representation $(1,1,1,1)$ and the complement decomposes into the remaining three non-isomorphic irreducible representations. Therefore, the $V\oplus V$ subspace cannot exist in the tangent space at the trivial equilibrium. Each cell in the network needs to be two-dimensional to hope to have a Hopf bifurcation with $\mathbb{D}_2$ symmetry in such a $4-$cell system.
\begin{priteo} For the $\mathbb{D}_2-$equivariant Hopf bifurcation in $\mathbb{C}$, there is a unique branch of periodic solutions consisting of rotating waves with $\mathbb{Z}_2$ spatial symmetry. \end{priteo} \begin{proof} First, unicity of a branch of periodic solutions is based on the fact that $\mathbb{D}_2$ is abelian, and we recall the Abelian Hopf Theorem in \cite{AbelianHopf}. Therefore, the periodic solution predicted by the Equivariant Hopf Theorem \cite{GS88} with the symmetry of the unique isotropy subgroup with a two-dimensional fixed point subspace is the only one. \end{proof}\\
\section{Normal form and generic linearized stability} We begin with recalling results on the normal form for the $\mathbb{D}_2-$equivariant bifurcation from \cite{GS85} on the $\mathbb{Z}_2\times\mathbb{Z}_2$ group, which is isomorphic to $\mathbb{D}_2.$ This means that we can adapt the theory of $\mathbb{Z}_2\times\mathbb{Z}_2$ to our case, withe merely a re-interpretation of the branches. We say that the bifurcation problem $g$ commutes with $\mathbb{D}_2$ if \begin{equation}\label{commut} g((\kappa,\zeta)\cdot(x,y),\lambda)=(\kappa,\zeta)\cdot g(x,y,\lambda). \end{equation} \begin{lema}\label{lema normal form} Let us consider the $g:\mathbb{C}\times\mathbb{R}\rightarrow\mathbb{R}$ bifurcation problem with $z=x+iy$ commuting with the action of $\mathbb{D}_2$. Then there exist smooth functions $p(u,v,\lambda),~q(u,v,\lambda)$ such that\\ \begin{equation}\label{eq lema normal form} \begin{array}{l} g(x,y,\lambda)=(p(x^2,y^2,\lambda)x,~q(x^2,y^2,\lambda)y),~~\mathrm{where}\\ \\ \hspace{2cm}p(0,0,0)=0,~~~q(0,0,0)=0. \end{array} \end{equation} \end{lema}
\begin{proof} The proof is almost identical to the one for $\mathbb{Z}_2\times\mathbb{Z}_2$ in \cite{GS85} but we prefer to give it adapted to the action of the group generators of $\mathbb{D}_2.$ We write $g$ as \begin{equation}\label{eq lema g} g(x,y,\lambda)=(a(x,y,\lambda),~b(x,y,\lambda)). \end{equation} Commutativity with equation \eqref{commut} implies \begin{equation}\label{eq lema normal form2} \begin{array}{l} a(\kappa x,\zeta y,\lambda)=\kappa a(x,y,\lambda)\\ \\ b(\kappa x,\zeta y,\lambda)=\zeta b(x,y,\lambda) \end{array} \end{equation} Now $\kappa$ transforms $z$ into $\bar{z}$, i.e. $(x,y)\rightarrow(x,-y)$ and if the action of $\zeta$ is $+1$ then we get that $a$ is odd in $x$ while $b$ is even in $x.$ When $\kappa$ acts as identity and $\zeta$ acts as $-1$ then $a$ is even in $y$ and $b$ is odd in $y.$ Now the rest of the proof is exactly the same as in \cite{GS85}. \end{proof}\\ \\ In the following we will follow again the route and results in \cite{GS85} to obtain the Jacobian matrix of equation \eqref{eq lema normal form} \begin{equation}\label{equation dg} dg= \begin{array}{l} \begin{bmatrix} p+2up_u&2p_vxy\\ 2q_uxy&q+2vq_v \end{bmatrix} \end{array}. \end{equation} What is more, we can reformulate or adapt to our case Lemma $3.1$ (page $427$ in \cite{GS85}) and add the proof again corresponding to our case, which in the cited reference has been left as an exercise for the reader. \begin{lema}\label{lema trivial} Let $(x,y,\lambda)$ be a solution to $g=0.$ \begin{itemize} \item[(a)] If $f(x,y,\lambda)$ is a trivial or a pure mode solution, then $dg$ in \eqref{equation dg} is diagonal and its eigenvalues which are real have the signs listed as follows:
\begin{itemize}
\item [(i)] Trivial solution: $\mathrm{sgn}~p(0,0,\lambda),~\mathrm{sgn}~q(0,0,\lambda);$
\item [(ii)] $x-$mode solution: $\mathrm{sgn}~p_u(x,0,\lambda),~\mathrm{sgn}~q(x,0,\lambda);$
\item [(iii)] $y-$mode solution: $\mathrm{sgn}~p(0,y,\lambda),~\mathrm{sgn}~q_v(0,y,\lambda);$
\end{itemize} \item[(b)] If $f(x,y,\lambda)$ is a mixed mode solution, then \begin{itemize} \item[(a)] $\mathrm{sgn~det}(dg)=\mathrm{sgn}(p_uq_v-p_vq_u)$ at $(x,y,\lambda);$ \item[(b)] $\mathrm{sgn~tr}(dg)=\mathrm{sgn}(up_u+vq_u)$ at $(x,y,\lambda).$ \end{itemize} \end{itemize} \end{lema} \begin{proof} By using the terminology in \cite{GS85}, we have \begin{itemize} \item[(a)] trivial solution when $x=y=0,$ \item[(b)] $x-$mode solution when $p(u,0,\lambda)=y=0,~x\neq0,$ \item[(c)] $y-$mode solution when $q(0,v,\lambda)=x=0,~y\neq0,$ \item[(d)] mixed-mode solution when $p(u,v,\lambda)=q(u,v,\lambda)=0,~x\neq0,~y\neq0.$ \end{itemize} Accordingly, if $x=y=0$ equation \eqref{equation dg} reduces to \begin{equation}\label{equation dg1} dg= \begin{array}{l} \begin{bmatrix} p&0\\ 0&q \end{bmatrix} \end{array} \end{equation} in the first case. Moreover, since $y=0$ for the $x-$mode solution, then $v=0,~x\neq0$ and therefore the Jacobian matrix becomes \begin{equation}\label{equation dg2} dg= \begin{array}{l} \begin{bmatrix} 2up_u&0\\ 0&q \end{bmatrix} \end{array}. \end{equation} Similarly, since $x=0$ for the $y-$mode solution, then $u=0,~y\neq0$ and therefore the Jacobian matrix becomes \begin{equation}\label{equation dg4} dg= \begin{array}{l} \begin{bmatrix} p&0\\ 0&2vq_v \end{bmatrix} \end{array}. \end{equation}
Finally, when we deal with the mixed-mode solutions then the Jacobian matrix becomes \begin{equation}\label{equation dg5} dg= \begin{array}{l} \begin{bmatrix} 2up_u&2p_vxy\\ 2q_uxy&2vq_v \end{bmatrix} \end{array}, \end{equation}
and the eigenvalues are \begin{equation}\label{eigenvalues normal form} \begin{array}{l} \lambda_{1,2}=\displaystyle{\frac{2up_u+2vq_v\pm2\sqrt{u^2p_u^2+6uvp_uq_v+v^2q_v^2}}{2}}=\\ \\ \hspace{3cm}up_u+vq_v\pm\sqrt{\left(up_u+vq_v\right)^2+4uvp_uq_v}. \end{array} \end{equation} Now since $u=x^2$ and $v=y^2$ we explicitly derive the conditions for stability/unstability of the linearized system, in Table \eqref{table sign jacobian}. \begin{table} \centering \begin{center} \caption{Mixed-mode solutions of the $\mathbb{D}_2-$equivariant normal of equation \eqref{eq lema normal form}.}\label{table sign jacobian}
\end{center} \begin{tabular}{cccc} \toprule
$(p_u,q_v)$ & $up_u+vq_v$& $\left(up_u+vq_v\right)^2+4uvp_uq_v$&Stability \\ \midrule \centering $(+,+)$ &+&+&unstable\\ $(-,-)$ &-&+&stable\\ $(+,-)$& +&+&unstable\\ $(+,-)$& +&-&unstable\\ $(+,-)$& -&+&stable\\ $(+,-)$& -&-&stable\\ $(-,+)$& +&+&unstable\\ $(-,+)$& +&-&unstable\\ $(-,+)$& -&+&stable\\ $(-,+)$& -&-&stable\\ \bottomrule \end{tabular} \end{table} \end{proof}
\section{Weak Coupling}\label{section Weak Coupling}
We can think on ideal physical systems as interactions between identical or nearly identical subsystems. Moreover, as in \cite{Ashwin_Swift}, the whole system can be described as being a perturbation of an uncoupled system. This can be thought of as continuous path between the system and a system that is a product of several dynamical systems. In the following we will assume that our system of oscillators has a weak coupling limit. In fact, as it has been shown in Ashwin and Swift \cite{Ashwin_Swift}, even strongly coupled oscillator systems must have a weakly coupled limit. Moreover, we will assume that our system is formed by dissipative oscillators, so that the periodic orbit is attracting and unique in some neighborhood. In the weak coupling case, we focus on the dynamics of the relative phases of the oscillators. This situation can be better understood when in the no coupling case there is an attracting $N-$torus with one angle for each oscillator. An apparent problem can arise when considering phase differences; however we can choose coordinates so that the dynamics is a linear flow in the direction and there is no change in the phase differences. The theory developed in \cite{Ashwin_Swift} shows that the torus is normally hyperbolic, so with small coupling, the torus persists and there is a slow evolution of the phase differences. Moreover, it has been pointed out that the weak coupling limit offers different information about the dynamics rather than analyzing only the small amplitude periodic solutions near the Hopf bifurcation point.
Our system \eqref{array 4 eq} can be rewritten under weak coupling case as an ODE of the form: \begin{equation}\label{generic weak coupling equation} \dot{x}_i=f(x_i)+\epsilon g_i(x_1,\ldots,x_4) \end{equation} for $i=1,\ldots,4,~x_i\in \mathcal{X}$ and commuting with the permutation action of $\mathbb{D}_2$ on $\mathcal{X}^4,$ both $f$ and $g_i$ being of the class $\mathcal{C}^{\infty}.$ The constant $\epsilon$ represents the coupling strength and it is assumed to have low values. As in \cite{Ashwin_Swift}, \cite{ADSW} or \cite{Stork} we assume $\dot{x}$ has an hyperbolic stable limit cycle.\\
As shown in \cite{Ashwin_Swift} in the case of weak coupling, there is a natural reason why we should not just look at irreducible representations of $\mathbb{D}_2$. In our case there are $4$ stable hyperbolic limit cycles in the limit of $\epsilon=0,$ which means that the asymptotic dynamics of the system factors into the asymptotic dynamics of four limit cycles. This way as it we show later, we can for example embed the flow of a $2-$dimensional torus on a four-dimensional torus $\mathbb{T}^4.$ Moreover we assume hyperbolicity of the individual limit cycles for small enough values of the coupling parameter, and this justifies expressing the dynamics of the system as an ODE in terms of four phases, i.e. an ODE on $\mathbb{T}^4$ which is $\mathbb{D}_2-$equivariant. \begin{table} \centering \begin{center} \caption{Isotropy subgroups and fixed point subspaces for the $\mathbb{D}_2\times\mathbb{S}^1$ action on $\mathbb{T}^4$. The generators are $\kappa=\{(0~1)(2~3)\}$ and $\zeta=\{(0~2)(1~3)\}.$ The isotropy subgroups act by reflection about the zero, one and two-dimensional manifolds on $\mathbb{T}^4.$ For example, by notation $\mathbb{Z}^{\phi}_2(\kappa_{0,0})$ we mean the group $\mathbb{Z}_2$ acting through reflection about the circle $x_0=x_1=0,x_2=x_3=\phi,$ while $\mathbb{Z}^{\phi_i}_2(\kappa_{0,0})$ whith $i=\{1,2\}$ acts by reflection about the disk $x_0=x_1=0,x_2=\phi_1,x_3=\phi_2$.$~(\phi,~\phi_1,~\phi_2\in\left[0,\pi\right])$.}\label{table grande}
\end{center} \begin{tabular}{ccccc} \toprule
$\Sigma$ & $\mathrm{Fix}(\Sigma)$ & Generators & $\mathrm{dim~Fix}(\Sigma)$\\ \midrule \centering $\mathbb{D}_2(0)$ & $(0,0,0,0)$ & $\{\kappa,~\zeta\}_{(0)}$ &0\\ $\mathbb{D}_2(\kappa,\mathrm{Id})$ & $(0,0,\pi,\pi)$ & $\{\kappa,~\mathrm{Id}\}_{(0,\pi)}$ &0\\ $\mathbb{D}_2(\kappa\zeta,\kappa)$ & $(0,\pi,0,\pi)$ & $\{\kappa\zeta,~\kappa\}_{(0,\pi)}$ &0\\ $\mathbb{D}_2(\pi)$& $(\pi,0,0,\pi)$ & $\{\kappa,~\zeta\}_{(\pi)}$ &0\\ $\mathbb{Z}^{\phi}_2(\kappa_{(0,0)})$ & $(0,0,\phi,\phi)$ & $\{(0~1)\}_{(0,0)}$ &1\\ $\mathbb{Z}^{\phi}_2(\kappa_{(\pi,\pi)})$ & $(\pi,\pi,\phi,\phi)$ & $\{(0~1)\}_{(\pi,\pi)}$ &1\\ $\mathbb{Z}^{\phi}_2(\zeta_{(\pi,\pi)})$ & $(\phi,\phi,\pi,\pi)$ & $\{(2~3)\}_{(\pi,\pi)}$ &1\\ $\mathbb{Z}^{\phi}_2(\zeta_{(0,0)})$ & $(\phi,\phi,0,0)$ & $\{(2~3)\}_{(0,0)}$ &1\\ $\mathbb{Z}^{\phi_i}_2(\kappa_{(0,0)})$ & $(0,0,\phi_1,\phi_2)$ & $\{(0~1)\}_{(0,0)}$ &2\\ $\mathbb{Z}^{\phi_i}_2(\kappa_{(\pi,\pi)})$ & $(\pi,\pi,\phi_1,\phi_2)$ & $\{(0~1)\}_{(\pi,\pi)}$ &2\\ $\mathbb{Z}^{\phi_i}_2(\zeta_{(\pi,\pi)})$ & $(\phi_1,\phi_2,\pi,\pi)$ & $\{(2~3)\}_{(\pi,\pi)}$ &2\\ $\mathbb{Z}^{\phi_i}_2(\zeta_{(0,0)})$ & $(\phi_1,\phi_2,0,0)$ & $\{(2~3)\}_{(0,0)}$ &2\\ \bottomrule \end{tabular} \end{table}
In addition, Ashwin and Swift showed in \cite{Ashwin_Swift} that for small enough values of the coupling parameter it is possible to average the equations and introduce an approximate decoupling between the fast variation of the phases and the slow variation of the phase differences. This can be seen as introducing and phase shift symmetry which acts on $\mathbb{T}^4$ by translation along the diagonal; $$R_{\theta}(\phi_1,\ldots,\phi_4):=(\phi_1+\theta,\ldots,\phi_4+\theta),$$ for $\theta\in\mathbb{S}^1.$
Now have an ODE on that is equivariant under the action of $\mathbb{D}_2\times\mathbb{S}^1;$ now we have to classify the isotropy types of points under this action.
\begin{priteo} The isotropy subgroups for the action of $\mathbb{D}_2\times\mathbb{S}^1$ on $\mathbb{T}^4$ together with their generators and dimension of their fixed-point subspaces are those listed in Table \eqref{table grande}. \end{priteo} \begin{proof} We will explicitly calculate two examples, for the zero and one-dimensional fixed-point subspaces, respectively, the other cases being treated similarly. $(a)$ Let's take the action of $\mathbb{D}_2(\kappa,\mathrm{Id})$ on $\mathbb{T}^4.$ We have \begin{equation}\label{teorema tabla example} \begin{array}{l} \begin{bmatrix} \cos\phi_1&-\sin\phi_1&0&0\\ \sin\phi_1&\cos\phi_1&0&0\\ 0&0&\cos\phi_2&-\sin\phi_2\\ 0&0&\sin\phi_2&\cos\phi_2\\ \end{bmatrix} \begin{bmatrix} a\\b\\c\\delta \end{bmatrix}= \begin{bmatrix} a\cos\phi_1-b\sin\phi_1\\ b\cos\phi_1+a\sin\phi_1\\ c\cos\phi_2-d\sin\phi_2\\ d\cos\phi_2+c\sin\phi_2\\ \end{bmatrix}= \begin{bmatrix} a\cos\phi\\ b\cos\phi\\ c\cos\phi\\ d\cos\phi\\ \end{bmatrix}= \begin{bmatrix} \pm a\\ \pm b\\ \pm c\\ \pm d\\ \end{bmatrix}. \end{array} \end{equation} because $\phi_1,\phi_2=\{0,\pi\}.$ Therefore the only possible values for any arbitrary point on $\mathbb{T}^4$ to be fixed by the group are $0$ and $\pi.$ The four choices for the first four lines are deduced from the action of the elements of $\mathbb{D}_2.$\\
$(b)$ Let's take for example to action of $\mathbb{Z}^{\phi}_2(\zeta_{(\pi,0)})$ on one dimensional manifolds on $\mathbb{T}^4.$ \begin{equation}\label{teorema tabla example2} \begin{array}{l} \begin{bmatrix} 1&0&0&0\\ 0&1&0&0\\ 0&0&\cos\phi&-\sin\phi\\ 0&0&\sin\phi&\cos\phi\\ \end{bmatrix} \begin{bmatrix} \phi\\ \phi\\ \pi\\ \pi\\ \end{bmatrix}= \begin{bmatrix} \phi\\ \phi\\ \pi\cos\phi\\ \pi\cos\phi\\ \end{bmatrix}= \begin{bmatrix} \phi\\ \phi\\ \pm\pi\\ \pm\pi \end{bmatrix}= \begin{bmatrix} \phi\\ \phi\\ \pi\\ \pi\\ \end{bmatrix}, \end{array} \end{equation} because the only options for $\phi$ are $0$ or $\pi$ radians.
\end{proof}
\subsection{Analysis of a family of vector fields in $\mathrm{Fix}(\mathbb{Z}_2)$}\label{onetorus_theor} We can define coordinates in $\mathrm{Fix}(\mathbb{Z}_2) $ by taking a basis \begin{equation}\label{basis} \begin{array}{l} e_1=\frac{1}{2}(-1,1,-1,-1)\\ e_2=\frac{1}{2}(-1,-1,1,-1) \end{array} \end{equation} and consider the space spanned by $\{e_1,e_2\}$ parameterized by $\{\phi_1,\phi_2\}:$ \begin{equation}\label{coordinates} \sum_{n=1}^2\phi_ne_n \end{equation} By using these coordinates, we construct the following family of two-dimensional differential systems which satisfies the symmetry of $\mathrm{Fix}(\mathbb{Z}_2)$. \begin{equation}\label{systema ejemplo} \left\{ \begin{array}{l} \dot{\phi_1}=a\sin{\phi_1}\cos{\phi_2}+\epsilon\sin{2\phi_1}\cos{2\phi_2}\\ \\ \dot{\phi_2}=-b\sin{\phi_2}\cos{\phi_1}+\epsilon\sin{2\phi_2}\cos{2\phi_1}+q(1+\cos\phi_1)\sin2\phi_2,\\ \end{array} \right. \end{equation} where $a,~b>0.$ We argue that the family of vector fields \eqref{systema ejemplo} exhibits structurally stable, attracting heteroclinic cycles, which are structurally stable or completely unstable, depending on parameters $a,~b,~\epsilon$ and $q,$ as we shall prove in Theorem \ref{teorema estabilidad heteroclinas}. In the following we will show that the planes $\phi_1=0~(\mathrm{mod}~\pi),~\phi_2=0~(\mathrm{mod}~\pi)$ are invariant under the flow of \eqref{systema ejemplo}.
Let $\mathcal{X}$ be the vector field of system \eqref{systema ejemplo}.\\ \begin{defi} We call a trigonometric invariant algebraic curve $h(\phi_1,\phi_2)=0,$ if it is invariant by the flow of \eqref{systema ejemplo}, i.e. there exists a function $K(\phi_1,\phi_2)$ such that \begin{equation}\label{campo} \mathcal{X}h=\frac{\partial h}{\partial\phi_1}\dot{\phi_1}+\frac{\partial h}{\partial\phi_2}\dot{\phi_2}=Kh. \end{equation} \end{defi}
\begin{lema}\label{lema invariant2} Functions $\sin\phi_1$ and $\sin\phi_2$ are trigonometric invariant algebraic curves for system \eqref{systema ejemplo}. \end{lema} \begin{proof} We can write the system \eqref{systema ejemplo} in the form \begin{equation}\label{systema ejemplo1} \left\{ \begin{array}{l} \dot{\phi_1}=\sin{\phi_1}\left(a\cos{\phi_2}+2\epsilon\cos{\phi_1}\cos{2\phi_2}\right)\\ \\ \dot{\phi_2}=\sin{\phi_2}\left(-b\cos{\phi_1}+2\epsilon\cos{\phi_2}\cos{2\phi_1}+2q(1+\cos\phi_1)\cos\phi_2\right)\\ \end{array} \right. \end{equation}
Now if we choose $h_1=\sin\phi_1,$ then $Xh_1=\cos{\phi_1}\sin{\phi_1}\left(a\cos{\phi_2}+2\epsilon\cos{\phi_1}\cos{2\phi_2}\right)$ so $K_1=\cos{\phi_1}\left(a\cos{\phi_2}+2\epsilon\cos{\phi_1}\cos{2\phi_2}\right).$ The second case follows similarly. \end{proof}\\
Since the planes $\phi_i=0(\mathrm{mod}~\pi),~i=1,2$ are invariant under the flow of \eqref{systema ejemplo}, it is clear that $(0,0),~(\pi,0),~(\pi,\pi),~(0,\pi)$ are equilibria for \eqref{systema ejemplo}. To check the possibility of heteroclinic cycles in system \eqref{systema ejemplo}, we linearize first about the equilibria (i.e. the zero-dimensional fixed points) and then about the one-dimensional manifolds connecting these equilibria. We can assume without loss of genericity that $\mathrm{Fix}\left(\mathbb{Z}_2\right)$ is attracting for the dynamics and therefore the stabilities. The idea is that the analysis of the dynamics within the fixed point space $\mathrm{Fix}\left(\mathbb{Z}_2\right)$ is crucial in determining the stabilities of the full system. In particular we will prove that eigenvalues of the linearization in each cases are of opposite signs, allowing the existence of such a heteroclinic network between the equilibria.
In the following we will show that there exists the possibility of a heteroclinic cycle in any of the two-dimensional fixed-point spaces in Table \eqref{table grande} and the connections between these zero dimensional fixed point spaces are possible within specific routes within the one-dimensional fixed points of \eqref{systema ejemplo}. For the proof we will need \begin{lema}\label{lema saddles}
Assume $|\epsilon|<\mathrm{min}\{\frac{a}{2},\frac{b}{2}\}$ and $|\epsilon+2q|<\frac{b}{2}.$ Then the four fixed points in Table \eqref{table fixed points} are saddles. \end{lema} \begin{proof} By inspection it is clear that the eigenvalues of these fixed points are of opposite signs in each case.
\begin{table} \centering \begin{center} \caption{Eigenvalues of the flow of equation \eqref{systema ejemplo}, at the four non-conjugate zero-dimensional fixed points.}\label{table fixed points}
\end{center} \begin{tabular}{cccc} \toprule
$\mathrm{Fix}(\Sigma)$ & $(\phi_1,\phi_2)$ & $\lambda_1$ & $\lambda_2$\\ \midrule \centering $\mathbb{D}_2(0)$& $(0,0)$ & $a+2\epsilon$ &$-b+2(\epsilon+2q)$\\ $\mathbb{D}_2(\kappa,\mathrm{Id})$ & $(0,\pi)$ & $-a+2\epsilon$ &$b+2(\epsilon+2q)$\\ $\mathbb{D}_2(\kappa\zeta,\kappa)$ & $(\pi,0)$ & $-a+2\epsilon$ &$b+2\epsilon$\\ $\mathbb{D}_2(\pi)$ & $(\pi,\pi)$ & $a+2\epsilon$ &$-b+2\epsilon$\\ \bottomrule \end{tabular} \end{table}
\end{proof} \begin{priteo} There exists the possibility of a heteroclinic cycle in any of the two-dimensional fixed point subspaces in Table \eqref{table grande} in the following way: \begin{equation}\label{flechas} \begin{array}{l} \cdots\xrightarrow{\mathbb{Z}^{\phi}_2(\zeta_{(0,0)})}\mathbb{D}_2(0)\xrightarrow{\mathbb{Z}^{\phi}_2(\kappa_{(0,0)})} \mathbb{D}_2(\kappa,\mathrm{Id})\xrightarrow{\mathbb{Z}^{\phi}_2(\zeta_{(\pi,\pi)})}\\ \\ \hspace{3cm}\mathbb{D}_2(\pi) \xrightarrow{\mathbb{Z}^{\phi}_2(\kappa_{(\pi,\pi)})}\mathbb{D}_2(\kappa\zeta,\kappa)\xrightarrow{\mathbb{Z}^{\phi}_2(\zeta_{(0,0)})}\cdots \end{array} \end{equation} where the connection between the four equilibria in the plane $\phi_1,\phi_2$ is carried out along the indicated one-dimensional manifolds. \end{priteo}
\begin{proof} From Lemma \eqref{lema saddles} we know that the four fixed points are saddles, so there is possible a heteroclinic connection between them. We linearize the system \eqref{systema ejemplo} at every point in the one-dimensional manifolds of the fixed point spaces in Table \eqref{table grande}. The Jacobian matrix of the system evaluated at these points has eigenvalues shown in Table \eqref{table fixed points repetition}. This way we obtain the four paths shown in Table \eqref{table fixed points}. Moreover, using the conditions for $\epsilon$ and $q$ in Lemma \eqref{lema saddles} these eigenvalues are clearly of opposite signs in each case. A schematic view of the heteroclinic cycle is offered in Figure \eqref{flechas repetition}. \begin{table} \centering \begin{center} \caption{Eigenvalues of the flow of equation \eqref{systema ejemplo}, at the four non-conjugate paths on the one-dimensional fixed-point subspaces on $\mathbb{T}^4$.}\label{table fixed points repetition} \end{center} \begin{tabular}{cccc} \toprule
$\mathrm{Fix}(\Sigma)$ & $(\phi_1,\phi_2)$ & $\lambda_1$ & $\lambda_2$\\ \midrule \centering $\mathbb{Z}^{\phi_1}_2(\kappa_{(0,0)})$ &$(\phi_1,0)$& $a\cos\phi_1+2\epsilon\cos2\phi_1$ &$-b\cos\phi_1+2q(1+\cos\phi_1)+2\epsilon\cos2\phi_1$\\ $\mathbb{Z}^{\phi}_2(\zeta_{(\pi,\pi)})$& $(\pi,\phi_2)$& $-a\cos\phi_2+2\epsilon\cos2\phi_2$ &$b\cos\phi_2+2\epsilon\cos2\phi_2$\\ $\mathbb{Z}^{\phi}_2(\kappa_{(\pi,\pi)})$ & $(\phi_1,\pi)$ &$-a\cos\phi_1+2\epsilon\cos2\phi_1$&$b\cos\phi_1+2q(1+\cos\phi_1)+2\epsilon\cos2\phi_1$\\ $\mathbb{Z}^{\phi}_2(\zeta_{(0,0)})$& $(0,\phi_2)$ & $a\cos\phi_2+2\epsilon\cos2\phi_2$ &$-b\cos\phi_2+2\epsilon\cos2\phi_2+4q\cos2\phi_2$\\ \bottomrule \end{tabular} \end{table}
\end{proof}
\begin{figure}
\caption{Schematic representation of a structurally stable heteroclinic cycle for the weak coupled oscillators within the two-dimensional space of the $\mathrm{Fix}(\mathbb{Z}_2)$ group.}
\label{flechas repetition}
\end{figure}
Under the conditions in Lemma \eqref{lema saddles}, the stability of the heteroclinic cycle is guaranteed by the fact that it is restricted to the dynamics of system \eqref{systema ejemplo} within the fixed point spaces. We use the criteria of Krupa and Melbourne \cite{Krupa} to study the stability of the heteroclinic cycle. We now recall these stability criteria, which are based on four hypotheses.\\ The first hypothesis guarantees that the heteroclinic cycle is robust. \begin{itemize} \item ($S_1$) There is an isotropy subgroup $\Sigma_j$ with the fixed-point subspace $P_j=\mathrm{Fix}(\Sigma_j)$ such that $W^u(\xi_j)\cap P_j\subset W^s(\xi_{j+1})$ and $\xi_{j+1}$ is a sink in $P_j.$\\ Corresponding to each isotropy subgroup $\Sigma_j$ in ($S_1$) is the isotypic decomposition $\mathbb{R}^n=W_0\oplus\ldots\oplus W_q$ of $\mathbb{R}^n$ into isotypic components. Let $W_0=P_j$ and $N(\Sigma_j)$ denote the normalizer of $\Sigma_j$ in $\Gamma.$\\ \item ($S_2$) The eigenspaces corresponding to $c_j,~t_j,~e_{j+1}$ and $t_{j+1}$ lie in the same $\Sigma_j-$isotypic component;\\ \item ($S_3$) $\mathrm{dim}W^u(\xi_j)\cap P_j=\mathrm{dim}(N(\Sigma_j)/\Sigma_j)+1;$\\ \item ($S_4$) All transverse eigenvalues of $\xi_j$ with positive real part lie in the same $\Sigma_j-$isotypic component. \end{itemize}
Set $\rho_j=\mathrm{min}(c_j/e_j,1-t_j/e_j)$ and define $\rho=\rho_1\cdot\cdot\cdot\rho_m.$
\begin{priteo} [Krupa, Melbourne] Let $\Gamma$ be a finite group acting on $\mathbb{R}^n$ and $f:\mathbb{R}^n\rightarrow\mathbb{R}^n$ be a $\Gamma-$equivariant vector field. Suppose that $X$ is a heteroclinic cycle for $f$ satisfying hypotheses $(S_1)-(S_4).$ Then generically the stability of $X$ is described by precisely one of the following possibilities. \begin{itemize} \item [(a)] asymptotically stable ($\rho>1$ and $t_j<0$ for each $j$), \item [(b)] unstable but essentially asymptotically stable ($\rho>1$ and $t_j<e_j$ for each $j$ and $t_j>0$ for some $j$), \item [(c)] almost completely unstable ($\rho<1$ or $t_j>e_j$ for some $j$), \item [(d)] completely unstable if $\rho<1.$ \end{itemize} \end{priteo}
Applying these criteria to our case, we have
\begin{priteo}\label{teorema estabilidad heteroclinas} Heteroclinic cycle \eqref{flechas} is: \begin{itemize} \item [(a)] asymptotically stable if \begin{equation}\label{eq1 teorema estabilidad} \begin{array}{l} -\mathrm{min}\{\frac{a}{2},\frac{b}{2}\}<\epsilon<0~~\mathrm{and}~~-\frac{b}{2}<\epsilon+2q<0,~~\mathrm{or}\\ \\ -\mathrm{min}\{\frac{a}{2},\frac{b}{2}\}<\epsilon<0~~\mathrm{and}~~0<\epsilon+2q<\frac{b}{2}~~\mathrm{and}\\ \\ \hspace{5cm}\displaystyle{\frac{b-2(\epsilon+2q)}{b+2(\epsilon+2q)}>\frac{(a+2\epsilon )^2(b+2\epsilon)}{(a-2\epsilon )^2(b-2\epsilon)}},~~\mathrm{or}\\ \\ 0<\epsilon<\mathrm{min}\{\frac{a}{2},\frac{b}{2}\}~~\mathrm{and}~~-\frac{b}{2}<\epsilon+2q<0~~\mathrm{and}\\ \\ \hspace{5cm}\displaystyle{\frac{b-2(\epsilon+2q)}{b+2(\epsilon+2q)}<\frac{(a+2\epsilon )^2(b+2\epsilon)}{(a-2\epsilon )^2(b-2\epsilon)}}; \end{array} \end{equation} \item [(b)] completely unstable if \begin{equation}\label{eq1 teorema estabilidad repetition} \begin{array}{l} 0<\epsilon<\mathrm{min}\{\frac{a}{2},\frac{b}{2}\}~~\mathrm{and}~~0<\epsilon+2q<\frac{b}{2},~~\mathrm{or}\\ \\ -\mathrm{min}\{\frac{a}{2},\frac{b}{2}\}<\epsilon<0~~\mathrm{and}~~0<\epsilon+2q<\frac{b}{2}~~\mathrm{and}\\ \\ \hspace{5cm}\displaystyle{\frac{b-2(\epsilon+2q)}{b+2(\epsilon+2q)}<\frac{(a+2\epsilon )^2(b+2\epsilon)}{(a-2\epsilon )^2(b-2\epsilon)}},~~\mathrm{or}\\ \\ 0<\epsilon<\mathrm{min}\{\frac{a}{2},\frac{b}{2}\}~~\mathrm{and}~~-\frac{b}{2}<\epsilon+2q<0~~\mathrm{and}\\ \\ \hspace{5cm}\displaystyle{\frac{b-2(\epsilon+2q)}{b+2(\epsilon+2q)}>\frac{(a+2\epsilon )^2(b+2\epsilon)}{(a-2\epsilon )^2(b-2\epsilon)}}. \end{array} \end{equation} \end{itemize} \end{priteo} \begin{proof}
The stability is expressed by \begin{equation}\label{stability krupa1} \rho=\prod_{j=1}^4\rho_j \end{equation} where \begin{equation}\label{stability krupa2} \rho_j=\mathrm{min}\{c_j/e_j,1-t_j/e_j\}. \end{equation} In equation \eqref{stability krupa2}, $e_i$ is the expanding eigenvector at the $i$th point of the cycle, $-c_i$ is the contracting eigenvector and $t_i$ is the tangential eigenvector of the linearization. For the heteroclinic cycle we have \begin{equation}\label{rho values1} \begin{array}{l} \rho_1=\displaystyle{\frac{b-2(\epsilon+2q)}{a+2\epsilon}}~~ \rho_2=\displaystyle{\frac{a-2\epsilon}{b+2(\epsilon+2q)}}~~ \rho_3=\displaystyle{\frac{a-2\epsilon}{b+2\epsilon}}~~ \rho_4=\displaystyle{\frac{b-2\epsilon}{a+2\epsilon}}, \end{array} \end{equation} so from equation \eqref{stability krupa1} we obtain \begin{equation}\label{rho values2} \begin{array}{l} \rho=\displaystyle{\frac{\left[b-2(\epsilon+2q)\right](a-2\epsilon )^2(b-2\epsilon)}{\left[b+2(\epsilon+2q)\right](a+2\epsilon )^2(b+2\epsilon)}}. \end{array} \end{equation} Then the proof follows by applying Theorem $2.4$ in \cite{Krupa}. \end{proof}\\ \paragraph{\bf Acknowledgements} The author would like to thank the referee for indications which improved the presentation of this paper. He also acknowledges financial support from FCT grant $SFRH/ BD/ 64374/ 2009.$
\end{document}
|
arXiv
|
{
"id": "1406.3856.tex",
"language_detection_score": 0.6383354067802429,
"char_num_after_normalized": null,
"contain_at_least_two_stop_words": null,
"ellipsis_line_ratio": null,
"idx": null,
"lines_start_with_bullet_point_ratio": null,
"mean_length_of_alpha_words": null,
"non_alphabetical_char_ratio": null,
"path": null,
"symbols_to_words_ratio": null,
"uppercase_word_ratio": null,
"type": null,
"book_name": null,
"mimetype": null,
"page_index": null,
"page_path": null,
"page_title": null
}
|
arXiv/math_arXiv_v0.2.jsonl
| null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.