-1.Motivation-
Large sieve a philosophy reflect as a large group of inequalities which is very effective on controlling some linear sum or square sum of some correlation of arithmetic function, some idea of which could have originated in harmonic analysis, merely rely on almost orthogonality.
One fundamental example is the estimate of the quality,
$latex \sum_{n\leq x}|\Lambda(n)\overline{\chi(n)}|$
One naive idea of control this quality is using Cauchy-schwarz inequality. But stupid use this we gain something even worse than trivial estimate. In fact by triangle inequality and trivial estimate we gain trivial bound: $latex \sum_{n\leq x}|\Lambda(n)\overline{\chi(n)}|\leq x$. But by stupid use Cauchy we get following,
$latex \sum_{n\leq x}|\Lambda(n)\overline{\chi(n)}|\leq ((\sum_{n\leq x}|\Lambda(n)|^2)(\sum_{n\leq x}|\chi(n)|^2))^{\frac{1}{2}}\leq xlog^{\frac{1}{2}}x$
But this does not mean Cauchy-Schwarz is useless on charge this quality, we careful look at the inequality and try to understand why the bound will be even worse. Every time we successful use Cauchy-Schwarz there are two main phenomenon, first, we lower down the complexity of the quantity we wish to bound, second we almost do not loss any thing at all. So we just reformulate the quantity and find it lower down the complexity and the change is compatible with the equivalent condition of Cauchy-Schwarz. For example we have following identity,
$latex \sum_{n\leq x}|\Lambda(n)\overline{\chi(n)}|=\sqrt{ \sum_{n\leq x}|\Lambda(n)\overline{\chi(n)}| \sum_{m\leq x}|\Lambda(m)\overline{\chi(m)}|}=\sqrt{ \sum_{k_1,k_2\in \mathbb F_p^{\times}}\sum_{n’,m’\leq \frac{x}{p}}|\Lambda(n’)\Lambda(m’)\overline{\chi(k_1)\chi(k_2)}| }$
So we could understand this quality as the Variation of primes in arithmetic profession constructed by $latex \{pn+b| b\in\{1,2,…,p-1\}\}$. But this is still difficult to estimate, merely because of we need to control the variation of convolution of $latex \Lambda$ with itself on $latex \mathbb F_p^{\times}\simeq \{pn+b| b\in\{1,2,…,p-1\}\}$.
Now we change our perspective, recall a variant of Cauchy-Schwarz inequality, which called Bessel inequality, as following,
Bessel inequality
Let $latex {g_1,\dots,g_J: {\bf N} \rightarrow {\bf C}}$ be finitely supported functions obeying the orthonormality relationship,
$latex \displaystyle \sum_n g_j(n) \overline{g_{j’}(n)} = 1_{j=j’}$
for all $latex {1 \leq j,j’ \leq J}$. Then for any function $latex {f: {\bf N} \rightarrow {\bf C}}$, we have,
$latex \displaystyle (\sum_{j=1}^J |\sum_{n} f(n) \overline{g_j(n)}|^2)^{1/2} \leq (\sum_n |f(n)|^2)^{1/2}$.
Pf: The proof is not very difficult, we just need to keep an orthogonal picture in our mind, consider $latex \{g_{j}(n)\}, 1\leq j\leq J$ to be a orthogonal basis on $latex l^2(\mathbb N)$, then this inequality is a natural corollary.
Have this inequality in mind, by the standard argument given by transform from version of orthogonal to almost orthogonal which was merely explained in the previous note. We could image the following corresponding almost orthogonal variate of “Bessel inequality” is true:
Generalised Bessel inequality
Let $latex {g_1,\dots,g_J: {\bf N} \rightarrow {\bf C}}$ be finitely supported functions, and let $latex {\nu: {\bf N} \rightarrow {\bf R}^+}$ be a non-negative function. Let $latex {f: {\bf N} \rightarrow {\bf C}}$ be such that $latex {f}$ vanishes whenever $latex {\nu}$ vanishes, we have
$latex \displaystyle (\sum_{j=1}^J |\sum_{n} f(n) \overline{g_j(n)}|^2)^{1/2} \leq (\sum_n |f(n)|^2 / \nu(n))^{1/2} \times ( \sum_{j=1}^J \sum_{j’=1}^J c_j \overline{c_{j’}} \sum_n \nu(n) g_j(n) \overline{g_{j’}(n)} )^{1/2}$
for some sequence $latex {c_1,\dots,c_J}$ of complex numbers with $latex {\sum_{j=1}^J |c_j|^2 = 1}$, with the convention that $latex {|f(n)|^2/\nu(n)}$ vanishes whenever $latex {f(n), \nu(n)}$ both vanish.