GIFT: Geometric Information Field Theory

17 Mollified Dirichlet Polynomial

The mollified Dirichlet polynomial \(S_w(T)\) provides a computable approximation to \(S(T) = \pi ^{-1}\arg \zeta (\frac{1}{2} + iT)\) using a finite sum over primes. This chapter formalizes the constructive core (Layer B): the mollifier kernel, the finite sum, and the adaptive cutoff. All results in this section are fully proven in Lean 4 with zero axioms.

17.1 Cosine-Squared Mollifier Kernel

Definition 17.1 Cosine Kernel
#

The cosine-squared mollifier kernel:

\[ w(x) = \begin{cases} \cos ^2\! \left(\frac{\pi x}{2}\right) & \text{if } x {\lt} 1 \\ 0 & \text{if } x \geq 1 \end{cases} \]
Theorem 17.2 Kernel Non-negativity

\(w(x) \geq 0\) for all \(x \in \mathbb {R}\).

Theorem 17.3 Kernel Bounded

\(w(x) \leq 1\) for all \(x \in \mathbb {R}\).

Theorem 17.4 Kernel at Zero

\(w(0) = 1\).

Theorem 17.5 Compact Support

\(w(x) = 0\) for all \(x \geq 1\).

17.2 Mollified Dirichlet Polynomial

Definition 17.6 Prime Power Term
#

For prime \(p\) and power \(m\):

\[ a_{p,m}(T; \theta ) = w\! \left(\frac{m\log p}{\theta \log T}\right) \cdot \frac{\sin (T \cdot m\log p)}{m \cdot p^{m/2}} \]
Definition 17.7 Mollified Sum
#
\[ S_w(T; \theta ) = -\frac{1}{\pi } \sum _{\substack {p \leq N \\ p\text{ prime}}} \sum _{m=1}^{K} a_{p,m}(T; \theta ) \]
Theorem 17.8 Well-Definedness

\(S_w(T; \theta )\) is a finite sum and therefore always a well-defined real number. Unlike the formal Dirichlet series \(-\zeta '/\zeta (s)\) which diverges on \(\mathrm{Re}(s) = \frac{1}{2}\), no convergence issues arise.

Remark 17.9 Prime Power Hierarchy

The \(R^2\) contribution by prime power order \(m\) is:

\(m\)

\(\Delta R^2\)

Fraction

1 (primes)

0.872

92.8%

2 (squares)

0.057

6.1%

3 (cubes)

0.011

1.1%

This justifies the standard choice \(K = 3\).

17.3 Adaptive Cutoff

Definition 17.10 Adaptive Theta
#

The adaptive cutoff exponent:

\[ \theta (T) = \theta _0 + \frac{\theta _1}{\log T} = 1.4091 + \frac{-3.9537}{\log T} \]

determined by the normalization constraint \(\alpha = 1\) uniformly across height windows.

Definition 17.11 Adaptive Sum

\(S_w(T) = S_w(T; \theta (T))\) using the adaptive cutoff.

Remark 17.12 Three Cutoff Parameterizations

Three related cutoff models appear across the paper:

Model

Value

Context

Constant \(\theta ^*\) (mpmath)

0.9941

§3.5 (100K zeros)

Constant \(\theta ^*\) (Odlyzko)

0.9640

§9 (2M zeros)

Adaptive \((\theta _0, \theta _1)\)

\((1.409, -3.954)\)

§4 (recommended)

17.4 Main Results (Numerical Certificates)

Theorem 17.13 Counting Accuracy

Over the first 100,000 zeros: \(\max _n |N_{\text{approx}}(T_n) - n| = 0.111 {\lt} 0.5\) (safety factor \(4.52\times \)).

Theorem 17.14 Localization Rate

98.0% of zeros are localized to their correct inter-Gram interval. The 2.0% failure rate is predicted (within 8%) by a simplified GUE model.

Theorem 17.15 Variance Explained

The adaptive formula explains 93.9% of the variance in zero corrections \(\delta _n\) (\(R^2 = 0.939\), out-of-sample \(R^2 = 0.919\)).

Complete verification of the constructive core: kernel properties + sum well-definedness + standard \(K = 3\).