Comparison of Greedy Angle Decompositions: Harmonic, Dyadic, and General Sequences

Victor Geere
March 2026

Table of Contents


Abstract

In a companion paper [Geere, 2026a] we introduced the greedy harmonic decomposition of angles, in which a target \(\theta \in [0,\pi]\) is expressed as a sub-sum of the harmonic angles \(\alpha_n = \pi/(n+2)\), and studied the correlation kernel of the resulting selection indicators. An open question posed there asks how the correlation structure depends on the choice of angle sequence. In this paper we give a systematic answer. We define a general greedy decomposition for any positive, decreasing, divergent sequence \((\alpha_n)\) and compare three canonical families: the harmonic sequence \(\alpha_n^{(\mathrm{H})} = \pi/(n+2)\), the dyadic (binary) sequence \(\alpha_n^{(\mathrm{D})} = \pi/2^{n+1}\), and the power-law family \(\alpha_n^{(p)} = \pi/(n+2)^p\) for \(p \in (0,1)\). We prove that the dyadic decomposition admits an explicit closed-form threshold sequence and a product-structure correlation kernel, in sharp contrast with the harmonic case, where thresholds obey a complex recursion. For general decreasing divergent sequences, we establish universal properties—monotonicity of selection, existence of thresholds, positive semi-definiteness of the kernel—and isolate the features that distinguish different sequences: convergence rate, threshold regularity, and spectral decay of the kernel. The power-law interpolation between harmonic (\(p=1\)) and slowly-decreasing (\(p \to 0^+\)) regimes reveals a phase transition in threshold complexity at \(p = 1\).


1. General Framework

Definition 1.1 (Angle sequence). An angle sequence is a sequence \((\alpha_n)_{n \geq 0}\) of positive reals satisfying: The divergence condition ensures that every target \(\theta \in [0,\pi]\) can be approximated arbitrarily well by partial sub-sums, while the decreasing condition guarantees that the greedy algorithm is well-defined and produces monotone selection indicators.
Definition 1.2 (General greedy decomposition). Given an angle sequence \((\alpha_n)\) and a target \(\theta \in [0,\pi]\), define recursively: \[ \begin{align} \theta_0 &= 0, \\[4pt] \delta_n(\theta) &= \begin{cases} 1 & \text{if } \theta - \theta_n \geq \alpha_n, \\ 0 & \text{otherwise,} \end{cases} \\[4pt] \theta_{n+1} &= \theta_n + \delta_n \,\alpha_n. \end{align} \] The residual at step \(n\) is \(r_n(\theta) = \theta - \theta_n\). The algorithm is "greedy" in that it selects \(\alpha_n\) whenever doing so does not cause the accumulated angle to exceed \(\theta\).
Definition 1.3 (Standard notation). Throughout this paper, for a given angle sequence \((\alpha_n)\):
Remark 1.4 (Three canonical sequences). The following families will serve as our primary objects of comparison:
Key distinction. The harmonic and power-law sequences have divergent partial sums and thus represent bona fide angle sequences. The dyadic sequence has a convergent sum equal to \(\pi\), so it decomposes any \(\theta \in [0,\pi]\) exactly (as a binary expansion) without needing to be "greedy" in the asymptotic sense. This fundamental difference underlies many of the structural contrasts developed below.

2. The Dyadic Decomposition

The dyadic (binary) decomposition is the most classical and best-understood angle decomposition. We develop its properties in full to provide a baseline for comparison with the harmonic case.

Definition 2.1 (Dyadic angle sequence). Define \(\alpha_n^{(\mathrm{D})} = \pi/2^{n+1}\) for \(n \geq 0\). Then \(\alpha_0^{(\mathrm{D})} = \pi/2\), \(\alpha_1^{(\mathrm{D})} = \pi/4\), \(\alpha_2^{(\mathrm{D})} = \pi/8\), etc. The partial sums satisfy \[ S_N^{(\mathrm{D})} = \sum_{n=0}^{N}\frac{\pi}{2^{n+1}} = \pi\!\left(1 - \frac{1}{2^{N+1}}\right). \]
Theorem 2.2 (Dyadic decomposition is binary expansion). For any \(\theta \in [0,\pi)\), the greedy decomposition with the dyadic sequence produces selection indicators \(\delta_n^{(\mathrm{D})}(\theta)\) that coincide with the binary digits of \(\theta/\pi\): \[ \frac{\theta}{\pi} = \sum_{n=0}^{\infty}\frac{\delta_n^{(\mathrm{D})}(\theta)}{2^{n+1}} = 0.\delta_0\delta_1\delta_2\cdots\;\text{(base 2)}. \] For \(\theta = \pi\), the greedy algorithm selects every index: \(\delta_n^{(\mathrm{D})}(\pi) = 1\) for all \(n\).
The greedy rule selects \(\alpha_n^{(\mathrm{D})} = \pi/2^{n+1}\) whenever the residual \(r_n \geq \pi/2^{n+1}\). This is precisely the standard greedy algorithm for binary expansion of a real number in \([0,1]\), applied to \(\theta/\pi\). At each step, \(\delta_n = 1\) if the residual is at least \(\pi/2^{n+1}\), and the new residual is \(r_{n+1} = r_n - \delta_n \cdot \pi/2^{n+1}\). The uniqueness of binary expansions (for non-dyadic rationals) gives the result. For dyadic rationals, the greedy algorithm produces the terminating expansion (with trailing zeros).

For \(\theta = \pi\): at step 0, \(r_0 = \pi \geq \pi/2\), so \(\delta_0 = 1\) and \(r_1 = \pi/2\). At step 1, \(r_1 = \pi/2 \geq \pi/4\), so \(\delta_1 = 1\) and \(r_2 = \pi/4\). By induction, \(\delta_n = 1\) and \(r_{n+1} = \pi/2^{n+1}\) for all \(n\).

Lemma 2.3 (Dyadic thresholds are explicit). The threshold angle for index \(n\) in the dyadic decomposition is \[ \theta_n^{*(\mathrm{D})} = \frac{\pi}{2^{n+1}}. \] That is, \(\delta_n^{(\mathrm{D})}(\theta) = 1\) if and only if the \((n+1)\)-th binary digit of \(\theta/\pi\) is 1, which occurs precisely when the residual at step \(n\) is at least \(\pi/2^{n+1}\).
Since the dyadic decomposition is the binary expansion (Theorem 2.2), the \(n\)-th digit is 1 for all \(\theta\) in a set that depends on the higher-order digits. However, by the monotonicity argument of [Geere, 2026a, Lemma 4.1] (which applies to any decreasing sequence), \(\delta_n^{(\mathrm{D})}(\theta)\) is non-decreasing in \(\theta\).

The smallest \(\theta\) for which \(\delta_n^{(\mathrm{D})}(\theta) = 1\) is achieved when all earlier digits \(\delta_0 = \cdots = \delta_{n-1} = 0\) and the residual at step \(n\) is exactly \(\alpha_n^{(\mathrm{D})}\). This gives \(\theta_n^{*(\mathrm{D})} = \alpha_n^{(\mathrm{D})} = \pi/2^{n+1}\).

To verify: if \(\theta = \pi/2^{n+1}\), then for all \(k < n\), \(\alpha_k^{(\mathrm{D})} = \pi/2^{k+1} > \pi/2^{n+1} = \theta \geq r_k\), so \(\delta_k = 0\) and the residual remains \(\theta\). At step \(n\), \(r_n = \theta = \pi/2^{n+1} = \alpha_n^{(\mathrm{D})}\), so \(\delta_n = 1\).

Remark 2.4 (Contrast with harmonic thresholds). The dyadic thresholds are strictly monotonically decreasing: \(\theta_0^{*(\mathrm{D})} > \theta_1^{*(\mathrm{D})} > \theta_2^{*(\mathrm{D})} > \cdots\). Moreover, they are completely determined by the angle sequence alone, with no interaction between indices. In the harmonic case, the threshold \(\theta_n^{*(\mathrm{H})}\) depends on all previous thresholds through the accumulated angle \(\theta_n(\theta_n^*)\), and the sequence \((\theta_n^{*(\mathrm{H})})_n\) is non-monotone (cf. [Geere, 2026a, Proposition 5.3(b)]). This reflects the fact that in the dyadic case, earlier non-selections leave the residual untouched, while in the harmonic case the density of selections creates complex feedback.
Lemma 2.5 (Independence of dyadic digits). For distinct indices \(n \neq m\), the centred selection indicators \(\delta_n^{(\mathrm{D})}(\theta) - \theta/\pi\) and \(\delta_m^{(\mathrm{D})}(\theta) - \theta/\pi\) satisfy \[ K_{\mathrm{D}}(n,m) = \int_0^\pi \!\left[\delta_n^{(\mathrm{D})}(\theta) - \frac{\theta}{\pi}\right]\!\left[\delta_m^{(\mathrm{D})}(\theta) - \frac{\theta}{\pi}\right]d\theta. \] However, unlike in the probabilistic setting (where binary digits of a uniform random variable are independent), the kernel \(K_{\mathrm{D}}(n,m)\) is not zero for \(n \neq m\), because the centring function \(\theta/\pi\) is not the conditional mean of \(\delta_n\) given the other digits.
The marginal "selection probability" (i.e., the Lebesgue measure of \(\{\theta \in [0,\pi] : \delta_n^{(\mathrm{D})}(\theta) = 1\}\)) is exactly \(\pi/2\) for every \(n\) (each binary digit is 1 on a set of measure \(1/2\) of \([0,1]\), scaled by \(\pi\)). Thus the Lebesgue-average of \(\delta_n^{(\mathrm{D})}\) over \([0,\pi]\) is \(1/2\), while the centring function \(\theta/\pi\) is not constant. This mismatch means the centred indicators are not orthogonal.

However, if we instead centre by the constant \(1/2\), the digits become exactly uncorrelated in the \(L^2([0,\pi])\) inner product, since the binary digits of a uniform random variable in \([0,1]\) are independent. We record this observation for comparison in Section 5.

Definition 2.6 (Constant-centred kernel). Define the alternative kernel \[ \widetilde{K}_\alpha(n,m) = \int_0^\pi \!\left[\delta_n(\theta) - \bar\delta_n\right]\!\left[\delta_m(\theta) - \bar\delta_m\right]d\theta, \] where \(\bar\delta_n = \frac{1}{\pi}\int_0^\pi \delta_n(\theta)\,d\theta = 1 - \theta_n^*/\pi\) is the Lebesgue-average selection rate. For the dyadic case, \(\bar\delta_n^{(\mathrm{D})} = 1/2\) for all \(n\).
Proposition 2.7 (Dyadic constant-centred kernel). With constant centring: \[ \widetilde{K}_{\mathrm{D}}(n,m) = \begin{cases} \pi/4 & \text{if } n = m, \\ 0 & \text{if } n \neq m. \end{cases} \] That is, the constant-centred dyadic kernel is a scalar multiple of the identity.
The binary digits of \(\theta/\pi\), viewed as a function from \([0,\pi]\) to \(\{0,1\}^\mathbb{N}\), push forward Lebesgue measure on \([0,\pi]\) to \(\pi\) times the fair-coin product measure on \(\{0,1\}^\mathbb{N}\). Under this product measure, the digits are independent with mean \(1/2\) and variance \(1/4\). Therefore \[ \widetilde{K}_{\mathrm{D}}(n,m) = \pi \cdot \operatorname{Cov}(\delta_n, \delta_m) = \begin{cases} \pi/4 & \text{if } n = m, \\ 0 & \text{if } n \neq m. \end{cases} \]
Key contrast. The dyadic decomposition has a diagonal constant-centred kernel (reflecting the independence of binary digits), while the harmonic decomposition has a dense, non-diagonal kernel (reflecting the complex dependencies between greedy selections). The choice of centring—by \(\theta/\pi\) or by the constant mean \(\bar\delta_n\)—is natural in different contexts and leads to different kernel structures.

3. Greedy Decompositions for General Sequences

We now establish the properties that hold universally for any angle sequence satisfying Definition 1.1, and identify the features that depend on the specific choice of sequence.

Theorem 3.1 (Universal properties). Let \((\alpha_n)\) be any angle sequence (Definition 1.1). Then:
  1. Convergence. For every \(\theta \in [0,\pi]\), the accumulated angle \(\theta_N \to \theta\) as \(N \to \infty\).
  2. Selection monotonicity. For each fixed \(n\), the map \(\theta \mapsto \delta_n(\theta)\) is non-decreasing.
  3. Threshold existence. For each \(n\), there exists \(\theta_n^* \in [0,\pi]\) such that \(\delta_n(\theta) = \mathbf{1}[\theta \geq \theta_n^*]\).
  4. Threshold lower bound. \(\theta_n^* \geq \alpha_n\) for all \(n\).
  5. Positive semi-definite kernel. The matrix \((K_\alpha(n,m))_{n,m}\) is positive semi-definite for any centring of the form \(\delta_n(\theta) - f(\theta)\) with \(f\) measurable.

(a) The residual bound \(r_N \leq \alpha_{N-1}\) follows from the greedy rule exactly as in [Geere, 2026a, Lemma 3.1]: if \(\delta_N = 0\), then \(r_N = r_{N-1} < \alpha_N \leq \alpha_{N-1}\); if \(\delta_N = 1\), then \(r_{N+1} = r_N - \alpha_N < \alpha_{N-1} - \alpha_N \leq \alpha_{N-1}\). Since \(\alpha_n \to 0\) (as a decreasing sequence whose terms must tend to zero for the sum to diverge, or even for a convergent sum), \(r_N \to 0\).

(b)–(c) The proof of [Geere, 2026a, Lemma 4.1] uses only the decreasing property of \((\alpha_n)\) and the greedy rule; it applies verbatim to any decreasing sequence.

(d) At the threshold \(\theta = \theta_n^*\), we have \(\theta_n^* = \alpha_n + \theta_n(\theta_n^*)\) with \(\theta_n(\theta_n^*) \geq 0\).

(e) The quadratic form \(\sum_{n,m} a_n\bar{a}_m K_\alpha(n,m) = \int_0^\pi |\sum_n a_n[\delta_n(\theta) - f(\theta)]|^2\,d\theta \geq 0\).

Definition 3.2 (Rate function). For an angle sequence \((\alpha_n)\), define the rate function \(\rho_\alpha(N) = \alpha_{N-1}\), so that \(|\theta - \theta_N| \leq \rho_\alpha(N)\) for all \(\theta\). The convergence rate of the greedy decomposition is determined by \(\rho_\alpha\):
Definition 3.3 (Selection density). The selection density is the function \[ d_\alpha(N,\theta) = \frac{D_\alpha(N,\theta)}{N+1}, \] measuring the fraction of the first \(N+1\) indices that are selected. Its behaviour depends on the relationship between the growth of \(S_N\) and the target \(\theta\).
Lemma 3.4 (Selection density asymptotics). Let \(\theta \in (0,\pi)\). Then:
In each case, the accumulated angle satisfies \(\theta_N = \sum_{n=0}^{N}\delta_n\alpha_n \to \theta\). The selection count is determined by inverting this relation.

Harmonic: \(\theta_N \approx \pi \sum_{n : \delta_n = 1} 1/(n+2)\). Since the selected terms contribute \(\theta/\pi\) of the harmonic sum asymptotically (cf. [Geere, 2026a, Lemma 3.3]), \(D_{\mathrm{H}}(N,\theta) \sim (\theta/\pi)\ln N\).

Dyadic: Given the binary expansion interpretation (Theorem 2.2), the density of 1-digits in the binary expansion of \(\theta/\pi\) depends on the specific value of \(\theta\). However, by the equidistribution of binary digits for Lebesgue-typical \(\theta\), the density is approximately \(1/2\) for typical targets. More precisely, \(D_{\mathrm{D}}(N,\theta)/(N+1)\) converges to the asymptotic density of 1-digits in the binary expansion of \(\theta/\pi\), which exists and equals \(1/2\) for Lebesgue-a.e. \(\theta\) (by the normal number theorem).

Power-law: \(\theta_N \approx \pi \sum_{n : \delta_n = 1}(n+2)^{-p}\). Since the full partial sum grows as \(\pi N^{1-p}/(1-p)\) and the selected terms accumulate to \(\theta\), the selection count satisfies \(D_p(N,\theta) \cdot \bar{\alpha}(N) \sim \theta\) where \(\bar{\alpha}(N) \sim \pi N^{-p}\) is the average term size near index \(N\). This gives \(D_p(N,\theta) \sim \theta N^p/\pi\), so \(d_p(N,\theta) \sim \theta N^{p-1}/\pi \to 0\).

Summary of Section 3. The universal properties (convergence, monotonicity, threshold existence, kernel semi-definiteness) hold for all angle sequences. The distinguishing features are the convergence rate \(\rho_\alpha(N)\), which ranges from polynomial (harmonic, power-law) to exponential (dyadic), and the selection density, which ranges from sub-linear growth (harmonic, power-law) to linear growth (dyadic). These differences propagate into the threshold structure and correlation kernel.

4. Threshold Structure Comparison

The threshold angles \(\theta_n^*\) encode the combinatorial complexity of a greedy decomposition. We compare their structure across the three canonical sequences.

Proposition 4.1 (Threshold characterisation). For the three canonical sequences:
  1. Dyadic: \(\theta_n^{*(\mathrm{D})} = \pi/2^{n+1}\). The thresholds are strictly decreasing, explicitly computable, and satisfy \(\theta_n^{*(\mathrm{D})} = \alpha_n^{(\mathrm{D})}\) (the threshold equals the angle). No interaction between indices.
  2. Harmonic: \(\theta_n^{*(\mathrm{H})} = \alpha_n^{(\mathrm{H})} + \theta_n(\theta_n^{*(\mathrm{H})})\). The thresholds satisfy a coupled recursion, are non-monotone in \(n\), and satisfy \(\theta_n^{*(\mathrm{H})} \geq \alpha_n^{(\mathrm{H})} = \pi/(n+2)\) with equality holding only when no earlier index is selected at \(\theta = \theta_n^{*(\mathrm{H})}\).
  3. Power-law (\(p \in (0,1)\)): \(\theta_n^{*(p)} = \alpha_n^{(p)} + \theta_n(\theta_n^{*(p)})\). The thresholds exhibit the same coupled recursion as the harmonic case, with non-monotonicity and interaction between indices. However, since the terms \(\alpha_n^{(p)} = \pi/(n+2)^p\) decrease more slowly than the harmonic terms, the accumulated angle \(\theta_n(\theta_n^{*(p)})\) is smaller relative to \(\alpha_n^{(p)}\) for large \(n\), so \(\theta_n^{*(p)}/\alpha_n^{(p)} \to 1\) more rapidly.

(a) Proved in Lemma 2.3. The key is that the dyadic terms decrease geometrically: \(\alpha_k^{(\mathrm{D})} = 2\alpha_{k+1}^{(\mathrm{D})}\), so at the threshold \(\theta = \pi/2^{n+1}\), all earlier terms exceed \(\theta\) and are not selected. Thus \(\theta_n(\theta_n^{*(\mathrm{D})}) = 0\).

(b) Established in [Geere, 2026a, Sections 4–5]. The non-monotonicity arises because the harmonic terms decrease polynomially: \(\alpha_k^{(\mathrm{H})}/\alpha_{k+1}^{(\mathrm{H})} = (k+3)/(k+2) \to 1\), so at the threshold of a later index, some earlier indices may already be selected, adding to the accumulated angle and pushing the threshold higher than \(\alpha_n^{(\mathrm{H})}\) alone.

(c) For the power-law sequence, the ratio \(\alpha_k^{(p)}/\alpha_{k+1}^{(p)} = ((k+3)/(k+2))^p \to 1\) as \(k \to \infty\), similarly to the harmonic case. The number of earlier indices selected at the threshold is of order \(D_p(n, \theta_n^{*(p)})\), which (by Lemma 3.4) grows sub-linearly. The accumulated angle from these selections is bounded by \(\theta_n^{*(p)}\), creating a self-consistent equation. For \(p < 1\), the terms \(\alpha_k^{(p)}\) decrease more slowly, so fewer selections are needed to accumulate a given angle, and the interaction effect is weaker.

Definition 4.2 (Threshold excess ratio). The threshold excess ratio for index \(n\) in the sequence \((\alpha_n)\) is \[ \varepsilon_n = \frac{\theta_n^*}{\alpha_n} - 1 = \frac{\theta_n(\theta_n^*)}{\alpha_n} \geq 0. \] This measures how much the threshold exceeds the angle itself, due to interactions with earlier selections. We have \(\varepsilon_n = 0\) when no earlier index is selected at the threshold, and \(\varepsilon_n > 0\) otherwise.
Proposition 4.3 (Threshold excess comparison).
  1. Dyadic: \(\varepsilon_n^{(\mathrm{D})} = 0\) for all \(n\). Every threshold equals the angle exactly.
  2. Harmonic: The set \(\{n : \varepsilon_n^{(\mathrm{H})} > 0\}\) has positive density. Moreover, \(\varepsilon_n^{(\mathrm{H})}\) can be arbitrarily large: for any \(C > 0\), there exist indices \(n\) with \(\varepsilon_n^{(\mathrm{H})} > C\).
  3. Power-law (\(p \in (0,1)\)): For each fixed \(p < 1\), the set \(\{n : \varepsilon_n^{(p)} > C\}\) has density tending to zero as \(C \to \infty\), and the decay is faster for smaller \(p\).

(a) Immediate from Lemma 2.3.

(b) For the harmonic sequence, at the threshold \(\theta_n^{*(\mathrm{H})}\), the accumulated angle from earlier selections is \(\theta_n(\theta_n^{*(\mathrm{H})}) = \sum_{k < n,\, \theta_k^{*(\mathrm{H})} \leq \theta_n^{*(\mathrm{H})}} \alpha_k^{(\mathrm{H})}\). When \(\theta_n^{*(\mathrm{H})}\) is not too small, many earlier indices with small angles (large \(k\)) satisfy \(\theta_k^{*(\mathrm{H})} \leq \theta_n^{*(\mathrm{H})}\) and are therefore selected, contributing a sum that can be much larger than \(\alpha_n^{(\mathrm{H})}\). For instance, the threshold \(\theta_0^{*(\mathrm{H})} = \pi/2\) has \(\varepsilon_0^{(\mathrm{H})} = 0\), but for any index \(n\) whose threshold happens to be near \(\pi/2\), many earlier small-angle terms are selected, producing a large excess.

(c) For the power-law sequence with \(p < 1\), the terms decrease more slowly, so at the threshold of index \(n\), fewer earlier indices have angles smaller than \(\theta_n^{*(p)}\) (since \(\alpha_k^{(p)} = \pi/(k+2)^p\) is larger for a given \(k\) compared to the harmonic case). The accumulated angle from the selected earlier indices grows more slowly relative to \(\alpha_n^{(p)}\), bounding the excess ratio.

Remark 4.4 (Threshold regularity hierarchy). The three families exhibit a clear hierarchy of threshold regularity: As \(p \to 1^-\), the power-law thresholds approach the harmonic case, and the excess ratios become increasingly large and irregular. This transition marks a qualitative change in combinatorial complexity.
Proposition 4.5 (Monotonicity criterion). The threshold sequence \((\theta_n^*)_{n \geq 0}\) is strictly decreasing if and only if \(\alpha_n/\alpha_{n-1} \to 0\) as \(n \to \infty\). In particular:
If \(\alpha_n/\alpha_{n-1} \to 0\), then for large \(n\), the angle \(\alpha_n\) is much smaller than \(\alpha_{n-1}\). At the threshold \(\theta = \theta_n^*\), which is at most \(\alpha_n + \sum_{kIf \(\alpha_n/\alpha_{n-1} \not\to 0\), then for infinitely many \(n\), the angle \(\alpha_n\) is comparable to \(\alpha_{n-1}\), and at the threshold of \(n\), the angle \(\alpha_{n-1}\) (or other nearby earlier terms) may also be selected, breaking monotonicity.

For the dyadic case: \(\alpha_n/\alpha_{n-1} = 1/2\), but \(\theta_n^{*(\mathrm{D})} = \pi/2^{n+1}\) while \(\alpha_{n-1}^{(\mathrm{D})} = \pi/2^n > \theta_n^{*(\mathrm{D})}\) for all \(n\), so no earlier term is selected. The threshold equals \(\alpha_n\) despite the ratio not tending to zero, because the geometric structure ensures exact separation.


5. Correlation Kernel Comparison

We now compare the correlation kernels \(K_\alpha(n,m)\) for the three canonical sequences. Recall from [Geere, 2026a, Definition 4.4] that \[ K_\alpha(n,m) = \int_0^\pi \!\left[\delta_n(\theta) - \frac{\theta}{\pi}\right]\!\left[\delta_m(\theta) - \frac{\theta}{\pi}\right]d\theta. \]

Theorem 5.1 (Dyadic kernel, explicit form). The dyadic correlation kernel with linear centring is: \[ K_{\mathrm{D}}(n,m) = \frac{(\theta_n^{*(\mathrm{D})})^3 + (\pi - \theta_m^{*(\mathrm{D})})^3}{3\pi^2} + \int_{\theta_n^{*(\mathrm{D})}}^{\theta_m^{*(\mathrm{D})}} \frac{-\theta(\pi-\theta)}{\pi^2}\,d\theta \] for \(\theta_n^{*(\mathrm{D})} \leq \theta_m^{*(\mathrm{D})}\), identical in form to the harmonic case [Geere, 2026a, Lemma 4.7] but with the explicit thresholds \(\theta_n^{*(\mathrm{D})} = \pi/2^{n+1}\). In particular, the diagonal entries are \[ K_{\mathrm{D}}(n,n) = \frac{(\pi/2^{n+1})^3 + (\pi - \pi/2^{n+1})^3}{3\pi^2} = \frac{\pi}{3}\cdot\frac{1 + (2^{n+1}-1)^3}{2^{3(n+1)}}. \]
This follows directly from [Geere, 2026a, Lemmas 4.6–4.7] applied with the dyadic thresholds. The integration is the same three-interval decomposition, with the specific threshold values substituted.

For the diagonal, substituting \(\theta_n^* = \pi/2^{n+1}\):

\[ K_{\mathrm{D}}(n,n) = \frac{(\pi/2^{n+1})^3 + (\pi(1 - 2^{-(n+1)}))^3}{3\pi^2} = \frac{\pi}{3}\left(\frac{1}{2^{3(n+1)}} + \left(1 - \frac{1}{2^{n+1}}\right)^3\right). \]
Lemma 5.2 (Dyadic diagonal asymptotics). As \(n \to \infty\): \[ K_{\mathrm{D}}(n,n) \to \frac{\pi}{3}. \] Contrast with the harmonic case, where \(K_{\mathrm{H}}(n,n)\) oscillates in \([\pi/12, \pi/3]\) depending on the threshold \(\theta_n^{*(\mathrm{H})}\).
Since \(\theta_n^{*(\mathrm{D})} = \pi/2^{n+1} \to 0\), by the formula in Theorem 5.1: \[ K_{\mathrm{D}}(n,n) = \frac{(\theta_n^{*(\mathrm{D})})^3 + (\pi - \theta_n^{*(\mathrm{D})})^3}{3\pi^2} \to \frac{0 + \pi^3}{3\pi^2} = \frac{\pi}{3}. \]
Proposition 5.3 (Kernel structure comparison). The key structural differences between the dyadic and harmonic kernels are:
Property Dyadic \(K_{\mathrm{D}}\) Harmonic \(K_{\mathrm{H}}\)
Diagonal values \(K_{\mathrm{D}}(n,n) \to \pi/3\) \(K_{\mathrm{H}}(n,n) \in [\pi/12, \pi/3]\), oscillating
Off-diagonal decay Rapid: \(|K_{\mathrm{D}}(n,m)|\) decreases as \(|n-m|\) increases Complex: depends on threshold spacing, non-monotone
Threshold separation \(|\theta_n^* - \theta_m^*| = \pi|2^{-(n+1)} - 2^{-(m+1)}|\) Non-monotone, complex spacing
Constant-centred kernel Diagonal (Proposition 2.7) Dense, non-diagonal
Kernel rank (finite truncation) Full rank for any \(N\) Full rank for any \(N\)
Proposition 5.4 (Kernel formula for general sequences). For any angle sequence \((\alpha_n)\) with thresholds \(\theta_n^*\), the kernel takes the same universal form as in [Geere, 2026a, Lemmas 4.6–4.7]: \[ K_\alpha(n,m) = \frac{(\theta_{\min}^*)^3}{3\pi^2} - \frac{1}{\pi^2}\left[\frac{\pi}{2}(\theta_{\max}^{*2} - \theta_{\min}^{*2}) - \frac{1}{3}(\theta_{\max}^{*3} - \theta_{\min}^{*3})\right] + \frac{(\pi - \theta_{\max}^*)^3}{3\pi^2}, \] where \(\theta_{\min}^* = \min(\theta_n^*, \theta_m^*)\) and \(\theta_{\max}^* = \max(\theta_n^*, \theta_m^*)\). This is a polynomial in the thresholds, and all differences between decompositions are encoded in the mapping \(n \mapsto \theta_n^*\).
The derivation in [Geere, 2026a, Lemma 4.7] uses only the fact that \(\delta_n(\theta) = \mathbf{1}[\theta \geq \theta_n^*]\) is a step function. This structure holds for any angle sequence by Theorem 3.1(c), so the same integration applies.
Key insight. The kernel formula is universal: it depends on the angle sequence only through the threshold map \(n \mapsto \theta_n^*\). All differences in correlation structure between different decompositions reduce to differences in threshold structure. The kernel is entirely determined by the thresholds.

6. Convergence Rate Comparison

The convergence rate of \(\sin(\theta_N)\) to \(\sin(\theta)\) depends on how quickly the residual \(r_N\) tends to zero. We compare the three families and derive the reconstruction error bounds.

Theorem 6.1 (Reconstruction error comparison). The sine reconstruction error \(|\sin\theta - \sin\theta_N|\) satisfies:
  1. Dyadic: \(|\sin\theta - \sin\theta_N| \leq \pi/2^{N+1}\). Exponential convergence.
  2. Harmonic: \(|\sin\theta - \sin\theta_N| \leq \pi/(N+2)\). Polynomial convergence, rate \(O(1/N)\).
  3. Power-law (\(p \in (0,1)\)): \(|\sin\theta - \sin\theta_N| \leq \pi/(N+1)^p\). Polynomial convergence, rate \(O(N^{-p})\).
In particular, the convergence rates satisfy \[ O(2^{-N}) \ll O(N^{-1}) \ll O(N^{-p}) \quad \text{for } p \in (0,1), \] showing that the dyadic decomposition converges fastest, the harmonic decomposition is intermediate, and the power-law decomposition converges slowest.
In all three cases, the Lipschitz continuity of sine gives \(|\sin\theta - \sin\theta_N| \leq |\theta - \theta_N| = r_N \leq \alpha_{N-1}\). Substituting the respective angle sequences yields the bounds.
Remark 6.2 (Convergence–complexity trade-off). There is a fundamental trade-off between convergence rate and threshold complexity: This suggests that the harmonic sequence sits at a critical boundary: it is the slowest-converging sequence (among polynomial families) for which the threshold complexity is maximal.
Proposition 6.3 (Efficiency: bits per unit angle). Define the efficiency of a decomposition as the number of angle terms needed to achieve a given accuracy \(\varepsilon\): \[ N_\alpha(\varepsilon) = \min\{N : \rho_\alpha(N) \leq \varepsilon\}. \] Then: The dyadic decomposition is exponentially more efficient than the harmonic or power-law decompositions.
In each case, solve \(\rho_\alpha(N) = \varepsilon\):

7. Spectral Properties

The positive semi-definite kernel matrices \((K_\alpha(n,m))_{n,m=0}^{N}\) have spectra (sets of eigenvalues) whose structure encodes global features of the decomposition. We compare the spectral properties of the dyadic and harmonic kernels.

Proposition 7.1 (Dyadic kernel spectral structure). For the dyadic decomposition:
  1. The constant-centred kernel \(\widetilde{K}_{\mathrm{D}}\) is \((\pi/4) I\), so all eigenvalues equal \(\pi/4\). The spectral measure is a single atom at \(\pi/4\).
  2. The linearly-centred kernel \(K_{\mathrm{D}}\) has eigenvalues that converge: the \((N+1) \times (N+1)\) matrix \((K_{\mathrm{D}}(n,m))_{n,m=0}^{N}\) has a bulk of eigenvalues near \(\pi/3\) (since \(K_{\mathrm{D}}(n,n) \to \pi/3\) and the off-diagonal entries are small for well-separated indices), with a bounded number of outlier eigenvalues arising from the first few indices where \(\theta_n^{*(\mathrm{D})}\) is not close to zero.

(a) Immediate from Proposition 2.7.

(b) The linearly-centred kernel can be written as \(K_{\mathrm{D}} = \widetilde{K}_{\mathrm{D}} + E\) where \(E\) is a correction matrix arising from the change of centring from constant \(1/2\) to linear \(\theta/\pi\). The correction involves the function \(1/2 - \theta/\pi\), whose \(L^2\) norm is bounded. By Weyl's inequality, the eigenvalues of \(K_{\mathrm{D}}\) differ from those of \(\widetilde{K}_{\mathrm{D}}\) by at most \(\|E\|_{\mathrm{op}}\), where the operator norm of \(E\) is bounded by the \(L^2\) norm of the centring difference times \(\sqrt{N+1}\). For large \(N\), the bulk eigenvalues approach \(\pi/4 + O(1)\).

Proposition 7.2 (Harmonic kernel spectral structure). For the harmonic decomposition:
  1. The eigenvalues of \((K_{\mathrm{H}}(n,m))_{n,m=0}^{N}\) are all positive (the matrix is strictly positive definite for any finite \(N\)).
  2. The largest eigenvalue is \(O(N)\) (growing linearly), while the smallest positive eigenvalue is \(O(1)\) (bounded away from zero).
  3. The spectral distribution does not concentrate at a single point, reflecting the non-trivial off-diagonal structure.

(a) Strict positive definiteness follows from the linear independence of the step functions \(\delta_n(\theta) - \theta/\pi\) in \(L^2([0,\pi])\). These are linearly independent because the step functions \(\mathbf{1}[\theta \geq \theta_n^*]\) have distinct jump points (the thresholds are all distinct, since each \(\theta_n^*\) depends on all previous selections in a way that generically produces distinct values).

(b) The trace is \(\sum_{n=0}^{N} K_{\mathrm{H}}(n,n) = \Theta(N)\) (since each diagonal entry is \(\Theta(1)\)), so the average eigenvalue is \(\Theta(1)\). The largest eigenvalue is bounded above by the trace, giving \(O(N)\). Due to the oscillatory structure of the thresholds, the off-diagonal mass is distributed, preventing any eigenvalue from growing faster than \(O(N)\).

(c) The non-concentration follows from the non-trivial correlation between different indices: unlike the dyadic case, \(K_{\mathrm{H}}(n,m)\) is not zero or negligible for many pairs \((n,m)\), so the kernel matrix is not close to a scalar times the identity.

Proposition 7.3 (Spectral comparison summary).
Spectral property Dyadic (constant-centred) Dyadic (linear-centred) Harmonic (linear-centred)
Eigenvalue spread All equal (\(\pi/4\)) Concentrated near \(\pi/3\) Spread over \([\Omega(1), O(N)]\)
Condition number 1 \(O(1)\) \(O(N)\)
Trace growth \(\Theta(N)\) \(\Theta(N)\) \(\Theta(N)\)
Off-diagonal mass Zero \(O(1)\) per row \(\Theta(1)\) per row

8. Worked Examples and Special Sequences

8.1. First three thresholds

We compute the first three thresholds for each canonical sequence to illustrate the structural differences.

Example 8.1 (Dyadic, \(n = 0, 1, 2\)). Thresholds: \(\pi/2 > \pi/4 > \pi/8\). Strictly decreasing. No interaction.
Example 8.2 (Harmonic, \(n = 0, 1, 2\)). Thresholds: \(\pi/2 > \pi/3 > \pi/4\). Strictly decreasing for these first three indices (the non-monotonicity appears later, as documented in [Geere, 2026a]).
Example 8.3 (Power-law with \(p = 1/2\), \(n = 0, 1, 2\)). Thresholds: \(\pi/\sqrt{2} > \pi/\sqrt{3} > \pi/2\). Strictly decreasing for these first indices.

8.2. Interaction effects in the harmonic case

Example 8.4 (Harmonic threshold interaction at \(n = 4\)). Consider the harmonic sequence with \(\alpha_n = \pi/(n+2)\). The first several thresholds are: At \(n = 4\), \(\alpha_4 = \pi/6 \approx 0.524\). We check if at \(\theta = \pi/6\), any earlier index is selected. Since \(\pi/6 < \pi/5 < \pi/4 < \pi/3 < \pi/2\), no earlier index is selected, so \(\theta_4^* = \pi/6\).

At \(n = 5\), \(\alpha_5 = \pi/7 \approx 0.449\). At \(\theta = \pi/7\): still below all previous thresholds, so \(\theta_5^* = \pi/7\).

The interaction effects become significant for larger \(n\), when the target angle \(\theta_n^*\) is large enough for some earlier small-angle terms to be selected. The first instance of a threshold exceeding the simple bound occurs when the greedy selections at the threshold create a nonzero accumulated angle.

8.3. A rapidly-decreasing sequence

Example 8.5 (Factorial sequence). Consider \(\alpha_n = \pi/n!\) (with \(\alpha_0 = \pi\), \(\alpha_1 = \pi\), \(\alpha_2 = \pi/2\), etc.). The sum \(\sum \pi/n! = \pi(e-1)\) converges, so this is not an angle sequence in the sense of Definition 1.1. However, since \(\pi(e-1) \approx 5.40 > \pi\), every \(\theta \in [0,\pi]\) can be represented (though the algorithm terminates in finitely many steps for each \(\theta\)). The thresholds decrease super-exponentially: \(\theta_n^* = \alpha_n = \pi/n!\) for sufficiently large \(n\). This sequence has no interaction effects and the simplest possible threshold structure.

9. Discussion

9.1. Summary of the comparison

The following table summarises the main comparison axes:

Feature Dyadic Harmonic Power-law \((p \in (0,1))\)
Sum convergence \(\sum \alpha_n = \pi\) (convergent) \(\sum \alpha_n = \infty\) \(\sum \alpha_n = \infty\)
Reconstruction rate \(O(2^{-N})\) \(O(1/N)\) \(O(N^{-p})\)
Efficiency \(N(\varepsilon)\) \(O(\log(1/\varepsilon))\) \(O(1/\varepsilon)\) \(O(\varepsilon^{-1/p})\)
Selection density \(\to \theta/\pi\) (const) \(\sim (\theta/\pi)\frac{\ln N}{N} \to 0\) \(\sim (\theta/\pi) N^{p-1} \to 0\)
Threshold sequence Monotone, explicit Non-monotone, recursive Non-monotone, recursive
Excess ratio \(\varepsilon_n\) 0 for all \(n\) Unbounded Bounded for most \(n\)
Kernel (constant-centred) Diagonal Dense Dense
Kernel (linear-centred) Near-diagonal Dense, oscillatory Dense
Spectral condition number \(O(1)\) \(O(N)\) Intermediate

9.2. The harmonic boundary

A recurring theme is that the harmonic sequence \(\alpha_n = \pi/(n+2)\) sits at a critical boundary:

This boundary character makes the harmonic decomposition the most interesting from a structural standpoint: it has the richest combinatorial features while still providing useful convergence.

9.3. The role of centring

The choice of centring function in the correlation kernel has a significant effect on the structure:

For the harmonic decomposition, both centrings produce dense kernels, and neither reveals a diagonal structure. This is a genuine structural feature of the harmonic case, not an artefact of the centring choice.

9.4. Open questions

This comparison suggests several further questions:

  1. Optimal angle sequences. Among all angle sequences with a given convergence rate \(\rho(N)\), which one has the simplest threshold structure (e.g., minimises the total excess \(\sum_{n=0}^{N} \varepsilon_n\))? Is there a trade-off between threshold simplicity and other desirable properties?
  2. Transition at \(p = 1\). The passage from \(p < 1\) (bounded excess, simpler thresholds) to \(p = 1\) (unbounded excess, complex thresholds) is qualitative. What is the quantitative behaviour of the excess ratios \(\varepsilon_n^{(p)}\) as \(p \to 1^-\)? Is there a scaling limit?
  3. Natural centring for general sequences. For the dyadic decomposition, the natural centring is the constant \(1/2\) (which diagonalises the kernel). For a general angle sequence, is there a centring function \(f_\alpha(\theta)\) that minimises the off-diagonal mass of the kernel \(K_\alpha\)?
  4. Non-greedy decompositions. All decompositions considered here are greedy. How do non-greedy selection rules (e.g., alternating, random, or optimisation-based) compare in terms of threshold structure and kernel properties?

References