← Back to Blog

Revisiting Sprecher's Proof of Kolmogorov's Superposition Theorem

Adnan Mahmud • July, 2025

This blog delves into David Sprecher's 1996 work on Kolmogorov's superposition theorem [1]. This pivotal research ignited a decade of intense mathematical investigation, culminating in accurate mathematical principles that now underpin advanced machine learning applications.

1. The Mathematical Challenge

Kolmogorov's 1957 superposition theorem [3] stands as one of the more counterintuitive results in analysis: any multivariate continuous function f(x₁, x₂, ..., xₙ) can be written as a finite sum of compositions involving only univariate functions. More precisely:

f(x₁,...,xₙ) = Σ(q=0 to 2n) Φq[Σ(p=1 to n) αₚψ(xₚ + qa)]

where ψ represents a single "universal" inner function that works for all possible f, whilst the outer functions Φq depend on the specific function being approximated. The theorem's proof, however, was entirely non-constructive—mathematicians could prove such functions existed without having the faintest idea how to compute them.

2. Sprecher's Construction: Building ψ from Scratch

Sprecher's revolutionary insight was deceptively simple: if one could explicitly construct the mysterious inner function ψ, Kolmogorov's abstract theorem would become a practical algorithm for neural networks [6,7]. His proof proceeded through several intricate steps, each building upon the previous in what appeared to be mathematical elegance.

2.1 The Foundation - Terminating Rationals

Sprecher began with terminating rational numbers—ordinary decimals like 0.123 or 0.5—represented as dk = Σ(r=1 to k) ir/γʳ, where ir denotes the r-th digit in base γ [1]. This choice was crucial: rational numbers form a dense subset of the real line, meaning they come arbitrarily close to any real number, yet remain computationally tractable.

His central construction defined:

ψ(dk) = Σ(r=1 to k) ĩᵣ2^(-mᵣ)γ^(-β(r-mᵣ))

where β(r) = (nʳ - 1)/(n - 1) relates to the dimension n of the problem. This formula embodies Sprecher's key insight: each digit ir should contribute independently to ψ(dk), with contributions carefully weighted to prevent unwanted clustering.

2.2 The Digit Modification Scheme

The construction's sophistication emerged through its treatment of individual digits. Sprecher defined modified digits via [1]:

ĩᵣ = iᵣ - (γ-2)⟨iᵣ⟩

where the bracket notation ⟨iᵣ⟩ equals 1 when iᵣ = γ-1 and 0 otherwise. This seemingly innocuous modification was designed to handle the special case when a digit reaches its maximum value γ-1, preventing certain pathological behaviours in the function's growth.

Simultaneously, he introduced shift parameters:

mᵣ = ⟨iᵣ⟩(1 + Σ(s=1 to r-1)[is] × ... × [ir-1])

where [iᵣ] equals 1 when iᵣ ∈ {γ-2, γ-1} and 0 otherwise. These parameters controlled the "shifting" of contributions, ensuring that the 2^(-mᵣ) terms would create appropriate spacing between function values at different rational points.

2.3 The Separation Strategy

Sprecher's proof required demonstrating that his construction satisfied two crucial properties [1]. First, the translation property:

ψ(dk + q/(γ(γ-1))) = ψ(dk) + q·Σ(r=k+1 to ∞) γ^(-β(r))

This lemma promised that shifting inputs by specific amounts would yield predictable output changes—essential for the superposition structure to work properly.

Second, the minimal separation lemma guaranteed:

min|μk| ≥ γ^(-nβ(k))

where μk = Σ(p=1 to n) αₚ[ψ(dk,p) - ψ(d'k,p)] measures differences between ψ values at nearby rational points. This separation bound ensured that different n-tuples of rational inputs would map to sufficiently distinct outputs, preventing the catastrophic clustering that would destroy the superposition's effectiveness.

2.4 The Algorithmic Implementation

With the inner function ψ constructed, Sprecher's algorithm for computing superpositions proceeded through three carefully orchestrated phases [1,2]:

Phase I - Discretisation: For any continuous function f, choose a fineness parameter k ensuring |f(x) - f(x')| ≤ ε‖f‖ whenever ‖x - x'‖ ≤ γ^(-k). This condition essentially makes the computational grid fine enough that f barely changes between nearby points, allowing discrete approximation without significant error.

Phase II - Inner Function Computation: For each q = 0,1,...,2n, evaluate ψ(dk,p + qa) at all grid points and form the crucial linear combinations:

ξ(xq) = Σ(p=1 to n) αₚψ(xₚ + qa)

These represent the "hidden layer" computations in Hecht-Nielsen's neural network interpretation [6,7], transforming n-dimensional inputs into 2n+1 one-dimensional values.

Phase III - Outer Function Construction: Define the outer functions through:

Φq(yq) = (1/(m+1))Σ f(dkᵣ)ω(dkᵣ; yq)

where ω serves as a carefully constructed window function—equalling 1 in specific intervals and 0 elsewhere—that enables the reconstruction of f from the transformed inner representations.

3. The Fatal Flaws: Where Sprecher Went Wrong

3.1 The Monotonicity Catastrophe

Sprecher's proof assumed without demonstration that his construction yielded a monotonic function. This assumption proved catastrophically false. Mario Köppen's devastating 2002 analysis [4] revealed the counter-example that shattered Sprecher's edifice: with γ = 10 and n = 2, ψ(0.58999) = 0.55175 whilst ψ(0.59) = 0.55.

This violated monotonicity since 0.58999 < 0.59 but ψ(0.58999) > ψ(0.59). The implication was severe: if ψ fails to preserve order, the entire superposition structure collapses, as the careful separation required by Kolmogorov's theorem vanishes.

3.2 The Continuity Crisis

Equally damaging, Sprecher assumed his function would extend continuously from rational numbers to real numbers. Investigation revealed discontinuous jumps at points like x = 0.i₁9 throughout the interval [0,1] [4,5]. These discontinuities arose because Sprecher's additive construction treated each digit independently, failing to account for the subtle interactions that occur in base-γ representations.

3.3 The Root Cause: Carry-Over Effects

The fundamental flaw lay in Sprecher's additive formula's inability to handle "carry-over" effects when digits transitioned from γ-2 to γ-1 [5]. His construction ψ(dk) = Σ(r=1 to k) ĩᵣ2^(-mᵣ)γ^(-β(r-mᵣ)) assumed each digit's contribution could be computed independently. However, the transition from, say, 0.589 to 0.590 involves complex interdependencies between digit positions that Sprecher's additive approach couldn't capture.

What appeared to be the construction's strength—the clean separation of digit contributions—proved to be its fundamental weakness. The modified digit scheme ĩᵣ = iᵣ - (γ-2)⟨iᵣ⟩ attempted to handle the γ-1 case, but did so inadequately, creating precisely the discontinuities and monotonicity violations that would doom the proof.

4. The Correction: Köppen's Recursive Remedy

Köppen's elegant solution [4] replaced Sprecher's additive formula with a recursive construction that handled boundary cases through averaging:

ψk(dk) = {
dk when k = 1
ψk-1(dk - ik/γᵏ) + ik/γᵝ⁽ᵏ⁾ when ik < γ-1
½[ψk-1(dk - ik/γᵏ) + ψk-1(dk + 1/γᵏ) + ik/γᵝ⁽ᵏ⁾] when ik = γ-1
}

The crucial insight lay in the third case: when ik = γ-1, instead of computing the contribution independently, Köppen averaged the function values at two nearby points. This averaging smoothed over the problematic transitions that caused Sprecher's discontinuities whilst preserving the essential separation properties required for the superposition theorem.

5. The Mathematical Resolution

The complete vindication arrived through Jürgen Braun and Michael Griebel's meticulous 2009 work [5]. Their proof strategy demonstrated that Köppen's corrected function possessed all required properties through three key steps:

Existence via Cauchy Sequences: They proved that the recursive definition converged by showing the sequence {ψk} formed a Cauchy sequence, with |ψk - ψk'| → 0 as k,k' → ∞, guaranteeing the existence of a well-defined limit function ψ.

Monotonicity through Careful Estimates: Using the averaging property in König's construction, they established that ψ⁺k ≥ ψk + 1/γᵝ⁽ᵏ⁾, where ψ⁺k represents the function value at the next rational point. This inequality propagated through the recursive structure to yield monotonicity of the limit function.

Continuity from Uniform Convergence: The recursive averaging ensured that the convergence was sufficiently uniform to preserve continuity, with explicit bounds on |ψ(x) - ψ(x₀)| in terms of |x - x₀|.

6. Conclusion: From Failure to Triumph

This mathematical journey from failure to triumph exemplifies how science advances through the interplay of ambition and rigour. Sprecher's flawed but visionary proof inspired a generation of researchers to bridge the gap between abstract mathematics and computational reality, ultimately leading to the 2024 breakthrough of Kolmogorov-Arnold Networks [9] that now power cutting-edge machine learning applications [10].

The irony of failures, remains delicious.

References

[1] D. A. Sprecher, "A numerical implementation of Kolmogorov's superpositions," Neural Networks, vol. 9, no. 5, pp. 765-772, 1996.

[3] A. N. Kolmogorov, "On the representation of continuous functions of many variables by superpositions of continuous functions of one variable and addition," Dokl. Akad. Nauk USSR, vol. 14, no. 5, pp. 953-956, 1957.

[4] M. Köppen, "On the training of a Kolmogorov network," in Artificial Neural Networks — ICANN 2002, vol. 2415, Lecture Notes in Computer Science, Berlin: Springer-Verlag, 2002, pp. 474-479.

[5] J. Braun and M. Griebel, "On a constructive proof of Kolmogorov's superposition theorem," Constructive Approximation, vol. 30, no. 3, pp. 653-675, 2009.

[9] Z. Liu et al., "KAN: Kolmogorov-Arnold networks," arXiv preprint arXiv:2404.19756, 2024.

[10] J. K. Actor, "Error bounds for deep ReLU networks using the Kolmogorov–Arnold superposition theorem," Neural Networks, vol. 129, pp. 1-6, 2019.