Next Article in Journal
On the Change of Measure for Brownian Processes
Previous Article in Journal
The Black Hole Information Problem
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Kolmogorov GAM Networks Are All You Need!

1
Oxford University, Oxford OX1 2JD, UK
2
Department of Systems Engineering and Operations Research, George Mason University, Fairfax, VA 22030, USA
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(6), 593; https://doi.org/10.3390/e27060593
Submission received: 3 January 2025 / Revised: 15 May 2025 / Accepted: 19 May 2025 / Published: 31 May 2025
(This article belongs to the Section Signal and Data Analysis)

Abstract

Kolmogorov GAM (K-GAM) networks have been shown to be an efficient architecture for both training and inference. They are additive models with embeddings that are independent of the target function of interest. They provide an alternative to Transformer architectures. They are the machine learning version of Kolmogorov’s superposition theorem (KST), which provides an efficient representation of multivariate functions. Such representations are useful in machine learning for encoding dictionaries (a.k.a. “look-up” tables). KST theory also provides a representation based on translates of the Köppen function. The goal of our paper is to interpret this representation in a machine learning context for applications in artificial intelligence (AI). Our architecture is equivalent to a topological embedding, which is independent of the function, together with an additive layer that uses a generalized additive model (GAM). This provides a class of learning procedures with far fewer parameters than current deep learning algorithms. Implementation can be parallelizable, which makes our algorithms computationally attractive. To illustrate our methodology, we use the iris data from statistical learning. We also show that our additive model with non-linear embedding provides an alternative to Transformer architectures, which, from a statistical viewpoint, are kernel smoothers. Additive KAN models, therefore, provide a natural alternative to Transformers. Finally, we conclude with directions for future research.

1. Introduction

The landscape of modern machine learning has been shaped by exponential growth in computational power, particularly through advances in GPU technology and frameworks like PyTorch. While Moore’s law continues to drive hardware improvements, and CUDA algorithms have revolutionized our ability to process vast amounts of internet data, we pose the following question: can we achieve superior performance through mathematically efficient representations of multivariate functions rather than raw computational power?
In this paper, we introduce Kolmogorov-generalized additive models (K-GAM), a novel neural network architecture whose additive structure enables simplified training and accelerated inference through a marked reduction in parameters compared to traditional approaches. K-GAMs leverage Kolmogorov’s superposition theorem by utilizing a composition of two key components, namely, a universal Köppen embedding (functioning as a space-filling curve), followed by a trainable outer function, g : [ 0 , 1 ] [ 0 , 1 ] . Unlike the original iterative look-up table approach proposed by Köppen and Sprecher, which faces NP-hard computational challenges, we implement the outer function using a ReLU neural network that can be efficiently trained using standard optimization techniques.
A fundamental challenge in machine learning lies in effectively handling high-dimensional input–output relationships. This challenge manifests itself in two distinct but related tasks. First, one task is to construct a “look-up” table (dictionary) for the fast search and retrieval of input–output examples. This is an encoding and can be thought of as a data compression problem. Second, and perhaps more importantly, we must develop prediction rules that can generalize beyond these examples to handle arbitrary inputs.
More formally, we seek to find a good predictor function f ( x ) that maps an input x to its output prediction y. In practice, the input x is typically a high-dimensional vector:
y = f ( x ) where x = ( x 1 , , x d )
Given a training dataset ( y i , x i ) i = 1 N of example input–output pairs, our goal is to train a model, i.e., to find the function f. The key question is as follows: how do we represent a multivariate function so as to obtain a desirable f?
We demonstrate that Transformer architectures can be decomposed into two fundamental operations: a particular form of embedding followed by kernel smoothing. Hence, K-GAM models are natural competitors to Transformers. The significance of this connection becomes clear when we consider that Transformers have largely supplanted alternative architectures across a wide range of applications. Previous work in this direction—[1,2]—demonstrated the potential of Köppen-based approaches only for 2 × 2 image reconstruction tasks. Our analysis extends this framework to encompass A wider spectrum of machine learning problems.
The rest of this paper is outlined as follows. Section 2 discusses the Kolmogorov superposition theorem. Section 3 introduces Kolmogorov-generalized additive models (K-GAMs). Section 4 describes Transformer architectures as kernel smoothers. Section 5 provides an application of GAM-Kolmogorov embeddings to the iris dataset. Section 6 concludes with directions for future research.

2. Kolmogorov Superposition Theorem (KST)

Kolmogorov demonstrated that any real-valued continuous function f ( x ) where x = ( x 1 , , x d ) defined on E d can be represented as a convolution of two single-variable functions:
f ( x 1 , , x d ) = q = 0 2 d g q Ψ q ( x 1 , , x d )
where g q are continuous single-variable functions defined on the d-dimensional unit cube E d = [ 0 , 1 ] d . Kolmogorov further showed that the ϕ q functions can be decomposed into sums of single-variable functions:
Ψ q ( x 1 , , x d ) = i = 1 d ψ q , i ( x i )
This result is known as the Kolmogorov representation theorem [3] and is often written in the following form:
f ( x 1 , , x d ) = q = 0 2 d g q i = 1 d ψ q , i ( x i )
Kolmogorov originally defined the inner functions as Ψ q ( x 1 , , x d ) , q = 0 , 1 , , 2 d , using the following property ψ q ( S q , k ) ψ q ( S q , k ) = , k k , q , = 0 , 1 , , 2 d . Then, for any integer d 2 , let S q , k be a family of mutually disjoint d-dimensional cubes S q , k S q , k = , k k , k , k = 1 , 2 , , q = 0 , 1 , , 2 d . The diameter of each cube approaches zero as k , and every point of a unit cube [ 0 , 1 ] d belongs to at least d + 1 cubes S q , k for each k. Unfortunately, this derivation is not constructive and does not provide a practical algorithm for constructing the inner functions ϕ q .
Our goal is to show that it applies directly to the problems of machine learning. Furthermore, we interpret the architecture as a statistical model that allows us to draw on uncertainty quantification and approximation bounds. To understand the power of this representation from a machine learning perspective, we can think of ψ q , i ( x i ) as features. Note that the number of features is then 2 d + 1 for the d-dimensional input. This representation is also remarkable in the sense that ψ q ( x ) can be viewed as an embedding, which is independent of f. For details, see Section 3.

2.1. Kolmogorov–Arnold Networks

A significant development has been the emergence of Kolmogorov–Arnold networks (KANs). The key innovation of KANs is their use of learnable functions rather than weights on the network edges. This replaces traditional linear weights with univariate functions, typically parametrized by splines, enhancing both representational capacity and interpretability. Ref. [4] established the first practical connection between KST and neural networks by showing that any KAN can be constructed as a three-layer MLP. Polson and Sokolov proposed Bayesian KAN networks.
Recently, Refs. [5,6] considered KST in the form of a sum of functions, represented as a two-layer model:
f ( x ) = f ( x 1 , , x d ) = ( g Φ ) ( x )
More explicitly,
f ( x 1 , , x d ) = q = 0 2 d p = 1 d ψ q , p ( x p )
Ref. [7] presented a significant theorem regarding the approximation of generalized bandlimited multivariate functions using deep ReLU networks. The key finding is that deep ReLU networks can approximate generalized bandlimited functions without suffering from the curse of dimensionality. The paper utilizes an extension of Carathéodory’s theorem to infinite-dimensional spaces, which is crucial for the approximation of bandlimited functions. As a result, this paper provides a theoretical foundation for the effectiveness of deep learning in high-dimensional problems, especially for a class of functions relevant to many scientific computing applications.
One can think of KAN networks as NNs in which functions ψ q , p are estimated at the nodes, rather than weights. This makes them very flexible. A natural choice involves the use of splines for ψ q , p ; Ref. [8] proposed the application of KSNs (Kolmogorov spline networks).
Another advantage is that they can be trained using fast, high-dimensional algorithms such as the Newton–Kaczmarz algorithm [9]. The authors showed that KAN is not only a superposition of functions but also a special case of a tree of discrete Urysohn operators:
U ( x 1 , , x d ) = p = 1 d g p ( x p )
This insight leads to a fast, scalable algorithm that avoids backpropagation, is applicable to any GAM model, and uses a projection descent method with a Newton–Kaczmarz scheme.
Recent research has explored various functional classes for the outer functions, including Wan-KAN [10], SineKAN [11], and functional combinations [12]. Ref. [13] demonstrated that KAN networks can outperform traditional MLPs, while [14] proved their optimality in terms of approximation errors. Ref. [15] discussed the Bayesian interpretation of the Kolmogorov–Arnold representation. KAN networks are widely trained with gradient descent. Ref. [9] proposed a Newton–Kaczmarz scheme. Bayesian methods are also available [16] and offer the potential for fewer parameters and more efficient inference. There are also optimal posterior concentration results [17].
Early theoretical groundwork for adaptive learning in KANs was laid by [18], who introduced the concept of counter-propagation networks, where outer functions self-organize in response to input–output pairs ( x , y ) . Theoretical understanding has continued to advance, with [19] showing that Kolmogorov networks with two hidden layers can precisely represent continuous, discontinuous, and unbounded multivariate functions, depending on the activation function choice in the second layer. Ref. [20] extended this to three-layer networks for discontinuous functions.

2.2. Inner and Outer Functions

The theorem has seen several refinements over time. Ref. [21] made a crucial advancement by proving that the inner functions could be Hölder continuous. Ref. [22] later strengthened this result, showing that these functions could be Lipschitz continuous, although this required modifications to both the outer and inner functions.
We focus on the Kolmogorov–Sprecher formulation, as modified by [23,24,25]. The first representation is as follows:
f ( x 1 , , x d ) = q = 0 2 d g q ( Ψ q ( x ) ) where Ψ q ( x ) = p = 1 d α p ψ ( x p + q α )
We will call this the Kq-GAM model to indicate the fact that there are q univariate outer functions. Ref. [7] provided an iterative algorithm to find g q based on the original Köppen formulation.
The second (and most parsimonious) representation is one in which there is only one univariate outer function! This expresses any multivariate continuous function as follows:
f ( x 1 , , x d ) = q = 0 2 d g ( Ψ q ( x ) ) where Ψ q ( x ) = p = 1 d α p ψ ( x p + q a ) + δ q .
Ref. [25] defines the δ k term.
We will refer to this as the K-GAM model in our ML implementation. Bryant discussed the choice of offset δ q . Note that the complexity reduction is that there is only one g : [ 0 , 1 ] [ 0 , 1 ] outer function that depends on f. The caveat involves finding a computation algorithm to compute g in polynomial time.
Here, the inner functions ψ q are constructed as sums of translated versions of a single univariate function ψ , known as the Köppen function. The Köppen function is monotone, Hölder continuous with smoothness parameters, and has fractal-like properties. Recent work by [26] provides a modern construction of the Köppen function and establishes theoretical approximation results with ReLU networks. Ref. [27] extended these theoretical results and showed that a function with a smaller number of parameters can be used. The author used λ -adic representation, where
f ( x 1 , , x d ) = g p = 1 d λ p ψ ( x p ) .
The structure of ψ q differs fundamentally from the approaches used in traditional neural networks, which rely on hyperplanes (as in ReLU networks) or decision trees. In ψ q , a slice in any given univariate direction cuts the regions defined by the embedding in half.
The inner functions, ψ q , partition the input space into distinct regions, and the outer function, g, must be constructed to provide the correct output values across the regions that the inner function defines. The outer function, g, can be determined via a computationally intensive process of averaging. For each input configuration, the inner functions, ψ q , generate a unique encoding, and g must map this encoding to the appropriate value of f ( x ) . This creates a dictionary-like structure that associates each region with its corresponding output value. Köppen made significant contributions by correcting Sprecher’s original proof of this construction process, with improvements to the computational algorithm later suggested by [28,29]. Braun further enhanced the understanding by providing precise definitions of the shift parameters δ k and characterizing the topological structure induced by ψ q .
A fundamental trade-off in KST exists between function smoothness and dimensionality. The inner functions, ψ p , q , can be chosen from two different function spaces, with each offering distinct advantages. The first option is to use functions from C 1 ( [ 0 , 1 ] ) , but this limits the network’s ability to handle higher dimensions effectively. The second option is to relax the smoothness requirement to Hölder-continuous functions ( ψ p , q Holder β ( [ 0 , 1 ] ) ), which satisfy the inequality | ψ ( x ) ψ ( y ) | < | x y | β . These functions are less smooth, but this “roughness” enables better approximation in higher dimensions.

2.3. Ridge and Projection Pursuit Regression

To understand the significance of this trade-off, we consider ridge functions, which represent a fundamental building block in multivariate analysis. Since our ultimate goal is to model arbitrary multivariate functions f, we need a way to reduce dimensionality while preserving the ability to capture nonlinear relationships. Ridge functions accomplish this by representing one of the simplest forms of nonlinear multivariate functions, requiring only a single linear projection and a univariate nonlinear transformation. Formally, a ridge function f : R d R takes the form f ( x ) = g ( w T x ) , where g is a univariate function and x , w R d . The non-zero vector w is called the direction. The term “ridge” reflects a key geometric property, that is, the function remains constant along any direction orthogonal to w. Specifically, for any direction u, such that w T u = 0 , we have the following:
f ( x + u ) = g ( w T ( x + u ) ) = g ( w T x ) = f ( x )
This structural simplicity makes ridge functions particularly useful as building blocks for high-dimensional approximation. Ref. [30] provides bounds on the approximation error when a function with a finite spectral norm is approximated by ridge functions.
Ridge functions play a central role in high-dimensional statistical analysis. For example, projection pursuit regression approximates input–output relations using a linear combination of ridge functions [31,32,33]:
f ( x ) = q = 0 2 d g q ( w q T x ) ,
where both the directions, w i , and functions, g i , are variables, and w i T x denotes one-dimensional projections of the input vector. The vector w i T x is a projection of the input vector x onto a one-dimensional space, and g i ( w i T x ) denotes a feature calculated from the data. Ref. [34] used nonlinear functions of linear combinations, laying important groundwork for deep learning.

3. Kolmogorov-Generalized Additive Models (K-GAMs)

We take a different approach. Rather than using learnable functions as network node activations, we directly use the KST representation. This is a two-layer network with a non-differentiable inner function. The network’s architecture can be expressed as follows:
f ( x 1 , , x d ) = q = 0 2 d g q ( z q )
where the inner layer performs an embedding from [ 0 , 1 ] d to R 2 d + 1 via the following;
z q = η q ( x 1 , , x d ) = p = 1 d λ p ψ ( x p + q a )
Here, λ p = r = 1 γ ( p 1 ) β ( r ) is a p-adic expansion with β ( r ) = ( d r 1 ) / ( d 1 ) and γ d + 2 , with a = ( γ ( γ 1 ) ) 1 .

3.1. Köppen Function

The Köppen function ψ straightforwardly computes and provides the basis for the inner architecture of the network. The function ψ is defined through a recursive limit:
ψ ( x ) = lim k ψ k l = 1 k i l γ l
where each x [ 0 , 1 ] has the following representation:
x = l = 1 i l γ l = lim k l = 1 k i l γ l
and ψ k is defined recursively as follows:
ψ k ( d ) = d , d D 1 ψ k 1 ( d i k γ k ) + i k γ β d ( k ) , d D k , k > 1 , i k < γ 1 1 2 ψ k 1 ( d i k γ k ) + ψ k 1 ( d + γ k ) + i k γ β d ( k ) , d D k , k > 1 , i k = γ 1
Here, the sets D k are defined as follows. D k is a uniform grid, with step size γ k , defined by the following:
D + k = { i γ k , 0 i γ k 1 } [ 0 , 1 )
There are γ k different points 0 x 1 γ k on each grid D k where each point is as follows: x = l = 1 l i l γ l , i { 0 , 1 , , γ 1 } .
To illustrate, we show how the iterates update. Figure 1 shows the plot of the Köppen function ψ k for k = 3 on intervals [ 0 , 1 ] and [ 0 , 0.2 ] (top row). The bottom row shows a “zoomed in” view of the function for k = 4 and k = 5 on the intervals [ 0 , 0.2 ] . The function has a fractal-like structure, with the number of discontinuities increasing with k. The Köppen function is a key component of the KST representation, providing a topological embedding of the input space that is independent of the target function f. This embedding is crucial for the network’s ability to capture complex multivariate relationships. Ref. [28] provided a detailed algorithm for computing the Köppen function and its properties. The author also provided a good, intuitive understanding of the function and its properties.
The most striking aspect of KST is that it leads to a generalized additive model (GAM) with fixed features that are independent of the target function f. These features, determined by the Köppen function, provide universal topological information about the input space, effectively implementing a k-nearest neighbors structure that is inherent to the representation. The outer function, g, is then responsible for learning the relationship between these features and the target function f. This separation of feature engineering and learning is a key advantage of K-GAM networks, enabling efficient training and inference.
Theorem 1
(K-GAM). Any function and dataset can be represented as a GAM with feature engineering (topological information) given by features z k in the hidden layer:
y ( i ) = q = 0 2 d g ( z q ( i ) )
z q ( i ) = p = 1 d λ q ψ ( x p ( i ) + q a ) + δ q
where ψ is a single activation function common to all nodes, known as the Köppen function, and g is a single outer function.
Proof. 
We can take the Köppen–Sprecher representation for y = f ( x ) and apply it N times to each output–input pair ( y ( i ) , x ( i ) ) and obtain a representation for y ( i ) = f ( x ( i ) ) . For example, for the K-GAM model and g-outer function, we have the following:
f ( x 1 , , x d ) = q = 0 2 d g p = 1 d λ q ψ ( x p + q a ) + δ q
where λ 1 = 1 > λ 2 > λ d and a = 1 / ( 2 d + 1 ) ( 2 d + 2 ) .
Now, for a dataset { ( y ( i ) , x ( i ) ) } i = 1 N where y ( i ) = f ( x ( i ) ) , we apply this representation to each data point:
y ( i ) = f ( x 1 ( i ) , , x d ( i ) ) = q = 0 2 d g p = 1 d λ q ϕ ( x p ( i ) + p a ) + δ q = q = 0 2 d g z q ( i )
where we define z q ( i ) = p = 1 d λ q ϕ ( x p ( i ) + p a ) + δ q .
Ref. [7] proved that the outer function g can be approximated by a ReLU network to arbitrary precision. Therefore, we can replace g with a single ReLU network:
g ( x ) = k = 1 K β k ReLU ( w k x + b k ) ,
where K is the number of neurons. We see that any function, f, and its corresponding dataset can be represented as a GAM with features engineered through the Köppen function ϕ .    □
Ref. [35] further developed this insight into a specialized two-layer ReLU network architecture. Specifically, the authors introduced a class of functions called Kolmogorov–Lipschitz continuous, where the outer function in the KST representation is Lipschitz continuous. Then they demonstrated that KL continuous functions can be approximated using a ReLU neural network with two hidden layers, achieving a dimension-independent approximation rate, and they introduced LKB-splines, created by using linear B-splines to replace the outer function in the KST representation.
As in [36], we can decompose with Lipschitz continuous functions. Note, the number of outer functions grows with the input dimension, d, not with the dataset size. The outer function, g, is shared across all terms, making this a true GAM representation with universal feature engineering provided by the inner Köppen function ϕ .
In a similar vein to [4], we show how this functional representation can be used to define a statistical model for applications in machine learning tasks [37,38]. We propose the following model:
f ( x 1 , , x d ) = q = 0 2 d g p = 1 d α p ψ ( x p + q α ) + δ q
g ( x ) = k = 1 K β k ReLU ( w k x + b k )
where K denotes the number of neurons in the outer function’s architecture. The parameters ( β k , w k , b k ) k = 1 K can be learned using L 2 -minimization and SGD. This architecture employs only two univariate activation functions, that is, a learned ReLU network for the outer layer and the Köppen function for the inner layer, which can be computed independently of f.

Note

A crucial aspect of KST-based architectures is their relationship to p-adic expansions and embeddings. This connection, first explored by [39] and recently advanced by [40], builds upon the foundational work of [41] on cellular neural networks in the 1980s, which included cellular automata as special cases.
For any input x, its p-adic expansion can be written as follows:
x = l = 1 i l γ l = lim k l = 1 k i l γ l
where γ serves as the base of the expansion and i l denotes the digits in this representation. This expansion provides a natural embedding of the input space into a higher-dimensional representation, resembling contemporary kernel embedding techniques in deep learning [15,37,38].

3.2. Inference

An important property of an algorithm is how many parameters exist in the evaluation of the network, the so-called inference problem. The K-GAM network is notable for its relatively sparse parameterization.
Consider the class that corresponds to kernel estimators of the following form:
f ( x ) = K ( x , t ) d μ ( t ) with d μ ( t ) K and x = ( x 1 , , x d )
We call K ( x , t ) = K t ( x ) an L -atom if | K t ( x ) | 1 . For such estimators, there exists an M-term approximation of f:
f M ( x ) = j = 1 M a j K ( x , t j ) such that f f N C M 1 2 ,
with C independent of dimension d. While general approximations suffer from the curse of dimensionality with error O ( M r / d ) [42], where r denotes isotropic smoothness, certain restrictions yield dimension-independent bounds. For instance, when f is in the L 1 -Fourier class, the error becomes O ( M 1 / 2 ) independent of d [43].
The class of functions that we consider can be represented as L 1 -combinations of L -atoms. For f M , we will use superpositions of functions. For example, with t : = ( x 0 , s ) and Gaussian atoms K ( x , t ) = exp x x 0 2 / s 2 , see [44]. Let t = ( x 0 , k ) , where k is an orthant indicator. Then the resultant approximation is also of order O ( M 1 / 2 ) , independent of d.
This framework naturally extends to incorporate dimension-reduction techniques like partial least squares (PLS). Given the input–output pairs ( x i , y i ) i = 1 N , we can construct models of the following form:
Y = g ( B ^ P L S X )
where B ^ P L S provides dimension reduction and g is a learnable transformation.
A more general class is found by the superposition of 2 d functions, each monotone for a different orthant. Then, the condition d | μ | ( t ) 1 manifests as a derivative condition:
d f x 1 x d L 1 .
This characterizes the space of functions with bounded mixed derivatives, which naturally leads to β -Hölder-smooth functions. For such β -Hölder-continuous functions, the network typically requires O ( M β ) parameters (see [45] for further results on additive models and approximation properties in high dimensions).
Kolmogorov spline networks, designed for differentiable functions with bounded derivatives, satisfy the stronger property:
f f M = O ( M 1 )
with the number of parameters given by O ( M 2 / 3 ) . This compares favorably to both the O ( M 1 / 2 ) approximation rate and the O ( M 2 ) parameter count of the general one-layer-hidden feed-forward networks [8].
Deep ReLU networks, using max ( x , 0 ) activation, create separating hyperplanes across L layers, unlike cylinder sets (a.k.a. trees), which are a special case. Ref. [46] showed that energy-norm-based sparse grids and Kolmogorov approximation schemes have bounds that are independent of d.
Ref. [47] showed that efficient approximation schemes can be developed by the use of superpositions. Every continuous function of several variables can be represented by the superposition of functions with only two variables. Ref. [20] considered discontinuous functions.
As Kolmogorov might have said—there are no true multivariate problems, only superpositions of univariate affine ones!

3.3. Kernel Smoothing: Interpolation

The theory of kernel methods was developed by Fredholm in the context of integral equations [48]. The idea is to represent a function as a linear combination of basis functions, which are called kernels.
f ( x ) = a b K ( x , x ) d μ ( x ) d x where x = ( x 1 , , x d )
Here, the unknown function, f ( x ) , is represented as a linear combination of kernels, K ( x , x ) , with unknown coefficients, ϕ ( x ) . The kernels are known, and the coefficients are unknown. The coefficients are found by solving the integral equation. The first work in this area was done by Abel [49], who considered equations of the form above.
Nowadays, we call those equations Volterra integral equations of the first kind. Integral equations typically arise in inverse problems. Their significance extends beyond their historical origins, as kernel methods have become instrumental in addressing one of the fundamental challenges in modern mathematics, that is, the curse of dimensionality.
Bartlett [50,51] proposed the use of kernels to estimate the regression function. The idea is to estimate the regression function, f ( x ) , at point x by averaging the values of the response variable, y i , at points x i that are close to x. The kernel is used to define the weights.
The regression function is estimated as follows:
f ^ ( x ) = i = 1 d y i K ( x , x i ) / i = 1 d K ( x , x i ) ,
where the kernel weights are normalized.
Both Nadaraya [50] and Watson [51] considered the symmetric kernel K ( x , x ) = K ( x x 2 ) , where | | · | | 2 is the Euclidean norm. The most popular kernel of that sort is the Gaussian kernel:
K ( x , x ) = exp x x 2 2 2 σ 2 .
Alternatively, the two-norm can be replaced by the inner product: K ( x , x ) = exp x T x / 2 σ 2 .
Later, Ref. [52] proposed using kernels to estimate the density function. The idea is to estimate the density function f ( x ) at point x by averaging the values of the kernel K ( x , x i ) at points x i that are close to x. This idea was applied in many contexts by statisticians [53,54], machine learners [55], and engineers [56]. K-GAM builds upon these foundations.
Kernel methods are supported by numerous generalization bounds, which often take the form of inequalities that describe the performance limits of kernel-based estimators. A particularly important example is the Bayes risk for k-nearest neighbors (k-NN), which can be expressed in a kernel framework as follows:
f ^ ( x ) = i = 1 N w i y i where w i : = K ( x i , x ) / i = 1 N K ( x i , x )
k-NN classifiers have been proven to converge to an error rate that is bounded in relation to the Bayes error rate, with the exact relationship depending on the number of classes. For binary classification, the asymptotic error rate of k-NN is, at most, 2 R * ( 1 R * ) , where R * is the Bayes error rate. This theoretical bound suggests potential for improvement in practice. Cover and Hart proved that interpolated k-NN schemes are consistent estimators, meaning that their performance improves with the increasing sample size.

3.4. Training Rates

Consider the non-parametric condition regression, y i = f ( x i ) + ϵ i where x i = ( x 1 i , , x d i ) . We wish to estimate f ( x 1 , , x d ) where x = ( x 1 , , x d ) [ 0 , 1 ] d . From a classical risk perspective, we define the following:
R ( f , f ^ N ) = E X , Y f f ^ N 2
where . denotes the L 2 ( P X ) -norm.
Under standard assumptions, we have an optimal minimax rate inf f ^ sup f R ( f , f ^ N ) of O p N 2 β / ( 2 β + d ) for β -Hölder-smooth functions f. This rate depends on the dimension d, which can be problematic in high-dimensional settings. By restricting the class of functions, better rates can be obtained, including ones that do not depend on d. In this sense, we avoid the curse of dimensionality. Common approaches include considering the class of linear superpositions (a.k.a. ridge functions) and projection pursuit models.
Another asymptotic result comes from a posterior concentration property. Here, f ^ N is constructed as a regularized MAP (maximum a posteriori) estimator, which solves the optimization problem:
f ^ N = arg min f ^ N 1 N i = 1 N ( y i f ^ N ( x i ) ) 2 + ϕ ( f ^ N )
where ϕ ( f ^ ) is a regularization term. Under appropriate conditions, the ensuing posterior distribution Π ( f | x , y ) can be shown to concentrate around the true function at the minimax rate (up to a log N factor).
A key result in the deep learning literature provides convergence rates for deep neural networks. Given a training dataset of input–output pairs ( x i , y i ) i = 1 N from the model y = f ( x ) + ϵ , where f is a deep learner (i.e., superposition of functions)
f = g L g 1 g 0
where each g i is a β i -smooth Hölder function with d i variables, we have the following condition:
| g i ( x ) g i ( y ) | < | x y | β .
Then, the estimator has an optimal rate:
O max 1 i L N 2 β * / ( 2 β * + d i ) where β i * = β i l = i + 1 L min ( β l , 1 )
This result can be applied to various function classes, including generalized additive models of the following form:
f 0 ( x ) = h p = 1 d f 0 , p ( x p )
where g 0 ( z ) = h ( z ) , g 1 ( x 1 , , x d ) = ( f 01 ( x 1 ) , , f 0 d ( x d ) ) and g 2 ( y 1 , , y d ) = i = 1 d y i . In this case, d 1 = d 2 = 1 , and assuming h is Lipschitz, we obtain an optimal rate of O ( N 1 / 3 ) , which is independent of d.
Ref. [57] shows that deep ReLU networks also have an optimal rate of O ( N 1 / 3 ) for certain function classes. For 3-times differentiable (e.g., cubic B-splines ), Ref. [58] found a rate of O ( N 3 / 7 ) = O ( N 3 / ( 2 × 3 + 1 ) ) . Ref. [8] found a rate O ( N 1 ) for Kolmogorov spline networks.
Finally, it is worth noting the relationship between expected risk and empirical risk. The expected risk, R, is typically bounded by the empirical risk plus a term of order 1 / N :
R ( y , f ) 1 N i = 1 N R ( y i , f ( x i ) ) + O f N
where f is the minimizer of the expected risk. However, in the case of interpolation, where the model perfectly fits the training data, the empirical risk term becomes zero, leaving only the O ( 1 / N ) term.

4. Transformers as Kernel Smoothing

The fundamental connection between K-GAMs and Transformers becomes clear when we decompose Transformer architectures into their core operations, that is, embedding and kernel smoothing. This decomposition reveals why K-GAMs serve as natural competitors to Transformers, as both architectures leverage these same fundamental principles, albeit through different mechanisms. In both architectures, the embedding phase transforms raw inputs into a higher-dimensional space, where relationships can be more easily captured, while the smoothing phase aggregates information across this embedded space to produce outputs.
Transformers have become a main building block for various natural language processing (NLP) tasks and have been extended to other domains as well due to their effectiveness. The Transformer architecture is primarily designed to handle sequential data, making it well-suited for tasks such as machine translation, language modeling, text generation, and more.
The Transformer approach to embedding begins by converting input tokens into vectors in a high-dimensional space, augmented with positional information to preserve sequential structure. K-GAMs, in contrast, employ the Köppen function to transform inputs into a universal embedding space. This difference in embedding strategy reflects a fundamental trade-off, that is, Transformers learn task-specific embeddings that can adapt to particular domains, while K-GAMs leverage a universal embedding that can theoretically capture any continuous function.
Next, we will discuss the role of kernel smoothing in Transformers. Ref. [59] first introduced kernel smoothing for sequence-to-sequence learning through what became known as the attention mechanism. This approach estimates the probability of the next word in the sequence using a so-called context vector c i , which is a weighted average of the vectors from the input sequence h j :
c i = j = 1 n α i j h j ,
where α i j denotes the weights. The weights are defined by a kernel function, specifically a normalized exponential (softmax):
α i j = exp e i j k = 1 n exp e i k .
When used for self-supervised learning, this attention mechanism is called self-attention. When a sequence is mapped to a matrix M, it is called multi-head attention.
This formulation mirrors the structure of traditional kernel smoothing methods, with the key innovation being that, rather than using a fixed similarity measure like Euclidean distance, the energy function e i j = a ( s i 1 , h j ) is learned through a neural network. This neural network adaptively measures the similarity between the last generated element of the output sequence s i 1 and the j-th element of the input sequence h j , allowing the model to learn context-dependent attention patterns.
Both architectures use these phases to reduce the dimensionality of the learning problem while preserving the ability to capture complex relationships. The key difference lies in how they implement these phases; Transformers learn both the embedding and smoothing operations, while K-GAMs use a fixed, universal embedding (Köppen function) followed by learned smoothing.
The main advantage of using smoothing techniques is that they are parallelizable. This is the key draw of Transformers. Current language models such as BERT, GPT, and T5 rely on the Transformer approach. Further, Transformers have also been applied to computer vision and other domains. Their ability to capture long-range dependencies and their scalability have made them powerful tools for a wide range of applications. See Ref. [60] for further details.

5. Application

5.1. Simulated Data

We also apply the K-GAM architecture to a simulated dataset to evaluate its performance on data with known structures and relationships. The dataset contains 100 observations generated from the following function:
y = μ ( x ) + ϵ , ϵ N ( 0 , 1 ) μ ( x ) = 10 sin ( π x 1 x 2 ) + 20 ( x 3 0.5 ) 2 + 10 x 4 + 5 x 5 .
The goal is to predict the function y ( x ) based on the input x. The dataset is often used as a benchmark dataset for regression algorithms due to its diverse mix of relationships (linear, quadratic, nonlinear, Gaussian random noise) between the input features and the target function. The plot of μ (no noise) vs. y (noise) is shown in Figure 2.
We use the Köppen function to transform the five-dimensional input into a set of 11 features ( 2 d + 1 ). We then learn the outer function g using a ReLU network. To thoroughly investigate the model’s capabilities, we implement two distinct approaches to learning the outer function. The first approach uses different g functions for each feature, following the original KST formulation. This allows each function to specialize in capturing specific patterns, but it might be more difficult to train and has more parameters. The second approach uses a single g function for all features, as proposed by [61], providing a more unified and parameter-efficient representation.
Figure 3 illustrates these architectural choices in detail, showing how the information flows through each version of the model. For the first model with multiple g i functions, the dimensions of each g i are as follows: W i 0 R 16 × 1 and for j = 1 , , 18 , W i j R 16 × 16 . Exemplary outer functions are shown in Figure 4.
The next architecture, which uses only one function, g, for all features, maintains a similar structure to the multiple-g function approach. The only difference is in the dimensionality of the inner layers—we increase the width from 16 to 200. This increased capacity allows the single function to learn more complex patterns and compensate for the constraint of using just one function instead of multiple specialized ones. The behavior of this unified outer function is shown in Figure 5, where we can observe how it adapts to handle multiple feature transformations simultaneously.

5.2. Iris Data

We apply the KST architecture to an iris dataset. The iris dataset is a classic dataset in machine learning and statistics. It contains 150 observations of iris flowers. Each observation contains four features, namely, sepal length, sepal width, petal length, and petal width. The goal is to predict the species of the iris flower based on these features. The dataset contains three classes of iris flowers, namely, setosa, versicolor, and virginica. The dataset is often used as a benchmark dataset for classification algorithms. The dataset has five variables, which include four characteristics of the iris flower and the species of the flower. Figure 6 shows the scatter plots of the iris dataset.
We calculate the mean μ SL of the sepal length column and use a binary variable y i = Sepal Length i > μ SL as the output. We use the other three flower characteristics, x 1 , x 2 , x 3 , as input variables. We use a classical GAM model to fit the data and compare it to KST-GAN. The classical generalized additive model is given by the following:
logit ( P ( y = 1 x ) ) = β 0 + f 1 ( x 2 ) + f 2 ( x 3 ) + f 3 ( x 4 ) ,
where f 1 , f 2 , f 3 are smooth functions of the input features. The KST-GAN model is given by the following:
logit ( P ( y = 1 ψ ) ) = β 0 + f 1 ( ψ 1 ) + + f 7 ( ψ 7 ) ,
where ψ q is the Köppen transformation function (in Appendix A) of the input features
ϕ q ( x 1 , x 2 , x 3 ) = i = 1 3 α i ψ ( x i + a q ) , q = 1 , , 7 .
We use m = 2 d + 1 = 7 and k = 6 to transform the input features into a set of seven features. We then learn the outer function g using a classical GAN approach. We use the mgcv package in R to fit the GAM model. This package uses a penalized likelihood approach to fit the model [62].
Table 1 compares the performances of the GAM and KST-GAM models. We also include a classical logistic regression model for comparison. The KST-GAM model has a higher AIC and BIC compared to the GAM model. The KST-GAM model has a comparable RMSE.
Table 2 and Table 3 show the confusion matrices for the GAM and KST GAM models, respectively. The KST GAM model has a lower accuracy compared to the GAM model.
Figure 7 shows the original features plotted against those fitted by the GAM function for both original inputs and the KST-transformed inputs.
Overall, the iris example demonstrates the inability of the GAM model to capture the complex relationships between the transformed features and the target variable.
The comparison between traditional GAM and K-GAM approaches in Figure 7 shows some notable differences. The GAM plots show slightly smoother relationships between individual features and the output. For instance, petal width and petal length show particularly strong linear relationships with the target variable. In comparison, the K-GAM architecture seems to capture different aspects of the feature relationships. Notably, while one of the features retains a largely linear relationship, the other two features encode more complicated relationships. This difference demonstrates the effect of the Köppen function’s mapping.
Next, the simulated data study explores how K-GAM handles data with known structures, as well as the differences between using multiple outer functions and a single shared outer function within the K-GAM framework. The comparison between multiple g i functions (Figure 4) and a single g function (Figure 5) demonstrates the architectural flexibility of the approach. The single g function variant shows a highly variable pattern across its input range, showing that the increased dimension compensates for the reduced flexibility of having a single function by developing more complex internal representations. The ability to choose between multiple specialized functions or a single shared function provides a useful degree of freedom in the model’s design.

5.3. Image Processing

A number of authors have proposed the use of Kolmogorov networks for image processing tasks (for example, see Refs. [1,2]). More recently, functional KANs were proposed in [12]. Figure 8 shows the error outputs for the MNIST dataset. Equivalent rates for the MNIST dataset can be found in [63].
Functional KAN methods have been shown to be effective for image processing tasks and have been shown to be promising in other tasks. Ref. [8] provided theoretical results for Kolmogorov networks applied to image segmentation tasks. Another alternative is deep partial least squares (DPLS) networks—a type of Kolmogorov network that uses partial least squares regression to learn the network weights [64].

6. Discussion

At its core, our work reinforces Kolmogorov’s profound insight, that is, there are no true multivariate problems, only superpositions of univariate affine ones. This principle guides our approach to efficiently decomposing complex multivariate functions into simpler univariate components through the Köppen transformation. The results demonstrate that while K-GAM can effectively model both real (iris) and simulated data, the internal representations it learns may be more complex than traditional GAM approaches. The approach appears particularly effective at capturing nonlinear patterns, although the interpretability of individual feature effects becomes more challenging due to the Köppen function transformation. Importantly, the K-GAM approach requires significantly fewer parameters compared to standard GAMs, as it leverages a shared embedding space through the Köppen function and can capture nonlinear relationships without requiring explicit interaction terms. However, most of the research since 2010 has been about increasing the depth of the networks and not about the width of the networks. The K-GAM approach is a step forward in the direction of increasing the width of the networks. The difference is that the Kolmogorov architecture is based on non-smooth functions, while the deep learning architecture is based on the composition of smooth functions. As shown by [65], it is impossible to obtain a universal approximation using a two-layer architecture with smooth functions.
Our findings suggest several promising directions for future research across both theoretical and practical domains. One priority involves the enhancement of the scalability of KST-based approaches, along with the characterization of the function classes for which K-GAM performs optimally over existing alternatives. Given K-GAM’s parameter efficiency, exploring Bayesian learning methods could provide principled approaches to uncertainty quantification and model regularization. An important optimization for K-GAM would be a specialized optimization algorithm capable of handling the discontinuities inherent in the Köppen function and the development of an efficient training algorithm specifically designed for high-dimensional problems.
These findings suggest that K-GAM networks represent a promising direction for efficient function approximation, particularly in scenarios where traditional deep learning approaches may be computationally intractable or parameter-inefficient. The combination of theoretical guarantees from Kolmogorov’s superposition theorem with modern machine learning techniques opens new avenues for both theoretical research and practical applications.

Author Contributions

Both authors contributed equally to this paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Function for Calculating the Köppen Transformation

def koppenFeatures(x,k):
qa = a*np.arange(m)
thf = koppenPhiv(np.add.outer(qa,x),k,gamma)
p = np.arange(1,n)
lmb = np.power(gamma,-np.multiply.outer(p,beta)).sum(axis=1)
lmb = np.insert(lmb,0,1)
thf = (thf@lmb).transpose()
return thf
def betaf(r):
  """
  Return the beta scalar for index r, as defined in Koppen paper
  Args:
    r (int): Index
  """
  print(f’n inside beta funciton is n’)
  return (pow(n,r)-1)/(n-1)
def lmb(beta,p):
  return pow(gamma,(1-p)*beta).sum()
def lambda_p(p):
  if p==1:
    return 1
  else:
    eps = 1
    tol = 10e-8
    total = 0
    r=1
    while eps>tol:
      eps = pow(gamma,(1-p)*beta[r])
      r+=1
      total += eps
  return total
def extractDigits(d,k,gamma):
  arr_d = []
  d=d*pow(gamma,k)
  d = round(d)
  for i in range(k):
    arr_d.insert(0,d % gamma)
    d = d//gamma
  return arr_d
def koppenPhiDigits(d,gamma):
  k = len(d)
  if k==1:
    return d[0]/gamma
  elif d[-1] < (gamma-1):
    return koppenPhiDigits(d[:-1],gamma) + d[-1]*pow(gamma,-beta[k])
.  else:
    d_prev = d[:-1]
    d_next = copy.deepcopy(d)
    while len(d_next)>0 and d_next[-1]==gamma-1:
      d_next = d_next[:-1]
    if len(d_next)==0:
      d_next = [gamma]
    else:
      d_next[-1] +=1
    return 0.5*(koppenPhiDigits(d_prev,gamma) + koppenPhiDigits
(d_next,gamma) + (gamma-2)*pow(gamma,-beta[k]))
    # return 0.5*(koppenPhiDigits(d[:-1],gamma) + koppenPhiDigits
(d_next,gamma) + (gamma-2)*pow(gamma,-beta[k]))
def koppenPhi(dk,k, gamma):
  assert dk<2
  assert dk>=0, f’Got negative input dk:.2e’
  if dk>1:
    return koppenPhi(dk-1,k,gamma)+1
  if k==1:
    return dk
  else:
    d = extractDigits(dk,k,gamma)
  return koppenPhiDigits(d,gamma)
koppenPhiv = np.vectorize(koppenPhi)

References

  1. Bryant, D.W. Analysis of Kolmogorov’s Superpostion Theorem and Its Implementation in Applications with Low and High Dimensional Data. Ph.D. Thesis, University of Central Florida, Orlando, FL, USA, 2008. [Google Scholar]
  2. Leni, P.E.; Fougerolle, Y.D.; Truchetet, F. Kolmogorov Superposition Theorem and Its Application to Multivariate Function Decompositions and Image Representation. In Proceedings of the 2008 IEEE International Conference on Signal Image Technology and Internet Based Systems, Bali, Indonesia, 30 November–3 December 2008; pp. 344–351. [Google Scholar]
  3. Kolmogorov, A.N. On the Representation of Continuous Functions of Several Variables as Superpositions of Functions of Smaller Number of Variables. Soviet. Math. Dokl. 1956, 108, 179–182. [Google Scholar]
  4. Hecht-Nielsen, R. Kolmogorov’s Mapping Neural Network Existence Theorem. In Proceedings of the International Conference on Neural Networks, San Diego, CA, USA, 21–24 June 1987; IEEE Press: New York, NY, USA, 1987; Volume 3, pp. 11–14. [Google Scholar]
  5. Liu, Z.; Wang, Y.; Vaidya, S.; Ruehle, F.; Halverson, J.; Soljačić, M.; Hou, T.Y.; Tegmark, M. KAN: Kolmogorov-Arnold Networks. arXiv 2024, arXiv:2404.19756. [Google Scholar] [CrossRef]
  6. Polar, A.; Poluektov, M. A Deep Machine Learning Algorithm for Construction of the Kolmogorov-Arnold Representation. arXiv 2020, arXiv:2001.04652. [Google Scholar] [CrossRef]
  7. Montanelli, H.; Yang, H.; Du, Q. Deep ReLU Networks Overcome the Curse of Dimensionality for Bandlimited Functions. arXiv 2020, arXiv:1903.00735. [Google Scholar] [CrossRef]
  8. Igelnik, B.; Parikh, N. Kolmogorov’s Spline Network. IEEE Trans. Neural Netw. 2003, 14, 725–733. [Google Scholar] [CrossRef] [PubMed]
  9. Poluektov, M.; Polar, A. Construction of the Kolmogorov-Arnold Representation Using the Newton-Kaczmarz Method. arXiv 2024, arXiv:2305.08194. [Google Scholar] [CrossRef]
  10. Bozorgasl, Z.; Chen, H. Wav-KAN: Wavelet Kolmogorov-Arnold Networks. arXiv 2024, arXiv:2405.12832. [Google Scholar] [CrossRef]
  11. Reinhardt, E.A.F.; Dinesh, P.R.; Gleyzer, S. SineKAN: Kolmogorov-Arnold Networks Using Sinusoidal Activation Functions. arXiv 2024, arXiv:2407.04149. [Google Scholar] [CrossRef]
  12. Ta, H.T.; Thai, D.Q.; Rahman, A.B.S.; Sidorov, G.; Gelbukh, A. FC-KAN: Function Combinations in Kolmogorov-Arnold Networks. arXiv 2024, arXiv:2409.01763. [Google Scholar] [CrossRef]
  13. Liu, Z.; Ma, P.; Wang, Y.; Matusik, W.; Tegmark, M. KAN 2.0: Kolmogorov-Arnold Networks Meet Science. arXiv 2024, arXiv:2408.10205. [Google Scholar] [CrossRef]
  14. Lai, M.J.; Shen, Z. The Optimal Rate for Linear KB-splines and LKB-splines Approximation of High Dimensional Continuous Functions and its Application. arXiv 2024, arXiv:2401.03956. [Google Scholar]
  15. Polson, N.G.; Sokolov, V. Deep Learning: A Bayesian Perspective. Bayesian Anal. 2017, 12, 1275–1304. [Google Scholar] [CrossRef]
  16. Wang, Y.; Polson, N.; Sokolov, V.O. Data Augmentation for Bayesian Deep Learning. Bayesian Anal. 2023, 18, 1041–1069. [Google Scholar] [CrossRef]
  17. Polson, N.G.; Ročková, V. Posterior Concentration for Sparse Deep Learning. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
  18. Hecht-Nielsen, R. Counterpropagation Networks. Appl. Opt. 1987, 26, 4979–4984. [Google Scholar] [CrossRef]
  19. Ismayilova, A.; Ismailov, V. On the Kolmogorov Neural Networks. arXiv 2023, arXiv:2311.00049. [Google Scholar] [CrossRef]
  20. Ismailov, V. A Three Layer Neural Network Can Represent Any Multivariate Function. arXiv 2022, arXiv:2012.03016. [Google Scholar] [CrossRef]
  21. Sprecher, D.A. On the Structure of Continuous Functions of Several Variables. Trans. Am. Math. Soc. 1965, 115, 340–355. [Google Scholar] [CrossRef]
  22. Fridman, B.L. An Improvement in the Smoothness of the Functions in AN Kolmogorov’s Theorem on Superpositions. Dokl. Akad. Nauk SSSR 1967, 177, 1019–1022. [Google Scholar]
  23. Köppen, M. On the Training of a Kolmogorov Network. In Proceedings of the Artificial Neural Networks—ICANN 2002, Madrid, Spain, 28–30 August 2002; Dorronsoro, J.R., Ed.; Lecture Notes in Computer Science. Springer: Berlin/Heidelberg, Germany, 2002; pp. 474–479. [Google Scholar]
  24. Köppen, M. The Curse of Dimensionality. In Proceedings of the 5th Online World Conference on Soft Computing in Industrial Applications (WSC5), Online, 4–18 September 2000; Volume 1, pp. 4–8. [Google Scholar]
  25. Braun, J. An Application of Kolmogorov’s Superposition Theorem to Function Reconstruction in Higher Dimensions. Ph.D. Thesis, Universitäts-und Landesbibliothek Bonn, Bonn, Germany, 2009. [Google Scholar]
  26. Montanelli, H.; Yang, H. Error Bounds for Deep ReLU Networks Using the Kolmogorov–Arnold Superposition Theorem. arXiv 2020, arXiv:1906.11945. [Google Scholar] [CrossRef]
  27. Schmidt-Hieber, J. Nonparametric Regression Using Deep Neural Networks with ReLU Activation Function. Ann. Stat. 2020, 48, 1875–1897. [Google Scholar]
  28. Actor, J. Computation for the Kolmogorov Superposition Theorem. Master’s Thesis, Rice, Houston, TX, USA, 2018. [Google Scholar]
  29. Demb, R.; Sprecher, D. A Note on Computing with Kolmogorov Superpositions without Iterations. Neural Netw. 2021, 144, 438–442. [Google Scholar] [CrossRef] [PubMed]
  30. Klusowski, J.M.; Barron, A.R. Risk bounds for high-dimensional ridge function combinations including neural networks. arXiv 2016, arXiv:1607.01434. [Google Scholar]
  31. Friedman, J.H.; Stuetzle, W. Projection pursuit regression. J. Am. Stat. Assoc. 1981, 76, 817–823. [Google Scholar] [CrossRef]
  32. Huber, P.J. Projection pursuit. Ann. Stat. 1985, 13, 435–475. [Google Scholar] [CrossRef]
  33. Jones, L.K. A Simple Lemma on Greedy Approximation in Hilbert Space and Convergence Rates for Projection Pursuit Regression and Neural Network Training. Ann. Statist. 1992, 20, 608–613. [Google Scholar] [CrossRef]
  34. Diaconis, P.; Shahshahani, M. On Nonlinear Functions of Linear Combinations. SIAM J. Sci. Stat. Comput. 1984, 5, 175–191. [Google Scholar] [CrossRef]
  35. Lai, M.J.; Shen, Z. The Kolmogorov Superposition Theorem Can Break the Curse of Dimensionality When Approximating High Dimensional Functions. arXiv 2023, arXiv:2112.09963. [Google Scholar] [CrossRef]
  36. Sprecher, D.A. An Improvement in the Superposition Theorem of Kolmogorov. J. Math. Anal. Appl. 1972, 38, 208–213. [Google Scholar] [CrossRef]
  37. Breiman, L. Statistical Modeling: The Two Cultures (with Comments and a Rejoinder by the Author). Stat. Sci. 2001, 16, 199–231. [Google Scholar] [CrossRef]
  38. Bhadra, A.; Datta, J.; Polson, N.G.; Sokolov, V.; Xu, J. Merging Two Cultures: Deep and Statistical Learning. WIREs Comput. Stat. 2024, 16, e1647. [Google Scholar] [CrossRef]
  39. Albeverio, S.; Khrennikov, A.; Kloeden, P.E. Memory Retrieval as a P-Adic Dynamical System. Biosystems 1999, 49, 105–115. [Google Scholar] [CrossRef] [PubMed]
  40. Zú niga-Galindo, W.A.; Zambrano-Luna, B.A.; Dibba, B. Hierarchical Neural Networks, p-Adic PDEs, and Applications to Image Processing. J. Nonlinear Math. Phys. 2024, 31, 63. [Google Scholar] [CrossRef]
  41. Chua, L.; Yang, L. Cellular Neural Networks: Theory. IEEE Trans. Circuits Syst. 1988, 35, 1257–1272. [Google Scholar] [CrossRef]
  42. Bellman, R. Adaptive Control Processes: A Guided Tour; Princeton University Press: Princeton, NJ, USA, 1961. [Google Scholar]
  43. Barron, A.R. Universal Approximation Bounds for Superpositions of a Sigmoidal Function. IEEE Trans. Inf. Theory 1993, 39, 930–945. [Google Scholar] [CrossRef]
  44. Niyogi, P.; Girosi, F.; Poggio, T. Incorporating Prior Information in Machine Learning by Creating Virtual Examples. Proc. IEEE 1998, 86, 2196–2209. [Google Scholar] [CrossRef]
  45. Hegland, M.; Pestov, V. Additive Models in High Dimensions. arXiv 2002, arXiv:9912020. [Google Scholar] [CrossRef]
  46. Bungartz, H.J.; Griebel, M. Sparse Grids. Acta Numer. 2004, 13, 147–269. [Google Scholar] [CrossRef]
  47. Khavinson, S. Best Approximation by Linear Superpositions (Approximate Nomography); American Mathematical Soc.: Providence, RI, USA, 1997; Volume 159. [Google Scholar]
  48. Fredholm, I. Sur Une Classe d’équations Fonctionnelles. Acta Math. 1903, 27, 365–390. [Google Scholar] [CrossRef]
  49. Abel, N.H. Oplösning af et Par Opgaver ved Hjælp af bestemte Integraler. Mag. Naturvidenskaberne 1823, 1, 11–27. [Google Scholar]
  50. Nadaraya, E.A. On Estimating Regression. Theory Probab. Its Appl. 1964, 9, 141–142. [Google Scholar] [CrossRef]
  51. Watson, G.S. Smooth Regression Analysis. Sankhyā Indian J. Stat. Ser. A (1961–2002) 1964, 26, 359–372. [Google Scholar]
  52. Parzen, E. On Estimation of a Probability Density Function and Mode. Ann. Math. Stat. 1962, 33, 1065–1076. [Google Scholar] [CrossRef]
  53. Gramacy, R.B.; Lee, H.K.H. Bayesian Treed Gaussian Process Models with an Application to Computer Modeling. J. Am. Stat. Assoc. 2008, 103, 1119–1130. [Google Scholar] [CrossRef]
  54. Higdon, D.; Gattiker, J.; Williams, B.; Rightley, M. Computer Model Calibration Using High-Dimensional Output. J. Am. Stat. Assoc. 2008, 103, 570–583. [Google Scholar] [CrossRef]
  55. Vapnik, V. The Nature of Statistical Learning Theory; Springer Science & Business Media: Berlin/Heidelberg, Germany, 1999. [Google Scholar]
  56. Mockus, J. Bayesian Approach to Global Optimization: Theory and Applications; Mathematics and Its Applications; Springer: Dordrecht, The Netherlands, 1989; Volume 37. [Google Scholar]
  57. Schmidt-Hieber, J. The Kolmogorov–Arnold Representation Theorem Revisited. Neural Netw. 2021, 137, 119–126. [Google Scholar] [CrossRef]
  58. Coppejans, M. On Kolmogorov’s Representation of Functions of Several Variables by Functions of One Variable. J. Econom. 2004, 123, 1–31. [Google Scholar] [CrossRef]
  59. Bahdanau, D.; Cho, K.; Bengio, Y. Neural Machine Translation by Jointly Learning to Align and Translate. arXiv 2014, arXiv:1409.0473. [Google Scholar] [CrossRef]
  60. Tsai, Y.H.H.; Bai, S.; Yamada, M.; Morency, L.P.; Salakhutdinov, R. Transformer Dissection: A Unified Understanding of Transformer’s Attention via the Lens of Kernel. arXiv 2019, arXiv:1908.11775. [Google Scholar] [CrossRef]
  61. Lorentz, G.G. The 13th Problem of Hilbert. In Proceedings of the Symposia in Pure Mathematics, Stanford, CA, USA, 2–21 August 1976; American Mathematical Society: Providence, RI, USA, 1976; Volume 28, pp. 419–430. [Google Scholar]
  62. Wood, S.N. Generalized Additive Models: An Introduction with R, 2nd ed.; Chapman and Hall/CRC: New York, NY, USA, 2017. [Google Scholar]
  63. Newton, M.A.; Polson, N.G.; Xu, J. Weighted Bayesian Bootstrap for Scalable Posterior Distributions. Can. J. Stat. 2021, 49, 421–437. [Google Scholar] [CrossRef]
  64. Polson, N.; Sokolov, V.; Xu, J. Deep Learning Partial Least Squares. arXiv 2021, arXiv:2106.14085. [Google Scholar] [CrossRef]
  65. Vitushkin, A.G. Some Properties of Linear Superpositions of Smooth Functions. Dokl. Akad. Nauk SSSR 1964, 156, 1003–1006. [Google Scholar]
Figure 1. Köppen function ψ k for k = 3 , 4 , 5 , γ = 10 .
Figure 1. Köppen function ψ k for k = 3 , 4 , 5 , γ = 10 .
Entropy 27 00593 g001
Figure 2. Scatter plot of the simulated dataset.
Figure 2. Scatter plot of the simulated dataset.
Entropy 27 00593 g002
Figure 3. KST architecture for the simulated dataset.
Figure 3. KST architecture for the simulated dataset.
Entropy 27 00593 g003
Figure 4. Examples of outer functions g 0 , g 2 , g 6 , g 8 for the simulated dataset.
Figure 4. Examples of outer functions g 0 , g 2 , g 6 , g 8 for the simulated dataset.
Entropy 27 00593 g004
Figure 5. Plot of the single outer function, g, for the simulated dataset.
Figure 5. Plot of the single outer function, g, for the simulated dataset.
Entropy 27 00593 g005
Figure 6. Scatter plots of the iris dataset.
Figure 6. Scatter plots of the iris dataset.
Entropy 27 00593 g006
Figure 7. Generalized additive model for the iris dataset versus KST.
Figure 7. Generalized additive model for the iris dataset versus KST.
Entropy 27 00593 g007
Figure 8. Functional KAN for the MNIST dataset.
Figure 8. Functional KAN for the MNIST dataset.
Entropy 27 00593 g008
Table 1. Summary of the classical GAM, KST GAM, and GLM models applied to the iris dataset.
Table 1. Summary of the classical GAM, KST GAM, and GLM models applied to the iris dataset.
Classical GAMKST GAMGLM
(Intercept)−1.75818.064−28.339
(1.435)(54.091)(8.371)
Sepal.Width 2.669
(1.499)
Petal.Length 6.377
(2.090)
Petal.Width −5.105
(2.292)
Num.Obs.105105105
R20.7900.591
AIC43.0207.350.6
BIC59.3258.161.2
Log.Lik. −21.281
RMSE0.220.290.26
Table 2. Confusion out-of-sample matrix for the GAM model applied to the iris dataset.
Table 2. Confusion out-of-sample matrix for the GAM model applied to the iris dataset.
Predicted 0Predicted 1
Actual 0211
Actual 1221
Table 3. Confusion out-of-sample matrix for the KST GAM model applied to the iris dataset.
Table 3. Confusion out-of-sample matrix for the KST GAM model applied to the iris dataset.
Predicted 0Predicted 1
Actual 0193
Actual 1815
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Polson, S.; Sokolov, V. Kolmogorov GAM Networks Are All You Need! Entropy 2025, 27, 593. https://doi.org/10.3390/e27060593

AMA Style

Polson S, Sokolov V. Kolmogorov GAM Networks Are All You Need! Entropy. 2025; 27(6):593. https://doi.org/10.3390/e27060593

Chicago/Turabian Style

Polson, Sarah, and Vadim Sokolov. 2025. "Kolmogorov GAM Networks Are All You Need!" Entropy 27, no. 6: 593. https://doi.org/10.3390/e27060593

APA Style

Polson, S., & Sokolov, V. (2025). Kolmogorov GAM Networks Are All You Need! Entropy, 27(6), 593. https://doi.org/10.3390/e27060593

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop