Next Article in Journal
Constructing Fuzzy Hypothesis Methods to Determine Critical-To-Quality Service Items
Next Article in Special Issue
Robust Stability of Complex-Valued Stochastic Neural Networks with Time-Varying Delays and Parameter Uncertainties
Previous Article in Journal
An Evolve-Then-Correct Reduced Order Model for Hidden Fluid Dynamics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Asymptotic Convergence of Soft-Constrained Neural Networks for Density Estimation

Dipartimento di Ingegneria dell’Informazione e Scienze Matematiche, Università di Siena, 53100 Siena, Italy
Mathematics 2020, 8(4), 572; https://doi.org/10.3390/math8040572
Submission received: 7 March 2020 / Revised: 7 April 2020 / Accepted: 8 April 2020 / Published: 12 April 2020
(This article belongs to the Special Issue Neural Networks and Learning Systems)

Abstract

:
A soft-constrained neural network for density estimation (SC-NN-4pdf) has recently been introduced to tackle the issues arising from the application of neural networks to density estimation problems (in particular, the satisfaction of the second Kolmogorov axiom). Although the SC-NN-4pdf has been shown to outperform parametric and non-parametric approaches (from both the machine learning and the statistics areas) over a variety of univariate and multivariate density estimation tasks, no clear rationale behind its performance has been put forward so far. Neither has there been any analysis of the fundamental theoretical properties of the SC-NN-4pdf. This paper narrows the gaps, delivering a formal statement of the class of density functions that can be modeled to any degree of precision by SC-NN-4pdfs, as well as a proof of asymptotic convergence in probability of the SC-NN-4pdf training algorithm under mild conditions for a popular class of neural architectures. These properties of the SC-NN-4pdf lay the groundwork for understanding the strong estimation capabilities that SC-NN-4pdfs have only exhibited empirically so far.

1. Introduction

Density estimation has long been a fundamental open issue in statistics and pattern classification. Implicitly or explicitly, it is at the core of statistical pattern recognition and unsupervised learning [1]. Applications embrace data compression and model selection [2], coding [3], and bioinformatics [4]. Density estimation was applied to the modeling of sequences [5,6] and structured data [7,8], as well. Finally, the task of estimating conditional probability distributions is fundamental to the broad area of probabilistic graphical models [9,10]. Statistical parametric and non-parametric techniques are available to practitioners [1], but they suffer from several significant shortcomings [11]. In fact, parametric techniques require a strong assumption on the form of the probability density function (pdf) at hand, while non-parametric approaches are memory-based (i.e., prone to overfitting), overly complex in time and space, and unreliable over small data samples. Consequently, density estimation via artificial neural networks (ANN) has been receiving increasing attention from researchers. Despite the ease of training ANNs for Bayes posterior probability estimation aimed at pattern classification [12,13], learning density functions raises problems entailed by the intrinsically unsupervised task and, above all, to the requirement of satisfying (at least numerically) the axioms of probability [14]. In particular, the integral of the function realized by the ANN shall be equal to one.
Traditional ANN-based attempts to tackle the pdf estimation problem could not overcome some major shortcomings [11]. First of all, when applied to pdf estimation, the so-called “probabilistic neural network” [15] reduces to a neural version of the Parzen window estimator [1], with a hidden neuron per each Parzen kernel, such that it inherits the aforementioned drawbacks of statistical non-parametric techniques. The maximum-likelihood framework [16] does not offer explicit solutions to the numeric computation of the integral of the neural estimate, and its application has been limited to univariate distributions. The extension of that framework handed out by [17] does not offer solutions to such issues, focusing only on the rate of convergence of the algorithm. A parametric approach is presented in [18], in the form of a self-organizing mixture network. The latter realizes a density estimator based on a self-organizing map (SOM) [19] lattice of units associated with a neuron-specific component density (e.g., a Gaussian distribution), resulting in a mixture of such component pdfs. Although the SOM-like learning procedure is original and effective, the model (being parametric) suffers from the aforementioned limitations of statistical parametric techniques. The approaches based on indirect pdf estimation by differentiation of an estimated cumulative distribution function [20,21] end up violating the axioms of probability (they do not ensure that a proper pdf is obtained) and can hardly be applied to multivariate data, as well. Given these drawbacks of traditional approaches, researchers have recently proposed more robust ANN-based techniques for multivariate density estimation [22,23,24]. In particular, soft-constrained neural networks for pdf estimation (SC-NN-4pdf) [25] realize a non-parametric neural density estimator that explicitly satisfies the second Kolmogorov axiom. The SC-NN-4pdf was shown to outperform both statistical estimation techniques (parametric and non-parametric) and its neural competitors over difficult multivariate, multimodal pdf estimation tasks [25]. No clear rationale behind such an empirical evidence has been put forward so far, neither has it any analysis of the fundamental theoretical properties of the SC-NN-4pdf. To this end, hereafter we deliver a mathematical study of the major theoretical properties of the SC-NN-4pdf. After introducing the pdf estimation setup (Section 2), followed by a concise review of the fundamentals of the model and its training algorithm (Section 2.1), the formal analysis is devised in Section 3. We first offer a formal statement of the class of pdfs that can be modeled by SC-NN-4pdf to any degree of precision (Section 3.1). Then, in Section 3.2 we hand out a proof of asymptotic convergence in probability under mild conditions of the SC-NN-4pdf training algorithm for a popular class of architectures, namely multilayer perceptrons (MLP) with logistic sigmoid activation functions. Conclusions are drawn in Section 4.

2. Materials and Methods

Let us write T = { x 1 , , x n } to represent the sample of n multivariate random vectors, ( x k R d for k = 1 , , n ) independently drawn from the unknown pdf p ( x ) . A feed-forward neural network (FFNN) with d input units, one output unit, one or more hidden layers is used to learn p ( x ) from T in an unsupervised fashion. We write f i ( a i ) to represent the activation function associated with the generic neuron i in the FFNN. The quantity a i is computed as a i = j w i j f j ( a j ) where w i j is the weight of the connection between neurons j and i. We write w to represent the weight vector embracing all the adaptive parameters of the FFNN (including the biases of sigmoids). When the FFNN is fed with a generic input x = ( x 1 , , x d ) , the i-th input neuron realizes the identity function f i ( a i ) = f i ( x i ) = x i . On the contrary, any nonlinearity can be associated with the hidden neurons, but for the intents and purposes of this paper we will assume that logistic sigmoids are used. Finally, the output neuron shall have an activation function whose counterdomain matches the notion of pdf, namely the range [ 0 , + ) . There are different choices for f i ( a i ) that end up satisfying this requirement, e.g., f i ( a i ) = exp ( a i ) . In this paper, we resort to the ReLU activation function, that is f i ( a i ) = max ( a i , 0 ) , which allows for both the respect of the axioms of probability and the nice modeling properties offered by linear combinations of sigmoids.
In summary, the FFNN computes a function φ ( x , w ) of its input x . In [25], without loss of generality, it is assumed that the random samples of interest are limited to a compact S R d (therefore, henceforth S can be treated as the domain of φ ( x , w ) ). As explained in [25], any data normalization method may be applied to the aim of satisfying this requirement. The training algorithm presented in [25] revolves around a learning rule capable of adapting w given T such that a proper estimate of p ( x ) is achieved via φ ( x , w ) . To this end, two purposes are pursued: (1) capitalizing on the information brought by T so as to approximate the unknown pdf; (2) holding the FFNN back from reaching degenerate solutions, such that the constraint S φ ( x , w ) d x = 1 holds true. Section 2.1 reviews the fundamentals of the training algorithm presented in [25], which actually reaches both of these purposes.

2.1. Review of the Fundamentals of the Training Algorithm

The present Section summarizes the main points of the SC-NN-4pdf training algorithm, as handed out in [25]. In so doing, the notation used in the remainder of the paper is introduced, and a common ground is fixed that allows the reader to make sense of the theoretical analysis presented in Section 3. The estimate p ˜ ( · ) of the unknown pdf p ( · ) is defined as
p ˜ ( x , w ) = φ ( x , w ) S φ ( x , w ) d x
for all x in S. The FFNN training algorithm is built on non-parametric statistics (see [1]). Given a generic pattern x ^ T , let us consider a d-ball B ( x ^ , T ) centered on x ^ and with the minimum radius r ( x ^ , T ) s.t. B ( x ^ , T ) T = k n + 1 where k n = k n and k N is a hyperparameter. Please note that x ^ is in B ( x ^ , T ) T by construction, therefore a proper estimate of p ( · ) over x ^ shall not involve the latter in the computation: this is the rationale behind considering k n + 1 patterns within the ball instead of k n .
We write P to represent the probability that a generic pattern drawn from p ( x ) is in B ( x ^ , T ) . By definition of pdf, P = B ( x ^ , T ) p ( x ) d x . It is seen that P k n / n and that B ( x ^ , T ) p ( x ) d x p ( x ^ ) V ( B ( x ^ , T ) ) where V ( . ) denotes the volume of its argument. As a consequence, p ( x ^ ) k n / n V ( B ( x ^ , T ) ) . Therefore, the SC-NN-4pdf training algorithm consists of a gradient-descent minimization of the unsupervised criterion function
L ( T , w ) = 1 2 x i T k n / n V ( B ( x i , T ) ) p ˜ ( x i , w ) 2 + ρ 2 1 S φ ( x , w ) d x 2
Regarding w (Equation (2) represents a batch criterion, although a stochastic approach can be taken as well by minimizing on-line a similar pattern-wise criterion upon presentation of individual training patterns one at a time [25]). The first term in the definition of L ( T , w ) aims at normalized FFNN outputs that approximate the aforementioned non-parametric estimate of p ( · ) . The second term results in a “soft” constraint enforcing a unit integral of p ˜ ( x , w ) on S. The quantity ρ R + weights the relative contribution from the constraint. In practice, ρ is realized as a vanishing quantity, such that ρ 0 as long as training proceeds, i.e., as long as the function realized by the FFNN approaches the pdf sought (a similar asymptotic convergence of ρ to zero is exploited shortly in the following theoretical analysis). The gradient-descent learning rule for the generic parameter w is devised in [25]. The learning rule prescribes a weight modification Δ w in the form Δ w = η L ( . ) w , where η R + is the learning rate. To calculate the partial derivative of L ( T , w ) regarding w we need to take the derivatives of both terms on the right of Equation (2). As shown in [25], the derivative of the first term can be written as
w 1 2 k n / n V ( B ( x ^ , T ) ) φ ( x ^ , w ) S φ ( x , w ) d x 2 = = 1 S φ ( x , w ) d x k n / n V ( B ( x ^ , T ) ) φ ( x ^ , w ) S φ ( x , w ) d x · φ ( x ^ , w ) w φ ( x ^ , w ) S φ ( x , w ) d x S φ ( x , w ) w d x
where Leibniz rule was applied. Similarly, the derivative of the second term can be written as [25]
w ρ 2 1 S φ ( x , w ) d x 2 = = ρ 1 S φ ( x , w ) d x S φ ( x , w ) w d x .
In the equations, the derivative φ ( x , w ) w of the FFNN output regarding one of its parameters is obtained via backpropagation (BP), as usual. Let w = w i j represent the weight of the connection holding between the j-th neuron in layer 1 and the i-th neuron in layer . We can write φ ( x , w ) w = δ i f j ( a j ) where δ i = f i ( a i ) if represents the output layer, or δ i = ( u + 1 w u i δ u ) f i ( a i ) otherwise, provided that the deltas for the neurons in layer + 1 had already been determined as prescribed by plain BP. Extension of the algorithm to the other adaptive parameters of the FFNN (e.g., the bias values) is straightforward.
To compute the right-hand side of Equations (3) and (4), efficient algorithms for the computation of S φ ( x , w ) d x and S w φ ( x , w ) d x are presented in [25] that scale properly with the dimensionality of the data and the size of the network. Eventually, the resulting overall algorithm takes the form of an iterative application of BP-based gradient-descent steps where the deltas at the output layer are properly defined, while the backpropagation of the deltas downward through the network takes place in a plain BP fashion. An illustrative example of the behavior of the SC-NN-4pdf is given in Figure 1. The figure represents the estimate obtained from a sample of n = 1000 random observations drawn from a mixture of three generalized extreme value (GEV) pdfs with different priors and parameters. The SC-NN-4pdf architecture and the hyperparameters are the same we used in [25] for the experiments with univariate data.

3. Results

We hereafter present the results of the theoretical analysis of SC-NN-4pdfs. Section 3.1 investigates the modeling capabilities of the machine, that is the class of pdfs that can actually be modeled by the machine to any degree of precision (an investigation that turns out to be straightforward in the light of [26,27]). Section 3.2 proves a Theorem of asymptotic convergence in probability of the SC-NN-4pdf training algorithm to the pdf sought under mild conditions. Both Sections rely on established results on the approximation capabilities of ANNs [26,27,28], as well as on non-parametric statistics [1,29,30].

3.1. Modeling Capabilities

This Section shows that for any nonpaltry pdf p ( · ) over X [11] and any ϵ R + a SC-NN-4pdf exists which computes φ ( ϵ ) ( · ) such that the corresponding density estimate p ˜ ( ϵ ) ( · ) = φ ( ϵ ) ( · ) X φ ( ϵ ) ( · ) d x approximates p ( · ) over X with precision ϵ . In qualitative terms, nonpaltry pdfs form the class of “interesting” pdfs p : R d R for which there is a compact subset X of I d (where I d = [ 0 , 1 ] d ) covering those finite regions of R d s.t. the integral of any such p ( · ) over R d \ X is < ϵ p for an arbitrarily small ϵ p R + . In practice, whenever the support of a pdf p ( · ) of interest is not a subset of I d , we expect that any suitable data normalization method can be applied so as to restrict the support of p ( · ) such that R d \ X p ( x ) d x < ϵ p . On the other hand, if the support of p ( · ) is a subset of I d then it can be extended to the whole I d by setting p ( x ) = 0 for all x I d \ X (note that this does not affect satisfaction of the axioms of probability). These notions are readily formalized in terms of the following, mildly stricter definition [11]:
Definition 1.
Let X I d be compact. A continuous pdf p : R d R such that X p ( x ) d x = 1 is said to be nonpaltry over X.
We write P ( X ) to represent the set of nonpaltry pdfs over X. As observed in [11], P ( X ) is not a linear space but a linear closure P ^ ( X ) of P ( X ) can be defined as P ^ ( X ) = { a p ( · ) + b q ( · ) a , b R , p ( · ) , q ( · ) P ( X ) } . The universe P ^ ( X ) is actually a linear subspace of C ( X ) , that is the set of all continuous, real-valued functions on X. Furthermore, if f ( · ) C ( X ) , α = X f ( x ) d x and α 0 , then p f ( · ) = f ( · ) / α is in P ^ ( X ) , i.e., any function in C ( X ) can be written as a linear combination of functions in P ^ ( X ) . Therefore, P ^ ( X ) turns out to be a Banach space with Chebyshev norm, as well. Then, since the functions in P ( X ) are bounded (in fact, they are continuous over a closed domain), the formal analysis presented in [26] applies, proving that for any nonpaltry pdf p ( · ) over X at least one SC-NN-4pdf exists that computes φ ( ϵ ) ( · ) s.t. p ˜ ( ϵ ) ( · ) p ( · ) , X < ϵ for any ϵ R + (from now on we write . , X to represent the Chebyshev norm over X). In point of fact, a SC-NN-4pdf computing such φ ( ϵ ) ( · ) is found in the class of multilayer perceptrons (MLP) with only one hidden layer with logistic sigmoid activation functions.
While Cybenko’s analysis [26] revolves around the uniform norm, the arguments put forward in [27] entail likewise conclusions as pertains the L l norm on X for l = 1 , 2 , . In the following, we make the dependence of this norm on X explicit by writing L l ( X ) . Since the functions in P ( X ) are bounded, non-negative, and l-integrable on X, Corollary 2.2 in [27] proves the following theorem.
Theorem 1.
Given any nonpaltry pdf p ( · ) over X, a SC-NN-4pdf with one hidden layer of sigmoid activation functions exists that computes φ ( ϵ ) ( · ) s.t. p ˜ ( ϵ ) ( · ) p ( · ) L l ( X ) < ϵ for any ϵ R + and for any l N .
Qualitatively speaking, these modeling capabilities of SC-NN-4pdfs (as well as the convergence Theorem presented in the next Section) can be seen as consequences of the fact that since φ ( ϵ ) ( · ) p ( · ) L 1 ( X ) is arbitrarily small [27] then X φ ( ϵ ) ( · ) d x is arbitrarily close to 1 (see Equation (1)).

3.2. Asymptotic Convergence in Probability

Aside from the theoretical modeling capabilities presented in Section 3.1, the SC-NN-4pdf training process can be shown to actually converge in probability to the true, nonpaltry, and unknown pdf p ^ ( · ) P ( X ) underlying the distribution of the data at hand, under mild condition, for the popular class of MLPs with a single hidden layer with logistic sigmoid activation functions. The formal proof is given hereafter. It is seen that convergence cannot be guaranteed in general, nonetheless convergence to a pdf that is arbitrarily close to p ^ ( · ) is proven under the mild conditions stated in [31,32].
Let φ ( · , A , w ) be the function on X computed by a SC-NN-4pdf with support X (i.e., S = X ), architecture A, and parameters w to be learned from the data. Accordingly, ϕ ( · , A , w n ) represents the estimate of p ^ ( · ) realize via the SC-NN-4pdf trained over a given training sample T n = { x 1 , , x n } of n independent random vectors drawn from p ^ ( · ) , that is ϕ ( · , A , w n ) = φ ( . , A , w n ) X φ ( . , A , w n ) d x . In fact, ϕ ( · , A , w n ) is a random quantity that depends on T n , i.e., on the value w n of the SC-NN-4pdf parameters upon completion of the training stage on T n . Hereafter, we analyze the behavior of SC-NN-4pdfs (that is, of the estimators ϕ ( · , A , w n ) they learn to compute) as n . First, note that k n / n V ( B ( x , T n ) ) C ( X ) by construction. Moreover, since k n = k n with k N , it is seen that V ( B ( x , T n ) ) converges asymptotically to zero, that is lim n V ( B ( x , T n ) ) = 0 , yet slower than 1 / n , that is lim n n V ( B ( x , T n ) ) = .
Let ϵ R + be any (small) positive real value, which is to say the “degree of precision” of the SC-NN-4pdf solution sought. Since T n is a random quantity, we study the convergence of the SC-NN-4pdf estimate ϕ ( · , A , w n ) to p ^ ( · ) in mean square (due to Markov’s inequality, this entails convergence in probability). Convergence in mean square prescribes [33] that an integer n 0 exists such that for all n > n 0 , the following inequalities hold:
E ϕ ( · , A , w n ) p ^ ( · ) L 2 ( X ) < ϵ
and
σ 2 ϕ ( · , A , w n ) p ^ ( · ) L 2 ( X ) < ϵ
where E [ · ] denotes the expectation of a random variable (such an expected value is ideally evaluated over all possible samples { x 1 , , x n } drawn from p ^ ( · ) ) and σ 2 [ · ] is the corresponding variance. Let ϵ 1 R + be s.t. ϵ 1 < ϵ . Theorem 1 guarantees that at least one SC-NN-4pdf with a single hidden layer of logistic sigmoid activation functions exists that computes φ ( ϵ 1 ) ( · , A ( ϵ 1 ) , w ( ϵ 1 ) ) s.t. the estimate ϕ ( ϵ 1 ) ( · , A ( ϵ 1 ) , w ( ϵ 1 ) ) = φ ( ϵ 1 ) ( · , A ( ϵ 1 ) , w ( ϵ 1 ) ) / X φ ( x , A ( ϵ 1 ) , w ( ϵ 1 ) ) d x satisfies p ^ ( · ) ϕ ( ϵ 1 ) ( · , A ( ϵ 1 ) , w ( ϵ 1 ) ) L 2 ( X ) < ϵ 1 . The actual convergence of a given SC-NN-4pdf depends on its architecture. Therefore (and, since ϵ and ϵ 1 have been fixed), the present analysis focuses on the class of SC-NN-4pdfs with architecture A ( ϵ 1 ) and whose parameters w are to be learned from the training sample. Of course, these SC-NN-4pdfs compute w -specific functions φ ( · , A ( ϵ 1 ) , w ) . In the light of the theoretical results found in [32], let us assume that ker [ X n ] D 1 Y = { 0 } for all the values of n. Herein, the writing X n denotes the transpose of the n × ( d + 1 ) matrix yielded by concatenation of the n training vectors x 1 , , x n (each of them being d-dimensional), where the “ + 1 ” accounts for the presence of the bias terms in the sigmoid activation functions of the hidden layer. The quantity D 1 Y represents the set of all delta layer traces (refer to [31]) Y 1 = C n ( w . ) a j ( m ) for the hidden layer of A ( ϵ 1 ) , generated by varying all the parameters w over their domains, where: C n ( w ) = 1 2 x i T n k n / n V ( B ( x i , T n ) ) φ ( x i , A ( ϵ 1 ) , w ) 2 (compare the latter quantity with the first term in Equation (2)), j ranges over all the hidden neurons in A ( ϵ 1 ) , m ranges over the training patterns ( m = 1 , , n ), and a j ( m ) is the input fed to j-th hidden neuron when the SC-NN-4pdf is fed with m-th input pattern. The following convergence Theorem can now be proven to hold true:
Theorem 2.
Let ϵ R + , ϵ 1 R + , and ϵ 1 < ϵ . Let φ ( · , A ( ϵ 1 ) , w ) be the function computed by a SC-NN-4pdf with architecture A ( ϵ 1 ) and parameters w . Let us assume that for all the values of n, BP training of the SC-NN-4pdf is applied to the minimization of criterion L ( T n , w ) with no early stopping, and that ker [ X n ] D 1 Y = { 0 } holds true. Let λ max be the largest eigenvalue of the Hessian of φ ( · , A ( ϵ 1 ) , w ) with respect to w . If the learning rate η satisfies | 1 η λ max | < 1 at each step of BP, then an integer n 0 exists s.t., for all n > n 0 , training the SC-NN-4pdf over T n converges to parameters w n with the result that ϕ ( · , A ( ϵ 1 ) , w n ) satisfies Equations (5) and (6).
Proof. 
Under the hypotheses, we prove the existence of an integer n 0 s.t. both Equations (5) and (6) hold true for n > n 0 . First, let us work out the expectation, that is Equation (5). Let ϵ 2 R + s.t. ϵ = ϵ 1 + ϵ 2 . Independently from the specific architecture A of the SC-NN-4pdf, we have:
ϕ ( · , A , w n ) p ^ ( · ) L 2 ( X ) ϕ ( · , A , w n ) k n / n V ( B ( · , T n ) ) L 2 ( X ) + k n / n V ( B ( · , T n ) ) p ^ ( · ) L 2 ( X )
From Equation (7) we can write:
E ϕ ( · , A , w n ) p ^ ( · ) L 2 ( X ) E ϕ ( · , A , w n ) k n / n V ( B ( · , T n ) ) L 2 ( X ) + E k n / n V ( B ( · , T n ) ) p ^ ( · ) L 2 ( X )
Given Equation (8), the existence of an integer n 0 s.t. Equation (5) holds true for n > n 0 is guaranteed if the following conditions hold:
E ϕ ( · , A , w n ) k n / n V ( B ( · , T n ) ) L 2 ( X ) < ϵ 1
and
E k n / n V ( B ( · , T n ) ) p ^ ( · ) L 2 ( X ) < ϵ 2
for all n > n 0 . It is convenient to focus on Equation (10) first. Bearing in mind that r ( x , T n ) is the radius of B ( x , T n ) , let us define the function ψ ( . ) of u X , v X , and T n as follows:
ψ ( u , v , T n ) = 1 if v B ( u , T n ) 0 otherwise
s.t. X ψ ( u , v , T n ) d v = V ( B ( u , T n ) ) . Accordingly, we have
E k n / n V ( B ( x , T n ) ) = 1 n x j T n E ψ ( x , x j , T n ) V ( B ( x , T n ) ) = X ψ ( x , v , T n ) V ( B ( x , T n ) ) p ^ ( v ) d v
which is a convolution of the unknown density and of the normalized ψ ( . ) function. Since ψ ( · ) is well-behaved by construction and since V ( B ( x , T n ) ) 0 for n s.t. X ψ ( x , v , T n ) d v 0 too, then in the infinite sample case ψ ( x , · , T n ) V ( B ( x , T n ) ) converges to a Dirac’s delta centered at x . Therefore, since p ^ ( · ) is continuous over its support (being nonpaltry over X), then lim n E k n / n V ( B ( · , T n ) ) = p ^ ( · ) . Consequently, there is an integer n p such that Equation (10) holds true for all n > n p .
Then, let us consider Equation (9). At any time during training, regardless of n and T n , the SC-NN-4pdf computing φ ( · , A ( ϵ 1 ) , w ) can be set and kept in canonical form (Theorem 3 in [32]) without affecting φ ( · , A ( ϵ 1 ) , w ) . In so doing, Theorem 4 in [32] ensures that the loss function C n ( w ) presents no local minima with respect to w (since by hypotheses ker [ X n ] D 1 Y = { 0 } ), therefore neither L ( T n , w ) does (as we let ρ 0 ). Consequently, for any value of n, BP training (without early stopping) applied to this SC-NN-4pdf converges to the global minimum w n * of the criterion L ( T n , w ) , i.e., w n = w n * , given the fact that the learning rate η was chosen (by hypotheses) s.t. | 1 η λ max | < 1 at each step of BP (see [12], Sec. 7.5.1, p. 266).
A paramount consequence of this argument is that if we let L ¯ n ( w ( ϵ 1 ) ) = 1 2 x i T n ϕ ( ϵ 1 ) ( x i , A ( ϵ 1 ) , w ( ϵ 1 ) ) k n / n V ( B ( x i , T n ) ) 2 + ρ 2 1 X φ ( ϵ 1 ) ( x , A ( ϵ 1 ) , w ( ϵ 1 ) ) d x 2 then L ( T n , w n ) L ¯ n ( w ( ϵ 1 ) ) for all the values of n. Ergo, bearing in mind that p ^ ( · ) = lim n E k n / n V ( B ( · , T n ) ) as well as that A ( ϵ 1 ) and w ( ϵ 1 ) were chosen in such a way that p ^ ( · ) ϕ ( ϵ 1 ) ( · , A ( ϵ 1 ) , w ( ϵ 1 ) ) L 2 ( X ) < ϵ 1 , it is seen that
lim n E L ( T n , w n ) L 2 ( X ) lim n E L ¯ n ( w ( ϵ 1 ) ) L 2 ( X ) < ϵ 1 2 < ϵ 1
where we exploited the definition of L 2 norm, as well as the fact that the integral of the function realized by the SC-NN-4pdf converges to 1 as φ ( . ) approaches a pdf. Equation (13) entails the existence of an integer n ϕ s.t. Equation (9) holds true for all n > n ϕ . Since Equation (10) was already proven for all n > n p , and since we fixed ϵ 1 and ϵ 2 s.t. ϵ 1 + ϵ 2 = ϵ , if we let n 0 = max ( n p , n ϕ ) then Equation (8) yields E ϕ ( · , A ( ϵ 1 ) , w n ) p ^ ( · ) L 2 ( X ) < ϵ for all n > n 0 , as sought.
Then, let us work out the convergence of the variance, aiming at satisfying the asymptotic condition (6). Given the fact that the variance is non-negative, and resorting to one of its computational formulae, it is possible to write
0 σ 2 ϕ ( · , A ( ϵ 1 ) , w n ) p ^ ( · ) L 2 ( X ) = E ϕ ( · , A ( ϵ 1 ) , w n ) p ^ ( · ) L 2 ( X ) 2 E ϕ ( · , A ( ϵ 1 ) , w n ) p ^ ( · ) L 2 ( X ) 2 E ϕ ( · , A ( ϵ 1 ) , w n ) p ^ ( · ) L 2 ( X ) 2
where we relied on the fact that { E [ ϕ ( · , A ( ϵ 1 ) , w n ) p ^ ( · ) L 2 ( X ) ] } 2 is non-negative. Following in the footsteps of the analysis we applied earlier to the convergence of the mean, it is seen that E [ ϕ ( · , A ( ϵ 1 ) , w n ) p ^ ( · ) L 2 ( X ) 2 ] < ϵ for all n > n 0 which, as a consequence of Equation (14), proves that also σ 2 [ ϕ ( · , A ( ϵ 1 ) , w n ) p ^ ( · ) L 2 ( X ) ] < ϵ for all n > n 0 , as desired. This completes the proof. □

4. Conclusions

Density estimation is at the core of many practical applications rooted in pattern recognition, unsupervised learning, statistical analysis. and coding. Despite its having long been investigated, it is still an open problem. On the one hand, statistical techniques suffer from severe drawbacks. On the other hand, ANN-based pdf estimation algorithms struggle to break-through due to the difficulties posed by the very nature of the unsupervised estimation task and the requirement of satisfying Kolmogorov’s axioms of probability. The topic has recently been receiving an increasing attention from the Community, and some algorithmic attempts to tackle the issues were presented in the literature. In particular, the SC-NN-4pdf was proposed and successfully applied over several univariate and multivariate density estimation tasks, yielding improvements over the established approaches. Despite the empirical evidence stressing its effectiveness, no theoretical analysis of its properties had been carried out so far. This paper contributed filling such a gap, along two major directions: (1) formal identification of the class of pdfs that can be modeled by SC-NN-4pdfs to any degree ϵ of precision, and (2) proof of the asymptotic convergence in probability of the SC-NN-4pdf training algorithm. In particular, it was shown that under the assumption ker [ X n ] D 1 Y = { 0 } , 2-layer SC-NN-4pdfs with proper architecture converge in probability to an expected solution that is close to the true pdf to the desired degree ϵ of precision. These theoretical properties of the SC-NN-4pdf lay the groundwork for understanding the strong estimation capabilities of the present family of ANN-based density estimators. The generality of the class of estimated pdfs (the nonpaltry pdfs) and the asymptotic convergence to the true pdf underlying a data distribution make the SC-NN-4pdf a promising practical tool especially over large data samples. In particular, in classification tasks involving c classes ω 1 , , ω c , the corresponding class-conditional pdfs p ( x | ω 1 ) , , p ( x | ω c ) can be estimated via c class-specific SC-NN-4pdfs that, as n increases, tend to the true pdfs such that the estimates of the class-posterior probabilities P ( ω 1 | x ) , , P ( ω c | x ) computed relying on these pdfs estimates will, in turn, converge to the true class-posteriors, and the maximum-a-posteriori decision rule will end up realizing the ideal Bayes decision rule, i.e., the classifier with minimum probability of error. An even more intriguing application can be found in semi-supervised learning, where a limited subsample of the overall training set is actually supervised, while a huge fraction of the training data is unlabeled. Again, SC-NN-4pdf models of the class-conditional pdfs can be estimated from the supervised subsample first, and then applied to the unsupervised data, where a maximum-likelihood criterion based on the SC-NN-4pdf outputs (due to their being proper pdfs) can be applied in order to label a fraction of the unlabeled patterns, enlarging the supervised subset, and so forth. Finally, since the SC-NN-4pdf training algorithm offers a technique for sampling from the neural network (a step required in the computation of the integrals involved), once trained the SC-NN-4pdf can be applied as a generative model in order to draw new data from the pdf learned. Nevertheless, like most statistical and neural density estimators, SC-NN-4pdfs still suffer from the curse of dimensionality [1]. Although the paper did not offer a general solution to such a problem, the asymptotic convergence for n has the obvious practical implication that the curse is less and less dramatic (in probability) as the cardinality of the training set increases. Yet, Theorem 2 proves convergence, but it does not offer insights on the convergence rate, i.e., convergence may be quite slow, especially over high-dimensional feature spaces.

Funding

This research received no external funding.

Conflicts of Interest

The author declares no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
pdfProbability density function
ANNArtificial neural network
MLPMultilayer perceptron
FFNNFeed-forward neural network
SC-NN-4pdfSoft-constrained neural network for pdfs
BPBackpropagation

References

  1. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification, 2nd ed.; Wiley-Interscience: New York, NY, USA, 2000. [Google Scholar]
  2. Liang, F.; Barron, A. Exact Minimax Strategies for Predictive Density Estimation, Data Compression, and Model Selection. IEEE Trans. Inf. Theory 2004, 50, 2708–2726. [Google Scholar] [CrossRef]
  3. Beirami, A.; Sardari, M.; Fekri, F. Wireless Network Compression Via Memory-Enabled Overhearing Helpers. IEEE Trans. Wirel. Commun. 2016, 15, 176–190. [Google Scholar] [CrossRef]
  4. Yang, Z. Machine Learning Approaches to Bioinformatics; World Scientific Publishing Company: Singapore, 2010. [Google Scholar]
  5. Rabiner, L.R. A tutorial on hidden Markov models and selected applications in speech recognition. Proc. IEEE 1989, 77, 257–286. [Google Scholar] [CrossRef]
  6. Trentin, E.; Scherer, S.; Schwenker, F. Emotion recognition from speech signals via a probabilistic echo-state network. Pattern Recognit. Lett. 2015, 66, 4–12. [Google Scholar] [CrossRef]
  7. Bongini, M.; Rigutini, L.; Trentin, E. Recursive Neural Networks for Density Estimation Over Generalized Random Graphs. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 5441–5458. [Google Scholar] [CrossRef] [PubMed]
  8. Trentin, E.; Di Iorio, E. Nonparametric small random networks for graph-structured pattern recognition. Neurocomputing 2018, 313, 14–24. [Google Scholar] [CrossRef]
  9. Baioletti, M.; Milani, A.; Santucci, V. Learning Bayesian Networks with Algebraic Differential Evolution. In Proceedings of the 15th International Conferenc on Parallel Problem Solving from Nature (PPSN XV), Coimbra, Portugal, 8–12 September 2018; Part II; Lecture Notes in Computer Science. Auger, A., Fonseca, C., Lourenço, N., Machado, P., Paquete, L., Whitley, D., Eds.; Springer: Cham, Switzerland, 2018; Volume 11102, pp. 436–448. [Google Scholar]
  10. Wang, C.; Xu, C.; Yao, X.; Tao, D. Evolutionary Generative Adversarial Networks. IEEE Trans. Evol. Comput. 2019, 23, 921–934. [Google Scholar] [CrossRef] [Green Version]
  11. Trentin, E.; Lusnig, L.; Cavalli, F. Parzen neural networks: Fundamentals, properties, and an application to forensic anthropology. Neural Netw. 2018, 97, 137–151. [Google Scholar] [CrossRef] [PubMed]
  12. Bishop, C.M. Neural Networks for Pattern Recognition; Oxford University Press: Oxford, UK, 1995. [Google Scholar]
  13. Trentin, E.; Freno, A. Probabilistic Interpretation of Neural Networks for the Classification of Vectors, Sequences and Graphs. In Innovations in Neural Information Paradigms and Applications; Springer: Berlin/Heidelberg, Germany, 2009; pp. 155–182. [Google Scholar]
  14. Trentin, E.; Bongini, M. Probabilistically Grounded Unsupervised Training of Neural Networks. In Unsupervised Learning Algorithms; Celebi, M., Aydin, K., Eds.; Springer: Cham, Switzerland, 2016; pp. 533–558. [Google Scholar]
  15. Specht, D. Probabilistic Neural Networks. Neural Netw. 1990, 3, 109–118. [Google Scholar] [CrossRef]
  16. Modha, D.S.; Fainman, Y. A learning law for density estimation. IEEE Trans. Neural Netw. 1994, 5, 519–523. [Google Scholar] [CrossRef] [PubMed]
  17. Modha, D.S.; Masry, E. Rate of convergence in density estimation using neural networks. Neural Comput. 1996, 8, 1107–1122. [Google Scholar] [CrossRef]
  18. Yin, H.; Allinson, N.M. Self-organizing mixture networks for probability density estimation. IEEE Trans. Neural Netw. 2001, 12, 405–411. [Google Scholar] [CrossRef] [PubMed]
  19. Kohonen, T. Self-organized formation of topologically correct feature maps. Biol. Cybern. 1982, 43, 59–69. [Google Scholar] [CrossRef]
  20. Vapnik, V.N.; Mukherjee, S. Support Vector Method for Multivariate Density Estimation. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2000; pp. 659–665. [Google Scholar]
  21. Magdon-Ismail, M.; Atiya, A. Density estimation and random variate generation using multilayer networks. IEEE Trans. Neural Netw. 2002, 13, 497–520. [Google Scholar] [CrossRef] [PubMed]
  22. Trentin, E. Soft-Constrained Nonparametric Density Estimation with Artificial Neural Networks. In Proceedings of the 7th Workshop on Artificial Neural Networks in Pattern Recognition (ANNPR), Ulm, Germany, 28–30 September 2016; Springer: Cham, Switzerland, 2016; pp. 68–79. [Google Scholar]
  23. Chilinski, P.; Silva, R. Neural Likelihoods via Cumulative Distribution Functions. arXiv 2018, arXiv:1811.00974. [Google Scholar]
  24. Trentin, E. Maximum-Likelihood Estimation of Neural Mixture Densities: Model, Algorithm, and Preliminary Experimental Evaluation. In Proceedings of the 8th IAPR TC3 Workshop on Artificial Neural Networks in Pattern Recognition, Siena, Italy, 19–21 September 2018; pp. 178–189. [Google Scholar]
  25. Trentin, E. Soft-Constrained Neural Networks for Nonparametric Density Estimation. Neural Process. Lett. 2018, 48, 915–932. [Google Scholar] [CrossRef]
  26. Cybenko, G. Approximation by superposition of sigmoidal functions. Math. Control. Signal Syst. 1989, 2, 303–314. [Google Scholar] [CrossRef]
  27. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  28. Kolmogorov, A.; Fomin, S. Elementy Teorii Funktsii I Funktsional’nogo Analiza; Nauka (MIR): Moskov, Russia, 1980. [Google Scholar]
  29. Fukunaga, K. Introduction to Statistical Pattern Recognition, 2nd ed.; Academic Press: San Diego, CA, USA, 1990. [Google Scholar]
  30. Dekking, F.; Kraaikamp, C.; Lopuhaä, H. A Modern Introduction to Probability and Statistics: Understanding Why and How; Springer: London, UK, 2005. [Google Scholar]
  31. Gori, M.; Tesi, A. On the Problem of Local Minima in Backpropagation. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 76–86. [Google Scholar] [CrossRef] [Green Version]
  32. Gori, M.; Tsoi, A.C. Comments on local minima free conditions in multilayer perceptrons. IEEE Trans. Neural Netw. 1998, 9, 1051–1053. [Google Scholar] [CrossRef] [PubMed]
  33. Parzen, E. Modern Probability Theory and its Applications; John Wiley & Sons: Hoboken, NJ, USA, 1962. [Google Scholar]
Figure 1. SC-NN-4pdf estimate of a mixture of GEV pdfs ( n = 1000 ).
Figure 1. SC-NN-4pdf estimate of a mixture of GEV pdfs ( n = 1000 ).
Mathematics 08 00572 g001

Share and Cite

MDPI and ACS Style

Trentin, E. Asymptotic Convergence of Soft-Constrained Neural Networks for Density Estimation. Mathematics 2020, 8, 572. https://doi.org/10.3390/math8040572

AMA Style

Trentin E. Asymptotic Convergence of Soft-Constrained Neural Networks for Density Estimation. Mathematics. 2020; 8(4):572. https://doi.org/10.3390/math8040572

Chicago/Turabian Style

Trentin, Edmondo. 2020. "Asymptotic Convergence of Soft-Constrained Neural Networks for Density Estimation" Mathematics 8, no. 4: 572. https://doi.org/10.3390/math8040572

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop